更新时间:2022-05-16 03:36:23
有 3 种可能的解决方案,
There is 3 possible solutions,
spark-submit
cli 中添加依赖项.您可以在 spark-submit
cli 中使用以下选项:
spark-submit
cli.You can use the following option in your spark-submit
cli :
--jars $(echo ./lib/*.jar | tr ' ' ',')
说明:假设您在项目根目录的 lib
目录中拥有所有 jar,这将读取所有库并将它们添加到应用程序提交中.
Explanation : Supposing that you have all your jars in a lib
directory in your project root, this will read all the libraries and add them to the application submit.
您也可以尝试在 SPARK_HOME/conf/spark 中配置这 2 个变量:
文件并将这些变量的值指定为jar文件的路径.确保工作节点上存在相同的路径.spark.driver.extraClassPath
和 spark.executor.extraClassPath
-default.conf
You can also try to configure these 2 variables : spark.driver.extraClassPath
and spark.executor.extraClassPath
in SPARK_HOME/conf/spark-default.conf
file and specify the value of these variables as the path of the jar file. Ensure that the same path exists on worker nodes.