简体   繁体   English

错误 SparkContext:无法将 None 添加到 Spark 环境

[英]ERROR SparkContext: Failed to add None to Spark environment

I submit a spark job first like this in a pyspark file我首先在 pyspark 文件中像这样提交一个 spark 作业

    os.system(f'spark-submit --master local --jars ./examples/lib/app.jar app.py')

Then in the submitted app.py file, I create a new SparkSession like this:然后在提交的 app.py 文件中,我创建了一个新的 SparkSession,如下所示:

  spark = SparkSession.builder.appName(appName) \
        .config('spark.jars') \
        .getOrCreate()

Error message:错误信息:

23/01/17 11:02:52 INFO SparkContext: Running Spark version 3.3.0
23/01/17 11:02:52 INFO ResourceUtils: ==============================================================
23/01/17 11:02:52 INFO ResourceUtils: No custom resources configured for spark.driver.
23/01/17 11:02:52 INFO ResourceUtils: ==============================================================
23/01/17 11:02:52 INFO SparkContext: Submitted application: symbolic_test
23/01/17 11:02:52 INFO ResourceProfile: Default ResourceProfile created, executor resources: Map(cores -> name: cores, amount: 1, script: , vendor: , memory -> name: memory, amount: 1024, script: , vendor: , offHeap -> name: offHeap, amount: 0, script: , vendor: ), task resources: Map(cpus -> name: cpus, amount: 1.0)
23/01/17 11:02:52 INFO ResourceProfile: Limiting resource is cpu
23/01/17 11:02:53 INFO ResourceProfileManager: Added ResourceProfile id: 0
23/01/17 11:02:53 INFO SecurityManager: Changing view acls to: annie
23/01/17 11:02:53 INFO SecurityManager: Changing modify acls to: annie
23/01/17 11:02:53 INFO SecurityManager: Changing view acls groups to: 
23/01/17 11:02:53 INFO SecurityManager: Changing modify acls groups to: 
23/01/17 11:02:53 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(annie); groups with view permissions: Set(); users  with modify permissions: Set(annie); groups with modify permissions: Set()
23/01/17 11:02:53 INFO Utils: Successfully started service 'sparkDriver' on port 42141.
23/01/17 11:02:53 INFO SparkEnv: Registering MapOutputTracker
23/01/17 11:02:53 INFO SparkEnv: Registering BlockManagerMaster
23/01/17 11:02:53 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
23/01/17 11:02:53 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
23/01/17 11:02:53 INFO SparkEnv: Registering BlockManagerMasterHeartbeat
23/01/17 11:02:53 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-e4cc3b01-a6d5-4454-ad2d-4d0f42066479
23/01/17 11:02:53 INFO MemoryStore: MemoryStore started with capacity 434.4 MiB
23/01/17 11:02:53 INFO SparkEnv: Registering OutputCommitCoordinator
23/01/17 11:02:53 INFO Utils: Successfully started service 'SparkUI' on port 4040.
23/01/17 11:02:53 ERROR SparkContext: Failed to add None to Spark environment
java.io.FileNotFoundException: Jar /home/annie/exampleApp/example/None not found
    at org.apache.spark.SparkContext.addLocalJarFile$1(SparkContext.scala:1949)
    at org.apache.spark.SparkContext.addJar(SparkContext.scala:2004)
    at org.apache.spark.SparkContext.$anonfun$new$12(SparkContext.scala:507)
    at org.apache.spark.SparkContext.$anonfun$new$12$adapted(SparkContext.scala:507)
    at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
    at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
    at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
    at org.apache.spark.SparkContext.<init>(SparkContext.scala:507)
    at org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:58)
    at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)
    at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)
    at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
    at py4j.Gateway.invoke(Gateway.java:238)
    at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)
    at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)
    at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)
    at py4j.ClientServerConnection.run(ClientServerConnection.java:106)
    at java.base/java.lang.Thread.run(Thread.java:829)

when creating spark session through pyspark, I get the following error messages, which only arise when I add.config('spark.jars').在创建 spark session 到 pyspark 时,我收到以下错误消息,这些消息仅在我添加.config('spark.jars') 时出现。

I've set my $SPARK_HOME variable correctly... Any help will be appreciated!我已经正确设置了我的 $SPARK_HOME 变量...任何帮助将不胜感激!

If your code sample is true you do not assign any value to spark.jars key while creating spark session. Assigning jar path as value may solve the error.如果您的代码示例为真,则在创建 spark session 时不会为 spark.jars 键分配任何值。将 jar 路径分配为值可能会解决错误。

SparkSession.builder.appName(appName) \
    .config('config_key', config_value) \

暂无
暂无

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 错误SparkContext无法在Apache Spark 2.1.1中添加文件 - ERROR SparkContext Failed to add file in Apache Spark 2.1.1 在本地创建 sparkcontext 时出错调用 None.org.apache.spark.api.java.JavaSparkContext 时出错 - Error creating sparkcontext locally An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext 错误SparkContext:初始化SparkContext时出错(在tensorflow上运行mnist示例在spark上) - ERROR SparkContext: Error initializing SparkContext(run mnist example with tensorflow on spark) spark-shell 命令抛出此错误:SparkContext:初始化 SparkContext 时出错 - spark-shell command throwing this error: SparkContext: Error initializing SparkContext 错误SparkContext:初始化SparkContext时出错-Java + Eclipse + Spark - ERROR SparkContext: Error initializing SparkContext - Java + Eclipse + Spark SparkContext:运行 Spark 作业时初始化 SparkContext 时出错 - SparkContext: Error initializing SparkContext While Running Spark Job Linux intellij idea spark 初始化 SparkContext 时出错 - Linux intellij idea spark Error initializing SparkContext 如果未启动SparkContext,则群集模式下的Spark会引发错误 - Spark in cluster mode throws error if a SparkContext is not started Mac spark-shell 初始化 SparkContext 时出错 - Mac spark-shell Error initializing SparkContext Apache Spark-JavaSparkContext无法转换为SparkContext错误 - Apache Spark - JavaSparkContext cannot be converted to SparkContext error
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM