简体   繁体   English

使用--master yarn-cluster运行spark-submit:spark-assembly问题

[英]Running spark-submit with --master yarn-cluster: issue with spark-assembly

I am running Spark 1.1.0, HDP 2.1, on a kerberized cluster. 我在kerberized集群上运行Spark 1.1.0,HDP 2.1。 I can successfully run spark-submit using --master yarn-client and the results are properly written to HDFS, however, the job doesn't show up on the Hadoop All Applications page. 我可以使用--master yarn-client成功运行spark-submit,并将结果正确写入HDFS,但是,该作业未显示在Hadoop All Applications页面上。 I want to run spark-submit using --master yarn-cluster but I continue to get this error: 我想使用--master yarn-cluster运行spark-submit但我继续收到此错误:

appDiagnostics: Application application_1417686359838_0012 failed 2 times due to AM Container
for appattempt_1417686359838_0012_000002 exited with  exitCode: -1000 due to: File does not
exist: hdfs://<HOST>/user/<username>/.sparkStaging/application_<numbers>_<more numbers>/spark-assembly-1.1.0-hadoop2.4.0.jar
.Failing this attempt.. Failing the application.

I've provisioned my account with access to the cluster. 我已经为我的帐户配置了对群集的访问权限。 I've configured yarn-site.xml. 我已经配置了yarn-site.xml。 I've cleared .sparkStaging. 我已经清除了.sparkStaging。 I've tried including --jars [path to my spark assembly in spark/lib]. 我尝试过包含--jars [spark / lib中我的spark程序集的路径]。 I've found this question that is very similar, yet unanswered. 我发现这个问题非常相似,但没有答案。 I can't tell if this is a 2.1 issue, spark 1.1.0, kerberized cluster, configurations, or what. 我不知道这是2.1问题,火花1.1.0,kerberized集群,配置,或什么。 Any help would be much appreciated. 任何帮助将非常感激。

这可能是因为你在代码中留下了sparkConf.setMaster("local[n]")

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM