简体   繁体   English

SPARK YARN:无法从客户端发送作业(org.apache.hadoop.ipc.Client-重试连接到服务器:0.0.0.0/0.0.0.0:8032)

[英]SPARK YARN: cannot send job from client (org.apache.hadoop.ipc.Client - Retrying connect to server: 0.0.0.0/0.0.0.0:8032)

I'm trying to send spark job to yarn (without HDFS) in HA mode. 我正在尝试在HA模式下将火花作业发送到纱线(没有HDFS)。

For submitting I'm using org.apache.spark.deploy.SparkSubmit . 对于提交,我使用的是org.apache.spark.deploy.SparkSubmit When I send request from machine with active Resource Manager, it works well. 当我从具有活动资源管理器的计算机发送请求时,它运行良好。 But if I' trying to send from machine with standby Resource Manager, job fails with error: 但是,如果我尝试使用备用资源管理器从计算机发送邮件,则作业将失败并显示以下错误:

DEBUG org.apache.hadoop.ipc.Client - Connecting to spark2-node-dev/10.10.10.167:8032
DEBUG org.apache.hadoop.ipc.Client - Connecting to /0.0.0.0:8032
org.apache.hadoop.ipc.Client - Retrying connect to server: 0.0.0.0/0.0.0.0:8032. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep

However, when I send request via command line (spark-submit), it works well through both active and standby machine. 但是,当我通过命令行(spark-submit)发送请求时,它在主用计算机和备用计算机上都能正常工作。

What can cause the problem? 是什么原因引起的?

PS Use the same parameters for both type of sending job: org.apache.spark.deploy.SparkSubmit and spark-submit command line request. PS对两种类型的发送作业使用相同的参数: org.apache.spark.deploy.SparkSubmitspark-submit命令行请求。 And properties yarn.resourcemanager.hostname.rm_id defined for all rm hosts 并为所有rm主机定义了属性yarn.resourcemanager.hostname.rm_id

The problem was with absence of yarn-site.xml within class path for spark-submitter jar. 问题是在火花发送器jar的类路径中缺少yarn-site.xml。 Actually spark submitter jar does not take to account YARN_CONF_DIR or HADOOP_CONF_DIR env var, so cannot see yarn-site. 实际上,火花提交器jar并没有考虑YARN_CONF_DIRHADOOP_CONF_DIR YARN_CONF_DIR ,因此看不到纱线位置。

One solution that I found was to put yarn-site into classpath of jar. 我发现的一种解决方案是将纱线位置放入jar的类路径中。

暂无
暂无

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 火花上的火花,在/0.0.0.0:8032连接到ResourceManager - spark on yarn, Connecting to ResourceManager at /0.0.0.0:8032 Hadoop 2.6在/0.0.0.0:8032连接到ResourceManager - Hadoop 2.6 Connecting to ResourceManager at /0.0.0.0:8032 为什么使用“java.lang.ClassNotFoundException:org.apache.hadoop.fs.FSDataInputStream”启动带有yarn-client的spark-shell失败? - Why does launching spark-shell with yarn-client fail with “java.lang.ClassNotFoundException: org.apache.hadoop.fs.FSDataInputStream”? 使用Yarn客户端或任何其他方式从Spring Web应用程序运行Apache spark作业 - Running Apache spark job from Spring Web application using Yarn client or any alternate way SPARK错误:服务器IPC版本9无法与客户端版本4通信 - SPARK error: Server IPC version 9 cannot communicate with client version 4 从主机未知主机的主机火花Spark YARN客户端作业? - Spark YARN client job from host with hostname unknown to cluster? 使用scala将eclipse中的spark作业提交给yarn-client - Submitting spark job from eclipse to yarn-client with scala 用于Hadoop EC2安装的Spark脚本:IPC客户端连接被拒绝 - Spark Script for Hadoop EC2 Installation: IPC client connection refused 如何在Azure HDInsight上将Apache Spark作业提交到Hadoop YARN - How to submit Apache Spark job to Hadoop YARN on Azure HDInsight 纱线客户端:用户类抛出异常:java.lang.NoClassDefFoundError:org/apache/spark/sql/DataFrame - Yarn Client: User class threw exception: java.lang.NoClassDefFoundError: org/apache/spark/sql/DataFrame
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM