简体   繁体   English

HDFS HA yarn 在 spark-submit 上的这个错误是什么

[英]What is this error on spark-submit by HDFS HA yarn

here is my error log:这是我的错误日志:

$ /spark-submit --master yarn --deploy-mode cluster pi.py
...
2021-12-23 01:31:04,330 INFO retry.RetryInvocationHandler: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException): Operation category WRITE is not supported in state standby. Visit https://s.apache.org/sbnn-error
    at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:88)
    at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1954)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1442)
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setPermission(FSNamesystem.java:1895)
    at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setPermission(NameNodeRpcServer.java:860)
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setPermission(ClientNamenodeProtocolServerSideTranslatorPB.java:526)
    at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:524)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1025)
    at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:876)
    at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:822)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2682)
, while invoking ClientNamenodeProtocolTranslatorPB.setPermission over master/172.17.0.2:8020. Trying to failover immediately.
...

Why I get this erorr??为什么我得到这个错误?

NOTE .注意 Spark master is run 'master', so spark-submit command run in 'master' Spark master 运行“master”,因此 spark-submit 命令在“master”中运行

NOTE .注意 Spark worker is run 'worker1' and 'worker2' and 'worker3' Spark 工作人员正在运行“worker1”和“worker2”和“worker3”

NOTE .注意 ResourceManager run in 'master' and 'master2' ResourceManager 在“master”和“master2”中运行

ADD .添加 When print above error log, master2's DFSZKFailoverController is disappeard to jps command result.打印上述错误日志时,master2 的 DFSZKFailoverController 消失为 jps 命令结果。

ADD .添加 When print above error log, master's Namenode is disappeard to jps command result.打印上面的错误日志时,master的Namenode消失到jps命令结果中。

It happens when Spark is unable to access HDFS.当 Spark 无法访问 HDFS 时会发生这种情况。

If configured correctly HDFS client will handle the StandbyException by attempting to fail itself over to the other NameNode in the HA, and then it will reattempt the operation.如果配置正确,HDFS 客户端将通过尝试将自身故障转移到 HA 中的另一个 NameNode 来处理 StandbyException,然后它将重新尝试该操作。

Replace active Namenode URI manually and check if you are still having the same error, if not HA is not properly configured.手动替换活动 Namenode URI 并检查您是否仍然遇到相同的错误,如果没有,则 HA 配置不正确。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM