简体   繁体   English

在EMR群集上运行Spark作业时发生异常“ java.io.IOException:所有数据节点均损坏”

[英]Exceptions while running Spark job on EMR cluster “java.io.IOException: All datanodes are bad”

We have AWS EMR setup to process jobs which are written in Scala. 我们拥有AWS EMR设置来处理用Scala编写的作业。 We are able to run the jobs on small dataset, but while running same job on large dataset I get exception "java.io.IOException: All datanodes are bad." 我们能够在小型数据集上运行作业,但是在大型数据集上运行相同作业时,出现异常“ java.io.IOException:所有数据节点均损坏。”

Setting spark.shuffle.service.enabled to true resolved this issue for me. 将spark.shuffle.service.enabled设置为true可以为我解决此问题。

The default configuration of AWS EMR has set spark.dynamicAllocation.enabled to true, but spark.shuffle.service.enabled is set to false . AWS EMR的默认配置已将spark.dynamicAllocation.enabled设置为true,但将spark.shuffle.service.enabled设置为false

spark.dynamicAllocation.enabled allows Spark to assign the executors dynamically to different task. spark.dynamicAllocation.enabled允许Spark将执行程序动态分配给其他任务。 The spark.shuffle.service.enabled when set to false disables the external shuffle service and data is stored only on executors. 如果将spark.shuffle.service.enabled设置为false ,则将禁用外部随机播放服务,并且数据仅存储在执行程序上。 When the executors is reassigned the data is lost and the exception "java.io.IOException: All datanodes are bad." 重新分配执行程序后,数据将丢失,并且出现异常“ java.io.IOException:所有数据节点均损坏”。 is thrown for data request. 引发数据请求。

暂无
暂无

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 线程“主”java.io.IOException 中的异常:作业失败 - Exception in thread "main" java.io.IOException: Job failed Spark&hbase:java.io.IOException:对等重置连接 - Spark&hbase: java.io.IOException: Connection reset by peer 运行sbt失败 - java.io.IOException:设备上没有剩余空间 - Running sbt fails - java.io.IOException: No space left on device 从 ensime 运行 sbt 时出现 java.io.IOException? - java.io.IOException when running sbt from ensime? sc.textFile() 在运行命令以获取文件权限时出错:java.io.IOException: (null) entry in command string - sc.textFile() gives Error while running command to get file permissions : java.io.IOException: (null) entry in command string java.io.IOException:方案的无文件系统:hdfs - java.io.IOException: No FileSystem for scheme : hdfs Spark在Gzip中编码并发送到S3 - java.io.IOException:设备上没有剩余空间 - Spark encode in Gzip and send to S3 - java.io.IOException: No space left on device 读取文件时 scala 项目出错:原因:java.io.IOException:方案没有文件系统:文件 - Error in scala project while reading file : Caused by: java.io.IOException: No FileSystem for scheme: file java.io.IOException:无法在 Hadoop 二进制文件中找到可执行文件 null\bin\winutils.exe。 火花 Eclipse 在 windows 7 - java.io.IOException: Could not locate executable null\bin\winutils.exe in the Hadoop binaries. spark Eclipse on windows 7 SBT发布未将jar文件上传到工件(发布)java.io.IOException:对URL的PUT操作失败,状态码为400:错误的请求 - SBT publish is not uploading jar file to artifactory (publish) java.io.IOException: PUT operation to URL failed with status code 400: Bad Request
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM