简体   繁体   English

EMR Hadoop并未利用所有群集节点

[英]EMR Hadoop does not utilize all cluster nodes

We are experimenting with Hadoop and processing of the Common Crawl. 我们正在尝试Hadoop和Common Crawl的处理。

Our problem is that if we create a cluster with 1 Master Node and 1 Core and 2 Task nodes, only one of the nodes per group will get high CPU/Network usage. 我们的问题是,如果创建一个具有1个Master节点,1个Core节点和2个Task节点的群集,则每个组中只有一个节点会获得较高的CPU /网络使用率。 We tried also with 2 Core and no Task nodes, but in this case also only one Core node was used. 我们还尝试了2个Core节点和没有Task节点,但是在这种情况下,也只使用了一个Core节点。

Following some screenshots of the Node/Cluster monitoring. 以下是“节点/集群”监视的一些屏幕截图。 The job was running all the time (in the first two parallel map phases), and should have used most of the available CPU power, as you can see in the screenshot of the working Task node. 作业一直在运行(在前两个并行映射阶段),并且应该已经使用了大部分可用的CPU能力,如工作Task节点的屏幕快照所示。 But why is the idle Task node not utilized? 但是,为什么没有利用空闲的Task节点呢?

Our hadoop job, running as an Jar step, has no limits for the map jobs. 我们的hadoop作业作为Jar步骤运行,对地图作业没有限制。 It consists of multiple map/reduce steps chained. 它由多个链接的映射/减少步骤组成。 The last reduce job is limited to one Reducer. 最后的reduce作业仅限于一个Reducer。

Screenshots: https://drive.google.com/drive/folders/1xwABYJMJAC_B0OuVpTQ9LNSj12TtbxI1?usp=sharing 截图: https//drive.google.com/drive/folders/1xwABYJMJAC_B0OuVpTQ9LNSj12TtbxI1?usp = sharing

ClusterId: j-3KAPYQ6UG9LU6 ClusterId: j-3KAPYQ6UG9LU6

StepId: s-2LY748QDLFLM9 步骤编号: s-2LY748QDLFLM9

We found the following in the System Logs of the idle Node during an other run, maybe it is an EMR problem? 在其他运行期间,我们在空闲节点的系统日志中找到以下内容,也许这是EMR问题?

ERROR main: Failed to fetch extraInstanceData from https://aws157-instance-data-1-prod-us-east-1.s3.amazonaws.com/j-2S62KOVL68GVK/ig-3QUKQSH7YJIAU.json?X-Amz-Algorithm=AWS4-HMAC-SHA256&X

Greetings Lukas 问候卢卡斯

Late to the party, but have you tried setting these properties as part of the spark submit command. 聚会晚了,但是您尝试将这些属性设置为spark Submit命令的一部分。

--conf 'spark.dynamicAllocation.enabled=true' 
--conf 'spark.dynamicAllocation.minExecutors=<MIN_NO_OF_CORE_OR_TASK_NODES_YOU_WANT>'

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM