简体   繁体   English

纱线-火花平行工作

[英]yarn - spark parallel job

I made yarn-cluster which has only 1 work node, and it seems to work fine when I submit my spark application job. 我制作了只有1个工作节点的yarn-cluster,提交我的spark应用程序作业时似乎工作正常。 When I submit job more than one, jobs are on hadoop queue and process submitted application one by one. 当我提交一份以上的作业时,作业处于hadoop队列中,并逐一处理提交的申请。 I want to process my applications parallelly, not one by one. 我想并行处理我的应用程序,而不是一个接一个地处理。 Is there any configuration for this? 有什么配置吗? or unable to do this on yarn? 还是无法在纱线上做到这一点?

Yarn submits jobs one by one by default. 默认情况下,Yarn逐一提交作业。 For submit multiple jobs you can change amount of your executor cores: 要提交多个作业,您可以更改执行者核心的数量:

spark-submit class /jar --executor-memory 2g --num-executors 15 --executor-cores 3 --master yarn --deploy-mode cluster 

You also can change this properties in your yarn-site.xml 您还可以在yarn-site.xml中更改此属性。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM