[英]Spark running in local machine instead of standalone cluster
I've created a spark cluster on ec2, after that, I installed Jupyter on the master node and started jupyter, after that I created sparkcontext using我在 ec2 上创建了一个 spark 集群,之后,我在主节点上安装了 Jupyter 并启动了 jupyter,之后我使用创建了 sparkcontext
findspark.init(spark_home='/home/ubuntu/spark')
import pyspark
from functools import partial
sc = pyspark.SparkContext(appName="Pi")
when I am trying to run any job, spark is only utilizing cores of the master machine, all the slaves are running and connected to master, but I am still not able to use the cores of any of the slave machines, anybody please help.当我尝试运行任何作业时,spark 仅使用主机的内核,所有从机都在运行并连接到主机,但我仍然无法使用任何从机的内核,请任何人帮忙。
创建SparkContext
时,您需要将主 url 设置为spark://...
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.