[英]SPARK SQL java.lang.OutOfMemoryError: GC overhead limit exceeded
I try to fetch the data for one day to test and create a temp table using Spark SQL.我尝试获取一天的数据以测试并使用 Spark SQL 创建一个临时表。 After that, I am performing my queries on temp table.
之后,我在临时表上执行我的查询。 So that process can be fast but still getting GC memory issue.
所以这个过程可以很快,但仍然会遇到 GC 内存问题。
Spark configuration:火花配置:
sparkConf = new SparkConf().setMaster("local[1]").setAppName("Scheduler");
sparkConf.set("spark.driver.memory", "2g") ;
sparkConf.set("spark.executor.memory", "1g");
sparkConf.set("spark.memory.onHeap.size","1g") ;
sparkConf.set("spark.num.executors", "4");
sparkConf.set("spark.executor.cores", "1");
sparkConf.set("spark.serializer", KryoSerializer.class.getName());
OoM can happen also when the GC activity consumes a certain percentage of CPU resource.当 GC 活动消耗一定百分比的 CPU 资源时,也会发生 OoM。 I would try to increase the heap and/or use a profiling tool jconsole or visualvm to check what is the root cause
我会尝试增加堆和/或使用分析工具 jconsole 或 visualvm 来检查根本原因是什么
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.