简体   繁体   中英

Spark SQL join on two RDDs created using java

I have created two RDDs and persisted them using java. I have submitted the jar to the spark master. Now when the spark shell prompts me to enter the query, I have given an SQL statement join condition on the two RDDs; then it throws an exception while performing collect() on the result set.

JOIN Condition : select a.ID from TABLE_1 as a JOIN TABLE_2 b ON a.NAME = b.NAME;

Exception : Null pointer Exception on resultSet.collect()

PS: I have checked if the result set is null before resultSet.collect(); but still it enters the condition and throws an NPE. I am using spark-sql-1.1.1 jar(latest).

Is it some issue with JOIN in SPARK SQL?

看起来Java Spark API的某些JOIN操作有问题。

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM