[英]spark:220: error: missing parameter type for “map”
Spark 2.1, scala: I am converting GDELT data to GraphX format. Spark 2.1,Scala:我正在将GDELT数据转换为GraphX格式。 However the example listed here fails while creating the hash value, using MurmurHash3: I dont know enough about scala types to diagnose this error message.
但是, 此处列出的示例在使用MurmurHash3创建哈希值时失败:我对scala类型了解不足,无法诊断此错误消息。
val eventsFromTo = gdelt.select("Actor1Name","Actor2Name").where("actor1Name is not null and actor2name is not null")
eventsFromTo.show(5)
+-------------+----------+
| Actor1Name|Actor2Name|
+-------------+----------+
| SENATE| RUSSIAN|
| MEXICO| TEXAS|
| RUSSIAN| SENATE|
| VERMONT| CANADA|
|UNITED STATES| POLICE|
+-------------+----------+
only showing top 5 rows
val eventActors = gdelt.select("Actor1Name","Actor2Name").where("actor1Name is not null and actor2name is not null").flatMap(x => Iterable(x(0).toString,x(1).toString))
eventActors.show(5)
+-------+
| value|
+-------+
| SENATE|
|RUSSIAN|
| MEXICO|
| TEXAS|
|RUSSIAN|
+-------+
Then I try to convert it for graphX: 然后,我尝试将其转换为graphX:
val eventVertices: RDD[(VertexId, String)] = eventActors.distinct().map(x => (MurmurHash3.stringHash((x),x)))
<console>:265: error: missing parameter type
If I add a type for i then I get this error: 如果我为我添加类型,则会出现此错误:
<console>:265: error: type mismatch;
found : String
required: Int
val eventVertices: RDD[(VertexId, String)] = eventActors.distinct().map((x:String) => (MurmurHash3.stringHash((x),x)))
I was missing ".rdd" to convert these to RDD's before executing the map() 我在执行map()之前缺少将这些转换为RDD的“ .rdd”
val eventsFromTo = gdelt.select("Actor1Name","Actor2Name").where("actor1Name is not null and actor2name is not null").rdd
val eventActors = gdelt.select("Actor1Name","Actor2Name").where("actor1Name is not null and actor2name is not null").flatMap(x => Iterable(x(0).toString,x(1).toString)).rdd
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.