简体   繁体   English

Scala Spark数据框按字段合并

[英]Scala spark data frames group by with field merging

I have data frame with 5 rows: 我有5行的数据框:

[user,sun,MainTest2,end,false]
[user,sun,MainTest2,start,false]
[user,sun,MainTest,end,true]
[user,sun,MainTest,start,true]
[user,sun,MainTest,start,true]

What i wan't to achieve is 2 aggregated rows with text merged on 4 field: 我将无法实现的是2个聚合行,并在4个字段上合并了文本:

[user,sun,MainTest2,end#some_union_symbol#start,false]
[user,sun,MainTest,end#some_union_symbol#start#some_union_symbol#start,true]

I think I need do something about reduce, but not quite sure. 我认为我需要做一些减少的事情,但还不是很确定。 Thank you! 谢谢!

Only one Answer from all stack that i can find and work with my 1.4.1 spark. 我能找到并使用我的1.4.1 spark的所有堆栈中只有一个Answer

sorry, for double post 抱歉,要重复

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM