简体   繁体   中英

Divide the values of two rdds in spark (python)

I have these two Key-value RDDs in spark:

rdd1 = [(u'Key1', 4), (u'Key2', 6), (u'Key3', 10)]
rdd2 = [(u'Key1', 4), (u'Key2', 3), (u'Key3', 2)]

And I looking the spark function to get the division of the values: (rdd3= (rdd1/rdd2))

In this case:

rdd3 = [(u'Key1', 1), (u'Key2', 2), (u'Key3', 5)]

你可以joinmapValues

rdd1.join(rdd2).mapValues(lambda x: x[0] / x[1])

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM