简体   繁体   English

我想将org.apache.spark.rdd.RDD [(Any,scala.collection.immutable.Iterable [String])]展平为Scala中的String

[英]I want to flatten org.apache.spark.rdd.RDD[(Any, scala.collection.immutable.Iterable[String])] to String in Scala

Example

(168,20874,List(, 33895, 2711)) (168,20874,List(,33895,2711))

to 168,20874| 到168,20874 | , 33895, 2711 ,33895、2711

Basically convert RDD[(Any, scala.collection.immutable.Iterable[String])] to String. 基本上将RDD [(Any,scala.collection.immutable.Iterable [String])]转换为String。

Thanks 谢谢

Method "map" with predefined formatter can be used: 可以使用带有预定义格式器的方法“ map”:

val rdd = sparkContext.parallelize(List((168, 20874, List(33895, 2711))))
val result = rdd.map { case (a, b, c) => s"$a,$b| ,${c.mkString(",")}" }
result.foreach(println)

Output: 输出:

168,20874| ,33895,2711

暂无
暂无

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 Scala / Spark中的异常org.apache.spark.rdd.RDD [(scala.collection.immutable.Map [String,Any],Int)] - Exception org.apache.spark.rdd.RDD[(scala.collection.immutable.Map[String,Any], Int)] in scala/spark Scala-如何过滤RDD org.apache.spark.rdd.RDD [String]] - Scala-How to filter an RDD org.apache.spark.rdd.RDD[String]] org.apache.spark.rdd.RDD [(((String,Double),(String,Double)))]到Scala中的数据框 - org.apache.spark.rdd.RDD[((String, Double), (String, Double))] to Dataframe in Scala ubuntu scala ide - spark - toDF 方法错误值 toDF is not member of org.apache.spark.rdd.RDD[String] - ubuntu scala ide - spark - toDF Method error value toDF is not a member of org.apache.spark.rdd.RDD[String] 来自org.apache.spark.rdd.RDD [(((Any,Any),Iterable [org.apache.spark.sql.Row])的Spark Sql数据 - Spark Sql data from org.apache.spark.rdd.RDD[((Any, Any), Iterable[org.apache.spark.sql.Row])] 将RDD从类型org.apache.spark.rdd.RDD [(((String,String),Double)]`转换为org.apache.spark.rdd.RDD [(((String),List [Double])]]` - Convert RDD from type `org.apache.spark.rdd.RDD[((String, String), Double)]` to `org.apache.spark.rdd.RDD[((String), List[Double])]` scala.collection.immutable.Iterable [org.apache.spark.sql.Row]到DataFrame吗? 错误:方法值重载createDataFrame及其它替代方法 - scala.collection.immutable.Iterable[org.apache.spark.sql.Row] to DataFrame ? error: overloaded method value createDataFrame with alternatives 如何映射类型为org.apache.spark.rdd.RDD [Array [String]]的RDD? - How to map a RDD of type org.apache.spark.rdd.RDD[Array[String]]? Scala:出现错误-mapPartitionsWithIndex不是org.apache.spark.rdd.RDD [Int]的成员 - Scala: Getting error - mapPartitionsWithIndex is not a member of org.apache.spark.rdd.RDD[Int] Spark 1.6-关联规则算法-无法应用于(org.apache.spark.rdd.RDD [Array [String]]) - Spark 1.6 - Association Rules algorithm - Cannot be applied to (org.apache.spark.rdd.RDD[Array[String]])
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM