简体   繁体   中英

convert a column with json value to a data frame using scala spark

I found several helpful answers but that were all converting son file to df, in my case, I have a df with columns with son in them, like this:

s-timestamp : 2019-10-10

content : {"META":{"testA":"1","TABLENAME":"some_table_name"},"PINACOLADA":{"sampleID":"0","itemInserted":"2019-10-10","sampleType":"BASE",}"

I need to normalize the content column, how can I do that.

Welcome. There are a few ways of dealing with JSON strings in Spark DF columns. You can use functions like get_json_object to extract specific fields from your JSON or from_json to transform the field into a StructType with a given schema. Another option is to use spark.read.json to parse and create a separate dataframe from the column's contents. Have a look at my solution here and let me know if it helps.

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM