[英]How to read in JSON file and write as JSON Lines file in Spark Scala?
我有一個 JSON 文件,我想將其轉換為 Spark scala 中的 JSON 行。 通過使用 Pandas read_json 方法,然后使用傳入的一些特殊行參數寫入它,我能夠在 Python 中弄清楚它。
說原始格式是:
{
"A": "400",
"B": "100",
"C": "DEM",
"D": "USD",
"E": "80029898",
"F": "1.64110-",
"G": "0 "
},
{
"A": "400",
"B": "100",
"C": "USD",
"D": "DEM",
"E": "80029898",
"F": "1.64110 ",
"G": "0 ",
},
我想寫的是:
{"A":"400","B":"100","C":"DEM","D":"USD","E":"80029898","F":"1.64110-","G":"0"}
{"A":"400","B":"100","C":"USD","D":"DEM","E":"80029898","F":"1.64110 ","G":"0"}
非常感謝,祝您有美好的一天!
如果您使用 > spark 2.2,則可以使用:
spark.read
.option("multiLine", true).option("mode", "PERMISSIVE")
.json("/path/to/user.json")
然后使用以所需的格式編寫它:
df.write.json()
聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.