![](/img/trans.png)
[英]Writing dataframe to Azure Blob Storage using Databricks generates an empty file
[英]Saving a dataframe as a csv file(processed in databricks) and uploading it to azure datalake blob storage
I had a csv file stored in azure datalake storage which i imported in databricks by mounting the datalake account in my databricks cluster, After doing preProcessing i wanted to store the csv back in the same datalakegen2 (blobstorage) account.Any leads and help on the問題表示贊賞。謝謝。
只需在相同的安裝位置寫入文件。 請參閱此處的示例: https://docs.databricks.com/spark/latest/data-sources/azure/azure-datalake-gen2.html#example-notebook
df.write.json("abfss://<file_system>@<storage-account-name>.dfs.core.windows.net/iot_devices.json")
只需將其直接保存到 Blob 存儲即可。
df.write.
format("com.databricks.spark.csv").
option("header", "true").
save("myfile.csv")
在本地保存文件然后將其推送到 Blob 中沒有意義。
聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.