[英]How to update CSV file placed at Blob Storage in Azure
I have Blob Storage in Azure that holds a CSV file.我在 Azure 中有 Blob 存储,其中包含一个 CSV 文件。 While incrementally updating it creates a new CSV file instead we need to update the existing CSV file.
在增量更新时,它会创建一个新的 CSV 文件,而我们需要更新现有的 CSV 文件。
Is there any solution in the Azure data factory to update incremental data into the existing CSV file? Azure数据工厂有没有办法把增量数据更新到现有的CSV文件中?
You can work on incremental updates to a Blob Storage file using Azure Databricks and in ADF you can call the databricks notebook activity您可以使用 Azure Databricks 对 Blob 存储文件进行增量更新,在 ADF 中您可以调用 databricks notebook 活动
In this you will have to mount the Azure storage on databricks cluster first在这里,您必须首先在数据块集群上安装 Azure 存储
Then you will have to read the original and incremental file or data as a dataframe然后你将不得不读取原始和增量文件或数据作为 dataframe
You can perform joins on the files and create an incremental data frame which you can overwrite to your original file which will be your incremental data and that too in a single file您可以对文件执行连接并创建一个增量数据框,您可以将其覆盖到您的原始文件,这将是您的增量数据,并且也在单个文件中
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.