[英]Azure Data Factory Copy Data From Blob Storage Only New Added file(s)
I would like to create a pipeline with Copy Data
(mapping JSON source to MySQL columns).我想创建一个带有Copy Data
的管道(将 JSON 源映射到 MySQL 列)。 At this moment it is working, but the tool is copying all the files
from my source container after the 'created' trigger
.目前它正在工作,但该工具正在'created' trigger
之后从我的源容器中复制all the files
。 Is it possible to retrieve only 'new added' files in my container, so I do not have double data in MySQL?是否可以在我的容器中只检索“新添加”文件,所以我在 MySQL 中没有双重数据?
Example current situation:当前情况示例:
first copy -> retrieve file 1
(added new file)
second copy -> retrieve file 1 + file 2
Example expected situation:示例预期情况:
first copy -> retrieve file 1
(added new file)
second copy -> retrieve file 2
This is actually a very used scenario , you should use LastModifiedDate for copying most recently added files and to prevent duplication.这实际上是一个非常常用的场景,您应该使用 LastModifiedDate 来复制最近添加的文件并防止重复。
Please check this link : https://docs.microsoft.com/en-us/azure/data-factory/tutorial-incremental-copy-overview请检查此链接: https ://docs.microsoft.com/en-us/azure/data-factory/tutorial-incremental-copy-overview
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.