[英]Spring Cloud DataFlow using launch task after new file in AWS S3 Bucket source
I'm trying to create a batch processing started by a new file in AWS S3.我正在尝试创建由 AWS S3 中的新文件启动的批处理。
So the flow is:所以流程是:
1 - A new file is uploaded to AWS S3 Bucket
2 - SCDF detects
3 - SCDF launch the task (Spring Batch application)
4 - Spring Batch application process the file and stores in a DB.
Something similar to this, but with S3 Bucket: https://dataflow.spring.io/docs/recipes/batch/sftp-to-jdbc/与此类似,但使用 S3 存储桶: https://dataflow.spring.io/docs/recipes/batch/sftp-to-jdbc/
Maybe is a misunderstood of the concept, but in SFTP Source I could set Port, Host, User and Pass, but in S3 Source I doesn't have region and credentials properties.也许是对这个概念的误解,但在 SFTP 源中我可以设置端口、主机、用户和密码,但在 S3 源中我没有区域和凭据属性。
Where do I set the AWS properties?我在哪里设置 AWS 属性?
There's an Amazon AWS common options
section in the README (see: old-app / new-app ), which includes the common AWS-specific properties one can override.自述文件中有一个Amazon AWS common options
部分(请参阅: old-app / new-app ),其中包括可以覆盖的常见 AWS 特定属性。
You can pass them as inline properties in the stream definition or when deploying the stream by following the deployer properties convention.您可以在 stream 定义中或在部署 stream 时按照部署程序属性约定将它们作为内联属性传递。
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.