简体   繁体   English

Spring Cloud DataFlow 在 AWS S3 存储桶源中的新文件后使用启动任务

[英]Spring Cloud DataFlow using launch task after new file in AWS S3 Bucket source

I'm trying to create a batch processing started by a new file in AWS S3.我正在尝试创建由 AWS S3 中的新文件启动的批处理。

So the flow is:所以流程是:

1 - A new file is uploaded to AWS S3 Bucket
2 - SCDF detects
3 - SCDF launch the task (Spring Batch application)
4 - Spring Batch application process the file and stores in a DB.

Something similar to this, but with S3 Bucket: https://dataflow.spring.io/docs/recipes/batch/sftp-to-jdbc/与此类似,但使用 S3 存储桶: https://dataflow.spring.io/docs/recipes/batch/sftp-to-jdbc/

Maybe is a misunderstood of the concept, but in SFTP Source I could set Port, Host, User and Pass, but in S3 Source I doesn't have region and credentials properties.也许是对这个概念的误解,但在 SFTP 源中我可以设置端口、主机、用户和密码,但在 S3 源中我没有区域和凭据属性。

Where do I set the AWS properties?我在哪里设置 AWS 属性?

There's an Amazon AWS common options section in the README (see: old-app / new-app ), which includes the common AWS-specific properties one can override.自述文件中有一个Amazon AWS common options部分(请参阅: old-app / new-app ),其中包括可以覆盖的常见 AWS 特定属性。

You can pass them as inline properties in the stream definition or when deploying the stream by following the deployer properties convention.您可以在 stream 定义中或在部署 stream 时按照部署程序属性约定将它们作为内联属性传递。

I provided a detailed example of this use case here (with Minio S3).在此处提供了此用例的详细示例(使用 Minio S3)。 This works out of the box with the latest release of stream applications, but will require some customization if you are using previous versions.这适用于最新版本的 stream 应用程序,开箱即用,但如果您使用以前的版本,则需要进行一些自定义。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM