简体   繁体   中英

How can I run a python script in s3 via airflow

We usually store our code in s3 and call it from airflow to add them as new steps in emr. spark-submit can read data from s3 (like jars and files). But how can i call a python script that is present in s3 from airflow? It is a simple boto3 script. Please provide me some guidance.

I would think that its a round about way of actually running an emr job via ariflow. Why would you even need to add steps.Your easiest option would be to use the ssh operator that connects to emr and then do a spark-submit via the ssh operator. spark-submit will get your code from s3 and and then run the jobs.

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM