简体   繁体   English

在 GCP 中使用 Cloud Functions 抓取页面并将其加载到 BigQuery

[英]Scrape page using Cloud Functions in GCP and load it to BigQuery

I have got question about GCP.我有关于 GCP 的问题。 Is it possible to scrape page on a daily basis using Cloud Functions and send it to BigQuery.是否可以使用 Cloud Functions 每天抓取页面并将其发送到 BigQuery。 I have already wrote working code in Python that scrapes page and save it to csv and now I would like to move it to cloud.我已经用 Python 编写了可抓取页面并将其保存到 csv 的工作代码,现在我想将其移动到云中。 How can I do that?我怎样才能做到这一点?

You can load your scraped ( csv ) data to Google Cloud Storage from python.您可以将抓取的 ( csv ) 数据从 python 加载到 Google Cloud Storage。 Bigquery supports loading data as a new table or appending new data to a table. Bigquery 支持将数据作为新表加载或将新数据附加到表中。 The latter requires you to do it in python, the former can be done using the UI in the cloud console.后者需要你在python中完成,前者可以使用云控制台中的UI来完成。

More info Here And Here更多信息在这里和这里

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM