[英]BigQuery error in load operation: URI not found
I have, in the same GCP project, a BigQuery dataset and a cloud storage bucket, both within the region us-central1.在同一个 GCP 项目中,我在 us-central1 区域内有一个 BigQuery 数据集和一个云存储桶。 The storage bucket has a single parquet file located in it.
存储桶中有一个 parquet 文件。 When I run the below command:
当我运行以下命令时:
bq load \
--project_id=myProject --location=us-central1 \
--source_format=PARQUET \
myDataSet:tableName \
gs://my-storage-bucket/my_parquet.parquet
It fails with the below error:它失败并出现以下错误:
BigQuery error in load operation: Error processing job '[job_no]': Not found: URI gs://my-storage-bucket/my_parquet.parquet
Removing the --project_id
or --location
tags don't affect the outcome.删除
--project_id
或--location
标签不会影响结果。
Figured it out - the documentation is incorrect, I actually had to declare the source as gs://my-storage-bucket/my_parquet.parquet/part*
and it loaded fine弄清楚了-文档不正确,实际上我必须将源声明为
gs://my-storage-bucket/my_parquet.parquet/part*
并且加载正常
There has been some internal issues with BigQuery on 3rd March and it has been fixed now. BigQuery 在 3 月 3 日出现了一些内部问题,现已修复。
I have confirmed and used the following command to upload successfully a parquet file from Cloud Storage to BigQuery Table using bq command:我已经确认并使用以下命令使用 bq 命令成功地将 parquet 文件从 Cloud Storage 上传到 BigQuery Table:
bq load --project_id=PROJECT_ID \
--source_format=PARQUET \
DATASET.TABLE_NAME gs://BUCKET/FILE.parquet
Please note that according to the BigQuery Official Documentation , you have to declare the name of the table as following DATASET.TABLE_NAME
( In the post, I can see: instead of. )请注意,根据BigQuery 官方文档,您必须将表的名称声明为以下
DATASET.TABLE_NAME
(在帖子中,我可以看到:而不是。)
声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.