简体   繁体   English

Bigquery 加载作业表示成功但数据未加载到表中

[英]Bigquery load job said successful but data did not get loaded into table

I submitted a Bigquery load job, it ran and returned with the status successful.我提交了一个 Bigquery 加载作业,它运行并以成功状态返回。 But the data didn't make into the destintation table.但是数据没有进入目标表。

Here was the command that was run:这是运行的命令:

/usr/local/bin/bq load --nosynchronous_mode --project_id=ardent-course-601 --job_id=logsToBq_load_impressions_20140816_1674a956_6c39_4859_bc45_eb09db7ef99a --source_format=NEWLINE_DELIMITED_JSON dw_logs_impressions.impressions_20140816 gs://sm-uk-hadoop/queries/logsToBq_transformLogs/impressions/20140816/9307f6e3-0b3a-44ca-8571-7107c399998c/part* /opt/sm-analytics/projects/logsTobqMR/jsonschema/impressionsSchema.txt

I checked the job status of the job logsToBq_load_impressions_20140816_1674a956_6c39_4859_bc45_eb09db7ef99a.我检查了作业 logsToBq_load_impressions_20140816_1674a956_6c39_4859_bc45_eb09db7ef99a 的作业状态。 The input file count and size showed the correct number of input files and total size.输入文件数和大小显示了正确的输入文件数和总大小。

Does anyone know why the data didn't make into the table but yet the job is reported as successful?有谁知道为什么数据没有进入表格但工作报告为成功?

Just in case this is not a mistake on our side, I ran the load job again but to a different destination table and this time the data made into the destination table fine.以防万一这不是我们这边的错误,我再次运行了加载作业,但加载到另一个目标表,这次数据进入目标表很好。

Thank you.谢谢你。

This is very surprising, but I've confirmed via the logs that this is indeed the case.这非常令人惊讶,但我已经通过日志确认情况确实如此。

Unfortunately, the detailed logs for this job, which ran on August 16, are no longer available.不幸的是,8 月 16 日运行的这项作业的详细日志不再可用。 We're investigating whether this may have affected other jobs more recently.我们正在调查这是否可能影响最近的其他工作。 Please ping this thread if you see this issue again.如果您再次看到此问题,请 ping 此线程。

I experienced this recently with BigQuery in sandbox mode without a billing account.我最近在沙盒模式下使用 BigQuery 时遇到了这种情况,而没有结算帐户。 In this mode the partition expiration is automatically set to 60 days.在这种模式下,分区过期时间自动设置为 60 天。 If you load data into the table where the partitioned column(eg date) is older than 60 days it won't show up in the table.如果您将数据加载到分区列(例如日期)早于 60 天的表中,它将不会显示在表中。 The load job still succeeds with the correct number of output rows.加载作业仍然成功,输出行数正确。

we had this issue in our system and the reason was like table was set with partition expiry for 30 days and table was partitioned on timestamp column.. Hence when someone was ingesting data which is older than partition expiry date bigquery load jobs were successfully completed in Spark but we see no data in ingestion tables.. since it was getting deleted moment after it was ingested.. due to partition expiry set on.我们的系统中遇到了这个问题,原因是表的分区到期时间设置为 30 天,并且表在时间戳列上进行了分区。因此,当有人摄取早于分区到期日期的数据时,bigquery 加载作业已成功完成Spark,但我们在摄取表中看不到任何数据……因为它在摄取后立即被删除……由于设置了分区到期。

Please check your bigquery table partition expiry parameters and see the partition column value of incoming data.请检查您的 bigquery 表分区到期参数并查看传入数据的分区列值。 If it value will be lower than partition expiry.. you wont see data in bigquery tables.. it will get deleted just after the ingestion.如果它的值低于分区到期时间。你将不会在 bigquery 表中看到数据。它会在摄取后立即被删除。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 Bigquery作业在通过streamingAPI加载了数据的表上运行的结果 - Result of Bigquery job running on a table in which data is loaded via streamingAPI 加载作业后过时的BigQuery表 - Stale BigQuery table after load job BigQuery转移:Google Ads(以前称为AdWords):转移作业成功但没有数据 - BigQuery Transfer: Google Ads (formerly AdWords): Transfer job is successful but no data BigQuery加载作业不会插入所有数据 - BigQuery load job does not insert all data 尽管表格未更新,但BigQuery载入CSV文件“成功” - BigQuery load CSV file 'successful' although the table does not update Bigquery - BigQuery 加载作业后数据可用性延迟 - Bigquery - delay in data availability after BigQuery load job 将JSON数据加载到BigQuery表中 - Load JSON data into BigQuery table 未创建表且未使用云函数通过 load_table_from_dataframe 将数据加载到 BigQuery 数据集表 - Table is not being created and data is not loaded to BigQuery Dataset Table through load_table_from_dataframe using cloud functions 获取 Job 创建的 BigQuery 临时表以更快地读取大数据的最佳方法 - Best way to get BigQuery temp table created by Job to read large data faster BigQuery 加载作业在 JSON 中的布尔数据类型字段上失败 - BigQuery load job failing on boolean data type fields from JSON
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM