简体   繁体   English

Bigquery 在流式传输时无法加载所有数据

[英]Bigquery cannot load all data when streaming

Hi i am streaming data from my apps to Bigquery by C++.嗨,我正在通过 C++ 将数据从我的应用程序流式传输到 Bigquery。 Thing got okay and the all are connectable, but the problem is the log file said there's 665 in streaming buffers enter image description here一切正常,一切都可以连接,但问题是日志文件说流缓冲区中有 665 在此处输入图像描述

However, final records in the table is just 4. Does anyone know to solve this?但是,表中的最终记录只有 4 条。有人知道解决这个问题吗? enter image description here在此处输入图像描述

"Estimated rows" are only an estimate. “估计行数”只是一个估计值。

Streaming data in BigQuery is available in real-time (though table copy commands can take up to 90 minutes). BigQuery 中的流式数据是实时可用的(尽管表复制命令最多可能需要 90 分钟)。 I recommend reading this article for more information.我建议阅读这篇文章以获取更多信息。

It sounds like you think you're losing data.听起来您认为自己正在丢失数据。 That's not likely.那不太可能。 I recommend checking what you believe is being inserted versus what's actually landing in the table.我建议检查您认为插入的内容与实际进入表格的内容。

Thank you for your great support.感谢您的大力支持。 We found out that the issue happen because in the code we wrote, "insertId" should be unique, otherwise, BQ will deduplicate row with similar insertID.我们发现问题的发生是因为在我们编写的代码中,“insertId”应该是唯一的,否则,BQ 会删除具有相似 insertID 的行。 In the code, it happened since we set insertID = 'rowdetail' only.在代码中,它发生是因为我们只设置了 insertID = 'rowdetail'。 After changing it to time inserted (ddmmyyhhmmss), the pipeline run normally.将其更改为时间插入(ddmmyyhhmmss)后,管道运行正常。 在此处输入图像描述

Hope it helpful for you and everyone希望对你和大家有帮助

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM