简体   繁体   English

谷歌云数据流 JDBC 到 BigQuery 作业只是挂起并且没有产生有用的日志/错误消息

[英]Google Cloud Data flow JDBC to BigQuery job just hanging and producing no helpful logs/error message

I am trying to use the Google provided Dataflow template JDBC to BigQuery.我正在尝试将 Google 提供的 Dataflow 模板 JDBC 用于 BigQuery。 I have an on-prem SQL Server that I want to connect to.我有一个要连接的本地 SQL 服务器。 I have configured all the necessary networking and firewall rules to allow the Dataflow workers to communicate with the on-prem server.我已经配置了所有必要的网络和防火墙规则,以允许 Dataflow 工作人员与本地服务器进行通信。 I have also confirmed that the provided connection string and credentials work.我还确认提供的连接字符串和凭据有效。

This issue is that when I run the job, it just hangs and doesn't produce any helpful log/errors.这个问题是当我运行该作业时,它只是挂起并且不会产生任何有用的日志/错误。 I have check the logs explorer and can't seem to find anything to help troubleshoot this.我检查了日志资源管理器,似乎找不到任何帮助解决此问题的方法。

If I let it run, eventually it hits a timeout and produces this: useless log如果我让它运行,最终它会超时并产生: useless log

I would suggest opening a support ticket, as it may be necessary to have the Job ID to see exactly what went wrong.我建议打开支持票,因为可能需要有作业 ID 才能准确查看问题所在。

Such symptom can be observed when created workers can't be reached by Dataflow -- either by having the subnetwork mismatching the job region, or because of strict firewall rules.当 Dataflow 无法访问创建的工作人员时,可以观察到这种症状——要么是由于子网与作业区域不匹配,要么是由于严格的防火墙规则。

Usually when I have dataflow job hanging, I open the full log query and I keep only the job id in the log query.通常当我挂起数据流作业时,我会打开完整的日志查询,并且只在日志查询中保留作业 ID。

The log query by default is:默认的日志查询是:

resource.type="dataflow_step"
resource.labels.job_id="job_id"
logName=("projects/project/logs/dataflow.googleapis.com%2Fjob-message" OR "projects/project/logs/dataflow.googleapis.com%2Flauncher")

I just change it to:我只是将其更改为:

resource.labels.job_id="job_id"

This usually gives me more details about the pipeline and what's wrong.这通常会为我提供有关管道和问题所在的更多详细信息。 It happened with me a lot when dealing with customer containers and flex templates etc.在处理客户容器和弹性模板等时,我经常遇到这种情况。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 从 Cloud 触发数据流作业时 dill 出错 Function - Error with dill when triggering a Data Flow job from Cloud Function 谷歌云数据流示例 - Google cloud data flow exmaple 使用 Golang 读取 Google Cloud Pubsub 消息并写入 BigQuery - Read Google Cloud Pubsub message and write to BigQuery using Golang 只需将 Cloud SQL 中的新数据导入 BigQuery? - Just import new data from Cloud SQL into BigQuery? Airflow 2:将数据从 BigQuery 传输到 Cloud Storage 时找不到作业 - Airflow 2: Job Not Found when transferring data from BigQuery into Cloud Storage 将表从 Google BigQuery 导出到 Google Cloud Storage 时出现权限错误 - Permissions Error Exporting a table from Google BigQuery to Google Cloud Storage 在谷歌云中运行 Apache Beam 作业时找不到错误模块 - Error module not found while running Apache beam job in google cloud 错误:模块“google.cloud.bigquery_storage”没有属性“BigQueryReadClient” - Error: module 'google.cloud.bigquery_storage' has no attribute 'BigQueryReadClient' 谷歌云 Function 不返回错误消息 - Google Cloud Function Not Returning Error Message Google Cloud DataStream to Bigquery 模板无法将数据同步到大查询 - Google Cloud DataStream to Bigquery template not able to sync data to big query
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM