简体   繁体   English

将数据从 Azure Synapse Serverless SQL 池批量加载到 Azure 存储或 Databricks Spark 的最佳方法

[英]Best way to bulk load data from Azure Synapse Serverless SQL pools into Azure storage or Databricks Spark

I am trying to bulk load data from Azure Synapse serverless SQL pools into Azure Storage or directly into Databricks Spark (using JDBC driver). I am trying to bulk load data from Azure Synapse serverless SQL pools into Azure Storage or directly into Databricks Spark (using JDBC driver). What is the best way to do this bulk loading assuming we only know the external table name and don't know the location of the file underneath?假设我们只知道外部表名并且不知道下面文件的位置,那么进行这种批量加载的最佳方法是什么? Is there any metadata query to know the location of the file as well?是否有任何元数据查询也可以知道文件的位置?

The files are already in Azure storage since Synapse Serverless SQL has no “in-database” storage.这些文件已经在 Azure 存储中,因为 Synapse Serverless SQL 没有“数据库内”存储。

Assuming it's an external table (not a view with OPENROWSET) then sys.external_tables has a location column with the path to the file or folder.假设它是一个外部表(不是带有 OPENROWSET 的视图),那么sys.external_tables有一个 location 列,其中包含文件或文件夹的路径。

If you don't already know which storage account and container it's in, you may need to join to sys.external_data_sources for that information.如果您还不知道它所在的存储帐户和容器,您可能需要加入sys.external_data_sources以获取该信息。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM