Unloading into Microsoft Azure¶
If you already have a Microsoft Azure account and use Azure containers for storing and managing your files, you can make use of your existing containers and folder paths when unloading data from Snowflake tables. This topic describes how to use the COPY command to unload data from a table into an Azure container. You can then download the unloaded data files to your local file system.
Snowflake supports the following types of blob storage accounts:
Data Lake Storage Gen2
Snowflake does not support Data Lake Storage Gen1.
As illustrated in the diagram below, unloading data into an Azure container is performed in two steps:
- Step 1
Use the COPY INTO <location> command to copy the data from the Snowflake database table into one or more files in an Azure container bucket. In the command, you specify a named external stage object that references the Azure container (recommended) or you can choose to unload directly to the container by specifying the URI and security credentials (if required) for the container.
Regardless of the method you use, this step requires a running, current virtual warehouse for the session. The warehouse provides the compute resources to write rows from the table.
- Step 2
Use the interfaces/tools provided by Microsoft to download the files from the Azure container.
The instructions in this set of topics assume you have read Preparing to Unload Data and have created a named file format, if desired.
Before you begin, you may also want to read Data Unloading Considerations for best practices, tips, and other guidance.
In this Topic:
Allowing the Azure Virtual Network Subnet IDs¶
If an Azure administrator in your organization has not explicitly granted Snowflake access to your Azure storage account, you can do so now. Follow the steps in Allowing the VNet Subnet IDs in the data loading configuration instructions.
Configuring an Azure Container for Unloading Data¶
For Snowflake to write to an Azure container, you must generate a shared access signature (SAS) token for your storage access account. For configuration instructions, see Configuring an Azure Container for Loading Data.
Unloading Data into an External Stage¶
External stages are named database objects that provide the greatest degree of flexibility for data unloading. Because they are database objects, privileges for named stages can be granted to any role.
You can create an external named stage using either the web interface or SQL:
- Classic Web Interface
Creating a Named Stage¶
The following example creates an external stage named
my_ext_unload_stage with a container named
mycontainer and a folder path named
unload. The stage references the named file format object called
my_csv_unload_format that was created in Preparing to Unload Data:
CREATE OR REPLACE STAGE my_ext_unload_stage URL='azure://myaccount.blob.core.windows.net/mycontainer/unload' CREDENTIALS=(AZURE_SAS_TOKEN='?sv=2016-05-31&ss=b&srt=sco&sp=rwdl&se=2018-06-27T10:05:50Z&st=2017-06-27T02:05:50Z&spr=https,http&sig=bgqQwoXwxzuD2GJfagRg7VOS8hzNr3QLT7rhS8OFRLQ%3D') ENCRYPTION=(TYPE='AZURE_CSE' MASTER_KEY = 'kPxX0jzYfIamtnJEUTHwq80Au6NbSgPH5r4BDDwOaO8=') FILE_FORMAT = my_csv_unload_format;
blob.core.windows.net endpoint for all supported types of Azure blob storage accounts, including Data Lake Storage Gen2.
Note that the AZURE_SAS_TOKEN and MASTER_KEY values used in this example are for illustration purposes only.
Unloading Data to the Named Stage¶
Use the COPY INTO <location> command to unload data from a table into an Azure container using the external stage.
The following example uses the
my_ext_unload_stagestage to unload all the rows in the
mytabletable into one or more files into the Azure container. A
d1filename prefix is applied to the files:
COPY INTO @my_ext_unload_stage/d1 from mytable;
Use the tools provided by Azure to retrieve the objects (i.e. files generated by the command) from the container.
Unloading Data Directly into an Azure Container¶
Use the COPY INTO <location> command to unload data from a table directly into a specified Azure container. This option works well for ad hoc unloading, when you aren’t planning regular data unloading with the same table and container parameters.
You must specify the URI for the Azure container and the security credentials for accessing the container in the COPY command.
The following example unloads all the rows in the
mytabletable into one or more files with the folder path prefix
unload/in an Azure container.
This example references a storage integration created using CREATE STORAGE INTEGRATION by an account administrator (i.e. a user with the ACCOUNTADMIN role) or a role with the global CREATE INTEGRATION privilege. A storage integration allows users to avoid supplying credentials to access a private storage location:
COPY INTO 'azure://myaccount.blob.core.windows.net/mycontainer/unload/' FROM mytable STORAGE_INTEGRATION = myint;
Use the Azure console (or equivalent client application) to retrieve the objects (i.e. files generated by the command) from the container.