Preparing to Load Data Using the Snowpipe REST API

This topic describes how to get started with Snowpipe when calling the REST API, including instructions for installing the required client SDK, creating a stage (if needed) and pipe, and the one-time security setup for each Snowpipe user.

In this Topic:


The instructions in this section assume you already have a target table in your Snowflake database where your data will be loaded.

Client Requirement (Java or Python SDK)

The Snowpipe service requires either the Java SDK or Python SDK. These SDKs are provided by Snowflake for your convenience.


The binaries are provided as Client Software under the terms of your master service agreement (MSA) with Snowflake.

Installing the Java SDK

  1. Download the Java SDK installer from the Maven Central Repository: (or

  2. Integrate the JAR file into an existing project.


The developer notes are hosted with the source code on GitHub.

Installing the Python SDK

Note that the Python SDK requires Python 3.6 or higher.

To install the SDK, execute the following command:

pip install snowflake-ingest

Alternatively, download the wheel file from PyPI and integrate it into an existing project.


The developer notes are hosted with the source code on GitHub.

Step 1: Create a Stage (If Needed)

Snowpipe supports loading from the following stage types:

  • Named internal (Snowflake) or external (Amazon S3, Google Cloud Storage, or Microsoft Azure) stages

  • Table stages

Create a named stage using the CREATE STAGE command, or you can choose to use an existing stage. You will stage your files temporarily before Snowpipe loads them into your target table.

Step 2: Create a Pipe

Create a new pipe in the system for defining the COPY INTO <table> statement used by Snowpipe to load data from an ingestion queue into tables. For information, see CREATE PIPE.


Creating a pipe requires the CREATE PIPE access control privilege, as well as the USAGE privilege on the database, schema and stage.

For example, create a pipe in the mydb.myschema schema that loads all the data from files staged in the mystage stage into the mytable table:

create pipe mydb.myschema.mypipe if not exists as copy into mydb.myschema.mytable from @mydb.myschema.mystage;

Step 3: Configure Security (Per User)

For each user who will execute continuous data loads using Snowpipe, generate a public-private key pair for making calls to the Snowpipe REST endpoints. In addition, grant sufficient privileges on the objects for the data load (i.e. the target database, schema, and table), the stage object, and the pipe.

If you plan to restrict Snowpipe data loads to a single user, you only need to configure key pair authentication for the user once. After that, you only need to grant access control privileges on the database objects used for each data load.


To follow the general principle of least privilege, we recommend creating a separate user and role to use for ingesting files using a pipe. The user should be created with this role as its default role.

Using Key Pair Authentication & Key Rotation

Snowflake supports using key pair authentication and key pair rotation.

Key pair authentication is an authentication option for supported Snowflake clients.

Key pair rotation provides enhanced key pair security.

To configure these features, see Key Pair Authentication & Key Pair Rotation.

Granting Access Privileges

Using Snowpipe requires a role with the following permissions:




Named pipe


Named stage

USAGE (external stages) , READ (internal stages)

Named file format


Optional; only needed if the either the stage (see Step 1: Create a Stage (If Needed)) or the pipe (see Step 2: Create a Pipe) references a named file format.

Target database


Target schema


Target table


Use the GRANT <privileges> … TO ROLE command to grant these privileges to the role.


Only security administrators (i.e. users with the SECURITYADMIN role) or higher, or another role with both the CREATE ROLE privilege on the account and the global MANAGE GRANTS privilege, can create roles and grant privileges.

For example, create a role named snowpipe1 that can access a set of mydb.myschema database objects as well as a pipe named mypipe. Grant the role to a user.

 -- Create a role to contain the Snowpipe privileges
use role securityadmin;

create or replace role snowpipe1;

-- Grant the required privileges on the database objects
grant usage on database mydb to role snowpipe1;

grant usage on schema mydb.myschema to role snowpipe1;

grant insert, select on mydb.myschema.mytable to role snowpipe1;

grant usage, read on stage mydb.myschema.mystage to role snowpipe1;

-- Grant the OWNERSHIP privilege on the pipe object
grant ownership on pipe mydb.myschema.mypipe to role snowpipe1;

-- Grant the role to a user
grant role snowpipe1 to user jsmith;

-- Set the role as the default role for the user
alter user jsmith set default_role = snowpipe1;

Step 4: Stage Data Files

Copy data files to the internal or external stage you created for loading files using Snowpipe.

  • Copy files to an external stage using the tools provided by the cloud storage service.

  • Copy files to an internal stage using the PUT command.


    If your Snowflake account is hosted on Amazon Web Services, we recommend always using the PUT … OVERWRITE = TRUE syntax.

    Amazon S3 provides read-after-write consistency for new objects created in a bucket. However, if a HEAD or GET request for an object is made before it is created, then S3 provides eventual consistency for the object. This means that an immediate request for a new object after it is created could return a file not found exception. Setting the OVERWRITE = TRUE parameter avoids the initiation of a HEAD request prior to the creation of the object in the S3 bucket.

    For more information about the S3 consistency model, see the S3 documentation.

Next: Learn how to call the public REST endpoints to load data and retrieve load history reports, in Calling Snowpipe REST Endpoints to Load Data.