Managing the Kafka Connector¶

This topic describes the administrative tasks associated with managing the Kafka connector.

In this Topic:

Dropping Snowflake Objects Used by the Kafka Connector¶

If you no longer plan to load data into Snowflake tables using the Kafka connector, you can shut down Kafka and drop the Snowflake objects used by the connector.

The connector uses Snowflake objects of the following types to ingest data:

• Named internal stages

• Pipes

• Tables

This section provides instructions for finding and dropping the Snowflake objects used by the Kafka connector.

Dropping Stages¶

The connector creates one named internal stage for each Kafka topic. The format of the stage name is:

SNOWFLAKE_KAFKA_CONNECTOR_connector_name_STAGE_table_name

Note that each internal stage stores not only files to be loaded into tables, but also “state” information that is used to ensure delivery of rows from Kafka to the table.

If a stage and its state information are preserved, then if the connector is stopped and restarted, the connector automatically tries to resume at the point where it left off. However, if a stage is removed, the connector cannot resume where it left off.

To drop the stages used by the Kafka connector:

1. Find the names of the stages by executing SHOW STAGES as the stages owner (i.e. the role with the OWNERSHIP privilege on the stages. This should be the default role of the user defined in the Kafka configuration file to run the Kafka connector).

2. Execute DROP STAGE to drop each stage you want to remove from the system.

Dropping Pipes¶

The connector creates one pipe for each partition in a Kafka topic. The format of the pipe name is:

SNOWFLAKE_KAFKA_CONNECTOR_connector_name_PIPE_table_name_partition_number

To drop the pipes used by the Kafka connector:

1. Find the names of the pipes by executing SHOW PIPES as the pipes owner (i.e. the role with the OWNERSHIP privilege on the pipes. This should be the default role of the user defined in the Kafka configuration file to run the Kafka connector).

2. Execute DROP PIPE to drop each pipe you want to remove from the system.

Dropping Tables¶

If the data loaded into your target tables is no longer needed, you can also drop these tables.

If you did not map Kafka topics to tables using the snowflake.topic2table.map parameter in the Kafka Configuration Properties, the Kafka connector created new tables using the topic names. The table name is in uppercase but is otherwise identical to the topic name, as long as the topic name does not violate Snowflake object naming rules. For example, Snowflake creates a table name TEMPERATURE_DATA for a Kafka topic named temperature_data.

To drop the tables used by the Kafka connector:

1. Find the names of the tables by executing SHOW TABLES as the tables owner (i.e. the role with the OWNERSHIP privilege on the tables. This should be the default role of the user defined in the Kafka configuration file to run the Kafka connector).

2. Execute DROP TABLE to drop each table you want to remove from the system.