Openflow Connector for Confluence Data Center: Set up a connector

Bemerkung

The connector is subject to the Snowflake Connector Terms.

This topic describes the steps required to configure the Openflow Connector for Confluence Data Center.

As a data engineer, perform the following tasks to install and configure the connector:

Install the connector

To install the connector, do the following as a data engineer:

  1. Navigate to the Openflow overview page. In the Featured connectors section, select View more connectors.

  2. On the Openflow connectors page, find the connector and select Add to runtime.

  3. In the Select runtime dialog, select your runtime from the Available runtimes drop-down list and click Add.

    Bemerkung

    Before you install the connector, ensure that you have created a database and schema in Snowflake for the connector to store ingested data.

  4. Authenticate to the deployment with your Snowflake account credentials and select Allow when prompted to allow the runtime application to access your Snowflake account. The connector installation process takes a few minutes to complete.

  5. Authenticate to the runtime with your Snowflake account credentials.

The Openflow canvas appears with the connector process group added to it.

Configure the connector

  1. Right-click on the imported process group and select Parameters.

  2. Enter the required parameter values as described in Flow parameters.

Flow parameters

This section describes the flow parameters that you can configure based on the following parameter contexts:

Atlassian Confluence source parameters

Parameter

Description

Example values

Confluence API Personal Access Token

Confluence API Personal Access Token used for authentication.

AAABBBCCC

Confluence Site URL

Confluence site from which the connector will ingest content.

https://your-confluence.atlassian.net

Additionally, for the use case with ACLs:

Parameter

Description

Example value

Confluence Admin API Personal Access Token

Confluence admin API Personal Access Token used for authentication. It’s issued to gather audit logs.

AAABBBCCCDDD

Atlassian Confluence ingestion parameters

Parameter

Description

Example value

CDC Refresh Frequency

Frequency of ingestion runs. Set a refresh frequency between 1 and 10080 minutes (seven days) that applies globally to all pages ingested. 1 day, 24 hours, 1440 minutes, 86400 seconds are all valid inputs and mean the same frequency.

1440 minutes

Confluence Space Keys

Comma-separated list of Confluence space keys to ingest. If not specified, all spaces will be ingested.

SPACEKEY1, SPACEKEY2

Confluence Page IDs

Comma-separated list of Confluence Page IDs to ingest. If not specified, all pages will be ingested.

87423,13342

Start Date

Optional: Filter content updated after specified date and (optionally) time (format: yyyy-mm-dd hh:mm:ss)

2025-07-30 14:30:05

Additionally, for the Cortex use case:

Parameter

Description

Example value

OCR Mode

The OCR Mode to use when parsing files with the Cortex AI_PARSE_DOCUMENT function. This can be OCR or LAYOUT. Choose OCR for simple text documents and LAYOUT for documents with tables and other structures that need to be preserved.

OCR

Snowflake Cortex Search Service User Role

The role that is assigned usage permissions on the Cortex Search service.

<CORTEX_ROLE>

Atlassian Confluence destination parameters

Parameter

Description

Required

Destination Database

The database where data will be persisted. It must already exist in Snowflake. The name is case-sensitive. For unquoted identifiers, provide the name in uppercase.

Yes

Destination Schema

The schema where data will be persisted, which must already exist in Snowflake. The name is case-sensitive. For unquoted identifiers, provide the name in uppercase.

See the following examples:

  • CREATE SCHEMA SCHEMA_NAME or CREATE SCHEMA schema_name: use SCHEMA_NAME

  • CREATE SCHEMA "schema_name" or CREATE SCHEMA "SCHEMA_NAME": use schema_name or SCHEMA_NAME, respectively

Yes

Snowflake Authentication Strategy

When using:

  • Snowflake Openflow Deployment or BYOC: Use SNOWFLAKE_MANAGED_TOKEN. This token is managed automatically by Snowflake. BYOC deployments must have previously configured runtime roles to use SNOWFLAKE_MANAGED_TOKEN.

  • BYOC: Alternatively BYOC can use KEY_PAIR as the value for authentication strategy.

Yes

Snowflake Account Identifier

When using:

  • Session Token Authentication Strategy: Must be blank.

  • KEY_PAIR: Snowflake account name formatted as [organization-name]-[account-name] where data will be persisted.

Yes

Snowflake Private Key

When using:

  • Session Token Authentication Strategy: Must be blank.

  • KEY_PAIR: Must be the RSA private key used for authentication.

    The RSA key must be formatted according to PKCS8 standards and have standard PEM headers and footers. Note that either a Snowflake Private Key File or a Snowflake Private Key must be defined.

No

Snowflake Private Key File

When using:

  • Session token authentication strategy: The private key file must be blank.

  • KEY_PAIR: Upload the file that contains the RSA private key used for authentication to Snowflake, formatted according to PKCS8 standards and including standard PEM headers and footers. The header line begins with -----BEGIN PRIVATE. To upload the private key file, select the Reference asset checkbox.

No

Snowflake Private Key Password

When using

  • Session Token Authentication Strategy: Must be blank.

  • KEY_PAIR: Provide the password associated with the Snowflake private key file.

No

Snowflake Role

When using

  • Session Token Authentication Strategy: Use your Snowflake role. You can find your Snowflake role in the Openflow UI, by navigating to View Details for your Runtime.

  • KEY_PAIR Authentication Strategy: Use a valid role configured for your service user.

Yes

Snowflake Username

When using

  • Session Token Authentication Strategy: Must be blank.

  • KEY_PAIR: Provide the user name used to connect to the Snowflake instance.

Yes

Oversized Value Strategy

Determines how the connector handles values that exceed its internal size limits (16 MB) during replication. Possible values are:

  • Fail Table (default): The table is marked as permanently failed, and replication stops for that table.

  • Set Null: The value is replaced with NULL in the destination table. Use this to prevent table failures when it is acceptable to lose data in tables beyond the oversized value.

No

Snowflake Warehouse

Snowflake warehouse used to run queries.

Yes

Run the flow

  1. Right-click in the control plane and select Enable all Controller Services.

  2. Right-click in the imported process group and select Start. The connector starts the data ingestion.

After starting the flow, the connector creates all required Snowflake objects, such as tables and the Cortex Search service (only for the Cortex use case). The connector then begins ingesting data into stages and tables in the specified destination schema.

Bemerkung

On the first connector run, the connector creates required Snowflake objects, including tables.

As a result of new object creation, transient errors may occur, which can be ignored. The most common examples of such errors are SQL or Snowpipe execution errors due to destination table not found. Once the objects are created, these errors will no longer occur.

When updating the connector configuration, you must reset the processor state.

To reset the processor state, perform the following steps:

  1. Right-click the Get Confluence Page Ids processor and select Stop.

  2. Right-click the Get Confluence Page Ids processor and select View State.

  3. In the State dialog box, select Clear State.

This task clears the state of the processor and allows it to fetch all pages again.

Next steps

Monitor the flow.