Openflow Connector for Confluence Data Center: Set up a connector¶
참고
The connector is subject to the Snowflake Connector Terms.
This topic describes the steps required to configure the Openflow Connector for Confluence Data Center.
As a data engineer, perform the following tasks to install and configure the connector:
Install the connector¶
To install the connector, do the following as a data engineer:
Navigate to the Openflow overview page. In the Featured connectors section, select View more connectors.
On the Openflow connectors page, find the connector and select Add to runtime.
In the Select runtime dialog, select your runtime from the Available runtimes drop-down list and click Add.
참고
Before you install the connector, ensure that you have created a database and schema in Snowflake for the connector to store ingested data.
Authenticate to the deployment with your Snowflake account credentials and select Allow when prompted to allow the runtime application to access your Snowflake account. The connector installation process takes a few minutes to complete.
Authenticate to the runtime with your Snowflake account credentials.
The Openflow canvas appears with the connector process group added to it.
Configure the connector¶
Right-click on the imported process group and select Parameters.
Enter the required parameter values as described in Flow parameters.
Flow parameters¶
This section describes the flow parameters that you can configure based on the following parameter contexts:
Atlassian Confluence source parameters: Used to establish a connection with the Confluence API.
Atlassian Confluence ingestion parameters: Contains all parameters from the other two parameter contexts and additional parameters specific to a given process group.
Atlassian Confluence destination parameters: Used to establish a connection with Snowflake.
Atlassian Confluence source parameters¶
Parameter |
Description |
Example values |
|---|---|---|
Confluence API Personal Access Token |
Confluence API Personal Access Token used for authentication. |
AAABBBCCC |
Confluence Site URL |
Confluence site from which the connector will ingest content. |
Additionally, for the use case with ACLs:
Parameter |
Description |
Example value |
|---|---|---|
Confluence Admin API Personal Access Token |
Confluence admin API Personal Access Token used for authentication. It’s issued to gather audit logs. |
AAABBBCCCDDD |
Atlassian Confluence ingestion parameters¶
Parameter |
Description |
Example value |
|---|---|---|
CDC Refresh Frequency |
Frequency of ingestion runs. Set a refresh frequency between 1 and 10080 minutes (seven days) that applies globally to all pages ingested. |
1440 minutes |
Confluence Space Keys |
Comma-separated list of Confluence space keys to ingest. If not specified, all spaces will be ingested. |
SPACEKEY1, SPACEKEY2 |
Confluence Page IDs |
Comma-separated list of Confluence Page IDs to ingest. If not specified, all pages will be ingested. |
87423,13342 |
Start Date |
Optional: Filter content updated after specified date and (optionally) time (format: yyyy-mm-dd hh:mm:ss) |
2025-07-30 14:30:05 |
Additionally, for the Cortex use case:
Parameter |
Description |
Example value |
|---|---|---|
OCR Mode |
The OCR Mode to use when parsing files with the Cortex AI_PARSE_DOCUMENT function. This can be OCR or LAYOUT. Choose OCR for simple text documents and LAYOUT for documents with tables and other structures that need to be preserved. |
OCR |
Snowflake Cortex Search Service User Role |
The role that is assigned usage permissions on the Cortex Search service. |
<CORTEX_ROLE> |
Atlassian Confluence destination parameters¶
Parameter |
Description |
Required |
|---|---|---|
Destination Database |
The database where data will be persisted. It must already exist in Snowflake. The name is case-sensitive. For unquoted identifiers, provide the name in uppercase. |
Yes |
Destination Schema |
The schema where data will be persisted, which must already exist in Snowflake. The name is case-sensitive. For unquoted identifiers, provide the name in uppercase. See the following examples:
|
Yes |
Snowflake Authentication Strategy |
When using:
|
Yes |
Snowflake Account Identifier |
When using:
|
Yes |
Snowflake Private Key |
When using:
|
No |
Snowflake Private Key File |
When using:
|
No |
Snowflake Private Key Password |
When using
|
No |
Snowflake Role |
When using
|
Yes |
Snowflake Username |
When using
|
Yes |
Oversized Value Strategy |
Determines how the connector handles values that exceed its internal size limits (16 MB) during replication. Possible values are:
|
No |
Snowflake Warehouse |
Snowflake warehouse used to run queries. |
Yes |
Run the flow¶
Right-click in the control plane and select Enable all Controller Services.
Right-click in the imported process group and select Start. The connector starts the data ingestion.
After starting the flow, the connector creates all required Snowflake objects, such as tables and the Cortex Search service (only for the Cortex use case). The connector then begins ingesting data into stages and tables in the specified destination schema.
참고
On the first connector run, the connector creates required Snowflake objects, including tables.
As a result of new object creation, transient errors may occur, which can be ignored. The most common examples of such errors are SQL or Snowpipe execution errors due to destination table not found. Once the objects are created, these errors will no longer occur.
When updating the connector configuration, you must reset the processor state.
To reset the processor state, perform the following steps:
Right-click the Get Confluence Page Ids processor and select Stop.
Right-click the Get Confluence Page Ids processor and select View State.
In the State dialog box, select Clear State.
This task clears the state of the processor and allows it to fetch all pages again.