All processors (alphabetical)¶
This topic provides a list of all Snowflake openflow processors in alphabetical order. The list includes:
The name of each processor
A summary of each processor
A¶
Processor |
Description |
|
---|---|---|
Generates a CSV representation of the input FlowFile Attributes. |
||
Generates a JSON representation of the input FlowFile Attributes. |
C¶
Processor |
Description |
|
---|---|---|
Counts the number of Records in a record set, optionally counting the number of elements per category, where the categories are defined by user-defined properties. |
||
Snowflake |
Reads CDC events from a MySQL database. |
|
Snowflake |
Reads CDC events from a PostgreSQL database. |
|
Snowflake |
Captures changes to a Shared Google Drive and emits a FlowFile for each change that occurs. |
|
Snowflake |
Captures Microsoft365 groups changes and emits a FlowFile for each change that occurs. |
|
Snowflake |
Captures changes from a Sharepoint Document Library and emits a FlowFile for each change that occurs. |
|
Snowflake |
Processor checking if the Meta Ads report is ready for download. |
|
Snowflake |
Given an input Openflow Document, chunks the data into segments that are more applicable for LLM synthesis or semantic embedding. |
|
Snowflake |
Chunks text with options for recursively splitting by delimiters and max character length. |
|
Snowflake |
Chunks text with options for recursively splitting by delimiters and max character length. |
|
Compresses or decompresses the contents of FlowFiles using a user-specified compression algorithm and updates the mime. |
||
Acts as a WebSocket client endpoint to interact with a remote WebSocket server. |
||
Consumes AMQP Messages from an AMQP Broker using the AMQP 0. |
||
Receives messages from Microsoft Azure Event Hubs with checkpointing to ensure consistent event processing. |
||
Consumes Enterprise Events from Box admin_logs_streaming Stream Type. |
||
Consumes all events from Box. |
||
A processor that repeatedly runs a paginated query against a field using a Range query to consume new Documents from an Elasticsearch index/query. |
||
Consumes messages from the configured Google Cloud PubSub subscription. |
||
Consumes messages from Email Server using IMAP protocol. |
||
Consumes JMS Message of type BytesMessage, TextMessage, ObjectMessage, MapMessage or StreamMessage transforming its content to a FlowFile and transitioning it to ‘success’ relationship. |
||
Snowflake |
Consumes messages from Apache Kafka Consumer API. |
|
Reads data from the specified AWS Kinesis stream and outputs a FlowFile for every processed Record (raw) or a FlowFile for a batch of processed records if a Record Reader and Record Writer are configured. |
||
Subscribes to a topic and receives messages from an MQTT broker |
||
Consumes messages from Email Server using POP3 protocol. |
||
Retrieves messages from one or more configured Slack channels. |
||
Snowflake |
Retrieves messages from Slack conversations available to the App. |
|
Snowflake |
Fetches historical messages from all Slack channels available to the App. |
|
Snowflake |
Fetches data from a Snowflake stream and writes it to a FlowFile. |
|
Streams tweets from Twitter’s streaming API v2. |
||
Controls the rate at which data is transferred to follow-on processors. |
||
Converts a FlowFile’s content from one character set to another |
||
Snowflake |
Converts a Open Office compatible file, to a PDF or Docx format. |
|
Snowflake |
Converts a PDF file into a series of images, one for each page. |
|
Converts records from one data format to another using configured Record Reader and Record Write Controller Services. |
||
Snowflake |
Converts the incoming database schema into the appropriate schema for a Snowflake CDC Journal table. |
|
Copies a blob in Azure Blob Storage from one account/container to another. |
||
Copies a file from one bucket and key to another in AWS S3 |
||
Counts various metrics on incoming text. |
||
Snowflake |
Processor which creates report configuration for Amazon Ads connector. |
|
Snowflake |
Uses Azure OpenAI to create embeddings for text. |
|
Snowflake |
Uses Cohere to create embeddings for text. |
|
Snowflake |
Processor which creates report configuration for Meta Ads connector. |
|
Snowflake |
Uses Ollama to create embeddings for text. |
|
Snowflake |
Uses OpenAI to create embeddings for text. |
|
Snowflake |
Create vector embeddings using Snowflake Cortex Large Language Model functions |
|
Snowflake |
Uses VertexAI to create embeddings for text. |
|
Calculates a cryptographic hash value for the flowfile content using the given algorithm and writes it to an output attribute. |
D¶
Processor |
Description |
|
---|---|---|
The DebugFlow processor aids testing and debugging the FlowFile framework by allowing various responses to be explicitly triggered in response to the receipt of a FlowFile or a timer event without a FlowFile if using timer or cron based scheduling. |
||
Decrypt content using the age-encryption. |
||
Decrypt contents of OpenPGP messages. |
||
This processor de-duplicates individual records within a record set. |
||
Deletes the specified blob from Azure Blob Storage. |
||
Deletes the provided file from Azure Data Lake Storage |
||
Delete from an Elasticsearch index using a query. |
||
Snowflake |
Delete a DBFS files and directories. |
|
Deletes a document from DynamoDB based on hash and range key. |
||
Deletes a file from the filesystem. |
||
Deletes objects from a Google Cloud Bucket. |
||
Deletes a file from GridFS using a file name or a query. |
||
Snowflake |
Deletes vectors from Milvus database from a collection by ID. |
|
Executes a delete query against a MongoDB collection. |
||
Snowflake |
Deletes vectors from a Pinecone index. |
|
Deletes a file from an Amazon S3 Bucket. |
||
Deletes a file residing on an SFTP server. |
||
Deletes a message from an Amazon Simple Queuing Service Queue |
||
Snowflake |
Delete a Unity Catalog file or directory. |
|
Caches a value, computed from FlowFile attributes, for each incoming FlowFile and determines if the cached value has already been seen. |
||
Distributes FlowFiles to downstream processors based on a Distribution Strategy. |
||
Intended for load testing, this processor will create the configured number of copies of each incoming FlowFile. |
E¶
Processor |
Description |
|
---|---|---|
Encode or decode the contents of a FlowFile using Base64, Base32, or hex encoding schemes |
||
Encrypt content using the age-encryption. |
||
Encrypt contents using OpenPGP. |
||
Enforces expected ordering of FlowFiles that belong to the same data group within a single node. |
||
Snowflake |
Looks up a value using the configured Lookup Service and adds the results to the FlowFile as one or more attributes. |
|
Snowflake |
Enriches incoming FlowFiles that come from CaptureChangePostgreSQL, etc. |
|
Evaluates one or more JsonPath expressions against the content of a FlowFile. |
||
Snowflake |
Evaluates the correctness of generated answers in a Retrieval-Augmented Generation (RAG) context by computing metrics such as F1 score, cosine similarity, and answer correctness. |
|
Snowflake |
Evaluates the faithfulness of generated answers in a Retrieval-Augmented Generation (RAG) system by analyzing responses using an LLM (e. |
|
Snowflake |
Calculates retrieval metrics (Precision@N, Recall@N, FScore@N, MAP@N, MRR) for a RAG system using an LLM as a judge. |
|
Evaluates one or more XPaths against the content of a FlowFile. |
||
Evaluates one or more XQueries against the content of a FlowFile. |
||
Experimental Extended Groovy script processor. |
||
Runs an operating system command specified by the user and writes the output of that command to a FlowFile. |
||
Experimental - Executes a script given the flow file and a process session. |
||
Executes provided SQL select query. |
||
Executes provided SQL select query. |
||
Snowflake |
Executes a SQL DDL or DML Statement against a database. |
|
The ExecuteStreamCommand processor provides a flexible way to integrate external commands and scripts into NiFi data flows. |
||
Extracts metadata from the header of an Avro datafile. |
||
Snowflake |
Extracts the text from a Document and writes it to the FlowFile content. |
|
Extract attachments from a mime formatted email file, splitting them into individual flowfiles. |
||
Using the flowfile content as source of data, extract header from an RFC compliant email file adding the relevant attributes to the flowfile. |
||
Evaluates one or more Grok Expressions against the content of a FlowFile, adding the results as attributes or replacing the content of the FlowFile with a JSON notation of the matched content |
||
Extracts the record schema from the FlowFile using the supplied Record Reader and writes it to the ‘avro. |
||
Evaluates one or more Regular Expressions against the content of a FlowFile. |
F¶
Processor |
Description |
|
---|---|---|
Retrieves the specified blob from Azure Blob Storage and writes its content to the content of the FlowFile. |
||
Fetch the specified file from Azure Data Lake Storage |
||
Fetches files from a Box Folder. |
||
Fetches metadata for files from Box and adds it to the FlowFile’s attributes. |
||
Fetches a Box file representation using a representation hint and writes it to the FlowFile content. |
||
Computes cache key(s) from FlowFile attributes, for each incoming FlowFile, and fetches the value(s) from the Distributed Map Cache associated with each key. |
||
Fetches files from Dropbox. |
||
Reads the contents of a file from disk and streams it into the contents of an incoming FlowFile. |
||
Fetches the content of a file from a remote FTP server and overwrites the contents of an incoming FlowFile with the content of the remote file. |
||
Fetches a file from a Google Cloud Bucket. |
||
Fetches files from a Google Drive Folder. |
||
Snowflake |
Fetches comments and their replies for a Google Drive file. |
|
Snowflake |
Fetches Google Drive file metadata. |
|
Retrieves one or more files from a GridFS bucket by file name or by a user-defined query. |
||
Snowflake |
Fetches issues from Jira Cloud using REST API v3 with configurable search options. |
|
Snowflake |
Fetch records from Microsoft Dataverse Tables |
|
Retrieves the contents of an S3 Object and writes it to the content of a FlowFile |
||
Fetches the content of a file from a remote SFTP server and overwrites the contents of an incoming FlowFile with the content of the remote file. |
||
Snowflake |
Fetches the contents of a file from a Sharepoint Drive, optionally downloading a PDF or HTML version of the file when applicable. |
|
Snowflake |
For each drive item retrieves its metadata and permissions and writes them as FlowFile attributes. |
|
Snowflake |
Fetches Slack conversation info and member emails |
|
Snowflake |
Downloads a file shared on Slack. |
|
Snowflake |
Fetches data about a single Slack message |
|
Fetches files from a SMB Share. |
||
Snowflake |
Reads properties from a table and stores them as flow file attributes. |
|
Snowflake |
Fetches the table schema (i. |
|
Snowflake |
Fetches a snapshot of a table from a database. |
|
Filters the attributes of a FlowFile by retaining specified attributes and removing the rest or by removing specified attributes and retaining the rest. |
||
Provides the user with the ability to take a nested JSON document and flatten it into a simple key/value pair document. |
||
Used in conjunction with the JoinEnrichment processor, this processor is responsible for adding the attributes that are necessary for the JoinEnrichment processor to perform its function. |
||
This processor allows the user to fork a record into multiple records. |
||
Snowflake |
Formats a MS Word docx file |
G¶
Processor |
Description |
|
---|---|---|
Snowflake |
Generates synthetic answers for each question present in the incoming records using a Large Language Model (LLM). |
|
Snowflake |
Generates synthetic answers for each question in the incoming records using an LLM. |
|
This processor creates FlowFiles with random data or custom content. |
||
This processor creates FlowFiles with records having random value for the specified fields. |
||
Generates SQL select queries that fetch “pages” of rows from a table. |
||
Looks up geolocation information for an IP address and adds the geo information to FlowFile attributes. |
||
Looks up geolocation information for an IP address and adds the geo information to FlowFile attributes. |
||
Snowflake |
Processor downloading report from Amazon Ads if ready. |
|
Retrieves the current status of an AWS Polly job. |
||
Retrieves the current status of an AWS Textract job. |
||
Retrieves the current status of an AWS Transcribe job. |
||
Retrieves the current status of an AWS Translate job. |
||
Receives messages from Microsoft Azure Event Hubs without reliable checkpoint tracking. |
||
Retrieves the messages from an Azure Queue Storage. |
||
Retrieves all collaborators on a Box file and adds the collaboration information to the FlowFile’s attributes. |
||
Retrieves members for a Box Group and writes their details in FlowFile attributes. |
||
Snowflake |
Read a DBFS file. |
|
Retrieves a document from DynamoDB based on hash and range key. |
||
Elasticsearch get processor that uses the official Elastic REST client libraries to fetch a single document from Elasticsearch by _id. |
||
Creates FlowFiles from files in a directory. |
||
This processor creates FlowFiles with the content of the configured File Resource. |
||
Fetches files from an FTP Server and creates FlowFiles from them |
||
Retrieves the current status of an Google Vision operation. |
||
Retrieves the current status of an Google Vision operation. |
||
Snowflake |
A processor which can interact with Google Ads Reporting API. |
|
Snowflake |
Retrieves the immediate (top-level) members of one or more Google Groups, specified as a comma-separated list of group IDs that is given as a FlowFile attribute. |
|
Snowflake |
Processor responsible for fetching data from Google Sheets. |
|
Retrieves JSON data from a private HubSpot application. |
||
Snowflake |
Get a HubSpot object and its associations by ID or unique value. |
|
Snowflake |
Retrieves Microsoft365 group members and emits a FlowFile for each change that occurs. |
|
Creates FlowFiles from documents in MongoDB loaded by a user-specified query. |
||
A record-based version of GetMongo that uses the Record writers to write the MongoDB result set. |
||
Check for the existence of an Object in S3 and fetch its Metadata without attempting to download it. |
||
Check for the existence of an Object in S3 and fetch its Tags without attempting to download it. |
||
Fetches files from an SFTP Server and creates FlowFiles from them |
||
Snowflake |
Retrieves all members of a SharePoint site group. |
|
Retrieves objects from a custom Shopify store. |
||
Reads file from a samba network location to FlowFiles. |
||
Retrieves data from Splunk Enterprise. |
||
Fetches messages from an Amazon Simple Queuing Service Queue |
||
Snowflake |
Read a Unity Catalog file up to 5 GiB. |
|
Snowflake |
Checks for Unity Catalog file metadata. |
|
A processor which can interact with a configurable Workday Report. |
||
Incrementally fetches data from Zendesk API. |
H¶
Processor |
Description |
|
---|---|---|
Starts an HTTP Server and listens for HTTP Requests. |
||
Sends an HTTP Response to the Requestor that generated a FlowFile. |
I¶
Processor |
Description |
|
---|---|---|
Attempts to identify the MIME Type used for a FlowFile. |
||
Snowflake |
Automatically infers and generates an Apache Avro schema from Jira issue data. |
|
An HTTP client processor which can interact with a configurable HTTP Endpoint. |
||
Experimental - Invokes a script engine for a Processor defined in the given script. |
||
Looks up ISP information for an IP address and adds the information to FlowFile attributes. |
J¶
Processor |
Description |
|
---|---|---|
Joins together Records from two different FlowFiles where one FlowFile, the ‘original’ contains arbitrary records and the second FlowFile, the ‘enrichment’ contains additional data that should be used to enrich the first. |
||
Applies a list of Jolt specifications to the flowfile JSON payload. |
||
Applies a JOLT specification to each record in the FlowFile payload. |
||
Applies a JSLT transformation to the FlowFile JSON payload. |
||
A processor that allows the user to run a query (with aggregations) written with the Elasticsearch JSON DSL. |
L¶
Processor |
Description |
|
---|---|---|
Snowflake |
Lists archived data from HubSpot for the chosen object type and generates one FlowFile per listed object with the corresponding metadata as FlowFile attributes. |
|
Lists blobs in an Azure Blob Storage container. |
||
Lists directory in an Azure Data Lake Storage Gen 2 filesystem |
||
Lists files in a Box folder. |
||
Generates a set of flow files, each containing attributes corresponding to metadata about a table from a database connection. |
||
Snowflake |
List file names in a DBFS directory and output a new FlowFile with the filename. |
|
Retrieves a listing of files from Dropbox (shortcuts are ignored). |
||
Starts an FTP server that listens on the specified port and transforms incoming files into FlowFiles. |
||
Starts an HTTP Server and listens on a given base path to transform incoming requests into FlowFiles. |
||
Collect OpenTelemetry messages over HTTP or gRPC. |
||
Retrieves real-time messages or Slack commands from one or more Slack conversations. |
||
Listens for Syslog messages being sent to a given port over TCP or UDP. |
||
Listens for incoming TCP connections and reads data from each connection using a line separator as the message demarcator. |
||
Listens for Datagram Packets on a given port. |
||
Listens for Datagram Packets on a given port and reads the content of each datagram using the configured Record Reader. |
||
Acts as a WebSocket server endpoint to accept client connections. |
||
Retrieves a listing of files from the input directory. |
||
Performs a listing of the files residing on an FTP server. |
||
Retrieves a listing of objects from a GCS bucket. |
||
Performs a listing of concrete files (shortcuts are ignored) in a Google Drive folder. |
||
Snowflake |
Lists all of the groups for a given domain in Google Workspace. |
|
Snowflake |
Fetches data from HubSpot for specified object types, and generates one FlowFile per listed object with the corresponding metadata as FlowFile attributes. |
|
Snowflake |
List Tables from Microsoft Dataverse environments |
|
Retrieves a listing of objects from an S3 bucket. |
||
Performs a listing of the files residing on an SFTP server. |
||
Snowflake |
Lists all SharePoint site groups available on a specified SharePoint site. |
|
Lists concrete files shared via SMB protocol. |
||
Snowflake |
Fetches all source table names and matches them with one of the possible configurations:- regexp expression e. |
|
Snowflake |
List file names in a Unity Catalog directory and output a new FlowFile with the filename. |
|
Emits attributes of the FlowFile at the specified log level |
||
Emits a log message at the specified log level |
||
Lookup attributes from a lookup service |
||
Extracts one or more fields from a Record and looks up a value for those fields in a LookupService. |
M¶
Processor |
Description |
|
---|---|---|
Merges a Group of FlowFiles together based on a user-defined strategy and packages them into a single FlowFile. |
||
Snowflake |
Given a FlowFile that contains a full Document and one more FlowFiles that contain additional data to merge into the Document, this Processor will merge the additional data into the Document. |
|
This Processor merges together multiple record-oriented FlowFiles into a single FlowFile that contains all of the Records of the input FlowFiles. |
||
Snowflake |
Triggers a merge operation on changes from journal table to a destination table in Snowflake. |
|
Discard byte range at the start and end or all content of a binary file. |
||
Changes the compression algorithm used to compress the contents of a FlowFile by decompressing the contents of FlowFiles using a user-specified compression algorithm and recompressing the contents using the specified compression format properties. |
||
Monitors the flow for activity and sends out an indicator when the flow has not had any data for some specified amount of time and again when the flow’s activity is restored |
||
Moves content within an Azure Data Lake Storage Gen 2. |
N¶
Processor |
Description |
|
---|---|---|
Caches a release signal identifier in the distributed cache, optionally along with the FlowFile’s attributes. |
O¶
Processor |
Description |
|
---|---|---|
Snowflake |
Transcribes audio into English text. |
P¶
Processor |
Description |
|
---|---|---|
This processor will package FlowFile attributes and content into an output FlowFile that can be exported from NiFi and imported back into NiFi, preserving the original attributes and content. |
||
A processor that allows the user to run a paginated query (with aggregations) written with the Elasticsearch JSON DSL. |
||
Parses the contents of a Windows Event Log file (evtx) and writes the resulting XML to the FlowFile |
||
Snowflake |
Parses a PDF file, extracting the text and additional information into a structured JSON document. |
|
Attempts to parses the contents of a Syslog message in accordance to RFC5424 and RFC3164 formats and adds attributes to the FlowFile for each of the parts of the Syslog message. |
||
Attempts to parse the contents of a well formed Syslog message in accordance to RFC5424 format and adds attributes to the FlowFile for each of the parts of the Syslog message, including Structured Data. |
||
Snowflake |
Extracts the text from a Table image and writes it to the FlowFile content in csv format. |
|
Splits, or partitions, record-oriented data based on the configured fields in the data. |
||
Snowflake |
Uses the Openflow Tesseract OCR Service to extract text from a PDF or image, optionally providing metadata including the bounding box, page numberand confidence level of the OCR. |
|
Snowflake |
Performs Optical Character Recognition (OCR) on PDF documents using Snowflake Cortex ML functions. |
|
Snowflake |
Sends a prompt to Anthropic, writing the response either as a FlowFile attribute or to the contents of the incoming FlowFile. |
|
Snowflake |
Sends a prompt to Azure’s OpenAI service, writing the response either as a FlowFile attribute or to the contents of the incoming FlowFile. |
|
Snowflake |
This processor sends a user defined prompt to a Large Language Model (LLM) to respond. |
|
Snowflake |
Sends a prompt to Ollama, writing the response either as a FlowFile attribute or to the contents of the incoming FlowFile. |
|
Snowflake |
Sends a prompt to OpenAI, writing the response either as a FlowFile attribute or to the contents of the incoming FlowFile. |
|
Snowflake |
Sends a prompt to Snowflake Cortex, writing the response either as a FlowFile attribute or to the contents of the incoming FlowFile. |
|
Snowflake |
Sends a prompt to VertexAI, writing the response either as a FlowFile attribute or to the contents of the incoming FlowFile. |
|
Creates an AMQP Message from the contents of a FlowFile and sends the message to an AMQP Exchange. |
||
Publishes the content of the incoming flowfile to the configured Google Cloud PubSub topic. |
||
Creates a JMS Message from the contents of a FlowFile and sends it to a JMS Destination (queue or topic) as JMS BytesMessage or TextMessage. |
||
Snowflake |
Sends the contents of a FlowFile as either a message or as individual records to Apache Kafka using the Kafka Producer API. |
|
Publishes a message to an MQTT topic |
||
Posts a message to the specified Slack channel. |
||
Puts content into a blob on Azure Blob Storage. |
||
This processor is a record-aware processor for inserting data into Cosmos DB with Core SQL API. |
||
Acts as an Azure Data Explorer sink which sends FlowFiles to the provided endpoint. |
||
Writes the contents of a FlowFile as a file on Azure Data Lake Storage Gen 2 |
||
Send FlowFile contents to Azure Event Hubs |
||
Writes the content of the incoming FlowFiles to the configured Azure Queue Storage. |
||
Writes the contents of a FlowFile to a Google BigQuery table. |
||
Puts content to a Box folder. |
||
Publishes metrics to Amazon CloudWatch. |
||
The PutDatabaseRecord processor uses a specified RecordReader to input (possibly multiple) records from an incoming flow file. |
||
Snowflake |
Submit a SQL Execution using Databricks REST API then write the JSON response to FlowFile Content. |
|
Snowflake |
Write FlowFile content to DBFS. |
|
Gets the content of a FlowFile and puts it to a distributed map cache, using a cache key computed from FlowFile attributes. |
||
Puts content to a Dropbox folder. |
||
Puts a document from DynamoDB based on hash and range key. |
||
Inserts items into DynamoDB based on record-oriented data. |
||
An Elasticsearch put processor that uses the official Elastic REST client libraries. |
||
A record-aware Elasticsearch put processor that uses the official Elastic REST client libraries. |
||
Sends an e-mail to configured recipients for each incoming FlowFile |
||
Writes the contents of a FlowFile to the local file system |
||
Sends FlowFiles to an FTP Server |
||
Writes the contents of a FlowFile as an object in a Google Cloud Storage. |
||
Writes the contents of a FlowFile as a file in Google Drive. |
||
Writes a file to a GridFS bucket. |
||
Snowflake |
Upsert a HubSpot object. |
|
Snowflake |
Store records in Iceberg using configurable Catalog for managing namespaces and tables. |
|
Sends the contents to a specified Amazon Kinesis Firehose. |
||
Sends the contents to a specified Amazon Kinesis. |
||
Sends the contents to a specified Amazon Lambda Function. |
||
Writes the contents of a FlowFile to MongoDB |
||
Writes the contents of a FlowFile to MongoDB as bulk-update |
||
This processor is a record-aware processor for inserting/upserting data into MongoDB. |
||
The PutRecord processor uses a specified RecordReader to input (possibly multiple) records from an incoming flow file, and sends them to a destination specified by a Record Destination Service (i. |
||
Puts record field data into Redis using a specified hash value, which is determined by a RecordPath to a field in each record containing the hash value. |
||
Writes the contents of a FlowFile as an S3 Object to an Amazon S3 Bucket. |
||
Creates new records for the specified Salesforce sObject. |
||
Sends FlowFiles to an SFTP Server |
||
Writes the contents of a FlowFile to a samba network location. |
||
Snowflake |
Puts files into a Snowflake internal stage. |
|
Snowflake |
Streams records into a Snowflake table. |
|
Sends the content of a FlowFile as a notification to the Amazon Simple Notification Service |
||
Sends logs to Splunk Enterprise over TCP, TCP + TLS/SSL, or UDP. |
||
Sends flow file content to the specified Splunk server over HTTP or HTTPS. |
||
Executes a SQL UPDATE or INSERT command. |
||
Publishes a message to an Amazon Simple Queuing Service Queue |
||
Sends Syslog messages to a given host and port over TCP or UDP. |
||
Sends serialized FlowFiles or Records over TCP to a configurable destination with optional support for TLS |
||
The PutUDP processor receives a FlowFile and packages the FlowFile content into a single UDP datagram packet which is then transmitted to the configured UDP server. |
||
Snowflake |
Write FlowFile content with max size of 5 GiB to Unity Catalog. |
|
Snowflake |
Generate and upload a JSON document to Vectara’s upload endpoint. |
|
Snowflake |
Upload a FlowFile content to Vectara’s index endpoint. |
|
Snowflake |
Uses Vespa document api to update a record in a specific namespace. |
|
Sends messages to a WebSocket remote endpoint using a WebSocket session that is established by either ListenWebSocket or ConnectWebSocket. |
||
Create Zendesk tickets using the Zendesk API. |
Q¶
Processor |
Description |
|
---|---|---|
Query Azure Data Explorer and stream JSON results to output FlowFiles |
||
Generates a SQL select query, or uses a provided statement, and executes it to fetch all rows whose values in the specified Maximum Value column(s) are larger than the previously-seen maxima. |
||
Generates a SQL select query, or uses a provided statement, and executes it to fetch all rows whose values in the specified Maximum Value column(s) are larger than the previously-seen maxima. |
||
Snowflake |
Evaluates a SQL-like query against the incoming Openflow Document JSON, producing the results on the outgoing FlowFile. |
|
Snowflake |
Queries a given collection in a Milvus database using vectors. |
|
Snowflake |
Queries Pinecone for vectors that are similar to the input vector, or retrieves a vector by ID. |
|
Evaluates one or more SQL queries against the contents of a FlowFile. |
||
Retrieves records from a Salesforce sObject. |
||
Queries Splunk server in order to acquire the status of indexing acknowledgement. |
R¶
Processor |
Description |
|
---|---|---|
Modifies the contents of a FlowFile that contains Record-oriented data (i. |
||
Renames one or more fields in each Record of a FlowFile. |
||
Updates the content of a FlowFile by searching for some textual value in the FlowFile content (via Regular Expression/regex, or literal value) and replacing the section of the content that matches with some alternate value. |
||
Updates the content of a FlowFile by evaluating a Regular Expression against it and replacing the section of the content that matches the Regular Expression with some alternate value provided in a mapping file. |
||
FlowFiles passed to this Processor have a ‘Retry Attribute’ value checked against a configured ‘Maximum Retries’ value. |
||
Routes FlowFiles based on their Attributes using the Attribute Expression Language |
||
Applies Regular Expressions to the content of a FlowFile and routes a copy of the FlowFile to each destination whose Regular Expression matches. |
||
Routes textual data based on a set of user-defined rules. |
||
Snowflake |
Triggers a pre-defined Databricks job to run with custom parameters. |
|
A processor that runs an aggregation query whenever a flowfile is received. |
S¶
Processor |
Description |
|
---|---|---|
Samples the records of a FlowFile based on a specified sampling strategy (such as Reservoir Sampling). |
||
Scans the specified attributes of FlowFiles, checking to see if any of their values are present within the specified dictionary of terms |
||
Scans the content of FlowFiles for terms that are found in a user-supplied dictionary. |
||
This processor provides the ability to filter records out from FlowFiles using the user-provided script. |
||
Receives Record-oriented data (i. |
||
Provides the ability to evaluate a simple script against each record in an incoming FlowFile. |
||
This processor provides the ability to validate records in FlowFiles using the user-provided script. |
||
A processor that allows the user to repeatedly run a paginated query (with aggregations) written with the Elasticsearch JSON DSL. |
||
Segments a FlowFile into multiple smaller segments on byte boundaries. |
||
Sign content using OpenPGP Private Keys |
||
Snowflake |
Checks if a FlowFile’s hash (provided as a FlowFile attribute) is already in a Snowflake table, and routes the FlowFile to ‘duplicate’ if found, ‘distinct’ if not found, or ‘failure’ on errors. |
|
Splits a binary encoded Avro datafile into smaller files based on the configured Output Size. |
||
Splits incoming FlowFiles by a specified byte sequence |
||
This processor splits a multi sheet Microsoft Excel spreadsheet into multiple Microsoft Excel spreadsheets where each sheet from the original file is converted to an individual spreadsheet in its own flow file. |
||
Splits a JSON File into multiple, separate FlowFiles for an array element specified by a JsonPath expression. |
||
Splits up an input FlowFile that is in a record-oriented data format into multiple smaller FlowFiles |
||
Splits a text file into multiple smaller text files on line boundaries limited by maximum number of lines or total size of fragment. |
||
Splits an XML File into multiple separate FlowFiles, each comprising a child or descendant of the original root element |
||
Trigger a AWS Polly job. |
||
Trigger a AWS Textract job. |
||
Trigger a AWS Transcribe job. |
||
Trigger a AWS Translate job. |
||
Trigger a Vision operation on file input. |
||
Trigger a Vision operation on image input. |
||
Snowflake |
This processor uses a Large Language Model (LLM) to summarize the content of a FlowFile. |
T¶
Processor |
Description |
|
---|---|---|
Adds or updates a tag on an Amazon S3 Object. |
||
“Tails” a file, or a list of files, ingesting data from the file as it is written to the file. |
||
Applies the provided XSLT file to the FlowFile XML payload. |
U¶
Processor |
Description |
|
---|---|---|
Unpacks the content of FlowFiles that have been packaged with one of several different Packaging Formats, emitting one to many FlowFiles for each input FlowFile. |
||
Updates the Attributes for a FlowFile by using the Attribute Expression Language and/or deletes the attributes based on a regular expression |
||
Update documents in an Elasticsearch index using a query. |
||
This processor allows users to set specific counters and key points in their flow. |
||
This processor uses a JDBC connection and incoming records to generate any database table changes needed to support the incoming records. |
||
Updates the contents of a FlowFile that contains Record-oriented data (i. |
||
Snowflake |
Updates the definition of a Snowflake table based on the schema provided in the incoming FlowFile. |
|
Snowflake |
Updates the state of a table in the Table State Service |
|
Snowflake |
Upserts vectors into Milvus database for a given collection |
|
Snowflake |
Publishes vectors, including metadata, and optionally text, to a Pinecone index. |
V¶
Processor |
Description |
|
---|---|---|
Validates the contents of FlowFiles or a FlowFile attribute value against a user-specified CSV schema. |
||
Validates the contents of FlowFiles against a configurable JSON Schema. |
||
Validates the Records of an incoming FlowFile against a given schema. |
||
Validates XML contained in a FlowFile. |
||
Calculates a Message Authentication Code using the provided Secret Key and compares it with the provided MAC property |
||
Verify signatures using OpenPGP Public Keys |
W¶
Processor |
Description |
|
---|---|---|
Routes incoming FlowFiles to the ‘wait’ relationship until a matching release signal is stored in the distributed cache from a corresponding Notify processor. |
||
Snowflake |
Blocks incoming FlowFiles until the corresponding table state is not equal to accepted state. |