Docs Connect Components Processors Processors Redpanda Connect processors are functions applied to messages passing through a pipeline. The function signature allows a processor to mutate or drop messages depending on the content of the message. There are many types on offer but the most powerful are the mapping and mutation processors. Processors are set via config, and depending on where in the config they are placed they will be run either immediately after a specific input (set in the input section), on all messages (set in the pipeline section) or before a specific output (set in the output section). Most processors apply to all messages and can be placed in the pipeline section: pipeline: threads: 1 processors: - label: my_cool_mapping mapping: | root.message = this root.meta.link_count = this.links.length() The threads field in the pipeline section determines how many parallel processing threads are created. You can read more about parallel processing in the pipeline guide. Labels Processors have an optional field label that can uniquely identify them in observability data such as metrics and logs. This can be useful when running configs with multiple nested processors, otherwise their metrics labels will be generated based on their composition. For more information check out the metrics documentation. Error handling Some processors have conditions whereby they might fail. Rather than throw these messages into the abyss Redpanda Connect still attempts to send these messages onwards, and has mechanisms for filtering, recovering or dead-letter queuing messages that have failed which can be read about here. Error logs Errors that occur during processing can be roughly separated into two groups; those that are unexpected intermittent errors such as connectivity problems, and those that are logical errors such as bad input data or unmatched schemas. All processing errors result in the messages being flagged as failed, error metrics increasing for the given errored processor, and debug level logs being emitted that describe the error. Only errors that are known to be intermittent are also logged at the error level. The reason for this behavior is to prevent noisy logging in cases where logical errors are expected and will likely be handled in config. However, this can also sometimes make it easy to miss logical errors in your configs when they lack error handling. If you suspect you are experiencing processing errors and do not wish to add error handling yet then a quick and easy way to expose those errors is to enable debug level logs with the cli flag --log.level=debug or by setting the level in config: logger: level: DEBUG Using processors as outputs It might be the case that a processor that results in a side effect, such as the sql_insert or redis processors, is the only side effect of a pipeline, and therefore could be considered the output. In such cases it’s possible to place these processors within a reject output so that they behave the same as regular outputs, where success results in dropping the message with an acknowledgement and failure results in a nack (or retry): output: reject: 'failed to send data: ${! error() }' processors: - try: - redis: url: tcp://localhost:6379 command: sadd args_mapping: 'root = [ this.key, this.value ]' - mapping: root = deleted() The way this works is that if your processor with the side effect (redis in this case) succeeds then the final mapping processor deletes the message which results in an acknowledgement. If the processor fails then the try block exits early without executing the mapping processor and instead the message is routed to the reject output, which nacks the message with an error message containing the error obtained from the redis processor. Categories Parsing Utility Mapping AI Integration Azure Composition Services Processors that specialize in translating messages from one format to another. archive Avro bloblang compress decompress Grok mapping MessagePack mutation parquet_decode parquet_encode parse_log Protobuf schema_registry_decode schema_registry_encode unarchive XML Processors that provide general utility or do not fit in another category. archive benchmark bounds_check cached dedupe log metric rate_limit redpanda_data_transform resource select_parts sleep split sync_response unarchive WebAssembly Processors that specialize in restructuring messages. AWK bloblang JavaScript JMESPath jq JSON Schema mapping mutation aws_bedrock_embeddings aws_bedrock_chat cohere_chat cohere_embeddings gcp_vertex_ai_chat gcp_vertex_ai_embeddings ollama_chat ollama_embeddings openai_chat_completion openai_embeddings openai_image_generation openai_speech openai_transcription openai_translation Processors that interact with external services. aws_dynamodb_partiql AWS Lambda cache command Couchbase gcp_bigquery_select HTTP Redis redis_script schema_registry_decode schema_registry_encode sql_insert sql_raw sql_select subprocess Azure Cosmos DB Higher level processors that compose other processors and modify their behavior. branch catch for_each group_by group_by_value insert_part parallel processors retry switch try while workflow MongoDB nats_kv nats_request_reply Batching and multiple-part messages All Redpanda Connect processors support multiple-part messages, which are synonymous with batches. This enables some cool windowed processing capabilities. Many processors are able to perform their behaviors on specific parts of a message batch, or on all parts, and have a field parts for specifying an array of part indexes they should apply to. If the list of target parts is empty these processors will be applied to all message parts. Part indexes can be negative, and if so the part will be selected from the end counting backwards starting from -1. E.g. if part = -1 then the selected part will be the last part of the message, if part = -2 then the part before the last element will be selected, and so on. Some processors such as dedupe act across an entire batch, when instead we might like to perform them on individual messages of a batch. In this case the for_each processor can be used. You can read more about batching in this document. Back to top × Simple online edits For simple changes, such as fixing a typo, you can edit the content directly on GitHub. Edit on GitHub Or, open an issue to let us know about something that you want us to change. Open an issue Contribution guide For extensive content updates, or if you prefer to work locally, read our contribution guide . Was this helpful? thumb_up thumb_down group Ask in the community mail Share your feedback group_add Make a contribution to_the_end archive