Docs Cloud Redpanda Connect Components Processors aws_bedrock_chat aws_bedrock_chat Beta Available in: Cloud, Self-Managed Generates responses to messages in a chat conversation, using the AWS Bedrock API. Common Advanced # Common configuration fields, showing default values label: "" aws_bedrock_chat: model: amazon.titan-text-express-v1 # No default (required) prompt: "" # No default (optional) system_prompt: "" # No default (optional) max_tokens: 0 # No default (optional) stop: 0 # No default (optional) # All configuration fields, showing default values label: "" aws_bedrock_chat: region: "" endpoint: "" credentials: profile: "" id: "" secret: "" token: "" from_ec2_role: false role: "" role_external_id: "" model: amazon.titan-text-express-v1 # No default (required) prompt: "" # No default (optional) system_prompt: "" # No default (optional) max_tokens: 0 # No default (optional) stop: 0 # No default (optional) temperature: [] # No default (optional) top_p: 0 # No default (optional) This processor sends prompts to your chosen large language model (LLM) and generates text from the responses, using the AWS Bedrock API. For more information, see the AWS Bedrock documentation. Fields model The model ID to use. For a full list, see the AWS Bedrock documentation. Type: string # Examples model: amazon.titan-text-express-v1 model: anthropic.claude-3-5-sonnet-20240620-v1:0 model: cohere.command-text-v14 model: meta.llama3-1-70b-instruct-v1:0 model: mistral.mistral-large-2402-v1:0 prompt The prompt you want to generate a response for. By default, the processor submits the entire payload as a string. Type: string system_prompt The system prompt to submit to the AWS Bedrock LLM. Type: string max_tokens The maximum number of tokens to allow in the generated response. Type: int stop The likelihood of the model selecting higher-probability options while generating a response. A lower value makes the model more likely to choose higher-probability options. A higher value makes the model more likely to choose lower-probability options. Type: float credentials Configure which AWS credentials to use (optional). For more information, see Amazon Web Services. Type: object credentials.profile The profile from ~/.aws/credentials to use. Type: string Default: "" credentials.id The ID of credentials to use. Type: string Default: "" credentials.secret The secret for the credentials you want to use. This field contains sensitive information. Review your cluster security before adding it to your configuration. Type: string Default: "" credentials.token The token for the credentials you want to use. You must enter this value when using short-term credentials. Type: string Default: "" credentials.from_ec2_role Use the credentials of a host EC2 machine configured to assume an IAM role associated with the instance. Type: bool Default: false credentials.role The role ARN to assume. Type: string Default: "" credentials.role_external_id The external ID to use when assuming a role. Type: string Default: "" region The AWS region to target. Type: string Default: "" endpoint Specify a custom endpoint for the AWS API. Type: string Default: "" temperature A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response. Type: array top_p The percentage of most-likely candidates that the model considers for the next token. For example, if you choose a value of 0.8, the model selects from the top 80% of the probability distribution of tokens that could be next in the sequence. Type: float Back to top × Simple online edits For simple changes, such as fixing a typo, you can edit the content directly on GitHub. Edit on GitHub Or, open an issue to let us know about something that you want us to change. Open an issue Contribution guide For extensive content updates, or if you prefer to work locally, read our contribution guide . Was this helpful? thumb_up thumb_down group Ask in the community mail Share your feedback group_add Make a contribution avro aws_bedrock_embeddings