OpenClaw can use Amazon Bedrock models via pi-ai’s Bedrock Converse streaming provider. Bedrock auth uses the AWS SDK default credential chain, not an API key.Documentation Index
Fetch the complete documentation index at: https://docs.openclaw.ai/llms.txt
Use this file to discover all available pages before exploring further.
| Property | Value |
|---|---|
| Provider | amazon-bedrock |
| API | bedrock-converse-stream |
| Auth | AWS credentials (env vars, shared config, or instance role) |
| Region | AWS_REGION or AWS_DEFAULT_REGION (default: us-east-1) |
Getting started
Choose your preferred auth method and follow the setup steps.- Access keys / env vars
- EC2 instance roles (IMDS)
Automatic model discovery
OpenClaw can automatically discover Bedrock models that support streaming and text output. Discovery usesbedrock:ListFoundationModels and
bedrock:ListInferenceProfiles, and results are cached (default: 1 hour).
How the implicit provider is enabled:
- If
plugins.entries.amazon-bedrock.config.discovery.enabledistrue, OpenClaw will try discovery even when no AWS env marker is present. - If
plugins.entries.amazon-bedrock.config.discovery.enabledis unset, OpenClaw only auto-adds the implicit Bedrock provider when it sees one of these AWS auth markers:AWS_BEARER_TOKEN_BEDROCK,AWS_ACCESS_KEY_ID+AWS_SECRET_ACCESS_KEY, orAWS_PROFILE. - The actual Bedrock runtime auth path still uses the AWS SDK default chain, so
shared config, SSO, and IMDS instance-role auth can work even when discovery
needed
enabled: trueto opt in.
For explicit
models.providers["amazon-bedrock"] entries, OpenClaw can still resolve Bedrock env-marker auth early from AWS env markers such as AWS_BEARER_TOKEN_BEDROCK without forcing full runtime auth loading. The actual model-call auth path still uses the AWS SDK default chain.Discovery config options
Discovery config options
Config options live under
plugins.entries.amazon-bedrock.config.discovery:| Option | Default | Description |
|---|---|---|
enabled | auto | In auto mode, OpenClaw only enables the implicit Bedrock provider when it sees a supported AWS env marker. Set true to force discovery. |
region | AWS_REGION / AWS_DEFAULT_REGION / us-east-1 | AWS region used for discovery API calls. |
providerFilter | (all) | Matches Bedrock provider names (for example anthropic, amazon). |
refreshInterval | 3600 | Cache duration in seconds. Set to 0 to disable caching. |
defaultContextWindow | 32000 | Context window used for discovered models (override if you know your model limits). |
defaultMaxTokens | 4096 | Max output tokens used for discovered models (override if you know your model limits). |
Quick setup (AWS path)
This walkthrough creates an IAM role, attaches Bedrock permissions, associates the instance profile, and enables OpenClaw discovery on the EC2 host.Advanced configuration
Inference profiles
Inference profiles
OpenClaw discovers regional and global inference profiles alongside
foundation models. When a profile maps to a known foundation model, the
profile inherits that model’s capabilities (context window, max tokens,
reasoning, vision) and the correct Bedrock request region is injected
automatically. This means cross-region Claude profiles work without manual
provider overrides.Inference profile IDs look like
us.anthropic.claude-opus-4-6-v1:0 (regional)
or anthropic.claude-opus-4-6-v1:0 (global). If the backing model is already
in the discovery results, the profile inherits its full capability set;
otherwise safe defaults apply.No extra configuration is needed. As long as discovery is enabled and the IAM
principal has bedrock:ListInferenceProfiles, profiles appear alongside
foundation models in openclaw models list.Claude Opus 4.7 temperature
Claude Opus 4.7 temperature
Bedrock rejects the
temperature parameter for Claude Opus 4.7. OpenClaw
omits temperature automatically for any Opus 4.7 Bedrock ref, including
foundation model ids, named inference profiles, application inference
profiles whose underlying model resolves to Opus 4.7 via
bedrock:GetInferenceProfile, and dotted opus-4.7 variants with
optional region prefixes (us., eu., ap., apac., au., jp.,
global.). No config knob is required, and the omission applies to both
the request options object and the inferenceConfig payload field.Guardrails
Guardrails
You can apply Amazon Bedrock Guardrails
to all Bedrock model invocations by adding a
guardrail object to the
amazon-bedrock plugin config. Guardrails let you enforce content filtering,
topic denial, word filters, sensitive information filters, and contextual
grounding checks.| Option | Required | Description |
|---|---|---|
guardrailIdentifier | Yes | Guardrail ID (e.g. abc123) or full ARN (e.g. arn:aws:bedrock:us-east-1:123456789012:guardrail/abc123). |
guardrailVersion | Yes | Published version number, or "DRAFT" for the working draft. |
streamProcessingMode | No | "sync" or "async" for guardrail evaluation during streaming. If omitted, Bedrock uses its default. |
trace | No | "enabled" or "enabled_full" for debugging; omit or set "disabled" for production. |
Embeddings for memory search
Embeddings for memory search
Bedrock can also serve as the embedding provider for
memory search. This is configured separately from the
inference provider — set Bedrock embeddings use the same AWS SDK credential chain as inference (instance
roles, SSO, access keys, shared config, and web identity). No API key is
needed. When
agents.defaults.memorySearch.provider to "bedrock":provider is "auto", Bedrock is auto-detected if that
credential chain resolves successfully.Supported embedding models include Amazon Titan Embed (v1, v2), Amazon Nova
Embed, Cohere Embed (v3, v4), and TwelveLabs Marengo. See
Memory configuration reference — Bedrock
for the full model list and dimension options.Notes and caveats
Notes and caveats
- Bedrock requires model access enabled in your AWS account/region.
- Automatic discovery needs the
bedrock:ListFoundationModelsandbedrock:ListInferenceProfilespermissions. - If you rely on auto mode, set one of the supported AWS auth env markers on the
gateway host. If you prefer IMDS/shared-config auth without env markers, set
plugins.entries.amazon-bedrock.config.discovery.enabled: true. - OpenClaw surfaces the credential source in this order:
AWS_BEARER_TOKEN_BEDROCK, thenAWS_ACCESS_KEY_ID+AWS_SECRET_ACCESS_KEY, thenAWS_PROFILE, then the default AWS SDK chain. - Reasoning support depends on the model; check the Bedrock model card for current capabilities.
- If you prefer a managed key flow, you can also place an OpenAI-compatible proxy in front of Bedrock and configure it as an OpenAI provider instead.
Related
Model selection
Choosing providers, model refs, and failover behavior.
Memory search
Bedrock embeddings for memory search configuration.
Memory config reference
Full Bedrock embedding model list and dimension options.
Troubleshooting
General troubleshooting and FAQ.