Documentation Index
Fetch the complete documentation index at: https://mintlify.com/portkey-AI/gateway/llms.txt
Use this file to discover all available pages before exploring further.
Overview
Amazon Bedrock provides access to foundation models from leading AI companies including Anthropic, Meta, Mistral, Cohere, and Amazon through a unified API with AWS security, compliance, and infrastructure. Service:bedrock (data plane) and bedrock-runtime (inference)
Supported Features
- ✅ Chat Completions (Converse API)
- ✅ Streaming
- ✅ Embeddings
- ✅ Image Generation (Stable Diffusion, Titan)
- ✅ Function Calling (via Converse API)
- ✅ Batch Inference
- ✅ Model Customization (Fine-tuning)
- ✅ Guardrails
- ✅ Multiple Authentication Methods
Quick Start
Basic Configuration
Available Models
Anthropic Claude
| Model ID | Model | Context | Best For |
|---|---|---|---|
anthropic.claude-3-5-sonnet-20241022-v2:0 | Claude 3.5 Sonnet | 200K | Most capable |
anthropic.claude-3-5-haiku-20241022-v1:0 | Claude 3.5 Haiku | 200K | Fast, efficient |
anthropic.claude-3-opus-20240229-v1:0 | Claude 3 Opus | 200K | Complex tasks |
anthropic.claude-3-sonnet-20240229-v1:0 | Claude 3 Sonnet | 200K | Balanced |
anthropic.claude-3-haiku-20240307-v1:0 | Claude 3 Haiku | 200K | Speed |
Meta Llama
| Model ID | Context | Description |
|---|---|---|
meta.llama3-3-70b-instruct-v1:0 | 128K | Latest Llama 3.3 |
meta.llama3-1-405b-instruct-v1:0 | 128K | Largest Llama 3.1 |
meta.llama3-1-70b-instruct-v1:0 | 128K | Efficient Llama 3.1 |
meta.llama3-1-8b-instruct-v1:0 | 128K | Fast, compact |
Mistral AI
| Model ID | Context | Description |
|---|---|---|
mistral.mistral-large-2407-v1:0 | 128K | Most capable |
mistral.mistral-large-2402-v1:0 | 32K | Previous generation |
mistral.mistral-small-2402-v1:0 | 32K | Cost-effective |
Amazon Titan
| Model ID | Type | Description |
|---|---|---|
amazon.titan-text-premier-v1:0 | Text | Premier text model |
amazon.titan-text-express-v1 | Text | Fast generation |
amazon.titan-embed-text-v2:0 | Embeddings | Text embeddings |
amazon.titan-image-generator-v2:0 | Image | Image generation |
Cohere
| Model ID | Type | Description |
|---|---|---|
cohere.command-r-plus-v1:0 | Chat | Most capable |
cohere.command-r-v1:0 | Chat | Balanced |
cohere.embed-english-v3 | Embeddings | English embeddings |
cohere.embed-multilingual-v3 | Embeddings | Multilingual |
AI21 Labs
| Model ID | Description |
|---|---|
ai21.jamba-1-5-large-v1:0 | Latest Jamba |
ai21.jamba-1-5-mini-v1:0 | Compact Jamba |
Stability AI
| Model ID | Type | Description |
|---|---|---|
stability.stable-diffusion-xl-v1 | Image | SDXL 1.0 |
stability.sd3-large-v1:0 | Image | Stable Diffusion 3 |
Authentication Methods
1. Access Keys (Default)
2. Assumed Role
3. IAM Role (EC2, ECS, Lambda)
4. Environment Variables
Advanced Features
Streaming
Function Calling (Converse API)
Embeddings
Image Generation
Batch Inference
Create batch jobs for cost-effective inference:Cross-Region Inference
Use inference profiles for cross-region routing:Multi-Region Configuration
Load balance across AWS regions:Fallback Configuration
Fallback from Bedrock Claude to Anthropic:Error Handling
Best Practices
- Use IAM roles - More secure than access keys
- Enable VPC endpoints - Private connectivity
- Request model access - Models require explicit access approval
- Use inference profiles - Better availability and routing
- Monitor with CloudWatch - Track usage and costs
- Set up guardrails - Content filtering and safety
- Use batch inference - Cost-effective for large workloads
- Implement retry logic - Handle throttling gracefully
Model Access
Before using models, request access in the AWS Console:- Go to AWS Bedrock Console
- Navigate to Model access
- Click Manage model access
- Select models and request access
- Wait for approval (usually instant)
Regional Availability
Bedrock is available in multiple AWS regions:- US: us-east-1, us-west-2
- Europe: eu-central-1, eu-west-1, eu-west-3
- Asia Pacific: ap-southeast-1, ap-northeast-1, ap-south-1
Pricing
Bedrock pricing includes:- On-demand: Pay per request/token
- Provisioned throughput: Reserved capacity
- Model customization: Additional costs for fine-tuning
AWS Bedrock Pricing
View detailed Bedrock pricing
Related Resources
Anthropic
Direct Anthropic integration
Load Balancing
Multi-region load balancing
Guardrails
Content filtering
Batch Processing
Batch inference guide