Get an estimate of what it costs to run an AI agent built on AgentCore in your AWS account.
Amazon Bedrock AgentCore uses consumption-based pricing across seven modular services. Runtime, Browser, and Code Interpreter bill by active CPU and memory. Gateway, Memory, and Policy bill by usage units like invocations and records. Most cost calculators stop there.
This one doesn't. Model inference is the largest cost driver in most agent deployments (often 50 to 70 percent of total spend) but can vary significatly based on model choice. This calculator includes inference cost estimates and let's you compare models, so the estimate reflects what you'll actually see on your AWS bill.
The total includes AgentCore Runtime, Memory, Gateway, Knowledge Base (optional), Observability (optional), and model inference for the foundation model you select. Bedrock Prompt Caching is not modeled, but can reduce input token costs by up to 90 percent for repeated context. CloudWatch charges for Observability route through your CloudWatch account and are estimated at 15 percent of AgentCore cost. Data transfer and ECR storage for container deployments are not included.
If you're using custom memory extraction strategies or overriding the default model for long-term memory consolidation, your Memory costs will be higher than shown. The default built-in strategies include model costs in the per-record price. based on the assumptions entered.
Your foundation model choice matters more than any other input. At typical session lengths, model inference accounts for the majority of the total. Switching from Claude Sonnet to a smaller model can cut that line item in half. The calculator lets you compare directly.
Deploy a production agent yourself in under 10 minutes with our free, open source product on AWS Marketplace.
Our AgentCore Accelerate program deploys a production agent in two weeks with our team alongside you, often at no cost through AWS funding*.
*Funding based on customer eligibility and available AWS programs.
AgentCore bills each service independently. Runtime charges $0.0895 per vCPU-hour and $0.00945 per GB-hour based on active resource use. Because agents spend 30 to 70 percent of session time waiting on LLM responses or tool calls, idle CPU accrues no charge. You pay only for active consumption. Gateway charges per tool invocation. Memory charges per event stored and per retrieval. Each component can be used independently or together.
This calculator does. Many AgentCore cost tools estimate infrastructure only and ask you to add Bedrock inference separately. This calculator includes model inference as a line item, using the current AWS pricing for the model you select. For many deployments, model costs represent 50 to 70 percent of total spend, but this depends on the model you're using.
Bedrock pricing covers foundation model inference, or what you pay per token to run Claude, Nova, Llama, or another model. AgentCore pricing covers the agent infrastructure around that model: the runtime environment, session isolation, memory, tool gateway, and observability. You incur both on any agent running in production. This calculator estimates both together.
Four things fall outside this estimate: Bedrock Prompt Caching savings (can reduce input token costs significantly for agents with repeated system prompts), data transfer charges at EC2 standard rates, ECR storage for container-deployed agents, and the cost of custom model overrides in Memory extraction strategies. For most POC and early-production workloads these are minor. At scale, Prompt Caching in particular is worth modeling separately.
For simple agents with predictable session volume, the estimate here is close enough to build a business case. Cost gets harder to model when you add persistent per-user memory, multi-agent orchestration, high tool call frequency, or RAG retrieval. Those architectures have compounding cost paths that depend heavily on design choices. If you're sizing a production deployment and the number matters to stakeholders, Tech 42's AgentCore Accelerate Program includes architecture review and AWS cost modeling as part of the two-week engagement, often funded through AWS.