AI Agent Cost Calculator
Cost Breakdown
Component Breakdown (per message)
Model Specifications
Complete LLM Model Comparison
Compare pricing, performance, and capabilities across 29+ LLM models from OpenAI, Anthropic, Google, DeepSeek, Meta, and xAI. Use this comprehensive comparison to choose the best model for your AI agent based on cost, speed, intelligence, and reasoning capabilities.
| Model | Provider |
Input Price (per 1M tokens) |
Output Price (per 1M tokens) |
Speed (tokens/sec) |
Intelligence (index) |
Context Window (tokens) |
TTFT (ms) | Reasoning |
Max Thinking (tokens) | Pricing Docs |
|---|---|---|---|---|---|---|---|---|---|---|
| GPT-4.1 | OpenAI | $2.00 | $8.00 | 38 | 53 | 1,000,000 | 850 | — | — | Link |
| GPT-4.1 mini | OpenAI | $0.40 | $1.60 | 940 | 65 | 1,000,000 | 950 | — | — | Link |
| GPT-4o | OpenAI | $5.00 | $15.00 | 191.3 | 40 | 128,000 | 400 | — | — | Link |
| GPT-4o mini | OpenAI | $0.15 | $0.60 | 200 | 40.24 | 128,000 | 450 | — | — | Link |
| GPT-5 | OpenAI | $1.25 | $10.00 | 126.2
(126.2 thinking)
| 44
(68 thinking)
| 400,000 | 460
(15,000 thinking)
| ✓ | 128,000 | Link |
| GPT-5 mini | OpenAI | $0.25 | $2.00 | 70.1
(170 thinking)
| 61
(64 thinking)
| 400,000 | 29,820
(14,600 thinking)
| ✓ | 128,000 | Link |
| GPT-5 nano | OpenAI | $0.05 | $0.40 | 219
(219 thinking)
| 29
(51 thinking)
| 400,000 | 900
(900 thinking)
| ✓ | 128,000 | Link |
| Claude 3.7 Sonnet | Anthropic | $3.00 | $15.00 | 75
(81 thinking)
| 50
(57 thinking)
| 200,000 | 1,690
(1,690 thinking)
| ✓ | 128,000 | Link |
| Claude 4 Opus | Anthropic | $15.00 | $75.00 | 46.1 | 54 | 200,000 | 1,590 | — | — | Link |
| Claude 4.1 Opus | Anthropic | $15.00 | $75.00 | 44.3 | 45 | 200,000 | 2,820 | — | — | Link |
| Claude 4.5 Haiku | Anthropic | $1.00 | $5.00 | 150
(150 thinking)
| 42
(55 thinking)
| 200,000 | 500
(500 thinking)
| ✓ | 128,000 | Link |
| Claude 4.5 Sonnet | Anthropic | $3.00 | $15.00 | 63
(63 thinking)
| 49
(61 thinking)
| 200,000 | 1,800
(1,800 thinking)
| ✓ | 64,000 | Link |
| Claude Sonnet 4 | Anthropic | $3.00 | $15.00 | 57.5
(55 thinking)
| 57
(59 thinking)
| 1,000,000 | 1,430
(1,430 thinking)
| ✓ | 64,000 | Link |
| Gemini 2.0 Flash | $0.10 | $0.40 | 165.9 | 34 | 1,000,000 | 350 | — | — | Link | |
| Gemini 2.0 Flash Lite | $0.10 | $0.40 | 470 | 30 | 1,000,000 | 400 | — | — | Link | |
| Gemini 2.5 Flash | $0.30 | $2.50 | 375.6
(375.6 thinking)
| 38
(60 thinking)
| 1,000,000 | 8,350
(8,350 thinking)
| ✓ | 24,576 | Link | |
| Gemini 2.5 Flash Lite | $0.10 | $0.40 | 470 | 48 | 1,000,000 | 1,000 | — | — | Link | |
| DeepSeek-R1 | DeepSeek | $0.56 | $1.68 | 45 | 68 | 128,000 | 500 | ✓ | 64,000 | Link |
| DeepSeek-V3 | DeepSeek | $0.56 | $1.68 | 18.6 | 44 | 130,000 | 2,760 | — | — | Link |
| DeepSeek-V3.1 | DeepSeek | $0.27 | $1.00 | 180 | 60 | 130,000 | 3,000 | — | — | Link |
| DeepSeek-V3.1 Terminus | DeepSeek | $0.23 | $0.90 | 200 | 46 | 130,000 | 2,900 | — | — | Link |
| Llama 3.1 405B | Meta | $3.75 | $3.75 | 30.4 | 26 | 128,000 | 770 | — | — | Link |
| LLaMA 3.3 70B | Meta | $0.59 | $0.99 | 276 | 74 | 128,000 | 400 | — | — | Link |
| Llama 4 Maverick | Meta | $0.24 | $0.77 | 124 | 50 | 1,000,000 | 340 | — | — | Link |
| Llama 4 Scout | Meta | $0.15 | $0.40 | 120 | 43 | 10,000,000 | 390 | — | — | Link |
| Grok-3 | xAI | $3.00 | $15.00 | 27.2 | 36 | 1,000,000 | 1,500 | — | — | Link |
| Grok-4 | xAI | $3.00 | $15.00 | 75 | 73 | 2,000,000 | 1,340 | — | — | Link |
| Grok-4 Fast | xAI | $0.20 | $0.50 | 344 | 60 | 2,000,000 | 2,550 | — | — | Link |
TTFT: Time to First Token (latency before model starts generating)
TTFAT: Time to First Answer Token (latency before model starts answering, after reasoning)
Intelligence Index: Relative capability score based on benchmark performance (higher is better)
Thinking/Reasoning: Models that show their reasoning process before answering (values in parentheses show performance when reasoning is enabled)
All pricing is per million tokens. Performance metrics are approximate and may vary based on query complexity and API load.
Last updated: November 5, 2025