Meta Llama Models - Token Counter & Cost Calculator
Calculate token usage and API costs for Meta's open-source Llama models. Get accurate pricing estimates for Llama 3.1 and 3.2 models via API providers.
Available Meta Llama Models
Llama 4 Maverick
Context: 1,000,000 tokens
Largest Llama 4 model with 400B total parameters, optimized for multimodal understanding, multilingual tasks, and agentic systems. Requires multiple GPUs.
Llama 4 Scout
Context: 10,000,000 tokens
Smaller Llama 4 model with 109B total parameters, can run on single GPU. Supports text + up to 5 images input, 10M token context.
Llama 3.3 70B Instruct
Context: 131,072 tokens
Enhanced performance relative to Llama 3.1 70B, approaches Llama 3.1 405B performance for some applications. Text-only model.
Llama 3 70B Instruct
Context: 8,192 tokens
Previous generation 70B model, cost-effective option via API providers like Together AI, Groq.
Llama 3 8B Instruct
Context: 8,192 tokens
Smaller, faster model starting as low as $0.05/1M tokens via providers like Replicate, DeepInfra.