This calculator uses published per-query energy and CO₂ estimates from peer-reviewed and industry research, scaled by your chosen model, prompts per day, and average tokens per prompt. We use a baseline of 1,000 tokens per query (~750 words) and scale linearly from there - a 2,000-token prompt produces roughly twice the energy of a 500-token one.
Per-query baselines
- GPT-4 / GPT-4o: ~0.0029 kWh, ~1.74 g CO₂ per query
- GPT-3.5 Turbo: ~0.0006 kWh, ~0.4 g CO₂ per query
- Claude (Sonnet / Opus): ~0.003 kWh, ~1.8 g CO₂ per query
- Gemini Pro: ~0.0025 kWh, ~1.5 g CO₂ per query
- Llama 3 8B (hosted): ~0.001 kWh, ~0.6 g CO₂ per query
- Llama 3 70B (self-hosted): ~0.005 kWh, ~3.0 g CO₂ per query
Sources
- Stanford AI Index 2024 - per-query energy comparisons across model classes.
- MIT Technology Review (August 2024) - the “Making an image with generative AI uses as much energy as charging your phone” coverage of frontier-model inference costs.
- Sasha Luccioni, Yacine Jernite, Emma Strubell - “Power Hungry Processing” (Hugging Face, 2023).
- Schwartz et al. - “Green AI” (2019, Allen Institute for AI).
- EPA Greenhouse Gas Equivalencies Calculator - real-world CO₂ comparisons (miles driven, phone charges).
Real footprints can vary by 2-5x based on data center efficiency (PUE), the carbon intensity of the local grid, request batching, and prompt length. Treat these numbers as a useful relative comparison, not a precise audit. Sources current as of April 2026.