Context · 128K+
Cheapest 128K context LLMs
Every LLM with a 128,000+ token context window. Ranked by input price per 1M tokens.
Models40
Cheapest$-1000000.00
Min context128K tokens
What this page is
128K is the modern baseline for LLM context. Every model priced per 1M tokens at 128K+ is listed here, cheapest first. For RAG pipelines, long conversations, and mid-sized documents, this is the sweet spot.
Ranked by input price
128K+ context models, cheapest first.
Top 3 cheapest 128K context LLMs
Cheapest 128K
Auto Router
input
$-1000000.00/M
output
$-1000000.00/M
Auto Router offers 2.0M context at $-1000000.00/M input. Baseline modern model.
Runner up · 128K
Body Builder (beta)
input
$-1000000.00/M
output
$-1000000.00/M
Body Builder (beta) offers 128K context at $-1000000.00/M input. Baseline modern model.
Third · 128K
Elephant
input
$0.00/M
output
$0.00/M
Elephant offers 262K context at $0.00/M input. Baseline modern model.
The price gap · cheapest vs most expensive
Cheapest
Auto Router
$-1000000.00/M
$ per 1M input tokens
Why the gap
At 128K, premium pricing pays for reasoning quality and vendor reliability, not window size. For RAG backbones, the cheap end almost always wins.
Most expensive
GPT-5 Nano
$0.05/M
$ per 1M input tokens
Frequently asked questions
For most production use cases, yes. 128K fits 300+ pages, a full API spec, or a large function library. Only reach for 200K+ when you regularly exceed 100K token prompts.