GPT
OpenAI's foundation-model family · started in 2018, scaled to frontier dominance with GPT-4 and beyond.
OpenAI's foundation-model family · started in 2018, scaled to frontier dominance with GPT-4 and beyond.
Basic
GPT stands for Generative Pre-trained Transformer. OpenAI published GPT-1 in 2018, GPT-2 in 2019, GPT-3 in 2020, GPT-4 in 2023, GPT-5 in 2025. Each generation scaled parameters, training compute, and training data. Modern GPT models (GPT-5 class) are multi-modal, support reasoning, and serve as the foundation for ChatGPT's consumer product.
Deep
GPT architecture: decoder-only transformer with causal attention. Pretraining on internet-scale text with next-token prediction. Post-training includes SFT, RLHF (later DPO), and safety tuning. GPT-4 class models are mixture-of-experts · GPT-5 likely continues this direction. OpenAI ships multiple tiers per generation: flagship (GPT-5), efficient variants (GPT-5 mini, GPT-5 nano), and reasoning-specialized (o1, o3, o4-mini). Pricing varies dramatically · frontier tier $10-60 per million output tokens, efficient tier $0.10-2.
Expert
GPT-4 is believed to be ~1.8T total parameters with MoE sparsity (estimates vary · OpenAI has not disclosed). Training FLOPs estimated at 2e25 for GPT-4, possibly 5-10× for GPT-5. Compute scaling followed Chinchilla ratios until GPT-4; post-GPT-4 shifted to train-for-longer on same parameters. Post-training is OpenAI's proprietary differentiator · data mix, RLHF/DPO recipes, and safety tuning are closely held. Reasoning models in the o-series are RL-tuned on verifiable-reward tasks and use hidden CoT at inference. OpenAI serves GPT models via the API, ChatGPT, Azure OpenAI (Microsoft), and select enterprise partnerships.
GPT-5 set the 2026 pricing ceiling · $10/M output for the flagship, dropping to $0.25/M for GPT-5 nano within 6 months of launch.
Depending on why you're here
- ·Decoder-only transformer · MoE at frontier tier (unconfirmed for latest)
- ·o-series uses RL on verifiable rewards for reasoning
- ·Training FLOPs doubling per generation
- ·Use GPT-5 nano/mini for cost-sensitive workloads
- ·GPT-5 full for frontier quality · o3 for hard reasoning
- ·See /family/gpt for all variants
- ·OpenAI sets the frontier pricing ceiling · competitors price below
- ·Azure partnership is OpenAI's distribution moat
- ·ChatGPT consumer franchise drives most of OpenAI's revenue
- ·The AI family behind ChatGPT
- ·Made by OpenAI · powered the AI boom
- ·Newer versions are smarter and cheaper each year
GPT is the brand that made AI mainstream. Every frontier generation drags the rest of the market along with it.