AI Agent
An AI that plans multi-step workflows, uses tools, and maintains state · not a single-turn chat.
An AI that plans multi-step workflows, uses tools, and maintains state · not a single-turn chat.
Basic
An AI agent takes a goal and works toward it over many steps. It decides which tool to call, reads the output, adjusts its plan, and continues. Examples: Claude Code resolves GitHub issues over dozens of edits, Cursor's Composer refactors across files, Operator browses the web and fills forms, Manus orchestrates research tasks. Agents need three things: planning, tool use, and memory across steps.
Deep
Agent architectures vary: ReAct (reason + act interleaved), function-calling loops, tree-search planners, and hybrid designs. The dominant 2026 pattern is a reasoning model (o3, Claude Opus with Extended Thinking, DeepSeek R1) plus tool definitions via function-calling or MCP. The agent runs in a loop: reason about next step → call tool → observe result → reason again → ... until done. State management includes conversation history, file system state, and task decomposition. Safety-critical: agents can make irreversible changes (send emails, modify files, spend money). Most production agents require explicit user approval for high-risk actions.
Expert
Formal frame: an agent is a policy π(a | s, g) mapping state + goal to actions. In LLM agents the policy is the LLM + a harness that parses tool calls, executes them, and feeds results back. Planning depth is bounded by effective context length; long-horizon agents use hierarchical task decomposition to stay within context. Error recovery: agents that can detect failed tool calls and replan outperform linear-execution agents by 30-50% on SWE-bench Verified. Reflexion, Self-Refine, and Tree-of-Thought extend single-shot reasoning into multi-iteration improvement. Cost is the dominant constraint · agent runs can consume 100K-10M tokens per task.
Agent ARR is the fastest-growing segment of AI revenue in 2026. Cursor, Cognition, Claude Code, and Replit Agent all pass $100M ARR.
Depending on why you're here
- ·ReAct loop, function-calling harness, MCP-based tool integration
- ·Planning depth bounded by effective context window
- ·Reflexion + Tree-of-Thought for error recovery and exploration
- ·Start with simple function-calling loops · don't over-engineer planning
- ·Budget tokens carefully · agent runs consume 100K-10M tokens
- ·Require explicit user approval for destructive actions
- ·Agent ARR outpaces chat ARR for the first time in 2026
- ·Agent infrastructure (orchestration, observability, memory) is the next open market
- ·Winner-take-most dynamics in specific verticals (coding, customer support)
- ·An AI that doesn't just chat · it gets things done
- ·Writes code, sends emails, researches topics, books flights
- ·The next step beyond ChatGPT
Often confused with
A reasoning model thinks harder. An agent acts on the world. Reasoning is internal, agents do things. Most modern agents use reasoning models as their brain.
A chatbot responds to a message. An agent takes a goal, plans, and executes · multi-step, tool-using, stateful.
Agents are where the frontier is in 2026. Chat is a solved problem. Agents are where the next decade of AI revenue compounds.