Embodies the Agentic AI paradigm as a concrete executable artifact - a single, autonomous instance with a goal, access to tools, memory, and a control loop that can be deployed, monitored, and versioned as a production software component.
Category
Abstraction level
Operation level
2022
Coding automation
Multi-file editing
Autonomous task resolutionComputer user agents β agents controlling the mouse, keyboard, and browser.Customer support agents with access to a knowledge base and actions.Researching and synthesizing information from multiple sourcesData analysis agents β iterative code execution and result visualizationVirtual assistants β scheduling, email management, calendar organization, shopping.High-level planning for robotic agents controlling robot actions
The agent receives a goal from the user and definitions of available tools (JSON/OpenAPI/MCP schemas) and system instructions (role, security policies). In each iteration of the loop: (1) the model analyzes the current context and decides on the next action - invoke a tool, ask a question, or terminate; (2) the host executes the chosen tool and returns the result; (3) the result is appended to the context as an observation; (4) the model decides whether to continue. The loop ends when the model determines the goal has been achieved, the max_steps limit is reached, or an error state requiring escalation to a human is detected. The agent can maintain short-term memory within the context window and long-term memory in an external storage (vector database, key-value store).
A single LLM invocation is unable to handle open-ended tasks, where the number of steps is not known in advance, requiring interaction with the environment, access to current data, code execution, or iterative result verification. The AI Agent solves this problem by embedding the model in a control loop with access to tools and memory, enabling autonomous end-to-end task execution.
01
Base language model
Reasoning and control loop engine
The agent's reasoning and decision engine. Generates plans, selects tools, interprets results, and decides termination. Typically an LLM post-trained with RLHF and tool-use, optionally a reasoning model (CoT with a dedicated reasoning token budget).
02
Tool Interface
Bridge between the agent and the external world.
Modular
Definitions of callable functions with their schemas (JSON Schema, OpenAPI) and documentation. Anthropic calls this the Agent-Computer Interface (ACI) β care in designing it is critical for agent reliability. Often exposed through Model Context Protocol.
03
Memory System
Persistence of task state and context
Modular
Short-term memory (conversation history, tool results in context) and optional long-term memory (vector store, key-value store, episodic structures) across sessions. Determines coherence and personalization in long-running tasks.
Defines the agent's identity and operational boundaries.
Constant instruction defining the agent's identity, goal, scope of responsibility, safety rules, response format, and termination criteria. First line of defense against misbehavior and prompt injection.
06
Guardrails / Safeguards
Safety and compliance oversight
Modular
Filters and validators operating before inference (input sanitization), during (tool call schema validation), and after (output control, PII redaction, blocking irreversible actions). Critical for production safety.
07
Evaluation and Observability
Production observability and quality
Modular
Step logging, traces (LangSmith, Arize, Helicone), metrics (success rate, tool error rate, average steps, cost per task), and automated evaluations against test sets. Essential for production maintenance of an agent.
Parallelism
Conditionally parallel
Parallelism is most often achieved inter-sessionally (multiple agents for different tasks) or in orchestrator-workers patterns (one orchestrator delegates to multiple worker agents simultaneously).
Paradigm
Conditional
Input dependent
An agent is distinct from a workflow, where the path is predefined in the code and the LLM only executes specific steps, whereas in an agent, the LLM controls the entire process.
Poziom autonomii
Critical
proposal_onlyAgent prepares, human executes.
confirm_irreversibleAutonomy with confirmation required for irreversible actions.
full_autonomy
Scope of decisions the agent makes without human approval β from proposal-only mode to full autonomy with rollback.
Toolkit
Critical
web_search + code_exec
file_read + file_write + bash + git
browser_use (computer use)
List of callable functions available to the agent. Defines the action space and is the strongest predictor of agent behavior.
Maximum Number of Steps
Standard
10Short tasks, low budget.
50-200Coding agents, long-term research.
Hard limit on loop iterations before forced termination. Safeguards against cost overrun and infinite loops.
Memory Strategy
Standard
in_context_only
in_context + summarization
in_context + vector_memory
How context is managed across steps and sessions: context window only, summarization, vector store, episodic structures.
Cost / Token Budget
Standard
10k_tokens
1M_tokens / 5_USD
Maximum computational cost or token count per agent run. Critical for production deployments with outcome-based billing.
Human fallback strategy
Standard
before_irreversible_actions
after_3_consecutive_failures
When the agent escalates to a human: never, on demand, after N failed steps, before irreversible action, based on uncertainty signals.
Common pitfalls
Poorly Designed Agent-Computer Interface (ACI)
CRITICAL
Unclear tool names, missing examples, ambiguous parameters β the same problems that affect junior developers affect the model. Anthropic reports spending more time optimizing tools than the agent prompt itself.
Each tool definition should include usage examples, edge cases, input format, and boundaries with other tools. Use absolute paths, clear types, and poka-yoke (error prevention through argument structure). Test iteratively on real-world examples.
Hallucinations in Action (Execution Claims)
CRITICAL
The agent may claim to have performed an action it didn't actually execute, or invoke tools with fabricated parameters β particularly dangerous in multi-step pipelines where errors propagate.
Validate all tool calls and schema; enforce ground truth through execution results; do not rely on the model's internal declarations about the system state β verify the actual state using tools.
Infinite loops and compounding errors
HIGH
Without a hard max_steps and repetition detection, the agent can loop indefinitely, generating wrong steps based on previously wrong observations. Costs grow linearly with the number of steps.
Set explicit max_steps, detect repeating action signatures, use an evaluator-optimizer to enforce termination conditions, and log costs in real-time.
Tool result prompt injection
CRITICAL
Malicious instructions embedded in web pages, documents, or emails the agent reads can hijack its behavior by impersonating system instructions.
Structurally isolate untrusted content by using role labeling and sandbox tags, require explicit user confirmation for actions resulting from observed content, and apply content filtering.
Irreversible actions taken without human oversight
CRITICAL
An agent with access to write-capable tools (delete, send_email, db_write, payment) can cause real damage based on flawed reasoning. Consequences may be irreversible.
Grant the agent minimal privileges, prefer reversible operations such as soft delete and drafts instead of send, and require human-in-the-loop before taking irreversible actions.
Context Window Overflow
HIGH
Accumulated action history and tool results can exceed the model's context window, causing silent truncation of earlier steps and loss of relevant information.
Implement compacting or summarizing of history, offload tool results to external storage with a reference, and monitor token budget per step.
Building an agent where a workflow would suffice
MEDIUM
Anthropic strongly recommends: don't build an agent when the task has a known, predefined structure. A workflow is cheaper, faster, more predictable, and easier to debug than an agent.
First, try a solution with a single LLM call plus retrieval, then workflow with chain, routing, or parallelization; introduce an agent only when the task is open-ended and the number of steps is unpredictable.
Russell and Norvig formalize rational agents; Belief-Desire-Intention architectures emerge (Rao and Georgeff). The canon is defined: an agent perceives the environment and takes goal-oriented actions.
2022
ReAct β LLM as Loop Reasoning-Action Engine
breakthrough
Yao et al. (2022) demonstrate that LLMs can interleave Chain-of-Thought with tool calls in a single loop. The practical definition of an LLM-based AI Agent.
AI Agents (Autonomous Agents) - First Wave of Autonomous Agents
Virally popular implementations show autonomous GPT-4 agents performing multi-step tasks. Despite limited reliability, they show the potential and popularize the term.
2023
Function calling in main model APIs
breakthrough
OpenAI (June 2023) introduces function calling in GPT-4; Anthropic and Google follow. First-class agent support at the level of commercial model APIs.
2024
Building Effective Agents β Formal Definition of Agent vs Workflow
breakthrough
Anthropic publishes (December 2024) guidelines distinguishing agents from workflows and five composition patterns. The canonical definition of an agent: a system in which an LLM dynamically directs its own process.
Anthropic introduces Computer Use in Claude (October 2024) β the agent clicks, types, and moves the mouse like a human. OpenAI's Operator (2025) follows. Opens a class of GUI-driven agents independent of APIs.
2025
Model Context Protocol β standard for agent tools interoperability
Anthropic releases MCP as an open standard for connecting LLMs to external tool servers. Enables an ecosystem of tools portable across model providers.
2026
Commercialization of agents in the Agents-as-a-Service model
Sierra (March 2026) announces the Agents-as-a-Service paradigm β customers buy outcomes delivered by an agent rather than SaaS access. Agents become the unit of product delivery, not just a technical library.
Base LLM inference dominates the agent's cost and latency; GPUs with tensor cores are the standard for all modern production-grade models.
TPUGOOD
Google deploys Gemini-based agents on TPUs; comparable throughput and cost to GPUs for most workloads.
Hardware agnosticGOOD
The control loop, tool parsing, and orchestration layer itself is lightweight and runs on CPU; hardware requirements stem from the base model, not from the agent's construction.
BUILT ON
LLM
A Large Language Model (LLM) is a class of machine learning models based on the Transformer architecture, trained on large text datasets via autoregressive language modeling (next-token prediction). These models have billions of parameters and can generate coherent text, answer questions, write code, translate languages, and perform many other language-cognitive tasks without task-specific fine-tuning. The term covers models such as GPT, LLaMA, Gemini, Claude, and Mistral. Most modern LLMs are instruction-tuned (SFT + RLHF) after the pre-training phase.
Tool-augmented LLM is an architectural pattern in which a large language model is equipped with access to one or more external tools that it can invoke during inference by generating structured function-call or API-call outputs. The model learns when and how to call tools by producing special tokens or structured output (e.g., JSON function calls) that are intercepted by a host runtime, executed against the tool, and whose results are returned to the model as new context for continued generation.
The canonical formalization appeared in the Toolformer paper (Schick et al., Meta AI, 2023), which demonstrated that LLMs can learn to self-supervise their own tool-use through API call annotation without requiring large labeled datasets. Toolformer showed that models trained this way can decide which tools to call, when, and with which arguments, and that tool use substantially improves performance on tasks requiring fresh information, arithmetic, multilingual lookup, and question answering.
The pattern encompasses several mechanisms: (1) in-context tool specification, where tool interfaces are described in the system prompt or context (JSON Schema, OpenAPI, natural language); (2) function calling APIs, where the model produces structured output matched to a defined schema and the host dispatches the call; (3) ReAct-style interleaving, where the model alternates reasoning traces with tool-use observations; and (4) parallel tool calling, where the model emits multiple tool calls simultaneously to be executed concurrently.
Key implementations include OpenAI function calling (GPT-4, June 2023), Anthropic tool use (Claude, 2023), Google Gemini function calling, and the Model Context Protocol (MCP, 2024) which standardizes tool server connectivity.
Chain-of-Thought (CoT) Reasoning is a prompting technique introduced by Wei et al. (2022) in which a large language model is induced to generate a series of intermediate natural-language reasoning steps as part of its output, prior to producing a final answer. The technique was shown to significantly improve LLM performance on arithmetic, commonsense, and symbolic reasoning benchmarks where standard few-shot prompting yields flat or poor results.
In the original formulation (few-shot CoT), a small number of exemplar question-answer pairs are included in the prompt, where each answer consists of a chain of thought followed by the final answer. The model learns from these demonstrations to produce its own reasoning chains. A subsequent zero-shot variant (Kojima et al., 2022) showed that appending the phrase 'Let's think step by step' to a question is sufficient to elicit reasoning chains from large models without any exemplars.
CoT is an emergent property: empirical results in the originating paper show that reasoning ability via CoT prompting appears only in models above a certain parameter threshold (approximately 100B parameters for the models tested in 2022), with smaller models not benefiting or performing worse. This relationship has been revisited by subsequent work as smaller models have been fine-tuned on CoT data.
Key extensions include Self-Consistency CoT (Wang et al., 2022), which samples multiple reasoning paths and selects the most frequent final answer; Tree of Thoughts (Yao et al., 2023), which frames reasoning as search over a tree of intermediate thoughts; and native reasoning models such as OpenAI o1 (2024) and DeepSeek-R1 (2025), which internalize extended reasoning through reinforcement learning on process reward signals rather than relying on prompting.
Agentic AI denotes an architectural transition from single-turn, stateless generative models toward goal-directed systems capable of autonomous perception, planning, action, and adaptation through iterative control loops. An agentic system wraps a large language model in a runtime that gives the model access to tools (web search, code execution, APIs, file I/O), persistent memory, and feedback from prior steps. The model then decides dynamically which tools to call, in what order, and whether to loop or stop, rather than following a predefined code path.
Two primary system types are commonly distinguished: (1) Workflows, in which LLMs and tools are orchestrated through predefined code paths, and (2) Agents, in which the LLM itself directs its process and tool usage dynamically. Both can be composed into multi-agent systems where specialized agents collaborate, with one acting as orchestrator and others as subagents. Key design patterns identified by Anthropic (2024) include prompt chaining, routing, parallelization, orchestrator-workers, and evaluator-optimizer loops. Andrew Ng's 2024 taxonomy describes four foundational patterns: Reflection, Tool Use, Planning, and Multi-Agent Collaboration.
Formal frameworks model agentic control loops as Partially Observable Markov Decision Processes (POMDPs). The control loop is: perceive state β reason/plan β select action β execute tool β observe result β update state β repeat. Agentic systems introduce risks not present in single-turn models, including hallucination in action, prompt injection through observed content, infinite loops, reward hacking, and tool misuse.
Tool-augmented LLM is an architectural pattern in which a large language model is equipped with access to one or more external tools that it can invoke during inference by generating structured function-call or API-call outputs. The model learns when and how to call tools by producing special tokens or structured output (e.g., JSON function calls) that are intercepted by a host runtime, executed against the tool, and whose results are returned to the model as new context for continued generation.
The canonical formalization appeared in the Toolformer paper (Schick et al., Meta AI, 2023), which demonstrated that LLMs can learn to self-supervise their own tool-use through API call annotation without requiring large labeled datasets. Toolformer showed that models trained this way can decide which tools to call, when, and with which arguments, and that tool use substantially improves performance on tasks requiring fresh information, arithmetic, multilingual lookup, and question answering.
The pattern encompasses several mechanisms: (1) in-context tool specification, where tool interfaces are described in the system prompt or context (JSON Schema, OpenAPI, natural language); (2) function calling APIs, where the model produces structured output matched to a defined schema and the host dispatches the call; (3) ReAct-style interleaving, where the model alternates reasoning traces with tool-use observations; and (4) parallel tool calling, where the model emits multiple tool calls simultaneously to be executed concurrently.
Key implementations include OpenAI function calling (GPT-4, June 2023), Anthropic tool use (Claude, 2023), Google Gemini function calling, and the Model Context Protocol (MCP, 2024) which standardizes tool server connectivity.
Chain-of-Thought (CoT) Reasoning is a prompting technique introduced by Wei et al. (2022) in which a large language model is induced to generate a series of intermediate natural-language reasoning steps as part of its output, prior to producing a final answer. The technique was shown to significantly improve LLM performance on arithmetic, commonsense, and symbolic reasoning benchmarks where standard few-shot prompting yields flat or poor results.
In the original formulation (few-shot CoT), a small number of exemplar question-answer pairs are included in the prompt, where each answer consists of a chain of thought followed by the final answer. The model learns from these demonstrations to produce its own reasoning chains. A subsequent zero-shot variant (Kojima et al., 2022) showed that appending the phrase 'Let's think step by step' to a question is sufficient to elicit reasoning chains from large models without any exemplars.
CoT is an emergent property: empirical results in the originating paper show that reasoning ability via CoT prompting appears only in models above a certain parameter threshold (approximately 100B parameters for the models tested in 2022), with smaller models not benefiting or performing worse. This relationship has been revisited by subsequent work as smaller models have been fine-tuned on CoT data.
Key extensions include Self-Consistency CoT (Wang et al., 2022), which samples multiple reasoning paths and selects the most frequent final answer; Tree of Thoughts (Yao et al., 2023), which frames reasoning as search over a tree of intermediate thoughts; and native reasoning models such as OpenAI o1 (2024) and DeepSeek-R1 (2025), which internalize extended reasoning through reinforcement learning on process reward signals rather than relying on prompting.
Multi-Agent Systems (MAS) are a paradigm in Distributed Artificial Intelligence in which multiple autonomous software entities β agents β interact within a shared environment to achieve individual or collective goals. Each agent perceives its environment through sensors or interfaces, reasons about its state, and acts through actuators or API calls. In the context of LLM-based MAS (emerging prominently from 2023 onward), agents are powered by large language models that provide the cognitive core (planning, reasoning, natural language communication), supplemented by memory modules, tool-use interfaces, and role-specific prompts. The system architecture defines how agents coordinate: coordination topologies include sequential pipelines, hierarchical orchestration (orchestrator-worker), parallel fan-out/fan-in, publish-subscribe messaging, and decentralized peer-to-peer communication. Core agent properties, as defined by Wooldridge and Jennings (1995), include autonomy, social ability, reactivity, and pro-activeness. In LLM-based systems, key components are: the agent (an LLM with a system prompt defining its role), a communication channel (natural language messages, structured function calls, or shared memory), an orchestrator or coordinator (managing task decomposition, routing, and state), tool-use interfaces (external APIs, code execution, web search), and a memory subsystem (short-term context, long-term vector storage). Prominent frameworks implementing LLM-based MAS include AutoGen (Microsoft, 2023), CAMEL (2023), MetaGPT (2023), CrewAI, and LangGraph.
Model Context Protocol (MCP) is an open protocol developed by Anthropic and released in November 2024. It addresses the MΓN integration problem in AI systems: connecting M different LLM applications to N different external tools previously required MΓN bespoke connectors. MCP defines a standardized client-host-server architecture where hosts (LLM applications) manage one or more clients, each maintaining a stateful session with a specific server. Servers expose capabilities as three primitives: Resources (structured data for context), Prompts (templated instructions), and Tools (executable functions). Clients expose two primitives: Roots (filesystem entry points) and Sampling (server-initiated LLM completions). Communication is based on JSON-RPC 2.0. Capability negotiation occurs at session initialization. The protocol is transport-agnostic and has been implemented in Python, TypeScript, C#, and Java SDKs. In December 2025, Anthropic donated MCP governance to the Agentic AI Foundation (AAIF) under the Linux Foundation.
Retrieval-Augmented Generation (RAG) was introduced by Lewis et al. (2020) as a general-purpose fine-tuning recipe combining pre-trained parametric memory (a seq2seq language model, specifically BART in the original paper) with non-parametric memory (a dense vector index of Wikipedia, accessed via Dense Passage Retrieval, DPR). In the original formulation, both the retriever and the generator are fine-tuned end-to-end: given an input query x, the retriever retrieves top-k documents z from the corpus, and the generator produces an output y conditioned on x and z. Two formulations were proposed: RAG-Sequence (the same retrieved documents condition the full output sequence) and RAG-Token (different documents may be used per generated token, marginalized during generation).
In widespread contemporary usage (post-2022, with the growth of LLM applications), 'RAG' has expanded to describe a broader class of retrieve-then-generate pipelines, typically with a frozen LLM, a vector store containing pre-computed dense embeddings of document chunks, and a retrieval step that fetches top-k relevant chunks based on embedding similarity to the query. The retrieved chunks are appended to the prompt as context before the LLM generates a response. This non-trainable pipeline variant is technically distinct from the original Lewis et al. formulation but is the dominant practical interpretation of RAG as of 2023β2025.
The canonical modern RAG pipeline consists of an offline indexing phase (document chunking, embedding computation, storage in a vector database) and an online query phase (query embedding, approximate nearest neighbor search, context-augmented generation). Key design decisions include: chunk size and overlap, embedding model choice, retrieval strategy (dense, sparse/BM25, or hybrid), number of retrieved documents k, and context integration method (prepend to prompt, cross-attention injection, or fusion-in-decoder).
RAG addresses two fundamental limitations of parametric-only LLMs: the knowledge cutoff problem (inability to access post-training information) and hallucination (generation of factually incorrect content). However, RAG introduces its own failure modes, including retrieval of irrelevant or misleading context and the LLM's susceptibility to being distracted by retrieved content that contradicts its parametric knowledge.
A reasoning model (also: large reasoning model, LRM, reasoning language model, RLM) is a type of large language model that has been specifically post-trained to solve complex multi-step problems by explicitly generating intermediate reasoning steps before committing to a final response. Unlike standard LLMs that generate a direct response in a single forward pass, reasoning models allocate additional computation at inference time β a property known as test-time compute scaling β by producing a long internal chain of thought (CoT). The reasoning trace typically includes steps such as problem decomposition, hypothesis generation, self-verification, reflection, and correction of errors.
The defining characteristics of reasoning models are: (1) post-training via large-scale reinforcement learning (RL) using reward signals based on final answer correctness (and sometimes intermediate step quality via process reward models); (2) the emergence of extended, often hidden, reasoning traces that precede the final answer; (3) a consistent empirical relationship between the length or computational budget allocated to the reasoning trace and final answer quality (test-time scaling law); (4) superior performance on verifiable tasks requiring multi-step logic, such as mathematics, competitive programming, and scientific reasoning.
The term 'reasoning model' was introduced as a product category by OpenAI in September 2024 with the release of the o1-preview model. OpenAI described o1 as trained via a large-scale RL algorithm teaching the model to use chain of thought productively. The approach does not rely on explicit tree search algorithms; instead, implicit search emerges via RL-trained CoT generation. In January 2025, DeepSeek published the first detailed open technical description of this class of models in the DeepSeek-R1 paper (arXiv:2501.12948), demonstrating that reasoning capabilities can be incentivized via pure RL without supervised fine-tuning, using Group Relative Policy Optimization (GRPO) as the RL framework.
Reasoning models typically employ the same base Transformer decoder architecture as standard LLMs, with the key difference residing entirely in the post-training pipeline: RL replaces or augments standard RLHF/SFT, and reward signals are grounded in verifiable outcomes. The resulting models generate substantially longer token sequences during inference (reasoning tokens), which are often hidden from end users but incur real compute costs. Performance consistently improves with both more training-time RL compute and more inference-time thinking budget.
Agents as a Service (AaaS) is a software delivery model in which a vendor provides the customer with an autonomous AI agent that performs concrete business tasks, instead of a traditional human-operated application. The term was publicly introduced on March 25, 2026, by Sierra co-founders Bret Taylor and Clay Bavor in a blog manifesto announcing their Ghostwriter agent as their own realization of this paradigm.
Unlike Software as a Service (SaaS), where customers buy access to an interface (menus, form fields, tables) and perform the work themselves through clicks, in AaaS the customer defines a desired outcome in natural language, and the vendor delivers an agent that builds and improves production agents or performs the work directly. The defining property is full autonomy: the agent has access to data, tools, a sandboxed test environment, and the deployment pipeline, while the human acts as a supervisor approving changes.
The key technical enabler is the agent harness β scaffolding of tools, memory, planning, and task context β combined with refactoring the platform into headless infrastructure that an agent can invoke programmatically rather than navigating through a UI. The work cycle includes analyzing interactions, identifying improvement opportunities, validating in a sandbox, and preparing for review β which Sierra calls an "agent assembly line." AaaS is tightly coupled with the Agentic AI paradigm (its technical foundation) and outcome-based billing models (its commercial superstructure).