A Field Guide to

Artificial
Intelligence
Agents

EDITION 2025

CLASSIFICATION

REFERENCE GUIDE

Vol. I — Foundations

An AI agent is a system that perceives its environment, reasons about goals, and takes actions autonomously to achieve them — often across multiple steps and tools.

Unlike a simple language model that responds to a single prompt, an agent operates in a loop: it observes, decides, acts, and reflects — repeating until a task is complete. This shifts AI from a passive oracle into an active participant in the world.


Agents can call APIs, browse the web, write and run code, manage files, coordinate with other agents, and adapt their strategy based on results. The defining characteristic is goal-directed autonomy over time.

The Agent Loop — Core Cycle

01
PERCEIVE Receive inputs & context
02
REASON Plan next action
03
ACT Use tools / call APIs
04
REFLECT Evaluate & continue
01

Type 01 · Foundation

Reactive Agents

The simplest form. Reactive agents map inputs directly to outputs using condition–action rules. They have no internal model of the world and no memory of past events. Fast, predictable, and brittle.

They excel in controlled, well-defined environments where every possible state can be anticipated and handled with a fixed rule.

Rule engines Chatbots (v1) Reflex systems

Type 02 · Memory

Model-Based Agents

These agents maintain an internal state — a partial model of the world — updated as they receive new observations. They can handle partially observable environments by reasoning about what they cannot directly sense.

This internal world-model enables richer, context-aware behavior beyond pure stimulus-response.

Conversational AI Navigation systems Diagnostic agents

Type 03 · Objectives

Goal-Based Agents

Beyond knowing the current state, goal-based agents consider desired future states. They search or plan through possible sequences of actions, choosing those that lead to the goal.

Planning and search algorithms (BFS, A*, MCTS) are the engines driving these agents toward defined objectives.

Path planners Game AI Task schedulers

Type 04 · Optimization

Utility-Based Agents

When multiple goals or paths compete, utility functions quantify how desirable each outcome is. The agent chooses the action that maximizes expected utility — balancing trade-offs under uncertainty.

Utility theory gives these agents nuanced decision-making: they don't just reach goals, they reach them well.

Recommendation engines Trading bots Resource allocators

Type 05 · Learning

Learning Agents

Learning agents improve through experience. A learning element updates the agent's knowledge using feedback from a critic that evaluates performance. A problem generator proposes exploratory actions to discover new information.

Reinforcement learning, fine-tuning, and RLHF all implement this architecture in modern AI systems.

RL agents (PPO, DQN) RLHF models Adaptive recommenders

Type 06 · Collaboration

Multi-Agent Systems

Networks of agents that communicate, coordinate, and divide labor. Each agent may be specialized — one plans, one searches, one critiques. Together they tackle tasks beyond any single agent's capacity.

Emergent behaviors, negotiation protocols, and swarm dynamics define this rapidly evolving frontier.

AutoGen CrewAI Swarm intelligence
02
Agent Type Memory Planning Learning Autonomy Best For
Reactive None None None Low Deterministic, fast responses
Model-Based Short-term state Implicit None Low–Medium Partially observable envs
Goal-Based State + goals Search / planning Limited Medium Navigation, logistics, games
Utility-Based State + preferences Optimization Possible Medium–High Multi-objective trade-offs
Learning Episodic + model Learned policy Core feature High Unknown, dynamic environments
Multi-Agent Distributed Collaborative Collective Very High Complex, large-scale tasks
03

Building Blocks

Every capable AI agent is assembled from a set of recurring components, regardless of the underlying model or framework.

  • LLM BrainThe reasoning core; language model driving decisions
  • Memory StoreShort + long-term context, vector databases
  • Tool LayerAPIs, code interpreters, browsers, file systems
  • PlannerTask decomposition and scheduling
  • EvaluatorChecks output quality and triggers retries
  • OrchestratorCoordinates sub-agents and merges results
COMPONENT — 01

Perception & Input Parsing

The agent ingests raw inputs — natural language, structured data, images, API responses — and converts them into a unified internal representation. Retrieval-augmented generation (RAG) extends perception by pulling relevant long-term memories or documents into the context window.

COMPONENT — 02

Reasoning & Planning

Using chain-of-thought, tree-of-thought, or ReAct prompting patterns, the agent breaks complex goals into sub-tasks. Modern LLM-based agents generate a plan, then execute and revise it step by step — maintaining a "scratchpad" of intermediate reasoning.

COMPONENT — 03

Tool Use & Action Execution

Agents extend their capabilities by calling external tools: web search, code execution (Python sandbox), database queries, REST APIs, browser automation, and more. The tool call is formatted as structured output, executed by a runtime, and results are fed back into the agent loop.

COMPONENT — 04

Memory & Context Management

Four layers of memory work together: in-context (the active prompt window), episodic (past interaction logs), semantic (vector-embedded knowledge), and procedural (learned skills / fine-tuned weights). Smart agents decide what to remember, compress, or forget.

COMPONENT — 05

Evaluation & Self-Correction

Before committing to a final output, the agent critiques its own reasoning, checks factual consistency, and reruns failed steps. Techniques like Reflexion and Constitutional AI build this self-improvement loop directly into the architecture.

04

Real-World Applications

  • Autonomous software engineering (write, test, deploy)
  • Scientific research acceleration
  • Enterprise workflow automation
  • Personal productivity assistants
  • Medical diagnosis and drug discovery
  • Financial analysis and trading
  • Autonomous customer support
  • Creative collaboration (writing, design)

Open Challenges

Despite remarkable progress, significant hurdles remain before agents can be trusted with high-stakes autonomy.


  • Hallucination and factual grounding
  • Long-horizon task coherence
  • Safe tool use and sandboxing
  • Alignment with user intent
  • Cost and latency at scale
  • Interpretability and auditability
  • Multi-agent coordination failures

The Road Ahead

The trajectory of AI agents is pointing toward greater autonomy, longer task horizons, and deeper integration with real-world systems.


Key research frontiers include world models (agents that simulate outcomes before acting), formal verification of agent behavior, and federated multi-agent ecosystems where specialized agents collaborate at internet scale.


The question is shifting from "can agents do this?" to "how do we govern what they do?"