Archive

Articles

The full archive. If you are new to the site, start with the cornerstone guides below or the Learning Path. Use the parallel hubs for Opinions, Tools, and Platforms once you have the main frame.

Cornerstone Guides

Cornerstone Guide

Introduction to AI Agents: What They Are, How They Work, and When to Use Them

AI agents are goal-directed software systems that can use models, tools, context, and control loops to work through tasks across multiple steps. This beginner guide explains the idea without the hype.

Cornerstone Guide

What Is Agent Engineering?

Agent engineering is the discipline of designing, building, evaluating, and operating goal-directed AI systems that can reason over state, use tools, and act inside real workflows under explicit control.

Cornerstone Guide

What Is an AI Agent?

An AI agent is a goal-directed system that can observe state, decide what to do next, use tools, and act across multiple steps. Here is the clean first-principles definition, plus how agents differ from LLMs and workflows.

Cornerstone Guide

LLMs, Workflows, and Agents: What Actually Changes?

The real shift from LLM to workflow to agent is not a buzzword change. It is a change in who owns the task, the execution path, and the next-step decisions.

Cornerstone Guide

When to Use a Workflow Instead of an Agent

Use a workflow when the valid path can be defined in advance, predictability matters more than flexibility, and the task does not need runtime path-finding.

Cornerstone Guide

Tool Use: How Agents Take Action

Tool use is how an agent leaves pure text generation and interacts with external systems. Reliable tool use depends on more than choosing a function name. It depends on arguments, execution control, permissions, and verification.

Cornerstone Guide

Structured Outputs, Guardrails, and Execution Boundaries

Structured outputs constrain shape, guardrails constrain policy, and execution boundaries constrain power. Safe agent systems need all three.

Cornerstone Guide

Tracing and Observability for Agent Systems

Tracing captures what happened inside a run. Observability is the broader operating discipline that makes agent behavior legible enough to debug, evaluate, and trust in production.

Cornerstone Guide

AgentOps: Running Agents in Production

AgentOps is the operating discipline for live agent systems. It turns traces, evaluations, guardrails, and human controls into an ongoing practice for running autonomous systems safely and reliably.

Cornerstone Guide

AI Agent Frameworks

Most framework comparisons are weaker than they look because they compare tools that live at different layers of the stack. The real decision is not just which framework is popular. It is which control surface your team actually needs.

Cornerstone Guide

Tool Integration Patterns for Real Agent Systems

Tool integration is a durable agent design problem about boundaries, trust, and execution control. MCP matters, but it is one interface pattern inside a much larger tool story.

May 5, 2026

How to Review an AI Agent Demo Without Getting Fooled

A 30-minute AI agent demo can prove or disprove production readiness if you know what to test live, what to ask the builder, and what to refuse to accept as proof. The D.E.M.O. lens gives you four tells.

AI Agents / Agent Engineering / Opinions / Buyer Skepticism / Reliability

April 23, 2026

Introduction to AI Agents: What They Are, How They Work, and When to Use Them

AI agents are goal-directed software systems that can use models, tools, context, and control loops to work through tasks across multiple steps. This beginner guide explains the idea without the hype.

AI Agents / Agent Engineering / Foundations / Beginners / Workflows

April 18, 2026

Structured Outputs Are Doing More Work Than Most Teams Realize

Structured outputs are not just a formatting upgrade. In real agent systems, they help define typed boundaries around tools, routing, approvals, workflows, and downstream state.

AI Agents / Agent Engineering / Tools / Structured Outputs / Guardrails

April 17, 2026

Tool Integration Patterns for Real Agent Systems

Tool integration is a durable agent design problem about boundaries, trust, and execution control. MCP matters, but it is one interface pattern inside a much larger tool story.

AI Agents / Agent Engineering / Tools / MCP / Tooling

April 17, 2026

AI Agent Frameworks

Most framework comparisons are weaker than they look because they compare tools that live at different layers of the stack. The real decision is not just which framework is popular. It is which control surface your team actually needs.

AI Agents / Agent Engineering / Platforms / Frameworks / Tooling

April 17, 2026

The Most Common Ways Agents Fail Silently

The most dangerous agent failures are often not dramatic incidents. They are quieter losses of trust: acceptable-looking outputs hiding weaker trajectories, more rescue, noisier grounding, and rising pressure on the system's real operating limits.

AI Agents / Agent Engineering / Reliability / Evaluation / AgentOps

April 17, 2026

Traces as Test Data: Using Production Runs to Improve Agent Quality

Production traces are not just for debugging. The best ones become future quality protection: regression fixtures, scenario cases, and stronger offline evals. The trick is knowing which traces deserve promotion.

AI Agents / Agent Engineering / Foundations / Evaluation / Reliability

April 14, 2026

Online Evals vs Offline Evals

Offline evals decide whether a change deserves release. Online evals judge how the live system is actually behaving under real traffic. Production agent teams need both, and they need them for different reasons.

AI Agents / Agent Engineering / Foundations / Reliability / Evaluation

April 14, 2026

Drift, Degradation, and Slow Failure in Long-Lived Agent Systems

Many agent systems do not fail all at once. They become less trustworthy gradually: shakier trajectories, rising rescue load, weaker recoveries, and more pressure on the operating envelope long before the output fully collapses.

AI Agents / Agent Engineering / Foundations / Reliability / AgentOps

April 13, 2026

What Is Agent Engineering?

Agent engineering is the discipline of designing, building, evaluating, and operating goal-directed AI systems that can reason over state, use tools, and act inside real workflows under explicit control.

Agent Engineering / AI Agents / Foundations / Systems Design / Prompt Engineering

April 13, 2026

AgentOps Is the Missing Layer Between an AI Demo and a Real Product

Your AI demo is not your product. AgentOps is the layer that turns agent capability into something reliable, observable, governable, and worth trusting in the real world.

AI Agents / Agent Engineering / Opinions / AgentOps / Reliability

April 13, 2026

How Good Agent Memory Actually Works in Production

Good agent memory is not one vector store plus chat history. It is a governed system for deciding what gets scoped, promoted, compressed, pinned, and retrieved.

AI Agents / Agent Engineering / Tools / Memory / Context Engineering

April 13, 2026

Agent Memory Is Growing Up - Why Agents Are Starting to Remember How, Not Just What

Agent memory is changing fast. The next wave of agents will not just remember facts. They will remember workflows, compress experience, and get better at solving the next problem.

AI Agents / Agent Engineering / Opinions / Memory / Research

April 6, 2026

Reliability Reviews for Agents

Regression tests protect the next release. Reliability reviews ask a broader question: is this live agent system still trustworthy enough to keep operating as designed?

AI Agents / Agent Engineering / Foundations / Reliability / AgentOps

April 6, 2026

Regression Testing for Agents

Regression testing is the release-gate discipline that checks whether an agent got worse after a change. For agent systems, that means testing not only outputs, but also trajectories, side effects, and operating envelopes.

AI Agents / Agent Engineering / Foundations / Reliability / Evaluation

March 25, 2026

AgentOps: Running Agents in Production

AgentOps is the operating discipline for live agent systems. It turns traces, evaluations, guardrails, and human controls into an ongoing practice for running autonomous systems safely and reliably.

AI Agents / Agent Engineering / Foundations / AgentOps / Reliability

March 23, 2026

Tracing and Observability for Agent Systems

Tracing captures what happened inside a run. Observability is the broader operating discipline that makes agent behavior legible enough to debug, evaluate, and trust in production.

AI Agents / Agent Engineering / Foundations / Observability / Reliability

March 23, 2026

OpenAI Codex as a Coding-Agent Platform

OpenAI Codex is easy to mistake for just a CLI or coding product. The more useful way to understand it is as a local-first coding-agent runtime built around a shared harness.

AI Agents / Agent Engineering / Platforms / OpenAI Codex / Coding Agents

March 22, 2026

Evaluating Agent Trajectories, Not Just Outputs

A correct final answer does not prove that an agent behaved well. Agent evaluation has to judge the run itself: the sequence, tool use, recovery behavior, and policy fit that produced the answer.

AI Agents / Agent Engineering / Foundations / Evaluation / Reliability

March 20, 2026

Human-in-the-Loop Control Design

Human-in-the-loop design is not about adding vague oversight. It is about deciding where human judgment should sit in an agent system and what type of checkpoint belongs there.

AI Agents / Agent Engineering / Foundations / Human-in-the-Loop / Control Design

March 19, 2026

Supervisor, Router, and Planner-Executor Patterns

Routers dispatch, planners break work into a roadmap, and supervisors retain control across the run. The right orchestration pattern depends on where authority should live.

AI Agents / Agent Engineering / Foundations / Orchestration / Multi-Agent Systems

March 19, 2026

Structured Outputs, Guardrails, and Execution Boundaries

Structured outputs constrain shape, guardrails constrain policy, and execution boundaries constrain power. Safe agent systems need all three.

AI Agents / Agent Engineering / Foundations / Guardrails / System Design

March 18, 2026

When to Use a Workflow Instead of an Agent

Use a workflow when the valid path can be defined in advance, predictability matters more than flexibility, and the task does not need runtime path-finding.

AI Agents / Agent Engineering / Foundations / Workflows / System Design

March 18, 2026

ReAct and the Basic Reasoning Loop

ReAct is a reasoning pattern where an agent thinks about the next move, takes an action, inspects the observation, and repeats. It is useful when the next step depends on what the last step discovered.

AI Agents / Agent Engineering / Foundations / ReAct / Reasoning Loops

March 18, 2026

Goals, Constraints, and Success Conditions

Goals tell an agent what outcome to pursue. Constraints define the boundaries on how it may pursue that outcome. Success conditions define what evidence lets the run stop. Real agents need all three.

AI Agents / Agent Engineering / Foundations / Goals / Guardrails

March 17, 2026

The Autonomy Spectrum: From Stateless Calls to Goal-Directed Systems

Autonomy is not a binary property that suddenly appears when a system uses tools or takes multiple steps. It is a spectrum shaped by who chooses goals, path, actions, and recovery behavior at runtime.

AI Agents / Agent Engineering / Foundations / Autonomy / Workflows

March 15, 2026

Context Engineering: The New Core Skill

Context engineering is not a replacement for prompt engineering. It is a specialization inside prompt engineering focused on constructing the dynamic, system-heavy parts of the final prompt payload.

AI Agents / Agent Engineering / Foundations / Context Engineering / Prompt Engineering

March 15, 2026

Short-Term Context, Retrieval, and Long-Term Memory

Agents do not just need more context. They need clean separation between what the model sees now, what the system can fetch now, and what the system should still know later.

AI Agents / Agent Engineering / Foundations / Memory / Retrieval / Context Engineering

March 15, 2026

Memory: Why Agents Need More Than Context Windows

A context window determines what a model can see right now. Memory determines what an agent can preserve across time. Reliable agent systems need more than long prompts. They need continuity.

AI Agents / Agent Engineering / Foundations / Memory / Context Engineering

March 13, 2026

What Stripe's Minions Reveal About Production Coding Agents

Stripe's Minions matter because they show what coding agents look like when they are treated as delegated workers inside a real engineering system. This case study extracts the reusable architecture patterns and compares Stripe's model with Devin and Claude Code.

AI Agents / Agent Engineering / Case Studies / Coding Agents / AgentOps

March 12, 2026

Tool Use: How Agents Take Action

Tool use is how an agent leaves pure text generation and interacts with external systems. Reliable tool use depends on more than choosing a function name. It depends on arguments, execution control, permissions, and verification.

AI Agents / Agent Engineering / Foundations / Tool Use / Function Calling

March 12, 2026

Planning and Task Decomposition

Planning chooses the path toward a goal. Task decomposition turns that path into executable, verifiable subtasks. In agent systems, the quality of that breakdown often determines whether the run succeeds.

AI Agents / Agent Engineering / Foundations / Planning / Task Decomposition

March 12, 2026

The Sense-Think-Act Loop

The sense-think-act loop is the runtime pattern that makes an AI agent agentic. It turns goals and changing state into repeated bounded actions instead of one-shot responses.

AI Agents / Agent Engineering / Foundations / Control Loops / ReAct

March 11, 2026

LLMs, Workflows, and Agents: What Actually Changes?

The real shift from LLM to workflow to agent is not a buzzword change. It is a change in who owns the task, the execution path, and the next-step decisions.

LLMs / Workflows / AI Agents / Agent Engineering / Foundations

March 11, 2026

Agentic Loops - What Are They and When to Use Them

Agentic loops are bounded feedback loops that can inspect state, choose the next action at runtime, learn from feedback, and continue toward a goal inside clear boundaries.

AI Agents / Agent Engineering / Foundations / Workflows / Control Loops

March 10, 2026

What Is an AI Agent?

An AI agent is a goal-directed system that can observe state, decide what to do next, use tools, and act across multiple steps. Here is the clean first-principles definition, plus how agents differ from LLMs and workflows.

AI Agents / Agent Engineering / Foundations / Workflows / LLMs

March 10, 2026

Why Agent Engineering Is Becoming Its Own Discipline

Agent engineering is emerging because the hard problem is no longer a single prompt. It is designing closed-loop systems that can reason, retrieve context, use tools, stay governable, and hold up in production.

AI Agents / Agent Engineering / Systems Design / Context Engineering / Evals