Yvo.Schedule
Back to archive012 entries

The index

A back-of-the-book index of the ideas the archive develops. Each concept points at the specific places where it lives across the essays. Unlike tags, which bucket whole posts, this indexes thinking.

A
Agents

A program that uses an LLM to take actions against a goal — not just generate text.

How to build your first AI agent in an afternoon
C
Client hype-cycle triage

Diagnosing whether a stalled project needs technical work, ownership work, or narrative work — and refusing to do all three at once.

Why your AI pilot is stuck (and what to do about it)
E
Evals as contracts

Reframing evaluation as a binding agreement between engineering, business, and compliance — rather than a test suite nobody owns.

Evals as contracts: a better way to know if your agent is working
F
Few-shot examples as spec

Using two or three input/output examples as a task specification, in place of rule-based instructions.

The three prompt patterns I actually use in production
G
Guardrails

Explicit constraints on an agent's behavior; the antidote to 'it should figure it out autonomously'.

How to build your first AI agent in an afternoon
I
Iteration over architecture

Shipping a v1 with hardcoded data beats designing a perfect v∞ that never ships.

How to build your first AI agent in an afternoon
L
LangGraph

A stateful workflow framework for LLM applications built around cyclic graphs.

How to build your first AI agent in an afternoon
P
Pilot purgatory

The state where an AI initiative has technically succeeded in a demo but has no clear path to production use.

Why your AI pilot is stuck (and what to do about it)
Q
Quality contract

A one-page document naming the specific failure modes, acceptable rates, measurement windows, and automatic actions for an AI system.

Evals as contracts: a better way to know if your agent is workingWhy your AI pilot is stuck (and what to do about it)
R
Role priming with negatives

System prompts that specify what the assistant is forbidden to do, with concrete examples. Does more work than telling it what to do.

The three prompt patterns I actually use in production
S
Structured output

Treating the model's response as a typed API payload rather than free-form text to be heuristically parsed.

The three prompt patterns I actually use in production
T
The three-failures rule

Start any eval design by listing the three failure modes you would personally be fired for. Those are the contract's load-bearing clauses.

Evals as contracts: a better way to know if your agent is working