The index
A back-of-the-book index of the ideas the archive develops. Each concept points at the specific places where it lives across the essays. Unlike tags, which bucket whole posts, this indexes thinking.
- Agents
A program that uses an LLM to take actions against a goal — not just generate text.
- →How to build your first AI agent in an afternoon
- Client hype-cycle triage
Diagnosing whether a stalled project needs technical work, ownership work, or narrative work — and refusing to do all three at once.
- →Why your AI pilot is stuck (and what to do about it)
- Evals as contracts
Reframing evaluation as a binding agreement between engineering, business, and compliance — rather than a test suite nobody owns.
- →Evals as contracts: a better way to know if your agent is working
- Few-shot examples as spec
Using two or three input/output examples as a task specification, in place of rule-based instructions.
- →The three prompt patterns I actually use in production
- Guardrails
Explicit constraints on an agent's behavior; the antidote to 'it should figure it out autonomously'.
- →How to build your first AI agent in an afternoon
- Iteration over architecture
Shipping a v1 with hardcoded data beats designing a perfect v∞ that never ships.
- →How to build your first AI agent in an afternoon
- LangGraph
A stateful workflow framework for LLM applications built around cyclic graphs.
- →How to build your first AI agent in an afternoon
- Pilot purgatory
The state where an AI initiative has technically succeeded in a demo but has no clear path to production use.
- →Why your AI pilot is stuck (and what to do about it)
- Quality contract
A one-page document naming the specific failure modes, acceptable rates, measurement windows, and automatic actions for an AI system.
- →Evals as contracts: a better way to know if your agent is working→Why your AI pilot is stuck (and what to do about it)
- Role priming with negatives
System prompts that specify what the assistant is forbidden to do, with concrete examples. Does more work than telling it what to do.
- →The three prompt patterns I actually use in production
- Structured output
Treating the model's response as a typed API payload rather than free-form text to be heuristically parsed.
- →The three prompt patterns I actually use in production
- The three-failures rule
Start any eval design by listing the three failure modes you would personally be fired for. Those are the contract's load-bearing clauses.
- →Evals as contracts: a better way to know if your agent is working