You direct the agent
Claude Code or Codex writes every line. You never edit source. Learning happens in the prompts you write and the decisions you make.
Capability Engineers
This is the starter program for capability engineers. One stack, one workflow, ten assignments. You are the architect and product owner. The agent writes the code. You learn by deciding, questioning, and verifying — not by reading source.
The program
Claude Code or Codex writes every line. You never edit source. Learning happens in the prompts you write and the decisions you make.
Each phase produces one human-readable artifact: a brief, a data model, a schema, an OpenAPI doc, a running app. That artifact is the contract between phases.
Every phase ends with a check you run yourself: a SQL query, an OpenAPI "Try it out", a user story walked in the browser.
The stack
No Docker. No cloud. Two terminals, two commands. The same stack across all ten assignments, so the only variable is the thing you are building.
Database
File-based, zero install, ships with Python. Schema in a single schema.sql. You write real SQL — no ORM.
Backend
Python 3.12+, managed with uv. Pydantic for validation. Automatic OpenAPI at /docs — your API contract, visualised.
Frontend
TypeScript, plain fetch, types generated from the OpenAPI spec. Minimal styling, no component library by default.
Glue
Vite dev server proxies /api/* to :8000. No CORS. No auth. One local user. One file on disk.
Prerequisites: Python 3.12+, uv, Node.js 20+ LTS. SQLite ships with Python.
The workflow
Each phase descends from the artifact of the last. If a phase reveals a gap, go back and fix the artifact before regenerating the code.
Prompt library
Each phase has one or two prompts you can paste into your agent session. Adapt them to your assignment, but keep the structure — each prompt names the artifact it produces and the artifact it descends from.
The ten
All ten are calibrated to the same envelope: four entities give or take one, three user stories, one aggregation report, one state transition. Five come from the clinical-operations domain. Five are general-purpose. Each teaches one identifiable concept beyond CRUD.
Ideas deliberately not on this list: external data reconciliation, clinical data quality checks, raw data to SDTM mapping, central monitoring with outlier detection, TLF generation, live clinicaltrials.gov feasibility, cross-registry benchmarking. Each is a good idea and worth building later. Each fails at least one of the starter criteria — typically the ones about specialist domain knowledge, external APIs, or statistics beyond aggregation. A future intermediary catalogue will revisit them.
When you're ready
uv, Node 20+ LTS.index.html. Paste the Phase-1 prompt into a fresh agent session. Your brainstorm becomes docs/01-brief.md.