Prompt Kit

Prompt Kit: The Harness Is the Story

This kit operationalizes the core insight from the article: AI's "jagged frontier" was an artifact of single-shot usage, not a property of AI intelligence. These three prompts help you audit your work for harness-readiness, simulate the Planner-Worker-Judge architecture on a real problem, and map the evaluation meta-skills you need to develop as execution gets automated.

How to use this kit

Prompt 1 is your starting point — run it to understand which parts of your work are ready for structured AI delegation today. Prompt 2 is the engine — use it whenever you face a complex task that would normally get a mediocre single-shot AI response. It forces the decompose-parallelize-verify-iterate pattern that the article identifies as the breakthrough. Prompt 3 is your career strategy — run it once, revisit quarterly.

All three work in any capable AI assistant (ChatGPT, Claude, Gemini). Prompt 2 benefits most from thinking-capable models and long conversations — don't rush it.


Prompt 1: Verifiability Audit — Map Your Work to the Harness Frontier

Job: Decomposes your actual work into sub-tasks and classifies each one on the article's verifiability spectrum (machine-checkable → expert-checkable → judgment-dependent), then produces a prioritized action plan for structured AI delegation.

When to use: When you want to figure out how much of your job is actually harness-ready — and you suspect the answer is "more than I think."

What you'll get: A detailed decomposition of your role into 15-30 sub-tasks, each classified by verifiability tier, with a priority matrix showing what to delegate now, what needs verification infrastructure first, and what remains genuinely human.

What the AI will ask you: Your role/title, the 3-5 major work products you produce, how quality is currently evaluated in your work, and what "getting it wrong" looks like in your domain.


Prompt 2: Harness Simulator — Planner-Worker-Judge for Complex Tasks

Job: Takes a complex task you'd normally handle in a single AI prompt (and get mediocre results) and works through it using the Planner-Worker-Judge pattern — decomposing, executing sub-tasks independently, verifying, and iterating — within a single conversation.

When to use: Whenever you have a meaty problem that you know a single-shot prompt won't handle well. Strategy documents, research synthesis, complex analysis, architectural decisions, thorough investigations. This is the prompt that operationalizes "the harness is the story."

What you'll get: A multi-phase output where the AI explicitly decomposes your problem, works each piece separately, evaluates the results critically, and revises — producing substantially better output than a single-shot attempt.

What the AI will ask you: The complex task you want to work through, what "done well" looks like, and your role in verifying the output.


Prompt 3: Evaluation Meta-Skills Map — What to Get Good At Next

Job: Identifies the specific evaluation competencies in your domain that become more valuable as AI handles more execution work — the "sniff-checking" skills the article identifies as the key differentiator.

When to use: For career planning and professional development. Run it when you want to understand what skills to invest in given the shift from execution to evaluation as the locus of professional value.

What you'll get: A concrete map of evaluation skills specific to your field, rated by current importance vs. projected importance, with specific ways to develop each one.

What the AI will ask you: Your field, your current skill set, and where you've seen AI produce "confidently wrong" output in your domain.