/challenge - Red-Team Review¶
| Runtime | ~5 minutes |
| Reads | Target document, theme claude.md, people.md, vault history |
| Writes | Challenge report to stdout, fixes to tasks.md |
| Model | Opus (5 parallel subagents) |
What It Does¶
Stress-tests any document, plan, or idea through five parallel analysis lenses before it leaves the building. Returns a verdict, the top issues to fix, and a pre-mortem of how it fails.
Why It Matters¶
Every important document has blind spots. The author can't see them because they built the argument. A single reviewer applies one perspective. Serial critique is slow and tends to anchor on the first problem found.
/challenge runs five independent reviewers simultaneously, each looking through a different lens. The result arrives in one pass, not five rounds of "what do you think?"
How It Works¶
graph TD
A[Target Document] --> B[Context Gathering]
B --> C1[Audience Fit]
B --> C2[Logic & Evidence]
B --> C3[Vault Contradictions]
B --> C4[Known Blind Spots]
B --> C5[Pre-Mortem]
C1 --> D[Consolidated Challenge Report]
C2 --> D
C3 --> D
C4 --> D
C5 --> D
D --> E[Top 3 Issues + Verdict]
- Identify target - Accepts a file path, topic keyword, or (with no arguments) challenges whatever you've been working on this session
- Gather context - Reads theme files, stakeholder profiles, and previous versions to understand what the document is trying to do and for whom
- Launch five parallel subagents - Each receives the full document plus context, each applies its specific lens
- Consolidate - Merges findings into a single report with a verdict (Strong / Needs Work / Weak)
- Extract tasks - Any actionable fixes go straight into
tasks.md
The Key Innovation¶
Parallel lenses, not sequential critique. The five subagents run simultaneously as Task workers:
- Audience Fit - Would the target reader actually finish this? Does it match their stated communication preferences from
people.md? Is the ask clear in the first 30 seconds? - Logic & Evidence - Are claims backed by specific numbers or just directional language? Would a sceptic find it compelling?
- Vault Contradictions - Searches the vault for conflicts. Has the positioning shifted without acknowledging the change? Do prior meeting notes contradict what's being proposed?
- Known Blind Spots - Applies your documented blind spots from CLAUDE.md. Architecture without execution mechanics? Culture angle missing? Greenfield bias showing?
- Pre-Mortem - How does this fail? Top 3 failure modes, strongest counter-argument, the question your audience will ask that isn't answered.
This matters because a single sequential review anchors on one perspective. Parallel execution means the pre-mortem doesn't know what the audience-fit lens found, and vice versa. Independent analysis produces genuinely different angles.
Example Usage¶
Or challenge by topic:
Or challenge what you've been working on:
Output includes a verdict, the top 3 fixes, and specific sections for each lens. The "What's Actually Good" section prevents the review from feeling purely negative - but it never contains filler praise.
Customisation Guide¶
- Blind spots - The fourth lens pulls from your Known Blind Spots section in CLAUDE.md. The more specific your blind spots, the more useful this lens becomes. Review quarterly.
- Stakeholder profiles - Audience Fit checks
people.mdfor communication preferences. Richer profiles produce sharper audience analysis. - Vault depth - The Vault Contradictions lens searches meeting notes, emails, and status docs. The more history in your vault, the more contradictions it catches.
- Stakes calibration - The skill scales effort to consequence. A Slack message gets a lighter treatment than a board paper. You can override by saying "full challenge" or "quick challenge".