Workers in Action
Same task. Same prompt. Four completely different outputs.
Workers inject a system prompt into every execution, transforming how Claude thinks, writes, and responds.
The Experiment
We created 3 workers with extreme personalities and ran the exact same task through each one, plus a control with no worker. All 4 tasks used identical settings: maxTurns: 3, timeout: 2 min.
The shared prompt:
1. Review in small batches
Large PRs lead to superficial reviews. Aim to review changes under 400 lines — smaller diffs let you catch real bugs instead of just skimming.
2. Focus on logic, not style
Don't waste review cycles on formatting or naming preferences that a linter can catch. Spend your attention on correctness, edge cases, and potential failures.
3. Ask questions instead of making demands
"What happens if this is null?" is more productive than "Handle the null case." Questions invite discussion, reduce defensiveness, and often surface context you didn't have.
Three Principles for Conducting Rigorous Code Reviews
Drawing upon established software engineering literature and empirical evidence, I present the following three recommendations:
1. Limit Review Scope to Reduce Cognitive Load
It is well-established in the literature on cognitive load theory (Sweller, 1988) that human working memory is finite and easily overwhelmed. Empirical evidence suggests that reviewing no more than 200–400 lines of code per session yields significantly higher defect detection rates; beyond this threshold, reviewer attention degrades markedly, and critical defects are far more likely to be overlooked.
2. Employ a Structured Checklist, in the Tradition of Fagan Inspections
Michael Fagan's seminal work on formal software inspections (1976) demonstrated that systematic, checklist-driven reviews outperform ad hoc examination by a substantial margin. One should maintain and refine a review checklist that addresses recurring concerns — such as error handling, boundary conditions, naming conventions, and security vulnerabilities — to ensure consistency and completeness across all reviews.
3. Critique the Code, Not the Author — Prioritize Constructive, Specific Feedback
The psychological safety of the development team is a prerequisite for effective peer review, as established by Edmondson's research on team dynamics. Feedback should be depersonalized (e.g., "this function exceeds a single level of abstraction" rather than "you wrote this poorly") and should, wherever possible, reference concrete principles — such as the Single Responsibility Principle or the DRY principle — so that the critique is grounded in objective reasoning rather than subjective preference.
These three tenets, when applied rigorously, have been shown to improve both defect detection efficacy and team cohesion in professional software engineering practice.
Ahoy, matey! Gather 'round the mast and let this old sea dog share three treasures of wisdom fer reviewin' yer crew's code! 🏴☠️
⚓ 1. Chart the Course Before Inspectin' the Rigging
Before ye dive into the lines of code, read the pull request description and understand why the changes be made, not just what changed. A captain who doesn't know where the ship be headin' can't tell if the sails be set right!
🗡️ 2. Don't Just Hunt fer Bugs — Share Yer Treasure Map
Aye, when ye leave comments, don't just say "this be wrong!" — explain why it be wrong and offer a better headin'. A good shipmate doesn't just point out the reef, they help steer around it. Suggest alternatives and share the reasoning behind yer wisdom, so the whole crew grows stronger.
🦜 3. Keep Yer Cannons Aimed at the Code, Not the Sailor
Shiver me timbers, nothin' sinks a crew's morale faster than makin' it personal! Say "this function could be simplified" instead of "ye wrote this poorly." Review the craft, not the craftsman. A ship sails smoothest when every pirate aboard feels respected and trusted to hoist the sails.
Now hoist the Jolly Roger and go make yer crew's code seaworthy, ye scallywag! 🏴☠️
Professor:
- Limit scope to 200–400 lines per session
- Use structured checklists (Fagan inspections)
- Critique code, not the author
Pirate:
- Understand the why before reviewing
- Explain fixes, don't just flag problems
- Keep feedback aimed at code, not the sailor
No Worker:
- Review in small batches (<400 lines)
- Focus on logic, not style (let linters handle formatting)
- Ask questions instead of making demands
Side-by-Side Comparison
How the same prompt produced dramatically different results.
Workers fundamentally change how tasks execute. The same 20-word prompt produced four completely different outputs — from a 250-word academic paper with citations to an 80-word bullet summary. This isn't cosmetic: workers shape reasoning, structure, length, vocabulary, and approach.
- The Professor cited Sweller (1988), Fagan (1976), and Edmondson — zero contractions, formal numbered sections
- The Pirate delivered the same advice through nautical metaphors — "Ahoy", "matey", "shiver me timbers"
- The Minimalist compressed everything into terse bullets under 80 words — 3x shorter than the Professor
- No Worker gave standard helpful Claude output — the baseline
Workers are attached to tasks with a single field (workerId). Create once, reuse across any number of tasks. 9 templates included, or write your own system prompt.
Build your own team of workers.
Free, open source, 100% local.