| name | dev-implement |
| version | 1 |
| description | REQUIRED Phase 5 of /dev workflow. Orchestrates per-task ralph loops with delegated TDD implementation. |
Announce: "I'm using dev-implement (Phase 5) to orchestrate implementation."
Where This Fits
Main Chat (you) Task Agent
─────────────────────────────────────────────────────
dev-implement (this skill)
→ dev-ralph-loop (per-task loops)
→ dev-delegate (spawn agents)
→ Task agent ──────────────→ follows dev-tdd
uses dev-test tools
Main chat orchestrates. Task agents implement.
Contents
- Prerequisites
- The Iron Law of Delegation
- The Process
- Sub-Skills Reference
- If Max Iterations Reached
- Phase Complete
Implementation (Orchestration)
Do NOT start implementation without these:
.claude/SPEC.mdexists with final requirements.claude/PLAN.mdexists with chosen approach- User explicitly approved in /dev-design phase
If any prerequisite is missing, STOP and complete the earlier phases.
Check PLAN.md for: files to modify, implementation order, testing strategy.
MAIN CHAT MUST NOT WRITE CODE. This is not negotiable.
Main chat orchestrates. Subagents implement. If you catch yourself about to use Write or Edit on a code file, STOP.
| Allowed in Main Chat | NOT Allowed in Main Chat |
|---|---|
| Spawn Task agents | Write/Edit code files |
| Review Task agent output | Direct implementation |
| Write to .claude/*.md files | "Quick fixes" |
| Run git commands | Any code editing |
| Start ralph loops | Bypassing delegation |
If you're about to edit code directly, STOP and spawn a Task agent instead.
Rationalization Prevention
These thoughts mean STOP—you're rationalizing:
| Thought | Reality |
|---|---|
| "It's just a small fix" | Small fixes become big mistakes. Delegate. |
| "I'll be quick" | Quick means sloppy. Delegate. |
| "The subagent will take too long" | Subagent time is cheap. Your context is expensive. |
| "I already know what to do" | Knowing ≠ doing it well. Delegate. |
| "Let me just do this one thing" | One thing leads to another. Delegate. |
| "This is too simple for a subagent" | Simple is exactly when delegation works best. |
| "I'm already here in the code" | Being there ≠ writing there. Delegate. |
| "The user is waiting" | User wants DONE, not fast. They won't debug your shortcuts. |
| "This is just porting/adapting code" | Porting = writing = code. Delegate. |
| "I already have context loaded" | Fresh context per task is the point. Delegate. |
| "It's config, not real code" | JSON/YAML/TOML = code. Delegate. |
| "I need to set things up first" | Setup IS implementation. Delegate. |
| "This is boilerplate" | Boilerplate = code = delegate. |
| "PLAN.md is detailed, just executing" | Execution IS implementation. Delegate. |
The Meta-Rationalization
If you're treating these rules as "guidelines for complex work" rather than "invariants for ALL work", you've already failed.
Simple work is EXACTLY when discipline matters most—because that's when you're most tempted to skip it.
The Process
For each task N in PLAN.md:
1. Start ralph loop for task N
→ Skill(skill="workflows:dev-ralph-loop")
2. Inside loop: spawn Task agent
→ Skill(skill="workflows:dev-delegate")
3. Task agent follows TDD (dev-tdd) using testing tools (dev-test)
4. Verify tests pass, output promise
5. Move to task N+1, start NEW ralph loop
Step 1: Start Ralph Loop for Each Task
REQUIRED SUB-SKILL:
Skill(skill="workflows:dev-ralph-loop")
Key points from dev-ralph-loop:
- ONE loop PER TASK (not one loop for feature)
- Each task gets its own completion promise
- Don't move to task N+1 until task N's loop completes
Step 2: Inside Loop - Spawn Task Agent
REQUIRED SUB-SKILL:
Skill(skill="workflows:dev-delegate")
Key points from dev-delegate:
- Implementer → Spec reviewer → Quality reviewer
- Task agent follows dev-tdd protocol
- Task agent uses dev-test tools
Step 3: Verify and Complete
After Task agent returns, verify:
- Tests EXECUTE code (not grep)
- Tests PASS (SKIP ≠ PASS)
- LEARNINGS.md has actual output
- Build succeeds
If ALL pass → output the promise. If ANY fail → iterate.
Sub-Skills Reference
| Skill | Purpose | Used By |
|---|---|---|
dev-ralph-loop |
Per-task loop pattern | Main chat |
dev-delegate |
Task agent templates | Main chat |
dev-tdd |
TDD protocol (RED-GREEN-REFACTOR) | Task agent |
dev-test |
Testing tools (pytest, Playwright, etc.) | Task agent |
Failure Recovery Protocol
Pattern from oh-my-opencode: After 3 consecutive implementation failures, escalate.
3-Failure Trigger
If you attempt 3 implementations and ALL fail tests:
Iteration 1: Implement approach A → tests fail
Iteration 2: Implement approach B → tests fail
Iteration 3: Implement approach C → tests fail
→ TRIGGER RECOVERY PROTOCOL
Recovery Steps
STOP all further implementation attempts
- No more "let me try a different approach"
- No guessing or throwing code at the problem
REVERT to last known working state
git checkout <last-passing-commit>- Or revert specific files
- Document what was attempted in
.claude/RECOVERY.md
DOCUMENT what was attempted
- All 3 approaches tried
- Test failures for each
- Why each approach failed
- What this reveals about the problem
CONSULT with user BEFORE continuing
- "I've tried 3 approaches. All fail tests. Here's what I've learned..."
- Present test failure patterns
- Request: requirements clarification, design input, or different strategy
ASK USER for direction
- Option A: Re-examine requirements (may need /dev-clarify)
- Option B: Try completely different design (may need /dev-design)
- Option C: Investigate why tests fail (may need /dev-debug)
- Option D: User provides domain knowledge
NO PASSING TESTS = NOT COMPLETE (hard rule)
Recovery Checklist
Before continuing after multiple failures:
- All 3 approaches documented with test failures
- Pattern in failures identified (same tests? different errors?)
- Current code reverted to clean state
- User consulted with specific question
- Clear direction from user before proceeding
Anti-Patterns After Failures
DON'T:
- Keep trying "just one more thing"
- Make larger and larger changes
- Skip TDD "to get it working first"
- Suppress test failures ("I'll fix them later")
- Blame the tests ("tests are wrong")
DO:
- Stop and analyze the failure pattern
- Revert to clean state
- Document what each approach revealed
- Consult user with specific findings
- Get clear direction before continuing
Example Recovery Flow
Loop 1: Implement with synchronous approach → Tests timeout
Loop 2: Implement with async/await → Tests hang
Loop 3: Implement with promises → Tests fail assertion
→ RECOVERY PROTOCOL:
1. STOP (no loop 4)
2. REVERT: git checkout HEAD -- src/feature.ts tests/
3. DOCUMENT in .claude/RECOVERY.md:
- Pattern: All async implementations cause timing issues
- Tests expect synchronous behavior
- Hypothesis: Requirements may need async, tests don't handle it
4. ASK USER:
"I've tried 3 async implementations. All cause timing issues.
Tests expect synchronous behavior.
This suggests either:
A) Feature should actually be synchronous (simpler)
B) Tests need updating for async behavior
Which direction should I take?"
When to Trigger Recovery
Trigger after 3 failures when:
- Same test keeps failing despite different approaches
- Different tests fail in pattern (suggests wrong approach)
- Tests pass locally but fail in CI
- Implementation works but breaks unrelated tests
Don't wait for max iterations - trigger early when pattern emerges.
If Max Iterations Reached
Ralph exits after max iterations. Still do NOT ask user to manually test.
Main chat should:
- Summarize what's failing (from LEARNINGS.md)
- Report which automated tests fail and why
- Ask user for direction:
- A) Start new loop with different approach
- B) Add more logging to debug
- C) User provides guidance
- D) User explicitly requests manual testing
Never default to "please test manually". Always exhaust automation first.
No Pause Between Tasks
| Thought | Reality |
|---|---|
| "Task done, let me check in with user" | NO. User wants ALL tasks done. Keep going. |
| "User might want to review" | User will review at the END. Continue. |
| "Natural pause point" | Only pause when ALL tasks complete or blocked. |
| "Let me summarize progress" | Summarize AFTER all tasks. Keep moving. |
| "User has been waiting" | User is waiting for COMPLETION, not updates. |
The promise signals task completion. After outputting promise, IMMEDIATELY start next task's loop.
Pausing between tasks is procrastination disguised as courtesy.
Phase Complete
REQUIRED SUB-SKILL: After ALL tasks complete with passing tests:
Skill(skill="workflows:dev-review")
Do NOT proceed until automated tests pass for every task.