Skip to content

๐Ÿ“‹ Full Skill Source โ€” This is the complete, unedited SKILL.md file. Nothing is hidden or summarized.

โ† Back to Skills Library

Execution โ€” Execute Plans at Scale โ€‹

Three modes, one skill. Choose based on task structure.

Mode Selection โ€‹

Have a plan with independent tasks?
โ”œโ”€โ”€ Stay in this session?
โ”‚   โ”œโ”€โ”€ YES โ†’ Mode B: Subagent-Driven
โ”‚   โ””โ”€โ”€ NO โ†’ Mode A: Batch Execution
โ””โ”€โ”€ Multiple independent failures/problems?
    โ””โ”€โ”€ YES โ†’ Mode C: Parallel Dispatch
ModeWhenStrategy
A: BatchPlan with checkpointsExecute 3 tasks โ†’ report โ†’ feedback โ†’ next batch
B: SubagentPlan with independent tasks, same sessionFresh subagent per task + 2-stage review
C: Parallel2+ independent problemsOne agent per problem domain

Mode A: Batch Execution โ€‹

Process โ€‹

  1. Load plan โ†’ review critically โ†’ raise concerns
  2. Execute batch (default: 3 tasks)
    • Mark in_progress โ†’ follow steps โ†’ verify โ†’ mark complete
  3. Report โ†’ show what was done + verification output
  4. Continue โ†’ apply feedback โ†’ next batch
  5. Complete โ†’ use cm-code-review to finish

Rules โ€‹

  • Follow plan steps exactly
  • Don't skip verifications
  • Between batches: report and wait
  • Stop when blocked, don't guess

Mode B: Subagent-Driven Development โ€‹

Process โ€‹

  1. Read plan โ†’ extract ALL tasks with full text
  2. Per task:
    • Dispatch implementer subagent with full task text
    • Answer subagent questions if any
    • Subagent implements, tests, commits, self-reviews
    • Dispatch spec reviewer โ†’ confirm matches spec
    • Dispatch code quality reviewer โ†’ confirm quality
    • If issues โ†’ implementer fixes โ†’ re-review โ†’ repeat
  3. After all tasks โ†’ final code review โ†’ cm-code-review

Prompt Template (Implementer) โ€‹

markdown
Implement [TASK_NAME]:

[Full task text from plan]

Context: [Where this fits in the project]

Rules:
- Follow TDD (cm-tdd)
- Commit when done
- Self-review before reporting
- Ask questions if unclear

Return: Summary of what you did + test results

Red Flags โ€‹

  • Never start on main/master without consent
  • Never skip reviews (spec OR quality)
  • Never dispatch parallel implementers (conflicts)
  • Never accept "close enough" on spec compliance

Mode C: Parallel Dispatch โ€‹

When โ€‹

  • 3+ test files failing with different root causes
  • Multiple subsystems broken independently
  • Each problem doesn't need context from others

Process โ€‹

  1. Group failures by independent domain
  2. Create focused agent prompt per domain:
    • Specific scope (one file/subsystem)
    • Clear goal
    • Constraints (don't change other code)
    • Expected output format
  3. Dispatch in parallel
  4. Review + integrate โ†’ verify no conflicts โ†’ run full suite

Common Mistakes โ€‹

  • โŒ Too broad: "Fix all the tests"
  • โœ… Specific: "Fix agent-tool-abort.test.ts"
  • โŒ No context: "Fix the race condition"
  • โœ… Context: Paste error messages + test names

Mode D: Autonomous RARV โ€‹

Self-driving execution. Tasks flow through Reason โ†’ Act โ†’ Reflect โ†’ Verify automatically.

When โ€‹

  • User runs /cm-start with a goal
  • cm-tasks.json exists with backlog items
  • You want continuous autonomous execution

Process (RARV Cycle) โ€‹

LOOP until backlog empty or user interrupts:
  1. REASON  โ†’ Read cm-tasks.json โ†’ pick highest-priority backlog task
                Update task status to "in_progress"
                Log: { phase: "REASON", message: "Selected: <title>" }

  2. ACT     โ†’ Execute using the task's assigned CM skill
                (cm-tdd, cm-debugging, cm-safe-deploy, etc.)
                Log: { phase: "ACT", message: "<what was done>" }

  3. REFLECT โ†’ Update cm-tasks.json with results
                Log: { phase: "REFLECT", message: "<outcome summary>" }

  4. VERIFY  โ†’ Run tests/checks (cm-quality-gate)
                If PASS โ†’ status = "done", completed_at = now()
                If FAIL โ†’ rarv_cycles++, log error, retry from REASON
                If rarv_cycles >= 3 โ†’ status = "blocked"
                Log: { phase: "VERIFY", message: "โœ… passed" or "โŒ <error>" }

  5. NEXT    โ†’ Recalculate stats, pick next task

cm-tasks.json Update Protocol โ€‹

After EVERY phase, you MUST:

  1. Read current cm-tasks.json
  2. Find the active task by id
  3. Update status, logs[], timestamps
  4. Recalculate stats object:
    stats.total = tasks.length
    stats.done = tasks.filter(t => t.status === 'done').length
    stats.in_progress = tasks.filter(t => t.status === 'in_progress').length
    stats.blocked = tasks.filter(t => t.status === 'blocked').length
    stats.backlog = tasks.filter(t => t.status === 'backlog').length
    stats.rarv_cycles_total = tasks.reduce((sum, t) => sum + (t.rarv_cycles || 0), 0)
  5. Set updated to current ISO timestamp
  6. Write back to cm-tasks.json

Rules โ€‹

  • Max 3 retries per task before marking "blocked"
  • Always log โ€” the dashboard reads logs in real-time
  • Don't batch-skip โ€” execute one task at a time through full RARV
  • Respect interrupts โ€” if user sends a message, pause and respond

Integration โ€‹

SkillWhen
cm-git-worktreesREQUIRED: isolated workspace before starting
cm-planningCreates the plan this skill executes
cm-code-reviewComplete development after all tasks
cm-tddSubagents follow TDD for each task
cm-quality-gateVERIFY phase uses this for validation
cm-ui-previewRECOMMENDED: Preview UI on Google Stitch before implementing frontend tasks

Workflows โ€‹

CommandPurpose
/cm-startCreate tasks + launch RARV + open dashboard
/cm-statusQuick terminal progress summary
/cm-dashboardOpen browser dashboard

The Bottom Line โ€‹

Choose your mode. Execute systematically. Review at every checkpoint.

Open Source AI Agent Skills Framework