Files
Claude-Code-Workflow/.codex/skills/spec-generator/phases/05-epics-stories.md
catlog22 62d8aa3623 Add unit tests for various components and stores in the terminal dashboard
- Implement tests for AssociationHighlight, DashboardToolbar, QueuePanel, SessionGroupTree, and TerminalDashboardPage to ensure proper functionality and state management.
- Create tests for cliSessionStore, issueQueueIntegrationStore, queueExecutionStore, queueSchedulerStore, sessionManagerStore, and terminalGridStore to validate state resets and workspace scoping.
- Mock necessary dependencies and state management hooks to isolate tests and ensure accurate behavior.
2026-03-08 21:38:20 +08:00

8.4 KiB

Phase 5: Epics & Stories

Execution Mode: Agent Delegated This phase is executed by a doc-generator agent. The orchestrator (SKILL.md) passes session context via the Task tool. The agent reads this file for instructions, executes all steps, writes output files, and returns a JSON summary.

Decompose the specification into executable Epics and Stories with dependency mapping.

Objective

  • Group requirements into 3-7 logical Epics
  • Tag MVP subset of Epics
  • Generate 2-5 Stories per Epic in standard user story format
  • Map cross-Epic dependencies (Mermaid diagram)
  • Generate epics.md using template

Input

  • Dependency: {workDir}/requirements/_index.md, {workDir}/architecture/_index.md (and individual files)
  • Reference: {workDir}/product-brief.md
  • Config: {workDir}/spec-config.json
  • Template: templates/epics-template.md (directory structure: _index.md + EPIC-*.md)

Execution Steps

Step 1: Load Phase 2-4 Context

const specConfig = JSON.parse(Read(`${workDir}/spec-config.json`));
const productBrief = Read(`${workDir}/product-brief.md`);
const requirements = Read(`${workDir}/requirements.md`);
const architecture = Read(`${workDir}/architecture.md`);

let glossary = null;
try {
  glossary = JSON.parse(Read(`${workDir}/glossary.json`));
} catch (e) { /* proceed without */ }

Step 2: Epic Decomposition via Gemini CLI

Bash({
  command: `ccw cli -p "PURPOSE: Decompose requirements into executable Epics and Stories for implementation planning.
Success: 3-7 Epics with prioritized Stories, dependency map, and MVP subset clearly defined.

PRODUCT BRIEF (summary):
${productBrief.slice(0, 2000)}

REQUIREMENTS:
${requirements.slice(0, 5000)}

ARCHITECTURE (summary):
${architecture.slice(0, 3000)}

TASK:
- Group requirements into 3-7 logical Epics:
  - Each Epic: EPIC-NNN ID, title, description, priority (Must/Should/Could)
  - Group by functional domain or user journey stage
  - Tag MVP Epics (minimum set for initial release)
  
- For each Epic, generate 2-5 Stories:
  - Each Story: STORY-{EPIC}-NNN ID, title
  - User story format: As a [persona], I want [action] so that [benefit]
  - 2-4 acceptance criteria per story (testable)
  - Relative size estimate: S/M/L/XL
  - Trace to source requirement(s): REQ-NNN
  
- Create dependency map:
  - Cross-Epic dependencies (which Epics block others)
  - Mermaid graph LR format
  - Recommended execution order with rationale
  
- Define MVP:
  - Which Epics are in MVP
  - MVP definition of done (3-5 criteria)
  - What is explicitly deferred post-MVP

MODE: analysis
EXPECTED: Structured output with: Epic list (ID, title, priority, MVP flag), Stories per Epic (ID, user story, AC, size, trace), dependency Mermaid diagram, execution order, MVP definition
CONSTRAINTS:
- Every Must-have requirement must appear in at least one Story
- Stories must be small enough to implement independently (no XL stories in MVP)
- Dependencies should be minimized across Epics
\${glossary ? \`- Maintain terminology consistency with glossary: \${glossary.terms.map(t => t.term).join(', ')}\` : ''}
" --tool gemini --mode analysis`,
  run_in_background: true
});

// Wait for CLI result

Step 2.5: Codex Epics Review

After receiving Gemini decomposition results, validate epic/story quality via Codex CLI:

Bash({
  command: `ccw cli -p "PURPOSE: Critical review of epic/story decomposition - validate coverage, sizing, and dependency structure.
Success: Actionable feedback on epic quality with specific issues identified.

GENERATED EPICS AND STORIES:
${geminiEpicsOutput.slice(0, 5000)}

REQUIREMENTS (Must-Have):
${mustHaveRequirements.slice(0, 2000)}

TASK:
- Verify Must-Have requirement coverage: every Must requirement appears in at least one Story
- Check MVP story sizing: no XL stories in MVP epics (too large to implement independently)
- Validate dependency graph: no circular dependencies between Epics
- Assess acceptance criteria: every Story AC is specific and testable
- Verify traceability: Stories trace back to specific REQ-NNN IDs
- Check Epic granularity: 3-7 epics (not too few/many), 2-5 stories each
- Rate overall decomposition quality: 1-5 with justification

MODE: analysis
EXPECTED: Epic review with: coverage gaps, oversized stories, dependency issues, traceability gaps, quality rating
CONSTRAINTS: Focus on issues that would block execution planning. Be specific about which Story/Epic has problems.
" --tool codex --mode analysis`,
  run_in_background: true
});

// Wait for Codex review result
// Integrate feedback into epics before writing files:
// - Add missing Stories for uncovered Must requirements
// - Split XL stories in MVP epics into smaller units
// - Fix dependency cycles identified by Codex
// - Improve vague acceptance criteria

Step 3: Interactive Validation (Optional)

if (!autoMode) {
  // Present Epic overview table and dependency diagram
  AskUserQuestion({
    questions: [
      {
        question: "Review the Epic breakdown. Any adjustments needed?",
        header: "Epics",
        multiSelect: false,
        options: [
          { label: "Looks good", description: "Epic structure is appropriate" },
          { label: "Merge epics", description: "Some epics should be combined" },
          { label: "Split epic", description: "An epic is too large, needs splitting" },
          { label: "Adjust MVP", description: "Change which epics are in MVP" }
        ]
      }
    ]
  });
  // Apply user adjustments
}

Step 4: Generate epics/ directory

const template = Read('templates/epics-template.md');

// Create epics directory
Bash(`mkdir -p "${workDir}/epics"`);

const status = autoMode ? 'complete' : 'draft';
const timestamp = new Date().toISOString();

// Parse CLI output into structured Epics
const epicsList = parseEpics(cliOutput);  // [{id, slug, title, priority, mvp, size, stories[], reqs[], adrs[], deps[]}]

// Step 4a: Write individual EPIC-*.md files (one per Epic, stories included)
epicsList.forEach(epic => {
  // Use EPIC-NNN-{slug}.md template from templates/epics-template.md
  // Fill: id, title, priority, mvp, size, description, requirements links,
  //       architecture links, dependency links, stories with user stories + AC
  Write(`${workDir}/epics/EPIC-${epic.id}-${epic.slug}.md`, epicContent);
});

// Step 4b: Write _index.md (overview + dependency map + MVP scope + traceability)
// Use _index.md template from templates/epics-template.md
// Fill: epic overview table (with links), dependency Mermaid diagram,
//       execution order, MVP scope, traceability matrix, estimation summary
Write(`${workDir}/epics/_index.md`, indexContent);

// Update spec-config.json
specConfig.phasesCompleted.push({
  phase: 5,
  name: "epics-stories",
  output_dir: "epics/",
  output_index: "epics/_index.md",
  file_count: epicsList.length + 1,
  completed_at: timestamp
});
Write(`${workDir}/spec-config.json`, JSON.stringify(specConfig, null, 2));

Output

  • Directory: epics/
    • _index.md — Overview table, dependency map, MVP scope, traceability matrix, links
    • EPIC-NNN-{slug}.md — Individual Epic with Stories (per Epic)
  • Format: Markdown with YAML frontmatter, cross-linked to requirements and architecture via relative paths

Quality Checklist

  • 3-7 Epic files with EPIC-NNN IDs
  • >= 1 Epic tagged as MVP in frontmatter
  • 2-5 Stories per Epic file
  • Stories use "As a...I want...So that..." format
  • _index.md has cross-Epic dependency map (Mermaid)
  • _index.md links to all individual Epic files
  • Relative sizing (S/M/L/XL) per Story
  • Epic files link to requirement files and ADR files
  • All files have valid YAML frontmatter

Next Phase

Proceed to Phase 6: Readiness Check to validate the complete specification package.


Agent Return Summary

When executed as a delegated agent, return the following JSON summary to the orchestrator:

{
  "phase": 5,
  "status": "complete",
  "files_created": ["epics/_index.md", "epics/EPIC-001-*.md", "..."],
  "file_count": 0,
  "codex_review_integrated": true,
  "mvp_epic_count": 0,
  "total_story_count": 0,
  "quality_notes": ["list of quality concerns or Codex feedback items addressed"],
  "key_decisions": ["MVP scope decisions", "dependency resolution choices"]
}

The orchestrator will:

  1. Validate that epics/ directory exists with _index.md and EPIC files
  2. Read spec-config.json to confirm phasesCompleted was updated
  3. Store the summary for downstream phase context