Files
Claude-Code-Workflow/.codex/skills/team-ultra-analyze/SKILL.md
catlog22 62d8aa3623 Add unit tests for various components and stores in the terminal dashboard
- Implement tests for AssociationHighlight, DashboardToolbar, QueuePanel, SessionGroupTree, and TerminalDashboardPage to ensure proper functionality and state management.
- Create tests for cliSessionStore, issueQueueIntegrationStore, queueExecutionStore, queueSchedulerStore, sessionManagerStore, and terminalGridStore to validate state resets and workspace scoping.
- Mock necessary dependencies and state management hooks to isolate tests and ensure accurate behavior.
2026-03-08 21:38:20 +08:00

32 KiB

name, description, argument-hint, allowed-tools
name description argument-hint allowed-tools
team-ultra-analyze Deep collaborative analysis pipeline. Multi-perspective exploration, deep analysis, user-driven discussion loops, and cross-perspective synthesis. Supports Quick, Standard, and Deep pipeline modes. [-y|--yes] [-c|--concurrency N] [--continue] [--mode quick|standard|deep] "analysis topic" spawn_agents_on_csv, spawn_agent, wait, send_input, close_agent, Read, Write, Edit, Bash, Glob, Grep, AskUserQuestion

Auto Mode

When --yes or -y: Auto-confirm task decomposition, skip interactive validation, use defaults.

Team Ultra Analyze

Usage

$team-ultra-analyze "Analyze authentication module architecture and security"
$team-ultra-analyze -c 4 --mode deep "Deep analysis of payment processing pipeline"
$team-ultra-analyze -y --mode quick "Quick overview of API endpoint structure"
$team-ultra-analyze --continue "uan-auth-analysis-20260308"

Flags:

  • -y, --yes: Skip all confirmations (auto mode)
  • -c, --concurrency N: Max concurrent agents within each wave (default: 3)
  • --mode: Pipeline mode override (quick|standard|deep)
  • --continue: Resume existing session

Output Directory: .workflow/.csv-wave/{session-id}/ Core Output: tasks.csv (master state) + results.csv (final) + discoveries.ndjson (shared exploration) + context.md (human-readable report)


Overview

Deep collaborative analysis with multi-perspective exploration, deep analysis, user-driven discussion loops, and cross-perspective synthesis. Each perspective gets its own explorer and analyst, working in parallel. Discussion rounds allow the user to steer analysis depth and direction.

Execution Model: Hybrid — CSV wave pipeline (primary) + individual agent spawn (secondary for discussion feedback loop)

┌─────────────────────────────────────────────────────────────────────────┐
│                    TEAM ULTRA ANALYZE WORKFLOW                           │
├─────────────────────────────────────────────────────────────────────────┤
│                                                                          │
│  Phase 0: Pre-Wave Interactive                                           │
│     ├─ Topic parsing + dimension detection                               │
│     ├─ Pipeline mode selection (quick/standard/deep)                     │
│     ├─ Perspective assignment                                            │
│     └─ Output: refined requirements for decomposition                    │
│                                                                          │
│  Phase 1: Requirement → CSV + Classification                             │
│     ├─ Parse topic into exploration + analysis + discussion + synthesis  │
│     ├─ Assign roles: explorer, analyst, discussant, synthesizer          │
│     ├─ Classify tasks: csv-wave | interactive (exec_mode)                │
│     ├─ Compute dependency waves (topological sort → depth grouping)      │
│     ├─ Generate tasks.csv with wave + exec_mode columns                  │
│     └─ User validates task breakdown (skip if -y)                        │
│                                                                          │
│  Phase 2: Wave Execution Engine (Extended)                               │
│     ├─ For each wave (1..N):                                             │
│     │   ├─ Build wave CSV (filter csv-wave tasks for this wave)          │
│     │   ├─ Inject previous findings into prev_context column             │
│     │   ├─ spawn_agents_on_csv(wave CSV)                                 │
│     │   ├─ Execute post-wave interactive tasks (if any)                  │
│     │   ├─ Merge all results into master tasks.csv                       │
│     │   └─ Check: any failed? → skip dependents                         │
│     └─ discoveries.ndjson shared across all modes (append-only)          │
│                                                                          │
│  Phase 3: Post-Wave Interactive (Discussion Loop)                        │
│     ├─ After discussant completes: user feedback gate                    │
│     ├─ User chooses: continue deeper | adjust direction | done           │
│     ├─ Creates dynamic tasks (DISCUSS-N, ANALYZE-fix-N) as needed        │
│     └─ Max discussion rounds: quick=0, standard=1, deep=5               │
│                                                                          │
│  Phase 4: Results Aggregation                                            │
│     ├─ Export final results.csv                                          │
│     ├─ Generate context.md with all findings                             │
│     ├─ Display summary: completed/failed/skipped per wave                │
│     └─ Offer: view results | export | archive                           │
│                                                                          │
└─────────────────────────────────────────────────────────────────────────┘

Task Classification Rules

Each task is classified by exec_mode:

exec_mode Mechanism Criteria
csv-wave spawn_agents_on_csv One-shot, structured I/O, no multi-round interaction
interactive spawn_agent/wait/send_input/close_agent Multi-round, user feedback, direction control

Classification Decision:

Task Property Classification
Codebase exploration (single perspective) csv-wave
Parallel exploration (multiple perspectives) csv-wave (parallel in same wave)
Deep analysis (single perspective) csv-wave
Parallel analysis (multiple perspectives) csv-wave (parallel in same wave)
Direction-fix analysis (adjusted focus) csv-wave
Discussion processing (aggregate results) csv-wave
Final synthesis (cross-perspective integration) csv-wave
Discussion feedback gate (user interaction) interactive
Topic clarification (Phase 0) interactive

CSV Schema

tasks.csv (Master State)

id,title,description,role,perspective,dimensions,discussion_round,discussion_type,deps,context_from,exec_mode,wave,status,findings,error
"EXPLORE-001","Explore from technical perspective","Search codebase from technical perspective. Collect files, patterns, findings.","explorer","technical","architecture;implementation","0","","","","csv-wave","1","pending","",""
"ANALYZE-001","Deep analysis from technical perspective","Analyze exploration results from technical perspective. Generate insights with confidence levels.","analyst","technical","architecture;implementation","0","","EXPLORE-001","EXPLORE-001","csv-wave","2","pending","",""
"DISCUSS-001","Initial discussion round","Aggregate all analysis results. Identify convergent themes, conflicts, top discussion points.","discussant","","","1","initial","ANALYZE-001;ANALYZE-002","ANALYZE-001;ANALYZE-002","csv-wave","3","pending","",""

Columns:

Column Phase Description
id Input Unique task identifier (string)
title Input Short task title
description Input Detailed task description
role Input Worker role: explorer, analyst, discussant, synthesizer
perspective Input Analysis perspective: technical, architectural, business, domain_expert
dimensions Input Analysis dimensions (semicolon-separated): architecture, implementation, performance, security, concept, comparison, decision
discussion_round Input Discussion round number (0 = N/A, 1+ = round number)
discussion_type Input Discussion type: initial, deepen, direction-adjusted, specific-questions
deps Input Semicolon-separated dependency task IDs
context_from Input Semicolon-separated task IDs whose findings this task needs
exec_mode Input csv-wave or interactive
wave Computed Wave number (computed by topological sort, 1-based)
status Output pendingcompleted / failed / skipped
findings Output Key discoveries or implementation notes (max 500 chars)
error Output Error message if failed (empty if success)

Per-Wave CSV (Temporary)

Each wave generates a temporary wave-{N}.csv with extra prev_context column (csv-wave tasks only).


Agent Registry (Interactive Agents)

Agent Role File Pattern Responsibility Position
discussion-feedback agents/discussion-feedback.md 2.3 (wait-respond) Collect user feedback after discussion round, create dynamic tasks post-wave (after discussant wave)
topic-analyzer agents/topic-analyzer.md 2.3 (wait-respond) Parse topic, detect dimensions, select pipeline mode and perspectives standalone (Phase 0)

COMPACT PROTECTION: Agent files are execution documents. When context compression occurs, you MUST immediately Read the corresponding agent.md to reload.


Output Artifacts

File Purpose Lifecycle
tasks.csv Master state — all tasks with status/findings Updated after each wave
wave-{N}.csv Per-wave input (temporary, csv-wave tasks only) Created before wave, deleted after
results.csv Final export of all task results Created in Phase 4
discoveries.ndjson Shared exploration board (all agents, both modes) Append-only, carries across waves
context.md Human-readable execution report Created in Phase 4
interactive/{id}-result.json Results from interactive tasks Created per interactive task

Session Structure

.workflow/.csv-wave/{session-id}/
├── tasks.csv                  # Master state (all tasks, both modes)
├── results.csv                # Final results export
├── discoveries.ndjson         # Shared discovery board (all agents)
├── context.md                 # Human-readable report
├── wave-{N}.csv               # Temporary per-wave input (csv-wave only)
└── interactive/               # Interactive task artifacts
    └── {id}-result.json       # Per-task results

Implementation

Session Initialization

const getUtc8ISOString = () => new Date(Date.now() + 8 * 60 * 60 * 1000).toISOString()

// Parse flags
const AUTO_YES = $ARGUMENTS.includes('--yes') || $ARGUMENTS.includes('-y')
const continueMode = $ARGUMENTS.includes('--continue')
const concurrencyMatch = $ARGUMENTS.match(/(?:--concurrency|-c)\s+(\d+)/)
const maxConcurrency = concurrencyMatch ? parseInt(concurrencyMatch[1]) : 3
const modeMatch = $ARGUMENTS.match(/--mode\s+(quick|standard|deep)/)
const explicitMode = modeMatch ? modeMatch[1] : null

// Clean requirement text (remove flags)
const topic = $ARGUMENTS
  .replace(/--yes|-y|--continue|--concurrency\s+\d+|-c\s+\d+|--mode\s+\w+/g, '')
  .trim()

const slug = topic.toLowerCase()
  .replace(/[^a-z0-9\u4e00-\u9fa5]+/g, '-')
  .substring(0, 40)
const dateStr = getUtc8ISOString().substring(0, 10).replace(/-/g, '')
let sessionId = `uan-${slug}-${dateStr}`
let sessionFolder = `.workflow/.csv-wave/${sessionId}`

// Continue mode: find existing session
if (continueMode) {
  const existing = Bash(`ls -t .workflow/.csv-wave/uan-* 2>/dev/null | head -1`).trim()
  if (existing) {
    sessionId = existing.split('/').pop()
    sessionFolder = existing
  }
}

Bash(`mkdir -p ${sessionFolder}/interactive`)

Phase 0: Pre-Wave Interactive

Objective: Parse topic, detect analysis dimensions, select pipeline mode, and assign perspectives.

Execution:

const analyzer = spawn_agent({
  message: `
## TASK ASSIGNMENT

### MANDATORY FIRST STEPS (Agent Execute)
1. **Read role definition**: .codex/skills/team-ultra-analyze/agents/topic-analyzer.md (MUST read first)
2. Read: .workflow/project-tech.json (if exists)

---

Goal: Analyze topic and recommend pipeline configuration
Topic: ${topic}
Explicit Mode: ${explicitMode || 'auto-detect'}

### Task
1. Detect analysis dimensions from topic keywords:
   - architecture, implementation, performance, security, concept, comparison, decision
2. Select perspectives based on dimensions:
   - technical, architectural, business, domain_expert
3. Determine pipeline mode (if not explicitly set):
   - Complexity 1-3 → quick, 4-6 → standard, 7+ → deep
4. Return structured configuration
`
})

const analyzerResult = wait({ ids: [analyzer], timeout_ms: 120000 })

if (analyzerResult.timed_out) {
  send_input({ id: analyzer, message: "Please finalize and output current findings." })
  wait({ ids: [analyzer], timeout_ms: 60000 })
}

close_agent({ id: analyzer })

// Parse result: pipeline_mode, perspectives[], dimensions[], depth
Write(`${sessionFolder}/interactive/topic-analyzer-result.json`, JSON.stringify({
  task_id: "topic-analysis",
  status: "completed",
  pipeline_mode: parsedMode,
  perspectives: parsedPerspectives,
  dimensions: parsedDimensions,
  depth: parsedDepth,
  timestamp: getUtc8ISOString()
}))

If not AUTO_YES, present user with configuration for confirmation:

if (!AUTO_YES) {
  const answer = AskUserQuestion({
    questions: [{
      question: `Topic: "${topic}"\nPipeline: ${pipeline_mode}\nPerspectives: ${perspectives.join(', ')}\nDimensions: ${dimensions.join(', ')}\n\nApprove?`,
      header: "Analysis Configuration",
      multiSelect: false,
      options: [
        { label: "Approve", description: `Use ${pipeline_mode} mode with ${perspectives.length} perspectives` },
        { label: "Quick", description: "1 explorer → 1 analyst → synthesizer (fast)" },
        { label: "Standard", description: "N explorers → N analysts → discussion → synthesizer" },
        { label: "Deep", description: "N explorers → N analysts → discussion loop (up to 5 rounds) → synthesizer" }
      ]
    }]
  })
}

Success Criteria:

  • Refined requirements available for Phase 1 decomposition
  • Interactive agents closed, results stored

Phase 1: Requirement → CSV + Classification

Objective: Build tasks.csv from selected pipeline mode and perspectives.

Decomposition Rules:

Pipeline Tasks Wave Structure
quick EXPLORE-001 → ANALYZE-001 → SYNTH-001 3 waves, serial, depth=1
standard EXPLORE-001..N → ANALYZE-001..N → DISCUSS-001 → SYNTH-001 4 wave groups, parallel explore+analyze
deep EXPLORE-001..N → ANALYZE-001..N → DISCUSS-001 (→ dynamic tasks) → SYNTH-001 3+ waves, SYNTH created after discussion loop

Where N = number of selected perspectives.

Classification Rules:

All work tasks (exploration, analysis, discussion processing, synthesis) are csv-wave. The discussion feedback gate (user interaction after discussant completes) is interactive.

Pipeline Task Definitions:

Quick Pipeline (3 csv-wave tasks)

Task ID Role Wave Deps Perspective Description
EXPLORE-001 explorer 1 (none) general Explore codebase structure for analysis topic
ANALYZE-001 analyst 2 EXPLORE-001 technical Deep analysis from technical perspective
SYNTH-001 synthesizer 3 ANALYZE-001 (all) Integrate analysis into final conclusions

Standard Pipeline (2N+2 tasks, parallel windows)

Task ID Role Wave Deps Perspective Description
EXPLORE-001..N explorer 1 (none) per-perspective Parallel codebase exploration, one per perspective
ANALYZE-001..N analyst 2 EXPLORE-N per-perspective Parallel deep analysis, one per perspective
DISCUSS-001 discussant 3 all ANALYZE-* (all) Aggregate analyses, identify themes and conflicts
FEEDBACK-001 (interactive) 4 DISCUSS-001 - User feedback: done → create SYNTH, continue → more discussion
SYNTH-001 synthesizer 5 FEEDBACK-001 (all) Cross-perspective integration and conclusions

Deep Pipeline (2N+1 initial tasks + dynamic)

Same as Standard, but SYNTH-001 is omitted initially. Created dynamically after the discussion loop (up to 5 rounds) completes. Additional dynamic tasks:

  • DISCUSS-N — subsequent discussion round
  • ANALYZE-fix-N — supplementary analysis with adjusted focus
  • SYNTH-001 — created after final discussion round

Wave Computation: Kahn's BFS topological sort with depth tracking (csv-wave tasks only).

User Validation: Display task breakdown with wave + exec_mode assignment (skip if AUTO_YES).

Success Criteria:

  • tasks.csv created with valid schema, wave, and exec_mode assignments
  • No circular dependencies
  • User approved (or AUTO_YES)

Phase 2: Wave Execution Engine (Extended)

Objective: Execute tasks wave-by-wave with hybrid mechanism support and cross-wave context propagation.

const failedIds = new Set()
const skippedIds = new Set()
let discussionRound = 0
const MAX_DISCUSSION_ROUNDS = pipeline_mode === 'deep' ? 5 : pipeline_mode === 'standard' ? 1 : 0

for (let wave = 1; wave <= maxWave; wave++) {
  console.log(`\n## Wave ${wave}/${maxWave}\n`)

  // 1. Read current master CSV
  const masterCsv = parseCsv(Read(`${sessionFolder}/tasks.csv`))

  // 2. Separate csv-wave and interactive tasks for this wave
  const waveTasks = masterCsv.filter(row => parseInt(row.wave) === wave)
  const csvTasks = waveTasks.filter(t => t.exec_mode === 'csv-wave')
  const interactiveTasks = waveTasks.filter(t => t.exec_mode === 'interactive')

  // 3. Skip tasks whose deps failed
  const executableCsvTasks = []
  for (const task of csvTasks) {
    const deps = task.deps.split(';').filter(Boolean)
    if (deps.some(d => failedIds.has(d) || skippedIds.has(d))) {
      skippedIds.add(task.id)
      updateMasterCsvRow(sessionFolder, task.id, {
        status: 'skipped', error: 'Dependency failed or skipped'
      })
      continue
    }
    executableCsvTasks.push(task)
  }

  // 4. Build prev_context for each csv-wave task
  for (const task of executableCsvTasks) {
    const contextIds = task.context_from.split(';').filter(Boolean)
    const prevFindings = contextIds
      .map(id => {
        const prevRow = masterCsv.find(r => r.id === id)
        if (prevRow && prevRow.status === 'completed' && prevRow.findings) {
          return `[Task ${id}: ${prevRow.title}] ${prevRow.findings}`
        }
        return null
      })
      .filter(Boolean)
      .join('\n')
    task.prev_context = prevFindings || 'No previous context available'
  }

  // 5. Write wave CSV and execute csv-wave tasks
  if (executableCsvTasks.length > 0) {
    const waveHeader = 'id,title,description,role,perspective,dimensions,discussion_round,discussion_type,deps,context_from,exec_mode,wave,prev_context'
    const waveRows = executableCsvTasks.map(t =>
      [t.id, t.title, t.description, t.role, t.perspective, t.dimensions,
       t.discussion_round, t.discussion_type, t.deps, t.context_from, t.exec_mode, t.wave, t.prev_context]
        .map(cell => `"${String(cell).replace(/"/g, '""')}"`)
        .join(',')
    )
    Write(`${sessionFolder}/wave-${wave}.csv`, [waveHeader, ...waveRows].join('\n'))

    const waveResult = spawn_agents_on_csv({
      csv_path: `${sessionFolder}/wave-${wave}.csv`,
      id_column: "id",
      instruction: buildAnalysisInstruction(sessionFolder, wave),
      max_concurrency: maxConcurrency,
      max_runtime_seconds: 600,
      output_csv_path: `${sessionFolder}/wave-${wave}-results.csv`,
      output_schema: {
        type: "object",
        properties: {
          id: { type: "string" },
          status: { type: "string", enum: ["completed", "failed"] },
          findings: { type: "string" },
          error: { type: "string" }
        },
        required: ["id", "status", "findings"]
      }
    })

    // Merge results into master CSV
    const waveResults = parseCsv(Read(`${sessionFolder}/wave-${wave}-results.csv`))
    for (const result of waveResults) {
      updateMasterCsvRow(sessionFolder, result.id, {
        status: result.status,
        findings: result.findings || '',
        error: result.error || ''
      })
      if (result.status === 'failed') failedIds.add(result.id)
    }

    Bash(`rm -f "${sessionFolder}/wave-${wave}.csv"`)
  }

  // 6. Execute post-wave interactive tasks (Discussion Feedback)
  for (const task of interactiveTasks) {
    if (task.status !== 'pending') continue
    const deps = task.deps.split(';').filter(Boolean)
    if (deps.some(d => failedIds.has(d) || skippedIds.has(d))) {
      skippedIds.add(task.id)
      continue
    }

    discussionRound++

    // Discussion Feedback Gate
    if (pipeline_mode === 'quick' || discussionRound > MAX_DISCUSSION_ROUNDS) {
      // No discussion or max rounds reached — proceed to synthesis
      if (!masterCsv.find(t => t.id === 'SYNTH-001')) {
        // Create SYNTH-001 dynamically
        const lastDiscuss = masterCsv.filter(t => t.id.startsWith('DISCUSS'))
          .sort((a, b) => b.id.localeCompare(a.id))[0]
        addTaskToMasterCsv(sessionFolder, {
          id: 'SYNTH-001', title: 'Final synthesis',
          description: 'Integrate all analysis into final conclusions',
          role: 'synthesizer', perspective: '', dimensions: '',
          discussion_round: '0', discussion_type: '',
          deps: lastDiscuss ? lastDiscuss.id : '', context_from: 'all',
          exec_mode: 'csv-wave', wave: String(wave + 1),
          status: 'pending', findings: '', error: ''
        })
        maxWave = wave + 1
      }
      updateMasterCsvRow(sessionFolder, task.id, {
        status: 'completed',
        findings: `Discussion round ${discussionRound}: proceeding to synthesis`
      })
      continue
    }

    // Spawn discussion feedback agent
    const feedbackAgent = spawn_agent({
      message: `
## TASK ASSIGNMENT

### MANDATORY FIRST STEPS (Agent Execute)
1. **Read role definition**: .codex/skills/team-ultra-analyze/agents/discussion-feedback.md (MUST read first)
2. Read: ${sessionFolder}/discoveries.ndjson (shared discoveries)

---

Goal: Collect user feedback on discussion round ${discussionRound}
Session: ${sessionFolder}
Discussion Round: ${discussionRound}/${MAX_DISCUSSION_ROUNDS}
Pipeline Mode: ${pipeline_mode}

### Context
The discussant has completed round ${discussionRound}. Present the user with discussion results and collect feedback on next direction.
`
    })

    const feedbackResult = wait({ ids: [feedbackAgent], timeout_ms: 300000 })
    if (feedbackResult.timed_out) {
      send_input({ id: feedbackAgent, message: "Please finalize: user did not respond, default to 'Done'." })
      wait({ ids: [feedbackAgent], timeout_ms: 60000 })
    }
    close_agent({ id: feedbackAgent })

    // Parse feedback decision: "continue_deeper" | "adjust_direction" | "done"
    Write(`${sessionFolder}/interactive/${task.id}-result.json`, JSON.stringify({
      task_id: task.id, status: "completed",
      discussion_round: discussionRound,
      feedback: feedbackDecision,
      timestamp: getUtc8ISOString()
    }))

    // Handle feedback
    if (feedbackDecision === 'done') {
      // Create SYNTH-001 blocked by last DISCUSS task
      addTaskToMasterCsv(sessionFolder, {
        id: 'SYNTH-001', deps: task.id.replace('FEEDBACK', 'DISCUSS'),
        role: 'synthesizer', exec_mode: 'csv-wave', wave: String(wave + 1)
      })
      maxWave = wave + 1
    } else if (feedbackDecision === 'adjust_direction') {
      // Create ANALYZE-fix-N and DISCUSS-N+1
      const fixId = `ANALYZE-fix-${discussionRound}`
      const nextDiscussId = `DISCUSS-${String(discussionRound + 1).padStart(3, '0')}`
      addTaskToMasterCsv(sessionFolder, {
        id: fixId, role: 'analyst', exec_mode: 'csv-wave', wave: String(wave + 1)
      })
      addTaskToMasterCsv(sessionFolder, {
        id: nextDiscussId, role: 'discussant', deps: fixId,
        exec_mode: 'csv-wave', wave: String(wave + 2)
      })
      addTaskToMasterCsv(sessionFolder, {
        id: `FEEDBACK-${String(discussionRound + 1).padStart(3, '0')}`,
        exec_mode: 'interactive', deps: nextDiscussId, wave: String(wave + 3)
      })
      maxWave = wave + 3
    } else {
      // continue_deeper: Create DISCUSS-N+1
      const nextDiscussId = `DISCUSS-${String(discussionRound + 1).padStart(3, '0')}`
      addTaskToMasterCsv(sessionFolder, {
        id: nextDiscussId, role: 'discussant', exec_mode: 'csv-wave', wave: String(wave + 1)
      })
      addTaskToMasterCsv(sessionFolder, {
        id: `FEEDBACK-${String(discussionRound + 1).padStart(3, '0')}`,
        exec_mode: 'interactive', deps: nextDiscussId, wave: String(wave + 2)
      })
      maxWave = wave + 2
    }

    updateMasterCsvRow(sessionFolder, task.id, {
      status: 'completed',
      findings: `Discussion feedback: ${feedbackDecision}, round ${discussionRound}`
    })
  }
}

Success Criteria:

  • All waves executed in order
  • Both csv-wave and interactive tasks handled per wave
  • Each wave's results merged into master CSV before next wave starts
  • Dependent tasks skipped when predecessor failed
  • discoveries.ndjson accumulated across all waves and mechanisms
  • Discussion loop controlled with proper round tracking
  • Dynamic tasks created correctly based on user feedback

Phase 3: Post-Wave Interactive

Objective: Handle discussion loop completion and ensure synthesis is triggered.

After all discussion rounds are exhausted or user chooses "done":

  1. Ensure SYNTH-001 exists in master CSV
  2. Ensure SYNTH-001 is unblocked (blocked by last completed discussion task)
  3. Execute remaining waves (synthesis)

Success Criteria:

  • Post-wave interactive processing complete
  • Interactive agents closed, results stored

Phase 4: Results Aggregation

Objective: Generate final results and human-readable report.

const masterCsv = Read(`${sessionFolder}/tasks.csv`)
Write(`${sessionFolder}/results.csv`, masterCsv)

const tasks = parseCsv(masterCsv)
const completed = tasks.filter(t => t.status === 'completed')
const failed = tasks.filter(t => t.status === 'failed')
const skipped = tasks.filter(t => t.status === 'skipped')

const contextContent = `# Ultra Analyze Report

**Session**: ${sessionId}
**Topic**: ${topic}
**Pipeline**: ${pipeline_mode}
**Perspectives**: ${perspectives.join(', ')}
**Discussion Rounds**: ${discussionRound}
**Completed**: ${getUtc8ISOString()}

---

## Summary

| Metric | Count |
|--------|-------|
| Total Tasks | ${tasks.length} |
| Completed | ${completed.length} |
| Failed | ${failed.length} |
| Skipped | ${skipped.length} |
| Discussion Rounds | ${discussionRound} |

---

## Wave Execution

${waveDetails}

---

## Analysis Artifacts

- Explorations: discoveries with type "exploration" in discoveries.ndjson
- Analyses: discoveries with type "analysis" in discoveries.ndjson
- Discussion: discoveries with type "discussion" in discoveries.ndjson
- Conclusions: discoveries with type "conclusion" in discoveries.ndjson

---

## Conclusions

${synthesisFindings}
`

Write(`${sessionFolder}/context.md`, contextContent)

If not AUTO_YES, offer completion options:

if (!AUTO_YES) {
  const answer = AskUserQuestion({
    questions: [{
      question: "Ultra-Analyze pipeline complete. What would you like to do?",
      header: "Completion",
      multiSelect: false,
      options: [
        { label: "Archive & Clean (Recommended)", description: "Archive session" },
        { label: "Keep Active", description: "Keep session for follow-up" },
        { label: "Export Results", description: "Export deliverables to specified location" }
      ]
    }]
  })
}

Success Criteria:

  • results.csv exported (all tasks, both modes)
  • context.md generated
  • All interactive agents closed
  • Summary displayed to user

Shared Discovery Board Protocol

All agents across all waves share discoveries.ndjson. This enables cross-role knowledge sharing.

Discovery Types:

Type Dedup Key Data Schema Description
exploration data.perspective+data.file {perspective, file, relevance, summary, patterns[]} Explored file/module
analysis data.perspective+data.insight {perspective, insight, confidence, evidence, file_ref} Analysis insight
pattern data.name {name, file, description, type} Code/architecture pattern
discussion_point data.topic {topic, perspectives[], convergence, open_questions[]} Discussion point
recommendation data.action {action, rationale, priority, confidence} Recommendation
conclusion data.point {point, evidence, confidence, perspectives_supporting[]} Final conclusion

Format: NDJSON, each line is self-contained JSON:

{"ts":"2026-03-08T10:00:00+08:00","worker":"EXPLORE-001","type":"exploration","data":{"perspective":"technical","file":"src/auth/index.ts","relevance":"high","summary":"Auth module entry point with OAuth and JWT exports","patterns":["module-pattern","strategy-pattern"]}}
{"ts":"2026-03-08T10:05:00+08:00","worker":"ANALYZE-001","type":"analysis","data":{"perspective":"technical","insight":"Auth module uses strategy pattern for provider switching","confidence":"high","evidence":"src/auth/strategies/*.ts","file_ref":"src/auth/index.ts:15"}}
{"ts":"2026-03-08T10:10:00+08:00","worker":"DISCUSS-001","type":"discussion_point","data":{"topic":"Authentication scalability","perspectives":["technical","architectural"],"convergence":"Both perspectives agree on stateless JWT approach","open_questions":["Token refresh strategy for long sessions"]}}

Protocol Rules:

  1. Read board before own exploration → skip covered areas
  2. Write discoveries immediately via echo >> → don't batch
  3. Deduplicate — check existing entries by type + dedup key
  4. Append-only — never modify or delete existing lines

Error Handling

Error Resolution
Circular dependency Detect in wave computation, abort with error message
CSV agent timeout Mark as failed in results, continue with wave
CSV agent failed Mark as failed, skip dependent tasks in later waves
Interactive agent timeout Urge convergence via send_input, then close if still timed out
Interactive agent failed Mark as failed, skip dependents
All agents in wave failed Log error, offer retry or abort
CSV parse error Validate CSV format before execution, show line number
discoveries.ndjson corrupt Ignore malformed lines, continue with valid entries
Discussion loop exceeds 5 rounds Force synthesis, offer continuation
Explorer finds nothing Continue with limited context, note limitation
CLI tool unavailable Fallback chain: gemini → codex → direct analysis
User timeout in discussion Save state, default to "done", proceed to synthesis
Continue mode: no session found List available sessions, prompt user to select

Core Rules

  1. Start Immediately: First action is session initialization, then Phase 0/1
  2. Wave Order is Sacred: Never execute wave N before wave N-1 completes and results are merged
  3. CSV is Source of Truth: Master tasks.csv holds all state (both csv-wave and interactive)
  4. CSV First: Default to csv-wave for tasks; only use interactive when user interaction is needed
  5. Context Propagation: prev_context built from master CSV, not from memory
  6. Discovery Board is Append-Only: Never clear, modify, or recreate discoveries.ndjson — both mechanisms share it
  7. Skip on Failure: If a dependency failed, skip the dependent task (regardless of mechanism)
  8. Lifecycle Balance: Every spawn_agent MUST have a matching close_agent
  9. Cleanup Temp Files: Remove wave-{N}.csv after results are merged
  10. DO NOT STOP: Continuous execution until all waves complete or all remaining tasks are skipped