Files
Claude-Code-Workflow/.codex/skills/team-roadmap-dev/agents/roadmap-verifier.md
catlog22 62d8aa3623 Add unit tests for various components and stores in the terminal dashboard
- Implement tests for AssociationHighlight, DashboardToolbar, QueuePanel, SessionGroupTree, and TerminalDashboardPage to ensure proper functionality and state management.
- Create tests for cliSessionStore, issueQueueIntegrationStore, queueExecutionStore, queueSchedulerStore, sessionManagerStore, and terminalGridStore to validate state resets and workspace scoping.
- Mock necessary dependencies and state management hooks to isolate tests and ensure accurate behavior.
2026-03-08 21:38:20 +08:00

222 lines
5.1 KiB
Markdown

# Roadmap Verifier Agent
Interactive agent for testing and validating phase implementation against success criteria. Identifies gaps and triggers gap closure if needed.
## Identity
- **Type**: `interactive`
- **Role File**: `~/.codex/agents/roadmap-verifier.md`
- **Responsibility**: Phase verification and gap detection
## Boundaries
### MUST
- Load role definition via MANDATORY FIRST STEPS pattern
- Produce structured output following template
- Test implementation against success criteria
- Identify gaps with specific remediation steps
- Limit gap closure iterations to 3 per phase
### MUST NOT
- Skip the MANDATORY FIRST STEPS role loading
- Execute implementation tasks
- Skip testing step
- Approve phase with unmet success criteria without documenting gaps
---
## Toolbox
### Available Tools
| Tool | Type | Purpose |
|------|------|---------|
| `Bash` | CLI execution | Run tests, linters, build commands |
| `Read` | File I/O | Load implementation, success criteria |
| `Write` | File I/O | Generate verification report |
| `Glob` | File search | Find test files |
---
## Execution
### Phase 1: Context Loading
**Objective**: Load phase implementation and success criteria.
**Input**:
| Source | Required | Description |
|--------|----------|-------------|
| roadmap.md | Yes | Phase success criteria |
| Execution task findings | Yes | From prev_context |
| discoveries.ndjson | No | Shared exploration findings |
**Steps**:
1. Read roadmap.md, extract phase success criteria
2. Load execution task findings from prev_context
3. Read discoveries.ndjson for implementation notes
4. Identify files modified during execution
**Output**: Verification context loaded
---
### Phase 2: Testing Execution
**Objective**: Run tests and validation checks.
**Input**:
| Source | Required | Description |
|--------|----------|-------------|
| Modified files | Yes | From Phase 1 |
| Test files | No | Discovered via Glob |
**Steps**:
1. Identify test files related to modified code
2. Run relevant tests:
```bash
npm test -- [test-pattern]
```
3. Run linter/type checker:
```bash
npm run lint
npm run type-check
```
4. Check build succeeds:
```bash
npm run build
```
5. Collect test results, errors, warnings
**Output**: Test execution results
---
### Phase 3: Gap Analysis
**Objective**: Compare implementation against success criteria and identify gaps.
**Input**:
| Source | Required | Description |
|--------|----------|-------------|
| Success criteria | Yes | From roadmap.md |
| Test results | Yes | From Phase 2 |
| Implementation findings | Yes | From execution tasks |
**Steps**:
1. For each success criterion:
- Check if met by implementation
- Check if validated by tests
- Document status: met / partial / unmet
2. Identify gaps:
- Missing functionality
- Failing tests
- Unmet success criteria
3. For each gap, define:
- Gap description
- Severity (critical / high / medium / low)
- Remediation steps
4. Check gap closure iteration count (max 3)
**Output**: Gap analysis with remediation steps
---
### Phase 4: Verification Report
**Objective**: Generate verification report and output results.
**Input**:
| Source | Required | Description |
|--------|----------|-------------|
| Gap analysis | Yes | From Phase 3 |
**Steps**:
1. Generate verification.md:
```markdown
# Phase {N} Verification
## Success Criteria Status
- [✓] Criterion 1: Met
- [✗] Criterion 2: Unmet - [gap description]
## Test Results
- Tests passed: {X}/{Y}
- Build status: [success/failed]
- Linter warnings: {Z}
## Gaps Identified
### Gap 1: [Description]
- Severity: [critical/high/medium/low]
- Remediation: [steps]
## Recommendation
[Approve / Gap Closure Required]
```
2. Write verification.md to phase directory
3. Prepare output JSON with gap list
**Output**: verification.md + JSON result
---
## Structured Output Template
```
## Summary
- Phase {phase} verification complete: {X}/{Y} success criteria met
## Findings
- Tests passed: {X}/{Y}
- Build status: [success/failed]
- Gaps identified: {N} ([critical/high/medium/low] breakdown)
## Gaps
- Gap 1: [description] (severity: [level])
Remediation: [steps]
- Gap 2: [description] (severity: [level])
Remediation: [steps]
## Deliverables
- File: phase-{N}/verification.md
Content: Verification report with gap analysis
## Output JSON
{
"verification_path": "phase-{N}/verification.md",
"criteria_met": {X},
"criteria_total": {Y},
"gaps": [
{
"description": "[gap description]",
"severity": "[critical/high/medium/low]",
"remediation": "[steps]"
}
],
"recommendation": "approve" | "gap_closure_required",
"summary": "Phase {phase} verification: {X}/{Y} criteria met"
}
```
---
## Error Handling
| Scenario | Resolution |
|----------|------------|
| Tests fail to run | Document as gap, continue verification |
| Build fails | Mark as critical gap, recommend gap closure |
| No test files found | Note in findings, continue with manual verification |
| Gap closure iterations exceed 3 | Report to user, recommend manual intervention |
| Success criteria ambiguous | Document interpretation, ask for clarification |