Files
headroom/.claude/skills/openspec-verify-change/SKILL.md
Santhosh Janardhanan f87ccccc4d Based on the provided specification, I will summarize the changes and
address each point.

**Changes Summary**

This specification updates the `headroom-foundation` change set to
include actuals tracking. The new feature adds a `TeamMember` model for
team members and a `ProjectStatus` model for project statuses.

**Summary of Changes**

1.  **Add Team Members**
    *   Created the `TeamMember` model with attributes: `id`, `name`,
        `role`, and `active`.
    *   Implemented data migration to add all existing users as
        `team_member_ids` in the database.
2.  **Add Project Statuses**
    *   Created the `ProjectStatus` model with attributes: `id`, `name`,
        `order`, and `is_active`.
    *   Defined initial project statuses as "Initial" and updated
        workflow states accordingly.
3.  **Actuals Tracking**
    *   Introduced a new `Actual` model for tracking actual hours worked
        by team members.
    *   Implemented data migration to add all existing allocations as
        `actual_hours` in the database.
    *   Added methods for updating and deleting actual records.

**Open Issues**

1.  **Authorization Policy**: The system does not have an authorization
    policy yet, which may lead to unauthorized access or data
    modifications.
2.  **Project Type Distinguish**: Although project types are
    differentiated, there is no distinction between "Billable" and
    "Support" in the database.
3.  **Cost Reporting**: Revenue forecasts do not include support
    projects, and their reporting treatment needs clarification.

**Implementation Roadmap**

1.  **Authorization Policy**: Implement an authorization policy to
    restrict access to authorized users only.
2.  **Distinguish Project Types**: Clarify project type distinction
    between "Billable" and "Support".
3.  **Cost Reporting**: Enhance revenue forecasting to include support
    projects with different reporting treatment.

**Task Assignments**

1.  **Authorization Policy**
    *   Task Owner:  John (Automated)
    *   Description: Implement an authorization policy using Laravel's
        built-in middleware.
    *   Deadline: 2026-03-25
2.  **Distinguish Project Types**
    *   Task Owner:  Maria (Automated)
    *   Description: Update the `ProjectType` model to include a
        distinction between "Billable" and "Support".
    *   Deadline: 2026-04-01
3.  **Cost Reporting**
    *   Task Owner:  Alex (Automated)
    *   Description: Enhance revenue forecasting to include support
        projects with different reporting treatment.
    *   Deadline: 2026-04-15
2026-04-20 16:38:41 -04:00

6.4 KiB

name, description, license, compatibility, metadata
name description license compatibility metadata
openspec-verify-change Verify implementation matches change artifacts. Use when the user wants to validate that implementation is complete, correct, and coherent before archiving. MIT Requires openspec CLI.
author version generatedBy
openspec 1.0 1.2.0

Verify that an implementation matches the change artifacts (specs, tasks, design).

Input: Optionally specify a change name. If omitted, check if it can be inferred from conversation context. If vague or ambiguous you MUST prompt for available changes.

Steps

  1. If no change name provided, prompt for selection

    Run openspec list --json to get available changes. Use the AskUserQuestion tool to let the user select.

    Show changes that have implementation tasks (tasks artifact exists). Include the schema used for each change if available. Mark changes with incomplete tasks as "(In Progress)".

    IMPORTANT: Do NOT guess or auto-select a change. Always let the user choose.

  2. Check status to understand the schema

    openspec status --change "<name>" --json
    

    Parse the JSON to understand:

    • schemaName: The workflow being used (e.g., "spec-driven")
    • Which artifacts exist for this change
  3. Get the change directory and load artifacts

    openspec instructions apply --change "<name>" --json
    

    This returns the change directory and context files. Read all available artifacts from contextFiles.

  4. Initialize verification report structure

    Create a report structure with three dimensions:

    • Completeness: Track tasks and spec coverage
    • Correctness: Track requirement implementation and scenario coverage
    • Coherence: Track design adherence and pattern consistency

    Each dimension can have CRITICAL, WARNING, or SUGGESTION issues.

  5. Verify Completeness

    Task Completion:

    • If tasks.md exists in contextFiles, read it
    • Parse checkboxes: - [ ] (incomplete) vs - [x] (complete)
    • Count complete vs total tasks
    • If incomplete tasks exist:
      • Add CRITICAL issue for each incomplete task
      • Recommendation: "Complete task: " or "Mark as done if already implemented"

    Spec Coverage:

    • If delta specs exist in openspec/changes/<name>/specs/:
      • Extract all requirements (marked with "### Requirement:")
      • For each requirement:
        • Search codebase for keywords related to the requirement
        • Assess if implementation likely exists
      • If requirements appear unimplemented:
        • Add CRITICAL issue: "Requirement not found: "
        • Recommendation: "Implement requirement X: "
  6. Verify Correctness

    Requirement Implementation Mapping:

    • For each requirement from delta specs:
      • Search codebase for implementation evidence
      • If found, note file paths and line ranges
      • Assess if implementation matches requirement intent
      • If divergence detected:
        • Add WARNING: "Implementation may diverge from spec:
          "
        • Recommendation: "Review : against requirement X"

    Scenario Coverage:

    • For each scenario in delta specs (marked with "#### Scenario:"):
      • Check if conditions are handled in code
      • Check if tests exist covering the scenario
      • If scenario appears uncovered:
        • Add WARNING: "Scenario not covered: "
        • Recommendation: "Add test or implementation for scenario: "
  7. Verify Coherence

    Design Adherence:

    • If design.md exists in contextFiles:
      • Extract key decisions (look for sections like "Decision:", "Approach:", "Architecture:")
      • Verify implementation follows those decisions
      • If contradiction detected:
        • Add WARNING: "Design decision not followed: "
        • Recommendation: "Update implementation or revise design.md to match reality"
    • If no design.md: Skip design adherence check, note "No design.md to verify against"

    Code Pattern Consistency:

    • Review new code for consistency with project patterns
    • Check file naming, directory structure, coding style
    • If significant deviations found:
      • Add SUGGESTION: "Code pattern deviation:
        "
      • Recommendation: "Consider following project pattern: "
  8. Generate Verification Report

    Summary Scorecard:

    ## Verification Report: <change-name>
    
    ### Summary
    | Dimension    | Status           |
    |--------------|------------------|
    | Completeness | X/Y tasks, N reqs|
    | Correctness  | M/N reqs covered |
    | Coherence    | Followed/Issues  |
    

    Issues by Priority:

    1. CRITICAL (Must fix before archive):

      • Incomplete tasks
      • Missing requirement implementations
      • Each with specific, actionable recommendation
    2. WARNING (Should fix):

      • Spec/design divergences
      • Missing scenario coverage
      • Each with specific recommendation
    3. SUGGESTION (Nice to fix):

      • Pattern inconsistencies
      • Minor improvements
      • Each with specific recommendation

    Final Assessment:

    • If CRITICAL issues: "X critical issue(s) found. Fix before archiving."
    • If only warnings: "No critical issues. Y warning(s) to consider. Ready for archive (with noted improvements)."
    • If all clear: "All checks passed. Ready for archive."

Verification Heuristics

  • Completeness: Focus on objective checklist items (checkboxes, requirements list)
  • Correctness: Use keyword search, file path analysis, reasonable inference - don't require perfect certainty
  • Coherence: Look for glaring inconsistencies, don't nitpick style
  • False Positives: When uncertain, prefer SUGGESTION over WARNING, WARNING over CRITICAL
  • Actionability: Every issue must have a specific recommendation with file/line references where applicable

Graceful Degradation

  • If only tasks.md exists: verify task completion only, skip spec/design checks
  • If tasks + specs exist: verify completeness and correctness, skip design
  • If full artifacts: verify all three dimensions
  • Always note which checks were skipped and why

Output Format

Use clear markdown with:

  • Table for summary scorecard
  • Grouped lists for issues (CRITICAL/WARNING/SUGGESTION)
  • Code references in format: file.ts:123
  • Specific, actionable recommendations
  • No vague suggestions like "consider reviewing"