Skip to main content

Quality Scaffolding and Completeness Rubric

This is a testing/completeness support page. Use it when you need a practical way to judge whether work is genuinely complete or only looks complete.

This rubric exists to make quality more structured and harder to fake.

AI can now help teams produce more artifacts, more summaries, more tests, and more implementation in less time. That does not automatically mean delivery is becoming more complete or more trustworthy.

VibeGov uses this rubric to separate trustworthy completion from appearance-of-completion.

Quality scaffolding

Quality scaffolding is the supporting artifact layer around a change that makes the work trustworthy, reviewable, maintainable, and releasable.

Typical quality scaffolding includes:

  • issue clarity
  • spec/requirement binding
  • tests and validation evidence
  • documentation updates
  • traceability links
  • PR and handoff clarity
  • release-readiness or operational evidence
  • explicit residual risk and follow-up debt

The practical rule is simple:

AI should help teams maintain quality scaffolding, not just accelerate implementation.

Completeness dimensions

For meaningful work, review the change across these dimensions where relevant.

1. Intent completeness

Questions:

  • Is the issue implementation-grade?
  • Are constraints, non-goals, and acceptance criteria clear?
  • Is the work bound to spec/requirement IDs?

Weak signal:

  • vague issue plus fast implementation

Strong signal:

  • clear intent and explicit requirement/spec binding before work starts

2. Change completeness

Questions:

  • Did the scoped change actually happen?
  • Is the diff coherent and in-scope?
  • Did the work avoid unrelated side changes?

Weak signal:

  • partial change reported as full change

Strong signal:

  • coherent, reviewable change unit aligned to the issue

3. Proof completeness

Questions:

  • What evidence proves the change?
  • Were tests/checks/manual proof run appropriately to risk?
  • Is there a clear line from requirement to evidence?

Weak signal:

  • build passed, therefore done

Strong signal:

  • evidence is explicit, relevant, and proportional to the claim

4. Explanation completeness

Questions:

  • Can another contributor understand what changed and why?
  • Are docs, PR summaries, and handoff notes clear?
  • Are known limits or residual risks explicit?

Weak signal:

  • the truth is trapped in chat or private memory

Strong signal:

  • the work is legible without archaeology

5. Traceability completeness

Questions:

  • Are issue, spec, implementation, and evidence linked?
  • Are follow-up gaps visible as tracked work?

Weak signal:

  • proof or rationale exists somewhere, but not in the governed system

Strong signal:

  • future contributors can follow the chain of intent to evidence to follow-up

6. Operational completeness

Questions:

  • Are release/deploy/startup/smoke implications explicit where relevant?
  • Is residual risk visible?
  • Is the board/PR/release state honest?

Weak signal:

  • merged means done, operationally unspecified

Strong signal:

  • the work is safer to promote or easier to hand off because the operational truth is visible

Fast review model

Use a simple three-state review per dimension:

  • Pass — strong enough for the scoped work
  • Weak — present but incomplete or easy to overclaim
  • Missing — absent for a dimension that matters to this work

This is not a vanity score. It is a review lens.

Proportional evidence principle

Not every change needs the same burden of proof.

Use evidence proportional to work type and risk:

  • low-risk docs/content changes -> lighter proof
  • normal behavior changes -> targeted + regression evidence
  • contract/API changes -> stronger compatibility/contract proof
  • release-critical or risky changes -> strongest operational evidence

The goal is not bureaucracy. The goal is honest completeness.

Anti-fake-completion rules

Do not confuse these with trustworthy completion:

  • passing build == verified behavior
  • merged PR == done
  • long summary == evidence
  • generated tests == meaningful coverage
  • closed issue == complete work

AI makes it easier to generate the appearance of completeness. Governance must separate appearance from proof.

Missing completeness becomes tracked work

When work cannot fully satisfy the relevant completeness dimensions, the missing part must become visible follow-up work.

Examples:

  • debt issue
  • missing-test issue
  • deferred-doc update
  • blocker artifact
  • explicit residual-risk note with linked follow-up issue

Missing completeness must not remain an invisible compromise.

Typical questions to ask before calling work done

  • What changed?
  • What proves it?
  • What explains it?
  • What links it back to intent?
  • What remains incomplete?
  • Did the change improve the quality state of the system, or just add output?