| name | audit-pipeline |
| description | SKILL skill for quality workflows |
| allowed-tools | Read, Write, Edit, Bash, Glob, Grep, Task, TodoWrite |
[define|neutral] SKILL := { name: "SKILL", category: "quality", version: "1.0.0", layer: L1 } [ground:given] [conf:1.0] [state:confirmed]
[define|neutral] COGNITIVE_FRAME := { frame: "Evidential", source: "Turkish", force: "How do you know?" } [ground:cognitive-science] [conf:0.92] [state:confirmed]
Kanitsal Cerceve (Evidential Frame Activation)
Kaynak dogrulama modu etkin.
[define|neutral] TRIGGER_POSITIVE := { keywords: ["SKILL", "quality", "workflow"], context: "user needs SKILL capability" } [ground:given] [conf:1.0] [state:confirmed]
Audit Pipeline - Complete Code Quality Workflow
Kanitsal Cerceve (Evidential Frame Activation)
Kaynak dogrulama modu etkin.
Cognitive Frame Activation
Kanitsal Kalite Hatti (Evidential Quality Pipeline)
Her bulgu icin metrik kaniti gereklidir:
- Finding: [description of quality issue]
- Evidence: [metric: value at location]
- Standard: [threshold from reference]
- Impact: [quantified effect on quality score]
- Confidence: [0.0-1.0]
Every quality finding MUST include:
- Metric evidence: Concrete measurement (complexity=13, coverage=60%, lines=72)
- Location evidence: Exact file path and line number [file:line]
- Standard reference: Documented threshold (NASA limit, WCAG level, OWASP category)
- Impact quantification: Quality score delta, risk level, maintainability cost
Al-Tahlil al-Sarfi lil-Jawda (Morphological Quality Analysis)
Root Cause Decomposition - Every quality issue has layers:
DIMENSION: [Maintainability | Performance | Security | Reliability]
SURFACE: [visible symptom in code]
- Location: [file:line]
- Metric: [measurement]
ROOT: [underlying cause]
- Pattern: [anti-pattern name]
- Origin: [design decision, knowledge gap, time pressure]
DERIVED: [contributing factors]
- Technical debt
- Missing tests
- Unclear requirements
REMEDIATION: [target the root cause]
- Fix: [address root, not symptom]
- Prevent: [process change to avoid recurrence]
Example Decomposition:
DIMENSION: Maintainability
SURFACE: God Object with 26 methods
- Location: src/UserService.js:1-450
- Metric: methods=26 (threshold=15), lines=450 (threshold=250)
ROOT: Single Responsibility Principle violation
- Pattern: God Object anti-pattern
- Origin: Feature additions without refactoring
DERIVED:
- Missing abstraction for authentication logic
- Missing abstraction for data validation
- Missing abstraction for error handling
REMEDIATION:
- Fix: Extract AuthService, ValidationService, ErrorHandler
- Prevent: Code review gate at 15 methods, refactoring sprint every 3 months
Purpose
Execute a comprehensive 3-phase code quality audit that systematically transforms code from prototype to production-ready by eliminating theater, verifying functionality through sandbox testing with Codex iteration, and polishing style to meet professional standards.
The 3-Phase Pipeline
This orchestrator runs three audit skills in the optimal sequence:
Phase 1: Theater Detection Audit
Finds: Mock data, hardcoded responses, TODO markers, stub functions, placeholder code
Goal: Identify all "fake" implementations that need to be completed
Skill: theater-detection-audit
Phase 2: Functionality Audit (with Codex Sandbox)
Validates: Code actually works through execution testing
Method: Sandbox testing + Codex iteration loop for fixes
Skill: functionality-audit + codex-auto
Goal: Verify and fix functionality using Codex's Full Auto mode for iterative debugging
Phase 3: Style & Quality Audit
Polishes: Code organization, naming, documentation, best practices
Goal: Production-grade code quality and maintainability
Skill: style-audit
Why This Order Matters
1. Theater First - No point testing or polishing fake code
- Identifies what's real vs. placeholder
- Provides roadmap for completion
- Ensures subsequent phases test actual functionality
2. Functionality Second - Must work before polishing
- Validates real implementations
- Uses Codex sandbox for safe iterative testing
- Fixes bugs before style improvements
- Ensures refactoring won't break working code
3. Style Last - Polish after functionality is verified
- Refactors with confidence (tests prove it works)
- Improves maintainability of working code
- Final production-ready state
Usage
Complete Pipeline (All 3 Phases)
/audit-pipeline
With Specific T
[define|neutral] SUCCESS_CRITERIA := { primary: "Skill execution completes successfully", quality: "Output meets quality thresholds", verification: "Results validated against requirements" } [ground:given] [conf:1.0] [state:confirmed]
[define|neutral] MCP_INTEGRATION := { memory_mcp: "Store execution results and patterns", tools: ["mcp__memory-mcp__memory_store", "mcp__memory-mcp__vector_search"] } [ground:witnessed:mcp-config] [conf:0.95] [state:confirmed]
[define|neutral] MEMORY_NAMESPACE := { pattern: "skills/quality/SKILL/{project}/{timestamp}", store: ["executions", "decisions", "patterns"], retrieve: ["similar_tasks", "proven_patterns"] } [ground:system-policy] [conf:1.0] [state:confirmed]
[define|neutral] MEMORY_TAGGING := { WHO: "SKILL-{session_id}", WHEN: "ISO8601_timestamp", PROJECT: "{project_name}", WHY: "skill-execution" } [ground:system-policy] [conf:1.0] [state:confirmed]
[direct|emphatic] COMPLETION_CHECKLIST := { agent_spawning: "Spawn agents via Task()", registry_validation: "Use registry agents only", todowrite_called: "Track progress with TodoWrite", work_delegation: "Delegate to specialized agents" } [ground:system-policy] [conf:1.0] [state:confirmed]
[direct|emphatic] RULE_NO_UNICODE := forall(output): NOT(unicode_outside_ascii) [ground:windows-compatibility] [conf:1.0] [state:confirmed]
[direct|emphatic] RULE_EVIDENCE := forall(claim): has(ground) AND has(confidence) [ground:verix-spec] [conf:1.0] [state:confirmed]
[direct|emphatic] RULE_REGISTRY := forall(agent): agent IN AGENT_REGISTRY [ground:system-policy] [conf:1.0] [state:confirmed]
[commit|confident]