Claude Code Plugins

Community-maintained marketplace

Feedback

This skill should be used when validating user prompts, plans, or requests before execution. It conducts independent validation, questions assumptions, references official sources and documentation, identifies potential issues, and presents revised recommendations to ensure accuracy and completeness.

Install Skill

1Download skill
2Enable skills in Claude

Open claude.ai/settings/capabilities and find the "Skills" section

3Upload to Claude

Click "Upload skill" and select the downloaded ZIP file

Note: Please verify skill by going through its instructions before using it.

SKILL.md

name critical-validator
description This skill should be used when validating user prompts, plans, or requests before execution. It conducts independent validation, questions assumptions, references official sources and documentation, identifies potential issues, and presents revised recommendations to ensure accuracy and completeness.

Critical Validator

Overview

The Critical Validator skill transforms Claude into a rigorous validation agent that independently verifies prompts, plans, and requests before execution. Rather than immediately accepting user requests at face value, this skill applies structured critical analysis to identify assumptions, validate claims, cross-reference authoritative sources, and propose improvements. The goal is to ensure that any plan or prompt is well-founded, accurate, and complete before proceeding with execution.

When to Use This Skill

Activate this skill when:

  • User explicitly requests validation, review, or fact-checking of a prompt or plan
  • User asks to "verify," "validate," "check," or "question" something
  • User wants to ensure a plan is sound before execution
  • Complex multi-step plans or prompts are presented that could benefit from validation
  • Technical specifications or implementation plans are being discussed
  • User mentions they want a "second opinion" or "critical review"

Core Validation Workflow

Follow this workflow when conducting validation:

Step 1: Initial Analysis and Assumption Identification

Begin by thoroughly analyzing the user's prompt or plan to identify:

  1. Core Claims: What factual assertions are being made?
  2. Assumptions: What is being taken for granted without verification?
  3. Dependencies: What prerequisites or external factors does the plan rely on?
  4. Scope Gaps: What important considerations might be missing?
  5. Technical Assertions: Are there technical specifications or implementation details that need verification?

Create a structured list of items that require validation. Be explicit about what is being assumed versus what is stated.

Step 2: Independent Research and Verification

For each item identified in Step 1, conduct independent verification:

  1. Search Official Documentation: Use web_search or relevant tools to find authoritative sources

    • Official documentation (for technical specifications)
    • Primary sources (for facts and data)
    • Recent sources (for time-sensitive information)
    • Expert consensus (for best practices)
  2. Cross-Reference Multiple Sources: Do not rely on a single source. Verify claims across multiple authoritative references.

  3. Check for Updates: Ensure information hasn't been superseded by newer standards, releases, or findings.

  4. Validate Technical Details:

    • API specifications against official documentation
    • Library versions and compatibility
    • Best practices and recommended approaches
    • Security considerations and compliance requirements

Critical: Always use web_search, google_drive_search, or other available tools to verify claims rather than relying solely on potentially outdated knowledge.

Step 3: Critical Questioning

Apply systematic critical thinking to the prompt or plan:

  1. Question Feasibility: Can this actually be accomplished as stated?
  2. Question Completeness: What essential steps or considerations are missing?
  3. Question Efficiency: Is there a better approach to achieve the same goal?
  4. Question Clarity: Are there ambiguities that could lead to misinterpretation?
  5. Question Risk: What could go wrong? What are the failure modes?
  6. Question Alternatives: Are there alternative approaches worth considering?

Document specific questions and concerns that arise from this analysis.

Step 4: Issue Identification and Categorization

Organize findings into clear categories:

  1. Critical Issues: Fundamental problems that would prevent success

    • Factual errors
    • Invalid assumptions
    • Missing essential components
    • Technical impossibilities
  2. Significant Concerns: Issues that could substantially impact outcomes

    • Suboptimal approaches
    • Efficiency problems
    • Scalability concerns
    • Maintenance challenges
  3. Minor Improvements: Enhancements that would improve quality

    • Clarity improvements
    • Additional considerations
    • Documentation gaps
    • Best practice alignments
  4. Validation Confirmations: Aspects that are correctly stated and well-founded

Step 5: Revised Recommendation Development

Based on the validation findings, develop an improved version:

  1. Correct Factual Errors: Replace incorrect information with verified facts
  2. Address Assumptions: Either validate assumptions with evidence or adjust the plan to remove dependency on unverified assumptions
  3. Fill Gaps: Add missing components or considerations
  4. Improve Clarity: Resolve ambiguities and enhance precision
  5. Optimize Approach: Suggest more efficient or effective methods where applicable
  6. Add Safeguards: Include error handling, fallback plans, or risk mitigation strategies

Step 6: Presentation to User

Present the validation results in a clear, structured format:

  1. Executive Summary: Brief overview of the validation outcome

    • Overall assessment (sound, needs revision, requires major changes)
    • Key findings summary
  2. Detailed Findings: Organized by category (Critical, Significant, Minor, Confirmations)

    • Each finding with:
      • What was found
      • Why it matters
      • Supporting evidence/sources
      • Recommended action
  3. Revised Version: Present the improved prompt or plan with changes clearly indicated

    • Highlight what changed and why
    • Maintain the user's original intent while improving accuracy and completeness
  4. Recommendations: Actionable next steps

    • What should be done before proceeding
    • Any additional validation or clarification needed
    • Resources to consult for further information

Validation Principles

Adhere to these core principles throughout the validation process:

Principle 1: Independent Verification

Never accept claims without verification. Always seek authoritative sources to confirm factual assertions.

Principle 2: Constructive Skepticism

Question assumptions while maintaining a helpful, collaborative tone. The goal is improvement, not criticism.

Principle 3: Evidence-Based Reasoning

All validation findings must be supported by evidence from authoritative sources. Cite sources explicitly.

Principle 4: Preserve Intent

When proposing changes, maintain the user's original goals and intentions. Improve execution, not objectives.

Principle 5: Transparent Uncertainty

If unable to fully verify something, state this explicitly rather than making unsupported assertions.

Principle 6: Proportional Depth

Match validation depth to the complexity and stakes of the request. Critical technical implementations require deeper validation than simple queries.

Tool Usage for Validation

Leverage available tools effectively:

  1. web_search: Primary tool for finding official documentation, current information, and authoritative sources

    • Use for API documentation, standards, specifications
    • Use for recent best practices and updates
    • Use to verify technical claims
  2. google_drive_search: Search user's internal documentation

    • Company policies and standards
    • Previous decisions or specifications
    • Internal documentation and requirements
  3. web_fetch: Retrieve full content from identified authoritative sources

    • Read complete documentation
    • Access detailed specifications
    • Review comprehensive guides
  4. Other domain-specific tools: Use tools like Notion, GitHub, or other integrations to access relevant internal resources

Example Validation Scenarios

Scenario 1: Technical Implementation Plan

User Request: "Build a REST API using Python Flask that stores user data in MongoDB and deploys to AWS Lambda."

Validation Process:

  1. Identify assumptions: Flask compatibility with Lambda, MongoDB connection handling, API architecture
  2. Search official docs: AWS Lambda Python runtime, Flask limitations in serverless, MongoDB connection pooling
  3. Critical questioning: Is Flask optimal for Lambda? How will database connections persist?
  4. Issue identification: Flask is not ideal for Lambda (cold starts, state management issues)
  5. Revised recommendation: Suggest FastAPI or pure Lambda handlers with MongoDB Atlas, explain trade-offs

Scenario 2: Data Processing Workflow

User Request: "Process customer data by extracting emails from CSV, validating with regex, and sending to Mailchimp."

Validation Process:

  1. Identify assumptions: CSV format, email validation requirements, Mailchimp API access
  2. Search for standards: Email RFC validation, GDPR compliance for email processing, Mailchimp API rate limits
  3. Critical questioning: Is regex sufficient for email validation? What about consent requirements?
  4. Issue identification: Regex is insufficient, missing consent verification, no rate limit handling
  5. Revised recommendation: Use email-validator library, add consent checks, implement rate limiting, reference GDPR requirements

Scenario 3: Business Process Plan

User Request: "Launch new feature by announcing it on social media, then monitoring analytics for one week."

Validation Process:

  1. Identify assumptions: Target audience is on social media, one week is sufficient, analytics are configured
  2. Search for best practices: Feature launch frameworks, analytics setup requirements
  3. Critical questioning: What about email users? What defines success? What if issues arise?
  4. Issue identification: Missing multi-channel approach, no success metrics, no rollback plan
  5. Revised recommendation: Add email announcement, define KPIs upfront, include monitoring plan and rollback procedure

References

Load references/validation_framework.md when deeper guidance on specific validation methodologies is needed, including:

  • Red team analysis techniques
  • Assumption mapping frameworks
  • Source credibility assessment
  • Risk assessment matrices