Claude Code Plugins

Community-maintained marketplace

Feedback

|

Install Skill

1Download skill
2Enable skills in Claude

Open claude.ai/settings/capabilities and find the "Skills" section

3Upload to Claude

Click "Upload skill" and select the downloaded ZIP file

Note: Please verify skill by going through its instructions before using it.

SKILL.md

name rl-optimization
description Apply reinforcement learning principles when working on spec-kit-extensions. Auto-activates when: (1) editing command prompts in commands/*.md, (2) modifying workflow templates in extensions/workflows/, (3) discussing user feedback about workflow friction, (4) reviewing issues or PRs mentioning prompt clarity or template problems, (5) analyzing chat logs or workflow usage from other repositories. Helps ensure changes improve prompt effectiveness and template utility.

RL Optimization Skill

Apply reinforcement learning principles to improve spec-kit prompts and templates.

When This Skill Activates

This skill provides background awareness when you're:

  • Editing commands/*.md files (prompt engineering)
  • Modifying extensions/workflows/*/ templates
  • Discussing friction points from real-world usage
  • Reviewing feedback about workflow effectiveness
  • Analyzing how workflows performed in other projects

Core Principles

Prompt Effectiveness Criteria

When editing command prompts, ensure they score well on:

Criterion Question to Ask
Initial Clarity Will the agent understand what to do immediately?
Step Sequence Are steps in logical order with clear transitions?
Action Specificity Are actions concrete and unambiguous?
Output Guidance Is the expected output format clear?
Error Recovery What happens if something goes wrong?
Completion Signal How does the agent know when it's done?

Template Effectiveness Criteria

When editing templates, ensure:

Criterion Question to Ask
Section Utility Will each section be filled with useful content?
Logical Order Does the structure flow naturally?
Placeholder Clarity Are placeholders self-explanatory?
Completeness Are all necessary sections present?
Conciseness Is there any redundancy to remove?

Red Flags to Watch For

In Prompts

  • Vague instructions: "Create the file" → "Create the file with ALL sections from the template"
  • Missing error handling: No guidance for when things fail
  • Assumed knowledge: References to concepts not explained
  • Ambiguous sequences: "Then do X or Y" without criteria for choosing
  • No completion criteria: Agent doesn't know when to stop

In Templates

  • Sections that are always skipped (remove or make optional)
  • Missing sections users frequently add manually
  • Placeholders that confuse more than help
  • Redundant information across sections
  • Poor ordering that breaks logical flow

Improvement Patterns

Pattern: Explicit Over Implicit

# Before (implicit)
Fill in the bug report template.

# After (explicit)
Fill in ALL sections of the bug report template. Do not skip any section,
even if information seems redundant. Pay special attention to:
- Reproduction Steps: Must be executable commands
- Root Cause: Use Five Whys analysis
- Prevention: Specific actions, not general statements

Pattern: Guided Decisions

# Before (ambiguous)
Choose the appropriate workflow.

# After (guided)
Choose the workflow based on the task:
- Bug with known cause → /speckit.bugfix
- Bug needing investigation → /speckit.bugfix (document investigation in root cause)
- Small improvement (<7 tasks) → /speckit.enhance
- Large feature → /speckit.specify

Pattern: Failure Recovery

# Before (no recovery)
Run the tests.

# After (with recovery)
Run the tests. If tests fail:
1. Check if failure is related to your changes
2. If yes, fix and re-run before proceeding
3. If no (pre-existing failure), document in notes and continue

When Making Changes

Before committing prompt or template changes:

  1. Check against criteria - Score the change on effectiveness criteria
  2. Look for red flags - Scan for patterns that cause friction
  3. Consider edge cases - What happens when things go wrong?
  4. Test mentally - Walk through as if you were the agent
  5. Compare before/after - Is the improvement clear?

Suggesting Intakes

When you notice patterns that suggest an RL intake would be valuable:

  • User describes repeated friction with a workflow
  • Multiple issues reference the same prompt confusion
  • A workflow was used extensively in another project
  • Post-mortem reveals systemic prompt issues

Suggest: "This sounds like good data for an RL intake. Want me to run /rl-intake to capture these patterns?"

Reference

See references/prompt-patterns.md for detailed examples of good and bad patterns.

Full process documentation: docs/rl-intake-process.md