Claude Code Plugins

Community-maintained marketplace

Feedback
909
13

Facilitate workshop sessions in a multi-turn, one-step flow with numbered recommendations at decision points and quick-select options for regular questions.

Install Skill

1Download skill
2Enable skills in Claude

Open claude.ai/settings/capabilities and find the "Skills" section

3Upload to Claude

Click "Upload skill" and select the downloaded ZIP file

Note: Please verify skill by going through its instructions before using it.

SKILL.md

name workshop-facilitation
description Facilitate workshop sessions in a one-step, multi-turn flow. Use when an interactive skill needs consistent pacing, options, and progress tracking.
intent Provide the canonical facilitation pattern for interactive skills: one step at a time, with clear progress, adaptive recommendations at decision points, and predictable interruption handling.
type interactive
theme workshops-facilitation
best_for Adding structured facilitation to any PM workshop or guided session, Running interactive sessions with numbered recommendations and progress tracking, Ensuring your workshops stay on track and end with actionable choices
scenarios I want to run a structured positioning workshop with my product team — set up the facilitation protocol, Help me facilitate a discovery sprint kickoff with clear questions, options, and progress labels
estimated_time varies by workshop

Purpose

Provide the canonical facilitation pattern for interactive skills: one step at a time, with clear progress, adaptive recommendations at decision points, and predictable interruption handling.

Key Concepts

  • One-step-at-a-time: Ask a single targeted question per turn.
  • Session heads-up + entry mode: Start by setting expectations and offering Guided, Context dump, or Best guess mode.
  • Progress visibility: Show user-facing progress labels like Context Qx/8 and Scoring Qx/5.
  • Decision-point recommendations: Use enumerated options only when a choice is needed, not after every answer.
  • Quick-select response options: For regular context/scoring questions, provide concise numbered answer options plus Other (specify) when useful.
  • Flexible selection parsing: Accept #1, 1, 1 and 3, 1,3, or custom text, then synthesize multi-select choices.
  • Context-aware progression: Build on previous answers and avoid re-asking resolved questions.
  • Interruption-safe flow: Answer meta questions directly (for example, "how many left?"), restate status, then resume.
  • Fast path: If the user requests a single-shot output, skip multi-turn facilitation and deliver a condensed result.

Application

  1. Start with a brief heads-up on estimated time and number of questions.
  2. Ask the user to choose an entry mode:
    • 1 Guided mode (one question at a time)
    • 2 Context dump (paste known context; skip redundancies)
    • 3 Best guess mode (infer missing details and label assumptions)
  3. Run one question per turn and wait for an answer before continuing.
  4. Keep questions plain-language; include a short example response format when helpful.
  5. Show progress each turn:
    • Context Qx/8 during context collection
    • Scoring Qx/5 during assessment/scoring
  6. Ask follow-up clarifications only when they materially improve recommendation quality.
  7. For regular context/scoring questions, offer quick-select numbered response options when practical:
    • Keep options concise and mutually exclusive when possible.
    • Include Other (specify) if likely answers are open-ended.
    • Accept multi-select responses like 1,3 or 1 and 3.
  8. Provide numbered recommendations only at decision points:
    • after context synthesis,
    • after maturity/profile synthesis,
    • during priority/action-plan selection.
  9. Accept numeric or custom choices, synthesize multi-select choices, and continue.
  10. If interrupted by a meta question, answer directly, then restate progress and pending question.
  11. If the user says stop/pause, halt immediately and wait for explicit resume.
  12. End with a clear summary, decisions made, and (if best guess mode was used) an Assumptions to Validate list.

Examples

Opening: "Quick heads-up: this should take about 7-10 minutes and around 10 questions. How do you want to start?

  1. Guided mode
  2. Context dump
  3. Best guess mode"

User: "2"

Facilitator: "Paste what you already know. I’ll skip answered areas and ask only what’s missing."

Decision point after synthesis:

  1. Prioritize Context Design (Recommended)
  2. Prioritize Agent Orchestration
  3. Prioritize Team-AI Facilitation

User: "1 and 3"

Facilitator: "Great. We’ll run Context Design first, with Team-AI Facilitation in parallel."

Common Pitfalls

  • Asking multiple questions in the same turn.
  • Offering recommendations after every answer (creates interaction drag).
  • Using shorthand labels without plain-language questions.
  • Hiding progress, so users don't know how much remains.
  • Ignoring the user's chosen option or custom direction.
  • Failing to label assumptions when running in best-guess mode.

References

  • Use as the source of truth for interactive facilitation behavior.
  • Apply alongside workshop skills in skills/*-workshop/SKILL.md and advisor-style interactive skills.