Claude Code Plugins

Community-maintained marketplace

Feedback

when-chaining-workflows-use-cascade-orchestrator

@DNYoussef/context-cascade
6
0

Create sophisticated workflow cascades with sequential pipelines, parallel execution, and conditional branching

Install Skill

1Download skill
2Enable skills in Claude

Open claude.ai/settings/capabilities and find the "Skills" section

3Upload to Claude

Click "Upload skill" and select the downloaded ZIP file

Note: Please verify skill by going through its instructions before using it.

SKILL.md

name when-chaining-workflows-use-cascade-orchestrator
description Create sophisticated workflow cascades with sequential pipelines, parallel execution, and conditional branching
version 1.0.0
tags cascade, workflows, orchestration, conditional, branching
category coordination
agents task-orchestrator, hierarchical-coordinator, memory-coordinator
complexity advanced
estimated_duration 45-90 minutes
prerequisites Claude Flow installed, Workflow design experience, Conditional logic knowledge
outputs Cascade workflow definitions, Execution results, Performance metrics, Workflow reports

Cascade Orchestrator SOP

Overview

Create sophisticated workflow cascades that chain multiple workflows together with sequential pipelines, parallel execution, conditional branching, and error handling.

Agents & Responsibilities

task-orchestrator

Role: Orchestrate workflow execution Responsibilities:

  • Coordinate workflow execution
  • Handle workflow transitions
  • Manage workflow state
  • Monitor progress

hierarchical-coordinator

Role: Manage workflow hierarchy Responsibilities:

  • Organize workflow levels
  • Handle parent-child workflows
  • Manage dependencies
  • Ensure proper ordering

memory-coordinator

Role: Manage workflow state and data flow Responsibilities:

  • Store workflow state
  • Pass data between workflows
  • Maintain execution history
  • Ensure data consistency

Phase 1: Design Cascade

Objective

Design cascade structure with workflows, dependencies, and branching logic.

Scripts

# Create cascade definition
cat > cascade-definition.yaml <<EOF
name: full-stack-development
workflows:
  - id: design
    type: sequential
    steps: [requirements, architecture, database-design]
  - id: backend
    type: parallel
    steps: [api-impl, auth-impl, db-impl]
    depends_on: [design]
  - id: frontend
    type: sequential
    steps: [ui-impl, integration]
    depends_on: [backend]
  - id: testing
    type: parallel
    steps: [unit-tests, integration-tests, e2e-tests]
    depends_on: [backend, frontend]
  - id: deployment
    type: conditional
    condition: "testing.success_rate > 0.95"
    steps: [build, deploy, verify]
    depends_on: [testing]
EOF

# Validate cascade
npx claude-flow@alpha cascade validate --definition cascade-definition.yaml

# Visualize cascade
npx claude-flow@alpha cascade visualize \
  --definition cascade-definition.yaml \
  --output cascade-diagram.png

# Store cascade definition
npx claude-flow@alpha memory store \
  --key "cascade/definition" \
  --file cascade-definition.yaml

Cascade Patterns

Sequential Cascade:

Workflow A → Workflow B → Workflow C

Parallel Cascade:

         ┌─ Workflow B ─┐
Workflow A ├─ Workflow C ─┤ Workflow E
         └─ Workflow D ─┘

Conditional Cascade:

Workflow A → Decision → [if true] Workflow B
                      → [if false] Workflow C

Hybrid Cascade:

Design → ┬─ Backend ─┐
         └─ Frontend─┴─ [if tests pass] → Deploy

Phase 2: Chain Workflows

Objective

Connect workflows with proper data flow and dependency management.

Scripts

# Initialize cascade
npx claude-flow@alpha cascade init \
  --definition cascade-definition.yaml

# Connect workflow stages
npx claude-flow@alpha cascade connect \
  --from design \
  --to backend \
  --data-flow "architecture-docs"

npx claude-flow@alpha cascade connect \
  --from design \
  --to frontend \
  --data-flow "ui-specs"

npx claude-flow@alpha cascade connect \
  --from backend \
  --to testing \
  --data-flow "api-endpoints"

# Setup conditional branching
npx claude-flow@alpha cascade branch \
  --workflow testing \
  --condition "success_rate > 0.95" \
  --true-path deployment \
  --false-path debugging

# Verify connections
npx claude-flow@alpha cascade status --show-connections

Data Flow Configuration

# Configure data passing
npx claude-flow@alpha cascade data-flow \
  --from "design.architecture" \
  --to "backend.api-spec" \
  --transform "extract-api-endpoints"

# Setup shared state
npx claude-flow@alpha memory store \
  --key "cascade/shared-state" \
  --value '{
    "project": "full-stack-app",
    "version": "1.0.0",
    "environment": "production"
  }'

Phase 3: Execute Cascade

Objective

Execute cascading workflows with proper sequencing and error handling.

Scripts

# Execute cascade
npx claude-flow@alpha cascade execute \
  --definition cascade-definition.yaml \
  --input initial-requirements.json \
  --strategy adaptive

# Monitor execution
npx claude-flow@alpha cascade monitor \
  --interval 10 \
  --show-current-workflow \
  --show-progress

# Handle workflow transitions
npx claude-flow@alpha cascade transitions \
  --show-pending \
  --show-completed

# Track workflow status
npx claude-flow@alpha cascade workflow-status \
  --workflow backend \
  --detailed

Execution Strategies

Adaptive Execution:

# Automatically adjusts based on performance
npx claude-flow@alpha cascade execute \
  --strategy adaptive \
  --optimize-for throughput

Fail-Fast Execution:

# Stops immediately on first error
npx claude-flow@alpha cascade execute \
  --strategy fail-fast \
  --error-action stop

Resilient Execution:

# Continues despite failures, retries failed workflows
npx claude-flow@alpha cascade execute \
  --strategy resilient \
  --retry-attempts 3 \
  --error-action continue

Phase 4: Monitor Progress

Objective

Track cascade execution with real-time monitoring and progress reporting.

Scripts

# Real-time monitoring
npx claude-flow@alpha cascade monitor \
  --interval 5 \
  --output cascade-monitor.log &

# Get progress summary
npx claude-flow@alpha cascade progress \
  --format json > progress.json

# Track workflow metrics
npx claude-flow@alpha cascade metrics \
  --per-workflow \
  --include-timing

# Detect bottlenecks
npx claude-flow@alpha cascade analyze \
  --detect-bottlenecks \
  --output analysis.json

# Generate progress report
npx claude-flow@alpha cascade report \
  --type progress \
  --include-timeline \
  --output progress-report.md

Progress Visualization

# Generate Gantt chart
npx claude-flow@alpha cascade gantt \
  --output cascade-gantt.png

# Generate flow diagram
npx claude-flow@alpha cascade flow-diagram \
  --show-current-state \
  --output cascade-flow.png

Phase 5: Optimize Flow

Objective

Optimize cascade performance through parallel execution and resource allocation.

Scripts

# Analyze cascade performance
npx claude-flow@alpha cascade analyze \
  --identify-parallelism \
  --output optimization-recommendations.json

# Apply optimizations
npx claude-flow@alpha cascade optimize \
  --enable-parallelism \
  --balance-load \
  --cache-results

# Rebalance workflows
npx claude-flow@alpha cascade rebalance \
  --strategy adaptive \
  --target-utilization 0.8

# Measure improvements
npx claude-flow@alpha cascade compare \
  --baseline baseline-metrics.json \
  --current current-metrics.json \
  --output improvements.json

# Generate optimization report
npx claude-flow@alpha cascade report \
  --type optimization \
  --show-improvements \
  --output optimization-report.md

Optimization Techniques

Parallel Execution:

  • Identify independent workflows
  • Execute simultaneously
  • Reduce total execution time

Workflow Caching:

  • Cache workflow results
  • Reuse for subsequent executions
  • Skip unnecessary recomputation

Resource Pooling:

  • Share resources across workflows
  • Reduce overhead
  • Improve efficiency

Smart Scheduling:

  • Schedule workflows based on dependencies
  • Optimize resource utilization
  • Minimize idle time

Success Criteria

  • Cascade designed with clear structure
  • Workflows chained correctly
  • Execution proceeding smoothly
  • Progress tracked continuously
  • Flow optimized for performance

Performance Targets

  • Cascade completion time: Within estimates ±15%
  • Workflow transition time: <5 seconds
  • Resource utilization: 70-85%
  • Error recovery time: <30 seconds
  • Overall success rate: >95%

Best Practices

  1. Clear Dependencies: Explicitly define workflow dependencies
  2. Error Handling: Implement retry and fallback mechanisms
  3. Data Validation: Validate data at workflow boundaries
  4. Progress Tracking: Monitor execution continuously
  5. Resource Management: Allocate resources efficiently
  6. State Management: Use shared memory for state
  7. Conditional Logic: Test all branches
  8. Performance Monitoring: Track metrics throughout

Common Issues & Solutions

Issue: Workflow Deadlock

Symptoms: Workflows waiting indefinitely Solution: Check for circular dependencies, implement timeout

Issue: Data Loss Between Workflows

Symptoms: Missing data in downstream workflows Solution: Verify data flow configuration, use persistent storage

Issue: Performance Degradation

Symptoms: Slow execution, increasing latency Solution: Analyze bottlenecks, optimize critical workflows

MCP Requirements

This skill requires the following MCP servers for optimal functionality:

ruv-swarm (25k tokens)

Purpose: Multi-agent swarm coordination for workflow execution

Tools Used:

  • mcp__ruv-swarm__swarm_init: Initialize swarm topology for cascade
  • mcp__ruv-swarm__task_orchestrate: Orchestrate tasks across cascade stages
  • mcp__ruv-swarm__task_status: Monitor cascade execution progress

Activation (PowerShell):

# Check if already active
claude mcp list

# Add if not present
claude mcp add ruv-swarm npx ruv-swarm mcp start

Usage Example:

// Initialize hierarchical swarm for cascade
mcp__ruv-swarm__swarm_init({
  topology: "hierarchical",
  maxAgents: 10,
  strategy: "adaptive"
})

// Orchestrate cascade task
mcp__ruv-swarm__task_orchestrate({
  task: "Execute design workflow, then backend workflow when complete",
  strategy: "sequential",
  priority: "high"
})

// Monitor progress
mcp__ruv-swarm__task_status({ detailed: true })

Token Cost: 25k tokens (12.5% of 200k context) When to Load: When coordinating multi-agent workflows

Flow Nexus (52k tokens)

Purpose: Cloud workflow orchestration with event-driven execution

Tools Used:

  • mcp__flow-nexus__workflow_create: Create cascade workflows
  • mcp__flow-nexus__workflow_execute: Execute workflows with dependencies

Activation (PowerShell):

# Check if already active
claude mcp list

# Add if not present
claude mcp add flow-nexus npx flow-nexus@latest mcp start

Usage Example:

// Create cascade workflow
mcp__flow-nexus__workflow_create({
  name: "backend-frontend-cascade",
  steps: [
    { name: "backend", agent_type: "backend-dev" },
    { name: "frontend", agent_type: "coder", depends_on: ["backend"] }
  ]
})

Token Cost: 52k tokens (26% of 200k context) When to Load: When using cloud-based workflow execution

Integration Points

  • swarm-orchestration: For agent coordination
  • pipeline-chaining: For data flow
  • performance-analysis: For optimization insights

References

  • Workflow Orchestration Patterns
  • Cascade Design Principles
  • Conditional Execution Strategies
  • Performance Optimization Techniques

Core Principles

1. Dependency-First Design

Principle: Always identify dependencies before execution order. Sequential tasks must complete before dependent tasks begin, while independent tasks should execute in parallel.

In practice:

  • Map workflow dependencies as a directed acyclic graph (DAG) before execution
  • Sequential cascades chain workflows where output of one feeds input of next
  • Parallel cascades spawn independent workflows concurrently for maximum throughput
  • Use conditional branching only when decision logic requires runtime evaluation
  • Never create circular dependencies - validate DAG before execution

2. Fail-Fast with Graceful Recovery

Principle: Detect failures immediately and provide multiple recovery strategies based on failure severity and workflow context.

In practice:

  • Implement timeout mechanisms for each workflow node to prevent indefinite blocking
  • Use retry logic with exponential backoff for transient failures (network, rate limits)
  • Provide fallback workflows for critical path failures (circuit breaker pattern)
  • Store intermediate state in memory coordination layer for recovery after crashes
  • Log failure context (stack trace, input data, workflow state) for post-mortem analysis

3. Observability Throughout Execution

Principle: Every workflow execution must be fully traceable, monitorable, and debuggable from start to finish.

In practice:

  • Emit structured logs at workflow boundaries (start, complete, error, transition)
  • Track performance metrics per workflow node (execution time, resource usage, throughput)
  • Store workflow execution history in memory for audit trails and replay
  • Generate real-time progress reports showing current state and completion percentage
  • Create visual workflow diagrams (Gantt charts, flow diagrams) updated during execution

Anti-Patterns

Anti-Pattern Problem Solution
Monolithic Workflows Single massive workflow with no decomposition leads to poor maintainability, difficult debugging, and inability to parallelize independent tasks. Break workflows into smaller, composable units. Use hierarchical cascades where high-level workflows orchestrate sub-workflows. Each workflow should have single responsibility.
Implicit Dependencies Workflows depend on undocumented side effects or shared global state, causing unexpected failures when execution order changes. Explicitly declare all dependencies in cascade definition. Use data flow configuration to pass outputs as inputs. Avoid shared mutable state - use message passing instead.
Ignoring Workflow Failures Cascade continues executing dependent workflows even when prerequisites fail, wasting resources and producing invalid results. Implement strict dependency checking. Use fail-fast strategy where failure of prerequisite immediately halts dependent workflows. Provide rollback or compensation logic for partial failures.

Common Anti-Patterns

Anti-Pattern Problem Solution
Monolithic Workflows Single massive workflow with no decomposition leads to poor maintainability, difficult debugging, and inability to parallelize independent tasks. Break workflows into smaller, composable units. Use hierarchical cascades where high-level workflows orchestrate sub-workflows. Each workflow should have single responsibility.
Implicit Dependencies Workflows depend on undocumented side effects or shared global state, causing unexpected failures when execution order changes. Explicitly declare all dependencies in cascade definition. Use data flow configuration to pass outputs as inputs. Avoid shared mutable state - use message passing instead.
Ignoring Workflow Failures Cascade continues executing dependent workflows even when prerequisites fail, wasting resources and producing invalid results. Implement strict dependency checking. Use fail-fast strategy where failure of prerequisite immediately halts dependent workflows. Provide rollback or compensation logic for partial failures.

Conclusion

The Cascade Orchestrator enables sophisticated workflow automation by combining sequential pipelines, parallel execution, and conditional branching into a unified coordination framework. Its power lies in three key capabilities: intelligent dependency management that maximizes parallelization while respecting execution constraints, robust failure handling that recovers gracefully from transient errors, and comprehensive observability that provides full visibility into complex multi-stage workflows.

When implemented correctly, cascade workflows dramatically improve system throughput by executing independent tasks concurrently while maintaining correctness through explicit dependency tracking. The memory coordination layer serves as the backbone for state management, enabling workflows to share data seamlessly, recover from failures by resuming from checkpoints, and provide audit trails for compliance and debugging. Performance optimization techniques like caching, resource pooling, and smart scheduling ensure cascades scale efficiently even with dozens of interconnected workflows.

The most critical success factor is designing workflows with clear separation of concerns - each workflow should focus on a single responsibility and communicate through well-defined interfaces rather than implicit side effects. By following the core principles of dependency-first design, fail-fast recovery, and observability throughout execution, teams can build reliable, maintainable workflow systems that handle complex business logic with confidence. The anti-patterns section serves as a reminder that workflow orchestration requires careful attention to failure modes, resource constraints, and state management - shortcuts in these areas inevitably lead to brittle, unreliable systems that fail under production workloads.