Claude Code Plugins

Community-maintained marketplace

Feedback

gpu-quality-priority

@smith6jt-cop/Skills_Registry
0
0

KINTSUGI processing principles: Never sacrifice quality for speed, always use GPU when available. Trigger: performance optimization, CPU/GPU choice, fast mode, quality vs speed.

Install Skill

1Download skill
2Enable skills in Claude

Open claude.ai/settings/capabilities and find the "Skills" section

3Upload to Claude

Click "Upload skill" and select the downloaded ZIP file

Note: Please verify skill by going through its instructions before using it.

SKILL.md

name gpu-quality-priority
description KINTSUGI processing principles: Never sacrifice quality for speed, always use GPU when available. Trigger: performance optimization, CPU/GPU choice, fast mode, quality vs speed.
author KINTSUGI Team
date Sun Dec 14 2025 00:00:00 GMT+0000 (Coordinated Universal Time)

GPU-Only and Quality-First Processing Principles

Experiment Overview

Item Details
Date 2025-12-14
Goal Establish processing principles for KINTSUGI batch processing
Environment HiPerGator, multi-GPU (NVIDIA), CuPy, KINTSUGI pipeline
Status Policy Established

Context

During performance optimization of Notebook 2 (Cycle Processing), a "fast mode" was proposed that would reduce BaSiC iteration parameters to speed up processing. The user explicitly rejected this approach, establishing core principles for KINTSUGI processing.

Scientific imaging requires quality-first processing. Unlike consumer applications where "good enough" may be acceptable, multiplex immunofluorescence analysis depends on accurate quantification. Quality degradation compounds through the pipeline: illumination correction errors affect stitching, which affects deconvolution, which affects segmentation, which affects all downstream analysis.

Core Principles

1. NEVER Sacrifice Quality for Speed

Quality parameters must remain at their scientifically-validated defaults unless the quality impact is negligible (verified, not assumed).

# CORRECT: Quality parameters (do not reduce)
BASIC_IF_DARKFIELD = True
BASIC_MAX_ITERATIONS = 500
BASIC_OPTIMIZATION_TOLERANCE = 1e-6
BASIC_MAX_REWEIGHT_ITERATIONS = 25
BASIC_REWEIGHT_TOLERANCE = 1e-3

2. ALWAYS Use GPU When Available - No CPU Fallback

If a GPU is available, it must be used. CPU fallback options should be disabled or removed.

# CORRECT: GPU enforcement
if not USE_GPU:
    raise RuntimeError(
        "GPU not available but required for processing.\n"
        "Check GPU status with: from kintsugi.gpu import get_gpu_manager; "
        "print(get_gpu_manager().summary())"
    )
use_gpu = True  # Always True - GPU required

3. Remove CPU Options When GPU Exists

Don't provide use_cpu or use_gpu=False options. If the system has a GPU, use it.

# WRONG: Providing CPU option
def process(use_gpu=True):  # Allows use_gpu=False
    ...

# CORRECT: GPU-only
def process(device_id=None):  # GPU assumed, only device selection
    if device_id is None:
        device_id = GPU_DEVICE_IDS[0]
    ...

Failed Attempts (Critical)

Attempt Why it Failed Lesson Learned
Added BASIC_FAST_MODE with reduced iterations (200/10) User rejected - quality is non-negotiable Never propose quality/speed tradeoffs without explicit request
Added use_gpu_basic=True/False parameter Creates temptation to use CPU Remove CPU options entirely when GPU is available
Proposed "fast mode for testing" Testing should use production parameters If testing finds issues, they should be found with real parameters
Suggested relaxed tolerances (1e-5, 1e-2) Even "slightly" relaxed tolerances compound errors Keep validated parameters exactly as specified

Acceptable Optimizations

These optimizations improve speed WITHOUT sacrificing quality:

Optimization Impact Safe?
Parallel image loading (ThreadPoolExecutor) 10-20x faster I/O YES - same data, faster loading
Parallel image resizing 10-20x faster preprocessing YES - same resize algorithm
GPU-accelerated computation 10-50x faster YES - same algorithm, faster hardware
Multi-GPU parallelism Linear scaling YES - same computation, more hardware
Optimized DCT (dctn vs sequential dct) 2-3x faster YES - mathematically equivalent
Power iteration for SVD 10x faster YES - sufficient for top singular value

Key Insights

  • Quality is non-negotiable - Scientific imaging requires accurate quantification
  • Speed comes from better hardware, not shortcuts - Invest in GPUs, not reduced iterations
  • Errors compound - A 5% error in illumination correction becomes 10%+ by segmentation
  • "Fast mode for testing" is a trap - Test with production parameters or you'll miss production issues
  • CPU fallback is never needed - If no GPU, the user should know immediately, not get silent degradation

Implementation Pattern

# GPU enforcement at module level
if not USE_GPU:
    raise RuntimeError("GPU required for KINTSUGI processing")

# Function signatures - no CPU options
def process_zplane(
    ...,
    device_id: int = None,  # Which GPU, not whether to use GPU
):
    """GPU is REQUIRED - no CPU fallback."""
    if device_id is None:
        device_id = GPU_DEVICE_IDS[0]

    # Use validated quality parameters
    corrector = KCorrectGPU(use_gpu=True, device_id=device_id)
    flatfield, darkfield = corrector.fit(
        images,
        max_iterations=500,           # Quality parameter - DO NOT REDUCE
        max_reweight_iterations=25,   # Quality parameter - DO NOT REDUCE
        optimization_tolerance=1e-6,  # Quality parameter - DO NOT REDUCE
    )

References

  • KINTSUGI Notebook 2: Cycle Processing
  • BaSiC paper: Peng et al., Nature Communications 2017
  • Skills Registry: basic-caching-evaluation (another quality-compromising approach that failed)