ray-train
@ovachiever/droid-tings5
0
Distributed training orchestration across clusters. Scales PyTorch/TensorFlow/HuggingFace from laptop to 1000s of nodes. Built-in hyperparameter tuning with Ray Tune, fault tolerance, elastic scaling. Use when training massive models across multiple machines or running distributed hyperparameter sweeps.
Install Skill
1Download skill
2Enable skills in Claude
Open claude.ai/settings/capabilities and find the "Skills" section
3Upload to Claude
Click "Upload skill" and select the downloaded ZIP file
Note: Please verify skill by going through its instructions before using it.