Welcome to Tensor Cloud
This documentation covers distributed deep learning training across multiple GPUs and nodes using the Tensor Cloud, and contains an introduction to distributed training with PyTorch.What You’ll Learn
This documentation is organized into two main sections:1. SF Tensor Library
Documentation for the SF Tensor Python library2. Distributed Training in PyTorch
A basic introduction to DDP and FSDP for training models in PyTorch across multiple GPUs.Quick Start
Install the SF Tensor library:Getting Started
SF Tensor Library
Learn the basics of the SF Tensor library and how it simplifies distributed training
Distributed Training
Understand DDP and FSDP for efficient multi-GPU training
Initialization Guide
Set up distributed training infrastructure automatically
DDP Training
Fast data-parallel training with DistributedDataParallel
FSDP Training
Train large models with FullyShardedDataParallel
Data Loading
Correctly load and distribute data across GPUs
Features
Simple Setup
Initialize distributed training with a single function call
Automatic Sync
Data download decorator prevents race conditions and duplicate downloads
Clean Logging
Rank-aware logging ensures messages print only once
Contents
SF Tensor Library Features
- ✅ One-line initialization -
sft.initialize_distributed_training() - ✅ Automatic device management -
sft.get_device() - ✅ Data downloading -
@dataDownloaddecorator - ✅ Persisted logging -
sft.log()andsft.logAccuracy()
Distributed Training guide
- ✅ DDP support - Fast training for standard models
- ✅ FSDP support - Memory-efficient training for large models
- ✅ DistributedSampler - Automatic data partitioning
- ✅ Mixed precision - FP16/BF16 for faster training
- ✅ Multi-node scaling - Training across multiple machines
Getting Started
Choose your starting point based on your needs:- New to distributed training? Start with SF Tensor Library Overview
- Ready to train models? Jump to DDP Training
- Training large models? Check out FSDP Training
- Need data loading help? See Data Loading Guide