Skip to content

nauyan/PyTorch-Distributed-Tutorials

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

PyTorch-Distributed-Tutorials

Detailed blog on various Distributed Training startegies can be read here.

To train standalone PyTorch script run:

python train.py

To train DataParallel PyTorch script run:

python train_dataparallel.py

To train DistributedDataParallel(DDP) PyTorch script run:

torchrun --nnodes=1 --nproc-per-node=4 train_ddp.py

To train FullyShardedDataParallel(FSDP) PyTorch script run:

torchrun --nnodes=1 --nproc-per-node=4 train_fsdp.py

About

Concise tutorials for distributed training using PyTorch

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages