Experiments with low level communication patterns that are useful for distributed training.
-
Updated
Nov 14, 2018 - Python
Experiments with low level communication patterns that are useful for distributed training.
An open collection of implementation tips, tricks and resources for training large language models
Distributed deep learning framework based on pytorch/numba/nccl and zeromq.
Python Distributed Non Negative Matrix Factorization with custom clustering
Federated Learning Utilities and Tools for Experimentation
An open collection of methodologies to help with successful training of large language models.
Distributed and decentralized training framework for PyTorch over graph
Best practices & guides on how to write distributed pytorch training code
Add a description, image, and links to the nccl topic page so that developers can more easily learn about it.
To associate your repository with the nccl topic, visit your repo's landing page and select "manage topics."