DeepSpeed v14.0 for PyTorch 2.2.1 & Python 3.11
Before you install DeepSpeed, its recommended you confirm AllTalk works without.
This version has been built for PyTorch 2.2.1 and also Python 3.11.x
For CUDA v12.1 - WINDOWS - Download
For CUDA v11.8 - WINDOWS -Download
For CUDA v12.1 - LINUX- Download
For versions that support PyTorch 2.1.x please look at the main releases page
If you need to check your CUDA version within Text-generation-webui run cmd_windows.bat
and then: python --version
to get the Python version and pip show torch
to get the CUDA version.
NOTE: You DO NOT need to set Text-generation-webUI's --deepspeed setting for AllTalk to be able to use DeepSpeed. These are two completely separate things and incorrectly setting that on Text-generation-webUI may cause other complications.