RFC: Improve the performance and usability of linear algebra on CUDA devices #78581
Labels
module: cuda
Related to torch.cuda, and CUDA support in general
module: linear algebra
Issues related to specialized linear algebra operations in PyTorch; includes matrix multiply matmul
module: magma
related to magma linear algebra cuda support
triaged
This issue has been looked at a team member, and triaged and prioritized into an appropriate module
馃殌 The feature, motivation and pitch
Currently, the
torch.linalg
(https://pytorch.org/docs/stable/linalg.html) package provides linear algebra functionalities in pytorch. The CUDA backend is supported by cuSOLVER and MAGMA libraries.For now, linear algebra operators in pytorch are implemented in either cuSOLVER or MAGMA, or both. Users can use
to prefer one of the two backends. Available options (python
str
) aredefault
(using heuristics),cusolver
, ormagma
. See doc for details https://pytorch.org/docs/stable/backends.html#torch.backends.cuda.preferred_linalg_library.However, libraries have limitations and heuristics can't be perfect on all devices, library versions, input batch sizes, and input shapes. We'd like to collect user feedbacks and feature requests on the performance and usability of pytorch linear algebra on CUDA devices. Please leave a comment if you have any suggestions. Thank you!
Alternatives
No response
Additional context
No response
cc @jianyuh @nikitaved @pearu @mruberry @walterddr @IvanYashchuk @xwang233 @lezcano @ptrblck @ngimel
The text was updated successfully, but these errors were encountered: