-
Notifications
You must be signed in to change notification settings - Fork 21.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.
Already on GitHub? Sign in to your account
BFloat16 on cuda: add triu/tril support #101932
Comments
Oh.. Fixed by merged #101414 : if I try my example in freshly build docker container, it will work >>> torch.__version__
'2.1.0a0+git22ca1a1'
>>> torch.arange(4).reshape(2,2).bfloat16().cuda().triu()
tensor([[0., 1.],
[0., 3.]], device='cuda:0', dtype=torch.bfloat16) |
Very nice idea, but are there any potential problems with this implementation? @Maykeye |
So |
How to install |
馃殌 The feature, motivation and pitch
Right now if you try to use torch.triu on bfloat16 tensor (I hit it when was training simple network with AMP) you'll get a error that triu is not support
It would be nice to have it.
Alternatives
Can be replaced with multiplication against torch.ones.triu()
Additional context
No response
The text was updated successfully, but these errors were encountered: