Skip to content

compare the theory attention gradient with PyTorch attention gradient

Notifications You must be signed in to change notification settings

Say-Hello2y/Transformer-attention

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Transformer-attention

The full derivation of Transformer gradient. compare the theory attention gradient with PyTorch attention gradient

  • If you want see the detail calcualtion,please see CN,EN

Citation

If you find this open source release useful, please cite in your paper:

@software{He_The_full_derivation_2022,
author = {He, Longxiang},
month = may,
title = {{The full derivation of Transformer gradient}},
url = {https://github.com/Say-Hello2y/Transformer-attention.git},
version = {0.0.0},
year = {2022}
}

About

compare the theory attention gradient with PyTorch attention gradient

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages