Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add double backward for LeakyReLU #1714

Merged
merged 1 commit into from
Jun 5, 2017

Conversation

caogang
Copy link
Contributor

@caogang caogang commented Jun 4, 2017

  • add double backward for LeakyReLU temporally (Future should be removed once Thnn Variable is enabled)

@apaszke
Copy link
Contributor

apaszke commented Jun 4, 2017

This is not a correct fix for ReLU. It's just a workaround for a bug in comparison functions, and this is what should be fixed. I already did it in my branch and I'll push it later today. Can you please revert this part?

@caogang caogang force-pushed the feature_leakyrelu_and_fixbug branch from bb89274 to b98f426 Compare June 4, 2017 12:01
@caogang
Copy link
Contributor Author

caogang commented Jun 4, 2017

Ok, I have reverted this part.

@caogang caogang changed the title Add double backward for LeakyReLU and fix bug of Threshold Add double backward for LeakyReLU Jun 4, 2017
@apaszke
Copy link
Contributor

apaszke commented Jun 4, 2017

@pytorchbot test this please

@soumith soumith merged commit 174c3cc into pytorch:master Jun 5, 2017
@caogang caogang deleted the feature_leakyrelu_and_fixbug branch June 8, 2017 08:10
houseroad added a commit to houseroad/pytorch that referenced this pull request Jan 19, 2019
…8bdbe7

Summary:
Previous import was fd60104394fa353e1762f44ecad1b2166e33deef

Included changes:
- **[c553fb3](onnx/onnx@c553fb3)**: Handle negative axis in scan shape inference (pytorch#1748) <G. Ramalingam>
- **[51b6ecc](onnx/onnx@51b6ecc)**: external_data: Store large tensor values in separate files (pytorch#678) <Michał Karzyński>
- **[ba05f26](onnx/onnx@ba05f26)**: Scan output axes (pytorch#1737) <G. Ramalingam>
- **[90920c0](onnx/onnx@90920c0)**: Add NonZero op. (pytorch#1714) <Sergii Dymchenko>
- **[c4cf112](onnx/onnx@c4cf112)**: fix the test cases for constantofshape (pytorch#1746) <Lu Fang>
- **[d902349](onnx/onnx@d902349)**: Add sample implementation support (pytorch#1712) <Lu Fang>

Differential Revision: D13745693

fbshipit-source-id: 057d827652e85ad19be8f0243d874e036bf69898
facebook-github-bot pushed a commit that referenced this pull request Jan 21, 2019
…8bdbe7 (#16190)

Summary:
Pull Request resolved: #16190

Previous import was fd60104394fa353e1762f44ecad1b2166e33deef

Included changes:
- **[c553fb3](onnx/onnx@c553fb3)**: Handle negative axis in scan shape inference (#1748) <G. Ramalingam>
- **[51b6ecc](onnx/onnx@51b6ecc)**: external_data: Store large tensor values in separate files (#678) <Michał Karzyński>
- **[ba05f26](onnx/onnx@ba05f26)**: Scan output axes (#1737) <G. Ramalingam>
- **[90920c0](onnx/onnx@90920c0)**: Add NonZero op. (#1714) <Sergii Dymchenko>
- **[c4cf112](onnx/onnx@c4cf112)**: fix the test cases for constantofshape (#1746) <Lu Fang>
- **[d902349](onnx/onnx@d902349)**: Add sample implementation support (#1712) <Lu Fang>

Differential Revision: D13745693

fbshipit-source-id: 05e2cce9ae1dfa2865db83840df64673d55cea57
jjsjann123 added a commit to jjsjann123/pytorch that referenced this pull request May 24, 2022
* Add a routine to query if a broadcast domain may be concretized to
multiple domains

* Don't group operations together that may have a broadcast that's being broadcasted to more than one size.

Co-authored-by: Christian Sarofeen <csarofeen@nvidia.com>
Co-authored-by: jjsjann123 <jiej@nvidia.com>
malfet pushed a commit that referenced this pull request Jun 8, 2022
Syncing nvfuser devel branch to upstream master. https://github.com/csarofeen/pytorch/

A few bigger updates:
1. Initial support of cp.async and cp.async.wait: csarofeen#1619
2. Emulate ampere's mma 16816 with Turing's mma 1688, for a unified interface: csarofeen#1643
3. Extending the infrastructure to support mma operators on turing and ampere arch: csarofeen#1440

Commits that's actually in this PR from the csarofeen branch
```
* dd23252 (csarofeen/devel) Fusion Segmenter: Unify single kernel and multi-kernel runtime path (#1710)
* b3d1c3f Fix missing cooperative launch (#1726)
* dc670a2 Async gmem copy support on sm80+ (#1619)
* 5e6a8da Add turing mma support and test (#1643)
* d6d6b7d Fix rFactor when there are indirect root domain(s), and refactor (#1723)
* 7093e39 Mma op integration on ampere (#1440)
* fade8da patch python test for bfloat16 (#1724)
* 8fbd0b1 Fine-grained kernel profiling (#1720)
* 77c1b4f Adding dry run mode to skip arch dependent checks (#1702)
* 151d95b More precise concretization analysis (#1719)
* f4d3630 Enable complex python tests (#1667)
* 4ceeee5 Minor bugfix in transform_rfactor.cpp (#1715)
* 3675c70 Separate root domain and rfactor domain in TransformPrinter (#1716)
* f68b830 Fix scheduling with polymorphic broadcast (#1714)
* 4ab5ef7 updating_ci_machine (#1718)
* 56585c5 Merge pull request #1711 from csarofeen/upstream_master_bump_0517
* 174d453 Allow using nvFuser on CUDA extension (#1701)
* 18bee67 Validate LOOP concrete IDs have complete IterDomains (#1676)
```
Pull Request resolved: #78244
Approved by: https://github.com/csarofeen, https://github.com/malfet
facebook-github-bot pushed a commit that referenced this pull request Jun 8, 2022
Summary:
Syncing nvfuser devel branch to upstream master. https://github.com/csarofeen/pytorch/

A few bigger updates:
1. Initial support of cp.async and cp.async.wait: csarofeen#1619
2. Emulate ampere's mma 16816 with Turing's mma 1688, for a unified interface: csarofeen#1643
3. Extending the infrastructure to support mma operators on turing and ampere arch: csarofeen#1440

Commits that's actually in this PR from the csarofeen branch
```
* dd23252 (csarofeen/devel) Fusion Segmenter: Unify single kernel and multi-kernel runtime path (#1710)
* b3d1c3f Fix missing cooperative launch (#1726)
* dc670a2 Async gmem copy support on sm80+ (#1619)
* 5e6a8da Add turing mma support and test (#1643)
* d6d6b7d Fix rFactor when there are indirect root domain(s), and refactor (#1723)
* 7093e39 Mma op integration on ampere (#1440)
* fade8da patch python test for bfloat16 (#1724)
* 8fbd0b1 Fine-grained kernel profiling (#1720)
* 77c1b4f Adding dry run mode to skip arch dependent checks (#1702)
* 151d95b More precise concretization analysis (#1719)
* f4d3630 Enable complex python tests (#1667)
* 4ceeee5 Minor bugfix in transform_rfactor.cpp (#1715)
* 3675c70 Separate root domain and rfactor domain in TransformPrinter (#1716)
* f68b830 Fix scheduling with polymorphic broadcast (#1714)
* 4ab5ef7 updating_ci_machine (#1718)
* 56585c5 Merge pull request #1711 from csarofeen/upstream_master_bump_0517
* 174d453 Allow using nvFuser on CUDA extension (#1701)
* 18bee67 Validate LOOP concrete IDs have complete IterDomains (#1676)
```

Pull Request resolved: #78244

Reviewed By: ejguan

Differential Revision: D36678948

Pulled By: davidberard98

fbshipit-source-id: 0ccde965acbd31da67d99c6adb2eaaa888948105
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants