[CUDA_FUSER] Expand operation support for cuda fuser#37849
Closed
jjsjann123 wants to merge 62 commits intopytorch:masterfrom
Closed
[CUDA_FUSER] Expand operation support for cuda fuser#37849jjsjann123 wants to merge 62 commits intopytorch:masterfrom
jjsjann123 wants to merge 62 commits intopytorch:masterfrom
Conversation
…our op declaration / or revisit type promotion
…ul, add_alpha, and sub_alpha.
… that Christian is working on a fix
…as to be solved either by defining them all internal to the fuser or somehow including a CUDA path.
Collaborator
Author
|
Windows test seems to be be jammed... I clicked the rerun that somehow spins up the passed tests 🤦 |
csarofeen
approved these changes
May 7, 2020
Contributor
facebook-github-bot
left a comment
There was a problem hiding this comment.
@soumith has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.
Contributor
glaringlee
pushed a commit
that referenced
this pull request
May 27, 2020
This is to reland #38675, and test cpp_extension compatible in _test only, this is enough, the purpose of this test is to make sure pytorch and cpp extension are compatible with xenial + cuda 9.2 + gcc 5.4 There are two non gcc5.4 (+ cuda9.2) compatible change introduced recently: #37849 #38627 which caused the following problems: https://app.circleci.com/pipelines/github/pytorch/pytorch/173756/workflows/7445e169-9c26-4ec4-a23a-ff6160d155b1/jobs/5582207/steps https://app.circleci.com/pipelines/github/pytorch/pytorch/173970/workflows/bf0de0f2-9156-4c8f-a097-53ca8e20d4b0/jobs/5589265/steps The root cause is that gcc 5.4 does not support uniform initialization list well, it can not deduce a correct type in some cases. It probably bugs in the gcc 5 compiler, I modified these code a little bit to make them compatible with cuda 9.2 + gcc 5.4. People are still using xenial + gcc5.4 + cuda 9.x, this env should be covered until xenial is deprecated. Differential Revision: [D21731026](https://our.internmc.facebook.com/intern/diff/D21731026) [ghstack-poisoned]
glaringlee
pushed a commit
that referenced
this pull request
May 27, 2020
This is to reland #38675, and test cpp_extension compatible in _test only, this is enough, the purpose of this test is to make sure pytorch and cpp extension are compatible with xenial + cuda 9.2 + gcc 5.4 There are two non gcc5.4 (+ cuda9.2) compatible change introduced recently: #37849 #38627 which caused the following problems: https://app.circleci.com/pipelines/github/pytorch/pytorch/173756/workflows/7445e169-9c26-4ec4-a23a-ff6160d155b1/jobs/5582207/steps https://app.circleci.com/pipelines/github/pytorch/pytorch/173970/workflows/bf0de0f2-9156-4c8f-a097-53ca8e20d4b0/jobs/5589265/steps The root cause is that gcc 5.4 does not support uniform initialization list well, it can not deduce a correct type in some cases. It probably bugs in the gcc 5 compiler, I modified these code a little bit to make them compatible with cuda 9.2 + gcc 5.4. People are still using xenial + gcc5.4 + cuda 9.x, this env should be covered until xenial is deprecated. Differential Revision: [D21731026](https://our.internmc.facebook.com/intern/diff/D21731026) [ghstack-poisoned]
glaringlee
pushed a commit
that referenced
this pull request
May 28, 2020
This is to reland #38675, and test cpp_extension compatible in _test only, this is enough, the purpose of this test is to make sure pytorch and cpp extension are compatible with xenial + cuda 9.2 + gcc 5.4 There are two non gcc5.4 (+ cuda9.2) compatible change introduced recently: #37849 #38627 which caused the following problems: https://app.circleci.com/pipelines/github/pytorch/pytorch/173756/workflows/7445e169-9c26-4ec4-a23a-ff6160d155b1/jobs/5582207/steps https://app.circleci.com/pipelines/github/pytorch/pytorch/173970/workflows/bf0de0f2-9156-4c8f-a097-53ca8e20d4b0/jobs/5589265/steps The root cause is that gcc 5.4 does not support uniform initialization list well, it can not deduce a correct type in some cases. It probably bugs in the gcc 5 compiler, I modified these code a little bit to make them compatible with cuda 9.2 + gcc 5.4. People are still using xenial + gcc5.4 + cuda 9.x, this env should be covered until xenial is deprecated. Differential Revision: [D21731026](https://our.internmc.facebook.com/intern/diff/D21731026) [ghstack-poisoned]
glaringlee
pushed a commit
that referenced
this pull request
May 28, 2020
This is to reland #38675, and test cpp_extension compatibility in _test only, this is enough, the purpose of this test is to make sure pytorch and cpp extension are compatible with xenial + cuda 9.2 + gcc 5.4 There are two non gcc5.4 (+ cuda9.2) compatible change introduced recently: #37849 #38627 which caused the following problems: https://app.circleci.com/pipelines/github/pytorch/pytorch/173756/workflows/7445e169-9c26-4ec4-a23a-ff6160d155b1/jobs/5582207/steps https://app.circleci.com/pipelines/github/pytorch/pytorch/173970/workflows/bf0de0f2-9156-4c8f-a097-53ca8e20d4b0/jobs/5589265/steps The root cause is that gcc 5.4 does not support uniform initialization list well, it can not deduce a correct type in some cases. It probably bugs in the gcc 5 compiler, I modified these code a little bit to make them compatible with cuda 9.2 + gcc 5.4. People are still using xenial + gcc5.4 + cuda 9.x, this env should be covered until xenial is deprecated. Differential Revision: [D21731026](https://our.internmc.facebook.com/intern/diff/D21731026) [ghstack-poisoned]
glaringlee
pushed a commit
that referenced
this pull request
May 28, 2020
This is to reland #38675, and test cpp_extension compatibility in _test only, this is enough, the purpose of this test is to make sure pytorch and cpp extension are compatible with xenial + cuda 9.2 + gcc 5.4 There are two non gcc5.4 (+ cuda9.2) compatible change introduced recently: #37849 #38627 which caused the following problems: https://app.circleci.com/pipelines/github/pytorch/pytorch/173756/workflows/7445e169-9c26-4ec4-a23a-ff6160d155b1/jobs/5582207/steps https://app.circleci.com/pipelines/github/pytorch/pytorch/173970/workflows/bf0de0f2-9156-4c8f-a097-53ca8e20d4b0/jobs/5589265/steps The root cause is that gcc 5.4 does not support uniform initialization list well, it can not deduce a correct type in some cases. It probably bugs in the gcc 5 compiler, I modified these code a little bit to make them compatible with cuda 9.2 + gcc 5.4. People are still using xenial + gcc5.4 + cuda 9.x, this env should be covered until xenial is deprecated. Differential Revision: [D21731026](https://our.internmc.facebook.com/intern/diff/D21731026) [ghstack-poisoned]
jjsjann123
added a commit
to jjsjann123/nvfuser
that referenced
this pull request
Oct 29, 2022
Summary: This PR added more supported operations in CUDA fuser. We are covering major point-wise operations supported in legacy fuser. In an attempt to adapt to legacy executor: 1. added an naive shape propagation pass on pytorch JIT IR; 2. small refactor on graph partitioning; 3. fallback interpreter execution of fusion group; Pull Request resolved: pytorch/pytorch#37849 Reviewed By: yf225 Differential Revision: D21444320 Pulled By: soumith fbshipit-source-id: 712e18ab8497f8d58a07e6f8d200cdab52cf0d74
jjsjann123
added a commit
to jjsjann123/nvfuser
that referenced
this pull request
Nov 10, 2022
Summary: This PR added more supported operations in CUDA fuser. We are covering major point-wise operations supported in legacy fuser. In an attempt to adapt to legacy executor: 1. added an naive shape propagation pass on pytorch JIT IR; 2. small refactor on graph partitioning; 3. fallback interpreter execution of fusion group; Pull Request resolved: pytorch/pytorch#37849 Reviewed By: yf225 Differential Revision: D21444320 Pulled By: soumith fbshipit-source-id: 712e18ab8497f8d58a07e6f8d200cdab52cf0d74
laurentdupin
pushed a commit
to laurentdupin/pytorch
that referenced
this pull request
Apr 24, 2026
Summary: This PR added more supported operations in CUDA fuser. We are covering major point-wise operations supported in legacy fuser. In an attempt to adapt to legacy executor: 1. added an naive shape propagation pass on pytorch JIT IR; 2. small refactor on graph partitioning; 3. fallback interpreter execution of fusion group; Pull Request resolved: pytorch#37849 Reviewed By: yf225 Differential Revision: D21444320 Pulled By: soumith fbshipit-source-id: 712e18ab8497f8d58a07e6f8d200cdab52cf0d74
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
This PR added more supported operations in CUDA fuser. We are covering major point-wise operations supported in legacy fuser.
In an attempt to adapt to legacy executor: