[cudnn] Support v8 API in fbcode#96512
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/96512
Note: Links to docs will display an error until the docs builds have been completed. ❌ 1 FailuresAs of commit 3862dc1: NEW FAILURES - The following jobs have failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
This pull request was exported from Phabricator. Differential Revision: D43784279 |
|
This pull request was exported from Phabricator. Differential Revision: D43784279 |
1b12f00 to
88103da
Compare
malfet
left a comment
There was a problem hiding this comment.
LGTM, though I'm surprised why size_t cast is needed
There was a problem hiding this comment.
Why this change is needed? If it is, then please use static_cast<size_t>(plan.getWorkspaceSize())
| if ((size_t) plan.getWorkspaceSize() <= max_workspace_size) { | |
| if (plan.getWorkspaceSize() <= max_workspace_size) { |
There was a problem hiding this comment.
That's our favorite signed-unsigned comparison probably, can we enable compilation flags to error on it in OSS builds? It's been a nuisance.
There was a problem hiding this comment.
Given that getWorkspaceSize() returns int64_t in cudnn frontend, a better fix would be to change our vars like max_worksplace_size and curr_workspace_size to int64_t instead of size_t
There was a problem hiding this comment.
sounds good! I'll make the change and see if I can make this from warning to error in oss :)
|
This pull request was exported from Phabricator. Differential Revision: D43784279 |
88103da to
e42ab56
Compare
Summary: As discussed in pytorch#96512, turn on sign-compare for OSS build Test Plan: pytorch CI Differential Revision: D44085536 fbshipit-source-id: d2093131d84230aed316f783198f9229f2a773dc
e42ab56 to
dcb3435
Compare
|
This pull request was exported from Phabricator. Differential Revision: D43784279 |
|
Looks this PR has a conflict with #96723, will probably wait a bit to get in sync |
|
This pull request was exported from Phabricator. Differential Revision: D43784279 |
dcb3435 to
a692d89
Compare
Summary: Pull Request resolved: pytorch#96512 It turns out we never turn on cudnn v8 API which blocks bf16 conv. Enable the new v8 API Test Plan: buck run mode/dev-nosan scripts/xdwang/example:fc_pytorch Reviewed By: ngimel Differential Revision: D43784279 fbshipit-source-id: 902a4e162807faae874cc9c4baaa90479cd72006
|
This pull request was exported from Phabricator. Differential Revision: D43784279 |
a692d89 to
3862dc1
Compare
|
@pytorchbot merge (Initiating merge automatically since Phabricator Diff has merged) |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Summary: It turns out we never turn on cudnn v8 API which blocks bf16 conv. Enable the new v8 API Test Plan: buck run mode/dev-nosan scripts/xdwang/example:fc_pytorch Reviewed By: ngimel Differential Revision: D43784279 Pull Request resolved: pytorch/pytorch#96512 Approved by: https://github.com/malfet
Summary: It turns out we never turn on cudnn v8 API which blocks bf16 conv. Enable the new v8 API Test Plan: buck run mode/dev-nosan scripts/xdwang/example:fc_pytorch Reviewed By: ngimel Differential Revision: D43784279 Pull Request resolved: pytorch/pytorch#96512 Approved by: https://github.com/malfet
Summary: It turns out we never turn on cudnn v8 API which blocks bf16 conv. Enable the new v8 API
Test Plan: buck run mode/dev-nosan scripts/xdwang/example:fc_pytorch
Reviewed By: ngimel
Differential Revision: D43784279