-
Notifications
You must be signed in to change notification settings - Fork 27.4k
[CUDA][CUDA 12] CUDA 12 Support Tracking Issue #91122
Copy link
Copy link
Closed
Labels
featureA request for a proper, new feature.A request for a proper, new feature.module: cudaRelated to torch.cuda, and CUDA support in generalRelated to torch.cuda, and CUDA support in generaltriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate moduleThis issue has been looked at a team member, and triaged and prioritized into an appropriate module
Metadata
Metadata
Assignees
Labels
featureA request for a proper, new feature.A request for a proper, new feature.module: cudaRelated to torch.cuda, and CUDA support in generalRelated to torch.cuda, and CUDA support in generaltriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate moduleThis issue has been looked at a team member, and triaged and prioritized into an appropriate module
CUDA 12 has been released, but we've identified several blocking issues (some code/API compatibility, some API functionality) that need to be addressed before a PyTorch + CUDA 12 build/environment could be considered usable by mainstream users. We're creating this issue to hopefully avoid duplicating work in identifying and resolving issues.
Required code changes for build-breaking changes:
cudaGraphInstantiateas the API has changed ([CUDA12] Make PyTorch compatible with CUDA 12 #91118)cudaProfilerInitialize([CUDA12] Make PyTorch compatible with CUDA 12 #91118, Follow up on CUDA 12 support for PyTorch/Caffe2 #95582)constCUDASparsedescriptors (Update cuSPARSE usage for CUDA 12.0 #90765, [CUDA 12] Fix the endif guard position for cusparse const descriptors #90897, [CUDA12] Clean up deprecated APIs #91050)Known API functionality issues:
cudaSetDeviceis currently eagerly creating a CUDA context when called, which is a significant departure from previous behavior. This potentially creates many unused contexts that waste GPU memory (especially in DDP workloads), and would either require framework or CUDA toolkit code changes to resolve. ([CUDA12] Conditionally set device in autograd engine #91191, [DO NOT MERGE] [CUDA12] Conditionally set device in device guard #91219, [CUDA12] set_device change #94864)Nightly CUDA 12.1 builds are now available for the adventurous:
@ https://pytorch.org/get-started/locally/
e.g.,
CC @ptrblck @crcrpar @IvanYashchuk @xwang233 @Aidyn-A @Fuzzkatt @syed-ahmed @puririshi98 @ngimel @malfet
Please feel free to update this issue as new items are discovered/resolved.
cc @ngimel