Skip to content

Miscellaneous updates for CUDA 10#12017

Closed
syed-ahmed wants to merge 2 commits intopytorch:masterfrom
syed-ahmed:cuda-10-updates
Closed

Miscellaneous updates for CUDA 10#12017
syed-ahmed wants to merge 2 commits intopytorch:masterfrom
syed-ahmed:cuda-10-updates

Conversation

@syed-ahmed
Copy link
Collaborator

This PR has some updates related to CUDA 10.

syed-ahmed and others added 2 commits September 24, 2018 08:16
Co-authored-by: Christian Sarofeen <csarofeen@nvidia.com>
Copy link
Contributor

@facebook-github-bot facebook-github-bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

soumith is landing this pull request. If you are a Facebook employee, you can view this diff on Phabricator.

zdevito pushed a commit to zdevito/ATen that referenced this pull request Sep 24, 2018
Summary:
This PR has some updates related to CUDA 10.

- pytorch/pytorch@c2195e9 ensures that the repo successfully builts on CUDA 10. Addresses pytorch/pytorch#11888
- pytorch/pytorch@423d8d3 follows up on the cufft max plan number bug: pytorch/pytorch#11089, which has been fixed in CUDA 10.
Pull Request resolved: pytorch/pytorch#12017

Differential Revision: D10013405

Pulled By: soumith

fbshipit-source-id: 5bc6d7f71d5133f7821b407b1ac6c51bef0f6fa8
// bug related to cuFFT plan cache max size has been fixed
// in CUDA 10. Hence, when compiling with CUDA 10, just
// don't do the erase.
#if CUDA_VERSION < 10000

This comment was marked as off-topic.

This comment was marked as off-topic.

This comment was marked as off-topic.

@@ -346,6 +346,7 @@ class CuFFTConfig {
// be fine for now.
// TODO: When CUDA 10 comes out, check if the bug is fixed or if we need another

This comment was marked as off-topic.

This comment was marked as off-topic.

This comment was marked as off-topic.

facebook-github-bot pushed a commit that referenced this pull request Oct 12, 2018
Summary:
SsnL As per your review in #12017, I added a max plan number for CUDA 10 path. Our internal cuFFT team couldn't suggest a number since the limit depends on host/device memory. That is, a plan allocates some buffers on the device and also creates objects for the plans on the host side. I raised this number to 4x arbitrarily per you suggestion.
Pull Request resolved: #12553

Differential Revision: D10320832

Pulled By: SsnL

fbshipit-source-id: 3148d45cd280dffb2039756e2f6a74fbc7aa086d
zdevito pushed a commit to zdevito/ATen that referenced this pull request Oct 12, 2018
Summary:
SsnL As per your review in pytorch/pytorch#12017, I added a max plan number for CUDA 10 path. Our internal cuFFT team couldn't suggest a number since the limit depends on host/device memory. That is, a plan allocates some buffers on the device and also creates objects for the plans on the host side. I raised this number to 4x arbitrarily per you suggestion.
Pull Request resolved: pytorch/pytorch#12553

Differential Revision: D10320832

Pulled By: SsnL

fbshipit-source-id: 3148d45cd280dffb2039756e2f6a74fbc7aa086d
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants