[CUDA][cuBLASLt] Fix scale setting for allowFP16AccumulationCuBLAS true case#153083
[CUDA][cuBLASLt] Fix scale setting for allowFP16AccumulationCuBLAS true case#153083eqy wants to merge 3 commits intopytorch:mainfrom
allowFP16AccumulationCuBLAS true case#153083Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/153083
Note: Links to docs will display an error until the docs builds have been completed. ✅ You can merge normally! (1 Unrelated Failure)As of commit 167c4a6 with merge base 590965f ( BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
| cudaDeviceProp* prop = at::cuda::getCurrentDeviceProperties(); | ||
| if (prop->major >= 7 && at::globalContext().allowFP16AccumulationCuBLAS()) { | ||
| computeType = CUBLAS_COMPUTE_16F; | ||
| scaleType = CUDA_R_16F; |
|
@pytorchmergebot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Merge failedReason: Command Details for Dev Infra teamRaised by workflow job |
|
@pytorchmergebot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Also add some missing
@onlyCUDA/ support check decorators intest_matmul_cuda.pyShould help resolve #151890
cc @ptrblck @msaroufim @jerryzh168 @csarofeen @xwang233