Disable autocast cache in torch.cuda.make_graphed_callables#84289
Disable autocast cache in torch.cuda.make_graphed_callables#84289Aidyn-A wants to merge 3 commits intopytorch:masterfrom
Conversation
🔗 Helpful links
✅ No Failures (0 Pending)As of commit d8b7efd (more details on the Dr. CI page): Expand to see more💚 💚 Looks good so far! There are no failures yet. 💚 💚 This comment was automatically generated by Dr. CI (expand for details).Please report bugs/suggestions to the (internal) Dr. CI Users group. |
…graphed_callables
|
Autocast caching has been added for weight reuse in training, but I'm not against removing it. |
|
Yep, we decided to remove caching because we realized its potential danger in @pytorchbot merge |
|
@pytorchbot successfully started a merge job. Check the current status here. |
|
Hey @Aidyn-A. |
…84289) Summary: There there are conflicts between `torch.clear_autocast_cache()` and `cudaMallocAsync` from #82682. Moreover, the use of autocast caching is not reasonable during training which is the main target of `make_graphed_callables`. cc eqy ptrblck Pull Request resolved: #84289 Approved by: https://github.com/ngimel Test Plan: contbuild & OSS CI, see https://hud.pytorch.org/commit/pytorch/pytorch/ce1b727e774c75f8e31b28ff5915851385c70dcf Reviewed By: mehtanirav, izaitsevfb Differential Revision: D39277326 fbshipit-source-id: aaa15276397f082bdc8d8eab08b653eeeb7e8fb7
There there are conflicts between
torch.clear_autocast_cache()andcudaMallocAsyncfrom #82682.Moreover, the use of autocast caching is not reasonable during training which is the main target of
make_graphed_callables.cc @eqy @ptrblck