Skip to content

[v0.3] Propagate volatile in zeros_like#3984

Merged
soumith merged 1 commit intopytorch:v0.3.0from
colesbury:double_backwards_leak
Dec 2, 2017
Merged

[v0.3] Propagate volatile in zeros_like#3984
soumith merged 1 commit intopytorch:v0.3.0from
colesbury:double_backwards_leak

Conversation

@colesbury
Copy link
Copy Markdown
Member

Gradients were becoming non-volatile because at::zeros_like returned a
Variable with volatile always set to false. The non-volatile gradients
accumulated history in the model which results in continuously
increasing memory usage,

See #3983, #3835, #3824

In v0.4 this will be more robustly solved by #3970

Gradients were becoming non-volatile because at::zeros_like returned a
Variable with volatile always set to false. The non-volatile gradients
accumulated history in the model which results in continuously
increasing memory usage,

See pytorch#3983, pytorch#3835, pytorch#3824

In v0.4 this will be more robustly solved by pytorch#3970
@colesbury colesbury requested a review from soumith December 2, 2017 22:14
@soumith soumith merged commit 4f90f78 into pytorch:v0.3.0 Dec 2, 2017
@colesbury colesbury deleted the double_backwards_leak branch December 2, 2017 23:27
peterjc123 pushed a commit to peterjc123/pytorch that referenced this pull request Dec 4, 2017
Gradients were becoming non-volatile because at::zeros_like returned a
Variable with volatile always set to false. The non-volatile gradients
accumulated history in the model which results in continuously
increasing memory usage,

See pytorch#3983, pytorch#3835, pytorch#3824

In v0.4 this will be more robustly solved by pytorch#3970
soumith pushed a commit that referenced this pull request Dec 4, 2017
Gradients were becoming non-volatile because at::zeros_like returned a
Variable with volatile always set to false. The non-volatile gradients
accumulated history in the model which results in continuously
increasing memory usage,

See #3983, #3835, #3824

In v0.4 this will be more robustly solved by #3970
@soumith soumith added the 0.3 label Feb 3, 2018
wuhuikx pushed a commit to wuhuikx/pytorch that referenced this pull request Jan 30, 2020
Gradients were becoming non-volatile because at::zeros_like returned a
Variable with volatile always set to false. The non-volatile gradients
accumulated history in the model which results in continuously
increasing memory usage,

See pytorch#3983, pytorch#3835, pytorch#3824

In v0.4 this will be more robustly solved by pytorch#3970
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants