Spilling on demand#756
Conversation
|
The failure is an OOM in the new |
181e1e5 to
b858620
Compare
Codecov Report
@@ Coverage Diff @@
## branch-21.12 #756 +/- ##
===============================================
Coverage ? 58.20%
===============================================
Files ? 21
Lines ? 2991
Branches ? 0
===============================================
Hits ? 1741
Misses ? 1250
Partials ? 0 Continue to review full report at Codecov.
|
41d8c19 to
7146d26
Compare
2ab25da to
a4d2b6f
Compare
|
I tested this on the spilling related workflow mentioned here and was able to increase the memory limit from CC: @randerzander |
In principle yes, a memory limit of |
We get similar performence in this case on PR and main. Agreed that PR itself does not cause a slowdown.
I did two runs so not a comprehensive test by any means. |
|
@gpucibot merge |
|
Thanks Mads for the PR! Also Peter and Vibhu for the review and testing 😄 |
|
Excited to try this out! |
|
Nightlies are up. So should be good to go Even 😀 |
Use rapidsai/rmm#892 to implement spilling on demand. Requires use of RMM and JIT-unspill enabled.
The
device_memory_limitstill works as usual -- when known allocations gets todevice_memory_limit, Dask-CUDA starts spilling preemptively. However, with this PR it is should be possible to increasedevice_memory_limitsignificantly since memory spikes will be handled by spilling on demand.Closes #755