Tune thread count for binning/grouping from scratch #3641
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
After #3638 reduced some baseline cost, increasing thread counts earlier pays off according to my local benchmarks (group two columns into 2M bins).
Results (blue: before #3638, orange: #3638, green: this)

I included
numpy.bincountfor comparison. It does something much simpler, but it nicely illustrates that our constant (in the number of rows) baseline cost is still quite substantial, significantly affecting timings up to 10M rows or so. I will have another look at this bit as a follow-up task.