Skip to content

Initial set of automated performance benchmarks (non-H2O) #191

@fjetter

Description

@fjetter

Over the past ~2 years we ran many manual benchmarks to confirm various improvements to our scheduling or stealing heuristics.

Recently, we started working on H20 benchmarks to have something to compare. The H20 benchmarks are however heavily biased towards shuffle operations since they are motivated by database benchmarks. While this use case is surely relevant, there are many dask-specific workloads these benchmarks do not cover properly.

AC


We should increase our benchmark coverage significantly to incorporate dask specific payload

Examples where this kind of automation would be helpful for future work or would've been helpful in the past. Some of these contain actual reproducers we could simply extract and adapt to Coiled

This would build up on the infrastructure created by #148

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions