Open
Description
Some benchmarks probably run for longer than they need to to provide useful information.
Given the data we already have, it should be easy enough to do an analysis of how much variability there is between runs of each benchmark (we usually run each 20 times), to determine which benchmarks could safely be run fewer times.
Within each benchmark, however, they could theoretically be doing less work and still providing the same value. That will require reducing data sizes and constants within the benchmarks themselves and measuring the impact on stability. It is probably worth focusing effort on the longest running benchmarks first.
Cc: @mpage, @colesbury, @brandtbucher
Metadata
Metadata
Assignees
Labels
No labels