-
Updated
Jul 22, 2020 - Python
benchmark
Here are 2,204 public repositories matching this topic...
-
Updated
Aug 11, 2020
For each Job, it adds plots about density, cumulative mean, and so on. But two files are named BenchmarkDotNet.Artifacts/results/MyBench.Sleeps-Time50--density.png and BenchmarkDotNet.Artifacts/results/MyBench.Sleeps-Time50--facetDensity.png, with the -- instead of single. Like some iteration variable is empty (since later there are names with -Default-


Is there currently any way to force hyperfine to write intermediate results, so that I can abort a benchmark without losing all the progress I've made so far? I'm asking, because I've been benchmarking several algorithms with exponential complexity recently, and so sometimes letting the benchmark finish isn't an option.