by anp on 10/1/18, 6:21 PM with 40 comments
by MikeHolman on 10/2/18, 12:41 AM
What I do now is run a script that averages results from the preceding 10 runs and compares that to the average of the following 5 runs to see if the regression is consistent or anomalous. If the regression is consistent, then the script automatically files a bug in our tracker.
There is still some noise in the results, but it cuts down on those one-off issues.
by chriswarbo on 10/2/18, 10:38 AM
It was intended for use with Python (virtualenv or anaconda), but I created a plugin ( http://chriswarbo.net/projects/nixos/asv_benchmarking.html ) which allows using Nix instead, so we can provide any commands/tools/build-products we like in the benchmarking environment (so far I've used it successfully with projects written in Racket and Haskell).
by anp on 10/1/18, 7:33 PM
by valarauca1 on 10/1/18, 8:04 PM
Assuming the compiling, and testing is done in the cloud how do you ensure the target platform (processor) doesn't change, and that you aren't being subjected to neighbors who are stealing RAM bandwidth, or CPU cache resources from your VM and impacting the results?
by panic on 10/2/18, 3:15 AM
by habitue on 10/1/18, 11:06 PM
How long do we expect it to take before "automagically" completely replaces "automatically" in English?
I am guessing less than a decade to go now
by hsivonen on 10/1/18, 8:08 PM
Do you track opt_level=2 (the Firefox Rust opt level) in addition to the default opt_level=3?
by thsowers on 10/1/18, 9:44 PM
by Twirrim on 10/1/18, 11:15 PM
by awake on 10/1/18, 7:44 PM