This issue tracker has been migrated to GitHub, and is currently read-only.
For more information, see the GitHub FAQs in the Python's Developer Guide.

Author vstinner
Recipients brett.cannon, fijall, ned.deily, pitrou, serhiy.storchaka, tim.peters, vstinner, yselivanov
Date 2016-09-21.14:20:43
SpamBayes Score -1.0
Marked as misclassified Yes
Message-id <1474467643.8.0.892269378533.issue28240@psf.upfronthosting.co.za>
In-reply-to
Content
Maciej Fijalkowski also sent me the following article a few months ago, it also explains indirectly why using the minimum for benchmarks is not reliable:

"Virtual Machine Warmup Blows Hot and Cold"
http://arxiv.org/pdf/1602.00602.pdf

Even if the article is more focused on JIT compilers, it shows that benchmarks are not straightforward but always full of bad surprises.

A benchmark doesn't have a single value but a *distribution*. The best question is how to summarize the full distribution without loosing too much information.

In the perf module I decided to not take a decision: a JSON file stores *all* data :-D But by default, perf displays mean +- std dev.
History
Date User Action Args
2016-09-21 14:20:43vstinnersetrecipients: + vstinner, tim.peters, brett.cannon, pitrou, ned.deily, fijall, serhiy.storchaka, yselivanov
2016-09-21 14:20:43vstinnersetmessageid: <1474467643.8.0.892269378533.issue28240@psf.upfronthosting.co.za>
2016-09-21 14:20:43vstinnerlinkissue28240 messages
2016-09-21 14:20:43vstinnercreate