[Python-Dev] [Python-checkins] BAD Benchmark Results for Python Default 2016-01-26 (original) (raw)

Stefan Krah stefan at bytereef.org
Thu Jan 28 08:49:50 EST 2016


IMO the timings of the benchmark suite are a bit unstable -- this is not the fault of Intel's setup, I noticed it also when running the suite myself.

On Tue, Jan 26, 2016 at 06:48:54PM +0000, Stewart, David C wrote:

Wow, what happened to Python default to cause such a regression?

On 1/26/16, 7:31 AM, "lpbenchmarkrobot" <lpbenchmarkrobot at intel.com> wrote: >Results for project Python default, build date 2016-01-26 03:07:40 +0000 >commit: cbd4a6a2657e >previous commit: f700bc0412bc >revision date: 2016-01-26 02:54:37 +0000 >environment: Haswell-EP > cpu: Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz 2x18 cores, stepping 2, LLC 45 MB > mem: 128 GB > os: CentOS 7.1 > kernel: Linux 3.10.0-229.4.2.el7.x8664 > >Baseline results were generated using release v3.4.3, with hash b4cbecbc0781 >from 2015-02-25 12:15:33+00:00 > >---------------------------------------------------------------------------------- > benchmark relative change since change since current rev run > stddev* last run baseline with PGO >---------------------------------------------------------------------------------- >:-) djangov2 0.21% -2.93% 8.95% 16.19% >:-| pybench 0.10% 0.05% -1.87% 5.40% >:-( regexv8 2.72% -0.02% -4.67% 4.57% >:-| nbody 0.13% -0.92% -1.33% 7.40% >:-| jsondumpv2 0.20% 0.87% -1.59% 11.48% >:-| normalstartup 0.90% -0.57% 0.10% 5.35% >---------------------------------------------------------------------------------- >* Relative Standard Deviation (Standard Deviation/Average) > >If this is not displayed properly please visit our results page here: http://languagesperformance.intel.com/bad-benchmark-results-for-python-default-2016-01-26/ > >Note: Benchmark results are measured in seconds. > >Subject Label Legend: >Attributes are determined based on the performance evolution of the workloads >compared to the previous measurement iteration. >NEUTRAL: performance did not change by more than 1% for any workload >GOOD: performance improved by more than 1% for at least one workload and there >is no regression greater than 1% >BAD: performance dropped by more than 1% for at least one workload and there is >no improvement greater than 1% >UGLY: performance improved by more than 1% for at least one workload and also >dropped by more than 1% for at least one workload > > >Our lab does a nightly source pull and build of the Python project and measures >performance changes against the previous stable version and the previous nightly >measurement. This is provided as a service to the community so that quality >issues with current hardware can be identified quickly. > >Intel technologies' features and benefits depend on system configuration and may >require enabled hardware, software or service activation. Performance varies >depending on system configuration.


Python-checkins mailing list Python-checkins at python.org https://mail.python.org/mailman/listinfo/python-checkins



More information about the Python-Dev mailing list