Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(348)

Issue 10828077: Rerun each CL 10 times to smooth out noise in perf testing. (Closed)

Created:
8 years, 4 months ago by Emily Fortuna
Modified:
8 years, 4 months ago
Reviewers:
ricow1, vsm
CC:
reviews_dartlang.org
Visibility:
Public.

Description

Rerun each CL 10 times to smooth out noise in perf testing. Committed: https://code.google.com/p/dart/source/detail?r=10086

Patch Set 1 : #

Patch Set 2 : #

Patch Set 3 : #

Total comments: 7

Patch Set 4 : #

Total comments: 2
Unified diffs Side-by-side diffs Delta from patch set Stats (+251 lines, -522 lines) Patch
A + tools/get_archive.py View 1 2 3 7 chunks +109 lines, -33 lines 0 comments Download
M tools/get_drt.py View 1 2 3 1 chunk +0 lines, -216 lines 0 comments Download
M tools/testing/dart/drt_updater.dart View 1 2 3 1 chunk +3 lines, -3 lines 0 comments Download
M tools/testing/perf_testing/run_perf_tests.py View 1 2 3 30 chunks +139 lines, -270 lines 2 comments Download

Messages

Total messages: 7 (0 generated)
Emily Fortuna
This change finally implements rerunning past CLs so what we have 10 results for each ...
8 years, 4 months ago (2012-07-30 21:07:05 UTC) #1
vsm
https://chromiumcodereview.appspot.com/10828077/diff/6001/tools/get_archive.py File tools/get_archive.py (right): https://chromiumcodereview.appspot.com/10828077/diff/6001/tools/get_archive.py#newcode269 tools/get_archive.py:269: If the revision is set and positional[0] != 'sdk', ...
8 years, 4 months ago (2012-07-30 23:22:05 UTC) #2
vsm
lgtm https://chromiumcodereview.appspot.com/10828077/diff/6001/tools/testing/perf_testing/run_perf_tests.py File tools/testing/perf_testing/run_perf_tests.py (right): https://chromiumcodereview.appspot.com/10828077/diff/6001/tools/testing/perf_testing/run_perf_tests.py#newcode1075 tools/testing/perf_testing/run_perf_tests.py:1075: # been checked in. Nevermind. I see what ...
8 years, 4 months ago (2012-07-30 23:27:53 UTC) #3
Emily Fortuna
https://chromiumcodereview.appspot.com/10828077/diff/6001/tools/get_archive.py File tools/get_archive.py (right): https://chromiumcodereview.appspot.com/10828077/diff/6001/tools/get_archive.py#newcode269 tools/get_archive.py:269: On 2012/07/30 23:22:05, vsm wrote: > If the revision ...
8 years, 4 months ago (2012-07-31 20:16:20 UTC) #4
ricow1
https://chromiumcodereview.appspot.com/10828077/diff/1005/tools/testing/perf_testing/run_perf_tests.py File tools/testing/perf_testing/run_perf_tests.py (right): https://chromiumcodereview.appspot.com/10828077/diff/1005/tools/testing/perf_testing/run_perf_tests.py#newcode1110 tools/testing/perf_testing/run_perf_tests.py:1110: while revision_num > 3000 and not has_run_extra: This will ...
8 years, 4 months ago (2012-08-06 11:10:03 UTC) #5
Emily Fortuna
https://chromiumcodereview.appspot.com/10828077/diff/1005/tools/testing/perf_testing/run_perf_tests.py File tools/testing/perf_testing/run_perf_tests.py (right): https://chromiumcodereview.appspot.com/10828077/diff/1005/tools/testing/perf_testing/run_perf_tests.py#newcode1110 tools/testing/perf_testing/run_perf_tests.py:1110: while revision_num > 3000 and not has_run_extra: On 2012/08/06 ...
8 years, 4 months ago (2012-08-06 17:39:34 UTC) #6
ricow1
8 years, 4 months ago (2012-08-06 18:51:58 UTC) #7
On 2012/08/06 17:39:34, Emily Fortuna wrote:
>
https://chromiumcodereview.appspot.com/10828077/diff/1005/tools/testing/perf_...
> File tools/testing/perf_testing/run_perf_tests.py (right):
> 
>
https://chromiumcodereview.appspot.com/10828077/diff/1005/tools/testing/perf_...
> tools/testing/perf_testing/run_perf_tests.py:1110: while revision_num > 3000
and
> not has_run_extra:
> On 2012/08/06 11:10:03, ricow1 wrote:
> > This will take us back in time really really slow, and this might very well
> > starve out even recent revisions since I don't think that the current setup
> with
> > just 2 macs can keep up with our commit pace. This will make it harder to
> track
> > down regressions, and I doubt that we will ever get back historic results.
How
> > does the distribution of download build/run benchmark look, i.e., if
fetching
> > the build is dominating the time to run the benchmarks I see why you would
do
> > this
> 
> Right, so when this was originally designed, I thought we had SDK builds going
> back for a long time so switching between revisions would be cheap. After
> writing this I discovered this is not the case -- we only have the last 100
> revisions saved on Google Storage, and building/switching between random
> revisions is much more expensive. So, I was planning on starting at revision
> 9000 instead, and from this point onward, we'll be storing SDK revisions, so
> switching between revisions in the future (should, I hope) be cheap. Does this
> sound reasonable?
Sure, I am fine with whatever design you have, I just thought this could be an
issue. You could consider having a local cache on the mac for cheap switching,
e.g, two caches:
one cache with the last 100 revisions
one cache with a window of older revisions for which you are currently producing
results

You can do simple split, i.e, use 70% of the time running on the new cache
entries and 30% of the time on recalculating older ones (and some heuristics for
deciding this window based on existing result counts)

If you really want to be efficient you can have another machine locally in the
lab be cache, i.e., if you have an old machine and buy an external drive you
have a pretty efficient cache setup - we do this in the aarhus lab. You might
want to tell the cache upfront what to fetch instead of just fetching on demand
as we do, but feel free to reuse our cache functionality if you go down this
road - ping me or anders if you need info on this.

Powered by Google App Engine
This is Rietveld 408576698