Index: tools/telemetry/telemetry/multi_page_benchmark_runner.py |
diff --git a/tools/telemetry/telemetry/multi_page_benchmark_runner.py b/tools/telemetry/telemetry/multi_page_benchmark_runner.py |
deleted file mode 100755 |
index e1952dd4b53619cad2655eaefdaaecc9dc29217a..0000000000000000000000000000000000000000 |
--- a/tools/telemetry/telemetry/multi_page_benchmark_runner.py |
+++ /dev/null |
@@ -1,107 +0,0 @@ |
-#!/usr/bin/env python |
-# Copyright (c) 2012 The Chromium Authors. All rights reserved. |
-# Use of this source code is governed by a BSD-style license that can be |
-# found in the LICENSE file. |
-import csv |
-import logging |
-import os |
-import sys |
- |
-from telemetry import all_page_actions # pylint: disable=W0611 |
-from telemetry import block_page_benchmark_results |
-from telemetry import browser_finder |
-from telemetry import browser_options |
-from telemetry import csv_page_benchmark_results |
-from telemetry import discover |
-from telemetry import multi_page_benchmark |
-from telemetry import page_runner |
-from telemetry import page_set |
- |
-def Main(benchmark_dir): |
- """Turns a MultiPageBenchmark into a command-line program. |
- |
- Args: |
- benchmark_dir: Path to directory containing MultiPageBenchmarks. |
- """ |
- benchmarks = discover.Discover(benchmark_dir, '', |
- multi_page_benchmark.MultiPageBenchmark) |
- |
- # Naively find the benchmark. If we use the browser options parser, we run |
- # the risk of failing to parse if we use a benchmark-specific parameter. |
- benchmark_name = None |
- for arg in sys.argv: |
- if arg in benchmarks: |
- benchmark_name = arg |
- |
- options = browser_options.BrowserOptions() |
- parser = options.CreateParser('%prog [options] <benchmark> <page_set>') |
- |
- page_runner.PageRunner.AddCommandLineOptions(parser) |
- parser.add_option('--output-format', |
- dest='output_format', |
- default='csv', |
- help='Output format. Can be "csv" or "block". ' |
- 'Defaults to "%default".') |
- parser.add_option('-o', '--output', |
- dest='output_file', |
- help='Redirects output to a file. Defaults to stdout.') |
- |
- benchmark = None |
- if benchmark_name is not None: |
- benchmark = benchmarks[benchmark_name]() |
- benchmark.AddCommandLineOptions(parser) |
- |
- _, args = parser.parse_args() |
- |
- if benchmark is None or len(args) != 2: |
- parser.print_usage() |
- import page_sets # pylint: disable=F0401 |
- print >> sys.stderr, 'Available benchmarks:\n%s\n' % ',\n'.join( |
- sorted(benchmarks.keys())) |
- print >> sys.stderr, 'Available page_sets:\n%s\n' % ',\n'.join( |
- sorted([os.path.relpath(f) |
- for f in page_sets.GetAllPageSetFilenames()])) |
- sys.exit(1) |
- |
- ps = page_set.PageSet.FromFile(args[1]) |
- |
- benchmark.CustomizeBrowserOptions(options) |
- possible_browser = browser_finder.FindBrowser(options) |
- if not possible_browser: |
- print >> sys.stderr, """No browser found.\n |
-Use --browser=list to figure out which are available.\n""" |
- sys.exit(1) |
- |
- if not options.output_file: |
- output_file = sys.stdout |
- elif options.output_file == '-': |
- output_file = sys.stdout |
- else: |
- output_file = open(os.path.expanduser(options.output_file), 'w') |
- |
- if options.output_format == 'csv': |
- results = csv_page_benchmark_results.CsvPageBenchmarkResults( |
- csv.writer(output_file), |
- benchmark.results_are_the_same_on_every_page) |
- elif options.output_format in ('block', 'terminal-block'): |
- results = block_page_benchmark_results.BlockPageBenchmarkResults( |
- output_file) |
- else: |
- raise Exception('Invalid --output-format value: "%s". Valid values are ' |
- '"csv" and "block".' |
- % options.output_format) |
- |
- with page_runner.PageRunner(ps) as runner: |
- runner.Run(options, possible_browser, benchmark, results) |
- # When using an exact executable, assume it is a reference build for the |
- # purpose of outputting the perf results. |
- results.PrintSummary(options.browser_executable and '_ref' or '') |
- |
- if len(results.page_failures): |
- logging.warning('Failed pages: %s', '\n'.join( |
- [failure['page'].url for failure in results.page_failures])) |
- |
- if len(results.skipped_pages): |
- logging.warning('Skipped pages: %s', '\n'.join( |
- [skipped['page'].url for skipped in results.skipped_pages])) |
- return min(255, len(results.page_failures)) |