| Index: tools/telemetry/telemetry/multi_page_benchmark_unittest_base.py
|
| diff --git a/tools/telemetry/telemetry/multi_page_benchmark_unittest_base.py b/tools/telemetry/telemetry/multi_page_benchmark_unittest_base.py
|
| deleted file mode 100644
|
| index eb9d5224c969b0d23127a0e64512660bbfe2c8e2..0000000000000000000000000000000000000000
|
| --- a/tools/telemetry/telemetry/multi_page_benchmark_unittest_base.py
|
| +++ /dev/null
|
| @@ -1,49 +0,0 @@
|
| -# Copyright (c) 2012 The Chromium Authors. All rights reserved.
|
| -# Use of this source code is governed by a BSD-style license that can be
|
| -# found in the LICENSE file.
|
| -import os
|
| -import unittest
|
| -
|
| -from telemetry import browser_finder
|
| -from telemetry import options_for_unittests
|
| -from telemetry import page_runner
|
| -from telemetry import page as page_module
|
| -from telemetry import page_benchmark_results
|
| -from telemetry import page_set
|
| -
|
| -class MultiPageBenchmarkUnitTestBase(unittest.TestCase):
|
| - """unittest.TestCase-derived class to help in the construction of unit tests
|
| - for a benchmark."""
|
| -
|
| - def CreatePageSetFromFileInUnittestDataDir(self, test_filename):
|
| - return self.CreatePageSet('file:///' + os.path.join('..', 'unittest_data',
|
| - test_filename))
|
| -
|
| - def CreatePageSet(self, test_filename):
|
| - base_dir = os.path.dirname(__file__)
|
| - ps = page_set.PageSet(file_path=os.path.join(base_dir, 'foo.json'))
|
| - page = page_module.Page(test_filename, ps, base_dir=base_dir)
|
| - setattr(page, 'smoothness', {'action': 'scrolling_action'})
|
| - ps.pages.append(page)
|
| - return ps
|
| -
|
| - def RunBenchmark(self, benchmark, ps, options=None):
|
| - """Runs a benchmark against a pageset, returning the rows its outputs."""
|
| - if options is None:
|
| - options = options_for_unittests.GetCopy()
|
| - assert options
|
| - temp_parser = options.CreateParser()
|
| - benchmark.AddCommandLineOptions(temp_parser)
|
| - defaults = temp_parser.get_default_values()
|
| - for k, v in defaults.__dict__.items():
|
| - if hasattr(options, k):
|
| - continue
|
| - setattr(options, k, v)
|
| -
|
| - benchmark.CustomizeBrowserOptions(options)
|
| - possible_browser = browser_finder.FindBrowser(options)
|
| -
|
| - results = page_benchmark_results.PageBenchmarkResults()
|
| - with page_runner.PageRunner(ps) as runner:
|
| - runner.Run(options, possible_browser, benchmark, results)
|
| - return results
|
|
|