Index: scripts/slave/telemetry_utils.py |
diff --git a/scripts/slave/telemetry_utils.py b/scripts/slave/telemetry_utils.py |
index cbe0a9df73e8b61bcd5c79e905f2be38a526ca0f..07c7307e5fc84f7dab56e5036b3660b67abd6727 100644 |
--- a/scripts/slave/telemetry_utils.py |
+++ b/scripts/slave/telemetry_utils.py |
@@ -15,6 +15,9 @@ import json |
import logging |
import os |
+from slave.performance_log_processor import _FormatHumanReadable |
+ |
+ |
class TelemetryResultsProcessor(object): |
def __init__(self, filename, is_ref): |
@@ -61,4 +64,20 @@ class TelemetryResultsProcessor(object): |
return [] |
def PerformanceSummary(self): |
- return '' |
+ """Writes the waterfall display text. |
+ |
+ The waterfall contains lines for each important trace, in the form |
+ tracename: value< (refvalue)> |
+ """ |
+ if self._is_reference_build: |
+ return [] |
+ |
+ results = [] |
+ for chart_name, chart_values in self.ChartJson()['charts'].iteritems(): |
+ if 'summary' in chart_values: |
+ summary = chart_values['summary'] |
shatch
2014/11/21 17:21:48
Are we only displaying summary results? This mostl
sullivan
2014/11/21 17:30:52
I think displaying summary results is fine.
|
+ if summary['important']: |
+ mean = sum(summary['values']) / float(len(summary['values'])) |
+ display = '%s: %s' % (chart_name, _FormatHumanReadable(mean)) |
+ results.append(display) |
+ return results |