1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187
|
#!/usr/bin/env vpython3
# Copyright 2018 The Chromium Authors
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import argparse
import collections
import json
import subprocess
import sys
_CLOUD_PROJECT_ID = 'chrome-luci-data'
QUERY_BY_BUILD_NUMBER = """
SELECT
test_id AS name,
SUM(duration) AS duration
FROM
`%s.chrome.ci_test_results`
WHERE
DATE(partition_time) = "{}"
AND exported.id like "{}"
GROUP BY
test_id
ORDER BY
test_id
LIMIT
1000
""" % _CLOUD_PROJECT_ID
QUERY_STORY_AVG_RUNTIME = """
SELECT
test_id AS name,
ROUND(AVG(duration)) AS duration
FROM
`%s.chrome.ci_test_results`
CROSS JOIN
UNNEST(variant) AS v
WHERE
v.key = 'builder'
AND v.value IN ({configuration_names})
AND duration IS NOT NULL
AND duration != 0
AND expected
AND DATE(partition_time) > DATE_SUB(CURRENT_DATE(), INTERVAL {num_last_days} DAY)
GROUP BY
test_id
ORDER BY
test_id
""" % _CLOUD_PROJECT_ID
QUERY_STORY_TOTAL_RUNTIME = """
SELECT
test_id AS name,
ROUND(AVG(time)) AS duration,
FROM (
SELECT
parent.id,
test_id,
SUM(duration) AS time
FROM
`%s.chrome.ci_test_results`
CROSS JOIN
UNNEST(variant) AS v
WHERE
v.key = 'builder'
AND v.value IN ({configuration_names})
AND duration IS NOT NULL
AND duration != 0
AND expected
AND DATE(partition_time) > DATE_SUB(CURRENT_DATE(), INTERVAL {num_last_days} DAY)
GROUP BY
parent.id,
test_id)
GROUP BY
test_id
ORDER BY
test_id
""" % _CLOUD_PROJECT_ID
_BQ_SETUP_INSTRUCTION = """
** NOTE: this script is only for Googlers to use. **
bq script isn't found on your machine. To run this script, you need to be able
to run bigquery in your terminal.
If this is the first time you run the script, do the following steps:
1) Follow the steps at https://cloud.google.com/sdk/docs/ to download and
unpack google-cloud-sdk in your home directory.
2) Run `gcloud auth login`
3) Run `gcloud config set project chrome-luci-data`
3a) If 'chrome-luci-data' does not show up, contact chops-data@
to be added as a user of the table
4) Run this script!
"""
def _run_query(query):
try:
subprocess.check_call(['which', 'bq'])
except subprocess.CalledProcessError:
raise RuntimeError(_BQ_SETUP_INSTRUCTION)
args = ["bq", "query", "--project_id="+_CLOUD_PROJECT_ID, "--format=json",
"--max_rows=100000", "--nouse_legacy_sql", query]
p = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
stdout, stderr = p.communicate()
if p.returncode == 0:
return json.loads(stdout)
raise RuntimeError(
'Error generating authentication token.\nStdout: %s\nStder:%s' %
(stdout, stderr))
def FetchStoryTimingDataForSingleBuild(date, build_number):
return _run_query(QUERY_BY_BUILD_NUMBER.format(date, build_number))
def FetchAverageStoryTimingData(configurations, num_last_days):
configurations_str = ','.join(repr(c) for c in configurations)
return _run_query(QUERY_STORY_AVG_RUNTIME.format(
configuration_names=configurations_str, num_last_days=num_last_days))
def FetchBenchmarkRuntime(configurations, num_last_days):
configurations_str = ','.join(repr(c) for c in configurations)
test_total_runtime = _run_query(QUERY_STORY_TOTAL_RUNTIME.format(
configuration_names=configurations_str, num_last_days=num_last_days))
benchmarks_data = collections.OrderedDict()
total_runtime = 0
total_num_stories = 0
for item in test_total_runtime:
duration = item['duration']
test_name = item['name']
benchmark_name, _ = test_name.split('/', 1)
if not benchmark_name in benchmarks_data:
benchmarks_data[benchmark_name] = {
'num_stories': 0,
'total_runtime_in_seconds': 0,
}
benchmarks_data[benchmark_name]['num_stories'] += 1
total_num_stories += 1
benchmarks_data[benchmark_name]['total_runtime_in_seconds'] += (
float(duration))
total_runtime += float(duration)
benchmarks_data['All benchmarks'] = {
'total_runtime_in_seconds': total_runtime,
'num_stories': total_num_stories
}
return benchmarks_data
_FETCH_BENCHMARK_RUNTIME = 'fetch-benchmark-runtime'
_FETCH_STORY_RUNTIME = 'fetch-story-runtime'
def main(args):
parser = argparse.ArgumentParser(
description='Retrieve story timing from bigquery.')
parser.add_argument('action',
choices=[_FETCH_BENCHMARK_RUNTIME, _FETCH_STORY_RUNTIME])
parser.add_argument(
'--output-file', '-o', action='store', required=True,
help='The filename to send the bigquery results to.')
parser.add_argument(
'--configurations', '-c', action='append', required=True,
help='The configuration(s) of the builder to query results from.')
parser.add_argument(
'--date', '-d',
help='The date to query results from.')
parser.add_argument(
'--build-number', action='store',
help='If specified, the build number to get timing data from.')
opts = parser.parse_args(args)
if opts.action == _FETCH_BENCHMARK_RUNTIME:
data = FetchBenchmarkRuntime(opts.configurations, num_last_days=5)
else:
if opts.build_number:
data = FetchStoryTimingDataForSingleBuild(opts.date, opts.build_number)
else:
data = FetchAverageStoryTimingData(opts.configurations, num_last_days=5)
with open(opts.output_file, 'w') as output_file:
json.dump(data, output_file, indent = 4, separators=(',', ': '))
if __name__ == '__main__':
sys.exit(main(sys.argv[1:]))
|