File: local_machine_hostside_test_run.py

package info (click to toggle)
chromium 139.0.7258.127-1
  • links: PTS, VCS
  • area: main
  • in suites:
  • size: 6,122,068 kB
  • sloc: cpp: 35,100,771; ansic: 7,163,530; javascript: 4,103,002; python: 1,436,920; asm: 946,517; xml: 746,709; pascal: 187,653; perl: 88,691; sh: 88,436; objc: 79,953; sql: 51,488; cs: 44,583; fortran: 24,137; makefile: 22,147; tcl: 15,277; php: 13,980; yacc: 8,984; ruby: 7,485; awk: 3,720; lisp: 3,096; lex: 1,327; ada: 727; jsp: 228; sed: 36
file content (342 lines) | stat: -rw-r--r-- 11,179 bytes parent folder | download | duplicates (9)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
# Copyright 2023 The Chromium Authors
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.

import collections
import contextlib
import dataclasses
import json
import logging
import os
import re
import subprocess
import sys
import time
from typing import Optional

from devil.android import logcat_monitor
from devil.android.tools import script_common
from devil.android.tools import webview_app
from pylib.base import base_test_result
from pylib.base import test_exception
from pylib.base import test_run
from pylib.local.machine import local_machine_junit_test_run as junitrun
from pylib.symbols import stack_symbolizer


_FAILURE_TYPES = (
    base_test_result.ResultType.FAIL,
    base_test_result.ResultType.CRASH,
    base_test_result.ResultType.TIMEOUT,
)

_TEST_START_RE = re.compile(r'.*=+ (\S+) STARTED: .*=+')
_TEST_END_RE = re.compile(r'.*=+ (\S+) ENDED: .*=+')
_JSON_RESULTS_RE = re.compile(
    r'.*D/LUCIResultReporter: JSON result for LUCI: (.*\.json)\b')

LOGCAT_FILTERS = [
  'chromium:v',
  'cr_*:v',
  'DEBUG:I',
  'StrictMode:D',
]

@dataclasses.dataclass
class _Job:
  shard_id: int
  cmd: str
  timeout: int
  env: Optional[dict] = None


class LocalMachineHostsideTestRun(test_run.TestRun):
  def __init__(self, env, test_instance):
    super().__init__(env, test_instance)
    self.webview_context = None
    self.device = None
    self.test_name_to_logcat = collections.defaultdict(collections.deque)

  # override
  def TestPackage(self):
    return self._test_instance.suite

  # override
  def GetTestsForListing(self):
    return self._GetTests()

  # override
  def SetUp(self):
    if self._test_instance.use_webview_provider:
      self.device = script_common.GetDevices(
          requested_devices=None,
          denylist_file=None)[0]
      for apk in self._test_instance.additional_apks:
        self.device.Install(apk)
      self.webview_context = webview_app.UseWebViewProvider(
          self.device,
          self._test_instance.use_webview_provider)
      # Pylint is not smart enough to realize that this field has
      # an __enter__ method, and will complain loudly.
      # pylint: disable=no-member
      self.webview_context.__enter__()
      # pylint: enable=no-member

  @staticmethod
  def _ApplyExternalSharding(tests, shard_index, total_shards):
    logging.info('Using external sharding settings. This is shard %d/%d',
                 shard_index, total_shards)

    if total_shards < 0 or shard_index < 0 or total_shards <= shard_index:
      raise test_exception.InvalidShardingSettings(shard_index, total_shards)

    return tests[shard_index::total_shards]

  def _MakeModeArgs(self):
    if self._test_instance.instant_mode:
      mode_args = [
          '--module-parameter',
          'INSTANT_APP',
      ]
    else:
      mode_args = [
          '--exclude-filter',
          f'{self.TestPackage()}[instant]',
      ]
    return mode_args

  def _MakeEnv(self):
    return dict(
        os.environ,
        PATH=':'.join([
          os.getenv('PATH'),
          self._test_instance.aapt_path,
          self._test_instance.adb_path]),
        CTS_ROOT=os.path.join(
          os.path.dirname(self._test_instance.tradefed_executable),
          os.pardir,
          os.pardir
        )
    )

  def _GetTests(self):
    filter_args = []
    # use passed in filters to list all filters tests
    for combined_filter in self._test_instance.test_filters:
      pattern_groups = combined_filter.split('-')
      negative_pattern = pattern_groups[1] if len(pattern_groups) > 1 else ''
      positive_pattern = pattern_groups[0]
      if negative_pattern:
        for exclude_filter in negative_pattern.split(':'):
          filter_args.extend([
              '--exclude-filter',
              self.TestPackage()
              + '[instant]' * self._test_instance.instant_mode
              + ' ' + '#'.join(exclude_filter.rsplit('.', 1)),
        ])
      if positive_pattern:
        for include_filter in positive_pattern.split(':'):
          filter_args.extend([
              '--include-filter',
              self.TestPackage()
              + '[instant]' * self._test_instance.instant_mode
              + ' ' + '#'.join(include_filter.rsplit('.', 1)),
        ])

    cmd = [
        self._test_instance.tradefed_executable,
        'run',
        'commandAndExit',
        'cts',
        '-m',
        self.TestPackage(),
    ] + self._MakeModeArgs() + filter_args + ['--collect-tests-only']

    logging.info('Getting tests from cts-tradefed using --collect_tests_only.')
    output = subprocess.check_output(
        cmd,
        timeout=600,
        env=self._MakeEnv(),
        universal_newlines=True,
    )

    tests = set()
    for line in output.splitlines():
      if test_start_match := _TEST_START_RE.match(line):
        tests.add(test_start_match.group(1))
    tests = self._ApplyExternalSharding(
        sorted(tests),
        self._test_instance.external_shard_index,
        self._test_instance.total_external_shards)
    return tests

  def _MakeJob(self):
    filter_args = []
    # only include current shard tests
    for test in self._GetTests():
      filter_args.extend([
          '--include-filter',
          self.TestPackage() + '[instant]' * self._test_instance.instant_mode
          + ' ' + test,
      ])

    cmd = [
        self._test_instance.tradefed_executable,
        'run',
        'commandAndExit',
        'cts',
        '-m',
        self.TestPackage(),
    ] + self._MakeModeArgs() + filter_args + [
        '--retry-strategy',
        'RETRY_ANY_FAILURE',
        '--max-testcase-run-count',
        str(self._test_instance.max_tries),
        '--template:map',
        'reporters=../../build/android/pylib/local/machine/'
        'local_machine_hostside_tradefed_config.xml',
    ]

    return _Job(
        # Note: the shard_id here is not to be confused with the external shard
        # index (set by swarming). This is a local shard index as we are reusing
        # more general-purpose code that allows for having multiple subprocesses
        # run tests locally in parallel. We are not doing that here, so sticking
        # to a fixed shard_id.
        shard_id=0,
        cmd=cmd,
        timeout=600,
        env=self._MakeEnv()
    )

  # override
  def RunTests(self, results, raw_logs_fh=None):
    job = self._MakeJob()

    per_test_logs = {}
    log_lines = []
    current_test = None
    json_results_path = None
    archive_logcat = None
    for line in junitrun.RunCommandsAndSerializeOutput([job], 1):
      if raw_logs_fh:
        raw_logs_fh.write(line)
      sys.stdout.write(line)

      # Collect log data between a test starting and the test failing.
      # There can be info after a test fails and before the next test starts
      # that we discard.
      if test_start_match := _TEST_START_RE.match(line):
        current_test = test_start_match.group(1)
        log_lines = [line]
        if archive_logcat is not None:
          archive_logcat.__exit__(None, None, None)
        archive_logcat = self._ArchiveLogcat(self.device, current_test)
        # Pylint is not smart enough to realize that this field has
        # an __enter__ method, and will complain loudly.
        # pylint: disable=no-member
        archive_logcat.__enter__()
        # pylint: enable=no-member
      else:
        log_lines.append(line)
        if _TEST_END_RE.match(line) and current_test:
          per_test_logs[current_test] = ''.join(log_lines)
          archive_logcat.__exit__(None, None, None)
          archive_logcat = None
          current_test = None
        elif json_results_path_match := _JSON_RESULTS_RE.match(line):
          json_results_path = json_results_path_match.group(1)

    sys.stdout.flush()
    if archive_logcat is not None:
      # Pylint is not smart enough to realize that this field has
      # an __exit__ method, and will complain loudly.
      # pylint: disable=no-member
      archive_logcat.__exit__(None, None, None)
      # pylint: enable=no-member

    result_list = []
    if json_results_path:
      with open(json_results_path, 'r') as f:
        json_results = json.load(f)
        parsed_results = _ParseResultsFromJson(json_results)
      for r in parsed_results:
        if r.GetType() in _FAILURE_TYPES:
          r.SetLog(per_test_logs.get(r.GetName(), ''))
      attempt_counter = collections.Counter(
          result.GetName() for result in parsed_results
      )
      for result in parsed_results:
        test_name = result.GetName()
        logcat_deque = self.test_name_to_logcat[test_name]
        if attempt_counter[test_name] == len(logcat_deque):
          # Set logcat link in FIFO order in case of multiple test attempts
          result.SetLink('logcat', logcat_deque.popleft())
          attempt_counter[test_name] -= 1

      result_list += parsed_results
    else:
      # In the case of a failure in the test runner
      # the output json file may never be written.
      result_list = [
          base_test_result.BaseTestResult(
              'Test Runner Failure',
              base_test_result.ResultType.UNKNOWN)
      ]

    test_run_results = base_test_result.TestRunResults()
    test_run_results.AddResults(result_list)
    results.append(test_run_results)

  # override
  def TearDown(self):
    if self._test_instance.use_webview_provider:
      # Pylint is not smart enough to realize that this field has
      # an __exit__ method, and will complain loudly.
      # pylint: disable=no-member
      self.webview_context.__exit__(*sys.exc_info())
      # pylint: enable=no-member

  @contextlib.contextmanager
  def _ArchiveLogcat(self, device, test_name):
    stream_name = 'logcat_%s_shard%s_%s_%s' % (
        test_name.replace('#', '.'), self._test_instance.external_shard_index,
        time.strftime('%Y%m%dT%H%M%S-UTC', time.gmtime()), device.serial)

    logcat_file = None
    logmon = None
    try:
      with self._env.output_manager.ArchivedTempfile(stream_name,
                                                     'logcat') as logcat_file:
        symbolizer = stack_symbolizer.PassThroughSymbolizerPool(
            device.product_cpu_abi)
        with symbolizer:
          with logcat_monitor.LogcatMonitor(
              device.adb,
              filter_specs=LOGCAT_FILTERS,
              output_file=logcat_file.name,
              check_error=False) as logmon:
              yield logcat_file
    finally:
      if logmon:
        logmon.Close()
      if logcat_file and logcat_file.Link():
        logging.critical('Logcat saved to %s', logcat_file.Link())
        self.test_name_to_logcat[test_name].append(logcat_file.Link())

def _ParseResultsFromJson(json_results):
  result_list = []
  result_list.extend([
      base_test_result.BaseTestResult(
          tr['testId'],
          getattr(
              base_test_result.ResultType,
              tr['status'],
              base_test_result.ResultType.UNKNOWN
          ),
          duration=int(tr['duration'] * 1000),
          failure_reason=tr.get('failureReason'),
      )
      for tr in json_results['tr']
  ])
  return result_list