blob: 35926e494495a8924ede644dee96ff777262f2d9 [file] [log] [blame]
maruel@chromium.org0437a732013-08-27 16:05:52 +00001#!/usr/bin/env python
maruelea586f32016-04-05 11:11:33 -07002# Copyright 2013 The LUCI Authors. All rights reserved.
maruelf1f5e2a2016-05-25 17:10:39 -07003# Use of this source code is governed under the Apache License, Version 2.0
4# that can be found in the LICENSE file.
maruel@chromium.org0437a732013-08-27 16:05:52 +00005
6"""Client tool to trigger tasks or retrieve results from a Swarming server."""
7
Robert Iannuccifafa7352018-06-13 17:08:17 +00008__version__ = '0.13'
maruel@chromium.org0437a732013-08-27 16:05:52 +00009
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -050010import collections
Marc-Antoine Ruel5e6ccdb2015-04-02 15:55:13 -040011import datetime
maruel@chromium.org0437a732013-08-27 16:05:52 +000012import json
13import logging
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -040014import optparse
maruel@chromium.org0437a732013-08-27 16:05:52 +000015import os
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +100016import re
maruel@chromium.org0437a732013-08-27 16:05:52 +000017import sys
maruel11e31af2017-02-15 07:30:50 -080018import textwrap
Vadim Shtayurab19319e2014-04-27 08:50:06 -070019import threading
maruel@chromium.org0437a732013-08-27 16:05:52 +000020import time
21import urllib
maruel@chromium.org0437a732013-08-27 16:05:52 +000022
23from third_party import colorama
24from third_party.depot_tools import fix_encoding
25from third_party.depot_tools import subcommand
vadimsh@chromium.org6b706212013-08-28 15:03:46 +000026
Marc-Antoine Ruel8806e622014-02-12 14:15:53 -050027from utils import file_path
maruel12e30012015-10-09 11:55:35 -070028from utils import fs
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -040029from utils import logging_utils
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -040030from third_party.chromium import natsort
vadimsh@chromium.org6b706212013-08-28 15:03:46 +000031from utils import net
Marc-Antoine Ruelcfb60852014-07-02 15:22:00 -040032from utils import on_error
maruel8e4e40c2016-05-30 06:21:07 -070033from utils import subprocess42
maruel@chromium.org0437a732013-08-27 16:05:52 +000034from utils import threading_utils
vadimsh@chromium.org6b706212013-08-28 15:03:46 +000035from utils import tools
maruel@chromium.org0437a732013-08-27 16:05:52 +000036
Vadim Shtayurae34e13a2014-02-02 11:23:26 -080037import auth
iannucci31ab9192017-05-02 19:11:56 -070038import cipd
maruel@chromium.org7b844a62013-09-17 13:04:59 +000039import isolateserver
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +000040import isolated_format
Marc-Antoine Ruel2666d9c2018-05-18 13:52:02 -040041import local_caching
maruelc070e672016-02-22 17:32:57 -080042import run_isolated
maruel@chromium.org0437a732013-08-27 16:05:52 +000043
44
tansella4949442016-06-23 22:34:32 -070045ROOT_DIR = os.path.dirname(os.path.abspath(
46 __file__.decode(sys.getfilesystemencoding())))
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -050047
48
49class Failure(Exception):
50 """Generic failure."""
51 pass
52
53
maruel0a25f6c2017-05-10 10:43:23 -070054def default_task_name(options):
55 """Returns a default task name if not specified."""
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -050056 if not options.task_name:
maruel0a25f6c2017-05-10 10:43:23 -070057 task_name = u'%s/%s' % (
marueld9cc8422017-05-09 12:07:02 -070058 options.user,
maruelaf6b06c2017-06-08 06:26:53 -070059 '_'.join('%s=%s' % (k, v) for k, v in options.dimensions))
maruel0a25f6c2017-05-10 10:43:23 -070060 if options.isolated:
61 task_name += u'/' + options.isolated
62 return task_name
63 return options.task_name
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -050064
65
66### Triggering.
67
68
maruel77f720b2015-09-15 12:35:22 -070069# See ../appengine/swarming/swarming_rpcs.py.
borenet02f772b2016-06-22 12:42:19 -070070CipdPackage = collections.namedtuple(
71 'CipdPackage',
72 [
73 'package_name',
74 'path',
75 'version',
76 ])
77
78
79# See ../appengine/swarming/swarming_rpcs.py.
80CipdInput = collections.namedtuple(
81 'CipdInput',
82 [
83 'client_package',
84 'packages',
85 'server',
86 ])
87
88
89# See ../appengine/swarming/swarming_rpcs.py.
maruel77f720b2015-09-15 12:35:22 -070090FilesRef = collections.namedtuple(
91 'FilesRef',
92 [
93 'isolated',
94 'isolatedserver',
95 'namespace',
96 ])
97
98
99# See ../appengine/swarming/swarming_rpcs.py.
Robert Iannuccibf5f84c2017-11-22 12:56:50 -0800100StringListPair = collections.namedtuple(
101 'StringListPair', [
102 'key',
103 'value', # repeated string
104 ]
105)
106
107
108# See ../appengine/swarming/swarming_rpcs.py.
maruel77f720b2015-09-15 12:35:22 -0700109TaskProperties = collections.namedtuple(
110 'TaskProperties',
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500111 [
maruel681d6802017-01-17 16:56:03 -0800112 'caches',
borenet02f772b2016-06-22 12:42:19 -0700113 'cipd_input',
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500114 'command',
Marc-Antoine Ruelba1bf222017-12-21 21:41:01 -0500115 'relative_cwd',
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500116 'dimensions',
117 'env',
Robert Iannuccibf5f84c2017-11-22 12:56:50 -0800118 'env_prefixes',
maruel77f720b2015-09-15 12:35:22 -0700119 'execution_timeout_secs',
120 'extra_args',
121 'grace_period_secs',
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500122 'idempotent',
maruel77f720b2015-09-15 12:35:22 -0700123 'inputs_ref',
124 'io_timeout_secs',
aludwincc5524e2016-10-28 10:25:24 -0700125 'outputs',
iannuccidc80dfb2016-10-28 12:50:20 -0700126 'secret_bytes',
maruel77f720b2015-09-15 12:35:22 -0700127 ])
128
129
130# See ../appengine/swarming/swarming_rpcs.py.
Marc-Antoine Ruel1f835c72018-05-25 12:29:42 -0400131TaskSlice = collections.namedtuple(
132 'TaskSlice',
133 [
134 'expiration_secs',
135 'properties',
136 'wait_for_capacity',
137 ])
138
139
140# See ../appengine/swarming/swarming_rpcs.py.
maruel77f720b2015-09-15 12:35:22 -0700141NewTaskRequest = collections.namedtuple(
142 'NewTaskRequest',
143 [
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500144 'name',
maruel77f720b2015-09-15 12:35:22 -0700145 'parent_task_id',
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500146 'priority',
Marc-Antoine Ruelb73066b2018-04-19 20:16:55 -0400147 'task_slices',
Vadim Shtayura2d83a942017-08-14 17:41:24 -0700148 'service_account',
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500149 'tags',
150 'user',
Robert Iannuccifafa7352018-06-13 17:08:17 +0000151 'pool_task_template',
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500152 ])
153
154
maruel77f720b2015-09-15 12:35:22 -0700155def namedtuple_to_dict(value):
156 """Recursively converts a namedtuple to a dict."""
Marc-Antoine Ruelb73066b2018-04-19 20:16:55 -0400157 if hasattr(value, '_asdict'):
158 return namedtuple_to_dict(value._asdict())
159 if isinstance(value, (list, tuple)):
160 return [namedtuple_to_dict(v) for v in value]
161 if isinstance(value, dict):
162 return {k: namedtuple_to_dict(v) for k, v in value.iteritems()}
163 return value
maruel77f720b2015-09-15 12:35:22 -0700164
165
Vadim Shtayura2d83a942017-08-14 17:41:24 -0700166def task_request_to_raw_request(task_request):
maruel71c61c82016-02-22 06:52:05 -0800167 """Returns the json-compatible dict expected by the server for new request.
maruelaf6269b2015-09-10 14:37:51 -0700168
169 This is for the v1 client Swarming API.
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500170 """
maruel77f720b2015-09-15 12:35:22 -0700171 out = namedtuple_to_dict(task_request)
Vadim Shtayura2d83a942017-08-14 17:41:24 -0700172 # Don't send 'service_account' if it is None to avoid confusing older
173 # version of the server that doesn't know about 'service_account' and don't
174 # use it at all.
175 if not out['service_account']:
176 out.pop('service_account')
Brad Hallf78187a2018-10-19 17:08:55 +0000177 for task_slice in out['task_slices']:
178 task_slice['properties']['env'] = [
179 {'key': k, 'value': v}
180 for k, v in task_slice['properties']['env'].iteritems()
181 ]
182 task_slice['properties']['env'].sort(key=lambda x: x['key'])
maruel77f720b2015-09-15 12:35:22 -0700183 return out
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500184
185
maruel77f720b2015-09-15 12:35:22 -0700186def swarming_trigger(swarming, raw_request):
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500187 """Triggers a request on the Swarming server and returns the json data.
188
189 It's the low-level function.
190
191 Returns:
192 {
193 'request': {
194 'created_ts': u'2010-01-02 03:04:05',
195 'name': ..
196 },
197 'task_id': '12300',
198 }
199 """
200 logging.info('Triggering: %s', raw_request['name'])
201
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500202 result = net.url_read_json(
smut281c3902018-05-30 17:50:05 -0700203 swarming + '/_ah/api/swarming/v1/tasks/new', data=raw_request)
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500204 if not result:
205 on_error.report('Failed to trigger task %s' % raw_request['name'])
206 return None
maruele557bce2015-11-17 09:01:27 -0800207 if result.get('error'):
marueld4d15312015-11-16 17:22:59 -0800208 # The reply is an error.
maruele557bce2015-11-17 09:01:27 -0800209 msg = 'Failed to trigger task %s' % raw_request['name']
210 if result['error'].get('errors'):
211 for err in result['error']['errors']:
212 if err.get('message'):
213 msg += '\nMessage: %s' % err['message']
214 if err.get('debugInfo'):
215 msg += '\nDebug info:\n%s' % err['debugInfo']
216 elif result['error'].get('message'):
maruelbf53e042015-12-01 15:00:51 -0800217 msg += '\nMessage: %s' % result['error']['message']
maruele557bce2015-11-17 09:01:27 -0800218
219 on_error.report(msg)
marueld4d15312015-11-16 17:22:59 -0800220 return None
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500221 return result
222
223
224def setup_googletest(env, shards, index):
225 """Sets googletest specific environment variables."""
226 if shards > 1:
maruel77f720b2015-09-15 12:35:22 -0700227 assert not any(i['key'] == 'GTEST_SHARD_INDEX' for i in env), env
228 assert not any(i['key'] == 'GTEST_TOTAL_SHARDS' for i in env), env
229 env = env[:]
230 env.append({'key': 'GTEST_SHARD_INDEX', 'value': str(index)})
231 env.append({'key': 'GTEST_TOTAL_SHARDS', 'value': str(shards)})
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500232 return env
233
234
235def trigger_task_shards(swarming, task_request, shards):
236 """Triggers one or many subtasks of a sharded task.
237
238 Returns:
239 Dict with task details, returned to caller as part of --dump-json output.
240 None in case of failure.
241 """
242 def convert(index):
Vadim Shtayura2d83a942017-08-14 17:41:24 -0700243 req = task_request_to_raw_request(task_request)
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500244 if shards > 1:
Marc-Antoine Ruelb73066b2018-04-19 20:16:55 -0400245 req['task_slices'][0]['properties']['env'] = setup_googletest(
246 req['task_slices'][0]['properties']['env'], shards, index)
maruel77f720b2015-09-15 12:35:22 -0700247 req['name'] += ':%s:%s' % (index, shards)
248 return req
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500249
250 requests = [convert(index) for index in xrange(shards)]
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500251 tasks = {}
252 priority_warning = False
253 for index, request in enumerate(requests):
maruel77f720b2015-09-15 12:35:22 -0700254 task = swarming_trigger(swarming, request)
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500255 if not task:
256 break
257 logging.info('Request result: %s', task)
258 if (not priority_warning and
Marc-Antoine Ruelb1216762017-08-17 10:07:49 -0400259 int(task['request']['priority']) != task_request.priority):
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500260 priority_warning = True
261 print >> sys.stderr, (
262 'Priority was reset to %s' % task['request']['priority'])
263 tasks[request['name']] = {
264 'shard_index': index,
265 'task_id': task['task_id'],
266 'view_url': '%s/user/task/%s' % (swarming, task['task_id']),
267 }
268
269 # Some shards weren't triggered. Abort everything.
270 if len(tasks) != len(requests):
271 if tasks:
272 print >> sys.stderr, 'Only %d shard(s) out of %d were triggered' % (
273 len(tasks), len(requests))
274 for task_dict in tasks.itervalues():
275 abort_task(swarming, task_dict['task_id'])
276 return None
277
278 return tasks
279
280
281### Collection.
maruel@chromium.org0437a732013-08-27 16:05:52 +0000282
283
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700284# How often to print status updates to stdout in 'collect'.
285STATUS_UPDATE_INTERVAL = 15 * 60.
286
Marc-Antoine Ruel2f6581a2014-10-03 11:09:53 -0400287
Marc-Antoine Ruel20b764d2018-06-22 18:08:37 +0000288class TaskState(object):
289 """Represents the current task state.
maruel@chromium.org0437a732013-08-27 16:05:52 +0000290
Marc-Antoine Ruel20b764d2018-06-22 18:08:37 +0000291 For documentation, see the comments in the swarming_rpcs.TaskState enum, which
292 is the source of truth for these values:
293 https://cs.chromium.org/chromium/infra/luci/appengine/swarming/swarming_rpcs.py?q=TaskState\(
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400294
Marc-Antoine Ruel20b764d2018-06-22 18:08:37 +0000295 It's in fact an enum.
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400296 """
297 RUNNING = 0x10
298 PENDING = 0x20
299 EXPIRED = 0x30
300 TIMED_OUT = 0x40
301 BOT_DIED = 0x50
302 CANCELED = 0x60
303 COMPLETED = 0x70
Marc-Antoine Ruel2e52c552018-03-26 19:27:36 -0400304 KILLED = 0x80
Marc-Antoine Ruelfc708352018-05-04 20:25:43 -0400305 NO_RESOURCE = 0x100
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400306
Marc-Antoine Ruel20b764d2018-06-22 18:08:37 +0000307 STATES_RUNNING = ('PENDING', 'RUNNING')
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400308
maruel77f720b2015-09-15 12:35:22 -0700309 _ENUMS = {
310 'RUNNING': RUNNING,
311 'PENDING': PENDING,
312 'EXPIRED': EXPIRED,
313 'TIMED_OUT': TIMED_OUT,
314 'BOT_DIED': BOT_DIED,
315 'CANCELED': CANCELED,
316 'COMPLETED': COMPLETED,
Marc-Antoine Ruel2e52c552018-03-26 19:27:36 -0400317 'KILLED': KILLED,
Marc-Antoine Ruelfc708352018-05-04 20:25:43 -0400318 'NO_RESOURCE': NO_RESOURCE,
maruel77f720b2015-09-15 12:35:22 -0700319 }
320
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400321 @classmethod
maruel77f720b2015-09-15 12:35:22 -0700322 def from_enum(cls, state):
323 """Returns int value based on the string."""
324 if state not in cls._ENUMS:
325 raise ValueError('Invalid state %s' % state)
326 return cls._ENUMS[state]
327
maruel@chromium.org0437a732013-08-27 16:05:52 +0000328
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700329class TaskOutputCollector(object):
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700330 """Assembles task execution summary (for --task-summary-json output).
331
332 Optionally fetches task outputs from isolate server to local disk (used when
333 --task-output-dir is passed).
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700334
335 This object is shared among multiple threads running 'retrieve_results'
336 function, in particular they call 'process_shard_result' method in parallel.
337 """
338
Takuto Ikuta1e6072c2018-11-06 20:42:43 +0000339 def __init__(self, task_output_dir, task_output_stdout, shard_count,
340 filter_cb):
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700341 """Initializes TaskOutputCollector, ensures |task_output_dir| exists.
342
343 Args:
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700344 task_output_dir: (optional) local directory to put fetched files to.
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700345 shard_count: expected number of task shards.
346 """
maruel12e30012015-10-09 11:55:35 -0700347 self.task_output_dir = (
348 unicode(os.path.abspath(task_output_dir))
349 if task_output_dir else task_output_dir)
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000350 self.task_output_stdout = task_output_stdout
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700351 self.shard_count = shard_count
Takuto Ikuta1e6072c2018-11-06 20:42:43 +0000352 self.filter_cb = filter_cb
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700353
354 self._lock = threading.Lock()
355 self._per_shard_results = {}
356 self._storage = None
357
nodire5028a92016-04-29 14:38:21 -0700358 if self.task_output_dir:
359 file_path.ensure_tree(self.task_output_dir)
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700360
Vadim Shtayurab450c602014-05-12 19:23:25 -0700361 def process_shard_result(self, shard_index, result):
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700362 """Stores results of a single task shard, fetches output files if necessary.
363
Marc-Antoine Ruele4dcbb82014-10-01 09:30:56 -0400364 Modifies |result| in place.
365
maruel77f720b2015-09-15 12:35:22 -0700366 shard_index is 0-based.
367
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700368 Called concurrently from multiple threads.
369 """
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700370 # Sanity check index is in expected range.
Vadim Shtayurab450c602014-05-12 19:23:25 -0700371 assert isinstance(shard_index, int)
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700372 if shard_index < 0 or shard_index >= self.shard_count:
373 logging.warning(
374 'Shard index %d is outside of expected range: [0; %d]',
375 shard_index, self.shard_count - 1)
376 return
377
maruel77f720b2015-09-15 12:35:22 -0700378 if result.get('outputs_ref'):
379 ref = result['outputs_ref']
380 result['outputs_ref']['view_url'] = '%s/browse?%s' % (
381 ref['isolatedserver'],
382 urllib.urlencode(
383 [('namespace', ref['namespace']), ('hash', ref['isolated'])]))
Kevin Graneyc2c3b9e2014-08-26 09:04:17 -0400384
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700385 # Store result dict of that shard, ignore results we've already seen.
386 with self._lock:
387 if shard_index in self._per_shard_results:
388 logging.warning('Ignoring duplicate shard index %d', shard_index)
389 return
390 self._per_shard_results[shard_index] = result
391
392 # Fetch output files if necessary.
maruel77f720b2015-09-15 12:35:22 -0700393 if self.task_output_dir and result.get('outputs_ref'):
Marc-Antoine Ruele4dcbb82014-10-01 09:30:56 -0400394 storage = self._get_storage(
maruel77f720b2015-09-15 12:35:22 -0700395 result['outputs_ref']['isolatedserver'],
396 result['outputs_ref']['namespace'])
Marc-Antoine Ruele4dcbb82014-10-01 09:30:56 -0400397 if storage:
398 # Output files are supposed to be small and they are not reused across
Marc-Antoine Ruel2666d9c2018-05-18 13:52:02 -0400399 # tasks. So use MemoryContentAddressedCache for them instead of on-disk
400 # cache. Make files writable, so that calling script can delete them.
Marc-Antoine Ruele4dcbb82014-10-01 09:30:56 -0400401 isolateserver.fetch_isolated(
maruel77f720b2015-09-15 12:35:22 -0700402 result['outputs_ref']['isolated'],
Marc-Antoine Ruele4dcbb82014-10-01 09:30:56 -0400403 storage,
Marc-Antoine Ruel2666d9c2018-05-18 13:52:02 -0400404 local_caching.MemoryContentAddressedCache(file_mode_mask=0700),
maruel4409e302016-07-19 14:25:51 -0700405 os.path.join(self.task_output_dir, str(shard_index)),
Takuto Ikuta1e6072c2018-11-06 20:42:43 +0000406 False, self.filter_cb)
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700407
408 def finalize(self):
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700409 """Assembles and returns task summary JSON, shutdowns underlying Storage."""
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700410 with self._lock:
411 # Write an array of shard results with None for missing shards.
412 summary = {
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700413 'shards': [
414 self._per_shard_results.get(i) for i in xrange(self.shard_count)
415 ],
416 }
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000417
418 # Don't store stdout in the summary if not requested too.
419 if "json" not in self.task_output_stdout:
420 for shard_json in summary['shards']:
421 if not shard_json:
422 continue
423 if "output" in shard_json:
424 del shard_json["output"]
425 if "outputs" in shard_json:
426 del shard_json["outputs"]
427
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700428 # Write summary.json to task_output_dir as well.
429 if self.task_output_dir:
430 tools.write_json(
maruel12e30012015-10-09 11:55:35 -0700431 os.path.join(self.task_output_dir, u'summary.json'),
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700432 summary,
433 False)
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700434 if self._storage:
435 self._storage.close()
436 self._storage = None
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700437 return summary
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700438
439 def _get_storage(self, isolate_server, namespace):
440 """Returns isolateserver.Storage to use to fetch files."""
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700441 assert self.task_output_dir
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700442 with self._lock:
443 if not self._storage:
444 self._storage = isolateserver.get_storage(isolate_server, namespace)
445 else:
446 # Shards must all use exact same isolate server and namespace.
447 if self._storage.location != isolate_server:
448 logging.error(
449 'Task shards are using multiple isolate servers: %s and %s',
450 self._storage.location, isolate_server)
451 return None
452 if self._storage.namespace != namespace:
453 logging.error(
454 'Task shards are using multiple namespaces: %s and %s',
455 self._storage.namespace, namespace)
456 return None
457 return self._storage
458
459
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500460def now():
461 """Exists so it can be mocked easily."""
462 return time.time()
463
464
maruel77f720b2015-09-15 12:35:22 -0700465def parse_time(value):
466 """Converts serialized time from the API to datetime.datetime."""
467 # When microseconds are 0, the '.123456' suffix is elided. This means the
468 # serialized format is not consistent, which confuses the hell out of python.
469 for fmt in ('%Y-%m-%dT%H:%M:%S.%f', '%Y-%m-%dT%H:%M:%S'):
470 try:
471 return datetime.datetime.strptime(value, fmt)
472 except ValueError:
473 pass
474 raise ValueError('Failed to parse %s' % value)
475
476
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700477def retrieve_results(
maruel9531ce02016-04-13 06:11:23 -0700478 base_url, shard_index, task_id, timeout, should_stop, output_collector,
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000479 include_perf, fetch_stdout):
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400480 """Retrieves results for a single task ID.
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700481
Vadim Shtayurab450c602014-05-12 19:23:25 -0700482 Returns:
483 <result dict> on success.
484 None on failure.
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700485 """
maruel71c61c82016-02-22 06:52:05 -0800486 assert timeout is None or isinstance(timeout, float), timeout
smut281c3902018-05-30 17:50:05 -0700487 result_url = '%s/_ah/api/swarming/v1/task/%s/result' % (base_url, task_id)
maruel9531ce02016-04-13 06:11:23 -0700488 if include_perf:
489 result_url += '?include_performance_stats=true'
smut281c3902018-05-30 17:50:05 -0700490 output_url = '%s/_ah/api/swarming/v1/task/%s/stdout' % (base_url, task_id)
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700491 started = now()
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -0400492 deadline = started + timeout if timeout > 0 else None
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700493 attempt = 0
494
495 while not should_stop.is_set():
496 attempt += 1
497
498 # Waiting for too long -> give up.
499 current_time = now()
500 if deadline and current_time >= deadline:
501 logging.error('retrieve_results(%s) timed out on attempt %d',
502 base_url, attempt)
503 return None
504
505 # Do not spin too fast. Spin faster at the beginning though.
506 # Start with 1 sec delay and for each 30 sec of waiting add another second
507 # of delay, until hitting 15 sec ceiling.
508 if attempt > 1:
509 max_delay = min(15, 1 + (current_time - started) / 30.0)
510 delay = min(max_delay, deadline - current_time) if deadline else max_delay
511 if delay > 0:
512 logging.debug('Waiting %.1f sec before retrying', delay)
513 should_stop.wait(delay)
514 if should_stop.is_set():
515 return None
516
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400517 # Disable internal retries in net.url_read_json, since we are doing retries
518 # ourselves.
519 # TODO(maruel): We'd need to know if it's a 404 and not retry at all.
maruel0eb1d1b2015-10-02 14:48:21 -0700520 # TODO(maruel): Sadly, we currently have to poll here. Use hanging HTTP
521 # request on GAE v2.
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -0400522 # Retry on 500s only if no timeout is specified.
523 result = net.url_read_json(result_url, retry_50x=bool(timeout == -1))
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400524 if not result:
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -0400525 if timeout == -1:
526 return None
Marc-Antoine Ruel200b3952014-08-14 11:07:44 -0400527 continue
maruel77f720b2015-09-15 12:35:22 -0700528
maruelbf53e042015-12-01 15:00:51 -0800529 if result.get('error'):
530 # An error occurred.
531 if result['error'].get('errors'):
532 for err in result['error']['errors']:
533 logging.warning(
534 'Error while reading task: %s; %s',
535 err.get('message'), err.get('debugInfo'))
536 elif result['error'].get('message'):
537 logging.warning(
538 'Error while reading task: %s', result['error']['message'])
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -0400539 if timeout == -1:
540 return result
maruelbf53e042015-12-01 15:00:51 -0800541 continue
542
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -0400543 # When timeout == -1, always return on first attempt. 500s are already
544 # retried in this case.
Marc-Antoine Ruel20b764d2018-06-22 18:08:37 +0000545 if result['state'] not in TaskState.STATES_RUNNING or timeout == -1:
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000546 if fetch_stdout:
547 out = net.url_read_json(output_url)
Vadim Shtayura6fd3c7b2017-11-03 15:32:51 -0700548 result['output'] = out.get('output', '') if out else ''
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700549 # Record the result, try to fetch attached output files (if any).
550 if output_collector:
551 # TODO(vadimsh): Respect |should_stop| and |deadline| when fetching.
Vadim Shtayurab450c602014-05-12 19:23:25 -0700552 output_collector.process_shard_result(shard_index, result)
maruel77f720b2015-09-15 12:35:22 -0700553 if result.get('internal_failure'):
554 logging.error('Internal error!')
555 elif result['state'] == 'BOT_DIED':
556 logging.error('Bot died!')
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700557 return result
maruel@chromium.org0437a732013-08-27 16:05:52 +0000558
559
maruel77f720b2015-09-15 12:35:22 -0700560def convert_to_old_format(result):
561 """Converts the task result data from Endpoints API format to old API format
562 for compatibility.
563
564 This goes into the file generated as --task-summary-json.
565 """
566 # Sets default.
567 result.setdefault('abandoned_ts', None)
568 result.setdefault('bot_id', None)
569 result.setdefault('bot_version', None)
570 result.setdefault('children_task_ids', [])
571 result.setdefault('completed_ts', None)
572 result.setdefault('cost_saved_usd', None)
573 result.setdefault('costs_usd', None)
574 result.setdefault('deduped_from', None)
575 result.setdefault('name', None)
576 result.setdefault('outputs_ref', None)
maruel77f720b2015-09-15 12:35:22 -0700577 result.setdefault('server_versions', None)
578 result.setdefault('started_ts', None)
579 result.setdefault('tags', None)
580 result.setdefault('user', None)
581
582 # Convertion back to old API.
583 duration = result.pop('duration', None)
584 result['durations'] = [duration] if duration else []
585 exit_code = result.pop('exit_code', None)
586 result['exit_codes'] = [int(exit_code)] if exit_code else []
587 result['id'] = result.pop('task_id')
588 result['isolated_out'] = result.get('outputs_ref', None)
589 output = result.pop('output', None)
590 result['outputs'] = [output] if output else []
maruel77f720b2015-09-15 12:35:22 -0700591 # server_version
592 # Endpoints result 'state' as string. For compatibility with old code, convert
593 # to int.
Marc-Antoine Ruel20b764d2018-06-22 18:08:37 +0000594 result['state'] = TaskState.from_enum(result['state'])
maruel77f720b2015-09-15 12:35:22 -0700595 result['try_number'] = (
maruela4e8d752015-09-16 18:03:20 -0700596 int(result['try_number']) if result.get('try_number') else None)
maruel8786f2b2015-09-18 06:03:56 -0700597 if 'bot_dimensions' in result:
598 result['bot_dimensions'] = {
vadimsh72bf2532016-06-07 18:06:17 -0700599 i['key']: i.get('value', []) for i in result['bot_dimensions']
maruel8786f2b2015-09-18 06:03:56 -0700600 }
601 else:
602 result['bot_dimensions'] = None
maruel77f720b2015-09-15 12:35:22 -0700603
604
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700605def yield_results(
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400606 swarm_base_url, task_ids, timeout, max_threads, print_status_updates,
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000607 output_collector, include_perf, fetch_stdout):
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -0500608 """Yields swarming task results from the swarming server as (index, result).
maruel@chromium.org0437a732013-08-27 16:05:52 +0000609
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700610 Duplicate shards are ignored. Shards are yielded in order of completion.
611 Timed out shards are NOT yielded at all. Caller can compare number of yielded
612 shards with len(task_keys) to verify all shards completed.
maruel@chromium.org0437a732013-08-27 16:05:52 +0000613
614 max_threads is optional and is used to limit the number of parallel fetches
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -0500615 done. Since in general the number of task_keys is in the range <=10, it's not
maruel@chromium.org0437a732013-08-27 16:05:52 +0000616 worth normally to limit the number threads. Mostly used for testing purposes.
Marc-Antoine Ruel5c720342014-02-21 14:46:14 -0500617
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700618 output_collector is an optional instance of TaskOutputCollector that will be
619 used to fetch files produced by a task from isolate server to the local disk.
620
Marc-Antoine Ruel5c720342014-02-21 14:46:14 -0500621 Yields:
622 (index, result). In particular, 'result' is defined as the
623 GetRunnerResults() function in services/swarming/server/test_runner.py.
maruel@chromium.org0437a732013-08-27 16:05:52 +0000624 """
maruel@chromium.org0437a732013-08-27 16:05:52 +0000625 number_threads = (
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400626 min(max_threads, len(task_ids)) if max_threads else len(task_ids))
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700627 should_stop = threading.Event()
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700628 results_channel = threading_utils.TaskChannel()
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700629
maruel@chromium.org0437a732013-08-27 16:05:52 +0000630 with threading_utils.ThreadPool(number_threads, number_threads, 0) as pool:
631 try:
Vadim Shtayurab450c602014-05-12 19:23:25 -0700632 # Adds a task to the thread pool to call 'retrieve_results' and return
633 # the results together with shard_index that produced them (as a tuple).
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400634 def enqueue_retrieve_results(shard_index, task_id):
Marc-Antoine Ruel486c9b52018-07-23 19:30:47 +0000635 # pylint: disable=no-value-for-parameter
Vadim Shtayurab450c602014-05-12 19:23:25 -0700636 task_fn = lambda *args: (shard_index, retrieve_results(*args))
maruel@chromium.org0437a732013-08-27 16:05:52 +0000637 pool.add_task(
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400638 0, results_channel.wrap_task(task_fn), swarm_base_url, shard_index,
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000639 task_id, timeout, should_stop, output_collector, include_perf,
640 fetch_stdout)
Vadim Shtayurab450c602014-05-12 19:23:25 -0700641
642 # Enqueue 'retrieve_results' calls for each shard key to run in parallel.
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400643 for shard_index, task_id in enumerate(task_ids):
644 enqueue_retrieve_results(shard_index, task_id)
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700645
646 # Wait for all of them to finish.
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400647 shards_remaining = range(len(task_ids))
648 active_task_count = len(task_ids)
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700649 while active_task_count:
Vadim Shtayurab450c602014-05-12 19:23:25 -0700650 shard_index, result = None, None
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700651 try:
Vadim Shtayurab450c602014-05-12 19:23:25 -0700652 shard_index, result = results_channel.pull(
653 timeout=STATUS_UPDATE_INTERVAL)
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700654 except threading_utils.TaskChannel.Timeout:
655 if print_status_updates:
656 print(
657 'Waiting for results from the following shards: %s' %
658 ', '.join(map(str, shards_remaining)))
659 sys.stdout.flush()
660 continue
661 except Exception:
662 logging.exception('Unexpected exception in retrieve_results')
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700663
664 # A call to 'retrieve_results' finished (successfully or not).
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700665 active_task_count -= 1
maruel@chromium.org0437a732013-08-27 16:05:52 +0000666 if not result:
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -0500667 logging.error('Failed to retrieve the results for a swarming key')
maruel@chromium.org0437a732013-08-27 16:05:52 +0000668 continue
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700669
Vadim Shtayurab450c602014-05-12 19:23:25 -0700670 # Yield back results to the caller.
671 assert shard_index in shards_remaining
672 shards_remaining.remove(shard_index)
673 yield shard_index, result
Vadim Shtayurab19319e2014-04-27 08:50:06 -0700674
maruel@chromium.org0437a732013-08-27 16:05:52 +0000675 finally:
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700676 # Done or aborted with Ctrl+C, kill the remaining threads.
maruel@chromium.org0437a732013-08-27 16:05:52 +0000677 should_stop.set()
678
679
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000680def decorate_shard_output(swarming, shard_index, metadata, include_stdout):
maruel@chromium.org0437a732013-08-27 16:05:52 +0000681 """Returns wrapped output for swarming task shard."""
maruel77f720b2015-09-15 12:35:22 -0700682 if metadata.get('started_ts') and not metadata.get('deduped_from'):
Marc-Antoine Ruel5e6ccdb2015-04-02 15:55:13 -0400683 pending = '%.1fs' % (
maruel77f720b2015-09-15 12:35:22 -0700684 parse_time(metadata['started_ts']) - parse_time(metadata['created_ts'])
685 ).total_seconds()
Marc-Antoine Ruel3f9931a2017-11-03 14:34:49 -0400686 elif (metadata.get('state') in ('BOT_DIED', 'CANCELED', 'EXPIRED') and
687 metadata.get('abandoned_ts')):
688 pending = '%.1fs' % (
689 parse_time(metadata['abandoned_ts']) -
690 parse_time(metadata['created_ts'])
691 ).total_seconds()
Marc-Antoine Ruel5e6ccdb2015-04-02 15:55:13 -0400692 else:
693 pending = 'N/A'
694
maruel77f720b2015-09-15 12:35:22 -0700695 if metadata.get('duration') is not None:
696 duration = '%.1fs' % metadata['duration']
Marc-Antoine Ruel5e6ccdb2015-04-02 15:55:13 -0400697 else:
698 duration = 'N/A'
699
maruel77f720b2015-09-15 12:35:22 -0700700 if metadata.get('exit_code') is not None:
701 # Integers are encoded as string to not loose precision.
702 exit_code = '%s' % metadata['exit_code']
Marc-Antoine Ruel5e6ccdb2015-04-02 15:55:13 -0400703 else:
704 exit_code = 'N/A'
705
706 bot_id = metadata.get('bot_id') or 'N/A'
707
maruel77f720b2015-09-15 12:35:22 -0700708 url = '%s/user/task/%s' % (swarming, metadata['task_id'])
Marc-Antoine Ruel4e6b73d2014-10-03 18:00:05 -0400709 tag_header = 'Shard %d %s' % (shard_index, url)
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000710 tag_footer1 = 'End of shard %d' % (shard_index)
Marc-Antoine Ruel3f9931a2017-11-03 14:34:49 -0400711 if metadata.get('state') == 'CANCELED':
712 tag_footer2 = ' Pending: %s CANCELED' % pending
713 elif metadata.get('state') == 'EXPIRED':
714 tag_footer2 = ' Pending: %s EXPIRED (lack of capacity)' % pending
Marc-Antoine Ruel2e52c552018-03-26 19:27:36 -0400715 elif metadata.get('state') in ('BOT_DIED', 'TIMED_OUT', 'KILLED'):
Marc-Antoine Ruel3f9931a2017-11-03 14:34:49 -0400716 tag_footer2 = ' Pending: %s Duration: %s Bot: %s Exit: %s %s' % (
717 pending, duration, bot_id, exit_code, metadata['state'])
718 else:
719 tag_footer2 = ' Pending: %s Duration: %s Bot: %s Exit: %s' % (
720 pending, duration, bot_id, exit_code)
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400721
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000722 tag_len = max(len(x) for x in [tag_header, tag_footer1, tag_footer2])
723 dash_pad = '+-%s-+' % ('-' * tag_len)
724 tag_header = '| %s |' % tag_header.ljust(tag_len)
725 tag_footer1 = '| %s |' % tag_footer1.ljust(tag_len)
726 tag_footer2 = '| %s |' % tag_footer2.ljust(tag_len)
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400727
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000728 if include_stdout:
729 return '\n'.join([
730 dash_pad,
731 tag_header,
732 dash_pad,
Marc-Antoine Ruel3f9931a2017-11-03 14:34:49 -0400733 (metadata.get('output') or '').rstrip(),
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000734 dash_pad,
735 tag_footer1,
736 tag_footer2,
737 dash_pad,
738 ])
739 else:
740 return '\n'.join([
741 dash_pad,
742 tag_header,
743 tag_footer2,
744 dash_pad,
745 ])
maruel@chromium.org0437a732013-08-27 16:05:52 +0000746
747
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700748def collect(
maruel0eb1d1b2015-10-02 14:48:21 -0700749 swarming, task_ids, timeout, decorate, print_status_updates,
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000750 task_summary_json, task_output_dir, task_output_stdout,
Takuto Ikuta1e6072c2018-11-06 20:42:43 +0000751 include_perf, filepath_filter):
maruela5490782015-09-30 10:56:59 -0700752 """Retrieves results of a Swarming task.
753
754 Returns:
755 process exit code that should be returned to the user.
756 """
Takuto Ikuta1e6072c2018-11-06 20:42:43 +0000757
758 filter_cb = None
759 if filepath_filter:
760 filter_cb = re.compile(filepath_filter).match
761
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700762 # Collect summary JSON and output files (if task_output_dir is not None).
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000763 output_collector = TaskOutputCollector(
Takuto Ikuta1e6072c2018-11-06 20:42:43 +0000764 task_output_dir, task_output_stdout, len(task_ids), filter_cb)
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700765
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700766 seen_shards = set()
maruela5490782015-09-30 10:56:59 -0700767 exit_code = None
Marc-Antoine Rueld59e8072014-10-21 18:54:45 -0400768 total_duration = 0
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700769 try:
Marc-Antoine Ruel5e6ccdb2015-04-02 15:55:13 -0400770 for index, metadata in yield_results(
Marc-Antoine Ruel4e6b73d2014-10-03 18:00:05 -0400771 swarming, task_ids, timeout, None, print_status_updates,
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000772 output_collector, include_perf,
773 (len(task_output_stdout) > 0),
774 ):
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700775 seen_shards.add(index)
Vadim Shtayura473455a2014-05-14 15:22:35 -0700776
Marc-Antoine Ruel5e6ccdb2015-04-02 15:55:13 -0400777 # Default to failure if there was no process that even started.
maruel77f720b2015-09-15 12:35:22 -0700778 shard_exit_code = metadata.get('exit_code')
779 if shard_exit_code:
maruela5490782015-09-30 10:56:59 -0700780 # It's encoded as a string, so bool('0') is True.
maruel77f720b2015-09-15 12:35:22 -0700781 shard_exit_code = int(shard_exit_code)
maruela5490782015-09-30 10:56:59 -0700782 if shard_exit_code or exit_code is None:
Marc-Antoine Ruel4e6b73d2014-10-03 18:00:05 -0400783 exit_code = shard_exit_code
maruel77f720b2015-09-15 12:35:22 -0700784 total_duration += metadata.get('duration', 0)
Vadim Shtayura473455a2014-05-14 15:22:35 -0700785
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700786 if decorate:
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000787 s = decorate_shard_output(
788 swarming, index, metadata,
789 "console" in task_output_stdout).encode(
790 'utf-8', 'replace')
leileied181762016-10-13 14:24:59 -0700791 print(s)
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400792 if len(seen_shards) < len(task_ids):
793 print('')
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700794 else:
maruel77f720b2015-09-15 12:35:22 -0700795 print('%s: %s %s' % (
796 metadata.get('bot_id', 'N/A'),
797 metadata['task_id'],
798 shard_exit_code))
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +1000799 if "console" in task_output_stdout and metadata['output']:
maruel77f720b2015-09-15 12:35:22 -0700800 output = metadata['output'].rstrip()
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400801 if output:
802 print(''.join(' %s\n' % l for l in output.splitlines()))
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700803 finally:
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700804 summary = output_collector.finalize()
805 if task_summary_json:
maruel77f720b2015-09-15 12:35:22 -0700806 # TODO(maruel): Make this optional.
807 for i in summary['shards']:
808 if i:
809 convert_to_old_format(i)
Vadim Shtayurac8437bf2014-07-09 19:45:36 -0700810 tools.write_json(task_summary_json, summary, False)
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700811
Marc-Antoine Rueld59e8072014-10-21 18:54:45 -0400812 if decorate and total_duration:
813 print('Total duration: %.1fs' % total_duration)
814
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -0400815 if len(seen_shards) != len(task_ids):
816 missing_shards = [x for x in range(len(task_ids)) if x not in seen_shards]
Vadim Shtayura86a2cef2014-04-18 11:13:39 -0700817 print >> sys.stderr, ('Results from some shards are missing: %s' %
818 ', '.join(map(str, missing_shards)))
Vadim Shtayurac524f512014-05-15 09:54:56 -0700819 return 1
Vadim Shtayurae3fbd102014-04-29 17:05:21 -0700820
maruela5490782015-09-30 10:56:59 -0700821 return exit_code if exit_code is not None else 1
maruel@chromium.org0437a732013-08-27 16:05:52 +0000822
823
maruel77f720b2015-09-15 12:35:22 -0700824### API management.
825
826
827class APIError(Exception):
828 pass
829
830
831def endpoints_api_discovery_apis(host):
832 """Uses Cloud Endpoints' API Discovery Service to returns metadata about all
833 the APIs exposed by a host.
834
835 https://developers.google.com/discovery/v1/reference/apis/list
836 """
maruel380e3262016-08-31 16:10:06 -0700837 # Uses the real Cloud Endpoints. This needs to be fixed once the Cloud
838 # Endpoints version is turned down.
maruel77f720b2015-09-15 12:35:22 -0700839 data = net.url_read_json(host + '/_ah/api/discovery/v1/apis')
840 if data is None:
841 raise APIError('Failed to discover APIs on %s' % host)
842 out = {}
843 for api in data['items']:
844 if api['id'] == 'discovery:v1':
845 continue
846 # URL is of the following form:
847 # url = host + (
848 # '/_ah/api/discovery/v1/apis/%s/%s/rest' % (api['id'], api['version'])
849 api_data = net.url_read_json(api['discoveryRestUrl'])
850 if api_data is None:
851 raise APIError('Failed to discover %s on %s' % (api['id'], host))
852 out[api['id']] = api_data
853 return out
854
855
maruelaf6b06c2017-06-08 06:26:53 -0700856def get_yielder(base_url, limit):
857 """Returns the first query and a function that yields following items."""
858 CHUNK_SIZE = 250
859
860 url = base_url
861 if limit:
862 url += '%slimit=%d' % ('&' if '?' in url else '?', min(CHUNK_SIZE, limit))
863 data = net.url_read_json(url)
864 if data is None:
865 # TODO(maruel): Do basic diagnostic.
866 raise Failure('Failed to access %s' % url)
867 org_cursor = data.pop('cursor', None)
868 org_total = len(data.get('items') or [])
869 logging.info('get_yielder(%s) returning %d items', base_url, org_total)
870 if not org_cursor or not org_total:
871 # This is not an iterable resource.
872 return data, lambda: []
873
874 def yielder():
875 cursor = org_cursor
876 total = org_total
877 # Some items support cursors. Try to get automatically if cursors are needed
878 # by looking at the 'cursor' items.
879 while cursor and (not limit or total < limit):
880 merge_char = '&' if '?' in base_url else '?'
881 url = base_url + '%scursor=%s' % (merge_char, urllib.quote(cursor))
882 if limit:
883 url += '&limit=%d' % min(CHUNK_SIZE, limit - total)
884 new = net.url_read_json(url)
885 if new is None:
886 raise Failure('Failed to access %s' % url)
887 cursor = new.get('cursor')
888 new_items = new.get('items')
889 nb_items = len(new_items or [])
890 total += nb_items
891 logging.info('get_yielder(%s) yielding %d items', base_url, nb_items)
892 yield new_items
893
894 return data, yielder
895
896
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500897### Commands.
898
899
900def abort_task(_swarming, _manifest):
901 """Given a task manifest that was triggered, aborts its execution."""
902 # TODO(vadimsh): No supported by the server yet.
903
904
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -0400905def add_filter_options(parser):
maruel681d6802017-01-17 16:56:03 -0800906 parser.filter_group = optparse.OptionGroup(parser, 'Bot selection')
Marc-Antoine Ruel5471e3d2013-11-11 19:10:32 -0500907 parser.filter_group.add_option(
Marc-Antoine Ruelb39e8cf2014-01-20 10:39:31 -0500908 '-d', '--dimension', default=[], action='append', nargs=2,
Marc-Antoine Ruel92f32422013-11-06 18:12:13 -0500909 dest='dimensions', metavar='FOO bar',
910 help='dimension to filter on')
Brad Hallf78187a2018-10-19 17:08:55 +0000911 parser.filter_group.add_option(
912 '--optional-dimension', default=[], action='append', nargs=3,
913 dest='optional_dimensions', metavar='key value expiration',
914 help='optional dimensions which will result in additional task slices ')
Marc-Antoine Ruel5471e3d2013-11-11 19:10:32 -0500915 parser.add_option_group(parser.filter_group)
916
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -0400917
Brad Hallf78187a2018-10-19 17:08:55 +0000918def _validate_filter_option(parser, key, value, expiration, argname):
919 if ':' in key:
920 parser.error('%s key cannot contain ":"' % argname)
921 if key.strip() != key:
922 parser.error('%s key has whitespace' % argname)
923 if not key:
924 parser.error('%s key is empty' % argname)
925
926 if value.strip() != value:
927 parser.error('%s value has whitespace' % argname)
928 if not value:
929 parser.error('%s value is empty' % argname)
930
931 if expiration is not None:
932 try:
933 expiration = int(expiration)
934 except ValueError:
935 parser.error('%s expiration is not an integer' % argname)
936 if expiration <= 0:
937 parser.error('%s expiration should be positive' % argname)
938 if expiration % 60 != 0:
939 parser.error('%s expiration is not divisible by 60' % argname)
940
941
maruelaf6b06c2017-06-08 06:26:53 -0700942def process_filter_options(parser, options):
943 for key, value in options.dimensions:
Brad Hallf78187a2018-10-19 17:08:55 +0000944 _validate_filter_option(parser, key, value, None, 'dimension')
945 for key, value, exp in options.optional_dimensions:
946 _validate_filter_option(parser, key, value, exp, 'optional-dimension')
maruelaf6b06c2017-06-08 06:26:53 -0700947 options.dimensions.sort()
948
949
Vadim Shtayurab450c602014-05-12 19:23:25 -0700950def add_sharding_options(parser):
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -0400951 parser.sharding_group = optparse.OptionGroup(parser, 'Sharding options')
Vadim Shtayurab450c602014-05-12 19:23:25 -0700952 parser.sharding_group.add_option(
maruel5475ba62017-05-31 15:35:47 -0700953 '--shards', type='int', default=1, metavar='NUMBER',
Vadim Shtayurab450c602014-05-12 19:23:25 -0700954 help='Number of shards to trigger and collect.')
955 parser.add_option_group(parser.sharding_group)
956
957
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -0400958def add_trigger_options(parser):
959 """Adds all options to trigger a task on Swarming."""
Marc-Antoine Ruelf7d737d2014-12-10 15:36:29 -0500960 isolateserver.add_isolate_server_options(parser)
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -0400961 add_filter_options(parser)
962
Marc-Antoine Ruel1f835c72018-05-25 12:29:42 -0400963 group = optparse.OptionGroup(parser, 'TaskSlice properties')
maruel681d6802017-01-17 16:56:03 -0800964 group.add_option(
maruel5475ba62017-05-31 15:35:47 -0700965 '-s', '--isolated', metavar='HASH',
Marc-Antoine Ruel185ded42015-01-28 20:49:18 -0500966 help='Hash of the .isolated to grab from the isolate server')
maruel681d6802017-01-17 16:56:03 -0800967 group.add_option(
Marc-Antoine Ruel5471e3d2013-11-11 19:10:32 -0500968 '-e', '--env', default=[], action='append', nargs=2, metavar='FOO bar',
Vadim Shtayurab450c602014-05-12 19:23:25 -0700969 help='Environment variables to set')
maruel681d6802017-01-17 16:56:03 -0800970 group.add_option(
Robert Iannuccibf5f84c2017-11-22 12:56:50 -0800971 '--env-prefix', default=[], action='append', nargs=2,
972 metavar='VAR local/path',
973 help='Prepend task-relative `local/path` to the task\'s VAR environment '
974 'variable using os-appropriate pathsep character. Can be specified '
975 'multiple times for the same VAR to add multiple paths.')
976 group.add_option(
Marc-Antoine Ruel02196392014-10-17 16:29:43 -0400977 '--idempotent', action='store_true', default=False,
978 help='When set, the server will actively try to find a previous task '
979 'with the same parameter and return this result instead if possible')
maruel681d6802017-01-17 16:56:03 -0800980 group.add_option(
maruel5475ba62017-05-31 15:35:47 -0700981 '--secret-bytes-path', metavar='FILE',
iannuccidc80dfb2016-10-28 12:50:20 -0700982 help='The optional path to a file containing the secret_bytes to use with'
983 'this task.')
maruel681d6802017-01-17 16:56:03 -0800984 group.add_option(
maruel5475ba62017-05-31 15:35:47 -0700985 '--hard-timeout', type='int', default=60*60, metavar='SECS',
Marc-Antoine Ruel2f6581a2014-10-03 11:09:53 -0400986 help='Seconds to allow the task to complete.')
maruel681d6802017-01-17 16:56:03 -0800987 group.add_option(
maruel5475ba62017-05-31 15:35:47 -0700988 '--io-timeout', type='int', default=20*60, metavar='SECS',
Marc-Antoine Ruel2f6581a2014-10-03 11:09:53 -0400989 help='Seconds to allow the task to be silent.')
maruel681d6802017-01-17 16:56:03 -0800990 group.add_option(
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -0500991 '--raw-cmd', action='store_true', default=False,
992 help='When set, the command after -- is used as-is without run_isolated. '
maruel0a25f6c2017-05-10 10:43:23 -0700993 'In this case, the .isolated file is expected to not have a command')
maruel681d6802017-01-17 16:56:03 -0800994 group.add_option(
Marc-Antoine Ruelba1bf222017-12-21 21:41:01 -0500995 '--relative-cwd',
996 help='Ignore the isolated \'relative_cwd\' and use this one instead; '
997 'requires --raw-cmd')
998 group.add_option(
maruel5475ba62017-05-31 15:35:47 -0700999 '--cipd-package', action='append', default=[], metavar='PKG',
1000 help='CIPD packages to install on the Swarming bot. Uses the format: '
borenet02f772b2016-06-22 12:42:19 -07001001 'path:package_name:version')
maruel681d6802017-01-17 16:56:03 -08001002 group.add_option(
1003 '--named-cache', action='append', nargs=2, default=[],
maruel5475ba62017-05-31 15:35:47 -07001004 metavar='NAME RELPATH',
maruel681d6802017-01-17 16:56:03 -08001005 help='"<name> <relpath>" items to keep a persistent bot managed cache')
1006 group.add_option(
vadimsh93d167c2016-09-13 11:31:51 -07001007 '--service-account',
Vadim Shtayura2d83a942017-08-14 17:41:24 -07001008 help='Email of a service account to run the task as, or literal "bot" '
1009 'string to indicate that the task should use the same account the '
1010 'bot itself is using to authenticate to Swarming. Don\'t use task '
1011 'service accounts if not given (default).')
maruel681d6802017-01-17 16:56:03 -08001012 group.add_option(
Robert Iannuccifafa7352018-06-13 17:08:17 +00001013 '--pool-task-template',
1014 choices=('AUTO', 'CANARY_PREFER', 'CANARY_NEVER', 'SKIP'),
1015 default='AUTO',
1016 help='Set how you want swarming to apply the pool\'s TaskTemplate. '
1017 'By default, the pool\'s TaskTemplate is automatically selected, '
1018 'according the pool configuration on the server. Choices are: '
1019 'AUTO, CANARY_PREFER, CANARY_NEVER, and SKIP (default: AUTO).')
1020 group.add_option(
maruel5475ba62017-05-31 15:35:47 -07001021 '-o', '--output', action='append', default=[], metavar='PATH',
1022 help='A list of files to return in addition to those written to '
1023 '${ISOLATED_OUTDIR}. An error will occur if a file specified by'
1024 'this option is also written directly to ${ISOLATED_OUTDIR}.')
Marc-Antoine Ruel1f835c72018-05-25 12:29:42 -04001025 group.add_option(
1026 '--wait-for-capacity', action='store_true', default=False,
1027 help='Instructs to leave the task PENDING even if there\'s no known bot '
1028 'that could run this task, otherwise the task will be denied with '
1029 'NO_RESOURCE')
maruel681d6802017-01-17 16:56:03 -08001030 parser.add_option_group(group)
1031
Marc-Antoine Ruel1f835c72018-05-25 12:29:42 -04001032 group = optparse.OptionGroup(parser, 'TaskRequest details')
maruel681d6802017-01-17 16:56:03 -08001033 group.add_option(
Marc-Antoine Ruel486c9b52018-07-23 19:30:47 +00001034 '--priority', type='int', default=200,
maruel681d6802017-01-17 16:56:03 -08001035 help='The lower value, the more important the task is')
1036 group.add_option(
maruel5475ba62017-05-31 15:35:47 -07001037 '-T', '--task-name', metavar='NAME',
maruel681d6802017-01-17 16:56:03 -08001038 help='Display name of the task. Defaults to '
1039 '<base_name>/<dimensions>/<isolated hash>/<timestamp> if an '
1040 'isolated file is provided, if a hash is provided, it defaults to '
1041 '<user>/<dimensions>/<isolated hash>/<timestamp>')
1042 group.add_option(
maruel5475ba62017-05-31 15:35:47 -07001043 '--tags', action='append', default=[], metavar='FOO:BAR',
maruel681d6802017-01-17 16:56:03 -08001044 help='Tags to assign to the task.')
1045 group.add_option(
1046 '--user', default='',
1047 help='User associated with the task. Defaults to authenticated user on '
1048 'the server.')
1049 group.add_option(
maruel5475ba62017-05-31 15:35:47 -07001050 '--expiration', type='int', default=6*60*60, metavar='SECS',
maruel681d6802017-01-17 16:56:03 -08001051 help='Seconds to allow the task to be pending for a bot to run before '
1052 'this task request expires.')
1053 group.add_option(
1054 '--deadline', type='int', dest='expiration',
1055 help=optparse.SUPPRESS_HELP)
1056 parser.add_option_group(group)
maruel@chromium.org0437a732013-08-27 16:05:52 +00001057
1058
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001059def process_trigger_options(parser, options, args):
Marc-Antoine Ruel3a030bc2018-04-23 10:31:25 -04001060 """Processes trigger options and does preparatory steps.
1061
1062 Returns:
1063 NewTaskRequest instance.
1064 """
maruelaf6b06c2017-06-08 06:26:53 -07001065 process_filter_options(parser, options)
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001066 options.env = dict(options.env)
maruel0a25f6c2017-05-10 10:43:23 -07001067 if args and args[0] == '--':
1068 args = args[1:]
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001069
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001070 if not options.dimensions:
1071 parser.error('Please at least specify one --dimension')
maruel0a25f6c2017-05-10 10:43:23 -07001072 if not all(len(t.split(':', 1)) == 2 for t in options.tags):
1073 parser.error('--tags must be in the format key:value')
1074 if options.raw_cmd and not args:
1075 parser.error(
1076 'Arguments with --raw-cmd should be passed after -- as command '
1077 'delimiter.')
1078 if options.isolate_server and not options.namespace:
1079 parser.error(
1080 '--namespace must be a valid value when --isolate-server is used')
1081 if not options.isolated and not options.raw_cmd:
1082 parser.error('Specify at least one of --raw-cmd or --isolated or both')
1083
1084 # Isolated
1085 # --isolated is required only if --raw-cmd wasn't provided.
1086 # TODO(maruel): --isolate-server may be optional as Swarming may have its own
1087 # preferred server.
1088 isolateserver.process_isolate_server_options(
1089 parser, options, False, not options.raw_cmd)
1090 inputs_ref = None
1091 if options.isolate_server:
1092 inputs_ref = FilesRef(
1093 isolated=options.isolated,
1094 isolatedserver=options.isolate_server,
1095 namespace=options.namespace)
1096
1097 # Command
1098 command = None
1099 extra_args = None
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001100 if options.raw_cmd:
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001101 command = args
Marc-Antoine Ruelba1bf222017-12-21 21:41:01 -05001102 if options.relative_cwd:
1103 a = os.path.normpath(os.path.abspath(options.relative_cwd))
1104 if not a.startswith(os.getcwd()):
1105 parser.error(
1106 '--relative-cwd must not try to escape the working directory')
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001107 else:
Marc-Antoine Ruelba1bf222017-12-21 21:41:01 -05001108 if options.relative_cwd:
1109 parser.error('--relative-cwd requires --raw-cmd')
maruel0a25f6c2017-05-10 10:43:23 -07001110 extra_args = args
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001111
maruel0a25f6c2017-05-10 10:43:23 -07001112 # CIPD
borenet02f772b2016-06-22 12:42:19 -07001113 cipd_packages = []
1114 for p in options.cipd_package:
1115 split = p.split(':', 2)
1116 if len(split) != 3:
1117 parser.error('CIPD packages must take the form: path:package:version')
1118 cipd_packages.append(CipdPackage(
1119 package_name=split[1],
1120 path=split[0],
1121 version=split[2]))
1122 cipd_input = None
1123 if cipd_packages:
1124 cipd_input = CipdInput(
1125 client_package=None,
1126 packages=cipd_packages,
1127 server=None)
1128
maruel0a25f6c2017-05-10 10:43:23 -07001129 # Secrets
iannuccidc80dfb2016-10-28 12:50:20 -07001130 secret_bytes = None
1131 if options.secret_bytes_path:
Marc-Antoine Ruel5c98fa72018-05-18 12:19:59 -04001132 with open(options.secret_bytes_path, 'rb') as f:
iannuccidc80dfb2016-10-28 12:50:20 -07001133 secret_bytes = f.read().encode('base64')
1134
maruel0a25f6c2017-05-10 10:43:23 -07001135 # Named caches
maruel681d6802017-01-17 16:56:03 -08001136 caches = [
1137 {u'name': unicode(i[0]), u'path': unicode(i[1])}
1138 for i in options.named_cache
1139 ]
maruel0a25f6c2017-05-10 10:43:23 -07001140
Robert Iannuccibf5f84c2017-11-22 12:56:50 -08001141 env_prefixes = {}
1142 for k, v in options.env_prefix:
1143 env_prefixes.setdefault(k, []).append(v)
1144
Brad Hallf78187a2018-10-19 17:08:55 +00001145 # Get dimensions into the key/value format we can manipulate later.
1146 orig_dims = [
1147 {'key': key, 'value': value} for key, value in options.dimensions]
1148 orig_dims.sort(key=lambda x: (x['key'], x['value']))
1149
1150 # Construct base properties that we will use for all the slices, adding in
1151 # optional dimensions for the fallback slices.
maruel77f720b2015-09-15 12:35:22 -07001152 properties = TaskProperties(
maruel681d6802017-01-17 16:56:03 -08001153 caches=caches,
borenet02f772b2016-06-22 12:42:19 -07001154 cipd_input=cipd_input,
maruel0a25f6c2017-05-10 10:43:23 -07001155 command=command,
Marc-Antoine Ruelba1bf222017-12-21 21:41:01 -05001156 relative_cwd=options.relative_cwd,
Brad Hallf78187a2018-10-19 17:08:55 +00001157 dimensions=orig_dims,
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001158 env=options.env,
Robert Iannuccibf5f84c2017-11-22 12:56:50 -08001159 env_prefixes=[StringListPair(k, v) for k, v in env_prefixes.iteritems()],
maruel77f720b2015-09-15 12:35:22 -07001160 execution_timeout_secs=options.hard_timeout,
maruel0a25f6c2017-05-10 10:43:23 -07001161 extra_args=extra_args,
maruel77f720b2015-09-15 12:35:22 -07001162 grace_period_secs=30,
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001163 idempotent=options.idempotent,
maruel77f720b2015-09-15 12:35:22 -07001164 inputs_ref=inputs_ref,
aludwincc5524e2016-10-28 10:25:24 -07001165 io_timeout_secs=options.io_timeout,
iannuccidc80dfb2016-10-28 12:50:20 -07001166 outputs=options.output,
1167 secret_bytes=secret_bytes)
Brad Hallf78187a2018-10-19 17:08:55 +00001168
1169 slices = []
1170
1171 # Group the optional dimensions by expiration.
1172 dims_by_exp = {}
1173 for key, value, exp_secs in options.optional_dimensions:
1174 dims_by_exp.setdefault(int(exp_secs), []).append(
1175 {'key': key, 'value': value})
1176
1177 # Create the optional slices with expiration deltas, we fix up the properties
1178 # below.
1179 last_exp = 0
1180 for expiration_secs in sorted(dims_by_exp):
1181 t = TaskSlice(
1182 expiration_secs=expiration_secs - last_exp,
1183 properties=properties,
1184 wait_for_capacity=False)
1185 slices.append(t)
1186 last_exp = expiration_secs
1187
1188 # Add back in the default slice (the last one).
1189 exp = max(int(options.expiration) - last_exp, 60)
1190 base_task_slice = TaskSlice(
1191 expiration_secs=exp,
Marc-Antoine Ruel1f835c72018-05-25 12:29:42 -04001192 properties=properties,
1193 wait_for_capacity=options.wait_for_capacity)
Brad Hallf78187a2018-10-19 17:08:55 +00001194 slices.append(base_task_slice)
1195
Brad Hall7f463e62018-11-16 16:13:30 +00001196 # Add optional dimensions to the task slices, replacing a dimension that
1197 # has the same key if it is a dimension where repeating isn't valid (otherwise
1198 # we append it). Currently the only dimension we can repeat is "caches"; the
1199 # rest (os, cpu, etc) shouldn't be repeated.
Brad Hallf78187a2018-10-19 17:08:55 +00001200 extra_dims = []
Brad Hall7f463e62018-11-16 16:13:30 +00001201 for i, (_, kvs) in enumerate(sorted(dims_by_exp.iteritems(), reverse=True)):
Brad Hallf78187a2018-10-19 17:08:55 +00001202 dims = list(orig_dims)
Brad Hall7f463e62018-11-16 16:13:30 +00001203 # Replace or append the key/value pairs for this expiration in extra_dims;
1204 # we keep extra_dims around because we are iterating backwards and filling
1205 # in slices with shorter expirations. Dimensions expire as time goes on so
1206 # the slices that expire earlier will generally have more dimensions.
1207 for kv in kvs:
1208 if kv['key'] == 'caches':
1209 extra_dims.append(kv)
1210 else:
1211 extra_dims = [x for x in extra_dims if x['key'] != kv['key']] + [kv]
1212 # Then, add all the optional dimensions to the original dimension set, again
1213 # replacing if needed.
1214 for kv in extra_dims:
1215 if kv['key'] == 'caches':
1216 dims.append(kv)
1217 else:
1218 dims = [x for x in dims if x['key'] != kv['key']] + [kv]
Brad Hallf78187a2018-10-19 17:08:55 +00001219 dims.sort(key=lambda x: (x['key'], x['value']))
1220 slice_properties = properties._replace(dimensions=dims)
1221 slices[-2 - i] = slices[-2 - i]._replace(properties=slice_properties)
1222
maruel77f720b2015-09-15 12:35:22 -07001223 return NewTaskRequest(
maruel0a25f6c2017-05-10 10:43:23 -07001224 name=default_task_name(options),
maruel77f720b2015-09-15 12:35:22 -07001225 parent_task_id=os.environ.get('SWARMING_TASK_ID', ''),
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001226 priority=options.priority,
Brad Hallf78187a2018-10-19 17:08:55 +00001227 task_slices=slices,
Vadim Shtayura2d83a942017-08-14 17:41:24 -07001228 service_account=options.service_account,
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001229 tags=options.tags,
Robert Iannuccifafa7352018-06-13 17:08:17 +00001230 user=options.user,
1231 pool_task_template=options.pool_task_template)
maruel@chromium.org0437a732013-08-27 16:05:52 +00001232
1233
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +10001234class TaskOutputStdoutOption(optparse.Option):
1235 """Where to output the each task's console output (stderr/stdout).
1236
1237 The output will be;
1238 none - not be downloaded.
1239 json - stored in summary.json file *only*.
1240 console - shown on stdout *only*.
1241 all - stored in summary.json and shown on stdout.
1242 """
1243
1244 choices = ['all', 'json', 'console', 'none']
1245
1246 def __init__(self, *args, **kw):
1247 optparse.Option.__init__(
1248 self,
1249 *args,
1250 choices=self.choices,
Marc-Antoine Ruel28488842017-09-12 18:09:17 -04001251 default=['console', 'json'],
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +10001252 help=re.sub('\s\s*', ' ', self.__doc__),
1253 **kw)
1254
1255 def convert_value(self, opt, value):
1256 if value not in self.choices:
1257 raise optparse.OptionValueError("%s must be one of %s not %r" % (
1258 self.get_opt_string(), self.choices, value))
1259 stdout_to = []
1260 if value == 'all':
Marc-Antoine Ruel28488842017-09-12 18:09:17 -04001261 stdout_to = ['console', 'json']
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +10001262 elif value != 'none':
1263 stdout_to = [value]
1264 return stdout_to
1265
1266
maruel@chromium.org0437a732013-08-27 16:05:52 +00001267def add_collect_options(parser):
Marc-Antoine Ruel5471e3d2013-11-11 19:10:32 -05001268 parser.server_group.add_option(
Marc-Antoine Ruele831f052018-04-20 15:01:03 -04001269 '-t', '--timeout', type='float', default=0.,
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001270 help='Timeout to wait for result, set to -1 for no timeout and get '
1271 'current state; defaults to waiting until the task completes')
Marc-Antoine Ruel5471e3d2013-11-11 19:10:32 -05001272 parser.group_logging.add_option(
1273 '--decorate', action='store_true', help='Decorate output')
Vadim Shtayura86a2cef2014-04-18 11:13:39 -07001274 parser.group_logging.add_option(
1275 '--print-status-updates', action='store_true',
1276 help='Print periodic status updates')
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -04001277 parser.task_output_group = optparse.OptionGroup(parser, 'Task output')
Vadim Shtayurae3fbd102014-04-29 17:05:21 -07001278 parser.task_output_group.add_option(
Vadim Shtayurac8437bf2014-07-09 19:45:36 -07001279 '--task-summary-json',
1280 metavar='FILE',
1281 help='Dump a summary of task results to this file as json. It contains '
1282 'only shards statuses as know to server directly. Any output files '
1283 'emitted by the task can be collected by using --task-output-dir')
1284 parser.task_output_group.add_option(
Vadim Shtayurae3fbd102014-04-29 17:05:21 -07001285 '--task-output-dir',
Vadim Shtayurac8437bf2014-07-09 19:45:36 -07001286 metavar='DIR',
Vadim Shtayurae3fbd102014-04-29 17:05:21 -07001287 help='Directory to put task results into. When the task finishes, this '
Vadim Shtayurac8437bf2014-07-09 19:45:36 -07001288 'directory contains per-shard directory with output files produced '
1289 'by shards: <task-output-dir>/<zero-based-shard-index>/.')
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +10001290 parser.task_output_group.add_option(TaskOutputStdoutOption(
Marc-Antoine Ruel28488842017-09-12 18:09:17 -04001291 '--task-output-stdout'))
maruel9531ce02016-04-13 06:11:23 -07001292 parser.task_output_group.add_option(
Takuto Ikuta1e6072c2018-11-06 20:42:43 +00001293 '--filepath-filter',
1294 help='This is regexp filter used to specify downloaded filepath when '
1295 'collecting isolated output.')
1296 parser.task_output_group.add_option(
maruel9531ce02016-04-13 06:11:23 -07001297 '--perf', action='store_true', default=False,
1298 help='Includes performance statistics')
Vadim Shtayurae3fbd102014-04-29 17:05:21 -07001299 parser.add_option_group(parser.task_output_group)
maruel@chromium.org0437a732013-08-27 16:05:52 +00001300
1301
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001302def process_collect_options(parser, options):
1303 # Only negative -1 is allowed, disallow other negative values.
1304 if options.timeout != -1 and options.timeout < 0:
1305 parser.error('Invalid --timeout value')
1306
1307
Marc-Antoine Ruel13e7c882015-03-26 18:19:10 -04001308@subcommand.usage('bots...')
1309def CMDbot_delete(parser, args):
1310 """Forcibly deletes bots from the Swarming server."""
1311 parser.add_option(
1312 '-f', '--force', action='store_true',
1313 help='Do not prompt for confirmation')
1314 options, args = parser.parse_args(args)
1315 if not args:
maruelfd0a90c2016-06-10 11:51:10 -07001316 parser.error('Please specify bots to delete')
Marc-Antoine Ruel13e7c882015-03-26 18:19:10 -04001317
1318 bots = sorted(args)
1319 if not options.force:
1320 print('Delete the following bots?')
1321 for bot in bots:
1322 print(' %s' % bot)
1323 if raw_input('Continue? [y/N] ') not in ('y', 'Y'):
1324 print('Goodbye.')
1325 return 1
1326
1327 result = 0
1328 for bot in bots:
smut281c3902018-05-30 17:50:05 -07001329 url = '%s/_ah/api/swarming/v1/bot/%s/delete' % (options.swarming, bot)
vadimshe4c0e242015-09-30 11:53:54 -07001330 if net.url_read_json(url, data={}, method='POST') is None:
1331 print('Deleting %s failed. Probably already gone' % bot)
Marc-Antoine Ruel13e7c882015-03-26 18:19:10 -04001332 result = 1
1333 return result
1334
1335
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001336def CMDbots(parser, args):
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -04001337 """Returns information about the bots connected to the Swarming server."""
1338 add_filter_options(parser)
1339 parser.filter_group.add_option(
Marc-Antoine Ruel28083112014-03-13 16:34:04 -04001340 '--dead-only', action='store_true',
maruelaf6b06c2017-06-08 06:26:53 -07001341 help='Filter out bots alive, useful to reap them and reimage broken bots')
Marc-Antoine Ruel28083112014-03-13 16:34:04 -04001342 parser.filter_group.add_option(
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -04001343 '-k', '--keep-dead', action='store_true',
maruelaf6b06c2017-06-08 06:26:53 -07001344 help='Keep both dead and alive bots')
1345 parser.filter_group.add_option(
1346 '--busy', action='store_true', help='Keep only busy bots')
1347 parser.filter_group.add_option(
1348 '--idle', action='store_true', help='Keep only idle bots')
1349 parser.filter_group.add_option(
1350 '--mp', action='store_true',
1351 help='Keep only Machine Provider managed bots')
1352 parser.filter_group.add_option(
1353 '--non-mp', action='store_true',
1354 help='Keep only non Machine Provider managed bots')
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -04001355 parser.filter_group.add_option(
1356 '-b', '--bare', action='store_true',
Marc-Antoine Ruele7b00162014-03-12 16:59:01 -04001357 help='Do not print out dimensions')
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -04001358 options, args = parser.parse_args(args)
maruelaf6b06c2017-06-08 06:26:53 -07001359 process_filter_options(parser, options)
Marc-Antoine Ruel28083112014-03-13 16:34:04 -04001360
1361 if options.keep_dead and options.dead_only:
maruelaf6b06c2017-06-08 06:26:53 -07001362 parser.error('Use only one of --keep-dead or --dead-only')
1363 if options.busy and options.idle:
1364 parser.error('Use only one of --busy or --idle')
1365 if options.mp and options.non_mp:
1366 parser.error('Use only one of --mp or --non-mp')
Vadim Shtayura6b555c12014-07-23 16:22:18 -07001367
smut281c3902018-05-30 17:50:05 -07001368 url = options.swarming + '/_ah/api/swarming/v1/bots/list?'
maruelaf6b06c2017-06-08 06:26:53 -07001369 values = []
1370 if options.dead_only:
1371 values.append(('is_dead', 'TRUE'))
1372 elif options.keep_dead:
1373 values.append(('is_dead', 'NONE'))
1374 else:
1375 values.append(('is_dead', 'FALSE'))
Marc-Antoine Ruelc6c579e2014-09-08 18:43:45 -04001376
maruelaf6b06c2017-06-08 06:26:53 -07001377 if options.busy:
1378 values.append(('is_busy', 'TRUE'))
1379 elif options.idle:
1380 values.append(('is_busy', 'FALSE'))
1381 else:
1382 values.append(('is_busy', 'NONE'))
1383
1384 if options.mp:
1385 values.append(('is_mp', 'TRUE'))
1386 elif options.non_mp:
1387 values.append(('is_mp', 'FALSE'))
1388 else:
1389 values.append(('is_mp', 'NONE'))
1390
1391 for key, value in options.dimensions:
1392 values.append(('dimensions', '%s:%s' % (key, value)))
1393 url += urllib.urlencode(values)
1394 try:
1395 data, yielder = get_yielder(url, 0)
1396 bots = data.get('items') or []
1397 for items in yielder():
1398 if items:
1399 bots.extend(items)
1400 except Failure as e:
1401 sys.stderr.write('\n%s\n' % e)
1402 return 1
maruel77f720b2015-09-15 12:35:22 -07001403 for bot in natsort.natsorted(bots, key=lambda x: x['bot_id']):
maruelaf6b06c2017-06-08 06:26:53 -07001404 print bot['bot_id']
1405 if not options.bare:
1406 dimensions = {i['key']: i.get('value') for i in bot.get('dimensions', {})}
1407 print ' %s' % json.dumps(dimensions, sort_keys=True)
1408 if bot.get('task_id'):
1409 print ' task: %s' % bot['task_id']
Marc-Antoine Ruel819fb162014-03-12 16:38:26 -04001410 return 0
1411
1412
maruelfd0a90c2016-06-10 11:51:10 -07001413@subcommand.usage('task_id')
1414def CMDcancel(parser, args):
1415 """Cancels a task."""
Marc-Antoine Ruel2e52c552018-03-26 19:27:36 -04001416 parser.add_option(
1417 '-k', '--kill-running', action='store_true', default=False,
1418 help='Kill the task even if it was running')
maruelfd0a90c2016-06-10 11:51:10 -07001419 options, args = parser.parse_args(args)
1420 if not args:
1421 parser.error('Please specify the task to cancel')
Marc-Antoine Ruel2e52c552018-03-26 19:27:36 -04001422 data = {'kill_running': options.kill_running}
maruelfd0a90c2016-06-10 11:51:10 -07001423 for task_id in args:
smut281c3902018-05-30 17:50:05 -07001424 url = '%s/_ah/api/swarming/v1/task/%s/cancel' % (options.swarming, task_id)
Marc-Antoine Ruel2e52c552018-03-26 19:27:36 -04001425 resp = net.url_read_json(url, data=data, method='POST')
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001426 if resp is None:
maruelfd0a90c2016-06-10 11:51:10 -07001427 print('Deleting %s failed. Probably already gone' % task_id)
1428 return 1
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001429 logging.info('%s', resp)
maruelfd0a90c2016-06-10 11:51:10 -07001430 return 0
1431
1432
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001433@subcommand.usage('--json file | task_id...')
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001434def CMDcollect(parser, args):
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001435 """Retrieves results of one or multiple Swarming task by its ID.
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001436
1437 The result can be in multiple part if the execution was sharded. It can
1438 potentially have retries.
1439 """
1440 add_collect_options(parser)
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001441 parser.add_option(
1442 '-j', '--json',
1443 help='Load the task ids from .json as saved by trigger --dump-json')
maruel77f720b2015-09-15 12:35:22 -07001444 options, args = parser.parse_args(args)
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001445 process_collect_options(parser, options)
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001446 if not args and not options.json:
1447 parser.error('Must specify at least one task id or --json.')
1448 if args and options.json:
1449 parser.error('Only use one of task id or --json.')
1450
1451 if options.json:
maruel1ceb3872015-10-14 06:10:44 -07001452 options.json = unicode(os.path.abspath(options.json))
Marc-Antoine Ruel9025a782015-03-17 16:42:59 -04001453 try:
maruel1ceb3872015-10-14 06:10:44 -07001454 with fs.open(options.json, 'rb') as f:
maruel71c61c82016-02-22 06:52:05 -08001455 data = json.load(f)
1456 except (IOError, ValueError):
1457 parser.error('Failed to open %s' % options.json)
1458 try:
1459 tasks = sorted(
1460 data['tasks'].itervalues(), key=lambda x: x['shard_index'])
1461 args = [t['task_id'] for t in tasks]
1462 except (KeyError, TypeError):
1463 parser.error('Failed to process %s' % options.json)
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001464 if not options.timeout:
Marc-Antoine Ruelb73066b2018-04-19 20:16:55 -04001465 # Take in account all the task slices.
1466 offset = 0
1467 for s in data['request']['task_slices']:
1468 m = (offset + s['properties']['execution_timeout_secs'] +
1469 s['expiration_secs'])
1470 if m > options.timeout:
1471 options.timeout = m
1472 offset += s['expiration_secs']
Marc-Antoine Ruel9fc42612018-04-20 08:34:22 -04001473 options.timeout += 10.
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001474 else:
1475 valid = frozenset('0123456789abcdef')
1476 if any(not valid.issuperset(task_id) for task_id in args):
1477 parser.error('Task ids are 0-9a-f.')
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001478
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001479 try:
1480 return collect(
1481 options.swarming,
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001482 args,
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001483 options.timeout,
1484 options.decorate,
1485 options.print_status_updates,
1486 options.task_summary_json,
maruel9531ce02016-04-13 06:11:23 -07001487 options.task_output_dir,
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +10001488 options.task_output_stdout,
Takuto Ikuta1e6072c2018-11-06 20:42:43 +00001489 options.perf,
1490 options.filepath_filter)
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001491 except Failure:
1492 on_error.report(None)
1493 return 1
1494
1495
maruel77f720b2015-09-15 12:35:22 -07001496@subcommand.usage('[method name]')
Marc-Antoine Ruel833f5eb2018-04-25 16:49:40 -04001497def CMDpost(parser, args):
1498 """Sends a JSON RPC POST to one API endpoint and prints out the raw result.
1499
1500 Input data must be sent to stdin, result is printed to stdout.
1501
1502 If HTTP response code >= 400, returns non-zero.
1503 """
1504 options, args = parser.parse_args(args)
1505 if len(args) != 1:
1506 parser.error('Must specify only API name')
smut281c3902018-05-30 17:50:05 -07001507 url = options.swarming + '/_ah/api/swarming/v1/' + args[0]
Marc-Antoine Ruel833f5eb2018-04-25 16:49:40 -04001508 data = sys.stdin.read()
1509 try:
1510 resp = net.url_read(url, data=data, method='POST')
1511 except net.TimeoutError:
1512 sys.stderr.write('Timeout!\n')
1513 return 1
1514 if not resp:
1515 sys.stderr.write('No response!\n')
1516 return 1
1517 sys.stdout.write(resp)
1518 return 0
1519
1520
1521@subcommand.usage('[method name]')
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001522def CMDquery(parser, args):
maruel77f720b2015-09-15 12:35:22 -07001523 """Returns raw JSON information via an URL endpoint. Use 'query-list' to
1524 gather the list of API methods from the server.
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001525
1526 Examples:
maruelaf6b06c2017-06-08 06:26:53 -07001527 Raw task request and results:
1528 swarming.py query -S server-url.com task/123456/request
1529 swarming.py query -S server-url.com task/123456/result
1530
maruel77f720b2015-09-15 12:35:22 -07001531 Listing all bots:
maruel84e77aa2015-10-21 06:37:24 -07001532 swarming.py query -S server-url.com bots/list
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001533
maruelaf6b06c2017-06-08 06:26:53 -07001534 Listing last 10 tasks on a specific bot named 'bot1':
1535 swarming.py query -S server-url.com --limit 10 bot/bot1/tasks
maruel84e77aa2015-10-21 06:37:24 -07001536
maruelaf6b06c2017-06-08 06:26:53 -07001537 Listing last 10 tasks with tags os:Ubuntu-14.04 and pool:Chrome. Note that
maruel84e77aa2015-10-21 06:37:24 -07001538 quoting is important!:
1539 swarming.py query -S server-url.com --limit 10 \\
maruelaf6b06c2017-06-08 06:26:53 -07001540 'tasks/list?tags=os:Ubuntu-14.04&tags=pool:Chrome'
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001541 """
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001542 parser.add_option(
1543 '-L', '--limit', type='int', default=200,
1544 help='Limit to enforce on limitless items (like number of tasks); '
1545 'default=%default')
Paweł Hajdan, Jr53ef0132015-03-20 17:49:18 +01001546 parser.add_option(
1547 '--json', help='Path to JSON output file (otherwise prints to stdout)')
maruel77f720b2015-09-15 12:35:22 -07001548 parser.add_option(
1549 '--progress', action='store_true',
1550 help='Prints a dot at each request to show progress')
1551 options, args = parser.parse_args(args)
marueld8aba222015-09-03 12:21:19 -07001552 if len(args) != 1:
maruel77f720b2015-09-15 12:35:22 -07001553 parser.error(
1554 'Must specify only method name and optionally query args properly '
1555 'escaped.')
smut281c3902018-05-30 17:50:05 -07001556 base_url = options.swarming + '/_ah/api/swarming/v1/' + args[0]
maruelaf6b06c2017-06-08 06:26:53 -07001557 try:
1558 data, yielder = get_yielder(base_url, options.limit)
1559 for items in yielder():
1560 if items:
1561 data['items'].extend(items)
maruel77f720b2015-09-15 12:35:22 -07001562 if options.progress:
maruelaf6b06c2017-06-08 06:26:53 -07001563 sys.stderr.write('.')
1564 sys.stderr.flush()
1565 except Failure as e:
1566 sys.stderr.write('\n%s\n' % e)
1567 return 1
maruel77f720b2015-09-15 12:35:22 -07001568 if options.progress:
maruelaf6b06c2017-06-08 06:26:53 -07001569 sys.stderr.write('\n')
1570 sys.stderr.flush()
Paweł Hajdan, Jr53ef0132015-03-20 17:49:18 +01001571 if options.json:
maruel1ceb3872015-10-14 06:10:44 -07001572 options.json = unicode(os.path.abspath(options.json))
1573 tools.write_json(options.json, data, True)
Paweł Hajdan, Jr53ef0132015-03-20 17:49:18 +01001574 else:
Marc-Antoine Ruelcda90ee2015-03-23 15:13:20 -04001575 try:
maruel77f720b2015-09-15 12:35:22 -07001576 tools.write_json(sys.stdout, data, False)
Marc-Antoine Ruelcda90ee2015-03-23 15:13:20 -04001577 sys.stdout.write('\n')
1578 except IOError:
1579 pass
Marc-Antoine Ruel79940ae2014-09-23 17:55:41 -04001580 return 0
1581
1582
maruel77f720b2015-09-15 12:35:22 -07001583def CMDquery_list(parser, args):
1584 """Returns list of all the Swarming APIs that can be used with command
1585 'query'.
1586 """
1587 parser.add_option(
1588 '--json', help='Path to JSON output file (otherwise prints to stdout)')
1589 options, args = parser.parse_args(args)
1590 if args:
1591 parser.error('No argument allowed.')
1592
1593 try:
1594 apis = endpoints_api_discovery_apis(options.swarming)
1595 except APIError as e:
1596 parser.error(str(e))
1597 if options.json:
maruel1ceb3872015-10-14 06:10:44 -07001598 options.json = unicode(os.path.abspath(options.json))
1599 with fs.open(options.json, 'wb') as f:
maruel77f720b2015-09-15 12:35:22 -07001600 json.dump(apis, f)
1601 else:
1602 help_url = (
1603 'https://apis-explorer.appspot.com/apis-explorer/?base=%s/_ah/api#p/' %
1604 options.swarming)
maruel11e31af2017-02-15 07:30:50 -08001605 for i, (api_id, api) in enumerate(sorted(apis.iteritems())):
1606 if i:
1607 print('')
maruel77f720b2015-09-15 12:35:22 -07001608 print api_id
maruel11e31af2017-02-15 07:30:50 -08001609 print ' ' + api['description'].strip()
1610 if 'resources' in api:
1611 # Old.
1612 for j, (resource_name, resource) in enumerate(
1613 sorted(api['resources'].iteritems())):
1614 if j:
1615 print('')
1616 for method_name, method in sorted(resource['methods'].iteritems()):
1617 # Only list the GET ones.
1618 if method['httpMethod'] != 'GET':
1619 continue
1620 print '- %s.%s: %s' % (
1621 resource_name, method_name, method['path'])
1622 print('\n'.join(
Sergey Berezina269e1a2018-05-16 16:55:12 -07001623 ' ' + l for l in textwrap.wrap(
1624 method.get('description', 'No description'), 78)))
maruel11e31af2017-02-15 07:30:50 -08001625 print ' %s%s%s' % (help_url, api['servicePath'], method['id'])
1626 else:
1627 # New.
1628 for method_name, method in sorted(api['methods'].iteritems()):
maruel77f720b2015-09-15 12:35:22 -07001629 # Only list the GET ones.
1630 if method['httpMethod'] != 'GET':
1631 continue
maruel11e31af2017-02-15 07:30:50 -08001632 print '- %s: %s' % (method['id'], method['path'])
1633 print('\n'.join(
1634 ' ' + l for l in textwrap.wrap(method['description'], 78)))
maruel77f720b2015-09-15 12:35:22 -07001635 print ' %s%s%s' % (help_url, api['servicePath'], method['id'])
1636 return 0
1637
1638
Vadim Shtayuraae8085b2014-05-02 17:13:10 -07001639@subcommand.usage('(hash|isolated) [-- extra_args]')
maruel@chromium.org0437a732013-08-27 16:05:52 +00001640def CMDrun(parser, args):
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001641 """Triggers a task and wait for the results.
maruel@chromium.org0437a732013-08-27 16:05:52 +00001642
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001643 Basically, does everything to run a command remotely.
maruel@chromium.org0437a732013-08-27 16:05:52 +00001644 """
1645 add_trigger_options(parser)
1646 add_collect_options(parser)
Vadim Shtayurab450c602014-05-12 19:23:25 -07001647 add_sharding_options(parser)
maruel@chromium.org0437a732013-08-27 16:05:52 +00001648 options, args = parser.parse_args(args)
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001649 process_collect_options(parser, options)
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001650 task_request = process_trigger_options(parser, options, args)
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001651 try:
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001652 tasks = trigger_task_shards(
1653 options.swarming, task_request, options.shards)
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001654 except Failure as e:
Marc-Antoine Ruelcfb60852014-07-02 15:22:00 -04001655 on_error.report(
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001656 'Failed to trigger %s(%s): %s' %
maruel0a25f6c2017-05-10 10:43:23 -07001657 (task_request.name, args[0], e.args[0]))
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001658 return 1
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001659 if not tasks:
Marc-Antoine Ruelcfb60852014-07-02 15:22:00 -04001660 on_error.report('Failed to trigger the task.')
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001661 return 1
maruel0a25f6c2017-05-10 10:43:23 -07001662 print('Triggered task: %s' % task_request.name)
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001663 task_ids = [
1664 t['task_id']
1665 for t in sorted(tasks.itervalues(), key=lambda x: x['shard_index'])
1666 ]
Marc-Antoine Ruelf24f09c2018-03-23 16:06:18 -04001667 if not options.timeout:
Marc-Antoine Ruel3a030bc2018-04-23 10:31:25 -04001668 offset = 0
1669 for s in task_request.task_slices:
Marc-Antoine Ruel1f835c72018-05-25 12:29:42 -04001670 m = (offset + s.properties.execution_timeout_secs +
1671 s.expiration_secs)
Marc-Antoine Ruel3a030bc2018-04-23 10:31:25 -04001672 if m > options.timeout:
1673 options.timeout = m
Marc-Antoine Ruel1f835c72018-05-25 12:29:42 -04001674 offset += s.expiration_secs
Marc-Antoine Ruel3a030bc2018-04-23 10:31:25 -04001675 options.timeout += 10.
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001676 try:
1677 return collect(
1678 options.swarming,
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001679 task_ids,
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001680 options.timeout,
Vadim Shtayura86a2cef2014-04-18 11:13:39 -07001681 options.decorate,
Vadim Shtayurae3fbd102014-04-29 17:05:21 -07001682 options.print_status_updates,
Vadim Shtayurac8437bf2014-07-09 19:45:36 -07001683 options.task_summary_json,
maruel9531ce02016-04-13 06:11:23 -07001684 options.task_output_dir,
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +10001685 options.task_output_stdout,
Takuto Ikuta1e6072c2018-11-06 20:42:43 +00001686 options.perf,
1687 options.filepath_filter)
Marc-Antoine Ruelcfb60852014-07-02 15:22:00 -04001688 except Failure:
1689 on_error.report(None)
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001690 return 1
maruel@chromium.org0437a732013-08-27 16:05:52 +00001691
1692
maruel18122c62015-10-23 06:31:23 -07001693@subcommand.usage('task_id -- <extra_args>')
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001694def CMDreproduce(parser, args):
1695 """Runs a task locally that was triggered on the server.
1696
1697 This running locally the same commands that have been run on the bot. The data
1698 downloaded will be in a subdirectory named 'work' of the current working
1699 directory.
maruel18122c62015-10-23 06:31:23 -07001700
1701 You can pass further additional arguments to the target command by passing
1702 them after --.
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001703 """
maruelc070e672016-02-22 17:32:57 -08001704 parser.add_option(
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001705 '--output', metavar='DIR', default='out',
maruelc070e672016-02-22 17:32:57 -08001706 help='Directory that will have results stored into')
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001707 parser.add_option(
1708 '--work', metavar='DIR', default='work',
1709 help='Directory to map the task input files into')
1710 parser.add_option(
1711 '--cache', metavar='DIR', default='cache',
1712 help='Directory that contains the input cache')
1713 parser.add_option(
1714 '--leak', action='store_true',
1715 help='Do not delete the working directory after execution')
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001716 options, args = parser.parse_args(args)
maruel18122c62015-10-23 06:31:23 -07001717 extra_args = []
1718 if not args:
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001719 parser.error('Must specify exactly one task id.')
maruel18122c62015-10-23 06:31:23 -07001720 if len(args) > 1:
1721 if args[1] == '--':
1722 if len(args) > 2:
1723 extra_args = args[2:]
1724 else:
1725 extra_args = args[1:]
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001726
smut281c3902018-05-30 17:50:05 -07001727 url = options.swarming + '/_ah/api/swarming/v1/task/%s/request' % args[0]
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001728 request = net.url_read_json(url)
1729 if not request:
1730 print >> sys.stderr, 'Failed to retrieve request data for the task'
1731 return 1
1732
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001733 workdir = unicode(os.path.abspath(options.work))
maruele7cd38e2016-03-01 19:12:48 -08001734 if fs.isdir(workdir):
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001735 parser.error('Please delete the directory %r first' % options.work)
maruele7cd38e2016-03-01 19:12:48 -08001736 fs.mkdir(workdir)
iannucci31ab9192017-05-02 19:11:56 -07001737 cachedir = unicode(os.path.abspath('cipd_cache'))
1738 if not fs.exists(cachedir):
1739 fs.mkdir(cachedir)
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001740
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001741 properties = request['properties']
iannucci31ab9192017-05-02 19:11:56 -07001742 env = os.environ.copy()
1743 env['SWARMING_BOT_ID'] = 'reproduce'
1744 env['SWARMING_TASK_ID'] = 'reproduce'
maruel29ab2fd2015-10-16 11:44:01 -07001745 if properties.get('env'):
Marc-Antoine Ruel119b0842014-12-19 15:27:58 -05001746 logging.info('env: %r', properties['env'])
maruelb76604c2015-11-11 11:53:44 -08001747 for i in properties['env']:
Marc-Antoine Ruel36e09792018-01-09 14:03:25 -05001748 key = i['key']
maruelb76604c2015-11-11 11:53:44 -08001749 if not i['value']:
1750 env.pop(key, None)
1751 else:
Marc-Antoine Ruel36e09792018-01-09 14:03:25 -05001752 env[key] = i['value']
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001753
Robert Iannuccibf5f84c2017-11-22 12:56:50 -08001754 if properties.get('env_prefixes'):
Marc-Antoine Ruel36e09792018-01-09 14:03:25 -05001755 env_prefixes = properties['env_prefixes']
Robert Iannuccibf5f84c2017-11-22 12:56:50 -08001756 logging.info('env_prefixes: %r', env_prefixes)
Marc-Antoine Ruel36e09792018-01-09 14:03:25 -05001757 for i in env_prefixes:
1758 key = i['key']
1759 paths = [os.path.normpath(os.path.join(workdir, p)) for p in i['value']]
Robert Iannuccibf5f84c2017-11-22 12:56:50 -08001760 cur = env.get(key)
1761 if cur:
1762 paths.append(cur)
1763 env[key] = os.path.pathsep.join(paths)
1764
iannucci31ab9192017-05-02 19:11:56 -07001765 command = []
nodir152cba62016-05-12 16:08:56 -07001766 if (properties.get('inputs_ref') or {}).get('isolated'):
maruel29ab2fd2015-10-16 11:44:01 -07001767 # Create the tree.
1768 with isolateserver.get_storage(
1769 properties['inputs_ref']['isolatedserver'],
1770 properties['inputs_ref']['namespace']) as storage:
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001771 # Do not use MemoryContentAddressedCache here, as on 32-bits python,
1772 # inputs larger than ~1GiB will not fit in memory. This is effectively a
1773 # leak.
1774 policies = local_caching.CachePolicies(0, 0, 0, 0)
1775 algo = isolated_format.get_hash_algo(
1776 properties['inputs_ref']['namespace'])
1777 cache = local_caching.DiskContentAddressedCache(
1778 unicode(os.path.abspath(options.cache)), policies, algo, False)
maruel29ab2fd2015-10-16 11:44:01 -07001779 bundle = isolateserver.fetch_isolated(
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001780 properties['inputs_ref']['isolated'], storage, cache, workdir, False)
maruel29ab2fd2015-10-16 11:44:01 -07001781 command = bundle.command
1782 if bundle.relative_cwd:
1783 workdir = os.path.join(workdir, bundle.relative_cwd)
maruela1b9e552016-01-06 12:42:03 -08001784 command.extend(properties.get('extra_args') or [])
iannucci31ab9192017-05-02 19:11:56 -07001785
1786 if properties.get('command'):
1787 command.extend(properties['command'])
1788
Marc-Antoine Ruelc7243592018-05-24 17:04:04 -04001789 # https://chromium.googlesource.com/infra/luci/luci-py.git/+/master/appengine/swarming/doc/Magic-Values.md
Robert Iannucci24ae76a2018-02-26 12:51:18 -08001790 command = tools.fix_python_cmd(command, env)
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001791 if not options.output:
Marc-Antoine Ruel88229872018-01-10 16:35:29 -05001792 new_command = run_isolated.process_command(command, 'invalid', None)
1793 if new_command != command:
Marc-Antoine Ruel29ba75c2018-01-10 15:04:14 -05001794 parser.error('The task has outputs, you must use --output-dir')
Marc-Antoine Ruel88229872018-01-10 16:35:29 -05001795 else:
1796 # Make the path absolute, as the process will run from a subdirectory.
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001797 options.output = os.path.abspath(options.output)
Marc-Antoine Ruel88229872018-01-10 16:35:29 -05001798 new_command = run_isolated.process_command(
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001799 command, options.output, None)
1800 if not os.path.isdir(options.output):
1801 os.makedirs(options.output)
iannucci31ab9192017-05-02 19:11:56 -07001802 command = new_command
1803 file_path.ensure_command_has_abs_path(command, workdir)
1804
1805 if properties.get('cipd_input'):
1806 ci = properties['cipd_input']
1807 cp = ci['client_package']
1808 client_manager = cipd.get_client(
1809 ci['server'], cp['package_name'], cp['version'], cachedir)
1810
1811 with client_manager as client:
1812 by_path = collections.defaultdict(list)
1813 for pkg in ci['packages']:
1814 path = pkg['path']
1815 # cipd deals with 'root' as ''
1816 if path == '.':
1817 path = ''
1818 by_path[path].append((pkg['package_name'], pkg['version']))
1819 client.ensure(workdir, by_path, cache_dir=cachedir)
1820
maruel77f720b2015-09-15 12:35:22 -07001821 try:
Marc-Antoine Ruel95c21872018-01-10 14:24:28 -05001822 return subprocess42.call(command + extra_args, env=env, cwd=workdir)
maruel77f720b2015-09-15 12:35:22 -07001823 except OSError as e:
maruel29ab2fd2015-10-16 11:44:01 -07001824 print >> sys.stderr, 'Failed to run: %s' % ' '.join(command)
maruel77f720b2015-09-15 12:35:22 -07001825 print >> sys.stderr, str(e)
1826 return 1
Marc-Antoine Ruel5aeb3bb2018-06-16 13:11:02 +00001827 finally:
1828 # Do not delete options.cache.
1829 if not options.leak:
1830 file_path.rmtree(workdir)
Marc-Antoine Ruel13a81272014-10-07 20:16:43 -04001831
1832
maruel0eb1d1b2015-10-02 14:48:21 -07001833@subcommand.usage('bot_id')
1834def CMDterminate(parser, args):
1835 """Tells a bot to gracefully shut itself down as soon as it can.
1836
1837 This is done by completing whatever current task there is then exiting the bot
1838 process.
1839 """
1840 parser.add_option(
1841 '--wait', action='store_true', help='Wait for the bot to terminate')
1842 options, args = parser.parse_args(args)
1843 if len(args) != 1:
1844 parser.error('Please provide the bot id')
smut281c3902018-05-30 17:50:05 -07001845 url = options.swarming + '/_ah/api/swarming/v1/bot/%s/terminate' % args[0]
maruel0eb1d1b2015-10-02 14:48:21 -07001846 request = net.url_read_json(url, data={})
1847 if not request:
1848 print >> sys.stderr, 'Failed to ask for termination'
1849 return 1
1850 if options.wait:
1851 return collect(
Tim 'mithro' Ansell5e8001d2017-09-08 09:32:52 +10001852 options.swarming,
1853 [request['task_id']],
1854 0.,
1855 False,
1856 False,
1857 None,
1858 None,
1859 [],
Takuto Ikuta1e6072c2018-11-06 20:42:43 +00001860 False,
1861 None)
maruelbfc5f872017-06-10 16:43:17 -07001862 else:
1863 print request['task_id']
maruel0eb1d1b2015-10-02 14:48:21 -07001864 return 0
1865
1866
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001867@subcommand.usage("(hash|isolated) [-- extra_args|raw command]")
maruel@chromium.org0437a732013-08-27 16:05:52 +00001868def CMDtrigger(parser, args):
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001869 """Triggers a Swarming task.
maruel@chromium.org0437a732013-08-27 16:05:52 +00001870
Vadim Shtayuraae8085b2014-05-02 17:13:10 -07001871 Passes all extra arguments provided after '--' as additional command line
1872 arguments for an isolated command specified in *.isolate file.
maruel@chromium.org0437a732013-08-27 16:05:52 +00001873 """
1874 add_trigger_options(parser)
Vadim Shtayurab450c602014-05-12 19:23:25 -07001875 add_sharding_options(parser)
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001876 parser.add_option(
1877 '--dump-json',
1878 metavar='FILE',
1879 help='Dump details about the triggered task(s) to this file as json')
Marc-Antoine Ruel7c543272013-11-26 13:26:15 -05001880 options, args = parser.parse_args(args)
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001881 task_request = process_trigger_options(parser, options, args)
maruel@chromium.org0437a732013-08-27 16:05:52 +00001882 try:
Marc-Antoine Ruelefdc5282014-12-12 19:31:00 -05001883 tasks = trigger_task_shards(
1884 options.swarming, task_request, options.shards)
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001885 if tasks:
maruel0a25f6c2017-05-10 10:43:23 -07001886 print('Triggered task: %s' % task_request.name)
Marc-Antoine Ruel2f6581a2014-10-03 11:09:53 -04001887 tasks_sorted = sorted(
1888 tasks.itervalues(), key=lambda x: x['shard_index'])
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001889 if options.dump_json:
1890 data = {
maruel0a25f6c2017-05-10 10:43:23 -07001891 'base_task_name': task_request.name,
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001892 'tasks': tasks,
Vadim Shtayura2d83a942017-08-14 17:41:24 -07001893 'request': task_request_to_raw_request(task_request),
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001894 }
maruel46b015f2015-10-13 18:40:35 -07001895 tools.write_json(unicode(options.dump_json), data, True)
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001896 print('To collect results, use:')
Bruce Dawsonf0a5ae42018-09-04 20:06:46 +00001897 print(' tools/swarming_client/swarming.py collect -S %s --json %s' %
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001898 (options.swarming, options.dump_json))
1899 else:
Marc-Antoine Ruel12a7da42014-10-01 08:29:47 -04001900 print('To collect results, use:')
Bruce Dawsonf0a5ae42018-09-04 20:06:46 +00001901 print(' tools/swarming_client/swarming.py collect -S %s %s' %
Marc-Antoine Ruel2f6581a2014-10-03 11:09:53 -04001902 (options.swarming, ' '.join(t['task_id'] for t in tasks_sorted)))
1903 print('Or visit:')
1904 for t in tasks_sorted:
1905 print(' ' + t['view_url'])
Marc-Antoine Rueld6dbe762014-06-18 13:49:42 -04001906 return int(not tasks)
Marc-Antoine Ruelcfb60852014-07-02 15:22:00 -04001907 except Failure:
1908 on_error.report(None)
vadimsh@chromium.orgd908a542013-10-30 01:36:17 +00001909 return 1
maruel@chromium.org0437a732013-08-27 16:05:52 +00001910
1911
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -04001912class OptionParserSwarming(logging_utils.OptionParserWithLogging):
maruel@chromium.org0437a732013-08-27 16:05:52 +00001913 def __init__(self, **kwargs):
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -04001914 logging_utils.OptionParserWithLogging.__init__(
maruel@chromium.org0437a732013-08-27 16:05:52 +00001915 self, prog='swarming.py', **kwargs)
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -04001916 self.server_group = optparse.OptionGroup(self, 'Server')
Marc-Antoine Ruel5471e3d2013-11-11 19:10:32 -05001917 self.server_group.add_option(
maruel@chromium.orge9403ab2013-09-20 18:03:49 +00001918 '-S', '--swarming',
Kevin Graney5346c162014-01-24 12:20:01 -05001919 metavar='URL', default=os.environ.get('SWARMING_SERVER', ''),
maruel@chromium.orge9403ab2013-09-20 18:03:49 +00001920 help='Swarming server to use')
Marc-Antoine Ruel5471e3d2013-11-11 19:10:32 -05001921 self.add_option_group(self.server_group)
Vadim Shtayurae34e13a2014-02-02 11:23:26 -08001922 auth.add_auth_options(self)
maruel@chromium.org0437a732013-08-27 16:05:52 +00001923
1924 def parse_args(self, *args, **kwargs):
Marc-Antoine Ruelf74cffe2015-07-15 15:21:34 -04001925 options, args = logging_utils.OptionParserWithLogging.parse_args(
maruel@chromium.org0437a732013-08-27 16:05:52 +00001926 self, *args, **kwargs)
Marc-Antoine Ruel012067b2014-12-10 15:45:42 -05001927 auth.process_auth_options(self, options)
1928 user = self._process_swarming(options)
1929 if hasattr(options, 'user') and not options.user:
1930 options.user = user
1931 return options, args
1932
1933 def _process_swarming(self, options):
1934 """Processes the --swarming option and aborts if not specified.
1935
1936 Returns the identity as determined by the server.
1937 """
maruel@chromium.org0437a732013-08-27 16:05:52 +00001938 if not options.swarming:
1939 self.error('--swarming is required.')
Marc-Antoine Ruel012067b2014-12-10 15:45:42 -05001940 try:
1941 options.swarming = net.fix_url(options.swarming)
1942 except ValueError as e:
1943 self.error('--swarming %s' % e)
1944 on_error.report_on_exception_exit(options.swarming)
Marc-Antoine Ruelf7d737d2014-12-10 15:36:29 -05001945 try:
1946 user = auth.ensure_logged_in(options.swarming)
1947 except ValueError as e:
1948 self.error(str(e))
Marc-Antoine Ruel012067b2014-12-10 15:45:42 -05001949 return user
maruel@chromium.org0437a732013-08-27 16:05:52 +00001950
1951
1952def main(args):
1953 dispatcher = subcommand.CommandDispatcher(__name__)
Marc-Antoine Ruelcfb60852014-07-02 15:22:00 -04001954 return dispatcher.execute(OptionParserSwarming(version=__version__), args)
maruel@chromium.org0437a732013-08-27 16:05:52 +00001955
1956
1957if __name__ == '__main__':
maruel8e4e40c2016-05-30 06:21:07 -07001958 subprocess42.inhibit_os_error_reporting()
maruel@chromium.org0437a732013-08-27 16:05:52 +00001959 fix_encoding.fix_encoding()
1960 tools.disable_buffering()
1961 colorama.init()
1962 sys.exit(main(sys.argv[1:]))