1 | #!/usr/bin/env python2
|
2 | """
|
3 | soil/web.py - Dashboard that uses the "Event Sourcing" Paradigm
|
4 |
|
5 | Given state like this:
|
6 |
|
7 | https://test.oils-for-unix.org/
|
8 | github-jobs/
|
9 | 1234/ # $GITHUB_RUN_NUMBER
|
10 | cpp-small.tsv # benchmarks/time.py output. Success/failure for each task.
|
11 | cpp-small.json # metadata when job is DONE
|
12 |
|
13 | (cpp-small.wwz is linked to, but not part of the state.)
|
14 |
|
15 | (cpp-small.state # maybe for more transient events)
|
16 |
|
17 | This script generates:
|
18 |
|
19 | https://test.oils-for-unix.org/
|
20 | github-jobs/
|
21 | tmp-$$.index.html # jobs for all runs
|
22 | 1234/
|
23 | tmp-$$.index.html # jobs and tasks for a given run
|
24 | tmp-$$.remove.txt # TODO: consolidate 'cleanup', to make it faster
|
25 |
|
26 | # For sourcehut
|
27 | git-0101abab/
|
28 | tmp-$$.index.html
|
29 |
|
30 | How to test changes to this file:
|
31 |
|
32 | $ soil/web-init.sh deploy-code
|
33 | $ soil/web-worker.sh remote-rewrite-jobs-index github- ${GITHUB_RUN_NUMBER}
|
34 | $ soil/web-worker.sh remote-rewrite-jobs-index srht- git-${commit_hash}
|
35 |
|
36 | """
|
37 | from __future__ import print_function
|
38 |
|
39 | import collections
|
40 | import csv
|
41 | import datetime
|
42 | import json
|
43 | import itertools
|
44 | import os
|
45 | import re
|
46 | import sys
|
47 | from doctools import html_head
|
48 | from vendor import jsontemplate
|
49 |
|
50 |
|
51 | def log(msg, *args):
|
52 | if args:
|
53 | msg = msg % args
|
54 | print(msg, file=sys.stderr)
|
55 |
|
56 |
|
57 | def PrettyTime(now, start_time):
|
58 | """
|
59 | Return a pretty string like 'an hour ago', 'Yesterday', '3 months ago', 'just
|
60 | now', etc
|
61 | """
|
62 | # *** UNUSED because it only makes sense on a dynamic web page! ***
|
63 | # Loosely based on
|
64 | # https://stackoverflow.com/questions/1551382/user-friendly-time-format-in-python
|
65 |
|
66 | return 'unused'
|
67 |
|
68 |
|
69 | def _MinutesSeconds(num_seconds):
|
70 | num_seconds = round(num_seconds) # round to integer
|
71 | minutes = num_seconds / 60
|
72 | seconds = num_seconds % 60
|
73 | return '%d:%02d' % (minutes, seconds)
|
74 |
|
75 |
|
76 | LINE_RE = re.compile(r'(\w+)[ ]+([\d.]+)')
|
77 |
|
78 | def _ParsePullTime(time_p_str):
|
79 | """
|
80 | Given time -p output like
|
81 |
|
82 | real 0.01
|
83 | user 0.02
|
84 | sys 0.02
|
85 |
|
86 | Return the real time as a string, or - if we don't know it.
|
87 | """
|
88 | for line in time_p_str.splitlines():
|
89 | m = LINE_RE.match(line)
|
90 | if m:
|
91 | name, value = m.groups()
|
92 | if name == 'real':
|
93 | return _MinutesSeconds(float(value))
|
94 |
|
95 | return '-' # Not found
|
96 |
|
97 |
|
98 | DETAILS_RUN_T = jsontemplate.Template('''\
|
99 |
|
100 | <table>
|
101 | <tr class="spacer">
|
102 | <td></td>
|
103 | </tr>
|
104 |
|
105 | <tr class="commit-row">
|
106 | <td>
|
107 | <code>
|
108 | {.section github-commit-link}
|
109 | <a href="https://github.com/oilshell/oil/commit/{commit-hash}">{commit-hash-short}</a>
|
110 | {.end}
|
111 |
|
112 | {.section sourcehut-commit-link}
|
113 | <a href="https://git.sr.ht/~andyc/oil/commit/{commit-hash}">{commit-hash-short}</a>
|
114 | {.end}
|
115 | </code>
|
116 | </td>
|
117 |
|
118 | <td class="commit-line">
|
119 | {.section github-pr}
|
120 | <i>
|
121 | PR <a href="https://github.com/oilshell/oil/pull/{pr-number}">#{pr-number}</a>
|
122 | from <a href="https://github.com/oilshell/oil/tree/{head-ref}">{head-ref}</a>
|
123 | </i>
|
124 | {.end}
|
125 | {.section commit-desc}
|
126 | {@|html}
|
127 | {.end}
|
128 |
|
129 | {.section git-branch}
|
130 | <br/>
|
131 | <div style="text-align: right; font-family: monospace">{@}</div>
|
132 | {.end}
|
133 | </td>
|
134 |
|
135 | </tr>
|
136 | <tr class="spacer">
|
137 | <td><td/>
|
138 | </tr>
|
139 |
|
140 | </table>
|
141 | ''')
|
142 |
|
143 |
|
144 | DETAILS_TABLE_T = jsontemplate.Template('''\
|
145 | <table class="col1-right col3-right col4-right col5-right col6-right">
|
146 |
|
147 | <thead>
|
148 | <tr>
|
149 | <td>ID</td>
|
150 | <td>Job Name</td>
|
151 | <td>Start Time</td>
|
152 | <td>Pull Time</td>
|
153 | <td>Run Time</td>
|
154 | <td>Status</td>
|
155 | </tr>
|
156 | </thead>
|
157 |
|
158 | {.repeated section jobs}
|
159 | <tr>
|
160 |
|
161 | <td>{job_num}</td>
|
162 |
|
163 | <!-- internal link -->
|
164 | <td> <code><a href="#job-{job-name}">{job-name}</a></code> </td>
|
165 |
|
166 | <td><a href="{job_url}">{start_time_str}</a></td>
|
167 | <td>
|
168 | {.section pull_time_str}
|
169 | <a href="{run_wwz_path}/_tmp/soil/image.html">{@}</a>
|
170 | {.or}
|
171 | -
|
172 | {.end}
|
173 | </td>
|
174 |
|
175 | <td>{run_time_str}</td>
|
176 |
|
177 | <td> <!-- status -->
|
178 | {.section passed}
|
179 | <span class="pass">pass</span>
|
180 | {.end}
|
181 |
|
182 | {.section failed}
|
183 | <span class="fail">FAIL</span><br/>
|
184 | <span class="fail-detail">
|
185 | {.section one-failure}
|
186 | task <code>{@}</code>
|
187 | {.end}
|
188 |
|
189 | {.section multiple-failures}
|
190 | {num-failures} of {num-tasks} tasks
|
191 | {.end}
|
192 | </span>
|
193 | {.end}
|
194 | </td>
|
195 |
|
196 | </tr>
|
197 | {.end}
|
198 |
|
199 | </table>
|
200 | ''')
|
201 |
|
202 |
|
203 | def ParseJobs(stdin):
|
204 | """
|
205 | Given the output of list-json, open JSON and corresponding TSV, and yield a
|
206 | list of JSON template rows.
|
207 | """
|
208 | for i, line in enumerate(stdin):
|
209 | json_path = line.strip()
|
210 |
|
211 | #if i % 20 == 0:
|
212 | # log('job %d = %s', i, json_path)
|
213 |
|
214 | with open(json_path) as f:
|
215 | meta = json.load(f)
|
216 | #print(meta)
|
217 |
|
218 | tsv_path = json_path[:-5] + '.tsv'
|
219 | #log('%s', tsv_path)
|
220 |
|
221 | all_tasks = []
|
222 | failed_tasks = []
|
223 | total_elapsed = 0.0
|
224 |
|
225 | with open(tsv_path) as f:
|
226 | reader = csv.reader(f, delimiter='\t')
|
227 |
|
228 | try:
|
229 | for row in reader:
|
230 | t = {}
|
231 | # Unpack, matching _tmp/soil/INDEX.tsv
|
232 | ( status, elapsed,
|
233 | t['name'], t['script_name'], t['func'], results_url) = row
|
234 |
|
235 | t['results_url'] = None if results_url == '-' else results_url
|
236 |
|
237 | status = int(status)
|
238 | elapsed = float(elapsed)
|
239 |
|
240 | t['elapsed_str'] = _MinutesSeconds(elapsed)
|
241 |
|
242 | all_tasks.append(t)
|
243 |
|
244 | t['status'] = status
|
245 | if status == 0:
|
246 | t['passed'] = True
|
247 | else:
|
248 | t['failed'] = True
|
249 | failed_tasks.append(t)
|
250 |
|
251 | total_elapsed += elapsed
|
252 |
|
253 | except (IndexError, ValueError) as e:
|
254 | raise RuntimeError('Error in %r: %s (%r)' % (tsv_path, e, row))
|
255 |
|
256 | # So we can print task tables
|
257 | meta['tasks'] = all_tasks
|
258 |
|
259 | num_failures = len(failed_tasks)
|
260 |
|
261 | if num_failures == 0:
|
262 | meta['passed'] = True
|
263 | else:
|
264 | failed = {}
|
265 | if num_failures == 1:
|
266 | failed['one-failure'] = failed_tasks[0]['name']
|
267 | else:
|
268 | failed['multiple-failures'] = {
|
269 | 'num-failures': num_failures,
|
270 | 'num-tasks': len(all_tasks),
|
271 | }
|
272 | meta['failed'] = failed
|
273 |
|
274 | meta['run_time_str'] = _MinutesSeconds(total_elapsed)
|
275 |
|
276 | pull_time = meta.get('image-pull-time')
|
277 | if pull_time is not None:
|
278 | meta['pull_time_str'] = _ParsePullTime(pull_time)
|
279 |
|
280 | start_time = meta.get('task-run-start-time')
|
281 | if start_time is None:
|
282 | start_time_str = '?'
|
283 | else:
|
284 | # Note: this is different clock! Could be desynchronized.
|
285 | # Doesn't make sense this is static!
|
286 | #now = time.time()
|
287 | start_time = int(start_time)
|
288 |
|
289 | t = datetime.datetime.fromtimestamp(start_time)
|
290 | # %-I avoids leading 0, and is 12 hour date.
|
291 | # lower() for 'pm' instead of 'PM'.
|
292 | start_time_str = t.strftime('%-m/%d at %-I:%M%p').lower()
|
293 |
|
294 | #start_time_str = PrettyTime(now, start_time)
|
295 |
|
296 | meta['start_time_str'] = start_time_str
|
297 |
|
298 | # Metadata for a "run". A run is for a single commit, and consists of many
|
299 | # jobs.
|
300 |
|
301 | meta['git-branch'] = meta.get('GITHUB_REF')
|
302 |
|
303 | # Show the branch ref/heads/soil-staging or ref/pull/1577/merge (linkified)
|
304 | pr_head_ref = meta.get('GITHUB_PR_HEAD_REF')
|
305 | pr_number = meta.get('GITHUB_PR_NUMBER')
|
306 |
|
307 | if pr_head_ref and pr_number:
|
308 | meta['github-pr'] = {
|
309 | 'head-ref': pr_head_ref,
|
310 | 'pr-number': pr_number,
|
311 | }
|
312 |
|
313 | # Show the user's commit, not the merge commit
|
314 | commit_hash = meta.get('GITHUB_PR_HEAD_SHA') or '?'
|
315 |
|
316 | else:
|
317 | # From soil/worker.sh save-metadata. This is intended to be
|
318 | # CI-independent, while the environment variables above are from Github.
|
319 | meta['commit-desc'] = meta.get('commit-line', '?')
|
320 | commit_hash = meta.get('commit-hash') or '?'
|
321 |
|
322 | commit_link = {
|
323 | 'commit-hash': commit_hash,
|
324 | 'commit-hash-short': commit_hash[:8],
|
325 | }
|
326 |
|
327 | meta['job-name'] = meta.get('job-name') or '?'
|
328 |
|
329 | # Metadata for "Job"
|
330 |
|
331 | # GITHUB_RUN_NUMBER (project-scoped) is shorter than GITHUB_RUN_ID (global
|
332 | # scope)
|
333 | github_run = meta.get('GITHUB_RUN_NUMBER')
|
334 |
|
335 | if github_run:
|
336 | meta['job_num'] = github_run
|
337 | meta['index_run_url'] = '%s/' % github_run
|
338 |
|
339 | meta['github-commit-link'] = commit_link
|
340 |
|
341 | run_url_prefix = ''
|
342 | else:
|
343 | sourcehut_job_id = meta['JOB_ID']
|
344 | meta['job_num'] = sourcehut_job_id
|
345 | meta['index_run_url'] = 'git-%s/' % meta['commit-hash']
|
346 |
|
347 | meta['sourcehut-commit-link'] = commit_link
|
348 |
|
349 | # sourcehut doesn't have RUN ID, so we're in
|
350 | # srht-jobs/git-ab01cd/index.html, and need to find srht-jobs/123/foo.wwz
|
351 | run_url_prefix = '../%s/' % sourcehut_job_id
|
352 |
|
353 | # For Github, we construct $JOB_URL in soil/github-actions.sh
|
354 | meta['job_url'] = meta.get('JOB_URL') or '?'
|
355 |
|
356 | prefix, _ = os.path.splitext(json_path) # x/y/123/myjob
|
357 | parts = prefix.split('/')
|
358 |
|
359 | # Paths relative to github-jobs/1234/
|
360 | meta['run_wwz_path'] = run_url_prefix + parts[-1] + '.wwz' # myjob.wwz
|
361 | meta['run_tsv_path'] = run_url_prefix + parts[-1] + '.tsv' # myjob.tsv
|
362 | meta['run_json_path'] = run_url_prefix + parts[-1] + '.json' # myjob.json
|
363 |
|
364 | # Relative to github-jobs/
|
365 | last_two_parts = parts[-2:] # ['123', 'myjob']
|
366 | meta['index_wwz_path'] = '/'.join(last_two_parts) + '.wwz' # 123/myjob.wwz
|
367 |
|
368 | yield meta
|
369 |
|
370 |
|
371 | HTML_BODY_TOP_T = jsontemplate.Template('''
|
372 | <body class="width50">
|
373 | <p id="home-link">
|
374 | <a href="..">Up</a>
|
375 | | <a href="/">Home</a>
|
376 | | <a href="//oilshell.org/">oilshell.org</a>
|
377 | </p>
|
378 |
|
379 | <h1>{title|html}</h1>
|
380 | ''')
|
381 |
|
382 | HTML_BODY_BOTTOM = '''\
|
383 | </body>
|
384 | </html>
|
385 | '''
|
386 |
|
387 | INDEX_HEADER = '''\
|
388 | <table>
|
389 | <thead>
|
390 | <tr>
|
391 | <td colspan=1> Commit </td>
|
392 | <td colspan=1> Description </td>
|
393 | </tr>
|
394 | </thead>
|
395 | '''
|
396 |
|
397 | INDEX_RUN_ROW_T = jsontemplate.Template('''\
|
398 | <tr class="spacer">
|
399 | <td colspan=2></td>
|
400 | </tr>
|
401 |
|
402 | <tr class="commit-row">
|
403 | <td>
|
404 | <code>
|
405 | {.section github-commit-link}
|
406 | <a href="https://github.com/oilshell/oil/commit/{commit-hash}">{commit-hash-short}</a>
|
407 | {.end}
|
408 |
|
409 | {.section sourcehut-commit-link}
|
410 | <a href="https://git.sr.ht/~andyc/oil/commit/{commit-hash}">{commit-hash-short}</a>
|
411 | {.end}
|
412 | </code>
|
413 |
|
414 | </td>
|
415 | </td>
|
416 |
|
417 | <td class="commit-line">
|
418 | {.section github-pr}
|
419 | <i>
|
420 | PR <a href="https://github.com/oilshell/oil/pull/{pr-number}">#{pr-number}</a>
|
421 | from <a href="https://github.com/oilshell/oil/tree/{head-ref}">{head-ref}</a>
|
422 | </i>
|
423 | {.end}
|
424 | {.section commit-desc}
|
425 | {@|html}
|
426 | {.end}
|
427 |
|
428 | {.section git-branch}
|
429 | <br/>
|
430 | <div style="text-align: right; font-family: monospace">{@}</div>
|
431 | {.end}
|
432 | </td>
|
433 |
|
434 | </tr>
|
435 | <tr class="spacer">
|
436 | <td colspan=2><td/>
|
437 | </tr>
|
438 | ''')
|
439 |
|
440 | INDEX_JOBS_T = jsontemplate.Template('''\
|
441 | <tr>
|
442 | <td>
|
443 | </td>
|
444 | <td>
|
445 | <a href="{index_run_url}">All Jobs and Tasks</a>
|
446 | </td>
|
447 | </tr>
|
448 |
|
449 | {.section jobs-passed}
|
450 | <tr>
|
451 | <td class="pass">
|
452 | Passed
|
453 | </td>
|
454 | <td>
|
455 | {.repeated section @}
|
456 | <code class="pass">{job-name}</code>
|
457 | <!--
|
458 | <span class="pass"> ✓ </span>
|
459 | -->
|
460 | {.alternates with}
|
461 |
|
462 | {.end}
|
463 | </td>
|
464 | </tr>
|
465 | {.end}
|
466 |
|
467 | {.section jobs-failed}
|
468 | <tr>
|
469 | <td class="fail">
|
470 | Failed
|
471 | </td>
|
472 | <td>
|
473 | {.repeated section @}
|
474 | <span class="fail"> ✗ </span>
|
475 | <code><a href="{index_run_url}#job-{job-name}">{job-name}</a></code>
|
476 |
|
477 | <span class="fail-detail">
|
478 | {.section failed}
|
479 | {.section one-failure}
|
480 | - task <code>{@}</code>
|
481 | {.end}
|
482 |
|
483 | {.section multiple-failures}
|
484 | - {num-failures} of {num-tasks} tasks
|
485 | {.end}
|
486 | {.end}
|
487 | </span>
|
488 |
|
489 | {.alternates with}
|
490 | <br />
|
491 | {.end}
|
492 | </td>
|
493 | </tr>
|
494 | {.end}
|
495 |
|
496 | <tr class="spacer">
|
497 | <td colspan=3> </td>
|
498 | </tr>
|
499 |
|
500 | ''')
|
501 |
|
502 | def PrintIndexHtml(title, groups, f=sys.stdout):
|
503 | # Bust cache (e.g. Safari iPad seems to cache aggressively and doesn't
|
504 | # have Ctrl-F5)
|
505 | html_head.Write(f, title,
|
506 | css_urls=['../web/base.css?cache=0', '../web/soil.css?cache=0'])
|
507 |
|
508 | d = {'title': title}
|
509 | print(HTML_BODY_TOP_T.expand(d), file=f)
|
510 |
|
511 | print(INDEX_HEADER, file=f)
|
512 |
|
513 | for key, jobs in groups.iteritems():
|
514 | # All jobs have run-level metadata, so just use the first
|
515 |
|
516 | print(INDEX_RUN_ROW_T.expand(jobs[0]), file=f)
|
517 |
|
518 | summary = {
|
519 | 'jobs-passed': [],
|
520 | 'jobs-failed': [],
|
521 | 'index_run_url': jobs[0]['index_run_url'],
|
522 | }
|
523 |
|
524 | for job in jobs:
|
525 | if job.get('passed'):
|
526 | summary['jobs-passed'].append(job)
|
527 | else:
|
528 | summary['jobs-failed'].append(job)
|
529 |
|
530 | print(INDEX_JOBS_T.expand(summary), file=f)
|
531 |
|
532 | print(' </table>', file=f)
|
533 | print(HTML_BODY_BOTTOM, file=f)
|
534 |
|
535 |
|
536 | TASK_TABLE_T = jsontemplate.Template('''\
|
537 |
|
538 | <h2>All Tasks</h2>
|
539 |
|
540 | <!-- right justify elapsed and status -->
|
541 | <table class="col2-right col3-right col4-right">
|
542 |
|
543 | {.repeated section jobs}
|
544 |
|
545 | <tr> <!-- link here -->
|
546 | <td colspan=4>
|
547 | <a name="job-{job-name}"></a>
|
548 | </td>
|
549 | </tr>
|
550 |
|
551 | <tr style="background-color: #EEE">
|
552 | <td colspan=3>
|
553 | <b>{job-name}</b>
|
554 |
|
555 |
|
556 |
|
557 | <a href="{run_wwz_path}/">wwz</a>
|
558 |
|
559 | <a href="{run_tsv_path}">TSV</a>
|
560 |
|
561 | <a href="{run_json_path}">JSON</a>
|
562 | <td>
|
563 | <a href="">Up</a>
|
564 | </td>
|
565 | </tr>
|
566 |
|
567 | <tr class="spacer">
|
568 | <td colspan=4> </td>
|
569 | </tr>
|
570 |
|
571 | <tr style="font-weight: bold">
|
572 | <td>Task</td>
|
573 | <td>Results</td>
|
574 | <td>Elapsed</td>
|
575 | <td>Status</td>
|
576 | </tr>
|
577 |
|
578 | {.repeated section tasks}
|
579 | <tr>
|
580 | <td>
|
581 | <a href="{run_wwz_path}/_tmp/soil/logs/{name}.txt">{name}</a> <br/>
|
582 | <code>{script_name} {func}</code>
|
583 | </td>
|
584 |
|
585 | <td>
|
586 | {.section results_url}
|
587 | <a href="{run_wwz_path}/{@}">Results</a>
|
588 | {.or}
|
589 | {.end}
|
590 | </td>
|
591 |
|
592 | <td>{elapsed_str}</td>
|
593 |
|
594 | {.section passed}
|
595 | <td>{status}</td>
|
596 | {.end}
|
597 | {.section failed}
|
598 | <td class="fail">status: {status}</td>
|
599 | {.end}
|
600 |
|
601 | </tr>
|
602 | {.end}
|
603 |
|
604 | <tr class="spacer">
|
605 | <td colspan=4> </td>
|
606 | </tr>
|
607 |
|
608 | {.end}
|
609 |
|
610 | </table>
|
611 |
|
612 | ''')
|
613 |
|
614 |
|
615 | def PrintRunHtml(title, jobs, f=sys.stdout):
|
616 | """Print index for jobs in a single run."""
|
617 |
|
618 | # Have to descend an extra level
|
619 | html_head.Write(f, title,
|
620 | css_urls=['../../web/base.css?cache=0', '../../web/soil.css?cache=0'])
|
621 |
|
622 | d = {'title': title}
|
623 | print(HTML_BODY_TOP_T.expand(d), file=f)
|
624 |
|
625 | print(DETAILS_RUN_T.expand(jobs[0]), file=f)
|
626 |
|
627 | d2 = {'jobs': jobs}
|
628 | print(DETAILS_TABLE_T.expand(d2), file=f)
|
629 |
|
630 | print(TASK_TABLE_T.expand(d2), file=f)
|
631 |
|
632 | print(HTML_BODY_BOTTOM, file=f)
|
633 |
|
634 |
|
635 | def GroupJobs(jobs, key_func):
|
636 | """
|
637 | Expands groupby result into a simple dict
|
638 | """
|
639 | groups = itertools.groupby(jobs, key=key_func)
|
640 |
|
641 | d = collections.OrderedDict()
|
642 |
|
643 | for key, job_iter in groups:
|
644 | jobs = list(job_iter)
|
645 |
|
646 | jobs.sort(key=ByTaskRunStartTime, reverse=True)
|
647 |
|
648 | d[key] = jobs
|
649 |
|
650 | return d
|
651 |
|
652 |
|
653 | def ByTaskRunStartTime(row):
|
654 | return int(row.get('task-run-start-time', 0))
|
655 |
|
656 | def ByCommitDate(row):
|
657 | # Written in the shell script
|
658 | # This is in ISO 8601 format (git log %aI), so we can sort by it.
|
659 | return row.get('commit-date', '?')
|
660 |
|
661 | def ByCommitHash(row):
|
662 | return row.get('commit-hash', '?')
|
663 |
|
664 | def ByGithubRun(row):
|
665 | # Written in the shell script
|
666 | # This is in ISO 8601 format (git log %aI), so we can sort by it.
|
667 | return int(row.get('GITHUB_RUN_NUMBER', 0))
|
668 |
|
669 |
|
670 | def main(argv):
|
671 | action = argv[1]
|
672 |
|
673 | if action == 'srht-index':
|
674 | index_out = argv[2]
|
675 | run_index_out = argv[3]
|
676 | run_id = argv[4] # looks like git-0101abab
|
677 |
|
678 | assert run_id.startswith('git-'), run_id
|
679 | commit_hash = run_id[4:]
|
680 |
|
681 | jobs = list(ParseJobs(sys.stdin))
|
682 |
|
683 | # sourcehut doesn't have a build number.
|
684 | # - Sort by descnding commit date. (Minor problem: Committing on a VM with
|
685 | # bad clock can cause commits "in the past")
|
686 | # - Group by commit HASH, because 'git rebase' can crate different commits
|
687 | # with the same date.
|
688 | jobs.sort(key=ByCommitDate, reverse=True)
|
689 | groups = GroupJobs(jobs, ByCommitHash)
|
690 |
|
691 | title = 'Recent Jobs (sourcehut)'
|
692 | with open(index_out, 'w') as f:
|
693 | PrintIndexHtml(title, groups, f=f)
|
694 |
|
695 | jobs = groups[commit_hash]
|
696 | title = 'Jobs for commit %s' % commit_hash
|
697 | with open(run_index_out, 'w') as f:
|
698 | PrintRunHtml(title, jobs, f=f)
|
699 |
|
700 | elif action == 'github-index':
|
701 |
|
702 | index_out = argv[2]
|
703 | run_index_out = argv[3]
|
704 | run_id = int(argv[4]) # compared as an integer
|
705 |
|
706 | jobs = list(ParseJobs(sys.stdin))
|
707 |
|
708 | jobs.sort(key=ByGithubRun, reverse=True) # ordered
|
709 | groups = GroupJobs(jobs, ByGithubRun)
|
710 |
|
711 | title = 'Recent Jobs (Github Actions)'
|
712 | with open(index_out, 'w') as f:
|
713 | PrintIndexHtml(title, groups, f=f)
|
714 |
|
715 | jobs = groups[run_id]
|
716 | title = 'Jobs for run %d' % run_id
|
717 |
|
718 | with open(run_index_out, 'w') as f:
|
719 | PrintRunHtml(title, jobs, f=f)
|
720 |
|
721 | elif action == 'cleanup':
|
722 | try:
|
723 | num_to_keep = int(argv[2])
|
724 | except IndexError:
|
725 | num_to_keep = 200
|
726 |
|
727 | prefixes = []
|
728 | for line in sys.stdin:
|
729 | json_path = line.strip()
|
730 |
|
731 | #log('%s', json_path)
|
732 | prefixes.append(json_path[:-5])
|
733 |
|
734 | log('%s cleanup: keep %d', sys.argv[0], num_to_keep)
|
735 | log('%s cleanup: got %d JSON paths', sys.argv[0], len(prefixes))
|
736 |
|
737 | # TODO: clean up git-$hash dirs
|
738 | #
|
739 | # github-jobs/
|
740 | # $GITHUB_RUN_NUMBER/
|
741 | # cpp-tarball.{json,wwz,tsv}
|
742 | # dummy.{json,wwz,tsv}
|
743 | # git-$hash/
|
744 | # oils-for-unix.tar
|
745 | #
|
746 | # srht-jobs/
|
747 | # 1234/
|
748 | # cpp-tarball.{json,wwz,tsv}
|
749 | # 1235/
|
750 | # dummy.{json,wwz,tsv}
|
751 | # git-$hash/
|
752 | # index.html # HTML for this job
|
753 | # oils-for-unix.tar
|
754 | #
|
755 | # We might have to read the most recent JSON, find the corresponding $hash,
|
756 | # and print that dir.
|
757 | #
|
758 | # Another option is to use a real database, rather than the file system!
|
759 |
|
760 | # Sort by 999 here
|
761 | # travis-ci.oilshell.org/github-jobs/999/foo.json
|
762 |
|
763 | prefixes.sort(key = lambda path: int(path.split('/')[-2]))
|
764 |
|
765 | prefixes = prefixes[:-num_to_keep]
|
766 |
|
767 | # Show what to delete. Then the user can pipe to xargs rm to remove it.
|
768 | for prefix in prefixes:
|
769 | print(prefix + '.json')
|
770 | print(prefix + '.tsv')
|
771 | print(prefix + '.wwz')
|
772 |
|
773 | else:
|
774 | raise RuntimeError('Invalid action %r' % action)
|
775 |
|
776 |
|
777 | if __name__ == '__main__':
|
778 | try:
|
779 | main(sys.argv)
|
780 | except RuntimeError as e:
|
781 | print('FATAL: %s' % e, file=sys.stderr)
|
782 | sys.exit(1)
|