Skip to content
Snippets Groups Projects
Forked from Iris / Iris
4630 commits behind the upstream repository.
gitlab-extract.py 5.09 KiB
#!/usr/bin/env python3
import argparse, pprint, sys, glob, zipfile, subprocess
import requests
import parse_log

def last(it):
    r = None
    for i in it:
        r = i
    return r

def first(it):
    for i in it:
        return i
    return None

def req(path):
    url = '%s/api/v3/%s' % (args.server, path)
    r = requests.get(url, headers={'PRIVATE-TOKEN': args.private_token})
    r.raise_for_status()
    return r

# read command-line arguments
parser = argparse.ArgumentParser(description='Extract iris-coq build logs from GitLab')
parser.add_argument("-t", "--private-token",
                    dest="private_token", required=True,
                    help="The private token used to authenticate access.")
parser.add_argument("-s", "--server",
                    dest="server", default="https://gitlab.mpi-sws.org/",
                    help="The GitLab server to contact.")
parser.add_argument("-p", "--project",
                    dest="project", default="FP/iris-coq",
                    help="The name of the project on GitLab.")
parser.add_argument("-f", "--file",
                    dest="file", required=True,
                    help="Filename to store the load in.")
parser.add_argument("-c", "--commits",
                    dest="commits",
                    help="The commits to fetch. Default is everything since the most recent entry in the log file.")
parser.add_argument("-a", "--artifacts",
                    dest="artifacts",
                    help="Location of the artifacts (following GitLab's folder structure).  If not given (which should be the common case), the artifacts will be downloaded from GitLab.")
parser.add_argument("-b", "--blacklist-branch",
                    dest="blacklist_branch",
                    help="Skip the commit if it is contained in the given branch.")
args = parser.parse_args()
log_file = sys.stdout if args.file == "-" else open(args.file, "a")

# determine commit, if missing
if args.commits is None:
    if args.file == "-":
        raise Exception("If you do not give explicit commits, you have to give a logfile so that we can determine the missing commits.")
    last_result = last(parse_log.parse(open(args.file, "r"), parse_times = parse_log.PARSE_NOT))
    args.commits = "{}..origin/master".format(last_result.commit)

projects = req("projects?per_page=512")
project = first(filter(lambda p: p['path_with_namespace'] == args.project, projects.json()))
if project is None:
    sys.stderr.write("Project not found.\n")
    sys.exit(1)

BREAK = False
for commit in parse_log.parse_git_commits(args.commits):
    if BREAK:
        break
    # test to skip the commit
    if args.blacklist_branch is not None:
        branches = subprocess.check_output(["git", "branch", "-r", "--contains", commit]).decode("utf-8")
        if args.blacklist_branch in map(lambda x: x.strip(), branches.split('\n')):
            continue
    # Find out more about the commit
    print("Fetching {}...".format(commit), end='')
    commit_data = req("/projects/{}/repository/commits/{}".format(project['id'], commit))
    if commit_data.status_code != 200:
        raise Exception("Commit not found?")
    builds = req("/projects/{}/repository/commits/{}/builds".format(project['id'], commit))
    if builds.status_code != 200:
        raise Exception("Build not found?")
    # iterate over builds by decreasing ID, and look for the artifact
    found_build = False
    for build in builds.json():
        if build['status'] in ('created', 'pending', 'running'):
            # build still not yet done, don't fetch this or any later commit
            BREAK = True
            print(" build still in progress, aborting")
            break
        if build['status'] != 'success':
            # build failed or cancelled, skip to next
            continue
        # now fetch the build times
        if args.artifacts:
            artifact_zip = glob.glob('{}/*/{}/{}/artifacts.zip'.format(args.artifacts, project['id'], build['id']))
            if not artifact_zip:
                # no artifact at this build, try another one
                continue
            assert len(artifact_zip) == 1, "Found too many artifacts"
            artifact_zip = artifact_zip[0]
            with zipfile.ZipFile(artifact_zip) as artifact:
                with artifact.open('build-time.txt') as build_times:
                    # Output into log file
                    log_file.write("# {}\n".format(commit))
                    log_file.write(build_times.read().decode('UTF-8'))
                    log_file.flush()
        else:
            build_times = requests.get("{}/builds/{}/artifacts/raw/build-time.txt".format(project['web_url'], build['id']))
            if build_times.status_code != 200:
                # no artifact at this build, try another one
                continue
            # Output in the log file format
            log_file.write("# {}\n".format(commit))
            log_file.write(build_times.text)
            log_file.flush()
            # don't fetch another build
            found_build = True
            print(" success")
            break
    if not found_build and not BREAK:
        print(" found no succeessful build")