Loading ci/get_id_of_last_job_occurence.py +8 −9 Original line number Diff line number Diff line Loading @@ -38,8 +38,8 @@ import requests PER_PAGE_SUFFIX = "?per_page=50" PAGE_SUFFIX = "&page={}" API_URL_TMPL = "https://forge.3gpp.org/rep/api/v4/projects/{}/pipelines" SCOPE_FAILED = "scope[]=failed" SCOPE_SUCCESS = "scope[]=success" SCOPE_FAILED = "&scope[]=failed" SCOPE_SUCCESS = "&scope[]=success" def get_job_id(branch_name, job_name, project_id, success_only): Loading @@ -53,16 +53,15 @@ def get_job_id(branch_name, job_name, project_id, success_only): resp_pls = requests.get(url_pls + suffix) for pl in resp_pls.json(): if pl["ref"] == branch_name: scope = f"?{SCOPE_SUCCESS}" if not success_only: scope += f"&{SCOPE_FAILED}" url_args = PER_PAGE_SUFFIX url_jobs = url_pls + f"/{pl['id']}/jobs{scope}" url_args += SCOPE_SUCCESS if not success_only: url_args += SCOPE_FAILED # only one of the suffixes here - this assumes only max of 50 jobs per pipeline # so only one page needed resp_jobs = requests.get(url_jobs + PER_PAGE_SUFFIX) url_jobs = url_pls + f"/{pl['id']}/jobs" + url_args resp_jobs = requests.get(url_jobs) if job_name not in resp_jobs.text: continue Loading Loading
ci/get_id_of_last_job_occurence.py +8 −9 Original line number Diff line number Diff line Loading @@ -38,8 +38,8 @@ import requests PER_PAGE_SUFFIX = "?per_page=50" PAGE_SUFFIX = "&page={}" API_URL_TMPL = "https://forge.3gpp.org/rep/api/v4/projects/{}/pipelines" SCOPE_FAILED = "scope[]=failed" SCOPE_SUCCESS = "scope[]=success" SCOPE_FAILED = "&scope[]=failed" SCOPE_SUCCESS = "&scope[]=success" def get_job_id(branch_name, job_name, project_id, success_only): Loading @@ -53,16 +53,15 @@ def get_job_id(branch_name, job_name, project_id, success_only): resp_pls = requests.get(url_pls + suffix) for pl in resp_pls.json(): if pl["ref"] == branch_name: scope = f"?{SCOPE_SUCCESS}" if not success_only: scope += f"&{SCOPE_FAILED}" url_args = PER_PAGE_SUFFIX url_jobs = url_pls + f"/{pl['id']}/jobs{scope}" url_args += SCOPE_SUCCESS if not success_only: url_args += SCOPE_FAILED # only one of the suffixes here - this assumes only max of 50 jobs per pipeline # so only one page needed resp_jobs = requests.get(url_jobs + PER_PAGE_SUFFIX) url_jobs = url_pls + f"/{pl['id']}/jobs" + url_args resp_jobs = requests.get(url_jobs) if job_name not in resp_jobs.text: continue Loading