Loading .gitlab-ci.yml +20 −4 Original line number Diff line number Diff line Loading @@ -413,6 +413,9 @@ external-renderer-pytest-on-merge-request: extends: - .test-job-linux - .rules-merge-request # TODO: remove, only for testing tags: - test-fhg-linux-runner1 needs: ["build-codec-linux-make"] # TODO: set reasonable timeout, will most likely take less timeout: "20 minutes" Loading Loading @@ -443,7 +446,7 @@ external-renderer-pytest-on-merge-request: # run test - exit_code=0 - python3 -m pytest -q --log-level ERROR -n auto -rA --junit-xml=report-junit.xml tests/renderer/test_renderer_be_comparison.py || exit_code=$? - python3 -m pytest -q --log-level ERROR -n auto -rA --html=report.html --junit-xml=report-junit.xml tests/renderer/test_renderer_be_comparison.py || exit_code=$? - zero_errors=$(cat report-junit.xml | grep -c 'errors="0"') || true - *merge-request-comparison-check Loading @@ -456,6 +459,7 @@ external-renderer-pytest-on-merge-request: when: always paths: - report-junit.xml - report.html expose_as: "pytest external renderer results" reports: junit: Loading @@ -466,6 +470,9 @@ ivas-pytest-on-merge-request: extends: - .test-job-linux - .rules-merge-request # TODO: remove, only for testing tags: - test-fhg-linux-runner1 stage: compare needs: ["build-codec-linux-cmake", "codec-smoke-test"] timeout: "10 minutes" Loading @@ -492,7 +499,7 @@ ivas-pytest-on-merge-request: ### run pytest - exit_code=0 - python3 -m pytest tests -v --junit-xml=report-junit.xml || exit_code=$? - python3 -m pytest tests -v --html=report.html --junit-xml=report-junit.xml || exit_code=$? - zero_errors=$(cat report-junit.xml | grep -c 'errors="0"') || true - *merge-request-comparison-check Loading @@ -505,6 +512,7 @@ ivas-pytest-on-merge-request: when: always paths: - report-junit.xml - report.html expose_as: "pytest ivas results" reports: junit: Loading @@ -514,6 +522,9 @@ evs-pytest-on-merge-request: extends: - .test-job-linux - .rules-merge-request # TODO: remove, only for testing tags: - test-fhg-linux-runner1 stage: compare needs: ["build-codec-linux-cmake", "codec-smoke-test"] timeout: "10 minutes" Loading @@ -537,7 +548,7 @@ evs-pytest-on-merge-request: ### run pytest for EVS cases - exit_code=0 - python3 -m pytest tests/test_param_file.py -v --param_file scripts/config/self_test_evs.prm --junit-xml=report-junit-evs.xml || exit_code=$? - python3 -m pytest tests/test_param_file.py -v --param_file scripts/config/self_test_evs.prm --html=report.html --junit-xml=report-junit-evs.xml || exit_code=$? - zero_errors=$(cat report-junit-evs.xml | grep -c 'errors="0"') || true - *merge-request-comparison-check Loading @@ -550,6 +561,7 @@ evs-pytest-on-merge-request: when: always paths: - report-junit-evs.xml - report.html expose_as: "pytest evs results" reports: junit: Loading Loading @@ -634,6 +646,9 @@ codec-comparison-on-main-push: extends: - .test-job-linux - .rules-main-push # TODO: remove, only for testing tags: - test-fhg-linux-runner1 stage: compare needs: ["build-codec-linux-cmake"] timeout: "30 minutes" # To be revisited Loading Loading @@ -691,7 +706,7 @@ codec-comparison-on-main-push: ### run pytest - exit_code=0 - python3 -m pytest tests -v --junit-xml=report-junit.xml || exit_code=$? - python3 -m pytest tests -v --html=report.html --junit-xml=report-junit.xml || exit_code=$? - if [ $exit_code -eq 1 ] && [ $non_be_flag == 0 ]; then echo "pytest run had failures and non-BE flag not present"; exit $EXIT_CODE_FAIL; fi - zero_errors=$(cat report-junit.xml | grep -c 'errors="0"') || true - if [ $exit_code -eq 1 ] && [ $zero_errors == 1 ]; then echo "pytest run had failures, but no errors and non-BE flag present"; exit $EXIT_CODE_NON_BE; fi Loading @@ -704,6 +719,7 @@ codec-comparison-on-main-push: when: always paths: - report-junit.xml - report.html expose_as: "Results of comparison to previous merge commit" reports: junit: report-junit.xml Loading Loading
.gitlab-ci.yml +20 −4 Original line number Diff line number Diff line Loading @@ -413,6 +413,9 @@ external-renderer-pytest-on-merge-request: extends: - .test-job-linux - .rules-merge-request # TODO: remove, only for testing tags: - test-fhg-linux-runner1 needs: ["build-codec-linux-make"] # TODO: set reasonable timeout, will most likely take less timeout: "20 minutes" Loading Loading @@ -443,7 +446,7 @@ external-renderer-pytest-on-merge-request: # run test - exit_code=0 - python3 -m pytest -q --log-level ERROR -n auto -rA --junit-xml=report-junit.xml tests/renderer/test_renderer_be_comparison.py || exit_code=$? - python3 -m pytest -q --log-level ERROR -n auto -rA --html=report.html --junit-xml=report-junit.xml tests/renderer/test_renderer_be_comparison.py || exit_code=$? - zero_errors=$(cat report-junit.xml | grep -c 'errors="0"') || true - *merge-request-comparison-check Loading @@ -456,6 +459,7 @@ external-renderer-pytest-on-merge-request: when: always paths: - report-junit.xml - report.html expose_as: "pytest external renderer results" reports: junit: Loading @@ -466,6 +470,9 @@ ivas-pytest-on-merge-request: extends: - .test-job-linux - .rules-merge-request # TODO: remove, only for testing tags: - test-fhg-linux-runner1 stage: compare needs: ["build-codec-linux-cmake", "codec-smoke-test"] timeout: "10 minutes" Loading @@ -492,7 +499,7 @@ ivas-pytest-on-merge-request: ### run pytest - exit_code=0 - python3 -m pytest tests -v --junit-xml=report-junit.xml || exit_code=$? - python3 -m pytest tests -v --html=report.html --junit-xml=report-junit.xml || exit_code=$? - zero_errors=$(cat report-junit.xml | grep -c 'errors="0"') || true - *merge-request-comparison-check Loading @@ -505,6 +512,7 @@ ivas-pytest-on-merge-request: when: always paths: - report-junit.xml - report.html expose_as: "pytest ivas results" reports: junit: Loading @@ -514,6 +522,9 @@ evs-pytest-on-merge-request: extends: - .test-job-linux - .rules-merge-request # TODO: remove, only for testing tags: - test-fhg-linux-runner1 stage: compare needs: ["build-codec-linux-cmake", "codec-smoke-test"] timeout: "10 minutes" Loading @@ -537,7 +548,7 @@ evs-pytest-on-merge-request: ### run pytest for EVS cases - exit_code=0 - python3 -m pytest tests/test_param_file.py -v --param_file scripts/config/self_test_evs.prm --junit-xml=report-junit-evs.xml || exit_code=$? - python3 -m pytest tests/test_param_file.py -v --param_file scripts/config/self_test_evs.prm --html=report.html --junit-xml=report-junit-evs.xml || exit_code=$? - zero_errors=$(cat report-junit-evs.xml | grep -c 'errors="0"') || true - *merge-request-comparison-check Loading @@ -550,6 +561,7 @@ evs-pytest-on-merge-request: when: always paths: - report-junit-evs.xml - report.html expose_as: "pytest evs results" reports: junit: Loading Loading @@ -634,6 +646,9 @@ codec-comparison-on-main-push: extends: - .test-job-linux - .rules-main-push # TODO: remove, only for testing tags: - test-fhg-linux-runner1 stage: compare needs: ["build-codec-linux-cmake"] timeout: "30 minutes" # To be revisited Loading Loading @@ -691,7 +706,7 @@ codec-comparison-on-main-push: ### run pytest - exit_code=0 - python3 -m pytest tests -v --junit-xml=report-junit.xml || exit_code=$? - python3 -m pytest tests -v --html=report.html --junit-xml=report-junit.xml || exit_code=$? - if [ $exit_code -eq 1 ] && [ $non_be_flag == 0 ]; then echo "pytest run had failures and non-BE flag not present"; exit $EXIT_CODE_FAIL; fi - zero_errors=$(cat report-junit.xml | grep -c 'errors="0"') || true - if [ $exit_code -eq 1 ] && [ $zero_errors == 1 ]; then echo "pytest run had failures, but no errors and non-BE flag present"; exit $EXIT_CODE_NON_BE; fi Loading @@ -704,6 +719,7 @@ codec-comparison-on-main-push: when: always paths: - report-junit.xml - report.html expose_as: "Results of comparison to previous merge commit" reports: junit: report-junit.xml Loading