Commit 32acfa9a authored by kinuthia's avatar kinuthia
Browse files

cleanup

parent 08b51ee7
Loading
Loading
Loading
Loading
Loading
+7 −35
Original line number Diff line number Diff line
@@ -1640,41 +1640,21 @@ voip-be-on-merge-request:

ivas-long-term-job-logs:
  rules:
    - if: $CI_PIPELINE_SOURCE == 'web' && $MANUAL_PIPELINE_TYPE == "long-term-logs" # change this to a scheduled job
    - if: $CI_PIPELINE_SOURCE == 'web' && $MANUAL_PIPELINE_TYPE == "long-term-logs"
    - if: $CI_PIPELINE_SOURCE == 'schedule'
  tags:
    - ivas-linux
  stage: .pre
  timeout: "20 minutes"
  before_script:
    - USE_LTV=1
    - DUT_DECODER_PATH=./$REF_DECODER_PATH
    - TEST_SUITE="$LONG_TEST_SUITE_ENCODER"
    - LEVEL_SCALING=1.0
    - SPLIT_COMPARISON="false"
    - GET_LOGS=1
  variables:
    # keep "mld" in artifact name for backwards compatibility reasons
    CSV_ARTIFACT_NAME: "mld--$CI_JOB_NAME-$CI_JOB_ID--sha-$CI_COMMIT_SHORT_SHA.csv"
    CSV_ARTIFACT_SPLIT: "mld--split--$CI_JOB_NAME-$CI_JOB_ID--sha-$CI_COMMIT_SHORT_SHA.csv"
    MERGED_CSV_ARTIFACT_NAME: "$CI_JOB_NAME--merged_csv--$CI_JOB_ID.csv"
    PAGES_HTML_ARTIFACT_NAME: "$CI_JOB_NAME-index.html"
    SUMMARY_HTML_ARTIFACT_NAME: "summary_$CI_JOB_NAME.html"
    SUMMARY_HTML_ARTIFACT_SPLIT: "summary_split_$CI_JOB_NAME.html"
    IMAGES_ARTIFACT_NAME: "images_$CI_JOB_NAME"
    IMAGES_ARTIFACT_SPLIT: "images_split_$CI_JOB_NAME"
  timeout: "10 minutes"
  script:
    - !reference [ .job-linux, before_script ]

    - set -euxo pipefail
    - bash "${CI_PROJECT_DIR}"/ivas-codec-ci/snippets/print-common-info.sh
    - bash "${CI_PROJECT_DIR}"/ivas-codec-ci/snippets/basop/update-scripts-repo.sh
    - if [ $USE_LTV -eq 1 ]; then
    -    bash "${CI_PROJECT_DIR}"/ivas-codec-ci/snippets/update-ltv-repo.sh
    -    bash "${CI_PROJECT_DIR}"/ivas-codec-ci/snippets/copy-ltv-files-to-testv-dir.sh
    -    testcase_timeout=$TESTCASE_TIMEOUT_LTV
    - else
    -    testcase_timeout=$TESTCASE_TIMEOUT_STV
    - fi

      # create logs dir if it doesn't exist
    - TODAY=$(date +'%Y-%m-%d')
    - mkdir -p logs/$TODAY

    # Aggregate job logs
    - |
@@ -1685,9 +1665,6 @@ ivas-long-term-job-logs:
        echo "Job ID from variables - "$job_name", Job ID from script - $id_previous"
        curl --request GET "https://forge.3gpp.org/rep/api/v4/projects/$CI_PROJECT_ID/jobs/$id_previous/artifacts" --output artifacts.zip
        unzip artifacts.zip -d previous_artifacts
        TODAY=$(date +'%Y-%m-%d')
        # create logs dir if it doesn't exist
        mkdir -p logs/$TODAY
        # This wildcard thingy relies on only one csv file being present per job
        mv previous_artifacts/mld--"$job_name"-$id_previous--sha-*.csv logs/$TODAY
        rm artifacts.zip
@@ -1704,13 +1681,8 @@ ivas-long-term-job-logs:
    expire_in: 1 week
    when: always
    paths:
      #- report-junit.xml
      #- report.html
      - logs
    expose_as: "ivas long term job logs results"
    #reports:
    #  junit:
    #    - report-junit.xml