Skip to content

UAT C1215803153-GES_DISC (ML2GPH) #34128

UAT C1215803153-GES_DISC (ML2GPH)

UAT C1215803153-GES_DISC (ML2GPH) #34128

Workflow file for this run

# This workflow ...
name: Verify New Associations
on:
pull_request:
types: [ synchronize ]
jobs:
verify_collection:
outputs:
any_test_failed: ${{ steps.test-result.outputs.any_test_failed }}
any_test_skipped: ${{ steps.test-result.outputs.any_test_skipped }}
any_test_error: ${{ steps.test-result.outputs.any_test_error }}
all_tests_passed: ${{ steps.test-result.outputs.all_tests_passed }}
permissions:
checks: write
# only needed unless run with comment_mode: off
pull-requests: write
if: |
${{ startsWith(github.head_ref, 'diff/uat') }} ||
${{ startsWith(github.head_ref, 'diff/ops') }} ||
contains(github.event.issue.labels.*.name, 'autotest')
name: Verify ${{ github.event.pull_request.title }}
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v3
- uses: actions/setup-python@v4
with:
python-version: "3.10"
- name: Install Poetry
uses: abatilo/actions-poetry@v2
with:
poetry-version: 1.5.1
- name: Get env and concept id
id: get-env-ccid
run: |
echo "test_env=$(python -c "print('${{ github.head_ref }}'.split('/')[1])")" >> $GITHUB_OUTPUT
echo "concept_id=$(python -c "print('${{ github.head_ref }}'.split('/')[2])")" >> $GITHUB_OUTPUT
- name: Execute tests
id: run-tests
env:
TEST_ENV: ${{ steps.get-env-ccid.outputs.test_env }}
CONCEPT_ID: ${{ steps.get-env-ccid.outputs.concept_id }}
CMR_USER: ${{ secrets.CMR_USER }}
CMR_PASS: ${{ secrets.CMR_PASS }}
run: |
poetry install
poetry run pytest tests/verify_collection.py --concept_id $CONCEPT_ID --env $TEST_ENV --junitxml=$GITHUB_WORKSPACE/test-results/test_report.xml --html=$GITHUB_WORKSPACE/test-results/test_report.html || true
- name: Publish Test Results
id: publish-test
uses: EnricoMi/publish-unit-test-result-action@v2
if: always()
with:
report_suite_logs: info
check_name: Tested with Harmony
json_file: test-results/test_results.json
comment_title: Test Results for ${{ steps.get-env-ccid.outputs.concept_id }}
files: test-results/test_report.xml
- name: Save Test Outputs
id: test-result
if: always()
run: |
echo "any_test_failed=${{ fromJSON( steps.publish-test.outputs.json ).stats.tests_fail > 0 }}" >> $GITHUB_OUTPUT
echo "any_test_skipped=${{ fromJSON( steps.publish-test.outputs.json ).stats.tests_skip > 0 }}" >> $GITHUB_OUTPUT
echo "any_test_error=${{ fromJSON( steps.publish-test.outputs.json ).stats.tests_error > 0 }}" >> $GITHUB_OUTPUT
echo "all_tests_passed=${{ fromJSON( steps.publish-test.outputs.json ).stats.tests_succ == fromJSON( steps.publish-test.outputs.json ).stats.tests }}" >> $GITHUB_OUTPUT
echo "all_tests_passed=${{ fromJSON( steps.publish-test.outputs.json ).stats.tests_succ == fromJSON( steps.publish-test.outputs.json ).stats.tests }}" >> $GITHUB_ENV
# echo "any_test_failed=false" >> $GITHUB_OUTPUT
# echo "any_test_skipped=false" >> $GITHUB_OUTPUT
# echo "any_test_error=false" >> $GITHUB_OUTPUT
# echo "all_test_passed=true" >> $GITHUB_OUTPUT
- uses: actions/upload-artifact@v3
if: always()
with:
name: test-results
path: test-results/**
- name: Succeeded in Completing All Tests
uses: LouisBrunner/checks-action@v1.6.1
if: env.all_tests_passed == 'true'
with:
token: ${{ secrets.GITHUB_TOKEN }}
name: All Tests Run And Passed
conclusion: success
output: |
{"summary":"No errors, skipped tests, or failed tests detected"}
tests_pass:
needs: verify_collection
name: Merge Pull Request
runs-on: ubuntu-latest
if: fromJSON( needs.verify_collection.outputs.all_tests_passed )
steps:
- uses: actions-ecosystem/action-remove-labels@v1
with:
labels: |
unverified
- name: add labels
uses: actions-ecosystem/action-add-labels@v1
with:
labels: |
verified
tests_fail:
needs: verify_collection
name: Fail Pull Request
runs-on: ubuntu-latest
if: |
fromJSON( needs.verify_collection.outputs.any_test_failed ) ||
fromJSON( needs.verify_collection.outputs.any_test_error )
steps:
- uses: actions-ecosystem/action-remove-labels@v1
with:
labels: |
unverified
- name: add labels
uses: actions-ecosystem/action-add-labels@v1
with:
labels: |
failed verification
bug
- uses: mshick/add-pr-comment@v2
with:
message: |
**Action Needed**
Tests have failed or encountered an unknown error. Open the status checks to view the logs and review the failure.
tests_skipped:
needs: verify_collection
name: Update Pull Request
runs-on: ubuntu-latest
if: fromJSON( needs.verify_collection.outputs.any_test_skipped )
steps:
- uses: actions-ecosystem/action-remove-labels@v1
with:
labels: |
unverified
- name: add labels
uses: actions-ecosystem/action-add-labels@v1
with:
labels: |
tests skipped
- uses: mshick/add-pr-comment@v2
with:
message: |
**Action Needed**
Unable to verify collection because test(s) were skipped.
Open the status check to view the logs from the test run and determine why tests were skipped.
If there is a justifiable reason, add a comment here and manually merge this PR to accept this unverified association.
Otherwise, fix the problem that caused the test to be skipped and rerun verification.