Skip to content

Linux unit tests

Linux unit tests #1561

Workflow file for this run

name: Linux unit tests
on:
workflow_dispatch:
inputs:
python_version:
required: true
type: string
ref:
required: false
type: string
check_run_id:
required: false
type: string
pr_repo:
required: false
type: string
push:
branches: devel
env:
COMMIT: ${{ inputs.ref || github.event.ref }}
PEM: ${{ secrets.BOT_PEM }}
GITHUB_RUN_ID: ${{ github.run_id }}
GITHUB_CHECK_RUN_ID: ${{ inputs.check_run_id }}
PR_REPO: ${{ inputs.pr_repo || github.repository }}
jobs:
matrix_prep:
runs-on: ubuntu-latest
if: github.event_name != 'push' || github.repository == 'pyccel/pyccel'
outputs:
matrix: ${{ steps.set-matrix.outputs.matrix }}
steps:
- id: push-matrix
if: github.event_name == 'push'
run: |
echo "matrix=['3.8', '3.9', '3.10', '3.11', '3.12']" >> $GITHUB_ENV
shell: bash
- id: dispatch-matrix
if: github.event_name == 'workflow_dispatch'
run: |
echo "matrix=[${{ inputs.python_version }}]" >> $GITHUB_ENV
shell: bash
- id: set-matrix
run: |
echo "$matrix"
echo "matrix={\"python_version\": $matrix}" >> $GITHUB_OUTPUT
shell: bash
Linux:
runs-on: ubuntu-latest
name: Unit tests
needs: matrix_prep
strategy:
matrix: ${{fromJson(needs.matrix_prep.outputs.matrix)}}
steps:
- uses: actions/checkout@v3
with:
ref: ${{ env.COMMIT }}
repository: ${{ env.PR_REPO }}
submodules: true
- name: Set up Python ${{ matrix.python_version }}
uses: actions/setup-python@v4
with:
python-version: ${{ matrix.python_version }}
- name: "Setup"
id: token
run: |
pip install jwt requests
python ci_tools/setup_check_run.py linux
- name: Install dependencies
uses: ./.github/actions/linux_install
- name: Install Pyccel with tests
run: |
python -m pip install --upgrade pip
python -m pip install .[test]
shell: bash
- name: Coverage install
uses: ./.github/actions/coverage_install
- name: Fortran/C tests with pytest
id: f_c_pytest
timeout-minutes: 60
uses: ./.github/actions/pytest_run
- name: Python tests with pytest
id: python_pytest
timeout-minutes: 20
uses: ./.github/actions/pytest_run_python
- name: Parallel tests with pytest
id: parallel
timeout-minutes: 20
uses: ./.github/actions/pytest_parallel
- name: Test with valgrind for memory leaks
id: valgrind
uses: ./.github/actions/valgrind_run
- name: Collect coverage information
continue-on-error: True
uses: ./.github/actions/coverage_collection
- name: Save code coverage report
uses: actions/upload-artifact@v3
with:
name: coverage-artifact
path: .coverage
retention-days: 1
- name: "Post completed"
if: always()
run:
python ci_tools/complete_check_run.py ${{ steps.f_c_pytest.outcome }} ${{ steps.python_pytest.outcome }} ${{ steps.parallel.outcome }} ${{ steps.valgrind.outcome }}