run all tests and only build docker image if tests pass w/ burnettk jbirddog

This commit is contained in:
jasquat 2022-11-11 12:39:46 -05:00
parent d5b7daf44e
commit 176f6a361f
1 changed files with 265 additions and 263 deletions

View File

@ -9,272 +9,273 @@ defaults:
working-directory: spiffworkflow-backend working-directory: spiffworkflow-backend
jobs: jobs:
# tests: tests:
# name: ${{ matrix.session }} ${{ matrix.python }} / ${{ matrix.os }} ${{ matrix.database }} name: ${{ matrix.session }} ${{ matrix.python }} / ${{ matrix.os }} ${{ matrix.database }}
# runs-on: ${{ matrix.os }} runs-on: ${{ matrix.os }}
# strategy: strategy:
# fail-fast: false fail-fast: false
# matrix: matrix:
# include: include:
# - { python: "3.11", os: "ubuntu-latest", session: "safety" } - { python: "3.11", os: "ubuntu-latest", session: "safety" }
# - { python: "3.11", os: "ubuntu-latest", session: "mypy" } - { python: "3.11", os: "ubuntu-latest", session: "mypy" }
# - { - {
# python: "3.11", python: "3.11",
# os: "ubuntu-latest", os: "ubuntu-latest",
# session: "tests", session: "tests",
# database: "mysql", database: "mysql",
# } }
# - { - {
# python: "3.11", python: "3.11",
# os: "ubuntu-latest", os: "ubuntu-latest",
# session: "tests", session: "tests",
# database: "postgres", database: "postgres",
# } }
# - { - {
# python: "3.11", python: "3.11",
# os: "ubuntu-latest", os: "ubuntu-latest",
# session: "tests", session: "tests",
# database: "sqlite", database: "sqlite",
# } }
# - { - {
# python: "3.11", python: "3.11",
# os: "macos-latest", os: "macos-latest",
# session: "tests", session: "tests",
# database: "sqlite", database: "sqlite",
# } }
# - { - {
# # typeguard 2.13.3 is broken with TypeDict in 3.11. # typeguard 2.13.3 is broken with TypeDict in 3.11.
# # probably the next release fixes it. # probably the next release fixes it.
# # https://github.com/agronholm/typeguard/issues/242 # https://github.com/agronholm/typeguard/issues/242
# python: "3.11", python: "3.11",
# os: "ubuntu-latest", os: "ubuntu-latest",
# session: "typeguard", session: "typeguard",
# database: "sqlite", database: "sqlite",
# } }
# - { python: "3.11", os: "ubuntu-latest", session: "xdoctest" } - { python: "3.11", os: "ubuntu-latest", session: "xdoctest" }
# - { python: "3.11", os: "ubuntu-latest", session: "docs-build" } - { python: "3.11", os: "ubuntu-latest", session: "docs-build" }
#
# env: env:
# NOXSESSION: ${{ matrix.session }} NOXSESSION: ${{ matrix.session }}
# SPIFF_DATABASE_TYPE: ${{ matrix.database }} SPIFF_DATABASE_TYPE: ${{ matrix.database }}
# FORCE_COLOR: "1" FORCE_COLOR: "1"
# PRE_COMMIT_COLOR: "always" PRE_COMMIT_COLOR: "always"
# DB_PASSWORD: password DB_PASSWORD: password
# FLASK_SESSION_SECRET_KEY: super_secret_key FLASK_SESSION_SECRET_KEY: super_secret_key
#
# steps: steps:
# - name: Check out the repository - name: Check out the repository
# uses: actions/checkout@v3.0.2 uses: actions/checkout@v3.0.2
#
# - name: Set up Python ${{ matrix.python }} - name: Set up Python ${{ matrix.python }}
# uses: actions/setup-python@v4.2.0 uses: actions/setup-python@v4.2.0
# with: with:
# python-version: ${{ matrix.python }} python-version: ${{ matrix.python }}
#
# - name: Upgrade pip - name: Upgrade pip
# run: | run: |
# pip install --constraint=.github/workflows/constraints.txt pip pip install --constraint=.github/workflows/constraints.txt pip
# pip --version pip --version
#
# - name: Upgrade pip in virtual environments - name: Upgrade pip in virtual environments
shell: python
run: |
import os
import pip
with open(os.environ["GITHUB_ENV"], mode="a") as io:
print(f"VIRTUALENV_PIP={pip.__version__}", file=io)
- name: Install Poetry
run: |
pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry
poetry --version
- name: Install Nox
run: |
pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox
pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry
nox --version
# - name: Compute pre-commit cache key
# if: matrix.session == 'pre-commit'
# id: pre-commit-cache
# shell: python # shell: python
# run: | # run: |
# import os # import hashlib
# import pip # import sys
# #
# with open(os.environ["GITHUB_ENV"], mode="a") as io: # python = "py{}.{}".format(*sys.version_info[:2])
# print(f"VIRTUALENV_PIP={pip.__version__}", file=io) # payload = sys.version.encode() + sys.executable.encode()
# digest = hashlib.sha256(payload).hexdigest()
# result = "${{ runner.os }}-{}-{}-pre-commit".format(python, digest[:8])
# #
# - name: Install Poetry # print("::set-output name=result::{}".format(result))
# run: |
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry
# poetry --version
# #
# - name: Install Nox # - name: Restore pre-commit cache
# run: | # uses: actions/cache@v3.0.11
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox # if: matrix.session == 'pre-commit'
# pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry
# nox --version
#
# # - name: Compute pre-commit cache key
# # if: matrix.session == 'pre-commit'
# # id: pre-commit-cache
# # shell: python
# # run: |
# # import hashlib
# # import sys
# #
# # python = "py{}.{}".format(*sys.version_info[:2])
# # payload = sys.version.encode() + sys.executable.encode()
# # digest = hashlib.sha256(payload).hexdigest()
# # result = "${{ runner.os }}-{}-{}-pre-commit".format(python, digest[:8])
# #
# # print("::set-output name=result::{}".format(result))
# #
# # - name: Restore pre-commit cache
# # uses: actions/cache@v3.0.11
# # if: matrix.session == 'pre-commit'
# # with:
# # path: ~/.cache/pre-commit
# # key: ${{ steps.pre-commit-cache.outputs.result }}-${{ hashFiles('.pre-commit-config.yaml') }}
# # restore-keys: |
# # ${{ steps.pre-commit-cache.outputs.result }}-
# - name: Setup Mysql
# uses: mirromutth/mysql-action@v1.1
# with: # with:
# host port: 3306 # path: ~/.cache/pre-commit
# container port: 3306 # key: ${{ steps.pre-commit-cache.outputs.result }}-${{ hashFiles('.pre-commit-config.yaml') }}
# mysql version: "8.0" # restore-keys: |
# mysql database: "spiffworkflow_backend_testing" # ${{ steps.pre-commit-cache.outputs.result }}-
# mysql root password: password - name: Setup Mysql
# if: matrix.database == 'mysql' uses: mirromutth/mysql-action@v1.1
# with:
# - name: Setup Postgres host port: 3306
# run: docker run --name postgres-spiff -p 5432:5432 -e POSTGRES_PASSWORD=spiffworkflow_backend -e POSTGRES_USER=spiffworkflow_backend -e POSTGRES_DB=spiffworkflow_backend_testing -d postgres container port: 3306
# if: matrix.database == 'postgres' mysql version: "8.0"
# mysql database: "spiffworkflow_backend_testing"
# - name: Run Nox mysql root password: password
# run: | if: matrix.database == 'mysql'
# nox --force-color --python=${{ matrix.python }}
# - name: Setup Postgres
# - name: Upload coverage data run: docker run --name postgres-spiff -p 5432:5432 -e POSTGRES_PASSWORD=spiffworkflow_backend -e POSTGRES_USER=spiffworkflow_backend -e POSTGRES_DB=spiffworkflow_backend_testing -d postgres
# # pin to upload coverage from only one matrix entry, otherwise coverage gets confused later if: matrix.database == 'postgres'
# if: always() && matrix.session == 'tests' && matrix.python == '3.11' && matrix.os == 'ubuntu-latest'
# uses: "actions/upload-artifact@v3.0.0" - name: Run Nox
# with: run: |
# name: coverage-data nox --force-color --python=${{ matrix.python }}
# path: ".coverage.*"
# - name: Upload coverage data
# - name: Upload documentation # pin to upload coverage from only one matrix entry, otherwise coverage gets confused later
# if: matrix.session == 'docs-build' if: always() && matrix.session == 'tests' && matrix.python == '3.11' && matrix.os == 'ubuntu-latest'
# uses: actions/upload-artifact@v3.0.0 uses: "actions/upload-artifact@v3.0.0"
# with: with:
# name: docs name: coverage-data
# path: docs/_build path: ".coverage.*"
#
# - name: Upload logs - name: Upload documentation
# if: failure() && matrix.session == 'tests' if: matrix.session == 'docs-build'
# uses: "actions/upload-artifact@v3.0.0" uses: actions/upload-artifact@v3.0.0
# with: with:
# name: logs-${{matrix.python}}-${{matrix.os}}-${{matrix.database}} name: docs
# path: "./log/*.log" path: docs/_build
#
# run_pre_commit_checks: - name: Upload logs
# runs-on: ubuntu-latest if: failure() && matrix.session == 'tests'
# defaults: uses: "actions/upload-artifact@v3.0.0"
# run: with:
# working-directory: . name: logs-${{matrix.python}}-${{matrix.os}}-${{matrix.database}}
# steps: path: "./log/*.log"
# - name: Check out the repository
# uses: actions/checkout@v3.0.2 run_pre_commit_checks:
# with: runs-on: ubuntu-latest
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud defaults:
# fetch-depth: 0 run:
# - name: Set up Python working-directory: .
# uses: actions/setup-python@v4.2.0 steps:
# with: - name: Check out the repository
# python-version: "3.11" uses: actions/checkout@v3.0.2
# - name: Install Poetry with:
# run: | # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
# pipx install poetry fetch-depth: 0
# poetry --version - name: Set up Python
# - name: Poetry Install uses: actions/setup-python@v4.2.0
# run: poetry install with:
# - name: run_pre_commit python-version: "3.11"
# run: ./bin/run_pre_commit_in_ci - name: Install Poetry
# run: |
# check_docker_start_script: pipx install poetry
# runs-on: ubuntu-latest poetry --version
# steps: - name: Poetry Install
# - name: Check out the repository run: poetry install
# uses: actions/checkout@v3.0.2 - name: run_pre_commit
# with: run: ./bin/run_pre_commit_in_ci
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
# fetch-depth: 0 check_docker_start_script:
# - name: start_backend runs-on: ubuntu-latest
# run: ./bin/build_and_run_with_docker_compose steps:
# timeout-minutes: 20 - name: Check out the repository
# env: uses: actions/checkout@v3.0.2
# SPIFFWORKFLOW_BACKEND_LOAD_FIXTURE_DATA: "true" with:
# - name: wait_for_backend # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
# run: ./bin/wait_for_server_to_be_up 5 fetch-depth: 0
# - name: start_backend
# coverage: run: ./bin/build_and_run_with_docker_compose
# runs-on: ubuntu-latest timeout-minutes: 20
# needs: tests env:
# steps: SPIFFWORKFLOW_BACKEND_LOAD_FIXTURE_DATA: "true"
# - name: Check out the repository - name: wait_for_backend
# uses: actions/checkout@v3.0.2 run: ./bin/wait_for_server_to_be_up 5
# with:
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud coverage:
# fetch-depth: 0 runs-on: ubuntu-latest
# needs: tests
# - name: Set up Python steps:
# uses: actions/setup-python@v4.2.0 - name: Check out the repository
# with: uses: actions/checkout@v3.0.2
# python-version: "3.11" with:
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
# - name: Upgrade pip fetch-depth: 0
# run: |
# pip install --constraint=.github/workflows/constraints.txt pip - name: Set up Python
# pip --version uses: actions/setup-python@v4.2.0
# with:
# - name: Install Poetry python-version: "3.11"
# run: |
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry - name: Upgrade pip
# poetry --version run: |
# pip install --constraint=.github/workflows/constraints.txt pip
# - name: Install Nox pip --version
# run: |
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox - name: Install Poetry
# pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry run: |
# nox --version pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry
# poetry --version
# - name: Download coverage data
# uses: actions/download-artifact@v3.0.1 - name: Install Nox
# with: run: |
# name: coverage-data pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox
# pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry
# - name: Combine coverage data and display human readable report nox --version
# run: |
# find . -name \*.pyc -delete - name: Download coverage data
# nox --force-color --session=coverage uses: actions/download-artifact@v3.0.1
# with:
# - name: Create coverage report name: coverage-data
# run: |
# nox --force-color --session=coverage -- xml - name: Combine coverage data and display human readable report
# run: |
# - name: Upload coverage report find . -name \*.pyc -delete
# uses: codecov/codecov-action@v3.1.0 nox --force-color --session=coverage
#
# - name: SonarCloud Scan - name: Create coverage report
# uses: sonarsource/sonarcloud-github-action@master run: |
# # thought about just skipping dependabot nox --force-color --session=coverage -- xml
# # if: ${{ github.actor != 'dependabot[bot]' }}
# # but figured all pull requests seems better, since none of them will have access to sonarcloud. - name: Upload coverage report
# # however, with just skipping pull requests, the build associated with "Triggered via push" is also associated with the pull request and also fails hitting sonarcloud uses: codecov/codecov-action@v3.1.0
# # if: ${{ github.event_name != 'pull_request' }}
# # so just skip everything but main - name: SonarCloud Scan
# if: github.ref_name == 'main' uses: sonarsource/sonarcloud-github-action@master
# env: # thought about just skipping dependabot
# GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} # if: ${{ github.actor != 'dependabot[bot]' }}
# SONAR_TOKEN: ${{ secrets.SONAR_TOKEN }} # but figured all pull requests seems better, since none of them will have access to sonarcloud.
# # part about saving PR number and then using it from auto-merge-dependabot-prs from: # however, with just skipping pull requests, the build associated with "Triggered via push" is also associated with the pull request and also fails hitting sonarcloud
# # https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#workflow_run # if: ${{ github.event_name != 'pull_request' }}
# - name: Save PR number # so just skip everything but main
# if: ${{ github.event_name == 'pull_request' }} if: github.ref_name == 'main'
# env: env:
# PR_NUMBER: ${{ github.event.number }} GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
# run: | SONAR_TOKEN: ${{ secrets.SONAR_TOKEN }}
# mkdir -p ./pr # part about saving PR number and then using it from auto-merge-dependabot-prs from:
# echo "$PR_NUMBER" > ./pr/pr_number # https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#workflow_run
# - uses: actions/upload-artifact@v3 - name: Save PR number
# with: if: ${{ github.event_name == 'pull_request' }}
# name: pr_number env:
# path: pr/ PR_NUMBER: ${{ github.event.number }}
run: |
mkdir -p ./pr
echo "$PR_NUMBER" > ./pr/pr_number
- uses: actions/upload-artifact@v3
with:
name: pr_number
path: pr/
build-and-push-image: build-and-push-image:
if: github.ref_name == 'main' && ${{ github.event_name == 'push' }} if: github.ref_name == 'main' && ${{ github.event_name == 'push' }}
needs: tests
env: env:
REGISTRY: ghcr.io REGISTRY: ghcr.io
IMAGE_NAME: sartography/spiffworkflow-backend IMAGE_NAME: sartography/spiffworkflow-backend
@ -312,6 +313,7 @@ jobs:
# like sub '/' with '-' # like sub '/' with '-'
TAG: ${{ github.ref_name }}-${{ steps.date.outputs.date }} TAG: ${{ github.ref_name }}-${{ steps.date.outputs.date }}
with: with:
# this action doesn't seem to respect working-directory so set context
context: spiffworkflow-backend context: spiffworkflow-backend
push: true push: true
tags: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}:${{ env.TAG }} tags: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}:${{ env.TAG }}