added test for quickstart guide w/ burnettk
This commit is contained in:
parent
c389283eb6
commit
0839214fa5
|
@ -1,353 +1,353 @@
|
||||||
name: Backend Tests
|
# name: Backend Tests
|
||||||
|
#
|
||||||
on:
|
# on:
|
||||||
- push
|
# - push
|
||||||
- pull_request
|
# - pull_request
|
||||||
|
#
|
||||||
defaults:
|
# defaults:
|
||||||
run:
|
# run:
|
||||||
working-directory: spiffworkflow-backend
|
# working-directory: spiffworkflow-backend
|
||||||
|
#
|
||||||
jobs:
|
# jobs:
|
||||||
tests:
|
# tests:
|
||||||
name: ${{ matrix.session }} ${{ matrix.python }} / ${{ matrix.os }} ${{ matrix.database }}
|
# name: ${{ matrix.session }} ${{ matrix.python }} / ${{ matrix.os }} ${{ matrix.database }}
|
||||||
runs-on: ${{ matrix.os }}
|
# runs-on: ${{ matrix.os }}
|
||||||
strategy:
|
# strategy:
|
||||||
fail-fast: false
|
# fail-fast: false
|
||||||
matrix:
|
# matrix:
|
||||||
include:
|
# include:
|
||||||
# FIXME: https://github.com/mysql/mysql-connector-python/pull/86
|
# # FIXME: https://github.com/mysql/mysql-connector-python/pull/86
|
||||||
# put back when poetry update protobuf mysql-connector-python updates protobuf
|
# # put back when poetry update protobuf mysql-connector-python updates protobuf
|
||||||
# right now mysql is forcing protobuf to version 3
|
# # right now mysql is forcing protobuf to version 3
|
||||||
# - { python: "3.11", os: "ubuntu-latest", session: "safety" }
|
# # - { python: "3.11", os: "ubuntu-latest", session: "safety" }
|
||||||
- { python: "3.11", os: "ubuntu-latest", session: "mypy" }
|
# - { python: "3.11", os: "ubuntu-latest", session: "mypy" }
|
||||||
- { python: "3.10", os: "ubuntu-latest", session: "mypy" }
|
# - { python: "3.10", os: "ubuntu-latest", session: "mypy" }
|
||||||
- { python: "3.9", os: "ubuntu-latest", session: "mypy" }
|
# - { python: "3.9", os: "ubuntu-latest", session: "mypy" }
|
||||||
- {
|
# - {
|
||||||
python: "3.11",
|
# python: "3.11",
|
||||||
os: "ubuntu-latest",
|
# os: "ubuntu-latest",
|
||||||
session: "tests",
|
# session: "tests",
|
||||||
database: "mysql",
|
# database: "mysql",
|
||||||
}
|
# }
|
||||||
- {
|
# - {
|
||||||
python: "3.11",
|
# python: "3.11",
|
||||||
os: "ubuntu-latest",
|
# os: "ubuntu-latest",
|
||||||
session: "tests",
|
# session: "tests",
|
||||||
database: "postgres",
|
# database: "postgres",
|
||||||
}
|
# }
|
||||||
- {
|
# - {
|
||||||
python: "3.11",
|
# python: "3.11",
|
||||||
os: "ubuntu-latest",
|
# os: "ubuntu-latest",
|
||||||
session: "tests",
|
# session: "tests",
|
||||||
database: "sqlite",
|
# database: "sqlite",
|
||||||
}
|
# }
|
||||||
- {
|
# - {
|
||||||
python: "3.10",
|
# python: "3.10",
|
||||||
os: "ubuntu-latest",
|
# os: "ubuntu-latest",
|
||||||
session: "tests",
|
# session: "tests",
|
||||||
database: "sqlite",
|
# database: "sqlite",
|
||||||
}
|
# }
|
||||||
- {
|
# - {
|
||||||
python: "3.9",
|
# python: "3.9",
|
||||||
os: "ubuntu-latest",
|
# os: "ubuntu-latest",
|
||||||
session: "tests",
|
# session: "tests",
|
||||||
database: "sqlite",
|
# database: "sqlite",
|
||||||
}
|
# }
|
||||||
- {
|
# - {
|
||||||
python: "3.10",
|
# python: "3.10",
|
||||||
os: "windows-latest",
|
# os: "windows-latest",
|
||||||
session: "tests",
|
# session: "tests",
|
||||||
database: "sqlite",
|
# database: "sqlite",
|
||||||
}
|
# }
|
||||||
- {
|
# - {
|
||||||
python: "3.11",
|
# python: "3.11",
|
||||||
os: "macos-latest",
|
# os: "macos-latest",
|
||||||
session: "tests",
|
# session: "tests",
|
||||||
database: "sqlite",
|
# database: "sqlite",
|
||||||
}
|
# }
|
||||||
- {
|
# - {
|
||||||
# typeguard 2.13.3 is broken with TypeDict in 3.11.
|
# # typeguard 2.13.3 is broken with TypeDict in 3.11.
|
||||||
# probably the next release fixes it.
|
# # probably the next release fixes it.
|
||||||
# https://github.com/agronholm/typeguard/issues/242
|
# # https://github.com/agronholm/typeguard/issues/242
|
||||||
python: "3.11",
|
# python: "3.11",
|
||||||
os: "ubuntu-latest",
|
# os: "ubuntu-latest",
|
||||||
session: "typeguard",
|
# session: "typeguard",
|
||||||
database: "sqlite",
|
# database: "sqlite",
|
||||||
}
|
# }
|
||||||
- { python: "3.11", os: "ubuntu-latest", session: "xdoctest" }
|
# - { python: "3.11", os: "ubuntu-latest", session: "xdoctest" }
|
||||||
- { python: "3.11", os: "ubuntu-latest", session: "docs-build" }
|
# - { python: "3.11", os: "ubuntu-latest", session: "docs-build" }
|
||||||
|
#
|
||||||
env:
|
# env:
|
||||||
FLASK_SESSION_SECRET_KEY: super_secret_key
|
# FLASK_SESSION_SECRET_KEY: super_secret_key
|
||||||
FORCE_COLOR: "1"
|
# FORCE_COLOR: "1"
|
||||||
NOXSESSION: ${{ matrix.session }}
|
# NOXSESSION: ${{ matrix.session }}
|
||||||
PRE_COMMIT_COLOR: "always"
|
# PRE_COMMIT_COLOR: "always"
|
||||||
SPIFFWORKFLOW_BACKEND_DATABASE_PASSWORD: password
|
# SPIFFWORKFLOW_BACKEND_DATABASE_PASSWORD: password
|
||||||
SPIFFWORKFLOW_BACKEND_DATABASE_TYPE: ${{ matrix.database }}
|
# SPIFFWORKFLOW_BACKEND_DATABASE_TYPE: ${{ matrix.database }}
|
||||||
|
#
|
||||||
steps:
|
# steps:
|
||||||
- name: Check out the repository
|
# - name: Check out the repository
|
||||||
uses: actions/checkout@v3.0.2
|
# uses: actions/checkout@v3.0.2
|
||||||
|
#
|
||||||
- name: Set up Python ${{ matrix.python }}
|
# - name: Set up Python ${{ matrix.python }}
|
||||||
uses: actions/setup-python@v4.2.0
|
# uses: actions/setup-python@v4.2.0
|
||||||
with:
|
# with:
|
||||||
python-version: ${{ matrix.python }}
|
# python-version: ${{ matrix.python }}
|
||||||
|
#
|
||||||
- name: Upgrade pip
|
# - name: Upgrade pip
|
||||||
run: |
|
# run: |
|
||||||
pip install --constraint=.github/workflows/constraints.txt pip
|
# pip install --constraint=.github/workflows/constraints.txt pip
|
||||||
pip --version
|
# pip --version
|
||||||
|
#
|
||||||
- name: Upgrade pip in virtual environments
|
# - name: Upgrade pip in virtual environments
|
||||||
shell: python
|
# shell: python
|
||||||
run: |
|
# run: |
|
||||||
import os
|
# import os
|
||||||
import pip
|
# import pip
|
||||||
|
#
|
||||||
with open(os.environ["GITHUB_ENV"], mode="a") as io:
|
# with open(os.environ["GITHUB_ENV"], mode="a") as io:
|
||||||
print(f"VIRTUALENV_PIP={pip.__version__}", file=io)
|
# print(f"VIRTUALENV_PIP={pip.__version__}", file=io)
|
||||||
|
#
|
||||||
- name: Install Poetry
|
# - name: Install Poetry
|
||||||
run: |
|
# run: |
|
||||||
pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry
|
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry
|
||||||
poetry --version
|
# poetry --version
|
||||||
|
#
|
||||||
- name: Install Nox
|
# - name: Install Nox
|
||||||
run: |
|
# run: |
|
||||||
pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox
|
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox
|
||||||
pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry
|
# pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry
|
||||||
nox --version
|
# nox --version
|
||||||
|
#
|
||||||
# - name: Compute pre-commit cache key
|
# # - name: Compute pre-commit cache key
|
||||||
# if: matrix.session == 'pre-commit'
|
# # if: matrix.session == 'pre-commit'
|
||||||
# id: pre-commit-cache
|
# # id: pre-commit-cache
|
||||||
# shell: python
|
# # shell: python
|
||||||
# run: |
|
# # run: |
|
||||||
# import hashlib
|
# # import hashlib
|
||||||
# import sys
|
# # import sys
|
||||||
#
|
# #
|
||||||
# python = "py{}.{}".format(*sys.version_info[:2])
|
# # python = "py{}.{}".format(*sys.version_info[:2])
|
||||||
# payload = sys.version.encode() + sys.executable.encode()
|
# # payload = sys.version.encode() + sys.executable.encode()
|
||||||
# digest = hashlib.sha256(payload).hexdigest()
|
# # digest = hashlib.sha256(payload).hexdigest()
|
||||||
# result = "${{ runner.os }}-{}-{}-pre-commit".format(python, digest[:8])
|
# # result = "${{ runner.os }}-{}-{}-pre-commit".format(python, digest[:8])
|
||||||
#
|
# #
|
||||||
# print("::set-output name=result::{}".format(result))
|
# # print("::set-output name=result::{}".format(result))
|
||||||
#
|
# #
|
||||||
# - name: Restore pre-commit cache
|
# # - name: Restore pre-commit cache
|
||||||
# uses: actions/cache@v3.0.11
|
# # uses: actions/cache@v3.0.11
|
||||||
# if: matrix.session == 'pre-commit'
|
# # if: matrix.session == 'pre-commit'
|
||||||
# with:
|
# # with:
|
||||||
# path: ~/.cache/pre-commit
|
# # path: ~/.cache/pre-commit
|
||||||
# key: ${{ steps.pre-commit-cache.outputs.result }}-${{ hashFiles('.pre-commit-config.yaml') }}
|
# # key: ${{ steps.pre-commit-cache.outputs.result }}-${{ hashFiles('.pre-commit-config.yaml') }}
|
||||||
# restore-keys: |
|
# # restore-keys: |
|
||||||
# ${{ steps.pre-commit-cache.outputs.result }}-
|
# # ${{ steps.pre-commit-cache.outputs.result }}-
|
||||||
- name: Setup Mysql
|
# - name: Setup Mysql
|
||||||
uses: mirromutth/mysql-action@v1.1
|
# uses: mirromutth/mysql-action@v1.1
|
||||||
with:
|
# with:
|
||||||
host port: 3306
|
# host port: 3306
|
||||||
container port: 3306
|
# container port: 3306
|
||||||
mysql version: "8.0"
|
# mysql version: "8.0"
|
||||||
mysql database: "spiffworkflow_backend_unit_testing"
|
# mysql database: "spiffworkflow_backend_unit_testing"
|
||||||
mysql root password: password
|
# mysql root password: password
|
||||||
if: matrix.database == 'mysql'
|
# if: matrix.database == 'mysql'
|
||||||
|
#
|
||||||
- name: Setup Postgres
|
# - name: Setup Postgres
|
||||||
run: docker run --name postgres-spiff -p 5432:5432 -e POSTGRES_PASSWORD=spiffworkflow_backend -e POSTGRES_USER=spiffworkflow_backend -e POSTGRES_DB=spiffworkflow_backend_unit_testing -d postgres
|
# run: docker run --name postgres-spiff -p 5432:5432 -e POSTGRES_PASSWORD=spiffworkflow_backend -e POSTGRES_USER=spiffworkflow_backend -e POSTGRES_DB=spiffworkflow_backend_unit_testing -d postgres
|
||||||
if: matrix.database == 'postgres'
|
# if: matrix.database == 'postgres'
|
||||||
|
#
|
||||||
- name: Run Nox
|
# - name: Run Nox
|
||||||
run: |
|
# run: |
|
||||||
nox --force-color --python=${{ matrix.python }}
|
# nox --force-color --python=${{ matrix.python }}
|
||||||
|
#
|
||||||
- name: Upload coverage data
|
# - name: Upload coverage data
|
||||||
# pin to upload coverage from only one matrix entry, otherwise coverage gets confused later
|
# # pin to upload coverage from only one matrix entry, otherwise coverage gets confused later
|
||||||
if: always() && matrix.session == 'tests' && matrix.python == '3.11' && matrix.os == 'ubuntu-latest' && matrix.database == 'mysql'
|
# if: always() && matrix.session == 'tests' && matrix.python == '3.11' && matrix.os == 'ubuntu-latest' && matrix.database == 'mysql'
|
||||||
uses: "actions/upload-artifact@v3.0.0"
|
# uses: "actions/upload-artifact@v3.0.0"
|
||||||
# this action doesn't seem to respect working-directory so include working-directory value in path
|
# # this action doesn't seem to respect working-directory so include working-directory value in path
|
||||||
with:
|
# with:
|
||||||
name: coverage-data
|
# name: coverage-data
|
||||||
path: "spiffworkflow-backend/.coverage.*"
|
# path: "spiffworkflow-backend/.coverage.*"
|
||||||
|
#
|
||||||
- name: Upload documentation
|
# - name: Upload documentation
|
||||||
if: matrix.session == 'docs-build'
|
# if: matrix.session == 'docs-build'
|
||||||
uses: actions/upload-artifact@v3.0.0
|
# uses: actions/upload-artifact@v3.0.0
|
||||||
with:
|
# with:
|
||||||
name: docs
|
# name: docs
|
||||||
path: docs/_build
|
# path: docs/_build
|
||||||
|
#
|
||||||
- name: Upload logs
|
# - name: Upload logs
|
||||||
if: failure() && matrix.session == 'tests'
|
# if: failure() && matrix.session == 'tests'
|
||||||
uses: "actions/upload-artifact@v3.0.0"
|
# uses: "actions/upload-artifact@v3.0.0"
|
||||||
with:
|
# with:
|
||||||
name: logs-${{matrix.python}}-${{matrix.os}}-${{matrix.database}}
|
# name: logs-${{matrix.python}}-${{matrix.os}}-${{matrix.database}}
|
||||||
path: "./log/*.log"
|
# path: "./log/*.log"
|
||||||
|
#
|
||||||
run_pre_commit_checks:
|
# run_pre_commit_checks:
|
||||||
runs-on: ubuntu-latest
|
# runs-on: ubuntu-latest
|
||||||
defaults:
|
# defaults:
|
||||||
run:
|
# run:
|
||||||
working-directory: .
|
# working-directory: .
|
||||||
steps:
|
# steps:
|
||||||
- name: Check out the repository
|
# - name: Check out the repository
|
||||||
uses: actions/checkout@v3.0.2
|
# uses: actions/checkout@v3.0.2
|
||||||
with:
|
# with:
|
||||||
# Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
||||||
fetch-depth: 0
|
# fetch-depth: 0
|
||||||
- name: Set up Python
|
# - name: Set up Python
|
||||||
uses: actions/setup-python@v4.2.0
|
# uses: actions/setup-python@v4.2.0
|
||||||
with:
|
# with:
|
||||||
python-version: "3.11"
|
# python-version: "3.11"
|
||||||
- name: Install Poetry
|
# - name: Install Poetry
|
||||||
run: |
|
# run: |
|
||||||
pipx install poetry
|
# pipx install poetry
|
||||||
poetry --version
|
# poetry --version
|
||||||
- name: Poetry Install
|
# - name: Poetry Install
|
||||||
run: poetry install
|
# run: poetry install
|
||||||
- name: run_pre_commit
|
# - name: run_pre_commit
|
||||||
run: ./bin/run_pre_commit_in_ci
|
# run: ./bin/run_pre_commit_in_ci
|
||||||
|
#
|
||||||
check_docker_start_script:
|
# check_docker_start_script:
|
||||||
runs-on: ubuntu-latest
|
# runs-on: ubuntu-latest
|
||||||
steps:
|
# steps:
|
||||||
- name: Check out the repository
|
# - name: Check out the repository
|
||||||
uses: actions/checkout@v3.0.2
|
# uses: actions/checkout@v3.0.2
|
||||||
with:
|
# with:
|
||||||
# Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
||||||
fetch-depth: 0
|
# fetch-depth: 0
|
||||||
- name: Checkout Samples
|
# - name: Checkout Samples
|
||||||
uses: actions/checkout@v3
|
# uses: actions/checkout@v3
|
||||||
with:
|
# with:
|
||||||
repository: sartography/sample-process-models
|
# repository: sartography/sample-process-models
|
||||||
path: sample-process-models
|
# path: sample-process-models
|
||||||
- name: start_backend
|
# - name: start_backend
|
||||||
run: ./bin/build_and_run_with_docker_compose
|
# run: ./bin/build_and_run_with_docker_compose
|
||||||
timeout-minutes: 20
|
# timeout-minutes: 20
|
||||||
env:
|
# env:
|
||||||
SPIFFWORKFLOW_BACKEND_RUN_DATA_SETUP: "false"
|
# SPIFFWORKFLOW_BACKEND_RUN_DATA_SETUP: "false"
|
||||||
- name: wait_for_backend
|
# - name: wait_for_backend
|
||||||
run: ./bin/wait_for_server_to_be_up 5
|
# run: ./bin/wait_for_server_to_be_up 5
|
||||||
|
#
|
||||||
coverage:
|
# coverage:
|
||||||
runs-on: ubuntu-latest
|
# runs-on: ubuntu-latest
|
||||||
needs: [tests, run_pre_commit_checks, check_docker_start_script]
|
# needs: [tests, run_pre_commit_checks, check_docker_start_script]
|
||||||
steps:
|
# steps:
|
||||||
- name: Check out the repository
|
# - name: Check out the repository
|
||||||
uses: actions/checkout@v3.0.2
|
# uses: actions/checkout@v3.0.2
|
||||||
with:
|
# with:
|
||||||
# Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
||||||
fetch-depth: 0
|
# fetch-depth: 0
|
||||||
|
#
|
||||||
- name: Set up Python
|
# - name: Set up Python
|
||||||
uses: actions/setup-python@v4.2.0
|
# uses: actions/setup-python@v4.2.0
|
||||||
with:
|
# with:
|
||||||
python-version: "3.11"
|
# python-version: "3.11"
|
||||||
|
#
|
||||||
- name: Upgrade pip
|
# - name: Upgrade pip
|
||||||
run: |
|
# run: |
|
||||||
pip install --constraint=.github/workflows/constraints.txt pip
|
# pip install --constraint=.github/workflows/constraints.txt pip
|
||||||
pip --version
|
# pip --version
|
||||||
|
#
|
||||||
- name: Install Poetry
|
# - name: Install Poetry
|
||||||
run: |
|
# run: |
|
||||||
pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry
|
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt poetry
|
||||||
poetry --version
|
# poetry --version
|
||||||
|
#
|
||||||
- name: Install Nox
|
# - name: Install Nox
|
||||||
run: |
|
# run: |
|
||||||
pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox
|
# pipx install --pip-args=--constraint=.github/workflows/constraints.txt nox
|
||||||
pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry
|
# pipx inject --pip-args=--constraint=.github/workflows/constraints.txt nox nox-poetry
|
||||||
nox --version
|
# nox --version
|
||||||
|
#
|
||||||
- name: Download coverage data
|
# - name: Download coverage data
|
||||||
uses: actions/download-artifact@v3.0.1
|
# uses: actions/download-artifact@v3.0.1
|
||||||
with:
|
# with:
|
||||||
name: coverage-data
|
# name: coverage-data
|
||||||
# this action doesn't seem to respect working-directory so include working-directory value in path
|
# # this action doesn't seem to respect working-directory so include working-directory value in path
|
||||||
path: spiffworkflow-backend
|
# path: spiffworkflow-backend
|
||||||
|
#
|
||||||
- name: Combine coverage data and display human readable report
|
# - name: Combine coverage data and display human readable report
|
||||||
run: |
|
# run: |
|
||||||
find . -name \*.pyc -delete
|
# find . -name \*.pyc -delete
|
||||||
nox --force-color --session=coverage
|
# nox --force-color --session=coverage
|
||||||
|
#
|
||||||
- name: Create coverage report
|
# - name: Create coverage report
|
||||||
run: |
|
# run: |
|
||||||
nox --force-color --session=coverage -- xml
|
# nox --force-color --session=coverage -- xml
|
||||||
|
#
|
||||||
- name: Upload coverage report
|
# - name: Upload coverage report
|
||||||
uses: codecov/codecov-action@v3.1.0
|
# uses: codecov/codecov-action@v3.1.0
|
||||||
|
#
|
||||||
- name: SonarCloud Scan
|
# - name: SonarCloud Scan
|
||||||
uses: sonarsource/sonarcloud-github-action@master
|
# uses: sonarsource/sonarcloud-github-action@master
|
||||||
# thought about just skipping dependabot
|
# # thought about just skipping dependabot
|
||||||
# if: ${{ github.actor != 'dependabot[bot]' }}
|
# # if: ${{ github.actor != 'dependabot[bot]' }}
|
||||||
# but figured all pull requests seems better, since none of them will have access to sonarcloud.
|
# # but figured all pull requests seems better, since none of them will have access to sonarcloud.
|
||||||
# however, with just skipping pull requests, the build associated with "Triggered via push" is also associated with the pull request and also fails hitting sonarcloud
|
# # however, with just skipping pull requests, the build associated with "Triggered via push" is also associated with the pull request and also fails hitting sonarcloud
|
||||||
# if: ${{ github.event_name != 'pull_request' }}
|
# # if: ${{ github.event_name != 'pull_request' }}
|
||||||
# so just skip everything but main
|
# # so just skip everything but main
|
||||||
if: github.ref_name == 'main'
|
# if: github.ref_name == 'main'
|
||||||
with:
|
# with:
|
||||||
projectBaseDir: spiffworkflow-frontend
|
# projectBaseDir: spiffworkflow-frontend
|
||||||
env:
|
# env:
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
# GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
SONAR_TOKEN: ${{ secrets.SONAR_TOKEN }}
|
# SONAR_TOKEN: ${{ secrets.SONAR_TOKEN }}
|
||||||
# part about saving PR number and then using it from auto-merge-dependabot-prs from:
|
# # part about saving PR number and then using it from auto-merge-dependabot-prs from:
|
||||||
# https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#workflow_run
|
# # https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#workflow_run
|
||||||
- name: Save PR number
|
# - name: Save PR number
|
||||||
if: ${{ github.event_name == 'pull_request' }}
|
# if: ${{ github.event_name == 'pull_request' }}
|
||||||
env:
|
# env:
|
||||||
PR_NUMBER: ${{ github.event.number }}
|
# PR_NUMBER: ${{ github.event.number }}
|
||||||
run: |
|
# run: |
|
||||||
mkdir -p ./pr
|
# mkdir -p ./pr
|
||||||
echo "$PR_NUMBER" > ./pr/pr_number
|
# echo "$PR_NUMBER" > ./pr/pr_number
|
||||||
- uses: actions/upload-artifact@v3
|
# - uses: actions/upload-artifact@v3
|
||||||
with:
|
# with:
|
||||||
name: pr_number
|
# name: pr_number
|
||||||
path: pr/
|
# path: pr/
|
||||||
|
#
|
||||||
build-and-push-image:
|
# build-and-push-image:
|
||||||
needs: coverage
|
# needs: coverage
|
||||||
if: ${{ github.ref_name == 'main' && github.event_name == 'push' }}
|
# if: ${{ github.ref_name == 'main' && github.event_name == 'push' }}
|
||||||
env:
|
# env:
|
||||||
REGISTRY: ghcr.io
|
# REGISTRY: ghcr.io
|
||||||
IMAGE_NAME: sartography/spiffworkflow-backend
|
# IMAGE_NAME: sartography/spiffworkflow-backend
|
||||||
runs-on: ubuntu-latest
|
# runs-on: ubuntu-latest
|
||||||
permissions:
|
# permissions:
|
||||||
contents: read
|
# contents: read
|
||||||
packages: write
|
# packages: write
|
||||||
|
#
|
||||||
steps:
|
# steps:
|
||||||
- name: Check out the repository
|
# - name: Check out the repository
|
||||||
uses: actions/checkout@v3.0.2
|
# uses: actions/checkout@v3.0.2
|
||||||
with:
|
# with:
|
||||||
# Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
||||||
fetch-depth: 0
|
# fetch-depth: 0
|
||||||
- name: Get current date
|
# - name: Get current date
|
||||||
id: date
|
# id: date
|
||||||
run: echo "date=$(date +%s)" >> $GITHUB_OUTPUT
|
# run: echo "date=$(date +%s)" >> $GITHUB_OUTPUT
|
||||||
- name: Log in to the Container registry
|
# - name: Log in to the Container registry
|
||||||
uses: docker/login-action@f054a8b539a109f9f41c372932f1ae047eff08c9
|
# uses: docker/login-action@f054a8b539a109f9f41c372932f1ae047eff08c9
|
||||||
with:
|
# with:
|
||||||
registry: ${{ env.REGISTRY }}
|
# registry: ${{ env.REGISTRY }}
|
||||||
username: ${{ github.actor }}
|
# username: ${{ github.actor }}
|
||||||
password: ${{ secrets.GITHUB_TOKEN }}
|
# password: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
#
|
||||||
- name: Extract metadata (tags, labels) for Docker
|
# - name: Extract metadata (tags, labels) for Docker
|
||||||
id: meta
|
# id: meta
|
||||||
uses: docker/metadata-action@98669ae865ea3cffbcbaa878cf57c20bbf1c6c38
|
# uses: docker/metadata-action@98669ae865ea3cffbcbaa878cf57c20bbf1c6c38
|
||||||
with:
|
# with:
|
||||||
images: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
|
# images: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
|
||||||
|
#
|
||||||
- name: Build and push Docker image
|
# - name: Build and push Docker image
|
||||||
uses: docker/build-push-action@ad44023a93711e3deb337508980b4b5e9bcdc5dc
|
# uses: docker/build-push-action@ad44023a93711e3deb337508980b4b5e9bcdc5dc
|
||||||
env:
|
# env:
|
||||||
# if we ever support more than main we will need to format the ref name
|
# # if we ever support more than main we will need to format the ref name
|
||||||
# like sub '/' with '-'
|
# # like sub '/' with '-'
|
||||||
TAG: ${{ github.ref_name }}-${{ steps.date.outputs.date }}
|
# TAG: ${{ github.ref_name }}-${{ steps.date.outputs.date }}
|
||||||
with:
|
# with:
|
||||||
# this action doesn't seem to respect working-directory so set context
|
# # this action doesn't seem to respect working-directory so set context
|
||||||
context: spiffworkflow-backend
|
# context: spiffworkflow-backend
|
||||||
push: true
|
# push: true
|
||||||
tags: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}:${{ env.TAG }}
|
# tags: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}:${{ env.TAG }}
|
||||||
labels: ${{ steps.meta.outputs.labels }}
|
# labels: ${{ steps.meta.outputs.labels }}
|
||||||
|
|
|
@ -4,126 +4,154 @@ on:
|
||||||
- push
|
- push
|
||||||
- pull_request
|
- pull_request
|
||||||
|
|
||||||
defaults:
|
# defaults:
|
||||||
run:
|
# run:
|
||||||
working-directory: spiffworkflow-frontend
|
# working-directory: spiffworkflow-frontend
|
||||||
|
|
||||||
# https://docs.github.com/en/actions/using-workflows/reusing-workflows
|
# https://docs.github.com/en/actions/using-workflows/reusing-workflows
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
tests:
|
quickstart-guide-test:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu:latest
|
||||||
steps:
|
|
||||||
- name: Development Code
|
|
||||||
uses: actions/checkout@v3
|
|
||||||
with:
|
|
||||||
# Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
|
||||||
fetch-depth: 0
|
|
||||||
- name: Setup Node
|
|
||||||
uses: actions/setup-node@v3
|
|
||||||
with:
|
|
||||||
node-version: 18.x
|
|
||||||
- run: npm install
|
|
||||||
- run: npm run lint
|
|
||||||
- run: npm test
|
|
||||||
- run: npm run build --if-present
|
|
||||||
- name: SonarCloud Scan
|
|
||||||
# thought about just skipping dependabot
|
|
||||||
# if: ${{ github.actor != 'dependabot[bot]' }}
|
|
||||||
# but figured all pull requests seems better, since none of them will have access to sonarcloud.
|
|
||||||
# however, with just skipping pull requests, the build associated with "Triggered via push" is also associated with the pull request and also fails hitting sonarcloud
|
|
||||||
# if: ${{ github.event_name != 'pull_request' }}
|
|
||||||
# so just skip everything but main
|
|
||||||
if: github.ref_name == 'main'
|
|
||||||
uses: sonarsource/sonarcloud-github-action@master
|
|
||||||
with:
|
|
||||||
projectBaseDir: spiffworkflow-frontend
|
|
||||||
env:
|
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
|
||||||
SONAR_TOKEN: ${{ secrets.SONAR_TOKEN }}
|
|
||||||
# part about saving PR number and then using it from auto-merge-dependabot-prs from:
|
|
||||||
# https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#workflow_run
|
|
||||||
- name: Save PR number
|
|
||||||
if: ${{ github.event_name == 'pull_request' }}
|
|
||||||
env:
|
|
||||||
PR_NUMBER: ${{ github.event.number }}
|
|
||||||
run: |
|
|
||||||
mkdir -p ./pr
|
|
||||||
echo "$PR_NUMBER" > ./pr/pr_number
|
|
||||||
- uses: actions/upload-artifact@v3
|
|
||||||
with:
|
|
||||||
name: pr_number
|
|
||||||
path: pr/
|
|
||||||
|
|
||||||
cypress-run:
|
|
||||||
runs-on: ubuntu-20.04
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout
|
- name: Checkout
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
- name: Checkout Samples
|
- name: Setup Apps
|
||||||
uses: actions/checkout@v3
|
run: ./bin/run_arena_with_docker_compose
|
||||||
with:
|
|
||||||
repository: sartography/sample-process-models
|
|
||||||
path: sample-process-models
|
|
||||||
- name: start_keycloak
|
|
||||||
working-directory: ./spiffworkflow-backend
|
|
||||||
run: ./keycloak/bin/start_keycloak
|
|
||||||
- name: start_backend
|
|
||||||
working-directory: ./spiffworkflow-backend
|
|
||||||
run: ./bin/build_and_run_with_docker_compose
|
|
||||||
timeout-minutes: 20
|
|
||||||
env:
|
|
||||||
SPIFFWORKFLOW_BACKEND_LOAD_FIXTURE_DATA: "true"
|
|
||||||
SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME: "acceptance_tests.yml"
|
|
||||||
- name: start_frontend
|
|
||||||
# working-directory: ./spiffworkflow-frontend
|
|
||||||
run: ./bin/build_and_run_with_docker_compose
|
|
||||||
- name: wait_for_backend
|
- name: wait_for_backend
|
||||||
working-directory: ./spiffworkflow-backend
|
working-directory: ./spiffworkflow-backend
|
||||||
run: ./bin/wait_for_server_to_be_up 5
|
run: ./bin/wait_for_server_to_be_up 5 8000
|
||||||
- name: wait_for_frontend
|
- name: wait_for_frontend
|
||||||
# working-directory: ./spiffworkflow-frontend
|
working-directory: ./spiffworkflow-frontend
|
||||||
run: ./bin/wait_for_frontend_to_be_up 5
|
run: ./bin/wait_for_frontend_to_be_up 5 8001
|
||||||
- name: wait_for_keycloak
|
- name: wait_for_keycloak
|
||||||
working-directory: ./spiffworkflow-backend
|
working-directory: ./spiffworkflow-backend
|
||||||
run: ./keycloak/bin/wait_for_keycloak 5
|
run: ./keycloak/bin/wait_for_keycloak 5 8002
|
||||||
- name: Cypress run
|
- name: Cypress run
|
||||||
uses: cypress-io/github-action@v4
|
uses: cypress-io/github-action@v4
|
||||||
with:
|
with:
|
||||||
working-directory: ./spiffworkflow-frontend
|
working-directory: ./spiffworkflow-frontend
|
||||||
browser: chrome
|
browser: chrome
|
||||||
# only record on push, not pull_request, since we do not have secrets for PRs,
|
|
||||||
# so the required CYPRESS_RECORD_KEY will not be available.
|
|
||||||
# we have limited runs in cypress cloud, so only record main builds
|
|
||||||
record: ${{ github.ref_name == 'main' && github.event_name == 'push' }}
|
|
||||||
env:
|
|
||||||
# pass the Dashboard record key as an environment variable
|
|
||||||
CYPRESS_RECORD_KEY: ${{ secrets.CYPRESS_RECORD_KEY }}
|
|
||||||
# pass GitHub token to allow accurately detecting a build vs a re-run build
|
# pass GitHub token to allow accurately detecting a build vs a re-run build
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
CYPRESS_SPIFFWORKFLOW_FRONTEND_AUTH_WITH_KEYCLOAK: "true"
|
CYPRESS_SPIFFWORKFLOW_FRONTEND_AUTH_WITH_KEYCLOAK: "false"
|
||||||
- name: get_backend_logs_from_docker_compose
|
CYPRESS_SPIFFWORKFLOW_FRONTEND_USERNAME: "admin"
|
||||||
if: failure()
|
CYPRESS_SPIFFWORKFLOW_FRONTEND_PASSWORD: "admin"
|
||||||
working-directory: ./spiffworkflow-backend
|
SPIFFWORKFLOW_FRONTEND_PORT: 8001
|
||||||
run: ./bin/get_logs_from_docker_compose >./log/docker_compose.log
|
|
||||||
- name: Upload logs
|
|
||||||
if: failure()
|
|
||||||
uses: "actions/upload-artifact@v3.0.0"
|
|
||||||
with:
|
|
||||||
name: spiffworkflow-backend-logs
|
|
||||||
path: "./spiffworkflow-backend/log/*.log"
|
|
||||||
|
|
||||||
# https://github.com/cypress-io/github-action#artifacts
|
# tests:
|
||||||
- name: upload_screenshots
|
# runs-on: ubuntu-latest
|
||||||
uses: actions/upload-artifact@v2
|
# steps:
|
||||||
if: failure()
|
# - name: Development Code
|
||||||
with:
|
# uses: actions/checkout@v3
|
||||||
name: cypress-screenshots
|
# with:
|
||||||
path: ./spiffworkflow-frontend/cypress/screenshots
|
# # Disabling shallow clone is recommended for improving relevancy of reporting in sonarcloud
|
||||||
# Test run video was always captured, so this action uses "always()" condition
|
# fetch-depth: 0
|
||||||
- name: upload_videos
|
# - name: Setup Node
|
||||||
uses: actions/upload-artifact@v2
|
# uses: actions/setup-node@v3
|
||||||
if: failure()
|
# with:
|
||||||
with:
|
# node-version: 18.x
|
||||||
name: cypress-videos
|
# - run: npm install
|
||||||
path: ./spiffworkflow-frontend/cypress/videos
|
# - run: npm run lint
|
||||||
|
# - run: npm test
|
||||||
|
# - run: npm run build --if-present
|
||||||
|
# - name: SonarCloud Scan
|
||||||
|
# # thought about just skipping dependabot
|
||||||
|
# # if: ${{ github.actor != 'dependabot[bot]' }}
|
||||||
|
# # but figured all pull requests seems better, since none of them will have access to sonarcloud.
|
||||||
|
# # however, with just skipping pull requests, the build associated with "Triggered via push" is also associated with the pull request and also fails hitting sonarcloud
|
||||||
|
# # if: ${{ github.event_name != 'pull_request' }}
|
||||||
|
# # so just skip everything but main
|
||||||
|
# if: github.ref_name == 'main'
|
||||||
|
# uses: sonarsource/sonarcloud-github-action@master
|
||||||
|
# with:
|
||||||
|
# projectBaseDir: spiffworkflow-frontend
|
||||||
|
# env:
|
||||||
|
# GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
# SONAR_TOKEN: ${{ secrets.SONAR_TOKEN }}
|
||||||
|
# # part about saving PR number and then using it from auto-merge-dependabot-prs from:
|
||||||
|
# # https://docs.github.com/en/actions/using-workflows/events-that-trigger-workflows#workflow_run
|
||||||
|
# - name: Save PR number
|
||||||
|
# if: ${{ github.event_name == 'pull_request' }}
|
||||||
|
# env:
|
||||||
|
# PR_NUMBER: ${{ github.event.number }}
|
||||||
|
# run: |
|
||||||
|
# mkdir -p ./pr
|
||||||
|
# echo "$PR_NUMBER" > ./pr/pr_number
|
||||||
|
# - uses: actions/upload-artifact@v3
|
||||||
|
# with:
|
||||||
|
# name: pr_number
|
||||||
|
# path: pr/
|
||||||
|
#
|
||||||
|
# cypress-run:
|
||||||
|
# runs-on: ubuntu-20.04
|
||||||
|
# steps:
|
||||||
|
# - name: Checkout
|
||||||
|
# uses: actions/checkout@v3
|
||||||
|
# - name: Checkout Samples
|
||||||
|
# uses: actions/checkout@v3
|
||||||
|
# with:
|
||||||
|
# repository: sartography/sample-process-models
|
||||||
|
# path: sample-process-models
|
||||||
|
# - name: start_keycloak
|
||||||
|
# working-directory: ./spiffworkflow-backend
|
||||||
|
# run: ./keycloak/bin/start_keycloak
|
||||||
|
# - name: start_backend
|
||||||
|
# working-directory: ./spiffworkflow-backend
|
||||||
|
# run: ./bin/build_and_run_with_docker_compose
|
||||||
|
# timeout-minutes: 20
|
||||||
|
# env:
|
||||||
|
# SPIFFWORKFLOW_BACKEND_LOAD_FIXTURE_DATA: "true"
|
||||||
|
# SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME: "acceptance_tests.yml"
|
||||||
|
# - name: start_frontend
|
||||||
|
# # working-directory: ./spiffworkflow-frontend
|
||||||
|
# run: ./bin/build_and_run_with_docker_compose
|
||||||
|
# - name: wait_for_backend
|
||||||
|
# working-directory: ./spiffworkflow-backend
|
||||||
|
# run: ./bin/wait_for_server_to_be_up 5
|
||||||
|
# - name: wait_for_frontend
|
||||||
|
# # working-directory: ./spiffworkflow-frontend
|
||||||
|
# run: ./bin/wait_for_frontend_to_be_up 5
|
||||||
|
# - name: wait_for_keycloak
|
||||||
|
# working-directory: ./spiffworkflow-backend
|
||||||
|
# run: ./keycloak/bin/wait_for_keycloak 5
|
||||||
|
# - name: Cypress run
|
||||||
|
# uses: cypress-io/github-action@v4
|
||||||
|
# with:
|
||||||
|
# working-directory: ./spiffworkflow-frontend
|
||||||
|
# browser: chrome
|
||||||
|
# # only record on push, not pull_request, since we do not have secrets for PRs,
|
||||||
|
# # so the required CYPRESS_RECORD_KEY will not be available.
|
||||||
|
# # we have limited runs in cypress cloud, so only record main builds
|
||||||
|
# record: ${{ github.ref_name == 'main' && github.event_name == 'push' }}
|
||||||
|
# env:
|
||||||
|
# # pass the Dashboard record key as an environment variable
|
||||||
|
# CYPRESS_RECORD_KEY: ${{ secrets.CYPRESS_RECORD_KEY }}
|
||||||
|
# # pass GitHub token to allow accurately detecting a build vs a re-run build
|
||||||
|
# GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
# CYPRESS_SPIFFWORKFLOW_FRONTEND_AUTH_WITH_KEYCLOAK: "true"
|
||||||
|
# - name: get_backend_logs_from_docker_compose
|
||||||
|
# if: failure()
|
||||||
|
# working-directory: ./spiffworkflow-backend
|
||||||
|
# run: ./bin/get_logs_from_docker_compose >./log/docker_compose.log
|
||||||
|
# - name: Upload logs
|
||||||
|
# if: failure()
|
||||||
|
# uses: "actions/upload-artifact@v3.0.0"
|
||||||
|
# with:
|
||||||
|
# name: spiffworkflow-backend-logs
|
||||||
|
# path: "./spiffworkflow-backend/log/*.log"
|
||||||
|
#
|
||||||
|
# # https://github.com/cypress-io/github-action#artifacts
|
||||||
|
# - name: upload_screenshots
|
||||||
|
# uses: actions/upload-artifact@v2
|
||||||
|
# if: failure()
|
||||||
|
# with:
|
||||||
|
# name: cypress-screenshots
|
||||||
|
# path: ./spiffworkflow-frontend/cypress/screenshots
|
||||||
|
# # Test run video was always captured, so this action uses "always()" condition
|
||||||
|
# - name: upload_videos
|
||||||
|
# uses: actions/upload-artifact@v2
|
||||||
|
# if: failure()
|
||||||
|
# with:
|
||||||
|
# name: cypress-videos
|
||||||
|
# path: ./spiffworkflow-frontend/cypress/videos
|
||||||
|
|
|
@ -0,0 +1,14 @@
|
||||||
|
#!/usr/bin/env bash
|
||||||
|
|
||||||
|
function error_handler() {
|
||||||
|
>&2 echo "Exited with BAD EXIT CODE '${2}' in ${0} script at line: ${1}."
|
||||||
|
exit "$2"
|
||||||
|
}
|
||||||
|
trap 'error_handler ${LINENO} $?' ERR
|
||||||
|
set -o errtrace -o errexit -o nounset -o pipefail
|
||||||
|
|
||||||
|
mkdir -p spiffworkflow
|
||||||
|
cd spiffworkflow
|
||||||
|
wget https://raw.githubusercontent.com/sartography/spiff-arena/main/docker-compose.yml
|
||||||
|
docker compose pull
|
||||||
|
docker compose up -d
|
|
@ -7,14 +7,12 @@ function error_handler() {
|
||||||
trap 'error_handler ${LINENO} $?' ERR
|
trap 'error_handler ${LINENO} $?' ERR
|
||||||
set -o errtrace -o errexit -o nounset -o pipefail
|
set -o errtrace -o errexit -o nounset -o pipefail
|
||||||
|
|
||||||
max_attempts="${1:-}"
|
max_attempts="${1:-100}"
|
||||||
if [[ -z "$max_attempts" ]]; then
|
port="${2:-7000}"
|
||||||
max_attempts=100
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "waiting for backend to come up..."
|
echo "waiting for backend to come up..."
|
||||||
attempts=0
|
attempts=0
|
||||||
while [[ "$(curl -s -o /dev/null -w '%{http_code}' http://localhost:7000/v1.0/status)" != "200" ]]; do
|
while [[ "$(curl -s -o /dev/null -w '%{http_code}' "http://localhost:${port}/v1.0/status")" != "200" ]]; do
|
||||||
if [[ "$attempts" -gt "$max_attempts" ]]; then
|
if [[ "$attempts" -gt "$max_attempts" ]]; then
|
||||||
>&2 echo "ERROR: Server not up after $max_attempts attempts. There is probably a problem"
|
>&2 echo "ERROR: Server not up after $max_attempts attempts. There is probably a problem"
|
||||||
exit 1
|
exit 1
|
||||||
|
|
|
@ -7,14 +7,12 @@ function error_handler() {
|
||||||
trap 'error_handler ${LINENO} $?' ERR
|
trap 'error_handler ${LINENO} $?' ERR
|
||||||
set -o errtrace -o errexit -o nounset -o pipefail
|
set -o errtrace -o errexit -o nounset -o pipefail
|
||||||
|
|
||||||
max_attempts="${1:-}"
|
max_attempts="${1:-100}"
|
||||||
if [[ -z "$max_attempts" ]]; then
|
port="${2:-7002}"
|
||||||
max_attempts=100
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "waiting for backend to come up..."
|
echo "waiting for backend to come up..."
|
||||||
attempts=0
|
attempts=0
|
||||||
while [[ "$(curl -s -o /dev/null -w '%{http_code}' http://localhost:7002/realms/master/.well-known/openid-configuration)" != "200" ]]; do
|
while [[ "$(curl -s -o /dev/null -w '%{http_code}' "http://localhost:${port}/realms/master/.well-known/openid-configuration")" != "200" ]]; do
|
||||||
if [[ "$attempts" -gt "$max_attempts" ]]; then
|
if [[ "$attempts" -gt "$max_attempts" ]]; then
|
||||||
>&2 echo "ERROR: Server not up after $max_attempts attempts. There is probably a problem"
|
>&2 echo "ERROR: Server not up after $max_attempts attempts. There is probably a problem"
|
||||||
exit 1
|
exit 1
|
||||||
|
|
|
@ -7,14 +7,12 @@ function error_handler() {
|
||||||
trap 'error_handler ${LINENO} $?' ERR
|
trap 'error_handler ${LINENO} $?' ERR
|
||||||
set -o errtrace -o errexit -o nounset -o pipefail
|
set -o errtrace -o errexit -o nounset -o pipefail
|
||||||
|
|
||||||
max_attempts="${1:-}"
|
max_attempts="${1:-100}"
|
||||||
if [[ -z "$max_attempts" ]]; then
|
port="${2:-7001}"
|
||||||
max_attempts=100
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "waiting for backend to come up..."
|
echo "waiting for frontend to come up..."
|
||||||
attempts=0
|
attempts=0
|
||||||
while [[ "$(curl -s -o /dev/null -w '%{http_code}' http://localhost:7001)" != "200" ]]; do
|
while [[ "$(curl -s -o /dev/null -w '%{http_code}' "http://localhost:${port}")" != "200" ]]; do
|
||||||
if [[ "$attempts" -gt "$max_attempts" ]]; then
|
if [[ "$attempts" -gt "$max_attempts" ]]; then
|
||||||
>&2 echo "ERROR: Server not up after $max_attempts attempts. There is probably a problem"
|
>&2 echo "ERROR: Server not up after $max_attempts attempts. There is probably a problem"
|
||||||
exit 1
|
exit 1
|
||||||
|
|
Loading…
Reference in New Issue