mirror of
https://github.com/status-im/spiff-arena.git
synced 2025-02-22 14:38:11 +00:00
0bfe4400f Merge branch 'main' of github.com:sartography/spiff-arena into main fbe97dd90 Point to the latest spiffworkflow 805400312 fixed conflict with db migrations w/ burnettk 6e48c8205 Merge remote-tracking branch 'origin/main' into feature/script_get_last_user_completing_task 3a6cc9c2b added script to get process initiator w/ burnettk 37770dbca run_pyl 8a068cdb2 Merging main bf25c1cef run_pyl 4ca6e3f27 Needed an additional check for empty correlation keys - which on a RECEIVE message, should always match anything. 0fb88c50d remove unwanted test files w/ burnettk 89d92687d script to get last user completing a task is working w/ burnettk 98f056e5c Merge remote-tracking branch 'origin/main' into feature/script_get_last_user_completing_task 46af55a60 poetry remove orjson 28a0db097 wip for get_last_user_completing_task script task 2596cfeb1 postgres really will just order however it wants if you do not specify an order_by clause 5e339b2fb add ppg.ba4.sme and ba5 873cfdfc1 fix postgres db name and comment out debug job f54d6ae4d removed some unused code from task and fixed the logs table a bit w/ burnettk 7c74e216a run_pyl f53c85924 skip failing test if postgres and added comment about cause w/ burnettk fbe2237f1 # SpiffWorkflow: 1) Type Safe checking on correlation properties (no more str()) 2) A running workflows Correlations are once again at the key level. 2677736c2 look users up by service and username instead of service_id since usernames have to be unique anyway w/ burnettk 22d1f8bbb avoid using task-data endpoint for task data and only use it to get tasks based on spiff step instead 698fb5d5c put back the task data code when getting tasks b52f97453 Merge remote-tracking branch 'origin/main' into feature/task_data_api_refactor 9c4a17c93 removed commented out code w/ burnettk 3357fbef4 removed task-data endpoints since we no longer need them w/ burnettk 3b64afb65 turn off profiling again fcbcfd0ae add two users and update one 2ee7cba09 BPMN Parser was returning all retrieval expressions, rather than the ones specific to a correlation property, as was intended. Adding a correlation cache - so we have a reference of all the messages and properties (though still lacking a description of keys) Adding yet another migration, maybe should squash em. 6658e26bb added api to get task data and do not return from task data list anymore w/ burnettk 1fd3261d7 run_pyl (part 2) f6c63eb1c Merge branch 'main' of github.com:sartography/spiff-arena 5c5262a31 added comment about refactoring getting task data w/ burnettk jbirddog 6d4aa9043 lint e4c0ed7e1 add test users adfb0644f Adding Migration. ec36290f2 remove task size check since it can take a long time to run and we do not do anything with it w/ burnettk jbirddog 9b0b95f95 Merge remote-tracking branch 'origin/main' into feature/message_fixes f3d124a70 run_pyl be6ac8743 BPMN.io -- Just show the message names not the ids - to assure we are only exposing the names. SpiffWorkflow - - start_messages function should return message names, not ids. - don't catch external thrown messages within the same workflow process - add an expected value to the Correlation Property Model so we can use this well defined class as an external communication tool (rather than building an arbitrary dictionary) - Added a "get_awaiting_correlations" to an event, so we can get a list of the correlation properties related to the workflows currently defined correlation values. - workflows.waiting_events() function now returns the above awaiting correlations as the value on returned message events Backend - Dropping MessageModel and MessageCorrelationProperties - at least for now. We don't need them to send / receive messages though we may eventually want to track the messages and correlations defined across the system - these things (which are ever changing) should not be directly connected to the Messages which may be in flux - and the cross relationships between the tables could cause unexpected and unceissary errors. Commented out the caching logic so we can turn this back on later. - Slight improvement to API Errors - MessageInstances are no longer in a many-to-many relationship with Correlations - Each message instance has a unique set of message correlations specific to the instance. - Message Instances have users, and can be linked through a "counterpart_id" so you can see what send is connected to what recieve. - Message Correlations are connected to recieving message instances. It is not to a process instance, and not to a message model. They now include the expected value and retrieval expression required to validate an incoming message. - A process instance is not connected to message correlations. - Message Instances are not always tied to a process instance (for example, a Send Message from an API) - API calls to create a message use the same logic as all other message catching code. - Make use of the new waiting_events() method to check for any new recieve messages in the workflow (much easier than churning through all of the tasks) - One giant mother of a migration. ce449971a do not call serialize if we can use the cached bpmn_json instead w/ burnettk 15d720f94 Merge branch 'main' of github.com:sartography/spiff-arena 8d8347068 turn on sentry detailed tracing for task-data w/ burnettk cdf5f4053 update spiff 5c1ea3c93 run_pyl 384c272af * SpiffWorkflow event_definitions wanted to return a message event's correlation properties mested within correlation keys. But messages are directly related to properties, not to keys - and it forced a number of conversions that made for tricky code. So Messages now contain a dictionary of correlation properties only. * SpiffWorkflow did not serialize correlations - so they were lost between save and retrieve. f45103406 Allow people to run commands like "flask db upgrade" without setting specific environment variables like FLASK_SESSION_SECRET_KEY everytime - they just need to add in their own /instance/config.py with their local configuration. b169c3a87 * Re-work message tests so I could wrap my simple head around what was happening - just needed an example that made sense to me. * Clear out complex get_message_instance_receive how that many-to-many works. * Create decent error messages when correlations fail * Move correlation checks into the MessageInstance class * The APIError could bomb out ugly if it hit a workflow exception with not Task Spec. a39590912 failing test. 6db600caa Merge branch 'main' into feature/message_fixes 4942a728b work in progress - * Link between message instance and correlations is now a link table and many-to-many relationships as recommended by SQLAlchemy * Use the correlation keys, not the process id when accepting api messages. git-subtree-dir: spiffworkflow-backend git-subtree-split: 0bfe4400f4191214b8972977438ceb35a9f5b3c3
70 lines
3.2 KiB
Bash
Executable File
70 lines
3.2 KiB
Bash
Executable File
#!/usr/bin/env bash
|
|
|
|
function error_handler() {
|
|
>&2 echo "Exited with BAD EXIT CODE '${2}' in ${0} script at line: ${1}."
|
|
exit "$2"
|
|
}
|
|
trap 'error_handler ${LINENO} $?' ERR
|
|
set -o errtrace -o errexit -o nounset -o pipefail
|
|
|
|
export FLASK_SESSION_SECRET_KEY="this_is_recreate_db_secret_key"
|
|
|
|
if [[ -z "${SPIFFWORKFLOW_BACKEND_BPMN_SPEC_ABSOLUTE_DIR:-}" ]]; then
|
|
script_dir="$( cd -- "$(dirname "$0")" >/dev/null 2>&1 ; pwd -P )"
|
|
|
|
SPIFFWORKFLOW_BACKEND_BPMN_SPEC_ABSOLUTE_DIR="${script_dir}/../../../sample-process-models"
|
|
if [[ ! -d "$SPIFFWORKFLOW_BACKEND_BPMN_SPEC_ABSOLUTE_DIR" ]]; then
|
|
SPIFFWORKFLOW_BACKEND_BPMN_SPEC_ABSOLUTE_DIR="${script_dir}/../../sample-process-models"
|
|
if [[ ! -d "$SPIFFWORKFLOW_BACKEND_BPMN_SPEC_ABSOLUTE_DIR" ]]; then
|
|
>&2 echo "ERROR: Could not find a location for the sample processes. Last tried: $SPIFFWORKFLOW_BACKEND_BPMN_SPEC_ABSOLUTE_DIR"
|
|
exit 1
|
|
fi
|
|
fi
|
|
export SPIFFWORKFLOW_BACKEND_BPMN_SPEC_ABSOLUTE_DIR
|
|
fi
|
|
|
|
tasks=""
|
|
if [[ "${1:-}" == "clean" ]]; then
|
|
subcommand="${2:-}"
|
|
if [[ "$subcommand" == "rmall" ]]; then
|
|
tasks="$tasks init migrate"
|
|
rm -rf migrations/
|
|
elif [[ -n "$subcommand" ]]; then
|
|
>&2 echo "ERROR: you passed a subcommand that was not rmall, and that is not supported: $subcommand"
|
|
exit 1
|
|
fi
|
|
|
|
rm -f ./src/instance/*.sqlite3
|
|
mysql -uroot -e "DROP DATABASE IF EXISTS spiffworkflow_backend_local_development"
|
|
mysql -uroot -e "DROP DATABASE IF EXISTS spiffworkflow_backend_unit_testing"
|
|
|
|
# TODO: check to see if the db already exists and we can connect to it. also actually clean it up.
|
|
# start postgres in background with one db
|
|
if [[ "${SPIFFWORKFLOW_BACKEND_DATABASE_TYPE:-}" == "postgres" ]]; then
|
|
if ! docker exec -it postgres-spiff psql -U spiffworkflow_backend spiffworkflow_backend_unit_testing -c "select 1"; then
|
|
docker run --name postgres-spiff -p 5432:5432 -e POSTGRES_PASSWORD=spiffworkflow_backend -e POSTGRES_USER=spiffworkflow_backend -e POSTGRES_DB=spiffworkflow_backend_unit_testing -d postgres
|
|
sleep 4 # classy
|
|
fi
|
|
if ! docker exec -it postgres-spiff psql -U spiffworkflow_backend spiffworkflow_backend_local_development -c "select 1"; then
|
|
# create other db. spiffworkflow_backend_unit_testing came with the docker run.
|
|
docker exec -it postgres-spiff psql -U spiffworkflow_backend spiffworkflow_backend_unit_testing -c "create database spiffworkflow_backend_local_development;"
|
|
fi
|
|
fi
|
|
elif [[ "${1:-}" == "migrate" ]]; then
|
|
tasks="$tasks migrate"
|
|
fi
|
|
tasks="$tasks upgrade"
|
|
|
|
mysql -uroot -e "CREATE DATABASE IF NOT EXISTS spiffworkflow_backend_local_development"
|
|
mysql -uroot -e "CREATE DATABASE IF NOT EXISTS spiffworkflow_backend_unit_testing"
|
|
|
|
for task in $tasks; do
|
|
SPIFFWORKFLOW_BACKEND_ENV=local_development FLASK_APP=src/spiffworkflow_backend poetry run flask db "$task"
|
|
done
|
|
|
|
SPIFFWORKFLOW_BACKEND_ENV=unit_testing FLASK_APP=src/spiffworkflow_backend poetry run flask db upgrade
|
|
if [[ -n "${SPIFFWORKFLOW_BACKEND_ENV:-}" ]] && ! grep -Eq '^(local_development|unit_testing)$' <<< "$SPIFFWORKFLOW_BACKEND_ENV"; then
|
|
mysql -uroot -e "CREATE DATABASE IF NOT EXISTS spiffworkflow_backend_$SPIFFWORKFLOW_BACKEND_ENV"
|
|
FLASK_APP=src/spiffworkflow_backend poetry run flask db upgrade
|
|
fi
|