Go to file
Aaron Louie fceda2c796 Tries defining commit in deploy script 2020-05-07 14:14:25 -04:00
config Setting JSON_SORT_KEYS to false, assuring that Flask does not resort all data returned to the front end. 2020-05-06 10:59:49 -04:00
crc No longer requires task_id 2020-05-06 14:27:28 -04:00
docs Cleanup. 2020-01-09 10:09:59 -05:00
migrations Removing duplicate migration. 2020-05-06 13:28:41 -04:00
postgres It became impossible to use the Swagger ui when we started adding authentication to all of the calls. I discovered Connexion and Swagger have a default way of handing JTW authentication and this cleans up our code quite a bit, moves the securing of endpoints into the API Definition, which is quite nice, and removes a whole library dependency (I never get to do that!) I added a SWAGGER_AUTH_KEY that can be used in non-production environments to allow users to quickly authenticate from the Swagger ui. I also removed all api calls to simple little happy api services, because that is just mean and pointless. 2020-03-24 14:15:21 -04:00
readme_images Updates to the README and removing a problematic dependency 2019-11-20 17:00:02 -05:00
schema Adds basic Flask, Connexion, and SpiffWorkflow structure 2019-11-19 15:57:46 -05:00
tests I had to give up and live with the idea that we can only render documentation on the current task, not on the previous or next tasks. I think this is ok. If you want to view a task, you need to make it the active task to assure all the parts and pieces are in place. 2020-05-06 13:01:38 -04:00
.gitignore Removes xlsx lock file. Adds LibreOffice lock files to gitignore 2020-04-27 22:58:12 -04:00
.sonarcloud.properties Generates coverage.xml report 2020-02-12 11:04:04 -05:00
.travis.yml Tries defining commit in deploy script 2020-05-07 14:14:25 -04:00
Dockerfile Trivial change to force travis docker push 2020-02-06 11:18:23 -05:00
LICENSE.md Adds basic Flask, Connexion, and SpiffWorkflow structure 2019-11-19 15:57:46 -05:00
Pipfile Ldap Service with Test and mocks. 2020-04-20 15:16:33 -04:00
Pipfile.lock Setting JSON_SORT_KEYS to false, assuring that Flask does not resort all data returned to the front end. 2020-05-06 10:59:49 -04:00
README.md Updates in travis to do a push to docker hub, so our code can be deployed to various environments. 2020-02-05 15:53:52 -05:00
deploy.sh Tries defining commit in deploy script 2020-05-07 14:14:25 -04:00
docker_run.sh Adds downgrade command 2020-04-21 19:37:56 -04:00
example_data.py Updates BPMN files 2020-05-05 16:15:38 -04:00
fact_runner.py Creating a dependency on the SpiffWorkflow fork we created. and placing the command line script into this code base and demonstrating that our workflow processor works at least on that level. Removing all kinds of garbage that had piled up before to keep this thing clean. 2019-12-11 11:45:44 -05:00
run.py Explicitly sets port. Prints some config settings for debugging. 2020-04-15 13:49:32 -04:00
sonar-project.properties Adds sonar properties 2020-02-08 14:00:37 -05:00
wait-for-it.sh Scripts should be executable. 2020-02-06 09:02:02 -05:00

README.md

CrConnectFrontend

Build Status

CR Connect Workflow Microservice

Development Setup

Tools

These instructions assume you're using these development and tools:

  • IDE: PyCharm Professional Edition
  • Operating System: Ubuntu

Environment Setup

Make sure all of the following are properly installed on your system:

  1. python3 & pip3:
  2. pipenv:

Project Initialization

  1. Clone this repository.

  2. In PyCharm:

    • Go to File > New Project...
    • Click Pure Python (NOT Flask!!)
    • Click the folder icon in the Location field.
    • Select the directory where you cloned this repository and click Ok.
    • Expand the Project Interpreter section.
    • Select the New environment using radio button and choose Pipenv in the dropdown.
    • Under Base interpreter, select Python 3.6
    • In the Pipenv executable field, enter /home/your_username_goes_here/.local/bin/pipenv
    • Click Create Project Interpreter
  3. PyCharm should automatically install the necessary packages via pipenv. For me, the project interpreter did not set set up for me correctly on first attempt. I had to go to File -> Settings -> Project Interpreter and again set the project to use the correct PipEnv environment. Be sure that your settings like simliar to this, or attempt to add the interpreter again by clicking on the gear icon. Project Interpreter Settings screen

  4. With this properly setup for the project, you can now right click on the run.py and set up a new run configuration and set up a run configuration that looks like the following (be sure to save this run configuration so it doesn't go away.) : Run Configuration Screenshot

Running the Web API

Just click the "Play" button next to RUN in the top right corner of the screen. The Swagger based view of the API will be avialable at http://0.0.0.0:5000/v1.0/ui/

Testing from the Shell

This app includes a command line interface that will read in BPMN files and let you play with it at the command line. To run it right click on app/command_line/joke.py and click run. Type "?" to get a list of commands. So far the joke system will work a little, when you file it up try these commands in this order:

> engine  (this will run all tasks up to first user task and should print a joke)
> answer clock (this is the correct answer)
> next (this completes the user task)
> engine (this runs the rest of the tasks, and should tell you that  you got the question right)

You can try re-running this and getting the question wrong. You might open up the Joke bpmn diagram so you can see what this looks like to draw out.

Documentation

Additional Documentation is available on ReadTheDocs

Additional Reading

  1. BPMN Is the tool we are using to create diagrams of our business processes. It's is a beautiful concise diagramming tool. We strongly recommend you read this complete tutorial, as this notation is the foundation on which this project as well as many other software systems for businesses are built. Know it well.

Notes on Creating Good BPMN Diagrams in Comunda

  1. Be sure to give each task a thoughtful (but unique!) id. This will make the command line and debugging far far easier. I've tended to pre-fix these, so task_ask_riddle if a task is asking a riddle for instance.