diff --git a/bin/import_tickets_for_command_line.py b/bin/import_tickets_for_command_line.py
index e193b599..cc94ba54 100644
--- a/bin/import_tickets_for_command_line.py
+++ b/bin/import_tickets_for_command_line.py
@@ -27,7 +27,6 @@ def main():
"""Main."""
app = get_hacked_up_app_for_script()
with app.app_context():
-
process_model_identifier_ticket = "ticket"
db.session.query(ProcessInstanceModel).filter(
ProcessInstanceModel.process_model_identifier
diff --git a/bin/keycloak_test_server.py b/bin/keycloak_test_server.py
index 59efd36c..3e933493 100644
--- a/bin/keycloak_test_server.py
+++ b/bin/keycloak_test_server.py
@@ -40,7 +40,8 @@ def hello_world():
return (
'Hello, %s, See private '
'Log out'
- ) % oidc.user_getfield("preferred_username")
+ % oidc.user_getfield("preferred_username")
+ )
else:
return 'Welcome anonymous, Log in'
diff --git a/src/spiffworkflow_backend/__init__.py b/src/spiffworkflow_backend/__init__.py
index c40b3e76..f1de793d 100644
--- a/src/spiffworkflow_backend/__init__.py
+++ b/src/spiffworkflow_backend/__init__.py
@@ -93,7 +93,8 @@ def create_app() -> flask.app.Flask:
if os.environ.get("FLASK_SESSION_SECRET_KEY") is None:
raise KeyError(
- "Cannot find the secret_key from the environment. Please set FLASK_SESSION_SECRET_KEY"
+ "Cannot find the secret_key from the environment. Please set"
+ " FLASK_SESSION_SECRET_KEY"
)
app.secret_key = os.environ.get("FLASK_SESSION_SECRET_KEY")
diff --git a/src/spiffworkflow_backend/config/__init__.py b/src/spiffworkflow_backend/config/__init__.py
index f7661bc1..fb5901f0 100644
--- a/src/spiffworkflow_backend/config/__init__.py
+++ b/src/spiffworkflow_backend/config/__init__.py
@@ -17,21 +17,21 @@ def setup_database_uri(app: Flask) -> None:
if app.config.get("SPIFFWORKFLOW_BACKEND_DATABASE_URI") is None:
database_name = f"spiffworkflow_backend_{app.config['ENV_IDENTIFIER']}"
if app.config.get("SPIFF_DATABASE_TYPE") == "sqlite":
- app.config[
- "SQLALCHEMY_DATABASE_URI"
- ] = f"sqlite:///{app.instance_path}/db_{app.config['ENV_IDENTIFIER']}.sqlite3"
+ app.config["SQLALCHEMY_DATABASE_URI"] = (
+ f"sqlite:///{app.instance_path}/db_{app.config['ENV_IDENTIFIER']}.sqlite3"
+ )
elif app.config.get("SPIFF_DATABASE_TYPE") == "postgres":
- app.config[
- "SQLALCHEMY_DATABASE_URI"
- ] = f"postgresql://spiffworkflow_backend:spiffworkflow_backend@localhost:5432/{database_name}"
+ app.config["SQLALCHEMY_DATABASE_URI"] = (
+ f"postgresql://spiffworkflow_backend:spiffworkflow_backend@localhost:5432/{database_name}"
+ )
else:
# use pswd to trick flake8 with hardcoded passwords
db_pswd = os.environ.get("DB_PASSWORD")
if db_pswd is None:
db_pswd = ""
- app.config[
- "SQLALCHEMY_DATABASE_URI"
- ] = f"mysql+mysqlconnector://root:{db_pswd}@localhost/{database_name}"
+ app.config["SQLALCHEMY_DATABASE_URI"] = (
+ f"mysql+mysqlconnector://root:{db_pswd}@localhost/{database_name}"
+ )
else:
app.config["SQLALCHEMY_DATABASE_URI"] = app.config.get(
"SPIFFWORKFLOW_BACKEND_DATABASE_URI"
@@ -91,10 +91,12 @@ def setup_config(app: Flask) -> None:
app.config["SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME"],
)
print(
- f"set permissions file name config: {app.config['SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME']}"
+ "set permissions file name config:"
+ f" {app.config['SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME']}"
)
print(
- f"set permissions file name full path: {app.config['PERMISSIONS_FILE_FULLPATH']}"
+ "set permissions file name full path:"
+ f" {app.config['PERMISSIONS_FILE_FULLPATH']}"
)
# unversioned (see .gitignore) config that can override everything and include secrets.
diff --git a/src/spiffworkflow_backend/models/message_instance.py b/src/spiffworkflow_backend/models/message_instance.py
index 2559a635..b0cc2aa3 100644
--- a/src/spiffworkflow_backend/models/message_instance.py
+++ b/src/spiffworkflow_backend/models/message_instance.py
@@ -86,5 +86,6 @@ def ensure_failure_cause_is_set_if_message_instance_failed(
if isinstance(instance, MessageInstanceModel):
if instance.status == "failed" and instance.failure_cause is None:
raise ValueError(
- f"{instance.__class__.__name__}: failure_cause must be set if status is failed"
+ f"{instance.__class__.__name__}: failure_cause must be set if"
+ " status is failed"
)
diff --git a/src/spiffworkflow_backend/models/process_instance.py b/src/spiffworkflow_backend/models/process_instance.py
index d1d63510..aa1440b4 100644
--- a/src/spiffworkflow_backend/models/process_instance.py
+++ b/src/spiffworkflow_backend/models/process_instance.py
@@ -62,7 +62,10 @@ class ProcessInstanceModel(SpiffworkflowBaseDBModel):
active_human_tasks = relationship(
"HumanTaskModel",
- primaryjoin="and_(HumanTaskModel.process_instance_id==ProcessInstanceModel.id, HumanTaskModel.completed == False)",
+ primaryjoin=(
+ "and_(HumanTaskModel.process_instance_id==ProcessInstanceModel.id,"
+ " HumanTaskModel.completed == False)"
+ ),
) # type: ignore
human_tasks = relationship(
diff --git a/src/spiffworkflow_backend/models/task.py b/src/spiffworkflow_backend/models/task.py
index 896c79e1..79814c1d 100644
--- a/src/spiffworkflow_backend/models/task.py
+++ b/src/spiffworkflow_backend/models/task.py
@@ -43,8 +43,8 @@ class Task:
FIELD_TYPE_EMAIL = "email" # email: Email address
FIELD_TYPE_URL = "url" # url: Website address
- FIELD_PROP_AUTO_COMPLETE_MAX = (
- "autocomplete_num" # Not used directly, passed in from the front end.
+ FIELD_PROP_AUTO_COMPLETE_MAX = ( # Not used directly, passed in from the front end.
+ "autocomplete_num"
)
# Required field
@@ -77,8 +77,8 @@ class Task:
# File specific field properties
FIELD_PROP_DOC_CODE = "doc_code" # to associate a file upload field with a doc code
- FIELD_PROP_FILE_DATA = (
- "file_data" # to associate a bit of data with a specific file upload file.
+ FIELD_PROP_FILE_DATA = ( # to associate a bit of data with a specific file upload file.
+ "file_data"
)
# Additional properties
diff --git a/src/spiffworkflow_backend/routes/messages_controller.py b/src/spiffworkflow_backend/routes/messages_controller.py
index c5238bad..51290770 100644
--- a/src/spiffworkflow_backend/routes/messages_controller.py
+++ b/src/spiffworkflow_backend/routes/messages_controller.py
@@ -131,8 +131,11 @@ def message_start(
raise (
ApiError(
error_code="cannot_find_waiting_message",
- message=f"Could not find waiting message for identifier {message_identifier} "
- f"and process instance {process_instance.id}",
+ message=(
+ "Could not find waiting message for identifier"
+ f" {message_identifier} and process instance"
+ f" {process_instance.id}"
+ ),
status_code=400,
)
)
@@ -151,7 +154,10 @@ def message_start(
raise (
ApiError(
error_code="cannot_start_message",
- message=f"Message with identifier cannot be start with message: {message_identifier}",
+ message=(
+ "Message with identifier cannot be start with message:"
+ f" {message_identifier}"
+ ),
status_code=400,
)
)
diff --git a/src/spiffworkflow_backend/routes/process_api_blueprint.py b/src/spiffworkflow_backend/routes/process_api_blueprint.py
index cff11eb9..6dcd79f7 100644
--- a/src/spiffworkflow_backend/routes/process_api_blueprint.py
+++ b/src/spiffworkflow_backend/routes/process_api_blueprint.py
@@ -43,7 +43,9 @@ def permissions_check(body: Dict[str, Dict[str, list[str]]]) -> flask.wrappers.R
raise (
ApiError(
error_code="could_not_requests_to_check",
- message="The key 'requests_to_check' not found at root of request body.",
+ message=(
+ "The key 'requests_to_check' not found at root of request body."
+ ),
status_code=400,
)
)
@@ -139,7 +141,8 @@ def task_data_update(
if process_instance:
if process_instance.status != "suspended":
raise ProcessInstanceTaskDataCannotBeUpdatedError(
- f"The process instance needs to be suspended to udpate the task-data. It is currently: {process_instance.status}"
+ "The process instance needs to be suspended to udpate the task-data."
+ f" It is currently: {process_instance.status}"
)
process_instance_bpmn_json_dict = json.loads(process_instance.bpmn_json)
@@ -163,12 +166,18 @@ def task_data_update(
else:
raise ApiError(
error_code="update_task_data_error",
- message=f"Could not find Task: {task_id} in Instance: {process_instance_id}.",
+ message=(
+ f"Could not find Task: {task_id} in Instance:"
+ f" {process_instance_id}."
+ ),
)
else:
raise ApiError(
error_code="update_task_data_error",
- message=f"Could not update task data for Instance: {process_instance_id}, and Task: {task_id}.",
+ message=(
+ f"Could not update task data for Instance: {process_instance_id}, and"
+ f" Task: {task_id}."
+ ),
)
return Response(
json.dumps(ProcessInstanceModelSchema().dump(process_instance)),
@@ -236,7 +245,9 @@ def manual_complete_task(
else:
raise ApiError(
error_code="complete_task",
- message=f"Could not complete Task {task_id} in Instance {process_instance_id}",
+ message=(
+ f"Could not complete Task {task_id} in Instance {process_instance_id}"
+ ),
)
return Response(
json.dumps(ProcessInstanceModelSchema().dump(process_instance)),
diff --git a/src/spiffworkflow_backend/routes/process_groups_controller.py b/src/spiffworkflow_backend/routes/process_groups_controller.py
index b0d43a4c..228be181 100644
--- a/src/spiffworkflow_backend/routes/process_groups_controller.py
+++ b/src/spiffworkflow_backend/routes/process_groups_controller.py
@@ -124,6 +124,7 @@ def process_group_move(
original_process_group_id, new_location
)
_commit_and_push_to_git(
- f"User: {g.user.username} moved process group {original_process_group_id} to {new_process_group.id}"
+ f"User: {g.user.username} moved process group {original_process_group_id} to"
+ f" {new_process_group.id}"
)
return make_response(jsonify(new_process_group), 200)
diff --git a/src/spiffworkflow_backend/routes/process_instances_controller.py b/src/spiffworkflow_backend/routes/process_instances_controller.py
index 08ec712e..50f1fc64 100644
--- a/src/spiffworkflow_backend/routes/process_instances_controller.py
+++ b/src/spiffworkflow_backend/routes/process_instances_controller.py
@@ -94,7 +94,10 @@ def process_instance_run(
if process_instance.status != "not_started":
raise ApiError(
error_code="process_instance_not_runnable",
- message=f"Process Instance ({process_instance.id}) is currently running or has already run.",
+ message=(
+ f"Process Instance ({process_instance.id}) is currently running or has"
+ " already run."
+ ),
status_code=400,
)
@@ -350,8 +353,8 @@ def process_instance_delete(
if not process_instance.has_terminal_status():
raise ProcessInstanceCannotBeDeletedError(
- f"Process instance ({process_instance.id}) cannot be deleted since it does not have a terminal status. "
- f"Current status is {process_instance.status}."
+ f"Process instance ({process_instance.id}) cannot be deleted since it does"
+ f" not have a terminal status. Current status is {process_instance.status}."
)
# (Pdb) db.session.delete
@@ -620,7 +623,8 @@ def _get_process_instance(
).first()
if spec_reference is None:
raise SpecReferenceNotFoundError(
- f"Could not find given process identifier in the cache: {process_identifier}"
+ "Could not find given process identifier in the cache:"
+ f" {process_identifier}"
)
process_model_with_diagram = ProcessModelService.get_process_model(
@@ -678,7 +682,10 @@ def _find_process_instance_for_me_or_raise(
raise (
ApiError(
error_code="process_instance_cannot_be_found",
- message=f"Process instance with id {process_instance_id} cannot be found that is associated with you.",
+ message=(
+ f"Process instance with id {process_instance_id} cannot be found"
+ " that is associated with you."
+ ),
status_code=400,
)
)
diff --git a/src/spiffworkflow_backend/routes/script_unit_tests_controller.py b/src/spiffworkflow_backend/routes/script_unit_tests_controller.py
index ddd48581..e97b26ae 100644
--- a/src/spiffworkflow_backend/routes/script_unit_tests_controller.py
+++ b/src/spiffworkflow_backend/routes/script_unit_tests_controller.py
@@ -40,7 +40,10 @@ def script_unit_test_create(
if file is None:
raise ApiError(
error_code="cannot_find_file",
- message=f"Could not find the primary bpmn file for process_model: {process_model.id}",
+ message=(
+ "Could not find the primary bpmn file for process_model:"
+ f" {process_model.id}"
+ ),
status_code=404,
)
diff --git a/src/spiffworkflow_backend/routes/tasks_controller.py b/src/spiffworkflow_backend/routes/tasks_controller.py
index a8f899b5..2c7ceeea 100644
--- a/src/spiffworkflow_backend/routes/tasks_controller.py
+++ b/src/spiffworkflow_backend/routes/tasks_controller.py
@@ -157,7 +157,10 @@ def task_show(process_instance_id: int, task_id: str) -> flask.wrappers.Response
raise (
ApiError(
error_code="no_human_task",
- message=f"Cannot find a task to complete for task id '{task_id}' and process instance {process_instance_id}.",
+ message=(
+ f"Cannot find a task to complete for task id '{task_id}' and"
+ f" process instance {process_instance_id}."
+ ),
status_code=500,
)
)
@@ -203,7 +206,10 @@ def task_show(process_instance_id: int, task_id: str) -> flask.wrappers.Response
raise (
ApiError(
error_code="missing_form_file",
- message=f"Cannot find a form file for process_instance_id: {process_instance_id}, task_id: {task_id}",
+ message=(
+ "Cannot find a form file for process_instance_id:"
+ f" {process_instance_id}, task_id: {task_id}"
+ ),
status_code=400,
)
)
@@ -221,7 +227,10 @@ def task_show(process_instance_id: int, task_id: str) -> flask.wrappers.Response
raise (
ApiError(
error_code="error_loading_form",
- message=f"Could not load form schema from: {form_schema_file_name}. Error was: {str(exception)}",
+ message=(
+ f"Could not load form schema from: {form_schema_file_name}."
+ f" Error was: {str(exception)}"
+ ),
status_code=400,
)
) from exception
@@ -285,8 +294,10 @@ def task_submit(
if not process_instance.can_submit_task():
raise ApiError(
error_code="process_instance_not_runnable",
- message=f"Process Instance ({process_instance.id}) has status "
- f"{process_instance.status} which does not allow tasks to be submitted.",
+ message=(
+ f"Process Instance ({process_instance.id}) has status "
+ f"{process_instance.status} which does not allow tasks to be submitted."
+ ),
status_code=400,
)
@@ -317,7 +328,10 @@ def task_submit(
raise (
ApiError(
error_code="no_human_task",
- message=f"Cannot find a task to complete for task id '{task_id}' and process instance {process_instance_id}.",
+ message=(
+ f"Cannot find a task to complete for task id '{task_id}' and"
+ f" process instance {process_instance_id}."
+ ),
status_code=500,
)
)
@@ -511,7 +525,10 @@ def _update_form_schema_with_task_data_as_needed(
raise (
ApiError(
error_code="missing_task_data_var",
- message=f"Task data is missing variable: {task_data_var}",
+ message=(
+ "Task data is missing variable:"
+ f" {task_data_var}"
+ ),
status_code=500,
)
)
diff --git a/src/spiffworkflow_backend/routes/user.py b/src/spiffworkflow_backend/routes/user.py
index 51301c70..1ac6207c 100644
--- a/src/spiffworkflow_backend/routes/user.py
+++ b/src/spiffworkflow_backend/routes/user.py
@@ -67,13 +67,16 @@ def verify_token(
user_model = get_user_from_decoded_internal_token(decoded_token)
except Exception as e:
current_app.logger.error(
- f"Exception in verify_token getting user from decoded internal token. {e}"
+ "Exception in verify_token getting user from decoded"
+ f" internal token. {e}"
)
elif "iss" in decoded_token.keys():
try:
if AuthenticationService.validate_id_token(token):
user_info = decoded_token
- except ApiError as ae: # API Error is only thrown in the token is outdated.
+ except (
+ ApiError
+ ) as ae: # API Error is only thrown in the token is outdated.
# Try to refresh the token
user = UserService.get_user_by_service_and_service_id(
decoded_token["iss"], decoded_token["sub"]
diff --git a/src/spiffworkflow_backend/routes/user_blueprint.py b/src/spiffworkflow_backend/routes/user_blueprint.py
index 29bbddcd..fd5c1ae9 100644
--- a/src/spiffworkflow_backend/routes/user_blueprint.py
+++ b/src/spiffworkflow_backend/routes/user_blueprint.py
@@ -26,6 +26,7 @@ user_blueprint = Blueprint("main", __name__)
# user = UserService.create_user('internal', username)
# return Response(json.dumps({"id": user.id}), status=201, mimetype=APPLICATION_JSON)
+
# def _create_user(username):
# user = UserModel.query.filter_by(username=username).first()
# if user is not None:
diff --git a/src/spiffworkflow_backend/scripts/fact_service.py b/src/spiffworkflow_backend/scripts/fact_service.py
index 6e8a23c2..c739d15a 100644
--- a/src/spiffworkflow_backend/scripts/fact_service.py
+++ b/src/spiffworkflow_backend/scripts/fact_service.py
@@ -35,7 +35,10 @@ class FactService(Script):
if fact == "cat":
details = "The cat in the hat" # self.get_cat()
elif fact == "norris":
- details = "Chuck Norris doesn’t read books. He stares them down until he gets the information he wants."
+ details = (
+ "Chuck Norris doesn’t read books. He stares them down until he gets the"
+ " information he wants."
+ )
elif fact == "buzzword":
details = "Move the Needle." # self.get_buzzword()
else:
diff --git a/src/spiffworkflow_backend/scripts/get_group_members.py b/src/spiffworkflow_backend/scripts/get_group_members.py
index 8b179a6d..0f20fbb3 100644
--- a/src/spiffworkflow_backend/scripts/get_group_members.py
+++ b/src/spiffworkflow_backend/scripts/get_group_members.py
@@ -32,7 +32,8 @@ class GetGroupMembers(Script):
group = GroupModel.query.filter_by(identifier=group_identifier).first()
if group is None:
raise GroupNotFoundError(
- f"Script 'get_group_members' could not find group with identifier '{group_identifier}'."
+ "Script 'get_group_members' could not find group with identifier"
+ f" '{group_identifier}'."
)
usernames = [u.username for u in group.users]
diff --git a/src/spiffworkflow_backend/scripts/get_process_info.py b/src/spiffworkflow_backend/scripts/get_process_info.py
index 138a19ac..99eb4ce2 100644
--- a/src/spiffworkflow_backend/scripts/get_process_info.py
+++ b/src/spiffworkflow_backend/scripts/get_process_info.py
@@ -28,5 +28,7 @@ class GetProcessInfo(Script):
"""Run."""
return {
"process_instance_id": script_attributes_context.process_instance_id,
- "process_model_identifier": script_attributes_context.process_model_identifier,
+ "process_model_identifier": (
+ script_attributes_context.process_model_identifier
+ ),
}
diff --git a/src/spiffworkflow_backend/scripts/script.py b/src/spiffworkflow_backend/scripts/script.py
index 9e5836d6..7ca79846 100644
--- a/src/spiffworkflow_backend/scripts/script.py
+++ b/src/spiffworkflow_backend/scripts/script.py
@@ -98,8 +98,9 @@ class Script:
).first()
if process_instance is None:
raise ProcessInstanceNotFoundError(
- f"Could not find a process instance with id '{script_attributes_context.process_instance_id}' "
- f"when running script '{script_function_name}'"
+ "Could not find a process instance with id"
+ f" '{script_attributes_context.process_instance_id}' when"
+ f" running script '{script_function_name}'"
)
user = process_instance.process_initiator
has_permission = AuthorizationService.user_has_permission(
@@ -107,7 +108,8 @@ class Script:
)
if not has_permission:
raise ScriptUnauthorizedForUserError(
- f"User {user.username} does not have access to run privileged script '{script_function_name}'"
+ f"User {user.username} does not have access to run"
+ f" privileged script '{script_function_name}'"
)
def run_script_if_allowed(*ar: Any, **kw: Any) -> Any:
@@ -149,7 +151,7 @@ class Script:
"""_get_all_subclasses."""
# hackish mess to make sure we have all the modules loaded for the scripts
pkg_dir = os.path.dirname(__file__)
- for (_module_loader, name, _ispkg) in pkgutil.iter_modules([pkg_dir]):
+ for _module_loader, name, _ispkg in pkgutil.iter_modules([pkg_dir]):
importlib.import_module("." + name, __package__)
"""Returns a list of all classes that extend this class."""
diff --git a/src/spiffworkflow_backend/services/acceptance_test_fixtures.py b/src/spiffworkflow_backend/services/acceptance_test_fixtures.py
index 81488910..6bbcad33 100644
--- a/src/spiffworkflow_backend/services/acceptance_test_fixtures.py
+++ b/src/spiffworkflow_backend/services/acceptance_test_fixtures.py
@@ -29,7 +29,6 @@ def load_acceptance_test_fixtures() -> list[ProcessInstanceModel]:
# suspended - 6 hours ago
process_instances = []
for i in range(len(statuses)):
-
process_instance = ProcessInstanceService.create_process_instance_from_process_model_identifier(
test_process_model_id, user
)
diff --git a/src/spiffworkflow_backend/services/authorization_service.py b/src/spiffworkflow_backend/services/authorization_service.py
index 5ee1bf62..69d19cb7 100644
--- a/src/spiffworkflow_backend/services/authorization_service.py
+++ b/src/spiffworkflow_backend/services/authorization_service.py
@@ -128,7 +128,8 @@ class AuthorizationService:
# to check for exact matches as well
# see test_user_can_access_base_path_when_given_wildcard_permission unit test
text(
- f"'{target_uri_normalized}' = replace(replace(permission_target.uri, '/%', ''), ':%', '')"
+ f"'{target_uri_normalized}' ="
+ " replace(replace(permission_target.uri, '/%', ''), ':%', '')"
),
)
)
@@ -200,7 +201,8 @@ class AuthorizationService:
if current_app.config["SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME"] is None:
raise (
PermissionsFileNotSetError(
- "SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME needs to be set in order to import permissions"
+ "SPIFFWORKFLOW_BACKEND_PERMISSIONS_FILE_NAME needs to be set in"
+ " order to import permissions"
)
)
@@ -280,9 +282,9 @@ class AuthorizationService:
"""Find_or_create_permission_target."""
uri_with_percent = re.sub(r"\*", "%", uri)
target_uri_normalized = uri_with_percent.removeprefix(V1_API_PATH_PREFIX)
- permission_target: Optional[
- PermissionTargetModel
- ] = PermissionTargetModel.query.filter_by(uri=target_uri_normalized).first()
+ permission_target: Optional[PermissionTargetModel] = (
+ PermissionTargetModel.query.filter_by(uri=target_uri_normalized).first()
+ )
if permission_target is None:
permission_target = PermissionTargetModel(uri=target_uri_normalized)
db.session.add(permission_target)
@@ -297,13 +299,13 @@ class AuthorizationService:
permission: str,
) -> PermissionAssignmentModel:
"""Create_permission_for_principal."""
- permission_assignment: Optional[
- PermissionAssignmentModel
- ] = PermissionAssignmentModel.query.filter_by(
- principal_id=principal.id,
- permission_target_id=permission_target.id,
- permission=permission,
- ).first()
+ permission_assignment: Optional[PermissionAssignmentModel] = (
+ PermissionAssignmentModel.query.filter_by(
+ principal_id=principal.id,
+ permission_target_id=permission_target.id,
+ permission=permission,
+ ).first()
+ )
if permission_assignment is None:
permission_assignment = PermissionAssignmentModel(
principal_id=principal.id,
@@ -403,7 +405,10 @@ class AuthorizationService:
raise ApiError(
error_code="unauthorized",
- message=f"User {g.user.username} is not authorized to perform requested action: {permission_string} - {request.path}",
+ message=(
+ f"User {g.user.username} is not authorized to perform requested action:"
+ f" {permission_string} - {request.path}"
+ ),
status_code=403,
)
@@ -482,7 +487,10 @@ class AuthorizationService:
except jwt.InvalidTokenError as exception:
raise ApiError(
"token_invalid",
- "The Authentication token you provided is invalid. You need a new token. ",
+ (
+ "The Authentication token you provided is invalid. You need a new"
+ " token. "
+ ),
) from exception
@staticmethod
@@ -504,8 +512,9 @@ class AuthorizationService:
if user not in human_task.potential_owners:
raise UserDoesNotHaveAccessToTaskError(
- f"User {user.username} does not have access to update task'{spiff_task.task_spec.name}'"
- f" for process instance '{process_instance_id}'"
+ f"User {user.username} does not have access to update"
+ f" task'{spiff_task.task_spec.name}' for process instance"
+ f" '{process_instance_id}'"
)
return True
@@ -723,8 +732,9 @@ class AuthorizationService:
)
else:
raise InvalidPermissionError(
- f"Target uri '{target}' with permission set '{permission_set}' is invalid. "
- f"The target uri must either be a macro of PG, PM, BASIC, or ALL or an api uri."
+ f"Target uri '{target}' with permission set '{permission_set}' is"
+ " invalid. The target uri must either be a macro of PG, PM, BASIC, or"
+ " ALL or an api uri."
)
return permissions_to_assign
diff --git a/src/spiffworkflow_backend/services/git_service.py b/src/spiffworkflow_backend/services/git_service.py
index 495603cf..43c18edc 100644
--- a/src/spiffworkflow_backend/services/git_service.py
+++ b/src/spiffworkflow_backend/services/git_service.py
@@ -173,13 +173,15 @@ class GitService:
if "repository" not in webhook or "clone_url" not in webhook["repository"]:
raise InvalidGitWebhookBodyError(
- f"Cannot find required keys of 'repository:clone_url' from webhook body: {webhook}"
+ "Cannot find required keys of 'repository:clone_url' from webhook"
+ f" body: {webhook}"
)
clone_url = webhook["repository"]["clone_url"]
if clone_url != current_app.config["GIT_CLONE_URL_FOR_PUBLISHING"]:
raise GitCloneUrlMismatchError(
- f"Configured clone url does not match clone url from webhook: {clone_url}"
+ "Configured clone url does not match clone url from webhook:"
+ f" {clone_url}"
)
if "ref" not in webhook:
@@ -189,8 +191,8 @@ class GitService:
if current_app.config["GIT_BRANCH"] is None:
raise MissingGitConfigsError(
- "Missing config for GIT_BRANCH. "
- "This is required for updating the repository as a result of the webhook"
+ "Missing config for GIT_BRANCH. This is required for updating the"
+ " repository as a result of the webhook"
)
ref = webhook["ref"]
diff --git a/src/spiffworkflow_backend/services/logging_service.py b/src/spiffworkflow_backend/services/logging_service.py
index dd34cb3f..599d5228 100644
--- a/src/spiffworkflow_backend/services/logging_service.py
+++ b/src/spiffworkflow_backend/services/logging_service.py
@@ -122,7 +122,8 @@ def setup_logger(app: Flask) -> None:
if upper_log_level_string not in log_levels:
raise InvalidLogLevelError(
- f"Log level given is invalid: '{upper_log_level_string}'. Valid options are {log_levels}"
+ f"Log level given is invalid: '{upper_log_level_string}'. Valid options are"
+ f" {log_levels}"
)
log_level = getattr(logging, upper_log_level_string)
@@ -176,7 +177,8 @@ def setup_logger(app: Flask) -> None:
spiff_logger = logging.getLogger("spiff")
spiff_logger.setLevel(spiff_log_level)
spiff_formatter = logging.Formatter(
- "%(asctime)s | %(levelname)s | %(message)s | %(action)s | %(task_type)s | %(process)s | %(processName)s | %(process_instance_id)s"
+ "%(asctime)s | %(levelname)s | %(message)s | %(action)s | %(task_type)s |"
+ " %(process)s | %(processName)s | %(process_instance_id)s"
)
# if you add a handler to spiff, it will be used/inherited by spiff.metrics
diff --git a/src/spiffworkflow_backend/services/message_service.py b/src/spiffworkflow_backend/services/message_service.py
index cfb42c83..b3d1e831 100644
--- a/src/spiffworkflow_backend/services/message_service.py
+++ b/src/spiffworkflow_backend/services/message_service.py
@@ -145,8 +145,11 @@ class MessageService:
if process_instance_receive is None:
raise MessageServiceError(
(
- f"Process instance cannot be found for queued message: {message_instance_receive.id}."
- f"Tried with id {message_instance_receive.process_instance_id}",
+ (
+ "Process instance cannot be found for queued message:"
+ f" {message_instance_receive.id}.Tried with id"
+ f" {message_instance_receive.process_instance_id}"
+ ),
)
)
@@ -182,7 +185,6 @@ class MessageService:
)
for message_instance_receive in message_instances_receive:
-
# sqlalchemy supports select / where statements like active record apparantly
# https://docs.sqlalchemy.org/en/14/core/tutorial.html#conjunctions
message_correlation_select = (
diff --git a/src/spiffworkflow_backend/services/process_instance_processor.py b/src/spiffworkflow_backend/services/process_instance_processor.py
index 6b03d4ce..74e4ab24 100644
--- a/src/spiffworkflow_backend/services/process_instance_processor.py
+++ b/src/spiffworkflow_backend/services/process_instance_processor.py
@@ -215,14 +215,14 @@ class CustomBpmnScriptEngine(PythonScriptEngine): # type: ignore
except Exception as exception:
if task is None:
raise ProcessInstanceProcessorError(
- "Error evaluating expression: "
- "'%s', exception: %s" % (expression, str(exception)),
+ "Error evaluating expression: '%s', exception: %s"
+ % (expression, str(exception)),
) from exception
else:
raise WorkflowTaskExecException(
task,
- "Error evaluating expression "
- "'%s', %s" % (expression, str(exception)),
+ "Error evaluating expression '%s', %s"
+ % (expression, str(exception)),
) from exception
def execute(
@@ -300,9 +300,7 @@ class ProcessInstanceProcessor:
tld.spiff_step = process_instance_model.spiff_step
# we want this to be the fully qualified path to the process model including all group subcomponents
- current_app.config[
- "THREAD_LOCAL_DATA"
- ].process_model_identifier = (
+ current_app.config["THREAD_LOCAL_DATA"].process_model_identifier = (
f"{process_instance_model.process_model_identifier}"
)
@@ -383,8 +381,10 @@ class ProcessInstanceProcessor:
except MissingSpecError as ke:
raise ApiError(
error_code="unexpected_process_instance_structure",
- message="Failed to deserialize process_instance"
- " '%s' due to a mis-placed or missing task '%s'"
+ message=(
+ "Failed to deserialize process_instance"
+ " '%s' due to a mis-placed or missing task '%s'"
+ )
% (self.process_model_identifier, str(ke)),
) from ke
@@ -400,7 +400,10 @@ class ProcessInstanceProcessor:
raise (
ApiError(
"process_model_not_found",
- f"The given process model was not found: {process_model_identifier}.",
+ (
+ "The given process model was not found:"
+ f" {process_model_identifier}."
+ ),
)
)
spec_files = SpecFileService.get_files(process_model_info)
@@ -530,8 +533,11 @@ class ProcessInstanceProcessor:
potential_owner_ids.append(lane_owner_user.id)
self.raise_if_no_potential_owners(
potential_owner_ids,
- f"No users found in task data lane owner list for lane: {task_lane}. "
- f"The user list used: {task.data['lane_owners'][task_lane]}",
+ (
+ "No users found in task data lane owner list for lane:"
+ f" {task_lane}. The user list used:"
+ f" {task.data['lane_owners'][task_lane]}"
+ ),
)
else:
group_model = GroupModel.query.filter_by(identifier=task_lane).first()
@@ -722,7 +728,8 @@ class ProcessInstanceProcessor:
if payload is not None:
event_definition.payload = payload
current_app.logger.info(
- f"Event of type {event_definition.event_type} sent to process instance {self.process_instance_model.id}"
+ f"Event of type {event_definition.event_type} sent to process instance"
+ f" {self.process_instance_model.id}"
)
self.bpmn_process_instance.catch(event_definition)
self.do_engine_steps(save=True)
@@ -732,12 +739,14 @@ class ProcessInstanceProcessor:
spiff_task = self.bpmn_process_instance.get_task(UUID(task_id))
if execute:
current_app.logger.info(
- f"Manually executing Task {spiff_task.task_spec.name} of process instance {self.process_instance_model.id}"
+ f"Manually executing Task {spiff_task.task_spec.name} of process"
+ f" instance {self.process_instance_model.id}"
)
spiff_task.complete()
else:
current_app.logger.info(
- f"Skipping Task {spiff_task.task_spec.name} of process instance {self.process_instance_model.id}"
+ f"Skipping Task {spiff_task.task_spec.name} of process instance"
+ f" {self.process_instance_model.id}"
)
spiff_task._set_state(TaskState.COMPLETED)
for child in spiff_task.children:
@@ -781,7 +790,8 @@ class ProcessInstanceProcessor:
"""Bpmn_file_full_path_from_bpmn_process_identifier."""
if bpmn_process_identifier is None:
raise ValueError(
- "bpmn_file_full_path_from_bpmn_process_identifier: bpmn_process_identifier is unexpectedly None"
+ "bpmn_file_full_path_from_bpmn_process_identifier:"
+ " bpmn_process_identifier is unexpectedly None"
)
spec_reference = SpecReferenceCache.query.filter_by(
@@ -803,7 +813,10 @@ class ProcessInstanceProcessor:
raise (
ApiError(
error_code="could_not_find_bpmn_process_identifier",
- message="Could not find the the given bpmn process identifier from any sources: %s"
+ message=(
+ "Could not find the the given bpmn process identifier from any"
+ " sources: %s"
+ )
% bpmn_process_identifier,
)
)
@@ -827,7 +840,6 @@ class ProcessInstanceProcessor:
new_bpmn_files = set()
for bpmn_process_identifier in processor_dependencies_new:
-
# ignore identifiers that spiff already knows about
if bpmn_process_identifier in bpmn_process_identifiers_in_parser:
continue
@@ -870,7 +882,10 @@ class ProcessInstanceProcessor:
raise (
ApiError(
error_code="no_primary_bpmn_error",
- message="There is no primary BPMN process id defined for process_model %s"
+ message=(
+ "There is no primary BPMN process id defined for"
+ " process_model %s"
+ )
% process_model_info.id,
)
)
@@ -931,7 +946,10 @@ class ProcessInstanceProcessor:
if not bpmn_message.correlations:
raise ApiError(
"message_correlations_missing",
- f"Could not find any message correlations bpmn_message: {bpmn_message.name}",
+ (
+ "Could not find any message correlations bpmn_message:"
+ f" {bpmn_message.name}"
+ ),
)
message_correlations = []
@@ -951,12 +969,16 @@ class ProcessInstanceProcessor:
if message_correlation_property is None:
raise ApiError(
"message_correlations_missing_from_process",
- "Could not find a known message correlation with identifier:"
- f"{message_correlation_property_identifier}",
+ (
+ "Could not find a known message correlation with"
+ f" identifier:{message_correlation_property_identifier}"
+ ),
)
message_correlations.append(
{
- "message_correlation_property": message_correlation_property,
+ "message_correlation_property": (
+ message_correlation_property
+ ),
"name": message_correlation_key,
"value": message_correlation_property_value,
}
@@ -1013,7 +1035,10 @@ class ProcessInstanceProcessor:
if message_model is None:
raise ApiError(
"invalid_message_name",
- f"Invalid message name: {waiting_task.task_spec.event_definition.name}.",
+ (
+ "Invalid message name:"
+ f" {waiting_task.task_spec.event_definition.name}."
+ ),
)
# Ensure we are only creating one message instance for each waiting message
diff --git a/src/spiffworkflow_backend/services/process_instance_report_service.py b/src/spiffworkflow_backend/services/process_instance_report_service.py
index 18c1bef6..7d725fa2 100644
--- a/src/spiffworkflow_backend/services/process_instance_report_service.py
+++ b/src/spiffworkflow_backend/services/process_instance_report_service.py
@@ -283,9 +283,9 @@ class ProcessInstanceReportService:
process_instance_dict = process_instance["ProcessInstanceModel"].serialized
for metadata_column in metadata_columns:
if metadata_column["accessor"] not in process_instance_dict:
- process_instance_dict[
- metadata_column["accessor"]
- ] = process_instance[metadata_column["accessor"]]
+ process_instance_dict[metadata_column["accessor"]] = (
+ process_instance[metadata_column["accessor"]]
+ )
results.append(process_instance_dict)
return results
diff --git a/src/spiffworkflow_backend/services/process_instance_service.py b/src/spiffworkflow_backend/services/process_instance_service.py
index c9722289..c6e3db42 100644
--- a/src/spiffworkflow_backend/services/process_instance_service.py
+++ b/src/spiffworkflow_backend/services/process_instance_service.py
@@ -85,7 +85,8 @@ class ProcessInstanceService:
db.session.add(process_instance)
db.session.commit()
error_message = (
- f"Error running waiting task for process_instance {process_instance.id}"
+ "Error running waiting task for process_instance"
+ f" {process_instance.id}"
+ f"({process_instance.process_model_identifier}). {str(e)}"
)
current_app.logger.error(error_message)
@@ -178,7 +179,10 @@ class ProcessInstanceService:
else:
raise ApiError.from_task(
error_code="task_lane_user_error",
- message="Spiff Task %s lane user dict must have a key called 'value' with the user's uid in it."
+ message=(
+ "Spiff Task %s lane user dict must have a key called"
+ " 'value' with the user's uid in it."
+ )
% spiff_task.task_spec.name,
task=spiff_task,
)
diff --git a/src/spiffworkflow_backend/services/process_model_service.py b/src/spiffworkflow_backend/services/process_model_service.py
index 714cd799..f9f34631 100644
--- a/src/spiffworkflow_backend/services/process_model_service.py
+++ b/src/spiffworkflow_backend/services/process_model_service.py
@@ -146,7 +146,10 @@ class ProcessModelService(FileSystemService):
if len(instances) > 0:
raise ApiError(
error_code="existing_instances",
- message=f"We cannot delete the model `{process_model_id}`, there are existing instances that depend on it.",
+ message=(
+ f"We cannot delete the model `{process_model_id}`, there are"
+ " existing instances that depend on it."
+ ),
)
process_model = self.get_process_model(process_model_id)
path = self.workflow_path(process_model)
@@ -339,8 +342,11 @@ class ProcessModelService(FileSystemService):
if len(problem_models) > 0:
raise ApiError(
error_code="existing_instances",
- message=f"We cannot delete the group `{process_group_id}`, "
- f"there are models with existing instances inside the group. {problem_models}",
+ message=(
+ f"We cannot delete the group `{process_group_id}`, there are"
+ " models with existing instances inside the group."
+ f" {problem_models}"
+ ),
)
shutil.rmtree(path)
self.cleanup_process_group_display_order()
@@ -392,7 +398,10 @@ class ProcessModelService(FileSystemService):
if process_group is None:
raise ApiError(
error_code="process_group_could_not_be_loaded_from_disk",
- message=f"We could not load the process_group from disk from: {dir_path}",
+ message=(
+ "We could not load the process_group from disk from:"
+ f" {dir_path}"
+ ),
)
else:
process_group_id = dir_path.replace(FileSystemService.root_path(), "")
@@ -457,7 +466,10 @@ class ProcessModelService(FileSystemService):
if process_model_info is None:
raise ApiError(
error_code="process_model_could_not_be_loaded_from_disk",
- message=f"We could not load the process_model from disk with data: {data}",
+ message=(
+ "We could not load the process_model from disk with data:"
+ f" {data}"
+ ),
)
else:
if name is None:
diff --git a/src/spiffworkflow_backend/services/script_unit_test_runner.py b/src/spiffworkflow_backend/services/script_unit_test_runner.py
index 9112e20f..ed331672 100644
--- a/src/spiffworkflow_backend/services/script_unit_test_runner.py
+++ b/src/spiffworkflow_backend/services/script_unit_test_runner.py
@@ -112,7 +112,10 @@ class ScriptUnitTestRunner:
except json.decoder.JSONDecodeError as ex:
return ScriptUnitTestResult(
result=False,
- error=f"Failed to parse expectedOutputJson: {unit_test['expectedOutputJson']}: {str(ex)}",
+ error=(
+ "Failed to parse expectedOutputJson:"
+ f" {unit_test['expectedOutputJson']}: {str(ex)}"
+ ),
)
script = task.task_spec.script
diff --git a/src/spiffworkflow_backend/services/secret_service.py b/src/spiffworkflow_backend/services/secret_service.py
index e4dee491..aa9e6d14 100644
--- a/src/spiffworkflow_backend/services/secret_service.py
+++ b/src/spiffworkflow_backend/services/secret_service.py
@@ -44,8 +44,10 @@ class SecretService:
except Exception as e:
raise ApiError(
error_code="create_secret_error",
- message=f"There was an error creating a secret with key: {key} and value ending with: {value[:-4]}. "
- f"Original error is {e}",
+ message=(
+ f"There was an error creating a secret with key: {key} and value"
+ f" ending with: {value[:-4]}. Original error is {e}"
+ ),
) from e
return secret_model
@@ -89,7 +91,9 @@ class SecretService:
else:
raise ApiError(
error_code="update_secret_error",
- message=f"Cannot update secret with key: {key}. Resource does not exist.",
+ message=(
+ f"Cannot update secret with key: {key}. Resource does not exist."
+ ),
status_code=404,
)
@@ -104,11 +108,16 @@ class SecretService:
except Exception as e:
raise ApiError(
error_code="delete_secret_error",
- message=f"Could not delete secret with key: {key}. Original error is: {e}",
+ message=(
+ f"Could not delete secret with key: {key}. Original error"
+ f" is: {e}"
+ ),
) from e
else:
raise ApiError(
error_code="delete_secret_error",
- message=f"Cannot delete secret with key: {key}. Resource does not exist.",
+ message=(
+ f"Cannot delete secret with key: {key}. Resource does not exist."
+ ),
status_code=404,
)
diff --git a/src/spiffworkflow_backend/services/spec_file_service.py b/src/spiffworkflow_backend/services/spec_file_service.py
index 72f59d1f..4fdfbd6d 100644
--- a/src/spiffworkflow_backend/services/spec_file_service.py
+++ b/src/spiffworkflow_backend/services/spec_file_service.py
@@ -192,7 +192,8 @@ class SpecFileService(FileSystemService):
full_file_path = SpecFileService.full_file_path(process_model_info, file_name)
if not os.path.exists(full_file_path):
raise ProcessModelFileNotFoundError(
- f"No file found with name {file_name} in {process_model_info.display_name}"
+ f"No file found with name {file_name} in"
+ f" {process_model_info.display_name}"
)
with open(full_file_path, "rb") as f_handle:
spec_file_data = f_handle.read()
@@ -314,8 +315,9 @@ class SpecFileService(FileSystemService):
).first()
if message_model is None:
raise ValidationException(
- f"Could not find message model with identifier '{message_model_identifier}'"
- f"Required by a Start Event in : {ref.file_name}"
+ "Could not find message model with identifier"
+ f" '{message_model_identifier}'Required by a Start Event in :"
+ f" {ref.file_name}"
)
message_triggerable_process_model = (
MessageTriggerableProcessModel.query.filter_by(
@@ -335,7 +337,8 @@ class SpecFileService(FileSystemService):
!= ref.process_model_id
):
raise ValidationException(
- f"Message model is already used to start process model {ref.process_model_id}"
+ "Message model is already used to start process model"
+ f" {ref.process_model_id}"
)
@staticmethod
@@ -353,8 +356,9 @@ class SpecFileService(FileSystemService):
).first()
if message_model is None:
raise ValidationException(
- f"Could not find message model with identifier '{message_model_identifier}'"
- f"specified by correlation property: {cpre}"
+ "Could not find message model with identifier"
+ f" '{message_model_identifier}'specified by correlation"
+ f" property: {cpre}"
)
# fixme: I think we are currently ignoring the correction properties.
message_correlation_property = (
diff --git a/tests/spiffworkflow_backend/helpers/base_test.py b/tests/spiffworkflow_backend/helpers/base_test.py
index 8a314f2c..47cf2d87 100644
--- a/tests/spiffworkflow_backend/helpers/base_test.py
+++ b/tests/spiffworkflow_backend/helpers/base_test.py
@@ -133,7 +133,6 @@ class BaseTest:
) -> TestResponse:
"""Create_process_model."""
if process_model_id is not None:
-
# make sure we have a group
process_group_id, _ = os.path.split(process_model_id)
modified_process_group_id = process_group_id.replace("/", ":")
@@ -141,7 +140,6 @@ class BaseTest:
os.path.join(FileSystemService.root_path(), process_group_id)
)
if ProcessModelService.is_group(process_group_path):
-
if exception_notification_addresses is None:
exception_notification_addresses = []
@@ -171,7 +169,8 @@ class BaseTest:
raise Exception("You must create the group first")
else:
raise Exception(
- "You must include the process_model_id, which must be a path to the model"
+ "You must include the process_model_id, which must be a path to the"
+ " model"
)
def get_test_data_file_contents(
diff --git a/tests/spiffworkflow_backend/integration/test_process_api.py b/tests/spiffworkflow_backend/integration/test_process_api.py
index 809063cb..461cdcd4 100644
--- a/tests/spiffworkflow_backend/integration/test_process_api.py
+++ b/tests/spiffworkflow_backend/integration/test_process_api.py
@@ -343,7 +343,8 @@ class TestProcessApi(BaseTest):
assert data["error_code"] == "existing_instances"
assert (
data["message"]
- == f"We cannot delete the model `{process_model_identifier}`, there are existing instances that depend on it."
+ == f"We cannot delete the model `{process_model_identifier}`, there are"
+ " existing instances that depend on it."
)
def test_process_model_update(
@@ -2058,7 +2059,6 @@ class TestProcessApi(BaseTest):
mail = app.config["MAIL_APP"]
with mail.record_messages() as outbox:
-
response = client.post(
f"/v1.0/process-instances/{self.modify_process_identifier_for_path_param(process_model_identifier)}/{process_instance_id}/run",
headers=self.logged_in_headers(with_super_admin_user),
@@ -2961,7 +2961,9 @@ class TestProcessApi(BaseTest):
) -> None:
"""Test_can_get_process_instance_list_with_report_metadata."""
process_model = load_test_spec(
- process_model_id="save_process_instance_metadata/save_process_instance_metadata",
+ process_model_id=(
+ "save_process_instance_metadata/save_process_instance_metadata"
+ ),
bpmn_file_name="save_process_instance_metadata.bpmn",
process_model_source_directory="save_process_instance_metadata",
)
@@ -3018,7 +3020,9 @@ class TestProcessApi(BaseTest):
) -> None:
"""Test_can_get_process_instance_list_with_report_metadata."""
process_model = load_test_spec(
- process_model_id="save_process_instance_metadata/save_process_instance_metadata",
+ process_model_id=(
+ "save_process_instance_metadata/save_process_instance_metadata"
+ ),
bpmn_file_name="save_process_instance_metadata.bpmn",
process_model_source_directory="save_process_instance_metadata",
)
diff --git a/tests/spiffworkflow_backend/scripts/test_save_process_instance_metadata.py b/tests/spiffworkflow_backend/scripts/test_save_process_instance_metadata.py
index 5f66fa72..738896cd 100644
--- a/tests/spiffworkflow_backend/scripts/test_save_process_instance_metadata.py
+++ b/tests/spiffworkflow_backend/scripts/test_save_process_instance_metadata.py
@@ -28,7 +28,9 @@ class TestSaveProcessInstanceMetadata(BaseTest):
client, with_super_admin_user, "test_group", "test_group"
)
process_model = load_test_spec(
- process_model_id="save_process_instance_metadata/save_process_instance_metadata",
+ process_model_id=(
+ "save_process_instance_metadata/save_process_instance_metadata"
+ ),
bpmn_file_name="save_process_instance_metadata.bpmn",
process_model_source_directory="save_process_instance_metadata",
)
diff --git a/tests/spiffworkflow_backend/unit/test_permissions.py b/tests/spiffworkflow_backend/unit/test_permissions.py
index b66f3237..a9698969 100644
--- a/tests/spiffworkflow_backend/unit/test_permissions.py
+++ b/tests/spiffworkflow_backend/unit/test_permissions.py
@@ -16,6 +16,7 @@ from spiffworkflow_backend.services.user_service import UserService
# we think we can get the list of roles for a user.
# spiff needs a way to determine what each role allows.
+
# user role allows list and read of all process groups/models
# super-admin role allows create, update, and delete of all process groups/models
# * super-admins users maybe conventionally get the user role as well
diff --git a/tests/spiffworkflow_backend/unit/test_process_instance_processor.py b/tests/spiffworkflow_backend/unit/test_process_instance_processor.py
index 4f7db3ad..b4a650dc 100644
--- a/tests/spiffworkflow_backend/unit/test_process_instance_processor.py
+++ b/tests/spiffworkflow_backend/unit/test_process_instance_processor.py
@@ -52,7 +52,8 @@ class TestProcessInstanceProcessor(BaseTest):
result = script_engine._evaluate("fact_service(type='norris')", {})
assert (
result
- == "Chuck Norris doesn’t read books. He stares them down until he gets the information he wants."
+ == "Chuck Norris doesn’t read books. He stares them down until he gets the"
+ " information he wants."
)
app.config["THREAD_LOCAL_DATA"].process_model_identifier = None
app.config["THREAD_LOCAL_DATA"].process_instance_id = None
diff --git a/tests/spiffworkflow_backend/unit/test_process_instance_report_service.py b/tests/spiffworkflow_backend/unit/test_process_instance_report_service.py
index 75ad3f28..b40412ff 100644
--- a/tests/spiffworkflow_backend/unit/test_process_instance_report_service.py
+++ b/tests/spiffworkflow_backend/unit/test_process_instance_report_service.py
@@ -880,7 +880,9 @@ class TestProcessInstanceReportService(BaseTest):
process_instance_report = ProcessInstanceReportService.report_with_identifier(
user=user_one,
- report_identifier="system_report_completed_instances_with_tasks_completed_by_me",
+ report_identifier=(
+ "system_report_completed_instances_with_tasks_completed_by_me"
+ ),
)
report_filter = (
ProcessInstanceReportService.filter_from_metadata_with_overrides(
@@ -983,7 +985,9 @@ class TestProcessInstanceReportService(BaseTest):
process_instance_report = ProcessInstanceReportService.report_with_identifier(
user=user_one,
- report_identifier="system_report_completed_instances_with_tasks_completed_by_my_groups",
+ report_identifier=(
+ "system_report_completed_instances_with_tasks_completed_by_my_groups"
+ ),
)
report_filter = (
ProcessInstanceReportService.filter_from_metadata_with_overrides(