mirror of
https://github.com/status-im/spiff-arena.git
synced 2025-01-28 10:45:07 +00:00
35fd8ffc0f
b439f69f2 Merge pull request #296 from sartography/bugfix/subprocess-access-to-data-objects 6d2a2031e update spiff subworkflow tasks too 992c38671 make data objects referenceable within subprocesses 6c8ff5cdf allow subprocesses & call activities to have different data copy policies 2b14f3a48 initialize subprocesses in _update_hook instead of _on_ready_before 791f335d5 Merge pull request #295 from sartography/improvement/remove-camunda-from-base-and-misc-cleanup 28b579beb remove a few unused, duplicative, and debugging methods 8f14d1098 remove some other unused diagrams and tests 408bc6734 rely on top level camunda parser for almost all namespace references 895b2cc9b remove camunda namespace from base bpmn parser 76ecbf7cc Merge pull request #294 from sartography/bugfix/reactivate-boundary-event 82b6c8ad4 hack to ensure timers (and other events) are reset if returned to via loop reset 590903f47 Merge pull request #292 from sartography/feature/multiinstance-refactor 537490043 fix bug & typo f31726db1 raise error on attempting to migrate workflows with MI 44e6d08d8 create spiff multiinstance task 2168c022b create camunda MI that approximates what it used to do 9894cea59 some improvements and bugfixes f857ad5d4 remove some now unused functionality & tests, create a few more tests 6fead9d04 updated serializer & fixes for most tests ec662ecdd add parallel multiinstance bd19b2a8a working sequential multiinstance 2f9c192b6 further cleanup around _update_hook 947792bf6 fix bug in exclusive gateway migration d3d87b28d add io spec to all tasks f1586e275 add support for standard loop tasks git-subtree-dir: SpiffWorkflow git-subtree-split: b439f69f23b547df4de1e8e0c636997f2fd4e33b
155 lines
6.0 KiB
Python
155 lines
6.0 KiB
Python
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright (C) 2007 Samuel Abels
|
|
#
|
|
# This library is free software; you can redistribute it and/or
|
|
# modify it under the terms of the GNU Lesser General Public
|
|
# License as published by the Free Software Foundation; either
|
|
# version 2.1 of the License, or (at your option) any later version.
|
|
#
|
|
# This library is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
# Lesser General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU Lesser General Public
|
|
# License along with this library; if not, write to the Free Software
|
|
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
|
|
# 02110-1301 USA
|
|
import os
|
|
|
|
from .StartTask import StartTask
|
|
from .base import TaskSpec
|
|
from ..task import TaskState
|
|
from ..exceptions import WorkflowException
|
|
from ..operators import valueof
|
|
|
|
|
|
class SubWorkflow(TaskSpec):
|
|
|
|
"""
|
|
A SubWorkflow is a task that wraps a WorkflowSpec, such that you can
|
|
re-use it in multiple places as if it were a task.
|
|
If more than one input is connected, the task performs an implicit
|
|
multi merge.
|
|
If more than one output is connected, the task performs an implicit
|
|
parallel split.
|
|
"""
|
|
|
|
def __init__(self,
|
|
wf_spec,
|
|
name,
|
|
file,
|
|
in_assign=None,
|
|
out_assign=None,
|
|
**kwargs):
|
|
"""
|
|
Constructor.
|
|
|
|
:type wf_spec: WorkflowSpec
|
|
:param wf_spec: A reference to the workflow specification.
|
|
:type name: str
|
|
:param name: The name of the task spec.
|
|
:type file: str
|
|
:param file: The name of a file containing a workflow.
|
|
:type in_assign: list(str)
|
|
:param in_assign: The names of data fields to carry over.
|
|
:type out_assign: list(str)
|
|
:param out_assign: The names of data fields to carry back.
|
|
:type kwargs: dict
|
|
:param kwargs: See :class:`SpiffWorkflow.specs.TaskSpec`.
|
|
"""
|
|
assert wf_spec is not None
|
|
assert name is not None
|
|
super(SubWorkflow, self).__init__(wf_spec, name, **kwargs)
|
|
self.file = None
|
|
self.in_assign = in_assign is not None and in_assign or []
|
|
self.out_assign = out_assign is not None and out_assign or []
|
|
if file is not None:
|
|
dirname = os.path.dirname(wf_spec.file)
|
|
self.file = os.path.join(dirname, file)
|
|
|
|
def test(self):
|
|
TaskSpec.test(self)
|
|
if self.file is not None and not os.path.exists(self.file):
|
|
raise WorkflowException('File does not exist: %s' % self.file, task_spec=self)
|
|
|
|
def _predict_hook(self, my_task):
|
|
# Modifying the task spec is a TERRIBLE idea, but if we don't do it, sync_children won't work
|
|
outputs = [task.task_spec for task in my_task.children]
|
|
for output in self.outputs:
|
|
if output not in outputs:
|
|
outputs.insert(0, output)
|
|
if my_task._is_definite():
|
|
# This prevents errors with sync children
|
|
my_task._sync_children(outputs, TaskState.LIKELY)
|
|
else:
|
|
my_task._sync_children(outputs, my_task.state)
|
|
|
|
def _create_subworkflow(self, my_task):
|
|
from ..serializer.prettyxml import XmlSerializer
|
|
from ..specs.WorkflowSpec import WorkflowSpec
|
|
from ..workflow import Workflow
|
|
file_name = valueof(my_task, self.file)
|
|
serializer = XmlSerializer()
|
|
with open(file_name) as fp:
|
|
xml = fp.read()
|
|
wf_spec = WorkflowSpec.deserialize(
|
|
serializer, xml, filename=file_name)
|
|
outer_workflow = my_task.workflow.outer_workflow
|
|
return Workflow(wf_spec, parent=outer_workflow)
|
|
|
|
def _on_ready_before_hook(self, my_task):
|
|
subworkflow = self._create_subworkflow(my_task)
|
|
subworkflow.completed_event.connect(
|
|
self._on_subworkflow_completed, my_task)
|
|
self._integrate_subworkflow_tree(my_task, subworkflow)
|
|
my_task._set_internal_data(subworkflow=subworkflow)
|
|
|
|
def _integrate_subworkflow_tree(self, my_task, subworkflow):
|
|
# Integrate the tree of the subworkflow into the tree of this workflow.
|
|
my_task._sync_children(self.outputs, TaskState.LIKELY)
|
|
for child in subworkflow.task_tree.children:
|
|
my_task.children.insert(0, child)
|
|
child.parent = my_task
|
|
|
|
def _on_ready_hook(self, my_task):
|
|
# Assign variables, if so requested.
|
|
subworkflow = my_task._get_internal_data('subworkflow')
|
|
for child in subworkflow.task_tree.children:
|
|
for assignment in self.in_assign:
|
|
assignment.assign(my_task, child)
|
|
child.task_spec._update(child)
|
|
# Instead of completing immediately, we'll wait for the subworkflow to complete
|
|
my_task._set_state(TaskState.WAITING)
|
|
|
|
def _update_hook(self, my_task):
|
|
|
|
super()._update_hook(my_task)
|
|
subworkflow = my_task._get_internal_data('subworkflow')
|
|
if subworkflow is None:
|
|
# On the first update, we have to create the subworkflow
|
|
return True
|
|
elif subworkflow.is_completed():
|
|
# Then wait until it finishes to complete
|
|
my_task.complete()
|
|
|
|
def _on_subworkflow_completed(self, subworkflow, my_task):
|
|
# Assign variables, if so requested.
|
|
for child in my_task.children:
|
|
if not isinstance(child.task_spec, StartTask):
|
|
if subworkflow.last_task is not None:
|
|
child.data = subworkflow.last_task.data
|
|
for assignment in self.out_assign:
|
|
assignment.assign(subworkflow, child)
|
|
|
|
# Alright, abusing that hook is just evil but it works.
|
|
child.task_spec._update_hook(child)
|
|
|
|
def serialize(self, serializer):
|
|
return serializer.serialize_sub_workflow(self)
|
|
|
|
@classmethod
|
|
def deserialize(self, serializer, wf_spec, s_state):
|
|
return serializer.deserialize_sub_workflow(wf_spec, s_state)
|