mistral/mistral/tests/unit/engine/test_javascript_action.py

110 lines
3.2 KiB
Python

# Copyright 2015 - Mirantis, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import mock
from oslo_config import cfg
from oslo_utils import importutils
import testtools
from mistral.db.v2 import api as db_api
from mistral.services import workflows as wf_service
from mistral.tests.unit.engine import base
from mistral.utils import javascript
from mistral.workflow import states
# Use the set_default method to set value otherwise in certain test cases
# the change in value is not permanent.
cfg.CONF.set_default('auth_enable', False, group='pecan')
JAVASCRIPT_WORKFLOW = """
version: "2.0"
wf:
input:
- length
tasks:
task1:
action: std.javascript
input:
script: |
let numberSequence = Array.from({length: $['length']},
(x, i) => i);
let evenNumbers = numberSequence.filter(x => x % 2 === 0);
return evenNumbers.length;
context: <% $ %>
publish:
res: <% task().result %>
"""
def fake_evaluate(_, context):
return context['length'] / 2
class JavaScriptEngineTest(base.EngineTestCase):
@testtools.skipIf(not importutils.try_import('py_mini_racer'),
'This test requires that py_mini_racer library was '
'installed')
def test_py_mini_racer_javascript_action(self):
cfg.CONF.set_default(
'js_implementation',
'py_mini_racer'
)
length = 1000
wf_service.create_workflows(JAVASCRIPT_WORKFLOW)
# Start workflow.
wf_ex = self.engine.start_workflow(
'wf',
wf_input={'length': length}
)
self.await_workflow_success(wf_ex.id)
with db_api.transaction():
wf_ex = db_api.get_workflow_execution(wf_ex.id)
task_ex = wf_ex.task_executions[0]
self.assertEqual(states.SUCCESS, task_ex.state)
self.assertDictEqual({}, task_ex.runtime_context)
self.assertEqual(length / 2, task_ex.published['res'])
@mock.patch.object(javascript, 'evaluate', fake_evaluate)
def test_fake_javascript_action_data_context(self):
length = 1000
wf_service.create_workflows(JAVASCRIPT_WORKFLOW)
# Start workflow.
wf_ex = self.engine.start_workflow(
'wf',
wf_input={'length': length}
)
self.await_workflow_success(wf_ex.id)
with db_api.transaction():
wf_ex = db_api.get_workflow_execution(wf_ex.id)
task_ex = wf_ex.task_executions[0]
self.assertEqual(states.SUCCESS, task_ex.state)
self.assertDictEqual({}, task_ex.runtime_context)
self.assertEqual(length / 2, task_ex.published['res'])