diff --git a/.github/workflows/tox.yaml b/.github/workflows/tox.yaml index 9379c9f94..311c12015 100644 --- a/.github/workflows/tox.yaml +++ b/.github/workflows/tox.yaml @@ -5,11 +5,35 @@ on: - pull_request jobs: + build_old_versions: + runs-on: ubuntu-20.04 + strategy: + matrix: + python-version: ['3.6'] + + steps: + - uses: actions/checkout@v1 + - name: Set up Python ${{ matrix.python-version }} + uses: actions/setup-python@v2 + with: + python-version: ${{ matrix.python-version }} + - name: Install dependencies + run: | + python -m pip install --upgrade pip + pip install codecov tox tox-gh-actions + - name: Lint with tox + run: tox -e pep8 + - name: Test with tox + run: tox -e py + - name: Codecov + run: | + set -euxo pipefail + codecov --verbose --gcov-glob unit_tests/* build: runs-on: ubuntu-latest strategy: matrix: - python-version: ['3.6', '3.7', '3.8', '3.9', '3.10'] + python-version: ['3.7', '3.8', '3.9', '3.10'] steps: - uses: actions/checkout@v1 diff --git a/requirements.txt b/requirements.txt index faffe82cc..5a5371b03 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,8 +1,7 @@ pyparsing<3.0.0 # pin for aodhclient which is held for py35 async_generator kubernetes<18.0.0; python_version < '3.6' # pined, as juju uses kubernetes -# pinned until 3.0 regressions are handled: https://github.com/openstack-charmers/zaza/issues/545 -juju<3.0 +juju juju_wait PyYAML>=3.0 pbr==5.6.0 diff --git a/setup.py b/setup.py index 87445e411..52a574275 100644 --- a/setup.py +++ b/setup.py @@ -36,7 +36,7 @@ 'hvac<0.7.0', 'jinja2', - 'juju<3.0', + 'juju', 'juju-wait', 'PyYAML', 'tenacity', diff --git a/unit_tests/test_zaza_model.py b/unit_tests/test_zaza_model.py index 0c8cf962e..8c75d95e7 100644 --- a/unit_tests/test_zaza_model.py +++ b/unit_tests/test_zaza_model.py @@ -142,6 +142,7 @@ async def _inner_is_leader(): self.run_action = mock.MagicMock() self.run_action.wait.side_effect = _wait self.action = mock.MagicMock() + self.action.wait.side_effect = _wait self.action.data = { 'model-uuid': '1a035018-71ff-473e-8aab-d1a8d6b6cda7', 'id': 'e26ffb69-6626-4e93-8840-07f7e041e99d', @@ -493,7 +494,9 @@ def test_block_until_auto_reconnect_model_disconnected_sync(self): with mock.patch.object(zaza, 'RUN_LIBJUJU_IN_THREAD', new=False): model.sync_wrapper(self._wrapper)() self.Model_mock.disconnect.assert_has_calls([mock.call()]) - self.Model_mock.connect_model.has_calls([mock.call('modelname')]) + self.Model_mock.connect_model.assert_has_calls( + [mock.call('testmodel')] + ) def test_block_until_auto_reconnect_model_disconnected_async(self): self._mocks_for_block_until_auto_reconnect_model( @@ -506,7 +509,9 @@ async def _async_true(): with mock.patch.object(zaza, 'RUN_LIBJUJU_IN_THREAD', new=False): model.sync_wrapper(self._wrapper)() self.Model_mock.disconnect.assert_has_calls([mock.call()]) - self.Model_mock.connect_model.has_calls([mock.call('modelname')]) + self.Model_mock.connect_model.assert_has_calls( + [mock.call('testmodel')] + ) def test_block_until_auto_reconnect_model_blocks_till_true(self): self._mocks_for_block_until_auto_reconnect_model(True, True) diff --git a/unit_tests/utilities/test_zaza_utilities.py b/unit_tests/utilities/test_zaza_utilities.py index b0a791817..038d53e0b 100644 --- a/unit_tests/utilities/test_zaza_utilities.py +++ b/unit_tests/utilities/test_zaza_utilities.py @@ -1,4 +1,4 @@ -# Copyright 2021 Canonical Ltd. +# Copyright 2018 Canonical Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -12,85 +12,633 @@ # See the License for the specific language governing permissions and # limitations under the License. -"""Unit tests for the __init__.py functions.""" +import mock +import subprocess -import unittest +import unit_tests.utils as ut_utils +import zaza +from zaza.utilities import generic as generic_utils +import zaza.utilities.exceptions as zaza_exceptions -import zaza.utilities +FAKE_STATUS = { + 'can-upgrade-to': '', + 'charm': 'local:trusty/app-136', + 'subordinate-to': [], + 'units': {'app/0': {'leader': True, + 'machine': '0', + 'subordinates': { + 'app-hacluster/0': { + 'charm': 'local:trusty/hacluster-0', + 'leader': True}}}, + 'app/1': {'machine': '1', + 'subordinates': { + 'app-hacluster/1': { + 'charm': 'local:trusty/hacluster-0'}}}, + 'app/2': {'machine': '2', + 'subordinates': { + 'app-hacluster/2': { + 'charm': 'local:trusty/hacluster-0'}}}}} -class TestConfigurableMixin(unittest.TestCase): +def tearDownModule(): + zaza.clean_up_libjuju_thread() - def test_sets_attributes(self): - class A(zaza.utilities.ConfigurableMixin): +class TestGenericUtils(ut_utils.BaseTestCase): - def __init__(self, **kwargs): - self.a = None - self.b = 1 - self.configure(**kwargs) + def setUp(self): + super(TestGenericUtils, self).setUp() + # Patch all subprocess calls + self.patch( + 'zaza.utilities.generic.subprocess', + new_callable=mock.MagicMock(), + name='subprocess' + ) - a = A(a=4) - self.assertEqual(a.a, 4) - self.assertEqual(a.b, 1) - a.configure(b=9) - self.assertEqual(a.b, 9) + # Juju Status Object and data + self.juju_status = mock.MagicMock() + self.juju_status.applications.__getitem__.return_value = FAKE_STATUS + self.patch_object(generic_utils, "model") + self.model.get_status.return_value = self.juju_status - def test_doesnt_instantiate_attributes(self): + def test_dict_to_yaml(self): + _dict_data = {"key": "value"} + _str_data = "key: value\n" + self.assertEqual(generic_utils.dict_to_yaml(_dict_data), + _str_data) - class A(zaza.utilities.ConfigurableMixin): + def test_get_network_config(self): + self.patch_object(generic_utils.os.path, "exists") + self.patch_object(generic_utils, "get_yaml_config") + self.patch_object(generic_utils, "get_undercloud_env_vars") + net_topology = "topo" + _data = {net_topology: {"network": "DATA"}} + self.get_yaml_config.return_value = _data - def __init__(self, **kwargs): - self.a = None - self.configure(**kwargs) + # YAML file does not exist + self.exists.return_value = False + with self.assertRaises(Exception): + generic_utils.get_network_config(net_topology) - with self.assertRaises(AttributeError): - A(a=5, b=10) + # No environmental variables + self.exists.return_value = True + self.assertEqual( + generic_utils.get_network_config(net_topology, + ignore_env_vars=True), + _data[net_topology]) + self.get_yaml_config.assert_called_once_with("network.yaml") + self.get_undercloud_env_vars.assert_not_called() - def test_cant_set_hidden(self): + # Update with environmental variables + _more_data = {"network": "NEW", + "other": "DATA"} + self.get_undercloud_env_vars.return_value = _more_data + _data[net_topology].update(_more_data) + self.assertEqual( + generic_utils.get_network_config(net_topology), + _data[net_topology]) + self.get_undercloud_env_vars.assert_called_once_with() - class A(zaza.utilities.ConfigurableMixin): + def test_get_pkg_version(self): + self.patch_object(generic_utils.model, "get_units") + self.patch_object(generic_utils.juju_utils, "remote_run") + _pkg = "os-thingy" + _version = "2:27.0.0-0ubuntu1~cloud0" + _dpkg_output = ("ii {} {} all OpenStack thingy\n" + .format(_pkg, _version)) + self.remote_run.return_value = _dpkg_output + _unit1 = mock.MagicMock() + _unit1.entity_id = "os-thingy/7" + _unit2 = mock.MagicMock() + _unit2.entity_id = "os-thingy/12" + _units = [_unit1, _unit2] + self.get_units.return_value = _units - def __init__(self, **kwargs): - self.a = None - self._b = 3 - self.configure(**kwargs) + # Matching + self.assertEqual(generic_utils.get_pkg_version(_pkg, _pkg), + _version) - with self.assertRaises(AttributeError): - a = A(_b=4) - a = A() - a.configure(a=10) - self.assertEqual(a.a, 10) - with self.assertRaises(AttributeError): - a.configure(_b=8) + # Mismatched + _different_dpkg_output = ("ii {} {} all OpenStack thingy\n" + .format(_pkg, "DIFFERENT")) + self.remote_run.side_effect = [_dpkg_output, _different_dpkg_output] + with self.assertRaises(Exception): + generic_utils.get_pkg_version(_pkg, _pkg) - def test_cant_set_properties(self): + def test_get_undercloud_env_vars(self): + self.patch_object(generic_utils.os.environ, "get") - class A(zaza.utilities.ConfigurableMixin): + def _get_env(key): + return _env.get(key) + self.get.side_effect = _get_env - def __init__(self): - self.a = 10 + # OSCI backward compatible env vars + _env = {"NET_ID": "netid", + "NAMESERVER": "10.0.0.10", + "GATEWAY": "10.0.0.1", + "CIDR_EXT": "10.0.0.0/24", + "FIP_RANGE": "10.0.200.0:10.0.200.254"} + _expected_result = {} + _expected_result["net_id"] = _env["NET_ID"] + _expected_result["external_dns"] = _env["NAMESERVER"] + _expected_result["default_gateway"] = _env["GATEWAY"] + _expected_result["external_net_cidr"] = _env["CIDR_EXT"] + _expected_result["start_floating_ip"] = _env["FIP_RANGE"].split(":")[0] + _expected_result["end_floating_ip"] = _env["FIP_RANGE"].split(":")[1] + self.assertEqual(generic_utils.get_undercloud_env_vars(), + _expected_result) - @property - def b(): - return 5 + # Overriding configure.network named variables + _override = {"start_floating_ip": "10.100.50.0", + "end_floating_ip": "10.100.50.254", + "default_gateway": "10.100.0.1", + "external_net_cidr": "10.100.0.0/16"} + _env.update(_override) + _expected_result.update(_override) + self.assertEqual(generic_utils.get_undercloud_env_vars(), + _expected_result) - a = A() - a.configure(a=5) - with self.assertRaises(AttributeError): - a.configure(a=3, b=10) + def test_get_yaml_config(self): + self.patch("builtins.open", + new_callable=mock.mock_open(), + name="_open") + _yaml = "data: somedata" + _yaml_dict = {"data": "somedata"} + _filename = "filename" + _fileobj = mock.MagicMock() + _fileobj.read.return_value = _yaml + self._open.return_value = _fileobj - def test_cant_set_methods(self): + self.assertEqual(generic_utils.get_yaml_config(_filename), + _yaml_dict) + self._open.assert_called_once_with(_filename, "r") - class A(zaza.utilities.ConfigurableMixin): + def test_do_release_upgrade(self): + _unit = "app/2" + generic_utils.do_release_upgrade(_unit) + self.subprocess.check_call.assert_called_once_with( + ['juju', 'ssh', _unit, 'sudo', 'DEBIAN_FRONTEND=noninteractive', + 'do-release-upgrade', '-f', 'DistUpgradeViewNonInteractive']) - def __init__(self): - self.a = 10 + def test_wrap_do_release_upgrade(self): + self.patch_object(generic_utils, "do_release_upgrade") + self.patch_object(generic_utils, "run_via_ssh") + self.patch_object(generic_utils.model, "scp_to_unit") + _unit = "app/2" + _from_series = "xenial" + _to_series = "bionic" + _workaround_script = "scriptname" + _files = ["filename", _workaround_script] + _scp_calls = [] + _run_calls = [ + mock.call(_unit, _workaround_script)] + for filename in _files: + _scp_calls.append(mock.call(_unit, filename, filename)) + generic_utils.wrap_do_release_upgrade( + _unit, to_series=_to_series, from_series=_from_series, + workaround_script=_workaround_script, files=_files) + self.scp_to_unit.assert_has_calls(_scp_calls) + self.run_via_ssh.assert_has_calls(_run_calls) + self.do_release_upgrade.assert_called_once_with(_unit) - def b(): - return 5 + def test_reboot(self): + _unit = "app/2" + generic_utils.reboot(_unit) + self.subprocess.check_call.assert_called_once_with( + ['juju', 'ssh', _unit, + 'sudo', 'reboot', '&&', 'exit']) - a = A() - a.configure(a=5) - with self.assertRaises(AttributeError): - a.configure(a=3, b=10) + def test_juju_reboot(self): + _unit = "app/2" + generic_utils.juju_reboot(_unit) + self.subprocess.check_call.assert_called_once_with( + ['juju', 'ssh', _unit, + f'sudo juju-run -u {_unit} "juju-reboot --now"']) + + def test_run_via_ssh(self): + _unit = "app/2" + _cmd = "hostname" + generic_utils.run_via_ssh(_unit, _cmd) + self.subprocess.check_call.assert_called_once_with( + ['juju', 'ssh', _unit, + 'sudo ' + _cmd]) + + def test_set_origin(self): + "application, origin='openstack-origin', pocket='distro'):" + self.patch_object(generic_utils.model, "set_application_config") + _application = "application" + _origin = "source" + _pocket = "cloud:fake-cloud" + generic_utils.set_origin(_application, origin=_origin, pocket=_pocket) + self.set_application_config.assert_called_once_with( + _application, {_origin: _pocket}) + + def test_series_upgrade(self): + self.patch_object(generic_utils.model, "block_until_all_units_idle") + self.patch_object(generic_utils.model, "block_until_unit_wl_status") + self.patch_object(generic_utils.model, "prepare_series_upgrade") + self.patch_object(generic_utils.model, "complete_series_upgrade") + self.patch_object(generic_utils.model, "set_series") + self.patch_object(generic_utils, "set_origin") + self.patch_object(generic_utils, "wrap_do_release_upgrade") + self.patch_object(generic_utils, "reboot") + _unit = "app/2" + _application = "app" + _machine_num = "4" + _from_series = "xenial" + _to_series = "bionic" + _origin = "source" + _files = ["filename", "scriptname"] + _workaround_script = "scriptname" + generic_utils.series_upgrade( + _unit, _machine_num, origin=_origin, + to_series=_to_series, from_series=_from_series, + workaround_script=_workaround_script, files=_files) + self.block_until_all_units_idle.assert_called_with() + self.prepare_series_upgrade.assert_called_once_with( + _machine_num, to_series=_to_series) + self.wrap_do_release_upgrade.assert_called_once_with( + _unit, to_series=_to_series, from_series=_from_series, + workaround_script=_workaround_script, files=_files) + self.complete_series_upgrade.assert_called_once_with(_machine_num) + self.set_series.assert_called_once_with(_application, _to_series) + self.set_origin.assert_called_once_with(_application, _origin) + self.reboot.assert_called_once_with(_unit) + + def test_series_upgrade_application_pause_peers_and_subordinates(self): + self.patch_object(generic_utils.model, "run_action") + self.patch_object(generic_utils, "series_upgrade") + _application = "app" + _from_series = "xenial" + _to_series = "bionic" + _origin = "source" + _files = ["filename", "scriptname"] + _workaround_script = "scriptname" + _completed_machines = [] + # Peers and Subordinates + _run_action_calls = [ + mock.call("{}-hacluster/1".format(_application), + "pause", action_params={}), + mock.call("{}/1".format(_application), "pause", action_params={}), + mock.call("{}-hacluster/2".format(_application), + "pause", action_params={}), + mock.call("{}/2".format(_application), "pause", action_params={}), + ] + _series_upgrade_calls = [] + for machine_num in ("0", "1", "2"): + _series_upgrade_calls.append( + mock.call("{}/{}".format(_application, machine_num), + machine_num, origin=_origin, + from_series=_from_series, to_series=_to_series, + workaround_script=_workaround_script, files=_files), + ) + + # Pause primary peers and subordinates + generic_utils.series_upgrade_application( + _application, origin=_origin, + to_series=_to_series, from_series=_from_series, + pause_non_leader_primary=True, + pause_non_leader_subordinate=True, + completed_machines=_completed_machines, + workaround_script=_workaround_script, files=_files), + self.run_action.assert_has_calls(_run_action_calls) + self.series_upgrade.assert_has_calls(_series_upgrade_calls) + + def test_series_upgrade_application_pause_subordinates(self): + self.patch_object(generic_utils.model, "run_action") + self.patch_object(generic_utils, "series_upgrade") + _application = "app" + _from_series = "xenial" + _to_series = "bionic" + _origin = "source" + _files = ["filename", "scriptname"] + _workaround_script = "scriptname" + _completed_machines = [] + # Subordinates only + _run_action_calls = [ + mock.call("{}-hacluster/1".format(_application), + "pause", action_params={}), + mock.call("{}-hacluster/2".format(_application), + "pause", action_params={}), + ] + _series_upgrade_calls = [] + + for machine_num in ("0", "1", "2"): + _series_upgrade_calls.append( + mock.call("{}/{}".format(_application, machine_num), + machine_num, origin=_origin, + from_series=_from_series, to_series=_to_series, + workaround_script=_workaround_script, files=_files), + ) + + # Pause subordinates + generic_utils.series_upgrade_application( + _application, origin=_origin, + to_series=_to_series, from_series=_from_series, + pause_non_leader_primary=False, + pause_non_leader_subordinate=True, + completed_machines=_completed_machines, + workaround_script=_workaround_script, files=_files), + self.run_action.assert_has_calls(_run_action_calls) + self.series_upgrade.assert_has_calls(_series_upgrade_calls) + + def test_series_upgrade_application_no_pause(self): + self.patch_object(generic_utils.model, "run_action") + self.patch_object(generic_utils, "series_upgrade") + _application = "app" + _from_series = "xenial" + _to_series = "bionic" + _origin = "source" + _series_upgrade_calls = [] + _files = ["filename", "scriptname"] + _workaround_script = "scriptname" + _completed_machines = [] + + for machine_num in ("0", "1", "2"): + _series_upgrade_calls.append( + mock.call("{}/{}".format(_application, machine_num), + machine_num, origin=_origin, + from_series=_from_series, to_series=_to_series, + workaround_script=_workaround_script, files=_files), + ) + + # No Pausiing + generic_utils.series_upgrade_application( + _application, origin=_origin, + to_series=_to_series, from_series=_from_series, + pause_non_leader_primary=False, + pause_non_leader_subordinate=False, + completed_machines=_completed_machines, + workaround_script=_workaround_script, files=_files) + self.run_action.assert_not_called() + self.series_upgrade.assert_has_calls(_series_upgrade_calls) + + def test_set_dpkg_non_interactive_on_unit(self): + self.patch_object(generic_utils, "model") + _unit_name = "app/1" + generic_utils.set_dpkg_non_interactive_on_unit(_unit_name) + self.model.run_on_unit.assert_called_with( + "app/1", + 'grep \'DPkg::options { "--force-confdef"; };\' ' + '/etc/apt/apt.conf.d/50unattended-upgrades || ' + 'echo \'DPkg::options { "--force-confdef"; };\' >> ' + '/etc/apt/apt.conf.d/50unattended-upgrades') + + def test_get_process_id_list(self): + self.patch( + "zaza.utilities.generic.model.run_on_unit", + new_callable=mock.MagicMock(), + name="_run" + ) + + # Return code is OK and STDOUT contains output + returns_ok = { + "Code": 0, + "Stdout": "1 2", + "Stderr": "" + } + self._run.return_value = returns_ok + p_id_list = generic_utils.get_process_id_list( + "ceph-osd/0", + "ceph-osd", + False + ) + expected = ["1", "2"] + cmd = 'pidof -x "ceph-osd" || exit 0 && exit 1' + self.assertEqual(p_id_list, expected) + self._run.assert_called_once_with(unit_name="ceph-osd/0", + command=cmd) + + # Return code is not OK + returns_nok = { + "Code": 1, + "Stdout": "", + "Stderr": "Something went wrong" + } + self._run.return_value = returns_nok + with self.assertRaises(zaza_exceptions.ProcessIdsFailed): + generic_utils.get_process_id_list("ceph-osd/0", "ceph") + cmd = 'pidof -x "ceph"' + self._run.assert_called_once_with(unit_name="ceph-osd/0", + command=cmd) + + def test_get_process_id_list_with_pgrep(self): + self.patch( + "zaza.utilities.generic.model.run_on_unit", + new_callable=mock.MagicMock(), + name="_run" + ) + + # Return code is OK and STDOUT contains output + returns_ok = { + "Code": 0, + "Stdout": "1 2", + "Stderr": "" + } + self._run.return_value = returns_ok + p_id_list = generic_utils.get_process_id_list( + "ceph-osd/0", + "ceph-osd", + False, + pgrep_full=True + ) + expected = ["1", "2"] + cmd = 'pgrep -f "ceph-osd" || exit 0 && exit 1' + self.assertEqual(p_id_list, expected) + self._run.assert_called_once_with(unit_name="ceph-osd/0", + command=cmd) + + # Return code is not OK + returns_nok = { + "Code": 1, + "Stdout": "", + "Stderr": "Something went wrong" + } + self._run.return_value = returns_nok + with self.assertRaises(zaza_exceptions.ProcessIdsFailed): + generic_utils.get_process_id_list("ceph-osd/0", "ceph") + cmd = 'pidof -x "ceph"' + self._run.assert_called_once_with(unit_name="ceph-osd/0", + command=cmd) + + def test_get_unit_process_ids(self): + self.patch( + "zaza.utilities.generic.get_process_id_list", + new_callable=mock.MagicMock(), + name="_get_pids" + ) + + pids = ["1", "2"] + self._get_pids.return_value = pids + unit_processes = { + "ceph-osd/0": { + "ceph-osd": 2 + }, + "unit/0": { + "pr1": 2, + "pr2": 2 + } + } + expected = { + "ceph-osd/0": { + "ceph-osd": ["1", "2"] + }, + "unit/0": { + "pr1": ["1", "2"], + "pr2": ["1", "2"] + } + } + result = generic_utils.get_unit_process_ids(unit_processes) + self.assertEqual(result, expected) + + def test_validate_unit_process_ids(self): + expected = { + "ceph-osd/0": { + "ceph-osd": 2 + }, + "unit/0": { + "pr1": 2, + "pr2": [1, 2] + } + } + + # Unit count mismatch + actual = {} + with self.assertRaises(zaza_exceptions.UnitCountMismatch): + generic_utils.validate_unit_process_ids(expected, actual) + + # Unit not found in actual dict + actual = { + "ceph-osd/0": { + "ceph-osd": ["1", "2"] + }, + # unit/0 not in the dict + "unit/1": { + "pr1": ["1", "2"], + "pr2": ["1", "2"] + } + } + with self.assertRaises(zaza_exceptions.UnitNotFound): + generic_utils.validate_unit_process_ids(expected, actual) + + # Process names count doesn't match + actual = { + "ceph-osd/0": { + "ceph-osd": ["1", "2"] + }, + "unit/0": { + # Only one process name instead of 2 expected + "pr1": ["1", "2"] + } + } + with self.assertRaises(zaza_exceptions.ProcessNameCountMismatch): + generic_utils.validate_unit_process_ids(expected, actual) + + # Process name doesn't match + actual = { + "ceph-osd/0": { + "ceph-osd": ["1", "2"] + }, + "unit/0": { + # Bad process name + "bad_name": ["1", "2"], + "pr2": ["1", "2"] + } + } + with self.assertRaises(zaza_exceptions.ProcessNameMismatch): + generic_utils.validate_unit_process_ids(expected, actual) + + # PID count doesn't match + actual = { + "ceph-osd/0": { + "ceph-osd": ["1", "2"] + }, + "unit/0": { + # Only one PID instead of 2 expected + "pr1": ["2"], + "pr2": ["1", "2"] + } + } + with self.assertRaises(zaza_exceptions.PIDCountMismatch): + generic_utils.validate_unit_process_ids(expected, actual) + + actual = { + "ceph-osd/0": { + "ceph-osd": ["1", "2"] + }, + "unit/0": { + "pr1": ["1", "2"], + # 3 PID instead of [1, 2] expected + "pr2": ["1", "2", "3"] + } + } + with self.assertRaises(zaza_exceptions.PIDCountMismatch): + generic_utils.validate_unit_process_ids(expected, actual) + + # It should work now... + actual = { + "ceph-osd/0": { + "ceph-osd": ["1", "2"] + }, + "unit/0": { + "pr1": ["1", "2"], + "pr2": ["1", "2"] + } + } + ret = generic_utils.validate_unit_process_ids(expected, actual) + self.assertTrue(ret) + + def test_check_call(self): + + async def async_check_output(*args, **kwargs): + return "hello" + + self.patch_object( + generic_utils, 'check_output', side_effect=async_check_output) + check_call = zaza.sync_wrapper(generic_utils.check_call) + check_call("a command") + self.check_output.assert_called_once_with( + 'a command', log_stdout=True, log_stderr=True) + self.check_output.reset_mock() + check_call("b command", log_stdout=False) + self.check_output.assert_called_once_with( + 'b command', log_stdout=False, log_stderr=True) + self.check_output.reset_mock() + check_call("c command", log_stderr=False) + self.check_output.assert_called_once_with( + 'c command', log_stdout=True, log_stderr=False) + + def test_check_output(self): + self.patch_object(generic_utils, 'logging', name='mock_logging') + + async def mock_communicate(): + return (b"output log", b"error log") + + mock_proc = mock.Mock( + communicate=mock_communicate, + returncode=0) + + async def mock_create_subprocess_exec(*args, **kwargs): + return mock_proc + + self.patch_object(generic_utils.asyncio, 'create_subprocess_exec', + side_effect=mock_create_subprocess_exec) + check_call = zaza.sync_wrapper(generic_utils.check_output) + expected = { + 'Code': str(mock_proc.returncode), + 'Stderr': "error log", + 'Stdout': "output log", + } + self.assertEqual(check_call(['a', 'command']), expected) + # check for raising an error + mock_proc.returncode = 5 + expected = { + 'Code': str(mock_proc.returncode), + 'Stderr': "error log", + 'Stdout': "output log", + } + self.subprocess.CalledProcessError = subprocess.CalledProcessError + try: + check_call(['a', 'command']) + except subprocess.CalledProcessError: + pass diff --git a/unit_tests/utilities/test_zaza_utilities_generic.py b/unit_tests/utilities/test_zaza_utilities_generic.py index 5d3fb7759..038d53e0b 100644 --- a/unit_tests/utilities/test_zaza_utilities_generic.py +++ b/unit_tests/utilities/test_zaza_utilities_generic.py @@ -254,7 +254,7 @@ def test_series_upgrade(self): _unit, _machine_num, origin=_origin, to_series=_to_series, from_series=_from_series, workaround_script=_workaround_script, files=_files) - self.block_until_all_units_idle.called_with() + self.block_until_all_units_idle.assert_called_with() self.prepare_series_upgrade.assert_called_once_with( _machine_num, to_series=_to_series) self.wrap_do_release_upgrade.assert_called_once_with( diff --git a/zaza/model.py b/zaza/model.py index e5539e15e..d05a1d7d6 100644 --- a/zaza/model.py +++ b/zaza/model.py @@ -545,6 +545,9 @@ def _normalise_action_results(results): results[old_key] = results[key] elif results.get(old_key) and not results.get(key): results[key] = results[old_key] + if 'return-code' in results: + results['Code'] = results.get('return-code') + del results['return-code'] return results else: return {} @@ -567,7 +570,10 @@ async def async_run_on_unit(unit_name, command, model_name=None, timeout=None): model = await get_model(model_name) unit = await async_get_unit_from_name(unit_name, model) action = await unit.run(command, timeout=timeout) + await action.wait() results = action.data.get('results') + if not results: + results = action.results return _normalise_action_results(results) run_on_unit = sync_wrapper(async_run_on_unit) @@ -593,7 +599,10 @@ async def async_run_on_leader(application_name, command, model_name=None, is_leader = await unit.is_leader_from_status() if is_leader: action = await unit.run(command, timeout=timeout) + await action.wait() results = action.data.get('results') + if not results: + results = action.results return _normalise_action_results(results) run_on_leader = sync_wrapper(async_run_on_leader) @@ -2049,7 +2058,13 @@ async def _check_file(): for unit in units: try: output = await unit.run('cat {}'.format(remote_file)) - contents = output.data.get('results').get('Stdout', '') + await output.wait() + results = output.data.get('results') + if not results: + results = output.results + contents = results.get('Stdout', '') + if not contents: + contents = results.get('stdout', '') if inspect.iscoroutinefunction(check_function): if not await check_function(contents): return False @@ -2184,7 +2199,13 @@ async def _check_for_file(model): for unit in units: try: output = await unit.run('test -e "{}"; echo $?'.format(path)) - contents = output.data.get('results')['Stdout'] + await output.wait() + output_result = output.data.get('results') + if not output_result: + output_result = output.results + contents = output_result.get('Stdout', '') + if not contents: + contents = output_result.get('stdout', '') results.append("1" in contents) # libjuju throws a generic error for connection failure. So we # cannot differentiate between a connectivity issue and a