diff options
author | Zuul <zuul@review.openstack.org> | 2018-07-25 12:58:24 +0000 |
---|---|---|
committer | Gerrit Code Review <review@openstack.org> | 2018-07-25 12:58:24 +0000 |
commit | 0d9d8264b19c6108c8093f3a47803dfc70d6bae6 (patch) | |
tree | fba66c06ff3a8f7aaf5b586626e3059853ca1e28 | |
parent | 191527ce8b4a994fbdae329e4a314ea7a28cbd01 (diff) | |
parent | 851b2621b6219d45cf8d2786d064e25e8fd49e8c (diff) | |
download | python-saharaclient-0d9d8264b19c6108c8093f3a47803dfc70d6bae6.tar.gz |
Merge "Adding boot from volume to osc"
25 files changed, 1125 insertions, 323 deletions
diff --git a/lower-constraints.txt b/lower-constraints.txt index 090f387..ce64b90 100644 --- a/lower-constraints.txt +++ b/lower-constraints.txt @@ -34,7 +34,7 @@ netifaces==0.10.4 openstacksdk==0.11.2 os-client-config==1.28.0 os-service-types==1.2.0 -osc-lib==1.8.0 +osc-lib==1.11.0 oslo.config==5.2.0 oslo.context==2.19.2 oslo.i18n==3.15.3 diff --git a/requirements.txt b/requirements.txt index 3701928..087b802 100644 --- a/requirements.txt +++ b/requirements.txt @@ -6,7 +6,7 @@ pbr!=2.1.0,>=2.0.0 # Apache-2.0 Babel!=2.4.0,>=2.3.4 # BSD keystoneauth1>=3.4.0 # Apache-2.0 -osc-lib>=1.8.0 # Apache-2.0 +osc-lib>=1.11.0 # Apache-2.0 oslo.log>=3.36.0 # Apache-2.0 oslo.serialization!=2.19.1,>=2.18.0 # Apache-2.0 oslo.i18n>=3.15.3 # Apache-2.0 diff --git a/saharaclient/osc/plugin.py b/saharaclient/osc/plugin.py index 52fb5cc..14f2b48 100644 --- a/saharaclient/osc/plugin.py +++ b/saharaclient/osc/plugin.py @@ -24,7 +24,8 @@ DEFAULT_DATA_PROCESSING_API_VERSION = "1.1" API_VERSION_OPTION = "os_data_processing_api_version" API_NAME = "data_processing" API_VERSIONS = { - "1.1": "saharaclient.api.client.Client" + "1.1": "saharaclient.api.client.Client", + "2": "saharaclient.api.client.ClientV2" } diff --git a/saharaclient/osc/utils.py b/saharaclient/osc/utils.py new file mode 100644 index 0000000..aebb628 --- /dev/null +++ b/saharaclient/osc/utils.py @@ -0,0 +1,373 @@ +# Copyright (c) 2015 Mirantis Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or +# implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import sys +import time + +from osc_lib import exceptions +from osc_lib import utils as osc_utils +from oslo_serialization import jsonutils as json +from oslo_utils import timeutils +from oslo_utils import uuidutils + +from saharaclient.api import base + + +def get_resource(manager, name_or_id, **kwargs): + if uuidutils.is_uuid_like(name_or_id): + return manager.get(name_or_id, **kwargs) + else: + resource = manager.find_unique(name=name_or_id) + if kwargs: + # we really need additional call to apply kwargs + resource = manager.get(resource.id, **kwargs) + return resource + + +def created_at_sorted(objs, reverse=False): + return sorted(objs, key=created_at_key, reverse=reverse) + + +def random_name(prefix=None): + return "%s-%s" % (prefix, uuidutils.generate_uuid()[:8]) + + +def created_at_key(obj): + return timeutils.parse_isotime(obj["created_at"]) + + +def get_resource_id(manager, name_or_id): + if uuidutils.is_uuid_like(name_or_id): + return name_or_id + else: + return manager.find_unique(name=name_or_id).id + + +def create_dict_from_kwargs(**kwargs): + return {k: v for (k, v) in kwargs.items() if v is not None} + + +def prepare_data(data, fields): + new_data = {} + for f in fields: + if f in data: + new_data[f.replace('_', ' ').capitalize()] = data[f] + + return new_data + + +def unzip(data): + return zip(*data) + + +def extend_columns(columns, items): + return unzip(list(unzip(columns)) + [('', '')] + items) + + +def prepare_column_headers(columns, remap=None): + remap = remap if remap else {} + new_columns = [] + for c in columns: + for old, new in remap.items(): + c = c.replace(old, new) + new_columns.append(c.replace('_', ' ').capitalize()) + + return new_columns + + +def get_by_name_substring(data, name): + return [obj for obj in data if name in obj.name] + + +def wait_for_delete(manager, obj_id, sleep_time=5, timeout=3000): + s_time = timeutils.utcnow() + while timeutils.delta_seconds(s_time, timeutils.utcnow()) < timeout: + try: + manager.get(obj_id) + except base.APIException as ex: + if ex.error_code == 404: + return True + raise + time.sleep(sleep_time) + + return False + + +def create_node_group_templates(client, app, parsed_args, flavor_id, configs, + shares): + if app.api_version['data_processing'] == '2': + data = client.node_group_templates.create( + name=parsed_args.name, + plugin_name=parsed_args.plugin, + plugin_version=parsed_args.plugin_version, + flavor_id=flavor_id, + description=parsed_args.description, + volumes_per_node=parsed_args.volumes_per_node, + volumes_size=parsed_args.volumes_size, + node_processes=parsed_args.processes, + floating_ip_pool=parsed_args.floating_ip_pool, + security_groups=parsed_args.security_groups, + auto_security_group=parsed_args.auto_security_group, + availability_zone=parsed_args.availability_zone, + volume_type=parsed_args.volumes_type, + is_proxy_gateway=parsed_args.proxy_gateway, + volume_local_to_instance=parsed_args.volumes_locality, + use_autoconfig=parsed_args.autoconfig, + is_public=parsed_args.public, + is_protected=parsed_args.protected, + node_configs=configs, + shares=shares, + volumes_availability_zone=( + parsed_args.volumes_availability_zone), + volume_mount_prefix=parsed_args.volumes_mount_prefix, + boot_from_volume=parsed_args.boot_from_volume).to_dict() + else: + data = client.node_group_templates.create( + name=parsed_args.name, + plugin_name=parsed_args.plugin, + hadoop_version=parsed_args.plugin_version, + flavor_id=flavor_id, + description=parsed_args.description, + volumes_per_node=parsed_args.volumes_per_node, + volumes_size=parsed_args.volumes_size, + node_processes=parsed_args.processes, + floating_ip_pool=parsed_args.floating_ip_pool, + security_groups=parsed_args.security_groups, + auto_security_group=parsed_args.auto_security_group, + availability_zone=parsed_args.availability_zone, + volume_type=parsed_args.volumes_type, + is_proxy_gateway=parsed_args.proxy_gateway, + volume_local_to_instance=parsed_args.volumes_locality, + use_autoconfig=parsed_args.autoconfig, + is_public=parsed_args.public, + is_protected=parsed_args.protected, + node_configs=configs, + shares=shares, + volumes_availability_zone=( + parsed_args.volumes_availability_zone), + volume_mount_prefix=parsed_args.volumes_mount_prefix).to_dict() + return data + + +class NodeGroupTemplatesUtils(object): + + def _create_take_action(self, client, app, parsed_args): + if parsed_args.json: + blob = osc_utils.read_blob_file_contents(parsed_args.json) + try: + template = json.loads(blob) + except ValueError as e: + raise exceptions.CommandError( + 'An error occurred when reading ' + 'template from file %s: %s' % (parsed_args.json, e)) + data = client.node_group_templates.create(**template).to_dict() + else: + if (not parsed_args.name or not parsed_args.plugin or + not parsed_args.plugin_version or not parsed_args.flavor or + not parsed_args.processes): + raise exceptions.CommandError( + 'At least --name, --plugin, --plugin-version, --processes,' + ' --flavor arguments should be specified or json template ' + 'should be provided with --json argument') + + configs = None + if parsed_args.configs: + blob = osc_utils.read_blob_file_contents(parsed_args.configs) + try: + configs = json.loads(blob) + except ValueError as e: + raise exceptions.CommandError( + 'An error occurred when reading ' + 'configs from file %s: %s' % (parsed_args.configs, e)) + + shares = None + if parsed_args.shares: + blob = osc_utils.read_blob_file_contents(parsed_args.shares) + try: + shares = json.loads(blob) + except ValueError as e: + raise exceptions.CommandError( + 'An error occurred when reading ' + 'shares from file %s: %s' % (parsed_args.shares, e)) + + compute_client = app.client_manager.compute + flavor_id = osc_utils.find_resource( + compute_client.flavors, parsed_args.flavor).id + + data = create_node_group_templates(client, app, parsed_args, + flavor_id, configs, shares) + + return data + + def _list_take_action(self, client, app, parsed_args): + search_opts = {} + if parsed_args.plugin: + search_opts['plugin_name'] = parsed_args.plugin + if parsed_args.plugin_version: + search_opts['hadoop_version'] = parsed_args.plugin_version + + data = client.node_group_templates.list(search_opts=search_opts) + + if parsed_args.name: + data = get_by_name_substring(data, parsed_args.name) + + if app.api_version['data_processing'] == '2': + if parsed_args.long: + columns = ('name', 'id', 'plugin_name', 'plugin_version', + 'node_processes', 'description') + column_headers = prepare_column_headers(columns) + + else: + columns = ('name', 'id', 'plugin_name', 'plugin_version') + column_headers = prepare_column_headers(columns) + else: + if parsed_args.long: + columns = ('name', 'id', 'plugin_name', 'hadoop_version', + 'node_processes', 'description') + column_headers = prepare_column_headers( + columns, {'hadoop_version': 'plugin_version'}) + + else: + columns = ('name', 'id', 'plugin_name', 'hadoop_version') + column_headers = prepare_column_headers( + columns, {'hadoop_version': 'plugin_version'}) + + return ( + column_headers, + (osc_utils.get_item_properties( + s, + columns, + formatters={ + 'node_processes': osc_utils.format_list + } + ) for s in data) + ) + + def _update_take_action(self, client, app, parsed_args): + ngt_id = get_resource_id( + client.node_group_templates, parsed_args.node_group_template) + + if parsed_args.json: + blob = osc_utils.read_blob_file_contents(parsed_args.json) + try: + template = json.loads(blob) + except ValueError as e: + raise exceptions.CommandError( + 'An error occurred when reading ' + 'template from file %s: %s' % (parsed_args.json, e)) + data = client.node_group_templates.update( + ngt_id, **template).to_dict() + else: + configs = None + if parsed_args.configs: + blob = osc_utils.read_blob_file_contents(parsed_args.configs) + try: + configs = json.loads(blob) + except ValueError as e: + raise exceptions.CommandError( + 'An error occurred when reading ' + 'configs from file %s: %s' % (parsed_args.configs, e)) + + shares = None + if parsed_args.shares: + blob = osc_utils.read_blob_file_contents(parsed_args.shares) + try: + shares = json.loads(blob) + except ValueError as e: + raise exceptions.CommandError( + 'An error occurred when reading ' + 'shares from file %s: %s' % (parsed_args.shares, e)) + + flavor_id = None + if parsed_args.flavor: + compute_client = self.app.client_manager.compute + flavor_id = osc_utils.find_resource( + compute_client.flavors, parsed_args.flavor).id + + update_dict = create_dict_from_kwargs( + name=parsed_args.name, + plugin_name=parsed_args.plugin, + hadoop_version=parsed_args.plugin_version, + flavor_id=flavor_id, + description=parsed_args.description, + volumes_per_node=parsed_args.volumes_per_node, + volumes_size=parsed_args.volumes_size, + node_processes=parsed_args.processes, + floating_ip_pool=parsed_args.floating_ip_pool, + security_groups=parsed_args.security_groups, + auto_security_group=parsed_args.use_auto_security_group, + availability_zone=parsed_args.availability_zone, + volume_type=parsed_args.volumes_type, + is_proxy_gateway=parsed_args.is_proxy_gateway, + volume_local_to_instance=parsed_args.volume_locality, + use_autoconfig=parsed_args.use_autoconfig, + is_public=parsed_args.is_public, + is_protected=parsed_args.is_protected, + node_configs=configs, + shares=shares, + volumes_availability_zone=( + parsed_args.volumes_availability_zone), + volume_mount_prefix=parsed_args.volumes_mount_prefix + ) + + if app.api_version['data_processing'] == '2': + if 'hadoop_version' in update_dict: + update_dict.pop('hadoop_version') + update_dict['plugin_version'] = parsed_args.plugin_version + if parsed_args.boot_from_volume is not None: + update_dict['boot_from_volume'] = ( + parsed_args.boot_from_volume) + data = client.node_group_templates.update( + ngt_id, **update_dict).to_dict() + + return data + + def _import_take_action(self, client, parsed_args): + if (not parsed_args.image_id or + not parsed_args.flavor_id): + raise exceptions.CommandError( + 'At least --image_id and --flavor_id should be specified') + blob = osc_utils.read_blob_file_contents(parsed_args.json) + try: + template = json.loads(blob) + except ValueError as e: + raise exceptions.CommandError( + 'An error occurred when reading ' + 'template from file %s: %s' % (parsed_args.json, e)) + template['node_group_template']['floating_ip_pool'] = ( + parsed_args.floating_ip_pool) + template['node_group_template']['image_id'] = ( + parsed_args.image_id) + template['node_group_template']['flavor_id'] = ( + parsed_args.flavor_id) + template['node_group_template']['security_groups'] = ( + parsed_args.security_groups) + if parsed_args.name: + template['node_group_template']['name'] = parsed_args.name + data = client.node_group_templates.create( + **template['node_group_template']).to_dict() + + return data + + def _export_take_action(self, client, parsed_args): + ngt_id = get_resource_id( + client.node_group_templates, parsed_args.node_group_template) + response = client.node_group_templates.export(ngt_id) + result = json.dumps(response._info, indent=4)+"\n" + if parsed_args.file: + with open(parsed_args.file, "w+") as file: + file.write(result) + else: + sys.stdout.write(result) diff --git a/saharaclient/osc/v1/cluster_templates.py b/saharaclient/osc/v1/cluster_templates.py index 6441c17..1442415 100644 --- a/saharaclient/osc/v1/cluster_templates.py +++ b/saharaclient/osc/v1/cluster_templates.py @@ -21,7 +21,7 @@ from osc_lib import utils as osc_utils from oslo_log import log as logging from oslo_serialization import jsonutils as json -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils CT_FIELDS = ['id', 'name', 'plugin_name', 'plugin_version', 'description', 'node_groups', 'anti_affinity', 'use_autoconfig', 'is_default', diff --git a/saharaclient/osc/v1/clusters.py b/saharaclient/osc/v1/clusters.py index b197c1d..6930bf3 100644 --- a/saharaclient/osc/v1/clusters.py +++ b/saharaclient/osc/v1/clusters.py @@ -21,7 +21,7 @@ from osc_lib import utils as osc_utils from oslo_log import log as logging from oslo_serialization import jsonutils -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils CLUSTER_FIELDS = ["cluster_template_id", "use_autoconfig", "user_keypair_id", "status", "image", "node_groups", "id", "info", diff --git a/saharaclient/osc/v1/data_sources.py b/saharaclient/osc/v1/data_sources.py index 2c296f0..05372dd 100644 --- a/saharaclient/osc/v1/data_sources.py +++ b/saharaclient/osc/v1/data_sources.py @@ -19,7 +19,7 @@ from osc_lib.command import command from osc_lib import utils as osc_utils from oslo_log import log as logging -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils DATA_SOURCE_FIELDS = ['name', 'id', 'type', 'url', 'description', 'is_public', 'is_protected'] diff --git a/saharaclient/osc/v1/images.py b/saharaclient/osc/v1/images.py index 4cab041..36ce753 100644 --- a/saharaclient/osc/v1/images.py +++ b/saharaclient/osc/v1/images.py @@ -19,7 +19,7 @@ from osc_lib.command import command from osc_lib import utils as osc_utils from oslo_log import log as logging -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils IMAGE_FIELDS = ['name', 'id', 'username', 'tags', 'status', 'description'] diff --git a/saharaclient/osc/v1/job_binaries.py b/saharaclient/osc/v1/job_binaries.py index 530d9c9..8333c27 100644 --- a/saharaclient/osc/v1/job_binaries.py +++ b/saharaclient/osc/v1/job_binaries.py @@ -23,7 +23,7 @@ from oslo_log import log as logging from oslo_serialization import jsonutils from saharaclient.api import base -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils JOB_BINARY_FIELDS = ['name', 'id', 'url', 'description', 'is_public', 'is_protected'] diff --git a/saharaclient/osc/v1/job_templates.py b/saharaclient/osc/v1/job_templates.py index 476a49e..9013563 100644 --- a/saharaclient/osc/v1/job_templates.py +++ b/saharaclient/osc/v1/job_templates.py @@ -21,7 +21,7 @@ from osc_lib import utils as osc_utils from oslo_log import log as logging from oslo_serialization import jsonutils -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils JOB_TEMPLATE_FIELDS = ['name', 'id', 'type', 'mains', 'libs', 'description', 'is_public', 'is_protected'] diff --git a/saharaclient/osc/v1/job_types.py b/saharaclient/osc/v1/job_types.py index 43fc0f6..2b778f3 100644 --- a/saharaclient/osc/v1/job_types.py +++ b/saharaclient/osc/v1/job_types.py @@ -22,8 +22,8 @@ from osc_lib import utils as osc_utils from oslo_log import log as logging from oslo_serialization import jsonutils +from saharaclient.osc import utils from saharaclient.osc.v1.job_templates import JOB_TYPES_CHOICES -from saharaclient.osc.v1 import utils class ListJobTypes(command.Lister): diff --git a/saharaclient/osc/v1/jobs.py b/saharaclient/osc/v1/jobs.py index ec9743d..27f1f88 100644 --- a/saharaclient/osc/v1/jobs.py +++ b/saharaclient/osc/v1/jobs.py @@ -21,7 +21,7 @@ from osc_lib import utils as osc_utils from oslo_log import log as logging from oslo_serialization import jsonutils -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils JOB_FIELDS = ['id', 'job_template_id', 'cluster_id', 'input_id', 'output_id', 'start_time', 'end_time', 'status', 'is_public', 'is_protected', diff --git a/saharaclient/osc/v1/node_group_templates.py b/saharaclient/osc/v1/node_group_templates.py index 9cfe282..590547e 100644 --- a/saharaclient/osc/v1/node_group_templates.py +++ b/saharaclient/osc/v1/node_group_templates.py @@ -16,12 +16,10 @@ import sys from osc_lib.command import command -from osc_lib import exceptions from osc_lib import utils as osc_utils from oslo_log import log as logging -from oslo_serialization import jsonutils as json -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils NGT_FIELDS = ['id', 'name', 'plugin_name', 'plugin_version', 'node_processes', 'description', 'auto_security_group', 'security_groups', @@ -43,7 +41,7 @@ def _format_ngt_output(data): del data['volumes_size'] -class CreateNodeGroupTemplate(command.ShowOne): +class CreateNodeGroupTemplate(command.ShowOne, utils.NodeGroupTemplatesUtils): """Creates node group template""" log = logging.getLogger(__name__ + ".CreateNodeGroupTemplate") @@ -202,73 +200,7 @@ class CreateNodeGroupTemplate(command.ShowOne): self.log.debug("take_action(%s)", parsed_args) client = self.app.client_manager.data_processing - if parsed_args.json: - blob = osc_utils.read_blob_file_contents(parsed_args.json) - try: - template = json.loads(blob) - except ValueError as e: - raise exceptions.CommandError( - 'An error occurred when reading ' - 'template from file %s: %s' % (parsed_args.json, e)) - data = client.node_group_templates.create(**template).to_dict() - else: - if (not parsed_args.name or not parsed_args.plugin or - not parsed_args.plugin_version or not parsed_args.flavor or - not parsed_args.processes): - raise exceptions.CommandError( - 'At least --name, --plugin, --plugin-version, --processes,' - ' --flavor arguments should be specified or json template ' - 'should be provided with --json argument') - - configs = None - if parsed_args.configs: - blob = osc_utils.read_blob_file_contents(parsed_args.configs) - try: - configs = json.loads(blob) - except ValueError as e: - raise exceptions.CommandError( - 'An error occurred when reading ' - 'configs from file %s: %s' % (parsed_args.configs, e)) - - shares = None - if parsed_args.shares: - blob = osc_utils.read_blob_file_contents(parsed_args.shares) - try: - shares = json.loads(blob) - except ValueError as e: - raise exceptions.CommandError( - 'An error occurred when reading ' - 'shares from file %s: %s' % (parsed_args.shares, e)) - - compute_client = self.app.client_manager.compute - flavor_id = osc_utils.find_resource( - compute_client.flavors, parsed_args.flavor).id - - data = client.node_group_templates.create( - name=parsed_args.name, - plugin_name=parsed_args.plugin, - hadoop_version=parsed_args.plugin_version, - flavor_id=flavor_id, - description=parsed_args.description, - volumes_per_node=parsed_args.volumes_per_node, - volumes_size=parsed_args.volumes_size, - node_processes=parsed_args.processes, - floating_ip_pool=parsed_args.floating_ip_pool, - security_groups=parsed_args.security_groups, - auto_security_group=parsed_args.auto_security_group, - availability_zone=parsed_args.availability_zone, - volume_type=parsed_args.volumes_type, - is_proxy_gateway=parsed_args.proxy_gateway, - volume_local_to_instance=parsed_args.volumes_locality, - use_autoconfig=parsed_args.autoconfig, - is_public=parsed_args.public, - is_protected=parsed_args.protected, - node_configs=configs, - shares=shares, - volumes_availability_zone=( - parsed_args.volumes_availability_zone), - volume_mount_prefix=parsed_args.volumes_mount_prefix - ).to_dict() + data = self._create_take_action(client, self.app, parsed_args) _format_ngt_output(data) data = utils.prepare_data(data, NGT_FIELDS) @@ -276,7 +208,7 @@ class CreateNodeGroupTemplate(command.ShowOne): return self.dict2columns(data) -class ListNodeGroupTemplates(command.Lister): +class ListNodeGroupTemplates(command.Lister, utils.NodeGroupTemplatesUtils): """Lists node group templates""" log = logging.getLogger(__name__ + ".ListNodeGroupTemplates") @@ -314,41 +246,10 @@ class ListNodeGroupTemplates(command.Lister): def take_action(self, parsed_args): self.log.debug("take_action(%s)", parsed_args) client = self.app.client_manager.data_processing - search_opts = {} - if parsed_args.plugin: - search_opts['plugin_name'] = parsed_args.plugin - if parsed_args.plugin_version: - search_opts['hadoop_version'] = parsed_args.plugin_version + return self._list_take_action(client, self.app, parsed_args) - data = client.node_group_templates.list(search_opts=search_opts) - if parsed_args.name: - data = utils.get_by_name_substring(data, parsed_args.name) - - if parsed_args.long: - columns = ('name', 'id', 'plugin_name', 'hadoop_version', - 'node_processes', 'description') - column_headers = utils.prepare_column_headers( - columns, {'hadoop_version': 'plugin_version'}) - - else: - columns = ('name', 'id', 'plugin_name', 'hadoop_version') - column_headers = utils.prepare_column_headers( - columns, {'hadoop_version': 'plugin_version'}) - - return ( - column_headers, - (osc_utils.get_item_properties( - s, - columns, - formatters={ - 'node_processes': osc_utils.format_list - } - ) for s in data) - ) - - -class ShowNodeGroupTemplate(command.ShowOne): +class ShowNodeGroupTemplate(command.ShowOne, utils.NodeGroupTemplatesUtils): """Display node group template details""" log = logging.getLogger(__name__ + ".ShowNodeGroupTemplate") @@ -378,7 +279,7 @@ class ShowNodeGroupTemplate(command.ShowOne): return self.dict2columns(data) -class DeleteNodeGroupTemplate(command.Command): +class DeleteNodeGroupTemplate(command.Command, utils.NodeGroupTemplatesUtils): """Deletes node group template""" log = logging.getLogger(__name__ + ".DeleteNodeGroupTemplate") @@ -406,7 +307,7 @@ class DeleteNodeGroupTemplate(command.Command): 'successfully.\n'.format(ngt=ngt)) -class UpdateNodeGroupTemplate(command.ShowOne): +class UpdateNodeGroupTemplate(command.ShowOne, utils.NodeGroupTemplatesUtils): """Updates node group template""" log = logging.getLogger(__name__ + ".UpdateNodeGroupTemplate") @@ -620,74 +521,7 @@ class UpdateNodeGroupTemplate(command.ShowOne): self.log.debug("take_action(%s)", parsed_args) client = self.app.client_manager.data_processing - ngt_id = utils.get_resource_id( - client.node_group_templates, parsed_args.node_group_template) - - if parsed_args.json: - blob = osc_utils.read_blob_file_contents(parsed_args.json) - try: - template = json.loads(blob) - except ValueError as e: - raise exceptions.CommandError( - 'An error occurred when reading ' - 'template from file %s: %s' % (parsed_args.json, e)) - data = client.node_group_templates.update( - ngt_id, **template).to_dict() - else: - configs = None - if parsed_args.configs: - blob = osc_utils.read_blob_file_contents(parsed_args.configs) - try: - configs = json.loads(blob) - except ValueError as e: - raise exceptions.CommandError( - 'An error occurred when reading ' - 'configs from file %s: %s' % (parsed_args.configs, e)) - - shares = None - if parsed_args.shares: - blob = osc_utils.read_blob_file_contents(parsed_args.shares) - try: - shares = json.loads(blob) - except ValueError as e: - raise exceptions.CommandError( - 'An error occurred when reading ' - 'shares from file %s: %s' % (parsed_args.shares, e)) - - flavor_id = None - if parsed_args.flavor: - compute_client = self.app.client_manager.compute - flavor_id = osc_utils.find_resource( - compute_client.flavors, parsed_args.flavor).id - - update_dict = utils.create_dict_from_kwargs( - name=parsed_args.name, - plugin_name=parsed_args.plugin, - hadoop_version=parsed_args.plugin_version, - flavor_id=flavor_id, - description=parsed_args.description, - volumes_per_node=parsed_args.volumes_per_node, - volumes_size=parsed_args.volumes_size, - node_processes=parsed_args.processes, - floating_ip_pool=parsed_args.floating_ip_pool, - security_groups=parsed_args.security_groups, - auto_security_group=parsed_args.use_auto_security_group, - availability_zone=parsed_args.availability_zone, - volume_type=parsed_args.volumes_type, - is_proxy_gateway=parsed_args.is_proxy_gateway, - volume_local_to_instance=parsed_args.volume_locality, - use_autoconfig=parsed_args.use_autoconfig, - is_public=parsed_args.is_public, - is_protected=parsed_args.is_protected, - node_configs=configs, - shares=shares, - volumes_availability_zone=( - parsed_args.volumes_availability_zone), - volume_mount_prefix=parsed_args.volumes_mount_prefix - ) - - data = client.node_group_templates.update( - ngt_id, **update_dict).to_dict() + data = self._update_take_action(client, self.app, parsed_args) _format_ngt_output(data) data = utils.prepare_data(data, NGT_FIELDS) @@ -695,7 +529,7 @@ class UpdateNodeGroupTemplate(command.ShowOne): return self.dict2columns(data) -class ImportNodeGroupTemplate(command.ShowOne): +class ImportNodeGroupTemplate(command.ShowOne, utils.NodeGroupTemplatesUtils): """Imports node group template""" log = logging.getLogger(__name__ + ".ImportNodeGroupTemplate") @@ -740,29 +574,8 @@ class ImportNodeGroupTemplate(command.ShowOne): def take_action(self, parsed_args): self.log.debug("take_action(%s)", parsed_args) client = self.app.client_manager.data_processing - if (not parsed_args.image_id or - not parsed_args.flavor_id): - raise exceptions.CommandError( - 'At least --image_id and --flavor_id should be specified') - blob = osc_utils.read_blob_file_contents(parsed_args.json) - try: - template = json.loads(blob) - except ValueError as e: - raise exceptions.CommandError( - 'An error occurred when reading ' - 'template from file %s: %s' % (parsed_args.json, e)) - template['node_group_template']['floating_ip_pool'] = ( - parsed_args.floating_ip_pool) - template['node_group_template']['image_id'] = ( - parsed_args.image_id) - template['node_group_template']['flavor_id'] = ( - parsed_args.flavor_id) - template['node_group_template']['security_groups'] = ( - parsed_args.security_groups) - if parsed_args.name: - template['node_group_template']['name'] = parsed_args.name - data = client.node_group_templates.create( - **template['node_group_template']).to_dict() + + data = self._import_take_action(client, parsed_args) _format_ngt_output(data) data = utils.prepare_data(data, NGT_FIELDS) @@ -770,7 +583,7 @@ class ImportNodeGroupTemplate(command.ShowOne): return self.dict2columns(data) -class ExportNodeGroupTemplate(command.Command): +class ExportNodeGroupTemplate(command.Command, utils.NodeGroupTemplatesUtils): """Export node group template to JSON""" log = logging.getLogger(__name__ + ".ExportNodeGroupTemplate") @@ -794,12 +607,4 @@ class ExportNodeGroupTemplate(command.Command): def take_action(self, parsed_args): self.log.debug("take_action(%s)", parsed_args) client = self.app.client_manager.data_processing - ngt_id = utils.get_resource_id( - client.node_group_templates, parsed_args.node_group_template) - response = client.node_group_templates.export(ngt_id) - result = json.dumps(response._info, indent=4)+"\n" - if parsed_args.file: - with open(parsed_args.file, "w+") as file: - file.write(result) - else: - sys.stdout.write(result) + self._export_take_action(client, parsed_args) diff --git a/saharaclient/osc/v1/plugins.py b/saharaclient/osc/v1/plugins.py index 8b10420..ac2d25a 100644 --- a/saharaclient/osc/v1/plugins.py +++ b/saharaclient/osc/v1/plugins.py @@ -22,7 +22,7 @@ from osc_lib import utils as osc_utils from oslo_log import log as logging from oslo_serialization import jsonutils -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils def _serialize_label_items(plugin): diff --git a/saharaclient/osc/v1/utils.py b/saharaclient/osc/v1/utils.py deleted file mode 100644 index 973f385..0000000 --- a/saharaclient/osc/v1/utils.py +++ /dev/null @@ -1,101 +0,0 @@ -# Copyright (c) 2015 Mirantis Inc. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or -# implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -import time - -from oslo_utils import timeutils -from oslo_utils import uuidutils - -from saharaclient.api import base - - -def get_resource(manager, name_or_id, **kwargs): - if uuidutils.is_uuid_like(name_or_id): - return manager.get(name_or_id, **kwargs) - else: - resource = manager.find_unique(name=name_or_id) - if kwargs: - # we really need additional call to apply kwargs - resource = manager.get(resource.id, **kwargs) - return resource - - -def created_at_sorted(objs, reverse=False): - return sorted(objs, key=created_at_key, reverse=reverse) - - -def random_name(prefix=None): - return "%s-%s" % (prefix, uuidutils.generate_uuid()[:8]) - - -def created_at_key(obj): - return timeutils.parse_isotime(obj["created_at"]) - - -def get_resource_id(manager, name_or_id): - if uuidutils.is_uuid_like(name_or_id): - return name_or_id - else: - return manager.find_unique(name=name_or_id).id - - -def create_dict_from_kwargs(**kwargs): - return {k: v for (k, v) in kwargs.items() if v is not None} - - -def prepare_data(data, fields): - new_data = {} - for f in fields: - if f in data: - new_data[f.replace('_', ' ').capitalize()] = data[f] - - return new_data - - -def unzip(data): - return zip(*data) - - -def extend_columns(columns, items): - return unzip(list(unzip(columns)) + [('', '')] + items) - - -def prepare_column_headers(columns, remap=None): - remap = remap if remap else {} - new_columns = [] - for c in columns: - for old, new in remap.items(): - c = c.replace(old, new) - new_columns.append(c.replace('_', ' ').capitalize()) - - return new_columns - - -def get_by_name_substring(data, name): - return [obj for obj in data if name in obj.name] - - -def wait_for_delete(manager, obj_id, sleep_time=5, timeout=3000): - s_time = timeutils.utcnow() - while timeutils.delta_seconds(s_time, timeutils.utcnow()) < timeout: - try: - manager.get(obj_id) - except base.APIException as ex: - if ex.error_code == 404: - return True - raise - time.sleep(sleep_time) - - return False diff --git a/saharaclient/osc/v2/__init__.py b/saharaclient/osc/v2/__init__.py new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/saharaclient/osc/v2/__init__.py diff --git a/saharaclient/osc/v2/node_group_templates.py b/saharaclient/osc/v2/node_group_templates.py new file mode 100644 index 0000000..c2aaf94 --- /dev/null +++ b/saharaclient/osc/v2/node_group_templates.py @@ -0,0 +1,176 @@ +# Copyright (c) 2018 Red Hat Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or +# implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import sys + +from osc_lib import utils as osc_utils + +from saharaclient.osc import utils +from saharaclient.osc.v1 import node_group_templates as ngt_v1 + +NGT_FIELDS = ['id', 'name', 'plugin_name', 'plugin_version', 'node_processes', + 'description', 'auto_security_group', 'security_groups', + 'availability_zone', 'flavor_id', 'floating_ip_pool', + 'volumes_per_node', 'volumes_size', + 'volume_type', 'volume_local_to_instance', 'volume_mount_prefix', + 'volumes_availability_zone', 'use_autoconfig', + 'is_proxy_gateway', 'is_default', 'is_protected', 'is_public', + 'boot_from_volume'] + + +def _format_ngt_output(data): + data['node_processes'] = osc_utils.format_list(data['node_processes']) + if data['volumes_per_node'] == 0: + del data['volume_local_to_instance'] + del data['volume_mount_prefix'] + del data['volume_type'], + del data['volumes_availability_zone'] + del data['volumes_size'] + + +class CreateNodeGroupTemplate(ngt_v1.CreateNodeGroupTemplate, + utils.NodeGroupTemplatesUtils): + """Creates node group template""" + + def get_parser(self, prog_name): + parser = super(CreateNodeGroupTemplate, self).get_parser(prog_name) + + parser.add_argument( + '--boot-from-volume', + action='store_true', + default=False, + help="Make the node group bootable from volume", + ) + return parser + + def take_action(self, parsed_args): + self.log.debug("take_action(%s)", parsed_args) + client = self.app.client_manager.data_processing + + data = self._create_take_action(client, self.app, parsed_args) + + _format_ngt_output(data) + data = utils.prepare_data(data, NGT_FIELDS) + + return self.dict2columns(data) + + +class ListNodeGroupTemplates(ngt_v1.ListNodeGroupTemplates, + utils.NodeGroupTemplatesUtils): + """Lists node group templates""" + + def take_action(self, parsed_args): + self.log.debug("take_action(%s)", parsed_args) + client = self.app.client_manager.data_processing + return self._list_take_action(client, self.app, parsed_args) + + +class ShowNodeGroupTemplate(ngt_v1.ShowNodeGroupTemplate, + utils.NodeGroupTemplatesUtils): + """Display node group template details""" + + def take_action(self, parsed_args): + self.log.debug("take_action(%s)", parsed_args) + client = self.app.client_manager.data_processing + + data = utils.get_resource( + client.node_group_templates, + parsed_args.node_group_template).to_dict() + + _format_ngt_output(data) + + data = utils.prepare_data(data, NGT_FIELDS) + + return self.dict2columns(data) + + +class DeleteNodeGroupTemplate(ngt_v1.DeleteNodeGroupTemplate, + utils.NodeGroupTemplatesUtils): + """Deletes node group template""" + + def take_action(self, parsed_args): + self.log.debug("take_action(%s)", parsed_args) + client = self.app.client_manager.data_processing + for ngt in parsed_args.node_group_template: + ngt_id = utils.get_resource_id( + client.node_group_templates, ngt) + client.node_group_templates.delete(ngt_id) + sys.stdout.write( + 'Node group template "{ngt}" has been removed ' + 'successfully.\n'.format(ngt=ngt)) + + +class UpdateNodeGroupTemplate(ngt_v1.UpdateNodeGroupTemplate, + utils.NodeGroupTemplatesUtils): + """Updates node group template""" + + def get_parser(self, prog_name): + parser = super(UpdateNodeGroupTemplate, self).get_parser(prog_name) + + bootfromvolume = parser.add_mutually_exclusive_group() + bootfromvolume.add_argument( + '--boot-from-volume-enable', + action='store_true', + help='Makes node group bootable from volume.', + dest='boot_from_volume' + ) + bootfromvolume.add_argument( + '--boot-from-volume-disable', + action='store_false', + help='Makes node group not bootable from volume.', + dest='boot_from_volume' + ) + parser.set_defaults(is_public=None, is_protected=None, + is_proxy_gateway=None, volume_locality=None, + use_auto_security_group=None, use_autoconfig=None, + boot_from_volume=None) + return parser + + def take_action(self, parsed_args): + self.log.debug("take_action(%s)", parsed_args) + client = self.app.client_manager.data_processing + + data = self._update_take_action(client, self.app, parsed_args) + + _format_ngt_output(data) + data = utils.prepare_data(data, NGT_FIELDS) + + return self.dict2columns(data) + + +class ImportNodeGroupTemplate(ngt_v1.ImportNodeGroupTemplate, + utils.NodeGroupTemplatesUtils): + """Imports node group template""" + + def take_action(self, parsed_args): + self.log.debug("take_action(%s)", parsed_args) + client = self.app.client_manager.data_processing + + data = self._import_take_action(client, parsed_args) + + _format_ngt_output(data) + data = utils.prepare_data(data, NGT_FIELDS) + + return self.dict2columns(data) + + +class ExportNodeGroupTemplate(ngt_v1.ExportNodeGroupTemplate, + utils.NodeGroupTemplatesUtils): + """Export node group template to JSON""" + + def take_action(self, parsed_args): + self.log.debug("take_action(%s)", parsed_args) + client = self.app.client_manager.data_processing + self._export_take_action(client, parsed_args) diff --git a/saharaclient/tests/unit/base.py b/saharaclient/tests/unit/base.py index ff23616..35f648b 100644 --- a/saharaclient/tests/unit/base.py +++ b/saharaclient/tests/unit/base.py @@ -30,6 +30,8 @@ class BaseTestCase(testtools.TestCase): super(BaseTestCase, self).setUp() self.responses = self.useFixture(fixture.Fixture()) self.client = client.Client(session=self.SESSION, sahara_url=self.URL) + self.client_v2 = client.ClientV2(session=self.SESSION, + sahara_url=self.URL) def assertFields(self, body, obj): for key, value in body.items(): diff --git a/saharaclient/tests/unit/osc/test_plugin.py b/saharaclient/tests/unit/osc/test_plugin.py index 5006c5c..cf5ac7d 100644 --- a/saharaclient/tests/unit/osc/test_plugin.py +++ b/saharaclient/tests/unit/osc/test_plugin.py @@ -36,3 +36,19 @@ class TestDataProcessingPlugin(base.BaseTestCase): region_name='region_name', sahara_url='url', endpoint_type='public') + + @mock.patch("saharaclient.api.client.ClientV2") + def test_make_client_v2(self, p_client): + + instance = mock.Mock() + instance._api_version = {"data_processing": '2'} + instance.session = 'session' + instance._region_name = 'region_name' + instance._cli_options.data_processing_url = 'url' + instance._interface = 'public' + + plugin.make_client(instance) + p_client.assert_called_with(session='session', + region_name='region_name', + sahara_url='url', + endpoint_type='public') diff --git a/saharaclient/tests/unit/osc/v1/test_node_group_templates.py b/saharaclient/tests/unit/osc/v1/test_node_group_templates.py index b23a4e4..10dc5aa 100644 --- a/saharaclient/tests/unit/osc/v1/test_node_group_templates.py +++ b/saharaclient/tests/unit/osc/v1/test_node_group_templates.py @@ -59,6 +59,7 @@ class TestNodeGroupTemplates(fakes.TestDataProcessing): self.ngt_mock = ( self.app.client_manager.data_processing.node_group_templates) self.ngt_mock.reset_mock() + self.app.api_version['data_processing'] = '1' class TestCreateNodeGroupTemplate(TestNodeGroupTemplates): diff --git a/saharaclient/tests/unit/osc/v1/test_utils.py b/saharaclient/tests/unit/osc/v1/test_utils.py index 7a9ccf1..59ddbdc 100644 --- a/saharaclient/tests/unit/osc/v1/test_utils.py +++ b/saharaclient/tests/unit/osc/v1/test_utils.py @@ -15,7 +15,7 @@ import mock -from saharaclient.osc.v1 import utils +from saharaclient.osc import utils from saharaclient.tests.unit import base diff --git a/saharaclient/tests/unit/osc/v2/__init__.py b/saharaclient/tests/unit/osc/v2/__init__.py new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/saharaclient/tests/unit/osc/v2/__init__.py diff --git a/saharaclient/tests/unit/osc/v2/test_node_group_templates.py b/saharaclient/tests/unit/osc/v2/test_node_group_templates.py new file mode 100644 index 0000000..df174dd --- /dev/null +++ b/saharaclient/tests/unit/osc/v2/test_node_group_templates.py @@ -0,0 +1,412 @@ +# Copyright (c) 2015 Mirantis Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or +# implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import mock +from osc_lib.tests import utils as osc_utils + +from saharaclient.api import node_group_templates as api_ngt +from saharaclient.osc.v2 import node_group_templates as osc_ngt +from saharaclient.tests.unit.osc.v1 import fakes + + +NGT_INFO = { + "node_processes": [ + "namenode", + "tasktracker" + ], + "name": "template", + "tenant_id": "tenant_id", + "availability_zone": 'av_zone', + "use_autoconfig": True, + "plugin_version": "0.1", + "shares": None, + "is_default": False, + "description": 'description', + "node_configs": {}, + "is_proxy_gateway": False, + "auto_security_group": True, + "volume_type": None, + "volumes_size": 2, + "volume_mount_prefix": "/volumes/disk", + "plugin_name": "fake", + "is_protected": False, + "security_groups": None, + "floating_ip_pool": "floating_pool", + "is_public": True, + "id": "ng_id", + "flavor_id": "flavor_id", + "volumes_availability_zone": None, + "volumes_per_node": 2, + "volume_local_to_instance": False, + "boot_from_volume": False +} + + +class TestNodeGroupTemplates(fakes.TestDataProcessing): + def setUp(self): + super(TestNodeGroupTemplates, self).setUp() + self.ngt_mock = ( + self.app.client_manager.data_processing.node_group_templates) + self.ngt_mock.reset_mock() + self.app.api_version['data_processing'] = '2' + + +class TestCreateNodeGroupTemplate(TestNodeGroupTemplates): + # TODO(apavlov): check for creation with --json + def setUp(self): + super(TestCreateNodeGroupTemplate, self).setUp() + self.ngt_mock.create.return_value = api_ngt.NodeGroupTemplate( + None, NGT_INFO) + + self.fl_mock = self.app.client_manager.compute.flavors + self.fl_mock.get.return_value = mock.Mock(id='flavor_id') + self.fl_mock.reset_mock() + + # Command to test + self.cmd = osc_ngt.CreateNodeGroupTemplate(self.app, None) + + def test_ngt_create_minimum_options(self): + arglist = ['--name', 'template', '--plugin', 'fake', + '--plugin-version', '0.1', '--processes', 'namenode', + 'tasktracker', '--flavor', 'flavor_id'] + verifylist = [('name', 'template'), ('plugin', 'fake'), + ('plugin_version', '0.1'), ('flavor', 'flavor_id'), + ('processes', ['namenode', 'tasktracker'])] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + self.cmd.take_action(parsed_args) + + # Check that correct arguments were passed + self.ngt_mock.create.assert_called_once_with( + auto_security_group=False, availability_zone=None, + description=None, flavor_id='flavor_id', floating_ip_pool=None, + plugin_version='0.1', is_protected=False, is_proxy_gateway=False, + is_public=False, name='template', + node_processes=['namenode', 'tasktracker'], plugin_name='fake', + security_groups=None, use_autoconfig=False, + volume_local_to_instance=False, + volume_type=None, volumes_availability_zone=None, + volumes_per_node=None, volumes_size=None, shares=None, + node_configs=None, volume_mount_prefix=None, + boot_from_volume=False) + + def test_ngt_create_all_options(self): + arglist = ['--name', 'template', '--plugin', 'fake', + '--plugin-version', '0.1', '--processes', 'namenode', + 'tasktracker', '--security-groups', 'secgr', + '--auto-security-group', '--availability-zone', 'av_zone', + '--flavor', 'flavor_id', '--floating-ip-pool', + 'floating_pool', '--volumes-per-node', + '2', '--volumes-size', '2', '--volumes-type', 'type', + '--volumes-availability-zone', 'vavzone', + '--volumes-mount-prefix', '/volume/asd', + '--volumes-locality', '--description', 'descr', + '--autoconfig', '--proxy-gateway', '--public', + '--protected', '--boot-from-volume'] + + verifylist = [('name', 'template'), ('plugin', 'fake'), + ('plugin_version', '0.1'), + ('processes', ['namenode', 'tasktracker']), + ('security_groups', ['secgr']), + ('auto_security_group', True), + ('availability_zone', 'av_zone'), + ('flavor', 'flavor_id'), + ('floating_ip_pool', 'floating_pool'), + ('volumes_per_node', 2), ('volumes_size', 2), + ('volumes_type', 'type'), + ('volumes_availability_zone', 'vavzone'), + ('volumes_mount_prefix', '/volume/asd'), + ('volumes_locality', True), ('description', 'descr'), + ('autoconfig', True), ('proxy_gateway', True), + ('public', True), ('protected', True), + ('boot_from_volume', True)] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + columns, data = self.cmd.take_action(parsed_args) + + # Check that correct arguments were passed + self.ngt_mock.create.assert_called_once_with( + auto_security_group=True, availability_zone='av_zone', + description='descr', flavor_id='flavor_id', + floating_ip_pool='floating_pool', plugin_version='0.1', + is_protected=True, is_proxy_gateway=True, is_public=True, + name='template', node_processes=['namenode', 'tasktracker'], + plugin_name='fake', security_groups=['secgr'], use_autoconfig=True, + volume_local_to_instance=True, volume_type='type', + volumes_availability_zone='vavzone', volumes_per_node=2, + volumes_size=2, shares=None, node_configs=None, + volume_mount_prefix='/volume/asd', boot_from_volume=True) + + # Check that columns are correct + expected_columns = ( + 'Auto security group', 'Availability zone', 'Boot from volume', + 'Description', 'Flavor id', 'Floating ip pool', 'Id', + 'Is default', 'Is protected', 'Is proxy gateway', 'Is public', + 'Name', 'Node processes', 'Plugin name', 'Plugin version', + 'Security groups', 'Use autoconfig', 'Volume local to instance', + 'Volume mount prefix', 'Volume type', 'Volumes availability zone', + 'Volumes per node', 'Volumes size') + self.assertEqual(expected_columns, columns) + + # Check that data is correct + expected_data = ( + True, 'av_zone', False, 'description', 'flavor_id', + 'floating_pool', 'ng_id', False, False, False, True, + 'template', 'namenode, tasktracker', 'fake', '0.1', None, True, + False, '/volumes/disk', None, None, 2, 2) + self.assertEqual(expected_data, data) + + +class TestListNodeGroupTemplates(TestNodeGroupTemplates): + def setUp(self): + super(TestListNodeGroupTemplates, self).setUp() + self.ngt_mock.list.return_value = [api_ngt.NodeGroupTemplate( + None, NGT_INFO)] + + # Command to test + self.cmd = osc_ngt.ListNodeGroupTemplates(self.app, None) + + def test_ngt_list_no_options(self): + arglist = [] + verifylist = [] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + columns, data = self.cmd.take_action(parsed_args) + + # Check that columns are correct + expected_columns = ['Name', 'Id', 'Plugin name', 'Plugin version'] + self.assertEqual(expected_columns, columns) + + # Check that data is correct + expected_data = [('template', 'ng_id', 'fake', '0.1')] + self.assertEqual(expected_data, list(data)) + + def test_ngt_list_long(self): + arglist = ['--long'] + verifylist = [('long', True)] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + columns, data = self.cmd.take_action(parsed_args) + + # Check that columns are correct + expected_columns = ['Name', 'Id', 'Plugin name', 'Plugin version', + 'Node processes', 'Description'] + self.assertEqual(expected_columns, columns) + + # Check that data is correct + expected_data = [('template', 'ng_id', 'fake', '0.1', + 'namenode, tasktracker', 'description')] + self.assertEqual(expected_data, list(data)) + + def test_ngt_list_extra_search_opts(self): + arglist = ['--plugin', 'fake', '--plugin-version', '0.1', '--name', + 'templ'] + verifylist = [('plugin', 'fake'), ('plugin_version', '0.1'), + ('name', 'templ')] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + columns, data = self.cmd.take_action(parsed_args) + + # Check that columns are correct + expected_columns = ['Name', 'Id', 'Plugin name', 'Plugin version'] + self.assertEqual(expected_columns, columns) + + # Check that data is correct + expected_data = [('template', 'ng_id', 'fake', '0.1')] + self.assertEqual(expected_data, list(data)) + + +class TestShowNodeGroupTemplate(TestNodeGroupTemplates): + def setUp(self): + super(TestShowNodeGroupTemplate, self).setUp() + self.ngt_mock.find_unique.return_value = api_ngt.NodeGroupTemplate( + None, NGT_INFO) + + # Command to test + self.cmd = osc_ngt.ShowNodeGroupTemplate(self.app, None) + + def test_ngt_show(self): + arglist = ['template'] + verifylist = [('node_group_template', 'template')] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + columns, data = self.cmd.take_action(parsed_args) + + # Check that correct arguments were passed + self.ngt_mock.find_unique.assert_called_once_with(name='template') + + # Check that columns are correct + expected_columns = ( + 'Auto security group', 'Availability zone', 'Boot from volume', + 'Description', 'Flavor id', 'Floating ip pool', 'Id', + 'Is default', 'Is protected', 'Is proxy gateway', 'Is public', + 'Name', 'Node processes', 'Plugin name', 'Plugin version', + 'Security groups', 'Use autoconfig', 'Volume local to instance', + 'Volume mount prefix', 'Volume type', 'Volumes availability zone', + 'Volumes per node', 'Volumes size') + self.assertEqual(expected_columns, columns) + + # Check that data is correct + expected_data = ( + True, 'av_zone', False, 'description', 'flavor_id', + 'floating_pool', 'ng_id', False, False, False, True, + 'template', 'namenode, tasktracker', 'fake', '0.1', None, True, + False, '/volumes/disk', None, None, 2, 2) + self.assertEqual(expected_data, data) + + +class TestDeleteNodeGroupTemplate(TestNodeGroupTemplates): + def setUp(self): + super(TestDeleteNodeGroupTemplate, self).setUp() + self.ngt_mock.find_unique.return_value = api_ngt.NodeGroupTemplate( + None, NGT_INFO) + + # Command to test + self.cmd = osc_ngt.DeleteNodeGroupTemplate(self.app, None) + + def test_ngt_delete(self): + arglist = ['template'] + verifylist = [('node_group_template', ['template'])] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + self.cmd.take_action(parsed_args) + + # Check that correct arguments were passed + self.ngt_mock.delete.assert_called_once_with('ng_id') + + +class TestUpdateNodeGroupTemplate(TestNodeGroupTemplates): + # TODO(apavlov): check for update with --json + def setUp(self): + super(TestUpdateNodeGroupTemplate, self).setUp() + self.ngt_mock.find_unique.return_value = api_ngt.NodeGroupTemplate( + None, NGT_INFO) + self.ngt_mock.update.return_value = api_ngt.NodeGroupTemplate( + None, NGT_INFO) + + self.fl_mock = self.app.client_manager.compute.flavors + self.fl_mock.get.return_value = mock.Mock(id='flavor_id') + self.fl_mock.reset_mock() + + # Command to test + self.cmd = osc_ngt.UpdateNodeGroupTemplate(self.app, None) + + def test_ngt_update_no_options(self): + arglist = [] + verifylist = [] + + self.assertRaises(osc_utils.ParserException, self.check_parser, + self.cmd, arglist, verifylist) + + def test_ngt_update_nothing_updated(self): + arglist = ['template'] + verifylist = [('node_group_template', 'template')] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + self.cmd.take_action(parsed_args) + + # Check that correct arguments were passed + self.ngt_mock.update.assert_called_once_with('ng_id') + + def test_ngt_update_all_options(self): + arglist = ['template', '--name', 'template', '--plugin', 'fake', + '--plugin-version', '0.1', '--processes', 'namenode', + 'tasktracker', '--security-groups', 'secgr', + '--auto-security-group-enable', + '--availability-zone', 'av_zone', '--flavor', 'flavor_id', + '--floating-ip-pool', 'floating_pool', '--volumes-per-node', + '2', '--volumes-size', '2', '--volumes-type', 'type', + '--volumes-availability-zone', 'vavzone', + '--volumes-mount-prefix', '/volume/asd', + '--volumes-locality-enable', '--description', 'descr', + '--autoconfig-enable', '--proxy-gateway-enable', '--public', + '--protected', '--boot-from-volume-enable'] + + verifylist = [('node_group_template', 'template'), + ('name', 'template'), ('plugin', 'fake'), + ('plugin_version', '0.1'), + ('processes', ['namenode', 'tasktracker']), + ('security_groups', ['secgr']), + ('use_auto_security_group', True), + ('availability_zone', 'av_zone'), + ('flavor', 'flavor_id'), + ('floating_ip_pool', 'floating_pool'), + ('volumes_per_node', 2), ('volumes_size', 2), + ('volumes_type', 'type'), + ('volumes_availability_zone', 'vavzone'), + ('volumes_mount_prefix', '/volume/asd'), + ('volume_locality', True), + ('description', 'descr'), ('use_autoconfig', True), + ('is_proxy_gateway', True), + ('is_public', True), ('is_protected', True), + ('boot_from_volume', True)] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + columns, data = self.cmd.take_action(parsed_args) + + # Check that correct arguments were passed + self.ngt_mock.update.assert_called_once_with( + 'ng_id', + auto_security_group=True, availability_zone='av_zone', + description='descr', flavor_id='flavor_id', + floating_ip_pool='floating_pool', plugin_version='0.1', + is_protected=True, is_proxy_gateway=True, is_public=True, + name='template', node_processes=['namenode', 'tasktracker'], + plugin_name='fake', security_groups=['secgr'], use_autoconfig=True, + volume_local_to_instance=True, volume_type='type', + volumes_availability_zone='vavzone', volumes_per_node=2, + volumes_size=2, volume_mount_prefix='/volume/asd', + boot_from_volume=True) + + # Check that columns are correct + expected_columns = ( + 'Auto security group', 'Availability zone', 'Boot from volume', + 'Description', 'Flavor id', 'Floating ip pool', 'Id', + 'Is default', 'Is protected', 'Is proxy gateway', 'Is public', + 'Name', 'Node processes', 'Plugin name', 'Plugin version', + 'Security groups', 'Use autoconfig', 'Volume local to instance', + 'Volume mount prefix', 'Volume type', 'Volumes availability zone', + 'Volumes per node', 'Volumes size') + self.assertEqual(expected_columns, columns) + + # Check that data is correct + expected_data = ( + True, 'av_zone', False, 'description', 'flavor_id', + 'floating_pool', 'ng_id', False, False, False, True, + 'template', 'namenode, tasktracker', 'fake', '0.1', None, True, + False, '/volumes/disk', None, None, 2, 2) + self.assertEqual(expected_data, data) + + def test_ngt_update_private_unprotected(self): + arglist = ['template', '--private', '--unprotected'] + verifylist = [('node_group_template', 'template'), + ('is_public', False), ('is_protected', False)] + + parsed_args = self.check_parser(self.cmd, arglist, verifylist) + + self.cmd.take_action(parsed_args) + + # Check that correct arguments were passed + self.ngt_mock.update.assert_called_once_with( + 'ng_id', is_protected=False, is_public=False) diff --git a/saharaclient/tests/unit/test_node_group_templates.py b/saharaclient/tests/unit/test_node_group_templates.py index f37416a..959bbee 100644 --- a/saharaclient/tests/unit/test_node_group_templates.py +++ b/saharaclient/tests/unit/test_node_group_templates.py @@ -154,3 +154,111 @@ class NodeGroupTemplateTest(base.BaseTestCase): self.assertEqual(url, self.responses.last_request.url) self.assertIsInstance(resp, ng.NodeGroupTemplate) self.assertDictsEqual(self.body, resp.__dict__[u'node_group_template']) + + +class NodeGroupTemplateTestV2(base.BaseTestCase): + body = { + "name": "name", + "plugin_name": "plugin", + "plugin_version": "1", + "flavor_id": "2", + "description": "description", + "volumes_per_node": "3", + "volumes_size": "4", + "node_processes": ["datanode"], + "use_autoconfig": True, + "volume_mount_prefix": '/volumes/disk', + "boot_from_volume": False + } + + update_json = { + "node_group_template": { + "name": "UpdatedName", + "plugin_name": "new_plugin", + "plugin_version": "2", + "flavor_id": "7", + "description": "description", + "volumes_per_node": "3", + "volumes_size": "4", + "node_processes": ["datanode", "namenode"], + "use_autoconfig": False, + "volume_mount_prefix": '/volumes/newdisk', + "boot_from_volume": True + } + } + + def test_create_node_group_template_v2(self): + url = self.URL + '/node-group-templates' + self.responses.post(url, status_code=202, + json={'node_group_template': self.body}) + + resp = self.client_v2.node_group_templates.create(**self.body) + + self.assertEqual(url, self.responses.last_request.url) + self.assertEqual(self.body, + json.loads(self.responses.last_request.body)) + self.assertIsInstance(resp, ng.NodeGroupTemplate) + self.assertFields(self.body, resp) + + def test_update_node_group_template_v2(self): + url = self.URL + '/node-group-templates' + self.responses.post(url, status_code=202, + json={'node_group_template': self.body}) + resp = self.client_v2.node_group_templates.create(**self.body) + + update_url = self.URL + '/node-group-templates/id' + self.responses.patch(update_url, status_code=202, + json=self.update_json) + + # check that all parameters will be updated + updated = self.client_v2.node_group_templates.update( + "id", + resp.name, + resp.plugin_name, + resp.plugin_version, + resp.flavor_id, + description=getattr(resp, "description", None), + volumes_per_node=getattr(resp, "volumes_per_node", None), + node_configs=getattr(resp, "node_configs", None), + floating_ip_pool=getattr(resp, "floating_ip_pool", None), + security_groups=getattr(resp, "security_groups", None), + auto_security_group=getattr(resp, "auto_security_group", None), + availability_zone=getattr(resp, "availability_zone", None), + volumes_availability_zone=getattr(resp, + "volumes_availability_zone", + None), + volume_type=getattr(resp, "volume_type", None), + image_id=getattr(resp, "image_id", None), + is_proxy_gateway=getattr(resp, "is_proxy_gateway", None), + volume_local_to_instance=getattr(resp, + "volume_local_to_instance", + None), + use_autoconfig=False, + boot_from_volume=getattr(resp, "boot_from_volume", None) + ) + self.assertIsInstance(updated, ng.NodeGroupTemplate) + self.assertFields(self.update_json["node_group_template"], updated) + + # check that parameters will not be updated + self.client_v2.node_group_templates.update("id") + self.assertEqual(update_url, self.responses.last_request.url) + self.assertEqual({}, + json.loads(self.responses.last_request.body)) + + # check that all parameters will be unset + unset_json = { + 'auto_security_group': None, 'availability_zone': None, + 'description': None, 'flavor_id': None, 'floating_ip_pool': None, + 'plugin_version': None, 'image_id': None, 'is_protected': None, + 'is_proxy_gateway': None, 'is_public': None, 'name': None, + 'node_configs': None, 'node_processes': None, 'plugin_name': None, + 'security_groups': None, 'shares': None, 'use_autoconfig': None, + 'volume_local_to_instance': None, 'volume_mount_prefix': None, + 'volume_type': None, 'volumes_availability_zone': None, + 'volumes_per_node': None, 'volumes_size': None, + 'boot_from_volume': None} + + self.client_v2.node_group_templates.update("id", **unset_json) + self.assertEqual(update_url, self.responses.last_request.url) + self.assertEqual(unset_json, + json.loads(self.responses.last_request.body)) @@ -96,5 +96,14 @@ openstack.data_processing.v1 = dataprocessing_job_binary_delete = saharaclient.osc.v1.job_binaries:DeleteJobBinary dataprocessing_job_binary_download = saharaclient.osc.v1.job_binaries:DownloadJobBinary +openstack.data_processing.v2 = + dataprocessing_node_group_template_create = saharaclient.osc.v2.node_group_templates:CreateNodeGroupTemplate + dataprocessing_node_group_template_list = saharaclient.osc.v2.node_group_templates:ListNodeGroupTemplates + dataprocessing_node_group_template_show = saharaclient.osc.v2.node_group_templates:ShowNodeGroupTemplate + dataprocessing_node_group_template_update = saharaclient.osc.v2.node_group_templates:UpdateNodeGroupTemplate + dataprocessing_node_group_template_delete = saharaclient.osc.v2.node_group_templates:DeleteNodeGroupTemplate + dataprocessing_node_group_template_import = saharaclient.osc.v2.node_group_templates:ImportNodeGroupTemplate + dataprocessing_node_group_template_export = saharaclient.osc.v2.node_group_templates:ExportNodeGroupTemplate + [wheel] universal = 1 |