1
0
Fork 0
mirror of https://github.com/ansible-collections/community.general.git synced 2024-09-14 20:13:21 +02:00
community.general/plugins/modules/storage/netapp/netapp_e_volume.py
Felix Fontein 07ecfc940c
Fix imports, part 2 (#42)
* Recover missing netapp.ontap module doc fragments from ansible/ansible@pre-ansible-base.

* Fix PEP8 issues.

* Remove netbox empty files, and test which shouldn't be here.

* Add forgotten file for kubevirt tests.

* Fix unit test imports.

* ansible/ansible#68415 has been fixed.

* Clean up/rearrange imports.

* Update ignore.txt, fix boilerplate.

* Netapp docs fragment: fix spacing

* Forgot to adjust kubevirt tests.
2020-03-25 11:43:51 +00:00

845 lines
42 KiB
Python

#!/usr/bin/python
# (c) 2016, NetApp, Inc
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
from __future__ import absolute_import, division, print_function
__metaclass__ = type
ANSIBLE_METADATA = {'metadata_version': '1.1',
'status': ['preview'],
'supported_by': 'community'}
DOCUMENTATION = '''
---
module: netapp_e_volume
short_description: NetApp E-Series manage storage volumes (standard and thin)
description:
- Create or remove volumes (standard and thin) for NetApp E/EF-series storage arrays.
author:
- Kevin Hulquest (@hulquest)
- Nathan Swartz (@ndswartz)
extends_documentation_fragment:
- community.general.netapp.eseries
options:
state:
description:
- Whether the specified volume should exist
required: true
choices: ['present', 'absent']
name:
description:
- The name of the volume to manage.
required: true
storage_pool_name:
description:
- Required only when requested I(state=='present').
- Name of the storage pool wherein the volume should reside.
required: false
size_unit:
description:
- The unit used to interpret the size parameter
choices: ['bytes', 'b', 'kb', 'mb', 'gb', 'tb', 'pb', 'eb', 'zb', 'yb']
default: 'gb'
size:
description:
- Required only when I(state=='present').
- Size of the volume in I(size_unit).
- Size of the virtual volume in the case of a thin volume in I(size_unit).
- Maximum virtual volume size of a thin provisioned volume is 256tb; however other OS-level restrictions may
exist.
required: true
segment_size_kb:
description:
- Segment size of the volume
- All values are in kibibytes.
- Some common choices include '8', '16', '32', '64', '128', '256', and '512' but options are system
dependent.
- Retrieve the definitive system list from M(netapp_e_facts) under segment_sizes.
- When the storage pool is a raidDiskPool then the segment size must be 128kb.
- Segment size migrations are not allowed in this module
default: '128'
thin_provision:
description:
- Whether the volume should be thin provisioned.
- Thin volumes can only be created when I(raid_level=="raidDiskPool").
- Generally, use of thin-provisioning is not recommended due to performance impacts.
type: bool
default: false
thin_volume_repo_size:
description:
- This value (in size_unit) sets the allocated space for the thin provisioned repository.
- Initial value must between or equal to 4gb and 256gb in increments of 4gb.
- During expansion operations the increase must be between or equal to 4gb and 256gb in increments of 4gb.
- This option has no effect during expansion if I(thin_volume_expansion_policy=="automatic").
- Generally speaking you should almost always use I(thin_volume_expansion_policy=="automatic).
required: false
thin_volume_max_repo_size:
description:
- This is the maximum amount the thin volume repository will be allowed to grow.
- Only has significance when I(thin_volume_expansion_policy=="automatic").
- When the percentage I(thin_volume_repo_size) of I(thin_volume_max_repo_size) exceeds
I(thin_volume_growth_alert_threshold) then a warning will be issued and the storage array will execute
the I(thin_volume_expansion_policy) policy.
- Expansion operations when I(thin_volume_expansion_policy=="automatic") will increase the maximum
repository size.
default: same as size (in size_unit)
thin_volume_expansion_policy:
description:
- This is the thin volume expansion policy.
- When I(thin_volume_expansion_policy=="automatic") and I(thin_volume_growth_alert_threshold) is exceed the
I(thin_volume_max_repo_size) will be automatically expanded.
- When I(thin_volume_expansion_policy=="manual") and I(thin_volume_growth_alert_threshold) is exceeded the
storage system will wait for manual intervention.
- The thin volume_expansion policy can not be modified on existing thin volumes in this module.
- Generally speaking you should almost always use I(thin_volume_expansion_policy=="automatic).
choices: ["automatic", "manual"]
default: "automatic"
thin_volume_growth_alert_threshold:
description:
- This is the thin provision repository utilization threshold (in percent).
- When the percentage of used storage of the maximum repository size exceeds this value then a alert will
be issued and the I(thin_volume_expansion_policy) will be executed.
- Values must be between or equal to 10 and 99.
default: 95
owning_controller:
description:
- Specifies which controller will be the primary owner of the volume
- Not specifying will allow the controller to choose ownership.
required: false
choices: ["A", "B"]
ssd_cache_enabled:
description:
- Whether an existing SSD cache should be enabled on the volume (fails if no SSD cache defined)
- The default value is to ignore existing SSD cache setting.
type: bool
default: false
data_assurance_enabled:
description:
- Determines whether data assurance (DA) should be enabled for the volume
- Only available when creating a new volume and on a storage pool with drives supporting the DA capability.
type: bool
default: false
read_cache_enable:
description:
- Indicates whether read caching should be enabled for the volume.
type: bool
default: true
read_ahead_enable:
description:
- Indicates whether or not automatic cache read-ahead is enabled.
- This option has no effect on thinly provisioned volumes since the architecture for thin volumes cannot
benefit from read ahead caching.
type: bool
default: true
write_cache_enable:
description:
- Indicates whether write-back caching should be enabled for the volume.
type: bool
default: true
cache_without_batteries:
description:
- Indicates whether caching should be used without battery backup.
- Warning, M(cache_without_batteries==true) and the storage system looses power and there is no battery backup, data will be lost!
type: bool
default: false
workload_name:
description:
- Label for the workload defined by the metadata.
- When I(workload_name) and I(metadata) are specified then the defined workload will be added to the storage
array.
- When I(workload_name) exists on the storage array but the metadata is different then the workload
definition will be updated. (Changes will update all associated volumes!)
- Existing workloads can be retrieved using M(netapp_e_facts).
required: false
metadata:
description:
- Dictionary containing meta data for the use, user, location, etc of the volume (dictionary is arbitrarily
defined for whatever the user deems useful)
- When I(workload_name) exists on the storage array but the metadata is different then the workload
definition will be updated. (Changes will update all associated volumes!)
- I(workload_name) must be specified when I(metadata) are defined.
type: dict
required: false
wait_for_initialization:
description:
- Forces the module to wait for expansion operations to complete before continuing.
type: bool
default: false
initialization_timeout:
description:
- Duration in seconds before the wait_for_initialization operation will terminate.
- M(wait_for_initialization==True) to have any effect on module's operations.
type: int
required: false
'''
EXAMPLES = """
- name: Create simple volume with workload tags (volume meta data)
netapp_e_volume:
ssid: "{{ ssid }}"
api_url: "{{ netapp_api_url }}"
api_username: "{{ netapp_api_username }}"
api_password: "{{ netapp_api_password }}"
validate_certs: "{{ netapp_api_validate_certs }}"
state: present
name: volume
storage_pool_name: storage_pool
size: 300
size_unit: gb
workload_name: volume_tag
metadata:
key1: value1
key2: value2
- name: Create a thin volume
netapp_e_volume:
ssid: "{{ ssid }}"
api_url: "{{ netapp_api_url }}"
api_username: "{{ netapp_api_username }}"
api_password: "{{ netapp_api_password }}"
validate_certs: "{{ netapp_api_validate_certs }}"
state: present
name: volume1
storage_pool_name: storage_pool
size: 131072
size_unit: gb
thin_provision: true
thin_volume_repo_size: 32
thin_volume_max_repo_size: 1024
- name: Expand thin volume's virtual size
netapp_e_volume:
ssid: "{{ ssid }}"
api_url: "{{ netapp_api_url }}"
api_username: "{{ netapp_api_username }}"
api_password: "{{ netapp_api_password }}"
validate_certs: "{{ netapp_api_validate_certs }}"
state: present
name: volume1
storage_pool_name: storage_pool
size: 262144
size_unit: gb
thin_provision: true
thin_volume_repo_size: 32
thin_volume_max_repo_size: 1024
- name: Expand thin volume's maximum repository size
netapp_e_volume:
ssid: "{{ ssid }}"
api_url: "{{ netapp_api_url }}"
api_username: "{{ netapp_api_username }}"
api_password: "{{ netapp_api_password }}"
validate_certs: "{{ netapp_api_validate_certs }}"
state: present
name: volume1
storage_pool_name: storage_pool
size: 262144
size_unit: gb
thin_provision: true
thin_volume_repo_size: 32
thin_volume_max_repo_size: 2048
- name: Delete volume
netapp_e_volume:
ssid: "{{ ssid }}"
api_url: "{{ netapp_api_url }}"
api_username: "{{ netapp_api_username }}"
api_password: "{{ netapp_api_password }}"
validate_certs: "{{ netapp_api_validate_certs }}"
state: absent
name: volume
"""
RETURN = """
msg:
description: State of volume
type: str
returned: always
sample: "Standard volume [workload_vol_1] has been created."
"""
from time import sleep
from ansible_collections.netapp.ontap.plugins.module_utils.netapp import NetAppESeriesModule
from ansible.module_utils._text import to_native
class NetAppESeriesVolume(NetAppESeriesModule):
VOLUME_CREATION_BLOCKING_TIMEOUT_SEC = 300
def __init__(self):
ansible_options = dict(
state=dict(required=True, choices=["present", "absent"]),
name=dict(required=True, type="str"),
storage_pool_name=dict(type="str"),
size_unit=dict(default="gb", choices=["bytes", "b", "kb", "mb", "gb", "tb", "pb", "eb", "zb", "yb"],
type="str"),
size=dict(type="float"),
segment_size_kb=dict(type="int", default=128),
owning_controller=dict(required=False, choices=['A', 'B']),
ssd_cache_enabled=dict(type="bool", default=False),
data_assurance_enabled=dict(type="bool", default=False),
thin_provision=dict(type="bool", default=False),
thin_volume_repo_size=dict(type="int"),
thin_volume_max_repo_size=dict(type="float"),
thin_volume_expansion_policy=dict(type="str", choices=["automatic", "manual"]),
thin_volume_growth_alert_threshold=dict(type="int", default=95),
read_cache_enable=dict(type="bool", default=True),
read_ahead_enable=dict(type="bool", default=True),
write_cache_enable=dict(type="bool", default=True),
cache_without_batteries=dict(type="bool", default=False),
workload_name=dict(type="str", required=False),
metadata=dict(type="dict", required=False),
wait_for_initialization=dict(type="bool", default=False),
initialization_timeout=dict(type="int", required=False))
required_if = [
["state", "present", ["storage_pool_name", "size"]],
["thin_provision", "true", ["thin_volume_repo_size"]]
]
super(NetAppESeriesVolume, self).__init__(ansible_options=ansible_options,
web_services_version="02.00.0000.0000",
supports_check_mode=True,
required_if=required_if)
args = self.module.params
self.state = args["state"]
self.name = args["name"]
self.storage_pool_name = args["storage_pool_name"]
self.size_unit = args["size_unit"]
self.segment_size_kb = args["segment_size_kb"]
if args["size"]:
self.size_b = self.convert_to_aligned_bytes(args["size"])
self.owning_controller_id = None
if args["owning_controller"]:
self.owning_controller_id = "070000000000000000000001" if args["owning_controller"] == "A" else "070000000000000000000002"
self.read_cache_enable = args["read_cache_enable"]
self.read_ahead_enable = args["read_ahead_enable"]
self.write_cache_enable = args["write_cache_enable"]
self.ssd_cache_enabled = args["ssd_cache_enabled"]
self.cache_without_batteries = args["cache_without_batteries"]
self.data_assurance_enabled = args["data_assurance_enabled"]
self.thin_provision = args["thin_provision"]
self.thin_volume_expansion_policy = args["thin_volume_expansion_policy"]
self.thin_volume_growth_alert_threshold = int(args["thin_volume_growth_alert_threshold"])
self.thin_volume_repo_size_b = None
self.thin_volume_max_repo_size_b = None
if args["thin_volume_repo_size"]:
self.thin_volume_repo_size_b = self.convert_to_aligned_bytes(args["thin_volume_repo_size"])
if args["thin_volume_max_repo_size"]:
self.thin_volume_max_repo_size_b = self.convert_to_aligned_bytes(args["thin_volume_max_repo_size"])
self.workload_name = args["workload_name"]
self.metadata = args["metadata"]
self.wait_for_initialization = args["wait_for_initialization"]
self.initialization_timeout = args["initialization_timeout"]
# convert metadata to a list of dictionaries containing the keys "key" and "value" corresponding to
# each of the workload attributes dictionary entries
metadata = []
if self.metadata:
if not self.workload_name:
self.module.fail_json(msg="When metadata is specified then the name for the workload must be specified."
" Array [%s]." % self.ssid)
for key in self.metadata.keys():
metadata.append(dict(key=key, value=self.metadata[key]))
self.metadata = metadata
if self.thin_provision:
if not self.thin_volume_max_repo_size_b:
self.thin_volume_max_repo_size_b = self.size_b
if not self.thin_volume_expansion_policy:
self.thin_volume_expansion_policy = "automatic"
if self.size_b > 256 * 1024 ** 4:
self.module.fail_json(msg="Thin provisioned volumes must be less than or equal to 256tb is size."
" Attempted size [%sg]" % (self.size_b * 1024 ** 3))
if (self.thin_volume_repo_size_b and self.thin_volume_max_repo_size_b and
self.thin_volume_repo_size_b > self.thin_volume_max_repo_size_b):
self.module.fail_json(msg="The initial size of the thin volume must not be larger than the maximum"
" repository size. Array [%s]." % self.ssid)
if self.thin_volume_growth_alert_threshold < 10 or self.thin_volume_growth_alert_threshold > 99:
self.module.fail_json(msg="thin_volume_growth_alert_threshold must be between or equal to 10 and 99."
"thin_volume_growth_alert_threshold [%s]. Array [%s]."
% (self.thin_volume_growth_alert_threshold, self.ssid))
self.volume_detail = None
self.pool_detail = None
self.workload_id = None
def convert_to_aligned_bytes(self, size):
"""Convert size to the truncated byte size that aligns on the segment size."""
size_bytes = int(size * self.SIZE_UNIT_MAP[self.size_unit])
segment_size_bytes = int(self.segment_size_kb * self.SIZE_UNIT_MAP["kb"])
segment_count = int(size_bytes / segment_size_bytes)
return segment_count * segment_size_bytes
def get_volume(self):
"""Retrieve volume details from storage array."""
volumes = list()
thin_volumes = list()
try:
rc, volumes = self.request("storage-systems/%s/volumes" % self.ssid)
except Exception as err:
self.module.fail_json(msg="Failed to obtain list of thick volumes. Array Id [%s]. Error[%s]."
% (self.ssid, to_native(err)))
try:
rc, thin_volumes = self.request("storage-systems/%s/thin-volumes" % self.ssid)
except Exception as err:
self.module.fail_json(msg="Failed to obtain list of thin volumes. Array Id [%s]. Error[%s]."
% (self.ssid, to_native(err)))
volume_detail = [volume for volume in volumes + thin_volumes if volume["name"] == self.name]
return volume_detail[0] if volume_detail else dict()
def wait_for_volume_availability(self, retries=VOLUME_CREATION_BLOCKING_TIMEOUT_SEC / 5):
"""Waits until volume becomes available.
:raises AnsibleFailJson when retries are exhausted.
"""
if retries == 0:
self.module.fail_json(msg="Timed out waiting for the volume %s to become available. Array [%s]."
% (self.name, self.ssid))
if not self.get_volume():
sleep(5)
self.wait_for_volume_availability(retries=retries - 1)
def wait_for_volume_action(self, timeout=None):
"""Waits until volume action is complete is complete.
:param: int timeout: Wait duration measured in seconds. Waits indefinitely when None.
"""
action = "unknown"
percent_complete = None
while action != "complete":
sleep(5)
try:
rc, operations = self.request("storage-systems/%s/symbol/getLongLivedOpsProgress" % self.ssid)
# Search long lived operations for volume
action = "complete"
for operation in operations["longLivedOpsProgress"]:
if operation["volAction"] is not None:
for key in operation.keys():
if (operation[key] is not None and "volumeRef" in operation[key] and
(operation[key]["volumeRef"] == self.volume_detail["id"] or
("storageVolumeRef" in self.volume_detail and operation[key]["volumeRef"] == self.volume_detail["storageVolumeRef"]))):
action = operation["volAction"]
percent_complete = operation["init"]["percentComplete"]
except Exception as err:
self.module.fail_json(msg="Failed to get volume expansion progress. Volume [%s]. Array Id [%s]."
" Error[%s]." % (self.name, self.ssid, to_native(err)))
if timeout is not None:
if timeout <= 0:
self.module.warn("Expansion action, %s, failed to complete during the allotted time. Time remaining"
" [%s]. Array Id [%s]." % (action, percent_complete, self.ssid))
self.module.fail_json(msg="Expansion action failed to complete. Time remaining [%s]. Array Id [%s]." % (percent_complete, self.ssid))
if timeout:
timeout -= 5
self.module.log("Expansion action, %s, is %s complete." % (action, percent_complete))
self.module.log("Expansion action is complete.")
def get_storage_pool(self):
"""Retrieve storage pool details from the storage array."""
storage_pools = list()
try:
rc, storage_pools = self.request("storage-systems/%s/storage-pools" % self.ssid)
except Exception as err:
self.module.fail_json(msg="Failed to obtain list of storage pools. Array Id [%s]. Error[%s]."
% (self.ssid, to_native(err)))
pool_detail = [storage_pool for storage_pool in storage_pools if storage_pool["name"] == self.storage_pool_name]
return pool_detail[0] if pool_detail else dict()
def check_storage_pool_sufficiency(self):
"""Perform a series of checks as to the sufficiency of the storage pool for the volume."""
if not self.pool_detail:
self.module.fail_json(msg='Requested storage pool (%s) not found' % self.storage_pool_name)
if not self.volume_detail:
if self.thin_provision and not self.pool_detail['diskPool']:
self.module.fail_json(msg='Thin provisioned volumes can only be created on raid disk pools.')
if (self.data_assurance_enabled and not
(self.pool_detail["protectionInformationCapabilities"]["protectionInformationCapable"] and
self.pool_detail["protectionInformationCapabilities"]["protectionType"] == "type2Protection")):
self.module.fail_json(msg="Data Assurance (DA) requires the storage pool to be DA-compatible."
" Array [%s]." % self.ssid)
if int(self.pool_detail["freeSpace"]) < self.size_b and not self.thin_provision:
self.module.fail_json(msg="Not enough storage pool free space available for the volume's needs."
" Array [%s]." % self.ssid)
else:
# Check for expansion
if (int(self.pool_detail["freeSpace"]) < int(self.volume_detail["totalSizeInBytes"]) - self.size_b and
not self.thin_provision):
self.module.fail_json(msg="Not enough storage pool free space available for the volume's needs."
" Array [%s]." % self.ssid)
def update_workload_tags(self, check_mode=False):
"""Check the status of the workload tag and update storage array definitions if necessary.
When the workload attributes are not provided but an existing workload tag name is, then the attributes will be
used.
:return bool: Whether changes were required to be made."""
change_required = False
workload_tags = None
request_body = None
ansible_profile_id = None
if self.workload_name:
try:
rc, workload_tags = self.request("storage-systems/%s/workloads" % self.ssid)
except Exception as error:
self.module.fail_json(msg="Failed to retrieve storage array workload tags. Array [%s]" % self.ssid)
# Generate common indexed Ansible workload tag
current_tag_index_list = [int(pair["value"].replace("ansible_workload_", ""))
for tag in workload_tags for pair in tag["workloadAttributes"]
if pair["key"] == "profileId" and "ansible_workload_" in pair["value"] and
str(pair["value"]).replace("ansible_workload_", "").isdigit()]
tag_index = 1
if current_tag_index_list:
tag_index = max(current_tag_index_list) + 1
ansible_profile_id = "ansible_workload_%d" % tag_index
request_body = dict(name=self.workload_name,
profileId=ansible_profile_id,
workloadInstanceIndex=None,
isValid=True)
# evaluate and update storage array when needed
for tag in workload_tags:
if tag["name"] == self.workload_name:
self.workload_id = tag["id"]
if not self.metadata:
break
# Determine if core attributes (everything but profileId) is the same
metadata_set = set(tuple(sorted(attr.items())) for attr in self.metadata)
tag_set = set(tuple(sorted(attr.items()))
for attr in tag["workloadAttributes"] if attr["key"] != "profileId")
if metadata_set != tag_set:
self.module.log("Workload tag change is required!")
change_required = True
# only perform the required action when check_mode==False
if change_required and not check_mode:
self.metadata.append(dict(key="profileId", value=ansible_profile_id))
request_body.update(dict(isNewWorkloadInstance=False,
isWorkloadDataInitialized=True,
isWorkloadCardDataToBeReset=True,
workloadAttributes=self.metadata))
try:
rc, resp = self.request("storage-systems/%s/workloads/%s" % (self.ssid, tag["id"]),
data=request_body, method="POST")
except Exception as error:
self.module.fail_json(msg="Failed to create new workload tag. Array [%s]. Error [%s]"
% (self.ssid, to_native(error)))
self.module.log("Workload tag [%s] required change." % self.workload_name)
break
# existing workload tag not found so create new workload tag
else:
change_required = True
self.module.log("Workload tag creation is required!")
if change_required and not check_mode:
if self.metadata:
self.metadata.append(dict(key="profileId", value=ansible_profile_id))
else:
self.metadata = [dict(key="profileId", value=ansible_profile_id)]
request_body.update(dict(isNewWorkloadInstance=True,
isWorkloadDataInitialized=False,
isWorkloadCardDataToBeReset=False,
workloadAttributes=self.metadata))
try:
rc, resp = self.request("storage-systems/%s/workloads" % self.ssid,
method="POST", data=request_body)
self.workload_id = resp["id"]
except Exception as error:
self.module.fail_json(msg="Failed to create new workload tag. Array [%s]. Error [%s]"
% (self.ssid, to_native(error)))
self.module.log("Workload tag [%s] was added." % self.workload_name)
return change_required
def get_volume_property_changes(self):
"""Retrieve the volume update request body when change(s) are required.
:raise AnsibleFailJson when attempting to change segment size on existing volume.
:return dict: request body when change(s) to a volume's properties are required.
"""
change = False
request_body = dict(flashCache=self.ssd_cache_enabled, metaTags=[],
cacheSettings=dict(readCacheEnable=self.read_cache_enable,
writeCacheEnable=self.write_cache_enable))
# check for invalid modifications
if self.segment_size_kb * 1024 != int(self.volume_detail["segmentSize"]):
self.module.fail_json(msg="Existing volume segment size is %s and cannot be modified."
% self.volume_detail["segmentSize"])
# common thick/thin volume properties
if (self.read_cache_enable != self.volume_detail["cacheSettings"]["readCacheEnable"] or
self.write_cache_enable != self.volume_detail["cacheSettings"]["writeCacheEnable"] or
self.ssd_cache_enabled != self.volume_detail["flashCached"]):
change = True
# controller ownership
if self.owning_controller_id and self.owning_controller_id != self.volume_detail["preferredManager"]:
change = True
request_body.update(dict(owningControllerId=self.owning_controller_id))
if self.workload_name:
request_body.update(dict(metaTags=[dict(key="workloadId", value=self.workload_id),
dict(key="volumeTypeId", value="volume")]))
if {"key": "workloadId", "value": self.workload_id} not in self.volume_detail["metadata"]:
change = True
elif self.volume_detail["metadata"]:
change = True
# thick/thin volume specific properties
if self.thin_provision:
if self.thin_volume_growth_alert_threshold != int(self.volume_detail["growthAlertThreshold"]):
change = True
request_body.update(dict(growthAlertThreshold=self.thin_volume_growth_alert_threshold))
if self.thin_volume_expansion_policy != self.volume_detail["expansionPolicy"]:
change = True
request_body.update(dict(expansionPolicy=self.thin_volume_expansion_policy))
else:
if self.read_ahead_enable != (int(self.volume_detail["cacheSettings"]["readAheadMultiplier"]) > 0):
change = True
request_body["cacheSettings"].update(dict(readAheadEnable=self.read_ahead_enable))
if self.cache_without_batteries != self.volume_detail["cacheSettings"]["cwob"]:
change = True
request_body["cacheSettings"].update(dict(cacheWithoutBatteries=self.cache_without_batteries))
return request_body if change else dict()
def get_expand_volume_changes(self):
"""Expand the storage specifications for the existing thick/thin volume.
:raise AnsibleFailJson when a thick/thin volume expansion request fails.
:return dict: dictionary containing all the necessary values for volume expansion request
"""
request_body = dict()
if self.size_b < int(self.volume_detail["capacity"]):
self.module.fail_json(msg="Reducing the size of volumes is not permitted. Volume [%s]. Array [%s]"
% (self.name, self.ssid))
if self.volume_detail["thinProvisioned"]:
if self.size_b > int(self.volume_detail["capacity"]):
request_body.update(dict(sizeUnit="bytes", newVirtualSize=self.size_b))
self.module.log("Thin volume virtual size have been expanded.")
if self.volume_detail["expansionPolicy"] == "automatic":
if self.thin_volume_max_repo_size_b > int(self.volume_detail["provisionedCapacityQuota"]):
request_body.update(dict(sizeUnit="bytes", newRepositorySize=self.thin_volume_max_repo_size_b))
self.module.log("Thin volume maximum repository size have been expanded (automatic policy).")
elif self.volume_detail["expansionPolicy"] == "manual":
if self.thin_volume_repo_size_b > int(self.volume_detail["currentProvisionedCapacity"]):
change = self.thin_volume_repo_size_b - int(self.volume_detail["currentProvisionedCapacity"])
if change < 4 * 1024 ** 3 or change > 256 * 1024 ** 3 or change % (4 * 1024 ** 3) != 0:
self.module.fail_json(msg="The thin volume repository increase must be between or equal to 4gb"
" and 256gb in increments of 4gb. Attempted size [%sg]."
% (self.thin_volume_repo_size_b * 1024 ** 3))
request_body.update(dict(sizeUnit="bytes", newRepositorySize=self.thin_volume_repo_size_b))
self.module.log("Thin volume maximum repository size have been expanded (manual policy).")
elif self.size_b > int(self.volume_detail["capacity"]):
request_body.update(dict(sizeUnit="bytes", expansionSize=self.size_b))
self.module.log("Volume storage capacities have been expanded.")
return request_body
def create_volume(self):
"""Create thick/thin volume according to the specified criteria."""
body = dict(name=self.name, poolId=self.pool_detail["id"], sizeUnit="bytes",
dataAssuranceEnabled=self.data_assurance_enabled)
if self.thin_provision:
body.update(dict(virtualSize=self.size_b,
repositorySize=self.thin_volume_repo_size_b,
maximumRepositorySize=self.thin_volume_max_repo_size_b,
expansionPolicy=self.thin_volume_expansion_policy,
growthAlertThreshold=self.thin_volume_growth_alert_threshold))
try:
rc, volume = self.request("storage-systems/%s/thin-volumes" % self.ssid, data=body, method="POST")
except Exception as error:
self.module.fail_json(msg="Failed to create thin volume. Volume [%s]. Array Id [%s]. Error[%s]."
% (self.name, self.ssid, to_native(error)))
self.module.log("New thin volume created [%s]." % self.name)
else:
body.update(dict(size=self.size_b, segSize=self.segment_size_kb))
try:
rc, volume = self.request("storage-systems/%s/volumes" % self.ssid, data=body, method="POST")
except Exception as error:
self.module.fail_json(msg="Failed to create volume. Volume [%s]. Array Id [%s]. Error[%s]."
% (self.name, self.ssid, to_native(error)))
self.module.log("New volume created [%s]." % self.name)
def update_volume_properties(self):
"""Update existing thin-volume or volume properties.
:raise AnsibleFailJson when either thick/thin volume update request fails.
:return bool: whether update was applied
"""
self.wait_for_volume_availability()
self.volume_detail = self.get_volume()
request_body = self.get_volume_property_changes()
if request_body:
if self.thin_provision:
try:
rc, resp = self.request("storage-systems/%s/thin-volumes/%s"
% (self.ssid, self.volume_detail["id"]), data=request_body, method="POST")
except Exception as error:
self.module.fail_json(msg="Failed to update thin volume properties. Volume [%s]. Array Id [%s]."
" Error[%s]." % (self.name, self.ssid, to_native(error)))
else:
try:
rc, resp = self.request("storage-systems/%s/volumes/%s" % (self.ssid, self.volume_detail["id"]),
data=request_body, method="POST")
except Exception as error:
self.module.fail_json(msg="Failed to update volume properties. Volume [%s]. Array Id [%s]."
" Error[%s]." % (self.name, self.ssid, to_native(error)))
return True
return False
def expand_volume(self):
"""Expand the storage specifications for the existing thick/thin volume.
:raise AnsibleFailJson when a thick/thin volume expansion request fails.
"""
request_body = self.get_expand_volume_changes()
if request_body:
if self.volume_detail["thinProvisioned"]:
try:
rc, resp = self.request("storage-systems/%s/thin-volumes/%s/expand"
% (self.ssid, self.volume_detail["id"]), data=request_body, method="POST")
except Exception as err:
self.module.fail_json(msg="Failed to expand thin volume. Volume [%s]. Array Id [%s]. Error[%s]."
% (self.name, self.ssid, to_native(err)))
self.module.log("Thin volume specifications have been expanded.")
else:
try:
rc, resp = self.request(
"storage-systems/%s/volumes/%s/expand" % (self.ssid, self.volume_detail['id']),
data=request_body, method="POST")
except Exception as err:
self.module.fail_json(msg="Failed to expand volume. Volume [%s]. Array Id [%s]. Error[%s]."
% (self.name, self.ssid, to_native(err)))
self.module.log("Volume storage capacities have been expanded.")
def delete_volume(self):
"""Delete existing thin/thick volume."""
if self.thin_provision:
try:
rc, resp = self.request("storage-systems/%s/thin-volumes/%s" % (self.ssid, self.volume_detail["id"]),
method="DELETE")
except Exception as error:
self.module.fail_json(msg="Failed to delete thin volume. Volume [%s]. Array Id [%s]. Error[%s]."
% (self.name, self.ssid, to_native(error)))
self.module.log("Thin volume deleted [%s]." % self.name)
else:
try:
rc, resp = self.request("storage-systems/%s/volumes/%s" % (self.ssid, self.volume_detail["id"]),
method="DELETE")
except Exception as error:
self.module.fail_json(msg="Failed to delete volume. Volume [%s]. Array Id [%s]. Error[%s]."
% (self.name, self.ssid, to_native(error)))
self.module.log("Volume deleted [%s]." % self.name)
def apply(self):
"""Determine and apply any changes necessary to satisfy the specified criteria.
:raise AnsibleExitJson when completes successfully"""
change = False
msg = None
self.volume_detail = self.get_volume()
self.pool_detail = self.get_storage_pool()
# Determine whether changes need to be applied to existing workload tags
if self.state == 'present' and self.update_workload_tags(check_mode=True):
change = True
# Determine if any changes need to be applied
if self.volume_detail:
if self.state == 'absent':
change = True
elif self.state == 'present':
if self.get_expand_volume_changes() or self.get_volume_property_changes():
change = True
elif self.state == 'present':
if self.thin_provision and (self.thin_volume_repo_size_b < 4 * 1024 ** 3 or
self.thin_volume_repo_size_b > 256 * 1024 ** 3 or
self.thin_volume_repo_size_b % (4 * 1024 ** 3) != 0):
self.module.fail_json(msg="The initial thin volume repository size must be between 4gb and 256gb in"
" increments of 4gb. Attempted size [%sg]."
% (self.thin_volume_repo_size_b * 1024 ** 3))
change = True
self.module.log("Update required: [%s]." % change)
# Apply any necessary changes
if change and not self.module.check_mode:
if self.state == 'present':
if self.update_workload_tags():
msg = "Workload tag change occurred."
if not self.volume_detail:
self.check_storage_pool_sufficiency()
self.create_volume()
self.update_volume_properties()
msg = msg[:-1] + " and volume [%s] was created." if msg else "Volume [%s] has been created."
else:
if self.update_volume_properties():
msg = "Volume [%s] properties were updated."
if self.get_expand_volume_changes():
self.expand_volume()
msg = msg[:-1] + " and was expanded." if msg else "Volume [%s] was expanded."
if self.wait_for_initialization:
self.module.log("Waiting for volume operation to complete.")
self.wait_for_volume_action(timeout=self.initialization_timeout)
elif self.state == 'absent':
self.delete_volume()
msg = "Volume [%s] has been deleted."
else:
msg = "Volume [%s] does not exist." if self.state == 'absent' else "Volume [%s] exists."
self.module.exit_json(msg=(msg % self.name if msg and "%s" in msg else msg), changed=change)
def main():
volume = NetAppESeriesVolume()
volume.apply()
if __name__ == '__main__':
main()