2021-08-08 10:40:22 +02:00
|
|
|
# -*- coding: utf-8 -*-
|
2020-03-09 10:11:07 +01:00
|
|
|
# Copyright (c) 2017-2018 Dell EMC Inc.
|
2022-08-05 12:28:29 +02:00
|
|
|
# GNU General Public License v3.0+ (see LICENSES/GPL-3.0-or-later.txt or https://www.gnu.org/licenses/gpl-3.0.txt)
|
|
|
|
# SPDX-License-Identifier: GPL-3.0-or-later
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
from __future__ import absolute_import, division, print_function
|
|
|
|
__metaclass__ = type
|
|
|
|
|
|
|
|
import json
|
2023-06-05 21:56:44 +02:00
|
|
|
import os
|
|
|
|
import random
|
|
|
|
import string
|
2023-10-09 09:14:47 +02:00
|
|
|
import gzip
|
2024-06-03 06:49:40 +02:00
|
|
|
import time
|
2023-10-09 09:14:47 +02:00
|
|
|
from io import BytesIO
|
2020-03-09 10:11:07 +01:00
|
|
|
from ansible.module_utils.urls import open_url
|
2021-06-26 23:59:11 +02:00
|
|
|
from ansible.module_utils.common.text.converters import to_native
|
|
|
|
from ansible.module_utils.common.text.converters import to_text
|
2023-06-05 21:56:44 +02:00
|
|
|
from ansible.module_utils.common.text.converters import to_bytes
|
|
|
|
from ansible.module_utils.six import text_type
|
2020-03-09 10:11:07 +01:00
|
|
|
from ansible.module_utils.six.moves import http_client
|
|
|
|
from ansible.module_utils.six.moves.urllib.error import URLError, HTTPError
|
2020-06-11 20:24:41 +02:00
|
|
|
from ansible.module_utils.six.moves.urllib.parse import urlparse
|
2024-02-23 06:35:30 +01:00
|
|
|
from ansible.module_utils.ansible_release import __version__ as ansible_version
|
|
|
|
from ansible_collections.community.general.plugins.module_utils.version import LooseVersion
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
GET_HEADERS = {'accept': 'application/json', 'OData-Version': '4.0'}
|
|
|
|
POST_HEADERS = {'content-type': 'application/json', 'accept': 'application/json',
|
|
|
|
'OData-Version': '4.0'}
|
|
|
|
PATCH_HEADERS = {'content-type': 'application/json', 'accept': 'application/json',
|
|
|
|
'OData-Version': '4.0'}
|
2022-12-05 18:31:50 +01:00
|
|
|
PUT_HEADERS = {'content-type': 'application/json', 'accept': 'application/json',
|
|
|
|
'OData-Version': '4.0'}
|
2020-03-09 10:11:07 +01:00
|
|
|
DELETE_HEADERS = {'accept': 'application/json', 'OData-Version': '4.0'}
|
|
|
|
|
2021-03-09 21:23:20 +01:00
|
|
|
FAIL_MSG = 'Issuing a data modification command without specifying the '\
|
|
|
|
'ID of the target %(resource)s resource when there is more '\
|
|
|
|
'than one %(resource)s is no longer allowed. Use the `resource_id` '\
|
|
|
|
'option to specify the target %(resource)s ID.'
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
|
|
|
|
class RedfishUtils(object):
|
|
|
|
|
|
|
|
def __init__(self, creds, root_uri, timeout, module, resource_id=None,
|
2024-06-20 21:03:28 +02:00
|
|
|
data_modification=False, strip_etag_quotes=False, ciphers=None):
|
2020-03-09 10:11:07 +01:00
|
|
|
self.root_uri = root_uri
|
|
|
|
self.creds = creds
|
|
|
|
self.timeout = timeout
|
|
|
|
self.module = module
|
|
|
|
self.service_root = '/redfish/v1/'
|
2023-01-30 21:05:22 +01:00
|
|
|
self.session_service_uri = '/redfish/v1/SessionService'
|
|
|
|
self.sessions_uri = '/redfish/v1/SessionService/Sessions'
|
2020-03-09 10:11:07 +01:00
|
|
|
self.resource_id = resource_id
|
|
|
|
self.data_modification = data_modification
|
2021-09-01 22:59:27 +02:00
|
|
|
self.strip_etag_quotes = strip_etag_quotes
|
2024-06-20 21:03:28 +02:00
|
|
|
self.ciphers = ciphers
|
2022-11-02 07:40:21 +01:00
|
|
|
self._vendor = None
|
2020-03-09 10:11:07 +01:00
|
|
|
self._init_session()
|
|
|
|
|
2021-03-19 21:14:33 +01:00
|
|
|
def _auth_params(self, headers):
|
|
|
|
"""
|
|
|
|
Return tuple of required authentication params based on the presence
|
|
|
|
of a token in the self.creds dict. If using a token, set the
|
|
|
|
X-Auth-Token header in the `headers` param.
|
|
|
|
|
|
|
|
:param headers: dict containing headers to send in request
|
|
|
|
:return: tuple of username, password and force_basic_auth
|
|
|
|
"""
|
|
|
|
if self.creds.get('token'):
|
|
|
|
username = None
|
|
|
|
password = None
|
|
|
|
force_basic_auth = False
|
|
|
|
headers['X-Auth-Token'] = self.creds['token']
|
|
|
|
else:
|
|
|
|
username = self.creds['user']
|
|
|
|
password = self.creds['pswd']
|
|
|
|
force_basic_auth = True
|
|
|
|
return username, password, force_basic_auth
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
def _check_request_payload(self, req_pyld, cur_pyld, uri):
|
|
|
|
"""
|
|
|
|
Checks the request payload with the values currently held by the
|
|
|
|
service. Will check if changes are needed and if properties are
|
|
|
|
supported by the service.
|
|
|
|
|
|
|
|
:param req_pyld: dict containing the properties to apply
|
|
|
|
:param cur_pyld: dict containing the properties currently set
|
|
|
|
:param uri: string containing the URI being modified
|
|
|
|
:return: dict containing response information
|
|
|
|
"""
|
|
|
|
|
|
|
|
change_required = False
|
|
|
|
for prop in req_pyld:
|
|
|
|
# Check if the property is supported by the service
|
|
|
|
if prop not in cur_pyld:
|
|
|
|
return {'ret': False,
|
|
|
|
'changed': False,
|
|
|
|
'msg': '%s does not support the property %s' % (uri, prop),
|
|
|
|
'changes_required': False}
|
|
|
|
|
|
|
|
# Perform additional checks based on the type of property
|
|
|
|
if isinstance(req_pyld[prop], dict) and isinstance(cur_pyld[prop], dict):
|
|
|
|
# If the property is a dictionary, check the nested properties
|
|
|
|
sub_resp = self._check_request_payload(req_pyld[prop], cur_pyld[prop], uri)
|
|
|
|
if not sub_resp['ret']:
|
|
|
|
# Unsupported property or other error condition; no change
|
|
|
|
return sub_resp
|
|
|
|
if sub_resp['changes_required']:
|
|
|
|
# Subordinate dictionary requires changes
|
|
|
|
change_required = True
|
|
|
|
|
|
|
|
else:
|
|
|
|
# For other properties, just compare the values
|
|
|
|
|
|
|
|
# Note: This is also a fallthrough for cases where the request
|
|
|
|
# payload and current settings do not match in their data type.
|
|
|
|
# There are cases where this can be expected, such as when a
|
|
|
|
# property is always 'null' in responses, so we want to attempt
|
|
|
|
# the PATCH request.
|
|
|
|
|
|
|
|
# Note: This is also a fallthrough for properties that are
|
|
|
|
# arrays of objects. Some services erroneously omit properties
|
|
|
|
# within arrays of objects when not configured, and it's
|
|
|
|
# expecting the client to provide them anyway.
|
|
|
|
|
|
|
|
if req_pyld[prop] != cur_pyld[prop]:
|
|
|
|
change_required = True
|
|
|
|
|
|
|
|
resp = {'ret': True, 'changes_required': change_required}
|
|
|
|
if not change_required:
|
|
|
|
# No changes required; all properties set
|
|
|
|
resp['changed'] = False
|
|
|
|
resp['msg'] = 'Properties in %s are already set' % uri
|
|
|
|
return resp
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
# The following functions are to send GET/POST/PATCH/DELETE requests
|
2024-06-03 06:49:40 +02:00
|
|
|
def get_request(self, uri, override_headers=None, allow_no_resp=False, timeout=None):
|
2021-03-19 21:14:33 +01:00
|
|
|
req_headers = dict(GET_HEADERS)
|
2023-10-09 09:14:47 +02:00
|
|
|
if override_headers:
|
|
|
|
req_headers.update(override_headers)
|
2021-03-19 21:14:33 +01:00
|
|
|
username, password, basic_auth = self._auth_params(req_headers)
|
2024-06-03 06:49:40 +02:00
|
|
|
if timeout is None:
|
|
|
|
timeout = self.timeout
|
2020-03-09 10:11:07 +01:00
|
|
|
try:
|
2023-01-30 21:05:22 +01:00
|
|
|
# Service root is an unauthenticated resource; remove credentials
|
|
|
|
# in case the caller will be using sessions later.
|
|
|
|
if uri == (self.root_uri + self.service_root):
|
|
|
|
basic_auth = False
|
2021-03-19 21:14:33 +01:00
|
|
|
resp = open_url(uri, method="GET", headers=req_headers,
|
|
|
|
url_username=username, url_password=password,
|
|
|
|
force_basic_auth=basic_auth, validate_certs=False,
|
2020-03-09 10:11:07 +01:00
|
|
|
follow_redirects='all',
|
2024-06-20 21:03:28 +02:00
|
|
|
use_proxy=True, timeout=timeout, ciphers=self.ciphers)
|
2024-09-08 14:22:37 +02:00
|
|
|
headers = {k.lower(): v for (k, v) in resp.info().items()}
|
2024-02-23 06:35:30 +01:00
|
|
|
try:
|
|
|
|
if headers.get('content-encoding') == 'gzip' and LooseVersion(ansible_version) < LooseVersion('2.14'):
|
|
|
|
# Older versions of Ansible do not automatically decompress the data
|
|
|
|
# Starting in 2.14, open_url will decompress the response data by default
|
|
|
|
data = json.loads(to_native(gzip.open(BytesIO(resp.read()), 'rt', encoding='utf-8').read()))
|
|
|
|
else:
|
|
|
|
data = json.loads(to_native(resp.read()))
|
|
|
|
except Exception as e:
|
|
|
|
# No response data; this is okay in certain cases
|
|
|
|
data = None
|
|
|
|
if not allow_no_resp:
|
|
|
|
raise
|
2020-03-09 10:11:07 +01:00
|
|
|
except HTTPError as e:
|
|
|
|
msg = self._get_extended_message(e)
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "HTTP Error %s on GET request to '%s', extended message: '%s'"
|
|
|
|
% (e.code, uri, msg),
|
|
|
|
'status': e.code}
|
|
|
|
except URLError as e:
|
|
|
|
return {'ret': False, 'msg': "URL Error on GET request to '%s': '%s'"
|
|
|
|
% (uri, e.reason)}
|
|
|
|
# Almost all errors should be caught above, but just in case
|
|
|
|
except Exception as e:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Failed GET request to '%s': '%s'" % (uri, to_text(e))}
|
2022-11-23 07:46:39 +01:00
|
|
|
return {'ret': True, 'data': data, 'headers': headers, 'resp': resp}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2023-06-05 21:56:44 +02:00
|
|
|
def post_request(self, uri, pyld, multipart=False):
|
2021-03-19 21:14:33 +01:00
|
|
|
req_headers = dict(POST_HEADERS)
|
|
|
|
username, password, basic_auth = self._auth_params(req_headers)
|
2020-03-09 10:11:07 +01:00
|
|
|
try:
|
2023-01-30 21:05:22 +01:00
|
|
|
# When performing a POST to the session collection, credentials are
|
|
|
|
# provided in the request body. Do not provide the basic auth
|
|
|
|
# header since this can cause conflicts with some services
|
|
|
|
if self.sessions_uri is not None and uri == (self.root_uri + self.sessions_uri):
|
|
|
|
basic_auth = False
|
2023-06-05 21:56:44 +02:00
|
|
|
if multipart:
|
|
|
|
# Multipart requests require special handling to encode the request body
|
|
|
|
multipart_encoder = self._prepare_multipart(pyld)
|
|
|
|
data = multipart_encoder[0]
|
|
|
|
req_headers['content-type'] = multipart_encoder[1]
|
|
|
|
else:
|
|
|
|
data = json.dumps(pyld)
|
|
|
|
resp = open_url(uri, data=data,
|
2021-03-19 21:14:33 +01:00
|
|
|
headers=req_headers, method="POST",
|
|
|
|
url_username=username, url_password=password,
|
|
|
|
force_basic_auth=basic_auth, validate_certs=False,
|
2020-03-09 10:11:07 +01:00
|
|
|
follow_redirects='all',
|
2024-06-20 21:03:28 +02:00
|
|
|
use_proxy=True, timeout=self.timeout, ciphers=self.ciphers)
|
2022-11-23 07:46:39 +01:00
|
|
|
try:
|
|
|
|
data = json.loads(to_native(resp.read()))
|
|
|
|
except Exception as e:
|
|
|
|
# No response data; this is okay in many cases
|
|
|
|
data = None
|
2024-09-08 14:22:37 +02:00
|
|
|
headers = {k.lower(): v for (k, v) in resp.info().items()}
|
2020-03-09 10:11:07 +01:00
|
|
|
except HTTPError as e:
|
|
|
|
msg = self._get_extended_message(e)
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "HTTP Error %s on POST request to '%s', extended message: '%s'"
|
|
|
|
% (e.code, uri, msg),
|
|
|
|
'status': e.code}
|
|
|
|
except URLError as e:
|
|
|
|
return {'ret': False, 'msg': "URL Error on POST request to '%s': '%s'"
|
|
|
|
% (uri, e.reason)}
|
|
|
|
# Almost all errors should be caught above, but just in case
|
|
|
|
except Exception as e:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Failed POST request to '%s': '%s'" % (uri, to_text(e))}
|
2022-11-23 07:46:39 +01:00
|
|
|
return {'ret': True, 'data': data, 'headers': headers, 'resp': resp}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
def patch_request(self, uri, pyld, check_pyld=False):
|
2021-03-19 21:14:33 +01:00
|
|
|
req_headers = dict(PATCH_HEADERS)
|
2020-03-09 10:11:07 +01:00
|
|
|
r = self.get_request(uri)
|
|
|
|
if r['ret']:
|
|
|
|
# Get etag from etag header or @odata.etag property
|
|
|
|
etag = r['headers'].get('etag')
|
|
|
|
if not etag:
|
|
|
|
etag = r['data'].get('@odata.etag')
|
|
|
|
if etag:
|
2021-09-01 22:59:27 +02:00
|
|
|
if self.strip_etag_quotes:
|
|
|
|
etag = etag.strip('"')
|
2021-03-19 21:14:33 +01:00
|
|
|
req_headers['If-Match'] = etag
|
2022-11-02 07:40:21 +01:00
|
|
|
|
|
|
|
if check_pyld:
|
|
|
|
# Check the payload with the current settings to see if changes
|
|
|
|
# are needed or if there are unsupported properties
|
|
|
|
if r['ret']:
|
|
|
|
check_resp = self._check_request_payload(pyld, r['data'], uri)
|
|
|
|
if not check_resp.pop('changes_required'):
|
|
|
|
check_resp['changed'] = False
|
|
|
|
return check_resp
|
|
|
|
else:
|
|
|
|
r['changed'] = False
|
|
|
|
return r
|
|
|
|
|
2021-03-19 21:14:33 +01:00
|
|
|
username, password, basic_auth = self._auth_params(req_headers)
|
2020-03-09 10:11:07 +01:00
|
|
|
try:
|
|
|
|
resp = open_url(uri, data=json.dumps(pyld),
|
2021-03-19 21:14:33 +01:00
|
|
|
headers=req_headers, method="PATCH",
|
|
|
|
url_username=username, url_password=password,
|
|
|
|
force_basic_auth=basic_auth, validate_certs=False,
|
2020-03-09 10:11:07 +01:00
|
|
|
follow_redirects='all',
|
2024-06-20 21:03:28 +02:00
|
|
|
use_proxy=True, timeout=self.timeout, ciphers=self.ciphers)
|
2020-03-09 10:11:07 +01:00
|
|
|
except HTTPError as e:
|
|
|
|
msg = self._get_extended_message(e)
|
2022-11-02 07:40:21 +01:00
|
|
|
return {'ret': False, 'changed': False,
|
2020-03-09 10:11:07 +01:00
|
|
|
'msg': "HTTP Error %s on PATCH request to '%s', extended message: '%s'"
|
|
|
|
% (e.code, uri, msg),
|
|
|
|
'status': e.code}
|
|
|
|
except URLError as e:
|
2022-11-02 07:40:21 +01:00
|
|
|
return {'ret': False, 'changed': False,
|
|
|
|
'msg': "URL Error on PATCH request to '%s': '%s'" % (uri, e.reason)}
|
2020-03-09 10:11:07 +01:00
|
|
|
# Almost all errors should be caught above, but just in case
|
|
|
|
except Exception as e:
|
2022-11-02 07:40:21 +01:00
|
|
|
return {'ret': False, 'changed': False,
|
2020-03-09 10:11:07 +01:00
|
|
|
'msg': "Failed PATCH request to '%s': '%s'" % (uri, to_text(e))}
|
2022-11-02 07:40:21 +01:00
|
|
|
return {'ret': True, 'changed': True, 'resp': resp, 'msg': 'Modified %s' % uri}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-12-05 18:31:50 +01:00
|
|
|
def put_request(self, uri, pyld):
|
|
|
|
req_headers = dict(PUT_HEADERS)
|
|
|
|
r = self.get_request(uri)
|
|
|
|
if r['ret']:
|
|
|
|
# Get etag from etag header or @odata.etag property
|
|
|
|
etag = r['headers'].get('etag')
|
|
|
|
if not etag:
|
|
|
|
etag = r['data'].get('@odata.etag')
|
|
|
|
if etag:
|
|
|
|
if self.strip_etag_quotes:
|
|
|
|
etag = etag.strip('"')
|
|
|
|
req_headers['If-Match'] = etag
|
|
|
|
username, password, basic_auth = self._auth_params(req_headers)
|
|
|
|
try:
|
|
|
|
resp = open_url(uri, data=json.dumps(pyld),
|
|
|
|
headers=req_headers, method="PUT",
|
|
|
|
url_username=username, url_password=password,
|
|
|
|
force_basic_auth=basic_auth, validate_certs=False,
|
|
|
|
follow_redirects='all',
|
2024-06-20 21:03:28 +02:00
|
|
|
use_proxy=True, timeout=self.timeout, ciphers=self.ciphers)
|
2022-12-05 18:31:50 +01:00
|
|
|
except HTTPError as e:
|
|
|
|
msg = self._get_extended_message(e)
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "HTTP Error %s on PUT request to '%s', extended message: '%s'"
|
|
|
|
% (e.code, uri, msg),
|
|
|
|
'status': e.code}
|
|
|
|
except URLError as e:
|
|
|
|
return {'ret': False, 'msg': "URL Error on PUT request to '%s': '%s'"
|
|
|
|
% (uri, e.reason)}
|
|
|
|
# Almost all errors should be caught above, but just in case
|
|
|
|
except Exception as e:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Failed PUT request to '%s': '%s'" % (uri, to_text(e))}
|
|
|
|
return {'ret': True, 'resp': resp}
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def delete_request(self, uri, pyld=None):
|
2021-03-19 21:14:33 +01:00
|
|
|
req_headers = dict(DELETE_HEADERS)
|
|
|
|
username, password, basic_auth = self._auth_params(req_headers)
|
2020-03-09 10:11:07 +01:00
|
|
|
try:
|
|
|
|
data = json.dumps(pyld) if pyld else None
|
|
|
|
resp = open_url(uri, data=data,
|
2021-03-19 21:14:33 +01:00
|
|
|
headers=req_headers, method="DELETE",
|
|
|
|
url_username=username, url_password=password,
|
|
|
|
force_basic_auth=basic_auth, validate_certs=False,
|
2020-03-09 10:11:07 +01:00
|
|
|
follow_redirects='all',
|
2024-06-20 21:03:28 +02:00
|
|
|
use_proxy=True, timeout=self.timeout, ciphers=self.ciphers)
|
2020-03-09 10:11:07 +01:00
|
|
|
except HTTPError as e:
|
|
|
|
msg = self._get_extended_message(e)
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "HTTP Error %s on DELETE request to '%s', extended message: '%s'"
|
|
|
|
% (e.code, uri, msg),
|
|
|
|
'status': e.code}
|
|
|
|
except URLError as e:
|
|
|
|
return {'ret': False, 'msg': "URL Error on DELETE request to '%s': '%s'"
|
|
|
|
% (uri, e.reason)}
|
|
|
|
# Almost all errors should be caught above, but just in case
|
|
|
|
except Exception as e:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Failed DELETE request to '%s': '%s'" % (uri, to_text(e))}
|
|
|
|
return {'ret': True, 'resp': resp}
|
|
|
|
|
2023-06-05 21:56:44 +02:00
|
|
|
@staticmethod
|
|
|
|
def _prepare_multipart(fields):
|
|
|
|
"""Prepares a multipart body based on a set of fields provided.
|
|
|
|
|
|
|
|
Ideally it would have been good to use the existing 'prepare_multipart'
|
|
|
|
found in ansible.module_utils.urls, but it takes files and encodes them
|
|
|
|
as Base64 strings, which is not expected by Redfish services. It also
|
|
|
|
adds escaping of certain bytes in the payload, such as inserting '\r'
|
2023-10-29 18:04:44 +01:00
|
|
|
any time it finds a standalone '\n', which corrupts the image payload
|
2023-06-05 21:56:44 +02:00
|
|
|
send to the service. This implementation is simplified to Redfish's
|
|
|
|
usage and doesn't necessarily represent an exhaustive method of
|
|
|
|
building multipart requests.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def write_buffer(body, line):
|
|
|
|
# Adds to the multipart body based on the provided data type
|
|
|
|
# At this time there is only support for strings, dictionaries, and bytes (default)
|
|
|
|
if isinstance(line, text_type):
|
|
|
|
body.append(to_bytes(line, encoding='utf-8'))
|
|
|
|
elif isinstance(line, dict):
|
|
|
|
body.append(to_bytes(json.dumps(line), encoding='utf-8'))
|
|
|
|
else:
|
|
|
|
body.append(line)
|
|
|
|
return
|
|
|
|
|
|
|
|
# Generate a random boundary marker; may need to consider probing the
|
|
|
|
# payload for potential conflicts in the future
|
|
|
|
boundary = ''.join(random.choice(string.digits + string.ascii_letters) for i in range(30))
|
|
|
|
body = []
|
|
|
|
for form in fields:
|
|
|
|
# Fill in the form details
|
|
|
|
write_buffer(body, '--' + boundary)
|
|
|
|
|
|
|
|
# Insert the headers (Content-Disposition and Content-Type)
|
|
|
|
if 'filename' in fields[form]:
|
|
|
|
name = os.path.basename(fields[form]['filename']).replace('"', '\\"')
|
|
|
|
write_buffer(body, u'Content-Disposition: form-data; name="%s"; filename="%s"' % (to_text(form), to_text(name)))
|
|
|
|
else:
|
|
|
|
write_buffer(body, 'Content-Disposition: form-data; name="%s"' % form)
|
|
|
|
write_buffer(body, 'Content-Type: %s' % fields[form]['mime_type'])
|
|
|
|
write_buffer(body, '')
|
|
|
|
|
|
|
|
# Insert the payload; read from the file if not given by the caller
|
|
|
|
if 'content' not in fields[form]:
|
|
|
|
with open(to_bytes(fields[form]['filename'], errors='surrogate_or_strict'), 'rb') as f:
|
|
|
|
fields[form]['content'] = f.read()
|
|
|
|
write_buffer(body, fields[form]['content'])
|
|
|
|
|
|
|
|
# Finalize the entire request
|
|
|
|
write_buffer(body, '--' + boundary + '--')
|
|
|
|
write_buffer(body, '')
|
|
|
|
return (b'\r\n'.join(body), 'multipart/form-data; boundary=' + boundary)
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
@staticmethod
|
|
|
|
def _get_extended_message(error):
|
|
|
|
"""
|
|
|
|
Get Redfish ExtendedInfo message from response payload if present
|
|
|
|
:param error: an HTTPError exception
|
|
|
|
:type error: HTTPError
|
|
|
|
:return: the ExtendedInfo message if present, else standard HTTP error
|
|
|
|
"""
|
|
|
|
msg = http_client.responses.get(error.code, '')
|
|
|
|
if error.code >= 400:
|
|
|
|
try:
|
|
|
|
body = error.read().decode('utf-8')
|
|
|
|
data = json.loads(body)
|
|
|
|
ext_info = data['error']['@Message.ExtendedInfo']
|
2022-05-05 07:56:26 +02:00
|
|
|
# if the ExtendedInfo contains a user friendly message send it
|
|
|
|
# otherwise try to send the entire contents of ExtendedInfo
|
|
|
|
try:
|
|
|
|
msg = ext_info[0]['Message']
|
|
|
|
except Exception:
|
|
|
|
msg = str(data['error']['@Message.ExtendedInfo'])
|
2020-03-09 10:11:07 +01:00
|
|
|
except Exception:
|
|
|
|
pass
|
|
|
|
return msg
|
|
|
|
|
|
|
|
def _init_session(self):
|
|
|
|
pass
|
|
|
|
|
2022-08-24 21:36:35 +02:00
|
|
|
def _get_vendor(self):
|
2022-11-02 07:40:21 +01:00
|
|
|
# If we got the vendor info once, don't get it again
|
|
|
|
if self._vendor is not None:
|
|
|
|
return {'ret': 'True', 'Vendor': self._vendor}
|
|
|
|
|
|
|
|
# Find the vendor info from the service root
|
2022-08-24 21:36:35 +02:00
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return {'ret': False, 'Vendor': ''}
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
|
2022-08-24 21:36:35 +02:00
|
|
|
if 'Vendor' in data:
|
2022-11-02 07:40:21 +01:00
|
|
|
# Extract the vendor string from the Vendor property
|
|
|
|
self._vendor = data["Vendor"]
|
2022-08-24 21:36:35 +02:00
|
|
|
return {'ret': True, 'Vendor': data["Vendor"]}
|
2022-11-02 07:40:21 +01:00
|
|
|
elif 'Oem' in data and len(data['Oem']) > 0:
|
|
|
|
# Determine the vendor from the OEM object if needed
|
|
|
|
vendor = list(data['Oem'].keys())[0]
|
|
|
|
if vendor == 'Hpe' or vendor == 'Hp':
|
|
|
|
# HPE uses Pascal-casing for their OEM object
|
|
|
|
# Older systems reported 'Hp' (pre-split)
|
|
|
|
vendor = 'HPE'
|
|
|
|
self._vendor = vendor
|
|
|
|
return {'ret': True, 'Vendor': vendor}
|
2022-08-24 21:36:35 +02:00
|
|
|
else:
|
2022-11-02 07:40:21 +01:00
|
|
|
# Could not determine; use an empty string
|
|
|
|
self._vendor = ''
|
2022-08-24 21:36:35 +02:00
|
|
|
return {'ret': True, 'Vendor': ''}
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def _find_accountservice_resource(self):
|
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'AccountService' not in data:
|
|
|
|
return {'ret': False, 'msg': "AccountService resource not found"}
|
|
|
|
else:
|
|
|
|
account_service = data["AccountService"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + account_service)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
accounts = data['Accounts']['@odata.id']
|
|
|
|
if accounts[-1:] == '/':
|
|
|
|
accounts = accounts[:-1]
|
|
|
|
self.accounts_uri = accounts
|
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def _find_sessionservice_resource(self):
|
2023-01-30 21:05:22 +01:00
|
|
|
# Get the service root
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2023-01-30 21:05:22 +01:00
|
|
|
|
|
|
|
# Check for the session service and session collection. Well-known
|
|
|
|
# defaults are provided in the constructor, but services that predate
|
|
|
|
# Redfish 1.6.0 might contain different values.
|
|
|
|
self.session_service_uri = data.get('SessionService', {}).get('@odata.id')
|
|
|
|
self.sessions_uri = data.get('Links', {}).get('Sessions', {}).get('@odata.id')
|
|
|
|
|
|
|
|
# If one isn't found, return an error
|
|
|
|
if self.session_service_uri is None:
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': False, 'msg': "SessionService resource not found"}
|
2023-01-30 21:05:22 +01:00
|
|
|
if self.sessions_uri is None:
|
|
|
|
return {'ret': False, 'msg': "SessionCollection resource not found"}
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def _get_resource_uri_by_id(self, uris, id_prop):
|
|
|
|
for uri in uris:
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
continue
|
|
|
|
data = response['data']
|
|
|
|
if id_prop == data.get('Id'):
|
|
|
|
return uri
|
|
|
|
return None
|
|
|
|
|
|
|
|
def _find_systems_resource(self):
|
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'Systems' not in data:
|
|
|
|
return {'ret': False, 'msg': "Systems resource not found"}
|
|
|
|
response = self.get_request(self.root_uri + data['Systems']['@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
self.systems_uris = [
|
|
|
|
i['@odata.id'] for i in response['data'].get('Members', [])]
|
|
|
|
if not self.systems_uris:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "ComputerSystem's Members array is either empty or missing"}
|
|
|
|
self.systems_uri = self.systems_uris[0]
|
|
|
|
if self.data_modification:
|
|
|
|
if self.resource_id:
|
|
|
|
self.systems_uri = self._get_resource_uri_by_id(self.systems_uris,
|
|
|
|
self.resource_id)
|
|
|
|
if not self.systems_uri:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "System resource %s not found" % self.resource_id}
|
|
|
|
elif len(self.systems_uris) > 1:
|
2021-03-09 21:23:20 +01:00
|
|
|
self.module.fail_json(msg=FAIL_MSG % {'resource': 'System'})
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def _find_updateservice_resource(self):
|
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'UpdateService' not in data:
|
|
|
|
return {'ret': False, 'msg': "UpdateService resource not found"}
|
|
|
|
else:
|
|
|
|
update = data["UpdateService"]["@odata.id"]
|
|
|
|
self.update_uri = update
|
|
|
|
response = self.get_request(self.root_uri + update)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
self.firmware_uri = self.software_uri = None
|
|
|
|
if 'FirmwareInventory' in data:
|
|
|
|
self.firmware_uri = data['FirmwareInventory'][u'@odata.id']
|
|
|
|
if 'SoftwareInventory' in data:
|
|
|
|
self.software_uri = data['SoftwareInventory'][u'@odata.id']
|
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def _find_chassis_resource(self):
|
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'Chassis' not in data:
|
|
|
|
return {'ret': False, 'msg': "Chassis resource not found"}
|
|
|
|
chassis = data["Chassis"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + chassis)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
self.chassis_uris = [
|
|
|
|
i['@odata.id'] for i in response['data'].get('Members', [])]
|
|
|
|
if not self.chassis_uris:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Chassis Members array is either empty or missing"}
|
|
|
|
self.chassis_uri = self.chassis_uris[0]
|
|
|
|
if self.data_modification:
|
|
|
|
if self.resource_id:
|
|
|
|
self.chassis_uri = self._get_resource_uri_by_id(self.chassis_uris,
|
|
|
|
self.resource_id)
|
|
|
|
if not self.chassis_uri:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "Chassis resource %s not found" % self.resource_id}
|
|
|
|
elif len(self.chassis_uris) > 1:
|
2021-03-09 21:23:20 +01:00
|
|
|
self.module.fail_json(msg=FAIL_MSG % {'resource': 'Chassis'})
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def _find_managers_resource(self):
|
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'Managers' not in data:
|
|
|
|
return {'ret': False, 'msg': "Manager resource not found"}
|
|
|
|
manager = data["Managers"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + manager)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
self.manager_uris = [
|
|
|
|
i['@odata.id'] for i in response['data'].get('Members', [])]
|
|
|
|
if not self.manager_uris:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Managers Members array is either empty or missing"}
|
|
|
|
self.manager_uri = self.manager_uris[0]
|
|
|
|
if self.data_modification:
|
|
|
|
if self.resource_id:
|
|
|
|
self.manager_uri = self._get_resource_uri_by_id(self.manager_uris,
|
|
|
|
self.resource_id)
|
|
|
|
if not self.manager_uri:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "Manager resource %s not found" % self.resource_id}
|
|
|
|
elif len(self.manager_uris) > 1:
|
2021-03-09 21:23:20 +01:00
|
|
|
self.module.fail_json(msg=FAIL_MSG % {'resource': 'Manager'})
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': True}
|
|
|
|
|
2020-06-11 20:24:41 +02:00
|
|
|
def _get_all_action_info_values(self, action):
|
|
|
|
"""Retrieve all parameter values for an Action from ActionInfo.
|
|
|
|
Fall back to AllowableValue annotations if no ActionInfo found.
|
|
|
|
Return the result in an ActionInfo-like dictionary, keyed
|
|
|
|
by the name of the parameter. """
|
|
|
|
ai = {}
|
|
|
|
if '@Redfish.ActionInfo' in action:
|
|
|
|
ai_uri = action['@Redfish.ActionInfo']
|
|
|
|
response = self.get_request(self.root_uri + ai_uri)
|
|
|
|
if response['ret'] is True:
|
|
|
|
data = response['data']
|
|
|
|
if 'Parameters' in data:
|
|
|
|
params = data['Parameters']
|
2024-09-08 14:22:37 +02:00
|
|
|
ai = {p['Name']: p for p in params if 'Name' in p}
|
2020-06-11 20:24:41 +02:00
|
|
|
if not ai:
|
2024-09-05 21:47:28 +02:00
|
|
|
ai = {
|
|
|
|
k[:-24]: {'AllowableValues': v}
|
|
|
|
for k, v in action.items()
|
|
|
|
if k.endswith('@Redfish.AllowableValues')
|
|
|
|
}
|
2020-06-11 20:24:41 +02:00
|
|
|
return ai
|
|
|
|
|
|
|
|
def _get_allowable_values(self, action, name, default_values=None):
|
|
|
|
if default_values is None:
|
|
|
|
default_values = []
|
|
|
|
ai = self._get_all_action_info_values(action)
|
|
|
|
allowable_values = ai.get(name, {}).get('AllowableValues')
|
|
|
|
# fallback to default values
|
|
|
|
if allowable_values is None:
|
|
|
|
allowable_values = default_values
|
|
|
|
return allowable_values
|
|
|
|
|
2024-06-03 06:49:40 +02:00
|
|
|
def check_service_availability(self):
|
|
|
|
"""
|
|
|
|
Checks if the service is accessible.
|
|
|
|
|
|
|
|
:return: dict containing the status of the service
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Get the service root
|
|
|
|
# Override the timeout since the service root is expected to be readily
|
|
|
|
# available.
|
|
|
|
service_root = self.get_request(self.root_uri + self.service_root, timeout=10)
|
|
|
|
if service_root['ret'] is False:
|
|
|
|
# Failed, either due to a timeout or HTTP error; not available
|
|
|
|
return {'ret': True, 'available': False}
|
|
|
|
|
|
|
|
# Successfully accessed the service root; available
|
|
|
|
return {'ret': True, 'available': True}
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def get_logs(self):
|
|
|
|
log_svcs_uri_list = []
|
|
|
|
list_of_logs = []
|
|
|
|
properties = ['Severity', 'Created', 'EntryType', 'OemRecordFormat',
|
|
|
|
'Message', 'MessageId', 'MessageArgs']
|
|
|
|
|
|
|
|
# Find LogService
|
|
|
|
response = self.get_request(self.root_uri + self.manager_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'LogServices' not in data:
|
|
|
|
return {'ret': False, 'msg': "LogServices resource not found"}
|
|
|
|
|
|
|
|
# Find all entries in LogServices
|
|
|
|
logs_uri = data["LogServices"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + logs_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
for log_svcs_entry in data.get('Members', []):
|
|
|
|
response = self.get_request(self.root_uri + log_svcs_entry[u'@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
_data = response['data']
|
|
|
|
if 'Entries' in _data:
|
|
|
|
log_svcs_uri_list.append(_data['Entries'][u'@odata.id'])
|
|
|
|
|
|
|
|
# For each entry in LogServices, get log name and all log entries
|
|
|
|
for log_svcs_uri in log_svcs_uri_list:
|
|
|
|
logs = {}
|
|
|
|
list_of_log_entries = []
|
|
|
|
response = self.get_request(self.root_uri + log_svcs_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
logs['Description'] = data.get('Description',
|
|
|
|
'Collection of log entries')
|
|
|
|
# Get all log entries for each type of log found
|
|
|
|
for logEntry in data.get('Members', []):
|
|
|
|
entry = {}
|
|
|
|
for prop in properties:
|
|
|
|
if prop in logEntry:
|
|
|
|
entry[prop] = logEntry.get(prop)
|
|
|
|
if entry:
|
|
|
|
list_of_log_entries.append(entry)
|
|
|
|
log_name = log_svcs_uri.split('/')[-1]
|
|
|
|
logs[log_name] = list_of_log_entries
|
|
|
|
list_of_logs.append(logs)
|
|
|
|
|
|
|
|
# list_of_logs[logs{list_of_log_entries[entry{}]}]
|
|
|
|
return {'ret': True, 'entries': list_of_logs}
|
|
|
|
|
|
|
|
def clear_logs(self):
|
|
|
|
# Find LogService
|
|
|
|
response = self.get_request(self.root_uri + self.manager_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'LogServices' not in data:
|
|
|
|
return {'ret': False, 'msg': "LogServices resource not found"}
|
|
|
|
|
|
|
|
# Find all entries in LogServices
|
|
|
|
logs_uri = data["LogServices"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + logs_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for log_svcs_entry in data[u'Members']:
|
|
|
|
response = self.get_request(self.root_uri + log_svcs_entry["@odata.id"])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
_data = response['data']
|
|
|
|
# Check to make sure option is available, otherwise error is ugly
|
|
|
|
if "Actions" in _data:
|
|
|
|
if "#LogService.ClearLog" in _data[u"Actions"]:
|
|
|
|
self.post_request(self.root_uri + _data[u"Actions"]["#LogService.ClearLog"]["target"], {})
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def aggregate(self, func, uri_list, uri_name):
|
|
|
|
ret = True
|
|
|
|
entries = []
|
|
|
|
for uri in uri_list:
|
|
|
|
inventory = func(uri)
|
|
|
|
ret = inventory.pop('ret') and ret
|
|
|
|
if 'entries' in inventory:
|
|
|
|
entries.append(({uri_name: uri},
|
|
|
|
inventory['entries']))
|
|
|
|
return dict(ret=ret, entries=entries)
|
|
|
|
|
|
|
|
def aggregate_chassis(self, func):
|
|
|
|
return self.aggregate(func, self.chassis_uris, 'chassis_uri')
|
|
|
|
|
|
|
|
def aggregate_managers(self, func):
|
|
|
|
return self.aggregate(func, self.manager_uris, 'manager_uri')
|
|
|
|
|
|
|
|
def aggregate_systems(self, func):
|
|
|
|
return self.aggregate(func, self.systems_uris, 'system_uri')
|
|
|
|
|
|
|
|
def get_storage_controller_inventory(self, systems_uri):
|
|
|
|
result = {}
|
|
|
|
controller_list = []
|
|
|
|
controller_results = []
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['CacheSummary', 'FirmwareVersion', 'Identifiers',
|
2021-01-29 07:17:57 +01:00
|
|
|
'Location', 'Manufacturer', 'Model', 'Name', 'Id',
|
2020-03-09 10:11:07 +01:00
|
|
|
'PartNumber', 'SerialNumber', 'SpeedGbps', 'Status']
|
2023-08-11 19:18:44 +02:00
|
|
|
key = "Controllers"
|
|
|
|
deprecated_key = "StorageControllers"
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Find Storage service
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if 'Storage' not in data:
|
|
|
|
return {'ret': False, 'msg': "Storage resource not found"}
|
|
|
|
|
|
|
|
# Get a list of all storage controllers and build respective URIs
|
|
|
|
storage_uri = data['Storage']["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + storage_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Loop through Members and their StorageControllers
|
|
|
|
# and gather properties from each StorageController
|
|
|
|
if data[u'Members']:
|
|
|
|
for storage_member in data[u'Members']:
|
|
|
|
storage_member_uri = storage_member[u'@odata.id']
|
|
|
|
response = self.get_request(self.root_uri + storage_member_uri)
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if key in data:
|
2023-08-11 19:18:44 +02:00
|
|
|
controllers_uri = data[key][u'@odata.id']
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + controllers_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if data[u'Members']:
|
|
|
|
for controller_member in data[u'Members']:
|
|
|
|
controller_member_uri = controller_member[u'@odata.id']
|
|
|
|
response = self.get_request(self.root_uri + controller_member_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
controller_result = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
controller_result[property] = data[property]
|
|
|
|
controller_results.append(controller_result)
|
|
|
|
elif deprecated_key in data:
|
|
|
|
controller_list = data[deprecated_key]
|
2020-03-09 10:11:07 +01:00
|
|
|
for controller in controller_list:
|
|
|
|
controller_result = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in controller:
|
|
|
|
controller_result[property] = controller[property]
|
|
|
|
controller_results.append(controller_result)
|
|
|
|
result['entries'] = controller_results
|
|
|
|
return result
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': "Storage resource not found"}
|
|
|
|
|
|
|
|
def get_multi_storage_controller_inventory(self):
|
|
|
|
return self.aggregate_systems(self.get_storage_controller_inventory)
|
|
|
|
|
|
|
|
def get_disk_inventory(self, systems_uri):
|
|
|
|
result = {'entries': []}
|
|
|
|
controller_list = []
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['BlockSizeBytes', 'CapableSpeedGbs', 'CapacityBytes',
|
|
|
|
'EncryptionAbility', 'EncryptionStatus',
|
|
|
|
'FailurePredicted', 'HotspareType', 'Id', 'Identifiers',
|
2023-08-21 21:19:53 +02:00
|
|
|
'Links', 'Manufacturer', 'MediaType', 'Model', 'Name',
|
2020-03-09 10:11:07 +01:00
|
|
|
'PartNumber', 'PhysicalLocation', 'Protocol', 'Revision',
|
|
|
|
'RotationSpeedRPM', 'SerialNumber', 'Status']
|
|
|
|
|
|
|
|
# Find Storage service
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if 'SimpleStorage' not in data and 'Storage' not in data:
|
|
|
|
return {'ret': False, 'msg': "SimpleStorage and Storage resource \
|
|
|
|
not found"}
|
|
|
|
|
|
|
|
if 'Storage' in data:
|
|
|
|
# Get a list of all storage controllers and build respective URIs
|
|
|
|
storage_uri = data[u'Storage'][u'@odata.id']
|
|
|
|
response = self.get_request(self.root_uri + storage_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if data[u'Members']:
|
|
|
|
for controller in data[u'Members']:
|
|
|
|
controller_list.append(controller[u'@odata.id'])
|
|
|
|
for c in controller_list:
|
|
|
|
uri = self.root_uri + c
|
|
|
|
response = self.get_request(uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
controller_name = 'Controller 1'
|
2023-08-11 19:18:44 +02:00
|
|
|
if 'Controllers' in data:
|
|
|
|
controllers_uri = data['Controllers'][u'@odata.id']
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + controllers_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
cdata = response['data']
|
|
|
|
|
|
|
|
if cdata[u'Members']:
|
|
|
|
controller_member_uri = cdata[u'Members'][0][u'@odata.id']
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + controller_member_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
cdata = response['data']
|
|
|
|
controller_name = cdata['Name']
|
|
|
|
elif 'StorageControllers' in data:
|
2020-03-09 10:11:07 +01:00
|
|
|
sc = data['StorageControllers']
|
|
|
|
if sc:
|
|
|
|
if 'Name' in sc[0]:
|
|
|
|
controller_name = sc[0]['Name']
|
|
|
|
else:
|
|
|
|
sc_id = sc[0].get('Id', '1')
|
|
|
|
controller_name = 'Controller %s' % sc_id
|
|
|
|
drive_results = []
|
|
|
|
if 'Drives' in data:
|
|
|
|
for device in data[u'Drives']:
|
|
|
|
disk_uri = self.root_uri + device[u'@odata.id']
|
|
|
|
response = self.get_request(disk_uri)
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
drive_result = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
if data[property] is not None:
|
2023-08-21 21:19:53 +02:00
|
|
|
if property == "Links":
|
|
|
|
if "Volumes" in data["Links"].keys():
|
|
|
|
volumes = [v["@odata.id"] for v in data["Links"]["Volumes"]]
|
|
|
|
drive_result["Volumes"] = volumes
|
|
|
|
else:
|
|
|
|
drive_result[property] = data[property]
|
2020-03-09 10:11:07 +01:00
|
|
|
drive_results.append(drive_result)
|
|
|
|
drives = {'Controller': controller_name,
|
|
|
|
'Drives': drive_results}
|
|
|
|
result["entries"].append(drives)
|
|
|
|
|
|
|
|
if 'SimpleStorage' in data:
|
|
|
|
# Get a list of all storage controllers and build respective URIs
|
|
|
|
storage_uri = data["SimpleStorage"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + storage_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for controller in data[u'Members']:
|
|
|
|
controller_list.append(controller[u'@odata.id'])
|
|
|
|
|
|
|
|
for c in controller_list:
|
|
|
|
uri = self.root_uri + c
|
|
|
|
response = self.get_request(uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'Name' in data:
|
|
|
|
controller_name = data['Name']
|
|
|
|
else:
|
|
|
|
sc_id = data.get('Id', '1')
|
|
|
|
controller_name = 'Controller %s' % sc_id
|
|
|
|
drive_results = []
|
|
|
|
for device in data[u'Devices']:
|
|
|
|
drive_result = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in device:
|
|
|
|
drive_result[property] = device[property]
|
|
|
|
drive_results.append(drive_result)
|
|
|
|
drives = {'Controller': controller_name,
|
|
|
|
'Drives': drive_results}
|
|
|
|
result["entries"].append(drives)
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_disk_inventory(self):
|
|
|
|
return self.aggregate_systems(self.get_disk_inventory)
|
|
|
|
|
|
|
|
def get_volume_inventory(self, systems_uri):
|
|
|
|
result = {'entries': []}
|
|
|
|
controller_list = []
|
|
|
|
volume_list = []
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['Id', 'Name', 'RAIDType', 'VolumeType', 'BlockSizeBytes',
|
|
|
|
'Capacity', 'CapacityBytes', 'CapacitySources',
|
|
|
|
'Encrypted', 'EncryptionTypes', 'Identifiers',
|
|
|
|
'Operations', 'OptimumIOSizeBytes', 'AccessCapabilities',
|
|
|
|
'AllocatedPools', 'Status']
|
|
|
|
|
|
|
|
# Find Storage service
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if 'SimpleStorage' not in data and 'Storage' not in data:
|
|
|
|
return {'ret': False, 'msg': "SimpleStorage and Storage resource \
|
|
|
|
not found"}
|
|
|
|
|
|
|
|
if 'Storage' in data:
|
|
|
|
# Get a list of all storage controllers and build respective URIs
|
|
|
|
storage_uri = data[u'Storage'][u'@odata.id']
|
|
|
|
response = self.get_request(self.root_uri + storage_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if data.get('Members'):
|
|
|
|
for controller in data[u'Members']:
|
|
|
|
controller_list.append(controller[u'@odata.id'])
|
2023-06-26 23:29:18 +02:00
|
|
|
for idx, c in enumerate(controller_list):
|
2020-03-09 10:11:07 +01:00
|
|
|
uri = self.root_uri + c
|
|
|
|
response = self.get_request(uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2023-06-26 23:29:18 +02:00
|
|
|
controller_name = 'Controller %s' % str(idx)
|
2023-08-11 19:18:44 +02:00
|
|
|
if 'Controllers' in data:
|
2023-06-26 23:29:18 +02:00
|
|
|
response = self.get_request(self.root_uri + data['Controllers'][u'@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
2023-07-15 16:28:27 +02:00
|
|
|
c_data = response['data']
|
2023-06-26 23:29:18 +02:00
|
|
|
|
2023-07-15 16:28:27 +02:00
|
|
|
if c_data.get('Members') and c_data['Members']:
|
|
|
|
response = self.get_request(self.root_uri + c_data['Members'][0][u'@odata.id'])
|
2023-06-26 23:29:18 +02:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
2023-07-15 16:28:27 +02:00
|
|
|
member_data = response['data']
|
2023-06-26 23:29:18 +02:00
|
|
|
|
2023-07-15 16:28:27 +02:00
|
|
|
if member_data:
|
|
|
|
if 'Name' in member_data:
|
|
|
|
controller_name = member_data['Name']
|
2023-06-26 23:29:18 +02:00
|
|
|
else:
|
2023-07-15 16:28:27 +02:00
|
|
|
controller_id = member_data.get('Id', '1')
|
2023-06-26 23:29:18 +02:00
|
|
|
controller_name = 'Controller %s' % controller_id
|
2023-08-11 19:18:44 +02:00
|
|
|
elif 'StorageControllers' in data:
|
|
|
|
sc = data['StorageControllers']
|
|
|
|
if sc:
|
|
|
|
if 'Name' in sc[0]:
|
|
|
|
controller_name = sc[0]['Name']
|
|
|
|
else:
|
|
|
|
sc_id = sc[0].get('Id', '1')
|
|
|
|
controller_name = 'Controller %s' % sc_id
|
2020-03-09 10:11:07 +01:00
|
|
|
volume_results = []
|
2023-06-26 23:29:18 +02:00
|
|
|
volume_list = []
|
2020-03-09 10:11:07 +01:00
|
|
|
if 'Volumes' in data:
|
|
|
|
# Get a list of all volumes and build respective URIs
|
|
|
|
volumes_uri = data[u'Volumes'][u'@odata.id']
|
|
|
|
response = self.get_request(self.root_uri + volumes_uri)
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if data.get('Members'):
|
|
|
|
for volume in data[u'Members']:
|
|
|
|
volume_list.append(volume[u'@odata.id'])
|
|
|
|
for v in volume_list:
|
|
|
|
uri = self.root_uri + v
|
|
|
|
response = self.get_request(uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
volume_result = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
if data[property] is not None:
|
|
|
|
volume_result[property] = data[property]
|
|
|
|
|
|
|
|
# Get related Drives Id
|
|
|
|
drive_id_list = []
|
|
|
|
if 'Links' in data:
|
|
|
|
if 'Drives' in data[u'Links']:
|
|
|
|
for link in data[u'Links'][u'Drives']:
|
|
|
|
drive_id_link = link[u'@odata.id']
|
|
|
|
drive_id = drive_id_link.split("/")[-1]
|
|
|
|
drive_id_list.append({'Id': drive_id})
|
|
|
|
volume_result['Linked_drives'] = drive_id_list
|
|
|
|
volume_results.append(volume_result)
|
|
|
|
volumes = {'Controller': controller_name,
|
|
|
|
'Volumes': volume_results}
|
|
|
|
result["entries"].append(volumes)
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': "Storage resource not found"}
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_volume_inventory(self):
|
|
|
|
return self.aggregate_systems(self.get_volume_inventory)
|
|
|
|
|
2022-04-11 20:10:18 +02:00
|
|
|
def manage_system_indicator_led(self, command):
|
|
|
|
return self.manage_indicator_led(command, self.systems_uri)
|
|
|
|
|
|
|
|
def manage_chassis_indicator_led(self, command):
|
|
|
|
return self.manage_indicator_led(command, self.chassis_uri)
|
|
|
|
|
|
|
|
def manage_indicator_led(self, command, resource_uri=None):
|
2022-11-02 07:40:21 +01:00
|
|
|
# If no resource is specified; default to the Chassis resource
|
2022-04-11 20:10:18 +02:00
|
|
|
if resource_uri is None:
|
|
|
|
resource_uri = self.chassis_uri
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Perform a PATCH on the IndicatorLED property based on the requested command
|
2020-03-09 10:11:07 +01:00
|
|
|
payloads = {'IndicatorLedOn': 'Lit', 'IndicatorLedOff': 'Off', "IndicatorLedBlink": 'Blinking'}
|
2022-11-02 07:40:21 +01:00
|
|
|
if command not in payloads.keys():
|
|
|
|
return {'ret': False, 'msg': 'Invalid command (%s)' % command}
|
|
|
|
payload = {'IndicatorLED': payloads[command]}
|
|
|
|
resp = self.patch_request(self.root_uri + resource_uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Set IndicatorLED to %s' % payloads[command]
|
|
|
|
return resp
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def _map_reset_type(self, reset_type, allowable_values):
|
|
|
|
equiv_types = {
|
|
|
|
'On': 'ForceOn',
|
|
|
|
'ForceOn': 'On',
|
|
|
|
'ForceOff': 'GracefulShutdown',
|
|
|
|
'GracefulShutdown': 'ForceOff',
|
|
|
|
'GracefulRestart': 'ForceRestart',
|
|
|
|
'ForceRestart': 'GracefulRestart'
|
|
|
|
}
|
|
|
|
|
|
|
|
if reset_type in allowable_values:
|
|
|
|
return reset_type
|
|
|
|
if reset_type not in equiv_types:
|
|
|
|
return reset_type
|
|
|
|
mapped_type = equiv_types[reset_type]
|
|
|
|
if mapped_type in allowable_values:
|
|
|
|
return mapped_type
|
|
|
|
return reset_type
|
|
|
|
|
|
|
|
def manage_system_power(self, command):
|
2020-09-23 08:42:56 +02:00
|
|
|
return self.manage_power(command, self.systems_uri,
|
|
|
|
'#ComputerSystem.Reset')
|
|
|
|
|
2024-06-03 06:49:40 +02:00
|
|
|
def manage_manager_power(self, command, wait=False, wait_timeout=120):
|
2020-09-23 08:42:56 +02:00
|
|
|
return self.manage_power(command, self.manager_uri,
|
2024-06-03 06:49:40 +02:00
|
|
|
'#Manager.Reset', wait, wait_timeout)
|
2020-09-23 08:42:56 +02:00
|
|
|
|
2024-06-03 06:49:40 +02:00
|
|
|
def manage_power(self, command, resource_uri, action_name, wait=False,
|
|
|
|
wait_timeout=120):
|
2020-03-09 10:11:07 +01:00
|
|
|
key = "Actions"
|
|
|
|
reset_type_values = ['On', 'ForceOff', 'GracefulShutdown',
|
|
|
|
'GracefulRestart', 'ForceRestart', 'Nmi',
|
|
|
|
'ForceOn', 'PushPowerButton', 'PowerCycle']
|
|
|
|
|
|
|
|
# command should be PowerOn, PowerForceOff, etc.
|
|
|
|
if not command.startswith('Power'):
|
|
|
|
return {'ret': False, 'msg': 'Invalid Command (%s)' % command}
|
2023-08-14 21:59:24 +02:00
|
|
|
|
|
|
|
# Commands (except PowerCycle) will be stripped of the 'Power' prefix
|
|
|
|
if command == 'PowerCycle':
|
|
|
|
reset_type = command
|
|
|
|
else:
|
|
|
|
reset_type = command[5:]
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# map Reboot to a ResetType that does a reboot
|
|
|
|
if reset_type == 'Reboot':
|
|
|
|
reset_type = 'GracefulRestart'
|
|
|
|
|
|
|
|
if reset_type not in reset_type_values:
|
|
|
|
return {'ret': False, 'msg': 'Invalid Command (%s)' % command}
|
|
|
|
|
2020-09-23 08:42:56 +02:00
|
|
|
# read the resource and get the current power state
|
|
|
|
response = self.get_request(self.root_uri + resource_uri)
|
2020-03-09 10:11:07 +01:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
power_state = data.get('PowerState')
|
|
|
|
|
|
|
|
# if power is already in target state, nothing to do
|
|
|
|
if power_state == "On" and reset_type in ['On', 'ForceOn']:
|
|
|
|
return {'ret': True, 'changed': False}
|
|
|
|
if power_state == "Off" and reset_type in ['GracefulShutdown', 'ForceOff']:
|
|
|
|
return {'ret': True, 'changed': False}
|
|
|
|
|
2020-09-23 08:42:56 +02:00
|
|
|
# get the reset Action and target URI
|
|
|
|
if key not in data or action_name not in data[key]:
|
|
|
|
return {'ret': False, 'msg': 'Action %s not found' % action_name}
|
|
|
|
reset_action = data[key][action_name]
|
2020-03-09 10:11:07 +01:00
|
|
|
if 'target' not in reset_action:
|
|
|
|
return {'ret': False,
|
2020-09-23 08:42:56 +02:00
|
|
|
'msg': 'target URI missing from Action %s' % action_name}
|
2020-03-09 10:11:07 +01:00
|
|
|
action_uri = reset_action['target']
|
|
|
|
|
2020-06-11 20:24:41 +02:00
|
|
|
# get AllowableValues
|
|
|
|
ai = self._get_all_action_info_values(reset_action)
|
|
|
|
allowable_values = ai.get('ResetType', {}).get('AllowableValues', [])
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# map ResetType to an allowable value if needed
|
|
|
|
if reset_type not in allowable_values:
|
|
|
|
reset_type = self._map_reset_type(reset_type, allowable_values)
|
|
|
|
|
|
|
|
# define payload
|
|
|
|
payload = {'ResetType': reset_type}
|
|
|
|
|
|
|
|
# POST to Action URI
|
|
|
|
response = self.post_request(self.root_uri + action_uri, payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
2024-06-03 06:49:40 +02:00
|
|
|
|
|
|
|
# If requested to wait for the service to be available again, block
|
|
|
|
# until it's ready
|
|
|
|
if wait:
|
|
|
|
elapsed_time = 0
|
|
|
|
start_time = time.time()
|
|
|
|
# Start with a large enough sleep. Some services will process new
|
|
|
|
# requests while in the middle of shutting down, thus breaking out
|
|
|
|
# early.
|
|
|
|
time.sleep(30)
|
|
|
|
|
|
|
|
# Periodically check for the service's availability.
|
|
|
|
while elapsed_time <= wait_timeout:
|
|
|
|
status = self.check_service_availability()
|
|
|
|
if status['available']:
|
|
|
|
# It's available; we're done
|
|
|
|
break
|
|
|
|
time.sleep(5)
|
|
|
|
elapsed_time = time.time() - start_time
|
|
|
|
|
|
|
|
if elapsed_time > wait_timeout:
|
|
|
|
# Exhausted the wait timer; error
|
|
|
|
return {'ret': False, 'changed': True,
|
|
|
|
'msg': 'The service did not become available after %d seconds' % wait_timeout}
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': True, 'changed': True}
|
|
|
|
|
2024-04-03 13:47:00 +02:00
|
|
|
def manager_reset_to_defaults(self, command):
|
|
|
|
return self.reset_to_defaults(command, self.manager_uri,
|
|
|
|
'#Manager.ResetToDefaults')
|
|
|
|
|
|
|
|
def reset_to_defaults(self, command, resource_uri, action_name):
|
|
|
|
key = "Actions"
|
|
|
|
reset_type_values = ['ResetAll',
|
|
|
|
'PreserveNetworkAndUsers',
|
|
|
|
'PreserveNetwork']
|
|
|
|
|
|
|
|
if command not in reset_type_values:
|
|
|
|
return {'ret': False, 'msg': 'Invalid Command (%s)' % command}
|
|
|
|
|
|
|
|
# read the resource and get the current power state
|
|
|
|
response = self.get_request(self.root_uri + resource_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# get the reset Action and target URI
|
|
|
|
if key not in data or action_name not in data[key]:
|
|
|
|
return {'ret': False, 'msg': 'Action %s not found' % action_name}
|
|
|
|
reset_action = data[key][action_name]
|
|
|
|
if 'target' not in reset_action:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': 'target URI missing from Action %s' % action_name}
|
|
|
|
action_uri = reset_action['target']
|
|
|
|
|
|
|
|
# get AllowableValues
|
|
|
|
ai = self._get_all_action_info_values(reset_action)
|
|
|
|
allowable_values = ai.get('ResetType', {}).get('AllowableValues', [])
|
|
|
|
|
|
|
|
# map ResetType to an allowable value if needed
|
|
|
|
if allowable_values and command not in allowable_values:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': 'Specified reset type (%s) not supported '
|
|
|
|
'by service. Supported types: %s' %
|
|
|
|
(command, allowable_values)}
|
|
|
|
|
|
|
|
# define payload
|
|
|
|
payload = {'ResetType': command}
|
|
|
|
|
|
|
|
# POST to Action URI
|
|
|
|
response = self.post_request(self.root_uri + action_uri, payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
return {'ret': True, 'changed': True}
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def _find_account_uri(self, username=None, acct_id=None):
|
|
|
|
if not any((username, acct_id)):
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide either account_id or account_username'}
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + self.accounts_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
uris = [a.get('@odata.id') for a in data.get('Members', []) if
|
|
|
|
a.get('@odata.id')]
|
|
|
|
for uri in uris:
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
continue
|
|
|
|
data = response['data']
|
|
|
|
headers = response['headers']
|
|
|
|
if username:
|
|
|
|
if username == data.get('UserName'):
|
|
|
|
return {'ret': True, 'data': data,
|
|
|
|
'headers': headers, 'uri': uri}
|
|
|
|
if acct_id:
|
|
|
|
if acct_id == data.get('Id'):
|
|
|
|
return {'ret': True, 'data': data,
|
|
|
|
'headers': headers, 'uri': uri}
|
|
|
|
|
|
|
|
return {'ret': False, 'no_match': True, 'msg':
|
|
|
|
'No account with the given account_id or account_username found'}
|
|
|
|
|
|
|
|
def _find_empty_account_slot(self):
|
|
|
|
response = self.get_request(self.root_uri + self.accounts_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
uris = [a.get('@odata.id') for a in data.get('Members', []) if
|
|
|
|
a.get('@odata.id')]
|
|
|
|
if uris:
|
|
|
|
# first slot may be reserved, so move to end of list
|
|
|
|
uris += [uris.pop(0)]
|
|
|
|
for uri in uris:
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
continue
|
|
|
|
data = response['data']
|
|
|
|
headers = response['headers']
|
|
|
|
if data.get('UserName') == "" and not data.get('Enabled', True):
|
|
|
|
return {'ret': True, 'data': data,
|
|
|
|
'headers': headers, 'uri': uri}
|
|
|
|
|
|
|
|
return {'ret': False, 'no_match': True, 'msg':
|
|
|
|
'No empty account slot found'}
|
|
|
|
|
|
|
|
def list_users(self):
|
|
|
|
result = {}
|
|
|
|
# listing all users has always been slower than other operations, why?
|
|
|
|
user_list = []
|
|
|
|
users_results = []
|
|
|
|
# Get these entries, but does not fail if not found
|
2023-07-12 19:48:29 +02:00
|
|
|
properties = ['Id', 'Name', 'UserName', 'RoleId', 'Locked', 'Enabled',
|
|
|
|
'AccountTypes', 'OEMAccountTypes']
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + self.accounts_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for users in data.get('Members', []):
|
|
|
|
user_list.append(users[u'@odata.id']) # user_list[] are URIs
|
|
|
|
|
|
|
|
# for each user, get details
|
|
|
|
for uri in user_list:
|
|
|
|
user = {}
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
user[property] = data[property]
|
|
|
|
|
2023-06-26 23:29:51 +02:00
|
|
|
# Filter out empty account slots
|
|
|
|
# An empty account slot can be detected if the username is an empty
|
|
|
|
# string and if the account is disabled
|
|
|
|
if user.get('UserName', '') == '' and not user.get('Enabled', False):
|
|
|
|
continue
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
users_results.append(user)
|
|
|
|
result["entries"] = users_results
|
|
|
|
return result
|
|
|
|
|
|
|
|
def add_user_via_patch(self, user):
|
|
|
|
if user.get('account_id'):
|
|
|
|
# If Id slot specified, use it
|
|
|
|
response = self._find_account_uri(acct_id=user.get('account_id'))
|
|
|
|
else:
|
|
|
|
# Otherwise find first empty slot
|
|
|
|
response = self._find_empty_account_slot()
|
|
|
|
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
uri = response['uri']
|
|
|
|
payload = {}
|
|
|
|
if user.get('account_username'):
|
|
|
|
payload['UserName'] = user.get('account_username')
|
|
|
|
if user.get('account_password'):
|
|
|
|
payload['Password'] = user.get('account_password')
|
|
|
|
if user.get('account_roleid'):
|
|
|
|
payload['RoleId'] = user.get('account_roleid')
|
2023-07-12 19:48:29 +02:00
|
|
|
if user.get('account_accounttypes'):
|
|
|
|
payload['AccountTypes'] = user.get('account_accounttypes')
|
|
|
|
if user.get('account_oemaccounttypes'):
|
|
|
|
payload['OEMAccountTypes'] = user.get('account_oemaccounttypes')
|
2022-11-02 07:40:21 +01:00
|
|
|
return self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def add_user(self, user):
|
|
|
|
if not user.get('account_username'):
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide account_username for AddUser command'}
|
|
|
|
|
|
|
|
response = self._find_account_uri(username=user.get('account_username'))
|
|
|
|
if response['ret']:
|
|
|
|
# account_username already exists, nothing to do
|
|
|
|
return {'ret': True, 'changed': False}
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + self.accounts_uri)
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
headers = response['headers']
|
|
|
|
|
|
|
|
if 'allow' in headers:
|
|
|
|
methods = [m.strip() for m in headers.get('allow').split(',')]
|
|
|
|
if 'POST' not in methods:
|
|
|
|
# if Allow header present and POST not listed, add via PATCH
|
|
|
|
return self.add_user_via_patch(user)
|
|
|
|
|
|
|
|
payload = {}
|
|
|
|
if user.get('account_username'):
|
|
|
|
payload['UserName'] = user.get('account_username')
|
|
|
|
if user.get('account_password'):
|
|
|
|
payload['Password'] = user.get('account_password')
|
|
|
|
if user.get('account_roleid'):
|
|
|
|
payload['RoleId'] = user.get('account_roleid')
|
2023-07-12 19:48:29 +02:00
|
|
|
if user.get('account_accounttypes'):
|
|
|
|
payload['AccountTypes'] = user.get('account_accounttypes')
|
|
|
|
if user.get('account_oemaccounttypes'):
|
|
|
|
payload['OEMAccountTypes'] = user.get('account_oemaccounttypes')
|
2021-09-16 22:20:49 +02:00
|
|
|
if user.get('account_id'):
|
|
|
|
payload['Id'] = user.get('account_id')
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
response = self.post_request(self.root_uri + self.accounts_uri, payload)
|
|
|
|
if not response['ret']:
|
|
|
|
if response.get('status') == 405:
|
|
|
|
# if POST returned a 405, try to add via PATCH
|
|
|
|
return self.add_user_via_patch(user)
|
|
|
|
else:
|
|
|
|
return response
|
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def enable_user(self, user):
|
|
|
|
response = self._find_account_uri(username=user.get('account_username'),
|
|
|
|
acct_id=user.get('account_id'))
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
uri = response['uri']
|
|
|
|
|
|
|
|
payload = {'Enabled': True}
|
2022-11-02 07:40:21 +01:00
|
|
|
return self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def delete_user_via_patch(self, user, uri=None, data=None):
|
|
|
|
if not uri:
|
|
|
|
response = self._find_account_uri(username=user.get('account_username'),
|
|
|
|
acct_id=user.get('account_id'))
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
uri = response['uri']
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
payload = {'UserName': ''}
|
|
|
|
if data.get('Enabled', False):
|
|
|
|
payload['Enabled'] = False
|
2022-11-02 07:40:21 +01:00
|
|
|
return self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def delete_user(self, user):
|
|
|
|
response = self._find_account_uri(username=user.get('account_username'),
|
|
|
|
acct_id=user.get('account_id'))
|
|
|
|
if not response['ret']:
|
|
|
|
if response.get('no_match'):
|
|
|
|
# account does not exist, nothing to do
|
|
|
|
return {'ret': True, 'changed': False}
|
|
|
|
else:
|
|
|
|
# some error encountered
|
|
|
|
return response
|
|
|
|
|
|
|
|
uri = response['uri']
|
|
|
|
headers = response['headers']
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if 'allow' in headers:
|
|
|
|
methods = [m.strip() for m in headers.get('allow').split(',')]
|
|
|
|
if 'DELETE' not in methods:
|
|
|
|
# if Allow header present and DELETE not listed, del via PATCH
|
|
|
|
return self.delete_user_via_patch(user, uri=uri, data=data)
|
|
|
|
|
|
|
|
response = self.delete_request(self.root_uri + uri)
|
|
|
|
if not response['ret']:
|
|
|
|
if response.get('status') == 405:
|
|
|
|
# if DELETE returned a 405, try to delete via PATCH
|
|
|
|
return self.delete_user_via_patch(user, uri=uri, data=data)
|
|
|
|
else:
|
|
|
|
return response
|
|
|
|
return {'ret': True}
|
|
|
|
|
|
|
|
def disable_user(self, user):
|
|
|
|
response = self._find_account_uri(username=user.get('account_username'),
|
|
|
|
acct_id=user.get('account_id'))
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
uri = response['uri']
|
2020-03-09 10:11:07 +01:00
|
|
|
payload = {'Enabled': False}
|
2022-11-02 07:40:21 +01:00
|
|
|
return self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def update_user_role(self, user):
|
|
|
|
if not user.get('account_roleid'):
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide account_roleid for UpdateUserRole command'}
|
|
|
|
|
|
|
|
response = self._find_account_uri(username=user.get('account_username'),
|
|
|
|
acct_id=user.get('account_id'))
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
uri = response['uri']
|
|
|
|
payload = {'RoleId': user['account_roleid']}
|
|
|
|
return self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def update_user_password(self, user):
|
2022-11-02 07:40:21 +01:00
|
|
|
if not user.get('account_password'):
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide account_password for UpdateUserPassword command'}
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self._find_account_uri(username=user.get('account_username'),
|
|
|
|
acct_id=user.get('account_id'))
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
2022-11-02 07:40:21 +01:00
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
uri = response['uri']
|
|
|
|
payload = {'Password': user['account_password']}
|
2022-11-02 07:40:21 +01:00
|
|
|
return self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def update_user_name(self, user):
|
|
|
|
if not user.get('account_updatename'):
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide account_updatename for UpdateUserName command'}
|
|
|
|
|
|
|
|
response = self._find_account_uri(username=user.get('account_username'),
|
|
|
|
acct_id=user.get('account_id'))
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
2022-11-02 07:40:21 +01:00
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
uri = response['uri']
|
|
|
|
payload = {'UserName': user['account_updatename']}
|
2022-11-02 07:40:21 +01:00
|
|
|
return self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def update_accountservice_properties(self, user):
|
2022-11-02 07:40:21 +01:00
|
|
|
account_properties = user.get('account_properties')
|
|
|
|
if account_properties is None:
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide account_properties for UpdateAccountServiceProperties command'}
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Find the AccountService resource
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + self.service_root)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
accountservice_uri = data.get("AccountService", {}).get("@odata.id")
|
|
|
|
if accountservice_uri is None:
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': False, 'msg': "AccountService resource not found"}
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Perform a PATCH on the AccountService resource with the requested properties
|
|
|
|
resp = self.patch_request(self.root_uri + accountservice_uri, account_properties, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Modified account service'
|
|
|
|
return resp
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def get_sessions(self):
|
|
|
|
result = {}
|
|
|
|
# listing all users has always been slower than other operations, why?
|
|
|
|
session_list = []
|
|
|
|
sessions_results = []
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['Description', 'Id', 'Name', 'UserName']
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + self.sessions_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for sessions in data[u'Members']:
|
|
|
|
session_list.append(sessions[u'@odata.id']) # session_list[] are URIs
|
|
|
|
|
|
|
|
# for each session, get details
|
|
|
|
for uri in session_list:
|
|
|
|
session = {}
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
session[property] = data[property]
|
|
|
|
|
|
|
|
sessions_results.append(session)
|
|
|
|
result["entries"] = sessions_results
|
|
|
|
return result
|
|
|
|
|
|
|
|
def clear_sessions(self):
|
|
|
|
response = self.get_request(self.root_uri + self.sessions_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# if no active sessions, return as success
|
|
|
|
if data['Members@odata.count'] == 0:
|
2022-11-02 07:40:21 +01:00
|
|
|
return {'ret': True, 'changed': False, 'msg': "There are no active sessions"}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# loop to delete every active session
|
|
|
|
for session in data[u'Members']:
|
|
|
|
response = self.delete_request(self.root_uri + session[u'@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
return {'ret': True, 'changed': True, 'msg': "Cleared all sessions successfully"}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2021-03-19 21:14:33 +01:00
|
|
|
def create_session(self):
|
|
|
|
if not self.creds.get('user') or not self.creds.get('pswd'):
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide the username and password parameters for '
|
|
|
|
'the CreateSession command'}
|
|
|
|
|
|
|
|
payload = {
|
|
|
|
'UserName': self.creds['user'],
|
|
|
|
'Password': self.creds['pswd']
|
|
|
|
}
|
|
|
|
response = self.post_request(self.root_uri + self.sessions_uri, payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
headers = response['headers']
|
|
|
|
if 'x-auth-token' not in headers:
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'The service did not return the X-Auth-Token header in '
|
|
|
|
'the response from the Sessions collection POST'}
|
|
|
|
|
|
|
|
if 'location' not in headers:
|
|
|
|
self.module.warn(
|
|
|
|
'The service did not return the Location header for the '
|
|
|
|
'session URL in the response from the Sessions collection '
|
|
|
|
'POST')
|
|
|
|
session_uri = None
|
|
|
|
else:
|
|
|
|
session_uri = urlparse(headers.get('location')).path
|
|
|
|
|
|
|
|
session = dict()
|
|
|
|
session['token'] = headers.get('x-auth-token')
|
|
|
|
session['uri'] = session_uri
|
|
|
|
return {'ret': True, 'changed': True, 'session': session,
|
|
|
|
'msg': 'Session created successfully'}
|
|
|
|
|
|
|
|
def delete_session(self, session_uri):
|
|
|
|
if not session_uri:
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide the session_uri parameter for the '
|
|
|
|
'DeleteSession command'}
|
|
|
|
|
|
|
|
response = self.delete_request(self.root_uri + session_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
return {'ret': True, 'changed': True,
|
|
|
|
'msg': 'Session deleted successfully'}
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def get_firmware_update_capabilities(self):
|
|
|
|
result = {}
|
|
|
|
response = self.get_request(self.root_uri + self.update_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
result['ret'] = True
|
|
|
|
|
|
|
|
result['entries'] = {}
|
|
|
|
|
|
|
|
data = response['data']
|
|
|
|
|
2024-04-12 07:42:55 +02:00
|
|
|
result['multipart_supported'] = 'MultipartHttpPushUri' in data
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
if "Actions" in data:
|
|
|
|
actions = data['Actions']
|
|
|
|
if len(actions) > 0:
|
|
|
|
for key in actions.keys():
|
|
|
|
action = actions.get(key)
|
|
|
|
if 'title' in action:
|
|
|
|
title = action['title']
|
|
|
|
else:
|
|
|
|
title = key
|
|
|
|
result['entries'][title] = action.get('TransferProtocol@Redfish.AllowableValues',
|
|
|
|
["Key TransferProtocol@Redfish.AllowableValues not found"])
|
|
|
|
else:
|
|
|
|
return {'ret': "False", 'msg': "Actions list is empty."}
|
|
|
|
else:
|
|
|
|
return {'ret': "False", 'msg': "Key Actions not found."}
|
|
|
|
return result
|
|
|
|
|
|
|
|
def _software_inventory(self, uri):
|
|
|
|
result = {}
|
|
|
|
result['entries'] = []
|
2023-07-28 21:01:10 +02:00
|
|
|
|
|
|
|
while uri:
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
2020-03-09 10:11:07 +01:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
2023-07-28 21:01:10 +02:00
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
data = response['data']
|
2023-07-28 21:01:10 +02:00
|
|
|
if data.get('Members@odata.nextLink'):
|
|
|
|
uri = data.get('Members@odata.nextLink')
|
|
|
|
else:
|
|
|
|
uri = None
|
|
|
|
|
|
|
|
for member in data[u'Members']:
|
|
|
|
fw_uri = self.root_uri + member[u'@odata.id']
|
|
|
|
# Get details for each software or firmware member
|
|
|
|
response = self.get_request(fw_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
software = {}
|
|
|
|
# Get these standard properties if present
|
|
|
|
for key in ['Name', 'Id', 'Status', 'Version', 'Updateable',
|
|
|
|
'SoftwareId', 'LowestSupportedVersion', 'Manufacturer',
|
|
|
|
'ReleaseDate']:
|
|
|
|
if key in data:
|
|
|
|
software[key] = data.get(key)
|
|
|
|
result['entries'].append(software)
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
return result
|
|
|
|
|
|
|
|
def get_firmware_inventory(self):
|
|
|
|
if self.firmware_uri is None:
|
|
|
|
return {'ret': False, 'msg': 'No FirmwareInventory resource found'}
|
|
|
|
else:
|
|
|
|
return self._software_inventory(self.firmware_uri)
|
|
|
|
|
|
|
|
def get_software_inventory(self):
|
|
|
|
if self.software_uri is None:
|
|
|
|
return {'ret': False, 'msg': 'No SoftwareInventory resource found'}
|
|
|
|
else:
|
|
|
|
return self._software_inventory(self.software_uri)
|
|
|
|
|
2022-11-23 07:46:39 +01:00
|
|
|
def _operation_results(self, response, data, handle=None):
|
|
|
|
"""
|
|
|
|
Builds the results for an operation from task, job, or action response.
|
|
|
|
|
|
|
|
:param response: HTTP response object
|
|
|
|
:param data: HTTP response data
|
|
|
|
:param handle: The task or job handle that was last used
|
|
|
|
:return: dict containing operation results
|
|
|
|
"""
|
|
|
|
|
|
|
|
operation_results = {'status': None, 'messages': [], 'handle': None, 'ret': True,
|
|
|
|
'resets_requested': []}
|
|
|
|
|
|
|
|
if response.status == 204:
|
|
|
|
# No content; successful, but nothing to return
|
|
|
|
# Use the Redfish "Completed" enum from TaskState for the operation status
|
|
|
|
operation_results['status'] = 'Completed'
|
|
|
|
else:
|
|
|
|
# Parse the response body for details
|
|
|
|
|
|
|
|
# Determine the next handle, if any
|
|
|
|
operation_results['handle'] = handle
|
|
|
|
if response.status == 202:
|
|
|
|
# Task generated; get the task monitor URI
|
|
|
|
operation_results['handle'] = response.getheader('Location', handle)
|
|
|
|
|
|
|
|
# Pull out the status and messages based on the body format
|
|
|
|
if data is not None:
|
|
|
|
response_type = data.get('@odata.type', '')
|
|
|
|
if response_type.startswith('#Task.') or response_type.startswith('#Job.'):
|
|
|
|
# Task and Job have similar enough structures to treat the same
|
|
|
|
operation_results['status'] = data.get('TaskState', data.get('JobState'))
|
|
|
|
operation_results['messages'] = data.get('Messages', [])
|
|
|
|
else:
|
|
|
|
# Error response body, which is a bit of a misnomer since it's used in successful action responses
|
|
|
|
operation_results['status'] = 'Completed'
|
|
|
|
if response.status >= 400:
|
|
|
|
operation_results['status'] = 'Exception'
|
|
|
|
operation_results['messages'] = data.get('error', {}).get('@Message.ExtendedInfo', [])
|
|
|
|
else:
|
|
|
|
# No response body (or malformed); build based on status code
|
|
|
|
operation_results['status'] = 'Completed'
|
|
|
|
if response.status == 202:
|
|
|
|
operation_results['status'] = 'New'
|
|
|
|
elif response.status >= 400:
|
|
|
|
operation_results['status'] = 'Exception'
|
|
|
|
|
|
|
|
# Clear out the handle if the operation is complete
|
|
|
|
if operation_results['status'] in ['Completed', 'Cancelled', 'Exception', 'Killed']:
|
|
|
|
operation_results['handle'] = None
|
|
|
|
|
|
|
|
# Scan the messages to see if next steps are needed
|
|
|
|
for message in operation_results['messages']:
|
2023-11-06 05:49:48 +01:00
|
|
|
message_id = message.get('MessageId')
|
|
|
|
if message_id is None:
|
|
|
|
# While this is invalid, treat the lack of a MessageId as "no message"
|
|
|
|
continue
|
2022-11-23 07:46:39 +01:00
|
|
|
|
|
|
|
if message_id.startswith('Update.1.') and message_id.endswith('.OperationTransitionedToJob'):
|
|
|
|
# Operation rerouted to a job; update the status and handle
|
|
|
|
operation_results['status'] = 'New'
|
|
|
|
operation_results['handle'] = message['MessageArgs'][0]
|
|
|
|
operation_results['resets_requested'] = []
|
|
|
|
# No need to process other messages in this case
|
|
|
|
break
|
|
|
|
|
|
|
|
if message_id.startswith('Base.1.') and message_id.endswith('.ResetRequired'):
|
|
|
|
# A reset to some device is needed to continue the update
|
|
|
|
reset = {'uri': message['MessageArgs'][0], 'type': message['MessageArgs'][1]}
|
|
|
|
operation_results['resets_requested'].append(reset)
|
|
|
|
|
|
|
|
return operation_results
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def simple_update(self, update_opts):
|
|
|
|
image_uri = update_opts.get('update_image_uri')
|
|
|
|
protocol = update_opts.get('update_protocol')
|
|
|
|
targets = update_opts.get('update_targets')
|
|
|
|
creds = update_opts.get('update_creds')
|
2022-11-23 07:46:39 +01:00
|
|
|
apply_time = update_opts.get('update_apply_time')
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
if not image_uri:
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must specify update_image_uri for the SimpleUpdate command'}
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + self.update_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'Actions' not in data:
|
|
|
|
return {'ret': False, 'msg': 'Service does not support SimpleUpdate'}
|
|
|
|
if '#UpdateService.SimpleUpdate' not in data['Actions']:
|
|
|
|
return {'ret': False, 'msg': 'Service does not support SimpleUpdate'}
|
|
|
|
action = data['Actions']['#UpdateService.SimpleUpdate']
|
|
|
|
if 'target' not in action:
|
|
|
|
return {'ret': False, 'msg': 'Service does not support SimpleUpdate'}
|
|
|
|
update_uri = action['target']
|
|
|
|
if protocol:
|
|
|
|
default_values = ['CIFS', 'FTP', 'SFTP', 'HTTP', 'HTTPS', 'NSF',
|
|
|
|
'SCP', 'TFTP', 'OEM', 'NFS']
|
|
|
|
allowable_values = self._get_allowable_values(action,
|
|
|
|
'TransferProtocol',
|
|
|
|
default_values)
|
|
|
|
if protocol not in allowable_values:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': 'Specified update_protocol (%s) not supported '
|
|
|
|
'by service. Supported protocols: %s' %
|
|
|
|
(protocol, allowable_values)}
|
|
|
|
if targets:
|
|
|
|
allowable_values = self._get_allowable_values(action, 'Targets')
|
|
|
|
if allowable_values:
|
|
|
|
for target in targets:
|
|
|
|
if target not in allowable_values:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': 'Specified target (%s) not supported '
|
|
|
|
'by service. Supported targets: %s' %
|
|
|
|
(target, allowable_values)}
|
|
|
|
|
|
|
|
payload = {
|
|
|
|
'ImageURI': image_uri
|
|
|
|
}
|
|
|
|
if protocol:
|
|
|
|
payload["TransferProtocol"] = protocol
|
|
|
|
if targets:
|
|
|
|
payload["Targets"] = targets
|
|
|
|
if creds:
|
|
|
|
if creds.get('username'):
|
|
|
|
payload["Username"] = creds.get('username')
|
|
|
|
if creds.get('password'):
|
|
|
|
payload["Password"] = creds.get('password')
|
2022-11-23 07:46:39 +01:00
|
|
|
if apply_time:
|
|
|
|
payload["@Redfish.OperationApplyTime"] = apply_time
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.post_request(self.root_uri + update_uri, payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
return {'ret': True, 'changed': True,
|
2022-11-23 07:46:39 +01:00
|
|
|
'msg': "SimpleUpdate requested",
|
|
|
|
'update_status': self._operation_results(response['resp'], response['data'])}
|
2023-06-05 21:56:44 +02:00
|
|
|
|
|
|
|
def multipath_http_push_update(self, update_opts):
|
|
|
|
"""
|
|
|
|
Provides a software update via the URI specified by the
|
|
|
|
MultipartHttpPushUri property. Callers should adjust the 'timeout'
|
|
|
|
variable in the base object to accommodate the size of the image and
|
|
|
|
speed of the transfer. For example, a 200MB image will likely take
|
|
|
|
more than the default 10 second timeout.
|
|
|
|
|
|
|
|
:param update_opts: The parameters for the update operation
|
|
|
|
:return: dict containing the response of the update request
|
|
|
|
"""
|
|
|
|
image_file = update_opts.get('update_image_file')
|
|
|
|
targets = update_opts.get('update_targets')
|
|
|
|
apply_time = update_opts.get('update_apply_time')
|
2023-10-02 21:46:29 +02:00
|
|
|
oem_params = update_opts.get('update_oem_params')
|
2023-06-05 21:56:44 +02:00
|
|
|
|
|
|
|
# Ensure the image file is provided
|
|
|
|
if not image_file:
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must specify update_image_file for the MultipartHTTPPushUpdate command'}
|
|
|
|
if not os.path.isfile(image_file):
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must specify a valid file for the MultipartHTTPPushUpdate command'}
|
|
|
|
try:
|
|
|
|
with open(image_file, 'rb') as f:
|
|
|
|
image_payload = f.read()
|
|
|
|
except Exception as e:
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Could not read file %s' % image_file}
|
|
|
|
|
|
|
|
# Check that multipart HTTP push updates are supported
|
|
|
|
response = self.get_request(self.root_uri + self.update_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'MultipartHttpPushUri' not in data:
|
|
|
|
return {'ret': False, 'msg': 'Service does not support MultipartHttpPushUri'}
|
|
|
|
update_uri = data['MultipartHttpPushUri']
|
|
|
|
|
|
|
|
# Assemble the JSON payload portion of the request
|
|
|
|
payload = {"@Redfish.OperationApplyTime": "Immediate"}
|
|
|
|
if targets:
|
|
|
|
payload["Targets"] = targets
|
|
|
|
if apply_time:
|
|
|
|
payload["@Redfish.OperationApplyTime"] = apply_time
|
2023-10-02 21:46:29 +02:00
|
|
|
if oem_params:
|
|
|
|
payload["Oem"] = oem_params
|
2023-06-05 21:56:44 +02:00
|
|
|
multipart_payload = {
|
|
|
|
'UpdateParameters': {'content': json.dumps(payload), 'mime_type': 'application/json'},
|
|
|
|
'UpdateFile': {'filename': image_file, 'content': image_payload, 'mime_type': 'application/octet-stream'}
|
|
|
|
}
|
|
|
|
response = self.post_request(self.root_uri + update_uri, multipart_payload, multipart=True)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
return {'ret': True, 'changed': True,
|
|
|
|
'msg': "MultipartHTTPPushUpdate requested",
|
|
|
|
'update_status': self._operation_results(response['resp'], response['data'])}
|
2022-11-23 07:46:39 +01:00
|
|
|
|
|
|
|
def get_update_status(self, update_handle):
|
|
|
|
"""
|
|
|
|
Gets the status of an update operation.
|
|
|
|
|
|
|
|
:param handle: The task or job handle tracking the update
|
|
|
|
:return: dict containing the response of the update status
|
|
|
|
"""
|
|
|
|
|
|
|
|
if not update_handle:
|
|
|
|
return {'ret': False, 'msg': 'Must provide a handle tracking the update.'}
|
|
|
|
|
|
|
|
# Get the task or job tracking the update
|
2024-02-23 06:35:30 +01:00
|
|
|
response = self.get_request(self.root_uri + update_handle, allow_no_resp=True)
|
2022-11-23 07:46:39 +01:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
# Inspect the response to build the update status
|
|
|
|
return self._operation_results(response['resp'], response['data'], update_handle)
|
|
|
|
|
|
|
|
def perform_requested_update_operations(self, update_handle):
|
|
|
|
"""
|
|
|
|
Performs requested operations to allow the update to continue.
|
|
|
|
|
|
|
|
:param handle: The task or job handle tracking the update
|
|
|
|
:return: dict containing the result of the operations
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Get the current update status
|
|
|
|
update_status = self.get_update_status(update_handle)
|
|
|
|
if update_status['ret'] is False:
|
|
|
|
return update_status
|
|
|
|
|
|
|
|
changed = False
|
|
|
|
|
|
|
|
# Perform any requested updates
|
|
|
|
for reset in update_status['resets_requested']:
|
|
|
|
resp = self.post_request(self.root_uri + reset['uri'], {'ResetType': reset['type']})
|
|
|
|
if resp['ret'] is False:
|
|
|
|
# Override the 'changed' indicator since other resets may have
|
|
|
|
# been successful
|
|
|
|
resp['changed'] = changed
|
|
|
|
return resp
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
msg = 'No operations required for the update'
|
|
|
|
if changed:
|
|
|
|
# Will need to consider finetuning this message if the scope of the
|
|
|
|
# requested operations grow over time
|
|
|
|
msg = 'One or more components reset to continue the update'
|
|
|
|
return {'ret': True, 'changed': changed, 'msg': msg}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def get_bios_attributes(self, systems_uri):
|
|
|
|
result = {}
|
|
|
|
bios_attributes = {}
|
|
|
|
key = "Bios"
|
|
|
|
|
|
|
|
# Search for 'key' entry and extract URI from it
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if key not in data:
|
|
|
|
return {'ret': False, 'msg': "Key %s not found" % key}
|
|
|
|
|
|
|
|
bios_uri = data[key]["@odata.id"]
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + bios_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
for attribute in data[u'Attributes'].items():
|
|
|
|
bios_attributes[attribute[0]] = attribute[1]
|
|
|
|
result["entries"] = bios_attributes
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_bios_attributes(self):
|
|
|
|
return self.aggregate_systems(self.get_bios_attributes)
|
|
|
|
|
|
|
|
def _get_boot_options_dict(self, boot):
|
|
|
|
# Get these entries from BootOption, if present
|
|
|
|
properties = ['DisplayName', 'BootOptionReference']
|
|
|
|
|
|
|
|
# Retrieve BootOptions if present
|
|
|
|
if 'BootOptions' in boot and '@odata.id' in boot['BootOptions']:
|
|
|
|
boot_options_uri = boot['BootOptions']["@odata.id"]
|
|
|
|
# Get BootOptions resource
|
|
|
|
response = self.get_request(self.root_uri + boot_options_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return {}
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Retrieve Members array
|
|
|
|
if 'Members' not in data:
|
|
|
|
return {}
|
|
|
|
members = data['Members']
|
|
|
|
else:
|
|
|
|
members = []
|
|
|
|
|
|
|
|
# Build dict of BootOptions keyed by BootOptionReference
|
|
|
|
boot_options_dict = {}
|
|
|
|
for member in members:
|
|
|
|
if '@odata.id' not in member:
|
|
|
|
return {}
|
|
|
|
boot_option_uri = member['@odata.id']
|
|
|
|
response = self.get_request(self.root_uri + boot_option_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return {}
|
|
|
|
data = response['data']
|
|
|
|
if 'BootOptionReference' not in data:
|
|
|
|
return {}
|
|
|
|
boot_option_ref = data['BootOptionReference']
|
|
|
|
|
|
|
|
# fetch the props to display for this boot device
|
|
|
|
boot_props = {}
|
|
|
|
for prop in properties:
|
|
|
|
if prop in data:
|
|
|
|
boot_props[prop] = data[prop]
|
|
|
|
|
|
|
|
boot_options_dict[boot_option_ref] = boot_props
|
|
|
|
|
|
|
|
return boot_options_dict
|
|
|
|
|
|
|
|
def get_boot_order(self, systems_uri):
|
|
|
|
result = {}
|
|
|
|
|
|
|
|
# Retrieve System resource
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Confirm needed Boot properties are present
|
|
|
|
if 'Boot' not in data or 'BootOrder' not in data['Boot']:
|
|
|
|
return {'ret': False, 'msg': "Key BootOrder not found"}
|
|
|
|
|
|
|
|
boot = data['Boot']
|
|
|
|
boot_order = boot['BootOrder']
|
|
|
|
boot_options_dict = self._get_boot_options_dict(boot)
|
|
|
|
|
|
|
|
# Build boot device list
|
|
|
|
boot_device_list = []
|
|
|
|
for ref in boot_order:
|
|
|
|
boot_device_list.append(
|
|
|
|
boot_options_dict.get(ref, {'BootOptionReference': ref}))
|
|
|
|
|
|
|
|
result["entries"] = boot_device_list
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_boot_order(self):
|
|
|
|
return self.aggregate_systems(self.get_boot_order)
|
|
|
|
|
|
|
|
def get_boot_override(self, systems_uri):
|
|
|
|
result = {}
|
|
|
|
|
|
|
|
properties = ["BootSourceOverrideEnabled", "BootSourceOverrideTarget",
|
|
|
|
"BootSourceOverrideMode", "UefiTargetBootSourceOverride", "BootSourceOverrideTarget@Redfish.AllowableValues"]
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if 'Boot' not in data:
|
|
|
|
return {'ret': False, 'msg': "Key Boot not found"}
|
|
|
|
|
|
|
|
boot = data['Boot']
|
|
|
|
|
|
|
|
boot_overrides = {}
|
|
|
|
if "BootSourceOverrideEnabled" in boot:
|
|
|
|
if boot["BootSourceOverrideEnabled"] is not False:
|
|
|
|
for property in properties:
|
|
|
|
if property in boot:
|
|
|
|
if boot[property] is not None:
|
|
|
|
boot_overrides[property] = boot[property]
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': "No boot override is enabled."}
|
|
|
|
|
|
|
|
result['entries'] = boot_overrides
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_boot_override(self):
|
|
|
|
return self.aggregate_systems(self.get_boot_override)
|
|
|
|
|
|
|
|
def set_bios_default_settings(self):
|
2022-11-02 07:40:21 +01:00
|
|
|
# Find the Bios resource from the requested ComputerSystem resource
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
bios_uri = data.get('Bios', {}).get('@odata.id')
|
|
|
|
if bios_uri is None:
|
|
|
|
return {'ret': False, 'msg': 'Bios resource not found'}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Find the URI of the ResetBios action
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + bios_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
reset_bios_uri = data.get('Actions', {}).get('#Bios.ResetBios', {}).get('target')
|
|
|
|
if reset_bios_uri is None:
|
|
|
|
return {'ret': False, 'msg': 'ResetBios action not found'}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Perform the ResetBios action
|
|
|
|
response = self.post_request(self.root_uri + reset_bios_uri, {})
|
2020-03-09 10:11:07 +01:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
2022-11-02 07:40:21 +01:00
|
|
|
return {'ret': True, 'changed': True, 'msg': "BIOS set to default settings"}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2020-09-10 17:25:48 +02:00
|
|
|
def set_boot_override(self, boot_opts):
|
2022-11-02 07:40:21 +01:00
|
|
|
# Extract the requested boot override options
|
2020-09-10 17:25:48 +02:00
|
|
|
bootdevice = boot_opts.get('bootdevice')
|
|
|
|
uefi_target = boot_opts.get('uefi_target')
|
|
|
|
boot_next = boot_opts.get('boot_next')
|
|
|
|
override_enabled = boot_opts.get('override_enabled')
|
2021-08-04 19:53:43 +02:00
|
|
|
boot_override_mode = boot_opts.get('boot_override_mode')
|
2020-09-10 17:25:48 +02:00
|
|
|
if not bootdevice and override_enabled != 'Disabled':
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': False,
|
2020-09-10 17:25:48 +02:00
|
|
|
'msg': "bootdevice option required for temporary boot override"}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Get the current boot override options from the Boot property
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
boot = data.get('Boot')
|
|
|
|
if boot is None:
|
|
|
|
return {'ret': False, 'msg': "Boot property not found"}
|
|
|
|
cur_override_mode = boot.get('BootSourceOverrideMode')
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Check if the requested target is supported by the system
|
2021-07-16 19:02:34 +02:00
|
|
|
if override_enabled != 'Disabled':
|
|
|
|
annotation = 'BootSourceOverrideTarget@Redfish.AllowableValues'
|
|
|
|
if annotation in boot:
|
|
|
|
allowable_values = boot[annotation]
|
|
|
|
if isinstance(allowable_values, list) and bootdevice not in allowable_values:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Boot device %s not in list of allowable values (%s)" %
|
|
|
|
(bootdevice, allowable_values)}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Build the request payload based on the desired boot override options
|
2020-09-10 17:25:48 +02:00
|
|
|
if override_enabled == 'Disabled':
|
|
|
|
payload = {
|
|
|
|
'Boot': {
|
2022-11-02 07:40:21 +01:00
|
|
|
'BootSourceOverrideEnabled': override_enabled,
|
|
|
|
'BootSourceOverrideTarget': 'None'
|
2020-09-10 17:25:48 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
elif bootdevice == 'UefiTarget':
|
2020-03-09 10:11:07 +01:00
|
|
|
if not uefi_target:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "uefi_target option required to SetOneTimeBoot for UefiTarget"}
|
|
|
|
payload = {
|
|
|
|
'Boot': {
|
2020-09-10 17:25:48 +02:00
|
|
|
'BootSourceOverrideEnabled': override_enabled,
|
2020-03-09 10:11:07 +01:00
|
|
|
'BootSourceOverrideTarget': bootdevice,
|
|
|
|
'UefiTargetBootSourceOverride': uefi_target
|
|
|
|
}
|
|
|
|
}
|
2022-11-02 07:40:21 +01:00
|
|
|
# If needed, also specify UEFI mode
|
|
|
|
if cur_override_mode == 'Legacy':
|
|
|
|
payload['Boot']['BootSourceOverrideMode'] = 'UEFI'
|
2020-03-09 10:11:07 +01:00
|
|
|
elif bootdevice == 'UefiBootNext':
|
|
|
|
if not boot_next:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "boot_next option required to SetOneTimeBoot for UefiBootNext"}
|
|
|
|
payload = {
|
|
|
|
'Boot': {
|
2020-09-10 17:25:48 +02:00
|
|
|
'BootSourceOverrideEnabled': override_enabled,
|
2020-03-09 10:11:07 +01:00
|
|
|
'BootSourceOverrideTarget': bootdevice,
|
|
|
|
'BootNext': boot_next
|
|
|
|
}
|
|
|
|
}
|
2022-11-02 07:40:21 +01:00
|
|
|
# If needed, also specify UEFI mode
|
|
|
|
if cur_override_mode == 'Legacy':
|
|
|
|
payload['Boot']['BootSourceOverrideMode'] = 'UEFI'
|
2020-03-09 10:11:07 +01:00
|
|
|
else:
|
|
|
|
payload = {
|
|
|
|
'Boot': {
|
2020-09-10 17:25:48 +02:00
|
|
|
'BootSourceOverrideEnabled': override_enabled,
|
2020-03-09 10:11:07 +01:00
|
|
|
'BootSourceOverrideTarget': bootdevice
|
|
|
|
}
|
|
|
|
}
|
2021-10-22 07:29:26 +02:00
|
|
|
if boot_override_mode:
|
|
|
|
payload['Boot']['BootSourceOverrideMode'] = boot_override_mode
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Apply the requested boot override request
|
|
|
|
resp = self.patch_request(self.root_uri + self.systems_uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] is False:
|
|
|
|
# WORKAROUND
|
|
|
|
# Older Dell systems do not allow BootSourceOverrideEnabled to be
|
|
|
|
# specified with UefiTarget as the target device
|
|
|
|
vendor = self._get_vendor()['Vendor']
|
|
|
|
if vendor == 'Dell':
|
|
|
|
if bootdevice == 'UefiTarget' and override_enabled != 'Disabled':
|
|
|
|
payload['Boot'].pop('BootSourceOverrideEnabled', None)
|
|
|
|
resp = self.patch_request(self.root_uri + self.systems_uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Updated the boot override settings'
|
|
|
|
return resp
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def set_bios_attributes(self, attributes):
|
2022-11-02 07:40:21 +01:00
|
|
|
# Find the Bios resource from the requested ComputerSystem resource
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
bios_uri = data.get('Bios', {}).get('@odata.id')
|
|
|
|
if bios_uri is None:
|
|
|
|
return {'ret': False, 'msg': 'Bios resource not found'}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Get the current BIOS settings
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + bios_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Make a copy of the attributes dict
|
|
|
|
attrs_to_patch = dict(attributes)
|
2021-05-11 19:30:09 +02:00
|
|
|
# List to hold attributes not found
|
|
|
|
attrs_bad = {}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Check the attributes
|
2021-05-11 19:30:09 +02:00
|
|
|
for attr_name, attr_value in attributes.items():
|
|
|
|
# Check if attribute exists
|
|
|
|
if attr_name not in data[u'Attributes']:
|
|
|
|
# Remove and proceed to next attribute if this isn't valid
|
|
|
|
attrs_bad.update({attr_name: attr_value})
|
|
|
|
del attrs_to_patch[attr_name]
|
|
|
|
continue
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
# If already set to requested value, remove it from PATCH payload
|
2024-09-05 21:47:28 +02:00
|
|
|
if data[u'Attributes'][attr_name] == attr_value:
|
2021-05-11 19:30:09 +02:00
|
|
|
del attrs_to_patch[attr_name]
|
|
|
|
|
|
|
|
warning = ""
|
|
|
|
if attrs_bad:
|
2022-11-02 07:40:21 +01:00
|
|
|
warning = "Unsupported attributes %s" % (attrs_bad)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Return success w/ changed=False if no attrs need to be changed
|
|
|
|
if not attrs_to_patch:
|
|
|
|
return {'ret': True, 'changed': False,
|
2021-05-11 19:30:09 +02:00
|
|
|
'msg': "BIOS attributes already set",
|
|
|
|
'warning': warning}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Get the SettingsObject URI to apply the attributes
|
|
|
|
set_bios_attr_uri = data.get("@Redfish.Settings", {}).get("SettingsObject", {}).get("@odata.id")
|
|
|
|
if set_bios_attr_uri is None:
|
|
|
|
return {'ret': False, 'msg': "Settings resource for BIOS attributes not found."}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Construct payload and issue PATCH command
|
|
|
|
payload = {"Attributes": attrs_to_patch}
|
|
|
|
response = self.patch_request(self.root_uri + set_bios_attr_uri, payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
2021-05-11 19:30:09 +02:00
|
|
|
return {'ret': True, 'changed': True,
|
|
|
|
'msg': "Modified BIOS attributes %s" % (attrs_to_patch),
|
|
|
|
'warning': warning}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def set_boot_order(self, boot_list):
|
|
|
|
if not boot_list:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "boot_order list required for SetBootOrder command"}
|
|
|
|
|
|
|
|
systems_uri = self.systems_uri
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Confirm needed Boot properties are present
|
|
|
|
if 'Boot' not in data or 'BootOrder' not in data['Boot']:
|
|
|
|
return {'ret': False, 'msg': "Key BootOrder not found"}
|
|
|
|
|
|
|
|
boot = data['Boot']
|
|
|
|
boot_order = boot['BootOrder']
|
|
|
|
boot_options_dict = self._get_boot_options_dict(boot)
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Verify the requested boot options are valid
|
2020-03-09 10:11:07 +01:00
|
|
|
if boot_options_dict:
|
|
|
|
boot_option_references = boot_options_dict.keys()
|
|
|
|
for ref in boot_list:
|
|
|
|
if ref not in boot_option_references:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "BootOptionReference %s not found in BootOptions" % ref}
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Apply the boot order
|
2020-03-09 10:11:07 +01:00
|
|
|
payload = {
|
|
|
|
'Boot': {
|
|
|
|
'BootOrder': boot_list
|
|
|
|
}
|
|
|
|
}
|
2022-11-02 07:40:21 +01:00
|
|
|
resp = self.patch_request(self.root_uri + systems_uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Modified the boot order'
|
|
|
|
return resp
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def set_default_boot_order(self):
|
|
|
|
systems_uri = self.systems_uri
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# get the #ComputerSystem.SetDefaultBootOrder Action and target URI
|
|
|
|
action = '#ComputerSystem.SetDefaultBootOrder'
|
|
|
|
if 'Actions' not in data or action not in data['Actions']:
|
|
|
|
return {'ret': False, 'msg': 'Action %s not found' % action}
|
|
|
|
if 'target' not in data['Actions'][action]:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': 'target URI missing from Action %s' % action}
|
|
|
|
action_uri = data['Actions'][action]['target']
|
|
|
|
|
|
|
|
# POST to Action URI
|
|
|
|
payload = {}
|
|
|
|
response = self.post_request(self.root_uri + action_uri, payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
return {'ret': True, 'changed': True,
|
|
|
|
'msg': "BootOrder set to default"}
|
|
|
|
|
|
|
|
def get_chassis_inventory(self):
|
|
|
|
result = {}
|
|
|
|
chassis_results = []
|
|
|
|
|
|
|
|
# Get these entries, but does not fail if not found
|
2021-01-29 07:17:57 +01:00
|
|
|
properties = ['Name', 'Id', 'ChassisType', 'PartNumber', 'AssetTag',
|
2020-03-09 10:11:07 +01:00
|
|
|
'Manufacturer', 'IndicatorLED', 'SerialNumber', 'Model']
|
|
|
|
|
|
|
|
# Go through list
|
|
|
|
for chassis_uri in self.chassis_uris:
|
|
|
|
response = self.get_request(self.root_uri + chassis_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
chassis_result = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
chassis_result[property] = data[property]
|
|
|
|
chassis_results.append(chassis_result)
|
|
|
|
|
|
|
|
result["entries"] = chassis_results
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_fan_inventory(self):
|
|
|
|
result = {}
|
|
|
|
fan_results = []
|
|
|
|
key = "Thermal"
|
|
|
|
# Get these entries, but does not fail if not found
|
2021-01-29 07:17:57 +01:00
|
|
|
properties = ['Name', 'FanName', 'Reading', 'ReadingUnits', 'Status']
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Go through list
|
|
|
|
for chassis_uri in self.chassis_uris:
|
|
|
|
response = self.get_request(self.root_uri + chassis_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
if key in data:
|
|
|
|
# match: found an entry for "Thermal" information = fans
|
|
|
|
thermal_uri = data[key]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + thermal_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
2021-12-11 21:27:09 +01:00
|
|
|
# Checking if fans are present
|
|
|
|
if u'Fans' in data:
|
|
|
|
for device in data[u'Fans']:
|
|
|
|
fan = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in device:
|
|
|
|
fan[property] = device[property]
|
|
|
|
fan_results.append(fan)
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': "No Fans present"}
|
2020-03-09 10:11:07 +01:00
|
|
|
result["entries"] = fan_results
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_chassis_power(self):
|
|
|
|
result = {}
|
|
|
|
key = "Power"
|
|
|
|
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['Name', 'PowerAllocatedWatts',
|
|
|
|
'PowerAvailableWatts', 'PowerCapacityWatts',
|
|
|
|
'PowerConsumedWatts', 'PowerMetrics',
|
|
|
|
'PowerRequestedWatts', 'RelatedItem', 'Status']
|
|
|
|
|
|
|
|
chassis_power_results = []
|
|
|
|
# Go through list
|
|
|
|
for chassis_uri in self.chassis_uris:
|
|
|
|
chassis_power_result = {}
|
|
|
|
response = self.get_request(self.root_uri + chassis_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
if key in data:
|
|
|
|
response = self.get_request(self.root_uri + data[key]['@odata.id'])
|
|
|
|
data = response['data']
|
|
|
|
if 'PowerControl' in data:
|
|
|
|
if len(data['PowerControl']) > 0:
|
|
|
|
data = data['PowerControl'][0]
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
chassis_power_result[property] = data[property]
|
|
|
|
chassis_power_results.append(chassis_power_result)
|
|
|
|
|
2022-06-30 20:36:01 +02:00
|
|
|
if len(chassis_power_results) > 0:
|
|
|
|
result['entries'] = chassis_power_results
|
|
|
|
return result
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': 'Power information not found.'}
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
def get_chassis_thermals(self):
|
|
|
|
result = {}
|
|
|
|
sensors = []
|
|
|
|
key = "Thermal"
|
|
|
|
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['Name', 'PhysicalContext', 'UpperThresholdCritical',
|
|
|
|
'UpperThresholdFatal', 'UpperThresholdNonCritical',
|
|
|
|
'LowerThresholdCritical', 'LowerThresholdFatal',
|
|
|
|
'LowerThresholdNonCritical', 'MaxReadingRangeTemp',
|
|
|
|
'MinReadingRangeTemp', 'ReadingCelsius', 'RelatedItem',
|
2021-08-19 21:13:10 +02:00
|
|
|
'SensorNumber', 'Status']
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Go through list
|
|
|
|
for chassis_uri in self.chassis_uris:
|
|
|
|
response = self.get_request(self.root_uri + chassis_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
if key in data:
|
|
|
|
thermal_uri = data[key]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + thermal_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
if "Temperatures" in data:
|
|
|
|
for sensor in data[u'Temperatures']:
|
|
|
|
sensor_result = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in sensor:
|
|
|
|
if sensor[property] is not None:
|
|
|
|
sensor_result[property] = sensor[property]
|
|
|
|
sensors.append(sensor_result)
|
|
|
|
|
|
|
|
if sensors is None:
|
|
|
|
return {'ret': False, 'msg': 'Key Temperatures was not found.'}
|
|
|
|
|
|
|
|
result['entries'] = sensors
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_cpu_inventory(self, systems_uri):
|
|
|
|
result = {}
|
|
|
|
cpu_list = []
|
|
|
|
cpu_results = []
|
|
|
|
key = "Processors"
|
|
|
|
# Get these entries, but does not fail if not found
|
2021-01-29 07:17:57 +01:00
|
|
|
properties = ['Id', 'Name', 'Manufacturer', 'Model', 'MaxSpeedMHz',
|
2023-07-07 22:24:45 +02:00
|
|
|
'ProcessorArchitecture', 'TotalCores', 'TotalThreads', 'Status']
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Search for 'key' entry and extract URI from it
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if key not in data:
|
|
|
|
return {'ret': False, 'msg': "Key %s not found" % key}
|
|
|
|
|
|
|
|
processors_uri = data[key]["@odata.id"]
|
|
|
|
|
|
|
|
# Get a list of all CPUs and build respective URIs
|
|
|
|
response = self.get_request(self.root_uri + processors_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for cpu in data[u'Members']:
|
|
|
|
cpu_list.append(cpu[u'@odata.id'])
|
|
|
|
|
|
|
|
for c in cpu_list:
|
|
|
|
cpu = {}
|
|
|
|
uri = self.root_uri + c
|
|
|
|
response = self.get_request(uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
cpu[property] = data[property]
|
|
|
|
|
|
|
|
cpu_results.append(cpu)
|
|
|
|
result["entries"] = cpu_results
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_cpu_inventory(self):
|
|
|
|
return self.aggregate_systems(self.get_cpu_inventory)
|
|
|
|
|
|
|
|
def get_memory_inventory(self, systems_uri):
|
|
|
|
result = {}
|
|
|
|
memory_list = []
|
|
|
|
memory_results = []
|
|
|
|
key = "Memory"
|
|
|
|
# Get these entries, but does not fail if not found
|
2021-01-29 07:17:57 +01:00
|
|
|
properties = ['Id', 'SerialNumber', 'MemoryDeviceType', 'PartNumber',
|
2020-03-09 10:11:07 +01:00
|
|
|
'MemoryLocation', 'RankCount', 'CapacityMiB', 'OperatingMemoryModes', 'Status', 'Manufacturer', 'Name']
|
|
|
|
|
|
|
|
# Search for 'key' entry and extract URI from it
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if key not in data:
|
|
|
|
return {'ret': False, 'msg': "Key %s not found" % key}
|
|
|
|
|
|
|
|
memory_uri = data[key]["@odata.id"]
|
|
|
|
|
|
|
|
# Get a list of all DIMMs and build respective URIs
|
|
|
|
response = self.get_request(self.root_uri + memory_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for dimm in data[u'Members']:
|
|
|
|
memory_list.append(dimm[u'@odata.id'])
|
|
|
|
|
|
|
|
for m in memory_list:
|
|
|
|
dimm = {}
|
|
|
|
uri = self.root_uri + m
|
|
|
|
response = self.get_request(uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if "Status" in data:
|
|
|
|
if "State" in data["Status"]:
|
|
|
|
if data["Status"]["State"] == "Absent":
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
continue
|
|
|
|
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
dimm[property] = data[property]
|
|
|
|
|
|
|
|
memory_results.append(dimm)
|
|
|
|
result["entries"] = memory_results
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_memory_inventory(self):
|
|
|
|
return self.aggregate_systems(self.get_memory_inventory)
|
|
|
|
|
2021-11-16 19:46:28 +01:00
|
|
|
def get_nic(self, resource_uri):
|
|
|
|
result = {}
|
|
|
|
properties = ['Name', 'Id', 'Description', 'FQDN', 'IPv4Addresses', 'IPv6Addresses',
|
|
|
|
'NameServers', 'MACAddress', 'PermanentMACAddress',
|
2023-09-28 21:01:05 +02:00
|
|
|
'SpeedMbps', 'MTUSize', 'AutoNeg', 'Status', 'LinkStatus']
|
2021-11-16 19:46:28 +01:00
|
|
|
response = self.get_request(self.root_uri + resource_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
nic = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
nic[property] = data[property]
|
|
|
|
result['entries'] = nic
|
2022-08-12 11:07:30 +02:00
|
|
|
return result
|
2021-11-16 19:46:28 +01:00
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def get_nic_inventory(self, resource_uri):
|
|
|
|
result = {}
|
|
|
|
nic_list = []
|
|
|
|
nic_results = []
|
|
|
|
key = "EthernetInterfaces"
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + resource_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if key not in data:
|
|
|
|
return {'ret': False, 'msg': "Key %s not found" % key}
|
|
|
|
|
|
|
|
ethernetinterfaces_uri = data[key]["@odata.id"]
|
|
|
|
|
|
|
|
# Get a list of all network controllers and build respective URIs
|
|
|
|
response = self.get_request(self.root_uri + ethernetinterfaces_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for nic in data[u'Members']:
|
|
|
|
nic_list.append(nic[u'@odata.id'])
|
|
|
|
|
|
|
|
for n in nic_list:
|
2021-11-16 19:46:28 +01:00
|
|
|
nic = self.get_nic(n)
|
|
|
|
if nic['ret']:
|
|
|
|
nic_results.append(nic['entries'])
|
2020-03-09 10:11:07 +01:00
|
|
|
result["entries"] = nic_results
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_nic_inventory(self, resource_type):
|
|
|
|
ret = True
|
|
|
|
entries = []
|
|
|
|
|
|
|
|
# Given resource_type, use the proper URI
|
|
|
|
if resource_type == 'Systems':
|
|
|
|
resource_uris = self.systems_uris
|
|
|
|
elif resource_type == 'Manager':
|
|
|
|
resource_uris = self.manager_uris
|
|
|
|
|
|
|
|
for resource_uri in resource_uris:
|
|
|
|
inventory = self.get_nic_inventory(resource_uri)
|
|
|
|
ret = inventory.pop('ret') and ret
|
|
|
|
if 'entries' in inventory:
|
|
|
|
entries.append(({'resource_uri': resource_uri},
|
|
|
|
inventory['entries']))
|
|
|
|
return dict(ret=ret, entries=entries)
|
|
|
|
|
|
|
|
def get_virtualmedia(self, resource_uri):
|
|
|
|
result = {}
|
|
|
|
virtualmedia_list = []
|
|
|
|
virtualmedia_results = []
|
|
|
|
key = "VirtualMedia"
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['Description', 'ConnectedVia', 'Id', 'MediaTypes',
|
|
|
|
'Image', 'ImageName', 'Name', 'WriteProtected',
|
|
|
|
'TransferMethod', 'TransferProtocolType']
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + resource_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if key not in data:
|
|
|
|
return {'ret': False, 'msg': "Key %s not found" % key}
|
|
|
|
|
|
|
|
virtualmedia_uri = data[key]["@odata.id"]
|
|
|
|
|
|
|
|
# Get a list of all virtual media and build respective URIs
|
|
|
|
response = self.get_request(self.root_uri + virtualmedia_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for virtualmedia in data[u'Members']:
|
|
|
|
virtualmedia_list.append(virtualmedia[u'@odata.id'])
|
|
|
|
|
|
|
|
for n in virtualmedia_list:
|
|
|
|
virtualmedia = {}
|
|
|
|
uri = self.root_uri + n
|
|
|
|
response = self.get_request(uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
virtualmedia[property] = data[property]
|
|
|
|
|
|
|
|
virtualmedia_results.append(virtualmedia)
|
|
|
|
result["entries"] = virtualmedia_results
|
|
|
|
return result
|
|
|
|
|
2022-08-24 21:36:35 +02:00
|
|
|
def get_multi_virtualmedia(self, resource_type='Manager'):
|
2020-03-09 10:11:07 +01:00
|
|
|
ret = True
|
|
|
|
entries = []
|
|
|
|
|
2022-08-24 21:36:35 +02:00
|
|
|
# Given resource_type, use the proper URI
|
|
|
|
if resource_type == 'Systems':
|
|
|
|
resource_uris = self.systems_uris
|
|
|
|
elif resource_type == 'Manager':
|
|
|
|
resource_uris = self.manager_uris
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
for resource_uri in resource_uris:
|
|
|
|
virtualmedia = self.get_virtualmedia(resource_uri)
|
|
|
|
ret = virtualmedia.pop('ret') and ret
|
|
|
|
if 'entries' in virtualmedia:
|
|
|
|
entries.append(({'resource_uri': resource_uri},
|
|
|
|
virtualmedia['entries']))
|
|
|
|
return dict(ret=ret, entries=entries)
|
|
|
|
|
2020-06-11 20:24:41 +02:00
|
|
|
@staticmethod
|
|
|
|
def _find_empty_virt_media_slot(resources, media_types,
|
2022-08-24 21:36:35 +02:00
|
|
|
media_match_strict=True, vendor=''):
|
2020-06-11 20:24:41 +02:00
|
|
|
for uri, data in resources.items():
|
|
|
|
# check MediaTypes
|
|
|
|
if 'MediaTypes' in data and media_types:
|
|
|
|
if not set(media_types).intersection(set(data['MediaTypes'])):
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
if media_match_strict:
|
|
|
|
continue
|
2022-08-24 21:36:35 +02:00
|
|
|
# Base on current Lenovo server capability, filter out slot RDOC1/2 and Remote1/2/3/4 which are not supported to Insert/Eject.
|
|
|
|
if vendor == 'Lenovo' and ('RDOC' in uri or 'Remote' in uri):
|
|
|
|
continue
|
|
|
|
# if ejected, 'Inserted' should be False and 'ImageName' cleared
|
|
|
|
if (not data.get('Inserted', False) and
|
|
|
|
not data.get('ImageName')):
|
2020-06-11 20:24:41 +02:00
|
|
|
return uri, data
|
|
|
|
return None, None
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _virt_media_image_inserted(resources, image_url):
|
|
|
|
for uri, data in resources.items():
|
|
|
|
if data.get('Image'):
|
|
|
|
if urlparse(image_url) == urlparse(data.get('Image')):
|
|
|
|
if data.get('Inserted', False) and data.get('ImageName'):
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _find_virt_media_to_eject(resources, image_url):
|
|
|
|
matched_uri, matched_data = None, None
|
|
|
|
for uri, data in resources.items():
|
|
|
|
if data.get('Image'):
|
|
|
|
if urlparse(image_url) == urlparse(data.get('Image')):
|
|
|
|
matched_uri, matched_data = uri, data
|
|
|
|
if data.get('Inserted', True) and data.get('ImageName', 'x'):
|
|
|
|
return uri, data, True
|
|
|
|
return matched_uri, matched_data, False
|
|
|
|
|
|
|
|
def _read_virt_media_resources(self, uri_list):
|
|
|
|
resources = {}
|
|
|
|
headers = {}
|
|
|
|
for uri in uri_list:
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
continue
|
|
|
|
resources[uri] = response['data']
|
|
|
|
headers[uri] = response['headers']
|
|
|
|
return resources, headers
|
|
|
|
|
|
|
|
@staticmethod
|
2022-11-02 07:40:21 +01:00
|
|
|
def _insert_virt_media_payload(options, param_map, data, ai):
|
2020-06-11 20:24:41 +02:00
|
|
|
payload = {
|
|
|
|
'Image': options.get('image_url')
|
|
|
|
}
|
|
|
|
for param, option in param_map.items():
|
|
|
|
if options.get(option) is not None and param in data:
|
|
|
|
allowable = ai.get(param, {}).get('AllowableValues', [])
|
|
|
|
if allowable and options.get(option) not in allowable:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Value '%s' specified for option '%s' not "
|
|
|
|
"in list of AllowableValues %s" % (
|
|
|
|
options.get(option), option,
|
|
|
|
allowable)}
|
|
|
|
payload[param] = options.get(option)
|
|
|
|
return payload
|
|
|
|
|
2022-05-05 07:56:26 +02:00
|
|
|
def virtual_media_insert_via_patch(self, options, param_map, uri, data, image_only=False):
|
2020-06-11 20:24:41 +02:00
|
|
|
# get AllowableValues
|
2024-09-05 21:47:28 +02:00
|
|
|
ai = {
|
|
|
|
k[:-24]: {'AllowableValues': v}
|
|
|
|
for k, v in data.items()
|
|
|
|
if k.endswith('@Redfish.AllowableValues')
|
|
|
|
}
|
2020-06-11 20:24:41 +02:00
|
|
|
# construct payload
|
2022-11-02 07:40:21 +01:00
|
|
|
payload = self._insert_virt_media_payload(options, param_map, data, ai)
|
2022-06-20 19:13:31 +02:00
|
|
|
if 'Inserted' not in payload and not image_only:
|
2022-11-02 07:40:21 +01:00
|
|
|
# Add Inserted to the payload if needed
|
2020-06-11 20:24:41 +02:00
|
|
|
payload['Inserted'] = True
|
2022-05-05 07:56:26 +02:00
|
|
|
|
2020-06-11 20:24:41 +02:00
|
|
|
# PATCH the resource
|
2022-11-02 07:40:21 +01:00
|
|
|
resp = self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] is False:
|
|
|
|
# WORKAROUND
|
|
|
|
# Older HPE systems with iLO 4 and Supermicro do not support
|
|
|
|
# specifying Inserted or WriteProtected
|
|
|
|
vendor = self._get_vendor()['Vendor']
|
|
|
|
if vendor == 'HPE' or vendor == 'Supermicro':
|
|
|
|
payload.pop('Inserted', None)
|
|
|
|
payload.pop('WriteProtected', None)
|
|
|
|
resp = self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'VirtualMedia inserted'
|
|
|
|
return resp
|
2020-06-11 20:24:41 +02:00
|
|
|
|
2022-08-24 21:36:35 +02:00
|
|
|
def virtual_media_insert(self, options, resource_type='Manager'):
|
2020-06-11 20:24:41 +02:00
|
|
|
param_map = {
|
|
|
|
'Inserted': 'inserted',
|
|
|
|
'WriteProtected': 'write_protected',
|
|
|
|
'UserName': 'username',
|
|
|
|
'Password': 'password',
|
|
|
|
'TransferProtocolType': 'transfer_protocol_type',
|
|
|
|
'TransferMethod': 'transfer_method'
|
|
|
|
}
|
|
|
|
image_url = options.get('image_url')
|
|
|
|
if not image_url:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "image_url option required for VirtualMediaInsert"}
|
|
|
|
media_types = options.get('media_types')
|
|
|
|
|
|
|
|
# locate and read the VirtualMedia resources
|
2022-08-24 21:36:35 +02:00
|
|
|
# Given resource_type, use the proper URI
|
|
|
|
if resource_type == 'Systems':
|
|
|
|
resource_uri = self.systems_uri
|
|
|
|
elif resource_type == 'Manager':
|
|
|
|
resource_uri = self.manager_uri
|
|
|
|
response = self.get_request(self.root_uri + resource_uri)
|
2020-06-11 20:24:41 +02:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'VirtualMedia' not in data:
|
|
|
|
return {'ret': False, 'msg': "VirtualMedia resource not found"}
|
2022-05-05 07:56:26 +02:00
|
|
|
|
2020-06-11 20:24:41 +02:00
|
|
|
virt_media_uri = data["VirtualMedia"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + virt_media_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
virt_media_list = []
|
|
|
|
for member in data[u'Members']:
|
|
|
|
virt_media_list.append(member[u'@odata.id'])
|
|
|
|
resources, headers = self._read_virt_media_resources(virt_media_list)
|
|
|
|
|
|
|
|
# see if image already inserted; if so, nothing to do
|
|
|
|
if self._virt_media_image_inserted(resources, image_url):
|
|
|
|
return {'ret': True, 'changed': False,
|
|
|
|
'msg': "VirtualMedia '%s' already inserted" % image_url}
|
|
|
|
|
|
|
|
# find an empty slot to insert the media
|
|
|
|
# try first with strict media_type matching
|
2022-08-24 21:36:35 +02:00
|
|
|
vendor = self._get_vendor()['Vendor']
|
2020-06-11 20:24:41 +02:00
|
|
|
uri, data = self._find_empty_virt_media_slot(
|
2022-08-24 21:36:35 +02:00
|
|
|
resources, media_types, media_match_strict=True, vendor=vendor)
|
2020-06-11 20:24:41 +02:00
|
|
|
if not uri:
|
|
|
|
# if not found, try without strict media_type matching
|
|
|
|
uri, data = self._find_empty_virt_media_slot(
|
2022-08-24 21:36:35 +02:00
|
|
|
resources, media_types, media_match_strict=False, vendor=vendor)
|
2020-06-11 20:24:41 +02:00
|
|
|
if not uri:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "Unable to find an available VirtualMedia resource "
|
|
|
|
"%s" % ('supporting ' + str(media_types)
|
|
|
|
if media_types else '')}
|
|
|
|
|
|
|
|
# confirm InsertMedia action found
|
|
|
|
if ('Actions' not in data or
|
|
|
|
'#VirtualMedia.InsertMedia' not in data['Actions']):
|
|
|
|
# try to insert via PATCH if no InsertMedia action found
|
|
|
|
h = headers[uri]
|
|
|
|
if 'allow' in h:
|
|
|
|
methods = [m.strip() for m in h.get('allow').split(',')]
|
|
|
|
if 'PATCH' not in methods:
|
|
|
|
# if Allow header present and PATCH missing, return error
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "%s action not found and PATCH not allowed"
|
|
|
|
% '#VirtualMedia.InsertMedia'}
|
|
|
|
return self.virtual_media_insert_via_patch(options, param_map,
|
2022-11-02 07:40:21 +01:00
|
|
|
uri, data)
|
2020-06-11 20:24:41 +02:00
|
|
|
|
|
|
|
# get the action property
|
|
|
|
action = data['Actions']['#VirtualMedia.InsertMedia']
|
|
|
|
if 'target' not in action:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "target URI missing from Action "
|
|
|
|
"#VirtualMedia.InsertMedia"}
|
|
|
|
action_uri = action['target']
|
|
|
|
# get ActionInfo or AllowableValues
|
|
|
|
ai = self._get_all_action_info_values(action)
|
|
|
|
# construct payload
|
2022-11-02 07:40:21 +01:00
|
|
|
payload = self._insert_virt_media_payload(options, param_map, data, ai)
|
2020-06-11 20:24:41 +02:00
|
|
|
# POST to action
|
|
|
|
response = self.post_request(self.root_uri + action_uri, payload)
|
2022-11-02 07:40:21 +01:00
|
|
|
if response['ret'] is False and ('Inserted' in payload or 'WriteProtected' in payload):
|
|
|
|
# WORKAROUND
|
|
|
|
# Older HPE systems with iLO 4 and Supermicro do not support
|
|
|
|
# specifying Inserted or WriteProtected
|
|
|
|
vendor = self._get_vendor()['Vendor']
|
|
|
|
if vendor == 'HPE' or vendor == 'Supermicro':
|
|
|
|
payload.pop('Inserted', None)
|
|
|
|
payload.pop('WriteProtected', None)
|
|
|
|
response = self.post_request(self.root_uri + action_uri, payload)
|
2020-06-11 20:24:41 +02:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
return {'ret': True, 'changed': True, 'msg': "VirtualMedia inserted"}
|
|
|
|
|
2022-05-05 07:56:26 +02:00
|
|
|
def virtual_media_eject_via_patch(self, uri, image_only=False):
|
2020-06-11 20:24:41 +02:00
|
|
|
# construct payload
|
|
|
|
payload = {
|
|
|
|
'Inserted': False,
|
|
|
|
'Image': None
|
|
|
|
}
|
2022-05-05 07:56:26 +02:00
|
|
|
|
|
|
|
# Inserted is not writable
|
|
|
|
if image_only:
|
|
|
|
del payload['Inserted']
|
|
|
|
|
2020-06-11 20:24:41 +02:00
|
|
|
# PATCH resource
|
2022-11-02 07:40:21 +01:00
|
|
|
resp = self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] is False and 'Inserted' in payload:
|
|
|
|
# WORKAROUND
|
|
|
|
# Older HPE systems with iLO 4 and Supermicro do not support
|
|
|
|
# specifying Inserted
|
|
|
|
vendor = self._get_vendor()['Vendor']
|
|
|
|
if vendor == 'HPE' or vendor == 'Supermicro':
|
|
|
|
payload.pop('Inserted', None)
|
|
|
|
resp = self.patch_request(self.root_uri + uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'VirtualMedia ejected'
|
|
|
|
return resp
|
2020-06-11 20:24:41 +02:00
|
|
|
|
2022-08-24 21:36:35 +02:00
|
|
|
def virtual_media_eject(self, options, resource_type='Manager'):
|
2020-06-11 20:24:41 +02:00
|
|
|
image_url = options.get('image_url')
|
|
|
|
if not image_url:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "image_url option required for VirtualMediaEject"}
|
|
|
|
|
|
|
|
# locate and read the VirtualMedia resources
|
2022-08-24 21:36:35 +02:00
|
|
|
# Given resource_type, use the proper URI
|
|
|
|
if resource_type == 'Systems':
|
|
|
|
resource_uri = self.systems_uri
|
|
|
|
elif resource_type == 'Manager':
|
|
|
|
resource_uri = self.manager_uri
|
|
|
|
response = self.get_request(self.root_uri + resource_uri)
|
2020-06-11 20:24:41 +02:00
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'VirtualMedia' not in data:
|
|
|
|
return {'ret': False, 'msg': "VirtualMedia resource not found"}
|
2022-05-05 07:56:26 +02:00
|
|
|
|
2020-06-11 20:24:41 +02:00
|
|
|
virt_media_uri = data["VirtualMedia"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + virt_media_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
virt_media_list = []
|
|
|
|
for member in data[u'Members']:
|
|
|
|
virt_media_list.append(member[u'@odata.id'])
|
|
|
|
resources, headers = self._read_virt_media_resources(virt_media_list)
|
|
|
|
|
|
|
|
# find the VirtualMedia resource to eject
|
|
|
|
uri, data, eject = self._find_virt_media_to_eject(resources, image_url)
|
|
|
|
if uri and eject:
|
|
|
|
if ('Actions' not in data or
|
|
|
|
'#VirtualMedia.EjectMedia' not in data['Actions']):
|
|
|
|
# try to eject via PATCH if no EjectMedia action found
|
|
|
|
h = headers[uri]
|
|
|
|
if 'allow' in h:
|
|
|
|
methods = [m.strip() for m in h.get('allow').split(',')]
|
|
|
|
if 'PATCH' not in methods:
|
|
|
|
# if Allow header present and PATCH missing, return error
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "%s action not found and PATCH not allowed"
|
|
|
|
% '#VirtualMedia.EjectMedia'}
|
2022-11-02 07:40:21 +01:00
|
|
|
return self.virtual_media_eject_via_patch(uri)
|
2020-06-11 20:24:41 +02:00
|
|
|
else:
|
|
|
|
# POST to the EjectMedia Action
|
|
|
|
action = data['Actions']['#VirtualMedia.EjectMedia']
|
|
|
|
if 'target' not in action:
|
|
|
|
return {'ret': False,
|
|
|
|
'msg': "target URI property missing from Action "
|
|
|
|
"#VirtualMedia.EjectMedia"}
|
|
|
|
action_uri = action['target']
|
|
|
|
# empty payload for Eject action
|
|
|
|
payload = {}
|
|
|
|
# POST to action
|
|
|
|
response = self.post_request(self.root_uri + action_uri,
|
|
|
|
payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
return {'ret': True, 'changed': True,
|
|
|
|
'msg': "VirtualMedia ejected"}
|
|
|
|
elif uri and not eject:
|
|
|
|
# already ejected: return success but changed=False
|
|
|
|
return {'ret': True, 'changed': False,
|
|
|
|
'msg': "VirtualMedia image '%s' already ejected" %
|
|
|
|
image_url}
|
|
|
|
else:
|
|
|
|
# return failure (no resources matching image_url found)
|
|
|
|
return {'ret': False, 'changed': False,
|
|
|
|
'msg': "No VirtualMedia resource found with image '%s' "
|
|
|
|
"inserted" % image_url}
|
|
|
|
|
2020-03-09 10:11:07 +01:00
|
|
|
def get_psu_inventory(self):
|
|
|
|
result = {}
|
|
|
|
psu_list = []
|
|
|
|
psu_results = []
|
|
|
|
key = "PowerSupplies"
|
|
|
|
# Get these entries, but does not fail if not found
|
|
|
|
properties = ['Name', 'Model', 'SerialNumber', 'PartNumber', 'Manufacturer',
|
|
|
|
'FirmwareVersion', 'PowerCapacityWatts', 'PowerSupplyType',
|
|
|
|
'Status']
|
|
|
|
|
|
|
|
# Get a list of all Chassis and build URIs, then get all PowerSupplies
|
|
|
|
# from each Power entry in the Chassis
|
2024-02-24 20:32:20 +01:00
|
|
|
for chassis_uri in self.chassis_uris:
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + chassis_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if 'Power' in data:
|
|
|
|
power_uri = data[u'Power'][u'@odata.id']
|
|
|
|
else:
|
|
|
|
continue
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + power_uri)
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if key not in data:
|
|
|
|
return {'ret': False, 'msg': "Key %s not found" % key}
|
|
|
|
|
|
|
|
psu_list = data[key]
|
|
|
|
for psu in psu_list:
|
|
|
|
psu_not_present = False
|
|
|
|
psu_data = {}
|
|
|
|
for property in properties:
|
|
|
|
if property in psu:
|
|
|
|
if psu[property] is not None:
|
|
|
|
if property == 'Status':
|
|
|
|
if 'State' in psu[property]:
|
|
|
|
if psu[property]['State'] == 'Absent':
|
|
|
|
psu_not_present = True
|
|
|
|
psu_data[property] = psu[property]
|
|
|
|
if psu_not_present:
|
|
|
|
continue
|
|
|
|
psu_results.append(psu_data)
|
|
|
|
|
|
|
|
result["entries"] = psu_results
|
|
|
|
if not result["entries"]:
|
|
|
|
return {'ret': False, 'msg': "No PowerSupply objects found"}
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_psu_inventory(self):
|
|
|
|
return self.aggregate_systems(self.get_psu_inventory)
|
|
|
|
|
|
|
|
def get_system_inventory(self, systems_uri):
|
|
|
|
result = {}
|
|
|
|
inventory = {}
|
|
|
|
# Get these entries, but does not fail if not found
|
2023-12-02 20:56:21 +01:00
|
|
|
properties = ['Status', 'HostName', 'PowerState', 'BootProgress', 'Model', 'Manufacturer',
|
2020-03-09 10:11:07 +01:00
|
|
|
'PartNumber', 'SystemType', 'AssetTag', 'ServiceTag',
|
|
|
|
'SerialNumber', 'SKU', 'BiosVersion', 'MemorySummary',
|
2021-01-29 07:17:57 +01:00
|
|
|
'ProcessorSummary', 'TrustedModules', 'Name', 'Id']
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
inventory[property] = data[property]
|
|
|
|
|
|
|
|
result["entries"] = inventory
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_system_inventory(self):
|
|
|
|
return self.aggregate_systems(self.get_system_inventory)
|
|
|
|
|
|
|
|
def get_network_protocols(self):
|
|
|
|
result = {}
|
|
|
|
service_result = {}
|
|
|
|
# Find NetworkProtocol
|
|
|
|
response = self.get_request(self.root_uri + self.manager_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'NetworkProtocol' not in data:
|
|
|
|
return {'ret': False, 'msg': "NetworkProtocol resource not found"}
|
|
|
|
networkprotocol_uri = data["NetworkProtocol"]["@odata.id"]
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + networkprotocol_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
protocol_services = ['SNMP', 'VirtualMedia', 'Telnet', 'SSDP', 'IPMI', 'SSH',
|
|
|
|
'KVMIP', 'NTP', 'HTTP', 'HTTPS', 'DHCP', 'DHCPv6', 'RDP',
|
|
|
|
'RFB']
|
|
|
|
for protocol_service in protocol_services:
|
|
|
|
if protocol_service in data.keys():
|
|
|
|
service_result[protocol_service] = data[protocol_service]
|
|
|
|
|
|
|
|
result['ret'] = True
|
|
|
|
result["entries"] = service_result
|
|
|
|
return result
|
|
|
|
|
|
|
|
def set_network_protocols(self, manager_services):
|
|
|
|
# Check input data validity
|
|
|
|
protocol_services = ['SNMP', 'VirtualMedia', 'Telnet', 'SSDP', 'IPMI', 'SSH',
|
|
|
|
'KVMIP', 'NTP', 'HTTP', 'HTTPS', 'DHCP', 'DHCPv6', 'RDP',
|
|
|
|
'RFB']
|
|
|
|
protocol_state_onlist = ['true', 'True', True, 'on', 1]
|
|
|
|
protocol_state_offlist = ['false', 'False', False, 'off', 0]
|
|
|
|
payload = {}
|
|
|
|
for service_name in manager_services.keys():
|
|
|
|
if service_name not in protocol_services:
|
|
|
|
return {'ret': False, 'msg': "Service name %s is invalid" % service_name}
|
|
|
|
payload[service_name] = {}
|
|
|
|
for service_property in manager_services[service_name].keys():
|
|
|
|
value = manager_services[service_name][service_property]
|
|
|
|
if service_property in ['ProtocolEnabled', 'protocolenabled']:
|
|
|
|
if value in protocol_state_onlist:
|
|
|
|
payload[service_name]['ProtocolEnabled'] = True
|
|
|
|
elif value in protocol_state_offlist:
|
|
|
|
payload[service_name]['ProtocolEnabled'] = False
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': "Value of property %s is invalid" % service_property}
|
|
|
|
elif service_property in ['port', 'Port']:
|
|
|
|
if isinstance(value, int):
|
|
|
|
payload[service_name]['Port'] = value
|
|
|
|
elif isinstance(value, str) and value.isdigit():
|
|
|
|
payload[service_name]['Port'] = int(value)
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': "Value of property %s is invalid" % service_property}
|
|
|
|
else:
|
|
|
|
payload[service_name][service_property] = value
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Find the ManagerNetworkProtocol resource
|
2020-03-09 10:11:07 +01:00
|
|
|
response = self.get_request(self.root_uri + self.manager_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
networkprotocol_uri = data.get("NetworkProtocol", {}).get("@odata.id")
|
|
|
|
if networkprotocol_uri is None:
|
2020-03-09 10:11:07 +01:00
|
|
|
return {'ret': False, 'msg': "NetworkProtocol resource not found"}
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Modify the ManagerNetworkProtocol resource
|
|
|
|
resp = self.patch_request(self.root_uri + networkprotocol_uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Modified manager network protocol settings'
|
|
|
|
return resp
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def to_singular(resource_name):
|
|
|
|
if resource_name.endswith('ies'):
|
|
|
|
resource_name = resource_name[:-3] + 'y'
|
|
|
|
elif resource_name.endswith('s'):
|
|
|
|
resource_name = resource_name[:-1]
|
|
|
|
return resource_name
|
|
|
|
|
|
|
|
def get_health_resource(self, subsystem, uri, health, expanded):
|
|
|
|
status = 'Status'
|
|
|
|
|
|
|
|
if expanded:
|
|
|
|
d = expanded
|
|
|
|
else:
|
|
|
|
r = self.get_request(self.root_uri + uri)
|
|
|
|
if r.get('ret'):
|
|
|
|
d = r.get('data')
|
|
|
|
else:
|
|
|
|
return
|
|
|
|
|
|
|
|
if 'Members' in d: # collections case
|
|
|
|
for m in d.get('Members'):
|
|
|
|
u = m.get('@odata.id')
|
|
|
|
r = self.get_request(self.root_uri + u)
|
|
|
|
if r.get('ret'):
|
|
|
|
p = r.get('data')
|
|
|
|
if p:
|
|
|
|
e = {self.to_singular(subsystem.lower()) + '_uri': u,
|
|
|
|
status: p.get(status,
|
|
|
|
"Status not available")}
|
|
|
|
health[subsystem].append(e)
|
|
|
|
else: # non-collections case
|
|
|
|
e = {self.to_singular(subsystem.lower()) + '_uri': uri,
|
|
|
|
status: d.get(status,
|
|
|
|
"Status not available")}
|
|
|
|
health[subsystem].append(e)
|
|
|
|
|
|
|
|
def get_health_subsystem(self, subsystem, data, health):
|
|
|
|
if subsystem in data:
|
|
|
|
sub = data.get(subsystem)
|
|
|
|
if isinstance(sub, list):
|
|
|
|
for r in sub:
|
|
|
|
if '@odata.id' in r:
|
|
|
|
uri = r.get('@odata.id')
|
|
|
|
expanded = None
|
|
|
|
if '#' in uri and len(r) > 1:
|
|
|
|
expanded = r
|
|
|
|
self.get_health_resource(subsystem, uri, health, expanded)
|
|
|
|
elif isinstance(sub, dict):
|
|
|
|
if '@odata.id' in sub:
|
|
|
|
uri = sub.get('@odata.id')
|
|
|
|
self.get_health_resource(subsystem, uri, health, None)
|
|
|
|
elif 'Members' in data:
|
|
|
|
for m in data.get('Members'):
|
|
|
|
u = m.get('@odata.id')
|
|
|
|
r = self.get_request(self.root_uri + u)
|
|
|
|
if r.get('ret'):
|
|
|
|
d = r.get('data')
|
|
|
|
self.get_health_subsystem(subsystem, d, health)
|
|
|
|
|
|
|
|
def get_health_report(self, category, uri, subsystems):
|
|
|
|
result = {}
|
|
|
|
health = {}
|
|
|
|
status = 'Status'
|
|
|
|
|
|
|
|
# Get health status of top level resource
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
health[category] = {status: data.get(status, "Status not available")}
|
|
|
|
|
|
|
|
# Get health status of subsystems
|
|
|
|
for sub in subsystems:
|
|
|
|
d = None
|
|
|
|
if sub.startswith('Links.'): # ex: Links.PCIeDevices
|
|
|
|
sub = sub[len('Links.'):]
|
|
|
|
d = data.get('Links', {})
|
|
|
|
elif '.' in sub: # ex: Thermal.Fans
|
|
|
|
p, sub = sub.split('.')
|
|
|
|
u = data.get(p, {}).get('@odata.id')
|
|
|
|
if u:
|
|
|
|
r = self.get_request(self.root_uri + u)
|
|
|
|
if r['ret']:
|
|
|
|
d = r['data']
|
|
|
|
if not d:
|
|
|
|
continue
|
|
|
|
else: # ex: Memory
|
|
|
|
d = data
|
|
|
|
health[sub] = []
|
|
|
|
self.get_health_subsystem(sub, d, health)
|
|
|
|
if not health[sub]:
|
|
|
|
del health[sub]
|
|
|
|
|
|
|
|
result["entries"] = health
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_system_health_report(self, systems_uri):
|
|
|
|
subsystems = ['Processors', 'Memory', 'SimpleStorage', 'Storage',
|
|
|
|
'EthernetInterfaces', 'NetworkInterfaces.NetworkPorts',
|
|
|
|
'NetworkInterfaces.NetworkDeviceFunctions']
|
|
|
|
return self.get_health_report('System', systems_uri, subsystems)
|
|
|
|
|
|
|
|
def get_multi_system_health_report(self):
|
|
|
|
return self.aggregate_systems(self.get_system_health_report)
|
|
|
|
|
|
|
|
def get_chassis_health_report(self, chassis_uri):
|
|
|
|
subsystems = ['Power.PowerSupplies', 'Thermal.Fans',
|
|
|
|
'Links.PCIeDevices']
|
|
|
|
return self.get_health_report('Chassis', chassis_uri, subsystems)
|
|
|
|
|
|
|
|
def get_multi_chassis_health_report(self):
|
|
|
|
return self.aggregate_chassis(self.get_chassis_health_report)
|
|
|
|
|
|
|
|
def get_manager_health_report(self, manager_uri):
|
|
|
|
subsystems = []
|
|
|
|
return self.get_health_report('Manager', manager_uri, subsystems)
|
|
|
|
|
|
|
|
def get_multi_manager_health_report(self):
|
|
|
|
return self.aggregate_managers(self.get_manager_health_report)
|
|
|
|
|
|
|
|
def set_manager_nic(self, nic_addr, nic_config):
|
2021-12-11 21:27:09 +01:00
|
|
|
# Get the manager ethernet interface uri
|
|
|
|
nic_info = self.get_manager_ethernet_uri(nic_addr)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
2021-12-11 21:27:09 +01:00
|
|
|
if nic_info.get('nic_addr') is None:
|
|
|
|
return nic_info
|
|
|
|
else:
|
|
|
|
target_ethernet_uri = nic_info['nic_addr']
|
|
|
|
target_ethernet_current_setting = nic_info['ethernet_setting']
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Convert input to payload and check validity
|
2022-11-02 07:40:21 +01:00
|
|
|
# Note: Some properties in the EthernetInterface resource are arrays of
|
|
|
|
# objects. The call into this module expects a flattened view, meaning
|
|
|
|
# the user specifies exactly one object for an array property. For
|
|
|
|
# example, if a user provides IPv4StaticAddresses in the request to this
|
|
|
|
# module, it will turn that into an array of one member. This pattern
|
|
|
|
# should be avoided for future commands in this module, but needs to be
|
|
|
|
# preserved here for backwards compatibility.
|
2020-03-09 10:11:07 +01:00
|
|
|
payload = {}
|
|
|
|
for property in nic_config.keys():
|
|
|
|
value = nic_config[property]
|
2022-11-02 07:40:21 +01:00
|
|
|
if property in target_ethernet_current_setting and isinstance(value, dict) and isinstance(target_ethernet_current_setting[property], list):
|
|
|
|
payload[property] = list()
|
|
|
|
payload[property].append(value)
|
2020-03-09 10:11:07 +01:00
|
|
|
else:
|
|
|
|
payload[property] = value
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Modify the EthernetInterface resource
|
|
|
|
resp = self.patch_request(self.root_uri + target_ethernet_uri, payload, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Modified manager NIC'
|
|
|
|
return resp
|
2021-11-13 13:59:29 +01:00
|
|
|
|
2021-12-11 21:27:09 +01:00
|
|
|
# A helper function to get the EthernetInterface URI
|
|
|
|
def get_manager_ethernet_uri(self, nic_addr='null'):
|
|
|
|
# Get EthernetInterface collection
|
|
|
|
response = self.get_request(self.root_uri + self.manager_uri)
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
if 'EthernetInterfaces' not in data:
|
|
|
|
return {'ret': False, 'msg': "EthernetInterfaces resource not found"}
|
|
|
|
ethernetinterfaces_uri = data["EthernetInterfaces"]["@odata.id"]
|
|
|
|
response = self.get_request(self.root_uri + ethernetinterfaces_uri)
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
uris = [a.get('@odata.id') for a in data.get('Members', []) if
|
|
|
|
a.get('@odata.id')]
|
|
|
|
|
|
|
|
# Find target EthernetInterface
|
|
|
|
target_ethernet_uri = None
|
|
|
|
target_ethernet_current_setting = None
|
|
|
|
if nic_addr == 'null':
|
|
|
|
# Find root_uri matched EthernetInterface when nic_addr is not specified
|
|
|
|
nic_addr = (self.root_uri).split('/')[-1]
|
|
|
|
nic_addr = nic_addr.split(':')[0] # split port if existing
|
|
|
|
for uri in uris:
|
|
|
|
response = self.get_request(self.root_uri + uri)
|
|
|
|
if not response['ret']:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
data_string = json.dumps(data)
|
|
|
|
if nic_addr.lower() in data_string.lower():
|
|
|
|
target_ethernet_uri = uri
|
|
|
|
target_ethernet_current_setting = data
|
|
|
|
break
|
|
|
|
|
|
|
|
nic_info = {}
|
|
|
|
nic_info['nic_addr'] = target_ethernet_uri
|
|
|
|
nic_info['ethernet_setting'] = target_ethernet_current_setting
|
|
|
|
|
|
|
|
if target_ethernet_uri is None:
|
|
|
|
return {}
|
|
|
|
else:
|
|
|
|
return nic_info
|
|
|
|
|
2021-11-13 13:59:29 +01:00
|
|
|
def set_hostinterface_attributes(self, hostinterface_config, hostinterface_id=None):
|
2022-11-02 07:40:21 +01:00
|
|
|
if hostinterface_config is None:
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide hostinterface_config for SetHostInterface command'}
|
|
|
|
|
|
|
|
# Find the HostInterfaceCollection resource
|
2021-11-13 13:59:29 +01:00
|
|
|
response = self.get_request(self.root_uri + self.manager_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
hostinterfaces_uri = data.get("HostInterfaces", {}).get("@odata.id")
|
|
|
|
if hostinterfaces_uri is None:
|
|
|
|
return {'ret': False, 'msg': "HostInterface resource not found"}
|
2021-11-13 13:59:29 +01:00
|
|
|
response = self.get_request(self.root_uri + hostinterfaces_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
uris = [a.get('@odata.id') for a in data.get('Members', []) if a.get('@odata.id')]
|
2022-11-02 07:40:21 +01:00
|
|
|
|
|
|
|
# Capture list of URIs that match a specified HostInterface resource Id
|
2021-11-13 13:59:29 +01:00
|
|
|
if hostinterface_id:
|
|
|
|
matching_hostinterface_uris = [uri for uri in uris if hostinterface_id in uri.split('/')[-1]]
|
|
|
|
if hostinterface_id and matching_hostinterface_uris:
|
|
|
|
hostinterface_uri = list.pop(matching_hostinterface_uris)
|
|
|
|
elif hostinterface_id and not matching_hostinterface_uris:
|
|
|
|
return {'ret': False, 'msg': "HostInterface ID %s not present." % hostinterface_id}
|
|
|
|
elif len(uris) == 1:
|
|
|
|
hostinterface_uri = list.pop(uris)
|
|
|
|
else:
|
|
|
|
return {'ret': False, 'msg': "HostInterface ID not defined and multiple interfaces detected."}
|
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
# Modify the HostInterface resource
|
|
|
|
resp = self.patch_request(self.root_uri + hostinterface_uri, hostinterface_config, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Modified host interface'
|
|
|
|
return resp
|
2021-11-16 19:46:28 +01:00
|
|
|
|
|
|
|
def get_hostinterfaces(self):
|
|
|
|
result = {}
|
|
|
|
hostinterface_results = []
|
|
|
|
properties = ['Id', 'Name', 'Description', 'HostInterfaceType', 'Status',
|
|
|
|
'InterfaceEnabled', 'ExternallyAccessible', 'AuthenticationModes',
|
|
|
|
'AuthNoneRoleId', 'CredentialBootstrapping']
|
|
|
|
manager_uri_list = self.manager_uris
|
|
|
|
for manager_uri in manager_uri_list:
|
|
|
|
response = self.get_request(self.root_uri + manager_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
2022-11-02 07:40:21 +01:00
|
|
|
hostinterfaces_uri = data.get("HostInterfaces", {}).get("@odata.id")
|
|
|
|
if hostinterfaces_uri is None:
|
2021-11-16 19:46:28 +01:00
|
|
|
continue
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + hostinterfaces_uri)
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
if 'Members' in data:
|
|
|
|
for hostinterface in data['Members']:
|
|
|
|
hostinterface_uri = hostinterface['@odata.id']
|
|
|
|
hostinterface_response = self.get_request(self.root_uri + hostinterface_uri)
|
|
|
|
# dictionary for capturing individual HostInterface properties
|
|
|
|
hostinterface_data_temp = {}
|
|
|
|
if hostinterface_response['ret'] is False:
|
|
|
|
return hostinterface_response
|
|
|
|
hostinterface_data = hostinterface_response['data']
|
|
|
|
for property in properties:
|
|
|
|
if property in hostinterface_data:
|
|
|
|
if hostinterface_data[property] is not None:
|
|
|
|
hostinterface_data_temp[property] = hostinterface_data[property]
|
|
|
|
# Check for the presence of a ManagerEthernetInterface
|
|
|
|
# object, a link to a _single_ EthernetInterface that the
|
|
|
|
# BMC uses to communicate with the host.
|
|
|
|
if 'ManagerEthernetInterface' in hostinterface_data:
|
|
|
|
interface_uri = hostinterface_data['ManagerEthernetInterface']['@odata.id']
|
|
|
|
interface_response = self.get_nic(interface_uri)
|
|
|
|
if interface_response['ret'] is False:
|
|
|
|
return interface_response
|
|
|
|
hostinterface_data_temp['ManagerEthernetInterface'] = interface_response['entries']
|
|
|
|
|
|
|
|
# Check for the presence of a HostEthernetInterfaces
|
|
|
|
# object, a link to a _collection_ of EthernetInterfaces
|
|
|
|
# that the host uses to communicate with the BMC.
|
|
|
|
if 'HostEthernetInterfaces' in hostinterface_data:
|
|
|
|
interfaces_uri = hostinterface_data['HostEthernetInterfaces']['@odata.id']
|
|
|
|
interfaces_response = self.get_request(self.root_uri + interfaces_uri)
|
|
|
|
if interfaces_response['ret'] is False:
|
|
|
|
return interfaces_response
|
|
|
|
interfaces_data = interfaces_response['data']
|
|
|
|
if 'Members' in interfaces_data:
|
|
|
|
for interface in interfaces_data['Members']:
|
|
|
|
interface_uri = interface['@odata.id']
|
|
|
|
interface_response = self.get_nic(interface_uri)
|
|
|
|
if interface_response['ret'] is False:
|
|
|
|
return interface_response
|
|
|
|
# Check if this is the first
|
|
|
|
# HostEthernetInterfaces item and create empty
|
|
|
|
# list if so.
|
|
|
|
if 'HostEthernetInterfaces' not in hostinterface_data_temp:
|
|
|
|
hostinterface_data_temp['HostEthernetInterfaces'] = []
|
|
|
|
|
|
|
|
hostinterface_data_temp['HostEthernetInterfaces'].append(interface_response['entries'])
|
|
|
|
|
|
|
|
hostinterface_results.append(hostinterface_data_temp)
|
|
|
|
else:
|
|
|
|
continue
|
|
|
|
result["entries"] = hostinterface_results
|
|
|
|
if not result["entries"]:
|
|
|
|
return {'ret': False, 'msg': "No HostInterface objects found"}
|
|
|
|
return result
|
2022-06-30 20:35:35 +02:00
|
|
|
|
|
|
|
def get_manager_inventory(self, manager_uri):
|
|
|
|
result = {}
|
|
|
|
inventory = {}
|
|
|
|
# Get these entries, but does not fail if not found
|
2023-08-25 20:50:28 +02:00
|
|
|
properties = ['Id', 'FirmwareVersion', 'ManagerType', 'Manufacturer', 'Model',
|
2024-01-23 07:28:54 +01:00
|
|
|
'PartNumber', 'PowerState', 'SerialNumber', 'ServiceIdentification',
|
|
|
|
'Status', 'UUID']
|
2022-06-30 20:35:35 +02:00
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + manager_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
for property in properties:
|
|
|
|
if property in data:
|
|
|
|
inventory[property] = data[property]
|
|
|
|
|
|
|
|
result["entries"] = inventory
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_multi_manager_inventory(self):
|
|
|
|
return self.aggregate_managers(self.get_manager_inventory)
|
2022-09-25 17:07:45 +02:00
|
|
|
|
2024-01-23 07:28:54 +01:00
|
|
|
def get_service_identification(self, manager):
|
|
|
|
result = {}
|
|
|
|
if manager is None:
|
|
|
|
if len(self.manager_uris) == 1:
|
|
|
|
manager = self.manager_uris[0].split('/')[-1]
|
|
|
|
elif len(self.manager_uris) > 1:
|
|
|
|
entries = self.get_multi_manager_inventory()['entries']
|
|
|
|
managers = [m[0]['manager_uri'] for m in entries if m[1].get('ServiceIdentification')]
|
|
|
|
if len(managers) == 1:
|
|
|
|
manager = managers[0].split('/')[-1]
|
|
|
|
else:
|
|
|
|
self.module.fail_json(msg=[
|
|
|
|
"Multiple managers with ServiceIdentification were found: %s" % str(managers),
|
|
|
|
"Please specify by using the 'manager' parameter in your playbook"])
|
|
|
|
elif len(self.manager_uris) == 0:
|
|
|
|
self.module.fail_json(msg="No manager identities were found")
|
|
|
|
response = self.get_request(self.root_uri + '/redfish/v1/Managers/' + manager, override_headers=None)
|
|
|
|
try:
|
|
|
|
result['service_identification'] = response['data']['ServiceIdentification']
|
|
|
|
except Exception as e:
|
|
|
|
self.module.fail_json(msg="Service ID not found for manager %s" % manager)
|
|
|
|
result['ret'] = True
|
|
|
|
return result
|
|
|
|
|
2024-02-11 13:16:07 +01:00
|
|
|
def set_service_identification(self, service_id):
|
|
|
|
data = {"ServiceIdentification": service_id}
|
|
|
|
resp = self.patch_request(self.root_uri + '/redfish/v1/Managers/' + self.resource_id, data, check_pyld=True)
|
|
|
|
return resp
|
|
|
|
|
2022-09-25 17:07:45 +02:00
|
|
|
def set_session_service(self, sessions_config):
|
2022-11-02 07:40:21 +01:00
|
|
|
if sessions_config is None:
|
|
|
|
return {'ret': False, 'msg':
|
|
|
|
'Must provide sessions_config for SetSessionService command'}
|
2022-09-25 17:07:45 +02:00
|
|
|
|
2022-11-02 07:40:21 +01:00
|
|
|
resp = self.patch_request(self.root_uri + self.session_service_uri, sessions_config, check_pyld=True)
|
|
|
|
if resp['ret'] and resp['changed']:
|
|
|
|
resp['msg'] = 'Modified session service'
|
|
|
|
return resp
|
2023-02-17 07:49:54 +01:00
|
|
|
|
|
|
|
def verify_bios_attributes(self, bios_attributes):
|
|
|
|
# This method verifies BIOS attributes against the provided input
|
|
|
|
server_bios = self.get_multi_bios_attributes()
|
|
|
|
if server_bios["ret"] is False:
|
|
|
|
return server_bios
|
|
|
|
|
|
|
|
bios_dict = {}
|
|
|
|
wrong_param = {}
|
|
|
|
|
|
|
|
# Verify bios_attributes with BIOS settings available in the server
|
|
|
|
for key, value in bios_attributes.items():
|
|
|
|
if key in server_bios["entries"][0][1]:
|
|
|
|
if server_bios["entries"][0][1][key] != value:
|
|
|
|
bios_dict.update({key: value})
|
|
|
|
else:
|
|
|
|
wrong_param.update({key: value})
|
|
|
|
|
|
|
|
if wrong_param:
|
|
|
|
return {
|
|
|
|
"ret": False,
|
|
|
|
"msg": "Wrong parameters are provided: %s" % wrong_param
|
|
|
|
}
|
|
|
|
|
|
|
|
if bios_dict:
|
|
|
|
return {
|
|
|
|
"ret": False,
|
|
|
|
"msg": "BIOS parameters are not matching: %s" % bios_dict
|
|
|
|
}
|
|
|
|
|
|
|
|
return {
|
|
|
|
"ret": True,
|
|
|
|
"changed": False,
|
|
|
|
"msg": "BIOS verification completed"
|
|
|
|
}
|
2023-02-17 13:24:35 +01:00
|
|
|
|
|
|
|
def enable_secure_boot(self):
|
|
|
|
# This function enable Secure Boot on an OOB controller
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if response["ret"] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
server_details = response["data"]
|
|
|
|
secure_boot_url = server_details["SecureBoot"]["@odata.id"]
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + secure_boot_url)
|
|
|
|
if response["ret"] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
body = {}
|
|
|
|
body["SecureBootEnable"] = True
|
|
|
|
|
|
|
|
return self.patch_request(self.root_uri + secure_boot_url, body, check_pyld=True)
|
2023-03-25 23:18:21 +01:00
|
|
|
|
2023-09-20 22:02:42 +02:00
|
|
|
def set_secure_boot(self, secure_boot_enable):
|
|
|
|
# This function enable Secure Boot on an OOB controller
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if response["ret"] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
server_details = response["data"]
|
|
|
|
secure_boot_url = server_details["SecureBoot"]["@odata.id"]
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + secure_boot_url)
|
|
|
|
if response["ret"] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
body = {}
|
|
|
|
body["SecureBootEnable"] = secure_boot_enable
|
|
|
|
|
|
|
|
return self.patch_request(self.root_uri + secure_boot_url, body, check_pyld=True)
|
|
|
|
|
2023-03-25 23:18:21 +01:00
|
|
|
def get_hpe_thermal_config(self):
|
|
|
|
result = {}
|
|
|
|
key = "Thermal"
|
|
|
|
# Go through list
|
2024-02-24 20:32:20 +01:00
|
|
|
for chassis_uri in self.chassis_uris:
|
2023-03-25 23:18:21 +01:00
|
|
|
response = self.get_request(self.root_uri + chassis_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
result['ret'] = True
|
|
|
|
data = response['data']
|
2024-02-24 20:32:20 +01:00
|
|
|
val = data.get('Oem', {}).get('Hpe', {}).get('ThermalConfiguration')
|
|
|
|
if val is not None:
|
|
|
|
return {"ret": True, "current_thermal_config": val}
|
|
|
|
return {"ret": False}
|
2023-03-25 23:18:21 +01:00
|
|
|
|
|
|
|
def get_hpe_fan_percent_min(self):
|
|
|
|
result = {}
|
|
|
|
key = "Thermal"
|
|
|
|
# Go through list
|
2024-02-24 20:32:20 +01:00
|
|
|
for chassis_uri in self.chassis_uris:
|
2023-03-25 23:18:21 +01:00
|
|
|
response = self.get_request(self.root_uri + chassis_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
2024-02-24 20:32:20 +01:00
|
|
|
val = data.get('Oem', {}).get('Hpe', {}).get('FanPercentMinimum')
|
|
|
|
if val is not None:
|
|
|
|
return {"ret": True, "fan_percent_min": val}
|
|
|
|
return {"ret": False}
|
2023-08-05 20:16:57 +02:00
|
|
|
|
|
|
|
def delete_volumes(self, storage_subsystem_id, volume_ids):
|
|
|
|
# Find the Storage resource from the requested ComputerSystem resource
|
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
storage_uri = data.get('Storage', {}).get('@odata.id')
|
|
|
|
if storage_uri is None:
|
|
|
|
return {'ret': False, 'msg': 'Storage resource not found'}
|
|
|
|
|
|
|
|
# Get Storage Collection
|
|
|
|
response = self.get_request(self.root_uri + storage_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Collect Storage Subsystems
|
|
|
|
self.storage_subsystems_uris = [i['@odata.id'] for i in response['data'].get('Members', [])]
|
|
|
|
if not self.storage_subsystems_uris:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "StorageCollection's Members array is either empty or missing"}
|
|
|
|
|
|
|
|
# Matching Storage Subsystem ID with user input
|
|
|
|
self.storage_subsystem_uri = ""
|
|
|
|
for storage_subsystem_uri in self.storage_subsystems_uris:
|
|
|
|
if storage_subsystem_uri.split("/")[-2] == storage_subsystem_id:
|
|
|
|
self.storage_subsystem_uri = storage_subsystem_uri
|
|
|
|
|
|
|
|
if not self.storage_subsystem_uri:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "Provided Storage Subsystem ID %s does not exist on the server" % storage_subsystem_id}
|
|
|
|
|
|
|
|
# Get Volume Collection
|
|
|
|
response = self.get_request(self.root_uri + self.storage_subsystem_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
response = self.get_request(self.root_uri + data['Volumes']['@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Collect Volumes
|
|
|
|
self.volume_uris = [i['@odata.id'] for i in response['data'].get('Members', [])]
|
|
|
|
if not self.volume_uris:
|
|
|
|
return {
|
|
|
|
'ret': True, 'changed': False,
|
|
|
|
'msg': "VolumeCollection's Members array is either empty or missing"}
|
|
|
|
|
|
|
|
# Delete each volume
|
|
|
|
for volume in self.volume_uris:
|
|
|
|
if volume.split("/")[-1] in volume_ids:
|
|
|
|
response = self.delete_request(self.root_uri + volume)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
return {'ret': True, 'changed': True,
|
|
|
|
'msg': "The following volumes were deleted: %s" % str(volume_ids)}
|
2023-10-06 23:11:38 +02:00
|
|
|
|
|
|
|
def create_volume(self, volume_details, storage_subsystem_id):
|
|
|
|
# Find the Storage resource from the requested ComputerSystem resource
|
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
storage_uri = data.get('Storage', {}).get('@odata.id')
|
|
|
|
if storage_uri is None:
|
|
|
|
return {'ret': False, 'msg': 'Storage resource not found'}
|
|
|
|
|
|
|
|
# Get Storage Collection
|
|
|
|
response = self.get_request(self.root_uri + storage_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Collect Storage Subsystems
|
|
|
|
self.storage_subsystems_uris = [i['@odata.id'] for i in response['data'].get('Members', [])]
|
|
|
|
if not self.storage_subsystems_uris:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "StorageCollection's Members array is either empty or missing"}
|
|
|
|
|
|
|
|
# Matching Storage Subsystem ID with user input
|
|
|
|
self.storage_subsystem_uri = ""
|
|
|
|
for storage_subsystem_uri in self.storage_subsystems_uris:
|
|
|
|
if storage_subsystem_uri.split("/")[-2] == storage_subsystem_id:
|
|
|
|
self.storage_subsystem_uri = storage_subsystem_uri
|
|
|
|
|
|
|
|
if not self.storage_subsystem_uri:
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "Provided Storage Subsystem ID %s does not exist on the server" % storage_subsystem_id}
|
|
|
|
|
|
|
|
# Validate input parameters
|
|
|
|
required_parameters = ['RAIDType', 'Drives', 'CapacityBytes']
|
|
|
|
allowed_parameters = ['DisplayName', 'InitializeMethod', 'MediaSpanCount',
|
|
|
|
'Name', 'ReadCachePolicy', 'StripSizeBytes', 'VolumeUsage', 'WriteCachePolicy']
|
|
|
|
|
|
|
|
for parameter in required_parameters:
|
|
|
|
if not volume_details.get(parameter):
|
|
|
|
return {
|
|
|
|
'ret': False,
|
|
|
|
'msg': "%s are required parameter to create a volume" % str(required_parameters)}
|
|
|
|
|
|
|
|
# Navigate to the volume uri of the correct storage subsystem
|
|
|
|
response = self.get_request(self.root_uri + self.storage_subsystem_uri)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
data = response['data']
|
|
|
|
|
|
|
|
# Deleting any volumes of RAIDType None present on the Storage Subsystem
|
|
|
|
response = self.get_request(self.root_uri + data['Volumes']['@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
volume_data = response['data']
|
|
|
|
|
|
|
|
if "Members" in volume_data:
|
|
|
|
for member in volume_data["Members"]:
|
|
|
|
response = self.get_request(self.root_uri + member['@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
member_data = response['data']
|
|
|
|
|
|
|
|
if member_data["RAIDType"] == "None":
|
|
|
|
response = self.delete_request(self.root_uri + member['@odata.id'])
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
# Construct payload and issue POST command to create volume
|
|
|
|
volume_details["Links"] = {}
|
|
|
|
volume_details["Links"]["Drives"] = []
|
|
|
|
for drive in volume_details["Drives"]:
|
|
|
|
volume_details["Links"]["Drives"].append({"@odata.id": drive})
|
|
|
|
del volume_details["Drives"]
|
|
|
|
payload = volume_details
|
|
|
|
response = self.post_request(self.root_uri + data['Volumes']['@odata.id'], payload)
|
|
|
|
if response['ret'] is False:
|
|
|
|
return response
|
|
|
|
|
|
|
|
return {'ret': True, 'changed': True,
|
|
|
|
'msg': "Volume Created"}
|
2023-10-09 09:14:47 +02:00
|
|
|
|
|
|
|
def get_bios_registries(self):
|
|
|
|
# Get /redfish/v1
|
|
|
|
response = self.get_request(self.root_uri + self.systems_uri)
|
|
|
|
if not response["ret"]:
|
|
|
|
return response
|
|
|
|
|
|
|
|
server_details = response["data"]
|
|
|
|
|
|
|
|
# Get Registries URI
|
|
|
|
if "Bios" not in server_details:
|
|
|
|
msg = "Getting BIOS URI failed, Key 'Bios' not found in /redfish/v1/Systems/1/ response: %s"
|
|
|
|
return {
|
|
|
|
"ret": False,
|
|
|
|
"msg": msg % str(server_details)
|
|
|
|
}
|
|
|
|
|
|
|
|
bios_uri = server_details["Bios"]["@odata.id"]
|
|
|
|
bios_resp = self.get_request(self.root_uri + bios_uri)
|
|
|
|
if not bios_resp["ret"]:
|
|
|
|
return bios_resp
|
|
|
|
|
|
|
|
bios_data = bios_resp["data"]
|
|
|
|
attribute_registry = bios_data["AttributeRegistry"]
|
|
|
|
|
|
|
|
reg_uri = self.root_uri + self.service_root + "Registries/" + attribute_registry
|
|
|
|
reg_resp = self.get_request(reg_uri)
|
|
|
|
if not reg_resp["ret"]:
|
|
|
|
return reg_resp
|
|
|
|
|
|
|
|
reg_data = reg_resp["data"]
|
|
|
|
|
|
|
|
# Get BIOS attribute registry URI
|
|
|
|
lst = []
|
|
|
|
|
|
|
|
# Get the location URI
|
|
|
|
response = self.check_location_uri(reg_data, reg_uri)
|
|
|
|
if not response["ret"]:
|
|
|
|
return response
|
|
|
|
|
|
|
|
rsp_data, rsp_uri = response["rsp_data"], response["rsp_uri"]
|
|
|
|
|
|
|
|
if "RegistryEntries" not in rsp_data:
|
|
|
|
return {
|
|
|
|
"msg": "'RegistryEntries' not present in %s response, %s" % (rsp_uri, str(rsp_data)),
|
|
|
|
"ret": False
|
|
|
|
}
|
|
|
|
|
|
|
|
return {
|
|
|
|
"bios_registry": rsp_data,
|
|
|
|
"bios_registry_uri": rsp_uri,
|
|
|
|
"ret": True
|
|
|
|
}
|
|
|
|
|
|
|
|
def check_location_uri(self, resp_data, resp_uri):
|
|
|
|
# Get the location URI response
|
|
|
|
# return {"msg": self.creds, "ret": False}
|
|
|
|
vendor = self._get_vendor()['Vendor']
|
|
|
|
rsp_uri = ""
|
|
|
|
for loc in resp_data['Location']:
|
2024-07-12 23:08:43 +02:00
|
|
|
if loc['Language'].startswith("en"):
|
2023-10-09 09:14:47 +02:00
|
|
|
rsp_uri = loc['Uri']
|
|
|
|
if vendor == 'HPE':
|
|
|
|
# WORKAROUND
|
2023-10-29 18:04:44 +01:00
|
|
|
# HPE systems with iLO 4 will have BIOS Attribute Registries location URI as a dictionary with key 'extref'
|
2023-10-09 09:14:47 +02:00
|
|
|
# Hence adding condition to fetch the Uri
|
2023-11-08 13:20:19 +01:00
|
|
|
if isinstance(loc['Uri'], dict) and "extref" in loc['Uri'].keys():
|
2023-10-09 09:14:47 +02:00
|
|
|
rsp_uri = loc['Uri']['extref']
|
|
|
|
if not rsp_uri:
|
2023-10-29 18:04:44 +01:00
|
|
|
msg = "Language 'en' not found in BIOS Attribute Registries location, URI: %s, response: %s"
|
2023-10-09 09:14:47 +02:00
|
|
|
return {
|
|
|
|
"ret": False,
|
|
|
|
"msg": msg % (resp_uri, str(resp_data))
|
|
|
|
}
|
|
|
|
|
|
|
|
res = self.get_request(self.root_uri + rsp_uri)
|
|
|
|
if res['ret'] is False:
|
|
|
|
# WORKAROUND
|
|
|
|
# HPE systems with iLO 4 or iLO5 compresses (gzip) for some URIs
|
|
|
|
# Hence adding encoding to the header
|
|
|
|
if vendor == 'HPE':
|
|
|
|
override_headers = {"Accept-Encoding": "gzip"}
|
|
|
|
res = self.get_request(self.root_uri + rsp_uri, override_headers=override_headers)
|
|
|
|
if res['ret']:
|
|
|
|
return {
|
|
|
|
"ret": True,
|
|
|
|
"rsp_data": res["data"],
|
|
|
|
"rsp_uri": rsp_uri
|
|
|
|
}
|
|
|
|
return res
|