mirror of
https://github.com/ansible-collections/community.general.git
synced 2024-09-14 20:13:21 +02:00
Misc. typo fixes (#4940)
Signed-off-by: Abhijeet Kasurde <akasurde@redhat.com>
This commit is contained in:
parent
35ddf31b5f
commit
bf94f08bc4
40 changed files with 55 additions and 55 deletions
2
.github/ISSUE_TEMPLATE/bug_report.yml
vendored
2
.github/ISSUE_TEMPLATE/bug_report.yml
vendored
|
@ -105,7 +105,7 @@ body:
|
||||||
attributes:
|
attributes:
|
||||||
label: Steps to Reproduce
|
label: Steps to Reproduce
|
||||||
description: |
|
description: |
|
||||||
Describe exactly how to reproduce the problem, using a minimal test-case. It would *really* help us understand your problem if you could also pased any playbooks, configs and commands you used.
|
Describe exactly how to reproduce the problem, using a minimal test-case. It would *really* help us understand your problem if you could also passed any playbooks, configs and commands you used.
|
||||||
|
|
||||||
**HINT:** You can paste https://gist.github.com links for larger files.
|
**HINT:** You can paste https://gist.github.com links for larger files.
|
||||||
value: |
|
value: |
|
||||||
|
|
|
@ -21,7 +21,7 @@ DOCUMENTATION = '''
|
||||||
'''
|
'''
|
||||||
|
|
||||||
EXAMPLES = '''
|
EXAMPLES = '''
|
||||||
- name: Count occurences
|
- name: Count occurrences
|
||||||
ansible.builtin.debug:
|
ansible.builtin.debug:
|
||||||
msg: >-
|
msg: >-
|
||||||
{{ [1, 'a', 2, 2, 'a', 'b', 'a'] | community.general.counter }}
|
{{ [1, 'a', 2, 2, 'a', 'b', 'a'] | community.general.counter }}
|
||||||
|
@ -30,7 +30,7 @@ EXAMPLES = '''
|
||||||
|
|
||||||
RETURN = '''
|
RETURN = '''
|
||||||
_value:
|
_value:
|
||||||
description: A dictionary with the elements of the sequence as keys, and their number of occurance in the sequence as values.
|
description: A dictionary with the elements of the sequence as keys, and their number of occurrences in the sequence as values.
|
||||||
type: dictionary
|
type: dictionary
|
||||||
'''
|
'''
|
||||||
|
|
||||||
|
|
|
@ -38,7 +38,7 @@ DOCUMENTATION = '''
|
||||||
parser:
|
parser:
|
||||||
description:
|
description:
|
||||||
- The correct parser for the input data.
|
- The correct parser for the input data.
|
||||||
- For exmaple C(ifconfig).
|
- For example C(ifconfig).
|
||||||
- See U(https://github.com/kellyjonbrazil/jc#parsers) for the latest list of parsers.
|
- See U(https://github.com/kellyjonbrazil/jc#parsers) for the latest list of parsers.
|
||||||
type: string
|
type: string
|
||||||
required: true
|
required: true
|
||||||
|
|
|
@ -213,7 +213,7 @@ class InventoryModule(BaseInventoryPlugin, Cacheable):
|
||||||
self.inventory.add_child(parent_group_name, group_name)
|
self.inventory.add_child(parent_group_name, group_name)
|
||||||
else:
|
else:
|
||||||
self.display.vvvv('Processing profile %s without parent\n' % profile['name'])
|
self.display.vvvv('Processing profile %s without parent\n' % profile['name'])
|
||||||
# Create a heirarchy of profile names
|
# Create a hierarchy of profile names
|
||||||
profile_elements = profile['name'].split('-')
|
profile_elements = profile['name'].split('-')
|
||||||
i = 0
|
i = 0
|
||||||
while i < len(profile_elements) - 1:
|
while i < len(profile_elements) - 1:
|
||||||
|
|
|
@ -522,7 +522,7 @@ class InventoryModule(BaseInventoryPlugin):
|
||||||
"""Helper to save data
|
"""Helper to save data
|
||||||
|
|
||||||
Helper to save the data in self.data
|
Helper to save the data in self.data
|
||||||
Detect if data is allready in branch and use dict_merge() to prevent that branch is overwritten.
|
Detect if data is already in branch and use dict_merge() to prevent that branch is overwritten.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
str(instance_name): name of instance
|
str(instance_name): name of instance
|
||||||
|
|
|
@ -351,7 +351,7 @@ def wait_to_finish(target, pending, refresh, timeout, min_interval=1, delay=3):
|
||||||
|
|
||||||
if pending and status not in pending:
|
if pending and status not in pending:
|
||||||
raise HwcModuleException(
|
raise HwcModuleException(
|
||||||
"unexpect status(%s) occured" % status)
|
"unexpect status(%s) occurred" % status)
|
||||||
|
|
||||||
if not is_last_time:
|
if not is_last_time:
|
||||||
wait *= 2
|
wait *= 2
|
||||||
|
|
|
@ -83,12 +83,12 @@ class OpenNebulaModule:
|
||||||
if self.module.params.get("api_username"):
|
if self.module.params.get("api_username"):
|
||||||
username = self.module.params.get("api_username")
|
username = self.module.params.get("api_username")
|
||||||
else:
|
else:
|
||||||
self.fail("Either api_username or the environment vairable ONE_USERNAME must be provided")
|
self.fail("Either api_username or the environment variable ONE_USERNAME must be provided")
|
||||||
|
|
||||||
if self.module.params.get("api_password"):
|
if self.module.params.get("api_password"):
|
||||||
password = self.module.params.get("api_password")
|
password = self.module.params.get("api_password")
|
||||||
else:
|
else:
|
||||||
self.fail("Either api_password or the environment vairable ONE_PASSWORD must be provided")
|
self.fail("Either api_password or the environment variable ONE_PASSWORD must be provided")
|
||||||
|
|
||||||
session = "%s:%s" % (username, password)
|
session = "%s:%s" % (username, password)
|
||||||
|
|
||||||
|
|
|
@ -691,7 +691,7 @@ def check_and_create_resource(
|
||||||
:param model: Model used to create a resource.
|
:param model: Model used to create a resource.
|
||||||
:param exclude_attributes: The attributes which should not be used to distinguish the resource. e.g. display_name,
|
:param exclude_attributes: The attributes which should not be used to distinguish the resource. e.g. display_name,
|
||||||
dns_label.
|
dns_label.
|
||||||
:param dead_states: List of states which can't transition to any of the usable states of the resource. This deafults
|
:param dead_states: List of states which can't transition to any of the usable states of the resource. This defaults
|
||||||
to ["TERMINATING", "TERMINATED", "FAULTY", "FAILED", "DELETING", "DELETED", "UNKNOWN_ENUM_VALUE"]
|
to ["TERMINATING", "TERMINATED", "FAULTY", "FAILED", "DELETING", "DELETED", "UNKNOWN_ENUM_VALUE"]
|
||||||
:param default_attribute_values: A dictionary containing default values for attributes.
|
:param default_attribute_values: A dictionary containing default values for attributes.
|
||||||
:return: A dictionary containing the resource & the "changed" status. e.g. {"vcn":{x:y}, "changed":True}
|
:return: A dictionary containing the resource & the "changed" status. e.g. {"vcn":{x:y}, "changed":True}
|
||||||
|
@ -1189,7 +1189,7 @@ def are_dicts_equal(
|
||||||
|
|
||||||
|
|
||||||
def should_dict_attr_be_excluded(map_option_name, option_key, exclude_list):
|
def should_dict_attr_be_excluded(map_option_name, option_key, exclude_list):
|
||||||
"""An entry for the Exclude list for excluding a map's key is specifed as a dict with the map option name as the
|
"""An entry for the Exclude list for excluding a map's key is specified as a dict with the map option name as the
|
||||||
key, and the value as a list of keys to be excluded within that map. For example, if the keys "k1" and "k2" of a map
|
key, and the value as a list of keys to be excluded within that map. For example, if the keys "k1" and "k2" of a map
|
||||||
option named "m1" needs to be excluded, the exclude list must have an entry {'m1': ['k1','k2']} """
|
option named "m1" needs to be excluded, the exclude list must have an entry {'m1': ['k1','k2']} """
|
||||||
for exclude_item in exclude_list:
|
for exclude_item in exclude_list:
|
||||||
|
|
|
@ -360,7 +360,7 @@ class LXDProfileManagement(object):
|
||||||
)
|
)
|
||||||
|
|
||||||
def _merge_dicts(self, source, destination):
|
def _merge_dicts(self, source, destination):
|
||||||
"""Merge Dictionarys
|
"""Merge Dictionaries
|
||||||
|
|
||||||
Get a list of filehandle numbers from logger to be handed to
|
Get a list of filehandle numbers from logger to be handed to
|
||||||
DaemonContext.files_preserve
|
DaemonContext.files_preserve
|
||||||
|
|
|
@ -303,7 +303,7 @@ class LXDProjectManagement(object):
|
||||||
)
|
)
|
||||||
|
|
||||||
def _merge_dicts(self, source, destination):
|
def _merge_dicts(self, source, destination):
|
||||||
""" Return a new dict taht merge two dict,
|
""" Return a new dict that merge two dict,
|
||||||
with values in source dict overwrite destination dict
|
with values in source dict overwrite destination dict
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
|
|
|
@ -1204,12 +1204,12 @@ def main():
|
||||||
# Ensure source VM id exists when cloning
|
# Ensure source VM id exists when cloning
|
||||||
proxmox.get_vm(vmid)
|
proxmox.get_vm(vmid)
|
||||||
|
|
||||||
# Ensure the choosen VM name doesn't already exist when cloning
|
# Ensure the chosen VM name doesn't already exist when cloning
|
||||||
existing_vmid = proxmox.get_vmid(name, ignore_missing=True)
|
existing_vmid = proxmox.get_vmid(name, ignore_missing=True)
|
||||||
if existing_vmid:
|
if existing_vmid:
|
||||||
module.exit_json(changed=False, vmid=existing_vmid, msg="VM with name <%s> already exists" % name)
|
module.exit_json(changed=False, vmid=existing_vmid, msg="VM with name <%s> already exists" % name)
|
||||||
|
|
||||||
# Ensure the choosen VM id doesn't already exist when cloning
|
# Ensure the chosen VM id doesn't already exist when cloning
|
||||||
if proxmox.get_vm(newid, ignore_missing=True):
|
if proxmox.get_vm(newid, ignore_missing=True):
|
||||||
module.exit_json(changed=False, vmid=vmid, msg="vmid %s with VM name %s already exists" % (newid, name))
|
module.exit_json(changed=False, vmid=vmid, msg="vmid %s with VM name %s already exists" % (newid, name))
|
||||||
|
|
||||||
|
|
|
@ -1207,7 +1207,7 @@ class XenServerVM(XenServerObject):
|
||||||
if (self.module.params['home_server'] and
|
if (self.module.params['home_server'] and
|
||||||
(not self.vm_params['affinity'] or self.module.params['home_server'] != self.vm_params['affinity']['name_label'])):
|
(not self.vm_params['affinity'] or self.module.params['home_server'] != self.vm_params['affinity']['name_label'])):
|
||||||
|
|
||||||
# Check existance only. Ignore return value.
|
# Check existence only. Ignore return value.
|
||||||
get_object_ref(self.module, self.module.params['home_server'], uuid=None, obj_type="home server", fail=True,
|
get_object_ref(self.module, self.module.params['home_server'], uuid=None, obj_type="home server", fail=True,
|
||||||
msg_prefix="VM check home_server: ")
|
msg_prefix="VM check home_server: ")
|
||||||
|
|
||||||
|
@ -1371,7 +1371,7 @@ class XenServerVM(XenServerObject):
|
||||||
disk_sr = disk_params.get('sr')
|
disk_sr = disk_params.get('sr')
|
||||||
|
|
||||||
if disk_sr_uuid is not None or disk_sr is not None:
|
if disk_sr_uuid is not None or disk_sr is not None:
|
||||||
# Check existance only. Ignore return value.
|
# Check existence only. Ignore return value.
|
||||||
get_object_ref(self.module, disk_sr, disk_sr_uuid, obj_type="SR", fail=True,
|
get_object_ref(self.module, disk_sr, disk_sr_uuid, obj_type="SR", fail=True,
|
||||||
msg_prefix="VM check disks[%s]: " % position)
|
msg_prefix="VM check disks[%s]: " % position)
|
||||||
elif self.default_sr_ref == 'OpaqueRef:NULL':
|
elif self.default_sr_ref == 'OpaqueRef:NULL':
|
||||||
|
@ -1448,7 +1448,7 @@ class XenServerVM(XenServerObject):
|
||||||
|
|
||||||
if cdrom_type == "iso":
|
if cdrom_type == "iso":
|
||||||
# Check if ISO exists.
|
# Check if ISO exists.
|
||||||
# Check existance only. Ignore return value.
|
# Check existence only. Ignore return value.
|
||||||
get_object_ref(self.module, cdrom_iso_name, uuid=None, obj_type="ISO image", fail=True,
|
get_object_ref(self.module, cdrom_iso_name, uuid=None, obj_type="ISO image", fail=True,
|
||||||
msg_prefix="VM check cdrom.iso_name: ")
|
msg_prefix="VM check cdrom.iso_name: ")
|
||||||
|
|
||||||
|
@ -1496,7 +1496,7 @@ class XenServerVM(XenServerObject):
|
||||||
self.module.fail_json(msg="VM check networks[%s]: network name cannot be an empty string!" % position)
|
self.module.fail_json(msg="VM check networks[%s]: network name cannot be an empty string!" % position)
|
||||||
|
|
||||||
if network_name:
|
if network_name:
|
||||||
# Check existance only. Ignore return value.
|
# Check existence only. Ignore return value.
|
||||||
get_object_ref(self.module, network_name, uuid=None, obj_type="network", fail=True,
|
get_object_ref(self.module, network_name, uuid=None, obj_type="network", fail=True,
|
||||||
msg_prefix="VM check networks[%s]: " % position)
|
msg_prefix="VM check networks[%s]: " % position)
|
||||||
|
|
||||||
|
|
|
@ -279,7 +279,7 @@ def do_ini(module, filename, section=None, option=None, values=None,
|
||||||
# handling multiple instances of option=value when state is 'present' with/without exclusive is a bit complex
|
# handling multiple instances of option=value when state is 'present' with/without exclusive is a bit complex
|
||||||
#
|
#
|
||||||
# 1. edit all lines where we have a option=value pair with a matching value in values[]
|
# 1. edit all lines where we have a option=value pair with a matching value in values[]
|
||||||
# 2. edit all the remaing lines where we have a matching option
|
# 2. edit all the remaining lines where we have a matching option
|
||||||
# 3. delete remaining lines where we have a matching option
|
# 3. delete remaining lines where we have a matching option
|
||||||
# 4. insert missing option line(s) at the end of the section
|
# 4. insert missing option line(s) at the end of the section
|
||||||
|
|
||||||
|
|
|
@ -207,7 +207,7 @@ def main():
|
||||||
changed = True
|
changed = True
|
||||||
else:
|
else:
|
||||||
changed = False
|
changed = False
|
||||||
out = "allready unpacked"
|
out = "already unpacked"
|
||||||
|
|
||||||
if remove:
|
if remove:
|
||||||
os.remove(path)
|
os.remove(path)
|
||||||
|
|
|
@ -45,7 +45,7 @@ EXAMPLES = r'''
|
||||||
ipa_user: admin
|
ipa_user: admin
|
||||||
ipa_pass: supersecret
|
ipa_pass: supersecret
|
||||||
|
|
||||||
- name: Ensure the TOTP syncronization window is set to 86400 seconds
|
- name: Ensure the TOTP synchronization window is set to 86400 seconds
|
||||||
community.general.ipa_otpconfig:
|
community.general.ipa_otpconfig:
|
||||||
ipatokentotpsyncwindow: '86400'
|
ipatokentotpsyncwindow: '86400'
|
||||||
ipa_host: localhost
|
ipa_host: localhost
|
||||||
|
@ -59,7 +59,7 @@ EXAMPLES = r'''
|
||||||
ipa_user: admin
|
ipa_user: admin
|
||||||
ipa_pass: supersecret
|
ipa_pass: supersecret
|
||||||
|
|
||||||
- name: Ensure the HOTP syncronization window is set to 100 hops
|
- name: Ensure the HOTP synchronization window is set to 100 hops
|
||||||
community.general.ipa_otpconfig:
|
community.general.ipa_otpconfig:
|
||||||
ipatokenhotpsyncwindow: '100'
|
ipatokenhotpsyncwindow: '100'
|
||||||
ipa_host: localhost
|
ipa_host: localhost
|
||||||
|
|
|
@ -63,7 +63,7 @@ options:
|
||||||
type: str
|
type: str
|
||||||
replace:
|
replace:
|
||||||
description:
|
description:
|
||||||
- Force replace the existant vault on IPA server.
|
- Force replace the existent vault on IPA server.
|
||||||
type: bool
|
type: bool
|
||||||
default: False
|
default: False
|
||||||
choices: ["True", "False"]
|
choices: ["True", "False"]
|
||||||
|
|
|
@ -64,7 +64,7 @@ msg:
|
||||||
|
|
||||||
realm_info:
|
realm_info:
|
||||||
description:
|
description:
|
||||||
- Representation of the realm public infomation.
|
- Representation of the realm public information.
|
||||||
returned: always
|
returned: always
|
||||||
type: dict
|
type: dict
|
||||||
contains:
|
contains:
|
||||||
|
|
|
@ -1501,7 +1501,7 @@ class Nmcli(object):
|
||||||
if self._hairpin is None:
|
if self._hairpin is None:
|
||||||
self.module.deprecate(
|
self.module.deprecate(
|
||||||
"Parameter 'hairpin' default value will change from true to false in community.general 7.0.0. "
|
"Parameter 'hairpin' default value will change from true to false in community.general 7.0.0. "
|
||||||
"Set the value explicitly to supress this warning.",
|
"Set the value explicitly to suppress this warning.",
|
||||||
version='7.0.0', collection_name='community.general',
|
version='7.0.0', collection_name='community.general',
|
||||||
)
|
)
|
||||||
# Should be False in 7.0.0 but then that should be in argument_specs
|
# Should be False in 7.0.0 but then that should be in argument_specs
|
||||||
|
|
|
@ -242,7 +242,7 @@ class AnsibleGalaxyInstall(CmdModuleHelper):
|
||||||
self.module.deprecate(
|
self.module.deprecate(
|
||||||
"Support for Ansible 2.9 and ansible-base 2.10 is being deprecated. "
|
"Support for Ansible 2.9 and ansible-base 2.10 is being deprecated. "
|
||||||
"At the same time support for them is ended, also the ack_ansible29 option will be removed. "
|
"At the same time support for them is ended, also the ack_ansible29 option will be removed. "
|
||||||
"Upgrading is strongly recommended, or set 'ack_min_ansiblecore211' to supress this message.",
|
"Upgrading is strongly recommended, or set 'ack_min_ansiblecore211' to suppress this message.",
|
||||||
version="8.0.0",
|
version="8.0.0",
|
||||||
collection_name="community.general",
|
collection_name="community.general",
|
||||||
)
|
)
|
||||||
|
|
|
@ -81,7 +81,7 @@ options:
|
||||||
classmap_authoritative:
|
classmap_authoritative:
|
||||||
description:
|
description:
|
||||||
- Autoload classes from classmap only.
|
- Autoload classes from classmap only.
|
||||||
- Implicitely enable optimize_autoloader.
|
- Implicitly enable optimize_autoloader.
|
||||||
- Recommended especially for production, but can take a bit of time to run.
|
- Recommended especially for production, but can take a bit of time to run.
|
||||||
default: false
|
default: false
|
||||||
type: bool
|
type: bool
|
||||||
|
|
|
@ -74,7 +74,7 @@ ilo_redfish_info:
|
||||||
type: dict
|
type: dict
|
||||||
contains:
|
contains:
|
||||||
ret:
|
ret:
|
||||||
description: Check variable to see if the information was succesfully retrived.
|
description: Check variable to see if the information was successfully retrieved.
|
||||||
type: bool
|
type: bool
|
||||||
msg:
|
msg:
|
||||||
description: Information of all active iLO sessions.
|
description: Information of all active iLO sessions.
|
||||||
|
|
|
@ -36,7 +36,7 @@ options:
|
||||||
name:
|
name:
|
||||||
description:
|
description:
|
||||||
- The name of the branch that needs to be protected.
|
- The name of the branch that needs to be protected.
|
||||||
- Can make use a wildcard charachter for like C(production/*) or just have C(main) or C(develop) as value.
|
- Can make use a wildcard character for like C(production/*) or just have C(main) or C(develop) as value.
|
||||||
required: true
|
required: true
|
||||||
type: str
|
type: str
|
||||||
merge_access_levels:
|
merge_access_levels:
|
||||||
|
|
|
@ -305,7 +305,7 @@ class GitLabUser(object):
|
||||||
# note: as we unfortunately have some uncheckable parameters
|
# note: as we unfortunately have some uncheckable parameters
|
||||||
# where it is not possible to determine if the update
|
# where it is not possible to determine if the update
|
||||||
# changed something or not, we must assume here that a
|
# changed something or not, we must assume here that a
|
||||||
# changed happend and that an user object update is needed
|
# changed happened and that an user object update is needed
|
||||||
potentionally_changed = True
|
potentionally_changed = True
|
||||||
|
|
||||||
# Assign ssh keys
|
# Assign ssh keys
|
||||||
|
|
|
@ -99,7 +99,7 @@ options:
|
||||||
- The size of namespace. This option supports the suffixes C(k) or C(K) or C(KB) for KiB,
|
- The size of namespace. This option supports the suffixes C(k) or C(K) or C(KB) for KiB,
|
||||||
C(m) or C(M) or C(MB) for MiB, C(g) or C(G) or C(GB) for GiB and C(t) or C(T) or C(TB) for TiB.
|
C(m) or C(M) or C(MB) for MiB, C(g) or C(G) or C(GB) for GiB and C(t) or C(T) or C(TB) for TiB.
|
||||||
- This option is required if multiple namespaces are configured.
|
- This option is required if multiple namespaces are configured.
|
||||||
- If this option is not set, all of the avaiable space of a region is configured.
|
- If this option is not set, all of the available space of a region is configured.
|
||||||
type: str
|
type: str
|
||||||
required: false
|
required: false
|
||||||
namespace_append:
|
namespace_append:
|
||||||
|
|
|
@ -57,7 +57,7 @@ options:
|
||||||
realname:
|
realname:
|
||||||
description:
|
description:
|
||||||
- The user's real ('human') name.
|
- The user's real ('human') name.
|
||||||
- This can also be used to add a comment to maintain compatability with C(useradd).
|
- This can also be used to add a comment to maintain compatibility with C(useradd).
|
||||||
aliases: [ 'comment' ]
|
aliases: [ 'comment' ]
|
||||||
type: str
|
type: str
|
||||||
realm:
|
realm:
|
||||||
|
|
|
@ -790,7 +790,7 @@ class AIXTimezone(Timezone):
|
||||||
inspects C(/etc/environment) to determine the current timezone.
|
inspects C(/etc/environment) to determine the current timezone.
|
||||||
|
|
||||||
While AIX time zones can be set using two formats (POSIX and
|
While AIX time zones can be set using two formats (POSIX and
|
||||||
Olson) the prefered method is Olson.
|
Olson) the preferred method is Olson.
|
||||||
See the following article for more information:
|
See the following article for more information:
|
||||||
https://developer.ibm.com/articles/au-aix-posix/
|
https://developer.ibm.com/articles/au-aix-posix/
|
||||||
|
|
||||||
|
|
|
@ -14,7 +14,7 @@
|
||||||
- "valid_comma_separated_spaces | community.general.from_csv(skipinitialspace=True) == expected_result"
|
- "valid_comma_separated_spaces | community.general.from_csv(skipinitialspace=True) == expected_result"
|
||||||
- "valid_comma_separated_spaces | community.general.from_csv != expected_result"
|
- "valid_comma_separated_spaces | community.general.from_csv != expected_result"
|
||||||
|
|
||||||
- name: Parse valid csv input with no headers with/without specifiying fieldnames
|
- name: Parse valid csv input with no headers with/without specifying fieldnames
|
||||||
assert:
|
assert:
|
||||||
that:
|
that:
|
||||||
- "valid_comma_separated_no_headers | community.general.from_csv(fieldnames=['id','name','role']) == expected_result"
|
- "valid_comma_separated_no_headers | community.general.from_csv(fieldnames=['id','name','role']) == expected_result"
|
||||||
|
|
|
@ -105,7 +105,7 @@
|
||||||
- iso_result is changed
|
- iso_result is changed
|
||||||
- iso_file.stat.exists == True
|
- iso_file.stat.exists == True
|
||||||
|
|
||||||
- name: Create iso file with Rock Ridge extention
|
- name: Create iso file with Rock Ridge extension
|
||||||
iso_create:
|
iso_create:
|
||||||
src_files:
|
src_files:
|
||||||
- "{{ remote_tmp_dir }}/test1.cfg"
|
- "{{ remote_tmp_dir }}/test1.cfg"
|
||||||
|
@ -124,7 +124,7 @@
|
||||||
- iso_result is changed
|
- iso_result is changed
|
||||||
- iso_file.stat.exists == True
|
- iso_file.stat.exists == True
|
||||||
|
|
||||||
- name: Create iso file with Joliet extention
|
- name: Create iso file with Joliet extension
|
||||||
iso_create:
|
iso_create:
|
||||||
src_files:
|
src_files:
|
||||||
- "{{ remote_tmp_dir }}/test1.cfg"
|
- "{{ remote_tmp_dir }}/test1.cfg"
|
||||||
|
|
|
@ -66,7 +66,7 @@
|
||||||
|
|
||||||
|
|
||||||
- name: Run tests
|
- name: Run tests
|
||||||
# Skip tests on Fedora 31 and 32 because dbus fails to start unless the container is run in priveleged mode.
|
# Skip tests on Fedora 31 and 32 because dbus fails to start unless the container is run in privileged mode.
|
||||||
# Even then, it starts unreliably. This may be due to the move to cgroup v2 in Fedora 31 and 32.
|
# Even then, it starts unreliably. This may be due to the move to cgroup v2 in Fedora 31 and 32.
|
||||||
# https://www.redhat.com/sysadmin/fedora-31-control-group-v2
|
# https://www.redhat.com/sysadmin/fedora-31-control-group-v2
|
||||||
when:
|
when:
|
||||||
|
|
|
@ -107,7 +107,7 @@ def test_validate_config(inventory):
|
||||||
}
|
}
|
||||||
with pytest.raises(AnsibleError) as error_message:
|
with pytest.raises(AnsibleError) as error_message:
|
||||||
inventory._validate_config(config)
|
inventory._validate_config(config)
|
||||||
assert "config missing client_secret, a required paramter" in error_message
|
assert "config missing client_secret, a required parameter" in error_message
|
||||||
|
|
||||||
config = {
|
config = {
|
||||||
"client_secret": "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx",
|
"client_secret": "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx",
|
||||||
|
@ -116,7 +116,7 @@ def test_validate_config(inventory):
|
||||||
}
|
}
|
||||||
with pytest.raises(AnsibleError) as error_message:
|
with pytest.raises(AnsibleError) as error_message:
|
||||||
inventory._validate_config(config)
|
inventory._validate_config(config)
|
||||||
assert "config missing client_id, a required paramter" in error_message
|
assert "config missing client_id, a required parameter" in error_message
|
||||||
|
|
||||||
|
|
||||||
def test_populate(inventory):
|
def test_populate(inventory):
|
||||||
|
|
|
@ -39,7 +39,7 @@ def XenAPI():
|
||||||
"""Imports and returns fake XenAPI module."""
|
"""Imports and returns fake XenAPI module."""
|
||||||
|
|
||||||
# Import of fake XenAPI module is wrapped by fixture so that it does not
|
# Import of fake XenAPI module is wrapped by fixture so that it does not
|
||||||
# affect other unit tests which could potentialy also use XenAPI module.
|
# affect other unit tests which could potentially also use XenAPI module.
|
||||||
|
|
||||||
# First we use importlib.import_module() to import the module and assign
|
# First we use importlib.import_module() to import the module and assign
|
||||||
# it to a local symbol.
|
# it to a local symbol.
|
||||||
|
|
|
@ -35,7 +35,7 @@ def XenAPI():
|
||||||
"""Imports and returns fake XenAPI module."""
|
"""Imports and returns fake XenAPI module."""
|
||||||
|
|
||||||
# Import of fake XenAPI module is wrapped by fixture so that it does not
|
# Import of fake XenAPI module is wrapped by fixture so that it does not
|
||||||
# affect other unit tests which could potentialy also use XenAPI module.
|
# affect other unit tests which could potentially also use XenAPI module.
|
||||||
|
|
||||||
# First we use importlib.import_module() to import the module and assign
|
# First we use importlib.import_module() to import the module and assign
|
||||||
# it to a local symbol.
|
# it to a local symbol.
|
||||||
|
|
|
@ -109,7 +109,7 @@ class TestPritunlOrg(ModuleTestCase):
|
||||||
idempotent_exc = idempotent_result.exception.args[0]
|
idempotent_exc = idempotent_result.exception.args[0]
|
||||||
|
|
||||||
# Ensure both calls resulted in the same returned value
|
# Ensure both calls resulted in the same returned value
|
||||||
# except for changed which sould be false the second time
|
# except for changed which should be false the second time
|
||||||
for k, v in iteritems(idempotent_exc):
|
for k, v in iteritems(idempotent_exc):
|
||||||
if k == "changed":
|
if k == "changed":
|
||||||
self.assertFalse(idempotent_exc[k])
|
self.assertFalse(idempotent_exc[k])
|
||||||
|
@ -158,7 +158,7 @@ class TestPritunlOrg(ModuleTestCase):
|
||||||
idempotent_exc = idempotent_result.exception.args[0]
|
idempotent_exc = idempotent_result.exception.args[0]
|
||||||
|
|
||||||
# Ensure both calls resulted in the same returned value
|
# Ensure both calls resulted in the same returned value
|
||||||
# except for changed which sould be false the second time
|
# except for changed which should be false the second time
|
||||||
self.assertFalse(idempotent_exc["changed"])
|
self.assertFalse(idempotent_exc["changed"])
|
||||||
self.assertEqual(idempotent_exc["response"], delete_exc["response"])
|
self.assertEqual(idempotent_exc["response"], delete_exc["response"])
|
||||||
|
|
||||||
|
|
|
@ -3309,7 +3309,7 @@ def test_gsm_connection_unchanged(mocked_gsm_connection_unchanged, capfd):
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP4_ADDRESSES, indirect=['patch_ansible_module'])
|
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP4_ADDRESSES, indirect=['patch_ansible_module'])
|
||||||
def test_create_ethernet_with_mulitple_ip4_addresses_static(mocked_generic_connection_create, capfd):
|
def test_create_ethernet_with_multiple_ip4_addresses_static(mocked_generic_connection_create, capfd):
|
||||||
"""
|
"""
|
||||||
Test : Create ethernet connection with static IP configuration
|
Test : Create ethernet connection with static IP configuration
|
||||||
"""
|
"""
|
||||||
|
@ -3349,7 +3349,7 @@ def test_create_ethernet_with_mulitple_ip4_addresses_static(mocked_generic_conne
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP6_ADDRESSES, indirect=['patch_ansible_module'])
|
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP6_ADDRESSES, indirect=['patch_ansible_module'])
|
||||||
def test_create_ethernet_with_mulitple_ip6_addresses_static(mocked_generic_connection_create, capfd):
|
def test_create_ethernet_with_multiple_ip6_addresses_static(mocked_generic_connection_create, capfd):
|
||||||
"""
|
"""
|
||||||
Test : Create ethernet connection with multiple IPv6 addresses configuration
|
Test : Create ethernet connection with multiple IPv6 addresses configuration
|
||||||
"""
|
"""
|
||||||
|
@ -3389,7 +3389,7 @@ def test_create_ethernet_with_mulitple_ip6_addresses_static(mocked_generic_conne
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP4_ADDRESSES, indirect=['patch_ansible_module'])
|
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP4_ADDRESSES, indirect=['patch_ansible_module'])
|
||||||
def test_ethernet_connection_static_with_mulitple_ip4_addresses_unchanged(mocked_ethernet_connection_static_multiple_ip4_addresses_unchanged, capfd):
|
def test_ethernet_connection_static_with_multiple_ip4_addresses_unchanged(mocked_ethernet_connection_static_multiple_ip4_addresses_unchanged, capfd):
|
||||||
"""
|
"""
|
||||||
Test : Ethernet connection with static IP configuration unchanged
|
Test : Ethernet connection with static IP configuration unchanged
|
||||||
"""
|
"""
|
||||||
|
@ -3403,7 +3403,7 @@ def test_ethernet_connection_static_with_mulitple_ip4_addresses_unchanged(mocked
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP6_ADDRESSES, indirect=['patch_ansible_module'])
|
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_MULTIPLE_IP6_ADDRESSES, indirect=['patch_ansible_module'])
|
||||||
def test_ethernet_connection_static_with_mulitple_ip6_addresses_unchanged(mocked_ethernet_connection_static_multiple_ip6_addresses_unchanged, capfd):
|
def test_ethernet_connection_static_with_multiple_ip6_addresses_unchanged(mocked_ethernet_connection_static_multiple_ip6_addresses_unchanged, capfd):
|
||||||
"""
|
"""
|
||||||
Test : Ethernet connection with multiple IPv6 addresses configuration unchanged
|
Test : Ethernet connection with multiple IPv6 addresses configuration unchanged
|
||||||
"""
|
"""
|
||||||
|
@ -3485,7 +3485,7 @@ def test_create_ethernet_addr_gen_mode_and_ip6_privacy_static(mocked_generic_con
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_IP6_PRIVACY_AND_ADDR_GEN_MODE, indirect=['patch_ansible_module'])
|
@pytest.mark.parametrize('patch_ansible_module', TESTCASE_ETHERNET_STATIC_IP6_PRIVACY_AND_ADDR_GEN_MODE, indirect=['patch_ansible_module'])
|
||||||
def test_ethernet_connection_static_with_mulitple_ip4_addresses_unchanged(mocked_ethernet_connection_static_ip6_privacy_and_addr_gen_mode_unchange, capfd):
|
def test_ethernet_connection_static_with_multiple_ip4_addresses_unchanged(mocked_ethernet_connection_static_ip6_privacy_and_addr_gen_mode_unchange, capfd):
|
||||||
"""
|
"""
|
||||||
Test : Ethernet connection with static IP configuration unchanged
|
Test : Ethernet connection with static IP configuration unchanged
|
||||||
"""
|
"""
|
||||||
|
|
|
@ -37,7 +37,7 @@ class TestSlackModule(ModuleTestCase):
|
||||||
with self.assertRaises(AnsibleFailJson):
|
with self.assertRaises(AnsibleFailJson):
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
def test_sucessful_message(self):
|
def test_successful_message(self):
|
||||||
"""tests sending a message. This is example 1 from the docs"""
|
"""tests sending a message. This is example 1 from the docs"""
|
||||||
set_module_args({
|
set_module_args({
|
||||||
'token': 'XXXX/YYYY/ZZZZ',
|
'token': 'XXXX/YYYY/ZZZZ',
|
||||||
|
|
|
@ -19,7 +19,7 @@ TESTED_MODULE = cpanm.__name__
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def patch_cpanm(mocker):
|
def patch_cpanm(mocker):
|
||||||
"""
|
"""
|
||||||
Function used for mocking some parts of redhat_subscribtion module
|
Function used for mocking some parts of redhat_subscription module
|
||||||
"""
|
"""
|
||||||
mocker.patch('ansible_collections.community.general.plugins.module_utils.module_helper.AnsibleModule.get_bin_path',
|
mocker.patch('ansible_collections.community.general.plugins.module_utils.module_helper.AnsibleModule.get_bin_path',
|
||||||
return_value='/testbin/cpanm')
|
return_value='/testbin/cpanm')
|
||||||
|
|
|
@ -3,7 +3,7 @@
|
||||||
## Tests structure
|
## Tests structure
|
||||||
|
|
||||||
- `input` directory contains interfaces configuration files
|
- `input` directory contains interfaces configuration files
|
||||||
- `test_interfaces_file.py` runs each hardcoded test agains all configurations in `input` directory and compares results with golden outputs in `golden_output`
|
- `test_interfaces_file.py` runs each hardcoded test against all configurations in `input` directory and compares results with golden outputs in `golden_output`
|
||||||
|
|
||||||
## Running unit tests with docker
|
## Running unit tests with docker
|
||||||
|
|
||||||
|
|
|
@ -17,7 +17,7 @@ TESTED_MODULE = gconftool2_info.__name__
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def patch_gconftool2_info(mocker):
|
def patch_gconftool2_info(mocker):
|
||||||
"""
|
"""
|
||||||
Function used for mocking some parts of redhat_subscribtion module
|
Function used for mocking some parts of redhat_subscription module
|
||||||
"""
|
"""
|
||||||
mocker.patch('ansible_collections.community.general.plugins.module_utils.mh.module_helper.AnsibleModule.get_bin_path',
|
mocker.patch('ansible_collections.community.general.plugins.module_utils.mh.module_helper.AnsibleModule.get_bin_path',
|
||||||
return_value='/testbin/gconftool-2')
|
return_value='/testbin/gconftool-2')
|
||||||
|
|
|
@ -19,7 +19,7 @@ TESTED_MODULE = xfconf.__name__
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def patch_xfconf(mocker):
|
def patch_xfconf(mocker):
|
||||||
"""
|
"""
|
||||||
Function used for mocking some parts of redhat_subscribtion module
|
Function used for mocking some parts of redhat_subscription module
|
||||||
"""
|
"""
|
||||||
mocker.patch('ansible_collections.community.general.plugins.module_utils.mh.module_helper.AnsibleModule.get_bin_path',
|
mocker.patch('ansible_collections.community.general.plugins.module_utils.mh.module_helper.AnsibleModule.get_bin_path',
|
||||||
return_value='/testbin/xfconf-query')
|
return_value='/testbin/xfconf-query')
|
||||||
|
|
|
@ -16,7 +16,7 @@ TESTED_MODULE = xfconf_info.__name__
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def patch_xfconf_info(mocker):
|
def patch_xfconf_info(mocker):
|
||||||
"""
|
"""
|
||||||
Function used for mocking some parts of redhat_subscribtion module
|
Function used for mocking some parts of redhat_subscription module
|
||||||
"""
|
"""
|
||||||
mocker.patch('ansible_collections.community.general.plugins.module_utils.mh.module_helper.AnsibleModule.get_bin_path',
|
mocker.patch('ansible_collections.community.general.plugins.module_utils.mh.module_helper.AnsibleModule.get_bin_path',
|
||||||
return_value='/testbin/xfconf-query')
|
return_value='/testbin/xfconf-query')
|
||||||
|
|
Loading…
Reference in a new issue