mirror of
https://github.com/ansible-collections/community.general.git
synced 2024-09-14 20:13:21 +02:00
parent
887456ab8e
commit
02f66b9369
38 changed files with 865 additions and 888 deletions
|
@ -249,9 +249,9 @@ class CLI(with_metaclass(ABCMeta, object)):
|
|||
|
||||
if runas_opts:
|
||||
# Check for privilege escalation conflicts
|
||||
if (op.su or op.su_user) and (op.sudo or op.sudo_user) or \
|
||||
(op.su or op.su_user) and (op.become or op.become_user) or \
|
||||
(op.sudo or op.sudo_user) and (op.become or op.become_user):
|
||||
if ((op.su or op.su_user) and (op.sudo or op.sudo_user) or
|
||||
(op.su or op.su_user) and (op.become or op.become_user) or
|
||||
(op.sudo or op.sudo_user) and (op.become or op.become_user)):
|
||||
|
||||
self.parser.error("Sudo arguments ('--sudo', '--sudo-user', and '--ask-sudo-pass') "
|
||||
"and su arguments ('-su', '--su-user', and '--ask-su-pass') "
|
||||
|
|
|
@ -99,11 +99,9 @@ class HostState:
|
|||
if not isinstance(other, HostState):
|
||||
return False
|
||||
|
||||
for attr in (
|
||||
'_blocks', 'cur_block', 'cur_regular_task', 'cur_rescue_task', 'cur_always_task',
|
||||
for attr in ('_blocks', 'cur_block', 'cur_regular_task', 'cur_rescue_task', 'cur_always_task',
|
||||
'run_state', 'fail_state', 'pending_setup', 'cur_dep_chain',
|
||||
'tasks_child_state', 'rescue_child_state', 'always_child_state'
|
||||
):
|
||||
'tasks_child_state', 'rescue_child_state', 'always_child_state'):
|
||||
if getattr(self, attr) != getattr(other, attr):
|
||||
return False
|
||||
|
||||
|
|
|
@ -2498,8 +2498,8 @@ class LinuxNetwork(Network):
|
|||
)
|
||||
interface = dict(v4 = {}, v6 = {})
|
||||
for v in 'v4', 'v6':
|
||||
if v == 'v6' and self.facts['os_family'] == 'RedHat' \
|
||||
and self.facts['distribution_version'].startswith('4.'):
|
||||
if (v == 'v6' and self.facts['os_family'] == 'RedHat' and
|
||||
self.facts['distribution_version'].startswith('4.')):
|
||||
continue
|
||||
if v == 'v6' and not socket.has_ipv6:
|
||||
continue
|
||||
|
|
|
@ -276,8 +276,8 @@ class SnsTopicManager(object):
|
|||
for sub in self.subscriptions_existing:
|
||||
sub_key = (sub['Protocol'], sub['Endpoint'])
|
||||
subscriptions_existing_list.append(sub_key)
|
||||
if self.purge_subscriptions and sub_key not in desired_subscriptions and \
|
||||
sub['SubscriptionArn'] not in ('PendingConfirmation', 'Deleted'):
|
||||
if (self.purge_subscriptions and sub_key not in desired_subscriptions and
|
||||
sub['SubscriptionArn'] not in ('PendingConfirmation', 'Deleted')):
|
||||
self.changed = True
|
||||
self.subscriptions_deleted.append(sub_key)
|
||||
if not self.check_mode:
|
||||
|
|
|
@ -847,13 +847,13 @@ class AzureRMVirtualMachine(AzureRMModuleBase):
|
|||
self.create_or_update_vm(vm_resource)
|
||||
|
||||
# Make sure we leave the machine in requested power state
|
||||
if powerstate_change == 'poweron' and \
|
||||
self.results['ansible_facts']['azure_vm']['powerstate'] != 'running':
|
||||
if (powerstate_change == 'poweron' and
|
||||
self.results['ansible_facts']['azure_vm']['powerstate'] != 'running'):
|
||||
# Attempt to power on the machine
|
||||
self.power_on_vm()
|
||||
|
||||
elif powerstate_change == 'poweroff' and \
|
||||
self.results['ansible_facts']['azure_vm']['powerstate'] == 'running':
|
||||
elif (powerstate_change == 'poweroff' and
|
||||
self.results['ansible_facts']['azure_vm']['powerstate'] == 'running'):
|
||||
# Attempt to power off the machine
|
||||
self.power_off_vm()
|
||||
|
||||
|
|
|
@ -275,8 +275,8 @@ class AnsibleCloudStackPortforwarding(AnsibleCloudStack):
|
|||
|
||||
if portforwarding_rules and 'portforwardingrule' in portforwarding_rules:
|
||||
for rule in portforwarding_rules['portforwardingrule']:
|
||||
if protocol == rule['protocol'] \
|
||||
and public_port == int(rule['publicport']):
|
||||
if (protocol == rule['protocol'] and
|
||||
public_port == int(rule['publicport'])):
|
||||
self.portforwarding_rule = rule
|
||||
break
|
||||
return self.portforwarding_rule
|
||||
|
|
|
@ -198,7 +198,14 @@ EXAMPLES = '''
|
|||
ostemplate: 'local:vztmpl/ubuntu-14.04-x86_64.tar.gz'
|
||||
|
||||
# Create new container automatically selecting the next available vmid.
|
||||
- proxmox: node='uk-mc02' api_user='root@pam' api_password='1q2w3e' api_host='node1' password='123456' hostname='example.org' ostemplate='local:vztmpl/ubuntu-14.04-x86_64.tar.gz'
|
||||
- proxmox:
|
||||
node: 'uk-mc02'
|
||||
api_user: 'root@pam'
|
||||
api_password: '1q2w3e'
|
||||
api_host: 'node1'
|
||||
password: '123456'
|
||||
hostname: 'example.org'
|
||||
ostemplate: 'local:vztmpl/ubuntu-14.04-x86_64.tar.gz'
|
||||
|
||||
# Create new container with minimal options with force(it will rewrite existing container)
|
||||
- proxmox:
|
||||
|
@ -354,13 +361,13 @@ def create_instance(module, proxmox, vmid, node, disk, storage, cpus, memory, sw
|
|||
taskid = getattr(proxmox_node, VZ_TYPE).create(vmid=vmid, storage=storage, memory=memory, swap=swap, **kwargs)
|
||||
|
||||
while timeout:
|
||||
if ( proxmox_node.tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox_node.tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
if (proxmox_node.tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox_node.tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
return True
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
module.fail_json(msg='Reached timeout while waiting for creating VM. Last line in task before timeout: %s'
|
||||
% proxmox_node.tasks(taskid).log.get()[:1])
|
||||
module.fail_json(msg='Reached timeout while waiting for creating VM. Last line in task before timeout: %s' %
|
||||
proxmox_node.tasks(taskid).log.get()[:1])
|
||||
|
||||
time.sleep(1)
|
||||
return False
|
||||
|
@ -368,13 +375,13 @@ def create_instance(module, proxmox, vmid, node, disk, storage, cpus, memory, sw
|
|||
def start_instance(module, proxmox, vm, vmid, timeout):
|
||||
taskid = getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.start.post()
|
||||
while timeout:
|
||||
if ( proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
return True
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
module.fail_json(msg='Reached timeout while waiting for starting VM. Last line in task before timeout: %s'
|
||||
% proxmox.nodes(vm[0]['node']).tasks(taskid).log.get()[:1])
|
||||
module.fail_json(msg='Reached timeout while waiting for starting VM. Last line in task before timeout: %s' %
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).log.get()[:1])
|
||||
|
||||
time.sleep(1)
|
||||
return False
|
||||
|
@ -385,13 +392,13 @@ def stop_instance(module, proxmox, vm, vmid, timeout, force):
|
|||
else:
|
||||
taskid = getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.shutdown.post()
|
||||
while timeout:
|
||||
if ( proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
return True
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
module.fail_json(msg='Reached timeout while waiting for stopping VM. Last line in task before timeout: %s'
|
||||
% proxmox_node.tasks(taskid).log.get()[:1])
|
||||
module.fail_json(msg='Reached timeout while waiting for stopping VM. Last line in task before timeout: %s' %
|
||||
proxmox_node.tasks(taskid).log.get()[:1])
|
||||
|
||||
time.sleep(1)
|
||||
return False
|
||||
|
@ -399,13 +406,13 @@ def stop_instance(module, proxmox, vm, vmid, timeout, force):
|
|||
def umount_instance(module, proxmox, vm, vmid, timeout):
|
||||
taskid = getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.umount.post()
|
||||
while timeout:
|
||||
if ( proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
return True
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
module.fail_json(msg='Reached timeout while waiting for unmounting VM. Last line in task before timeout: %s'
|
||||
% proxmox_node.tasks(taskid).log.get()[:1])
|
||||
module.fail_json(msg='Reached timeout while waiting for unmounting VM. Last line in task before timeout: %s' %
|
||||
proxmox_node.tasks(taskid).log.get()[:1])
|
||||
|
||||
time.sleep(1)
|
||||
return False
|
||||
|
@ -562,12 +569,12 @@ def main():
|
|||
vm = get_instance(proxmox, vmid)
|
||||
if not vm:
|
||||
module.fail_json(msg='VM with vmid = %s not exists in cluster' % vmid)
|
||||
if ( getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.current.get()['status'] == 'stopped'
|
||||
or getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.current.get()['status'] == 'mounted' ):
|
||||
if (getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.current.get()['status'] == 'stopped' or
|
||||
getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.current.get()['status'] == 'mounted'):
|
||||
module.exit_json(changed=False, msg="VM %s is not running" % vmid)
|
||||
|
||||
if ( stop_instance(module, proxmox, vm, vmid, timeout, force = module.params['force']) and
|
||||
start_instance(module, proxmox, vm, vmid, timeout) ):
|
||||
if (stop_instance(module, proxmox, vm, vmid, timeout, force = module.params['force']) and
|
||||
start_instance(module, proxmox, vm, vmid, timeout)):
|
||||
module.exit_json(changed=True, msg="VM %s is restarted" % vmid)
|
||||
except Exception as e:
|
||||
module.fail_json(msg="restarting of VM %s failed with exception: %s" % ( vmid, e ))
|
||||
|
@ -586,8 +593,8 @@ def main():
|
|||
|
||||
taskid = getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE).delete(vmid)
|
||||
while timeout:
|
||||
if ( proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
module.exit_json(changed=True, msg="VM %s removed" % vmid)
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
|
|
|
@ -858,13 +858,13 @@ def create_vm(module, proxmox, vmid, newid, node, name, memory, cpu, cores, sock
|
|||
taskid = getattr(proxmox_node, VZ_TYPE).create(vmid=vmid, name=name, memory=memory, cpu=cpu, cores=cores, sockets=sockets, **kwargs)
|
||||
|
||||
while timeout:
|
||||
if (proxmox_node.tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox_node.tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
if (proxmox_node.tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox_node.tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
return True
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
module.fail_json(msg='Reached timeout while waiting for creating VM. Last line in task before timeout: %s'
|
||||
% proxmox_node.tasks(taskid).log.get()[:1])
|
||||
module.fail_json(msg='Reached timeout while waiting for creating VM. Last line in task before timeout: %s' %
|
||||
proxmox_node.tasks(taskid).log.get()[:1])
|
||||
time.sleep(1)
|
||||
return False
|
||||
|
||||
|
@ -872,8 +872,8 @@ def create_vm(module, proxmox, vmid, newid, node, name, memory, cpu, cores, sock
|
|||
def start_vm(module, proxmox, vm, vmid, timeout):
|
||||
taskid = getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.start.post()
|
||||
while timeout:
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
return True
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
|
@ -890,8 +890,8 @@ def stop_vm(module, proxmox, vm, vmid, timeout, force):
|
|||
else:
|
||||
taskid = getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE)(vmid).status.shutdown.post()
|
||||
while timeout:
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK'):
|
||||
return True
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
|
@ -1207,8 +1207,8 @@ def main():
|
|||
|
||||
taskid = getattr(proxmox.nodes(vm[0]['node']), VZ_TYPE).delete(vmid)
|
||||
while timeout:
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped'
|
||||
and proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
if (proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['status'] == 'stopped' and
|
||||
proxmox.nodes(vm[0]['node']).tasks(taskid).status.get()['exitstatus'] == 'OK' ):
|
||||
module.exit_json(changed=True, msg="VM %s removed" % vmid)
|
||||
timeout = timeout - 1
|
||||
if timeout == 0:
|
||||
|
|
|
@ -90,10 +90,10 @@ def _needs_update(module, aggregate):
|
|||
new_metadata = (module.params['metadata'] or {})
|
||||
new_metadata['availability_zone'] = module.params['availability_zone']
|
||||
|
||||
if (module.params['name'] != aggregate.name) or \
|
||||
(module.params['hosts'] is not None and module.params['hosts'] != aggregate.hosts) or \
|
||||
(module.params['availability_zone'] is not None and module.params['availability_zone'] != aggregate.availability_zone) or \
|
||||
(module.params['metadata'] is not None and new_metadata != aggregate.metadata):
|
||||
if ((module.params['name'] != aggregate.name) or
|
||||
(module.params['hosts'] is not None and module.params['hosts'] != aggregate.hosts) or
|
||||
(module.params['availability_zone'] is not None and module.params['availability_zone'] != aggregate.availability_zone) or
|
||||
(module.params['metadata'] is not None and new_metadata != aggregate.metadata)):
|
||||
return True
|
||||
|
||||
return False
|
||||
|
|
|
@ -1006,7 +1006,8 @@ def reconfigure_vm(vsphere_client, vm, module, esxi, resource_pool, cluster_name
|
|||
dev = dev_list[disk_num]
|
||||
if disksize < int(dev.capacityInKB):
|
||||
vsphere_client.disconnect()
|
||||
module.fail_json(msg="Error in '%s' definition. New size needs to be higher than the current value (%s GB)." % (disk, int(dev.capacityInKB) / 1024 / 1024))
|
||||
module.fail_json(msg="Error in '%s' definition. New size needs to be higher than the current value (%s GB)." %
|
||||
(disk, int(dev.capacityInKB) / 1024 / 1024))
|
||||
|
||||
# Set the new disk size
|
||||
elif disksize > int(dev.capacityInKB):
|
||||
|
|
|
@ -216,8 +216,8 @@ def present(schema_facts, cursor, schema, usage_roles, create_roles, owner):
|
|||
"Changing schema owner is not supported. "
|
||||
"Current owner: {0}."
|
||||
).format(schema_facts[schema_key]['owner']))
|
||||
if cmp(sorted(usage_roles), sorted(schema_facts[schema_key]['usage_roles'])) != 0 or \
|
||||
cmp(sorted(create_roles), sorted(schema_facts[schema_key]['create_roles'])) != 0:
|
||||
if (cmp(sorted(usage_roles), sorted(schema_facts[schema_key]['usage_roles'])) != 0 or
|
||||
cmp(sorted(create_roles), sorted(schema_facts[schema_key]['create_roles'])) != 0):
|
||||
update_roles(schema_facts, cursor, schema,
|
||||
schema_facts[schema_key]['usage_roles'], usage_roles,
|
||||
schema_facts[schema_key]['create_roles'], create_roles)
|
||||
|
|
|
@ -205,11 +205,11 @@ def check(user_facts, user, profile, resource_pool,
|
|||
return False
|
||||
if password and password != user_facts[user_key]['password']:
|
||||
return False
|
||||
if expired is not None and expired != (user_facts[user_key]['expired'] == 'True') or \
|
||||
ldap is not None and ldap != (user_facts[user_key]['expired'] == 'True'):
|
||||
if (expired is not None and expired != (user_facts[user_key]['expired'] == 'True') or
|
||||
ldap is not None and ldap != (user_facts[user_key]['expired'] == 'True')):
|
||||
return False
|
||||
if roles and (cmp(sorted(roles), sorted(user_facts[user_key]['roles'])) != 0 or \
|
||||
cmp(sorted(roles), sorted(user_facts[user_key]['default_roles'])) != 0):
|
||||
if (roles and (cmp(sorted(roles), sorted(user_facts[user_key]['roles'])) != 0 or
|
||||
cmp(sorted(roles), sorted(user_facts[user_key]['default_roles'])) != 0)):
|
||||
return False
|
||||
return True
|
||||
|
||||
|
@ -275,8 +275,8 @@ def present(user_facts, cursor, user, profile, resource_pool,
|
|||
changed = True
|
||||
if changed:
|
||||
cursor.execute(' '.join(query_fragments))
|
||||
if roles and (cmp(sorted(roles), sorted(user_facts[user_key]['roles'])) != 0 or \
|
||||
cmp(sorted(roles), sorted(user_facts[user_key]['default_roles'])) != 0):
|
||||
if (roles and (cmp(sorted(roles), sorted(user_facts[user_key]['roles'])) != 0 or
|
||||
cmp(sorted(roles), sorted(user_facts[user_key]['default_roles'])) != 0)):
|
||||
update_roles(user_facts, cursor, user,
|
||||
user_facts[user_key]['roles'], user_facts[user_key]['default_roles'], roles)
|
||||
changed = True
|
||||
|
|
|
@ -317,8 +317,7 @@ def main():
|
|||
if commands:
|
||||
if (existing.get('route_distinguisher') and
|
||||
proposed.get('route_distinguisher')):
|
||||
if (existing['route_distinguisher'] != proposed[
|
||||
'route_distinguisher'] and
|
||||
if (existing['route_distinguisher'] != proposed['route_distinguisher'] and
|
||||
proposed['route_distinguisher'] != 'default'):
|
||||
WARNINGS.append('EVPN RD {0} was automatically removed. '
|
||||
'It is highly recommended to use a task '
|
||||
|
|
|
@ -236,9 +236,7 @@ def get_commands(module, state, mode):
|
|||
elif module.params['system_mode_maintenance_on_reload_reset_reason']:
|
||||
reset_reasons = get_reset_reasons(module)
|
||||
if (state == 'present' and
|
||||
module.params[
|
||||
'system_mode_maintenance_on_reload_reset_reason'].lower() not
|
||||
in reset_reasons.lower()):
|
||||
module.params['system_mode_maintenance_on_reload_reset_reason'].lower() not in reset_reasons.lower()):
|
||||
commands.append('system mode maintenance on-reload '
|
||||
'reset-reason {0}'.format(
|
||||
module.params[
|
||||
|
|
|
@ -407,8 +407,7 @@ def main():
|
|||
if (existing.get('mapped_vni') == '0' and
|
||||
proposed.get('mapped_vni') == 'default'):
|
||||
proposed.pop('mapped_vni')
|
||||
delta = dict(set(
|
||||
proposed.items()).difference(existing.items()))
|
||||
delta = dict(set(proposed.items()).difference(existing.items()))
|
||||
if delta or not existing:
|
||||
commands = get_vlan_config_commands(delta, vlan_id)
|
||||
|
||||
|
|
|
@ -622,8 +622,7 @@ class NetAppESeriesStoragePool(object):
|
|||
|
||||
@property
|
||||
def reserved_drive_count_differs(self):
|
||||
if int(self.pool_detail['volumeGroupData']['diskPoolData'][
|
||||
'reconstructionReservedDriveCount']) != self.reserve_drive_count:
|
||||
if int(self.pool_detail['volumeGroupData']['diskPoolData']['reconstructionReservedDriveCount']) != self.reserve_drive_count:
|
||||
return True
|
||||
return False
|
||||
|
||||
|
|
|
@ -347,9 +347,9 @@ class Block(Base, Become, Conditional, Taggable):
|
|||
for task in target:
|
||||
if isinstance(task, Block):
|
||||
tmp_list.append(evaluate_block(task))
|
||||
elif task.action == 'meta' \
|
||||
or (task.action == 'include' and task.evaluate_tags([], play_context.skip_tags, all_vars=all_vars)) \
|
||||
or task.evaluate_tags(play_context.only_tags, play_context.skip_tags, all_vars=all_vars):
|
||||
elif (task.action == 'meta' or
|
||||
(task.action == 'include' and task.evaluate_tags([], play_context.skip_tags, all_vars=all_vars)) or
|
||||
task.evaluate_tags(play_context.only_tags, play_context.skip_tags, all_vars=all_vars)):
|
||||
tmp_list.append(task)
|
||||
return tmp_list
|
||||
|
||||
|
|
|
@ -178,9 +178,9 @@ def _private_query(v, value):
|
|||
|
||||
def _public_query(v, value):
|
||||
v_ip = netaddr.IPAddress(str(v.ip))
|
||||
if v_ip.is_unicast() and not v_ip.is_private() and \
|
||||
not v_ip.is_loopback() and not v_ip.is_netmask() and \
|
||||
not v_ip.is_hostmask():
|
||||
if (v_ip.is_unicast() and not v_ip.is_private() and
|
||||
not v_ip.is_loopback() and not v_ip.is_netmask() and
|
||||
not v_ip.is_hostmask()):
|
||||
return value
|
||||
|
||||
def _revdns_query(v):
|
||||
|
|
|
@ -1,7 +1,6 @@
|
|||
lib/ansible/cli/__init__.py
|
||||
lib/ansible/cli/galaxy.py
|
||||
lib/ansible/cli/playbook.py
|
||||
lib/ansible/compat/six/__init__.py
|
||||
lib/ansible/constants.py
|
||||
lib/ansible/errors/__init__.py
|
||||
lib/ansible/executor/play_iterator.py
|
||||
|
@ -59,7 +58,6 @@ lib/ansible/modules/cloud/amazon/s3.py
|
|||
lib/ansible/modules/cloud/amazon/s3_lifecycle.py
|
||||
lib/ansible/modules/cloud/amazon/s3_sync.py
|
||||
lib/ansible/modules/cloud/amazon/s3_website.py
|
||||
lib/ansible/modules/cloud/amazon/sns_topic.py
|
||||
lib/ansible/modules/cloud/amazon/sts_assume_role.py
|
||||
lib/ansible/modules/cloud/amazon/sts_session_token.py
|
||||
lib/ansible/modules/cloud/azure/azure.py
|
||||
|
@ -73,18 +71,14 @@ lib/ansible/modules/cloud/cloudscale/cloudscale_server.py
|
|||
lib/ansible/modules/cloud/cloudstack/cs_host.py
|
||||
lib/ansible/modules/cloud/cloudstack/cs_instance.py
|
||||
lib/ansible/modules/cloud/cloudstack/cs_iso.py
|
||||
lib/ansible/modules/cloud/cloudstack/cs_portforward.py
|
||||
lib/ansible/modules/cloud/digital_ocean/digital_ocean.py
|
||||
lib/ansible/modules/cloud/google/gc_storage.py
|
||||
lib/ansible/modules/cloud/google/gce_tag.py
|
||||
lib/ansible/modules/cloud/google/gcpubsub.py
|
||||
lib/ansible/modules/cloud/misc/ovirt.py
|
||||
lib/ansible/modules/cloud/misc/proxmox.py
|
||||
lib/ansible/modules/cloud/misc/proxmox_kvm.py
|
||||
lib/ansible/modules/cloud/misc/proxmox_template.py
|
||||
lib/ansible/modules/cloud/misc/serverless.py
|
||||
lib/ansible/modules/cloud/openstack/_nova_compute.py
|
||||
lib/ansible/modules/cloud/openstack/os_nova_host_aggregate.py
|
||||
lib/ansible/modules/cloud/packet/packet_device.py
|
||||
lib/ansible/modules/cloud/packet/packet_sshkey.py
|
||||
lib/ansible/modules/cloud/profitbricks/profitbricks.py
|
||||
|
@ -111,8 +105,6 @@ lib/ansible/modules/database/postgresql/postgresql_ext.py
|
|||
lib/ansible/modules/database/postgresql/postgresql_privs.py
|
||||
lib/ansible/modules/database/postgresql/postgresql_schema.py
|
||||
lib/ansible/modules/database/postgresql/postgresql_user.py
|
||||
lib/ansible/modules/database/vertica/vertica_schema.py
|
||||
lib/ansible/modules/database/vertica/vertica_user.py
|
||||
lib/ansible/modules/files/acl.py
|
||||
lib/ansible/modules/files/archive.py
|
||||
lib/ansible/modules/files/copy.py
|
||||
|
@ -139,16 +131,6 @@ lib/ansible/modules/network/f5/bigip_pool_member.py
|
|||
lib/ansible/modules/network/f5/bigip_virtual_server.py
|
||||
lib/ansible/modules/network/haproxy.py
|
||||
lib/ansible/modules/network/nmcli.py
|
||||
lib/ansible/modules/network/nxos/nxos_bgp.py
|
||||
lib/ansible/modules/network/nxos/nxos_bgp_af.py
|
||||
lib/ansible/modules/network/nxos/nxos_bgp_neighbor.py
|
||||
lib/ansible/modules/network/nxos/nxos_bgp_neighbor_af.py
|
||||
lib/ansible/modules/network/nxos/nxos_evpn_vni.py
|
||||
lib/ansible/modules/network/nxos/nxos_feature.py
|
||||
lib/ansible/modules/network/nxos/nxos_gir.py
|
||||
lib/ansible/modules/network/nxos/nxos_interface_ospf.py
|
||||
lib/ansible/modules/network/nxos/nxos_snapshot.py
|
||||
lib/ansible/modules/network/nxos/nxos_vlan.py
|
||||
lib/ansible/modules/network/panos/panos_nat_policy.py
|
||||
lib/ansible/modules/network/snmp_facts.py
|
||||
lib/ansible/modules/notification/hall.py
|
||||
|
@ -210,7 +192,6 @@ lib/ansible/modules/windows/win_uri.py
|
|||
lib/ansible/modules/windows/win_webpicmd.py
|
||||
lib/ansible/parsing/mod_args.py
|
||||
lib/ansible/playbook/attribute.py
|
||||
lib/ansible/playbook/block.py
|
||||
lib/ansible/playbook/role/__init__.py
|
||||
lib/ansible/playbook/role/metadata.py
|
||||
lib/ansible/plugins/action/set_fact.py
|
||||
|
@ -223,7 +204,6 @@ lib/ansible/plugins/connection/accelerate.py
|
|||
lib/ansible/plugins/connection/paramiko_ssh.py
|
||||
lib/ansible/plugins/connection/ssh.py
|
||||
lib/ansible/plugins/connection/winrm.py
|
||||
lib/ansible/plugins/filter/ipaddr.py
|
||||
lib/ansible/plugins/lookup/first_found.py
|
||||
lib/ansible/plugins/shell/fish.py
|
||||
lib/ansible/plugins/shell/sh.py
|
||||
|
|
|
@ -1,5 +1 @@
|
|||
E111
|
||||
E114
|
||||
E125
|
||||
E129
|
||||
E501
|
||||
|
|
Loading…
Reference in a new issue