1
0
Fork 0
mirror of https://github.com/ansible-collections/community.general.git synced 2024-09-14 20:13:21 +02:00

Fix merge conflicts from #5519

This commit is contained in:
James Tanner 2014-01-21 13:35:06 -05:00
commit 61d283e2ad
3 changed files with 31 additions and 15 deletions

View file

@ -296,17 +296,17 @@ class PlayBook(object):
# *****************************************************
def _list_available_hosts(self, *args):
def _trim_unavailable_hosts(self, hostlist=[]):
''' returns a list of hosts that haven't failed and aren't dark '''
return [ h for h in self.inventory.list_hosts(*args) if (h not in self.stats.failures) and (h not in self.stats.dark)]
return [ h for h in hostlist if (h not in self.stats.failures) and (h not in self.stats.dark)]
# *****************************************************
def _run_task_internal(self, task):
''' run a particular module step in a playbook '''
hosts = self._list_available_hosts()
hosts = self._trim_unavailable_hosts(task.play._play_hosts)
self.inventory.restrict_to(hosts)
runner = ansible.runner.Runner(
@ -342,7 +342,8 @@ class PlayBook(object):
error_on_undefined_vars=C.DEFAULT_UNDEFINED_VAR_BEHAVIOR,
su=task.su,
su_user=task.su_user,
su_pass=task.su_pass
su_pass=task.su_pass,
run_hosts=hosts
)
if task.async_seconds == 0:
@ -462,7 +463,7 @@ class PlayBook(object):
if play.gather_facts is False:
return {}
host_list = self._list_available_hosts(play.hosts)
host_list = self._trim_unavailable_hosts(play._play_hosts)
self.callbacks.on_setup()
self.inventory.restrict_to(host_list)
@ -542,8 +543,10 @@ class PlayBook(object):
''' run a list of tasks for a given pattern, in order '''
self.callbacks.on_play_start(play.name)
# Get the hosts for this play
play._play_hosts = self.inventory.list_hosts(play.hosts)
# if no hosts matches this play, drop out
if not self.inventory.list_hosts(play.hosts):
if not play._play_hosts:
self.callbacks.on_no_hosts_matched()
return True
@ -552,8 +555,9 @@ class PlayBook(object):
# now with that data, handle contentional variable file imports!
all_hosts = self._list_available_hosts(play.hosts)
all_hosts = self._trim_unavailable_hosts(play._play_hosts)
play.update_vars_files(all_hosts)
hosts_count = len(all_hosts)
serialized_batch = []
if play.serial <= 0:
@ -569,6 +573,9 @@ class PlayBook(object):
for on_hosts in serialized_batch:
# restrict the play to just the hosts we have in our on_hosts block that are
# available.
play._play_hosts = self._trim_unavailable_hosts(on_hosts)
self.inventory.also_restrict_to(on_hosts)
for task in play.tasks():
@ -591,7 +598,7 @@ class PlayBook(object):
# prevent duplicate handler includes from running more than once
fired_names[handler_name] = 1
host_list = self._list_available_hosts(play.hosts)
host_list = self._trim_unavailable_hosts(play._play_hosts)
if handler.any_errors_fatal and len(host_list) < hosts_count:
play.max_fail_pct = 0
if (hosts_count - len(host_list)) > int((play.max_fail_pct)/100.0 * hosts_count):
@ -610,8 +617,6 @@ class PlayBook(object):
continue
hosts_count = len(self._list_available_hosts(play.hosts))
# only run the task if the requested tags match
should_run = False
for x in self.only_tags:
@ -634,7 +639,9 @@ class PlayBook(object):
# just didn't match anything and that's ok
return False
host_list = self._list_available_hosts(play.hosts)
# Get a new list of what hosts are left as available, the ones that
# did not go fail/dark during the task
host_list = self._trim_unavailable_hosts(play._play_hosts)
# Set max_fail_pct to 0, So if any hosts fails, bail out
if task.any_errors_fatal and len(host_list) < hosts_count:

View file

@ -34,7 +34,7 @@ class Play(object):
'handlers', 'remote_user', 'remote_port', 'included_roles', 'accelerate',
'accelerate_port', 'accelerate_ipv6', 'sudo', 'sudo_user', 'transport', 'playbook',
'tags', 'gather_facts', 'serial', '_ds', '_handlers', '_tasks',
'basedir', 'any_errors_fatal', 'roles', 'max_fail_pct', 'su', 'su_user'
'basedir', 'any_errors_fatal', 'roles', 'max_fail_pct', '_play_hosts', 'su', 'su_user'
]
# to catch typos and so forth -- these are userland names
@ -137,6 +137,8 @@ class Play(object):
if self.sudo_user != 'root':
self.sudo = True
# place holder for the discovered hosts to be used in this play
self._play_hosts = None
# *************************************************

View file

@ -143,7 +143,8 @@ class Runner(object):
accelerate_port=None, # port to use with accelerated connection
su=False, # Are we running our command via su?
su_user=None, # User to su to when running command, ex: 'root'
su_pass=C.DEFAULT_SU_PASS
su_pass=C.DEFAULT_SU_PASS,
run_hosts=None, # an optional list of pre-calculated hosts to run on
):
# used to lock multiprocess inputs and outputs at various levels
@ -211,6 +212,10 @@ class Runner(object):
# don't override subset when passed from playbook
self.inventory.subset(subset)
# If we get a pre-built list of hosts to run on, from say a playbook, use them.
# Also where we will store the hosts to run on once discovered
self.run_hosts = run_hosts
if self.transport == 'local':
self.remote_user = pwd.getpwuid(os.geteuid())[0]
@ -1034,7 +1039,7 @@ class Runner(object):
results2["dark"][host] = result.result
# hosts which were contacted but never got a chance to return
for host in self.inventory.list_hosts(self.pattern):
for host in self.run_hosts:
if not (host in results2['dark'] or host in results2['contacted']):
results2["dark"][host] = {}
return results2
@ -1045,7 +1050,9 @@ class Runner(object):
''' xfer & run module on all matched hosts '''
# find hosts that match the pattern
hosts = self.inventory.list_hosts(self.pattern)
if not self.run_hosts:
self.run_hosts = self.inventory.list_hosts(self.pattern)
hosts = self.run_hosts
if len(hosts) == 0:
self.callbacks.on_no_hosts()
return dict(contacted={}, dark={})