1
0
Fork 0
mirror of https://github.com/ansible-collections/community.general.git synced 2024-09-14 20:13:21 +02:00

Add a new vmware inventory script backed by pyvmomi (#15967)

Add a new dynamic vmware inventory script backed by pyvmomi
This commit is contained in:
jctanner 2016-05-30 20:41:48 -04:00
parent d81b9ca29e
commit 018d3c3118
5 changed files with 761 additions and 0 deletions

View file

@ -0,0 +1,68 @@
# Ansible VMware external inventory script settings
[vmware]
# The resolvable hostname or ip address of the vsphere
server=vcenter
# The port for the vsphere API
#password=443
# The username with access to the vsphere API
username=administrator@vsphere.local
# The password for the vsphere API
password=vmware
# Specify the number of seconds to use the inventory cache before it is
# considered stale. If not defined, defaults to 0 seconds.
#cache_max_age = 3600
# Specify the directory used for storing the inventory cache. If not defined,
# caching will be disabled.
#cache_dir = ~/.cache/ansible
# Max object level refers to the level of recursion the script will delve into
# the objects returned from pyvomi to find serializable facts. The default
# level of 0 is sufficient for most tasks and will be the most performant.
# Beware that the recursion can exceed python's limit (causing traceback),
# cause sluggish script performance and return huge blobs of facts.
# If you do not know what you are doing, leave this set to 1.
#max_object_level=1
# Lower the keynames for facts to make addressing them easier.
#lower_var_keys=True
# Host alias for objects in the inventory. VMWare allows duplicate VM names
# so they can not be considered unique. Use this setting to alter the alias
# returned for the hosts. Any atributes for the guest can be used to build
# this alias. The default combines the config name and the config uuid and
# expects that the ansible_host will be set by the host_pattern.
#alias_pattern={{ config.name + '_' + config.uuid }}
# Host pattern is the value set for ansible_host and ansible_ssh_host, which
# needs to be a hostname or ipaddress the ansible controlhost can reach.
#host_pattern={{ guest.ipaddress }}
# Host filters are a comma separated list of jinja patterns to remove
# non-matching hosts from the final result.
# EXAMPLES:
# host_filters={{ config.guestid == 'rhel7_64Guest' }}
# host_filters={{ config.cpuhotremoveenabled != False }},{{ runtime.maxmemoryusage >= 512 }}
# host_filters={{ config.cpuhotremoveenabled != False }},{{ runtime.maxmemoryusage >= 512 }}
# The default is only gueststate of 'running'
#host_filters={{ guest.gueststate == "running" }}
# Groupby patterns enable the user to create groups via any possible jinja
# expression. The resulting value will the groupname and the host will be added
# to that group. Be careful to not make expressions that simply return True/False
# because those values will become the literal group name. The patterns can be
# comma delimited to create as many groups as necessary
#groupby_patterns={{ guest.guestid }},{{ 'templates' if config.template else 'guests'}}

View file

@ -0,0 +1,571 @@
#!/usr/bin/env python
# Requirements
# - pyvmomi >= 6.0.0.2016.4
# TODO:
# * more jq examples
# * optional folder heirarchy
"""
$ jq '._meta.hostvars[].config' data.json | head
{
"alternateguestname": "",
"instanceuuid": "5035a5cd-b8e8-d717-e133-2d383eb0d675",
"memoryhotaddenabled": false,
"guestfullname": "Red Hat Enterprise Linux 7 (64-bit)",
"changeversion": "2016-05-16T18:43:14.977925Z",
"uuid": "4235fc97-5ddb-7a17-193b-9a3ac97dc7b4",
"cpuhotremoveenabled": false,
"vpmcenabled": false,
"firmware": "bios",
"""
from __future__ import print_function
import argparse
import atexit
import datetime
import getpass
import jinja2
import os
import six
import ssl
import sys
import uuid
from collections import defaultdict
from six.moves import configparser
from time import time
HAS_PYVMOMI = False
try:
from pyVmomi import vim
from pyVim.connect import SmartConnect, Disconnect
HAS_PYVMOMI = True
except ImportError:
pass
try:
import json
except ImportError:
import simplejson as json
hasvcr = False
try:
import vcr
hasvcr = True
except ImportError:
pass
class VMWareInventory(object):
__name__ = 'VMWareInventory'
instances = []
debug = False
load_dumpfile = None
write_dumpfile = None
maxlevel = 1
lowerkeys = True
config = None
cache_max_age = None
cache_path_cache = None
cache_path_index = None
server = None
port = None
username = None
password = None
host_filters = []
groupby_patterns = []
bad_types = ['Array']
if (sys.version_info > (3, 0)):
safe_types = [int, bool, str, float, None]
else:
safe_types = [int, long, bool, str, float, None]
iter_types = [dict, list]
skip_keys = ['dynamicproperty', 'dynamictype', 'managedby', 'childtype']
def _empty_inventory(self):
return {"_meta" : {"hostvars" : {}}}
def __init__(self, load=True):
self.inventory = self._empty_inventory()
if load:
# Read settings and parse CLI arguments
self.parse_cli_args()
self.read_settings()
# Check the cache
cache_valid = self.is_cache_valid()
# Handle Cache
if self.args.refresh_cache or not cache_valid:
self.do_api_calls_update_cache()
else:
self.inventory = self.get_inventory_from_cache()
def debugl(self, text):
if self.args.debug:
print(text)
def show(self):
# Data to print
data_to_print = None
if self.args.host:
data_to_print = self.get_host_info(self.args.host)
elif self.args.list:
# Display list of instances for inventory
data_to_print = self.inventory
return json.dumps(data_to_print, indent=2)
def is_cache_valid(self):
''' Determines if the cache files have expired, or if it is still valid '''
valid = False
if os.path.isfile(self.cache_path_cache):
mod_time = os.path.getmtime(self.cache_path_cache)
current_time = time()
if (mod_time + self.cache_max_age) > current_time:
valid = True
return valid
def do_api_calls_update_cache(self):
''' Get instances and cache the data '''
instances = self.get_instances()
self.instances = instances
self.inventory = self.instances_to_inventory(instances)
self.write_to_cache(self.inventory, self.cache_path_cache)
def write_to_cache(self, data, cache_path):
''' Dump inventory to json file '''
with open(self.cache_path_cache, 'wb') as f:
f.write(json.dumps(data))
def get_inventory_from_cache(self):
''' Read in jsonified inventory '''
jdata = None
with open(self.cache_path_cache, 'rb') as f:
jdata = f.read()
return json.loads(jdata)
def read_settings(self):
''' Reads the settings from the vmware_inventory.ini file '''
scriptbasename = os.path.realpath(__file__)
scriptbasename = os.path.basename(scriptbasename)
scriptbasename = scriptbasename.replace('.py', '')
defaults = {'vmware': {
'server': '',
'port': 443,
'username': '',
'password': '',
'ini_path': os.path.join(os.path.dirname(os.path.realpath(__file__)), '%s.ini' % scriptbasename),
'cache_name': 'ansible-vmware',
'cache_path': '~/.ansible/tmp',
'cache_max_age': 3600,
'max_object_level': 1,
'alias_pattern': '{{ config.name + "_" + config.uuid }}',
'host_pattern': '{{ guest.ipaddress }}',
'host_filters': '{{ guest.gueststate == "running" }}',
'groupby_patterns': '{{ guest.guestid }},{{ "templates" if config.template else "guests"}}',
'lower_var_keys': True }
}
if six.PY3:
config = configparser.ConfigParser()
else:
config = configparser.SafeConfigParser()
# where is the config?
vmware_ini_path = os.environ.get('VMWARE_INI_PATH', defaults['vmware']['ini_path'])
vmware_ini_path = os.path.expanduser(os.path.expandvars(vmware_ini_path))
config.read(vmware_ini_path)
# apply defaults
for k,v in defaults['vmware'].iteritems():
if not config.has_option('vmware', k):
config.set('vmware', k, str(v))
# where is the cache?
self.cache_dir = os.path.expanduser(config.get('vmware', 'cache_path'))
if self.cache_dir and not os.path.exists(self.cache_dir):
os.makedirs(self.cache_dir)
# set the cache filename and max age
cache_name = config.get('vmware', 'cache_name')
self.cache_path_cache = self.cache_dir + "/%s.cache" % cache_name
self.cache_max_age = int(config.getint('vmware', 'cache_max_age'))
# mark the connection info
self.server = os.environ.get('VMWARE_SERVER', config.get('vmware', 'server'))
self.port = int(os.environ.get('VMWARE_PORT', config.get('vmware', 'port')))
self.username = os.environ.get('VMWARE_USERNAME', config.get('vmware', 'username'))
self.password = os.environ.get('VMWARE_PASSWORD', config.get('vmware', 'password'))
# behavior control
self.maxlevel = int(config.get('vmware', 'max_object_level'))
self.lowerkeys = config.get('vmware', 'lower_var_keys')
if type(self.lowerkeys) != bool:
if str(self.lowerkeys).lower() in ['yes', 'true', '1']:
self.lowerkeys = True
else:
self.lowerkeys = False
self.host_filters = list(config.get('vmware', 'host_filters').split(','))
self.groupby_patterns = list(config.get('vmware', 'groupby_patterns').split(','))
# save the config
self.config = config
def parse_cli_args(self):
''' Command line argument processing '''
parser = argparse.ArgumentParser(description='Produce an Ansible Inventory file based on PyVmomi')
parser.add_argument('--debug', action='store_true', default=False,
help='show debug info')
parser.add_argument('--list', action='store_true', default=True,
help='List instances (default: True)')
parser.add_argument('--host', action='store',
help='Get all the variables about a specific instance')
parser.add_argument('--refresh-cache', action='store_true', default=False,
help='Force refresh of cache by making API requests to VSphere (default: False - use cache files)')
parser.add_argument('--max-instances', default=None, type=int,
help='maximum number of instances to retrieve')
self.args = parser.parse_args()
def get_instances(self):
''' Get a list of vm instances with pyvmomi '''
instances = []
kwargs = {'host': self.server,
'user': self.username,
'pwd': self.password,
'port': int(self.port) }
if hasattr(ssl, 'SSLContext'):
# older ssl libs do not have an SSLContext method:
# context = ssl.SSLContext(ssl.PROTOCOL_TLSv1)
# AttributeError: 'module' object has no attribute 'SSLContext'
# older pyvmomi version also do not have an sslcontext kwarg:
# https://github.com/vmware/pyvmomi/commit/92c1de5056be7c5390ac2a28eb08ad939a4b7cdd
context = ssl.SSLContext(ssl.PROTOCOL_TLSv1)
context.verify_mode = ssl.CERT_NONE
kwargs['sslContext'] = context
self.debugl("### RUNNING WITHOUT VCR")
instances = self._get_instances(kwargs)
self.debugl("### INSTANCES RETRIEVED")
return instances
def _get_instances(self, inkwargs):
''' Make API calls without VCR fixtures '''
instances = []
si = SmartConnect(**inkwargs)
if not si:
print("Could not connect to the specified host using specified "
"username and password")
return -1
atexit.register(Disconnect, si)
content = si.RetrieveContent()
for child in content.rootFolder.childEntity:
instances += self._get_instances_from_children(child)
if self.args.max_instances:
if len(instances) >= (self.args.max_instances+1):
instances = instances[0:(self.args.max_instances+1)]
instance_tuples = []
for instance in sorted(instances):
ifacts = self.facts_from_vobj(instance)
instance_tuples.append((instance, ifacts))
return instance_tuples
def _get_instances_from_children(self, child):
instances = []
if hasattr(child, 'childEntity'):
self.debugl("CHILDREN: %s" % str(child.childEntity))
instances += self._get_instances_from_children(child.childEntity)
elif hasattr(child, 'vmFolder'):
self.debugl("FOLDER: %s" % str(child))
instances += self._get_instances_from_children(child.vmFolder)
elif hasattr(child, 'index'):
self.debugl("LIST: %s" % str(child))
for x in sorted(child):
self.debugl("LIST_ITEM: %s" % x)
instances += self._get_instances_from_children(x)
elif hasattr(child, 'guest'):
self.debugl("GUEST: %s" % str(child))
instances.append(child)
elif hasattr(child, 'vm'):
# resource pools
self.debugl("RESOURCEPOOL: %s" % child.vm)
if child.vm:
instances += self._get_instances_from_children(child.vm)
else:
self.debugl("ELSE ...")
try:
self.debugl(child.__dict__)
except Exception as e:
pass
self.debugl(child)
return instances
def instances_to_inventory(self, instances):
''' Convert a list of vm objects into a json compliant inventory '''
inventory = self._empty_inventory()
inventory['all'] = {}
inventory['all']['hosts'] = []
last_idata = None
total = len(instances)
for idx,instance in enumerate(instances):
# make a unique id for this object to avoid vmware's
# numerous uuid's which aren't all unique.
thisid = str(uuid.uuid4())
idata = instance[1]
# Put it in the inventory
inventory['all']['hosts'].append(thisid)
inventory['_meta']['hostvars'][thisid] = idata.copy()
inventory['_meta']['hostvars'][thisid]['ansible_uuid'] = thisid
# Make a map of the uuid to the name the user wants
name_mapping = self.create_template_mapping(inventory,
self.config.get('vmware', 'alias_pattern'))
# Make a map of the uuid to the ssh hostname the user wants
host_mapping = self.create_template_mapping(inventory,
self.config.get('vmware', 'host_pattern'))
# Reset the inventory keys
for k,v in name_mapping.iteritems():
# set ansible_host (2.x)
inventory['_meta']['hostvars'][k]['ansible_host'] = host_mapping[k]
# 1.9.x backwards compliance
inventory['_meta']['hostvars'][k]['ansible_ssh_host'] = host_mapping[k]
if k == v:
continue
# add new key
inventory['all']['hosts'].append(v)
inventory['_meta']['hostvars'][v] = inventory['_meta']['hostvars'][k]
# cleanup old key
inventory['all']['hosts'].remove(k)
inventory['_meta']['hostvars'].pop(k, None)
self.debugl('PREFILTER_HOSTS:')
for i in inventory['all']['hosts']:
self.debugl(i)
# Apply host filters
for hf in self.host_filters:
if not hf:
continue
filter_map = self.create_template_mapping(inventory, hf, dtype='boolean')
for k,v in filter_map.iteritems():
if not v:
# delete this host
inventory['all']['hosts'].remove(k)
inventory['_meta']['hostvars'].pop(k, None)
self.debugl('POSTFILTER_HOSTS:')
for i in inventory['all']['hosts']:
self.debugl(i)
# Create groups
for gbp in self.groupby_patterns:
groupby_map = self.create_template_mapping(inventory, gbp)
for k,v in groupby_map.iteritems():
if v not in inventory:
inventory[v] = {}
inventory[v]['hosts'] = []
if k not in inventory[v]['hosts']:
inventory[v]['hosts'].append(k)
return inventory
def create_template_mapping(self, inventory, pattern, dtype='string'):
''' Return a hash of uuid to templated string from pattern '''
mapping = {}
for k,v in inventory['_meta']['hostvars'].iteritems():
t = jinja2.Template(pattern)
newkey = None
try:
newkey = t.render(v)
newkey = newkey.strip()
except Exception as e:
self.debugl(str(e))
#import epdb; epdb.st()
if dtype == 'integer':
newkey = int(newkey)
elif dtype == 'boolean':
if newkey.lower() == 'false':
newkey = False
elif newkey.lower() == 'true':
newkey = True
elif dtype == 'string':
pass
mapping[k] = newkey
return mapping
def facts_from_vobj(self, vobj, level=0):
''' Traverse a VM object and return a json compliant data structure '''
# pyvmomi objects are not yet serializable, but may be one day ...
# https://github.com/vmware/pyvmomi/issues/21
rdata = {}
# Do not serialize self
if hasattr(vobj, '__name__'):
if vobj.__name__ == 'VMWareInventory':
return rdata
# Exit early if maxlevel is reached
if level > self.maxlevel:
return rdata
# Objects usually have a dict property
if hasattr(vobj, '__dict__') and not level == 0:
keys = sorted(vobj.__dict__.keys())
for k in keys:
v = vobj.__dict__[k]
# Skip private methods
if k.startswith('_'):
continue
if k.lower() in self.skip_keys:
continue
if self.lowerkeys:
k = k.lower()
rdata[k] = self._process_object_types(v, level=level)
else:
methods = dir(vobj)
methods = [str(x) for x in methods if not x.startswith('_')]
methods = [x for x in methods if not x in self.bad_types]
methods = sorted(methods)
for method in methods:
if method in rdata:
continue
# Attempt to get the method, skip on fail
try:
methodToCall = getattr(vobj, method)
except Exception as e:
continue
# Skip callable methods
if callable(methodToCall):
continue
if self.lowerkeys:
method = method.lower()
rdata[method] = self._process_object_types(methodToCall, level=level)
return rdata
def _process_object_types(self, vobj, level=0):
rdata = {}
self.debugl("PROCESSING: %s" % str(vobj))
if type(vobj) in self.safe_types:
try:
rdata = vobj
except Exception as e:
self.debugl(str(e))
elif hasattr(vobj, 'append'):
rdata = []
for vi in sorted(vobj):
if type(vi) in self.safe_types:
rdata.append(vi)
else:
if (level+1 <= self.maxlevel):
vid = self.facts_from_vobj(vi, level=(level+1))
if vid:
rdata.append(vid)
elif hasattr(vobj, '__dict__'):
if (level+1 <= self.maxlevel):
md = None
md = self.facts_from_vobj(vobj, level=(level+1))
if md:
rdata = md
elif not vobj or type(vobj) in self.safe_types:
rdata = vobj
elif type(vobj) == datetime.datetime:
rdata = str(vobj)
else:
self.debugl("unknown datatype: %s" % type(vobj))
if not rdata:
rdata = None
return rdata
def get_host_info(self, host):
''' Return hostvars for a single host '''
return self.inventory['_meta']['hostvars'][host]
if __name__ == "__main__":
# Run the script
print(VMWareInventory().show())

View file

View file

View file

@ -0,0 +1,122 @@
#!/usr/bin/env python
import json
import os
import pickle
import unittest
import sys
# contrib's dirstruct doesn't contain __init__.py files
checkout_path = os.path.dirname(__file__)
checkout_path = checkout_path.replace('/test/units/contrib/inventory', '')
inventory_dir = os.path.join(checkout_path, 'contrib', 'inventory')
sys.path.append(os.path.abspath(inventory_dir))
from vmware_inventory import VMWareInventory
# cleanup so that nose's path is not polluted with other inv scripts
sys.path.remove(os.path.abspath(inventory_dir))
BASICINVENTORY = {'all': {'hosts': ['foo', 'bar']},
'_meta': { 'hostvars': { 'foo': {'hostname': 'foo'},
'bar': {'hostname': 'bar'}}
}
}
class FakeArgs(object):
debug = False
write_dumpfile = None
load_dumpfile = None
host = False
list = True
class TestVMWareInventory(unittest.TestCase):
def test_host_info_returns_single_host(self):
vmw = VMWareInventory(load=False)
vmw.inventory = BASICINVENTORY
foo = vmw.get_host_info('foo')
bar = vmw.get_host_info('bar')
assert foo == {'hostname': 'foo'}
assert bar == {'hostname': 'bar'}
def test_show_returns_serializable_data(self):
fakeargs = FakeArgs()
vmw = VMWareInventory(load=False)
vmw.args = fakeargs
vmw.inventory = BASICINVENTORY
showdata = vmw.show()
serializable = False
try:
json.loads(showdata)
serializable = True
except:
pass
assert serializable
#import epdb; epdb.st()
def test_show_list_returns_serializable_data(self):
fakeargs = FakeArgs()
vmw = VMWareInventory(load=False)
vmw.args = fakeargs
vmw.args.list = True
vmw.inventory = BASICINVENTORY
showdata = vmw.show()
serializable = False
try:
json.loads(showdata)
serializable = True
except:
pass
assert serializable
#import epdb; epdb.st()
def test_show_list_returns_all_data(self):
fakeargs = FakeArgs()
vmw = VMWareInventory(load=False)
vmw.args = fakeargs
vmw.args.list = True
vmw.inventory = BASICINVENTORY
showdata = vmw.show()
expected = json.dumps(BASICINVENTORY, indent=2)
assert showdata == expected
def test_show_host_returns_serializable_data(self):
fakeargs = FakeArgs()
vmw = VMWareInventory(load=False)
vmw.args = fakeargs
vmw.args.host = 'foo'
vmw.inventory = BASICINVENTORY
showdata = vmw.show()
serializable = False
try:
json.loads(showdata)
serializable = True
except:
pass
assert serializable
#import epdb; epdb.st()
def test_show_host_returns_just_host(self):
fakeargs = FakeArgs()
vmw = VMWareInventory(load=False)
vmw.args = fakeargs
vmw.args.list = False
vmw.args.host = 'foo'
vmw.inventory = BASICINVENTORY
showdata = vmw.show()
expected = BASICINVENTORY['_meta']['hostvars']['foo']
expected = json.dumps(expected, indent=2)
#import epdb; epdb.st()
assert showdata == expected
if __name__ == '__main__':
unittest.main()