diff --git a/legacy/gce_tests.py b/legacy/gce_tests.py deleted file mode 100644 index 3f0a4273b0..0000000000 --- a/legacy/gce_tests.py +++ /dev/null @@ -1,748 +0,0 @@ -#!/usr/bin/env python -# Copyright 2013 Google Inc. -# -# This file is part of Ansible -# -# Ansible is free software: you can redistribute it and/or modify -# it under the terms of the GNU General Public License as published by -# the Free Software Foundation, either version 3 of the License, or -# (at your option) any later version. -# -# Ansible is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -# GNU General Public License for more details. -# -# You should have received a copy of the GNU General Public License -# along with Ansible. If not, see . - -# This is a custom functional test script for the Google Compute Engine -# ansible modules. In order to run these tests, you must: -# 1) Create a Google Cloud Platform account and enable the Google -# Compute Engine service and billing -# 2) Download, install, and configure 'gcutil' -# see [https://developers.google.com/compute/docs/gcutil/] -# 3) Convert your GCE Service Account private key from PKCS12 to PEM format -# $ openssl pkcs12 -in pkey.pkcs12 -passin pass:notasecret \ -# > -nodes -nocerts | openssl rsa -out pkey.pem -# 4) Make sure you have libcloud 0.13.3 or later installed. -# 5) Make sure you have a libcloud 'secrets.py' file in your PYTHONPATH -# 6) Set GCE_PARAMS and GCE_KEYWORD_PARMS in your 'secrets.py' file. -# 7) Set up a simple hosts file -# $ echo 127.0.0.1 > ~/ansible_hosts -# $ echo "export ANSIBLE_HOSTS='~/ansible_hosts'" >> ~/.bashrc -# $ . ~/.bashrc -# 8) Set up your ansible 'hacking' environment -# $ cd ~/ansible -# $ . hacking/env-setup -# $ export ANSIBLE_HOST_KEY_CHECKING=no -# $ ansible all -m ping -# 9) Set your PROJECT variable below -# 10) Run and time the tests and log output, take ~30 minutes to run -# $ time stdbuf -oL python test/gce_tests.py 2>&1 | tee log -# -# Last update: gcutil-1.11.0 and v1beta16 - -# Set this to your test Project ID -PROJECT="google.com:erjohnso" - -# debugging -DEBUG=False # lots of debugging output -VERBOSE=True # on failure, display ansible command and expected/actual result - -# location - note that some tests rely on the module's 'default' -# region/zone, which should match the settings below. -REGION="us-central1" -ZONE="%s-a" % REGION - -# Peeking is a way to trigger looking at a specified set of resources -# before and/or after a test run. The 'test_cases' data structure below -# has a few tests with 'peek_before' and 'peek_after'. When those keys -# are set and PEEKING_ENABLED is True, then these steps will be executed -# to aid in debugging tests. Normally, this is not needed. -PEEKING_ENABLED=False - -# disks -DNAME="aaaaa-ansible-disk" -DNAME2="aaaaa-ansible-disk2" -DNAME6="aaaaa-ansible-inst6" -DNAME7="aaaaa-ansible-inst7" -USE_PD="true" -KERNEL="https://www.googleapis.com/compute/v1beta16/projects/google/global/kernels/gce-no-conn-track-v20130813" - -# instances -INAME="aaaaa-ansible-inst" -INAME2="aaaaa-ansible-inst2" -INAME3="aaaaa-ansible-inst3" -INAME4="aaaaa-ansible-inst4" -INAME5="aaaaa-ansible-inst5" -INAME6="aaaaa-ansible-inst6" -INAME7="aaaaa-ansible-inst7" -TYPE="n1-standard-1" -IMAGE="https://www.googleapis.com/compute/v1beta16/projects/debian-cloud/global/images/debian-7-wheezy-v20131014" -NETWORK="default" -SCOPES="https://www.googleapis.com/auth/userinfo.email,https://www.googleapis.com/auth/compute,https://www.googleapis.com/auth/devstorage.full_control" - -# networks / firewalls -NETWK1="ansible-network1" -NETWK2="ansible-network2" -NETWK3="ansible-network3" -CIDR1="10.240.16.0/24" -CIDR2="10.240.32.0/24" -CIDR3="10.240.64.0/24" -GW1="10.240.16.1" -GW2="10.240.32.1" -FW1="ansible-fwrule1" -FW2="ansible-fwrule2" -FW3="ansible-fwrule3" -FW4="ansible-fwrule4" - -# load-balancer tests -HC1="ansible-hc1" -HC2="ansible-hc2" -HC3="ansible-hc3" -LB1="ansible-lb1" -LB2="ansible-lb2" - -from commands import getstatusoutput as run -import sys - -test_cases = [ - {'id': '01', 'desc': 'Detach / Delete disk tests', - 'setup': ['gcutil addinstance "%s" --wait_until_running --zone=%s --machine_type=%s --network=%s --service_account_scopes="%s" --image="%s" --persistent_boot_disk=%s' % (INAME, ZONE, TYPE, NETWORK, SCOPES, IMAGE, USE_PD), - 'gcutil adddisk "%s" --size_gb=2 --zone=%s --wait_until_complete' % (DNAME, ZONE)], - - 'tests': [ - {'desc': 'DETACH_ONLY but disk not found [success]', - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s detach_only=yes state=absent' % ("missing-disk", INAME, ZONE), - 'r': '127.0.0.1 | success >> {"changed": false, "detach_only": true, "detached_from_instance": "%s", "name": "missing-disk", "state": "absent", "zone": "%s"}' % (INAME, ZONE), - }, - {'desc': 'DETACH_ONLY but instance not found [success]', - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s detach_only=yes state=absent' % (DNAME, "missing-instance", ZONE), - 'r': '127.0.0.1 | success >> {"changed": false, "detach_only": true, "detached_from_instance": "missing-instance", "name": "%s", "size_gb": 2, "state": "absent", "zone": "%s"}' % (DNAME, ZONE), - }, - {'desc': 'DETACH_ONLY but neither disk nor instance exists [success]', - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s detach_only=yes state=absent' % ("missing-disk", "missing-instance", ZONE), - 'r': '127.0.0.1 | success >> {"changed": false, "detach_only": true, "detached_from_instance": "missing-instance", "name": "missing-disk", "state": "absent", "zone": "%s"}' % (ZONE), - }, - {'desc': 'DETACH_ONLY but disk is not currently attached [success]', - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s detach_only=yes state=absent' % (DNAME, INAME, ZONE), - 'r': '127.0.0.1 | success >> {"changed": false, "detach_only": true, "detached_from_instance": "%s", "name": "%s", "size_gb": 2, "state": "absent", "zone": "%s"}' % (INAME, DNAME, ZONE), - }, - {'desc': 'DETACH_ONLY disk is attached and should be detached [success]', - 'setup': ['gcutil attachdisk --disk="%s,mode=READ_ONLY" --zone=%s %s' % (DNAME, ZONE, INAME), 'sleep 10'], - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s detach_only=yes state=absent' % (DNAME, INAME, ZONE), - 'r': '127.0.0.1 | success >> {"attached_mode": "READ_ONLY", "attached_to_instance": "%s", "changed": true, "detach_only": true, "detached_from_instance": "%s", "name": "%s", "size_gb": 2, "state": "absent", "zone": "%s"}' % (INAME, INAME, DNAME, ZONE), - 'teardown': ['gcutil detachdisk --zone=%s --device_name=%s %s' % (ZONE, DNAME, INAME)], - }, - {'desc': 'DETACH_ONLY but not instance specified [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s zone=%s detach_only=yes state=absent' % (DNAME, ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Must specify an instance name when detaching a disk"}', - }, - {'desc': 'DELETE but disk not found [success]', - 'm': 'gce_pd', - 'a': 'name=%s zone=%s state=absent' % ("missing-disk", ZONE), - 'r': '127.0.0.1 | success >> {"changed": false, "name": "missing-disk", "state": "absent", "zone": "%s"}' % (ZONE), - }, - {'desc': 'DELETE but disk is attached [FAIL]', - 'setup': ['gcutil attachdisk --disk="%s,mode=READ_ONLY" --zone=%s %s' % (DNAME, ZONE, INAME), 'sleep 10'], - 'm': 'gce_pd', - 'a': 'name=%s zone=%s state=absent' % (DNAME, ZONE), - 'r': "127.0.0.1 | FAILED >> {\"changed\": false, \"failed\": true, \"msg\": \"The disk resource 'projects/%s/zones/%s/disks/%s' is already being used by 'projects/%s/zones/%s/instances/%s'\"}" % (PROJECT, ZONE, DNAME, PROJECT, ZONE, INAME), - 'teardown': ['gcutil detachdisk --zone=%s --device_name=%s %s' % (ZONE, DNAME, INAME)], - }, - {'desc': 'DELETE disk [success]', - 'm': 'gce_pd', - 'a': 'name=%s zone=%s state=absent' % (DNAME, ZONE), - 'r': '127.0.0.1 | success >> {"changed": true, "name": "%s", "size_gb": 2, "state": "absent", "zone": "%s"}' % (DNAME, ZONE), - }, - ], - 'teardown': ['gcutil deleteinstance -f "%s" --zone=%s' % (INAME, ZONE), - 'sleep 15', - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME, ZONE), - 'sleep 10', - 'gcutil deletedisk -f "%s" --zone=%s' % (DNAME, ZONE), - 'sleep 10'], - }, - - {'id': '02', 'desc': 'Create disk but do not attach (e.g. no instance_name param)', - 'setup': [], - 'tests': [ - {'desc': 'CREATE_NO_ATTACH "string" for size_gb [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s size_gb="foo" zone=%s' % (DNAME, ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Must supply a size_gb larger than 1 GB"}', - }, - {'desc': 'CREATE_NO_ATTACH negative size_gb [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s size_gb=-2 zone=%s' % (DNAME, ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Must supply a size_gb larger than 1 GB"}', - }, - {'desc': 'CREATE_NO_ATTACH size_gb exceeds quota [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s size_gb=9999 zone=%s' % ("big-disk", ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Requested disk size exceeds quota"}', - }, - {'desc': 'CREATE_NO_ATTACH create the disk [success]', - 'm': 'gce_pd', - 'a': 'name=%s zone=%s' % (DNAME, ZONE), - 'r': '127.0.0.1 | success >> {"changed": true, "name": "%s", "size_gb": 10, "state": "present", "zone": "%s"}' % (DNAME, ZONE), - }, - {'desc': 'CREATE_NO_ATTACH but disk already exists [success]', - 'm': 'gce_pd', - 'a': 'name=%s zone=%s' % (DNAME, ZONE), - 'r': '127.0.0.1 | success >> {"changed": false, "name": "%s", "size_gb": 10, "state": "present", "zone": "%s"}' % (DNAME, ZONE), - }, - ], - 'teardown': ['gcutil deletedisk -f "%s" --zone=%s' % (DNAME, ZONE), - 'sleep 10'], - }, - - {'id': '03', 'desc': 'Create and attach disk', - 'setup': ['gcutil addinstance "%s" --zone=%s --machine_type=%s --network=%s --service_account_scopes="%s" --image="%s" --persistent_boot_disk=%s' % (INAME2, ZONE, TYPE, NETWORK, SCOPES, IMAGE, USE_PD), - 'gcutil addinstance "%s" --zone=%s --machine_type=%s --network=%s --service_account_scopes="%s" --image="%s" --persistent_boot_disk=%s' % (INAME, ZONE, "g1-small", NETWORK, SCOPES, IMAGE, USE_PD), - 'gcutil adddisk "%s" --size_gb=2 --zone=%s' % (DNAME, ZONE), - 'gcutil adddisk "%s" --size_gb=2 --zone=%s --wait_until_complete' % (DNAME2, ZONE),], - 'tests': [ - {'desc': 'CREATE_AND_ATTACH "string" for size_gb [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s size_gb="foo" instance_name=%s zone=%s' % (DNAME, INAME, ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Must supply a size_gb larger than 1 GB"}', - }, - {'desc': 'CREATE_AND_ATTACH negative size_gb [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s size_gb=-2 instance_name=%s zone=%s' % (DNAME, INAME, ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Must supply a size_gb larger than 1 GB"}', - }, - {'desc': 'CREATE_AND_ATTACH size_gb exceeds quota [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s size_gb=9999 instance_name=%s zone=%s' % ("big-disk", INAME, ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Requested disk size exceeds quota"}', - }, - {'desc': 'CREATE_AND_ATTACH missing instance [FAIL]', - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s' % (DNAME, "missing-instance", ZONE), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Instance %s does not exist in zone %s"}' % ("missing-instance", ZONE), - }, - {'desc': 'CREATE_AND_ATTACH disk exists but not attached [success]', - 'peek_before': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s' % (DNAME, INAME, ZONE), - 'r': '127.0.0.1 | success >> {"attached_mode": "READ_ONLY", "attached_to_instance": "%s", "changed": true, "name": "%s", "size_gb": 2, "state": "present", "zone": "%s"}' % (INAME, DNAME, ZONE), - 'peek_after': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - }, - {'desc': 'CREATE_AND_ATTACH disk exists already attached [success]', - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s' % (DNAME, INAME, ZONE), - 'r': '127.0.0.1 | success >> {"attached_mode": "READ_ONLY", "attached_to_instance": "%s", "changed": false, "name": "%s", "size_gb": 2, "state": "present", "zone": "%s"}' % (INAME, DNAME, ZONE), - }, - {'desc': 'CREATE_AND_ATTACH attached RO, attempt RO to 2nd inst [success]', - 'peek_before': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s' % (DNAME, INAME2, ZONE), - 'r': '127.0.0.1 | success >> {"attached_mode": "READ_ONLY", "attached_to_instance": "%s", "changed": true, "name": "%s", "size_gb": 2, "state": "present", "zone": "%s"}' % (INAME2, DNAME, ZONE), - 'peek_after': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - }, - {'desc': 'CREATE_AND_ATTACH attached RO, attach RW to self [FAILED no-op]', - 'peek_before': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s mode=READ_WRITE' % (DNAME, INAME, ZONE), - 'r': '127.0.0.1 | success >> {"attached_mode": "READ_ONLY", "attached_to_instance": "%s", "changed": false, "name": "%s", "size_gb": 2, "state": "present", "zone": "%s"}' % (INAME, DNAME, ZONE), - }, - {'desc': 'CREATE_AND_ATTACH attached RW, attach RW to other [FAIL]', - 'setup': ['gcutil attachdisk --disk=%s,mode=READ_WRITE --zone=%s %s' % (DNAME2, ZONE, INAME), 'sleep 10'], - 'peek_before': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s mode=READ_WRITE' % (DNAME2, INAME2, ZONE), - 'r': "127.0.0.1 | FAILED >> {\"changed\": false, \"failed\": true, \"msg\": \"Unexpected response: HTTP return_code[200], API error code[RESOURCE_IN_USE] and message: The disk resource 'projects/%s/zones/%s/disks/%s' is already being used in read-write mode\"}" % (PROJECT, ZONE, DNAME2), - 'peek_after': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - }, - {'desc': 'CREATE_AND_ATTACH attach too many disks to inst [FAIL]', - 'setup': ['gcutil adddisk aa-disk-dummy --size_gb=2 --zone=%s' % (ZONE), - 'gcutil adddisk aa-disk-dummy2 --size_gb=2 --zone=%s --wait_until_complete' % (ZONE), - 'gcutil attachdisk --disk=aa-disk-dummy --zone=%s %s' % (ZONE, INAME), - 'sleep 5'], - 'peek_before': ["gcutil --format=csv listinstances --zone=%s --filter=\"name eq 'aaaa.*'\"" % (ZONE)], - 'm': 'gce_pd', - 'a': 'name=%s instance_name=%s zone=%s' % ("aa-disk-dummy2", INAME, ZONE), - 'r': "127.0.0.1 | FAILED >> {\"changed\": false, \"failed\": true, \"msg\": \"Unexpected response: HTTP return_code[200], API error code[LIMIT_EXCEEDED] and message: Exceeded limit 'maximum_persistent_disks' on resource 'projects/%s/zones/%s/instances/%s'. Limit: 4\"}" % (PROJECT, ZONE, INAME), - 'teardown': ['gcutil detachdisk --device_name=aa-disk-dummy --zone=%s %s' % (ZONE, INAME), - 'sleep 3', - 'gcutil deletedisk -f aa-disk-dummy --zone=%s' % (ZONE), - 'sleep 10', - 'gcutil deletedisk -f aa-disk-dummy2 --zone=%s' % (ZONE), - 'sleep 10'], - }, - ], - 'teardown': ['gcutil deleteinstance -f "%s" --zone=%s' % (INAME2, ZONE), - 'sleep 15', - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME, ZONE), - 'sleep 15', - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME, ZONE), - 'sleep 10', - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME2, ZONE), - 'sleep 10', - 'gcutil deletedisk -f "%s" --zone=%s' % (DNAME, ZONE), - 'sleep 10', - 'gcutil deletedisk -f "%s" --zone=%s' % (DNAME2, ZONE), - 'sleep 10'], - }, - - {'id': '04', 'desc': 'Delete / destroy instances', - 'setup': ['gcutil addinstance "%s" --zone=%s --machine_type=%s --image="%s" --persistent_boot_disk=false' % (INAME, ZONE, TYPE, IMAGE), - 'gcutil addinstance "%s" --zone=%s --machine_type=%s --image="%s" --persistent_boot_disk=false' % (INAME2, ZONE, TYPE, IMAGE), - 'gcutil addinstance "%s" --zone=%s --machine_type=%s --image="%s" --persistent_boot_disk=false' % (INAME3, ZONE, TYPE, IMAGE), - 'gcutil addinstance "%s" --zone=%s --machine_type=%s --image="%s" --persistent_boot_disk=false' % (INAME4, ZONE, TYPE, IMAGE), - 'gcutil addinstance "%s" --wait_until_running --zone=%s --machine_type=%s --image="%s" --persistent_boot_disk=false' % (INAME5, ZONE, TYPE, IMAGE)], - 'tests': [ - {'desc': 'DELETE instance, bad zone param [FAIL]', - 'm': 'gce', - 'a': 'name=missing-inst zone=bogus state=absent', - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "value of zone must be one of: us-central1-a,us-central1-b,us-central2-a,europe-west1-a,europe-west1-b, got: bogus"}', - }, - {'desc': 'DELETE non-existent instance, no-op [success]', - 'm': 'gce', - 'a': 'name=missing-inst zone=%s state=absent' % (ZONE), - 'r': '127.0.0.1 | success >> {"changed": false, "name": "missing-inst", "state": "absent", "zone": "%s"}' % (ZONE), - }, - {'desc': 'DELETE an existing named instance [success]', - 'm': 'gce', - 'a': 'name=%s zone=%s state=absent' % (INAME, ZONE), - 'r': '127.0.0.1 | success >> {"changed": true, "name": "%s", "state": "absent", "zone": "%s"}' % (INAME, ZONE), - }, - {'desc': 'DELETE list of instances with a non-existent one [success]', - 'm': 'gce', - 'a': 'instance_names=%s,missing,%s zone=%s state=absent' % (INAME2,INAME3, ZONE), - 'r': '127.0.0.1 | success >> {"changed": true, "instance_names": ["%s", "%s"], "state": "absent", "zone": "%s"}' % (INAME2, INAME3, ZONE), - }, - {'desc': 'DELETE list of instances all pre-exist [success]', - 'm': 'gce', - 'a': 'instance_names=%s,%s zone=%s state=absent' % (INAME4,INAME5, ZONE), - 'r': '127.0.0.1 | success >> {"changed": true, "instance_names": ["%s", "%s"], "state": "absent", "zone": "%s"}' % (INAME4, INAME5, ZONE), - }, - ], - 'teardown': ['gcutil deleteinstance -f "%s" --zone=%s' % (INAME, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME2, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME3, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME4, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME5, ZONE), - 'sleep 10'], - }, - - {'id': '05', 'desc': 'Create instances', - 'setup': ['gcutil adddisk --source_image=%s --zone=%s %s --wait_until_complete' % (IMAGE, ZONE, DNAME7), - 'gcutil addinstance boo --wait_until_running --zone=%s --machine_type=%s --network=%s --disk=%s,mode=READ_WRITE,boot --kernel=%s' % (ZONE,TYPE,NETWORK,DNAME7,KERNEL), - ], - 'tests': [ - {'desc': 'CREATE_INSTANCE invalid image arg [FAIL]', - 'm': 'gce', - 'a': 'name=foo image=foo', - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Missing required create instance variable"}', - }, - {'desc': 'CREATE_INSTANCE metadata a list [FAIL]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s zone=%s metadata=\'[\\"foo\\":\\"bar\\",\\"baz\\":1]\'' % (INAME,ZONE), - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "bad metadata syntax"}', - }, - {'desc': 'CREATE_INSTANCE metadata not a dict [FAIL]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s zone=%s metadata=\\"foo\\":\\"bar\\",\\"baz\\":1' % (INAME,ZONE), - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "bad metadata syntax"}', - }, - {'desc': 'CREATE_INSTANCE with metadata form1 [FAIL]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s zone=%s metadata=\'{"foo":"bar","baz":1}\'' % (INAME,ZONE), - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "bad metadata: malformed string"}', - }, - {'desc': 'CREATE_INSTANCE with metadata form2 [FAIL]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s zone=%s metadata={\'foo\':\'bar\',\'baz\':1}' % (INAME,ZONE), - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "bad metadata: malformed string"}', - }, - {'desc': 'CREATE_INSTANCE with metadata form3 [FAIL]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s zone=%s metadata="foo:bar" '% (INAME,ZONE), - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "bad metadata syntax"}', - }, - {'desc': 'CREATE_INSTANCE with metadata form4 [FAIL]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s zone=%s metadata="{\'foo\':\'bar\'}"'% (INAME,ZONE), - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "bad metadata: malformed string"}', - }, - {'desc': 'CREATE_INSTANCE invalid image arg [FAIL]', - 'm': 'gce', - 'a': 'instance_names=foo,bar image=foo', - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Missing required create instance variable"}', - }, - {'desc': 'CREATE_INSTANCE single inst, using defaults [success]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s' % (INAME), - 'r': '127.0.0.1 | success >> {"changed": true, "instance_data": [{"image": "debian-7-wheezy-v20130816", "machine_type": "n1-standard-1", "metadata": {}, "name": "%s", "network": "default", "private_ip": "10.240.175.15", "public_ip": "173.255.120.190", "status": "RUNNING", "tags": [], "zone": "%s"}], "name": "%s", "state": "present", "zone": "%s"}' % (INAME, ZONE, INAME, ZONE), - }, - {'desc': 'CREATE_INSTANCE the same instance again, no-op [success]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s' % (INAME), - 'r': '127.0.0.1 | success >> {"changed": false, "instance_data": [{"image": "debian-7-wheezy-v20130816", "machine_type": "n1-standard-1", "metadata": {}, "name": "%s", "network": "default", "private_ip": "10.240.175.15", "public_ip": "173.255.120.190", "status": "RUNNING", "tags": [], "zone": "%s"}], "name": "%s", "state": "present", "zone": "%s"}' % (INAME, ZONE, INAME, ZONE), - }, - {'desc': 'CREATE_INSTANCE instance with alt type [success]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s machine_type=n1-standard-2' % (INAME2), - 'r': '127.0.0.1 | success >> {"changed": true, "instance_data": [{"image": "debian-7-wheezy-v20130816", "machine_type": "n1-standard-2", "metadata": {}, "name": "%s", "network": "default", "private_ip": "10.240.192.227", "public_ip": "173.255.121.233", "status": "RUNNING", "tags": [], "zone": "%s"}], "name": "%s", "state": "present", "zone": "%s"}' % (INAME2, ZONE, INAME2, ZONE), - }, - {'desc': 'CREATE_INSTANCE instance with root pd [success]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s persistent_boot_disk=yes' % (INAME3), - 'r': '127.0.0.1 | success >> {"changed": true, "instance_data": [{"image": null, "machine_type": "n1-standard-1", "metadata": {}, "name": "%s", "network": "default", "private_ip": "10.240.178.140", "public_ip": "173.255.121.176", "status": "RUNNING", "tags": [], "zone": "%s"}], "name": "%s", "state": "present", "zone": "%s"}' % (INAME3, ZONE, INAME3, ZONE), - }, - {'desc': 'CREATE_INSTANCE instance with root pd, that already exists [success]', - 'setup': ['gcutil adddisk --source_image=%s --zone=%s %s --wait_until_complete' % (IMAGE, ZONE, DNAME6),], - 'strip_numbers': True, - 'm': 'gce', - 'a': 'name=%s zone=%s persistent_boot_disk=yes' % (INAME6, ZONE), - 'r': '127.0.0.1 | success >> {"changed": true, "instance_data": [{"image": null, "machine_type": "n1-standard-1", "metadata": {}, "name": "%s", "network": "default", "private_ip": "10.240.178.140", "public_ip": "173.255.121.176", "status": "RUNNING", "tags": [], "zone": "%s"}], "name": "%s", "state": "present", "zone": "%s"}' % (INAME6, ZONE, INAME6, ZONE), - }, - {'desc': 'CREATE_INSTANCE instance with root pd attached to other inst [FAIL]', - 'm': 'gce', - 'a': 'name=%s zone=%s persistent_boot_disk=yes' % (INAME7, ZONE), - 'r': '127.0.0.1 | FAILED >> {"failed": true, "msg": "Unexpected error attempting to create instance %s, error: The disk resource \'projects/%s/zones/%s/disks/%s\' is already being used in read-write mode"}' % (INAME7,PROJECT,ZONE,DNAME7), - }, - {'desc': 'CREATE_INSTANCE use *all* the options! [success]', - 'strip_numbers': True, - 'm': 'gce', - 'a': 'instance_names=%s,%s metadata=\'{\\"foo\\":\\"bar\\", \\"baz\\":1}\' tags=t1,t2,t3 zone=%s image=centos-6-v20130731 persistent_boot_disk=yes' % (INAME4,INAME5,ZONE), - 'r': '127.0.0.1 | success >> {"changed": true, "instance_data": [{"image": null, "machine_type": "n1-standard-1", "metadata": {"baz": "1", "foo": "bar"}, "name": "%s", "network": "default", "private_ip": "10.240.130.4", "public_ip": "173.255.121.97", "status": "RUNNING", "tags": ["t1", "t2", "t3"], "zone": "%s"}, {"image": null, "machine_type": "n1-standard-1", "metadata": {"baz": "1", "foo": "bar"}, "name": "%s", "network": "default", "private_ip": "10.240.207.226", "public_ip": "173.255.121.85", "status": "RUNNING", "tags": ["t1", "t2", "t3"], "zone": "%s"}], "instance_names": ["%s", "%s"], "state": "present", "zone": "%s"}' % (INAME4, ZONE, INAME5, ZONE, INAME4, INAME5, ZONE), - }, - ], - 'teardown': ['gcutil deleteinstance -f "%s" --zone=%s' % (INAME, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME2, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME3, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME4, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME5, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME6, ZONE), - 'gcutil deleteinstance -f "%s" --zone=%s' % (INAME7, ZONE), - 'gcutil deleteinstance -f boo --zone=%s' % (ZONE), - 'sleep 10', - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME3, ZONE), - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME4, ZONE), - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME5, ZONE), - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME6, ZONE), - 'gcutil deletedisk -f "%s" --zone=%s' % (INAME7, ZONE), - 'sleep 10'], - }, - - {'id': '06', 'desc': 'Delete / destroy networks and firewall rules', - 'setup': ['gcutil addnetwork --range="%s" --gateway="%s" %s' % (CIDR1, GW1, NETWK1), - 'gcutil addnetwork --range="%s" --gateway="%s" %s' % (CIDR2, GW2, NETWK2), - 'sleep 5', - 'gcutil addfirewall --allowed="tcp:80" --network=%s %s' % (NETWK1, FW1), - 'gcutil addfirewall --allowed="tcp:80" --network=%s %s' % (NETWK2, FW2), - 'sleep 5'], - 'tests': [ - {'desc': 'DELETE bogus named firewall [success]', - 'm': 'gce_net', - 'a': 'fwname=missing-fwrule state=absent', - 'r': '127.0.0.1 | success >> {"changed": false, "fwname": "missing-fwrule", "state": "absent"}', - }, - {'desc': 'DELETE bogus named network [success]', - 'm': 'gce_net', - 'a': 'name=missing-network state=absent', - 'r': '127.0.0.1 | success >> {"changed": false, "name": "missing-network", "state": "absent"}', - }, - {'desc': 'DELETE named firewall rule [success]', - 'm': 'gce_net', - 'a': 'fwname=%s state=absent' % (FW1), - 'r': '127.0.0.1 | success >> {"changed": true, "fwname": "%s", "state": "absent"}' % (FW1), - 'teardown': ['sleep 5'], # pause to give GCE time to delete fwrule - }, - {'desc': 'DELETE unused named network [success]', - 'm': 'gce_net', - 'a': 'name=%s state=absent' % (NETWK1), - 'r': '127.0.0.1 | success >> {"changed": true, "name": "%s", "state": "absent"}' % (NETWK1), - }, - {'desc': 'DELETE named network *and* fwrule [success]', - 'm': 'gce_net', - 'a': 'name=%s fwname=%s state=absent' % (NETWK2, FW2), - 'r': '127.0.0.1 | success >> {"changed": true, "fwname": "%s", "name": "%s", "state": "absent"}' % (FW2, NETWK2), - }, - ], - 'teardown': ['gcutil deletenetwork -f %s' % (NETWK1), - 'gcutil deletenetwork -f %s' % (NETWK2), - 'sleep 5', - 'gcutil deletefirewall -f %s' % (FW1), - 'gcutil deletefirewall -f %s' % (FW2)], - }, - - {'id': '07', 'desc': 'Create networks and firewall rules', - 'setup': ['gcutil addnetwork --range="%s" --gateway="%s" %s' % (CIDR1, GW1, NETWK1), - 'sleep 5', - 'gcutil addfirewall --allowed="tcp:80" --network=%s %s' % (NETWK1, FW1), - 'sleep 5'], - 'tests': [ - {'desc': 'CREATE network without specifying ipv4_range [FAIL]', - 'm': 'gce_net', - 'a': 'name=fail', - 'r': "127.0.0.1 | FAILED >> {\"changed\": false, \"failed\": true, \"msg\": \"Missing required 'ipv4_range' parameter\"}", - }, - {'desc': 'CREATE network with specifying bad ipv4_range [FAIL]', - 'm': 'gce_net', - 'a': 'name=fail ipv4_range=bad_value', - 'r': "127.0.0.1 | FAILED >> {\"changed\": false, \"failed\": true, \"msg\": \"Unexpected response: HTTP return_code[400], API error code[None] and message: Invalid value for field 'resource.IPv4Range': 'bad_value'. Must be a CIDR address range that is contained in the RFC1918 private address blocks: [10.0.0.0/8, 172.16.0.0/12, 192.168.0.0/16]\"}", - }, - {'desc': 'CREATE existing network, not changed [success]', - 'm': 'gce_net', - 'a': 'name=%s ipv4_range=%s' % (NETWK1, CIDR1), - 'r': '127.0.0.1 | success >> {"changed": false, "ipv4_range": "%s", "name": "%s", "state": "present"}' % (CIDR1, NETWK1), - }, - {'desc': 'CREATE new network, changed [success]', - 'm': 'gce_net', - 'a': 'name=%s ipv4_range=%s' % (NETWK2, CIDR2), - 'r': '127.0.0.1 | success >> {"changed": true, "ipv4_range": "10.240.32.0/24", "name": "%s", "state": "present"}' % (NETWK2), - }, - {'desc': 'CREATE new fw rule missing params [FAIL]', - 'm': 'gce_net', - 'a': 'name=%s fwname=%s' % (NETWK1, FW1), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Missing required firewall rule parameter(s)"}', - }, - {'desc': 'CREATE new fw rule bad params [FAIL]', - 'm': 'gce_net', - 'a': 'name=%s fwname=broken allowed=blah src_tags="one,two"' % (NETWK1), - 'r': "127.0.0.1 | FAILED >> {\"changed\": false, \"failed\": true, \"msg\": \"Unexpected response: HTTP return_code[400], API error code[None] and message: Invalid value for field 'resource.allowed[0].IPProtocol': 'blah'. Must be one of [\\\"tcp\\\", \\\"udp\\\", \\\"icmp\\\"] or an IP protocol number between 0 and 255\"}", - }, - {'desc': 'CREATE existing fw rule [success]', - 'm': 'gce_net', - 'a': 'name=%s fwname=%s allowed="tcp:80" src_tags="one,two"' % (NETWK1, FW1), - 'r': '127.0.0.1 | success >> {"allowed": "tcp:80", "changed": false, "fwname": "%s", "ipv4_range": "%s", "name": "%s", "src_range": null, "src_tags": ["one", "two"], "state": "present"}' % (FW1, CIDR1, NETWK1), - }, - {'desc': 'CREATE new fw rule [success]', - 'm': 'gce_net', - 'a': 'name=%s fwname=%s allowed="tcp:80" src_tags="one,two"' % (NETWK1, FW3), - 'r': '127.0.0.1 | success >> {"allowed": "tcp:80", "changed": true, "fwname": "%s", "ipv4_range": "%s", "name": "%s", "src_range": null, "src_tags": ["one", "two"], "state": "present"}' % (FW3, CIDR1, NETWK1), - }, - {'desc': 'CREATE new network *and* fw rule [success]', - 'm': 'gce_net', - 'a': 'name=%s ipv4_range=%s fwname=%s allowed="tcp:80" src_tags="one,two"' % (NETWK3, CIDR3, FW4), - 'r': '127.0.0.1 | success >> {"allowed": "tcp:80", "changed": true, "fwname": "%s", "ipv4_range": "%s", "name": "%s", "src_range": null, "src_tags": ["one", "two"], "state": "present"}' % (FW4, CIDR3, NETWK3), - }, - ], - 'teardown': ['gcutil deletefirewall -f %s' % (FW1), - 'gcutil deletefirewall -f %s' % (FW2), - 'gcutil deletefirewall -f %s' % (FW3), - 'gcutil deletefirewall -f %s' % (FW4), - 'sleep 5', - 'gcutil deletenetwork -f %s' % (NETWK1), - 'gcutil deletenetwork -f %s' % (NETWK2), - 'gcutil deletenetwork -f %s' % (NETWK3), - 'sleep 5'], - }, - - {'id': '08', 'desc': 'Create load-balancer resources', - 'setup': ['gcutil addinstance "%s" --zone=%s --machine_type=%s --network=%s --service_account_scopes="%s" --image="%s" --nopersistent_boot_disk' % (INAME, ZONE, TYPE, NETWORK, SCOPES, IMAGE), - 'gcutil addinstance "%s" --wait_until_running --zone=%s --machine_type=%s --network=%s --service_account_scopes="%s" --image="%s" --nopersistent_boot_disk' % (INAME2, ZONE, TYPE, NETWORK, SCOPES, IMAGE), - ], - 'tests': [ - {'desc': 'Do nothing [FAIL]', - 'm': 'gce_lb', - 'a': 'httphealthcheck_port=7', - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Nothing to do, please specify a \\\"name\\\" or \\\"httphealthcheck_name\\\" parameter"}', - }, - {'desc': 'CREATE_HC create basic http healthcheck [success]', - 'm': 'gce_lb', - 'a': 'httphealthcheck_name=%s' % (HC1), - 'r': '127.0.0.1 | success >> {"changed": true, "httphealthcheck_healthy_count": 2, "httphealthcheck_host": null, "httphealthcheck_interval": 5, "httphealthcheck_name": "%s", "httphealthcheck_path": "/", "httphealthcheck_port": 80, "httphealthcheck_timeout": 5, "httphealthcheck_unhealthy_count": 2, "name": null, "state": "present"}' % (HC1), - }, - {'desc': 'CREATE_HC (repeat, no-op) create basic http healthcheck [success]', - 'm': 'gce_lb', - 'a': 'httphealthcheck_name=%s' % (HC1), - 'r': '127.0.0.1 | success >> {"changed": false, "httphealthcheck_healthy_count": 2, "httphealthcheck_host": null, "httphealthcheck_interval": 5, "httphealthcheck_name": "%s", "httphealthcheck_path": "/", "httphealthcheck_port": 80, "httphealthcheck_timeout": 5, "httphealthcheck_unhealthy_count": 2, "name": null, "state": "present"}' % (HC1), - }, - {'desc': 'CREATE_HC create custom http healthcheck [success]', - 'm': 'gce_lb', - 'a': 'httphealthcheck_name=%s httphealthcheck_port=1234 httphealthcheck_path="/whatup" httphealthcheck_host="foo" httphealthcheck_interval=300' % (HC2), - 'r': '127.0.0.1 | success >> {"changed": true, "httphealthcheck_healthy_count": 2, "httphealthcheck_host": "foo", "httphealthcheck_interval": 300, "httphealthcheck_name": "%s", "httphealthcheck_path": "/whatup", "httphealthcheck_port": 1234, "httphealthcheck_timeout": 5, "httphealthcheck_unhealthy_count": 2, "name": null, "state": "present"}' % (HC2), - }, - {'desc': 'CREATE_HC create (broken) custom http healthcheck [FAIL]', - 'm': 'gce_lb', - 'a': 'httphealthcheck_name=%s httphealthcheck_port="string" httphealthcheck_path=7' % (HC3), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Unexpected response: HTTP return_code[400], API error code[None] and message: Invalid value for: Expected a signed integer, got \'string\' (class java.lang.String)"}', - }, - {'desc': 'CREATE_LB create lb, missing region [FAIL]', - 'm': 'gce_lb', - 'a': 'name=%s' % (LB1), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Missing required region name"}', - }, - {'desc': 'CREATE_LB create lb, bogus region [FAIL]', - 'm': 'gce_lb', - 'a': 'name=%s region=bogus' % (LB1), - 'r': '127.0.0.1 | FAILED >> {"changed": false, "failed": true, "msg": "Unexpected response: HTTP return_code[404], API error code[None] and message: The resource \'projects/%s/regions/bogus\' was not found"}' % (PROJECT), - }, - {'desc': 'CREATE_LB create lb, minimal params [success]', - 'strip_numbers': True, - 'm': 'gce_lb', - 'a': 'name=%s region=%s' % (LB1, REGION), - 'r': '127.0.0.1 | success >> {"changed": true, "external_ip": "173.255.123.245", "httphealthchecks": [], "members": [], "name": "%s", "port_range": "1-65535", "protocol": "tcp", "region": "%s", "state": "present"}' % (LB1, REGION), - }, - {'desc': 'CREATE_LB create lb full params [success]', - 'strip_numbers': True, - 'm': 'gce_lb', - 'a': 'httphealthcheck_name=%s httphealthcheck_port=5055 httphealthcheck_path="/howami" name=%s port_range=8000-8888 region=%s members=%s/%s,%s/%s' % (HC3,LB2,REGION,ZONE,INAME,ZONE,INAME2), - 'r': '127.0.0.1 | success >> {"changed": true, "external_ip": "173.255.126.81", "httphealthcheck_healthy_count": 2, "httphealthcheck_host": null, "httphealthcheck_interval": 5, "httphealthcheck_name": "%s", "httphealthcheck_path": "/howami", "httphealthcheck_port": 5055, "httphealthcheck_timeout": 5, "httphealthcheck_unhealthy_count": 2, "httphealthchecks": ["%s"], "members": ["%s/%s", "%s/%s"], "name": "%s", "port_range": "8000-8888", "protocol": "tcp", "region": "%s", "state": "present"}' % (HC3,HC3,ZONE,INAME,ZONE,INAME2,LB2,REGION), - }, - ], - 'teardown': [ - 'gcutil deleteinstance --zone=%s -f %s %s' % (ZONE, INAME, INAME2), - 'gcutil deleteforwardingrule --region=%s -f %s %s' % (REGION, LB1, LB2), - 'sleep 10', - 'gcutil deletetargetpool --region=%s -f %s-tp %s-tp' % (REGION, LB1, LB2), - 'sleep 10', - 'gcutil deletehttphealthcheck -f %s %s %s' % (HC1, HC2, HC3), - ], - }, - - {'id': '09', 'desc': 'Destroy load-balancer resources', - 'setup': ['gcutil addhttphealthcheck %s' % (HC1), - 'sleep 5', - 'gcutil addhttphealthcheck %s' % (HC2), - 'sleep 5', - 'gcutil addtargetpool --health_checks=%s --region=%s %s-tp' % (HC1, REGION, LB1), - 'sleep 5', - 'gcutil addforwardingrule --target=%s-tp --region=%s %s' % (LB1, REGION, LB1), - 'sleep 5', - 'gcutil addtargetpool --region=%s %s-tp' % (REGION, LB2), - 'sleep 5', - 'gcutil addforwardingrule --target=%s-tp --region=%s %s' % (LB2, REGION, LB2), - 'sleep 5', - ], - 'tests': [ - {'desc': 'DELETE_LB: delete a non-existent LB [success]', - 'm': 'gce_lb', - 'a': 'name=missing state=absent', - 'r': '127.0.0.1 | success >> {"changed": false, "name": "missing", "state": "absent"}', - }, - {'desc': 'DELETE_LB: delete a non-existent LB+HC [success]', - 'm': 'gce_lb', - 'a': 'name=missing httphealthcheck_name=alsomissing state=absent', - 'r': '127.0.0.1 | success >> {"changed": false, "httphealthcheck_name": "alsomissing", "name": "missing", "state": "absent"}', - }, - {'desc': 'DELETE_LB: destroy standalone healthcheck [success]', - 'm': 'gce_lb', - 'a': 'httphealthcheck_name=%s state=absent' % (HC2), - 'r': '127.0.0.1 | success >> {"changed": true, "httphealthcheck_name": "%s", "name": null, "state": "absent"}' % (HC2), - }, - {'desc': 'DELETE_LB: destroy standalone balancer [success]', - 'm': 'gce_lb', - 'a': 'name=%s state=absent' % (LB2), - 'r': '127.0.0.1 | success >> {"changed": true, "name": "%s", "state": "absent"}' % (LB2), - }, - {'desc': 'DELETE_LB: destroy LB+HC [success]', - 'm': 'gce_lb', - 'a': 'name=%s httphealthcheck_name=%s state=absent' % (LB1, HC1), - 'r': '127.0.0.1 | success >> {"changed": true, "httphealthcheck_name": "%s", "name": "%s", "state": "absent"}' % (HC1,LB1), - }, - ], - 'teardown': [ - 'gcutil deleteforwardingrule --region=%s -f %s %s' % (REGION, LB1, LB2), - 'sleep 10', - 'gcutil deletetargetpool --region=%s -f %s-tp %s-tp' % (REGION, LB1, LB2), - 'sleep 10', - 'gcutil deletehttphealthcheck -f %s %s' % (HC1, HC2), - ], - }, -] - -def main(tests_to_run=[]): - for test in test_cases: - if tests_to_run and test['id'] not in tests_to_run: - continue - print "=> starting/setup '%s:%s'"% (test['id'], test['desc']) - if DEBUG: print "=debug>", test['setup'] - for c in test['setup']: - (s,o) = run(c) - test_i = 1 - for t in test['tests']: - if DEBUG: print "=>debug>", test_i, t['desc'] - # run any test-specific setup commands - if t.has_key('setup'): - for setup in t['setup']: - (status, output) = run(setup) - - # run any 'peek_before' commands - if t.has_key('peek_before') and PEEKING_ENABLED: - for setup in t['peek_before']: - (status, output) = run(setup) - - # run the ansible test if 'a' exists, otherwise - # an empty 'a' directive allows test to run - # setup/teardown for a subsequent test. - if t['a']: - if DEBUG: print "=>debug>", t['m'], t['a'] - acmd = "ansible all -o -m %s -a \"%s\"" % (t['m'],t['a']) - #acmd = "ANSIBLE_KEEP_REMOTE_FILES=1 ansible all -vvv -m %s -a \"%s\"" % (t['m'],t['a']) - (s,o) = run(acmd) - - # check expected output - if DEBUG: print "=debug>", o.strip(), "!=", t['r'] - print "=> %s.%02d '%s':" % (test['id'], test_i, t['desc']), - if t.has_key('strip_numbers'): - # strip out all numbers so we don't trip over different - # IP addresses - is_good = (o.strip().translate(None, "0123456789") == t['r'].translate(None, "0123456789")) - else: - is_good = (o.strip() == t['r']) - - if is_good: - print "PASS" - else: - print "FAIL" - if VERBOSE: - print "=>", acmd - print "=> Expected:", t['r'] - print "=> Got:", o.strip() - - # run any 'peek_after' commands - if t.has_key('peek_after') and PEEKING_ENABLED: - for setup in t['peek_after']: - (status, output) = run(setup) - - # run any test-specific teardown commands - if t.has_key('teardown'): - for td in t['teardown']: - (status, output) = run(td) - test_i += 1 - - print "=> completing/teardown '%s:%s'" % (test['id'], test['desc']) - if DEBUG: print "=debug>", test['teardown'] - for c in test['teardown']: - (s,o) = run(c) - - -if __name__ == '__main__': - tests_to_run = [] - if len(sys.argv) == 2: - if sys.argv[1] in ["--help", "--list"]: - print "usage: %s [id1,id2,...,idN]" % sys.argv[0] - print " * An empty argument list will execute all tests" - print " * Do not need to specify tests in numerical order" - print " * List test categories with --list or --help" - print "" - for test in test_cases: - print "\t%s:%s" % (test['id'], test['desc']) - sys.exit(0) - else: - tests_to_run = sys.argv[1].split(',') - main(tests_to_run)