diff --git a/lib/ansible/modules/cloud/azure/azure_rm_aks.py b/lib/ansible/modules/cloud/azure/azure_rm_aks.py new file mode 100644 index 0000000000..f15b83b230 --- /dev/null +++ b/lib/ansible/modules/cloud/azure/azure_rm_aks.py @@ -0,0 +1,605 @@ +#!/usr/bin/python +# +# Copyright (c) 2018 Sertac Ozercan, +# +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +ANSIBLE_METADATA = {'metadata_version': '1.1', + 'status': ['preview'], + 'supported_by': 'community'} + +DOCUMENTATION = ''' +--- +module: azure_rm_aks +version_added: "2.6" +short_description: Manage a managed Azure Container Service (AKS) Instance. +description: + - Create, update and delete a managed Azure Container Service (AKS) Instance. + +options: + resource_group: + description: + - Name of a resource group where the managed Azure Container Services (AKS) exists or will be created. + required: true + name: + description: + - Name of the managed Azure Container Services (AKS) instance. + required: true + state: + description: + - Assert the state of the AKS. Use C(present) to create or update an AKS and C(absent) to delete it. + default: present + choices: + - absent + - present + location: + description: + - Valid azure location. Defaults to location of the resource group. + dns_prefix: + description: + - DNS prefix specified when creating the managed cluster. + kubernetes_version: + description: + - Version of Kubernetes specified when creating the managed cluster. + linux_profile: + description: + - The linux profile suboptions. + suboptions: + admin_username: + description: + - The Admin Username for the Cluster. + required: true + ssh_key: + description: + - The Public SSH Key used to access the cluster. + required: true + agent_pool_profiles: + description: + - The agent pool profile suboptions. + suboptions: + name: + description: + - Unique name of the agent pool profile in the context of the subscription and resource group. + required: true + count: + description: + - Number of agents (VMs) to host docker containers. + - Allowed values must be in the range of 1 to 100 (inclusive). + required: true + vm_size: + description: + - The VM Size of each of the Agent Pool VM's (e.g. Standard_F1 / Standard_D2v2). + required: true + os_disk_size_gb: + description: + - Size of the OS disk. + service_principal: + description: + - The service principal suboptions. + suboptions: + client_id: + description: + - The ID for the Service Principal. + required: true + client_secret: + description: + - The secret password associated with the service principal. + required: true + +extends_documentation_fragment: + - azure + - azure_tags + +author: + - "Sertac Ozercan (@sozercan)" + - "Yuwei Zhou (@yuwzho)" + +''' + +EXAMPLES = ''' + - name: Create a managed Azure Container Services (AKS) instance + azure_rm_aks: + name: acctestaks1 + location: eastus + resource_group: Testing + dns_prefix: akstest + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAA... + service_principal: + client_id: "cf72ca99-f6b9-4004-b0e0-bee10c521948" + client_secret: "mySPNp@ssw0rd!" + agent_pool_profiles: + - name: default + count: 5 + vm_size: Standard_D2_v2 + tags: + Environment: Production + + - name: Remove a managed Azure Container Services (AKS) instance + azure_rm_aks: + name: acctestaks3 + location: eastus + dns_prefix: akstest + resource_group: Testing + state: absent + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAA... + agent_pool_profiles: + - name: default + count: 4 + vm_size: Standard_A0 + tags: + Ansible: azure_rm_aks +''' +RETURN = ''' +state: + description: Current state of the Azure Container Service (AKS) + returned: always + type: dict + example: + agent_pool_profiles: + - count: 1 + dns_prefix: Null + name: default + os_disk_size_gb: Null + os_type: Linux + ports: Null + storage_profile: ManagedDisks + vm_size: Standard_DS1_v2 + vnet_subnet_id: Null + changed: false + dns_prefix: aks9860bdcd89 + id: "/subscriptions/XXXXXXXX-XXXX-XXXX-XXXX-XXXXXXXXXXXX/resourcegroups/yuwzhoaks/providers/Microsoft.ContainerService/managedClusters/aks9860bdc" + kube_config: "......" + kubernetes_version: 1.7.7 + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADA..... + location: eastus + name: aks9860bdc + provisioning_state: Succeeded + service_principal_profile: + client_id: XXXXXXXX-XXXX-XXXX-XXXX-XXXXXXXXXXXX + tags: {} + type: Microsoft.ContainerService/ManagedClusters +''' +from ansible.module_utils.azure_rm_common import AzureRMModuleBase +import base64 + +try: + from msrestazure.azure_exceptions import CloudError + from azure.mgmt.containerservice.models import ( + ManagedCluster, ContainerServiceServicePrincipalProfile, + ContainerServiceAgentPoolProfile, ContainerServiceLinuxProfile, + ContainerServiceSshConfiguration, ContainerServiceSshPublicKey + ) +except ImportError: + # This is handled in azure_rm_common + pass + + +def create_agent_pool_profile_instance(agentpoolprofile): + ''' + Helper method to serialize a dict to a ContainerServiceAgentPoolProfile + :param: agentpoolprofile: dict with the parameters to setup the ContainerServiceAgentPoolProfile + :return: ContainerServiceAgentPoolProfile + ''' + return ContainerServiceAgentPoolProfile( + name=agentpoolprofile['name'], + count=agentpoolprofile['count'], + vm_size=agentpoolprofile['vm_size'], + os_disk_size_gb=agentpoolprofile['os_disk_size_gb'], + dns_prefix=agentpoolprofile.get('dns_prefix'), + ports=agentpoolprofile.get('ports'), + storage_profile=agentpoolprofile.get('storage_profile'), + vnet_subnet_id=agentpoolprofile.get('vnet_subnet_id'), + os_type=agentpoolprofile.get('os_type') + ) + + +def create_service_principal_profile_instance(spnprofile): + ''' + Helper method to serialize a dict to a ContainerServiceServicePrincipalProfile + :param: spnprofile: dict with the parameters to setup the ContainerServiceServicePrincipalProfile + :return: ContainerServiceServicePrincipalProfile + ''' + return ContainerServiceServicePrincipalProfile( + client_id=spnprofile['client_id'], + secret=spnprofile['client_secret'] + ) + + +def create_linux_profile_instance(linuxprofile): + ''' + Helper method to serialize a dict to a ContainerServiceLinuxProfile + :param: linuxprofile: dict with the parameters to setup the ContainerServiceLinuxProfile + :return: ContainerServiceLinuxProfile + ''' + return ContainerServiceLinuxProfile( + admin_username=linuxprofile['admin_username'], + ssh=create_ssh_configuration_instance(linuxprofile['ssh_key']) + ) + + +def create_ssh_configuration_instance(sshconf): + ''' + Helper method to serialize a dict to a ContainerServiceSshConfiguration + :param: sshconf: dict with the parameters to setup the ContainerServiceSshConfiguration + :return: ContainerServiceSshConfiguration + ''' + return ContainerServiceSshConfiguration( + public_keys=[ContainerServiceSshPublicKey(key_data=str(sshconf))] + ) + + +def create_aks_dict(aks): + ''' + Helper method to deserialize a ContainerService to a dict + :param: aks: ContainerService or AzureOperationPoller with the Azure callback object + :return: dict with the state on Azure + ''' + + return dict( + id=aks.id, + name=aks.name, + location=aks.location, + dns_prefix=aks.dns_prefix, + kubernetes_version=aks.kubernetes_version, + tags=aks.tags, + linux_profile=create_linux_profile_dict(aks.linux_profile), + service_principal_profile=create_service_principal_profile_dict( + aks.service_principal_profile), + provisioning_state=aks.provisioning_state, + agent_pool_profiles=create_agent_pool_profiles_dict( + aks.agent_pool_profiles), + type=aks.type, + kube_config=aks.kube_config + ) + + +def create_linux_profile_dict(linuxprofile): + ''' + Helper method to deserialize a ContainerServiceLinuxProfile to a dict + :param: linuxprofile: ContainerServiceLinuxProfile with the Azure callback object + :return: dict with the state on Azure + ''' + return dict( + ssh_key=linuxprofile.ssh.public_keys[0].key_data, + admin_username=linuxprofile.admin_username + ) + + +def create_service_principal_profile_dict(serviceprincipalprofile): + ''' + Helper method to deserialize a ContainerServiceServicePrincipalProfile to a dict + Note: For security reason, the service principal secret is skipped on purpose. + :param: serviceprincipalprofile: ContainerServiceServicePrincipalProfile with the Azure callback object + :return: dict with the state on Azure + ''' + return dict( + client_id=serviceprincipalprofile.client_id + ) + + +def create_agent_pool_profiles_dict(agentpoolprofiles): + ''' + Helper method to deserialize a ContainerServiceAgentPoolProfile to a dict + :param: agentpoolprofiles: ContainerServiceAgentPoolProfile with the Azure callback object + :return: dict with the state on Azure + ''' + return [dict( + count=profile.count, + vm_size=profile.vm_size, + name=profile.name, + os_disk_size_gb=profile.os_disk_size_gb, + dns_prefix=profile.dns_prefix, + ports=profile.ports, + storage_profile=profile.storage_profile, + vnet_subnet_id=profile.vnet_subnet_id, + os_type=profile.os_type + ) for profile in agentpoolprofiles] if agentpoolprofiles else None + + +linux_profile_spec = dict( + admin_username=dict(type='str', required=True), + ssh_key=dict(type='str', required=True) +) + + +service_principal_spec = dict( + client_id=dict(type='str', required=True), + client_secret=dict(type='str') +) + + +agent_pool_profile_spec = dict( + name=dict(type='str', required=True), + count=dict(type='int', required=True), + vm_size=dict(type='str', required=True), + os_disk_size_gb=dict(type='int'), + dns_prefix=dict(type='str'), + ports=dict(type='list', elements='int'), + storage_profiles=dict(type='str', choices=[ + 'StorageAccount', 'ManagedDisks']), + vnet_subnet_id=dict(type='str'), + os_type=dict(type='str', choices=['Linux', 'Windows']) +) + + +class AzureRMManagedCluster(AzureRMModuleBase): + """Configuration class for an Azure RM container service (AKS) resource""" + + def __init__(self): + self.module_arg_spec = dict( + resource_group=dict( + type='str', + required=True + ), + name=dict( + type='str', + required=True + ), + state=dict( + type='str', + default='present', + choices=['present', 'absent'] + ), + location=dict( + type='str' + ), + dns_prefix=dict( + type='str' + ), + kubernetes_version=dict( + type='str' + ), + linux_profile=dict( + type='dict', + options=linux_profile_spec + ), + agent_pool_profiles=dict( + type='list', + elements='dict', + options=agent_pool_profile_spec + ), + service_principal=dict( + type='dict', + options=service_principal_spec + ), + ) + + self.resource_group = None + self.name = None + self.location = None + self.dns_prefix = None + self.kubernetes_version = None + self.tags = None + self.state = None + self.linux_profile = None + self.agent_pool_profiles = None + self.service_principal = None + + required_if = [ + ('state', 'present', [ + 'dns_prefix', 'linux_profile', 'agent_pool_profiles', 'service_principal']) + ] + + self.results = dict(changed=False) + + super(AzureRMManagedCluster, self).__init__(derived_arg_spec=self.module_arg_spec, + supports_check_mode=True, + supports_tags=True, + required_if=required_if) + + def exec_module(self, **kwargs): + """Main module execution method""" + + for key in list(self.module_arg_spec.keys()) + ['tags']: + setattr(self, key, kwargs[key]) + + resource_group = None + to_be_updated = False + + resource_group = self.get_resource_group(self.resource_group) + if not self.location: + self.location = resource_group.location + + response = self.get_aks() + + # Check if the AKS instance already present in the RG + if self.state == 'present': + # For now Agent Pool cannot be more than 1, just remove this part in the future if it change + agentpoolcount = len(self.agent_pool_profiles) + if agentpoolcount > 1: + self.fail( + 'You cannot specify more than one agent_pool_profiles currently') + + if response: + self.results = response + if not response: + to_be_updated = True + + else: + self.log('Results : {0}'.format(response)) + update_tags, response['tags'] = self.update_tags( + response['tags']) + + if response['provisioning_state'] == "Succeeded": + if update_tags: + to_be_updated = True + + def is_property_changed(profile, property, ignore_case=False): + base = response[profile].get(property) + new = getattr(self, profile).get(property) + if ignore_case: + return base.lower() != new.lower() + else: + return base != new + + # Cannot Update the SSH Key for now // Let service to handle it + if is_property_changed('linux_profile', 'ssh_key'): + self.log(("Linux Profile Diff SSH, Was {0} / Now {1}" + .format(response['linux_profile']['ssh_key'], self.linux_profile.get('ssh_key')))) + to_be_updated = True + # self.module.warn("linux_profile.ssh_key cannot be updated") + + # self.log("linux_profile response : {0}".format(response['linux_profile'].get('admin_username'))) + # self.log("linux_profile self : {0}".format(self.linux_profile[0].get('admin_username'))) + # Cannot Update the Username for now // Let service to handle it + if is_property_changed('linux_profile', 'admin_username'): + self.log(("Linux Profile Diff User, Was {0} / Now {1}" + .format(response['linux_profile']['admin_username'], self.linux_profile.get('admin_username')))) + to_be_updated = True + # self.module.warn("linux_profile.admin_username cannot be updated") + + # Cannot have more that one agent pool profile for now + if len(response['agent_pool_profiles']) != len(self.agent_pool_profiles): + self.log("Agent Pool count is diff, need to updated") + to_be_updated = True + + if response['kubernetes_version'] != self.kubernetes_version: + to_be_updated = True + + for profile_result in response['agent_pool_profiles']: + matched = False + for profile_self in self.agent_pool_profiles: + if profile_result['name'] == profile_self['name']: + matched = True + if profile_result['count'] != profile_self['count'] \ + or profile_result['vm_size'] != profile_self['vm_size'] \ + or profile_result['os_disk_size_gb'] != profile_self['os_disk_size_gb'] \ + or profile_result['dns_prefix'] != profile_self['dns_prefix'] \ + or profile_result['vnet_subnet_id'] != profile_self.get('vnet_subnet_id') \ + or set(profile_result['ports'] or []) != set(profile_self.get('ports') or []): + self.log( + ("Agent Profile Diff - Origin {0} / Update {1}".format(str(profile_result), str(profile_self)))) + to_be_updated = True + if not matched: + self.log("Agent Pool not found") + to_be_updated = True + + if to_be_updated: + self.log("Need to Create / Update the AKS instance") + + if not self.check_mode: + self.results = self.create_update_aks() + self.log("Creation / Update done") + + self.results['changed'] = True + return self.results + + elif self.state == 'absent' and response: + self.log("Need to Delete the AKS instance") + self.results['changed'] = True + + if self.check_mode: + return self.results + + self.delete_aks() + + self.log("AKS instance deleted") + + return self.results + + def create_update_aks(self): + ''' + Creates or updates a managed Azure container service (AKS) with the specified configuration of agents. + + :return: deserialized AKS instance state dictionary + ''' + self.log("Creating / Updating the AKS instance {0}".format(self.name)) + + agentpools = [] + + if self.agent_pool_profiles: + agentpools = [create_agent_pool_profile_instance( + profile) for profile in self.agent_pool_profiles] + + service_principal_profile = create_service_principal_profile_instance( + self.service_principal) + + parameters = ManagedCluster( + location=self.location, + dns_prefix=self.dns_prefix, + kubernetes_version=self.kubernetes_version, + tags=self.tags, + service_principal_profile=service_principal_profile, + agent_pool_profiles=agentpools, + linux_profile=create_linux_profile_instance(self.linux_profile) + ) + + # self.log("service_principal_profile : {0}".format(parameters.service_principal_profile)) + # self.log("linux_profile : {0}".format(parameters.linux_profile)) + # self.log("ssh from yaml : {0}".format(results.get('linux_profile')[0])) + # self.log("ssh : {0}".format(parameters.linux_profile.ssh)) + # self.log("agent_pool_profiles : {0}".format(parameters.agent_pool_profiles)) + + try: + poller = self.containerservice_client.managed_clusters.create_or_update(self.resource_group, self.name, parameters) + response = self.get_poller_result(poller) + response.kube_config = self.get_aks_kubeconfig() + return create_aks_dict(response) + except CloudError as exc: + self.log('Error attempting to create the AKS instance.') + self.fail("Error creating the AKS instance: {0}".format(exc.message)) + + def delete_aks(self): + ''' + Deletes the specified managed container service (AKS) in the specified subscription and resource group. + + :return: True + ''' + self.log("Deleting the AKS instance {0}".format(self.name)) + try: + poller = self.containerservice_client.managed_clusters.delete( + self.resource_group, self.name) + self.get_poller_result(poller) + return True + except CloudError as e: + self.log('Error attempting to delete the AKS instance.') + self.fail("Error deleting the AKS instance: {0}".format(e.message)) + return False + + def get_aks(self): + ''' + Gets the properties of the specified container service. + + :return: deserialized AKS instance state dictionary + ''' + self.log( + "Checking if the AKS instance {0} is present".format(self.name)) + try: + response = self.containerservice_client.managed_clusters.get( + self.resource_group, self.name) + self.log("Response : {0}".format(response)) + self.log("AKS instance : {0} found".format(response.name)) + response.kube_config = self.get_aks_kubeconfig() + return create_aks_dict(response) + except CloudError: + self.log('Did not find the AKS instance.') + return False + + def get_aks_kubeconfig(self): + ''' + Gets kubeconfig for the specified AKS instance. + + :return: AKS instance kubeconfig + ''' + access_profile = self.containerservice_client.managed_clusters.get_access_profiles( + self.resource_group, self.name, "clusterUser") + return base64.b64decode(access_profile.kube_config) + + +def main(): + """Main execution""" + AzureRMManagedCluster() + + +if __name__ == '__main__': + main() diff --git a/lib/ansible/modules/cloud/azure/azure_rm_aks_facts.py b/lib/ansible/modules/cloud/azure/azure_rm_aks_facts.py new file mode 100644 index 0000000000..c1a793aa69 --- /dev/null +++ b/lib/ansible/modules/cloud/azure/azure_rm_aks_facts.py @@ -0,0 +1,161 @@ +#!/usr/bin/python +# +# Copyright (c) 2018 Yuwei Zhou, +# +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function +__metaclass__ = type + + +ANSIBLE_METADATA = {'metadata_version': '1.1', + 'status': ['preview'], + 'supported_by': 'community'} + +DOCUMENTATION = ''' +--- +module: azure_rm_aks_facts + +version_added: "2.6" + +short_description: Get Azure Kubernetes Service facts. + +description: + - Get facts for a specific Azure Kubernetes Service or all Azure Kubernetes Services. + +options: + name: + description: + - Limit results to a specific resource group. + resource_group: + description: + - The resource group to search for the desired Azure Kubernetes Service + tags: + description: + - Limit results by providing a list of tags. Format tags as 'key' or 'key:value'. + +extends_documentation_fragment: + - azure + +author: + - "Yuwei Zhou (@yuwzho)" +''' + +EXAMPLES = ''' + - name: Get facts for one Azure Kubernetes Service + azure_rm_aks_facts: + name: Testing + resource_group: TestRG + + - name: Get facts for all Azure Kubernetes Services + azure_rm_aks_facts: + + - name: Get facts by tags + azure_rm_aks_facts: + tags: + - testing +''' + +RETURN = ''' +azure_aks: + description: List of Azure Kubernetes Service dicts. + returned: always + type: list +''' + +from ansible.module_utils.azure_rm_common import AzureRMModuleBase + +try: + from msrestazure.azure_exceptions import CloudError + from azure.common import AzureHttpError +except: + # handled in azure_rm_common + pass + +AZURE_OBJECT_CLASS = 'managedClusters' + + +class AzureRMManagedClusterFacts(AzureRMModuleBase): + """Utility class to get Azure Kubernetes Service facts""" + + def __init__(self): + + self.module_args = dict( + name=dict(type='str'), + resource_group=dict(type='str'), + tags=dict(type='list') + ) + + self.results = dict( + changed=False, + aks=[] + ) + + self.name = None + self.resource_group = None + self.tags = None + + super(AzureRMManagedClusterFacts, self).__init__( + derived_arg_spec=self.module_args, + supports_tags=False, + facts_module=True + ) + + def exec_module(self, **kwargs): + + for key in self.module_args: + setattr(self, key, kwargs[key]) + + self.results['aks'] = ( + self.get_item() if self.name + else self.list_items() + ) + + return self.results + + def get_item(self): + """Get a single Azure Kubernetes Service""" + + self.log('Get properties for {0}'.format(self.name)) + + item = None + result = [] + + try: + item = self.containerservice_client.managed_clusters.get( + self.resource_group, self.name) + except CloudError: + pass + + if item and self.has_tags(item.tags, self.tags): + result = [self.serialize_obj(item, AZURE_OBJECT_CLASS)] + + return result + + def list_items(self): + """Get all Azure Kubernetes Services""" + + self.log('List all Azure Kubernetes Services') + + try: + response = self.containerservice_client.managed_clusters.list( + self.resource_group) + except AzureHttpError as exc: + self.fail('Failed to list all items - {0}'.format(str(exc))) + + results = [] + for item in response: + if self.has_tags(item.tags, self.tags): + results.append(self.serialize_obj(item, AZURE_OBJECT_CLASS)) + + return results + + +def main(): + """Main module execution code path""" + + AzureRMManagedClusterFacts() + + +if __name__ == '__main__': + main() diff --git a/packaging/requirements/requirements-azure.txt b/packaging/requirements/requirements-azure.txt index e3345c6787..802cb5d0d9 100644 --- a/packaging/requirements/requirements-azure.txt +++ b/packaging/requirements/requirements-azure.txt @@ -13,7 +13,7 @@ azure-mgmt-keyvault>=0.40.0,<0.41 azure-mgmt-batch>=4.1.0,<5 azure-mgmt-sql>=0.7.1,<0.8 azure-mgmt-web>=0.32.0,<0.33 -azure-mgmt-containerservice>=2.0.0,<3.0.0 +azure-mgmt-containerservice>=3.0.1 azure-mgmt-containerregistry>=1.0.1 azure-mgmt-rdbms>=0.2.0rc1,<0.3.0 azure-mgmt-containerinstance>=0.3.1,<0.4.0 diff --git a/test/integration/targets/azure_rm_aks/aliases b/test/integration/targets/azure_rm_aks/aliases new file mode 100644 index 0000000000..e340069a05 --- /dev/null +++ b/test/integration/targets/azure_rm_aks/aliases @@ -0,0 +1,3 @@ +cloud/azure +unsupported +destructive \ No newline at end of file diff --git a/test/integration/targets/azure_rm_aks/meta/main.yml b/test/integration/targets/azure_rm_aks/meta/main.yml new file mode 100644 index 0000000000..95e1952f98 --- /dev/null +++ b/test/integration/targets/azure_rm_aks/meta/main.yml @@ -0,0 +1,2 @@ +dependencies: + - setup_azure diff --git a/test/integration/targets/azure_rm_aks/tasks/main.yml b/test/integration/targets/azure_rm_aks/tasks/main.yml new file mode 100644 index 0000000000..5a38045d54 --- /dev/null +++ b/test/integration/targets/azure_rm_aks/tasks/main.yml @@ -0,0 +1,172 @@ + - name: Create an AKS instance (check mode) + azure_rm_aks: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + location: eastus + dns_prefix: "aks{{ resource_group | hash('md5') | truncate(10, True, '') }}" + kubernetes_version: 1.7.7 + service_principal: + client_id: "{{ lookup('env', 'AZURE_CLIENT_ID') }}" + client_secret: "{{ lookup('env', 'AZURE_SECRET') }}" + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDSPmiqkvDH1/+MDAVDZT8381aYqp73Odz8cnD5hegNhqtXajqtiH0umVg7HybX3wt1HjcrwKJovZURcIbbcDvzdH2bnYbF93T4OLXA0bIfuIp6M86x1iutFtXdpN3TTicINrmSXEE2Ydm51iMu77B08ZERjVaToya2F7vC+egfoPvibf7OLxE336a5tPCywavvNihQjL8sjgpDT5AAScjb3YqK/6VLeQ18Ggt8/ufINsYkb+9/Ji/3OcGFeflnDXq80vPUyF3u4iIylob6RSZenC38cXmQB05tRNxS1B6BXCjMRdy0v4pa7oKM2GA4ADKpNrr0RI9ed+peRFwmsclH test@ansible + agent_pool_profiles: + - name: default + count: 1 + vm_size: Standard_DS1_v2 + check_mode: yes + + - name: Check there is no AKS created + azure_rm_aks_facts: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + register: fact + + - name: Check there is no AKS created + assert: + that: + - "fact.aks | length == 0" + + - name: Create an AKS instance + azure_rm_aks: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + location: eastus + dns_prefix: "aks{{ resource_group | hash('md5') | truncate(10, True, '') }}" + kubernetes_version: 1.7.7 + service_principal: + client_id: "{{ lookup('env', 'AZURE_CLIENT_ID') }}" + client_secret: "{{ lookup('env', 'AZURE_SECRET') }}" + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDSPmiqkvDH1/+MDAVDZT8381aYqp73Odz8cnD5hegNhqtXajqtiH0umVg7HybX3wt1HjcrwKJovZURcIbbcDvzdH2bnYbF93T4OLXA0bIfuIp6M86x1iutFtXdpN3TTicINrmSXEE2Ydm51iMu77B08ZERjVaToya2F7vC+egfoPvibf7OLxE336a5tPCywavvNihQjL8sjgpDT5AAScjb3YqK/6VLeQ18Ggt8/ufINsYkb+9/Ji/3OcGFeflnDXq80vPUyF3u4iIylob6RSZenC38cXmQB05tRNxS1B6BXCjMRdy0v4pa7oKM2GA4ADKpNrr0RI9ed+peRFwmsclH test@ansible + agent_pool_profiles: + - name: default + count: 1 + vm_size: Standard_DS1_v2 + register: output + + - name: Assert the AKS instance is well created + assert: + that: + - output.changed + - output.provisioning_state == 'Succeeded' + + - name: Get AKS fact + azure_rm_aks_facts: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + register: fact + + - name: Assert fact returns the created one + assert: + that: + - "fact.aks | length == 1" + - fact.aks[0].id == output.id + + - name: Create an AKS instance (idempotent) + azure_rm_aks: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + location: eastus + dns_prefix: "aks{{ resource_group | hash('md5') | truncate(10, True, '') }}" + kubernetes_version: 1.7.7 + service_principal: + client_id: "{{ lookup('env', 'AZURE_CLIENT_ID') }}" + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDSPmiqkvDH1/+MDAVDZT8381aYqp73Odz8cnD5hegNhqtXajqtiH0umVg7HybX3wt1HjcrwKJovZURcIbbcDvzdH2bnYbF93T4OLXA0bIfuIp6M86x1iutFtXdpN3TTicINrmSXEE2Ydm51iMu77B08ZERjVaToya2F7vC+egfoPvibf7OLxE336a5tPCywavvNihQjL8sjgpDT5AAScjb3YqK/6VLeQ18Ggt8/ufINsYkb+9/Ji/3OcGFeflnDXq80vPUyF3u4iIylob6RSZenC38cXmQB05tRNxS1B6BXCjMRdy0v4pa7oKM2GA4ADKpNrr0RI9ed+peRFwmsclH test@ansible + agent_pool_profiles: + - name: default + count: 1 + vm_size: Standard_DS1_v2 + register: output + + - name: Assert idempotent + assert: + that: + - not output.changed + + - name: Upgrade the AKS instance + azure_rm_aks: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + location: eastus + dns_prefix: "aks{{ resource_group | hash('md5') | truncate(10, True, '') }}" + kubernetes_version: 1.8.1 + service_principal: + client_id: "{{ lookup('env', 'AZURE_CLIENT_ID') }}" + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDSPmiqkvDH1/+MDAVDZT8381aYqp73Odz8cnD5hegNhqtXajqtiH0umVg7HybX3wt1HjcrwKJovZURcIbbcDvzdH2bnYbF93T4OLXA0bIfuIp6M86x1iutFtXdpN3TTicINrmSXEE2Ydm51iMu77B08ZERjVaToya2F7vC+egfoPvibf7OLxE336a5tPCywavvNihQjL8sjgpDT5AAScjb3YqK/6VLeQ18Ggt8/ufINsYkb+9/Ji/3OcGFeflnDXq80vPUyF3u4iIylob6RSZenC38cXmQB05tRNxS1B6BXCjMRdy0v4pa7oKM2GA4ADKpNrr0RI9ed+peRFwmsclH test@ansible + agent_pool_profiles: + - name: default + count: 1 + vm_size: Standard_DS1_v2 + register: output + + - name: Assert the AKS instance is upgraded + assert: + that: + - output.changed + - output.kubernetes_version == '1.8.1' + + - name: Scale the AKS instance agent count from 1 to 2 + azure_rm_aks: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + location: eastus + dns_prefix: "aks{{ resource_group | hash('md5') | truncate(10, True, '') }}" + kubernetes_version: 1.8.1 + service_principal: + client_id: "{{ lookup('env', 'AZURE_CLIENT_ID') }}" + linux_profile: + admin_username: azureuser + ssh_key: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDSPmiqkvDH1/+MDAVDZT8381aYqp73Odz8cnD5hegNhqtXajqtiH0umVg7HybX3wt1HjcrwKJovZURcIbbcDvzdH2bnYbF93T4OLXA0bIfuIp6M86x1iutFtXdpN3TTicINrmSXEE2Ydm51iMu77B08ZERjVaToya2F7vC+egfoPvibf7OLxE336a5tPCywavvNihQjL8sjgpDT5AAScjb3YqK/6VLeQ18Ggt8/ufINsYkb+9/Ji/3OcGFeflnDXq80vPUyF3u4iIylob6RSZenC38cXmQB05tRNxS1B6BXCjMRdy0v4pa7oKM2GA4ADKpNrr0RI9ed+peRFwmsclH test@ansible + agent_pool_profiles: + - name: default + count: 2 + vm_size: Standard_DS1_v2 + register: output + + - name: Assert the AKS instance is well scaled + assert: + that: + - output.changed + - output.agent_pool_profiles[0].count == 2 + + - name: Delete the AKS instance + azure_rm_aks: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + state: absent + register: output + + - name: Assert the AKS instance is well deleted + assert: + that: + - output.changed + + - name: Delete the AKS instance (idempotent) + azure_rm_aks: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + state: absent + register: output + + - name: Assert idempotent + assert: + that: + - not output.changed + + - name: Get AKS fact + azure_rm_aks_facts: + name: "aks{{ resource_group | hash('md5') | truncate(7, True, '') }}" + resource_group: "{{ resource_group }}" + register: fact + + - name: Assert fact returns empty + assert: + that: + - "fact.aks | length == 0" diff --git a/test/runner/requirements/integration.cloud.azure.txt b/test/runner/requirements/integration.cloud.azure.txt index e3345c6787..802cb5d0d9 100644 --- a/test/runner/requirements/integration.cloud.azure.txt +++ b/test/runner/requirements/integration.cloud.azure.txt @@ -13,7 +13,7 @@ azure-mgmt-keyvault>=0.40.0,<0.41 azure-mgmt-batch>=4.1.0,<5 azure-mgmt-sql>=0.7.1,<0.8 azure-mgmt-web>=0.32.0,<0.33 -azure-mgmt-containerservice>=2.0.0,<3.0.0 +azure-mgmt-containerservice>=3.0.1 azure-mgmt-containerregistry>=1.0.1 azure-mgmt-rdbms>=0.2.0rc1,<0.3.0 azure-mgmt-containerinstance>=0.3.1,<0.4.0