From 47467814236bc79bea724044215254eb164c9084 Mon Sep 17 00:00:00 2001 From: Alex Stephen Date: Wed, 16 Jan 2019 11:16:51 -0800 Subject: [PATCH] New GCP module: gcp_bigquery_dataset (#50676) --- .../cloud/google/gcp_bigquery_dataset.py | 555 ++++++++++++++++++ .../targets/gcp_bigquery_dataset/aliases | 2 + .../gcp_bigquery_dataset/defaults/main.yml | 3 + .../gcp_bigquery_dataset/meta/main.yml | 0 .../gcp_bigquery_dataset/tasks/main.yml | 112 ++++ 5 files changed, 672 insertions(+) create mode 100644 lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py create mode 100644 test/integration/targets/gcp_bigquery_dataset/aliases create mode 100644 test/integration/targets/gcp_bigquery_dataset/defaults/main.yml create mode 100644 test/integration/targets/gcp_bigquery_dataset/meta/main.yml create mode 100644 test/integration/targets/gcp_bigquery_dataset/tasks/main.yml diff --git a/lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py b/lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py new file mode 100644 index 0000000000..cd6eb1e9e3 --- /dev/null +++ b/lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py @@ -0,0 +1,555 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- +# +# Copyright (C) 2017 Google +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) +# ---------------------------------------------------------------------------- +# +# *** AUTO GENERATED CODE *** AUTO GENERATED CODE *** +# +# ---------------------------------------------------------------------------- +# +# This file is automatically generated by Magic Modules and manual +# changes will be clobbered when the file is regenerated. +# +# Please read more about how to change this file at +# https://www.github.com/GoogleCloudPlatform/magic-modules +# +# ---------------------------------------------------------------------------- + +from __future__ import absolute_import, division, print_function +__metaclass__ = type + +################################################################################ +# Documentation +################################################################################ + +ANSIBLE_METADATA = {'metadata_version': '1.1', + 'status': ["preview"], + 'supported_by': 'community'} + +DOCUMENTATION = ''' +--- +module: gcp_bigquery_dataset +description: +- Datasets allow you to organize and control access to your tables. +short_description: Creates a GCP Dataset +version_added: 2.8 +author: Google Inc. (@googlecloudplatform) +requirements: +- python >= 2.6 +- requests >= 2.18.4 +- google-auth >= 1.3.0 +options: + state: + description: + - Whether the given object should exist in GCP + choices: + - present + - absent + default: present + name: + description: + - Dataset name. + required: false + access: + description: + - Access controls on the bucket. + required: false + suboptions: + domain: + description: + - A domain to grant access to. Any users signed in with the domain specified + will be granted the specified access . + required: false + group_by_email: + description: + - An email address of a Google Group to grant access to. + required: false + role: + description: + - Describes the rights granted to the user specified by the other member of + the access object . + required: false + choices: + - READER + - WRITER + - OWNER + special_group: + description: + - A special group to grant access to. + required: false + user_by_email: + description: + - 'An email address of a user to grant access to. For example: fred@example.com + .' + required: false + view: + description: + - A view from a different dataset to grant access to. Queries executed against + that view will have read access to tables in this dataset. The role field + is not required when this field is set. If that view is updated by any user, + access to the view needs to be granted again via an update operation. + required: false + suboptions: + dataset_id: + description: + - The ID of the dataset containing this table. + required: true + project_id: + description: + - The ID of the project containing this table. + required: true + table_id: + description: + - The ID of the table. The ID must contain only letters (a-z, A-Z), numbers + (0-9), or underscores. The maximum length is 1,024 characters. + required: true + dataset_reference: + description: + - A reference that identifies the dataset. + required: true + suboptions: + dataset_id: + description: + - A unique ID for this dataset, without the project name. The ID must contain + only letters (a-z, A-Z), numbers (0-9), or underscores. The maximum length + is 1,024 characters. + required: true + project_id: + description: + - The ID of the project containing this dataset. + required: false + default_table_expiration_ms: + description: + - The default lifetime of all tables in the dataset, in milliseconds . + required: false + description: + description: + - A user-friendly description of the dataset. + required: false + friendly_name: + description: + - A descriptive name for the dataset. + required: false + labels: + description: + - The labels associated with this dataset. You can use these to organize and group + your datasets . + required: false + location: + description: + - The geographic location where the dataset should reside. Possible values include + EU and US. The default value is US. + required: false + default: US +extends_documentation_fragment: gcp +''' + +EXAMPLES = ''' +- name: create a dataset + gcp_bigquery_dataset: + name: my_example_dataset + dataset_reference: + dataset_id: my_example_dataset + project: "test_project" + auth_kind: "serviceaccount" + service_account_file: "/tmp/auth.pem" + state: present +''' + +RETURN = ''' +name: + description: + - Dataset name. + returned: success + type: str +access: + description: + - Access controls on the bucket. + returned: success + type: complex + contains: + domain: + description: + - A domain to grant access to. Any users signed in with the domain specified + will be granted the specified access . + returned: success + type: str + groupByEmail: + description: + - An email address of a Google Group to grant access to. + returned: success + type: str + role: + description: + - Describes the rights granted to the user specified by the other member of + the access object . + returned: success + type: str + specialGroup: + description: + - A special group to grant access to. + returned: success + type: str + userByEmail: + description: + - 'An email address of a user to grant access to. For example: fred@example.com + .' + returned: success + type: str + view: + description: + - A view from a different dataset to grant access to. Queries executed against + that view will have read access to tables in this dataset. The role field + is not required when this field is set. If that view is updated by any user, + access to the view needs to be granted again via an update operation. + returned: success + type: complex + contains: + datasetId: + description: + - The ID of the dataset containing this table. + returned: success + type: str + projectId: + description: + - The ID of the project containing this table. + returned: success + type: str + tableId: + description: + - The ID of the table. The ID must contain only letters (a-z, A-Z), numbers + (0-9), or underscores. The maximum length is 1,024 characters. + returned: success + type: str +creationTime: + description: + - The time when this dataset was created, in milliseconds since the epoch. + returned: success + type: int +datasetReference: + description: + - A reference that identifies the dataset. + returned: success + type: complex + contains: + datasetId: + description: + - A unique ID for this dataset, without the project name. The ID must contain + only letters (a-z, A-Z), numbers (0-9), or underscores. The maximum length + is 1,024 characters. + returned: success + type: str + projectId: + description: + - The ID of the project containing this dataset. + returned: success + type: str +defaultTableExpirationMs: + description: + - The default lifetime of all tables in the dataset, in milliseconds . + returned: success + type: int +description: + description: + - A user-friendly description of the dataset. + returned: success + type: str +friendlyName: + description: + - A descriptive name for the dataset. + returned: success + type: str +id: + description: + - The fully-qualified unique name of the dataset in the format projectId:datasetId. + The dataset name without the project name is given in the datasetId field . + returned: success + type: str +labels: + description: + - The labels associated with this dataset. You can use these to organize and group + your datasets . + returned: success + type: dict +lastModifiedTime: + description: + - The date when this dataset or any of its tables was last modified, in milliseconds + since the epoch. + returned: success + type: int +location: + description: + - The geographic location where the dataset should reside. Possible values include + EU and US. The default value is US. + returned: success + type: str +''' + +################################################################################ +# Imports +################################################################################ + +from ansible.module_utils.gcp_utils import navigate_hash, GcpSession, GcpModule, GcpRequest, remove_nones_from_dict, replace_resource_dict +import json + +################################################################################ +# Main +################################################################################ + + +def main(): + """Main function""" + + module = GcpModule( + argument_spec=dict( + state=dict(default='present', choices=['present', 'absent'], type='str'), + name=dict(type='str'), + access=dict(type='list', elements='dict', options=dict( + domain=dict(type='str'), + group_by_email=dict(type='str'), + role=dict(type='str', choices=['READER', 'WRITER', 'OWNER']), + special_group=dict(type='str'), + user_by_email=dict(type='str'), + view=dict(type='dict', options=dict( + dataset_id=dict(required=True, type='str'), + project_id=dict(required=True, type='str'), + table_id=dict(required=True, type='str') + )) + )), + dataset_reference=dict(required=True, type='dict', options=dict( + dataset_id=dict(required=True, type='str'), + project_id=dict(type='str') + )), + default_table_expiration_ms=dict(type='int'), + description=dict(type='str'), + friendly_name=dict(type='str'), + labels=dict(type='dict'), + location=dict(default='US', type='str') + ) + ) + + if not module.params['scopes']: + module.params['scopes'] = ['https://www.googleapis.com/auth/bigquery'] + + state = module.params['state'] + kind = 'bigquery#dataset' + + fetch = fetch_resource(module, self_link(module), kind) + changed = False + + if fetch: + if state == 'present': + if is_different(module, fetch): + update(module, self_link(module), kind) + fetch = fetch_resource(module, self_link(module), kind) + changed = True + else: + delete(module, self_link(module), kind) + fetch = {} + changed = True + else: + if state == 'present': + fetch = create(module, collection(module), kind) + changed = True + else: + fetch = {} + + fetch.update({'changed': changed}) + + module.exit_json(**fetch) + + +def create(module, link, kind): + auth = GcpSession(module, 'bigquery') + return return_if_object(module, auth.post(link, resource_to_request(module)), kind) + + +def update(module, link, kind): + auth = GcpSession(module, 'bigquery') + return return_if_object(module, auth.put(link, resource_to_request(module)), kind) + + +def delete(module, link, kind): + auth = GcpSession(module, 'bigquery') + return return_if_object(module, auth.delete(link), kind) + + +def resource_to_request(module): + request = { + u'kind': 'bigquery#dataset', + u'name': module.params.get('name'), + u'access': DatasetAccessArray(module.params.get('access', []), module).to_request(), + u'datasetReference': DatasetDatasetreference(module.params.get('dataset_reference', {}), module).to_request(), + u'defaultTableExpirationMs': module.params.get('default_table_expiration_ms'), + u'description': module.params.get('description'), + u'friendlyName': module.params.get('friendly_name'), + u'labels': module.params.get('labels'), + u'location': module.params.get('location') + } + return_vals = {} + for k, v in request.items(): + if v or v is False: + return_vals[k] = v + + return return_vals + + +def fetch_resource(module, link, kind, allow_not_found=True): + auth = GcpSession(module, 'bigquery') + return return_if_object(module, auth.get(link), kind, allow_not_found) + + +def self_link(module): + return "https://www.googleapis.com/bigquery/v2/projects/{project}/datasets/{name}".format(**module.params) + + +def collection(module): + return "https://www.googleapis.com/bigquery/v2/projects/{project}/datasets".format(**module.params) + + +def return_if_object(module, response, kind, allow_not_found=False): + # If not found, return nothing. + if allow_not_found and response.status_code == 404: + return None + + # If no content, return nothing. + if response.status_code == 204: + return None + + try: + module.raise_for_status(response) + result = response.json() + except getattr(json.decoder, 'JSONDecodeError', ValueError) as inst: + module.fail_json(msg="Invalid JSON response with error: %s" % inst) + + if navigate_hash(result, ['error', 'errors']): + module.fail_json(msg=navigate_hash(result, ['error', 'errors'])) + + return result + + +def is_different(module, response): + request = resource_to_request(module) + response = response_to_hash(module, response) + + # Remove all output-only from response. + response_vals = {} + for k, v in response.items(): + if k in request: + response_vals[k] = v + + request_vals = {} + for k, v in request.items(): + if k in response: + request_vals[k] = v + + return GcpRequest(request_vals) != GcpRequest(response_vals) + + +# Remove unnecessary properties from the response. +# This is for doing comparisons with Ansible's current parameters. +def response_to_hash(module, response): + return { + u'name': response.get(u'name'), + u'access': DatasetAccessArray(response.get(u'access', []), module).from_response(), + u'creationTime': response.get(u'creationTime'), + u'datasetReference': DatasetDatasetreference(response.get(u'datasetReference', {}), module).from_response(), + u'defaultTableExpirationMs': response.get(u'defaultTableExpirationMs'), + u'description': response.get(u'description'), + u'friendlyName': response.get(u'friendlyName'), + u'id': response.get(u'id'), + u'labels': response.get(u'labels'), + u'lastModifiedTime': response.get(u'lastModifiedTime'), + u'location': response.get(u'location') + } + + +class DatasetAccessArray(object): + def __init__(self, request, module): + self.module = module + if request: + self.request = request + else: + self.request = [] + + def to_request(self): + items = [] + for item in self.request: + items.append(self._request_for_item(item)) + return items + + def from_response(self): + items = [] + for item in self.request: + items.append(self._response_from_item(item)) + return items + + def _request_for_item(self, item): + return remove_nones_from_dict({ + u'domain': item.get('domain'), + u'groupByEmail': item.get('group_by_email'), + u'role': item.get('role'), + u'specialGroup': item.get('special_group'), + u'userByEmail': item.get('user_by_email'), + u'view': DatasetView(item.get('view', {}), self.module).to_request() + }) + + def _response_from_item(self, item): + return remove_nones_from_dict({ + u'domain': item.get(u'domain'), + u'groupByEmail': item.get(u'groupByEmail'), + u'role': item.get(u'role'), + u'specialGroup': item.get(u'specialGroup'), + u'userByEmail': item.get(u'userByEmail'), + u'view': DatasetView(item.get(u'view', {}), self.module).from_response() + }) + + +class DatasetView(object): + def __init__(self, request, module): + self.module = module + if request: + self.request = request + else: + self.request = {} + + def to_request(self): + return remove_nones_from_dict({ + u'datasetId': self.request.get('dataset_id'), + u'projectId': self.request.get('project_id'), + u'tableId': self.request.get('table_id') + }) + + def from_response(self): + return remove_nones_from_dict({ + u'datasetId': self.request.get(u'datasetId'), + u'projectId': self.request.get(u'projectId'), + u'tableId': self.request.get(u'tableId') + }) + + +class DatasetDatasetreference(object): + def __init__(self, request, module): + self.module = module + if request: + self.request = request + else: + self.request = {} + + def to_request(self): + return remove_nones_from_dict({ + u'datasetId': self.request.get('dataset_id'), + u'projectId': self.request.get('project_id') + }) + + def from_response(self): + return remove_nones_from_dict({ + u'datasetId': self.request.get(u'datasetId'), + u'projectId': self.request.get(u'projectId') + }) + + +if __name__ == '__main__': + main() diff --git a/test/integration/targets/gcp_bigquery_dataset/aliases b/test/integration/targets/gcp_bigquery_dataset/aliases new file mode 100644 index 0000000000..9812f019ca --- /dev/null +++ b/test/integration/targets/gcp_bigquery_dataset/aliases @@ -0,0 +1,2 @@ +cloud/gcp +unsupported diff --git a/test/integration/targets/gcp_bigquery_dataset/defaults/main.yml b/test/integration/targets/gcp_bigquery_dataset/defaults/main.yml new file mode 100644 index 0000000000..aa87a2a8e0 --- /dev/null +++ b/test/integration/targets/gcp_bigquery_dataset/defaults/main.yml @@ -0,0 +1,3 @@ +--- +# defaults file +resource_name: '{{resource_prefix}}' diff --git a/test/integration/targets/gcp_bigquery_dataset/meta/main.yml b/test/integration/targets/gcp_bigquery_dataset/meta/main.yml new file mode 100644 index 0000000000..e69de29bb2 diff --git a/test/integration/targets/gcp_bigquery_dataset/tasks/main.yml b/test/integration/targets/gcp_bigquery_dataset/tasks/main.yml new file mode 100644 index 0000000000..db480b71f6 --- /dev/null +++ b/test/integration/targets/gcp_bigquery_dataset/tasks/main.yml @@ -0,0 +1,112 @@ +--- +# ---------------------------------------------------------------------------- +# +# *** AUTO GENERATED CODE *** AUTO GENERATED CODE *** +# +# ---------------------------------------------------------------------------- +# +# This file is automatically generated by Magic Modules and manual +# changes will be clobbered when the file is regenerated. +# +# Please read more about how to change this file at +# https://www.github.com/GoogleCloudPlatform/magic-modules +# +# ---------------------------------------------------------------------------- +# Pre-test setup +- name: delete a dataset + gcp_bigquery_dataset: + name: my_example_dataset + dataset_reference: + dataset_id: my_example_dataset + project: "{{ gcp_project }}" + auth_kind: "{{ gcp_cred_kind }}" + service_account_file: "{{ gcp_cred_file }}" + state: absent +#---------------------------------------------------------- +- name: create a dataset + gcp_bigquery_dataset: + name: my_example_dataset + dataset_reference: + dataset_id: my_example_dataset + project: "{{ gcp_project }}" + auth_kind: "{{ gcp_cred_kind }}" + service_account_file: "{{ gcp_cred_file }}" + state: present + register: result +- name: assert changed is true + assert: + that: + - result.changed == true + - "result.kind == 'bigquery#dataset'" +- name: verify that dataset was created + gcp_bigquery_dataset_facts: + project: "{{ gcp_project }}" + auth_kind: "{{ gcp_cred_kind }}" + service_account_file: "{{ gcp_cred_file }}" + scopes: + - https://www.googleapis.com/auth/bigquery + register: results +- name: verify that command succeeded + assert: + that: + - results['items'] | length >= 1 +# ---------------------------------------------------------------------------- +- name: create a dataset that already exists + gcp_bigquery_dataset: + name: my_example_dataset + dataset_reference: + dataset_id: my_example_dataset + project: "{{ gcp_project }}" + auth_kind: "{{ gcp_cred_kind }}" + service_account_file: "{{ gcp_cred_file }}" + state: present + register: result +- name: assert changed is false + assert: + that: + - result.changed == false + - "result.kind == 'bigquery#dataset'" +#---------------------------------------------------------- +- name: delete a dataset + gcp_bigquery_dataset: + name: my_example_dataset + dataset_reference: + dataset_id: my_example_dataset + project: "{{ gcp_project }}" + auth_kind: "{{ gcp_cred_kind }}" + service_account_file: "{{ gcp_cred_file }}" + state: absent + register: result +- name: assert changed is true + assert: + that: + - result.changed == true + - result.has_key('kind') == False +- name: verify that dataset was deleted + gcp_bigquery_dataset_facts: + project: "{{ gcp_project }}" + auth_kind: "{{ gcp_cred_kind }}" + service_account_file: "{{ gcp_cred_file }}" + scopes: + - https://www.googleapis.com/auth/bigquery + register: results +- name: verify that command succeeded + assert: + that: + - results['items'] | length == 0 +# ---------------------------------------------------------------------------- +- name: delete a dataset that does not exist + gcp_bigquery_dataset: + name: my_example_dataset + dataset_reference: + dataset_id: my_example_dataset + project: "{{ gcp_project }}" + auth_kind: "{{ gcp_cred_kind }}" + service_account_file: "{{ gcp_cred_file }}" + state: absent + register: result +- name: assert changed is false + assert: + that: + - result.changed == false + - result.has_key('kind') == False