summaryrefslogtreecommitdiff
path: root/lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py
diff options
context:
space:
mode:
Diffstat (limited to 'lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py')
-rw-r--r--lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py745
1 files changed, 0 insertions, 745 deletions
diff --git a/lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py b/lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py
deleted file mode 100644
index db27ce2b42..0000000000
--- a/lib/ansible/modules/cloud/google/gcp_bigquery_dataset.py
+++ /dev/null
@@ -1,745 +0,0 @@
-#!/usr/bin/python
-# -*- coding: utf-8 -*-
-#
-# Copyright (C) 2017 Google
-# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
-# ----------------------------------------------------------------------------
-#
-# *** AUTO GENERATED CODE *** AUTO GENERATED CODE ***
-#
-# ----------------------------------------------------------------------------
-#
-# This file is automatically generated by Magic Modules and manual
-# changes will be clobbered when the file is regenerated.
-#
-# Please read more about how to change this file at
-# https://www.github.com/GoogleCloudPlatform/magic-modules
-#
-# ----------------------------------------------------------------------------
-
-from __future__ import absolute_import, division, print_function
-
-__metaclass__ = type
-
-################################################################################
-# Documentation
-################################################################################
-
-ANSIBLE_METADATA = {'metadata_version': '1.1', 'status': ["preview"], 'supported_by': 'community'}
-
-DOCUMENTATION = '''
----
-module: gcp_bigquery_dataset
-description:
-- Datasets allow you to organize and control access to your tables.
-short_description: Creates a GCP Dataset
-version_added: '2.8'
-author: Google Inc. (@googlecloudplatform)
-requirements:
-- python >= 2.6
-- requests >= 2.18.4
-- google-auth >= 1.3.0
-options:
- state:
- description:
- - Whether the given object should exist in GCP
- choices:
- - present
- - absent
- default: present
- type: str
- name:
- description:
- - Dataset name.
- required: false
- type: str
- access:
- description:
- - An array of objects that define dataset access for one or more entities.
- required: false
- type: list
- suboptions:
- domain:
- description:
- - A domain to grant access to. Any users signed in with the domain specified
- will be granted the specified access .
- required: false
- type: str
- group_by_email:
- description:
- - An email address of a Google Group to grant access to.
- required: false
- type: str
- role:
- description:
- - Describes the rights granted to the user specified by the other member of
- the access object. Primitive, Predefined and custom roles are supported.
- Predefined roles that have equivalent primitive roles are swapped by the
- API to their Primitive counterparts, and will show a diff post-create. See
- [official docs](U(https://cloud.google.com/bigquery/docs/access-control)).
- required: false
- type: str
- special_group:
- description:
- - A special group to grant access to.
- - 'Possible values include: * `projectOwners`: Owners of the enclosing project.'
- - "* `projectReaders`: Readers of the enclosing project."
- - "* `projectWriters`: Writers of the enclosing project."
- - "* `allAuthenticatedUsers`: All authenticated BigQuery users. ."
- required: false
- type: str
- user_by_email:
- description:
- - 'An email address of a user to grant access to. For example: fred@example.com
- .'
- required: false
- type: str
- view:
- description:
- - A view from a different dataset to grant access to. Queries executed against
- that view will have read access to tables in this dataset. The role field
- is not required when this field is set. If that view is updated by any user,
- access to the view needs to be granted again via an update operation.
- required: false
- type: dict
- suboptions:
- dataset_id:
- description:
- - The ID of the dataset containing this table.
- required: true
- type: str
- project_id:
- description:
- - The ID of the project containing this table.
- required: true
- type: str
- table_id:
- description:
- - The ID of the table. The ID must contain only letters (a-z, A-Z), numbers
- (0-9), or underscores. The maximum length is 1,024 characters.
- required: true
- type: str
- dataset_reference:
- description:
- - A reference that identifies the dataset.
- required: true
- type: dict
- suboptions:
- dataset_id:
- description:
- - A unique ID for this dataset, without the project name. The ID must contain
- only letters (a-z, A-Z), numbers (0-9), or underscores. The maximum length
- is 1,024 characters.
- required: true
- type: str
- project_id:
- description:
- - The ID of the project containing this dataset.
- required: false
- type: str
- default_table_expiration_ms:
- description:
- - The default lifetime of all tables in the dataset, in milliseconds.
- - The minimum value is 3600000 milliseconds (one hour).
- - Once this property is set, all newly-created tables in the dataset will have
- an `expirationTime` property set to the creation time plus the value in this
- property, and changing the value will only affect new tables, not existing ones.
- When the `expirationTime` for a given table is reached, that table will be deleted
- automatically.
- - If a table's `expirationTime` is modified or removed before the table expires,
- or if you provide an explicit `expirationTime` when creating a table, that value
- takes precedence over the default expiration time indicated by this property.
- required: false
- type: int
- default_partition_expiration_ms:
- description:
- - The default partition expiration for all partitioned tables in the dataset,
- in milliseconds.
- - Once this property is set, all newly-created partitioned tables in the dataset
- will have an `expirationMs` property in the `timePartitioning` settings set
- to this value, and changing the value will only affect new tables, not existing
- ones. The storage in a partition will have an expiration time of its partition
- time plus this value.
- - 'Setting this property overrides the use of `defaultTableExpirationMs` for partitioned
- tables: only one of `defaultTableExpirationMs` and `defaultPartitionExpirationMs`
- will be used for any new partitioned table. If you provide an explicit `timePartitioning.expirationMs`
- when creating or updating a partitioned table, that value takes precedence over
- the default partition expiration time indicated by this property.'
- required: false
- type: int
- version_added: '2.9'
- description:
- description:
- - A user-friendly description of the dataset.
- required: false
- type: str
- friendly_name:
- description:
- - A descriptive name for the dataset.
- required: false
- type: str
- labels:
- description:
- - The labels associated with this dataset. You can use these to organize and group
- your datasets .
- required: false
- type: dict
- location:
- description:
- - The geographic location where the dataset should reside.
- - See [official docs](U(https://cloud.google.com/bigquery/docs/dataset-locations)).
- - There are two types of locations, regional or multi-regional. A regional location
- is a specific geographic place, such as Tokyo, and a multi-regional location
- is a large geographic area, such as the United States, that contains at least
- two geographic places.
- - 'Possible regional values include: `asia-east1`, `asia-northeast1`, `asia-southeast1`,
- `australia-southeast1`, `europe-north1`, `europe-west2` and `us-east4`.'
- - 'Possible multi-regional values: `EU` and `US`.'
- - The default value is multi-regional location `US`.
- - Changing this forces a new resource to be created.
- required: false
- default: US
- type: str
- default_encryption_configuration:
- description:
- - The default encryption key for all tables in the dataset. Once this property
- is set, all newly-created partitioned tables in the dataset will have encryption
- key set to this value, unless table creation request (or query) overrides the
- key.
- required: false
- type: dict
- version_added: '2.10'
- suboptions:
- kms_key_name:
- description:
- - Describes the Cloud KMS encryption key that will be used to protect destination
- BigQuery table. The BigQuery Service Account associated with your project
- requires access to this encryption key.
- required: true
- type: str
- project:
- description:
- - The Google Cloud Platform project to use.
- type: str
- auth_kind:
- description:
- - The type of credential used.
- type: str
- required: true
- choices:
- - application
- - machineaccount
- - serviceaccount
- service_account_contents:
- description:
- - The contents of a Service Account JSON file, either in a dictionary or as a
- JSON string that represents it.
- type: jsonarg
- service_account_file:
- description:
- - The path of a Service Account JSON file if serviceaccount is selected as type.
- type: path
- service_account_email:
- description:
- - An optional service account email address if machineaccount is selected and
- the user does not wish to use the default email.
- type: str
- scopes:
- description:
- - Array of scopes to be used
- type: list
- env_type:
- description:
- - Specifies which Ansible environment you're running this module within.
- - This should not be set unless you know what you're doing.
- - This only alters the User Agent string for any API requests.
- type: str
-'''
-
-EXAMPLES = '''
-- name: create a dataset
- gcp_bigquery_dataset:
- name: my_example_dataset
- dataset_reference:
- dataset_id: my_example_dataset
- project: test_project
- auth_kind: serviceaccount
- service_account_file: "/tmp/auth.pem"
- state: present
-'''
-
-RETURN = '''
-name:
- description:
- - Dataset name.
- returned: success
- type: str
-access:
- description:
- - An array of objects that define dataset access for one or more entities.
- returned: success
- type: complex
- contains:
- domain:
- description:
- - A domain to grant access to. Any users signed in with the domain specified
- will be granted the specified access .
- returned: success
- type: str
- groupByEmail:
- description:
- - An email address of a Google Group to grant access to.
- returned: success
- type: str
- role:
- description:
- - Describes the rights granted to the user specified by the other member of
- the access object. Primitive, Predefined and custom roles are supported. Predefined
- roles that have equivalent primitive roles are swapped by the API to their
- Primitive counterparts, and will show a diff post-create. See [official docs](U(https://cloud.google.com/bigquery/docs/access-control)).
- returned: success
- type: str
- specialGroup:
- description:
- - A special group to grant access to.
- - 'Possible values include: * `projectOwners`: Owners of the enclosing project.'
- - "* `projectReaders`: Readers of the enclosing project."
- - "* `projectWriters`: Writers of the enclosing project."
- - "* `allAuthenticatedUsers`: All authenticated BigQuery users. ."
- returned: success
- type: str
- userByEmail:
- description:
- - 'An email address of a user to grant access to. For example: fred@example.com
- .'
- returned: success
- type: str
- view:
- description:
- - A view from a different dataset to grant access to. Queries executed against
- that view will have read access to tables in this dataset. The role field
- is not required when this field is set. If that view is updated by any user,
- access to the view needs to be granted again via an update operation.
- returned: success
- type: complex
- contains:
- datasetId:
- description:
- - The ID of the dataset containing this table.
- returned: success
- type: str
- projectId:
- description:
- - The ID of the project containing this table.
- returned: success
- type: str
- tableId:
- description:
- - The ID of the table. The ID must contain only letters (a-z, A-Z), numbers
- (0-9), or underscores. The maximum length is 1,024 characters.
- returned: success
- type: str
-creationTime:
- description:
- - The time when this dataset was created, in milliseconds since the epoch.
- returned: success
- type: int
-datasetReference:
- description:
- - A reference that identifies the dataset.
- returned: success
- type: complex
- contains:
- datasetId:
- description:
- - A unique ID for this dataset, without the project name. The ID must contain
- only letters (a-z, A-Z), numbers (0-9), or underscores. The maximum length
- is 1,024 characters.
- returned: success
- type: str
- projectId:
- description:
- - The ID of the project containing this dataset.
- returned: success
- type: str
-defaultTableExpirationMs:
- description:
- - The default lifetime of all tables in the dataset, in milliseconds.
- - The minimum value is 3600000 milliseconds (one hour).
- - Once this property is set, all newly-created tables in the dataset will have an
- `expirationTime` property set to the creation time plus the value in this property,
- and changing the value will only affect new tables, not existing ones. When the
- `expirationTime` for a given table is reached, that table will be deleted automatically.
- - If a table's `expirationTime` is modified or removed before the table expires,
- or if you provide an explicit `expirationTime` when creating a table, that value
- takes precedence over the default expiration time indicated by this property.
- returned: success
- type: int
-defaultPartitionExpirationMs:
- description:
- - The default partition expiration for all partitioned tables in the dataset, in
- milliseconds.
- - Once this property is set, all newly-created partitioned tables in the dataset
- will have an `expirationMs` property in the `timePartitioning` settings set to
- this value, and changing the value will only affect new tables, not existing ones.
- The storage in a partition will have an expiration time of its partition time
- plus this value.
- - 'Setting this property overrides the use of `defaultTableExpirationMs` for partitioned
- tables: only one of `defaultTableExpirationMs` and `defaultPartitionExpirationMs`
- will be used for any new partitioned table. If you provide an explicit `timePartitioning.expirationMs`
- when creating or updating a partitioned table, that value takes precedence over
- the default partition expiration time indicated by this property.'
- returned: success
- type: int
-description:
- description:
- - A user-friendly description of the dataset.
- returned: success
- type: str
-etag:
- description:
- - A hash of the resource.
- returned: success
- type: str
-friendlyName:
- description:
- - A descriptive name for the dataset.
- returned: success
- type: str
-id:
- description:
- - The fully-qualified unique name of the dataset in the format projectId:datasetId.
- The dataset name without the project name is given in the datasetId field .
- returned: success
- type: str
-labels:
- description:
- - The labels associated with this dataset. You can use these to organize and group
- your datasets .
- returned: success
- type: dict
-lastModifiedTime:
- description:
- - The date when this dataset or any of its tables was last modified, in milliseconds
- since the epoch.
- returned: success
- type: int
-location:
- description:
- - The geographic location where the dataset should reside.
- - See [official docs](U(https://cloud.google.com/bigquery/docs/dataset-locations)).
- - There are two types of locations, regional or multi-regional. A regional location
- is a specific geographic place, such as Tokyo, and a multi-regional location is
- a large geographic area, such as the United States, that contains at least two
- geographic places.
- - 'Possible regional values include: `asia-east1`, `asia-northeast1`, `asia-southeast1`,
- `australia-southeast1`, `europe-north1`, `europe-west2` and `us-east4`.'
- - 'Possible multi-regional values: `EU` and `US`.'
- - The default value is multi-regional location `US`.
- - Changing this forces a new resource to be created.
- returned: success
- type: str
-defaultEncryptionConfiguration:
- description:
- - The default encryption key for all tables in the dataset. Once this property is
- set, all newly-created partitioned tables in the dataset will have encryption
- key set to this value, unless table creation request (or query) overrides the
- key.
- returned: success
- type: complex
- contains:
- kmsKeyName:
- description:
- - Describes the Cloud KMS encryption key that will be used to protect destination
- BigQuery table. The BigQuery Service Account associated with your project
- requires access to this encryption key.
- returned: success
- type: str
-'''
-
-################################################################################
-# Imports
-################################################################################
-
-from ansible.module_utils.gcp_utils import navigate_hash, GcpSession, GcpModule, GcpRequest, remove_nones_from_dict, replace_resource_dict
-import json
-
-################################################################################
-# Main
-################################################################################
-
-
-def main():
- """Main function"""
-
- module = GcpModule(
- argument_spec=dict(
- state=dict(default='present', choices=['present', 'absent'], type='str'),
- name=dict(type='str'),
- access=dict(
- type='list',
- elements='dict',
- options=dict(
- domain=dict(type='str'),
- group_by_email=dict(type='str'),
- role=dict(type='str'),
- special_group=dict(type='str'),
- user_by_email=dict(type='str'),
- view=dict(
- type='dict',
- options=dict(
- dataset_id=dict(required=True, type='str'), project_id=dict(required=True, type='str'), table_id=dict(required=True, type='str')
- ),
- ),
- ),
- ),
- dataset_reference=dict(required=True, type='dict', options=dict(dataset_id=dict(required=True, type='str'), project_id=dict(type='str'))),
- default_table_expiration_ms=dict(type='int'),
- default_partition_expiration_ms=dict(type='int'),
- description=dict(type='str'),
- friendly_name=dict(type='str'),
- labels=dict(type='dict'),
- location=dict(default='US', type='str'),
- default_encryption_configuration=dict(type='dict', options=dict(kms_key_name=dict(required=True, type='str'))),
- )
- )
-
- if not module.params['scopes']:
- module.params['scopes'] = ['https://www.googleapis.com/auth/bigquery']
-
- state = module.params['state']
- kind = 'bigquery#dataset'
-
- fetch = fetch_resource(module, self_link(module), kind)
- changed = False
-
- if fetch:
- if state == 'present':
- if is_different(module, fetch):
- update(module, self_link(module), kind)
- fetch = fetch_resource(module, self_link(module), kind)
- changed = True
- else:
- delete(module, self_link(module), kind)
- fetch = {}
- changed = True
- else:
- if state == 'present':
- fetch = create(module, collection(module), kind)
- changed = True
- else:
- fetch = {}
-
- fetch.update({'changed': changed})
-
- module.exit_json(**fetch)
-
-
-def create(module, link, kind):
- auth = GcpSession(module, 'bigquery')
- return return_if_object(module, auth.post(link, resource_to_request(module)), kind)
-
-
-def update(module, link, kind):
- auth = GcpSession(module, 'bigquery')
- return return_if_object(module, auth.put(link, resource_to_request(module)), kind)
-
-
-def delete(module, link, kind):
- auth = GcpSession(module, 'bigquery')
- return return_if_object(module, auth.delete(link), kind)
-
-
-def resource_to_request(module):
- request = {
- u'kind': 'bigquery#dataset',
- u'name': module.params.get('name'),
- u'access': DatasetAccessArray(module.params.get('access', []), module).to_request(),
- u'datasetReference': DatasetDatasetreference(module.params.get('dataset_reference', {}), module).to_request(),
- u'defaultTableExpirationMs': module.params.get('default_table_expiration_ms'),
- u'defaultPartitionExpirationMs': module.params.get('default_partition_expiration_ms'),
- u'description': module.params.get('description'),
- u'friendlyName': module.params.get('friendly_name'),
- u'labels': module.params.get('labels'),
- u'location': module.params.get('location'),
- u'defaultEncryptionConfiguration': DatasetDefaultencryptionconfiguration(
- module.params.get('default_encryption_configuration', {}), module
- ).to_request(),
- }
- return_vals = {}
- for k, v in request.items():
- if v or v is False:
- return_vals[k] = v
-
- return return_vals
-
-
-def fetch_resource(module, link, kind, allow_not_found=True):
- auth = GcpSession(module, 'bigquery')
- return return_if_object(module, auth.get(link), kind, allow_not_found)
-
-
-def self_link(module):
- return "https://www.googleapis.com/bigquery/v2/projects/{project}/datasets/{name}".format(**module.params)
-
-
-def collection(module):
- return "https://www.googleapis.com/bigquery/v2/projects/{project}/datasets".format(**module.params)
-
-
-def return_if_object(module, response, kind, allow_not_found=False):
- # If not found, return nothing.
- if allow_not_found and response.status_code == 404:
- return None
-
- # If no content, return nothing.
- if response.status_code == 204:
- return None
-
- try:
- module.raise_for_status(response)
- result = response.json()
- except getattr(json.decoder, 'JSONDecodeError', ValueError):
- module.fail_json(msg="Invalid JSON response with error: %s" % response.text)
-
- if navigate_hash(result, ['error', 'errors']):
- module.fail_json(msg=navigate_hash(result, ['error', 'errors']))
-
- return result
-
-
-def is_different(module, response):
- request = resource_to_request(module)
- response = response_to_hash(module, response)
-
- # Remove all output-only from response.
- response_vals = {}
- for k, v in response.items():
- if k in request:
- response_vals[k] = v
-
- request_vals = {}
- for k, v in request.items():
- if k in response:
- request_vals[k] = v
-
- return GcpRequest(request_vals) != GcpRequest(response_vals)
-
-
-# Remove unnecessary properties from the response.
-# This is for doing comparisons with Ansible's current parameters.
-def response_to_hash(module, response):
- return {
- u'name': response.get(u'name'),
- u'access': DatasetAccessArray(response.get(u'access', []), module).from_response(),
- u'creationTime': response.get(u'creationTime'),
- u'datasetReference': DatasetDatasetreference(response.get(u'datasetReference', {}), module).from_response(),
- u'defaultTableExpirationMs': response.get(u'defaultTableExpirationMs'),
- u'defaultPartitionExpirationMs': response.get(u'defaultPartitionExpirationMs'),
- u'description': response.get(u'description'),
- u'etag': response.get(u'etag'),
- u'friendlyName': response.get(u'friendlyName'),
- u'id': response.get(u'id'),
- u'labels': response.get(u'labels'),
- u'lastModifiedTime': response.get(u'lastModifiedTime'),
- u'location': response.get(u'location'),
- u'defaultEncryptionConfiguration': DatasetDefaultencryptionconfiguration(response.get(u'defaultEncryptionConfiguration', {}), module).from_response(),
- }
-
-
-class DatasetAccessArray(object):
- def __init__(self, request, module):
- self.module = module
- if request:
- self.request = request
- else:
- self.request = []
-
- def to_request(self):
- items = []
- for item in self.request:
- items.append(self._request_for_item(item))
- return items
-
- def from_response(self):
- items = []
- for item in self.request:
- items.append(self._response_from_item(item))
- return items
-
- def _request_for_item(self, item):
- return remove_nones_from_dict(
- {
- u'domain': item.get('domain'),
- u'groupByEmail': item.get('group_by_email'),
- u'role': item.get('role'),
- u'specialGroup': item.get('special_group'),
- u'userByEmail': item.get('user_by_email'),
- u'view': DatasetView(item.get('view', {}), self.module).to_request(),
- }
- )
-
- def _response_from_item(self, item):
- return remove_nones_from_dict(
- {
- u'domain': item.get(u'domain'),
- u'groupByEmail': item.get(u'groupByEmail'),
- u'role': item.get(u'role'),
- u'specialGroup': item.get(u'specialGroup'),
- u'userByEmail': item.get(u'userByEmail'),
- u'view': DatasetView(item.get(u'view', {}), self.module).from_response(),
- }
- )
-
-
-class DatasetView(object):
- def __init__(self, request, module):
- self.module = module
- if request:
- self.request = request
- else:
- self.request = {}
-
- def to_request(self):
- return remove_nones_from_dict(
- {u'datasetId': self.request.get('dataset_id'), u'projectId': self.request.get('project_id'), u'tableId': self.request.get('table_id')}
- )
-
- def from_response(self):
- return remove_nones_from_dict(
- {u'datasetId': self.request.get(u'datasetId'), u'projectId': self.request.get(u'projectId'), u'tableId': self.request.get(u'tableId')}
- )
-
-
-class DatasetDatasetreference(object):
- def __init__(self, request, module):
- self.module = module
- if request:
- self.request = request
- else:
- self.request = {}
-
- def to_request(self):
- return remove_nones_from_dict({u'datasetId': self.request.get('dataset_id'), u'projectId': self.request.get('project_id')})
-
- def from_response(self):
- return remove_nones_from_dict({u'datasetId': self.request.get(u'datasetId'), u'projectId': self.request.get(u'projectId')})
-
-
-class DatasetDefaultencryptionconfiguration(object):
- def __init__(self, request, module):
- self.module = module
- if request:
- self.request = request
- else:
- self.request = {}
-
- def to_request(self):
- return remove_nones_from_dict({u'kmsKeyName': self.request.get('kms_key_name')})
-
- def from_response(self):
- return remove_nones_from_dict({u'kmsKeyName': self.request.get(u'kmsKeyName')})
-
-
-if __name__ == '__main__':
- main()