mirror of
https://github.com/ansible-collections/community.general.git
synced 2024-09-14 20:13:21 +02:00
510ad54062
Various docs improvements (#8664)
Various docs improvements.
(cherry picked from commit c0fd10e793
)
Co-authored-by: Felix Fontein <felix@fontein.de>
762 lines
32 KiB
Python
762 lines
32 KiB
Python
#!/usr/bin/python
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright (c) 2014, Chris Schmidt <chris.schmidt () contrastsecurity.com>
|
|
#
|
|
# Built using https://github.com/hamnis/useful-scripts/blob/master/python/download-maven-artifact
|
|
# as a reference and starting point.
|
|
#
|
|
# GNU General Public License v3.0+ (see LICENSES/GPL-3.0-or-later.txt or https://www.gnu.org/licenses/gpl-3.0.txt)
|
|
# SPDX-License-Identifier: GPL-3.0-or-later
|
|
from __future__ import absolute_import, division, print_function
|
|
__metaclass__ = type
|
|
|
|
DOCUMENTATION = r'''
|
|
---
|
|
module: maven_artifact
|
|
short_description: Downloads an Artifact from a Maven Repository
|
|
description:
|
|
- Downloads an artifact from a maven repository given the maven coordinates provided to the module.
|
|
- Can retrieve snapshots or release versions of the artifact and will resolve the latest available
|
|
version if one is not available.
|
|
author: "Chris Schmidt (@chrisisbeef)"
|
|
requirements:
|
|
- lxml
|
|
- boto if using a S3 repository (V(s3://...))
|
|
attributes:
|
|
check_mode:
|
|
support: none
|
|
diff_mode:
|
|
support: none
|
|
options:
|
|
group_id:
|
|
type: str
|
|
description:
|
|
- The Maven groupId coordinate.
|
|
required: true
|
|
artifact_id:
|
|
type: str
|
|
description:
|
|
- The maven artifactId coordinate.
|
|
required: true
|
|
version:
|
|
type: str
|
|
description:
|
|
- The maven version coordinate.
|
|
- Mutually exclusive with O(version_by_spec).
|
|
version_by_spec:
|
|
type: str
|
|
description:
|
|
- The maven dependency version ranges.
|
|
- See supported version ranges on U(https://cwiki.apache.org/confluence/display/MAVENOLD/Dependency+Mediation+and+Conflict+Resolution)
|
|
- The range type V((,1.0],[1.2,\)) and V((,1.1\),(1.1,\)) is not supported.
|
|
- Mutually exclusive with O(version).
|
|
version_added: '0.2.0'
|
|
classifier:
|
|
type: str
|
|
description:
|
|
- The maven classifier coordinate.
|
|
default: ''
|
|
extension:
|
|
type: str
|
|
description:
|
|
- The maven type/extension coordinate.
|
|
default: jar
|
|
repository_url:
|
|
type: str
|
|
description:
|
|
- The URL of the Maven Repository to download from.
|
|
- Use V(s3://...) if the repository is hosted on Amazon S3.
|
|
- Use V(file://...) if the repository is local.
|
|
default: https://repo1.maven.org/maven2
|
|
username:
|
|
type: str
|
|
description:
|
|
- The username to authenticate as to the Maven Repository. Use AWS secret key of the repository is hosted on S3.
|
|
aliases: [ "aws_secret_key" ]
|
|
password:
|
|
type: str
|
|
description:
|
|
- The password to authenticate with to the Maven Repository. Use AWS secret access key of the repository is hosted on S3.
|
|
aliases: [ "aws_secret_access_key" ]
|
|
headers:
|
|
description:
|
|
- Add custom HTTP headers to a request in hash/dict format.
|
|
type: dict
|
|
force_basic_auth:
|
|
description:
|
|
- httplib2, the library used by the uri module only sends authentication information when a webservice
|
|
responds to an initial request with a 401 status. Since some basic auth services do not properly
|
|
send a 401, logins will fail. This option forces the sending of the Basic authentication header
|
|
upon initial request.
|
|
default: false
|
|
type: bool
|
|
version_added: '0.2.0'
|
|
dest:
|
|
type: path
|
|
description:
|
|
- The path where the artifact should be written to.
|
|
- If file mode or ownerships are specified and destination path already exists, they affect the downloaded file.
|
|
required: true
|
|
state:
|
|
type: str
|
|
description:
|
|
- The desired state of the artifact.
|
|
default: present
|
|
choices: [present,absent]
|
|
timeout:
|
|
type: int
|
|
description:
|
|
- Specifies a timeout in seconds for the connection attempt.
|
|
default: 10
|
|
validate_certs:
|
|
description:
|
|
- If V(false), SSL certificates will not be validated. This should only be set to V(false) when no other option exists.
|
|
type: bool
|
|
default: true
|
|
client_cert:
|
|
description:
|
|
- PEM formatted certificate chain file to be used for SSL client authentication.
|
|
- This file can also include the key as well, and if the key is included, O(client_key) is not required.
|
|
type: path
|
|
version_added: '1.3.0'
|
|
client_key:
|
|
description:
|
|
- PEM formatted file that contains your private key to be used for SSL client authentication.
|
|
- If O(client_cert) contains both the certificate and key, this option is not required.
|
|
type: path
|
|
version_added: '1.3.0'
|
|
keep_name:
|
|
description:
|
|
- If V(true), the downloaded artifact's name is preserved, i.e the version number remains part of it.
|
|
- This option only has effect when O(dest) is a directory and O(version) is set to V(latest) or O(version_by_spec)
|
|
is defined.
|
|
type: bool
|
|
default: false
|
|
verify_checksum:
|
|
type: str
|
|
description:
|
|
- If V(never), the MD5/SHA1 checksum will never be downloaded and verified.
|
|
- If V(download), the MD5/SHA1 checksum will be downloaded and verified only after artifact download. This is the default.
|
|
- If V(change), the MD5/SHA1 checksum will be downloaded and verified if the destination already exist,
|
|
to verify if they are identical. This was the behaviour before 2.6. Since it downloads the checksum before (maybe)
|
|
downloading the artifact, and since some repository software, when acting as a proxy/cache, return a 404 error
|
|
if the artifact has not been cached yet, it may fail unexpectedly.
|
|
If you still need it, you should consider using V(always) instead - if you deal with a checksum, it is better to
|
|
use it to verify integrity after download.
|
|
- V(always) combines V(download) and V(change).
|
|
required: false
|
|
default: 'download'
|
|
choices: ['never', 'download', 'change', 'always']
|
|
checksum_alg:
|
|
type: str
|
|
description:
|
|
- If V(md5), checksums will use the MD5 algorithm. This is the default.
|
|
- If V(sha1), checksums will use the SHA1 algorithm. This can be used on systems configured to use
|
|
FIPS-compliant algorithms, since MD5 will be blocked on such systems.
|
|
default: 'md5'
|
|
choices: ['md5', 'sha1']
|
|
version_added: 3.2.0
|
|
unredirected_headers:
|
|
type: list
|
|
elements: str
|
|
version_added: 5.2.0
|
|
description:
|
|
- A list of headers that should not be included in the redirection. This headers are sent to the C(fetch_url) function.
|
|
- On ansible-core version 2.12 or later, the default of this option is V([Authorization, Cookie]).
|
|
- Useful if the redirection URL does not need to have sensitive headers in the request.
|
|
- Requires ansible-core version 2.12 or later.
|
|
directory_mode:
|
|
type: str
|
|
description:
|
|
- Filesystem permission mode applied recursively to O(dest) when it is a directory.
|
|
extends_documentation_fragment:
|
|
- ansible.builtin.files
|
|
- community.general.attributes
|
|
'''
|
|
|
|
EXAMPLES = '''
|
|
- name: Download the latest version of the JUnit framework artifact from Maven Central
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
dest: /tmp/junit-latest.jar
|
|
|
|
- name: Download JUnit 4.11 from Maven Central
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
version: 4.11
|
|
dest: /tmp/junit-4.11.jar
|
|
|
|
- name: Download an artifact from a private repository requiring authentication
|
|
community.general.maven_artifact:
|
|
group_id: com.company
|
|
artifact_id: library-name
|
|
repository_url: 'https://repo.company.com/maven'
|
|
username: user
|
|
password: pass
|
|
dest: /tmp/library-name-latest.jar
|
|
|
|
- name: Download an artifact from a private repository requiring certificate authentication
|
|
community.general.maven_artifact:
|
|
group_id: com.company
|
|
artifact_id: library-name
|
|
repository_url: 'https://repo.company.com/maven'
|
|
client_cert: /path/to/cert.pem
|
|
client_key: /path/to/key.pem
|
|
dest: /tmp/library-name-latest.jar
|
|
|
|
- name: Download a WAR File to the Tomcat webapps directory to be deployed
|
|
community.general.maven_artifact:
|
|
group_id: com.company
|
|
artifact_id: web-app
|
|
extension: war
|
|
repository_url: 'https://repo.company.com/maven'
|
|
dest: /var/lib/tomcat7/webapps/web-app.war
|
|
|
|
- name: Keep a downloaded artifact's name, i.e. retain the version
|
|
community.general.maven_artifact:
|
|
version: latest
|
|
artifact_id: spring-core
|
|
group_id: org.springframework
|
|
dest: /tmp/
|
|
keep_name: true
|
|
|
|
- name: Download the latest version of the JUnit framework artifact from Maven local
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
dest: /tmp/junit-latest.jar
|
|
repository_url: "file://{{ lookup('env','HOME') }}/.m2/repository"
|
|
|
|
- name: Download the latest version between 3.8 and 4.0 (exclusive) of the JUnit framework artifact from Maven Central
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
version_by_spec: "[3.8,4.0)"
|
|
dest: /tmp/
|
|
'''
|
|
|
|
import hashlib
|
|
import os
|
|
import posixpath
|
|
import shutil
|
|
import io
|
|
import tempfile
|
|
import traceback
|
|
import re
|
|
|
|
from ansible_collections.community.general.plugins.module_utils.version import LooseVersion
|
|
from ansible.module_utils.ansible_release import __version__ as ansible_version
|
|
from re import match
|
|
|
|
LXML_ETREE_IMP_ERR = None
|
|
try:
|
|
from lxml import etree
|
|
HAS_LXML_ETREE = True
|
|
except ImportError:
|
|
LXML_ETREE_IMP_ERR = traceback.format_exc()
|
|
HAS_LXML_ETREE = False
|
|
|
|
BOTO_IMP_ERR = None
|
|
try:
|
|
import boto3
|
|
HAS_BOTO = True
|
|
except ImportError:
|
|
BOTO_IMP_ERR = traceback.format_exc()
|
|
HAS_BOTO = False
|
|
|
|
SEMANTIC_VERSION_IMP_ERR = None
|
|
try:
|
|
from semantic_version import Version, Spec
|
|
HAS_SEMANTIC_VERSION = True
|
|
except ImportError:
|
|
SEMANTIC_VERSION_IMP_ERR = traceback.format_exc()
|
|
HAS_SEMANTIC_VERSION = False
|
|
|
|
|
|
from ansible.module_utils.basic import AnsibleModule, missing_required_lib
|
|
from ansible.module_utils.six.moves.urllib.parse import urlparse
|
|
from ansible.module_utils.urls import fetch_url
|
|
from ansible.module_utils.common.text.converters import to_bytes, to_native, to_text
|
|
|
|
|
|
def split_pre_existing_dir(dirname):
|
|
'''
|
|
Return the first pre-existing directory and a list of the new directories that will be created.
|
|
'''
|
|
head, tail = os.path.split(dirname)
|
|
b_head = to_bytes(head, errors='surrogate_or_strict')
|
|
if not os.path.exists(b_head):
|
|
if head == dirname:
|
|
return None, [head]
|
|
else:
|
|
(pre_existing_dir, new_directory_list) = split_pre_existing_dir(head)
|
|
else:
|
|
return head, [tail]
|
|
new_directory_list.append(tail)
|
|
return pre_existing_dir, new_directory_list
|
|
|
|
|
|
def adjust_recursive_directory_permissions(pre_existing_dir, new_directory_list, module, directory_args, changed):
|
|
'''
|
|
Walk the new directories list and make sure that permissions are as we would expect
|
|
'''
|
|
if new_directory_list:
|
|
first_sub_dir = new_directory_list.pop(0)
|
|
if not pre_existing_dir:
|
|
working_dir = first_sub_dir
|
|
else:
|
|
working_dir = os.path.join(pre_existing_dir, first_sub_dir)
|
|
directory_args['path'] = working_dir
|
|
changed = module.set_fs_attributes_if_different(directory_args, changed)
|
|
changed = adjust_recursive_directory_permissions(working_dir, new_directory_list, module, directory_args, changed)
|
|
return changed
|
|
|
|
|
|
class Artifact(object):
|
|
def __init__(self, group_id, artifact_id, version, version_by_spec, classifier='', extension='jar'):
|
|
if not group_id:
|
|
raise ValueError("group_id must be set")
|
|
if not artifact_id:
|
|
raise ValueError("artifact_id must be set")
|
|
|
|
self.group_id = group_id
|
|
self.artifact_id = artifact_id
|
|
self.version = version
|
|
self.version_by_spec = version_by_spec
|
|
self.classifier = classifier
|
|
|
|
if not extension:
|
|
self.extension = "jar"
|
|
else:
|
|
self.extension = extension
|
|
|
|
def is_snapshot(self):
|
|
return self.version and self.version.endswith("SNAPSHOT")
|
|
|
|
def path(self, with_version=True):
|
|
base = posixpath.join(self.group_id.replace(".", "/"), self.artifact_id)
|
|
if with_version and self.version:
|
|
timestamp_version_match = re.match("^(.*-)?([0-9]{8}\\.[0-9]{6}-[0-9]+)$", self.version)
|
|
if timestamp_version_match:
|
|
base = posixpath.join(base, timestamp_version_match.group(1) + "SNAPSHOT")
|
|
else:
|
|
base = posixpath.join(base, self.version)
|
|
return base
|
|
|
|
def _generate_filename(self):
|
|
filename = self.artifact_id + "-" + self.classifier + "." + self.extension
|
|
if not self.classifier:
|
|
filename = self.artifact_id + "." + self.extension
|
|
return filename
|
|
|
|
def get_filename(self, filename=None):
|
|
if not filename:
|
|
filename = self._generate_filename()
|
|
elif os.path.isdir(filename):
|
|
filename = os.path.join(filename, self._generate_filename())
|
|
return filename
|
|
|
|
def __str__(self):
|
|
result = "%s:%s:%s" % (self.group_id, self.artifact_id, self.version)
|
|
if self.classifier:
|
|
result = "%s:%s:%s:%s:%s" % (self.group_id, self.artifact_id, self.extension, self.classifier, self.version)
|
|
elif self.extension != "jar":
|
|
result = "%s:%s:%s:%s" % (self.group_id, self.artifact_id, self.extension, self.version)
|
|
return result
|
|
|
|
@staticmethod
|
|
def parse(input):
|
|
parts = input.split(":")
|
|
if len(parts) >= 3:
|
|
g = parts[0]
|
|
a = parts[1]
|
|
v = parts[-1]
|
|
t = None
|
|
c = None
|
|
if len(parts) == 4:
|
|
t = parts[2]
|
|
if len(parts) == 5:
|
|
t = parts[2]
|
|
c = parts[3]
|
|
return Artifact(g, a, v, c, t)
|
|
else:
|
|
return None
|
|
|
|
|
|
class MavenDownloader:
|
|
def __init__(self, module, base, local=False, headers=None):
|
|
self.module = module
|
|
if base.endswith("/"):
|
|
base = base.rstrip("/")
|
|
self.base = base
|
|
self.local = local
|
|
self.headers = headers
|
|
self.user_agent = "Ansible {0} maven_artifact".format(ansible_version)
|
|
self.latest_version_found = None
|
|
self.metadata_file_name = "maven-metadata-local.xml" if local else "maven-metadata.xml"
|
|
|
|
def find_version_by_spec(self, artifact):
|
|
path = "/%s/%s" % (artifact.path(False), self.metadata_file_name)
|
|
content = self._getContent(self.base + path, "Failed to retrieve the maven metadata file: " + path)
|
|
xml = etree.fromstring(content)
|
|
original_versions = xml.xpath("/metadata/versioning/versions/version/text()")
|
|
versions = []
|
|
for version in original_versions:
|
|
try:
|
|
versions.append(Version.coerce(version))
|
|
except ValueError:
|
|
# This means that version string is not a valid semantic versioning
|
|
pass
|
|
|
|
parse_versions_syntax = {
|
|
# example -> (,1.0]
|
|
r"^\(,(?P<upper_bound>[0-9.]*)]$": "<={upper_bound}",
|
|
# example -> 1.0
|
|
r"^(?P<version>[0-9.]*)$": "~={version}",
|
|
# example -> [1.0]
|
|
r"^\[(?P<version>[0-9.]*)\]$": "=={version}",
|
|
# example -> [1.2, 1.3]
|
|
r"^\[(?P<lower_bound>[0-9.]*),\s*(?P<upper_bound>[0-9.]*)\]$": ">={lower_bound},<={upper_bound}",
|
|
# example -> [1.2, 1.3)
|
|
r"^\[(?P<lower_bound>[0-9.]*),\s*(?P<upper_bound>[0-9.]+)\)$": ">={lower_bound},<{upper_bound}",
|
|
# example -> [1.5,)
|
|
r"^\[(?P<lower_bound>[0-9.]*),\)$": ">={lower_bound}",
|
|
}
|
|
|
|
for regex, spec_format in parse_versions_syntax.items():
|
|
regex_result = match(regex, artifact.version_by_spec)
|
|
if regex_result:
|
|
spec = Spec(spec_format.format(**regex_result.groupdict()))
|
|
selected_version = spec.select(versions)
|
|
|
|
if not selected_version:
|
|
raise ValueError("No version found with this spec version: {0}".format(artifact.version_by_spec))
|
|
|
|
# To deal when repos on maven don't have patch number on first build (e.g. 3.8 instead of 3.8.0)
|
|
if str(selected_version) not in original_versions:
|
|
selected_version.patch = None
|
|
|
|
return str(selected_version)
|
|
|
|
raise ValueError("The spec version {0} is not supported! ".format(artifact.version_by_spec))
|
|
|
|
def find_latest_version_available(self, artifact):
|
|
if self.latest_version_found:
|
|
return self.latest_version_found
|
|
path = "/%s/%s" % (artifact.path(False), self.metadata_file_name)
|
|
content = self._getContent(self.base + path, "Failed to retrieve the maven metadata file: " + path)
|
|
xml = etree.fromstring(content)
|
|
v = xml.xpath("/metadata/versioning/versions/version[last()]/text()")
|
|
if v:
|
|
self.latest_version_found = v[0]
|
|
return v[0]
|
|
|
|
def find_uri_for_artifact(self, artifact):
|
|
if artifact.version_by_spec:
|
|
artifact.version = self.find_version_by_spec(artifact)
|
|
|
|
if artifact.version == "latest":
|
|
artifact.version = self.find_latest_version_available(artifact)
|
|
|
|
if artifact.is_snapshot():
|
|
if self.local:
|
|
return self._uri_for_artifact(artifact, artifact.version)
|
|
path = "/%s/%s" % (artifact.path(), self.metadata_file_name)
|
|
content = self._getContent(self.base + path, "Failed to retrieve the maven metadata file: " + path)
|
|
xml = etree.fromstring(content)
|
|
|
|
for snapshotArtifact in xml.xpath("/metadata/versioning/snapshotVersions/snapshotVersion"):
|
|
classifier = snapshotArtifact.xpath("classifier/text()")
|
|
artifact_classifier = classifier[0] if classifier else ''
|
|
extension = snapshotArtifact.xpath("extension/text()")
|
|
artifact_extension = extension[0] if extension else ''
|
|
if artifact_classifier == artifact.classifier and artifact_extension == artifact.extension:
|
|
return self._uri_for_artifact(artifact, snapshotArtifact.xpath("value/text()")[0])
|
|
timestamp_xmlpath = xml.xpath("/metadata/versioning/snapshot/timestamp/text()")
|
|
if timestamp_xmlpath:
|
|
timestamp = timestamp_xmlpath[0]
|
|
build_number = xml.xpath("/metadata/versioning/snapshot/buildNumber/text()")[0]
|
|
return self._uri_for_artifact(artifact, artifact.version.replace("SNAPSHOT", timestamp + "-" + build_number))
|
|
|
|
return self._uri_for_artifact(artifact, artifact.version)
|
|
|
|
def _uri_for_artifact(self, artifact, version=None):
|
|
if artifact.is_snapshot() and not version:
|
|
raise ValueError("Expected uniqueversion for snapshot artifact " + str(artifact))
|
|
elif not artifact.is_snapshot():
|
|
version = artifact.version
|
|
if artifact.classifier:
|
|
return posixpath.join(self.base, artifact.path(), artifact.artifact_id + "-" + version + "-" + artifact.classifier + "." + artifact.extension)
|
|
|
|
return posixpath.join(self.base, artifact.path(), artifact.artifact_id + "-" + version + "." + artifact.extension)
|
|
|
|
# for small files, directly get the full content
|
|
def _getContent(self, url, failmsg, force=True):
|
|
if self.local:
|
|
parsed_url = urlparse(url)
|
|
if os.path.isfile(parsed_url.path):
|
|
with io.open(parsed_url.path, 'rb') as f:
|
|
return f.read()
|
|
if force:
|
|
raise ValueError(failmsg + " because can not find file: " + url)
|
|
return None
|
|
response = self._request(url, failmsg, force)
|
|
if response:
|
|
return response.read()
|
|
return None
|
|
|
|
# only for HTTP request
|
|
def _request(self, url, failmsg, force=True):
|
|
url_to_use = url
|
|
parsed_url = urlparse(url)
|
|
|
|
if parsed_url.scheme == 's3':
|
|
parsed_url = urlparse(url)
|
|
bucket_name = parsed_url.netloc
|
|
key_name = parsed_url.path[1:]
|
|
client = boto3.client('s3', aws_access_key_id=self.module.params.get('username', ''), aws_secret_access_key=self.module.params.get('password', ''))
|
|
url_to_use = client.generate_presigned_url('get_object', Params={'Bucket': bucket_name, 'Key': key_name}, ExpiresIn=10)
|
|
|
|
req_timeout = self.module.params.get('timeout')
|
|
|
|
# Hack to add parameters in the way that fetch_url expects
|
|
self.module.params['url_username'] = self.module.params.get('username', '')
|
|
self.module.params['url_password'] = self.module.params.get('password', '')
|
|
self.module.params['http_agent'] = self.user_agent
|
|
|
|
kwargs = {}
|
|
if self.module.params['unredirected_headers']:
|
|
kwargs['unredirected_headers'] = self.module.params['unredirected_headers']
|
|
|
|
response, info = fetch_url(
|
|
self.module,
|
|
url_to_use,
|
|
timeout=req_timeout,
|
|
headers=self.headers,
|
|
**kwargs
|
|
)
|
|
|
|
if info['status'] == 200:
|
|
return response
|
|
if force:
|
|
raise ValueError(failmsg + " because of " + info['msg'] + "for URL " + url_to_use)
|
|
return None
|
|
|
|
def download(self, tmpdir, artifact, verify_download, filename=None, checksum_alg='md5'):
|
|
if (not artifact.version and not artifact.version_by_spec) or artifact.version == "latest":
|
|
artifact = Artifact(artifact.group_id, artifact.artifact_id, self.find_latest_version_available(artifact), None,
|
|
artifact.classifier, artifact.extension)
|
|
url = self.find_uri_for_artifact(artifact)
|
|
tempfd, tempname = tempfile.mkstemp(dir=tmpdir)
|
|
|
|
try:
|
|
# copy to temp file
|
|
if self.local:
|
|
parsed_url = urlparse(url)
|
|
if os.path.isfile(parsed_url.path):
|
|
shutil.copy2(parsed_url.path, tempname)
|
|
else:
|
|
return "Can not find local file: " + parsed_url.path
|
|
else:
|
|
response = self._request(url, "Failed to download artifact " + str(artifact))
|
|
with os.fdopen(tempfd, 'wb') as f:
|
|
shutil.copyfileobj(response, f)
|
|
|
|
if verify_download:
|
|
invalid_checksum = self.is_invalid_checksum(tempname, url, checksum_alg)
|
|
if invalid_checksum:
|
|
# if verify_change was set, the previous file would be deleted
|
|
os.remove(tempname)
|
|
return invalid_checksum
|
|
except Exception as e:
|
|
os.remove(tempname)
|
|
raise e
|
|
|
|
# all good, now copy temp file to target
|
|
shutil.move(tempname, artifact.get_filename(filename))
|
|
return None
|
|
|
|
def is_invalid_checksum(self, file, remote_url, checksum_alg='md5'):
|
|
if os.path.exists(file):
|
|
local_checksum = self._local_checksum(checksum_alg, file)
|
|
if self.local:
|
|
parsed_url = urlparse(remote_url)
|
|
remote_checksum = self._local_checksum(checksum_alg, parsed_url.path)
|
|
else:
|
|
try:
|
|
remote_checksum = to_text(self._getContent(remote_url + '.' + checksum_alg, "Failed to retrieve checksum", False), errors='strict')
|
|
except UnicodeError as e:
|
|
return "Cannot retrieve a valid %s checksum from %s: %s" % (checksum_alg, remote_url, to_native(e))
|
|
if not remote_checksum:
|
|
return "Cannot find %s checksum from %s" % (checksum_alg, remote_url)
|
|
try:
|
|
# Check if remote checksum only contains md5/sha1 or md5/sha1 + filename
|
|
_remote_checksum = remote_checksum.split(None, 1)[0]
|
|
remote_checksum = _remote_checksum
|
|
# remote_checksum is empty so we continue and keep original checksum string
|
|
# This should not happen since we check for remote_checksum before
|
|
except IndexError:
|
|
pass
|
|
if local_checksum.lower() == remote_checksum.lower():
|
|
return None
|
|
else:
|
|
return "Checksum does not match: we computed " + local_checksum + " but the repository states " + remote_checksum
|
|
|
|
return "Path does not exist: " + file
|
|
|
|
def _local_checksum(self, checksum_alg, file):
|
|
if checksum_alg.lower() == 'md5':
|
|
hash = hashlib.md5()
|
|
elif checksum_alg.lower() == 'sha1':
|
|
hash = hashlib.sha1()
|
|
else:
|
|
raise ValueError("Unknown checksum_alg %s" % checksum_alg)
|
|
with io.open(file, 'rb') as f:
|
|
for chunk in iter(lambda: f.read(8192), b''):
|
|
hash.update(chunk)
|
|
return hash.hexdigest()
|
|
|
|
|
|
def main():
|
|
module = AnsibleModule(
|
|
argument_spec=dict(
|
|
group_id=dict(required=True),
|
|
artifact_id=dict(required=True),
|
|
version=dict(default=None),
|
|
version_by_spec=dict(default=None),
|
|
classifier=dict(default=''),
|
|
extension=dict(default='jar'),
|
|
repository_url=dict(default='https://repo1.maven.org/maven2'),
|
|
username=dict(default=None, aliases=['aws_secret_key']),
|
|
password=dict(default=None, no_log=True, aliases=['aws_secret_access_key']),
|
|
headers=dict(type='dict'),
|
|
force_basic_auth=dict(default=False, type='bool'),
|
|
state=dict(default="present", choices=["present", "absent"]), # TODO - Implement a "latest" state
|
|
timeout=dict(default=10, type='int'),
|
|
dest=dict(type="path", required=True),
|
|
validate_certs=dict(required=False, default=True, type='bool'),
|
|
client_cert=dict(type="path", required=False),
|
|
client_key=dict(type="path", required=False),
|
|
keep_name=dict(required=False, default=False, type='bool'),
|
|
verify_checksum=dict(required=False, default='download', choices=['never', 'download', 'change', 'always']),
|
|
checksum_alg=dict(required=False, default='md5', choices=['md5', 'sha1']),
|
|
unredirected_headers=dict(type='list', elements='str', required=False),
|
|
directory_mode=dict(type='str'),
|
|
),
|
|
add_file_common_args=True,
|
|
mutually_exclusive=([('version', 'version_by_spec')])
|
|
)
|
|
|
|
if LooseVersion(ansible_version) < LooseVersion("2.12") and module.params['unredirected_headers']:
|
|
module.fail_json(msg="Unredirected Headers parameter provided, but your ansible-core version does not support it. Minimum version is 2.12")
|
|
|
|
if LooseVersion(ansible_version) >= LooseVersion("2.12") and module.params['unredirected_headers'] is None:
|
|
# if the user did not supply unredirected params, we use the default, ONLY on ansible core 2.12 and above
|
|
module.params['unredirected_headers'] = ['Authorization', 'Cookie']
|
|
|
|
if not HAS_LXML_ETREE:
|
|
module.fail_json(msg=missing_required_lib('lxml'), exception=LXML_ETREE_IMP_ERR)
|
|
|
|
if module.params['version_by_spec'] and not HAS_SEMANTIC_VERSION:
|
|
module.fail_json(msg=missing_required_lib('semantic_version'), exception=SEMANTIC_VERSION_IMP_ERR)
|
|
|
|
repository_url = module.params["repository_url"]
|
|
if not repository_url:
|
|
repository_url = "https://repo1.maven.org/maven2"
|
|
try:
|
|
parsed_url = urlparse(repository_url)
|
|
except AttributeError as e:
|
|
module.fail_json(msg='url parsing went wrong %s' % e)
|
|
|
|
local = parsed_url.scheme == "file"
|
|
|
|
if parsed_url.scheme == 's3' and not HAS_BOTO:
|
|
module.fail_json(msg=missing_required_lib('boto3', reason='when using s3:// repository URLs'),
|
|
exception=BOTO_IMP_ERR)
|
|
|
|
group_id = module.params["group_id"]
|
|
artifact_id = module.params["artifact_id"]
|
|
version = module.params["version"]
|
|
version_by_spec = module.params["version_by_spec"]
|
|
classifier = module.params["classifier"]
|
|
extension = module.params["extension"]
|
|
headers = module.params['headers']
|
|
state = module.params["state"]
|
|
dest = module.params["dest"]
|
|
b_dest = to_bytes(dest, errors='surrogate_or_strict')
|
|
keep_name = module.params["keep_name"]
|
|
verify_checksum = module.params["verify_checksum"]
|
|
verify_download = verify_checksum in ['download', 'always']
|
|
verify_change = verify_checksum in ['change', 'always']
|
|
checksum_alg = module.params["checksum_alg"]
|
|
|
|
downloader = MavenDownloader(module, repository_url, local, headers)
|
|
|
|
if not version_by_spec and not version:
|
|
version = "latest"
|
|
|
|
try:
|
|
artifact = Artifact(group_id, artifact_id, version, version_by_spec, classifier, extension)
|
|
except ValueError as e:
|
|
module.fail_json(msg=e.args[0])
|
|
|
|
changed = False
|
|
prev_state = "absent"
|
|
|
|
if dest.endswith(os.sep):
|
|
b_dest = to_bytes(dest, errors='surrogate_or_strict')
|
|
if not os.path.exists(b_dest):
|
|
(pre_existing_dir, new_directory_list) = split_pre_existing_dir(dest)
|
|
os.makedirs(b_dest)
|
|
directory_args = module.load_file_common_arguments(module.params)
|
|
directory_mode = module.params["directory_mode"]
|
|
if directory_mode is not None:
|
|
directory_args['mode'] = directory_mode
|
|
else:
|
|
directory_args['mode'] = None
|
|
changed = adjust_recursive_directory_permissions(pre_existing_dir, new_directory_list, module, directory_args, changed)
|
|
|
|
if os.path.isdir(b_dest):
|
|
version_part = version
|
|
if version == 'latest':
|
|
version_part = downloader.find_latest_version_available(artifact)
|
|
elif version_by_spec:
|
|
version_part = downloader.find_version_by_spec(artifact)
|
|
|
|
filename = "{artifact_id}{version_part}{classifier}.{extension}".format(
|
|
artifact_id=artifact_id,
|
|
version_part="-{0}".format(version_part) if keep_name else "",
|
|
classifier="-{0}".format(classifier) if classifier else "",
|
|
extension=extension
|
|
)
|
|
dest = posixpath.join(dest, filename)
|
|
|
|
b_dest = to_bytes(dest, errors='surrogate_or_strict')
|
|
|
|
if os.path.lexists(b_dest) and ((not verify_change) or not downloader.is_invalid_checksum(dest, downloader.find_uri_for_artifact(artifact), checksum_alg)):
|
|
prev_state = "present"
|
|
|
|
if prev_state == "absent":
|
|
try:
|
|
download_error = downloader.download(module.tmpdir, artifact, verify_download, b_dest, checksum_alg)
|
|
if download_error is None:
|
|
changed = True
|
|
else:
|
|
module.fail_json(msg="Cannot retrieve the artifact to destination: " + download_error)
|
|
except ValueError as e:
|
|
module.fail_json(msg=e.args[0])
|
|
|
|
file_args = module.load_file_common_arguments(module.params, path=dest)
|
|
changed = module.set_fs_attributes_if_different(file_args, changed)
|
|
if changed:
|
|
module.exit_json(state=state, dest=dest, group_id=group_id, artifact_id=artifact_id, version=version, classifier=classifier,
|
|
extension=extension, repository_url=repository_url, changed=changed)
|
|
else:
|
|
module.exit_json(state=state, dest=dest, changed=changed)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
main()
|