mirror of
https://github.com/ansible-collections/community.general.git
synced 2024-09-14 20:13:21 +02:00
9787e8a6bf
* Adjust path for time stamped SNAPSHOT versions Test the version string for the timestamp SNAPSHOT pattern. If it matches, substitute the SNAPSHOT version in the artifact path rather than the timestamped version string. * Add changelog fragment for PR 713 * Update changelogs/fragments/713-maven-timestamp-snapshot.yml Co-authored-by: Felix Fontein <felix@fontein.de> * C Fix problems reported by shippable Fix problems reported by pull request checks * Resolve additional shipping errors Co-authored-by: jweber <jweber@seastreet.com> Co-authored-by: Felix Fontein <felix@fontein.de>
712 lines
29 KiB
Python
712 lines
29 KiB
Python
#!/usr/bin/python
|
|
# -*- coding: utf-8 -*-
|
|
|
|
# Copyright (c) 2014, Chris Schmidt <chris.schmidt () contrastsecurity.com>
|
|
#
|
|
# Built using https://github.com/hamnis/useful-scripts/blob/master/python/download-maven-artifact
|
|
# as a reference and starting point.
|
|
#
|
|
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
|
from __future__ import absolute_import, division, print_function
|
|
__metaclass__ = type
|
|
|
|
DOCUMENTATION = '''
|
|
---
|
|
module: maven_artifact
|
|
short_description: Downloads an Artifact from a Maven Repository
|
|
description:
|
|
- Downloads an artifact from a maven repository given the maven coordinates provided to the module.
|
|
- Can retrieve snapshots or release versions of the artifact and will resolve the latest available
|
|
version if one is not available.
|
|
author: "Chris Schmidt (@chrisisbeef)"
|
|
requirements:
|
|
- lxml
|
|
- boto if using a S3 repository (s3://...)
|
|
options:
|
|
group_id:
|
|
type: str
|
|
description:
|
|
- The Maven groupId coordinate
|
|
required: true
|
|
artifact_id:
|
|
type: str
|
|
description:
|
|
- The maven artifactId coordinate
|
|
required: true
|
|
version:
|
|
type: str
|
|
description:
|
|
- The maven version coordinate
|
|
- Mutually exclusive with I(version_by_spec).
|
|
version_by_spec:
|
|
type: str
|
|
description:
|
|
- The maven dependency version ranges.
|
|
- See supported version ranges on U(https://cwiki.apache.org/confluence/display/MAVENOLD/Dependency+Mediation+and+Conflict+Resolution)
|
|
- The range type "(,1.0],[1.2,)" and "(,1.1),(1.1,)" is not supported.
|
|
- Mutually exclusive with I(version).
|
|
version_added: '0.2.0'
|
|
classifier:
|
|
type: str
|
|
description:
|
|
- The maven classifier coordinate
|
|
extension:
|
|
type: str
|
|
description:
|
|
- The maven type/extension coordinate
|
|
default: jar
|
|
repository_url:
|
|
type: str
|
|
description:
|
|
- The URL of the Maven Repository to download from.
|
|
- Use s3://... if the repository is hosted on Amazon S3, added in version 2.2.
|
|
- Use file://... if the repository is local, added in version 2.6
|
|
default: https://repo1.maven.org/maven2
|
|
username:
|
|
type: str
|
|
description:
|
|
- The username to authenticate as to the Maven Repository. Use AWS secret key of the repository is hosted on S3
|
|
aliases: [ "aws_secret_key" ]
|
|
password:
|
|
type: str
|
|
description:
|
|
- The password to authenticate with to the Maven Repository. Use AWS secret access key of the repository is hosted on S3
|
|
aliases: [ "aws_secret_access_key" ]
|
|
headers:
|
|
description:
|
|
- Add custom HTTP headers to a request in hash/dict format.
|
|
type: dict
|
|
force_basic_auth:
|
|
description:
|
|
- httplib2, the library used by the uri module only sends authentication information when a webservice
|
|
responds to an initial request with a 401 status. Since some basic auth services do not properly
|
|
send a 401, logins will fail. This option forces the sending of the Basic authentication header
|
|
upon initial request.
|
|
default: 'no'
|
|
type: bool
|
|
version_added: '0.2.0'
|
|
dest:
|
|
type: path
|
|
description:
|
|
- The path where the artifact should be written to
|
|
- If file mode or ownerships are specified and destination path already exists, they affect the downloaded file
|
|
required: true
|
|
state:
|
|
type: str
|
|
description:
|
|
- The desired state of the artifact
|
|
default: present
|
|
choices: [present,absent]
|
|
timeout:
|
|
type: int
|
|
description:
|
|
- Specifies a timeout in seconds for the connection attempt
|
|
default: 10
|
|
validate_certs:
|
|
description:
|
|
- If C(no), SSL certificates will not be validated. This should only be set to C(no) when no other option exists.
|
|
type: bool
|
|
default: 'yes'
|
|
client_cert:
|
|
description:
|
|
- PEM formatted certificate chain file to be used for SSL client authentication.
|
|
- This file can also include the key as well, and if the key is included, I(client_key) is not required.
|
|
type: path
|
|
version_added: '1.3.0'
|
|
client_key:
|
|
description:
|
|
- PEM formatted file that contains your private key to be used for SSL client authentication.
|
|
- If I(client_cert) contains both the certificate and key, this option is not required.
|
|
type: path
|
|
version_added: '1.3.0'
|
|
keep_name:
|
|
description:
|
|
- If C(yes), the downloaded artifact's name is preserved, i.e the version number remains part of it.
|
|
- This option only has effect when C(dest) is a directory and C(version) is set to C(latest) or C(version_by_spec)
|
|
is defined.
|
|
type: bool
|
|
default: 'no'
|
|
verify_checksum:
|
|
type: str
|
|
description:
|
|
- If C(never), the md5 checksum will never be downloaded and verified.
|
|
- If C(download), the md5 checksum will be downloaded and verified only after artifact download. This is the default.
|
|
- If C(change), the md5 checksum will be downloaded and verified if the destination already exist,
|
|
to verify if they are identical. This was the behaviour before 2.6. Since it downloads the md5 before (maybe)
|
|
downloading the artifact, and since some repository software, when acting as a proxy/cache, return a 404 error
|
|
if the artifact has not been cached yet, it may fail unexpectedly.
|
|
If you still need it, you should consider using C(always) instead - if you deal with a checksum, it is better to
|
|
use it to verify integrity after download.
|
|
- C(always) combines C(download) and C(change).
|
|
required: false
|
|
default: 'download'
|
|
choices: ['never', 'download', 'change', 'always']
|
|
extends_documentation_fragment:
|
|
- files
|
|
'''
|
|
|
|
EXAMPLES = '''
|
|
- name: Download the latest version of the JUnit framework artifact from Maven Central
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
dest: /tmp/junit-latest.jar
|
|
|
|
- name: Download JUnit 4.11 from Maven Central
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
version: 4.11
|
|
dest: /tmp/junit-4.11.jar
|
|
|
|
- name: Download an artifact from a private repository requiring authentication
|
|
community.general.maven_artifact:
|
|
group_id: com.company
|
|
artifact_id: library-name
|
|
repository_url: 'https://repo.company.com/maven'
|
|
username: user
|
|
password: pass
|
|
dest: /tmp/library-name-latest.jar
|
|
|
|
- name: Download an artifact from a private repository requiring certificate authentication
|
|
community.general.maven_artifact:
|
|
group_id: com.company
|
|
artifact_id: library-name
|
|
repository_url: 'https://repo.company.com/maven'
|
|
client_cert: /path/to/cert.pem
|
|
client_key: /path/to/key.pem
|
|
dest: /tmp/library-name-latest.jar
|
|
|
|
- name: Download a WAR File to the Tomcat webapps directory to be deployed
|
|
community.general.maven_artifact:
|
|
group_id: com.company
|
|
artifact_id: web-app
|
|
extension: war
|
|
repository_url: 'https://repo.company.com/maven'
|
|
dest: /var/lib/tomcat7/webapps/web-app.war
|
|
|
|
- name: Keep a downloaded artifact's name, i.e. retain the version
|
|
community.general.maven_artifact:
|
|
version: latest
|
|
artifact_id: spring-core
|
|
group_id: org.springframework
|
|
dest: /tmp/
|
|
keep_name: yes
|
|
|
|
- name: Download the latest version of the JUnit framework artifact from Maven local
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
dest: /tmp/junit-latest.jar
|
|
repository_url: "file://{{ lookup('env','HOME') }}/.m2/repository"
|
|
|
|
- name: Download the latest version between 3.8 and 4.0 (exclusive) of the JUnit framework artifact from Maven Central
|
|
community.general.maven_artifact:
|
|
group_id: junit
|
|
artifact_id: junit
|
|
version_by_spec: "[3.8,4.0)"
|
|
dest: /tmp/
|
|
'''
|
|
|
|
import hashlib
|
|
import os
|
|
import posixpath
|
|
import shutil
|
|
import io
|
|
import tempfile
|
|
import traceback
|
|
import re
|
|
|
|
from ansible.module_utils.ansible_release import __version__ as ansible_version
|
|
from re import match
|
|
|
|
LXML_ETREE_IMP_ERR = None
|
|
try:
|
|
from lxml import etree
|
|
HAS_LXML_ETREE = True
|
|
except ImportError:
|
|
LXML_ETREE_IMP_ERR = traceback.format_exc()
|
|
HAS_LXML_ETREE = False
|
|
|
|
BOTO_IMP_ERR = None
|
|
try:
|
|
import boto3
|
|
HAS_BOTO = True
|
|
except ImportError:
|
|
BOTO_IMP_ERR = traceback.format_exc()
|
|
HAS_BOTO = False
|
|
|
|
SEMANTIC_VERSION_IMP_ERR = None
|
|
try:
|
|
from semantic_version import Version, Spec
|
|
HAS_SEMANTIC_VERSION = True
|
|
except ImportError:
|
|
SEMANTIC_VERSION_IMP_ERR = traceback.format_exc()
|
|
HAS_SEMANTIC_VERSION = False
|
|
|
|
|
|
from ansible.module_utils.basic import AnsibleModule, missing_required_lib
|
|
from ansible.module_utils.six.moves.urllib.parse import urlparse
|
|
from ansible.module_utils.urls import fetch_url
|
|
from ansible.module_utils._text import to_bytes, to_native, to_text
|
|
|
|
|
|
def split_pre_existing_dir(dirname):
|
|
'''
|
|
Return the first pre-existing directory and a list of the new directories that will be created.
|
|
'''
|
|
head, tail = os.path.split(dirname)
|
|
b_head = to_bytes(head, errors='surrogate_or_strict')
|
|
if not os.path.exists(b_head):
|
|
if head == dirname:
|
|
return None, [head]
|
|
else:
|
|
(pre_existing_dir, new_directory_list) = split_pre_existing_dir(head)
|
|
else:
|
|
return head, [tail]
|
|
new_directory_list.append(tail)
|
|
return pre_existing_dir, new_directory_list
|
|
|
|
|
|
def adjust_recursive_directory_permissions(pre_existing_dir, new_directory_list, module, directory_args, changed):
|
|
'''
|
|
Walk the new directories list and make sure that permissions are as we would expect
|
|
'''
|
|
if new_directory_list:
|
|
first_sub_dir = new_directory_list.pop(0)
|
|
if not pre_existing_dir:
|
|
working_dir = first_sub_dir
|
|
else:
|
|
working_dir = os.path.join(pre_existing_dir, first_sub_dir)
|
|
directory_args['path'] = working_dir
|
|
changed = module.set_fs_attributes_if_different(directory_args, changed)
|
|
changed = adjust_recursive_directory_permissions(working_dir, new_directory_list, module, directory_args, changed)
|
|
return changed
|
|
|
|
|
|
class Artifact(object):
|
|
def __init__(self, group_id, artifact_id, version, version_by_spec, classifier='', extension='jar'):
|
|
if not group_id:
|
|
raise ValueError("group_id must be set")
|
|
if not artifact_id:
|
|
raise ValueError("artifact_id must be set")
|
|
|
|
self.group_id = group_id
|
|
self.artifact_id = artifact_id
|
|
self.version = version
|
|
self.version_by_spec = version_by_spec
|
|
self.classifier = classifier
|
|
|
|
if not extension:
|
|
self.extension = "jar"
|
|
else:
|
|
self.extension = extension
|
|
|
|
def is_snapshot(self):
|
|
return self.version and self.version.endswith("SNAPSHOT")
|
|
|
|
def path(self, with_version=True):
|
|
base = posixpath.join(self.group_id.replace(".", "/"), self.artifact_id)
|
|
if with_version and self.version:
|
|
timestamp_version_match = re.match("^(.*-)?([0-9]{8}\\.[0-9]{6}-[0-9]+)$", self.version)
|
|
if timestamp_version_match:
|
|
base = posixpath.join(base, timestamp_version_match.group(1) + "SNAPSHOT")
|
|
else:
|
|
base = posixpath.join(base, self.version)
|
|
return base
|
|
|
|
def _generate_filename(self):
|
|
filename = self.artifact_id + "-" + self.classifier + "." + self.extension
|
|
if not self.classifier:
|
|
filename = self.artifact_id + "." + self.extension
|
|
return filename
|
|
|
|
def get_filename(self, filename=None):
|
|
if not filename:
|
|
filename = self._generate_filename()
|
|
elif os.path.isdir(filename):
|
|
filename = os.path.join(filename, self._generate_filename())
|
|
return filename
|
|
|
|
def __str__(self):
|
|
result = "%s:%s:%s" % (self.group_id, self.artifact_id, self.version)
|
|
if self.classifier:
|
|
result = "%s:%s:%s:%s:%s" % (self.group_id, self.artifact_id, self.extension, self.classifier, self.version)
|
|
elif self.extension != "jar":
|
|
result = "%s:%s:%s:%s" % (self.group_id, self.artifact_id, self.extension, self.version)
|
|
return result
|
|
|
|
@staticmethod
|
|
def parse(input):
|
|
parts = input.split(":")
|
|
if len(parts) >= 3:
|
|
g = parts[0]
|
|
a = parts[1]
|
|
v = parts[len(parts) - 1]
|
|
t = None
|
|
c = None
|
|
if len(parts) == 4:
|
|
t = parts[2]
|
|
if len(parts) == 5:
|
|
t = parts[2]
|
|
c = parts[3]
|
|
return Artifact(g, a, v, c, t)
|
|
else:
|
|
return None
|
|
|
|
|
|
class MavenDownloader:
|
|
def __init__(self, module, base, local=False, headers=None):
|
|
self.module = module
|
|
if base.endswith("/"):
|
|
base = base.rstrip("/")
|
|
self.base = base
|
|
self.local = local
|
|
self.headers = headers
|
|
self.user_agent = "Ansible {0} maven_artifact".format(ansible_version)
|
|
self.latest_version_found = None
|
|
self.metadata_file_name = "maven-metadata-local.xml" if local else "maven-metadata.xml"
|
|
|
|
def find_version_by_spec(self, artifact):
|
|
path = "/%s/%s" % (artifact.path(False), self.metadata_file_name)
|
|
content = self._getContent(self.base + path, "Failed to retrieve the maven metadata file: " + path)
|
|
xml = etree.fromstring(content)
|
|
original_versions = xml.xpath("/metadata/versioning/versions/version/text()")
|
|
versions = []
|
|
for version in original_versions:
|
|
try:
|
|
versions.append(Version.coerce(version))
|
|
except ValueError:
|
|
# This means that version string is not a valid semantic versioning
|
|
pass
|
|
|
|
parse_versions_syntax = {
|
|
# example -> (,1.0]
|
|
r"^\(,(?P<upper_bound>[0-9.]*)]$": "<={upper_bound}",
|
|
# example -> 1.0
|
|
r"^(?P<version>[0-9.]*)$": "~={version}",
|
|
# example -> [1.0]
|
|
r"^\[(?P<version>[0-9.]*)\]$": "=={version}",
|
|
# example -> [1.2, 1.3]
|
|
r"^\[(?P<lower_bound>[0-9.]*),\s*(?P<upper_bound>[0-9.]*)\]$": ">={lower_bound},<={upper_bound}",
|
|
# example -> [1.2, 1.3)
|
|
r"^\[(?P<lower_bound>[0-9.]*),\s*(?P<upper_bound>[0-9.]+)\)$": ">={lower_bound},<{upper_bound}",
|
|
# example -> [1.5,)
|
|
r"^\[(?P<lower_bound>[0-9.]*),\)$": ">={lower_bound}",
|
|
}
|
|
|
|
for regex, spec_format in parse_versions_syntax.items():
|
|
regex_result = match(regex, artifact.version_by_spec)
|
|
if regex_result:
|
|
spec = Spec(spec_format.format(**regex_result.groupdict()))
|
|
selected_version = spec.select(versions)
|
|
|
|
if not selected_version:
|
|
raise ValueError("No version found with this spec version: {0}".format(artifact.version_by_spec))
|
|
|
|
# To deal when repos on maven don't have patch number on first build (e.g. 3.8 instead of 3.8.0)
|
|
if str(selected_version) not in original_versions:
|
|
selected_version.patch = None
|
|
|
|
return str(selected_version)
|
|
|
|
raise ValueError("The spec version {0} is not supported! ".format(artifact.version_by_spec))
|
|
|
|
def find_latest_version_available(self, artifact):
|
|
if self.latest_version_found:
|
|
return self.latest_version_found
|
|
path = "/%s/%s" % (artifact.path(False), self.metadata_file_name)
|
|
content = self._getContent(self.base + path, "Failed to retrieve the maven metadata file: " + path)
|
|
xml = etree.fromstring(content)
|
|
v = xml.xpath("/metadata/versioning/versions/version[last()]/text()")
|
|
if v:
|
|
self.latest_version_found = v[0]
|
|
return v[0]
|
|
|
|
def find_uri_for_artifact(self, artifact):
|
|
if artifact.version_by_spec:
|
|
artifact.version = self.find_version_by_spec(artifact)
|
|
|
|
if artifact.version == "latest":
|
|
artifact.version = self.find_latest_version_available(artifact)
|
|
|
|
if artifact.is_snapshot():
|
|
if self.local:
|
|
return self._uri_for_artifact(artifact, artifact.version)
|
|
path = "/%s/%s" % (artifact.path(), self.metadata_file_name)
|
|
content = self._getContent(self.base + path, "Failed to retrieve the maven metadata file: " + path)
|
|
xml = etree.fromstring(content)
|
|
|
|
for snapshotArtifact in xml.xpath("/metadata/versioning/snapshotVersions/snapshotVersion"):
|
|
classifier = snapshotArtifact.xpath("classifier/text()")
|
|
artifact_classifier = classifier[0] if classifier else ''
|
|
extension = snapshotArtifact.xpath("extension/text()")
|
|
artifact_extension = extension[0] if extension else ''
|
|
if artifact_classifier == artifact.classifier and artifact_extension == artifact.extension:
|
|
return self._uri_for_artifact(artifact, snapshotArtifact.xpath("value/text()")[0])
|
|
timestamp_xmlpath = xml.xpath("/metadata/versioning/snapshot/timestamp/text()")
|
|
if timestamp_xmlpath:
|
|
timestamp = timestamp_xmlpath[0]
|
|
build_number = xml.xpath("/metadata/versioning/snapshot/buildNumber/text()")[0]
|
|
return self._uri_for_artifact(artifact, artifact.version.replace("SNAPSHOT", timestamp + "-" + build_number))
|
|
|
|
return self._uri_for_artifact(artifact, artifact.version)
|
|
|
|
def _uri_for_artifact(self, artifact, version=None):
|
|
if artifact.is_snapshot() and not version:
|
|
raise ValueError("Expected uniqueversion for snapshot artifact " + str(artifact))
|
|
elif not artifact.is_snapshot():
|
|
version = artifact.version
|
|
if artifact.classifier:
|
|
return posixpath.join(self.base, artifact.path(), artifact.artifact_id + "-" + version + "-" + artifact.classifier + "." + artifact.extension)
|
|
|
|
return posixpath.join(self.base, artifact.path(), artifact.artifact_id + "-" + version + "." + artifact.extension)
|
|
|
|
# for small files, directly get the full content
|
|
def _getContent(self, url, failmsg, force=True):
|
|
if self.local:
|
|
parsed_url = urlparse(url)
|
|
if os.path.isfile(parsed_url.path):
|
|
with io.open(parsed_url.path, 'rb') as f:
|
|
return f.read()
|
|
if force:
|
|
raise ValueError(failmsg + " because can not find file: " + url)
|
|
return None
|
|
response = self._request(url, failmsg, force)
|
|
if response:
|
|
return response.read()
|
|
return None
|
|
|
|
# only for HTTP request
|
|
def _request(self, url, failmsg, force=True):
|
|
url_to_use = url
|
|
parsed_url = urlparse(url)
|
|
|
|
if parsed_url.scheme == 's3':
|
|
parsed_url = urlparse(url)
|
|
bucket_name = parsed_url.netloc
|
|
key_name = parsed_url.path[1:]
|
|
client = boto3.client('s3', aws_access_key_id=self.module.params.get('username', ''), aws_secret_access_key=self.module.params.get('password', ''))
|
|
url_to_use = client.generate_presigned_url('get_object', Params={'Bucket': bucket_name, 'Key': key_name}, ExpiresIn=10)
|
|
|
|
req_timeout = self.module.params.get('timeout')
|
|
|
|
# Hack to add parameters in the way that fetch_url expects
|
|
self.module.params['url_username'] = self.module.params.get('username', '')
|
|
self.module.params['url_password'] = self.module.params.get('password', '')
|
|
self.module.params['http_agent'] = self.user_agent
|
|
|
|
response, info = fetch_url(self.module, url_to_use, timeout=req_timeout, headers=self.headers)
|
|
if info['status'] == 200:
|
|
return response
|
|
if force:
|
|
raise ValueError(failmsg + " because of " + info['msg'] + "for URL " + url_to_use)
|
|
return None
|
|
|
|
def download(self, tmpdir, artifact, verify_download, filename=None):
|
|
if (not artifact.version and not artifact.version_by_spec) or artifact.version == "latest":
|
|
artifact = Artifact(artifact.group_id, artifact.artifact_id, self.find_latest_version_available(artifact), None,
|
|
artifact.classifier, artifact.extension)
|
|
url = self.find_uri_for_artifact(artifact)
|
|
tempfd, tempname = tempfile.mkstemp(dir=tmpdir)
|
|
|
|
try:
|
|
# copy to temp file
|
|
if self.local:
|
|
parsed_url = urlparse(url)
|
|
if os.path.isfile(parsed_url.path):
|
|
shutil.copy2(parsed_url.path, tempname)
|
|
else:
|
|
return "Can not find local file: " + parsed_url.path
|
|
else:
|
|
response = self._request(url, "Failed to download artifact " + str(artifact))
|
|
with os.fdopen(tempfd, 'wb') as f:
|
|
shutil.copyfileobj(response, f)
|
|
|
|
if verify_download:
|
|
invalid_md5 = self.is_invalid_md5(tempname, url)
|
|
if invalid_md5:
|
|
# if verify_change was set, the previous file would be deleted
|
|
os.remove(tempname)
|
|
return invalid_md5
|
|
except Exception as e:
|
|
os.remove(tempname)
|
|
raise e
|
|
|
|
# all good, now copy temp file to target
|
|
shutil.move(tempname, artifact.get_filename(filename))
|
|
return None
|
|
|
|
def is_invalid_md5(self, file, remote_url):
|
|
if os.path.exists(file):
|
|
local_md5 = self._local_md5(file)
|
|
if self.local:
|
|
parsed_url = urlparse(remote_url)
|
|
remote_md5 = self._local_md5(parsed_url.path)
|
|
else:
|
|
try:
|
|
remote_md5 = to_text(self._getContent(remote_url + '.md5', "Failed to retrieve MD5", False), errors='strict')
|
|
except UnicodeError as e:
|
|
return "Cannot retrieve a valid md5 from %s: %s" % (remote_url, to_native(e))
|
|
if(not remote_md5):
|
|
return "Cannot find md5 from " + remote_url
|
|
try:
|
|
# Check if remote md5 only contains md5 or md5 + filename
|
|
_remote_md5 = remote_md5.split(None)[0]
|
|
remote_md5 = _remote_md5
|
|
# remote_md5 is empty so we continue and keep original md5 string
|
|
# This should not happen since we check for remote_md5 before
|
|
except IndexError:
|
|
pass
|
|
if local_md5.lower() == remote_md5.lower():
|
|
return None
|
|
else:
|
|
return "Checksum does not match: we computed " + local_md5 + " but the repository states " + remote_md5
|
|
|
|
return "Path does not exist: " + file
|
|
|
|
def _local_md5(self, file):
|
|
md5 = hashlib.md5()
|
|
with io.open(file, 'rb') as f:
|
|
for chunk in iter(lambda: f.read(8192), b''):
|
|
md5.update(chunk)
|
|
return md5.hexdigest()
|
|
|
|
|
|
def main():
|
|
module = AnsibleModule(
|
|
argument_spec=dict(
|
|
group_id=dict(required=True),
|
|
artifact_id=dict(required=True),
|
|
version=dict(default=None),
|
|
version_by_spec=dict(default=None),
|
|
classifier=dict(default=''),
|
|
extension=dict(default='jar'),
|
|
repository_url=dict(default='https://repo1.maven.org/maven2'),
|
|
username=dict(default=None, aliases=['aws_secret_key']),
|
|
password=dict(default=None, no_log=True, aliases=['aws_secret_access_key']),
|
|
headers=dict(type='dict'),
|
|
force_basic_auth=dict(default=False, type='bool'),
|
|
state=dict(default="present", choices=["present", "absent"]), # TODO - Implement a "latest" state
|
|
timeout=dict(default=10, type='int'),
|
|
dest=dict(type="path", required=True),
|
|
validate_certs=dict(required=False, default=True, type='bool'),
|
|
client_cert=dict(type="path", required=False),
|
|
client_key=dict(type="path", required=False),
|
|
keep_name=dict(required=False, default=False, type='bool'),
|
|
verify_checksum=dict(required=False, default='download', choices=['never', 'download', 'change', 'always']),
|
|
directory_mode=dict(type='str'), # Used since https://github.com/ansible/ansible/pull/24965, not sure
|
|
# if this should really be here.
|
|
),
|
|
add_file_common_args=True,
|
|
mutually_exclusive=([('version', 'version_by_spec')])
|
|
)
|
|
|
|
if not HAS_LXML_ETREE:
|
|
module.fail_json(msg=missing_required_lib('lxml'), exception=LXML_ETREE_IMP_ERR)
|
|
|
|
if module.params['version_by_spec'] and not HAS_SEMANTIC_VERSION:
|
|
module.fail_json(msg=missing_required_lib('semantic_version'), exception=SEMANTIC_VERSION_IMP_ERR)
|
|
|
|
repository_url = module.params["repository_url"]
|
|
if not repository_url:
|
|
repository_url = "https://repo1.maven.org/maven2"
|
|
try:
|
|
parsed_url = urlparse(repository_url)
|
|
except AttributeError as e:
|
|
module.fail_json(msg='url parsing went wrong %s' % e)
|
|
|
|
local = parsed_url.scheme == "file"
|
|
|
|
if parsed_url.scheme == 's3' and not HAS_BOTO:
|
|
module.fail_json(msg=missing_required_lib('boto3', reason='when using s3:// repository URLs'),
|
|
exception=BOTO_IMP_ERR)
|
|
|
|
group_id = module.params["group_id"]
|
|
artifact_id = module.params["artifact_id"]
|
|
version = module.params["version"]
|
|
version_by_spec = module.params["version_by_spec"]
|
|
classifier = module.params["classifier"]
|
|
extension = module.params["extension"]
|
|
headers = module.params['headers']
|
|
state = module.params["state"]
|
|
dest = module.params["dest"]
|
|
b_dest = to_bytes(dest, errors='surrogate_or_strict')
|
|
keep_name = module.params["keep_name"]
|
|
verify_checksum = module.params["verify_checksum"]
|
|
verify_download = verify_checksum in ['download', 'always']
|
|
verify_change = verify_checksum in ['change', 'always']
|
|
|
|
downloader = MavenDownloader(module, repository_url, local, headers)
|
|
|
|
if not version_by_spec and not version:
|
|
version = "latest"
|
|
|
|
try:
|
|
artifact = Artifact(group_id, artifact_id, version, version_by_spec, classifier, extension)
|
|
except ValueError as e:
|
|
module.fail_json(msg=e.args[0])
|
|
|
|
changed = False
|
|
prev_state = "absent"
|
|
|
|
if dest.endswith(os.sep):
|
|
b_dest = to_bytes(dest, errors='surrogate_or_strict')
|
|
if not os.path.exists(b_dest):
|
|
(pre_existing_dir, new_directory_list) = split_pre_existing_dir(dest)
|
|
os.makedirs(b_dest)
|
|
directory_args = module.load_file_common_arguments(module.params)
|
|
directory_mode = module.params["directory_mode"]
|
|
if directory_mode is not None:
|
|
directory_args['mode'] = directory_mode
|
|
else:
|
|
directory_args['mode'] = None
|
|
changed = adjust_recursive_directory_permissions(pre_existing_dir, new_directory_list, module, directory_args, changed)
|
|
|
|
if os.path.isdir(b_dest):
|
|
version_part = version
|
|
if version == 'latest':
|
|
version_part = downloader.find_latest_version_available(artifact)
|
|
elif version_by_spec:
|
|
version_part = downloader.find_version_by_spec(artifact)
|
|
|
|
filename = "{artifact_id}{version_part}{classifier}.{extension}".format(
|
|
artifact_id=artifact_id,
|
|
version_part="-{0}".format(version_part) if keep_name else "",
|
|
classifier="-{0}".format(classifier) if classifier else "",
|
|
extension=extension
|
|
)
|
|
dest = posixpath.join(dest, filename)
|
|
|
|
b_dest = to_bytes(dest, errors='surrogate_or_strict')
|
|
|
|
if os.path.lexists(b_dest) and ((not verify_change) or not downloader.is_invalid_md5(dest, downloader.find_uri_for_artifact(artifact))):
|
|
prev_state = "present"
|
|
|
|
if prev_state == "absent":
|
|
try:
|
|
download_error = downloader.download(module.tmpdir, artifact, verify_download, b_dest)
|
|
if download_error is None:
|
|
changed = True
|
|
else:
|
|
module.fail_json(msg="Cannot retrieve the artifact to destination: " + download_error)
|
|
except ValueError as e:
|
|
module.fail_json(msg=e.args[0])
|
|
|
|
try:
|
|
file_args = module.load_file_common_arguments(module.params, path=dest)
|
|
except TypeError:
|
|
# The path argument is only supported in Ansible-base 2.10+. Fall back to
|
|
# pre-2.10 behavior for older Ansible versions.
|
|
module.params['path'] = dest
|
|
file_args = module.load_file_common_arguments(module.params)
|
|
changed = module.set_fs_attributes_if_different(file_args, changed)
|
|
if changed:
|
|
module.exit_json(state=state, dest=dest, group_id=group_id, artifact_id=artifact_id, version=version, classifier=classifier,
|
|
extension=extension, repository_url=repository_url, changed=changed)
|
|
else:
|
|
module.exit_json(state=state, dest=dest, changed=changed)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
main()
|