2020-03-09 10:11:07 +01:00
|
|
|
#!/usr/bin/python
|
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
|
|
|
# Copyright: (c) 2016, Ben Doherty <bendohmv@gmail.com>
|
|
|
|
# Sponsored by Oomph, Inc. http://www.oomphinc.com
|
|
|
|
# Copyright: (c) 2017, Ansible Project
|
|
|
|
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
|
|
|
|
|
|
|
from __future__ import absolute_import, division, print_function
|
|
|
|
__metaclass__ = type
|
|
|
|
|
|
|
|
DOCUMENTATION = r'''
|
|
|
|
---
|
|
|
|
module: archive
|
|
|
|
short_description: Creates a compressed archive of one or more files or trees
|
|
|
|
extends_documentation_fragment: files
|
|
|
|
description:
|
|
|
|
- Creates or extends an archive.
|
|
|
|
- The source and archive are on the remote host, and the archive I(is not) copied to the local host.
|
|
|
|
- Source files can be deleted after archival by specifying I(remove=True).
|
|
|
|
options:
|
|
|
|
path:
|
|
|
|
description:
|
|
|
|
- Remote absolute path, glob, or list of paths or globs for the file or files to compress or archive.
|
|
|
|
type: list
|
2020-11-24 08:33:19 +01:00
|
|
|
elements: path
|
2020-03-09 10:11:07 +01:00
|
|
|
required: true
|
|
|
|
format:
|
|
|
|
description:
|
|
|
|
- The type of compression to use.
|
|
|
|
- Support for xz was added in Ansible 2.5.
|
|
|
|
type: str
|
|
|
|
choices: [ bz2, gz, tar, xz, zip ]
|
|
|
|
default: gz
|
|
|
|
dest:
|
|
|
|
description:
|
|
|
|
- The file name of the destination archive. The parent directory must exists on the remote host.
|
|
|
|
- This is required when C(path) refers to multiple files by either specifying a glob, a directory or multiple paths in a list.
|
2021-03-22 20:55:47 +01:00
|
|
|
- If the destination archive already exists, it will be truncated and overwritten.
|
2020-03-09 10:11:07 +01:00
|
|
|
type: path
|
|
|
|
exclude_path:
|
|
|
|
description:
|
2020-07-28 10:42:17 +02:00
|
|
|
- Remote absolute path, glob, or list of paths or globs for the file or files to exclude from I(path) list and glob expansion.
|
2020-03-09 10:11:07 +01:00
|
|
|
type: list
|
2020-11-24 08:33:19 +01:00
|
|
|
elements: path
|
2020-03-09 10:11:07 +01:00
|
|
|
force_archive:
|
|
|
|
description:
|
2021-03-22 20:55:35 +01:00
|
|
|
- Allows you to force the module to treat this as an archive even if only a single file is specified.
|
|
|
|
- By default when a single file is specified it is compressed only (not archived).
|
|
|
|
- Enable this if you want to use M(ansible.builtin.unarchive) on an archive of a single file created with this module.
|
2020-03-09 10:11:07 +01:00
|
|
|
type: bool
|
|
|
|
default: false
|
|
|
|
remove:
|
|
|
|
description:
|
|
|
|
- Remove any added source files and trees after adding to archive.
|
|
|
|
type: bool
|
|
|
|
default: no
|
|
|
|
notes:
|
|
|
|
- Requires tarfile, zipfile, gzip and bzip2 packages on target host.
|
|
|
|
- Requires lzma or backports.lzma if using xz format.
|
|
|
|
- Can produce I(gzip), I(bzip2), I(lzma) and I(zip) compressed files or archives.
|
|
|
|
seealso:
|
2020-06-26 13:52:01 +02:00
|
|
|
- module: ansible.builtin.unarchive
|
2020-03-09 10:11:07 +01:00
|
|
|
author:
|
|
|
|
- Ben Doherty (@bendoh)
|
|
|
|
'''
|
|
|
|
|
|
|
|
EXAMPLES = r'''
|
|
|
|
- name: Compress directory /path/to/foo/ into /path/to/foo.tgz
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path: /path/to/foo
|
|
|
|
dest: /path/to/foo.tgz
|
|
|
|
|
|
|
|
- name: Compress regular file /path/to/foo into /path/to/foo.gz and remove it
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path: /path/to/foo
|
|
|
|
remove: yes
|
|
|
|
|
|
|
|
- name: Create a zip archive of /path/to/foo
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path: /path/to/foo
|
|
|
|
format: zip
|
|
|
|
|
|
|
|
- name: Create a bz2 archive of multiple files, rooted at /path
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path:
|
|
|
|
- /path/to/foo
|
|
|
|
- /path/wong/foo
|
|
|
|
dest: /path/file.tar.bz2
|
|
|
|
format: bz2
|
|
|
|
|
|
|
|
- name: Create a bz2 archive of a globbed path, while excluding specific dirnames
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path:
|
|
|
|
- /path/to/foo/*
|
|
|
|
dest: /path/file.tar.bz2
|
|
|
|
exclude_path:
|
|
|
|
- /path/to/foo/bar
|
|
|
|
- /path/to/foo/baz
|
|
|
|
format: bz2
|
|
|
|
|
|
|
|
- name: Create a bz2 archive of a globbed path, while excluding a glob of dirnames
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path:
|
|
|
|
- /path/to/foo/*
|
|
|
|
dest: /path/file.tar.bz2
|
|
|
|
exclude_path:
|
|
|
|
- /path/to/foo/ba*
|
|
|
|
format: bz2
|
|
|
|
|
|
|
|
- name: Use gzip to compress a single archive (i.e don't archive it first with tar)
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path: /path/to/foo/single.file
|
|
|
|
dest: /path/file.gz
|
|
|
|
format: gz
|
|
|
|
|
|
|
|
- name: Create a tar.gz archive of a single file.
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.archive:
|
2020-03-09 10:11:07 +01:00
|
|
|
path: /path/to/foo/single.file
|
|
|
|
dest: /path/file.tar.gz
|
|
|
|
format: gz
|
|
|
|
force_archive: true
|
|
|
|
'''
|
|
|
|
|
|
|
|
RETURN = r'''
|
|
|
|
state:
|
|
|
|
description:
|
|
|
|
The current state of the archived file.
|
|
|
|
If 'absent', then no source files were found and the archive does not exist.
|
|
|
|
If 'compress', then the file source file is in the compressed state.
|
|
|
|
If 'archive', then the source file or paths are currently archived.
|
|
|
|
If 'incomplete', then an archive was created, but not all source paths were found.
|
|
|
|
type: str
|
|
|
|
returned: always
|
|
|
|
missing:
|
|
|
|
description: Any files that were missing from the source.
|
|
|
|
type: list
|
|
|
|
returned: success
|
|
|
|
archived:
|
|
|
|
description: Any files that were compressed or added to the archive.
|
|
|
|
type: list
|
|
|
|
returned: success
|
|
|
|
arcroot:
|
|
|
|
description: The archive root.
|
|
|
|
type: str
|
|
|
|
returned: always
|
|
|
|
expanded_paths:
|
|
|
|
description: The list of matching paths from paths argument.
|
|
|
|
type: list
|
|
|
|
returned: always
|
|
|
|
expanded_exclude_paths:
|
|
|
|
description: The list of matching exclude paths from the exclude_path argument.
|
|
|
|
type: list
|
|
|
|
returned: always
|
|
|
|
'''
|
|
|
|
|
|
|
|
import bz2
|
|
|
|
import filecmp
|
|
|
|
import glob
|
|
|
|
import gzip
|
|
|
|
import io
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import tarfile
|
|
|
|
import zipfile
|
|
|
|
from traceback import format_exc
|
|
|
|
|
|
|
|
from ansible.module_utils.basic import AnsibleModule, missing_required_lib
|
|
|
|
from ansible.module_utils._text import to_bytes, to_native
|
|
|
|
from ansible.module_utils.six import PY3
|
|
|
|
|
|
|
|
|
|
|
|
LZMA_IMP_ERR = None
|
|
|
|
if PY3:
|
|
|
|
try:
|
|
|
|
import lzma
|
|
|
|
HAS_LZMA = True
|
|
|
|
except ImportError:
|
|
|
|
LZMA_IMP_ERR = format_exc()
|
|
|
|
HAS_LZMA = False
|
|
|
|
else:
|
|
|
|
try:
|
|
|
|
from backports import lzma
|
|
|
|
HAS_LZMA = True
|
|
|
|
except ImportError:
|
|
|
|
LZMA_IMP_ERR = format_exc()
|
|
|
|
HAS_LZMA = False
|
|
|
|
|
|
|
|
|
|
|
|
def main():
|
|
|
|
module = AnsibleModule(
|
|
|
|
argument_spec=dict(
|
2020-11-24 08:33:19 +01:00
|
|
|
path=dict(type='list', elements='path', required=True),
|
2020-03-09 10:11:07 +01:00
|
|
|
format=dict(type='str', default='gz', choices=['bz2', 'gz', 'tar', 'xz', 'zip']),
|
|
|
|
dest=dict(type='path'),
|
2020-11-24 08:33:19 +01:00
|
|
|
exclude_path=dict(type='list', elements='path'),
|
2020-03-09 10:11:07 +01:00
|
|
|
force_archive=dict(type='bool', default=False),
|
|
|
|
remove=dict(type='bool', default=False),
|
|
|
|
),
|
|
|
|
add_file_common_args=True,
|
|
|
|
supports_check_mode=True,
|
|
|
|
)
|
|
|
|
|
|
|
|
params = module.params
|
|
|
|
check_mode = module.check_mode
|
|
|
|
paths = params['path']
|
|
|
|
dest = params['dest']
|
|
|
|
b_dest = None if not dest else to_bytes(dest, errors='surrogate_or_strict')
|
|
|
|
exclude_paths = params['exclude_path']
|
|
|
|
remove = params['remove']
|
|
|
|
|
|
|
|
b_expanded_paths = []
|
|
|
|
b_expanded_exclude_paths = []
|
|
|
|
fmt = params['format']
|
|
|
|
b_fmt = to_bytes(fmt, errors='surrogate_or_strict')
|
|
|
|
force_archive = params['force_archive']
|
|
|
|
globby = False
|
|
|
|
changed = False
|
|
|
|
state = 'absent'
|
|
|
|
|
|
|
|
# Simple or archive file compression (inapplicable with 'zip' since it's always an archive)
|
|
|
|
archive = False
|
|
|
|
b_successes = []
|
|
|
|
|
|
|
|
# Fail early
|
|
|
|
if not HAS_LZMA and fmt == 'xz':
|
|
|
|
module.fail_json(msg=missing_required_lib("lzma or backports.lzma", reason="when using xz format"),
|
|
|
|
exception=LZMA_IMP_ERR)
|
|
|
|
module.fail_json(msg="lzma or backports.lzma is required when using xz format.")
|
|
|
|
|
|
|
|
for path in paths:
|
2020-11-24 08:33:19 +01:00
|
|
|
b_path = to_bytes(path, errors='surrogate_or_strict')
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Expand any glob characters. If found, add the expanded glob to the
|
|
|
|
# list of expanded_paths, which might be empty.
|
|
|
|
if (b'*' in b_path or b'?' in b_path):
|
|
|
|
b_expanded_paths.extend(glob.glob(b_path))
|
|
|
|
globby = True
|
|
|
|
|
|
|
|
# If there are no glob characters the path is added to the expanded paths
|
|
|
|
# whether the path exists or not
|
|
|
|
else:
|
|
|
|
b_expanded_paths.append(b_path)
|
|
|
|
|
|
|
|
# Only attempt to expand the exclude paths if it exists
|
|
|
|
if exclude_paths:
|
|
|
|
for exclude_path in exclude_paths:
|
2020-11-24 08:33:19 +01:00
|
|
|
b_exclude_path = to_bytes(exclude_path, errors='surrogate_or_strict')
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
# Expand any glob characters. If found, add the expanded glob to the
|
|
|
|
# list of expanded_paths, which might be empty.
|
|
|
|
if (b'*' in b_exclude_path or b'?' in b_exclude_path):
|
|
|
|
b_expanded_exclude_paths.extend(glob.glob(b_exclude_path))
|
|
|
|
|
|
|
|
# If there are no glob character the exclude path is added to the expanded
|
|
|
|
# exclude paths whether the path exists or not.
|
|
|
|
else:
|
|
|
|
b_expanded_exclude_paths.append(b_exclude_path)
|
|
|
|
|
|
|
|
if not b_expanded_paths:
|
|
|
|
return module.fail_json(
|
|
|
|
path=', '.join(paths),
|
|
|
|
expanded_paths=to_native(b', '.join(b_expanded_paths), errors='surrogate_or_strict'),
|
|
|
|
msg='Error, no source paths were found'
|
|
|
|
)
|
|
|
|
|
|
|
|
# Only try to determine if we are working with an archive or not if we haven't set archive to true
|
|
|
|
if not force_archive:
|
|
|
|
# If we actually matched multiple files or TRIED to, then
|
|
|
|
# treat this as a multi-file archive
|
|
|
|
archive = globby or os.path.isdir(b_expanded_paths[0]) or len(b_expanded_paths) > 1
|
|
|
|
else:
|
|
|
|
archive = True
|
|
|
|
|
|
|
|
# Default created file name (for single-file archives) to
|
|
|
|
# <file>.<format>
|
|
|
|
if not b_dest and not archive:
|
|
|
|
b_dest = b'%s.%s' % (b_expanded_paths[0], b_fmt)
|
|
|
|
|
|
|
|
# Force archives to specify 'dest'
|
|
|
|
if archive and not b_dest:
|
|
|
|
module.fail_json(dest=dest, path=', '.join(paths), msg='Error, must specify "dest" when archiving multiple files or trees')
|
|
|
|
|
|
|
|
b_sep = to_bytes(os.sep, errors='surrogate_or_strict')
|
|
|
|
|
|
|
|
b_archive_paths = []
|
|
|
|
b_missing = []
|
|
|
|
b_arcroot = b''
|
|
|
|
|
|
|
|
for b_path in b_expanded_paths:
|
|
|
|
# Use the longest common directory name among all the files
|
|
|
|
# as the archive root path
|
|
|
|
if b_arcroot == b'':
|
|
|
|
b_arcroot = os.path.dirname(b_path) + b_sep
|
|
|
|
else:
|
|
|
|
for i in range(len(b_arcroot)):
|
|
|
|
if b_path[i] != b_arcroot[i]:
|
|
|
|
break
|
|
|
|
|
|
|
|
if i < len(b_arcroot):
|
|
|
|
b_arcroot = os.path.dirname(b_arcroot[0:i + 1])
|
|
|
|
|
|
|
|
b_arcroot += b_sep
|
|
|
|
|
|
|
|
# Don't allow archives to be created anywhere within paths to be removed
|
|
|
|
if remove and os.path.isdir(b_path):
|
|
|
|
b_path_dir = b_path
|
|
|
|
if not b_path.endswith(b'/'):
|
|
|
|
b_path_dir += b'/'
|
|
|
|
|
|
|
|
if b_dest.startswith(b_path_dir):
|
|
|
|
module.fail_json(
|
|
|
|
path=', '.join(paths),
|
|
|
|
msg='Error, created archive can not be contained in source paths when remove=True'
|
|
|
|
)
|
|
|
|
|
|
|
|
if os.path.lexists(b_path) and b_path not in b_expanded_exclude_paths:
|
|
|
|
b_archive_paths.append(b_path)
|
|
|
|
else:
|
|
|
|
b_missing.append(b_path)
|
|
|
|
|
|
|
|
# No source files were found but the named archive exists: are we 'compress' or 'archive' now?
|
|
|
|
if len(b_missing) == len(b_expanded_paths) and b_dest and os.path.exists(b_dest):
|
|
|
|
# Just check the filename to know if it's an archive or simple compressed file
|
|
|
|
if re.search(br'(\.tar|\.tar\.gz|\.tgz|\.tbz2|\.tar\.bz2|\.tar\.xz|\.zip)$', os.path.basename(b_dest), re.IGNORECASE):
|
|
|
|
state = 'archive'
|
|
|
|
else:
|
|
|
|
state = 'compress'
|
|
|
|
|
|
|
|
# Multiple files, or globbiness
|
|
|
|
elif archive:
|
|
|
|
if not b_archive_paths:
|
|
|
|
# No source files were found, but the archive is there.
|
|
|
|
if os.path.lexists(b_dest):
|
|
|
|
state = 'archive'
|
|
|
|
elif b_missing:
|
|
|
|
# SOME source files were found, but not all of them
|
|
|
|
state = 'incomplete'
|
|
|
|
|
|
|
|
archive = None
|
|
|
|
size = 0
|
|
|
|
errors = []
|
|
|
|
|
|
|
|
if os.path.lexists(b_dest):
|
|
|
|
size = os.path.getsize(b_dest)
|
|
|
|
|
|
|
|
if state != 'archive':
|
|
|
|
if check_mode:
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
else:
|
|
|
|
try:
|
|
|
|
# Slightly more difficult (and less efficient!) compression using zipfile module
|
|
|
|
if fmt == 'zip':
|
|
|
|
arcfile = zipfile.ZipFile(
|
|
|
|
to_native(b_dest, errors='surrogate_or_strict', encoding='ascii'),
|
|
|
|
'w',
|
|
|
|
zipfile.ZIP_DEFLATED,
|
|
|
|
True
|
|
|
|
)
|
|
|
|
|
|
|
|
# Easier compression using tarfile module
|
|
|
|
elif fmt == 'gz' or fmt == 'bz2':
|
|
|
|
arcfile = tarfile.open(to_native(b_dest, errors='surrogate_or_strict', encoding='ascii'), 'w|' + fmt)
|
|
|
|
|
|
|
|
# python3 tarfile module allows xz format but for python2 we have to create the tarfile
|
|
|
|
# in memory and then compress it with lzma.
|
|
|
|
elif fmt == 'xz':
|
|
|
|
arcfileIO = io.BytesIO()
|
|
|
|
arcfile = tarfile.open(fileobj=arcfileIO, mode='w')
|
|
|
|
|
|
|
|
# Or plain tar archiving
|
|
|
|
elif fmt == 'tar':
|
|
|
|
arcfile = tarfile.open(to_native(b_dest, errors='surrogate_or_strict', encoding='ascii'), 'w')
|
|
|
|
|
|
|
|
b_match_root = re.compile(br'^%s' % re.escape(b_arcroot))
|
|
|
|
for b_path in b_archive_paths:
|
|
|
|
if os.path.isdir(b_path):
|
|
|
|
# Recurse into directories
|
|
|
|
for b_dirpath, b_dirnames, b_filenames in os.walk(b_path, topdown=True):
|
|
|
|
if not b_dirpath.endswith(b_sep):
|
|
|
|
b_dirpath += b_sep
|
|
|
|
|
|
|
|
for b_dirname in b_dirnames:
|
|
|
|
b_fullpath = b_dirpath + b_dirname
|
|
|
|
n_fullpath = to_native(b_fullpath, errors='surrogate_or_strict', encoding='ascii')
|
|
|
|
n_arcname = to_native(b_match_root.sub(b'', b_fullpath), errors='surrogate_or_strict')
|
|
|
|
|
|
|
|
try:
|
|
|
|
if fmt == 'zip':
|
|
|
|
arcfile.write(n_fullpath, n_arcname)
|
|
|
|
else:
|
|
|
|
arcfile.add(n_fullpath, n_arcname, recursive=False)
|
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
errors.append('%s: %s' % (n_fullpath, to_native(e)))
|
|
|
|
|
|
|
|
for b_filename in b_filenames:
|
|
|
|
b_fullpath = b_dirpath + b_filename
|
|
|
|
n_fullpath = to_native(b_fullpath, errors='surrogate_or_strict', encoding='ascii')
|
|
|
|
n_arcname = to_native(b_match_root.sub(b'', b_fullpath), errors='surrogate_or_strict')
|
|
|
|
|
|
|
|
try:
|
|
|
|
if fmt == 'zip':
|
|
|
|
arcfile.write(n_fullpath, n_arcname)
|
|
|
|
else:
|
|
|
|
arcfile.add(n_fullpath, n_arcname, recursive=False)
|
|
|
|
|
|
|
|
b_successes.append(b_fullpath)
|
|
|
|
except Exception as e:
|
|
|
|
errors.append('Adding %s: %s' % (to_native(b_path), to_native(e)))
|
|
|
|
else:
|
|
|
|
path = to_native(b_path, errors='surrogate_or_strict', encoding='ascii')
|
|
|
|
arcname = to_native(b_match_root.sub(b'', b_path), errors='surrogate_or_strict')
|
|
|
|
if fmt == 'zip':
|
|
|
|
arcfile.write(path, arcname)
|
|
|
|
else:
|
|
|
|
arcfile.add(path, arcname, recursive=False)
|
|
|
|
|
|
|
|
b_successes.append(b_path)
|
|
|
|
|
|
|
|
except Exception as e:
|
|
|
|
expanded_fmt = 'zip' if fmt == 'zip' else ('tar.' + fmt)
|
|
|
|
module.fail_json(
|
|
|
|
msg='Error when writing %s archive at %s: %s' % (expanded_fmt, dest, to_native(e)),
|
|
|
|
exception=format_exc()
|
|
|
|
)
|
|
|
|
|
|
|
|
if arcfile:
|
|
|
|
arcfile.close()
|
|
|
|
state = 'archive'
|
|
|
|
|
|
|
|
if fmt == 'xz':
|
|
|
|
with lzma.open(b_dest, 'wb') as f:
|
|
|
|
f.write(arcfileIO.getvalue())
|
|
|
|
arcfileIO.close()
|
|
|
|
|
|
|
|
if errors:
|
|
|
|
module.fail_json(msg='Errors when writing archive at %s: %s' % (dest, '; '.join(errors)))
|
|
|
|
|
|
|
|
if state in ['archive', 'incomplete'] and remove:
|
|
|
|
for b_path in b_successes:
|
|
|
|
try:
|
|
|
|
if os.path.isdir(b_path):
|
|
|
|
shutil.rmtree(b_path)
|
|
|
|
elif not check_mode:
|
|
|
|
os.remove(b_path)
|
|
|
|
except OSError as e:
|
|
|
|
errors.append(to_native(b_path))
|
|
|
|
|
|
|
|
if errors:
|
|
|
|
module.fail_json(dest=dest, msg='Error deleting some source files: ', files=errors)
|
|
|
|
|
|
|
|
# Rudimentary check: If size changed then file changed. Not perfect, but easy.
|
|
|
|
if not check_mode and os.path.getsize(b_dest) != size:
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
if b_successes and state != 'incomplete':
|
|
|
|
state = 'archive'
|
|
|
|
|
|
|
|
# Simple, single-file compression
|
|
|
|
else:
|
|
|
|
b_path = b_expanded_paths[0]
|
|
|
|
|
|
|
|
# No source or compressed file
|
|
|
|
if not (os.path.exists(b_path) or os.path.lexists(b_dest)):
|
|
|
|
state = 'absent'
|
|
|
|
|
|
|
|
# if it already exists and the source file isn't there, consider this done
|
|
|
|
elif not os.path.lexists(b_path) and os.path.lexists(b_dest):
|
|
|
|
state = 'compress'
|
|
|
|
|
|
|
|
else:
|
|
|
|
if module.check_mode:
|
|
|
|
if not os.path.exists(b_dest):
|
|
|
|
changed = True
|
|
|
|
else:
|
|
|
|
size = 0
|
|
|
|
f_in = f_out = arcfile = None
|
|
|
|
|
|
|
|
if os.path.lexists(b_dest):
|
|
|
|
size = os.path.getsize(b_dest)
|
|
|
|
|
|
|
|
try:
|
|
|
|
if fmt == 'zip':
|
|
|
|
arcfile = zipfile.ZipFile(
|
|
|
|
to_native(b_dest, errors='surrogate_or_strict', encoding='ascii'),
|
|
|
|
'w',
|
|
|
|
zipfile.ZIP_DEFLATED,
|
|
|
|
True
|
|
|
|
)
|
|
|
|
arcfile.write(
|
|
|
|
to_native(b_path, errors='surrogate_or_strict', encoding='ascii'),
|
|
|
|
to_native(b_path[len(b_arcroot):], errors='surrogate_or_strict')
|
|
|
|
)
|
|
|
|
arcfile.close()
|
|
|
|
state = 'archive' # because all zip files are archives
|
|
|
|
elif fmt == 'tar':
|
|
|
|
arcfile = tarfile.open(to_native(b_dest, errors='surrogate_or_strict', encoding='ascii'), 'w')
|
|
|
|
arcfile.add(to_native(b_path, errors='surrogate_or_strict', encoding='ascii'))
|
|
|
|
arcfile.close()
|
|
|
|
else:
|
|
|
|
f_in = open(b_path, 'rb')
|
|
|
|
|
|
|
|
n_dest = to_native(b_dest, errors='surrogate_or_strict', encoding='ascii')
|
|
|
|
if fmt == 'gz':
|
|
|
|
f_out = gzip.open(n_dest, 'wb')
|
|
|
|
elif fmt == 'bz2':
|
|
|
|
f_out = bz2.BZ2File(n_dest, 'wb')
|
|
|
|
elif fmt == 'xz':
|
|
|
|
f_out = lzma.LZMAFile(n_dest, 'wb')
|
|
|
|
else:
|
|
|
|
raise OSError("Invalid format")
|
|
|
|
|
|
|
|
shutil.copyfileobj(f_in, f_out)
|
|
|
|
|
|
|
|
b_successes.append(b_path)
|
|
|
|
|
|
|
|
except OSError as e:
|
|
|
|
module.fail_json(
|
|
|
|
path=to_native(b_path),
|
|
|
|
dest=dest,
|
|
|
|
msg='Unable to write to compressed file: %s' % to_native(e), exception=format_exc()
|
|
|
|
)
|
|
|
|
|
|
|
|
if arcfile:
|
|
|
|
arcfile.close()
|
|
|
|
if f_in:
|
|
|
|
f_in.close()
|
|
|
|
if f_out:
|
|
|
|
f_out.close()
|
|
|
|
|
|
|
|
# Rudimentary check: If size changed then file changed. Not perfect, but easy.
|
|
|
|
if os.path.getsize(b_dest) != size:
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
state = 'compress'
|
|
|
|
|
|
|
|
if remove and not check_mode:
|
|
|
|
try:
|
|
|
|
os.remove(b_path)
|
|
|
|
|
|
|
|
except OSError as e:
|
|
|
|
module.fail_json(
|
|
|
|
path=to_native(b_path),
|
|
|
|
msg='Unable to remove source file: %s' % to_native(e), exception=format_exc()
|
|
|
|
)
|
|
|
|
|
2020-05-08 15:38:23 +02:00
|
|
|
try:
|
|
|
|
file_args = module.load_file_common_arguments(params, path=b_dest)
|
|
|
|
except TypeError:
|
2020-06-16 21:24:42 +02:00
|
|
|
# The path argument is only supported in Ansible-base 2.10+. Fall back to
|
2020-05-08 15:38:23 +02:00
|
|
|
# pre-2.10 behavior for older Ansible versions.
|
|
|
|
params['path'] = b_dest
|
|
|
|
file_args = module.load_file_common_arguments(params)
|
2020-03-09 10:11:07 +01:00
|
|
|
|
|
|
|
if not check_mode:
|
|
|
|
changed = module.set_fs_attributes_if_different(file_args, changed)
|
|
|
|
|
|
|
|
module.exit_json(
|
|
|
|
archived=[to_native(p, errors='surrogate_or_strict') for p in b_successes],
|
|
|
|
dest=dest,
|
|
|
|
changed=changed,
|
|
|
|
state=state,
|
|
|
|
arcroot=to_native(b_arcroot, errors='surrogate_or_strict'),
|
|
|
|
missing=[to_native(p, errors='surrogate_or_strict') for p in b_missing],
|
|
|
|
expanded_paths=[to_native(p, errors='surrogate_or_strict') for p in b_expanded_paths],
|
|
|
|
expanded_exclude_paths=[to_native(p, errors='surrogate_or_strict') for p in b_expanded_exclude_paths],
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
main()
|