1
0
Fork 0
mirror of https://github.com/ansible-collections/community.general.git synced 2024-09-14 20:13:21 +02:00

Initial commit of extras/archive module.

This manages compressed files or archives of many compressed files. You can maintain or update .gz, .bz2 compressed files, .zip archives, or tarballs compressed with gzip or bzip2.

Possible use cases:

* Back up user home directories
* Ensure large text files are always compressed
* Archive trees for distribution
This commit is contained in:
Ben Doherty 2016-05-26 19:54:30 -04:00 committed by Matt Clay
parent f408a4db02
commit 2e1ce01dff

View file

@ -0,0 +1,285 @@
#!/usr/bin/python
# -*- coding: utf-8 -*-
DOCUMENTATION = '''
---
module: archive
version_added: 2.2
short_description: Creates a compressed archive of one or more files or trees.
extends_documentation_fragment: files
description:
- The M(archive) module packs an archive. It is the opposite of the unarchive module. By default, it assumes the compression source exists on the target. It will not copy the source file from the local system to the target before archiving - set copy=yes to pack an archive which does not already exist on the target. The source files are deleted after archiving.
options:
path:
description:
- Remote absolute path, glob, or list of paths or globs for the file or files to archive or compress.
required: false
default: null
compression:
description:
- "The type of compression to use. Can be 'gz', 'bz2', or 'zip'.
choices: [ 'gz', 'bz2', 'zip' ]
creates:
description:
- The file name of the destination archive. When it already exists, this step will B(not) be run. This is required when 'path' refers to multiple files by either specifying a glob, a directory or multiple paths in a list.
required: false
default: null
author: "Ben Doherty (@bendoh)"
notes:
- requires tarfile, zipfile, gzip, and bzip2 packages on target host
- can product I(gzip), I(bzip2) and I(zip) compressed files or archives
- removes source files by default
'''
EXAMPLES = '''
# Compress directory /path/to/foo/ into /path/to/foo.tgz
- archive: path=/path/to/foo creates=/path/to/foo.tgz
# Compress regular file /path/to/foo into /path/to/foo.gz
- archive: path=/path/to/foo
# Create a zip archive of /path/to/foo
- archive: path=/path/to/foo compression=zip
# Create a bz2 archive of multiple files, rooted at /path
- archive:
path:
- /path/to/foo
- /path/wong/foo
creates: /path/file.tar.bz2
compression: bz2
'''
import stat
import os
import errno
import glob
import shutil
import gzip
import bz2
import zipfile
import tarfile
def main():
module = AnsibleModule(
argument_spec = dict(
path = dict(required=True),
compression = dict(choices=['gz', 'bz2', 'zip'], default='gz', required=False),
creates = dict(required=False),
remove = dict(required=False, default=True, type='bool'),
),
add_file_common_args=True,
supports_check_mode=True,
)
params = module.params
paths = params['path']
creates = params['creates']
remove = params['remove']
expanded_paths = []
compression = params['compression']
globby = False
changed = False
state = 'absent'
# Simple or archive file compression (inapplicable with 'zip')
archive = False
successes = []
if isinstance(paths, basestring):
paths = [paths]
for i, path in enumerate(paths):
path = os.path.expanduser(params['path'])
# Detect glob-like characters
if any((c in set('*?')) for c in path):
expanded_paths = expanded_paths + glob.glob(path)
else:
expanded_paths.append(path)
if len(expanded_paths) == 0:
module.fail_json(path, msg='Error, no source paths were found')
# If we actually matched multiple files or TRIED to, then
# treat this as a multi-file archive
archive = globby or len(expanded_paths) > 1 or any(os.path.isdir(path) for path in expanded_paths)
# Default created file name (for single-file archives) to
# <file>.<compression>
if not archive and not creates:
creates = '%s.%s' % (expanded_paths[0], compression)
# Force archives to specify 'creates'
if archive and not creates:
module.fail_json(creates=creates, path=', '.join(paths), msg='Error, must specify "creates" when archiving multiple files or trees')
archive_paths = []
missing = []
arcroot = ''
for path in expanded_paths:
# Use the longest common directory name among all the files
# as the archive root path
if arcroot == '':
arcroot = os.path.dirname(path)
else:
for i in xrange(len(arcroot)):
if path[i] != arcroot[i]:
break
if i < len(arcroot):
arcroot = os.path.dirname(arcroot[0:i+1])
if path == creates:
# Don't allow the archive to specify itself! this is an error.
module.fail_json(path=', '.join(paths), msg='Error, created archive would be included in archive')
if os.path.lexists(path):
archive_paths.append(path)
else:
missing.append(path)
# No source files were found but the named archive exists: are we 'compress' or 'archive' now?
if len(missing) == len(expanded_paths) and creates and os.path.exists(creates):
# Just check the filename to know if it's an archive or simple compressed file
if re.search(r'(\.tar\.gz|\.tgz|.tbz2|\.tar\.bz2|\.zip)$', os.path.basename(creates), re.IGNORECASE):
state = 'archive'
else:
state = 'compress'
# Multiple files, or globbiness
elif archive:
if len(archive_paths) == 0:
# No source files were found, but the archive is there.
if os.path.lexists(creates):
state = 'archive'
elif len(missing) > 0:
# SOME source files were found, but not all of them
state = 'incomplete'
archive = None
size = 0
errors = []
if os.path.lexists(creates):
size = os.path.getsize(creates)
if state != 'archive':
try:
if compression == 'gz' or compression == 'bz2':
archive = tarfile.open(creates, 'w|' + compression)
for path in archive_paths:
archive.add(path, path[len(arcroot):])
successes.append(path)
elif compression == 'zip':
archive = zipfile.ZipFile(creates, 'wb')
for path in archive_paths:
archive.write(path, path[len(arcroot):])
successes.append(path)
except OSError:
e = get_exception()
module.fail_json(msg='Error when writing zip archive at %s: %s' % (creates, str(e)))
if archive:
archive.close()
if state != 'archive' and remove:
for path in successes:
try:
if os.path.isdir(path):
shutil.rmtree(path)
else:
os.remove(path)
except OSError:
e = get_exception()
errors.append(path)
if len(errors) > 0:
module.fail_json(creates=creates, msg='Error deleting some source files: ' + str(e), files=errors)
# Rudimentary check: If size changed then file changed. Not perfect, but easy.
if os.path.getsize(creates) != size:
changed = True
if len(successes) and state != 'incomplete':
state = 'archive'
# Simple, single-file compression
else:
path = expanded_paths[0]
# No source or compressed file
if not (os.path.exists(path) or os.path.lexists(creates)):
state = 'absent'
# if it already exists and the source file isn't there, consider this done
elif not os.path.lexists(path) and os.path.lexists(creates):
state = 'compress'
else:
if module.check_mode:
if not os.path.exists(creates):
changed = True
else:
size = 0
f_in = f_out = archive = None
if os.path.lexists(creates):
size = os.path.getsize(creates)
try:
if compression == 'zip':
archive = zipfile.ZipFile(creates, 'wb')
archive.write(path, path[len(arcroot):])
archive.close()
state = 'archive' # because all zip files are archives
else:
f_in = open(path, 'rb')
if compression == 'gz':
f_out = gzip.open(creates, 'wb')
elif compression == 'bz2':
f_out = bz2.BZ2File(creates, 'wb')
else:
raise OSError("Invalid compression")
shutil.copyfileobj(f_in, f_out)
except OSError:
e = get_exception()
module.fail_json(path=path, creates=creates, msg='Unable to write to compressed file: %s' % str(e))
if archive:
archive.close()
if f_in:
f_in.close()
if f_out:
f_out.close()
# Rudimentary check: If size changed then file changed. Not perfect, but easy.
if os.path.getsize(creates) != size:
changed = True
state = 'compress'
if remove:
try:
os.remove(path)
except OSError:
e = get_exception()
module.fail_json(path=path, msg='Unable to remove source file: %s' % str(e))
module.exit_json(archived=successes, creates=creates, changed=changed, state=state, arcroot=arcroot, missing=missing, expanded_paths=expanded_paths)
# import module snippets
from ansible.module_utils.basic import *
if __name__ == '__main__':
main()