#!/usr/bin/python # -*- coding: utf-8 -*- # Copyright: (c) 2016, Ben Doherty # Sponsored by Oomph, Inc. http://www.oomphinc.com # Copyright: (c) 2017, Ansible Project # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import absolute_import, division, print_function __metaclass__ = type DOCUMENTATION = r''' --- module: archive short_description: Creates a compressed archive of one or more files or trees extends_documentation_fragment: files description: - Creates or extends an archive. - The source and archive are on the remote host, and the archive I(is not) copied to the local host. - Source files can be deleted after archival by specifying I(remove=True). options: path: description: - Remote absolute path, glob, or list of paths or globs for the file or files to compress or archive. type: list elements: path required: true format: description: - The type of compression to use. - Support for xz was added in Ansible 2.5. type: str choices: [ bz2, gz, tar, xz, zip ] default: gz dest: description: - The file name of the destination archive. The parent directory must exists on the remote host. - This is required when C(path) refers to multiple files by either specifying a glob, a directory or multiple paths in a list. - If the destination archive already exists, it will be truncated and overwritten. type: path exclude_path: description: - Remote absolute path, glob, or list of paths or globs for the file or files to exclude from I(path) list and glob expansion. - Use I(exclusion_patterns) to instead exclude files or subdirectories below any of the paths from the I(path) list. type: list elements: path default: [] exclusion_patterns: description: - Glob style patterns to exclude files or directories from the resulting archive. - This differs from I(exclude_path) which applies only to the source paths from I(path). type: list elements: path version_added: 3.2.0 force_archive: description: - Allows you to force the module to treat this as an archive even if only a single file is specified. - By default when a single file is specified it is compressed only (not archived). - Enable this if you want to use M(ansible.builtin.unarchive) on an archive of a single file created with this module. type: bool default: false remove: description: - Remove any added source files and trees after adding to archive. type: bool default: no notes: - Requires tarfile, zipfile, gzip and bzip2 packages on target host. - Requires lzma or backports.lzma if using xz format. - Can produce I(gzip), I(bzip2), I(lzma) and I(zip) compressed files or archives. seealso: - module: ansible.builtin.unarchive author: - Ben Doherty (@bendoh) ''' EXAMPLES = r''' - name: Compress directory /path/to/foo/ into /path/to/foo.tgz community.general.archive: path: /path/to/foo dest: /path/to/foo.tgz - name: Compress regular file /path/to/foo into /path/to/foo.gz and remove it community.general.archive: path: /path/to/foo remove: yes - name: Create a zip archive of /path/to/foo community.general.archive: path: /path/to/foo format: zip - name: Create a bz2 archive of multiple files, rooted at /path community.general.archive: path: - /path/to/foo - /path/wong/foo dest: /path/file.tar.bz2 format: bz2 - name: Create a bz2 archive of a globbed path, while excluding specific dirnames community.general.archive: path: - /path/to/foo/* dest: /path/file.tar.bz2 exclude_path: - /path/to/foo/bar - /path/to/foo/baz format: bz2 - name: Create a bz2 archive of a globbed path, while excluding a glob of dirnames community.general.archive: path: - /path/to/foo/* dest: /path/file.tar.bz2 exclude_path: - /path/to/foo/ba* format: bz2 - name: Use gzip to compress a single archive (i.e don't archive it first with tar) community.general.archive: path: /path/to/foo/single.file dest: /path/file.gz format: gz - name: Create a tar.gz archive of a single file. community.general.archive: path: /path/to/foo/single.file dest: /path/file.tar.gz format: gz force_archive: true ''' RETURN = r''' state: description: The state of the input C(path). type: str returned: always dest_state: description: - The state of the I(dest) file. - C(absent) when the file does not exist. - C(archive) when the file is an archive. - C(compress) when the file is compressed, but not an archive. - C(incomplete) when the file is an archive, but some files under I(path) were not found. type: str returned: success version_added: 3.4.0 missing: description: Any files that were missing from the source. type: list returned: success archived: description: Any files that were compressed or added to the archive. type: list returned: success arcroot: description: The archive root. type: str returned: always expanded_paths: description: The list of matching paths from paths argument. type: list returned: always expanded_exclude_paths: description: The list of matching exclude paths from the exclude_path argument. type: list returned: always ''' import abc import bz2 import glob import gzip import io import os import re import shutil import tarfile import zipfile from fnmatch import fnmatch from sys import version_info from traceback import format_exc from zlib import crc32 from ansible.module_utils.basic import AnsibleModule, missing_required_lib from ansible.module_utils.common.text.converters import to_bytes, to_native from ansible.module_utils import six LZMA_IMP_ERR = None if six.PY3: try: import lzma HAS_LZMA = True except ImportError: LZMA_IMP_ERR = format_exc() HAS_LZMA = False else: try: from backports import lzma HAS_LZMA = True except ImportError: LZMA_IMP_ERR = format_exc() HAS_LZMA = False PY27 = version_info[0:2] >= (2, 7) STATE_ABSENT = 'absent' STATE_ARCHIVED = 'archive' STATE_COMPRESSED = 'compress' STATE_INCOMPLETE = 'incomplete' def common_path(paths): empty = b'' if paths and isinstance(paths[0], six.binary_type) else '' return os.path.join( os.path.dirname(os.path.commonprefix([os.path.join(os.path.dirname(p), empty) for p in paths])), empty ) def expand_paths(paths): expanded_path = [] is_globby = False for path in paths: b_path = _to_bytes(path) if b'*' in b_path or b'?' in b_path: e_paths = glob.glob(b_path) is_globby = True else: e_paths = [b_path] expanded_path.extend(e_paths) return expanded_path, is_globby def matches_exclusion_patterns(path, exclusion_patterns): return any(fnmatch(path, p) for p in exclusion_patterns) def is_archive(path): return re.search(br'\.(tar|tar\.(gz|bz2|xz)|tgz|tbz2|zip)$', os.path.basename(path), re.IGNORECASE) def strip_prefix(prefix, string): return string[len(prefix):] if string.startswith(prefix) else string def _to_bytes(s): return to_bytes(s, errors='surrogate_or_strict') def _to_native(s): return to_native(s, errors='surrogate_or_strict') def _to_native_ascii(s): return to_native(s, errors='surrogate_or_strict', encoding='ascii') @six.add_metaclass(abc.ABCMeta) class Archive(object): def __init__(self, module): self.module = module self.destination = _to_bytes(module.params['dest']) if module.params['dest'] else None self.exclusion_patterns = module.params['exclusion_patterns'] or [] self.format = module.params['format'] self.must_archive = module.params['force_archive'] self.remove = module.params['remove'] self.changed = False self.destination_state = STATE_ABSENT self.errors = [] self.file = None self.successes = [] self.targets = [] self.not_found = [] paths = module.params['path'] self.expanded_paths, has_globs = expand_paths(paths) self.expanded_exclude_paths = expand_paths(module.params['exclude_path'])[0] self.paths = sorted(set(self.expanded_paths) - set(self.expanded_exclude_paths)) if not self.paths: module.fail_json( path=', '.join(paths), expanded_paths=_to_native(b', '.join(self.expanded_paths)), expanded_exclude_paths=_to_native(b', '.join(self.expanded_exclude_paths)), msg='Error, no source paths were found' ) self.root = common_path(self.paths) if not self.must_archive: self.must_archive = any([has_globs, os.path.isdir(self.paths[0]), len(self.paths) > 1]) if not self.destination and not self.must_archive: self.destination = b'%s.%s' % (self.paths[0], _to_bytes(self.format)) if self.must_archive and not self.destination: module.fail_json( dest=_to_native(self.destination), path=', '.join(paths), msg='Error, must specify "dest" when archiving multiple files or trees' ) if self.remove: self._check_removal_safety() self.original_checksums = self.destination_checksums() self.original_size = self.destination_size() def add(self, path, archive_name): try: self._add(_to_native_ascii(path), _to_native(archive_name)) if self.contains(_to_native(archive_name)): self.successes.append(path) except Exception as e: self.errors.append('%s: %s' % (_to_native_ascii(path), _to_native(e))) def add_single_target(self, path): if self.format in ('zip', 'tar'): self.open() self.add(path, strip_prefix(self.root, path)) self.close() self.destination_state = STATE_ARCHIVED else: try: f_out = self._open_compressed_file(_to_native_ascii(self.destination), 'wb') with open(path, 'rb') as f_in: shutil.copyfileobj(f_in, f_out) f_out.close() self.successes.append(path) self.destination_state = STATE_COMPRESSED except (IOError, OSError) as e: self.module.fail_json( path=_to_native(path), dest=_to_native(self.destination), msg='Unable to write to compressed file: %s' % _to_native(e), exception=format_exc() ) def add_targets(self): self.open() try: for target in self.targets: if os.path.isdir(target): for directory_path, directory_names, file_names in os.walk(target, topdown=True): for directory_name in directory_names: full_path = os.path.join(directory_path, directory_name) self.add(full_path, strip_prefix(self.root, full_path)) for file_name in file_names: full_path = os.path.join(directory_path, file_name) self.add(full_path, strip_prefix(self.root, full_path)) else: self.add(target, strip_prefix(self.root, target)) except Exception as e: if self.format in ('zip', 'tar'): archive_format = self.format else: archive_format = 'tar.' + self.format self.module.fail_json( msg='Error when writing %s archive at %s: %s' % ( archive_format, _to_native(self.destination), _to_native(e) ), exception=format_exc() ) self.close() if self.errors: self.module.fail_json( msg='Errors when writing archive at %s: %s' % (_to_native(self.destination), '; '.join(self.errors)) ) def is_different_from_original(self): if self.original_checksums is None: return self.original_size != self.destination_size() else: return self.original_checksums != self.destination_checksums() def destination_checksums(self): if self.destination_exists() and self.destination_readable(): return self._get_checksums(self.destination) return None def destination_exists(self): return self.destination and os.path.exists(self.destination) def destination_readable(self): return self.destination and os.access(self.destination, os.R_OK) def destination_size(self): return os.path.getsize(self.destination) if self.destination_exists() else 0 def find_targets(self): for path in self.paths: if not os.path.lexists(path): self.not_found.append(path) else: self.targets.append(path) def has_targets(self): return bool(self.targets) def has_unfound_targets(self): return bool(self.not_found) def remove_single_target(self, path): try: os.remove(path) except OSError as e: self.module.fail_json( path=_to_native(path), msg='Unable to remove source file: %s' % _to_native(e), exception=format_exc() ) def remove_targets(self): for path in self.successes: if os.path.exists(path): try: if os.path.isdir(path): shutil.rmtree(path) else: os.remove(path) except OSError: self.errors.append(_to_native(path)) for path in self.paths: try: if os.path.isdir(path): shutil.rmtree(path) except OSError: self.errors.append(_to_native(path)) if self.errors: self.module.fail_json( dest=_to_native(self.destination), msg='Error deleting some source files: ', files=self.errors ) def update_permissions(self): try: file_args = self.module.load_file_common_arguments(self.module.params, path=self.destination) except TypeError: # The path argument is only supported in Ansible-base 2.10+. Fall back to # pre-2.10 behavior for older Ansible versions. self.module.params['path'] = self.destination file_args = self.module.load_file_common_arguments(self.module.params) self.changed = self.module.set_fs_attributes_if_different(file_args, self.changed) @property def result(self): return { 'archived': [_to_native(p) for p in self.successes], 'dest': _to_native(self.destination), 'dest_state': self.destination_state, 'changed': self.changed, 'arcroot': _to_native(self.root), 'missing': [_to_native(p) for p in self.not_found], 'expanded_paths': [_to_native(p) for p in self.expanded_paths], 'expanded_exclude_paths': [_to_native(p) for p in self.expanded_exclude_paths], } def _check_removal_safety(self): for path in self.paths: if os.path.isdir(path) and self.destination.startswith(os.path.join(path, b'')): self.module.fail_json( path=b', '.join(self.paths), msg='Error, created archive can not be contained in source paths when remove=true' ) def _open_compressed_file(self, path, mode): f = None if self.format == 'gz': f = gzip.open(path, mode) elif self.format == 'bz2': f = bz2.BZ2File(path, mode) elif self.format == 'xz': f = lzma.LZMAFile(path, mode) else: self.module.fail_json(msg="%s is not a valid format" % self.format) return f @abc.abstractmethod def close(self): pass @abc.abstractmethod def contains(self, name): pass @abc.abstractmethod def open(self): pass @abc.abstractmethod def _add(self, path, archive_name): pass @abc.abstractmethod def _get_checksums(self, path): pass class ZipArchive(Archive): def __init__(self, module): super(ZipArchive, self).__init__(module) def close(self): self.file.close() def contains(self, name): try: self.file.getinfo(name) except KeyError: return False return True def open(self): self.file = zipfile.ZipFile(_to_native_ascii(self.destination), 'w', zipfile.ZIP_DEFLATED, True) def _add(self, path, archive_name): if not matches_exclusion_patterns(path, self.exclusion_patterns): self.file.write(path, archive_name) def _get_checksums(self, path): try: archive = zipfile.ZipFile(_to_native_ascii(path), 'r') checksums = set((info.filename, info.CRC) for info in archive.infolist()) archive.close() except zipfile.BadZipfile: checksums = set() return checksums class TarArchive(Archive): def __init__(self, module): super(TarArchive, self).__init__(module) self.fileIO = None def close(self): self.file.close() if self.format == 'xz': with lzma.open(_to_native(self.destination), 'wb') as f: f.write(self.fileIO.getvalue()) self.fileIO.close() def contains(self, name): try: self.file.getmember(name) except KeyError: return False return True def open(self): if self.format in ('gz', 'bz2'): self.file = tarfile.open(_to_native_ascii(self.destination), 'w|' + self.format) # python3 tarfile module allows xz format but for python2 we have to create the tarfile # in memory and then compress it with lzma. elif self.format == 'xz': self.fileIO = io.BytesIO() self.file = tarfile.open(fileobj=self.fileIO, mode='w') elif self.format == 'tar': self.file = tarfile.open(_to_native_ascii(self.destination), 'w') else: self.module.fail_json(msg="%s is not a valid archive format" % self.format) def _add(self, path, archive_name): def py27_filter(tarinfo): return None if matches_exclusion_patterns(tarinfo.name, self.exclusion_patterns) else tarinfo def py26_filter(path): return matches_exclusion_patterns(path, self.exclusion_patterns) if PY27: self.file.add(path, archive_name, recursive=False, filter=py27_filter) else: self.file.add(path, archive_name, recursive=False, exclude=py26_filter) def _get_checksums(self, path): try: if self.format == 'xz': with lzma.open(_to_native_ascii(path), 'r') as f: archive = tarfile.open(fileobj=f) checksums = set((info.name, info.chksum) for info in archive.getmembers()) archive.close() else: archive = tarfile.open(_to_native_ascii(path), 'r|' + self.format) checksums = set((info.name, info.chksum) for info in archive.getmembers()) archive.close() except (lzma.LZMAError, tarfile.ReadError, tarfile.CompressionError): try: # The python implementations of gzip, bz2, and lzma do not support restoring compressed files # to their original names so only file checksum is returned f = self._open_compressed_file(_to_native_ascii(path), 'r') checksums = set([(b'', crc32(f.read()))]) f.close() except Exception: checksums = set() return checksums def get_archive(module): if module.params['format'] == 'zip': return ZipArchive(module) else: return TarArchive(module) def main(): module = AnsibleModule( argument_spec=dict( path=dict(type='list', elements='path', required=True), format=dict(type='str', default='gz', choices=['bz2', 'gz', 'tar', 'xz', 'zip']), dest=dict(type='path'), exclude_path=dict(type='list', elements='path', default=[]), exclusion_patterns=dict(type='list', elements='path'), force_archive=dict(type='bool', default=False), remove=dict(type='bool', default=False), ), add_file_common_args=True, supports_check_mode=True, ) if not HAS_LZMA and module.params['format'] == 'xz': module.fail_json( msg=missing_required_lib("lzma or backports.lzma", reason="when using xz format"), exception=LZMA_IMP_ERR ) check_mode = module.check_mode archive = get_archive(module) archive.find_targets() if not archive.has_targets(): if archive.destination_exists(): archive.destination_state = STATE_ARCHIVED if is_archive(archive.destination) else STATE_COMPRESSED elif archive.has_targets() and archive.must_archive: if check_mode: archive.changed = True else: archive.add_targets() archive.destination_state = STATE_INCOMPLETE if archive.has_unfound_targets() else STATE_ARCHIVED archive.changed |= archive.is_different_from_original() if archive.remove: archive.remove_targets() else: if check_mode: if not archive.destination_exists(): archive.changed = True else: path = archive.paths[0] archive.add_single_target(path) archive.changed |= archive.is_different_from_original() if archive.remove: archive.remove_single_target(path) if archive.destination_exists(): archive.update_permissions() module.exit_json(**archive.result) if __name__ == '__main__': main()