Source code for lago.export

from __future__ import absolute_import

from future.builtins import super
from abc import ABCMeta, abstractmethod
import logging
import functools
import copy
import os
from os import path
import time
import json
import shutil

from six import with_metaclass

from . import log_utils, utils

LOGGER = logging.getLogger(__name__)
LogTask = functools.partial(log_utils.LogTask, logger=LOGGER)


[docs]class DiskExportManager(with_metaclass(ABCMeta, object)): """ DiskExportManager object is responsible on the export process of an image from the current Lago prefix. DiskExportManger is the base class of specific DiskExportManger. Each specific DiskExportManger is responsible on the export process of an image with a specific type (e.g template, file...) Attributes: src (str): Path to the image that should be exported name (str): The name of the exported disk dst (str): The absolute path of the exported disk disk_type (str): The type of the image e.g template, file, empty... disk (dict): Disk attributes (of the disk that should be exported) as found in workdir/current/virt/VM-NAME exported_metadata(dict): A copy of the source disk metadata, this dict should be updated with new values during the export process. do_compress(bool): If true, apply compression to the exported disk. """ def __init__(self, dst, disk_type, disk, do_compress): """ Args: dst (str): The absolute path of the exported disk disk_type (str): The type of the image e.g template, file, empty... disk (dict): Disk attributes as found in workdir/current/virt/VM-NAME do_compress(bool): If true, apply compression to the exported disk. """ self.src = path.expandvars(disk['path']) self.name = path.basename(self.src) self.dst = path.join(path.expandvars(dst), self.name) self.dist_type = disk_type self.disk = disk self.exported_metadata = copy.deepcopy(disk['metadata']) self.do_compress = do_compress
[docs] @staticmethod def get_instance_by_type(dst, disk, do_compress, *args, **kwargs): """ Args: dst (str): The path of the new exported disk. can contain env variables. disk (dict): Disk attributes (of the disk that should be exported) as found in workdir/current/virt/VM-NAME do_compress(bool): If true, apply compression to the exported disk. Returns An instance of a subclass of DiskExportManager which matches the disk type. """ disk_type = disk['type'] cls = HANDLERS.get(disk_type) if not cls: raise utils.LagoUserException( 'Export is not supported for disk type {}'.format(disk_type) ) return cls(dst, disk_type, disk, do_compress, *args, **kwargs)
[docs] @abstractmethod def export(self): """ This method will handle the export process and should implemented in each subclass. """ raise NotImplementedError( 'export is an abstract method that should' 'be implemented in subclass' )
[docs] def copy(self): """ Copy the disk using cp in order to preserves the 'sparse' structure of the file """ with LogTask('Copying disk'): utils.cp(self.src, self.dst)
[docs] def sparse(self): """ Make the exported images more compact by removing unused space. Please refer to 'virt-sparsify' for more info. """ with LogTask('Making disk sparse'): # Removed unused space from the disk utils.sparse(self.dst, self.disk['format'])
[docs] def calc_sha(self, checksum): """ Calculate the checksum of the new exported disk, write it to a file, and update this managers 'exported_metadata'. Args: checksum(str): The type of the checksum """ with LogTask('Calculating {}'.format(checksum)): with open(self.dst + '.hash', 'wt') as f: sha = utils.get_hash(self.dst, checksum) f.write(sha) self.exported_metadata[checksum] = sha
[docs] def compress(self): """ Compress the new exported image, Block size was taken from virt-builder page """ if not self.do_compress: return with LogTask('Compressing disk'): utils.compress(self.dst, 16777216) os.unlink(self.dst)
[docs] def update_lago_metadata(self): with LogTask('Updating Lago metadata'): self.exported_metadata['size'] = os.stat(self.dst).st_size self.exported_metadata['name'] = self.name self.exported_metadata['version'] = time.strftime("%Y%m%d.0")
[docs] def write_lago_metadata(self): with LogTask('Writing Lago metadata'): with open(self.dst + '.metadata', 'wt') as f: json.dump(self.exported_metadata, f)
[docs]class TemplateExportManager(DiskExportManager): """ TemplateExportManager is responsible exporting images of type template. Attributes: See superclass """ def __init__(self, dst, disk_type, disk, do_compress, *args, **kwargs): super().__init__(dst, disk_type, disk, do_compress) self.standalone = kwargs['standalone'] self.src_qemu_info = utils.get_qemu_info(self.src, backing_chain=True)
[docs] def rebase(self): """ Change the backing-file entry of the exported disk. Please refer to 'qemu-img rebase' manual for more info. """ if self.standalone: rebase_msg = 'Merging layered image with base' else: rebase_msg = 'Rebase' with LogTask(rebase_msg): if len(self.src_qemu_info) == 1: # Base image (doesn't have predecessors) return if self.standalone: # Consolidate the layers and base image utils.qemu_rebase(target=self.dst, backing_file="") else: if len(self.src_qemu_info) > 2: raise utils.LagoUserException( 'Layered export is currently supported for one ' 'layer only. You can try to use Standalone export.' ) # Put an identifier in the metadata of the copied layer, # this identifier will be used later by Lago in order # to resolve and download the base image parent = self.src_qemu_info[0]['backing-filename'] # Hack for working with lago images naming convention # For example: /var/lib/lago/store/phx_repo:el7.3-base:v1 # Extract only the image name and the version # (in the example el7.3-base:v1) parent = os.path.basename(parent) try: parent = parent.split(':', 1)[1] except IndexError: pass parent = './{}'.format(parent) utils.qemu_rebase( target=self.dst, backing_file=parent, safe=False )
[docs] def update_lago_metadata(self): super().update_lago_metadata() if self.standalone: self.exported_metadata['base'] = 'None' else: self.exported_metadata['base'] = os.path.basename(self.src)
[docs] def export(self): """ See DiskExportManager.export """ with LogTask('Exporting disk {} to {}'.format(self.name, self.dst)): with utils.RollbackContext() as rollback: rollback.prependDefer( shutil.rmtree, self.dst, ignore_errors=True ) self.copy() self.sparse() self.rebase() self.calc_sha('sha1') self.update_lago_metadata() self.write_lago_metadata() self.compress() rollback.clear()
[docs]class FileExportManager(DiskExportManager): """ FileExportManager is responsible exporting images of type file and empty. Attributes: standalone (bool): If true, create a new image which is the result of merging all the layers of src (the image that we want to export). src_qemu_info (dict): Metadata on src which was generated by qemu-img. """ def __init__(self, dst, disk_type, disk, do_compress, *args, **kwargs): super().__init__(dst, disk_type, disk, do_compress)
[docs] def export(self): """ See DiskExportManager.export """ with LogTask('Exporting disk {} to {}'.format(self.name, self.dst)): with utils.RollbackContext() as rollback: rollback.prependDefer( shutil.rmtree, self.dst, ignore_errors=True ) self.copy() if not self.disk['format'] == 'iso': self.sparse() self.calc_sha('sha1') self.update_lago_metadata() self.write_lago_metadata() self.compress() rollback.clear()
[docs]class VMExportManager(object): """ VMExportManager object is responsible on the export process of a list of disks. Attributes: disks (list of dicts): Disks to export. dst (str): Where to place the exported disks. compress(bool): If True compress each exported disk. with_threads(bool): If True, run the export in parallel *args(list): Extra args, will be passed to each DiskExportManager **kwargs(dict): Extra args, will be passed to each DiskExportManager """ def __init__( self, disks, dst, compress, with_threads=True, *args, **kwargs ): self._disks = disks self._dst = os.path.expandvars(os.path.realpath(dst)) self._compress = compress self._with_threads = with_threads self._args = args self._kwargs = kwargs
[docs] def _collect(self): """ Returns: (generator of dicts): The disks that needed to be exported """ return (disk for disk in self._disks if not disk.get('skip-export'))
[docs] def collect_paths(self): """ Returns: (list of str): The path of the disks that will be exported. """ return [os.path.expandvars(disk['path']) for disk in self._collect()]
[docs] def exported_disks_paths(self): """ Returns: (list of str): The path of the exported disks. """ return [ os.path.join(self._dst, os.path.basename(disk['path'])) for disk in self._collect() ]
[docs] def _get_export_mgr(self): """ Returns: (DiskExportManager): Handler for each disk """ return ( DiskExportManager.get_instance_by_type( dst=self._dst, disk=disk, do_compress=self._compress, *self._args, **self._kwargs ) for disk in self._collect() )
[docs] def export(self): """ Run the export process Returns: (list of str): The path of the exported disks. """ if self._with_threads: utils.invoke_different_funcs_in_parallel( *[mgr.export for mgr in self._get_export_mgr()] ) else: for mgr in self._get_export_mgr(): mgr.export() return self.exported_disks_paths()
HANDLERS = { 'file': FileExportManager, 'empty': FileExportManager, 'template': TemplateExportManager }