aboutsummaryrefslogtreecommitdiffstats
path: root/bin
diff options
context:
space:
mode:
Diffstat (limited to 'bin')
-rwxr-xr-xbin/backup.py273
-rwxr-xr-xbin/dar_backup.py315
2 files changed, 0 insertions, 588 deletions
diff --git a/bin/backup.py b/bin/backup.py
deleted file mode 100755
index f204f13..0000000
--- a/bin/backup.py
+++ /dev/null
@@ -1,273 +0,0 @@
-#!/usr/bin/env python3
-#
-# Copyright (c) 2017 Aaron LI
-# MIT license
-#
-# 2017-05-11
-
-"""
-Backup the specified files & directories with the metadata (e.g., ownership,
-permission) preserved as much as possible.
-
-The extended attributes (i.e., xattr) are only available on Linux only.
-
-
-Example configuration (YAML format)
------------------------------------
-src_root : /
-sources : []
-dest_root : /backup
-dest_remove : []
-syspath : []
------------------------------------
-
-References
-----------
-"""
-
-import os
-import sys
-import stat
-import argparse
-import subprocess
-import logging
-from fnmatch import fnmatch
-from datetime import datetime
-
-try:
- from yaml import load as yaml_load
-except ImportError:
- from ruamel.yaml import safe_load as yaml_load
-
-
-logging.basicConfig(level=logging.INFO,
- format="%(asctime)s [%(levelname)s] %(message)s",
- datefmt="%Y-%m-%dT%H:%M:%S")
-progname = os.path.basename(sys.argv[0])
-logger = logging.getLogger(progname)
-
-# Protected system paths
-# The destination backup path is checked against these paths, and cannot
-# located where matches these paths, thus to avoid the system files being
-# overwritten or removed in accidence.
-syspath = set(["/", "/home", "/media", "/mnt", "/root",
- "/bin/*", "/boot/*", "/dev/*", "/etc/*", "/lib/*", "/lib64/*",
- "/proc/*", "/sbin/*", "/sys/*", "/usr/*", "/var/*"])
-
-
-class Backup:
- """
- Backup files with metadata preserved.
- """
- def __init__(self, configfile, dryrun=False, debug=False):
- self.configfile = configfile
- self.config = yaml_load(open(configfile))
- logger.info("Loaded configuration file: %s" % configfile)
- src_root = self.config.get("src_root", "/")
- if os.path.isabs(src_root):
- self.src_root = src_root
- logger.info("Source root directory: %s" % self.src_root)
- else:
- raise ValueError("Source root must be an absolute path")
- self.syspath = syspath.union(self.config.get("syspath", []))
- logger.info("Protected system paths: {0}".format(self.syspath))
- dest_root = os.path.expanduser(self.config["dest_root"])
- logger.info("Check backup destination against protected paths ...")
- self.dest_root = self.check_dest(dest_root)
- logger.info("Backup destination: %s" % self.dest_root)
- self.dryrun = dryrun
- logger.info("Dry run mode: %s" % dryrun)
- self.debug = debug
- logger.info("Show DEBUG information: %s" % debug)
-
- def check_dest(self, dest):
- """
- Check the given destination backup directory against the protected
- system paths.
- """
- dest = dest.rstrip("/")
- if not os.path.isabs(dest):
- raise ValueError("Destination must be an absolute path")
- for sp in self.syspath:
- if fnmatch(dest, sp):
- raise ValueError("Destination cannot under "
- "protected paths: %s" % dest)
- return dest
-
- def backup(self):
- """
- Do backup sources to the destination.
-
- The necessary destination directories are created before copying
- files.
- """
- if not os.path.exists(self.dest_root):
- logger.info("Create destination root: " % self.dest_root)
- self.mkdir(self.dest_root, dryrun=self.dryrun)
- for p in self.config["sources"]:
- src = os.path.join(self.src_root, p.lstrip("/"))
- dest = os.path.join(self.dest_root, p.lstrip("/"))
- if os.path.isdir(src):
- src_dir = src
- dest_dir = dest
- else:
- src_dir = os.path.dirname(src)
- dest_dir = os.path.dirname(dest)
- self.mkdirs(dest_dir, ref=src_dir, dryrun=self.dryrun)
- self.copy(src, dest, dryrun=self.dryrun, debug=self.debug)
-
- def cleanup(self):
- """
- Remove the obsolete files/directories from the destination.
- """
- for p in self.config.get("dest_remove", []):
- path = os.path.join(self.dest_root, p.lstrip("/"))
- self.remove(path, dest_root=self.dest_root,
- dryrun=self.dryrun, debug=self.debug)
-
- @staticmethod
- def copy(src, dest, dryrun=False, debug=False):
- """
- Copy file/directory using `rsync` with metadata preserved, and to
- keep directory contents in sync.
-
- Use `rsync --version` to check the available capabilities.
- """
- if os.path.isdir(src):
- src = src.rstrip("/") + "/"
- dest = dest.rstrip("/") + "/"
- logger.info("Copy & sync: %s -> %s" % (src, dest))
- args = ["--archive", "--hard-links", "--numeric-ids",
- "--delete", "--delete-after"]
- if os.uname().sysname == "Linux":
- args += ["--acls", "--xattrs"]
- if debug:
- args += ["--verbose"]
- cmd = ["rsync"] + args + [src, dest]
- if not dryrun:
- subprocess.check_call(cmd)
-
- @staticmethod
- def remove(path, dest_root, dryrun=False, debug=False):
- """
- Remove the specified file/directory using `rm -rf`, to clean
- up the destination backup.
-
- The specified path must locate under the `dest_root` for safety.
- """
- if not fnmatch(path, dest_root+"/*"):
- raise ValueError("Not allowed to remove file/directory "
- "outside destination: %s" % path)
- if not os.path.exists(path):
- return
- logger.info("Remove: %s" % path)
- args = ["-r", "-f"]
- if debug:
- args += ["-v"]
- cmd = ["rm"] + args + [path]
- if not dryrun:
- subprocess.check_call(cmd)
-
- @classmethod
- def mkdirs(cls, dest, ref=None, dryrun=False):
- """
- Recursively create the destination directories.
- """
- head, tail = os.path.split(dest)
- head_ref, tail_ref = os.path.split(ref)
- if not tail:
- head, tail = os.path.split(head)
- head_ref, tail_ref = os.path.split(head_ref)
- if head and tail and not os.path.exists(head):
- cls.mkdirs(head, ref=head_ref, dryrun=dryrun)
- if tail == os.path.curdir:
- # `xxx/newdir/.` exists if `xxx/newdir` exists
- return
- cls.mkdir(dest, ref=ref, dryrun=dryrun)
-
- @classmethod
- def mkdir(cls, dest, ref=None, dryrun=False):
- """
- Make the destination directory with the same metadata w.r.t. the
- reference directory.
-
- Parameters
- ----------
- dest : str
- The destination directory to be created
- ref : str, optional
- The reference directory, whose metadata will be mirrored to
- the destination directory.
- dryrun : bool
- If True, do not actually create the directory
- """
- if os.path.exists(dest):
- return
- logger.info("Create directory: %s" % dest)
- if not dryrun:
- os.mkdir(dest)
- if ref and not dryrun:
- cls.copystat(ref, dest)
-
- @classmethod
- def copystat(cls, src, dest, copy_own=True, copy_xattr=True):
- """
- Copy all stat info (mode bits, atime, mtime, flags) from `src` to
- `dest`. If `copy_own=True`, the uid and gid are also copied.
- If `copy_xattr=True`, the extended attributes are also copied
- (only available on Linux).
- """
- st = os.stat(src)
- mode = stat.S_IMODE(st.st_mode)
- os.chmod(dest, mode=mode)
- os.utime(dest, ns=(st.st_atime_ns, st.st_mtime_ns))
- if hasattr(st, "st_flags"):
- os.chflags(dest, flags=st.st_flags)
- if copy_own:
- os.chown(dest, uid=st.st_uid, gid=st.st_gid)
- if copy_xattr:
- cls.copyxattr(src, dest)
-
- @staticmethod
- def copyxattr(src, dest):
- """
- Copy the extended attributes (xattr) from `src` to `dest`.
-
- NOTE: xattr only available on Linux.
- """
- if not hasattr(os, "listxattr"):
- return
- for name in os.listxattr(src):
- value = os.getxattr(src, name)
- os.setxattr(dest, name, value)
-
-
-def main():
- parser = argparse.ArgumentParser(
- description="Backup files preserving metadata")
- parser.add_argument("-n", "--dry-run", dest="dryrun", action="store_true",
- help="dry run, do not perform actual action")
- parser.add_argument("-q", "--quiet", dest="quiet", action="store_true",
- help="be quiet")
- parser.add_argument("-d", "--debug", dest="debug", action="store_true",
- help="show verbose debug information")
- parser.add_argument("config", help="configuration file")
- args = parser.parse_args()
-
- if args.quiet and not args.dryrun:
- logging.basicConfig(level=logging.WARNING)
- if args.debug:
- logging.basicConfig(level=logging.DEBUG)
-
- now = datetime.now()
- logger.info("=== %s @ %s ===" % (" ".join(sys.argv), now.isoformat()))
- if args.dryrun:
- logger.info("*** DRY RUN ***")
- backup = Backup(args.config, dryrun=args.dryrun, debug=args.debug)
- backup.backup()
- backup.cleanup()
- logger.info("=== Backup Finished! @ %s ===" % datetime.now().isoformat())
-
-
-if __name__ == "__main__":
- main()
diff --git a/bin/dar_backup.py b/bin/dar_backup.py
deleted file mode 100755
index 47ff815..0000000
--- a/bin/dar_backup.py
+++ /dev/null
@@ -1,315 +0,0 @@
-#!/usr/bin/env python3
-#
-# Copyright (c) 2017 Aaron LI
-# MIT license
-#
-# 2017-03-08
-
-"""
-Full/differential backup using 'dar' with 'par2'.
-
-This script accepts a configuration file, which is an simple YAML file
-and defines all necessary options for 'dar'.
-The 'dar'-created archives are stored under the same directory as the
-configuration file, and 'par2' is used to create redundancy data for
-possible repairs at the same time, and finally the archives are
-tested/verified by 'dar'.
-
-
-Example configuration file for system/root backup
--------------------------------------------------
-# recommended format: <hostname>-root-<distro>-{date}
-name: "office-root-opensuse-{date}"
-fs_root: "/"
-is_system: true
-prune: []
-exclude: []
-exclude_compression: []
-
-
-Example configuration file for user home backup
------------------------------------------------
-# recommended format: <hostname>-home-<user>-{date}
-name: "office-home-aly-{date}"
-fs_root: "/home/aly"
-prune: [".cache",]
-exclude: []
-exclude_compression: []
-
-
-Credit/references
------------------
-[1] http://www.halfgaar.net/backing-up-unix
-[2] http://dar.linux.free.fr/doc/man/dar.html
-[3] https://github.com/Parchive/par2cmdline
-[4] http://dar.linux.free.fr/doc/samples/dar_par_create.duc
-[5] http://dar.linux.free.fr/doc/samples/dar_par_test.duc
-"""
-
-import os
-import sys
-import argparse
-import subprocess
-import logging
-from datetime import datetime
-from glob import glob
-
-import yaml
-
-
-progname = os.path.basename(sys.argv[0])
-logger = logging.getLogger(progname)
-
-
-class DarSettings:
- """
- dar settings
- """
- # directory (relative path) for storing the isolated catalogs
- catalog_path = "catalogs"
-
- # date format for output archive name
- date_fmt = "%Y%m%dT%H%M"
-
- # Default settings
- args_common = [
- "--min-digits", "3,3,3",
- "--noconf", # do not try to read /etc/darrc or ~/.darrc
- ]
- args_default = [
- "--alter=atime", # do not preserve atime
- "--alter=no-case", # case insensitive mode
- "--alter=glob", # glob expression mode (instead of regex mode)
- "--alter=binary", # use 2^10 instead of 10^3
- "--compression=bzip2",
- "--empty-dir", # create empty directories for the excluded
- "--hash", "sha512", # calculate the hash of slices on the fly
- "--no-overwrite",
- "--no-mount-points", # stay in the same filesystem
- ]
- size_slice = "2G"
- exclude = set(["*~", ".*~", ".~*"])
- exclude_compression = set([
- "*.7z", "*.ape", "*.avi", "*.bz2", "*.deb", "*.exe", "*.flac",
- "*.flv", "*.gz", "*.iso", "*.jar", "*.jpg", "*.jpeg",
- "*.m4a", "*.m4v", "*.mkv", "*.mov", "*.mp3", "*.mp4", "*.ogg",
- "*.rar", "*.rpm", "*.tar.bz2", "*.tar.gz", "*.tar.xz",
- "*.tbz", "*.tgz", "*.txz", "*.wmv", "*.xz", "*.zip",
- ])
- prune = set()
- verbose = set()
-
- # Parchive
- redundancy = 5 # 5% of redundancy for par2
- cmd_par_create = ("par2 create -r{redundancy} -n1 "
- "'%p/%b.%N.par2' '%p/%b.%N.%e'")
- cmd_par_test = "par2 verify '%p/%b.%N.par2'"
-
- # Whether it is a backup of system root
- is_system = False
- prune_system = set(["dev/pts", "dev/shm", "home", "lost+found",
- "media", "mnt", "proc", "run", "sys",
- "tmp", "var/cache", "var/tmp"])
-
- def __init__(self, configfile, verbose=False, dry_run=False):
- if verbose:
- self.verbose = set(["treated", "messages"])
- else:
- self.verbose = set()
- if dry_run:
- self.args_common += ["--dry-run"]
-
- self.path = os.path.dirname(os.path.abspath(configfile))
- settings = yaml.load(open(configfile))
- self.merge_settings(settings)
-
- def merge_settings(self, settings):
- self.name = settings["name"]
- self.fs_root = settings["fs_root"]
- self.is_system = settings.get("is_system", self.is_system)
- self.date_fmt = settings.get("date_fmt", self.date_fmt)
- self.catalog_path = settings.get("catalog_path", self.catalog_path)
- self.size_slice = settings.get("size_slice", self.size_slice)
- self.redundancy = settings.get("redundancy", self.redundancy)
- self.verbose = self.verbose.union(settings.get("verbose", set()))
- self.prune = self.prune.union(settings.get("prune", set()))
- self.exclude = self.exclude.union(settings.get("exclude", set()))
- self.exclude_compression = self.exclude_compression.union(
- settings.get("exclude_compression", set()))
-
- def archive_name(self, date=None):
- # Return the archive name with date substituted
- if date is None:
- date = datetime.now().strftime(self.date_fmt)
- name = os.path.join(self.path, self.name.format(date=date))
- return name
-
- def last_archive_name(self):
- # Get the last archive (backup data or isolated catalog)
- # used as the reference archive for a differential backup.
- name = self.last_backup()
- if name is None:
- name = self.last_catalog()
- return name
-
- def last_backup(self):
- # Return the name of last backup if found, otherwise, return None
- backup_glob = self.archive_name(date="*") + ".dar"
- try:
- backup = sorted(glob(backup_glob))[-1]
- name = os.path.splitext(os.path.splitext(backup)[0])[0]
- logger.info("Found last backup: {0}".format(name))
- return name
- except IndexError:
- logger.info("No existing backups found!")
- return None
-
- def last_catalog(self):
- # Return the name of last isolated catalog if found, otherwise,
- # return None
- basename = os.path.basename(self.archive_name(date="*")) + ".dar"
- catalog_glob = os.path.join(self.path, self.catalog_path, basename)
- try:
- catalog = sorted(glob(catalog_glob))[-1]
- name = os.path.splitext(os.path.splitext(catalog)[0])[0]
- logger.info("Found last catalog: {0}".format(name))
- return name
- except IndexError:
- logger.info("No existing catalogs found!")
- return None
-
- @property
- def args_create(self):
- cmd_par = self.cmd_par_create.format(redundancy=self.redundancy)
- args = self.args_default + [
- "--execute", cmd_par,
- "--fs-root", self.fs_root,
- "--slice", self.size_slice,
- ]
- args += ["--verbose=%s" % item for item in self.verbose]
- for item in self.exclude:
- args += ["--exclude", item]
- for item in self.exclude_compression:
- args += ["-Z", item]
- for item in self.prune:
- args += ["--prune", item]
- if self.is_system:
- for item in self.prune_system:
- args += ["--prune", item]
- logger.info("args_create: {0}".format(args))
- return args
-
- @property
- def args_test(self):
- args = ["--execute", self.cmd_par_test]
- args += ["--verbose=%s" % item for item in self.verbose]
- logger.info("args_test: {0}".format(args))
- return args
-
-
-class DarBackup:
- """
- dar backup with settings
- """
- def __init__(self, settings):
- self.settings = settings
- self.is_system = settings.is_system
- self.path = settings.path
- self.catalog_path = settings.catalog_path
- self.archive_name = settings.archive_name()
- self.last_archive_name = settings.last_archive_name()
- self.args_common = settings.args_common
- self.args_create = settings.args_create
- self.args_test = settings.args_test
-
- def run(self, dry_run=False):
- if self.is_system and (not self.is_running_as_root()):
- raise RuntimeError("Running as root to backup the system root")
-
- logger.info("Output archive name: {0}".format(self.archive_name))
- if self.last_archive_name is not None:
- self.backup_diff()
- else:
- self.backup_full()
-
- if not dry_run:
- self.test_backup()
- self.isolate_catalog()
-
- def backup_full(self):
- logger.info("Start full backup ...")
- cmd = [
- "dar",
- "--create", self.archive_name,
- ] + self.args_common + self.args_create
- logger.info("Command: {0}".format(cmd))
- subprocess.run(cmd, check=True)
- logger.info("Full backup: DONE!")
-
- def backup_diff(self):
- logger.info("Start differential backup ...")
- cmd = [
- "dar",
- "--create", self.archive_name,
- "--ref", self.last_archive_name,
- ] + self.args_common + self.args_create
- logger.info("Command: {0}".format(cmd))
- subprocess.run(cmd, check=True)
- logger.info("Differential backup: DONE!")
-
- def test_backup(self):
- logger.info("Test created backup ...")
- cmd = [
- "dar",
- "--test", self.archive_name,
- ] + self.args_common + self.args_test
- logger.info("Command: {0}".format(cmd))
- subprocess.run(cmd, check=True)
- logger.info("Test backup: DONE!")
-
- def isolate_catalog(self):
- logger.info("Isolate the catalog from backup ...")
- catalog_dir = os.path.join(self.path, self.catalog_path)
- if not os.path.exists(catalog_dir):
- os.mkdir(catalog_dir)
- catalog = os.path.join(catalog_dir,
- os.path.basename(self.archive_name))
- logger.info("Output catalog: {0}".format(catalog))
- cmd = [
- "dar",
- "--isolate", catalog,
- "--ref", self.archive_name,
- ] + self.args_common
- logger.info("Command: {0}".format(cmd))
- subprocess.run(cmd, check=True)
- logger.info("Isolate backup catalog: DONE!")
-
- @staticmethod
- def is_running_as_root():
- return os.getuid() == 0
-
-
-def main():
- parser = argparse.ArgumentParser(
- description="Backup system/data using dar and par2")
- parser.add_argument("-c", "--config", dest="config", required=True,
- help="configuration file for dar and archive. " +
- "NOTE: the backup archive will be placed under " +
- "the same directory as this configuration file")
- parser.add_argument("-n", "--dry-run", dest="dry_run", action="store_true",
- help="dry run, do not perform any action")
- parser.add_argument("-v", "--verbose", dest="verbose", action="store_true",
- help="show verbose information")
- args = parser.parse_args()
-
- if args.verbose:
- logging.basicConfig(level=logging.INFO)
-
- settings = DarSettings(args.config, verbose=args.verbose,
- dry_run=args.dry_run)
- dar = DarBackup(settings)
- dar.run(dry_run=args.dry_run)
-
-
-if __name__ == "__main__":
- main()