Skip to content

A small/simple python script to back up folders and databases.

Notifications You must be signed in to change notification settings

rossigee/backups

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Backup scripts

This python script, given a backup configuration, will cycle through a list of backup 'sources', packing them up and encrypting them into a temporary file, which it will upload to each of a list of destinations, finally reporting success or failure via a list of notification handlers.

Overview

Roughly speaking there are 'sources', 'destinations' and 'notifications'.

Currently implemented sources are:

  • folders (using tar)
  • folders via SSH (using tar)
  • MySQL databases (using mysqldump)
  • MySQL databases via SSH (using mysqldump)
  • RDS database snapshots (using mysqldump)
  • PostgreSQL databases (using pg_dump)
  • Azure Managed Disks

Currently implemented destinations are:

  • an S3 bucket (uses aws-cli)
  • a GS bucket (uses gsutil)
  • a Samba share (uses pysmbc)

Currently implemented notifications are:

  • an e-mail (via smtplib)
  • a HipChat room notification
  • a Discord room notification
  • a Slack notification
  • a Telegram notification
  • a simple flag file
  • a Prometheus push gateway
  • an Elasticsearch index

Hopefully, it's fairly straightforward to extend or add to the above.

The sources will be used to generate dump files in a temporary working area. By default, this is '/var/tmp', but for large DB dumps, you may need to specify an alternative folder mounted somewhere with enough space to store a compressed dump, and it's encrypted equivalent, temporarily.

Backups will be encrypted with a given passphrase (using GnuPG), and put into a folder on the destination using the following filename pattern...

/{hostname}/{yyyy-mm-dd}/{dumpfile_id}.{sql|tar}.gpg

If the backup configuration specifies a retention policy, then any copies that exist on the backup destination that fall outside that scope are deleted. Typically, a retention policy will specify to keep a copies for a certain number of days, or just a number of the most recent copies.

If encryption is not needed or desired, you may specify that only compression is used. Note that this will not safeguard the contents of the file should someone gain access to it. Compression is performed using gzip --fast and can be specfied in the source configuration using...

"compress_only": 1

Also, a word to the wise about scheduling. Backing up large chunks of data can use a lot of resources, and even if set to run after 'work' hours, can often run for on into the morning and indeed the next working day. Break larger backups down into smaller chunks where possible. If you are backing up mission criticial servers over mission critical network connections, you may need to take additional precautions to ensure that your backup scripts are not going to cause knock-on problems. You have been warned.

For security purposes, the script is designed to run as a non-privileged user. By default, it expects to be run as the 'backups' user, but this can be overridden using the 'RUN_AS_USER' environment variable.

Installation

  1. Install me. Unpack me, run 'python setup.py' as root.

  2. Create a configuration file listing the folders/databases you want to back up, an encryption passphrase, and one or more places you want the encrypted dump file uploaded to, and details of any notification methods you wish to use. See docs further down for examples. You can install the dependencies now, or you can wait until you get errors later :)

  3. Create a 'backups' user to run the script as. If you need folders to be backed up that require superuser privileges, you will also need a 'sudoers' directive to allow the backups user to run 'tar' as root to do it's work.

# cat >/etc/sudoers.d/99-backups <<EOF
backups ALL=(ALL) NOPASSWD: /bin/tar
EOF
  1. Test your configuration by doing an initial run of the backup.
# sudo -H -u backups /usr/local/bin/backup -v /home/backups/mynightlybackup.cfg
  1. Add your job to cron.
# cat >/etc/cron.d/nightly-backups.conf <<EOF
0 2 * * * backups /usr/local/bin/backup /home/backups/mynightlybackup.cfg
EOF

You're done.

Example configuration

IMPORTANT: The configuration file may contain sensitive information. Be sure to keep it somewhere safe, with suitable permissions.

To keep things simple, backup jobs are defined as JSON files, and are passed in at runtime.

The root object will contain the default settings that are not specific to a particular source, destination or notification.

{
  "hostname": "my.hostname.com",
  "tmpdir": "/var/tmp/backups",
  "modules": [],
  "sources": [],
  "destinations": [],
  "notifications": [],
}

Parameters available in the root object:

Config key Purpose
hostname Used to help identify the folder the backup should be stored in on the destination.
tmpdir Somewhere with enough temporary space to contain the compressed backup and it's encrypted copy as it's being built.
modules A list of modules to load to handle this job.
sources A list of sources to collect backup data from.
destinations A list of destinations to push backup data to.
notifications A list of mechanisms by which to report success/failure.

Common Source parameters

Config key Purpose
passphrase A passphrase to encrypt the backup with (using GnuPG).
recipients (Optional) If recipients are provided, the encrypted backup is encrypted for the list of recipients specified. If recipients are not provided, the backup symetrically encrypted using the given passphrase

Source - Folder

You can specify one or more folders to be backed up.

The 'modules' paramter must contain the backups.sources.folder module and there should be one or more source objects such as:

{
  "id": "accountsdata",
  "name": "Live Accounts Data",
  "type": "folder",
  "path": "/var/lib/myaccountspkg/data"
}

You can specify an array of --exclude parameters to tell tar which folders/files to exclude too.

{
  "id": "accountsdata",
  "name": "Live Accounts Data",
  "type": "folder",
  "path": "/var/lib/myaccountspkg/data",
  "excludes": [
    "tmp",
    "*~"
  ]
}

Parameters available in 'folder':

Config key Purpose
name Description of data being backed up (for reporting purposes).
path Name of file or folder to be backed up.
excludes One or more files/paths to be excluded from the backup.

Source - MySQL Database

You can specify one or more mySQL databases to be backed up.

The 'modules' paramter must contain the backups.sources.mysql module and there should be one or more source objects such as:

{
  "id": "livecompanydb",
  "name": "Live Company Data",
  "type": "mysql",
  "dbhost": "localhost",
  "dbname": "livecompany",
  "dbuser": "backups",
  "dbpass": "zzzuserwithreadonlyperms",
  "passphrase": "64b0c7405f2d8051e2b9f02aa4898acc",
  "compress_only": 1
}

The script creates a temporary 'mysqlclient' credentials file using the given details, and passes it as the '--defaults-file' argument.

By default, the '--events' flag is passed to mysqldump. This may break older versions of mysqldump (prior to version 5.1, IIRC), so you can disable this flag with the 'noevents' parameter. I hope you aren't using versions of MySQL that old!

[client]
host=specific.host.database.com
...
noevents=1

If supplied, directives should be a single string with all options seperated with a space. For example:

{
  "options": "--all-databases --single-transaction --skip-triggers"
}

NOTE: If you pass --all-databases as a directive, dbname will be ignored.

Parameters available in 'mysql':

Config key Purpose
name Description of data being backed up (for reporting purposes).
dbhost Name of the mySQL host to connect to.
dbname Name of the mySQL database to back up. You may specify more than one.
dbuser Username to connect to mySQL as.
dbpass Password to connect to mySQL with.
defaults The location of an 'mysqlclient' credentials file to use instead of creating a temporary one with using above 'db*' variables.
options An optional string of options to pass to 'mysqldump'.
noevents Don't pass the '--events' flag to 'mysqldump'.

Source - PostgreSQL Database

You can specify one of more PostgreSQL databases to be backed up.

{
  "id": "livecompanydb",
  "name": "Live Company Data",
  "type": "pgsql",
  "dbhost": "localhost",
  "dbname": "livecompany",
  "dbuser": "backups",
  "dbpass": "zzzuserwithreadonlyperms",
  "passphrase": "64b0c7405f2d8051e2b9f02aa4898acc"
}

The script creates a temporary 'pgpass' credentials file using the given details.

Parameters available in 'pgsql':

Config key Purpose
name Description of data being backed up (for reporting purposes).
dbhost Name of the PostgreSQL host to connect to.
dbname Name of the PostgreSQL database to back up.
dbuser Username to connect to PostgreSQL as.
dbpass Password to connect to PostgreSQL with.

Source - RDS Database Snapshots

You can specify one or more Amazon RDS databases to be backed up, where they have automated rolling backups enabled. Currently, this only supports MySQL-based RDS instances.

The last automatic snapshot of the given database will be identified using the AWS credentials provided. That will then be restored to a temporary instance, in a security group that allows access to the agent machine. The agent machine (running this script), will then 'mysqldump' the data to a '.sql.gz' file, and destroy the temporary instance. This avoids the need to run any backup queries that would adversely affect the performance of the live database.

WARNING: This can take a very long time in many cases.

{
  "id": "livecompanydb",
  "name": "Live Company Data",
  "type": "pgsql",
  "dbhost": "localhost",
  "dbname": "livecompany",
  "dbuser": "backups",
  "dbpass": "zzzuserwithreadonlyperms",
  "instancename": "livedb1",
  "region": "eu-west-1",
  "security_group": "livedbbackup",
  "instance_class": "db.m1.small",
  "passphrase": "64b0c7405f2d8051e2b9f02aa4898acc"
}

This module uses the 'boto' package, and expects auth credentials to be provided in the '~/.boto' file. This needs to be configured:

# cat >/home/backups/.boto <<EOF
[Credentials]
aws_access_key_id = YOURACCESSKEY
aws_secret_access_key = YOURSECRETKEY
EOF
# chown backups /home/backups/.boto
# chmod 400 /home/backups/.boto

By default, the '--events' flag is passed to mysqldump. This may break older versions of mysqldump (prior to version 5.1, IIRC), so you can disable this flag with the 'noevents' parameter.

The 'security_group' must allow access from the host running mysqldump.

Parameters available in 'rds':

Config key Purpose
name Description of data being backed up (for reporting purposes).
dbhost Name of the mySQL host to connect to.
dbname Name of the mySQL database to back up.
dbuser Username to connect to mySQL as.
dbpass Password to connect to mySQL with.
defaults The location of an 'mysqlclient' credentials file to use instead of creating a temporary one with using above 'db*' variables.
noevents Don't pass the '--events' flag to 'mysqldump'.
aws_access_key_id AWS access key
aws_secret_access_key AWS secret access key
instancename RDS instance name to clone backup of.
region AWS hosting region of RDS database.
security_group VPC security group for replica to provide access to.
instance_class RDS instance class to create replica as (defaults to 'db.m1.small').

Source - Volume snapshots

This method allows us to specify a set of RDS/EC2 volumes. When sourced, this will issue snapshot commands for the volumes or instances specified.

{
  "id": "livecompanydb",
  "name": "Live Company Data",
  "type": "snapshot",
  "instancename": "maindb1",
  "credentials": {
    "aws_access_key_id": "AKIAJPG7RJVVKWT3UX3A",
    "aws_secret_access_key": "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"
  }
}

It leaves a temporary file with status details as a 'dumpfile'. This gives the destination plugin something to work with when reporting success.

This source plugin is best used in a configuration on it's own, or with other volume snapshot definitions only, as used with a destination but no payload the configuration wouldn't make much sense!

NOTE: This plugin is largely unused and unmaintained at present.

Parameters available in 'snapshot':

Config key Purpose
name Description of data being backed up (for reporting purposes).
volume_id ID of the EBS volume to be snapshotted.
availability_zone AWS availability zone.
aws_access_key_id AWS access key.
aws_secret_access_key AWS secret access key.

Source - Azure Managed Disks

This method uses the Azure API to create a snapshot of the specified managed disk.

{
  "id": "mydisk1",
  "type": "azure-managed-disk",
  "name": "My Disk 1",
  "subscription_id": "285adaab-abcd-41cc-1234-9f256543d534",
  "source_resource_group": "MC_KUBERNETES_AZURECLUSTER1_NORTHEUROPE",
  "destination_resource_group": "MyBackupsResourceGroup",
  "disk_name": "mydisk1",
  "retain_snapshots": 3
}

The retain_snapshots ensures that a certain number of recent snapshots are left on Azure.

Parameters available in 'azure-managed-disk':

Config key Purpose
subscription_id The Azure subscription ID
source_resource_group The RG that the managed disk resides in
destination_resource_group The RG that the snapshot should be created in
disk_name The name of the managed disk in the source RG
retain_snapshots How many snapshots to leave in place on Azure

Additionally, the following environment variables must be defined for authentication (from AD service principal):

Env var Description
TENANT_ID Azure Tenant ID
CLIENT_ID Azure client/app ID
CLIENT_KEY Azure client secret key

Source - LVM over SSH

This method calls LVM commands over SSH to obtain a snapshot of an LVM volume.

{
  "id": "mydisk1",
  "type": "lvm-ssh",
  "name": "My Disk 1",
  "sshhost": "hostname.domain.com",
  "sshuser": "root",
  "vg_name": "ubuntu-vg",
  "lv_name": "mydisk1",
  "size": "5G",
  "retain_snapshots": 3
}

The retain_snapshots ensures that a certain number of snapshot volumes are retained on the host.

Parameters available in 'lvm-ssh':

Config key Purpose
sshhost The hostname to SSH to
sshuser The username to SSH with
vg_name The volume group name
lv_name The logical volume name
size The logical volume size
retain_snapshots How many snapshot volumes to leave in place

Destination - S3

You can specify an S3 bucket to back up to.

{
  "bucket": "backups-123456789",
}

The 'aws' CLI client gets it's authentication credentials and other configuration from the 'backups' user's '~/.aws/config' file. This needs to be configured as per instructions in the AWS CLI documentation.

Additionally, the S3 destination provides some simple backup rotation options. After a successful backup, the backup files are listed and the 'retention_copies' and 'retention_days' options, if present, are applied to identify and remove any backups that are no longer required.

Parameters available in 's3':

Config key Purpose
bucket S3 bucket to dump files to.
region AWS availability zone.
aws_access_key_id AWS access key.
aws_secret_access_key AWS secret access key.
retention_copies How many copies of older backups to keep.
retention_days How many days of backups to keep.
endpoint_url (optional) Endpoint URL for S3 service

Destination - GS

You can specify a GS bucket to back up to.

{
  "bucket": "backups-123456789",
}

The 'gs' destination module uses the boto library in conjunction with 'gsutil.' The 'gsutil' CLI client gets it's authentication credentials and other configuration from the 'backups' user's '~/.boto' file.

The GS module requires a GCP service account to be created with appropriate permissions to write and delete from GS buckets. The key file needs to be in P12 format. IMPORTANT: Properly secure this file and related information.

More information on configuring gsutil and boto as well as preparing a service account can be found at https://cloud.google.com/storage/docs/boto-plugin.

The boto file should contain entries similar to:

[Credentials]
gs_service_client_id = [email protected]
gs_service_key_file = /some/path/to/your/service-account-credential-file.p12
gs_service_key_file_password = asecretpassword

[GSUtil]
default_api_version = 2

AWS and GCP credential data can happily share the same section.

Additionally, the GS destination provides some simple backup rotation options. After a successful backup, the backup files are listed and the 'retention_copies' and 'retention_days' options, if present, are applied to identify and remove any backups that are no longer required.

Parameters available in 'gs':

Config key Purpose
bucket GS bucket to dump files to.
gs_service_client_id GCP service account.
gs_service_key_file Location of service account file (in P12 format).
gs_service_key_file_password Password for service account file.
retention_copies How many copies of older backups to keep.
retention_days How many days of backups to keep.

Destination - Samba

You can specify a Samba share to back up to.

{
  "id": "bigshare",
  "type": "samba",
  "host": "server1.local.lan",
  "share": "TheShare",
  "workgroup": "WORKGROUP",
  "credentials": {
    "username": "backups",
    "password": "getyourown"
  },
  "suffix": "gpg"
}

A temporary authentication credentials file will be created and the 'smbclient' client will be instructed to use it.

Parameters available in 'samba':

Config key Purpose
host Samba (DNS) hostname.
share Samba share to manage dump files on
username Samba username.
password Samba password.
workgroup Samba domain/workgroup.
suffix Suffix for created files.

Notification - Email

You can specify the SMTP configuration for notifications.

{
  "host": "smtp.mycompany.com",
  "port": 587,
  "credentials": {}
    "username": "backups",
    "password": "relaypassword"
  },
  "use_tls": 1,
  "use_ssl": 0,
  "success_to": "[email protected]",
  "failure_to": "[email protected]",
}

Parameters available in 'smtp':

Config key Purpose
host SMTP hostname.
port SMTP port.
username SMTP username.
password SMTP password.
use_tls If using STARTTLS, '1'.
use_ssl If using plain SSL, '1'.
success_to If set, success notification will be sent to the address specified. If not set, success notification will not be sent.
failure_to If set, failure notification will be sent to the address specified. If not set, failure notification will not be sent.

Notification - HipChat

You can specify that a notification of success or failure (or just failure) is posted as a HipChat room notification by providing the HipChat API credentials.

{
  "type": "hipchat",
  "url": "https://www.hipchat.com/...",
  "notify_on_success": 0,
  "notify_on_failure": 1
}

Parameters available in 'hipchat':

Config key Purpose
url URL provided by Hipchat for v2 integration.

Notification - Discord

You can specify that a notification of success or failure (or just failure) is posted as a Discord channel notification by providing the Discord webhook URL.

{
  "type": "discord",
  "url": "https://discordapp.com/api/webhooks/...",
  "notify_on_success": 0,
  "notify_on_failure": 1
}

Parameters available in 'discord':

Config key Purpose
url Discord webhook URL.

Notification - Matrix

You can specify that a notification of success or failure (or just failure) is posted as a Matrix room notification by providing Matrix service details.

{
  "type": "matrix",
  "url": "https://matrix.org",
  "access_token": "ALongStringOfCharacters",
  "room_id": "[email protected]",
  "notify_on_success": 0,
  "notify_on_failure": 1
}

Parameters available in 'matrix':

Config key Purpose
url Matrix home server URL.
access_token Pre-obtained access token.
room_id Matrix room ID for notifications.

Notification - Slack

You can specify that a notification of success or failure (or just failure) is posted as a Slack channel notification by providing a Slack URL.

{
  "type": "slack",
  "url": "https://hooks.slack.com/services/Z03JDZYT4/B03JEZGFG/jqWw4Zr5xPuArlZbbUexIaxP",
  "notify_on_success": 0,
  "notify_on_failure": 1
}

Parameters available in 'slack':

Config key Purpose
url Slack integration URL.

Notification - Telegram

You can specify that a notification of success or failure (or just failure) is posted as Telegram message by providing an API token and chat id.

{
  "type": "telegram",
  "api_token": "295432120:use_your_own_token_here",
  "chat_id": "38123456",
  "notify_on_success": 0,
  "notify_on_failure": 1
}

Parameters available in 'telegram':

Config key Purpose
api_token Telegram API token
chat_id Telegram Chat ID

Notification - Flag file

This simply updates a flag file in a specific location if the backup is successful.

{
  "type": "flagfile",
  "flagfile": "/backups/gitdata.ok"
}

We use this on our backup agents in conjunction with a simple 'file_age_secs' script running via SNMP. The following line in the 'snmpd.conf' file allows us to poll how long ago the last known successful backup was run.

exec check_git_backups /usr/local/bin/file_age_secs /backups/gitdata.ok

We then configure our monitoring/alerting service, Icinga, to poll this value and alert is if it exceeds a certain time.

(for file_age_secs source see https://gist.github.com/rossigee/44b9287e95068ebb9ae1)

Parameters available in 'flagfile':

Config key Purpose
filename Filename for file to create on successful backup completion.

Notification - Prometheus

Pushes a few statistics regarding the backup to a Prometheus push gateway.

{
  "type": "prometheus",
  "url": "http://pushgw:9091/",
  "credentials": {
    "username": "prometheus",
    "password": "yourpasswordhere"
  }
}

If the 'username and 'password' parameters are set, they are used to create the HTTP Basic Auth header for the connection.

Parameters available in 'prometheus':

Config key Purpose
url URL of Prometheus push gateway (not inc '/metrics' part).
username Push gateway auth username.
password Push gateway auth password.

Notification - Elasticsearch

You can specify a configuration to store backup statistics as an entry in an Elasticsearch index.

{
  "id": "logging",
  "type": "elasticsearch",
  "url": "https://logging-es-http.elasticsearch:9200",
  "credentials": {
    "username": "backups",
    "password": "yourpasswordhere"
  },
  "indexpattern": "backups-%Y"
}

Parameters available in 'elasticsearch':

Config key Purpose
url Elasticsearch URL, including authentication if required
indexpattern Pattern to use to determine the index to store the document to.

Complete Example

This simple example backs up some folders and a database, and deposits them to a Samba share, notifying sysadmin on failure only.

{
  "tmpdir": "/var/tmp/folderdumps",
  "modules": [
    "backups.sources.folder",
    "backups.sources.mysql",
    "backups.destinations.s3",
    "backups.notifications.smtp"
  ],
  "sources": [
    {
      "id": "config",
      "type": "folder",
      "name": "Live server config",
      "path": "/etc",
      "passphrase": "youllneverguesswhat"
    },
    {
      "id": "data",
      "type": "folder",
      "name": "Live server data folder",
      "path": "/data",
      "passphrase": "youllneverguesswhat"
    },
    {
      "id": "db",
      "type": "mysql",
      "name": "Live server database",
      "path": "/data",
      "passphrase": "youllneverguesswhat"
    },
  ],
  "destinations": [
    {
      "id": "myamazon",
      "type": "s3",
      "bucket": "mybucketnamehere",
      "region": "eu-west-1",
      "credentials": {
        "aws_access_key_id": "AKIAJPG7RJVVKWT3UX3A",
        "aws_secret_access_key": "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"
      }
    }
  ],
  "notifications": []
}

Credits

Ross Golder [email protected]

About

A small/simple python script to back up folders and databases.

Resources

Stars

Watchers

Forks

Packages

No packages published