title | description | author | resourceType | personas | labels | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Backup Container |
A simple containerized backup solution for backing up one or more supported databases to a secondary location. |
WadeBarnes |
Components |
|
|
Backup Container is a simple containerized backup solution for backing up one or more supported databases to a secondary location. Code and documentation was originally pulled from the HETS Project
MongoDB MSSQL - Currently MSSQL requires that the nfs db volume be shared with the database for backups to function correctly. PostgresSQL
You can run the Backup Container for supported databases separately or in a mixed environment. For a mixed environment:
- You MUST use the recommended
backup.conf
configuration. - Within the
backup.conf
, you MUST specify theDatabaseType
for each listed database. - You will need to create two builds and two deployment configs. One for each type of supported backup container in use.
- Mount the same
backup.conf
file (ConfigMap) to each deployed container.
This project provides you with a starting point for integrating backups into your OpenShift projects. The scripts and templates provided in the openshift directory are compatible with the openshift-developer-tools scripts. They help you create an OpenShift deployment or cronjob called backup
in your projects that runs backups on databases within the project environment. You only need to integrate the scripts and templates into your project(s), the builds can be done with this repository as the source.
Following are the instructions for running the backups and a restore.
Before we get too far into the the details, we're going to take a moment to discuss the most important part of the whole process - The Storage. The backup container uses two volumes, one for storing the backups and the other for restore/verification testing. The deployment template separates them intentionally.
The following sections on storage discuss the recommendations and limitations of the storage classes created specifically for the BC Government's PathFinder environment.
The recommended storage class for the backup volume is nfs-backup
. This class of storage cannot be auto-provisioned through the use of a deployment template. The PersistentVolumeClaim
declared in the supplied deployment template for the backup volume will purposely fail to properly provision and wire an nfs-backup
volume if published before you manually provision your nfs-backup
claim.
When using nfs-backup
you will need to provision your claims before you publish your deployment configuration, through either the service catalog using the BC Gov NFS Storage wizard, or by using the svcat cli.
You'll note the name of the resulting storage claim has a random component to it (example, bk-devex-von-bc-tob-test-xjrmkhsnshay
). This name needs to be injected into the default value of the BACKUP_VOLUME_NAME
parameter of the template before publishing the deployment configuration in order for the storage to be correctly mounted to the /backups/
directory of the container.
nfs-backup
storageClass is a lower tier of storage and not considered highly available. read: don't use this for live application storage
. The storageClass IS covered by the default enterprise backup policies, and can be directly referenced for restores using the PVC name when opening a restore ticket with 7700.
nfs-backup
PVCs cannot be used for restore/verification. The permissions on the underlying volume do not allow the PostgreSql server to host it's configuration and data files on a directory backed by this class of storage.
Ensure you review and plan your storage requirements before provisioning.
More information on provisioning nfs-backup
storage here; provision-nfs-apb
NFS backed storage is covered by the following backup and retention policies:
- Backup
- Daily: Incremental
- Monthly: Full
- Retention
- 90 days
The default storage class for the restore/verification volume is netapp-file-standard
. The supplied deployment template will auto-provision this volume for you with it is published. Refer to the Storage Performance section for performance considerations.
This volume should be large enough to host your largest database. Set the size by updating/overriding the VERIFICATION_VOLUME_SIZE
value within the template.
The performance of netapp-block-standard
for restore/verification is far superior to that of netapp-file-standard
, however it should only be used in cases where the time it takes to verify a backup begins to encroach on the over-all timing and verification cycle. You want the verification(s) to complete before another backup and verification cycle begins and you want a bit of idle time between the end of one cycle and the beginning of another in case things take a little longer now and again.
There are currently no performance stats for the netapp
storage types.
Together, the scripts and templates provided in the openshift directory will automatically deploy the backup
app as described below. The backup-deploy.overrides.sh script generates the deployment configuration necessary for the backup.conf file to be mounted as a ConfigMap by the backup
container.
The following environment variables are defaults used by the backup
app.
NOTE: These environment variables MUST MATCH those used by the database container(s) you are planning to backup.
Name | Default (if not set) | Purpose |
---|---|---|
BACKUP_STRATEGY | rolling | To control the backup strategy used for backups. This is explained more below. |
BACKUP_DIR | /backups/ | The directory under which backups will be stored. The deployment configuration mounts the persistent volume claim to this location when first deployed. |
NUM_BACKUPS | 31 | Used for backward compatibility only, this value is used with the daily backup strategy to set the number of backups to retain before pruning. |
DAILY_BACKUPS | 6 | When using the rolling backup strategy this value is used to determine the number of daily (Mon-Sat) backups to retain before pruning. |
WEEKLY_BACKUPS | 4 | When using the rolling backup strategy this value is used to determine the number of weekly (Sun) backups to retain before pruning. |
MONTHLY_BACKUPS | 1 | When using the rolling backup strategy this value is used to determine the number of monthly (last day of the month) backups to retain before pruning. |
BACKUP_PERIOD | 1d | Only used for Legacy Mode. Ignored when running in Cron Mode. The schedule on which to run the backups. The value is used by a sleep command and can be defined in d, h, m, or s. |
DATABASE_SERVICE_NAME | postgresql | Used for backward compatibility only. The name of the service/host for the default database target. |
DATABASE_USER_KEY_NAME | database-user | The database user key name stored in database deployment resources specified by DATABASE_DEPLOYMENT_NAME. |
DATABASE_PASSWORD_KEY_NAME | database-password | The database password key name stored in database deployment resources specified by DATABASE_DEPLOYMENT_NAME. |
DATABASE_NAME | my_postgres_db | Used for backward compatibility only. The name of the default database target; the name of the database you want to backup. |
DATABASE_USER | wired to a secret | The username for the database(s) hosted by the database server. The deployment configuration makes the assumption you have your database credentials stored in secrets (which you should), and the key for the username is database-user . The name of the secret must be provided as the DATABASE_DEPLOYMENT_NAME parameter to the deployment configuration template. |
DATABASE_PASSWORD | wired to a secret | The password for the database(s) hosted by the database server. The deployment configuration makes the assumption you have your database credentials stored in secrets (which you should), and the key for the username is database-password . The name of the secret must be provided as the DATABASE_DEPLOYMENT_NAME parameter to the deployment configuration template. |
FTP_URL | The FTP server URL. If not specified, the FTP backup feature is disabled. The default value in the deployment configuration is an empty value - not specified. | |
FTP_USER | wired to a secret | The username for the FTP server. The deployment configuration creates a secret with the name specified in the FTP_SECRET_KEY parameter (default: ftp-secret ). The key for the username is ftp-user and the value is an empty value by default. |
FTP_PASSWORD | wired to a secret | The password for the FTP server. The deployment configuration creates a secret with the name specified in the FTP_SECRET_KEY parameter (default: ftp-secret ). The key for the password is ftp-password and the value is an empty value by default. |
WEBHOOK_URL | The URL of the webhook endpoint to use for notifications. If not specified, the webhook integration feature is disabled. The default value in the deployment configuration is an empty value - not specified. | |
ENVIRONMENT_FRIENDLY_NAME | A friendly (human readable) name of the environment. This variable is used by the webhook integration to identify the environment from which the backup notifications originate. The default value in the deployment configuration is an empty value - not specified. | |
ENVIRONMENT_NAME | A name or ID of the environment. This variable is used by the webhook integration to identify the environment from which the backup notifications originate. The default value in the deployment configuration is an empty value - not specified. |
Using this default configuration you can easily back up a single postgres database, however we recommend you extend the configuration and use the backup.conf
file to list a number of databases for backup and even set a cron schedule for the backups.
When using the backup.conf
file the following environment variables are ignored, since you list all of your host
/database
pairs in the file; DATABASE_SERVICE_NAME
, DATABASE_NAME
. To provide the credentials needed for the listed databases you extend the deployment configuration to include hostname_USER
and hostname_PASSWORD
credential pairs which are wired to the appropriate secrets (where hostname matches the hostname/servicename, in all caps and underscores, of the database). For example, if you are backing up a database named wallet-db/my_wallet
, you would have to extend the deployment configuration to include a WALLET_DB_USER
and WALLET_DB_PASSWORD
credential pair, wired to the appropriate secrets, to access the database(s) on the wallet-db
server.
The backup
container supports running the backups on a cron schedule. The schedule is specified in the backup.conf
file. Refer to the backup.conf file for additional details and examples.
This section describes the configuration of an OpenShift CronJob this is different than the Cron Mode supported by the container when deployed in "long running" mode.
The cronjob object can be deployed in the same manner as the application, and will also have a dependency on the image built by the build config. The main constraint for the cronjob objects is that they will require a configmap in place of environment variables and does not support the backup.conf
for multiple database backups in the same job. In order to backup multiple databases, create multiple cronjob objects with their associated configmaps and secrets.
The following variables are supported in the first iteration of the backup cronjob:
Name | Default (if not set) | Purpose |
---|---|---|
BACKUP_STRATEGY | daily | To control the backup strategy used for backups. This is explained more below. |
BACKUP_DIR | /backups/ | The directory under which backups will be stored. The deployment configuration mounts the persistent volume claim to this location when first deployed. |
SCHEDULE | 0 1 * * * | Cron Schedule to Execute the Job (using local cluster system TZ). |
NUM_BACKUPS | 31 | For backward compatibility this value is used with the daily backup strategy to set the number of backups to retain before pruning. |
DAILY_BACKUPS | 6 | When using the rolling backup strategy this value is used to determine the number of daily (Mon-Sat) backups to retain before pruning. |
WEEKLY_BACKUPS | 4 | When using the rolling backup strategy this value is used to determine the number of weekly (Sun) backups to retain before pruning. |
MONTHLY_BACKUPS | 1 | When using the rolling backup strategy this value is used to determine the number of monthly (last day of the month) backups to retain before pruning. |
DATABASE_SERVICE_NAME | postgresql | The name of the service/host for the default database target. |
DATABASE_USER_KEY_NAME | database-user | The database user key name stored in database deployment resources specified by DATABASE_DEPLOYMENT_NAME. |
DATABASE_PASSWORD_KEY_NAME | database-password | The database password key name stored in database deployment resources specified by DATABASE_DEPLOYMENT_NAME. |
POSTGRESQL_DATABASE | my_postgres_db | The name of the default database target; the name of the database you want to backup. |
POSTGRESQL_USER | wired to a secret | The username for the database(s) hosted by the postgresql Postgres server. The deployment configuration makes the assumption you have your database credentials stored in secrets (which you should), and the key for the username is database-user . The name of the secret must be provided as the DATABASE_DEPLOYMENT_NAME parameter to the deployment configuration template. |
POSTGRESQL_PASSWORD | wired to a secret | The password for the database(s) hosted by the postgresql Postgres server. The deployment configuration makes the assumption you have your database credentials stored in secrets (which you should), and the key for the username is database-password . The name of the secret must be provided as the DATABASE_DEPLOYMENT_NAME parameter to the deployment configuration template. |
The following variables are NOT supported:
Name | Default (if not set) | Purpose |
---|---|---|
BACKUP_PERIOD | 1d | The schedule on which to run the backups. The value is replaced by the cron schedule variable (SCHEDULE) |
The scheduled job does not yet support the FTP environment variables.
Name |
---|
FTP_URL |
FTP_USER |
FTP_PASSWORD |
The backup-container is assigned with Best-effort
resource type (setting zero for request and limit), which allows the resources to scale up and down without an explicit limit as resource on the node allow. It benefits from large bursts of recourses for short periods of time to get things more quickly. After some time of running the backup-container, you could then set the request and limit according to the average resource consumption.
When backing up multiple databases, the retention settings apply to each database individually. For instance if you use the daily
strategy and set the retention number(s) to 5, you will retain 5 copies of each database. So plan your backup storage accordingly.
An example of the backup container in action can be found here; example log output
The backup
app supports two backup strategies, each are explained below. Regardless of the strategy backups are identified using a core name derived from the host/database
specification and a timestamp. All backups are compressed using gzip.
The daily backup strategy is very simple. Backups are created in dated folders under the top level /backups/
folder. When the maximum number of backups (NUM_BACKUPS
) is exceeded, the oldest ones are pruned from disk.
For example (faked):
================================================================================================================================
Current Backups:
--------------------------------------------------------------------------------------------------------------------------------
1.0K 2018-10-03 22:16 ./backups/2018-10-03/postgresql-TheOrgBook_Database_2018-10-03_22-16-11.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/postgresql-TheOrgBook_Database_2018-10-03_22-16-28.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/postgresql-TheOrgBook_Database_2018-10-03_22-16-46.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/wallet-db-tob_holder_2018-10-03_22-16-13.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/wallet-db-tob_holder_2018-10-03_22-16-31.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/wallet-db-tob_holder_2018-10-03_22-16-48.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/wallet-db-tob_verifier_2018-10-03_22-16-08.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/wallet-db-tob_verifier_2018-10-03_22-16-25.sql.gz
1.0K 2018-10-03 22:16 ./backups/2018-10-03/wallet-db-tob_verifier_2018-10-03_22-16-43.sql.gz
13K 2018-10-03 22:16 ./backups/2018-10-03
...
61K 2018-10-04 10:43 ./backups/
================================================================================================================================
The rolling backup strategy provides a bit more flexibility. It allows you to keep a number of recent daily
backups, a number of weekly
backups, and a number of monthly
backups.
- Daily backups are any backups done Monday through Saturday.
- Weekly backups are any backups done at the end of the week, which we're calling Sunday.
- Monthly backups are any backups done on the last day of a month.
There are retention settings you can set for each. The defaults provide you with a week's worth of daily
backups, a month's worth of weekly
backups, and a single backup for the previous month.
Although the example does not show any weekly
or monthly
backups, you can see from the example that the folders are further broken down into the backup type.
For example (faked):
================================================================================================================================
Current Backups:
--------------------------------------------------------------------------------------------------------------------------------
0 2018-10-03 22:16 ./backups/daily/2018-10-03
1.0K 2018-10-04 09:29 ./backups/daily/2018-10-04/postgresql-TheOrgBook_Database_2018-10-04_09-29-52.sql.gz
1.0K 2018-10-04 10:37 ./backups/daily/2018-10-04/postgresql-TheOrgBook_Database_2018-10-04_10-37-15.sql.gz
1.0K 2018-10-04 09:29 ./backups/daily/2018-10-04/wallet-db-tob_holder_2018-10-04_09-29-55.sql.gz
1.0K 2018-10-04 10:37 ./backups/daily/2018-10-04/wallet-db-tob_holder_2018-10-04_10-37-18.sql.gz
1.0K 2018-10-04 09:29 ./backups/daily/2018-10-04/wallet-db-tob_verifier_2018-10-04_09-29-49.sql.gz
1.0K 2018-10-04 10:37 ./backups/daily/2018-10-04/wallet-db-tob_verifier_2018-10-04_10-37-12.sql.gz
22K 2018-10-04 10:43 ./backups/daily/2018-10-04
22K 2018-10-04 10:43 ./backups/daily
4.0K 2018-10-03 22:16 ./backups/monthly/2018-10-03
4.0K 2018-10-03 22:16 ./backups/monthly
4.0K 2018-10-03 22:16 ./backups/weekly/2018-10-03
4.0K 2018-10-03 22:16 ./backups/weekly
61K 2018-10-04 10:43 ./backups/
================================================================================================================================
The backup script has a few utility features built into it. For a full list of features and documentation run backup.sh -h
.
Features include:
- The ability to list the existing backups,
backup.sh -l
- Listing the current configuration,
backup.sh -c
- Running a single backup cycle,
backup.sh -1
- Restoring a database from backup,
backup.sh -r <databaseSpec/> [-f <backupFileFilter>]
- Restore mode will allow you to restore a database to a different location (host, and/or database name) provided it can contact the host and you can provide the appropriate credentials.
- Verifying backups,
backup.sh [-s] -v <databaseSpec/> [-f <backupFileFilter>]
- Verify mode will restore a backup to the local server to ensure it can be restored without error. Once restored a table query is performed to ensure there was at least one table restored and queries against the database succeed without error. All database files and configuration are destroyed following the tests.
The backup script supports running manual or scheduled verifications on your backups; backup.sh [-s] -v <databaseSpec/> [-f <backupFileFilter>]
. Refer to the script documentation backup.sh -h
, and the configuration documentation, backup.conf, for additional details on how to use this feature.
- The FTP backup feature is enabled by specifying the FTP server URL
FTP_URL
. - The FTP server must support FTPS.
- Path can be added to the URL. For example, the URL can be
ftp://ftp.gov.bc.ca/schoolbus-db-backup/
. Note that when adding path, the URL must be ended with/
as the example. - The username and password must be populated in the secret key. Refer to the deployment configuration section.
- There is a known issue for FTPS with Windows 2012 FTP. http://redoubtsolutions.com/fix-the-supplied-message-is-incomplete-error-when-you-use-an-ftps-client-to-upload-a-file-in-windows/
The Webhook integration feature is enabled by specifying the webhook URL, WEBHOOK_URL
, in your configuration. It's recommended that you also provide values for ENVIRONMENT_FRIENDLY_NAME
and ENVIRONMENT_NAME
, so you can better identify the environment from which the messages originate and do things like produce links to the environment.
The Webhook integration feature was built with Rocket.Chat in mind and an integration script for Rocket.Chat can be found in rocket.chat.integration.js. This script was developed to support the BC OpenShift Pathfinder environment and will format the notifications from the backup script into Rocket.Chat messages (examples below). If you provide values for the environment name (ENVIRONMENT_FRIENDLY_NAME
and ENVIRONMENT_NAME
) hyperlinks will be added to the messages to link you to the pathfinder project console.
Sample Message:
Sample Error Message:
For information on how setup a webhook in Rocket.Chat refer to Incoming WebHook Scripting. The Webhook URL created during this process is the URL you use for WEBHOOK_URL
to enable the Webhook integration feature.
The backup container uses a plugin architecture to perform the database specific operations needed to support various database types.
The plugins are loaded dynamically based on the container type. By default the backup.null.plugin
will be loaded when the container type is not recognized.
To add support for a new database type:
- Update the
getContainerType
function in backup.container.utils to detect the new type of database. - Using the existing plugins as reference, implement the database specific scripts for the new database type.
- Using the existing docker files as reference, create a new one to build the new container type.
- Update the build and deployment templates and their documentation as needed.
- Update the project documentation as needed.
- Test, test, test.
- Submit a PR.
Plugin Examples:
-
- Postgres backup implementation.
-
- Mongo backup implementation.
-
- MSSQL backup implementation.
-
- Sample/Template backup implementation that simply outputs log messages for the various operations.
The following sections describes (some) postgres specific implementation, however the steps are generally the same between database implementations.
The purpose of the backup app is to do automatic backups. Deploy the Backup app to do daily backups. Viewing the Logs for the Backup App will show a record of backups that have been completed.
The Backup app performs the following sequence of operations:
- Create a directory that will be used to store the backup.
- Use the
pg_dump
andgzip
commands to make a backup. - Cull backups more than $NUM_BACKUPS (default 31 - configured in deployment script)
- Wait/Sleep for a period of time and repeat
Note that with the pod deployment, we support cron schedule(s) or the legacy mode (which uses a simple "sleep") to run the backup periodically. With the OpenShift Scheduled Job deployment, use the backup-cronjob.yaml template and set the schedule via the OpenShift cronjob object SCHEDULE template parameter.
A separate pod is used vs. having the backups run from the Postgres Pod for fault tolerant purposes - to keep the backups separate from the database storage. We don't want to, for example, lose the storage of the database, or have the database and backups storage fill up, and lose both the database and the backups.
- Check the logs of the Backup pod to make sure a backup isn't run right now (pretty unlikely...)
- Open a terminal window to the pod
- Run
backup.sh -1
- This will run a single backup cycle and exit.
- Run the following:
oc create job ${SOMEJOBNAME} --from=cronjob/${BACKUP_CRONJOB_NAME}
- example:
oc create job my-backup-1 --from=cronjob/backup-postgresql
- this will run a single backup job and exit.
- note: the jobs created in this manner are NOT cleaned up by the scheduler like the automated jobs are.
- example:
The backup.sh
script's restore mode makes it very simple to restore the most recent backup of a particular database. It's as simple as running a the following command, for example (run backup.sh -h
for full details on additional options);
backup.sh -r postgresql/TheOrgBook_Database
Following are more detailed steps to perform a restore of a backup.
- Log into the OpenShift Console and log into OpenShift on the command shell window.
- The instructions here use a mix of the console and command line, but all could be done from a command shell using "oc" commands.
- Scale to 0 all Apps that use the database connection.
- This is necessary as the Apps will need to restart to pull data from the restored backup.
- It is recommended that you also scale down to 0 your client application so that users know the application is unavailable while the database restore is underway.
- A nice addition to this would be a user-friendly "This application is offline" message - not yet implemented.
- Restart the database pod as a quick way of closing any other database connections from users using port forward or that have rsh'd to directly connect to the database.
- Open an rsh into the backup pod:
- Open a command prompt connection to OpenShift using
oc login
with parameters appropriate for your OpenShift host. - Change to the OpenShift project containing the Backup App
oc project <Project Name>
- List pods using
oc get pods
- Open a remote shell connection to the backup pod.
oc rsh <Backup Pod Name>
- Open a command prompt connection to OpenShift using
- In the rsh run the backup script in restore mode,
./backup.sh -r <DatabaseSpec/>
, to restore the desired backup file. For full information on how to use restore mode, refer to the script documentation,./backup.sh -h
. Have the Admin password for the database handy, the script will ask for it during the restore process.- The restore script will automatically grant the database user access to the restored database. If there are other users needing access to the database, such as the DBA group, you will need to additionally run the following commands on the database pod itself using
psql
:- Get a list of the users by running the command
\du
- For each user that is not "postgres" and $POSTGRESQL_USER, execute the command
GRANT SELECT ON ALL TABLES IN SCHEMA public TO "<name of user>";
- Get a list of the users by running the command
- If users have been set up with other grants, set them up as well.
- The restore script will automatically grant the database user access to the restored database. If there are other users needing access to the database, such as the DBA group, you will need to additionally run the following commands on the database pod itself using
- Verify that the database restore worked
- On the database pod, query a table - e.g the USER table:
SELECT * FROM "SBI_USER";
- you can look at other tables if you want. - Verify the expected data is shown.
- On the database pod, query a table - e.g the USER table:
- Exit remote shells back to your local command line
- From the Openshift Console restart the app:
- Scale up any pods you scaled down and wait for them to finish starting up. View the logs to verify there were no startup issues.
- Verify full application functionality.
Done!
Please refer to the Tips and Tricks document for solutions to known issues.
To report bugs/issues/feature requests, please file an issue.
If you would like to contribute, please see our CONTRIBUTING guidelines.
Please note that this project is released with a Contributor Code of Conduct. By participating in this project you agree to abide by its terms.