mirror of
https://github.com/jkaninda/mysql-bkup.git
synced 2025-12-08 14:39:41 +01:00
Compare commits
55 Commits
b6192f4c42
...
dependabot
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
197f1c1a8d | ||
| 8ac6ffc394 | |||
| 59296c0c17 | |||
| 19b64bbad1 | |||
| 640e11b924 | |||
|
|
a746366348 | ||
|
|
64d57a077e | ||
|
|
995db83b7c | ||
| 5d5cbea60b | |||
|
|
2cb5603b88 | ||
| 548de725dd | |||
|
|
37b2867974 | ||
| d6c01c9a4a | |||
|
|
bd92cc6844 | ||
| 3006abac3e | |||
|
|
3a16f6929c | ||
| 2f3ed7d0d8 | |||
|
|
a92bba05e4 | ||
| 7b565d54bd | |||
|
|
d10321dac6 | ||
| d151c50caa | |||
| 07ecec57b3 | |||
| bc4aab6ed0 | |||
| 12c17c18d6 | |||
| 80cd70e153 | |||
|
|
80abdb4299 | ||
| 72b21f00e6 | |||
| 3439c74257 | |||
| 4110ff4e64 | |||
| cb299a35bf | |||
| ba7c096bf3 | |||
| 6ac8dcef9e | |||
| 6bcc5d6bd4 | |||
| fd35fabf97 | |||
| 23cce10e8c | |||
| e666466d27 | |||
| 36bca254a9 | |||
| ad18d42145 | |||
| d9d44c2798 | |||
| 300a592508 | |||
| be82e841e7 | |||
| a73a365ebf | |||
| 75e965c0c5 | |||
| fc60ddb308 | |||
| 573ef15ef3 | |||
| b1776d3689 | |||
| 376d47f738 | |||
| eb6268f8ec | |||
| 731e2d789d | |||
| 6300a8f2dd | |||
| cd827a9277 | |||
| 71cf3fae85 | |||
| 528282bbd4 | |||
| 002c93a796 | |||
|
|
907e70d552 |
6
.github/workflows/deploy-docs.yml
vendored
6
.github/workflows/deploy-docs.yml
vendored
@@ -32,14 +32,14 @@ jobs:
|
||||
working-directory: docs
|
||||
- name: Setup Pages
|
||||
id: pages
|
||||
uses: actions/configure-pages@v2
|
||||
uses: actions/configure-pages@v5
|
||||
- name: Build with Jekyll
|
||||
working-directory: docs
|
||||
run: bundle exec jekyll build --baseurl "${{ steps.pages.outputs.base_path }}"
|
||||
env:
|
||||
JEKYLL_ENV: production
|
||||
- name: Upload artifact
|
||||
uses: actions/upload-pages-artifact@v1
|
||||
uses: actions/upload-pages-artifact@v3
|
||||
with:
|
||||
path: 'docs/_site/'
|
||||
|
||||
@@ -52,4 +52,4 @@ jobs:
|
||||
steps:
|
||||
- name: Deploy to GitHub Pages
|
||||
id: deployment
|
||||
uses: actions/deploy-pages@v1
|
||||
uses: actions/deploy-pages@v4
|
||||
23
.github/workflows/lint.yml
vendored
Normal file
23
.github/workflows/lint.yml
vendored
Normal file
@@ -0,0 +1,23 @@
|
||||
name: Lint
|
||||
|
||||
on:
|
||||
push:
|
||||
pull_request:
|
||||
|
||||
jobs:
|
||||
lint:
|
||||
name: Run on Ubuntu
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Clone the code
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Setup Go
|
||||
uses: actions/setup-go@v5
|
||||
with:
|
||||
go-version: '~1.23'
|
||||
|
||||
- name: Run linter
|
||||
uses: golangci/golangci-lint-action@v6
|
||||
with:
|
||||
version: v1.61
|
||||
289
.github/workflows/tests.yml
vendored
Normal file
289
.github/workflows/tests.yml
vendored
Normal file
@@ -0,0 +1,289 @@
|
||||
name: Tests
|
||||
|
||||
on:
|
||||
push:
|
||||
branches:
|
||||
- main
|
||||
- nightly
|
||||
pull_request:
|
||||
branches:
|
||||
- main
|
||||
env:
|
||||
IMAGE_NAME: mysql-bkup
|
||||
|
||||
jobs:
|
||||
test:
|
||||
runs-on: ubuntu-latest
|
||||
services:
|
||||
mysql:
|
||||
image: mysql:9
|
||||
env:
|
||||
MYSQL_ROOT_PASSWORD: password
|
||||
MYSQL_DATABASE: testdb
|
||||
MYSQL_USER: user
|
||||
MYSQL_PASSWORD: password
|
||||
ports:
|
||||
- 3306:3306
|
||||
options: >-
|
||||
--health-cmd="mysqladmin ping -h 127.0.0.1 -uuser -ppassword"
|
||||
--health-interval=10s
|
||||
--health-timeout=5s
|
||||
--health-retries=5
|
||||
mysql8:
|
||||
image: mysql:8
|
||||
env:
|
||||
MYSQL_ROOT_PASSWORD: password
|
||||
MYSQL_DATABASE: testdb
|
||||
MYSQL_USER: user
|
||||
MYSQL_PASSWORD: password
|
||||
ports:
|
||||
- 3308:3306
|
||||
options: >-
|
||||
--health-cmd="mysqladmin ping -h 127.0.0.1 -uuser -ppassword"
|
||||
--health-interval=10s
|
||||
--health-timeout=5s
|
||||
--health-retries=5
|
||||
mysql5:
|
||||
image: mysql:5
|
||||
env:
|
||||
MYSQL_ROOT_PASSWORD: password
|
||||
MYSQL_DATABASE: testdb
|
||||
MYSQL_USER: user
|
||||
MYSQL_PASSWORD: password
|
||||
ports:
|
||||
- 3305:3306
|
||||
options: >-
|
||||
--health-cmd="mysqladmin ping -h 127.0.0.1 -uuser -ppassword"
|
||||
--health-interval=10s
|
||||
--health-timeout=5s
|
||||
--health-retries=5
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Set up Docker Buildx
|
||||
uses: docker/setup-buildx-action@v3
|
||||
- name: Create Minio container
|
||||
run: |
|
||||
docker run -d --rm --name minio \
|
||||
--network host \
|
||||
-p 9000:9000 \
|
||||
-e MINIO_ACCESS_KEY=minioadmin \
|
||||
-e MINIO_SECRET_KEY=minioadmin \
|
||||
-e MINIO_REGION_NAME="eu" \
|
||||
minio/minio server /data
|
||||
echo "Create Minio container completed"
|
||||
- name: Install MinIO Client (mc)
|
||||
run: |
|
||||
curl -O https://dl.min.io/client/mc/release/linux-amd64/mc
|
||||
chmod +x mc
|
||||
sudo mv mc /usr/local/bin/
|
||||
|
||||
- name: Wait for MinIO to be ready
|
||||
run: sleep 5
|
||||
|
||||
- name: Configure MinIO Client
|
||||
run: |
|
||||
mc alias set local http://localhost:9000 minioadmin minioadmin
|
||||
mc alias list
|
||||
|
||||
- name: Create MinIO Bucket
|
||||
run: |
|
||||
mc mb local/backups
|
||||
echo "Bucket backups created successfully."
|
||||
# Build the Docker image
|
||||
- name: Build Docker Image
|
||||
run: |
|
||||
docker buildx build --build-arg appVersion=test -t ${{ env.IMAGE_NAME }}:latest --load .
|
||||
|
||||
- name: Verify Docker images
|
||||
run: |
|
||||
docker images
|
||||
|
||||
- name: Wait for MySQL to be ready
|
||||
run: |
|
||||
docker run --rm --network host mysql:9 mysqladmin ping -h 127.0.0.1 -uuser -ppassword --wait
|
||||
- name: Test restore
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=root \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest restore -f init.sql
|
||||
echo "Database restore completed"
|
||||
- name: Test restore Mysql8
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_PORT=3308 \
|
||||
-e DB_USERNAME=root \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest restore -f init.sql
|
||||
echo "Test restore Mysql8 completed"
|
||||
- name: Test restore Mysql5
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_PORT=3305 \
|
||||
-e DB_USERNAME=root \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest restore -f init.sql
|
||||
echo "Test restore Mysql5 completed"
|
||||
- name: Test backup
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=user \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest backup
|
||||
echo "Database backup completed"
|
||||
- name: Test backup Mysql8
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_PORT=3308 \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=user \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest backup
|
||||
echo "Test backup Mysql8 completed"
|
||||
- name: Test backup Mysql5
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_PORT=3305 \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=user \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest backup
|
||||
echo "Test backup Mysql5 completed"
|
||||
- name: Test encrypted backup
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=user \
|
||||
-e DB_PASSWORD=password \
|
||||
-e GPG_PASSPHRASE=password \
|
||||
${{ env.IMAGE_NAME }}:latest backup -d testdb --disable-compression --custom-name encrypted-bkup
|
||||
echo "Database encrypted backup completed"
|
||||
- name: Test restore encrypted backup | testdb -> testdb2
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=root \
|
||||
-e DB_PASSWORD=password \
|
||||
-e GPG_PASSPHRASE=password \
|
||||
-e DB_NAME=testdb2 \
|
||||
${{ env.IMAGE_NAME }}:latest restore -f /backup/encrypted-bkup.sql.gpg
|
||||
echo "Test restore encrypted backup completed"
|
||||
- name: Test migrate database testdb -> testdb3
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=root \
|
||||
-e DB_PASSWORD=password \
|
||||
-e GPG_PASSPHRASE=password \
|
||||
-e DB_NAME=testdb \
|
||||
-e TARGET_DB_HOST=127.0.0.1 \
|
||||
-e TARGET_DB_PORT=3306 \
|
||||
-e TARGET_DB_NAME=testdb3 \
|
||||
-e TARGET_DB_USERNAME=root \
|
||||
-e TARGET_DB_PASSWORD=password \
|
||||
${{ env.IMAGE_NAME }}:latest migrate
|
||||
echo "Test migrate database testdb -> testdb3 completed"
|
||||
- name: Test backup all databases
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=root \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest backup --all-databases
|
||||
echo "Database backup completed"
|
||||
- name: Test multiple backup
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e TESTDB2_DB_USERNAME=root \
|
||||
-e TESTDB2_DB_PASSWORD=password \
|
||||
-e TESTDB2_DB_HOST=127.0.0.1 \
|
||||
${{ env.IMAGE_NAME }}:latest backup -c /backup/test_config.yaml
|
||||
echo "Database backup completed"
|
||||
- name: Test backup Minio (s3)
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=user \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
-e AWS_S3_ENDPOINT="http://127.0.0.1:9000" \
|
||||
-e AWS_S3_BUCKET_NAME=backups \
|
||||
-e AWS_ACCESS_KEY=minioadmin \
|
||||
-e AWS_SECRET_KEY=minioadmin \
|
||||
-e AWS_DISABLE_SSL="true" \
|
||||
-e AWS_REGION="eu" \
|
||||
-e AWS_FORCE_PATH_STYLE="true" ${{ env.IMAGE_NAME }}:latest backup -s s3 --custom-name minio-backup
|
||||
echo "Test backup Minio (s3) completed"
|
||||
- name: Test restore Minio (s3)
|
||||
run: |
|
||||
docker run --rm --name ${{ env.IMAGE_NAME }} \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=user \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
-e AWS_S3_ENDPOINT="http://127.0.0.1:9000" \
|
||||
-e AWS_S3_BUCKET_NAME=backups \
|
||||
-e AWS_ACCESS_KEY=minioadmin \
|
||||
-e AWS_SECRET_KEY=minioadmin \
|
||||
-e AWS_DISABLE_SSL="true" \
|
||||
-e AWS_REGION="eu" \
|
||||
-e AWS_FORCE_PATH_STYLE="true" ${{ env.IMAGE_NAME }}:latest restore -s s3 -f minio-backup.sql.gz
|
||||
echo "Test backup Minio (s3) completed"
|
||||
- name: Test scheduled backup
|
||||
run: |
|
||||
docker run -d --rm --name ${{ env.IMAGE_NAME }} \
|
||||
-v ./migrations:/backup/ \
|
||||
--network host \
|
||||
-e DB_HOST=127.0.0.1 \
|
||||
-e DB_USERNAME=user \
|
||||
-e DB_PASSWORD=password \
|
||||
-e DB_NAME=testdb \
|
||||
${{ env.IMAGE_NAME }}:latest backup -e "@every 10s"
|
||||
|
||||
echo "Waiting for backup to be done..."
|
||||
sleep 25
|
||||
docker logs ${{ env.IMAGE_NAME }}
|
||||
echo "Test scheduled backup completed"
|
||||
# Cleanup: Stop and remove containers
|
||||
- name: Clean up
|
||||
run: |
|
||||
docker stop ${{ env.IMAGE_NAME }} || true
|
||||
docker rm ${{ env.IMAGE_NAME }} || true
|
||||
13
Dockerfile
13
Dockerfile
@@ -1,4 +1,4 @@
|
||||
FROM golang:1.24.0 AS build
|
||||
FROM golang:1.25.5 AS build
|
||||
WORKDIR /app
|
||||
ARG appVersion=""
|
||||
|
||||
@@ -10,7 +10,7 @@ RUN go mod download
|
||||
# Build
|
||||
RUN CGO_ENABLED=0 GOOS=linux go build -ldflags="-X 'github.com/jkaninda/mysql-bkup/utils.Version=${appVersion}'" -o /app/mysql-bkup
|
||||
|
||||
FROM alpine:3.21.3
|
||||
FROM alpine:3.22.1
|
||||
ENV TZ=UTC
|
||||
ARG WORKDIR="/config"
|
||||
ARG BACKUPDIR="/backup"
|
||||
@@ -18,9 +18,12 @@ ARG BACKUP_TMP_DIR="/tmp/backup"
|
||||
ARG TEMPLATES_DIR="/config/templates"
|
||||
ARG appVersion=""
|
||||
ENV VERSION=${appVersion}
|
||||
LABEL author="Jonas Kaninda"
|
||||
LABEL version=${appVersion}
|
||||
LABEL github="github.com/jkaninda/mysql-bkup"
|
||||
LABEL org.opencontainers.image.title="mysql-bkup"
|
||||
LABEL org.opencontainers.image.description="A lightweight MySQL backup and restore tool"
|
||||
LABEL org.opencontainers.image.licenses="MIT"
|
||||
LABEL org.opencontainers.image.authors="Jonas Kaninda <me@jonaskaninda.com>"
|
||||
LABEL org.opencontainers.image.version=${appVersion}
|
||||
LABEL org.opencontainers.image.source="https://github.com/jkaninda/mysql-bkup"
|
||||
|
||||
RUN apk --update add --no-cache mysql-client mariadb-connector-c tzdata ca-certificates
|
||||
RUN mkdir -p $WORKDIR $BACKUPDIR $TEMPLATES_DIR $BACKUP_TMP_DIR && \
|
||||
|
||||
203
README.md
203
README.md
@@ -3,6 +3,14 @@
|
||||
**MYSQL-BKUP** is a Docker container image designed to **backup, restore, and migrate MySQL databases**.
|
||||
It supports a variety of storage options and ensures data security through GPG encryption.
|
||||
|
||||
MYSQL-BKUP is designed for seamless deployment on **Docker** and **Kubernetes**, simplifying MySQL backup, restoration, and migration across environments.
|
||||
It is a lightweight, multi-architecture solution compatible with **Docker**, **Docker Swarm**, **Kubernetes**, and other container orchestration platforms.
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
[](https://github.com/jkaninda/mysql-bkup/actions/workflows/tests.yml)
|
||||
[](https://github.com/jkaninda/mysql-bkup/actions/workflows/release.yml)
|
||||
[](https://goreportcard.com/report/github.com/jkaninda/mysql-bkup)
|
||||

|
||||
@@ -11,9 +19,9 @@ It supports a variety of storage options and ensures data security through GPG e
|
||||
|
||||
## Features
|
||||
|
||||
- **Storage Options:**
|
||||
- Local storage
|
||||
- AWS S3 or any S3-compatible object storage
|
||||
- **Flexible Storage Backends:**
|
||||
- Local filesystem
|
||||
- Amazon S3 & S3-compatible storage (e.g., MinIO, Wasabi)
|
||||
- FTP
|
||||
- SSH-compatible storage
|
||||
- Azure Blob storage
|
||||
@@ -33,16 +41,19 @@ It supports a variety of storage options and ensures data security through GPG e
|
||||
- **Telegram**
|
||||
- **Email**
|
||||
|
||||
## Use Cases
|
||||
## 💡Use Cases
|
||||
|
||||
- **Automated Recurring Backups:** Schedule regular backups for MySQL databases.
|
||||
- **Cross-Environment Migration:** Easily migrate your MySQL databases across different environments using supported storage options.
|
||||
- **Secure Backup Management:** Protect your data with GPG encryption.
|
||||
- **Scheduled Backups**: Automate recurring backups using Docker or Kubernetes.
|
||||
- **Disaster Recovery:** Quickly restore backups to a clean MySQL instance.
|
||||
- **Database Migration**: Seamlessly move data across environments using the built-in `migrate` feature.
|
||||
- **Secure Archiving:** Keep backups encrypted and safely stored in the cloud or remote servers.
|
||||
|
||||
|
||||
Successfully tested on:
|
||||
## ✅ Verified Platforms:
|
||||
MYSQL-BKUP has been tested and runs successfully on:
|
||||
|
||||
- Docker
|
||||
- Docker in Swarm mode
|
||||
- Docker Swarm
|
||||
- Kubernetes
|
||||
- OpenShift
|
||||
|
||||
@@ -58,17 +69,12 @@ Successfully tested on:
|
||||
|
||||
- [PostgreSQL](https://github.com/jkaninda/pg-bkup)
|
||||
|
||||
## Storage:
|
||||
- Local
|
||||
- AWS S3 or any S3 Alternatives for Object Storage
|
||||
- SSH remote storage server
|
||||
- FTP remote storage server
|
||||
- Azure Blob storage
|
||||
|
||||
## Quickstart
|
||||
|
||||
### Simple backup using Docker CLI
|
||||
### Simple Backup Using Docker CLI
|
||||
|
||||
To run a one time backup, bind your local volume to `/backup` in the container and run the `backup` command:
|
||||
To perform a one-time backup, bind your local volume to `/backup` in the container and run the `backup` command:
|
||||
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
@@ -80,17 +86,36 @@ To run a one time backup, bind your local volume to `/backup` in the container a
|
||||
jkaninda/mysql-bkup backup -d database_name
|
||||
```
|
||||
|
||||
Alternatively, pass a `--env-file` in order to use a full config as described below.
|
||||
Alternatively, use an environment file (`--env-file`) for configuration:
|
||||
|
||||
```yaml
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
--env-file your-env-file \
|
||||
-v $PWD/backup:/backup/ \
|
||||
jkaninda/mysql-bkup backup -d database_name
|
||||
```
|
||||
### Simple restore using Docker CLI
|
||||
|
||||
To restore a database, bind your local volume to `/backup` in the container and run the `restore` command:
|
||||
### Backup All Databases
|
||||
|
||||
To back up all databases on the server, use the `--all-databases` or `-a` flag. By default, this creates individual backup files for each database.
|
||||
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
-v $PWD/backup:/backup/ \
|
||||
-e "DB_HOST=dbhost" \
|
||||
-e "DB_PORT=3306" \
|
||||
-e "DB_USERNAME=username" \
|
||||
-e "DB_PASSWORD=password" \
|
||||
jkaninda/mysql-bkup backup --all-databases --disable-compression
|
||||
```
|
||||
|
||||
> **Note:** Use the `--all-in-one` or `-A` flag to combine backups into a single file.
|
||||
|
||||
---
|
||||
|
||||
### Simple Restore Using Docker CLI
|
||||
|
||||
To restore a database, bind your local volume to `/backup` and run the `restore` command:
|
||||
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
@@ -101,15 +126,18 @@ To restore a database, bind your local volume to `/backup` in the container and
|
||||
-e "DB_PASSWORD=password" \
|
||||
jkaninda/mysql-bkup restore -d database_name -f backup_file.sql.gz
|
||||
```
|
||||
### Simple backup in docker compose file
|
||||
|
||||
---
|
||||
|
||||
### Backup with Docker Compose
|
||||
|
||||
Below is an example of a `docker-compose.yml` file for running a one-time backup:
|
||||
|
||||
```yaml
|
||||
services:
|
||||
mysql-bkup:
|
||||
# In production, it is advised to lock your image tag to a proper
|
||||
# release version instead of using `latest`.
|
||||
# Check https://github.com/jkaninda/mysql-bkup/releases
|
||||
# for a list of available releases.
|
||||
pg-bkup:
|
||||
# In production, pin your image tag to a specific release version instead of `latest`.
|
||||
# See available releases: https://github.com/jkaninda/mysql-bkup/releases
|
||||
image: jkaninda/mysql-bkup
|
||||
container_name: mysql-bkup
|
||||
command: backup
|
||||
@@ -122,13 +150,18 @@ services:
|
||||
- DB_USERNAME=bar
|
||||
- DB_PASSWORD=password
|
||||
- TZ=Europe/Paris
|
||||
# mysql-bkup container must be connected to the same network with your database
|
||||
networks:
|
||||
- web
|
||||
|
||||
networks:
|
||||
web:
|
||||
```
|
||||
### Docker recurring backup
|
||||
|
||||
---
|
||||
|
||||
### Recurring Backups with Docker
|
||||
|
||||
You can schedule recurring backups using the `--cron-expression` or `-e` flag:
|
||||
|
||||
```shell
|
||||
docker run --rm --network network_name \
|
||||
@@ -136,15 +169,20 @@ networks:
|
||||
-e "DB_HOST=hostname" \
|
||||
-e "DB_USERNAME=user" \
|
||||
-e "DB_PASSWORD=password" \
|
||||
jkaninda/mysql-bkup backup -d dbName --cron-expression "@every 15m" #@midnight
|
||||
jkaninda/mysql-bkup backup -d dbName --cron-expression "@every 15m"
|
||||
```
|
||||
See: https://jkaninda.github.io/mysql-bkup/reference/#predefined-schedules
|
||||
|
||||
For predefined schedules, refer to the [documentation](https://jkaninda.github.io/mysql-bkup/reference/#predefined-schedules).
|
||||
|
||||
---
|
||||
|
||||
## Deploy on Kubernetes
|
||||
|
||||
For Kubernetes, you don't need to run it in scheduled mode. You can deploy it as Job or CronJob.
|
||||
For Kubernetes, you can deploy `mysql-bkup` as a Job or CronJob. Below are examples for both.
|
||||
|
||||
### Simple Kubernetes backup Job :
|
||||
### Kubernetes Backup Job
|
||||
|
||||
This example defines a one-time backup job:
|
||||
|
||||
```yaml
|
||||
apiVersion: batch/v1
|
||||
@@ -157,15 +195,10 @@ spec:
|
||||
spec:
|
||||
containers:
|
||||
- name: mysql-bkup
|
||||
# In production, it is advised to lock your image tag to a proper
|
||||
# release version instead of using `latest`.
|
||||
# Check https://github.com/jkaninda/mysql-bkup/releases
|
||||
# for a list of available releases.
|
||||
# Pin the image tag to a specific release version in production.
|
||||
# See available releases: https://github.com/jkaninda/mysql-bkup/releases
|
||||
image: jkaninda/mysql-bkup
|
||||
command:
|
||||
- /bin/sh
|
||||
- -c
|
||||
- backup -d dbname
|
||||
command: ["backup", "-d", "dbname"]
|
||||
resources:
|
||||
limits:
|
||||
memory: "128Mi"
|
||||
@@ -183,10 +216,94 @@ spec:
|
||||
volumes:
|
||||
- name: backup
|
||||
hostPath:
|
||||
path: /home/toto/backup # directory location on host
|
||||
type: Directory # this field is optional
|
||||
path: /home/toto/backup # Directory location on the host
|
||||
type: Directory # Optional field
|
||||
restartPolicy: Never
|
||||
```
|
||||
|
||||
### Kubernetes CronJob for Scheduled Backups
|
||||
|
||||
For scheduled backups, use a `CronJob`:
|
||||
|
||||
```yaml
|
||||
apiVersion: batch/v1
|
||||
kind: CronJob
|
||||
metadata:
|
||||
name: pg-bkup-cronjob
|
||||
spec:
|
||||
schedule: "0 2 * * *" # Runs daily at 2 AM
|
||||
jobTemplate:
|
||||
spec:
|
||||
template:
|
||||
spec:
|
||||
containers:
|
||||
- name: pg-bkup
|
||||
image: jkaninda/mysql-bkup
|
||||
command: ["backup", "-d", "dbname"]
|
||||
env:
|
||||
- name: DB_HOST
|
||||
value: "mysql"
|
||||
- name: DB_USERNAME
|
||||
value: "user"
|
||||
- name: DB_PASSWORD
|
||||
value: "password"
|
||||
volumeMounts:
|
||||
- mountPath: /backup
|
||||
name: backup
|
||||
volumes:
|
||||
- name: backup
|
||||
hostPath:
|
||||
path: /home/toto/backup
|
||||
type: Directory
|
||||
restartPolicy: OnFailure
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## 🚀 Why Use MYSQL-BKUP?
|
||||
|
||||
**MYSQL-BKUP** isn't just another MySQL backup tool, it's a robust, production-ready solution purpose-built for modern DevOps workflows.
|
||||
|
||||
Here’s why developers, sysadmins, and DevOps choose **MYSQL-BKUP**:
|
||||
|
||||
### ✅ All-in-One Backup, Restore & Migration
|
||||
|
||||
Whether you're backing up a single database, restoring critical data, or migrating across environments, MYSQL-BKUP handles it all with a **single, unified CLI** no scripting gymnastics required.
|
||||
|
||||
|
||||
### 🔄 Works Everywhere You Deploy
|
||||
|
||||
Designed to be cloud-native:
|
||||
|
||||
* **Runs seamlessly on Docker, Docker Swarm, and Kubernetes**
|
||||
* Supports **CronJobs** for automated scheduled backups
|
||||
* Compatible with GitOps and CI/CD workflows
|
||||
|
||||
### ☁️ Flexible Storage Integrations
|
||||
|
||||
Store your backups **anywhere**:
|
||||
|
||||
* Local disks
|
||||
* Amazon S3, MinIO, Wasabi, Azure Blob, FTP, SSH
|
||||
|
||||
### 🔒 Enterprise-Grade Security
|
||||
|
||||
* **GPG Encryption**: Protect sensitive data with optional encryption before storing backups locally or in the cloud.
|
||||
* **Secure Storage** Options: Supports S3, Azure Blob, SFTP, and SSH with encrypted transfers, keeping backups safe from unauthorized access.
|
||||
|
||||
### 📬 Instant Notifications
|
||||
|
||||
Stay in the loop with real-time notifications via **Telegram** and **Email**. Know immediately when a backup succeeds—or fails.
|
||||
|
||||
### 🏃♂️ Lightweight and Fast
|
||||
|
||||
Written in **Go**, MYSQL-BKUP is fast, multi-arch compatible (`amd64`, `arm64`, `arm/v7`), and optimized for minimal memory and CPU usage. Ideal for both cloud and edge deployments.
|
||||
|
||||
### 🧪 Tested. Verified. Trusted.
|
||||
|
||||
Actively maintained with **automated testing**, **Docker image size optimizations**, and verified support across major container platforms.
|
||||
|
||||
---
|
||||
## Available image registries
|
||||
|
||||
This Docker image is published to both Docker Hub and the GitHub container registry.
|
||||
|
||||
@@ -50,6 +50,8 @@ func init() {
|
||||
BackupCmd.PersistentFlags().StringP("cron-expression", "e", "", "Backup cron expression (e.g., `0 0 * * *` or `@daily`)")
|
||||
BackupCmd.PersistentFlags().StringP("config", "c", "", "Configuration file for multi database backup. (e.g: `/backup/config.yaml`)")
|
||||
BackupCmd.PersistentFlags().BoolP("disable-compression", "", false, "Disable backup compression")
|
||||
BackupCmd.PersistentFlags().BoolP("all", "a", false, "Backup all databases")
|
||||
BackupCmd.PersistentFlags().BoolP("all-databases", "a", false, "Backup all databases")
|
||||
BackupCmd.PersistentFlags().BoolP("all-in-one", "A", false, "Backup all databases in a single file")
|
||||
BackupCmd.PersistentFlags().StringP("custom-name", "", "", "Custom backup name")
|
||||
|
||||
}
|
||||
|
||||
61
docs/how-tos/backup-all.md
Normal file
61
docs/how-tos/backup-all.md
Normal file
@@ -0,0 +1,61 @@
|
||||
---
|
||||
title: Backup all databases in the server
|
||||
layout: default
|
||||
parent: How Tos
|
||||
nav_order: 12
|
||||
---
|
||||
|
||||
# Backup All Databases
|
||||
|
||||
MySQL-Bkup supports backing up all databases on the server using the `--all-databases` (`-a`) flag. By default, this creates separate backup files for each database. If you prefer a single backup file, you can use the `--all-in-one` (`-A`) flag.
|
||||
|
||||
Backing up all databases is useful for creating a snapshot of the entire database server, whether for disaster recovery or migration purposes.
|
||||
## Backup Modes
|
||||
|
||||
### Separate Backup Files (Default)
|
||||
|
||||
Using --all-databases without --all-in-one creates individual backup files for each database.
|
||||
|
||||
- Creates separate backup files for each database.
|
||||
- Provides more flexibility in restoring individual databases or tables.
|
||||
- Can be more manageable in cases where different databases have different retention policies.
|
||||
- Might take slightly longer due to multiple file operations.
|
||||
- It is the default behavior when using the `--all-databases` flag.
|
||||
- It does not backup system databases (`information_schema`, `performance_schema`, `mysql`, `sys`, `innodb`,...).
|
||||
|
||||
**Command:**
|
||||
|
||||
```bash
|
||||
docker run --rm --network your_network_name \
|
||||
-v $PWD/backup:/backup/ \
|
||||
-e "DB_HOST=dbhost" \
|
||||
-e "DB_PORT=3306" \
|
||||
-e "DB_USERNAME=username" \
|
||||
-e "DB_PASSWORD=password" \
|
||||
jkaninda/mysql-bkup backup --all-databases
|
||||
```
|
||||
### Single Backup File
|
||||
|
||||
Using --all-in-one (-A) creates a single backup file containing all databases.
|
||||
|
||||
- Creates a single backup file containing all databases.
|
||||
- Easier to manage if you need to restore everything at once.
|
||||
- Faster to back up and restore in bulk.
|
||||
- Can be problematic if you only need to restore a specific database or table.
|
||||
- It is recommended to use this option for disaster recovery purposes.
|
||||
- It backups system databases as well.
|
||||
|
||||
```bash
|
||||
docker run --rm --network your_network_name \
|
||||
-v $PWD/backup:/backup/ \
|
||||
-e "DB_HOST=dbhost" \
|
||||
-e "DB_PORT=3306" \
|
||||
-e "DB_USERNAME=username" \
|
||||
-e "DB_PASSWORD=password" \
|
||||
jkaninda/mysql-bkup backup --all-in-one
|
||||
```
|
||||
|
||||
### When to Use Which?
|
||||
|
||||
- Use `--all-in-one` if you want a quick, simple backup for disaster recovery where you'll restore everything at once.
|
||||
- Use `--all-databases` if you need granularity in restoring specific databases or tables without affecting others.
|
||||
@@ -19,7 +19,7 @@ The configuration file can be mounted into the container at `/config/config.yaml
|
||||
|
||||
### Key Features:
|
||||
- **Global Environment Variables**: Use these for databases that share the same configuration.
|
||||
- **Database-Specific Overrides**: Override global settings for individual databases by specifying them in the configuration file or using the database name as a suffix in the variable name (e.g., `DB_HOST_DATABASE1`).
|
||||
- **Database-Specific Overrides**: Override global settings for individual databases by specifying them in the configuration file or using the database name as a prefix or suffix in the variable name (e.g., `DB_HOST_DATABASENAME` or `DATABASENAME_DB_HOST`).
|
||||
- **Global Cron Expression**: Define a global `cronExpression` in the configuration file to schedule backups for all databases. If omitted, backups will run immediately.
|
||||
- **Configuration File Path**: Specify the configuration file path using:
|
||||
- The `BACKUP_CONFIG_FILE` environment variable.
|
||||
@@ -34,8 +34,8 @@ Below is an example configuration file (`config.yaml`) that defines multiple dat
|
||||
```yaml
|
||||
# Optional: Define a global cron expression for scheduled backups.
|
||||
# Example: "@every 20m" (runs every 20 minutes). If omitted, backups run immediately.
|
||||
cronExpression: ""
|
||||
|
||||
cronExpression: "" # Optional: Define a global cron expression for scheduled backups.
|
||||
backupRescueMode: false # Optional: Set to true to enable rescue mode for backups.
|
||||
databases:
|
||||
- host: mysql1 # Optional: Overrides DB_HOST or uses DB_HOST_DATABASE1.
|
||||
port: 3306 # Optional: Default is 5432. Overrides DB_PORT or uses DB_PORT_DATABASE1.
|
||||
@@ -89,7 +89,7 @@ services:
|
||||
environment:
|
||||
## Specify the path to the configuration file
|
||||
- BACKUP_CONFIG_FILE=/backup/config.yaml
|
||||
# Ensure the pg-bkup container is connected to the same network as your database
|
||||
# Ensure the mysql-bkup container is connected to the same network as your database
|
||||
networks:
|
||||
- web
|
||||
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
title: Receive notifications
|
||||
layout: default
|
||||
parent: How Tos
|
||||
nav_order: 12
|
||||
nav_order: 13
|
||||
---
|
||||
|
||||
# Receive Notifications
|
||||
|
||||
@@ -10,6 +10,8 @@ nav_order: 1
|
||||
**MYSQL-BKUP** is a Docker container image designed to **backup, restore, and migrate MySQL databases**.
|
||||
It supports a variety of storage options and ensures data security through GPG encryption.
|
||||
|
||||
**MYSQL-BKUP** is designed for seamless deployment on **Docker** and **Kubernetes**, simplifying MySQL backup, restoration, and migration across environments.
|
||||
It is a lightweight, multi-architecture solution compatible with **Docker**, **Docker Swarm**, **Kubernetes**, and other container orchestration platforms.
|
||||
---
|
||||
|
||||
## Key Features
|
||||
@@ -39,11 +41,21 @@ It supports a variety of storage options and ensures data security through GPG e
|
||||
|
||||
---
|
||||
|
||||
## Use Cases
|
||||
## 💡Use Cases
|
||||
|
||||
- **Automated Recurring Backups:** Schedule regular backups for MySQL databases.
|
||||
- **Cross-Environment Migration:** Easily migrate MySQL databases across different environments using supported storage options.
|
||||
- **Secure Backup Management:** Protect your data with GPG encryption.
|
||||
- **Scheduled Backups**: Automate recurring backups using Docker or Kubernetes.
|
||||
- **Disaster Recovery:** Quickly restore backups to a clean MySQL instance.
|
||||
- **Database Migration**: Seamlessly move data across environments using the built-in `migrate` feature.
|
||||
- **Secure Archiving:** Keep backups encrypted and safely stored in the cloud or remote servers.
|
||||
|
||||
|
||||
## ✅ Verified Platforms:
|
||||
MYSQL-BKUP has been tested and runs successfully on:
|
||||
|
||||
- Docker
|
||||
- Docker Swarm
|
||||
- Kubernetes
|
||||
- OpenShift
|
||||
|
||||
---
|
||||
|
||||
|
||||
@@ -10,11 +10,11 @@ This guide provides quick examples for running backups using Docker CLI, Docker
|
||||
|
||||
---
|
||||
|
||||
## Simple Backup Using Docker CLI
|
||||
### Simple Backup Using Docker CLI
|
||||
|
||||
To run a one-time backup, bind your local volume to `/backup` in the container and execute the `backup` command:
|
||||
To perform a one-time backup, bind your local volume to `/backup` in the container and run the `backup` command:
|
||||
|
||||
```bash
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
-v $PWD/backup:/backup/ \
|
||||
-e "DB_HOST=dbhost" \
|
||||
@@ -24,20 +24,36 @@ docker run --rm --network your_network_name \
|
||||
jkaninda/mysql-bkup backup -d database_name
|
||||
```
|
||||
|
||||
### Using an Environment File
|
||||
Alternatively, use an environment file (`--env-file`) for configuration:
|
||||
|
||||
Alternatively, you can use an `--env-file` to pass a full configuration:
|
||||
|
||||
```bash
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
--env-file your-env-file \
|
||||
-v $PWD/backup:/backup/ \
|
||||
jkaninda/mysql-bkup backup -d database_name
|
||||
```
|
||||
|
||||
### Simple restore using Docker CLI
|
||||
### Backup All Databases
|
||||
|
||||
To restore a database, bind your local volume to `/backup` in the container and run the `restore` command:
|
||||
To back up all databases on the server, use the `--all-databases` or `-a` flag. By default, this creates individual backup files for each database.
|
||||
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
-v $PWD/backup:/backup/ \
|
||||
-e "DB_HOST=dbhost" \
|
||||
-e "DB_PORT=3306" \
|
||||
-e "DB_USERNAME=username" \
|
||||
-e "DB_PASSWORD=password" \
|
||||
jkaninda/mysql-bkup backup --all-databases --disable-compression
|
||||
```
|
||||
|
||||
> **Note:** Use the `--all-in-one` or `-A` flag to combine backups into a single file.
|
||||
|
||||
---
|
||||
|
||||
### Simple Restore Using Docker CLI
|
||||
|
||||
To restore a database, bind your local volume to `/backup` and run the `restore` command:
|
||||
|
||||
```shell
|
||||
docker run --rm --network your_network_name \
|
||||
@@ -48,17 +64,18 @@ To restore a database, bind your local volume to `/backup` in the container and
|
||||
-e "DB_PASSWORD=password" \
|
||||
jkaninda/mysql-bkup restore -d database_name -f backup_file.sql.gz
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Simple Backup Using Docker Compose
|
||||
### Backup with Docker Compose
|
||||
|
||||
Below is an example `docker-compose.yml` configuration for running a backup:
|
||||
Below is an example of a `docker-compose.yml` file for running a one-time backup:
|
||||
|
||||
```yaml
|
||||
services:
|
||||
mysql-bkup:
|
||||
# In production, lock the image tag to a specific release version.
|
||||
# Check https://github.com/jkaninda/mysql-bkup/releases for available releases.
|
||||
pg-bkup:
|
||||
# In production, pin your image tag to a specific release version instead of `latest`.
|
||||
# See available releases: https://github.com/jkaninda/mysql-bkup/releases
|
||||
image: jkaninda/mysql-bkup
|
||||
container_name: mysql-bkup
|
||||
command: backup
|
||||
@@ -71,7 +88,6 @@ services:
|
||||
- DB_USERNAME=bar
|
||||
- DB_PASSWORD=password
|
||||
- TZ=Europe/Paris
|
||||
# Ensure the mysql-bkup container is connected to the same network as your database.
|
||||
networks:
|
||||
- web
|
||||
|
||||
@@ -81,11 +97,11 @@ networks:
|
||||
|
||||
---
|
||||
|
||||
## Recurring Backup with Docker
|
||||
### Recurring Backups with Docker
|
||||
|
||||
To schedule recurring backups, use the `--cron-expression` flag:
|
||||
You can schedule recurring backups using the `--cron-expression` or `-e` flag:
|
||||
|
||||
```bash
|
||||
```shell
|
||||
docker run --rm --network network_name \
|
||||
-v $PWD/backup:/backup/ \
|
||||
-e "DB_HOST=hostname" \
|
||||
@@ -98,9 +114,13 @@ For predefined schedules, refer to the [documentation](https://jkaninda.github.i
|
||||
|
||||
---
|
||||
|
||||
## Backup Using Kubernetes
|
||||
## Deploy on Kubernetes
|
||||
|
||||
Below is an example Kubernetes `Job` configuration for running a backup:
|
||||
For Kubernetes, you can deploy `mysql-bkup` as a Job or CronJob. Below are examples for both.
|
||||
|
||||
### Kubernetes Backup Job
|
||||
|
||||
This example defines a one-time backup job:
|
||||
|
||||
```yaml
|
||||
apiVersion: batch/v1
|
||||
@@ -113,8 +133,8 @@ spec:
|
||||
spec:
|
||||
containers:
|
||||
- name: mysql-bkup
|
||||
# In production, lock the image tag to a specific release version.
|
||||
# Check https://github.com/jkaninda/mysql-bkup/releases for available releases.
|
||||
# Pin the image tag to a specific release version in production.
|
||||
# See available releases: https://github.com/jkaninda/mysql-bkup/releases
|
||||
image: jkaninda/mysql-bkup
|
||||
command:
|
||||
- /bin/sh
|
||||
@@ -128,7 +148,7 @@ spec:
|
||||
- name: DB_HOST
|
||||
value: "mysql"
|
||||
- name: DB_USERNAME
|
||||
value: "postgres"
|
||||
value: "user"
|
||||
- name: DB_PASSWORD
|
||||
value: "password"
|
||||
volumeMounts:
|
||||
@@ -142,6 +162,46 @@ spec:
|
||||
restartPolicy: Never
|
||||
```
|
||||
|
||||
### Kubernetes CronJob for Scheduled Backups
|
||||
|
||||
For scheduled backups, use a `CronJob`:
|
||||
|
||||
```yaml
|
||||
apiVersion: batch/v1
|
||||
kind: CronJob
|
||||
metadata:
|
||||
name: pg-bkup-cronjob
|
||||
spec:
|
||||
schedule: "0 2 * * *" # Runs daily at 2 AM
|
||||
jobTemplate:
|
||||
spec:
|
||||
template:
|
||||
spec:
|
||||
containers:
|
||||
- name: pg-bkup
|
||||
image: jkaninda/mysql-bkup
|
||||
command:
|
||||
- /bin/sh
|
||||
- -c
|
||||
- backup -d dbname
|
||||
env:
|
||||
- name: DB_HOST
|
||||
value: "mysql"
|
||||
- name: DB_USERNAME
|
||||
value: "user"
|
||||
- name: DB_PASSWORD
|
||||
value: "password"
|
||||
volumeMounts:
|
||||
- mountPath: /backup
|
||||
name: backup
|
||||
volumes:
|
||||
- name: backup
|
||||
hostPath:
|
||||
path: /home/toto/backup
|
||||
type: Directory
|
||||
restartPolicy: OnFailure
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Key Notes
|
||||
|
||||
@@ -6,28 +6,31 @@ nav_order: 3
|
||||
|
||||
# Configuration Reference
|
||||
|
||||
Backup, restore, and migration targets, schedules, and retention policies are configured using **environment variables** or **CLI flags**.
|
||||
|
||||
---
|
||||
MySQL backup, restore, and migration processes can be configured using **environment variables** or **CLI flags**.
|
||||
|
||||
## CLI Utility Usage
|
||||
|
||||
The `mysql-bkup` CLI provides commands and options to manage MySQL backups efficiently.
|
||||
|
||||
| Option | Short Flag | Description |
|
||||
|-------------------------|------------|-------------------------------------------------------------------------------|
|
||||
| `pg-bkup` | `bkup` | CLI utility for managing PostgreSQL backups. |
|
||||
| `backup` | | Perform a backup operation. |
|
||||
| `restore` | | Perform a restore operation. |
|
||||
| `migrate` | | Migrate a database from one instance to another. |
|
||||
| `--storage` | `-s` | Storage type (`local`, `s3`, `ssh`, etc.). Default: `local`. |
|
||||
| `--file` | `-f` | File name for restoration. |
|
||||
| `--path` | | Path for storage (e.g., `/custom_path` for S3 or `/home/foo/backup` for SSH). |
|
||||
| `--config` | `-c` | Configuration file for multi database backup. (e.g: `/backup/config.yaml`). |
|
||||
| `--dbname` | `-d` | Database name. |
|
||||
| `--port` | `-p` | Database port. Default: `3306`. |
|
||||
| `--disable-compression` | | Disable compression for database backups. |
|
||||
| `--cron-expression` | `-e` | Cron expression for scheduled backups (e.g., `0 0 * * *` or `@daily`). |
|
||||
| `--help` | `-h` | Display help message and exit. |
|
||||
| `--version` | `-V` | Display version information and exit. |
|
||||
|-------------------------|------------|-----------------------------------------------------------------------------------------|
|
||||
| `mysql-bkup` | `bkup` | CLI tool for managing MySQL backups, restoration, and migration. |
|
||||
| `backup` | | Executes a backup operation. |
|
||||
| `restore` | | Restores a database from a backup file. |
|
||||
| `migrate` | | Migrates a database from one instance to another. |
|
||||
| `--storage` | `-s` | Specifies the storage type (`local`, `s3`, `ssh`, etc.). Default: `local`. |
|
||||
| `--file` | `-f` | Defines the backup file name for restoration. |
|
||||
| `--path` | | Sets the storage path (e.g., `/custom_path` for S3 or `/home/foo/backup` for SSH). |
|
||||
| `--config` | `-c` | Provides a configuration file for multi-database backups (e.g., `/backup/config.yaml`). |
|
||||
| `--dbname` | `-d` | Specifies the database name to back up or restore. |
|
||||
| `--port` | `-p` | Defines the database port. Default: `3306`. |
|
||||
| `--disable-compression` | | Disables compression for database backups. |
|
||||
| `--cron-expression` | `-e` | Schedules backups using a cron expression (e.g., `0 0 * * *` or `@daily`). |
|
||||
| `--all-databases` | `-a` | Backs up all databases separately (e.g., `backup --all-databases`). |
|
||||
| `--all-in-one` | `-A` | Backs up all databases in a single file (e.g., `backup --all-databases --single-file`). |
|
||||
| `--custom-name` | `` | Sets custom backup name for one time backup |
|
||||
| `--help` | `-h` | Displays the help message and exits. |
|
||||
| `--version` | `-V` | Shows version information and exits. |
|
||||
|
||||
---
|
||||
|
||||
@@ -40,6 +43,8 @@ Backup, restore, and migration targets, schedules, and retention policies are co
|
||||
| `DB_NAME` | Optional (if provided via `-d` flag) | Database name. |
|
||||
| `DB_USERNAME` | Required | Database username. |
|
||||
| `DB_PASSWORD` | Required | Database password. |
|
||||
| `DB_SSL_CA` | Optional | Database client CA certificate file |
|
||||
| `DB_SSL_MODE` | Optional(`0 or 1`) default: `0` | Database client Enable CA validation |
|
||||
| `AWS_ACCESS_KEY` | Required for S3 storage | AWS S3 Access Key. |
|
||||
| `AWS_SECRET_KEY` | Required for S3 storage | AWS S3 Secret Key. |
|
||||
| `AWS_BUCKET_NAME` | Required for S3 storage | AWS S3 Bucket Name. |
|
||||
|
||||
8
go.mod
8
go.mod
@@ -1,16 +1,16 @@
|
||||
module github.com/jkaninda/mysql-bkup
|
||||
|
||||
go 1.23.2
|
||||
go 1.24.5
|
||||
|
||||
require github.com/spf13/pflag v1.0.6 // indirect
|
||||
require github.com/spf13/pflag v1.0.9 // indirect
|
||||
|
||||
require (
|
||||
github.com/go-mail/mail v2.3.1+incompatible
|
||||
github.com/jkaninda/encryptor v0.0.0-20241111100652-926393c9437e
|
||||
github.com/jkaninda/go-storage v0.1.3
|
||||
github.com/jkaninda/go-utils v0.1.1
|
||||
github.com/jkaninda/go-utils v0.1.3
|
||||
github.com/robfig/cron/v3 v3.0.1
|
||||
github.com/spf13/cobra v1.9.1
|
||||
github.com/spf13/cobra v1.10.1
|
||||
gopkg.in/yaml.v3 v3.0.1
|
||||
)
|
||||
|
||||
|
||||
12
go.sum
12
go.sum
@@ -43,8 +43,8 @@ github.com/jkaninda/encryptor v0.0.0-20241111100652-926393c9437e h1:jtFKZHt/PLGQ
|
||||
github.com/jkaninda/encryptor v0.0.0-20241111100652-926393c9437e/go.mod h1:Y1EXpPWQ9PNd7y7E6ez3xgnzZc8fuDWXwX/1/dXNCE4=
|
||||
github.com/jkaninda/go-storage v0.1.3 h1:lEpHVgFLKSvjsi/6tAek96Y07za3vxmsXF2/+jiCMZU=
|
||||
github.com/jkaninda/go-storage v0.1.3/go.mod h1:zVRnLprBk/9AUz2+za6Y03MgoNYrqKLy3edVtjqMaps=
|
||||
github.com/jkaninda/go-utils v0.1.1 h1:PMrtXR9d51YzHo85y9Z6YVL0YyBURbRTPemHVbFDqZg=
|
||||
github.com/jkaninda/go-utils v0.1.1/go.mod h1:pf0/U6k4JbxlablM2G4eSTZdQ2LFshfAsCK5Q8qNfGo=
|
||||
github.com/jkaninda/go-utils v0.1.3 h1:BhVRHpR6JIGhE7JVgqOgiFketIqmNtvMvYpveVwDqqY=
|
||||
github.com/jkaninda/go-utils v0.1.3/go.mod h1:Aa54jEAcDykc3CnOdreqZG80UfSZOvrYecyusu+oPb4=
|
||||
github.com/jlaffaye/ftp v0.2.0 h1:lXNvW7cBu7R/68bknOX3MrRIIqZ61zELs1P2RAiA3lg=
|
||||
github.com/jlaffaye/ftp v0.2.0/go.mod h1:is2Ds5qkhceAPy2xD6RLI6hmp/qysSoymZ+Z2uTnspI=
|
||||
github.com/jmespath/go-jmespath v0.4.0 h1:BEgLn5cpjn8UN1mAw4NjwDrS35OdebyEtFe+9YPoQUg=
|
||||
@@ -68,10 +68,10 @@ github.com/robfig/cron/v3 v3.0.1/go.mod h1:eQICP3HwyT7UooqI/z+Ov+PtYAWygg1TEWWzG
|
||||
github.com/rogpeppe/go-internal v1.12.0 h1:exVL4IDcn6na9z1rAb56Vxr+CgyK3nn3O+epU5NdKM8=
|
||||
github.com/rogpeppe/go-internal v1.12.0/go.mod h1:E+RYuTGaKKdloAfM02xzb0FW3Paa99yedzYV+kq4uf4=
|
||||
github.com/russross/blackfriday/v2 v2.1.0/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM=
|
||||
github.com/spf13/cobra v1.9.1 h1:CXSaggrXdbHK9CF+8ywj8Amf7PBRmPCOJugH954Nnlo=
|
||||
github.com/spf13/cobra v1.9.1/go.mod h1:nDyEzZ8ogv936Cinf6g1RU9MRY64Ir93oCnqb9wxYW0=
|
||||
github.com/spf13/pflag v1.0.6 h1:jFzHGLGAlb3ruxLB8MhbI6A8+AQX/2eW4qeyNZXNp2o=
|
||||
github.com/spf13/pflag v1.0.6/go.mod h1:McXfInJRrz4CZXVZOBLb0bTZqETkiAhM9Iw0y3An2Bg=
|
||||
github.com/spf13/cobra v1.10.1 h1:lJeBwCfmrnXthfAupyUTzJ/J4Nc1RsHC/mSRU2dll/s=
|
||||
github.com/spf13/cobra v1.10.1/go.mod h1:7SmJGaTHFVBY0jW4NXGluQoLvhqFQM+6XSKD+P4XaB0=
|
||||
github.com/spf13/pflag v1.0.9 h1:9exaQaMOCwffKiiiYk6/BndUBv+iRViNW+4lEMi0PvY=
|
||||
github.com/spf13/pflag v1.0.9/go.mod h1:McXfInJRrz4CZXVZOBLb0bTZqETkiAhM9Iw0y3An2Bg=
|
||||
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
||||
github.com/stretchr/testify v1.9.0 h1:HtqpIVDClZ4nwg75+f6Lvsy/wHu+3BoSGCbBAcpTsTg=
|
||||
github.com/stretchr/testify v1.9.0/go.mod h1:r2ic/lqez/lEtzL7wO/rwa5dbSLXVDPFyf8C91i36aY=
|
||||
|
||||
35
migrations/init.sql
Normal file
35
migrations/init.sql
Normal file
@@ -0,0 +1,35 @@
|
||||
-- Create the database testdb2 and testdb3
|
||||
CREATE DATABASE IF NOT EXISTS testdb2;
|
||||
CREATE DATABASE IF NOT EXISTS testdb3;
|
||||
CREATE DATABASE IF NOT EXISTS fakedb;
|
||||
USE testdb;
|
||||
|
||||
-- Create the 'users' table
|
||||
CREATE TABLE users (
|
||||
id INT AUTO_INCREMENT PRIMARY KEY,
|
||||
name VARCHAR(100) NOT NULL,
|
||||
email VARCHAR(100) NOT NULL UNIQUE,
|
||||
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
-- Create the 'orders' table
|
||||
CREATE TABLE orders (
|
||||
id INT AUTO_INCREMENT PRIMARY KEY,
|
||||
user_id INT NOT NULL,
|
||||
amount DECIMAL(10,2) NOT NULL,
|
||||
status ENUM('pending', 'completed', 'canceled') NOT NULL DEFAULT 'pending',
|
||||
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP,
|
||||
FOREIGN KEY (user_id) REFERENCES users(id) ON DELETE CASCADE
|
||||
);
|
||||
|
||||
-- Insert fake users
|
||||
INSERT INTO users (name, email) VALUES
|
||||
('Alice Smith', 'alice@example.com'),
|
||||
('Bob Johnson', 'bob@example.com'),
|
||||
('Charlie Brown', 'charlie@example.com');
|
||||
|
||||
-- Insert fake orders
|
||||
INSERT INTO orders (user_id, amount, status) VALUES
|
||||
(1, 100.50, 'completed'),
|
||||
(2, 200.75, 'pending'),
|
||||
(3, 50.00, 'canceled');
|
||||
13
migrations/test_config.yaml
Normal file
13
migrations/test_config.yaml
Normal file
@@ -0,0 +1,13 @@
|
||||
#cronExpression: "@every 20s"
|
||||
#backupRescueMode: false
|
||||
databases:
|
||||
- host: 127.0.0.1
|
||||
port: 3306
|
||||
name: testdb
|
||||
user: user
|
||||
password: password
|
||||
- name: testdb2
|
||||
# database credentials from environment variables
|
||||
#TESTDB2_DB_USERNAME
|
||||
#TESTDB2_DB_PASSWORD
|
||||
#TESTDB2_DB_HOST
|
||||
@@ -39,7 +39,7 @@ func azureBackup(db *dbConfig, config *BackupConfig) {
|
||||
utils.Info("Backup database to Azure Blob Storage")
|
||||
|
||||
// Backup database
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all)
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all, config.allInOne)
|
||||
if err != nil {
|
||||
recoverMode(err, "Error backing up database")
|
||||
return
|
||||
@@ -104,7 +104,7 @@ func azureBackup(db *dbConfig, config *BackupConfig) {
|
||||
})
|
||||
// Delete temp
|
||||
deleteTemp()
|
||||
utils.Info("Backup successfully completed in %s", duration)
|
||||
utils.Info("The backup of the %s database has been completed in %s", db.dbName, duration)
|
||||
}
|
||||
func azureRestore(db *dbConfig, conf *RestoreConfig) {
|
||||
utils.Info("Restore database from Azure Blob storage")
|
||||
|
||||
103
pkg/backup.go
103
pkg/backup.go
@@ -26,6 +26,8 @@ SOFTWARE.
|
||||
package pkg
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"errors"
|
||||
"fmt"
|
||||
"github.com/jkaninda/encryptor"
|
||||
"github.com/jkaninda/go-storage/pkg/local"
|
||||
@@ -36,6 +38,7 @@ import (
|
||||
"os"
|
||||
"os/exec"
|
||||
"path/filepath"
|
||||
"strings"
|
||||
"time"
|
||||
)
|
||||
|
||||
@@ -48,7 +51,8 @@ func StartBackup(cmd *cobra.Command) {
|
||||
if err != nil {
|
||||
dbConf = initDbConfig(cmd)
|
||||
if config.cronExpression == "" {
|
||||
BackupTask(dbConf, config)
|
||||
config.allowCustomName = true
|
||||
createBackupTask(dbConf, config)
|
||||
} else {
|
||||
if utils.IsValidCronExpression(config.cronExpression) {
|
||||
scheduledMode(dbConf, config)
|
||||
@@ -82,7 +86,7 @@ func scheduledMode(db *dbConfig, config *BackupConfig) {
|
||||
c := cron.New()
|
||||
|
||||
_, err = c.AddFunc(config.cronExpression, func() {
|
||||
BackupTask(db, config)
|
||||
createBackupTask(db, config)
|
||||
utils.Info("Next backup time is: %v", utils.CronNextTime(config.cronExpression).Format(timeFormat))
|
||||
|
||||
})
|
||||
@@ -104,32 +108,69 @@ func multiBackupTask(databases []Database, bkConfig *BackupConfig) {
|
||||
if db.Path != "" {
|
||||
bkConfig.remotePath = db.Path
|
||||
}
|
||||
BackupTask(getDatabase(db), bkConfig)
|
||||
createBackupTask(getDatabase(db), bkConfig)
|
||||
}
|
||||
}
|
||||
|
||||
// BackupTask backups database
|
||||
func BackupTask(db *dbConfig, config *BackupConfig) {
|
||||
// createBackupTask backup task
|
||||
func createBackupTask(db *dbConfig, config *BackupConfig) {
|
||||
if config.all && !config.allInOne {
|
||||
backupAll(db, config)
|
||||
} else {
|
||||
if db.dbName == "" && !config.all {
|
||||
utils.Fatal("Database name is required, use DB_NAME environment variable or -d flag")
|
||||
}
|
||||
backupTask(db, config)
|
||||
}
|
||||
}
|
||||
|
||||
// backupAll backup all databases
|
||||
func backupAll(db *dbConfig, config *BackupConfig) {
|
||||
databases, err := listDatabases(*db)
|
||||
if err != nil {
|
||||
utils.Fatal("Error listing databases: %s", err)
|
||||
}
|
||||
for _, dbName := range databases {
|
||||
if dbName == "information_schema" || dbName == "performance_schema" || dbName == "mysql" || dbName == "sys" || dbName == "innodb" || dbName == "Database" {
|
||||
continue
|
||||
}
|
||||
db.dbName = dbName
|
||||
config.backupFileName = fmt.Sprintf("%s_%s.sql.gz", dbName, time.Now().Format("20060102_150405"))
|
||||
backupTask(db, config)
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
// backupTask backup task
|
||||
func backupTask(db *dbConfig, config *BackupConfig) {
|
||||
utils.Info("Starting backup task...")
|
||||
startTime = time.Now()
|
||||
prefix := db.dbName
|
||||
if config.all {
|
||||
if config.all && config.allInOne {
|
||||
prefix = "all_databases"
|
||||
}
|
||||
|
||||
// Generate file name
|
||||
backupFileName := fmt.Sprintf("%s_%s.sql.gz", prefix, time.Now().Format("20060102_150405"))
|
||||
if config.disableCompression {
|
||||
backupFileName = fmt.Sprintf("%s_%s.sql", prefix, time.Now().Format("20060102_150405"))
|
||||
}
|
||||
if config.customName != "" && config.allowCustomName && !config.all {
|
||||
backupFileName = fmt.Sprintf("%s.sql.gz", config.customName)
|
||||
if config.disableCompression {
|
||||
backupFileName = fmt.Sprintf("%s.sql", config.customName)
|
||||
}
|
||||
}
|
||||
config.backupFileName = backupFileName
|
||||
switch config.storage {
|
||||
s := strings.ToLower(config.storage)
|
||||
switch s {
|
||||
case "local":
|
||||
localBackup(db, config)
|
||||
case "s3", "S3":
|
||||
case "s3":
|
||||
s3Backup(db, config)
|
||||
case "ssh", "SSH", "remote", "sftp":
|
||||
case "ssh", "remote", "sftp":
|
||||
sshBackup(db, config)
|
||||
case "ftp", "FTP":
|
||||
case "ftp":
|
||||
ftpBackup(db, config)
|
||||
case "azure":
|
||||
azureBackup(db, config)
|
||||
@@ -137,6 +178,8 @@ func BackupTask(db *dbConfig, config *BackupConfig) {
|
||||
localBackup(db, config)
|
||||
}
|
||||
}
|
||||
|
||||
// startMultiBackup start multi backup
|
||||
func startMultiBackup(bkConfig *BackupConfig, configFile string) {
|
||||
utils.Info("Starting Multi backup task...")
|
||||
conf, err := readConf(configFile)
|
||||
@@ -199,7 +242,7 @@ func startMultiBackup(bkConfig *BackupConfig, configFile string) {
|
||||
}
|
||||
|
||||
// BackupDatabase backup database
|
||||
func BackupDatabase(db *dbConfig, backupFileName string, disableCompression, all bool) error {
|
||||
func BackupDatabase(db *dbConfig, backupFileName string, disableCompression, all, singleFile bool) error {
|
||||
storagePath = os.Getenv("STORAGE_PATH")
|
||||
utils.Info("Starting database backup...")
|
||||
|
||||
@@ -208,9 +251,11 @@ func BackupDatabase(db *dbConfig, backupFileName string, disableCompression, all
|
||||
}
|
||||
|
||||
dumpArgs := []string{fmt.Sprintf("--defaults-file=%s", mysqlClientConfig)}
|
||||
if all {
|
||||
if all && singleFile {
|
||||
utils.Info("Backing up all databases...")
|
||||
dumpArgs = append(dumpArgs, "--all-databases", "--single-transaction", "--routines", "--triggers")
|
||||
} else {
|
||||
utils.Info("Backing up %s database...", db.dbName)
|
||||
dumpArgs = append(dumpArgs, db.dbName)
|
||||
}
|
||||
|
||||
@@ -221,6 +266,7 @@ func BackupDatabase(db *dbConfig, backupFileName string, disableCompression, all
|
||||
return runCommandWithCompression("mysqldump", dumpArgs, backupPath)
|
||||
}
|
||||
|
||||
// runCommandAndSaveOutput runs a command and saves the output to a file
|
||||
func runCommandAndSaveOutput(command string, args []string, outputPath string) error {
|
||||
cmd := exec.Command(command, args...)
|
||||
output, err := cmd.Output()
|
||||
@@ -231,6 +277,7 @@ func runCommandAndSaveOutput(command string, args []string, outputPath string) e
|
||||
return os.WriteFile(outputPath, output, 0644)
|
||||
}
|
||||
|
||||
// runCommandWithCompression runs a command and compresses the output
|
||||
func runCommandWithCompression(command string, args []string, outputPath string) error {
|
||||
cmd := exec.Command(command, args...)
|
||||
stdout, err := cmd.StdoutPipe()
|
||||
@@ -265,9 +312,11 @@ func runCommandWithCompression(command string, args []string, outputPath string)
|
||||
utils.Info("Database has been backed up")
|
||||
return nil
|
||||
}
|
||||
|
||||
// localBackup backup database to local storage
|
||||
func localBackup(db *dbConfig, config *BackupConfig) {
|
||||
utils.Info("Backup database to local storage")
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all)
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all, config.allInOne)
|
||||
if err != nil {
|
||||
recoverMode(err, "Error backing up database")
|
||||
return
|
||||
@@ -314,9 +363,10 @@ func localBackup(db *dbConfig, config *BackupConfig) {
|
||||
}
|
||||
// Delete temp
|
||||
deleteTemp()
|
||||
utils.Info("Backup successfully completed in %s", duration)
|
||||
utils.Info("The backup of the %s database has been completed in %s", db.dbName, duration)
|
||||
}
|
||||
|
||||
// encryptBackup encrypt backup
|
||||
func encryptBackup(config *BackupConfig) {
|
||||
backupFile, err := os.ReadFile(filepath.Join(tmpPath, config.backupFileName))
|
||||
outputFile := fmt.Sprintf("%s.%s", filepath.Join(tmpPath, config.backupFileName), gpgExtension)
|
||||
@@ -346,6 +396,31 @@ func encryptBackup(config *BackupConfig) {
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
// listDatabases list all databases
|
||||
func listDatabases(db dbConfig) ([]string, error) {
|
||||
databases := []string{}
|
||||
// Create the mysql client config file
|
||||
if err := createMysqlClientConfigFile(db); err != nil {
|
||||
return databases, errors.New(err.Error())
|
||||
}
|
||||
utils.Info("Listing databases...")
|
||||
// Step 1: List all databases
|
||||
cmd := exec.Command("mariadb", fmt.Sprintf("--defaults-file=%s", mysqlClientConfig), "-e", "SHOW DATABASES;")
|
||||
var out bytes.Buffer
|
||||
cmd.Stdout = &out
|
||||
err := cmd.Run()
|
||||
if err != nil {
|
||||
return databases, fmt.Errorf("failed to list databases: %s", err)
|
||||
}
|
||||
// Step 2: Parse the output
|
||||
for _, _db := range strings.Split(out.String(), "\n") {
|
||||
if _db != "" {
|
||||
databases = append(databases, _db)
|
||||
}
|
||||
}
|
||||
return databases, nil
|
||||
}
|
||||
func recoverMode(err error, msg string) {
|
||||
if err != nil {
|
||||
if backupRescueMode {
|
||||
|
||||
@@ -78,6 +78,9 @@ type BackupConfig struct {
|
||||
storage string
|
||||
cronExpression string
|
||||
all bool
|
||||
allInOne bool
|
||||
customName string
|
||||
allowCustomName bool
|
||||
}
|
||||
type FTPConfig struct {
|
||||
host string
|
||||
@@ -257,7 +260,12 @@ func initBackupConfig(cmd *cobra.Command) *BackupConfig {
|
||||
prune = true
|
||||
}
|
||||
disableCompression, _ = cmd.Flags().GetBool("disable-compression")
|
||||
all, _ := cmd.Flags().GetBool("all")
|
||||
customName, _ := cmd.Flags().GetString("custom-name")
|
||||
all, _ := cmd.Flags().GetBool("all-databases")
|
||||
allInOne, _ := cmd.Flags().GetBool("all-in-one")
|
||||
if allInOne {
|
||||
all = true
|
||||
}
|
||||
_, _ = cmd.Flags().GetString("mode")
|
||||
passphrase := os.Getenv("GPG_PASSPHRASE")
|
||||
_ = utils.GetEnv(cmd, "path", "AWS_S3_PATH")
|
||||
@@ -284,6 +292,8 @@ func initBackupConfig(cmd *cobra.Command) *BackupConfig {
|
||||
config.usingKey = usingKey
|
||||
config.cronExpression = cronExpression
|
||||
config.all = all
|
||||
config.allInOne = allInOne
|
||||
config.customName = customName
|
||||
return &config
|
||||
}
|
||||
|
||||
|
||||
@@ -26,6 +26,7 @@ package pkg
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"errors"
|
||||
"fmt"
|
||||
goutils "github.com/jkaninda/go-utils"
|
||||
"github.com/jkaninda/mysql-bkup/utils"
|
||||
@@ -37,7 +38,7 @@ import (
|
||||
)
|
||||
|
||||
func intro() {
|
||||
fmt.Println("Starting MySQL Backup...")
|
||||
fmt.Println("Starting MYSQL-BKUP...")
|
||||
fmt.Printf("Version: %s\n", utils.Version)
|
||||
fmt.Println("Copyright (c) 2024 Jonas Kaninda")
|
||||
}
|
||||
@@ -70,14 +71,13 @@ func deleteTemp() {
|
||||
func testDatabaseConnection(db *dbConfig) error {
|
||||
// Create the mysql client config file
|
||||
if err := createMysqlClientConfigFile(*db); err != nil {
|
||||
return fmt.Errorf(err.Error())
|
||||
return errors.New(err.Error())
|
||||
}
|
||||
utils.Info("Connecting to %s database ...", db.dbName)
|
||||
// Set database name for notification error
|
||||
utils.DatabaseName = db.dbName
|
||||
|
||||
// Prepare the command to test the database connection
|
||||
//cmd := exec.Command("mariadb", "-h", db.dbHost, "-P", db.dbPort, "-u", db.dbUserName, db.dbName, "-e", "quit")
|
||||
cmd := exec.Command("mariadb", fmt.Sprintf("--defaults-file=%s", mysqlClientConfig), db.dbName, "-e", "quit")
|
||||
// Capture the output
|
||||
var out bytes.Buffer
|
||||
@@ -195,8 +195,8 @@ func createMysqlClientConfigFile(db dbConfig) error {
|
||||
sslMode := goutils.GetStringEnvWithDefault("DB_SSL_MODE", "0")
|
||||
// Create the mysql client config file
|
||||
mysqlClientConfigFile := filepath.Join(tmpPath, "my.cnf")
|
||||
mysqlClientConfig := fmt.Sprintf("[client]\nhost=%s\nport=%s\nuser=%s\npassword=%s\nssl-ca=%s\nssl=%s\n", db.dbHost, db.dbPort, db.dbUserName, db.dbPassword, caCertPath, sslMode)
|
||||
if err := os.WriteFile(mysqlClientConfigFile, []byte(mysqlClientConfig), 0644); err != nil {
|
||||
mysqlCl := fmt.Sprintf("[client]\nhost=%s\nport=%s\nuser=%s\npassword=%s\nssl-ca=%s\nssl=%s\n", db.dbHost, db.dbPort, db.dbUserName, db.dbPassword, caCertPath, sslMode)
|
||||
if err := os.WriteFile(mysqlClientConfigFile, []byte(mysqlCl), 0644); err != nil {
|
||||
return fmt.Errorf("failed to create mysql client config file: %v", err)
|
||||
}
|
||||
return nil
|
||||
|
||||
@@ -51,7 +51,7 @@ func StartMigration(cmd *cobra.Command) {
|
||||
conf := &RestoreConfig{}
|
||||
conf.file = backupFileName
|
||||
// Backup source Database
|
||||
err := BackupDatabase(dbConf, backupFileName, true, false)
|
||||
err := BackupDatabase(dbConf, backupFileName, true, false, false)
|
||||
if err != nil {
|
||||
utils.Fatal("Error backing up database: %s", err)
|
||||
}
|
||||
|
||||
@@ -39,7 +39,7 @@ import (
|
||||
func sshBackup(db *dbConfig, config *BackupConfig) {
|
||||
utils.Info("Backup database to Remote server")
|
||||
// Backup database
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all)
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all, config.allInOne)
|
||||
if err != nil {
|
||||
recoverMode(err, "Error backing up database")
|
||||
return
|
||||
@@ -108,7 +108,7 @@ func sshBackup(db *dbConfig, config *BackupConfig) {
|
||||
})
|
||||
// Delete temp
|
||||
deleteTemp()
|
||||
utils.Info("Backup successfully completed in %s", duration)
|
||||
utils.Info("The backup of the %s database has been completed in %s", db.dbName, duration)
|
||||
|
||||
}
|
||||
func remoteRestore(db *dbConfig, conf *RestoreConfig) {
|
||||
@@ -160,7 +160,7 @@ func ftpBackup(db *dbConfig, config *BackupConfig) {
|
||||
utils.Info("Backup database to the remote FTP server")
|
||||
|
||||
// Backup database
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all)
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all, config.allInOne)
|
||||
if err != nil {
|
||||
recoverMode(err, "Error backing up database")
|
||||
return
|
||||
@@ -224,5 +224,5 @@ func ftpBackup(db *dbConfig, config *BackupConfig) {
|
||||
})
|
||||
// Delete temp
|
||||
deleteTemp()
|
||||
utils.Info("Backup successfully completed in %s", duration)
|
||||
utils.Info("The backup of the %s database has been completed in %s", db.dbName, duration)
|
||||
}
|
||||
|
||||
112
pkg/restore.go
112
pkg/restore.go
@@ -57,11 +57,17 @@ func StartRestore(cmd *cobra.Command) {
|
||||
}
|
||||
func localRestore(dbConf *dbConfig, restoreConf *RestoreConfig) {
|
||||
utils.Info("Restore database from local")
|
||||
basePath := filepath.Dir(restoreConf.file)
|
||||
fileName := filepath.Base(restoreConf.file)
|
||||
restoreConf.file = fileName
|
||||
if basePath == "" || basePath == "." {
|
||||
basePath = storagePath
|
||||
}
|
||||
localStorage := local.NewStorage(local.Config{
|
||||
RemotePath: storagePath,
|
||||
RemotePath: basePath,
|
||||
LocalPath: tmpPath,
|
||||
})
|
||||
err := localStorage.CopyFrom(restoreConf.file)
|
||||
err := localStorage.CopyFrom(fileName)
|
||||
if err != nil {
|
||||
utils.Fatal("Error copying backup file: %s", err)
|
||||
}
|
||||
@@ -69,87 +75,79 @@ func localRestore(dbConf *dbConfig, restoreConf *RestoreConfig) {
|
||||
|
||||
}
|
||||
|
||||
// RestoreDatabase restore database
|
||||
// RestoreDatabase restores the database from a backup file
|
||||
func RestoreDatabase(db *dbConfig, conf *RestoreConfig) {
|
||||
if conf.file == "" {
|
||||
utils.Fatal("Error, file required")
|
||||
}
|
||||
extension := filepath.Ext(filepath.Join(tmpPath, conf.file))
|
||||
rFile, err := os.ReadFile(filepath.Join(tmpPath, conf.file))
|
||||
outputFile := RemoveLastExtension(filepath.Join(tmpPath, conf.file))
|
||||
|
||||
filePath := filepath.Join(tmpPath, conf.file)
|
||||
rFile, err := os.ReadFile(filePath)
|
||||
if err != nil {
|
||||
utils.Fatal("Error reading backup file: %s ", err)
|
||||
utils.Fatal("Error reading backup file: %v", err)
|
||||
}
|
||||
|
||||
extension := filepath.Ext(filePath)
|
||||
outputFile := RemoveLastExtension(filePath)
|
||||
|
||||
if extension == ".gpg" {
|
||||
decryptBackup(conf, rFile, outputFile)
|
||||
}
|
||||
|
||||
restorationFile := filepath.Join(tmpPath, conf.file)
|
||||
if !utils.FileExists(restorationFile) {
|
||||
utils.Fatal("File not found: %s", restorationFile)
|
||||
}
|
||||
|
||||
if err := testDatabaseConnection(db); err != nil {
|
||||
utils.Fatal("Error connecting to the database: %v", err)
|
||||
}
|
||||
|
||||
utils.Info("Restoring database...")
|
||||
restoreDatabaseFile(db, restorationFile)
|
||||
}
|
||||
|
||||
func decryptBackup(conf *RestoreConfig, rFile []byte, outputFile string) {
|
||||
if conf.usingKey {
|
||||
utils.Info("Decrypting backup using private key...")
|
||||
utils.Warn("Backup decryption using a private key is not fully supported")
|
||||
prKey, err := os.ReadFile(conf.privateKey)
|
||||
if err != nil {
|
||||
utils.Fatal("Error reading public key: %s ", err)
|
||||
utils.Fatal("Error reading private key: %v", err)
|
||||
}
|
||||
err = encryptor.DecryptWithPrivateKey(rFile, outputFile, prKey, conf.passphrase)
|
||||
if err != nil {
|
||||
utils.Fatal("error during decrypting backup %v", err)
|
||||
if err := encryptor.DecryptWithPrivateKey(rFile, outputFile, prKey, conf.passphrase); err != nil {
|
||||
utils.Fatal("Error decrypting backup: %v", err)
|
||||
}
|
||||
utils.Info("Decrypting backup using private key...done")
|
||||
} else {
|
||||
if conf.passphrase == "" {
|
||||
utils.Error("Error, passphrase or private key required")
|
||||
utils.Fatal("Your file seems to be a GPG file.\nYou need to provide GPG keys. GPG_PASSPHRASE or GPG_PRIVATE_KEY environment variable is required.")
|
||||
} else {
|
||||
utils.Fatal("Passphrase or private key required for GPG file.")
|
||||
}
|
||||
utils.Info("Decrypting backup using passphrase...")
|
||||
// decryptWithGPG file
|
||||
err := encryptor.Decrypt(rFile, outputFile, conf.passphrase)
|
||||
if err != nil {
|
||||
utils.Fatal("Error decrypting file %s %v", file, err)
|
||||
if err := encryptor.Decrypt(rFile, outputFile, conf.passphrase); err != nil {
|
||||
utils.Fatal("Error decrypting file: %v", err)
|
||||
}
|
||||
utils.Info("Decrypting backup using passphrase...done")
|
||||
// Update file name
|
||||
conf.file = RemoveLastExtension(file)
|
||||
conf.file = RemoveLastExtension(conf.file)
|
||||
}
|
||||
}
|
||||
|
||||
func restoreDatabaseFile(db *dbConfig, restorationFile string) {
|
||||
extension := filepath.Ext(restorationFile)
|
||||
var cmdStr string
|
||||
|
||||
switch extension {
|
||||
case ".gz":
|
||||
cmdStr = fmt.Sprintf("zcat %s | mariadb --defaults-file=%s %s", restorationFile, mysqlClientConfig, db.dbName)
|
||||
case ".sql":
|
||||
cmdStr = fmt.Sprintf("cat %s | mariadb --defaults-file=%s %s", restorationFile, mysqlClientConfig, db.dbName)
|
||||
default:
|
||||
utils.Fatal("Unknown file extension: %s", extension)
|
||||
}
|
||||
|
||||
if utils.FileExists(filepath.Join(tmpPath, conf.file)) {
|
||||
err = testDatabaseConnection(db)
|
||||
cmd := exec.Command("sh", "-c", cmdStr)
|
||||
output, err := cmd.CombinedOutput()
|
||||
if err != nil {
|
||||
utils.Fatal("Error connecting to the database %v", err)
|
||||
utils.Fatal("Error restoring database: %v\nOutput: %s", err, string(output))
|
||||
}
|
||||
utils.Info("Restoring database...")
|
||||
|
||||
extension := filepath.Ext(filepath.Join(tmpPath, conf.file))
|
||||
// Restore from a compressed file / .sql.gz
|
||||
if extension == ".gz" {
|
||||
str := "zcat " + filepath.Join(tmpPath, conf.file) + " | mariadb " + fmt.Sprintf("--defaults-file=%s", mysqlClientConfig) + " " + db.dbName
|
||||
output, err := exec.Command("sh", "-c", str).Output()
|
||||
if err != nil {
|
||||
utils.Fatal("Error, in restoring the database %v output: %v", err, output)
|
||||
}
|
||||
utils.Info("Restoring database... done")
|
||||
utils.Info("Database has been restored")
|
||||
// Delete temp
|
||||
utils.Info("Database has been restored successfully.")
|
||||
deleteTemp()
|
||||
|
||||
} else if extension == ".sql" {
|
||||
// Restore from SQL file
|
||||
str := "cat " + filepath.Join(tmpPath, conf.file) + " | mariadb " + fmt.Sprintf("--defaults-file=%s", mysqlClientConfig) + " " + db.dbName
|
||||
output, err := exec.Command("sh", "-c", str).Output()
|
||||
if err != nil {
|
||||
utils.Fatal("Error, in restoring the database %v output: %v", err, output)
|
||||
}
|
||||
utils.Info("Restoring database... done")
|
||||
utils.Info("Database has been restored")
|
||||
// Delete temp
|
||||
deleteTemp()
|
||||
} else {
|
||||
utils.Fatal("Unknown file extension %s", extension)
|
||||
}
|
||||
|
||||
} else {
|
||||
utils.Fatal("File not found in %s", filepath.Join(tmpPath, conf.file))
|
||||
}
|
||||
}
|
||||
|
||||
@@ -39,7 +39,7 @@ func s3Backup(db *dbConfig, config *BackupConfig) {
|
||||
|
||||
utils.Info("Backup database to s3 storage")
|
||||
// Backup database
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all)
|
||||
err := BackupDatabase(db, config.backupFileName, disableCompression, config.all, config.allInOne)
|
||||
if err != nil {
|
||||
recoverMode(err, "Error backing up database")
|
||||
return
|
||||
@@ -107,7 +107,7 @@ func s3Backup(db *dbConfig, config *BackupConfig) {
|
||||
})
|
||||
// Delete temp
|
||||
deleteTemp()
|
||||
utils.Info("Backup successfully completed in %s", duration)
|
||||
utils.Info("The backup of the %s database has been completed in %s", db.dbName, duration)
|
||||
|
||||
}
|
||||
func s3Restore(db *dbConfig, conf *RestoreConfig) {
|
||||
|
||||
@@ -54,7 +54,6 @@ var dbHVars = []string{
|
||||
"DB_HOST",
|
||||
"DB_PASSWORD",
|
||||
"DB_USERNAME",
|
||||
"DB_NAME",
|
||||
}
|
||||
var tdbRVars = []string{
|
||||
"TARGET_DB_HOST",
|
||||
|
||||
Reference in New Issue
Block a user