From 34e01fa303dff79eed2f6d935a563c6fd6422557 Mon Sep 17 00:00:00 2001 From: Frederik Ring Date: Sun, 22 Aug 2021 18:07:32 +0200 Subject: [PATCH] adapt repo layout to go --- Dockerfile | 6 +- {src => cmd/backup}/main.go | 18 +-- src/entrypoint.sh => entrypoint.sh | 0 src/backup.sh | 181 ----------------------------- 4 files changed, 13 insertions(+), 192 deletions(-) rename {src => cmd/backup}/main.go (96%) rename src/entrypoint.sh => entrypoint.sh (100%) delete mode 100644 src/backup.sh diff --git a/Dockerfile b/Dockerfile index 65090fb..8e6a603 100644 --- a/Dockerfile +++ b/Dockerfile @@ -5,8 +5,8 @@ FROM golang:1.17-alpine as builder WORKDIR /app COPY go.mod go.sum ./ -COPY src/main.go ./src/main.go -RUN go build -o backup src/main.go +COPY cmd/backup/main.go ./cmd/backup/main.go +RUN go build -o backup cmd/backup/main.go FROM alpine:3.14 @@ -16,7 +16,7 @@ RUN apk add --update ca-certificates COPY --from=builder /app/backup /usr/bin/backup -COPY src/entrypoint.sh /root/ +COPY ./entrypoint.sh /root/ RUN chmod +x entrypoint.sh ENTRYPOINT ["/root/entrypoint.sh"] diff --git a/src/main.go b/cmd/backup/main.go similarity index 96% rename from src/main.go rename to cmd/backup/main.go index 148ad3d..f7dc2b4 100644 --- a/src/main.go +++ b/cmd/backup/main.go @@ -1,3 +1,6 @@ +// Copyright 2021 - Offen Authors +// SPDX-License-Identifier: MPL-2.0 + package main import ( @@ -28,7 +31,7 @@ import ( ) func main() { - unlock, err := lock() + unlock, err := lock("/var/dockervolumebackup.lock") if err != nil { panic(err) } @@ -57,10 +60,9 @@ type script struct { archive string } -// lock opens a lock file without releasing it and returns a function that -// can be called once the lock shall be released again. -func lock() (func() error, error) { - lockfile := "/var/dockervolumebackup.lock" +// lock opens a lockfile, keeping it open until the caller invokes the returned +// release func. +func lock(lockfile string) (func() error, error) { lf, err := os.OpenFile(lockfile, os.O_CREATE, os.ModeAppend) if err != nil { return nil, fmt.Errorf("lock: error opening lock file: %w", err) @@ -316,7 +318,7 @@ func (s *script) cleanBackup() error { if err := os.Remove(s.file); err != nil { return fmt.Errorf("cleanBackup: error removing file: %w", err) } - s.logger.Info("Successfully cleaned local backup.") + s.logger.Info("Successfully cleaned up local artifacts.") return nil } @@ -391,7 +393,7 @@ func (s *script) pruneOldBackups() error { } else if len(matches) != 0 && len(matches) == lenCandidates { s.logger.Warnf("The current configuration would delete all %d remote backups. Refusing to do so.", len(matches)) } else { - s.logger.Info("No remote backups were pruned.") + s.logger.Infof("None of %d remote backups were pruned.", lenCandidates) } } @@ -443,7 +445,7 @@ func (s *script) pruneOldBackups() error { } else if len(matches) != 0 && len(matches) == len(candidates) { s.logger.Warnf("The current configuration would delete all %d local backups. Refusing to do so.", len(matches)) } else { - s.logger.Info("No local backups were pruned.") + s.logger.Infof("None of %d local backups were pruned.", len(candidates)) } } return nil diff --git a/src/entrypoint.sh b/entrypoint.sh similarity index 100% rename from src/entrypoint.sh rename to entrypoint.sh diff --git a/src/backup.sh b/src/backup.sh deleted file mode 100644 index 08b600f..0000000 --- a/src/backup.sh +++ /dev/null @@ -1,181 +0,0 @@ -#!/bin/sh - -# Copyright 2021 - Offen Authors -# SPDX-License-Identifier: MPL-2.0 - -# Portions of this file are taken from github.com/futurice/docker-volume-backup -# See NOTICE for information about authors and licensing. - -source env.sh - -function info { - echo -e "\n[INFO] $1\n" -} - -info "Preparing backup" -DOCKER_SOCK="/var/run/docker.sock" - -if [ -S "$DOCKER_SOCK" ]; then - TEMPFILE="$(mktemp)" - docker ps -q \ - --filter "label=docker-volume-backup.stop-during-backup=$BACKUP_STOP_CONTAINER_LABEL" \ - > "$TEMPFILE" - CONTAINERS_TO_STOP="$(cat $TEMPFILE | tr '\n' ' ')" - CONTAINERS_TO_STOP_TOTAL="$(cat $TEMPFILE | wc -l)" - CONTAINERS_TOTAL="$(docker ps -q | wc -l)" - rm "$TEMPFILE" - echo "$CONTAINERS_TOTAL containers running on host in total." - echo "$CONTAINERS_TO_STOP_TOTAL containers marked to be stopped during backup." -else - CONTAINERS_TO_STOP_TOTAL="0" - CONTAINERS_TOTAL="0" - echo "Cannot access \"$DOCKER_SOCK\", won't look for containers to stop." -fi - -if [ "$CONTAINERS_TO_STOP_TOTAL" != "0" ]; then - info "Stopping containers" - docker stop $CONTAINERS_TO_STOP -fi - -info "Creating backup" -BACKUP_FILENAME="$(date +"$BACKUP_FILENAME")" -tar -czvf "$BACKUP_FILENAME" $BACKUP_SOURCES # allow the var to expand, in case we have multiple sources - -if [ ! -z "$GPG_PASSPHRASE" ]; then - info "Encrypting backup" - gpg --symmetric --cipher-algo aes256 --batch --passphrase "$GPG_PASSPHRASE" \ - -o "${BACKUP_FILENAME}.gpg" $BACKUP_FILENAME - rm $BACKUP_FILENAME - BACKUP_FILENAME="${BACKUP_FILENAME}.gpg" -fi - -if [ "$CONTAINERS_TO_STOP_TOTAL" != "0" ]; then - info "Starting containers/services back up" - # The container might be part of a stack when running in swarm mode, so - # its parent service needs to be restarted instead once backup is finished. - SERVICES_REQUIRING_UPDATE="" - for CONTAINER_ID in $CONTAINERS_TO_STOP; do - SWARM_SERVICE_NAME=$( - docker inspect \ - --format "{{ index .Config.Labels \"com.docker.swarm.service.name\" }}" \ - $CONTAINER_ID - ) - if [ -z "$SWARM_SERVICE_NAME" ]; then - echo "Restarting $(docker start $CONTAINER_ID)" - else - echo "Removing $(docker rm $CONTAINER_ID)" - # Multiple containers might belong to the same service, so they will - # be restarted only after all names are known. - SERVICES_REQUIRING_UPDATE="${SERVICES_REQUIRING_UPDATE} ${SWARM_SERVICE_NAME}" - fi - done - - if [ -n "$SERVICES_REQUIRING_UPDATE" ]; then - for SERVICE_NAME in $(echo -n "$SERVICES_REQUIRING_UPDATE" | tr ' ' '\n' | sort -u); do - docker service update --force $SERVICE_NAME - done - fi -fi - -copy_backup () { - mc cp $MC_GLOBAL_OPTIONS "$BACKUP_FILENAME" "$1" -} - -if [ ! -z "$AWS_S3_BUCKET_NAME" ]; then - info "Uploading backup to remote storage" - echo "Will upload to bucket \"$AWS_S3_BUCKET_NAME\"." - copy_backup "backup-target/$AWS_S3_BUCKET_NAME" - echo "Upload finished." -fi - -if [ -d "$BACKUP_ARCHIVE" ]; then - info "Copying backup to local archive" - echo "Will copy to \"$BACKUP_ARCHIVE\"." - copy_backup "$BACKUP_ARCHIVE" - echo "Finished copying." -fi - -if [ -f "$BACKUP_FILENAME" ]; then - info "Cleaning up" - rm -vf "$BACKUP_FILENAME" -fi - -info "Backup finished" -echo "Will wait for next scheduled backup." - -probe_expired () { - local target=$1 - local is_local=$2 - if [ -z "$is_local" ]; then - if [ ! -z "$BACKUP_PRUNING_PREFIX" ]; then - target="${target}/${BACKUP_PRUNING_PREFIX}" - fi - mc rm $MC_GLOBAL_OPTIONS --fake --recursive --force \ - --older-than "${BACKUP_RETENTION_DAYS}d" \ - "$target" - else - find $target -name "${BACKUP_PRUNING_PREFIX:-*}" -type f -mtime "+${BACKUP_RETENTION_DAYS}" - fi -} - -probe_all () { - local target=$1 - local is_local=$2 - if [ -z "$is_local" ]; then - if [ ! -z "$BACKUP_PRUNING_PREFIX" ]; then - target="${target}/${BACKUP_PRUNING_PREFIX}" - fi - mc ls $MC_GLOBAL_OPTIONS "$target" - else - find $target -name "${BACKUP_PRUNING_PREFIX:-*}" -type f - fi -} - -delete () { - local target=$1 - local is_local=$2 - if [ -z "$is_local" ]; then - if [ ! -z "$BACKUP_PRUNING_PREFIX" ]; then - target="${target}/${BACKUP_PRUNING_PREFIX}" - fi - mc rm $MC_GLOBAL_OPTIONS --recursive --force \ - --older-than "${BACKUP_RETENTION_DAYS}d" \ - "$target" - else - find $target -name "${BACKUP_PRUNING_PREFIX:-*}" -type f -mtime "+${BACKUP_RETENTION_DAYS}" -delete - fi -} - -prune () { - local target=$1 - local is_local=$2 - rule_applies_to=$(probe_expired "$target" "$is_local" | wc -l) - if [ "$rule_applies_to" == "0" ]; then - echo "No backups found older than the configured retention period of ${BACKUP_RETENTION_DAYS} days." - echo "Doing nothing." - else - total=$(probe_all "$target" "$is_local" | wc -l) - - if [ "$rule_applies_to" == "$total" ]; then - echo "Using a retention of ${BACKUP_RETENTION_DAYS} days would prune all currently existing backups, will not continue." - echo "If this is what you want, please remove files manually instead of using this script." - else - delete "$target" "$is_local" - echo "Successfully pruned ${rule_applies_to} backups older than ${BACKUP_RETENTION_DAYS} days." - fi - fi -} - -if [ ! -z "$BACKUP_RETENTION_DAYS" ]; then - info "Pruning old backups" - echo "Sleeping ${BACKUP_PRUNING_LEEWAY} before checking eligibility." - sleep "$BACKUP_PRUNING_LEEWAY" - if [ ! -z "$AWS_S3_BUCKET_NAME" ]; then - info "Pruning old backups from remote storage" - prune "backup-target/$AWS_S3_BUCKET_NAME" - fi - if [ -d "$BACKUP_ARCHIVE" ]; then - info "Pruning old backups from local archive" - prune "$BACKUP_ARCHIVE" "local" - fi -fi