mirror of
https://github.com/offen/docker-volume-backup.git
synced 2024-12-23 17:30:20 +01:00
192 lines
5.4 KiB
Go
192 lines
5.4 KiB
Go
// Copyright 2022 - offen.software <hioffen@posteo.de>
|
|
// SPDX-License-Identifier: MPL-2.0
|
|
|
|
package azure
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"os"
|
|
"path/filepath"
|
|
"strings"
|
|
"sync"
|
|
"text/template"
|
|
"time"
|
|
|
|
"github.com/Azure/azure-sdk-for-go/sdk/azidentity"
|
|
"github.com/Azure/azure-sdk-for-go/sdk/storage/azblob"
|
|
"github.com/Azure/azure-sdk-for-go/sdk/storage/azblob/blob"
|
|
"github.com/Azure/azure-sdk-for-go/sdk/storage/azblob/blockblob"
|
|
"github.com/Azure/azure-sdk-for-go/sdk/storage/azblob/container"
|
|
"github.com/offen/docker-volume-backup/internal/errwrap"
|
|
"github.com/offen/docker-volume-backup/internal/storage"
|
|
)
|
|
|
|
type azureBlobStorage struct {
|
|
*storage.StorageBackend
|
|
client *azblob.Client
|
|
uploadStreamOptions *blockblob.UploadStreamOptions
|
|
containerName string
|
|
}
|
|
|
|
// Config contains values that define the configuration of an Azure Blob Storage.
|
|
type Config struct {
|
|
AccountName string
|
|
ContainerName string
|
|
PrimaryAccountKey string
|
|
ConnectionString string
|
|
Endpoint string
|
|
RemotePath string
|
|
AccessTier string
|
|
}
|
|
|
|
// NewStorageBackend creates and initializes a new Azure Blob Storage backend.
|
|
func NewStorageBackend(opts Config, logFunc storage.Log) (storage.Backend, error) {
|
|
if opts.PrimaryAccountKey != "" && opts.ConnectionString != "" {
|
|
return nil, errwrap.Wrap(nil, "using primary account key and connection string are mutually exclusive")
|
|
}
|
|
|
|
endpointTemplate, err := template.New("endpoint").Parse(opts.Endpoint)
|
|
if err != nil {
|
|
return nil, errwrap.Wrap(err, "error parsing endpoint template")
|
|
}
|
|
var ep bytes.Buffer
|
|
if err := endpointTemplate.Execute(&ep, opts); err != nil {
|
|
return nil, errwrap.Wrap(err, "error executing endpoint template")
|
|
}
|
|
normalizedEndpoint := fmt.Sprintf("%s/", strings.TrimSuffix(ep.String(), "/"))
|
|
|
|
var client *azblob.Client
|
|
if opts.PrimaryAccountKey != "" {
|
|
cred, err := azblob.NewSharedKeyCredential(opts.AccountName, opts.PrimaryAccountKey)
|
|
if err != nil {
|
|
return nil, errwrap.Wrap(err, "error creating shared key Azure credential")
|
|
}
|
|
|
|
client, err = azblob.NewClientWithSharedKeyCredential(normalizedEndpoint, cred, nil)
|
|
if err != nil {
|
|
return nil, errwrap.Wrap(err, "error creating azure client from primary account key")
|
|
}
|
|
} else if opts.ConnectionString != "" {
|
|
client, err = azblob.NewClientFromConnectionString(opts.ConnectionString, nil)
|
|
if err != nil {
|
|
return nil, errwrap.Wrap(err, "error creating azure client from connection string")
|
|
}
|
|
} else {
|
|
cred, err := azidentity.NewManagedIdentityCredential(nil)
|
|
if err != nil {
|
|
return nil, errwrap.Wrap(err, "error creating managed identity credential")
|
|
}
|
|
client, err = azblob.NewClient(normalizedEndpoint, cred, nil)
|
|
if err != nil {
|
|
return nil, errwrap.Wrap(err, "error creating azure client from managed identity")
|
|
}
|
|
}
|
|
|
|
var uploadStreamOptions *blockblob.UploadStreamOptions
|
|
if opts.AccessTier != "" {
|
|
var found bool
|
|
for _, t := range blob.PossibleAccessTierValues() {
|
|
if string(t) == opts.AccessTier {
|
|
found = true
|
|
uploadStreamOptions = &blockblob.UploadStreamOptions{
|
|
AccessTier: &t,
|
|
}
|
|
}
|
|
}
|
|
if !found {
|
|
return nil, errwrap.Wrap(nil, fmt.Sprintf("%s is not a possible access tier value", opts.AccessTier))
|
|
}
|
|
}
|
|
|
|
storage := azureBlobStorage{
|
|
client: client,
|
|
uploadStreamOptions: uploadStreamOptions,
|
|
containerName: opts.ContainerName,
|
|
StorageBackend: &storage.StorageBackend{
|
|
DestinationPath: opts.RemotePath,
|
|
Log: logFunc,
|
|
},
|
|
}
|
|
return &storage, nil
|
|
}
|
|
|
|
// Name returns the name of the storage backend
|
|
func (b *azureBlobStorage) Name() string {
|
|
return "Azure"
|
|
}
|
|
|
|
// Copy copies the given file to the storage backend.
|
|
func (b *azureBlobStorage) Copy(file string) error {
|
|
fileReader, err := os.Open(file)
|
|
if err != nil {
|
|
return errwrap.Wrap(err, fmt.Sprintf("error opening file %s", file))
|
|
}
|
|
|
|
_, err = b.client.UploadStream(
|
|
context.Background(),
|
|
b.containerName,
|
|
filepath.Join(b.DestinationPath, filepath.Base(file)),
|
|
fileReader,
|
|
b.uploadStreamOptions,
|
|
)
|
|
if err != nil {
|
|
return errwrap.Wrap(err, fmt.Sprintf("error uploading file %s", file))
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Prune rotates away backups according to the configuration and provided
|
|
// deadline for the Azure Blob storage backend.
|
|
func (b *azureBlobStorage) Prune(deadline time.Time, pruningPrefix string) (*storage.PruneStats, error) {
|
|
lookupPrefix := filepath.Join(b.DestinationPath, pruningPrefix)
|
|
pager := b.client.NewListBlobsFlatPager(b.containerName, &container.ListBlobsFlatOptions{
|
|
Prefix: &lookupPrefix,
|
|
})
|
|
var matches []string
|
|
var totalCount uint
|
|
for pager.More() {
|
|
resp, err := pager.NextPage(context.Background())
|
|
if err != nil {
|
|
return nil, errwrap.Wrap(err, "error paging over blobs")
|
|
}
|
|
for _, v := range resp.Segment.BlobItems {
|
|
totalCount++
|
|
if v.Properties.LastModified.Before(deadline) {
|
|
matches = append(matches, *v.Name)
|
|
}
|
|
}
|
|
}
|
|
|
|
stats := &storage.PruneStats{
|
|
Total: totalCount,
|
|
Pruned: uint(len(matches)),
|
|
}
|
|
|
|
pruneErr := b.DoPrune(b.Name(), len(matches), int(totalCount), deadline, func() error {
|
|
wg := sync.WaitGroup{}
|
|
wg.Add(len(matches))
|
|
var errs []error
|
|
|
|
for _, match := range matches {
|
|
name := match
|
|
go func() {
|
|
_, err := b.client.DeleteBlob(context.Background(), b.containerName, name, nil)
|
|
if err != nil {
|
|
errs = append(errs, err)
|
|
}
|
|
wg.Done()
|
|
}()
|
|
}
|
|
wg.Wait()
|
|
if len(errs) != 0 {
|
|
return errors.Join(errs...)
|
|
}
|
|
return nil
|
|
})
|
|
|
|
return stats, pruneErr
|
|
}
|