bacula/docker/etc/samples/bacula-dir-cloud-aws.conf.sample
2023-08-23 07:33:42 -03:00

73 lines
1.5 KiB
Plaintext
Executable File

# bacula-dir-cloud.conf
#
# JobDefs
# Job
# Restore
# Pool
# Autochanger
#
# Template to store in cloud
JobDefs {
Name = "DefaultJobToCloudAWS"
Type = Backup
Level = Incremental
Client = bacula-fd
FileSet = "Full Set"
Schedule = "WeeklyCycle"
Storage = "CloudS3AWS"
Messages = Standard
Pool = CloudAWS
SpoolAttributes = yes
Priority = 10
Write Bootstrap = "/opt/bacula/working/%c.bsr"
}
# Jobs
Job {
Name = "BackupClient1ToCloudAWS"
JobDefs = "DefaultJobToCloudAWS"
}
# Restore
Job {
Name = "RestoreFromCloudAWS"
Type = Restore
Client=bacula-fd
Storage = CloudS3AWS
FileSet="Full Set"
Pool = CloudAWS
Messages = Standard
Where = /tmp/bacula-restores
}
# Cloud Pool definition
Pool {
Name = CloudAWS
Pool Type = Backup
Recycle = no # Bacula can automatically recycle Volumes
AutoPrune = yes # Prune expired volumes
Volume Retention = 365 days # one year
Maximum Volume Jobs = 1 #
# Maximum Volume Bytes = 100M # Limit Volume size to something reasonable
Label Format = "Vol-JobId-${JobId}" # Auto label
}
# Autochanger definition
Autochanger {
Name = "CloudS3AWS"
# Do not use "localhost" here
Address = bacula-sd # N.B. Use a fully qualified name here
SDPort = 9103
Password = "TS8EQJ99iLFSK39oJy33YqkZ98v4ZapjRcA+j1N6ED1n"
Device = "CloudAutoChangerS3"
Media Type = "CloudType"
Maximum Concurrent Jobs = 10 # run up to 10 jobs a the same time
}