Michael Grote
304eda8371
Reviewed-on: #160 Co-authored-by: Michael Grote <michael.grote@posteo.de> Co-committed-by: Michael Grote <michael.grote@posteo.de>
157 lines
5.1 KiB
YAML
157 lines
5.1 KiB
YAML
---
|
|
# pbs_*
|
|
pbs_datastores:
|
|
- name: zfs_backup
|
|
path: /backup/pbs_data
|
|
gc_schedule: "sat 19:00"
|
|
|
|
pbs_prune_jobs:
|
|
- name: standard
|
|
schedule: "sat 18:15"
|
|
store: zfs_backup
|
|
keep_last: 1
|
|
keep_hourly: 3
|
|
keep_daily: 3
|
|
|
|
pbs_permissions:
|
|
- user: user_pve5@pbs
|
|
datastore: zfs_backup
|
|
role: DatastoreBackup
|
|
|
|
pbs_users:
|
|
- name: user_pve5
|
|
password: "{{ lookup('viczem.keepass.keepass', 'pbs_pve_user', 'password') }}"
|
|
realm: pbs
|
|
# rpool ist unverschlüsselt als Boot-Medium
|
|
# entschlüsseln nach Boot mit: sudo zpool import -d /dev/disk/by-id/ -a && sudo zfs mount -a -l
|
|
|
|
## backup
|
|
### sudo zpool create -o ashift=12 -o feature@encryption=enabled -O encryption=on -O keylocation=prompt -O keyformat=passphrase backup /dev/disk/by-id/ata-TOSHIBA_MG09ACA18TE_Z1B0A28LFJDH
|
|
|
|
# mgrote.zfs_manage_datasets
|
|
### mgrote_zfs_extra
|
|
# Variablen für mgrote.zfs_health/trim/scrub/zed/arc_mem/ sind zusammengefasst unter zfs_extra_*
|
|
zfs_datasets: # DatenPools werden hier nicht verwaltet
|
|
# rpool - System-Datasets
|
|
- dataset: rpool
|
|
state: present
|
|
compression: zstd
|
|
sync: disabled
|
|
xattr: sa
|
|
dnodesize: auto
|
|
atime: on # noqa yaml[truthy]
|
|
snapdir: hidden
|
|
reservation: 1G
|
|
refreservation: 10G
|
|
acltype: posix
|
|
- dataset: rpool/ROOT
|
|
state: present
|
|
refreservation: 10G
|
|
- dataset: rpool/ROOT/pbs-1
|
|
state: present
|
|
refreservation: 10G
|
|
acltype: posix # https://docs.ansible.com/ansible-core/2.14/playbook_guide/playbooks_privilege_escalation.html#risks-of-becoming-an-unprivileged-user ; sonst kann die dotfiles-Rolle kein setfacl machen
|
|
# backup-pool
|
|
- dataset: backup/pbs_data
|
|
state: present
|
|
quota: 1TB
|
|
- dataset: backup/pve5
|
|
state: present
|
|
canmount: off # noqa yaml[truthy]
|
|
# Variablen für mgrote.zfs_health/trim/scrub/zed/arc_mem/ sind zusammengefasst unter zfs_extra_*
|
|
zfs_extra_arc_max_size: "4294967296" # 4GB in Bytes
|
|
zfs_extra_zfs_pools:
|
|
- name: "rpool"
|
|
systemd_timer_schedule: "*-01,04,07,10-01 23:00" # jeden ersten eines jeden Quartals
|
|
- name: "backup"
|
|
systemd_timer_schedule: "*-01,04,07,10-01 23:00"
|
|
|
|
### mgrote_zfs_sanoid
|
|
sanoid_snaps_enable: true
|
|
|
|
## syncoid
|
|
sanoid_syncoid_destination_host: true
|
|
sanoid_syncoid_ssh_privkey: "{{ lookup('viczem.keepass.keepass', 'sanoid_syncoid_private_key', 'notes') }}"
|
|
sanoid_syncoid_timer: '*-*-* *:00:00' # jede Stunde
|
|
sanoid_syncoid_bwlimit: 50M # 30MB/s
|
|
sanoid_syncoid_datasets_sync:
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/pve_backup
|
|
source_dataset: hdd_data/pve_backup
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/videos
|
|
source_dataset: hdd_data/videos
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/music
|
|
source_dataset: hdd_data/music
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/tmp
|
|
source_dataset: hdd_data/tmp
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/archiv
|
|
source_dataset: hdd_data/archiv
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/bilder
|
|
source_dataset: hdd_data/bilder
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/scans
|
|
source_dataset: hdd_data/scans
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/restic
|
|
source_dataset: hdd_data/restic
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/backup
|
|
source_dataset: hdd_data/backup
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/buecher
|
|
source_dataset: hdd_data/buecher
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/programme
|
|
source_dataset: hdd_data/programme
|
|
|
|
- source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann
|
|
destination_mount_check: backup
|
|
destination_dataset: backup/pve5/vm
|
|
source_dataset: hdd_data/vm
|
|
|
|
# sanoid
|
|
sanoid_datasets:
|
|
### rpool
|
|
- path: rpool
|
|
recursive: 'no'
|
|
snapshots: true
|
|
template: 'pve3tage'
|
|
- path: rpool/ROOT
|
|
recursive: 'no'
|
|
snapshots: true
|
|
template: 'pve3tage'
|
|
- path: rpool/ROOT/pbs-1
|
|
recursive: 'no'
|
|
snapshots: true
|
|
template: 'pve3tage'
|
|
### backup
|
|
- path: backup/pbs_data
|
|
recursive: 'no'
|
|
snapshots: true
|
|
template: '3tage'
|