From 7fba03c73a2137f0543be01fd1c569d78bec6fa1 Mon Sep 17 00:00:00 2001 From: mg Date: Mon, 17 Jul 2023 22:20:44 +0200 Subject: [PATCH] neuer Backup-Server (Umzug) (#552) Co-authored-by: Michael Grote Reviewed-on: https://git.mgrote.net/mg/homeserver/pulls/552 --- docker-compose/homer/assets/mgmt.yml | 4 +-- docker-compose/munin/docker-compose.yml.j2 | 2 +- host_vars/pbs.grote.lan.yml | 33 +++++++++++-------- host_vars/pve5.grote.lan.yml | 2 +- inventory | 1 - .../templates/prune-jobs.j2 | 1 - 6 files changed, 24 insertions(+), 19 deletions(-) diff --git a/docker-compose/homer/assets/mgmt.yml b/docker-compose/homer/assets/mgmt.yml index 5ff56524..532d516e 100644 --- a/docker-compose/homer/assets/mgmt.yml +++ b/docker-compose/homer/assets/mgmt.yml @@ -126,11 +126,11 @@ services: target: "_blank" - name: "Proxmox Backup Server" logo: "assets/icons/pve.png" - url: "https://192.168.2.239:8007" + url: "https://pbs.grote.lan:8007" target: "_blank" subtitle: "M93 Tiny" - name: "Proxmox Backup Server - Test" logo: "assets/icons/pve.png" - url: "https://192.168.2.18:8007" + url: "https://pbs-test.grote.lan:8007" target: "_blank" subtitle: "Test" diff --git a/docker-compose/munin/docker-compose.yml.j2 b/docker-compose/munin/docker-compose.yml.j2 index aff2b9ea..a9232f77 100644 --- a/docker-compose/munin/docker-compose.yml.j2 +++ b/docker-compose/munin/docker-compose.yml.j2 @@ -22,7 +22,7 @@ services: dokuwiki2.grote.lan:dokuwiki2.grote.lan gitea.grote.lan:gitea.grote.lan docker10.grote.lan:docker10.grote.lan - pbs.grote.lan:192.168.2.239 + pbs.grote.lan:pbs.grote.lan # z.B. # computer-test.grote.lan.test:192.68.2.4 # computer.grote.lan:computer.grote.lan diff --git a/host_vars/pbs.grote.lan.yml b/host_vars/pbs.grote.lan.yml index 09fb2107..23cbb8b2 100644 --- a/host_vars/pbs.grote.lan.yml +++ b/host_vars/pbs.grote.lan.yml @@ -73,67 +73,67 @@ sanoid_syncoid_timer: '*-*-* *:00:00' # jede Stunde sanoid_syncoid_bwlimit: 50M # 30MB/s sanoid_syncoid_datasets_sync: - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/pve_backup source_dataset: hdd_data_raidz/pve_backup - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/videos source_dataset: hdd_data_raidz/videos - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/music source_dataset: hdd_data_raidz/music - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/tmp source_dataset: hdd_data_raidz/tmp - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/archiv source_dataset: hdd_data_raidz/archiv - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/bilder source_dataset: hdd_data_raidz/bilder - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/hm source_dataset: hdd_data_raidz/hm - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/scans source_dataset: hdd_data_raidz/scans - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/restic source_dataset: hdd_data_raidz/restic - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/backup source_dataset: hdd_data_raidz/backup - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/buecher source_dataset: hdd_data_raidz/buecher - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/programme source_dataset: hdd_data_raidz/programme - - source_host: pve5.grote.lan + - source_host: 192.168.2.16 # pve5, weil pbs den fqdn nicht auflösen kann destination_mount_check: backup destination_dataset: backup/pve5/vm source_dataset: hdd_data_raidz/vm @@ -159,3 +159,10 @@ recursive: 'no' snapshots: true template: '3tage' + + ### mgrote.apt_manage_packages + apt_packages_internet: + - http://192.168.2.43:3344/bash-helper-scripts-mgrote-latest.deb # weil pbs den fqdn nicht auflösen kann + + ### mgrote.zfs_sanoid + sanoid_deb_url: http://192.168.2.43:3344/sanoid_3.0.0.deb # weil pbs den fqdn nicht auflösen kann diff --git a/host_vars/pve5.grote.lan.yml b/host_vars/pve5.grote.lan.yml index 1ccb1e70..88043986 100644 --- a/host_vars/pve5.grote.lan.yml +++ b/host_vars/pve5.grote.lan.yml @@ -236,7 +236,7 @@ # mgrote.pbs_pve_integration pve_pbs_datastore: - name: pbs - server: 192.168.2.239 + server: pbs.grote.lan datastore: zfs_backup username: user_pve5@pbs password: "{{ lookup('keepass', 'pbs_pve_user', 'password') }}" diff --git a/inventory b/inventory index 3dc76b78..54b1c07d 100644 --- a/inventory +++ b/inventory @@ -32,7 +32,6 @@ all: hosts: pve5.grote.lan: pbs.grote.lan: - ansible_host: 192.168.2.239 gitea: hosts: gitea.grote.lan: diff --git a/roles/mgrote.pbs_datastores/templates/prune-jobs.j2 b/roles/mgrote.pbs_datastores/templates/prune-jobs.j2 index e8cbc95f..35dbe087 100644 --- a/roles/mgrote.pbs_datastores/templates/prune-jobs.j2 +++ b/roles/mgrote.pbs_datastores/templates/prune-jobs.j2 @@ -5,7 +5,6 @@ prune: {{ item.name }} keep-daily {{ item.keep_daily }} keep-hourly {{ item.keep_hourly }} keep-last {{ item.keep_last }} - ns schedule {{ item.schedule }} store {{ item.store }}