--- ### mgrote.zfs_manage_datasets # "Daten"-Datasets verwalte ich händisch! # zfs create -o encryption=aes-256-gcm -o keyformat=passphrase hdd_data_raidz/data_crypt # zfs create hdd_data_raidz/videos # Pools müssen vorher erstellt sein! # rpool # zpool create -f -o ashift=12 rpool mirror /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi8 /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi7 # HDD_DATA_RAIDZ # zpool create -f -o ashift=12 hdd_data_raidz raidz /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi9 /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi3 /dev/disk/by-id/scsi-0QEMU_QEMU_HARDDISK_drive-scsi4 zfs_datasets: # DatenPools werden hier nicht verwaltet - dataset: rpool/vm state: present compression: lz4 sync: disabled xattr: sa dnodesize: auto atime: on snapdir: hidden - dataset: rpool/vm/dir state: present compression: lz4 sync: disabled xattr: sa dnodesize: auto atime: on snapdir: hidden - dataset: rpool/vm/zvol state: present compression: lz4 sync: disabled xattr: sa dnodesize: auto atime: on snapdir: hidden - dataset: rpool/vm/qcow state: present compression: lz4 sync: disabled xattr: sa dnodesize: auto atime: on snapdir: hidden ### mgrote.apcupsd apcupsd_nis_master: true apcupsd_nis_master_hostname: pve2-test.grote.lan ### mgrote.zfs_extra # Variablen für mgrote.zfs_health/trim/scrub/zed/arc_mem/ sind zusammengefasst unter zfs_extra_* zfs_extra_arc_max_size: "2147483648" # 2GB in Bytes zfs_extra_zfs_pools: - name: "ssd_vm_mirror" type: "ssd" cron_minute_zfs_trim: "5" cron_hour_zfs_trim: "22" cron_month_zfs_trim: "4,8,12" cron_day_zfs_trim: "2" cron_weekday_zfs_scrub: "6" cron_minutes_zfs_scrub: "0" cron_hour_zfs_scrub: "23" - name: "hdd_vm_mirror" type: "hdd" cron_minute_zfs_trim: "5" cron_hour_zfs_trim: "23" cron_month_zfs_trim: "4,8,12" cron_day_zfs_trim: "2" cron_weekday_zfs_scrub: "6" cron_minutes_zfs_scrub: "0" cron_hour_zfs_scrub: "22"