I’ve started encountering a problem that I should use some assistance troubleshooting. I’ve got a Proxmox system that hosts, primarily, my Opnsense router. I’ve had this specific setup for about a year.

Recently, I’ve been experiencing sluggishness and noticed that the IO wait is through the roof. Rebooting the Opnsense VM, which normally only takes a few minutes is now taking upwards of 15-20. The entire time my IO wait sits between 50-80%.

The system has 1 disk in it that is formatted ZFS. I’ve checked dmesg, and the syslog for indications of disk errors (this feels like a failing disk) and found none. I also checked the smart statistics and they all “PASSED”.

Any pointers would be appreciated.

Example of my most recent host reboot.

Edit: I believe I’ve found the root cause of the change in performance and it was a bit of shooting myself in the foot. I’ve been experimenting with different tools for log collection and the most recent one is a SIEM tool called Wazuh. I didn’t realize that upon reboot it runs an integrity check that generates a ton of disk I/O. So when I rebooted this proxmox server, that integrity check was running on proxmox, my pihole, and (I think) opnsense concurrently. All against a single consumer grade HDD.

Thanks to everyone who responded. I really appreciate all the performance tuning guidance. I’ve also made the following changes:

  1. Added a 2nd drive (I have several of these lying around, don’t ask) converting the zfs pool into a mirror. This gives me both redundancy and should improve read performance.
  2. Configured a 2nd storage target on the same zpool with compression enabled and a 64k block size in proxmox. I then migrated the 2 VMs to that storage.
  3. Since I’m collecting logs in Wazuh I set Opnsense to use ram disks for /tmp and /var/log.

Rebooted Opensense and it was back up in 1:42 min.

    • Pyrosis@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 months ago

      Upgrading a ZFS pool itself shouldn’t make a system unbootable even if an rpool (root pool) exists on it.

      That could only happen if the upgrade took a shit during a power outage or something like that. The upgrade itself usually only takes a few seconds from the command line.

      If it makes you feel better I upgraded mine with an rpool on it and it was painless. I do have a everything backed up tho so I rarely worry. However ai understand being hesitant.

      • 𝓢𝓮𝓮𝓙𝓪𝔂𝓔𝓶𝓶OP
        link
        fedilink
        English
        arrow-up
        1
        ·
        7 months ago

        I’m referring to this.

        … using grub to directly boot from ZFS - such setups are in general not safe to run zpool upgrade on!

        $ sudo proxmox-boot-tool status
        Re-executing '/usr/sbin/proxmox-boot-tool' in new private mount namespace..
        System currently booted with legacy bios
        8357-FBD5 is configured with: grub (versions: 6.5.11-7-pve, 6.5.13-5-pve, 6.8.4-2-pve)
        

        Unless I’m misunderstanding the guidance.

        • Pyrosis@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          7 months ago

          It looks like you are using legacy bios. mine is using uefi with a zfs rpool

          proxmox-boot-tool status
          Re-executing '/usr/sbin/proxmox-boot-tool' in new private mount namespace..
          System currently booted with uefi
          31FA-87E2 is configured with: uefi (versions: 6.5.11-8-pve, 6.5.13-5-pve)
          

          However, like with everything a method always exists to get it done. Or not if you are concerned.

          If you are interested it would look like…

          Pool Upgrade

          sudo zpool upgrade <pool_name>
          

          Confirm Upgrade

          sudo zpool status
          
          

          Refresh boot config

          sudo pveboot-tool refresh
          
          

          Confirm Boot configuration

          cat /boot/grub/grub.cfg
          

          You are looking for directives like this to see if they are indeed pointing at your existing rpool

          root=ZFS=rpool/ROOT/pve-1 boot=zfs quiet
          

          here is my file if it helps you compare…

          #
          # DO NOT EDIT THIS FILE
          #
          # It is automatically generated by grub-mkconfig using templates
          # from /etc/grub.d and settings from /etc/default/grub
          #
          
          ### BEGIN /etc/grub.d/000_proxmox_boot_header ###
          #
          # This system is booted via proxmox-boot-tool! The grub-config used when
          # booting from the disks configured with proxmox-boot-tool resides on the vfat
          # partitions with UUIDs listed in /etc/kernel/proxmox-boot-uuids.
          # /boot/grub/grub.cfg is NOT read when booting from those disk!
          ### END /etc/grub.d/000_proxmox_boot_header ###
          
          ### BEGIN /etc/grub.d/00_header ###
          if [ -s $prefix/grubenv ]; then
            set have_grubenv=true
            load_env
          fi
          if [ "${next_entry}" ] ; then
             set default="${next_entry}"
             set next_entry=
             save_env next_entry
             set boot_once=true
          else
             set default="0"
          fi
          
          if [ x"${feature_menuentry_id}" = xy ]; then
            menuentry_id_option="--id"
          else
            menuentry_id_option=""
          fi
          
          export menuentry_id_option
          
          if [ "${prev_saved_entry}" ]; then
            set saved_entry="${prev_saved_entry}"
            save_env saved_entry
            set prev_saved_entry=
            save_env prev_saved_entry
            set boot_once=true
          fi
          
          function savedefault {
            if [ -z "${boot_once}" ]; then
              saved_entry="${chosen}"
              save_env saved_entry
            fi
          }
          function load_video {
            if [ x$feature_all_video_module = xy ]; then
              insmod all_video
            else
              insmod efi_gop
              insmod efi_uga
              insmod ieee1275_fb
              insmod vbe
              insmod vga
              insmod video_bochs
              insmod video_cirrus
            fi
          }
          
          if loadfont unicode ; then
            set gfxmode=auto
            load_video
            insmod gfxterm
            set locale_dir=$prefix/locale
            set lang=en_US
            insmod gettext
          fi
          terminal_output gfxterm
          if [ "${recordfail}" = 1 ] ; then
            set timeout=30
          else
            if [ x$feature_timeout_style = xy ] ; then
              set timeout_style=menu
              set timeout=5
            # Fallback normal timeout code in case the timeout_style feature is
            # unavailable.
            else
              set timeout=5
            fi
          fi
          ### END /etc/grub.d/00_header ###
          
          ### BEGIN /etc/grub.d/05_debian_theme ###
          set menu_color_normal=cyan/blue
          set menu_color_highlight=white/blue
          ### END /etc/grub.d/05_debian_theme ###
          
          ### BEGIN /etc/grub.d/10_linux ###
          function gfxmode {
                  set gfxpayload="${1}"
          }
          set linux_gfx_mode=
          export linux_gfx_mode
          menuentry 'Proxmox VE GNU/Linux' --class proxmox --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-simple-/dev/sdc3' {
                  load_video
                  insmod gzio
                  if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
                  insmod part_gpt
                  echo    'Loading Linux 6.5.13-5-pve ...'
                  linux   /ROOT/pve-1@/boot/vmlinuz-6.5.13-5-pve root=ZFS=/ROOT/pve-1 ro       root=ZFS=rpool/ROOT/pve-1 boot=zfs quiet
                  echo    'Loading initial ramdisk ...'
                  initrd  /ROOT/pve-1@/boot/initrd.img-6.5.13-5-pve
          }
          submenu 'Advanced options for Proxmox VE GNU/Linux' $menuentry_id_option 'gnulinux-advanced-/dev/sdc3' {
                  menuentry 'Proxmox VE GNU/Linux, with Linux 6.5.13-5-pve' --class proxmox --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-6.5.13-5-pve-advanced-/dev/sdc3' {
                          load_video
                          insmod gzio
                          if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
                          insmod part_gpt
                          echo    'Loading Linux 6.5.13-5-pve ...'
                          linux   /ROOT/pve-1@/boot/vmlinuz-6.5.13-5-pve root=ZFS=/ROOT/pve-1 ro       root=ZFS=rpool/ROOT/pve-1 boot=zfs quiet
                          echo    'Loading initial ramdisk ...'
                          initrd  /ROOT/pve-1@/boot/initrd.img-6.5.13-5-pve
                  }
                  menuentry 'Proxmox VE GNU/Linux, with Linux 6.5.13-5-pve (recovery mode)' --class proxmox --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-6.5.13-5-pve-recovery-/dev/sdc3' {
                          load_video
                          insmod gzio
                          if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
                          insmod part_gpt
                          echo    'Loading Linux 6.5.13-5-pve ...'
                          linux   /ROOT/pve-1@/boot/vmlinuz-6.5.13-5-pve root=ZFS=/ROOT/pve-1 ro single       root=ZFS=rpool/ROOT/pve-1 boot=zfs
                          echo    'Loading initial ramdisk ...'
                          initrd  /ROOT/pve-1@/boot/initrd.img-6.5.13-5-pve
                  }
                  menuentry 'Proxmox VE GNU/Linux, with Linux 6.5.11-8-pve' --class proxmox --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-6.5.11-8-pve-advanced-/dev/sdc3' {
                          load_video
                          insmod gzio
                          if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
                          insmod part_gpt
                          echo    'Loading Linux 6.5.11-8-pve ...'
                          linux   /ROOT/pve-1@/boot/vmlinuz-6.5.11-8-pve root=ZFS=/ROOT/pve-1 ro       root=ZFS=rpool/ROOT/pve-1 boot=zfs quiet
                          echo    'Loading initial ramdisk ...'
                          initrd  /ROOT/pve-1@/boot/initrd.img-6.5.11-8-pve
                  }
                  menuentry 'Proxmox VE GNU/Linux, with Linux 6.5.11-8-pve (recovery mode)' --class proxmox --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-6.5.11-8-pve-recovery-/dev/sdc3' {
                          load_video
                          insmod gzio
                          if [ x$grub_platform = xxen ]; then insmod xzio; insmod lzopio; fi
                          insmod part_gpt
                          echo    'Loading Linux 6.5.11-8-pve ...'
                          linux   /ROOT/pve-1@/boot/vmlinuz-6.5.11-8-pve root=ZFS=/ROOT/pve-1 ro single       root=ZFS=rpool/ROOT/pve-1 boot=zfs
                          echo    'Loading initial ramdisk ...'
                          initrd  /ROOT/pve-1@/boot/initrd.img-6.5.11-8-pve
                  }
          }
          
          ### END /etc/grub.d/10_linux ###
          
          ### BEGIN /etc/grub.d/20_linux_xen ###
          
          ### END /etc/grub.d/20_linux_xen ###
          
          ### BEGIN /etc/grub.d/20_memtest86+ ###
          ### END /etc/grub.d/20_memtest86+ ###
          
          ### BEGIN /etc/grub.d/30_os-prober ###
          ### END /etc/grub.d/30_os-prober ###
          
          ### BEGIN /etc/grub.d/30_uefi-firmware ###
          menuentry 'UEFI Firmware Settings' $menuentry_id_option 'uefi-firmware' {
                  fwsetup
          }
          ### END /etc/grub.d/30_uefi-firmware ###
          
          ### BEGIN /etc/grub.d/40_custom ###
          # This file provides an easy way to add custom menu entries.  Simply type the
          # menu entries you want to add after this comment.  Be careful not to change
          # the 'exec tail' line above.
          ### END /etc/grub.d/40_custom ###
          
          ### BEGIN /etc/grub.d/41_custom ###
          if [ -f  ${config_directory}/custom.cfg ]; then
            source ${config_directory}/custom.cfg
          elif [ -z "${config_directory}" -a -f  $prefix/custom.cfg ]; then
            source $prefix/custom.cfg
          fi
          ### END /etc/grub.d/41_custom ###
          

          You can see the lines by the linux sections.