proxmox

What to do with a Degraded ZFS Pool

What to do with a Degraded ZFS Pool

#Degraded #ZFS #Pool

“Craft Computing”

Thanks to Vultr for sponsoring today’s video. Visit to start your free trial, and receive $250 in credit for your first 30 days!

This video is going to focus on a ZFS Mirror inside Proxmox, but the commands and steps apply to nearly every type of ZFS array, regardless…

source

 

To see the full content, share this page by clicking one of the buttons below

Related Articles

25 Comments

  1. I love the show. I actually have thought about trying prox-mox for a while, just haven't yet. I hope in the future prox-mox can incorporate some sort of warning to let you know of ZFS issues, as I assume that zfs recently was added to prox-mox. Also, I would hope for the less tech types that having a zfs menu with these commands would do wonders so one would not have to goto command line and try to remember these commands to do a hd swap (or atleast show the command in a zfs menu.

    Does prox-mox have a email option to email you when things go wrong? I run 2 truenas scale box's and I have email enabled to let me know when there are problems.
    Maybe this is more of a wish list hoping prox-mox adds to future releases.

  2. As someone who's never used ProxMox, I was wondering how this was a 15 minute video given that ProxMox has always appeared to have a nice WebUI. And then the command line tools came out and I went "Oh. Ohhh. OHHHHHH."

  3. Imho this video is a bit rough, a bit rushed. No mention about setting up email notifications for events and zfs (very important to get notified of failures), no mention of the fact that this is a root pool so you have to follow a different procedure to create the boot partition and update the config so proxmox re-generates the boot files there too.

  4. I don’t like zfs on proxmox in particular. Every install in keeps about 50% of the drive for the system and the other half for the pool. Since I don’t need 500G for proxmox itself I ether remove the zfs pool entirely or extend it which is definitely not easy. I rather go through a Debian net install and have a clean lvm install and then install proxmox over it instead of dealing with zfs with the ISO.

    ZFS on a dedicated NAS is a completely different thing tho so don’t get me wrong. I like zfs but not in combination with proxmox.

  5. I've been casually looking for a while to try to find a disk format/system that will take care of a bunch of drives, allow you to add or remove a drive at will, of various sizes, and report any errors as they occur. It should try to maximise throughput (read and write) while having redundancy, so long as there are at least 2 drives. Ideally with a web interface.

    I know, moon on a stick.

  6. I've had errors on my large pool and I reset the disk twice in many years. It's still going and it's been months without a single error on that disk. Really interesting. I have a replacement new in box ready if I ever do feel the need to replace a disk but so far… nothing. That's over an array of 12 disks. I mean sure, my disks aren't worked very hard but still I'm impressed.

  7. Good video, but didn't you miss to replicate the partition table of the old / known good device (-part3 for rpool) and instead used the whole new NVMe device for rpool. If your now old NVMe device fails you have now no Boot-Partition to boot from on your newly added NVMe. Or am i completely wrong here?

    Those are the steps, i usally take to replace a failed Proxmox ZFS rpool disc:

    1) Replace the physical failed/offline drive with /dev/sdc (for example)

    Initialize Disk
    2) From the WebUI, Servername -> Disks -> Initialize Disk with GPT (/dev/sdc) OR gdisk /dev/sdc –> o (new empty GPT) –> w (write)

    Copy the partition table from /dev/sda (known good device) to /dev/sdc
    3) sgdisk –replicate=/dev/sdc /dev/sda

    Ensure the GUIDs are randomized
    4) sgdisk –randomize-guids /dev/sdc

    Install the Grub on the new disk
    5) grub-install /dev/sdc

    Then replace the disk in the ZFS pool
    6) zpool replace rpool /dev/sdc3
    OR zpool attach rpool /dev/sda3 /dev/sdc3 –> sda3 known good devcie / sdc3 new device

    Maybe detach old disk from the ZFS pool
    7) zpool detach rpool /dev/sdx3

    Maybe install Proxmox Boot Tool to new device
    8) proxmox-boot-tool status
    proxmox-boot-tool format /dev/sdc2
    proxmox-boot-tool init /dev/sdc2
    proxmox-boot-tool clean

  8. Do NOT use the simple /dev/ path when replacing a disk. there's no guarantee Linux will always use the same number for the same drive. There's a reason why rpool is set up with disk by id.

  9. Hey, @CraftComputing I do not think it is fair to share with the community how to block notifications that proxmox has non-enterprise repo enabled. Staff of Proxmox Server Solutions GmbH has to pay their bills to continue developing products.

  10. While software RAID is the future, I really did like my old Compaq hardware RAID servers. See a drive with a yellow light, pull it and replace it with a new one, job done. The controller does everything in the background, no need to even log in. The green light starts blinking as it rebuilds, when it's solid, it's done. I could even run redundant hot swap controllers in the cluster boxes.

  11. ZFS may be better, but replacing a drive on a RAID controller is so much easier. I was stumped when I had my first ever drive failure, and all the instructions said to pull the drive and insert a new one. Didn't realize it was that easy.

Leave a Reply