I know this is an old conversation but I was hoping for a little help with my own HA cluster ZFS. I have built what you originally wrote up at github dot com/ewwhite/zfs-ha/wiki and it is partially functional but seeing some strange behavior and was wondering if anyone has any ideas.
I have 2 dell servers that are my Centos7 HA cluster, and have connected those via 2 HBA’s to a HP D2700 JBOD. When my primary controller is up, my zpool status shows perfect, all drives are great. When I do a test failover, and my secondary controller is up, my zpool shows a degraded state and shows that 3 drives (out of 25) are bad. I have tried using the zpool clear option to clear the status of the drives but they stay in a degraded state. Not sure why one controller shows all disks are good and one controller shows 3 bad drives. Any thoughts? Could this be caused by a bad HBA? Mini-SAS cable? Configuration issue?
Thanks in advance
A ZFS/cluster NOOB!!
P.S. - This is just a test lab, not production so there is no emergency, more of a curiosity.