P acemaker + corosync cluster with 2  virtual machines (ubuntu 22.04, 16 Gb 
RAM, 8 CPU each) are assembled into a cluster, an HBA is forwarded to each of 
them to connect to a disk shelf according to the instructions  
https://netbergtw.com/top-support/articles/zfs-cib /. A ZFS pool was assembled 
from 4 disks in draid1, resources were configured - virtual IP, iSCSITarget, 
iSCSILun. LUN connected in VMware. During an abnormal shutdown of the node, 
resources move, but at the moment this happens, VMware loses contact with the 
LUN, which should not happen. The journalctl log at the time of the move is 
here:  https://pastebin.com/eLj8DdtY . I also tried to build a common storage 
on drbd with cloned VIP and Target resources, but this also does not work, 
besides, every time I move, there are always some problems with the start of 
resources. Any ideas what can be done about this? Loss of communication with 
the LUN even for a couple of seconds is already critical.
 
corosync-qdevice/jammy,now 3.0.1-1 amd64 [installed]
corosync-qnetd/jammy,now 3.0.1-1 amd64 [installed]
corosync/jammy,now 3.1.6-1ubuntu1 amd64 [installed]
pacemaker-cli-utils/jammy,now 2.1.2-1ubuntu3 amd64 [installed,automatic]
pacemaker-common/jammy,now 2.1.2-1ubuntu3 all [installed,automatic]
pacemaker-resource-agents/jammy,now 2.1.2-1ubuntu3 all [installed,automatic]
pacemaker/jammy,now 2.1.2-1ubuntu3 amd64 [installed]
pcs/jammy,now 0.10.11-2ubuntu3 all [installed]
_______________________________________________
Manage your subscription:
https://lists.clusterlabs.org/mailman/listinfo/users

ClusterLabs home: https://www.clusterlabs.org/

Reply via email to