Hi! It should be easy (I guess), but when requiring both masters to be on the same node, can't you do with one DRBD device (something like putting a LVM VG on that and proivide two LVs)?
Regards, Ulrich >>> <fatcha...@gmx.de> schrieb am 18.01.2021 um 18:43 in Nachricht <trinity-8d5b7455-9df6-4cfa-9c24-931a12d0e322-1610991804216@3c-app-gmx-bs17>: > Hi again, > > I need some help to figure out how to let a two node cluster with two > drbd‑devices start the master‑devices on the same node. > How can I configure colocation to work that way ? I tried to bond one > drbd‑device with the other but that didn't work out quite well. > > This is my config: >> > I'm installing a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm using > this versions: >> > >> > kmod‑drbd90‑9.0.25‑2.el8_3.elrepo.x86_64 >> > drbd90‑utils‑9.13.1‑1.el8.elrepo.x86_64 >> > >> > pacemaker‑cluster‑libs‑2.0.4‑6.el8.x86_64 >> > pacemaker‑cli‑2.0.4‑6.el8.x86_64 >> > pacemaker‑schemas‑2.0.4‑6.el8.noarch >> > pacemaker‑2.0.4‑6.el8.x86_64 >> > pacemaker‑libs‑2.0.4‑6.el8.x86_64 >> > >> > clusternode‑names are lisbon and susanne >> > >> > Status of the cluster: >> > Cluster Summary: >> > * Stack: corosync >> > * Current DC: lisbon (version 2.0.4‑6.el8‑2deceaa3ae) ‑ partition with > quorum >> > * Last updated: Mon Jan 18 16:30:21 2021 >> > * Last change: Mon Jan 18 16:30:17 2021 by root via cibadmin on lisbon >> > * 2 nodes configured >> > * 7 resource instances configured >> > >> > Node List: >> > * Online: [ lisbon susanne ] >> > >> > Active Resources: >> > * HA‑IP_1 (ocf::heartbeat:IPaddr2): Started susanne >> > * Clone Set: drbd_database‑clone [drbd_database] (promotable): >> > * Masters: [ susanne ] >> > * Slaves: [ lisbon ] >> > * fs_database (ocf::heartbeat:Filesystem): Started susanne >> > * Clone Set: drbd_logsfiles‑clone [drbd_logsfiles] (promotable): >> > * Masters: [ susanne ] >> > * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne >> > >> > drbdadm status >> > >> > [root@susanne ~]# drbdadm status >> > drbd1 role:Primary >> > disk:UpToDate >> > lisbon role:Secondary >> > peer‑disk:UpToDate >> > >> > drbd2 role:Primary >> > disk:UpToDate >> > lisbon connection:Connecting >> > >> > [root@lisbon ~]# drbdadm status >> > drbd1 role:Secondary >> > disk:UpToDate >> > susanne role:Primary >> > peer‑disk:UpToDate >> > >> > >> > cluster‑config: >> > Cluster Name: mysql_cluster >> > Corosync Nodes: >> > susanne lisbon >> > Pacemaker Nodes: >> > lisbon susanne >> > >> > Resources: >> > Resource: HA‑IP_1 (class=ocf provider=heartbeat type=IPaddr2) >> > Attributes: cidr_netmask=24 ip=192.168.18.150 >> > Operations: monitor interval=15s (HA‑IP_1‑monitor‑interval‑15s) >> > start interval=0s timeout=20s (HA‑IP_1‑start‑interval‑0s) >> > stop interval=0s timeout=20s (HA‑IP_1‑stop‑interval‑0s) >> > Clone: drbd_database‑clone >> > Meta Attrs: clone‑max=2 clone‑node‑max=1 notify=true promotable=true > promoted‑max=1 promoted‑node‑max=1 >> > Resource: drbd_database (class=ocf provider=linbit type=drbd) >> > Attributes: drbd_resource=drbd1 >> > Operations: demote interval=0s timeout=90 > (drbd_database‑demote‑interval‑0s) >> > monitor interval=20 role=Slave timeout=20 > (drbd_database‑monitor‑interval‑20) >> > monitor interval=10 role=Master timeout=20 > (drbd_database‑monitor‑interval‑10) >> > notify interval=0s timeout=90 > (drbd_database‑notify‑interval‑0s) >> > promote interval=0s timeout=90 > (drbd_database‑promote‑interval‑0s) >> > reload interval=0s timeout=30 > (drbd_database‑reload‑interval‑0s) >> > start interval=0s timeout=240 (drbd_database‑start‑interval‑0s) >> > stop interval=0s timeout=100 (drbd_database‑stop‑interval‑0s) >> > Resource: fs_database (class=ocf provider=heartbeat type=Filesystem) >> > Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4 >> > Operations: monitor interval=20s timeout=40s > (fs_database‑monitor‑interval‑20s) >> > start interval=0s timeout=60s (fs_database‑start‑interval‑0s) >> > stop interval=0s timeout=60s (fs_database‑stop‑interval‑0s) >> > Clone: drbd_logsfiles‑clone >> > Meta Attrs: clone‑max=2 clone‑node‑max=1 notify=true promotable=true > promoted‑max=1 promoted‑node‑max=1 >> > Resource: drbd_logsfiles (class=ocf provider=linbit type=drbd) >> > Attributes: drbd_resource=drbd2 >> > Operations: demote interval=0s timeout=90 > (drbd_logsfiles‑demote‑interval‑0s) >> > monitor interval=20 role=Slave timeout=20 > (drbd_logsfiles‑monitor‑interval‑20) >> > monitor interval=10 role=Master timeout=20 > (drbd_logsfiles‑monitor‑interval‑10) >> > notify interval=0s timeout=90 > (drbd_logsfiles‑notify‑interval‑0s) >> > promote interval=0s timeout=90 > (drbd_logsfiles‑promote‑interval‑0s) >> > reload interval=0s timeout=30 > (drbd_logsfiles‑reload‑interval‑0s) >> > start interval=0s timeout=240 > (drbd_logsfiles‑start‑interval‑0s) >> > stop interval=0s timeout=100 (drbd_logsfiles‑stop‑interval‑0s) >> > Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem) >> > Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4 >> > Operations: monitor interval=20s timeout=40s > (fs_logfiles‑monitor‑interval‑20s) >> > start interval=0s timeout=60s (fs_logfiles‑start‑interval‑0s) >> > stop interval=0s timeout=60s (fs_logfiles‑stop‑interval‑0s) >> > >> > Stonith Devices: >> > Fencing Levels: >> > >> > Location Constraints: >> > Ordering Constraints: >> > start drbd_database‑clone then start fs_database (kind:Mandatory) > (id:order‑drbd_database‑clone‑fs_database‑mandatory) >> > start drbd_logsfiles‑clone then start fs_logfiles (kind:Mandatory) > (id:order‑drbd_logsfiles‑clone‑fs_logfiles‑mandatory) >> > Colocation Constraints: >> > fs_database with drbd_database‑clone (score:INFINITY) (with‑rsc‑role:Master) > (id:colocation‑fs_database‑drbd_database‑clone‑INFINITY) >> > fs_logfiles with drbd_logsfiles‑clone (score:INFINITY) > (with‑rsc‑role:Master) > (id:colocation‑fs_logfiles‑drbd_logsfiles‑clone‑INFINITY) > ERROR:>>>>> drbd_logsfiles‑clone with drbd_database‑clone (score:INFINITY) > (with‑rsc‑role:Master) > (id:colocation‑drbd_logsfiles‑clone‑drbd_database‑clone‑INFINITY) >> > Ticket Constraints: >> > >> > Alerts: >> > No alerts defined >> > >> > Resources Defaults: >> > No defaults set >> > Operations Defaults: >> > No defaults set >> > >> > Cluster Properties: >> > cluster‑infrastructure: corosync >> > cluster‑name: mysql_cluster >> > dc‑version: 2.0.4‑6.el8‑2deceaa3ae >> > have‑watchdog: false >> > last‑lrm‑refresh: 1610382881 >> > stonith‑enabled: false >> > >> > Tags: >> > No tags defined >> > >> > Quorum: >> > Options: > > > > Any suggestions are welcome > > stay safe and healty > > fatcharly > > _______________________________________________ > Manage your subscription: > https://lists.clusterlabs.org/mailman/listinfo/users > > ClusterLabs home: https://www.clusterlabs.org/ _______________________________________________ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/