Hi Brent, I now use the drbd without starting the drbd service. All is managed by the pcs, everything gets mounted and unmounted as expected. In the beginning I was dealing with a lot of problems caused by the selinux and a wrong used constrain. Now everything works as I know it from my other cluster which are build on CentOS 5/6/7. Now in the beginning I also work with a disabled firewall because the system is in a secured dmz. Is there anything I can do for you ?
best regards and stay healthy fatcharly > Gesendet: Montag, 18. Januar 2021 um 22:16 Uhr > Von: "Brent Jensen" <[email protected]> > An: [email protected] > Betreff: Re: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started > > Are you getting the the cluster to switch over when doing a 'pcs node > standby' on the promoted node? I have a super basic config w/ your same > software versions and I cannot get the slave to promote (constantly > looping errors such as "Refusing to be Primary while peer is not > outdated", until I do a drbdadm up <resource> on the standby node). I > appreciate your input. Brent > > On 1/18/2021 9:49 AM, [email protected] wrote: > > Sorry Guys, > > > > problem found, it was a colocation constrain of the drbd-devices. > > > > best regards > > > > fatcharly > > > > > > > >> Gesendet: Montag, 18. Januar 2021 um 16:52 Uhr > >> Von: [email protected] > >> An: "clusterlabs" <[email protected]> > >> Betreff: [ClusterLabs] CentOS 8 & drbd 9 second slave is not started > >> > >> Hi, > >> > >> I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm > >> using this versions: > >> > >> kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64 > >> drbd90-utils-9.13.1-1.el8.elrepo.x86_64 > >> > >> pacemaker-cluster-libs-2.0.4-6.el8.x86_64 > >> pacemaker-cli-2.0.4-6.el8.x86_64 > >> pacemaker-schemas-2.0.4-6.el8.noarch > >> pacemaker-2.0.4-6.el8.x86_64 > >> pacemaker-libs-2.0.4-6.el8.x86_64 > >> > >> clusternode-names are lisbon and susanne > >> > >> There are two drbd-resources configured which are paired with > >> filesystem-resources. Both are working in simple master/slave > >> configuration. > >> When I start up the cluster one resource is promoted to a master and a > >> slave, the other is just getting on one the master started, but the slave > >> is not starting up. > >> > >> Status of the cluster: > >> Cluster Summary: > >> * Stack: corosync > >> * Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) - partition with > >> quorum > >> * Last updated: Mon Jan 18 16:30:21 2021 > >> * Last change: Mon Jan 18 16:30:17 2021 by root via cibadmin on lisbon > >> * 2 nodes configured > >> * 7 resource instances configured > >> > >> Node List: > >> * Online: [ lisbon susanne ] > >> > >> Active Resources: > >> * HA-IP_1 (ocf::heartbeat:IPaddr2): Started susanne > >> * Clone Set: drbd_database-clone [drbd_database] (promotable): > >> * Masters: [ susanne ] > >> * Slaves: [ lisbon ] > >> * fs_database (ocf::heartbeat:Filesystem): Started susanne > >> * Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable): > >> * Masters: [ susanne ] > >> * fs_logfiles (ocf::heartbeat:Filesystem): Started susanne > >> > >> drbdadm status > >> > >> [root@susanne ~]# drbdadm status > >> drbd1 role:Primary > >> disk:UpToDate > >> lisbon role:Secondary > >> peer-disk:UpToDate > >> > >> drbd2 role:Primary > >> disk:UpToDate > >> lisbon connection:Connecting > >> > >> [root@lisbon ~]# drbdadm status > >> drbd1 role:Secondary > >> disk:UpToDate > >> susanne role:Primary > >> peer-disk:UpToDate > >> > >> > >> cluster-config: > >> Cluster Name: mysql_cluster > >> Corosync Nodes: > >> susanne lisbon > >> Pacemaker Nodes: > >> lisbon susanne > >> > >> Resources: > >> Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2) > >> Attributes: cidr_netmask=24 ip=192.168.18.150 > >> Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s) > >> start interval=0s timeout=20s (HA-IP_1-start-interval-0s) > >> stop interval=0s timeout=20s (HA-IP_1-stop-interval-0s) > >> Clone: drbd_database-clone > >> Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true > >> promoted-max=1 promoted-node-max=1 > >> Resource: drbd_database (class=ocf provider=linbit type=drbd) > >> Attributes: drbd_resource=drbd1 > >> Operations: demote interval=0s timeout=90 > >> (drbd_database-demote-interval-0s) > >> monitor interval=20 role=Slave timeout=20 > >> (drbd_database-monitor-interval-20) > >> monitor interval=10 role=Master timeout=20 > >> (drbd_database-monitor-interval-10) > >> notify interval=0s timeout=90 > >> (drbd_database-notify-interval-0s) > >> promote interval=0s timeout=90 > >> (drbd_database-promote-interval-0s) > >> reload interval=0s timeout=30 > >> (drbd_database-reload-interval-0s) > >> start interval=0s timeout=240 > >> (drbd_database-start-interval-0s) > >> stop interval=0s timeout=100 > >> (drbd_database-stop-interval-0s) > >> Resource: fs_database (class=ocf provider=heartbeat type=Filesystem) > >> Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4 > >> Operations: monitor interval=20s timeout=40s > >> (fs_database-monitor-interval-20s) > >> start interval=0s timeout=60s > >> (fs_database-start-interval-0s) > >> stop interval=0s timeout=60s (fs_database-stop-interval-0s) > >> Clone: drbd_logsfiles-clone > >> Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true > >> promoted-max=1 promoted-node-max=1 > >> Resource: drbd_logsfiles (class=ocf provider=linbit type=drbd) > >> Attributes: drbd_resource=drbd2 > >> Operations: demote interval=0s timeout=90 > >> (drbd_logsfiles-demote-interval-0s) > >> monitor interval=20 role=Slave timeout=20 > >> (drbd_logsfiles-monitor-interval-20) > >> monitor interval=10 role=Master timeout=20 > >> (drbd_logsfiles-monitor-interval-10) > >> notify interval=0s timeout=90 > >> (drbd_logsfiles-notify-interval-0s) > >> promote interval=0s timeout=90 > >> (drbd_logsfiles-promote-interval-0s) > >> reload interval=0s timeout=30 > >> (drbd_logsfiles-reload-interval-0s) > >> start interval=0s timeout=240 > >> (drbd_logsfiles-start-interval-0s) > >> stop interval=0s timeout=100 > >> (drbd_logsfiles-stop-interval-0s) > >> Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem) > >> Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4 > >> Operations: monitor interval=20s timeout=40s > >> (fs_logfiles-monitor-interval-20s) > >> start interval=0s timeout=60s > >> (fs_logfiles-start-interval-0s) > >> stop interval=0s timeout=60s (fs_logfiles-stop-interval-0s) > >> > >> Stonith Devices: > >> Fencing Levels: > >> > >> Location Constraints: > >> Ordering Constraints: > >> start drbd_database-clone then start fs_database (kind:Mandatory) > >> (id:order-drbd_database-clone-fs_database-mandatory) > >> start drbd_logsfiles-clone then start fs_logfiles (kind:Mandatory) > >> (id:order-drbd_logsfiles-clone-fs_logfiles-mandatory) > >> Colocation Constraints: > >> fs_database with drbd_database-clone (score:INFINITY) > >> (with-rsc-role:Master) > >> (id:colocation-fs_database-drbd_database-clone-INFINITY) > >> fs_logfiles with drbd_logsfiles-clone (score:INFINITY) > >> (with-rsc-role:Master) > >> (id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY) > >> drbd_logsfiles-clone with drbd_database-clone (score:INFINITY) > >> (with-rsc-role:Master) > >> (id:colocation-drbd_logsfiles-clone-drbd_database-clone-INFINITY) > >> Ticket Constraints: > >> > >> Alerts: > >> No alerts defined > >> > >> Resources Defaults: > >> No defaults set > >> Operations Defaults: > >> No defaults set > >> > >> Cluster Properties: > >> cluster-infrastructure: corosync > >> cluster-name: mysql_cluster > >> dc-version: 2.0.4-6.el8-2deceaa3ae > >> have-watchdog: false > >> last-lrm-refresh: 1610382881 > >> stonith-enabled: false > >> > >> Tags: > >> No tags defined > >> > >> Quorum: > >> Options: > >> > >> > >> Any suggestions are welcome > >> > >> stay safe and healty > >> > >> fatcharly > >> _______________________________________________ > >> Manage your subscription: > >> https://lists.clusterlabs.org/mailman/listinfo/users > >> > >> ClusterLabs home: https://www.clusterlabs.org/ > >> > > _______________________________________________ > > Manage your subscription: > > https://lists.clusterlabs.org/mailman/listinfo/users > > > > ClusterLabs home: https://www.clusterlabs.org/ > > -- > This email has been checked for viruses by Avast antivirus software. > https://www.avast.com/antivirus > > _______________________________________________ Manage your subscription: https://lists.clusterlabs.org/mailman/listinfo/users ClusterLabs home: https://www.clusterlabs.org/
