[ClusterLabs] Antw: [EXT] Re: Antw: Instable SLES15 SP3 kernel

2022-04-27 Thread Ulrich Windl
>>> "Gao,Yan" schrieb am 27.04.2022 um 14:31 in Nachricht <90862536-bbfb-f8b3-1a80-d8e9c1022...@suse.com>: > Hi Ulrich, > > On 2022/4/27 11:13, Ulrich Windl wrote: >> Update for the Update: >> >> I had installed SLES Updates in one VM and rebooted it via cluster. While >> installing the updates

Re: [ClusterLabs] OT: Linstor/DRBD Problem

2022-04-27 Thread Strahil Nikolov via Users
Why do you use Linstor and not DRBD ?As far as I know Linstor is more suitable for Kubernetes/Openshift . Best Regards,Strahil Nikolov On Thu, Apr 28, 2022 at 8:19, Eric Robinson wrote: This is probably off-topic but I’ll try anyway. Do we have any Linstor gurus around here? I’ve read

[ClusterLabs] OT: Linstor/DRBD Problem

2022-04-27 Thread Eric Robinson
This is probably off-topic but I'll try anyway. Do we have any Linstor gurus around here? I've read through the Linstor User Guide and all the help screens, but I don't see an answer to this question. We added a new physical drive to each of our cluster nodes and extended the LVM volume groups.

Re: [ClusterLabs] How many nodes redhat cluster does supports

2022-04-27 Thread Umar Draz
@Ken Now what is the solution? Do I need to communicate with Redhat commercial support on this issue? Regards, Umar On Thu, Apr 28, 2022 at 3:28 AM Ken Gaillot wrote: > On Wed, 2022-04-27 at 22:10 +0200, Valentin Vidić via Users wrote: > > On Thu, Apr 28, 2022 at 12:25:37AM +0500, Umar Draz

Re: [ClusterLabs] How many nodes redhat cluster does supports

2022-04-27 Thread Strahil Nikolov via Users
What is the output of 'gfs2_edit -p jindex /dev/shared_vg1/shared_lv1 |grep journal Source: https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/8/html-single/configuring_gfs2_file_systems#proc_adding-gfs2-journal-creating-mounting-gfs2 Best Regards,Strahil Nikolov On

Re: [ClusterLabs] How many nodes redhat cluster does supports

2022-04-27 Thread Ken Gaillot
On Wed, 2022-04-27 at 22:10 +0200, Valentin Vidić via Users wrote: > On Thu, Apr 28, 2022 at 12:25:37AM +0500, Umar Draz wrote: > > * sharedfs1_start_0 on g2fs-1 'error' (1): call=158, > > status='complete', > > exitreason='Couldn't mount device [/dev/shared_vg1/shared_lv1] as > > /mnt/webgfs',

Re: [ClusterLabs] OCF_TIMEOUT - Does it recover by itself?

2022-04-27 Thread Strahil Nikolov via Users
You can use a meta attribute to expire failures . The attribute name is 'failure-timeout'I have used it for my fencing devices as during the night the network was quite busy. Best Regards,Strahil Nikolov On Tue, Apr 26, 2022 at 23:54, Hayden, Robert via Users wrote: Robert Hayden |

Re: [ClusterLabs] How many nodes redhat cluster does supports

2022-04-27 Thread Valentin Vidić via Users
On Thu, Apr 28, 2022 at 12:25:37AM +0500, Umar Draz wrote: > * sharedfs1_start_0 on g2fs-1 'error' (1): call=158, status='complete', > exitreason='Couldn't mount device [/dev/shared_vg1/shared_lv1] as > /mnt/webgfs', last-rc-change='Tue Apr 26 01:07:45 2022', queued=0ms, > exec=806ms Maybe the

[ClusterLabs] How many nodes redhat cluster does supports

2022-04-27 Thread Umar Draz
Hi I am running a 3 nodes cluster on my AWS vms where I plan to use 3 nodes for my websites. Now the issue is only 2 nodes at a time can mount the lvm not all the 3 nodes. Here is the pcs status output. [root@g2fs-1 ~]# pcs status --full Cluster name: wp-cluster Cluster Summary: * Stack:

Re: [ClusterLabs] Antw: [EXT] Re: OCF_TIMEOUT ‑ Does it recover by itself?

2022-04-27 Thread Ken Gaillot
On Wed, 2022-04-27 at 08:49 +0200, Ulrich Windl wrote: > > > > Ken Gaillot schrieb am 26.04.2022 um > > > > 21:24 in > Nachricht > : > > On Tue, 2022‑04‑26 at 15:20 ‑0300, Salatiel Filho wrote: > > > I have a question about OCF_TIMEOUT. Some times my cluster shows > > > me > > > this on pcs

Re: [ClusterLabs] Antw: Instable SLES15 SP3 kernel

2022-04-27 Thread Gao,Yan via Users
Hi Ulrich, On 2022/4/27 11:13, Ulrich Windl wrote: Update for the Update: I had installed SLES Updates in one VM and rebooted it via cluster. While installing the updates in the VM the Xen host got RAM corruption (it seems any disk I/O on the host, either locally or via a VM image causes RAM

[ClusterLabs] Antw: Instable SLES15 SP3 kernel

2022-04-27 Thread Ulrich Windl
Update for the Update: I had installed SLES Updates in one VM and rebooted it via cluster. While installing the updates in the VM the Xen host got RAM corruption (it seems any disk I/O on the host, either locally or via a VM image causes RAM corruption): Apr 27 10:56:44 h19 kernel:

[ClusterLabs] Antw: [EXT] Re: OCF_TIMEOUT ‑ Does it recover by itself?

2022-04-27 Thread Ulrich Windl
>>> Ken Gaillot schrieb am 26.04.2022 um 21:24 in Nachricht : > On Tue, 2022‑04‑26 at 15:20 ‑0300, Salatiel Filho wrote: >> I have a question about OCF_TIMEOUT. Some times my cluster shows me >> this on pcs status: >> Failed Resource Actions: >> * fence‑server02_monitor_6 on server01

[ClusterLabs] Antw: Instable SLES15 SP3 kernel

2022-04-27 Thread Ulrich Windl
Hi! I want to give a non-update on the issue: The kernel still segfaults random processes, and there is really nothing from support within two months that could help improve the situation. The cluster is logging all kinds on non-funny messages like these: Apr 27 02:20:49 h18