>>> "Gao,Yan" schrieb am 27.04.2022 um 14:31 in Nachricht
<90862536-bbfb-f8b3-1a80-d8e9c1022...@suse.com>:
> Hi Ulrich,
>
> On 2022/4/27 11:13, Ulrich Windl wrote:
>> Update for the Update:
>>
>> I had installed SLES Updates in one VM and rebooted it via cluster. While
>> installing the updates
Why do you use Linstor and not DRBD ?As far as I know Linstor is more suitable
for Kubernetes/Openshift .
Best Regards,Strahil Nikolov
On Thu, Apr 28, 2022 at 8:19, Eric Robinson wrote:
This is probably off-topic but I’ll try anyway. Do we have any Linstor gurus
around here? I’ve read
This is probably off-topic but I'll try anyway. Do we have any Linstor gurus
around here? I've read through the Linstor User Guide and all the help screens,
but I don't see an answer to this question. We added a new physical drive to
each of our cluster nodes and extended the LVM volume groups.
@Ken
Now what is the solution? Do I need to communicate with Redhat commercial
support on this issue?
Regards,
Umar
On Thu, Apr 28, 2022 at 3:28 AM Ken Gaillot wrote:
> On Wed, 2022-04-27 at 22:10 +0200, Valentin Vidić via Users wrote:
> > On Thu, Apr 28, 2022 at 12:25:37AM +0500, Umar Draz
What is the output of 'gfs2_edit -p jindex /dev/shared_vg1/shared_lv1 |grep
journal
Source:
https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/8/html-single/configuring_gfs2_file_systems#proc_adding-gfs2-journal-creating-mounting-gfs2
Best Regards,Strahil Nikolov
On
On Wed, 2022-04-27 at 22:10 +0200, Valentin Vidić via Users wrote:
> On Thu, Apr 28, 2022 at 12:25:37AM +0500, Umar Draz wrote:
> > * sharedfs1_start_0 on g2fs-1 'error' (1): call=158,
> > status='complete',
> > exitreason='Couldn't mount device [/dev/shared_vg1/shared_lv1] as
> > /mnt/webgfs',
You can use a meta attribute to expire failures . The attribute name is
'failure-timeout'I have used it for my fencing devices as during the night the
network was quite busy.
Best Regards,Strahil Nikolov
On Tue, Apr 26, 2022 at 23:54, Hayden, Robert via
Users wrote:
Robert Hayden |
On Thu, Apr 28, 2022 at 12:25:37AM +0500, Umar Draz wrote:
> * sharedfs1_start_0 on g2fs-1 'error' (1): call=158, status='complete',
> exitreason='Couldn't mount device [/dev/shared_vg1/shared_lv1] as
> /mnt/webgfs', last-rc-change='Tue Apr 26 01:07:45 2022', queued=0ms,
> exec=806ms
Maybe the
Hi
I am running a 3 nodes cluster on my AWS vms where I plan to use 3 nodes
for my websites. Now the issue is only 2 nodes at a time can mount the lvm
not all the 3 nodes. Here is the pcs status output.
[root@g2fs-1 ~]# pcs status --full
Cluster name: wp-cluster
Cluster Summary:
* Stack:
On Wed, 2022-04-27 at 08:49 +0200, Ulrich Windl wrote:
> > > > Ken Gaillot schrieb am 26.04.2022 um
> > > > 21:24 in
> Nachricht
> :
> > On Tue, 2022‑04‑26 at 15:20 ‑0300, Salatiel Filho wrote:
> > > I have a question about OCF_TIMEOUT. Some times my cluster shows
> > > me
> > > this on pcs
Hi Ulrich,
On 2022/4/27 11:13, Ulrich Windl wrote:
Update for the Update:
I had installed SLES Updates in one VM and rebooted it via cluster. While
installing the updates in the VM the Xen host got RAM corruption (it seems any
disk I/O on the host, either locally or via a VM image causes RAM
Update for the Update:
I had installed SLES Updates in one VM and rebooted it via cluster. While
installing the updates in the VM the Xen host got RAM corruption (it seems any
disk I/O on the host, either locally or via a VM image causes RAM corruption):
Apr 27 10:56:44 h19 kernel:
>>> Ken Gaillot schrieb am 26.04.2022 um 21:24 in
Nachricht
:
> On Tue, 2022‑04‑26 at 15:20 ‑0300, Salatiel Filho wrote:
>> I have a question about OCF_TIMEOUT. Some times my cluster shows me
>> this on pcs status:
>> Failed Resource Actions:
>> * fence‑server02_monitor_6 on server01
Hi!
I want to give a non-update on the issue:
The kernel still segfaults random processes, and there is really nothing from
support within two months that could help improve the situation.
The cluster is logging all kinds on non-funny messages like these:
Apr 27 02:20:49 h18
14 matches
Mail list logo