Re: [Linux-cluster] sparse-file clone breaks on GFS2

2020-09-24 Thread Pierre-Philipp Braun
It seem that 4.18.20 is missing the relevant parts of the following fixes:   a27a0c9b6a2 ("gfs2: gfs2_walk_metadata fix")   566a2ab3c90 ("gfs2: Another gfs2_walk_metadata fix") Without those fixes, lseek SEEK_HOLE/SEEK_DATa will sometimes report garbage, and so cp doesn't know what to copy.

Re: [Linux-cluster] sparse-file clone breaks on GFS2

2020-09-21 Thread Gionatan Danti
Il 2020-09-21 15:45 Andreas Gruenbacher ha scritto: On Mon, Sep 21, 2020 at 3:40 PM Pierre-Philipp Braun wrote: In case it matters, I am using vanilla Linux 4.18.20 and not the RHEL nor CentOS with patches. It seem that 4.18.20 is missing the relevant parts of the following fixes:

Re: [Linux-cluster] sparse-file clone breaks on GFS2

2020-09-20 Thread Pierre-Philipp Braun
>> cd573cfaace07e7949bc0c46028904ff /root/dummy-on-ext4.ext4 >> cd573cfaace07e7949bc0c46028904ff /data2/dummy-clone-from-ext4.ext4 >> >> and... WOW! Cloning it yet again, it finally remains intact! >> >> cd573cfaace07e7949bc0c46028904ff dummy-clone-from-ext4-and-clone.ext4 >>

Re: [Linux-cluster] sparse-file clone breaks on GFS2

2020-09-19 Thread Gionatan Danti
Il 2020-09-19 18:57 Pierre-Philipp Braun ha scritto: On 19.09.2020 14:14, Gionatan Danti wrote: cp --sparse=never Hello neighbor (ciao from France). I am not sure what you mean, as there **was both** the original and the copied file sums in my message. Sorry, I somewhat managed to miss

Re: [Linux-cluster] sparse-file clone breaks on GFS2

2020-09-19 Thread Pierre-Philipp Braun
On 19.09.2020 14:14, Gionatan Danti wrote: cp --sparse=never Hello neighbor (ciao from France). I am not sure what you mean, as there **was both** the original and the copied file sums in my message. Anyhow, thank you and good idea about --sparse=never. I applied the checksum multiple

Re: [Linux-cluster] sparse-file clone breaks on GFS2

2020-09-19 Thread Gionatan Danti
Il 2020-09-19 13:21 Pierre-Philipp Braun ha scritto: Hello #linux-cluster I am attempting to host thin-provisioned virtual disks on GFS2. In that regard, I experience a weird and unexpected issue: when copying (or packing/extracting) a sparse file with a file-system on it, and which lives and

[Linux-cluster] sparse-file clone breaks on GFS2

2020-09-19 Thread Pierre-Philipp Braun
Hello #linux-cluster I am attempting to host thin-provisioned virtual disks on GFS2. In that regard, I experience a weird and unexpected issue: when copying (or packing/extracting) a sparse file with a file-system on it, and which lives and gets cloned on a GFS2 mount point, it results in a

Re: [Linux-cluster] Using gfs2/dlm directly and from outer space

2020-08-23 Thread Pierre-Philipp Braun
It's been almost a week, so in the meanwhile I found a workaround. I just downgraded the kernel to 4.18.x to match RHEL8's maintained version. I did not even have to apply all the RHEL/CentOS patches I found. Just building the latest 4.18 vanilla kernel (4.18.20) did the job. I suppose it

[Linux-cluster] Using gfs2/dlm directly and from outer space

2020-08-16 Thread Pierre-Philipp Braun
Hello #linux-cluster David's guide was quite helpful to attempt to run GFS2 with a minimal amount of cluster setup, and on top of a DRBD mirror (http://people.redhat.com/teigland/cluster4-gfs2-dlm.txt). Instead of using RHEL, I built all the required pieces from scratch on Slackware Linux

[Linux-cluster] clvmd deactive all LVM when an node reboot

2019-10-15 Thread Domenico Pastore
Hi, I have configure an pacemaker cluster with corosync. The cluster in compose by 4 node now. I have an problem with CLVMD, when a node leaves the cluster, all services are correctly relocated to other cluster nodes. But when the node re-joins the cluster, all services on all nodes get

Re: [Linux-cluster] PCS Cluster

2019-04-25 Thread Jan Pokorný
Nicola, On 01/02/19 09:59 +, Nicola Contu wrote: > any news on this? > > Is there anything else we can do to troubleshoot before this happens > again? I don't lively attend this list anymore, and so the most of the pacemaker/corosync cluster stack audience unless they are interested in

Re: [Linux-cluster] PCS Cluster

2019-02-01 Thread Nicola Contu
Hello, any news on this? Is there anything else we can do to troubleshoot before this happens again? From: Nicola Contu Sent: Wednesday, January 30, 2019 4:08:20 PM To: linux-cluster@redhat.com Subject: Re: [Linux-cluster] PCS Cluster I can give you the

Re: [Linux-cluster] PCS Cluster

2019-01-30 Thread Nicola Contu
I can give you the current arp table, but we do not log ARP so there’s no way to tell you what the table looked like at that particular time. Or we need to wait for the next time if someone in Engineering is available to get an ARP (I am not maning routers unfortunately) Do you want me to send

Re: [Linux-cluster] PCS Cluster

2019-01-30 Thread Michael Schwartzkopff
Am 30.01.19 um 09:59 schrieb Nicola Contu: > Hello, > I've setup a cluster using pacemaker and corosync for HA using HAProxy. > Two resources : Haproxy and a floating IP. > They are running on Centos7 on vmware. > > The strange thing that happens sometime is that the floating IP won't be >

[Linux-cluster] PCS Cluster

2019-01-30 Thread Nicola Contu
Hello, I've setup a cluster using pacemaker and corosync for HA using HAProxy. Two resources : Haproxy and a floating IP. They are running on Centos7 on vmware. The strange thing that happens sometime is that the floating IP won't be pingable from outside the LAN. So for example, a server on the

[Linux-cluster] gfs2-utils 3.2.0 released

2018-05-24 Thread Andrew Price
Hi all, I am happy to announce the 3.2.0 release of gfs2-utils. This is an important release adding support for new on-disk features introduced in the 4.16 kernel. In fact it is required when building against 4.16 and later kernel headers due to poor assumptions made by earlier gfs2-utils

Re: [Linux-cluster] DLM won't (stay) running

2018-05-09 Thread Jason Gauthier
On Wed, May 9, 2018 at 6:26 AM, Andrew Price wrote: > [linux-cluster@ isn't really used nowadays; CCing users@clusterlabs] > > On 08/05/18 12:18, Jason Gauthier wrote: >> >> Greetings, >> >> I'm working on a setup of a two-node cluster with shared storage. >> I've been

Re: [Linux-cluster] DLM won't (stay) running

2018-05-09 Thread Andrew Price
[linux-cluster@ isn't really used nowadays; CCing users@clusterlabs] On 08/05/18 12:18, Jason Gauthier wrote: Greetings, I'm working on a setup of a two-node cluster with shared storage. I've been able to see the storage on both nodes, and appropriate configuration for fencing the bock

Re: [Linux-cluster] DLM won't (stay) running

2018-05-08 Thread Jason Gauthier
On Tue, May 8, 2018 at 10:50 AM, David Teigland wrote: > On Tue, May 08, 2018 at 07:18:17AM -0400, Jason Gauthier wrote: >> node 1084772368: alpha >> node 1084772369: beta >> primitive p_dlm_controld ocf:pacemaker:controld \ >> op monitor interval=60 timeout=60 \ >>

Re: [Linux-cluster] DLM won't (stay) running

2018-05-08 Thread David Teigland
On Tue, May 08, 2018 at 07:18:17AM -0400, Jason Gauthier wrote: > node 1084772368: alpha > node 1084772369: beta > primitive p_dlm_controld ocf:pacemaker:controld \ > op monitor interval=60 timeout=60 \ > meta target-role=Started args=-K > primitive p_gfs_controld

[Linux-cluster] DLM won't (stay) running

2018-05-08 Thread Jason Gauthier
Greetings, I'm working on a setup of a two-node cluster with shared storage. I've been able to see the storage on both nodes, and appropriate configuration for fencing the bock device. The next step was getting DLM and GFS2 in a clone group to mount the FS on both drives. This is where I am

[Linux-cluster] 电子邮件升级(不要忽略)!

2017-12-08 Thread Admin Localhost
NB2r31v84 linux-cluster@redhat.com, �D9f2Fb11NECv84�EARA8[89Q68|FB~DFv84[89Q68�1AwE5�0C`A8v84u35[50�AENF6�1alinux-clus...@redhat.com N0D[89Q68 SEF�FDg09uC5kD2002 ODDbA4`A8v84�AE{B1Q4DSD751�EF�93�7CT8CuC5kD2e3BQFB002 W2824\0FeF6Q85S47~A7`A8v84u35[50�AENF6[89Q68002

Re: [Linux-cluster] GFS2 DLM problem on NVMes

2017-11-22 Thread Eric H. Chang
Hi Dave, When errors started to come out, the system got slower (perf degraded) and lots of error messages showed up repeatedly. Specifically, when the large amount of slab memory was reclaimed such as 9GB to 6GB, the about 30 error messages came out. ‘send_repeat_remove’ messages were

Re: [Linux-cluster] GFS2 DLM problem on NVMes

2017-11-22 Thread David Teigland
On Wed, Nov 22, 2017 at 04:32:13AM +, Eric H. Chang wrote: > We��ve tested with different ��toss_secs�� as advised. When we > configured it as 1000, we saw the ��send_repeat_remove�� log after > 1000sec. We can test with other values on ��toss_secs��, but we think it > would have the same

Re: [Linux-cluster] GFS2 DLM problem on NVMes

2017-11-21 Thread Eric H. Chang
Hi Dave and Steven, Thank you for the assistance. We made some progress here and would like to share with you. #1. We’ve set ‘vm.vfs_cache_pressure’ to zero and ran tests. As a result, we couldn’t see the same problem happening and observed that the slab grew slowly and saturated to 25GB

Re: [Linux-cluster] GFS2 DLM problem on NVMes

2017-11-20 Thread David Teigland
> We are developing storage systems using 10 NVMes (current test set). > Using MD RAID10 + CLVM/GFS2 over four hosts achieves 22 GB/s (Max. on Reads). Does MD RAID10 work correctly under GFS2? Does the RAID10 make use of the recent md-cluster enhancements (which also use the dlm)? > However, a

Re: [Linux-cluster] GFS2 DLM problem on NVMes

2017-11-20 Thread Steven Whitehouse
Hi, On 20/11/17 04:23, 성백재 wrote: Hello, List. We are developing storage systems using 10 NVMes (current test set). Using MD RAID10 + CLVM/GFS2 over four hosts achieves 22 GB/s (Max. on Reads). However, a GFS2 DLM problem occurred. The problem is that each host frequently reports “dlm:

[Linux-cluster] GFS2 DLM problem on NVMes

2017-11-19 Thread 성백재
Hello, List. We are developing storage systems using 10 NVMes (current test set). Using MD RAID10 + CLVM/GFS2 over four hosts achieves 22 GB/s (Max. on Reads). However, a GFS2 DLM problem occurred. The problem is that each host frequently reports “dlm: gfs2: send_repeat_remove” kernel messages,

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Gionatan Danti
Il 29-08-2017 13:28 Steven Whitehouse ha scritto: There is no siginificant overhead when reading the same file on multiple nodes. The overhead mostly applies when writes are involved in some form, whether mixed with other writes or reads. GFS2 does ensure cache coherency, but in order to do that

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Steven Whitehouse
Hi, On 29/08/17 12:26, Gionatan Danti wrote: Il 29-08-2017 13:13 Steven Whitehouse ha scritto: Whatever kind of storage is being used with GFS2, it needs to act as if there was no cache or as if there is a common cache between all nodes - what we want to avoid is caches which are specific to

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Gionatan Danti
Il 29-08-2017 13:13 Steven Whitehouse ha scritto: Whatever kind of storage is being used with GFS2, it needs to act as if there was no cache or as if there is a common cache between all nodes - what we want to avoid is caches which are specific to each node. Using individual node caching will

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Steven Whitehouse
Hi, On 29/08/17 12:07, Gionatan Danti wrote: Il 29-08-2017 12:59 Steven Whitehouse ha scritto: Yes, it definitely needs to be set to cache=none mode. Barrier passing is only one issue, and as you say it is down to the cache coherency, since the block layer is not aware of the caching

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Gionatan Danti
Il 29-08-2017 12:59 Steven Whitehouse ha scritto: Yes, it definitely needs to be set to cache=none mode. Barrier passing is only one issue, and as you say it is down to the cache coherency, since the block layer is not aware of the caching requirements of the upper layers in this case. Ok.

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Steven Whitehouse
On 29/08/17 11:54, Gionatan Danti wrote: Hi Steven, Il 29-08-2017 11:45 Steven Whitehouse ha scritto: Yes, there is some additional overhead due to the clustering. You can however usually organise things so that the overheads are minimised as you mentioned above by being careful about the

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Gionatan Danti
Hi Steven, Il 29-08-2017 11:45 Steven Whitehouse ha scritto: Yes, there is some additional overhead due to the clustering. You can however usually organise things so that the overheads are minimised as you mentioned above by being careful about the workload. No. You want to use the default

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-29 Thread Steven Whitehouse
Hi, On 26/08/17 07:11, Gionatan Danti wrote: Hi list, I am evaluating how to refresh my "standard" cluster configuration and GFS2 clearly is on the table ;) GOAL: to have a 2-node HA cluster running DRBD (active/active), GFS2 (to store disk image) and KVM (as hypervisor). The cluster had

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-26 Thread Gionatan Danti
Il 26-08-2017 11:34 Kristián Feldsam ha scritto: Hello, accroding to red hat documentation "smaller is better". I personaly use 1TB volumes with 256MB journal https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html-single/Global_File_System_2/index.html#s1-formatting-gfs2

Re: [Linux-cluster] GFS2 as virtual machine disk store

2017-08-26 Thread Kristián Feldsam
Hello, accroding to red hat documentation "smaller is better". I personaly use 1TB volumes with 256MB journal https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html-single/Global_File_System_2/index.html#s1-formatting-gfs2 S pozdravem Kristián Feldsam Tel.: +420 773 303

[Linux-cluster] GFS2 as virtual machine disk store

2017-08-26 Thread Gionatan Danti
Hi list, I am evaluating how to refresh my "standard" cluster configuration and GFS2 clearly is on the table ;) GOAL: to have a 2-node HA cluster running DRBD (active/active), GFS2 (to store disk image) and KVM (as hypervisor). The cluster had to support live migration, but manual failover

Re: [Linux-cluster] Need advice Redhat Clusters

2017-07-30 Thread Digimer
On 2017-07-30 02:03 PM, deepesh kumar wrote: > I need to set up 2 node HA Active Passive redhat cluster on rhel 6.9. > > Should I start with rgmanager or pacemaker .. > > Do I need Quorum disk ..(mandatory ) and what fence method should I use. > > Thanks to great friends..!!! > > -- >

[Linux-cluster] Need advice Redhat Clusters

2017-07-30 Thread deepesh kumar
I need to set up 2 node HA Active Passive redhat cluster on rhel 6.9. Should I start with rgmanager or pacemaker .. Do I need Quorum disk ..(mandatory ) and what fence method should I use. Thanks to great friends..!!! -- DEEPESH KUMAR -- Linux-cluster mailing list

Re: [Linux-cluster] GFS2 Errors

2017-07-19 Thread Kristián Feldsam
Hello, kernel actually running in nodes is 4.11.1-1.el7.elrepo.x86_64 use case: 3 compute nodes in cluster corosync/pacemaker resources dlm clvm gfs volume 1 gfs volume 2 volume journal size 256MB please tell me cmds for more informations Thank you! S pozdravem Kristián Feldsam Tel.: +420

Re: [Linux-cluster] GFS2 Errors

2017-07-19 Thread Steven Whitehouse
Hi, On 19/07/17 00:39, Digimer wrote: On 2017-07-18 07:25 PM, Kristián Feldsam wrote: Hello, I see today GFS2 errors in log and nothing about that is on net, so I writing to this mailing list. node2 19.07.2017 01:11:55 kernel kernerr vmscan: shrink_slab:

Re: [Linux-cluster] GFS2 Errors

2017-07-18 Thread Digimer
On 2017-07-18 07:25 PM, Kristián Feldsam wrote: > Hello, I see today GFS2 errors in log and nothing about that is on net, > so I writing to this mailing list. > > node2 19.07.2017 01:11:55 kernel kernerr vmscan: shrink_slab: > gfs2_glock_shrink_scan+0x0/0x2f0 [gfs2] negative objects

Re: [Linux-cluster] missing tags on dlm git repo at pagure.io

2017-06-19 Thread Jan Pokorný
On 01/04/17 00:46 +0200, Jean-Marc Saffroy wrote: > I just noticed that git tags are gone on the repo at pagure.io. Good catch, Jean-Marc. As you can see, this list doesn't enjoy much attention these days, as you can also observe with the delay of my response. FYI, quite a bit of service-level

[Linux-cluster] Error while doing failover of cluster nodes - Rhel7.2 - Gfs2 over nfs

2017-05-25 Thread Dawood Munavar S M
Hello, I have configured a Rhel 7.2 pacemaker cluster and created nfs resources over gfs2 file system & Mounting of file system is successful on NFS client. I started I/O on NFS client and when I failover one of the cluster, I see the below message and the I/O terminates, 05/12/17 20:15:26

Re: [Linux-cluster] HA cluster 6.5 redhat active passive Error

2017-04-29 Thread Digimer
On 29/04/17 07:34 AM, deepesh kumar wrote: > Hi , > > Currently I am testing 2 node active passive cluster on RHEL 6.5 with > ext4 fs. > > I have set up all basic settings ..like Luci , ricci and other services > running and both the nodes are part of cluster. > > I am not able to make the

[Linux-cluster] HA cluster 6.5 redhat active passive Error

2017-04-29 Thread deepesh kumar
Hi , Currently I am testing 2 node active passive cluster on RHEL 6.5 with ext4 fs. I have set up all basic settings ..like Luci , ricci and other services running and both the nodes are part of cluster. I am not able to make the clvmd configuration attributes for my vy . All required seetings

Re: [Linux-cluster] Rhel 7.2 Pacemaker cluster - gfs2 file system- NFS document

2017-04-29 Thread Dawood Munavar S M
Hi Emmanuel, Please find the status below, [root@node1-emulex ~]# pcs status Cluster name: mycluster Stack: corosync Current DC: node2-atto (version 1.1.15-11.el7_3.4-e174ec8) - partition with quorum Last updated: Sat Apr 29 03:01:41 2017 Last change: Fri Apr 28 11:09:25 2017 by root

Re: [Linux-cluster] Rhel 7.2 Pacemaker cluster - gfs2 file system- NFS document

2017-04-28 Thread emmanuel segura
can you provide pcs status ? 2017-04-28 17:51 GMT+02:00 Dawood Munavar S M : > Hi, > > Thanks for your reply and the valuable comments. > > Curently I am only sanity testing the Rhel 7.2 cluster configuration with > our FC storage arrays, so performance really

Re: [Linux-cluster] Rhel 7.2 Pacemaker cluster - gfs2 file system- NFS document

2017-04-28 Thread Dawood Munavar S M
Hi, Thanks for your reply and the valuable comments. Curently I am only sanity testing the Rhel 7.2 cluster configuration with our FC storage arrays, so performance really doesn't matters for us at the moment. Also with regard to creating nfs cluster resources over gfs2, we followed the below

Re: [Linux-cluster] Rhel 7.2 Pacemaker cluster - gfs2 file system- NFS document

2017-04-28 Thread Digimer
On 28/04/17 06:34 AM, Dawood Munavar S M wrote: > Hello All, > > Could you please share any links/documents to create NFS HA cluster over > gfs2 file system using Pacemaker. > > Currently I have completed till mounting of gfs2 file systems on cluster > nodes and now I need to create cluster

[Linux-cluster] Rhel 7.2 Pacemaker cluster - gfs2 file system- NFS document

2017-04-28 Thread Dawood Munavar S M
Hello All, Could you please share any links/documents to create NFS HA cluster over gfs2 file system using Pacemaker. Currently I have completed till mounting of gfs2 file systems on cluster nodes and now I need to create cluster resources for NFS server, exports and mount on client. Thanks,

[Linux-cluster] clvmd issue with RH/Centos 6.9 update

2017-04-11 Thread Andrea Costantino
Hello fellow cluster guys, I just upgraded my CentOS cluster to 6.9 and after a node reboot I discovered that the clvm functionality was broken. Long story made short, I tracked it back to clvmd process being hung with one CPU stuck. The quick and dirty solution was to rollback kernel

Re: [Linux-cluster] Mount error: gfs2: rhel 7.2: Transport endpoint is not connected

2017-04-04 Thread Bob Peterson
- Original Message - | Hello Team, | | I am working on creating cluster configurations using pacemaker on Rhel 7.2 | server connected to a SAN storage controller and currently I am struck with | the below issue. | It would be great if you can help me in resolving this issue. | | |

[Linux-cluster] Mount error: gfs2: rhel 7.2: Transport endpoint is not connected

2017-04-04 Thread Dawood Munavar S M
Hello Team, I am working on creating cluster configurations using pacemaker on Rhel 7.2 server connected to a SAN storage controller and currently I am struck with the below issue. It would be great if you can help me in resolving this issue. *Node1: hostname: node1-atto* *Node2: hostname:

[Linux-cluster] missing tags on dlm git repo at pagure.io

2017-03-31 Thread Jean-Marc Saffroy
Hi, I just noticed that git tags are gone on the repo at pagure.io. In case someone wants to restore them, in my older checkout the tags are: $ git log --no-walk --tags --pretty="%h %d %s" --decorate=full e9302c0 (tag: refs/tags/dlm-4.0.7, refs/remotes/origin/master, refs/remotes/origin/HEAD)

[Linux-cluster] gfs2-utils 3.1.10 released

2017-03-28 Thread Andrew Price
Hi all, I am happy to announce the 3.1.10 release of gfs2-utils. This is the first release of gfs2-utils since the project was moved to Pagure git hosting and is a relatively small release, mainly adding polish and stability improvements over 3.1.9. This release includes the following

Re: [Linux-cluster] Active/passive cluster between physical and VM

2017-03-22 Thread Digimer
On 22/03/17 03:11 AM, Amjad Syed wrote: > Hello, > > We are planning to build a 2 node Active/passive cluster using pacemaker. > Can the cluster be build between one physical and one VM machine in > Centos 7.x? > If yes, what can be used as fencing agent ? So long as the traffic between the

[Linux-cluster] Active/passive cluster between physical and VM

2017-03-22 Thread Amjad Syed
Hello, We are planning to build a 2 node Active/passive cluster using pacemaker. Can the cluster be build between one physical and one VM machine in Centos 7.x? If yes, what can be used as fencing agent ? -- Linux-cluster mailing list Linux-cluster@redhat.com

Re: [Linux-cluster] Active passive cluster with shared storage without LVM

2017-03-03 Thread Stefan Fuhrmann
Hi Have a look to drbd http://www.drbd.org/en/ At active/passive setup you can use ext filesystem. Stefan Am Freitag, 3. März 2017, 12:56:17 CET schrieb Amjad Syed: > Hello, > > We are using RHEL 7.2 to create an active/passive cluster using pacemaker. > The cluster will have shared storage.

[Linux-cluster] Active passive cluster with shared storage without LVM

2017-03-03 Thread Amjad Syed
Hello, We are using RHEL 7.2 to create an active/passive cluster using pacemaker. The cluster will have shared storage. Our servers have standard partition. Not LVM . We are using only 2 servers for this cluster. Can it be possible to create shared stroage resource in pacemaker without using

[Linux-cluster] Pagure.io as legacy codebases/distribution files/documentation hosting (Was: Moving cluster project)

2017-03-02 Thread Jan Pokorný
[I've realized I should give a heads up to linux-cluster as well, beside cluster-devel and developers-at-clusterlabs lists, especially since I am reusing it's name for a projects group/namespace in pagure.io] On 28/02/17 03:18 +0100, Jan Pokorný wrote: > On 17/01/17 22:27 +0100, Jan Pokorný

Re: [Linux-cluster] gfs2-utils.git moving on Feb 3

2017-02-03 Thread Andrew Price
Hi, This is now complete and pushes to the old Fedora Hosted git repository should no longer work. Please update your gfs2-utils trees as below. Thanks, Andy On 31/01/17 15:15, Andrew Price wrote: Hi, This is a heads-up that from Friday February 3 the gfs2-utils git repository will be

[Linux-cluster] gfs2-utils.git moving on Feb 3

2017-01-31 Thread Andrew Price
Hi, This is a heads-up that from Friday February 3 the gfs2-utils git repository will be hosted on Pagure. We are migrating it away from Fedora Hosted as that service is due to be discontinued at the end of February. The gfs2-utils git repository on Pagure

Re: [Linux-cluster] Protocol compatibility of DLM/Corosync across versions

2016-12-20 Thread David Teigland
On Tue, Dec 20, 2016 at 02:56:51AM +0100, Jean-Marc Saffroy wrote: > Ah, so that could be a serious problem for me. I hoped to be able to use > dlm across distributions without having to qualify each possible > combination... > > > Within the context of one distribution things shouldn't break

Re: [Linux-cluster] DLM user API for blocking AST

2016-12-20 Thread David Teigland
> > As a workaround to avoid an unwanted bast callback after a completion, I > > wonder if you could make a no-op call with NULL astaddr/astarg to prevent > > any further callback using those? > > I assume that what you call a no-op is a lock conversion towards the same > mode as before,

Re: [Linux-cluster] Protocol compatibility of DLM/Corosync across versions

2016-12-20 Thread Jean-Marc Saffroy
Thanks Jan for these details! Cheers, JM -- saff...@gmail.com -- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster

Re: [Linux-cluster] Protocol compatibility of DLM/Corosync across versions

2016-12-20 Thread Jan Friesse
Hi (again), Another question I have regarding DLM and Corosync (because Corosync is required to use DLM): should I expect compatibility across versions? I will add just Corosync information. Corosync with same major version is compatible (ie 2.3.0 works with 2.4.0) but if major version

Re: [Linux-cluster] Protocol compatibility of DLM/Corosync across versions

2016-12-19 Thread Jean-Marc Saffroy
On Mon, 19 Dec 2016, David Teigland wrote: > On Sun, Dec 18, 2016 at 08:46:25PM +0100, Jean-Marc Saffroy wrote: > > Hi (again), > > > > Another question I have regarding DLM and Corosync (because Corosync is > > required to use DLM): should I expect compatibility across versions? > > > > I did

Re: [Linux-cluster] DLM user API for blocking AST

2016-12-19 Thread Jean-Marc Saffroy
On Mon, 19 Dec 2016, David Teigland wrote: > As a workaround to avoid an unwanted bast callback after a completion, I > wonder if you could make a no-op call with NULL astaddr/astarg to prevent > any further callback using those? I assume that what you call a no-op is a lock conversion towards

Re: [Linux-cluster] Protocol compatibility of DLM/Corosync across versions

2016-12-19 Thread David Teigland
On Sun, Dec 18, 2016 at 08:46:25PM +0100, Jean-Marc Saffroy wrote: > Hi (again), > > Another question I have regarding DLM and Corosync (because Corosync is > required to use DLM): should I expect compatibility across versions? > > I did a quick test between distributions running different

Re: [Linux-cluster] DLM user API for blocking AST

2016-12-19 Thread David Teigland
On Sun, Dec 18, 2016 at 08:42:49PM +0100, Jean-Marc Saffroy wrote: > Hi, > > Continuing with my experiments with the DLM user API, I am trying to use > blocking AST callbacks, and find that the rules for the lifetime and > ownership of the dlm_lksb struct are a bit surprising. This led me to

[Linux-cluster] Protocol compatibility of DLM/Corosync across versions

2016-12-18 Thread Jean-Marc Saffroy
Hi (again), Another question I have regarding DLM and Corosync (because Corosync is required to use DLM): should I expect compatibility across versions? I did a quick test between distributions running different kernels (CentOS 6, Centos7 and Ubuntu 14) but rather close versions of Corosync,

[Linux-cluster] DLM user API for blocking AST

2016-12-18 Thread Jean-Marc Saffroy
Hi, Continuing with my experiments with the DLM user API, I am trying to use blocking AST callbacks, and find that the rules for the lifetime and ownership of the dlm_lksb struct are a bit surprising. This led me to some investigations, and the question at the end of this email. It looks like

[Linux-cluster] Status of Git DLM

2016-12-17 Thread Ferenc Wágner
Hi David, Is the current DLM HEAD (d5d7b8dd) stable enough for packaging? If so, could you please tag and release it? -- Thanks, Feri -- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster

Re: [Linux-cluster] DLM user API for lock value block

2016-12-06 Thread Jean-Marc Saffroy
On Tue, 6 Dec 2016, David Teigland wrote: > I haven't looked at your test to check if you're actually seeing this bug, > but you'll want this fix in any case: > > http://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git/commit/fs/dlm/user.c?id=b96f465035f9fae83c1d8de3e80eecfe6877608c

[Linux-cluster] DLM user API for lock value block

2016-12-06 Thread Jean-Marc Saffroy
Hi, I am trying to use the DLM userland API (libdlm3), and while I was able to do plain lock acquisitions and conversions, I am stuck trying to update and then read the lock value block. Does anyone have working examples of this? I did look at the rhdlmbook doc, but couldn't fine one.

[Linux-cluster] Implementing Active/Passive RHEL cluster with shared storage on RHEL 7

2016-09-21 Thread Amjad Syed
Hello, We have implemented Active/passive (2 node cluster) with shared storage using RHEL 6.5. However now we would like to implement same in RHEL 7.x. Are there any documents which will show how to implement the above. thanks Amjad -- Linux-cluster mailing list Linux-cluster@redhat.com

[Linux-cluster] problem with virbro, virbr-nic

2016-09-12 Thread Andrew Kerber
I have gfs2 shared file system configured and I am using fench-virsh as the fencing system on RHEL 7. On virtuals. One of our technicians made a mistake and destroyed on of the VM's, which re-created by cloning the second node of the cluster. I reconfigured everything by removing the cluster

[Linux-cluster] Scrpit Resource with Separate Start and Stop file

2016-09-01 Thread jasim.alam
Hi, Normally in RHEL 6, I use script resource like following But now I have to add a service which don't have single control file. Rather it is as following Star_Service.sh Stop_Service.sh Is there any way I can add this service as script resource ? Additionally, the

Re: [Linux-cluster] Two node failover with shared storage and Luns in active/passive

2016-08-29 Thread Debjyoti Banerjee
Hi Amjad, You should try glusterfs. Regards, DJ On Aug 29, 2016 6:59 PM, "Amjad Syed" wrote: > Hello, > We have implemented a two node failover cluster using RHEL pcs in active > passive mode > We have shared storage with one LUN, > Our concern is what if this LUN fails or

Re: [Linux-cluster] Two node failover with shared storage and Luns in active/passive

2016-08-29 Thread Rick Stevens
On 08/29/2016 06:23 AM, Amjad Syed wrote: > Hello, > We have implemented a two node failover cluster using RHEL pcs in > active passive mode > We have shared storage with one LUN, > Our concern is what if this LUN fails or there is hardware problem > accessing the data in it. > What is the best

[Linux-cluster] Two node failover with shared storage and Luns in active/passive

2016-08-29 Thread Amjad Syed
Hello, We have implemented a two node failover cluster using RHEL pcs in active passive mode We have shared storage with one LUN, Our concern is what if this LUN fails or there is hardware problem accessing the data in it. What is the best way to synchronize data between two LUNS ? any

Re: [Linux-cluster] unable to start mysql as a clustered service, OK stand-alone

2016-08-08 Thread Digimer
Please ask again on the Clusterlabs - Users list. This list is (quite) deprecated now. http://clusterlabs.org/mailman/listinfo/users digimer On 08/08/16 06:40 PM, berg...@merctech.com wrote: > I've got a 3-node CentOS6 cluster and I'm trying to add mysql 5.1 as a new > service. Other cluster >

[Linux-cluster] unable to start mysql as a clustered service, OK stand-alone

2016-08-08 Thread bergman
I've got a 3-node CentOS6 cluster and I'm trying to add mysql 5.1 as a new service. Other cluster services (IP addresses, Postgresql, applications) work fine. The mysql config file and data files are located on shared, cluster-wide storage (GPFS). On each node, I can successfully start mysql

[Linux-cluster] les commérages

2016-07-02 Thread bunkertor
Salut, Avez-vous entendu les derniers potins? Voici ce que vous devez lire Bisouxx, bunker...@tiscali.it-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster

[Linux-cluster] gfs2-utils 3.1.9 released

2016-06-07 Thread Andrew Price
Hi all, I am happy to announce the 3.1.9 release of gfs2-utils. This release includes the following notable changes: * fsck.gfs2 now uses less memory * Improvements and fixes to fsck.gfs2's xattr and resource group checking * mkfs.gfs2 reports progress so that you can tell it's still

Re: [Linux-cluster] Fencing Question

2016-06-06 Thread Andrew Kerber
I think you have identified the information I am missing. Is there documentation on configuring virsh to manage VMware workstation VMs? For all of my research, I have not seen such documentation. Sent from my iPad > On Jun 6, 2016, at 5:05 PM, emmanuel segura wrote: > >

Re: [Linux-cluster] Fencing Question

2016-06-06 Thread Digimer
On 06/06/16 05:37 PM, Andrew Kerber wrote: > I am doing some experimentation with Linux clustering, and still fairly > new on it. I have built a cluster as a proof of concept running a > PostgreSQL 9.5 database on gfs2 using VMware workstation 12.0 and > RHEL7. GFS2 requires a fencing resource,

Re: [Linux-cluster] Fencing Question

2016-06-06 Thread emmanuel segura
do you configured virsh to manager your vmware vms? fence_virsh is an I/O Fencing agent which can be used with the virtual machines managed by libvirt. It logs via ssh to a dom0 and there run virsh command, which does all work. By default, virsh needs root account to do properly work. So you

[Linux-cluster] Fencing Question

2016-06-06 Thread Andrew Kerber
I am doing some experimentation with Linux clustering, and still fairly new on it. I have built a cluster as a proof of concept running a PostgreSQL 9.5 database on gfs2 using VMware workstation 12.0 and RHEL7. GFS2 requires a fencing resource, which I have managed to create using fence_virsh.

[Linux-cluster] Σχετ: How to add an unimportant resource to an important cluster?

2016-05-24 Thread Θεοφάνης Κοντογιάννης
Hi, I believe that what you need is the rgmanager's __independent_subtree tag. Check here: https://fedorahosted.org/cluster/wiki/ResourceTrees Specifically chech the section called "Independent Subtrees" Hope it helpsRegards,Theophanis Kontogiannis Στις 12:22 μ.μ. Τρίτη, 24 Μαΐου 2016,

Re: [Linux-cluster] How to add an unimportant resource to an important cluster?

2016-05-24 Thread emmanuel segura
You need to use __independent_subtree if the resource that you want to use is in the same group of the important resource 2016-05-24 10:45 GMT+02:00 Eivind Olsen : > I have a cluster, running RHEL 6.7, with Ricci, Luci, rgmanager etc. > This is a 2 node cluster, where services

[Linux-cluster] How to add an unimportant resource to an important cluster?

2016-05-24 Thread Eivind Olsen
I have a cluster, running RHEL 6.7, with Ricci, Luci, rgmanager etc. This is a 2 node cluster, where services are running on one node. The service is an Oracle database, and the cluster controls several resources: * LVM volume (using clvmd) * file system on a logical volume * IP address *

Re: [Linux-cluster] I/O to gfs2 hanging or not hanging after heartbeat loss

2016-04-18 Thread Jonathan Davies
On 15/04/16 17:14, David Teigland wrote: However, on some occasions, I observe that node A continues in the loop believing that it is successfully writing to the file node A has the exclusive lock, so it continues writing... but, according to node C, the file stops being updated.

Re: [Linux-cluster] I/O to gfs2 hanging or not hanging after heartbeat loss

2016-04-15 Thread David Teigland
> > However, on some occasions, I observe that node A continues in the loop > > believing that it is successfully writing to the file node A has the exclusive lock, so it continues writing... > > but, according to > > node C, the file stops being updated. (Meanwhile, the file written by > > node

Re: [Linux-cluster] I/O to gfs2 hanging or not hanging after heartbeat loss

2016-04-15 Thread Bob Peterson
- Original Message - > Dear linux-cluster, > > I have made some observations about the behaviour of gfs2 and would > appreciate confirmation of whether this is expected behaviour or > something has gone wrong. > > I have a three-node cluster -- let's call the nodes A, B and C. On each >

Re: [Linux-cluster] Help with corosync and GFS2 on multi network setup

2016-04-12 Thread Christine Caulfield
On 12/04/16 15:02, Stefano Panella wrote: > Hi Christine, > > thanks for your input. I have checked and in the configuration with only one > network I have debugging turned on as well (same corosync.conf files). > > These messages are repeating every 1-2 seconds and the reason why I think >

Re: [Linux-cluster] Help with corosync and GFS2 on multi network setup

2016-04-12 Thread Stefano Panella
Hi Christine, thanks for your input. I have checked and in the configuration with only one network I have debugging turned on as well (same corosync.conf files). These messages are repeating every 1-2 seconds and the reason why I think there is something wrong is that if I do operation on a

  1   2   3   4   5   6   7   8   9   10   >