I am having this exact problem with an up-to-date 5 node RHEL5.1 cluster using an EVA8100 for storage. The sequence is this.
- create an additional vdisk on the eva and present it to the cluster nodes - rescan the scsi bus on all nodes to pick up the new device. - add the new device to my multipath config and activate it. - run pvcreate on the new multipath device - extend the VG using the new device - extend an existing volume with the add VG extents (+100%FREE)) <BOOM> Everything works as expected until I try to extend a volume. I get the same error until I restart clvmd on all the nodes. Once the VG is extended, gfs2_grow works fine and things proceed normally. Also, i am _not_ using any non-clustered lvm on any of these machines. Standing by to see the results of your testing .. I hope it all works out as the idea of having to restart clvmd on a production cluster to extend a volume is a bit unnerving. :) On Mon, May 12, 2008 at 10:41 PM, Collins, Kevin [Beeline] <[EMAIL PROTECTED]> wrote: > FYI, I am still working the ticket with RedHat. Looks like a known bug. > I am currently testing updated versions of lvm2, lvm2-cluster and > device-mapper, which so far seems promising. > > Thanks, > > Kevin > > -----Original Message----- > From: [EMAIL PROTECTED] > [mailto:[EMAIL PROTECTED] On Behalf Of Collins, Kevin > [Beeline] > Sent: Tuesday, May 06, 2008 8:17 AM > To: Red Hat Enterprise Linux 5 (Tikanga) discussion mailing-list > Subject: RE: [rhelv5-list] Clustered LVM > > I don't know - I was hoping Nuno would respond and say what version he > was running so we could find out :) > > Kevin > > -----Original Message----- > From: [EMAIL PROTECTED] > [mailto:[EMAIL PROTECTED] On Behalf Of Zavodsky, Daniel > (GE Money) > Sent: Monday, May 05, 2008 11:31 PM > To: Red Hat Enterprise Linux 5 (Tikanga) discussion mailing-list > Subject: RE: [rhelv5-list] Clustered LVM > > Hello, > I am using the latest stable one. I cannot really use beta > channels on production machines. Is it fixed there? > > Regards, > Daniel > > > -----Original Message----- > From: [EMAIL PROTECTED] > [mailto:[EMAIL PROTECTED] On Behalf Of Collins, Kevin > [Beeline] > Sent: Monday, May 05, 2008 7:02 PM > To: Red Hat Enterprise Linux 5 (Tikanga) discussion mailing-list > Subject: RE: [rhelv5-list] Clustered LVM > > Which version of lvm2-cluster are you using? Are you subscribed to the > Beta channel for Cluster Storage? > > I am running the "standard" version, but there is a later one in the > Beta channel: > > root# rpm -q lvm2-cluster > lvm2-cluster-2.02.26-1.el5 > > Thanks, > > Kevin > > -----Original Message----- > From: [EMAIL PROTECTED] > [mailto:[EMAIL PROTECTED] On Behalf Of Nuno Fernandes > Sent: Saturday, May 03, 2008 2:30 PM > To: [email protected] > Subject: Re: [rhelv5-list] Clustered LVM > > On Monday 28 April 2008 07:57:39 Zavodsky, Daniel (GE Money) wrote: >> Hello, >> I have been experiencing exactly the same issues... and I found > the >> trick with restarting clvmd myself too. :-) It looks like the problem >> happens when you are using both non-clustered and clustered LVM. > I'm also using both clustered and non-clustered lvm on the same machine > and i don't have that problem.. Everything is working fine. > > Best regards, > Nuno Fernandes > >> I don't think you are doing anything wrong. >> >> Regards, >> Daniel >> >> >> >> _____ >> >> From: [EMAIL PROTECTED] >> [mailto:[EMAIL PROTECTED] On Behalf Of Collins, Kevin >> [Beeline] >> Sent: Friday, April 25, 2008 9:08 PM >> To: [email protected] >> Subject: [rhelv5-list] Clustered LVM >> >> >> >> Hi, >> >> I am working through the process of building my first GFS >> installation and cluster. So far, things are going well. I have shared > >> SAN disk that is visible to both nodes, I've got the basic cluster up >> and I am now starting to create my LVM structure. >> >> I spent a lot of time trying to determine why I was seeing errors >> similar to the following when trying to create LVs in a new VG: >> >> root# lvcreate -L 50M /dev/vgtest >> Rounding up size to full physical extent 52.00 MB >> Error locking on node cpafisxb: Volume group for uuid not found: >> IC070PzNGG68uEesi33dH902E4GeGEvwcAccZY4AdnJRkPHUL7EYJzL0Xxkg3eqV >> >> Error locking on node cpafisxa: Volume group for uuid not found: >> IC070PzNGG68uEesi33dH902E4GeGEvwcAccZY4AdnJRkPHUL7EYJzL0Xxkg3eqV >> >> Failed to activate new LV. >> >> I noticed that I did not have a /dev/vgtest on the node where I > created >> the VG. I discovered (by googling the error message above) that > another >> person had seen a similar problem and resolved it by restarting the >> clvmd service. I did the same thing and the problem was resolved - I >> don't see that error when creating LVs in that VG, and I now have a >> /dev/vgtest and associated files (only on the node I created it on). >> >> Now, I created a new VG, try to create a new LV within it and see the >> same error! I see in the clvmd man page that there is a "-R" option, >> which sounds like maybe what should be done instead of restarting > clvmd. >> Tried it, but no luck - same error and no /dev/ directory created. >> Restart clvmd, instant fix. >> >> So - what am I doing wrong here? I can't imagine I am supposed to >> restart clvmd every time I create a new VG, am I? >> >> Additionally, I can see the clustered VGs and LVs on my other node, > but >> there are not any files in /dev for the clustered VGs - is this > normal? >> I have also restarted clvmd there with no effect. >> >> Thanks, >> >> Kevin > > > _______________________________________________ > rhelv5-list mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/rhelv5-list > > _______________________________________________ > rhelv5-list mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/rhelv5-list > > > > _______________________________________________ > rhelv5-list mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/rhelv5-list > > _______________________________________________ > rhelv5-list mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/rhelv5-list > > _______________________________________________ > rhelv5-list mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/rhelv5-list > _______________________________________________ rhelv5-list mailing list [email protected] https://www.redhat.com/mailman/listinfo/rhelv5-list
