Hi Honza, Just checking if you have the official patch available for this issue.
As far as I am concerned barring couple of issues, everything seems to be working fine on our big-endian system. Much relieved. :) -Thanks Nikhil On Thu, May 5, 2016 at 3:24 PM, Nikhil Utane <nikhil.subscri...@gmail.com> wrote: > It worked for me. :) > I'll wait for your formal patch but until then I am able to proceed > further. (Don't know if I'll run into something else) > > However now encountering issue in pacemaker. > > May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit: > The cib process (15224) can no longer be respawned, shutting the cluster > down. > May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit: > The stonith-ng process (15225) can no longer be respawned, shutting the > cluster down. > May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit: > The lrmd process (15226) can no longer be respawned, shutting the cluster > down. > May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit: > The crmd process (15229) can no longer be respawned, shutting the cluster > down. > May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit: > The pengine process (15228) can no longer be respawned, shutting the > cluster down. > May 05 09:35:53 [15184] airv_cu pacemakerd: warning: pcmk_child_exit: > The attrd process (15227) can no longer be respawned, shutting the cluster > down. > > Looking into it. > > -Thanks > Nikhil > > On Thu, May 5, 2016 at 2:58 PM, Jan Friesse <jfrie...@redhat.com> wrote: > >> Nikhil >> >> Found the root-cause. >>> In file schedwrk.c, the function handle2void() uses a union which was not >>> initialized. >>> Because of that the handle value was computed incorrectly (lower half was >>> garbage). >>> >>> 56 static hdb_handle_t >>> 57 void2handle (const void *v) { union u u={}; u.v = v; return u.h; } >>> 58 static const void * >>> 59 handle2void (hdb_handle_t h) { union u u={}; u.h = h; return u.v; } >>> >>> After initializing (as highlighted), the corosync initialization seems to >>> be going through fine. Will check other things. >>> >> >> Your patch is incorrect and actually doesn't work. As I said (when >> pointing you to schedwrk.c), I will send you proper patch, but fix that >> issue correctly is not easy. >> >> Regards, >> Honza >> >> >>> -Regards >>> Nikhil >>> >>> On Tue, May 3, 2016 at 7:04 PM, Nikhil Utane < >>> nikhil.subscri...@gmail.com> >>> wrote: >>> >>> Thanks for your response Dejan. >>>> >>>> I do not know yet whether this has anything to do with endianness. >>>> FWIW, there could be something quirky with the system so keeping all >>>> options open. :) >>>> >>>> I added some debug prints to understand what's happening under the hood. >>>> >>>> *Success case: (on x86 machine): * >>>> [TOTEM ] entering OPERATIONAL state. >>>> [TOTEM ] A new membership (10.206.1.7:137220) was formed. Members >>>> joined: >>>> 181272839 >>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0, >>>> my_high_delivered=0 >>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1, >>>> my_high_delivered=0 >>>> [TOTEM ] Delivering 0 to 1 >>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue >>>> [SYNC ] Nikhil: Inside sync_deliver_fn. header->id=1 >>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=2, >>>> my_high_delivered=1 >>>> [TOTEM ] Delivering 1 to 2 >>>> [TOTEM ] Delivering MCAST message with seq 2 to pending delivery queue >>>> [SYNC ] Nikhil: Inside sync_deliver_fn. header->id=0 >>>> [SYNC ] Nikhil: Entering sync_barrier_handler >>>> [SYNC ] Committing synchronization for corosync configuration map >>>> access >>>> . >>>> [TOTEM ] Delivering 2 to 4 >>>> [TOTEM ] Delivering MCAST message with seq 3 to pending delivery queue >>>> [TOTEM ] Delivering MCAST message with seq 4 to pending delivery queue >>>> [CPG ] comparing: sender r(0) ip(10.206.1.7) ; members(old:0 left:0) >>>> [CPG ] chosen downlist: sender r(0) ip(10.206.1.7) ; members(old:0 >>>> left:0) >>>> [SYNC ] Committing synchronization for corosync cluster closed process >>>> group service v1.01 >>>> *[MAIN ] Completed service synchronization, ready to provide service.* >>>> >>>> >>>> *Failure case: (on ppc)*: >>>> >>>> [TOTEM ] entering OPERATIONAL state. >>>> [TOTEM ] A new membership (10.207.24.101:16) was formed. Members >>>> joined: >>>> 181344357 >>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=0, >>>> my_high_delivered=0 >>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1, >>>> my_high_delivered=0 >>>> [TOTEM ] Delivering 0 to 1 >>>> [TOTEM ] Delivering MCAST message with seq 1 to pending delivery queue >>>> [SYNC ] Nikhil: Inside sync_deliver_fn header->id=1 >>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1, >>>> my_high_delivered=1 >>>> [TOTEM ] Nikhil: Inside messages_deliver_to_app. end_point=1, >>>> my_high_delivered=1 >>>> Above message repeats continuously. >>>> >>>> So it appears that in failure case I do not receive messages with >>>> sequence >>>> number 2-4. >>>> If somebody can throw some ideas that'll help a lot. >>>> >>>> -Thanks >>>> Nikhil >>>> >>>> On Tue, May 3, 2016 at 5:26 PM, Dejan Muhamedagic <deja...@fastmail.fm> >>>> wrote: >>>> >>>> Hi, >>>>> >>>>> On Mon, May 02, 2016 at 08:54:09AM +0200, Jan Friesse wrote: >>>>> >>>>>> As your hardware is probably capable of running ppcle and if you have >>>>>>> >>>>>> an >>>>> >>>>>> environment >>>>>>> at hand without too much effort it might pay off to try that. >>>>>>> There are of course distributions out there support corosync on >>>>>>> big-endian architectures >>>>>>> but I don't know if there is an automatized regression for corosync >>>>>>> on >>>>>>> big-endian that >>>>>>> would catch big-endian-issues right away with something as current as >>>>>>> your 2.3.5. >>>>>>> >>>>>> >>>>>> No we are not testing big-endian. >>>>>> >>>>>> So totally agree with Klaus. Give a try to ppcle. Also make sure all >>>>>> nodes are little-endian. Corosync should work in mixed BE/LE >>>>>> environment but because it's not tested, it may not work (and it's a >>>>>> bug, so if ppcle works I will try to fix BE). >>>>>> >>>>> >>>>> I tested a cluster consisting of big endian/little endian nodes >>>>> (s390 and x86-64), but that was a while ago. IIRC, all relevant >>>>> bugs in corosync got fixed at that time. Don't know what is the >>>>> situation with the latest version. >>>>> >>>>> Thanks, >>>>> >>>>> Dejan >>>>> >>>>> Regards, >>>>>> Honza >>>>>> >>>>>> >>>>>>> Regards, >>>>>>> Klaus >>>>>>> >>>>>>> On 05/02/2016 06:44 AM, Nikhil Utane wrote: >>>>>>> >>>>>>>> Re-sending as I don't see my post on the thread. >>>>>>>> >>>>>>>> On Sun, May 1, 2016 at 4:21 PM, Nikhil Utane >>>>>>>> <nikhil.subscri...@gmail.com <mailto:nikhil.subscri...@gmail.com>> >>>>>>>> >>>>>>> wrote: >>>>> >>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> Looking for some guidance here as we are completely blocked >>>>>>>> otherwise :(. >>>>>>>> >>>>>>>> -Regards >>>>>>>> Nikhil >>>>>>>> >>>>>>>> On Fri, Apr 29, 2016 at 6:11 PM, Sriram <sriram...@gmail.com >>>>>>>> <mailto:sriram...@gmail.com>> wrote: >>>>>>>> >>>>>>>> Corrected the subject. >>>>>>>> >>>>>>>> We went ahead and captured corosync debug logs for our ppc >>>>>>>> >>>>>>> board. >>>>> >>>>>> After log analysis and comparison with the sucessful logs( >>>>>>>> from x86 machine) , >>>>>>>> we didnt find *"[ MAIN ] Completed service >>>>>>>> synchronization, >>>>>>>> ready to provide service.*" in ppc logs. >>>>>>>> So, looks like corosync is not in a position to accept >>>>>>>> connection from Pacemaker. >>>>>>>> Even I tried with the new corosync.conf with no success. >>>>>>>> >>>>>>>> Any hints on this issue would be really helpful. >>>>>>>> >>>>>>>> Attaching ppc_notworking.log, x86_working.log, >>>>>>>> corosync.conf. >>>>>>>> >>>>>>>> Regards, >>>>>>>> Sriram >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Fri, Apr 29, 2016 at 2:44 PM, Sriram < >>>>>>>> sriram...@gmail.com >>>>>>>> <mailto:sriram...@gmail.com>> wrote: >>>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> I went ahead and made some changes in file system(Like >>>>>>>> I >>>>>>>> brought in /etc/init.d/corosync and >>>>>>>> >>>>>>> /etc/init.d/pacemaker, >>>>> >>>>>> /etc/sysconfig ), After that I was able to run "pcs >>>>>>>> cluster start". >>>>>>>> But it failed with the following error >>>>>>>> # pcs cluster start >>>>>>>> Starting Cluster... >>>>>>>> Starting Pacemaker Cluster Manager[FAILED] >>>>>>>> Error: unable to start pacemaker >>>>>>>> >>>>>>>> And in the /var/log/pacemaker.log, I saw these errors >>>>>>>> pacemakerd: info: mcp_read_config: cmap connection >>>>>>>> setup failed: CS_ERR_TRY_AGAIN. Retrying in 4s >>>>>>>> Apr 29 08:53:47 [15863] node_cu pacemakerd: info: >>>>>>>> mcp_read_config: cmap connection setup failed: >>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 5s >>>>>>>> Apr 29 08:53:52 [15863] node_cu pacemakerd: warning: >>>>>>>> mcp_read_config: Could not connect to Cluster >>>>>>>> Configuration Database API, error 6 >>>>>>>> Apr 29 08:53:52 [15863] node_cu pacemakerd: notice: >>>>>>>> main: Could not obtain corosync config data, >>>>>>>> exiting >>>>>>>> Apr 29 08:53:52 [15863] node_cu pacemakerd: info: >>>>>>>> crm_xml_cleanup: Cleaning up memory from libxml2 >>>>>>>> >>>>>>>> >>>>>>>> And in the /var/log/Debuglog, I saw these errors coming >>>>>>>> from corosync >>>>>>>> 20160429 085347.487050 <tel:085347.487050> airv_cu >>>>>>>> daemon.warn corosync[12857]: [QB ] Denied >>>>>>>> >>>>>>> connection, >>>>> >>>>>> is not ready (12857-15863-14) >>>>>>>> 20160429 085347.487067 <tel:085347.487067> airv_cu >>>>>>>> daemon.info <http://daemon.info> corosync[12857]: >>>>>>>> [QB >>>>>>>> ] Denied connection, is not ready (12857-15863-14) >>>>>>>> >>>>>>>> >>>>>>>> I browsed the code of libqb to find that it is failing >>>>>>>> in >>>>>>>> >>>>>>>> >>>>>>>> https://github.com/ClusterLabs/libqb/blob/master/lib/ipc_setup.c >>>>> >>>>>> >>>>>>>> Line 600 : >>>>>>>> handle_new_connection function >>>>>>>> >>>>>>>> Line 637: >>>>>>>> if (auth_result == 0 && >>>>>>>> c->service->serv_fns.connection_accept) { >>>>>>>> res = c->service->serv_fns.connection_accept(c, >>>>>>>> c->euid, c->egid); >>>>>>>> } >>>>>>>> if (res != 0) { >>>>>>>> goto send_response; >>>>>>>> } >>>>>>>> >>>>>>>> Any hints on this issue would be really helpful for me >>>>>>>> to >>>>>>>> go ahead. >>>>>>>> Please let me know if any logs are required, >>>>>>>> >>>>>>>> Regards, >>>>>>>> Sriram >>>>>>>> >>>>>>>> On Thu, Apr 28, 2016 at 2:42 PM, Sriram >>>>>>>> <sriram...@gmail.com <mailto:sriram...@gmail.com>> >>>>>>>> >>>>>>> wrote: >>>>> >>>>>> >>>>>>>> Thanks Ken and Emmanuel. >>>>>>>> Its a big endian machine. I will try with running >>>>>>>> >>>>>>> "pcs >>>>> >>>>>> cluster setup" and "pcs cluster start" >>>>>>>> Inside cluster.py, "service pacemaker start" and >>>>>>>> "service corosync start" are executed to bring up >>>>>>>> pacemaker and corosync. >>>>>>>> Those service scripts and the infrastructure needed >>>>>>>> >>>>>>> to >>>>> >>>>>> bring up the processes in the above said manner >>>>>>>> doesn't exist in my board. >>>>>>>> As it is a embedded board with the limited memory, >>>>>>>> full fledged linux is not installed. >>>>>>>> Just curious to know, what could be reason the >>>>>>>> pacemaker throws that error. >>>>>>>> >>>>>>>> /"cmap connection setup failed: CS_ERR_TRY_AGAIN. >>>>>>>> Retrying in 1s" >>>>>>>> >>>>>>>> / >>>>>>>> Thanks for response. >>>>>>>> >>>>>>>> Regards, >>>>>>>> Sriram. >>>>>>>> >>>>>>>> On Thu, Apr 28, 2016 at 8:55 AM, Ken Gaillot >>>>>>>> <kgail...@redhat.com <mailto:kgail...@redhat.com>> >>>>>>>> >>>>>>> wrote: >>>>> >>>>>> >>>>>>>> On 04/27/2016 11:25 AM, emmanuel segura wrote: >>>>>>>> > you need to use pcs to do everything, pcs >>>>>>>> cluster setup and pcs >>>>>>>> > cluster start, try to use the redhat docs for >>>>>>>> more information. >>>>>>>> >>>>>>>> Agreed -- pcs cluster setup will create a >>>>>>>> proper >>>>>>>> corosync.conf for you. >>>>>>>> Your corosync.conf below uses corosync 1 >>>>>>>> syntax, >>>>>>>> and there were >>>>>>>> significant changes in corosync 2. In >>>>>>>> particular, >>>>>>>> you don't need the >>>>>>>> file created in step 4, because pacemaker is no >>>>>>>> longer launched via a >>>>>>>> corosync plugin. >>>>>>>> >>>>>>>> > 2016-04-27 17:28 GMT+02:00 Sriram >>>>>>>> <sriram...@gmail.com <mailto: >>>>>>>> sriram...@gmail.com >>>>>>>> >>>>>>> : >>>>>>> >>>>>>>> >> Dear All, >>>>>>>> >> >>>>>>>> >> I m trying to use pacemaker and corosync for >>>>>>>> the clustering requirement that >>>>>>>> >> came up recently. >>>>>>>> >> We have cross compiled corosync, pacemaker >>>>>>>> and >>>>>>>> pcs(python) for ppc >>>>>>>> >> environment (Target board where pacemaker >>>>>>>> and >>>>>>>> corosync are supposed to run) >>>>>>>> >> I m having trouble bringing up pacemaker in >>>>>>>> that environment, though I could >>>>>>>> >> successfully bring up corosync. >>>>>>>> >> Any help is welcome. >>>>>>>> >> >>>>>>>> >> I m using these versions of pacemaker and >>>>>>>> >>>>>>> corosync >>>>> >>>>>> >> [root@node_cu pacemaker]# corosync -v >>>>>>>> >> Corosync Cluster Engine, version '2.3.5' >>>>>>>> >> Copyright (c) 2006-2009 Red Hat, Inc. >>>>>>>> >> [root@node_cu pacemaker]# pacemakerd -$ >>>>>>>> >> Pacemaker 1.1.14 >>>>>>>> >> Written by Andrew Beekhof >>>>>>>> >> >>>>>>>> >> For running corosync, I did the following. >>>>>>>> >> 1. Created the following directories, >>>>>>>> >> /var/lib/pacemaker >>>>>>>> >> /var/lib/corosync >>>>>>>> >> /var/lib/pacemaker >>>>>>>> >> /var/lib/pacemaker/cores >>>>>>>> >> /var/lib/pacemaker/pengine >>>>>>>> >> /var/lib/pacemaker/blackbox >>>>>>>> >> /var/lib/pacemaker/cib >>>>>>>> >> >>>>>>>> >> >>>>>>>> >> 2. Created a file called corosync.conf under >>>>>>>> /etc/corosync folder with the >>>>>>>> >> following contents >>>>>>>> >> >>>>>>>> >> totem { >>>>>>>> >> >>>>>>>> >> version: 2 >>>>>>>> >> token: 5000 >>>>>>>> >> token_retransmits_before_loss_const: >>>>>>>> >>>>>>> 20 >>>>> >>>>>> >> join: 1000 >>>>>>>> >> consensus: 7500 >>>>>>>> >> vsftype: none >>>>>>>> >> max_messages: 20 >>>>>>>> >> secauth: off >>>>>>>> >> cluster_name: mycluster >>>>>>>> >> transport: udpu >>>>>>>> >> threads: 0 >>>>>>>> >> clear_node_high_bit: yes >>>>>>>> >> >>>>>>>> >> interface { >>>>>>>> >> ringnumber: 0 >>>>>>>> >> # The following three values >>>>>>>> need to be set based on your >>>>>>>> >> environment >>>>>>>> >> bindnetaddr: 10.x.x.x >>>>>>>> >> mcastaddr: 226.94.1.1 >>>>>>>> >> mcastport: 5405 >>>>>>>> >> } >>>>>>>> >> } >>>>>>>> >> >>>>>>>> >> logging { >>>>>>>> >> fileline: off >>>>>>>> >> to_syslog: yes >>>>>>>> >> to_stderr: no >>>>>>>> >> to_syslog: yes >>>>>>>> >> logfile: /var/log/corosync.log >>>>>>>> >> syslog_facility: daemon >>>>>>>> >> debug: on >>>>>>>> >> timestamp: on >>>>>>>> >> } >>>>>>>> >> >>>>>>>> >> amf { >>>>>>>> >> mode: disabled >>>>>>>> >> } >>>>>>>> >> >>>>>>>> >> quorum { >>>>>>>> >> provider: corosync_votequorum >>>>>>>> >> } >>>>>>>> >> >>>>>>>> >> nodelist { >>>>>>>> >> node { >>>>>>>> >> ring0_addr: node_cu >>>>>>>> >> nodeid: 1 >>>>>>>> >> } >>>>>>>> >> } >>>>>>>> >> >>>>>>>> >> 3. Created authkey under /etc/corosync >>>>>>>> >> >>>>>>>> >> 4. Created a file called pcmk under >>>>>>>> /etc/corosync/service.d and contents as >>>>>>>> >> below, >>>>>>>> >> cat pcmk >>>>>>>> >> service { >>>>>>>> >> # Load the Pacemaker Cluster >>>>>>>> Resource >>>>>>>> Manager >>>>>>>> >> name: pacemaker >>>>>>>> >> ver: 1 >>>>>>>> >> } >>>>>>>> >> >>>>>>>> >> 5. Added the node name "node_cu" in >>>>>>>> /etc/hosts >>>>>>>> with 10.X.X.X ip >>>>>>>> >> >>>>>>>> >> 6. ./corosync -f -p & --> this step started >>>>>>>> corosync >>>>>>>> >> >>>>>>>> >> [root@node_cu pacemaker]# netstat -alpn | >>>>>>>> >>>>>>> grep >>>>> >>>>>> -i coros >>>>>>>> >> udp 0 0 10.X.X.X:61841 >>>>>>>> >>>>>>> 0.0.0.0:* >>>>> >>>>>> >> 9133/corosync >>>>>>>> >> udp 0 0 10.X.X.X:5405 >>>>>>>> >>>>>>> 0.0.0.0:* >>>>> >>>>>> >> 9133/corosync >>>>>>>> >> unix 2 [ ACC ] STREAM >>>>>>>> LISTENING >>>>>>>> 148888 9133/corosync >>>>>>>> >> @quorum >>>>>>>> >> unix 2 [ ACC ] STREAM >>>>>>>> LISTENING >>>>>>>> 148884 9133/corosync >>>>>>>> >> @cmap >>>>>>>> >> unix 2 [ ACC ] STREAM >>>>>>>> LISTENING >>>>>>>> 148887 9133/corosync >>>>>>>> >> @votequorum >>>>>>>> >> unix 2 [ ACC ] STREAM >>>>>>>> LISTENING >>>>>>>> 148885 9133/corosync >>>>>>>> >> @cfg >>>>>>>> >> unix 2 [ ACC ] STREAM >>>>>>>> LISTENING >>>>>>>> 148886 9133/corosync >>>>>>>> >> @cpg >>>>>>>> >> unix 2 [ ] DGRAM >>>>>>>> 148840 9133/corosync >>>>>>>> >> >>>>>>>> >> 7. ./pacemakerd -f & gives the following >>>>>>>> error >>>>>>>> and exits. >>>>>>>> >> [root@node_cu pacemaker]# pacemakerd -f >>>>>>>> >> cmap connection setup failed: >>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 1s >>>>>>>> >> cmap connection setup failed: >>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 2s >>>>>>>> >> cmap connection setup failed: >>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 3s >>>>>>>> >> cmap connection setup failed: >>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 4s >>>>>>>> >> cmap connection setup failed: >>>>>>>> CS_ERR_TRY_AGAIN. Retrying in 5s >>>>>>>> >> Could not connect to Cluster Configuration >>>>>>>> Database API, error 6 >>>>>>>> >> >>>>>>>> >> Can you please point me, what is missing in >>>>>>>> these steps ? >>>>>>>> >> >>>>>>>> >> Before trying these steps, I tried running >>>>>>>> >>>>>>> "pcs >>>>> >>>>>> cluster start", but that >>>>>>>> >> command fails with "service" script not >>>>>>>> found. >>>>>>>> As the root filesystem >>>>>>>> >> doesn't contain either /etc/init.d/ or >>>>>>>> /sbin/service >>>>>>>> >> >>>>>>>> >> So, the plan is to bring up corosync and >>>>>>>> pacemaker manually, later do the >>>>>>>> >> cluster configuration using "pcs" commands. >>>>>>>> >> >>>>>>>> >> Regards, >>>>>>>> >> Sriram >>>>>>>> >> >>>>>>>> >> >>>>>>>> >>>>>>> _______________________________________________ >>>>> >>>>>> >> Users mailing list: Users@clusterlabs.org >>>>>>>> <mailto:Users@clusterlabs.org> >>>>>>>> >> >>>>>>>> http://clusterlabs.org/mailman/listinfo/users >>>>>>>> >> >>>>>>>> >> Project Home: http://www.clusterlabs.org >>>>>>>> >> Getting started: >>>>>>>> >>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> >>>>>> >> Bugs: http://bugs.clusterlabs.org >>>>>>>> >> >>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list: Users@clusterlabs.org >>>>>>>> <mailto:Users@clusterlabs.org> >>>>>>>> http://clusterlabs.org/mailman/listinfo/users >>>>>>>> >>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>> Getting started: >>>>>>>> >>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> >>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list: Users@clusterlabs.org >>>>>>>> <mailto:Users@clusterlabs.org> >>>>>>>> http://clusterlabs.org/mailman/listinfo/users >>>>>>>> >>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>> Getting started: >>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list: Users@clusterlabs.org >>>>>>>> http://clusterlabs.org/mailman/listinfo/users >>>>>>>> >>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>> Getting started: >>>>>>>> >>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> >>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> Users mailing list: Users@clusterlabs.org >>>>>>> http://clusterlabs.org/mailman/listinfo/users >>>>>>> >>>>>>> Project Home: http://www.clusterlabs.org >>>>>>> Getting started: >>>>>>> >>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> >>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>> >>>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list: Users@clusterlabs.org >>>>>> http://clusterlabs.org/mailman/listinfo/users >>>>>> >>>>>> Project Home: http://www.clusterlabs.org >>>>>> Getting started: >>>>>> >>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> >>>>>> Bugs: http://bugs.clusterlabs.org >>>>>> >>>>> >>>>> _______________________________________________ >>>>> Users mailing list: Users@clusterlabs.org >>>>> http://clusterlabs.org/mailman/listinfo/users >>>>> >>>>> Project Home: http://www.clusterlabs.org >>>>> Getting started: >>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> Bugs: http://bugs.clusterlabs.org >>>>> >>>>> >>>> >>>> >>> >>> >>> _______________________________________________ >>> Users mailing list: Users@clusterlabs.org >>> http://clusterlabs.org/mailman/listinfo/users >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >>> >> >> _______________________________________________ >> Users mailing list: Users@clusterlabs.org >> http://clusterlabs.org/mailman/listinfo/users >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> > >
_______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org