Hi, Can any one help me here ??
On Wed, Jan 14, 2015 at 3:25 PM, Punit Dambiwal <hypu...@gmail.com> wrote: > Hi Donny, > > I am not using gluster for the NFS mount...no volume has been created > because of those errors.... > > On Wed, Jan 14, 2015 at 9:47 AM, Donny Davis <do...@cloudspin.me> wrote: > >> And >> >> >> >> rpcbind is running >> >> >> >> can you do a regular nfs mount of the gluster volume? >> >> gluster volume info {your volume name here} >> >> >> >> >> >> Just gathering intel to hopefully provide a solution. I just deployed >> gluster with hosted engine today, and I did get some of the same errors as >> you when I was bringing everything up >> >> Did you follow a guide, or are you craving your own? >> >> Are you using swift for anything… that is usually for openstack to my >> knowledge? I guess you could use it for ovirt, but I didn’t >> >> >> >> Donny D >> >> >> >> *From:* Punit Dambiwal [mailto:hypu...@gmail.com] >> *Sent:* Tuesday, January 13, 2015 6:41 PM >> *To:* Donny Davis >> *Cc:* users@ovirt.org >> >> *Subject:* Re: [ovirt-users] Failed to create volume in OVirt with >> gluster >> >> >> >> Hi Donny, >> >> >> >> No I am not using CTDB....it's totally new deployment... >> >> >> >> On Wed, Jan 14, 2015 at 1:50 AM, Donny Davis <do...@cloudspin.me> wrote: >> >> Are you using ctdb??? and did you specify Lock=False in /etc/nfsmount.conf >> >> >> >> Can you give a full run down of topology, and has this ever been working >> or is it a new deployment? >> >> >> >> Donny D >> >> >> >> *From:* users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] *On >> Behalf Of *Punit Dambiwal >> *Sent:* Monday, January 12, 2015 7:18 PM >> *To:* Kanagaraj Mayilsamy >> *Cc:* gluster-us...@gluster.org; Kaushal M; users@ovirt.org >> *Subject:* Re: [ovirt-users] Failed to create volume in OVirt with >> gluster >> >> >> >> Hi, >> >> >> >> Please find the more details on this ....can anybody from gluster will >> help me here :- >> >> >> >> >> >> Gluster CLI Logs :- /var/log/glusterfs/cli.log >> >> >> >> [2015-01-13 02:06:23.071969] T [cli.c:264:cli_rpc_notify] 0-glusterfs: >> got RPC_CLNT_CONNECT >> >> [2015-01-13 02:06:23.072012] T [cli-quotad-client.c:94:cli_quotad_notify] >> 0-glusterfs: got RPC_CLNT_CONNECT >> >> [2015-01-13 02:06:23.072024] I [socket.c:2344:socket_event_handler] >> 0-transport: disconnecting now >> >> [2015-01-13 02:06:23.072055] T >> [cli-quotad-client.c:100:cli_quotad_notify] 0-glusterfs: got >> RPC_CLNT_DISCONNECT >> >> [2015-01-13 02:06:23.072131] T [rpc-clnt.c:1381:rpc_clnt_record] >> 0-glusterfs: Auth Info: pid: 0, uid: 0, gid: 0, owner: >> >> [2015-01-13 02:06:23.072176] T >> [rpc-clnt.c:1238:rpc_clnt_record_build_header] 0-rpc-clnt: Request fraglen >> 128, payload: 64, rpc hdr: 64 >> >> [2015-01-13 02:06:23.072572] T [socket.c:2863:socket_connect] (--> >> /usr/lib64/libglusterfs.so.0(_gf_log_callingfn+0x1e0)[0x7fed02f15420] (--> >> /usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0x7293)[0x7fed001a4293] >> (--> /usr/lib64/libgfrpc.so.0(rpc_clnt_submit+0x468)[0x7fed0266df98] (--> >> /usr/sbin/gluster(cli_submit_request+0xdb)[0x40a9bb] (--> >> /usr/sbin/gluster(cli_cmd_submit+0x8e)[0x40b7be] ))))) 0-glusterfs: connect >> () called on transport already connected >> >> [2015-01-13 02:06:23.072616] T [rpc-clnt.c:1573:rpc_clnt_submit] >> 0-rpc-clnt: submitted request (XID: 0x1 Program: Gluster CLI, ProgVers: 2, >> Proc: 27) to rpc-transport (glusterfs) >> >> [2015-01-13 02:06:23.072633] D [rpc-clnt-ping.c:231:rpc_clnt_start_ping] >> 0-glusterfs: ping timeout is 0, returning >> >> [2015-01-13 02:06:23.075930] T [rpc-clnt.c:660:rpc_clnt_reply_init] >> 0-glusterfs: received rpc message (RPC XID: 0x1 Program: Gluster CLI, >> ProgVers: 2, Proc: 27) from rpc-transport (glusterfs) >> >> [2015-01-13 02:06:23.075976] D [cli-rpc-ops.c:6548:gf_cli_status_cbk] >> 0-cli: Received response to status cmd >> >> [2015-01-13 02:06:23.076025] D [cli-cmd.c:384:cli_cmd_submit] 0-cli: >> Returning 0 >> >> [2015-01-13 02:06:23.076049] D [cli-rpc-ops.c:6811:gf_cli_status_volume] >> 0-cli: Returning: 0 >> >> [2015-01-13 02:06:23.076192] D [cli-xml-output.c:84:cli_begin_xml_output] >> 0-cli: Returning 0 >> >> [2015-01-13 02:06:23.076244] D >> [cli-xml-output.c:131:cli_xml_output_common] 0-cli: Returning 0 >> >> [2015-01-13 02:06:23.076256] D >> [cli-xml-output.c:1375:cli_xml_output_vol_status_begin] 0-cli: Returning 0 >> >> [2015-01-13 02:06:23.076437] D [cli-xml-output.c:108:cli_end_xml_output] >> 0-cli: Returning 0 >> >> [2015-01-13 02:06:23.076459] D >> [cli-xml-output.c:1398:cli_xml_output_vol_status_end] 0-cli: Returning 0 >> >> [2015-01-13 02:06:23.076490] I [input.c:36:cli_batch] 0-: Exiting with: 0 >> >> >> >> Command log :- /var/log/glusterfs/.cmd_log_history >> >> >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:10:35.836676] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:16:25.956514] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:17:36.977833] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:21:07.048053] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:26:57.168661] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:28:07.194428] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:30:27.256667] : volume status vol01 : FAILED : Locking >> failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. >> >> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for >> details. >> >> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for >> details. >> >> [2015-01-13 01:34:58.350748] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:36:08.375326] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:36:08.386470] : volume status vol01 : FAILED : Locking >> failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. >> >> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for >> details. >> >> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for >> details. >> >> [2015-01-13 01:42:59.524215] : volume stop vol01 : FAILED : Locking >> failed on cpu02.zne01.hkg1.stack.com. Please check log file for details. >> >> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file for >> details. >> >> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file for >> details. >> >> [2015-01-13 01:45:10.550659] : volume status all tasks : FAILED : >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> Staging failed on 00000000-0000-0000-0000-000000000000. Please check log >> file for details. >> >> [2015-01-13 01:46:10.656802] : volume status all tasks : SUCCESS >> >> [2015-01-13 01:51:02.796031] : volume status all tasks : SUCCESS >> >> [2015-01-13 01:52:02.897804] : volume status all tasks : SUCCESS >> >> [2015-01-13 01:55:25.841070] : system:: uuid get : SUCCESS >> >> [2015-01-13 01:55:26.752084] : system:: uuid get : SUCCESS >> >> [2015-01-13 01:55:32.499049] : system:: uuid get : SUCCESS >> >> [2015-01-13 01:55:38.716907] : system:: uuid get : SUCCESS >> >> [2015-01-13 01:56:52.905899] : volume status all tasks : SUCCESS >> >> [2015-01-13 01:58:53.109613] : volume status all tasks : SUCCESS >> >> [2015-01-13 02:03:26.769430] : system:: uuid get : SUCCESS >> >> [2015-01-13 02:04:22.859213] : volume status all tasks : SUCCESS >> >> [2015-01-13 02:05:22.970393] : volume status all tasks : SUCCESS >> >> [2015-01-13 02:06:23.075823] : volume status all tasks : SUCCESS >> >> >> >> >> >> On Mon, Jan 12, 2015 at 10:53 PM, Kanagaraj Mayilsamy < >> kmayi...@redhat.com> wrote: >> >> I can see the failures in glusterd log. >> >> Can someone from glusterfs dev pls help on this? >> >> Thanks, >> Kanagaraj >> >> >> ----- Original Message ----- >> > From: "Punit Dambiwal" <hypu...@gmail.com> >> > To: "Kanagaraj" <kmayi...@redhat.com> >> > Cc: "Martin Pavlík" <mpav...@redhat.com>, "Vijay Bellur" < >> vbel...@redhat.com>, "Kaushal M" <kshlms...@gmail.com>, >> > users@ovirt.org, gluster-us...@gluster.org >> > Sent: Monday, January 12, 2015 3:36:43 PM >> > Subject: Re: Failed to create volume in OVirt with gluster >> > >> > Hi Kanagaraj, >> > >> > Please find the logs from here :- http://ur1.ca/jeszc >> > >> > [image: Inline image 1] >> > >> > [image: Inline image 2] >> > >> > On Mon, Jan 12, 2015 at 1:02 PM, Kanagaraj <kmayi...@redhat.com> wrote: >> > >> > > Looks like there are some failures in gluster. >> > > Can you send the log output from glusterd log file from the relevant >> hosts? >> > > >> > > Thanks, >> > > Kanagaraj >> > > >> > > >> > > On 01/12/2015 10:24 AM, Punit Dambiwal wrote: >> > > >> > > Hi, >> > > >> > > Is there any one from gluster can help me here :- >> > > >> > > Engine logs :- >> > > >> > > 2015-01-12 12:50:33,841 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:34,725 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:36,824 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:36,853 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:36,866 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:37,751 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:39,849 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:39,878 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:39,890 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:40,776 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:42,878 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:42,903 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:42,916 INFO >> > > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > > (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait >> lock >> > > EngineLock [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 >> > > value: GLUSTER >> > > , sharedLocks= ] >> > > 2015-01-12 12:50:43,771 INFO >> > > >> [org.ovirt.engine.core.vdsbroker.gluster.CreateGlusterVolumeVDSCommand] >> > > (ajp--127.0.0.1-8702-1) [330ace48] FINISH, >> CreateGlusterVolumeVDSCommand, >> > > log id: 303e70a4 >> > > 2015-01-12 12:50:43,780 ERROR >> > > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >> > > (ajp--127.0.0.1-8702-1) [330ace48] Correlation ID: 330ace48, Job ID: >> > > 896a69b3-a678-40a7-bceb-3635e4062aa0, Call Stack: null, Custom Event >> ID: >> > > -1, Message: Creation of Gluster Volume vol01 failed. >> > > 2015-01-12 12:50:43,785 INFO >> > > [org.ovirt.engine.core.bll.gluster.CreateGlusterVolumeCommand] >> > > (ajp--127.0.0.1-8702-1) [330ace48] Lock freed to object EngineLock >> > > [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 value: >> GLUSTER >> > > , sharedLocks= ] >> > > >> > > [image: Inline image 2] >> > > >> > > >> > > On Sun, Jan 11, 2015 at 6:48 PM, Martin Pavlík <mpav...@redhat.com> >> wrote: >> > > >> > >> Hi Punit, >> > >> >> > >> unfortunately I’am not that good with the gluster, I was just >> following >> > >> the obvious clue from the log. Could you try on the nodes if the >> packages >> > >> are even available for installation >> > >> >> > >> yum install gluster-swift gluster-swift-object gluster-swift-plugin >> > >> gluster-swift-account >> > >> gluster-swift-proxy gluster-swift-doc gluster-swift-container >> > >> glusterfs-geo-replication >> > >> >> > >> if not you could try to get them in official gluster repo. >> > >> >> http://download.gluster.org/pub/gluster/glusterfs/LATEST/CentOS/glusterfs-epel.repo >> > >> >> > >> HTH >> > >> >> > >> M. >> > >> >> > >> >> > >> >> > >> >> > >> On 10 Jan 2015, at 04:35, Punit Dambiwal <hypu...@gmail.com> >> wrote: >> > >> >> > >> Hi Martin, >> > >> >> > >> I installed gluster from ovirt repo....is it require to install >> those >> > >> packages manually ?? >> > >> >> > >> On Fri, Jan 9, 2015 at 7:19 PM, Martin Pavlík <mpav...@redhat.com> >> wrote: >> > >> >> > >>> Hi Punit, >> > >>> >> > >>> can you verify that nodes contain cluster packages from the >> following >> > >>> log? >> > >>> >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,823::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('gluster-swift',) not found >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,825::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('gluster-swift-object',) not found >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,826::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('gluster-swift-plugin',) not found >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('gluster-swift-account',) not found >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('gluster-swift-proxy',) not found >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('gluster-swift-doc',) not found >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('gluster-swift-container',) not found >> > >>> Thread-14::DEBUG::2015-01-09 >> > >>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package >> > >>> ('glusterfs-geo-replication',) not found >> > >>> >> > >>> >> > >>> M. >> > >>> >> > >>> On 09 Jan 2015, at 11:13, Punit Dambiwal <hypu...@gmail.com> >> wrote: >> > >>> >> > >>> Hi Kanagaraj, >> > >>> >> > >>> Please find the attached logs :- >> > >>> >> > >>> Engine Logs :- http://ur1.ca/jdopt >> > >>> VDSM Logs :- http://ur1.ca/jdoq9 >> > >>> >> > >>> >> > >>> >> > >>> On Thu, Jan 8, 2015 at 6:05 PM, Kanagaraj <kmayi...@redhat.com> >> wrote: >> > >>> >> > >>>> Do you see any errors in the UI? >> > >>>> >> > >>>> Also please provide the engine.log and vdsm.log when the failure >> > >>>> occured. >> > >>>> >> > >>>> Thanks, >> > >>>> Kanagaraj >> > >>>> >> > >>>> >> > >>>> On 01/08/2015 02:25 PM, Punit Dambiwal wrote: >> > >>>> >> > >>>> Hi Martin, >> > >>>> >> > >>>> The steps are below :- >> > >>>> >> > >>>> 1. Step the ovirt engine on the one server... >> > >>>> 2. Installed centos 7 on 4 host node servers.. >> > >>>> 3. I am using host node (compute+storage)....now i have added all 4 >> > >>>> nodes to engine... >> > >>>> 4. Create the gluster volume from GUI... >> > >>>> >> > >>>> Network :- >> > >>>> eth0 :- public network (1G) >> > >>>> eth1+eth2=bond0= VM public network (1G) >> > >>>> eth3+eth4=bond1=ovirtmgmt+storage (10G private network) >> > >>>> >> > >>>> every hostnode has 24 bricks=24*4(distributed replicated) >> > >>>> >> > >>>> Thanks, >> > >>>> Punit >> > >>>> >> > >>>> >> > >>>> On Thu, Jan 8, 2015 at 3:20 PM, Martin Pavlík <mpav...@redhat.com> >> > >>>> wrote: >> > >>>> >> > >>>>> Hi Punit, >> > >>>>> >> > >>>>> can you please provide also errors from /var/log/vdsm/vdsm.log and >> > >>>>> /var/log/vdsm/vdsmd.log >> > >>>>> >> > >>>>> it would be really helpful if you provided exact steps how to >> > >>>>> reproduce the problem. >> > >>>>> >> > >>>>> regards >> > >>>>> >> > >>>>> Martin Pavlik - rhev QE >> > >>>>> > On 08 Jan 2015, at 03:06, Punit Dambiwal <hypu...@gmail.com> >> wrote: >> > >>>>> > >> > >>>>> > Hi, >> > >>>>> > >> > >>>>> > I try to add gluster volume but it failed... >> > >>>>> > >> > >>>>> > Ovirt :- 3.5 >> > >>>>> > VDSM :- vdsm-4.16.7-1.gitdb83943.el7 >> > >>>>> > KVM :- 1.5.3 - 60.el7_0.2 >> > >>>>> > libvirt-1.1.1-29.el7_0.4 >> > >>>>> > Glusterfs :- glusterfs-3.5.3-1.el7 >> > >>>>> > >> > >>>>> > Engine Logs :- >> > >>>>> > >> > >>>>> > 2015-01-08 09:57:52,569 INFO >> > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and >> wait lock >> > >>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >> > >>>>> value: GLUSTER >> > >>>>> > , sharedLocks= ] >> > >>>>> > 2015-01-08 09:57:52,609 INFO >> > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and >> wait lock >> > >>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >> > >>>>> value: GLUSTER >> > >>>>> > , sharedLocks= ] >> > >>>>> > 2015-01-08 09:57:55,582 INFO >> > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and >> wait lock >> > >>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >> > >>>>> value: GLUSTER >> > >>>>> > , sharedLocks= ] >> > >>>>> > 2015-01-08 09:57:55,591 INFO >> > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and >> wait lock >> > >>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >> > >>>>> value: GLUSTER >> > >>>>> > , sharedLocks= ] >> > >>>>> > 2015-01-08 09:57:55,596 INFO >> > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and >> wait lock >> > >>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >> > >>>>> value: GLUSTER >> > >>>>> > , sharedLocks= ] >> > >>>>> > 2015-01-08 09:57:55,633 INFO >> > >>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >> > >>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and >> wait lock >> > >>>>> EngineLock [exclusiveLocks= key: >> 00000001-0001-0001-0001-000000000300 >> > >>>>> value: GLUSTER >> > >>>>> > , sharedLocks= ] >> > >>>>> > ^C >> > >>>>> > >> > >>>>> > >> > >>>>> >> > >>>>> >> > >>>> >> > >>>> >> > >>> <216 09-Jan-15.jpg><217 09-Jan-15.jpg> >> > >>> >> > >>> >> > >>> >> > >> >> > >> >> > > >> > > >> > >> >> >> >> >> > >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users