[ovirt-users] [Gluster-users] Failed to create volume in OVirt with gluster

Punit Dambiwal hypunit at gmail.com
Tue Jan 13 10:15:56 UTC 2015


Hi Atin,

What about if i will use glusterfs 3.5 ?? is this bug will affect 3.5 also
??

On Tue, Jan 13, 2015 at 3:00 PM, Atin Mukherjee <amukherj at redhat.com> wrote:

>
>
> On 01/13/2015 12:12 PM, Punit Dambiwal wrote:
> > Hi Atin,
> >
> > Please find the output from here :- http://ur1.ca/jf4bs
> >
> Looks like http://review.gluster.org/#/c/9269/ should solve this issue.
> Please note this patch has not been taken in 3.6 release. Would you be
> able to apply this patch on the source and re-test?
>
> ~Atin
> > On Tue, Jan 13, 2015 at 12:37 PM, Atin Mukherjee <amukherj at redhat.com>
> > wrote:
> >
> >> Punit,
> >>
> >> cli log wouldn't help much here. To debug this issue further can you
> >> please let us know the following:
> >>
> >> 1. gluster peer status output
> >> 2. gluster volume status output
> >> 3. gluster --version output.
> >> 4. Which command got failed
> >> 5. glusterd log file of all the nodes
> >>
> >> ~Atin
> >>
> >>
> >> On 01/13/2015 07:48 AM, Punit Dambiwal wrote:
> >>> Hi,
> >>>
> >>> Please find the more details on this ....can anybody from gluster will
> >> help
> >>> me here :-
> >>>
> >>>
> >>> Gluster CLI Logs :- /var/log/glusterfs/cli.log
> >>>
> >>> [2015-01-13 02:06:23.071969] T [cli.c:264:cli_rpc_notify] 0-glusterfs:
> >> got
> >>> RPC_CLNT_CONNECT
> >>> [2015-01-13 02:06:23.072012] T
> [cli-quotad-client.c:94:cli_quotad_notify]
> >>> 0-glusterfs: got RPC_CLNT_CONNECT
> >>> [2015-01-13 02:06:23.072024] I [socket.c:2344:socket_event_handler]
> >>> 0-transport: disconnecting now
> >>> [2015-01-13 02:06:23.072055] T
> >> [cli-quotad-client.c:100:cli_quotad_notify]
> >>> 0-glusterfs: got RPC_CLNT_DISCONNECT
> >>> [2015-01-13 02:06:23.072131] T [rpc-clnt.c:1381:rpc_clnt_record]
> >>> 0-glusterfs: Auth Info: pid: 0, uid: 0, gid: 0, owner:
> >>> [2015-01-13 02:06:23.072176] T
> >>> [rpc-clnt.c:1238:rpc_clnt_record_build_header] 0-rpc-clnt: Request
> >> fraglen
> >>> 128, payload: 64, rpc hdr: 64
> >>> [2015-01-13 02:06:23.072572] T [socket.c:2863:socket_connect] (-->
> >>> /usr/lib64/libglusterfs.so.0(_gf_log_callingfn+0x1e0)[0x7fed02f15420]
> >> (-->
> >>>
> >>
> /usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0x7293)[0x7fed001a4293]
> >>> (--> /usr/lib64/libgfrpc.so.0(rpc_clnt_submit+0x468)[0x7fed0266df98]
> (-->
> >>> /usr/sbin/gluster(cli_submit_request+0xdb)[0x40a9bb] (-->
> >>> /usr/sbin/gluster(cli_cmd_submit+0x8e)[0x40b7be] ))))) 0-glusterfs:
> >> connect
> >>> () called on transport already connected
> >>> [2015-01-13 02:06:23.072616] T [rpc-clnt.c:1573:rpc_clnt_submit]
> >>> 0-rpc-clnt: submitted request (XID: 0x1 Program: Gluster CLI, ProgVers:
> >> 2,
> >>> Proc: 27) to rpc-transport (glusterfs)
> >>> [2015-01-13 02:06:23.072633] D
> [rpc-clnt-ping.c:231:rpc_clnt_start_ping]
> >>> 0-glusterfs: ping timeout is 0, returning
> >>> [2015-01-13 02:06:23.075930] T [rpc-clnt.c:660:rpc_clnt_reply_init]
> >>> 0-glusterfs: received rpc message (RPC XID: 0x1 Program: Gluster CLI,
> >>> ProgVers: 2, Proc: 27) from rpc-transport (glusterfs)
> >>> [2015-01-13 02:06:23.075976] D [cli-rpc-ops.c:6548:gf_cli_status_cbk]
> >>> 0-cli: Received response to status cmd
> >>> [2015-01-13 02:06:23.076025] D [cli-cmd.c:384:cli_cmd_submit] 0-cli:
> >>> Returning 0
> >>> [2015-01-13 02:06:23.076049] D
> [cli-rpc-ops.c:6811:gf_cli_status_volume]
> >>> 0-cli: Returning: 0
> >>> [2015-01-13 02:06:23.076192] D
> [cli-xml-output.c:84:cli_begin_xml_output]
> >>> 0-cli: Returning 0
> >>> [2015-01-13 02:06:23.076244] D
> >> [cli-xml-output.c:131:cli_xml_output_common]
> >>> 0-cli: Returning 0
> >>> [2015-01-13 02:06:23.076256] D
> >>> [cli-xml-output.c:1375:cli_xml_output_vol_status_begin] 0-cli:
> Returning
> >> 0
> >>> [2015-01-13 02:06:23.076437] D
> [cli-xml-output.c:108:cli_end_xml_output]
> >>> 0-cli: Returning 0
> >>> [2015-01-13 02:06:23.076459] D
> >>> [cli-xml-output.c:1398:cli_xml_output_vol_status_end] 0-cli: Returning
> 0
> >>> [2015-01-13 02:06:23.076490] I [input.c:36:cli_batch] 0-: Exiting
> with: 0
> >>>
> >>> Command log :- /var/log/glusterfs/.cmd_log_history
> >>>
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:10:35.836676]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:16:25.956514]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:17:36.977833]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:21:07.048053]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:26:57.168661]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:28:07.194428]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:30:27.256667]  : volume status vol01 : FAILED : Locking
> >>> failed on cpu02.zne01.hkg1.stack.com. Please check log file for
> details.
> >>> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file
> for
> >>> details.
> >>> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file
> for
> >>> details.
> >>> [2015-01-13 01:34:58.350748]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:36:08.375326]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:36:08.386470]  : volume status vol01 : FAILED : Locking
> >>> failed on cpu02.zne01.hkg1.stack.com. Please check log file for
> details.
> >>> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file
> for
> >>> details.
> >>> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file
> for
> >>> details.
> >>> [2015-01-13 01:42:59.524215]  : volume stop vol01 : FAILED : Locking
> >> failed
> >>> on cpu02.zne01.hkg1.stack.com. Please check log file for details.
> >>> Locking failed on cpu03.zne01.hkg1.stack.com. Please check log file
> for
> >>> details.
> >>> Locking failed on cpu04.zne01.hkg1.stack.com. Please check log file
> for
> >>> details.
> >>> [2015-01-13 01:45:10.550659]  : volume status all tasks : FAILED :
> >> Staging
> >>> failed on 00000000-0000-0000-0000-000000000000. Please check log file
> for
> >>> details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> Staging failed on 00000000-0000-0000-0000-000000000000. Please check
> log
> >>> file for details.
> >>> [2015-01-13 01:46:10.656802]  : volume status all tasks : SUCCESS
> >>> [2015-01-13 01:51:02.796031]  : volume status all tasks : SUCCESS
> >>> [2015-01-13 01:52:02.897804]  : volume status all tasks : SUCCESS
> >>> [2015-01-13 01:55:25.841070]  : system:: uuid get : SUCCESS
> >>> [2015-01-13 01:55:26.752084]  : system:: uuid get : SUCCESS
> >>> [2015-01-13 01:55:32.499049]  : system:: uuid get : SUCCESS
> >>> [2015-01-13 01:55:38.716907]  : system:: uuid get : SUCCESS
> >>> [2015-01-13 01:56:52.905899]  : volume status all tasks : SUCCESS
> >>> [2015-01-13 01:58:53.109613]  : volume status all tasks : SUCCESS
> >>> [2015-01-13 02:03:26.769430]  : system:: uuid get : SUCCESS
> >>> [2015-01-13 02:04:22.859213]  : volume status all tasks : SUCCESS
> >>> [2015-01-13 02:05:22.970393]  : volume status all tasks : SUCCESS
> >>> [2015-01-13 02:06:23.075823]  : volume status all tasks : SUCCESS
> >>>
> >>>
> >>> On Mon, Jan 12, 2015 at 10:53 PM, Kanagaraj Mayilsamy <
> >> kmayilsa at redhat.com>
> >>> wrote:
> >>>
> >>>> I can see the failures in glusterd log.
> >>>>
> >>>> Can someone from glusterfs dev pls help on this?
> >>>>
> >>>> Thanks,
> >>>> Kanagaraj
> >>>>
> >>>> ----- Original Message -----
> >>>>> From: "Punit Dambiwal" <hypunit at gmail.com>
> >>>>> To: "Kanagaraj" <kmayilsa at redhat.com>
> >>>>> Cc: "Martin Pavlík" <mpavlik at redhat.com>, "Vijay Bellur" <
> >>>> vbellur at redhat.com>, "Kaushal M" <kshlmster at gmail.com>,
> >>>>> users at ovirt.org, gluster-users at gluster.org
> >>>>> Sent: Monday, January 12, 2015 3:36:43 PM
> >>>>> Subject: Re: Failed to create volume in OVirt with gluster
> >>>>>
> >>>>> Hi Kanagaraj,
> >>>>>
> >>>>> Please find the logs from here :- http://ur1.ca/jeszc
> >>>>>
> >>>>> [image: Inline image 1]
> >>>>>
> >>>>> [image: Inline image 2]
> >>>>>
> >>>>> On Mon, Jan 12, 2015 at 1:02 PM, Kanagaraj <kmayilsa at redhat.com>
> >> wrote:
> >>>>>
> >>>>>>  Looks like there are some failures in gluster.
> >>>>>> Can you send the log output from glusterd log file from the relevant
> >>>> hosts?
> >>>>>>
> >>>>>> Thanks,
> >>>>>> Kanagaraj
> >>>>>>
> >>>>>>
> >>>>>> On 01/12/2015 10:24 AM, Punit Dambiwal wrote:
> >>>>>>
> >>>>>> Hi,
> >>>>>>
> >>>>>>  Is there any one from gluster can help me here :-
> >>>>>>
> >>>>>>  Engine logs :-
> >>>>>>
> >>>>>>  2015-01-12 12:50:33,841 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:34,725 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:36,824 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:36,853 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:36,866 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:37,751 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:39,849 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:39,878 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:39,890 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:40,776 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:42,878 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:42,903 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:42,916 INFO
> >>>>>>  [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>> (DefaultQuartzScheduler_Worker-12) Failed to acquire lock and wait
> >> lock
> >>>>>> EngineLock [exclusiveLocks= key:
> 00000001-0001-0001-0001-000000000300
> >>>>>> value: GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>> 2015-01-12 12:50:43,771 INFO
> >>>>>>
> >>>>
> [org.ovirt.engine.core.vdsbroker.gluster.CreateGlusterVolumeVDSCommand]
> >>>>>> (ajp--127.0.0.1-8702-1) [330ace48] FINISH,
> >>>> CreateGlusterVolumeVDSCommand,
> >>>>>> log id: 303e70a4
> >>>>>> 2015-01-12 12:50:43,780 ERROR
> >>>>>>
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> >>>>>> (ajp--127.0.0.1-8702-1) [330ace48] Correlation ID: 330ace48, Job ID:
> >>>>>> 896a69b3-a678-40a7-bceb-3635e4062aa0, Call Stack: null, Custom Event
> >>>> ID:
> >>>>>> -1, Message: Creation of Gluster Volume vol01 failed.
> >>>>>> 2015-01-12 12:50:43,785 INFO
> >>>>>>  [org.ovirt.engine.core.bll.gluster.CreateGlusterVolumeCommand]
> >>>>>> (ajp--127.0.0.1-8702-1) [330ace48] Lock freed to object EngineLock
> >>>>>> [exclusiveLocks= key: 00000001-0001-0001-0001-000000000300 value:
> >>>> GLUSTER
> >>>>>> , sharedLocks= ]
> >>>>>>
> >>>>>>  [image: Inline image 2]
> >>>>>>
> >>>>>>
> >>>>>> On Sun, Jan 11, 2015 at 6:48 PM, Martin Pavlík <mpavlik at redhat.com>
> >>>> wrote:
> >>>>>>
> >>>>>>> Hi Punit,
> >>>>>>>
> >>>>>>>  unfortunately I’am not that good with the gluster, I was just
> >>>> following
> >>>>>>> the obvious clue from the log. Could you try on the nodes if the
> >>>> packages
> >>>>>>> are even available for installation
> >>>>>>>
> >>>>>>>  yum install gluster-swift gluster-swift-object
> gluster-swift-plugin
> >>>>>>>  gluster-swift-account
> >>>>>>> gluster-swift-proxy gluster-swift-doc gluster-swift-container
> >>>>>>> glusterfs-geo-replication
> >>>>>>>
> >>>>>>>  if not you could try to get them in official gluster repo.
> >>>>>>>
> >>>>
> >>
> http://download.gluster.org/pub/gluster/glusterfs/LATEST/CentOS/glusterfs-epel.repo
> >>>>>>>
> >>>>>>>  HTH
> >>>>>>>
> >>>>>>>  M.
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>   On 10 Jan 2015, at 04:35, Punit Dambiwal <hypunit at gmail.com>
> >> wrote:
> >>>>>>>
> >>>>>>>  Hi Martin,
> >>>>>>>
> >>>>>>>  I installed gluster from ovirt repo....is it require to install
> >> those
> >>>>>>> packages manually ??
> >>>>>>>
> >>>>>>> On Fri, Jan 9, 2015 at 7:19 PM, Martin Pavlík <mpavlik at redhat.com>
> >>>> wrote:
> >>>>>>>
> >>>>>>>>  Hi Punit,
> >>>>>>>>
> >>>>>>>> can you verify that nodes contain cluster packages from the
> >> following
> >>>>>>>> log?
> >>>>>>>>
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,823::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('gluster-swift',) not found
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,825::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('gluster-swift-object',) not found
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,826::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('gluster-swift-plugin',) not found
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('gluster-swift-account',) not found
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('gluster-swift-proxy',) not found
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,829::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('gluster-swift-doc',) not found
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('gluster-swift-container',) not found
> >>>>>>>> Thread-14::DEBUG::2015-01-09
> >>>>>>>> 18:06:28,830::caps::716::root::(_getKeyPackages) rpm package
> >>>>>>>> ('glusterfs-geo-replication',) not found
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>  M.
> >>>>>>>>
> >>>>>>>>    On 09 Jan 2015, at 11:13, Punit Dambiwal <hypunit at gmail.com>
> >>>> wrote:
> >>>>>>>>
> >>>>>>>>    Hi Kanagaraj,
> >>>>>>>>
> >>>>>>>>  Please find the attached logs :-
> >>>>>>>>
> >>>>>>>>  Engine Logs :- http://ur1.ca/jdopt
> >>>>>>>> VDSM Logs :- http://ur1.ca/jdoq9
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> On Thu, Jan 8, 2015 at 6:05 PM, Kanagaraj <kmayilsa at redhat.com>
> >>>> wrote:
> >>>>>>>>
> >>>>>>>>>  Do you see any errors in the UI?
> >>>>>>>>>
> >>>>>>>>> Also please provide the engine.log and vdsm.log when the failure
> >>>>>>>>> occured.
> >>>>>>>>>
> >>>>>>>>> Thanks,
> >>>>>>>>> Kanagaraj
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> On 01/08/2015 02:25 PM, Punit Dambiwal wrote:
> >>>>>>>>>
> >>>>>>>>> Hi Martin,
> >>>>>>>>>
> >>>>>>>>>  The steps are below :-
> >>>>>>>>>
> >>>>>>>>>  1. Step the ovirt engine on the one server...
> >>>>>>>>> 2. Installed centos 7 on 4 host node servers..
> >>>>>>>>> 3. I am using host node (compute+storage)....now i have added
> all 4
> >>>>>>>>> nodes to engine...
> >>>>>>>>> 4. Create the gluster volume from GUI...
> >>>>>>>>>
> >>>>>>>>>  Network :-
> >>>>>>>>> eth0 :- public network (1G)
> >>>>>>>>> eth1+eth2=bond0= VM public network (1G)
> >>>>>>>>> eth3+eth4=bond1=ovirtmgmt+storage (10G private network)
> >>>>>>>>>
> >>>>>>>>>  every hostnode has 24 bricks=24*4(distributed replicated)
> >>>>>>>>>
> >>>>>>>>>  Thanks,
> >>>>>>>>> Punit
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> On Thu, Jan 8, 2015 at 3:20 PM, Martin Pavlík <
> mpavlik at redhat.com>
> >>>>>>>>> wrote:
> >>>>>>>>>
> >>>>>>>>>> Hi Punit,
> >>>>>>>>>>
> >>>>>>>>>> can you please provide also errors from /var/log/vdsm/vdsm.log
> and
> >>>>>>>>>> /var/log/vdsm/vdsmd.log
> >>>>>>>>>>
> >>>>>>>>>> it would be really helpful if you provided exact steps how to
> >>>>>>>>>> reproduce the problem.
> >>>>>>>>>>
> >>>>>>>>>> regards
> >>>>>>>>>>
> >>>>>>>>>> Martin Pavlik - rhev QE
> >>>>>>>>>>  > On 08 Jan 2015, at 03:06, Punit Dambiwal <hypunit at gmail.com>
> >>>> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>> Hi,
> >>>>>>>>>>>
> >>>>>>>>>>> I try to add gluster volume but it failed...
> >>>>>>>>>>>
> >>>>>>>>>>> Ovirt :- 3.5
> >>>>>>>>>>> VDSM :- vdsm-4.16.7-1.gitdb83943.el7
> >>>>>>>>>>> KVM :- 1.5.3 - 60.el7_0.2
> >>>>>>>>>>> libvirt-1.1.1-29.el7_0.4
> >>>>>>>>>>> Glusterfs :- glusterfs-3.5.3-1.el7
> >>>>>>>>>>>
> >>>>>>>>>>> Engine Logs :-
> >>>>>>>>>>>
> >>>>>>>>>>> 2015-01-08 09:57:52,569 INFO
> >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and
> wait
> >>>> lock
> >>>>>>>>>> EngineLock [exclusiveLocks= key:
> >>>> 00000001-0001-0001-0001-000000000300
> >>>>>>>>>> value: GLUSTER
> >>>>>>>>>>> , sharedLocks= ]
> >>>>>>>>>>> 2015-01-08 09:57:52,609 INFO
> >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and
> wait
> >>>> lock
> >>>>>>>>>> EngineLock [exclusiveLocks= key:
> >>>> 00000001-0001-0001-0001-000000000300
> >>>>>>>>>> value: GLUSTER
> >>>>>>>>>>> , sharedLocks= ]
> >>>>>>>>>>> 2015-01-08 09:57:55,582 INFO
> >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and
> wait
> >>>> lock
> >>>>>>>>>> EngineLock [exclusiveLocks= key:
> >>>> 00000001-0001-0001-0001-000000000300
> >>>>>>>>>> value: GLUSTER
> >>>>>>>>>>> , sharedLocks= ]
> >>>>>>>>>>> 2015-01-08 09:57:55,591 INFO
> >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and
> wait
> >>>> lock
> >>>>>>>>>> EngineLock [exclusiveLocks= key:
> >>>> 00000001-0001-0001-0001-000000000300
> >>>>>>>>>> value: GLUSTER
> >>>>>>>>>>> , sharedLocks= ]
> >>>>>>>>>>> 2015-01-08 09:57:55,596 INFO
> >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and
> wait
> >>>> lock
> >>>>>>>>>> EngineLock [exclusiveLocks= key:
> >>>> 00000001-0001-0001-0001-000000000300
> >>>>>>>>>> value: GLUSTER
> >>>>>>>>>>> , sharedLocks= ]
> >>>>>>>>>>> 2015-01-08 09:57:55,633 INFO
> >>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> >>>>>>>>>> (DefaultQuartzScheduler_Worker-16) Failed to acquire lock and
> wait
> >>>> lock
> >>>>>>>>>> EngineLock [exclusiveLocks= key:
> >>>> 00000001-0001-0001-0001-000000000300
> >>>>>>>>>> value: GLUSTER
> >>>>>>>>>>> , sharedLocks= ]
> >>>>>>>>>>> ^C
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>   <216 09-Jan-15.jpg><217 09-Jan-15.jpg>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>>
> >>>
> >>>
> >>> _______________________________________________
> >>> Gluster-users mailing list
> >>> Gluster-users at gluster.org
> >>> http://www.gluster.org/mailman/listinfo/gluster-users
> >>>
> >>
> >
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150113/7af78887/attachment-0001.html>


More information about the Users mailing list