[ovirt-users] Select As SPM Fails

Michael Watters wattersm at watters.ws
Thu Jan 19 20:16:29 UTC 2017


Anything is possible however I haven't had any issues since upgrading to
vdsm 4.17.35.

 

On 01/19/2017 02:58 PM, Beau Sapach wrote:
> Hmmm, makes sense, thanks for the info!  I'm not enthusiastic about
> installing packages outside of the ovirt repos so will probably look
> into an upgrade regardless.  I noticed that ovirt 4 only lists support
> for RHEL/CentOS 7.2, will a situation such as this crop up again
> eventually as incremental updates for the OS continue to push it past
> the supported version?  I've been running oVirt for less than a year
> now so I'm curious what to expect.
>
> On Thu, Jan 19, 2017 at 10:42 AM, Michael Watters
> <Michael.Watters at dart.biz <mailto:Michael.Watters at dart.biz>> wrote:
>
>     You can upgrade vdsm without upgrading to ovirt 4.  I went through the
>     same issue on our cluster a few weeks ago and the process was pretty
>     simple.
>
>     You'll need to do this on each of your hosts.
>
>     yum --enablerepo=extras install -y epel-release git
>     git clone https://github.com/oVirt/vdsm.git
>     <https://github.com/oVirt/vdsm.git>
>     cd  vdsm
>     git checkout v4.17.35
>     yum install -y `cat ./automation/build-artifacts.packages`
>     ./automation/build-artifacts.sh
>
>     cd /root/rpmbuild/RPMS/noarch
>     yum --enablerepo=extras install centos-release-qemu-ev
>     yum localinstall vdsm-4.17.35-1.el7.centos.noarch.rpm 
>     vdsm-hook-vmfex-dev-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-infra-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-jsonrpc-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-python-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-xmlrpc-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-yajsonrpc-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-python-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-xmlrpc-4.17.35-1.el7.centos.noarch.rpm
>     vdsm-cli-4.17.35-1.el7.centos.noarch.rpm
>     systemctl restart vdsmd
>
>     The qemu-ev repo is needed to avoid dependency errors.
>
>
>     On Thu, 2017-01-19 at 09:16 -0700, Beau Sapach wrote:
>     > Uh oh, looks like an upgrade to version 4 is the only option
>     then....
>     > unless I'm missing something.
>     >
>     > On Thu, Jan 19, 2017 at 1:36 AM, Pavel Gashev <Pax at acronis.com
>     <mailto:Pax at acronis.com>>
>     > wrote:
>     > > Beau,
>     > >  
>     > > Looks like you have upgraded to CentOS 7.3. Now you have to update
>     > > the vdsm package to 4.17.35.
>     > >  
>     > >  
>     > > From: <users-bounces at ovirt.org
>     <mailto:users-bounces at ovirt.org>> on behalf of Beau Sapach <bsapach at u
>     > > alberta.ca <http://alberta.ca>>
>     > > Date: Wednesday 18 January 2017 at 23:56
>     > > To: "users at ovirt.org <mailto:users at ovirt.org>"
>     <users at ovirt.org <mailto:users at ovirt.org>>
>     > > Subject: [ovirt-users] Select As SPM Fails
>     > >  
>     > > Hello everyone,
>     > >  
>     > > I'm about to start digging through the mailing list archives in
>     > > search of a solution but thought I would post to the list as
>     well. 
>     > > I'm running oVirt 3.6 on a 2 node CentOS7 cluster backed by fiber
>     > > channel storage and with a separate engine VM running outside of
>     > > the cluster (NOT  hosted-engine).
>     > >  
>     > > When I try to move the SPM role from one node to the other I get
>     > > the following in the web interface:
>     > >  
>     > >
>     > >  
>     > > When I look into /var/log/ovirt-engine/engine.log I see the
>     > > following:
>     > >  
>     > > 2017-01-18 13:35:09,332 ERROR
>     > >
>     [org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVD
>     > > SCommand] (default task-26) [6990cfca] Failed in
>     > > 'HSMGetAllTasksStatusesVDS' method
>     > > 2017-01-18 13:35:09,340 ERROR
>     > >
>     [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirect
>     > > or] (default task-26) [6990cfca] Correlation ID: null, Call Stack:
>     > > null, Custom Event ID: -1, Message: VDSM v6 command failed:
>     Logical
>     > > Volume extend failed
>     > >  
>     > > When I look at the task list on the host currently holding the SPM
>     > > role (in this case 'v6'), using: vdsClient -s 0 getAllTasks, I see
>     > > a long list like this:
>     > >  
>     > > dc75d3e7-cea7-449b-9a04-76fd8ef0f82b :
>     > >          verb = downloadImageFromStream
>     > >          code = 554
>     > >          state = recovered
>     > >          tag = spm
>     > >          result =
>     > >          message = Logical Volume extend failed
>     > >          id = dc75d3e7-cea7-449b-9a04-76fd8ef0f82b
>     > >  
>     > > When I look at /var/log/vdsm/vdsm.log on the host in question (v6)
>     > > I see messages like this:
>     > >  
>     > > '531dd533-22b1-47a0-aae8-76c1dd7d9a56': {'code': 554, 'tag':
>     > > u'spm', 'state': 'recovered', 'verb': 'downloadImageFromStreaam',
>     > > 'result': '', 'message': 'Logical Volume extend failed', 'id':
>     > > '531dd533-22b1-47a0-aae8-76c1dd7d9a56'}
>     > >  
>     > > As well as the error from the attempted extend of the logical
>     > > volume:
>     > >  
>     > > e980df5f-d068-4c84-8aa7-9ce792690562::ERROR::2017-01-18
>     > > 13:24:50,710::task::866::Storage.TaskManager.Task::(_setError)
>     > > Task=`e980df5f-d068-4c84-8aa7-9ce792690562`::Unexpected error
>     > > Traceback (most recent call last):
>     > >   File "/usr/share/vdsm/storage/task.py", line 873, in _run
>     > >     return fn(*args, **kargs)
>     > >   File "/usr/share/vdsm/storage/task.py", line 332, in run
>     > >     return self.cmd(*self.argslist, **self.argsdict)
>     > >   File "/usr/share/vdsm/storage/securable.py", line 77, in wrapper
>     > >     return method(self, *args, **kwargs)
>     > >   File "/usr/share/vdsm/storage/sp.py", line 1776, in
>     > > downloadImageFromStream
>     > >     .copyToImage(methodArgs, sdUUID, imgUUID, volUUID)
>     > >   File "/usr/share/vdsm/storage/image.py", line 1373, in
>     > > copyToImage
>     > >     / volume.BLOCK_SIZE)
>     > >   File "/usr/share/vdsm/storage/blockVolume.py", line 310, in
>     > > extend
>     > >     lvm.extendLV(self.sdUUID, self.volUUID, sizemb)
>     > >   File "/usr/share/vdsm/storage/lvm.py", line 1179, in extendLV
>     > >     _resizeLV("lvextend", vgName, lvName, size)
>     > >   File "/usr/share/vdsm/storage/lvm.py", line 1175, in _resizeLV
>     > >     raise se.LogicalVolumeExtendError(vgName, lvName, "%sM" %
>     > > (size, ))
>     > > LogicalVolumeExtendError:
>     > > Logical Volume extend failed: 'vgname=ae05947f-875c-4507-ad51-
>     > > 62b0d35ef567 lvname=caaef597-eddd-4c24-8df2-a61f35f744f8
>     > > newsize=1M'
>     > > e980df5f-d068-4c84-8aa7-9ce792690562::DEBUG::2017-01-18
>     > > 13:24:50,711::task::885::Storage.TaskManager.Task::(_run)
>     > > Task=`e980df5f-d068-4c84-8aa7-9ce792690562`::Task._run: e980df5f-
>     > > d068-4c84-8aa7-9ce792690562 () {} failed - stopping task
>     > >  
>     > > The logical volume in question is an OVF_STORE disk that lives on
>     > > one of the fiber channel backed LUNs.  If I run:
>     > >  
>     > > vdsClient -s 0 ClearTask TASK-UUID-HERE
>     > >  
>     > > for each task that appears in the:
>     > >  
>     > > vdsClient -s 0 getAllTasks 
>     > >  
>     > > output then they disappear and I'm able to move the SPM role
>     to the
>     > > other host.
>     > >  
>     > > This problem then crops up again on the new host once the SPM role
>     > > is moved.  What's going on here?  Does anyone have any insight as
>     > > to how to prevent this task from re-appearing?  Or why it's
>     failing
>     > > in the first place?
>     > >  
>     > > Beau
>     > >  
>     > >  
>     > >  
>     > >
>     >
>     >
>     >
>     > -- 
>     > Beau Sapach
>     > System Administrator | Information Technology Services | University
>     > of Alberta Libraries
>     > Phone: 780.492.4181 | Email: Beau.Sapach at ualberta.ca
>     <mailto:Beau.Sapach at ualberta.ca>
>     >
>
>
>
>
> -- 
> Beau Sapach
> *System Administrator | Information Technology Services | University
> of Alberta Libraries*
> *Phone: 780.492.4181 | Email: Beau.Sapach at ualberta.ca
> <mailto:Beau.Sapach at ualberta.ca>*
>
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170119/13b4ade1/attachment-0001.html>


More information about the Users mailing list