Users
Threads by month
- ----- 2026 -----
- January
- ----- 2025 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2024 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2023 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2022 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2021 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2020 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2019 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2018 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2017 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2016 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2015 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2014 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2013 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2012 -----
- December
- November
- October
- September
- August
- July
- June
- May
- April
- March
- February
- January
- ----- 2011 -----
- December
- November
- October
- 12 participants
- 19168 discussions
Hello,
I hope you got my irc message that I got my exams earlier(which are now
over) because of which I was not able to spent much time on project related
things.
since I am free now,I have read about ovirt architecture and documentation
to get more familiar with ovirt.
Now the purpose of the mail is that I am having some issue while installing
ovirt from source.(
http://www.ovirt.org/develop/developer-guide/engine/engine-development-envi…
)
when I am trying to do
$ make install-dev PREFIX="$HOME/ovirt-engine"
..
..
..
[ERROR] Failed to execute goal
org.codehaus.mojo:gwt-maven-plugin:2.6.1:compile (gwtcompile) on
project webadmin: Command [[
[ERROR] /bin/sh -c
/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.121-8.b14.fc25.x86_64/jre/bin/java
-javaagent:/home/0day/.m2/repository/org/aspectj/aspectjweaver/1.8.2/aspectjweaver-1.8.2.jar
-Dgwt.jjs.permutationWorkerFactory=com.google.gwt.dev.ThreadedPermutationWorkerFactory
\
[ERROR] -Dgwt.jjs.maxThreads=4 \
[ERROR] -Djava.io.tmpdir="/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/target/tmp"
\
[ERROR] -Djava.util.prefs.systemRoot="/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/target/tmp"
\
[ERROR] -Djava.util.prefs.userRoot="/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/target/tmp"
\
[ERROR] -Djava.util.logging.config.class=org.ovirt.engine.ui.gwtextension.JavaLoggingConfig
\
[ERROR] -Xms1024M -Xmx8192M
'-Dgwt.dontPrune=org\.ovirt\.engine\.core\.(common|compat)\..*'
-classpath /home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/target/webadmin-4.2.0-SNAPSHOT/WEB-INF/classes:/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/src/main/java:/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/target/generated-sources/annotations:/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/target/generated-sources/gwt:/home/0day/.m2/repository/com/gwtplatform/gwtp-mvp-client/1.3.1/gwtp-mvp-client-1.3.1.jar:/home/0day/.m2/repository/com/gwtplatform/gwtp-clients-common/1.3.1/gwtp-clients-common-1.3.1.jar:/home/0day/.m2/repository/com/gwtplatform/gwtp-mvp-shared/1.3.1/gwtp-mvp-shared-1.3.1.jar:/home/0day/.m2/repository/org/apache/velocity/velocity/1.7/velocity-1.7.jar:/home/0day/.m2/repository/com/gwtplatform/gwtp-processors/1.3.1/gwtp-processors-1.3.1.jar:/home/0day/.m2/repository/com/google/inject/guice/3.0/guice-3.0.jar:/home/0day/.m2/repository/javax/inject/javax.inject/1/javax.inject-1.jar:/home/0day/.m2/repository/aopalliance/aopalliance/1.0/aopalliance-1.0.jar:/home/0day/.m2/repository/com/google/inject/extensions/guice-assistedinject/3.0/guice-assistedinject-3.0.jar:/home/0day/.m2/repository/com/google/gwt/inject/gin/2.1.2/gin-2.1.2.jar:/home/0day/.m2/repository/com/google/gwt/gwt-user/2.6.1/gwt-user-2.6.1.jar:/home/0day/.m2/repository/javax/validation/validation-api/1.0.0.GA/validation-api-1.0.0.GA.jar:/home/0day/.m2/repository/javax/validation/validation-api/1.0.0.GA/validation-api-1.0.0.GA-sources.jar:/home/0day/.m2/repository/org/json/json/20090211/json-20090211.jar:/home/0day/.m2/repository/com/google/gwt/gwt-dev/2.6.1/gwt-dev-2.6.1.jar:/home/0day/.m2/repository/org/gwtbootstrap3/gwtbootstrap3/0.9.3/gwtbootstrap3-0.9.3.jar:/home/0day/.m2/repository/org/gwtbootstrap3/gwtbootstrap3-extras/0.9.3/gwtbootstrap3-extras-0.9.3.jar:/home/0day/git/ovirt-engine/frontend/webadmin/modules/uicommonweb/target/uicommonweb-4.2.0-SNAPSHOT.jar:/home/0day/git/ovirt-engine/backend/manager/modules/searchbackend/target/searchbackend-4.2.0-SNAPSHOT.jar:/home/0day/git/ovirt-engine/frontend/webadmin/modules/uicompat/target/uicompat-4.2.0-SNAPSHOT.jar:/home/0day/git/ovirt-engine/frontend/webadmin/modules/gwt-common/target/gwt-common-4.2.0-SNAPSHOT.jar:/home/0day/git/ovirt-engine/backend/manager/modules/common/target/common-4.2.0-SNAPSHOT.jar:/home/0day/.m2/repository/org/hibernate/hibernate-validator/4.2.0.Final/hibernate-validator-4.2.0.Final.jar:/home/0day/.m2/repository/org/jboss/modules/jboss-modules/1.1.1.GA/jboss-modules-1.1.1.GA.jar:/home/0day/git/ovirt-engine/backend/manager/modules/common/target/common-4.2.0-SNAPSHOT-sources.jar:/home/0day/git/ovirt-engine/backend/manager/modules/compat/target/compat-4.2.0-SNAPSHOT-sources.jar:/home/0day/git/ovirt-engine/backend/manager/modules/searchbackend/target/searchbackend-4.2.0-SNAPSHOT-sources.jar:/home/0day/git/ovirt-engine/frontend/webadmin/modules/frontend/target/frontend-4.2.0-SNAPSHOT.jar:/home/0day/git/ovirt-engine/frontend/webadmin/modules/gwt-extension/target/ovirt-engine-gwt-extension-4.2.0-SNAPSHOT.jar:/home/0day/.m2/repository/javax/xml/bind/jaxb-api/2.1/jaxb-api-2.1.jar:/home/0day/.m2/repository/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar:/home/0day/.m2/repository/javax/activation/activation/1.1/activation-1.1.jar:/home/0day/.m2/repository/org/jboss/spec/javax/servlet/jboss-servlet-api_3.1_spec/1.0.0.Final/jboss-servlet-api_3.1_spec-1.0.0.Final.jar:/home/0day/.m2/repository/org/jboss/spec/javax/servlet/jstl/jboss-jstl-api_1.2_spec/1.0.3.Final/jboss-jstl-api_1.2_spec-1.0.3.Final.jar:/home/0day/.m2/repository/org/jboss/spec/javax/el/jboss-el-api_2.2_spec/1.0.1.Final/jboss-el-api_2.2_spec-1.0.1.Final.jar:/home/0day/.m2/repository/org/jboss/spec/javax/servlet/jboss-servlet-api_3.0_spec/1.0.1.Final/jboss-servlet-api_3.0_spec-1.0.1.Final.jar:/home/0day/.m2/repository/org/jboss/spec/javax/servlet/jsp/jboss-jsp-api_2.2_spec/1.0.1.Final/jboss-jsp-api_2.2_spec-1.0.1.Final.jar:/home/0day/.m2/repository/xalan/xalan/2.7.1.jbossorg-2/xalan-2.7.1.jbossorg-2.jar:/home/0day/.m2/repository/xalan/serializer/2.7.1.jbossorg-2/serializer-2.7.1.jbossorg-2.jar:/home/0day/.m2/repository/org/jboss/spec/javax/ejb/jboss-ejb-api_3.1_spec/1.0.2.Final/jboss-ejb-api_3.1_spec-1.0.2.Final.jar:/home/0day/.m2/repository/org/jboss/spec/javax/enterprise/concurrent/jboss-concurrency-api_1.0_spec/1.0.0.Final/jboss-concurrency-api_1.0_spec-1.0.0.Final.jar:/home/0day/.m2/repository/org/aspectj/aspectjweaver/1.8.2/aspectjweaver-1.8.2.jar:/home/0day/.m2/repository/org/slf4j/slf4j-api/1.7.5/slf4j-api-1.7.5.jar:/home/0day/git/ovirt-engine/backend/manager/modules/utils/target/utils-4.2.0-SNAPSHOT.jar:/home/0day/git/ovirt-engine/backend/manager/modules/extensions-api-root/extensions-api/target/ovirt-engine-extensions-api-0.0.0-SNAPSHOT.jar:/home/0day/git/ovirt-engine/backend/manager/modules/extensions-manager/target/extensions-manager-4.2.0-SNAPSHOT.jar:/home/0day/.m2/repository/commons-beanutils/commons-beanutils/1.9.2/commons-beanutils-1.9.2.jar:/home/0day/.m2/repository/commons-logging/commons-logging/1.1.1/commons-logging-1.1.1.jar:/home/0day/.m2/repository/commons-collections/commons-collections/3.2.2/commons-collections-3.2.2.jar:/home/0day/git/ovirt-engine/backend/manager/modules/compat/target/compat-4.2.0-SNAPSHOT.jar:/home/0day/.m2/repository/org/ovirt/otopi/otopi/1.5.2/otopi-1.5.2.jar:/home/0day/.m2/repository/org/ovirt/ovirt-host-deploy/ovirt-host-deploy/1.5.0/ovirt-host-deploy-1.5.0.jar:/home/0day/.m2/repository/org/apache/commons/commons-compress/1.5/commons-compress-1.5.jar:/home/0day/.m2/repository/org/tukaani/xz/1.2/xz-1.2.jar:/home/0day/git/ovirt-engine/backend/manager/modules/uutils/target/uutils-4.2.0-SNAPSHOT.jar:/home/0day/.m2/repository/org/apache/sshd/sshd-core/0.12.0/sshd-core-0.12.0.jar:/home/0day/.m2/repository/org/apache/mina/mina-core/2.0.7/mina-core-2.0.7.jar:/home/0day/.m2/repository/org/apache/httpcomponents/httpclient/4.5/httpclient-4.5.jar:/home/0day/.m2/repository/org/apache/httpcomponents/httpcore/4.4.1/httpcore-4.4.1.jar:/home/0day/.m2/repository/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/home/0day/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar:/home/0day/.m2/repository/commons-httpclient/commons-httpclient/3.1/commons-httpclient-3.1.jar:/home/0day/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.9/jackson-core-asl-1.9.9.jar:/home/0day/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.9/jackson-mapper-asl-1.9.9.jar:/home/0day/.m2/repository/org/springframework/spring-core/4.2.4.RELEASE/spring-core-4.2.4.RELEASE.jar:/home/0day/.m2/repository/org/infinispan/infinispan-core/5.2.5.Final/infinispan-core-5.2.5.Final.jar:/home/0day/.m2/repository/org/jgroups/jgroups/3.2.7.Final/jgroups-3.2.7.Final.jar:/home/0day/.m2/repository/org/jboss/spec/javax/transaction/jboss-transaction-api_1.1_spec/1.0.0.Final/jboss-transaction-api_1.1_spec-1.0.0.Final.jar:/home/0day/.m2/repository/org/jboss/marshalling/jboss-marshalling-river/1.3.15.GA/jboss-marshalling-river-1.3.15.GA.jar:/home/0day/.m2/repository/org/jboss/marshalling/jboss-marshalling/1.3.15.GA/jboss-marshalling-1.3.15.GA.jar:/home/0day/.m2/repository/org/jboss/logging/jboss-logging/3.1.1.GA/jboss-logging-3.1.1.GA.jar:/home/0day/.m2/repository/org/jboss/staxmapper/1.1.0.Final/staxmapper-1.1.0.Final.jar:/home/0day/.m2/repository/com/woorea/keystone-client/3.1.2/keystone-client-3.1.2.jar:/home/0day/.m2/repository/com/woorea/openstack-client/3.1.2/openstack-client-3.1.2.jar:/home/0day/.m2/repository/com/woorea/keystone-model/3.1.2/keystone-model-3.1.2.jar:/home/0day/.m2/repository/com/woorea/glance-client/3.1.2/glance-client-3.1.2.jar:/home/0day/.m2/repository/com/woorea/glance-model/3.1.2/glance-model-3.1.2.jar:/home/0day/.m2/repository/com/woorea/cinder-client/3.1.2/cinder-client-3.1.2.jar:/home/0day/.m2/repository/com/woorea/cinder-model/3.1.2/cinder-model-3.1.2.jar:/home/0day/.m2/repository/org/ebaysf/web/cors-filter/1.0.1/cors-filter-1.0.1.jar:/home/0day/git/ovirt-engine/backend/manager/modules/branding/target/branding-4.2.0-SNAPSHOT.jar:/home/0day/.m2/repository/com/google/gwt/gwt-user/2.6.1/gwt-user-2.6.1.jar:/home/0day/.m2/repository/com/google/gwt/gwt-dev/2.6.1/gwt-dev-2.6.1.jar:/home/0day/git/ovirt-engine/frontend/webadmin/modules/gwt-extension/target/ovirt-engine-gwt-extension-4.2.0-SNAPSHOT-sources.jar:/home/0day/git/ovirt-engine/frontend/webadmin/modules/uicommonweb/target/uicommonweb-4.2.0-SNAPSHOT-sources.jar
com.google.gwt.dev.Compiler -logLevel INFO -style OBF -war
/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/target/generated-gwt
-localWorkers 4 -failOnError -XfragmentCount -1 -sourceLevel auto -gen
/home/0day/git/ovirt-engine/frontend/webadmin/modules/webadmin/gen
org.ovirt.engine.ui.webadmin.WebAdmin
[ERROR] ]] failed with status 1
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with
the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions,
please read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :webadmin
Makefile:257: recipe for target 'maven' failed
make[2]: *** [maven] Error 1
make[2]: Leaving directory '/home/0day/git/ovirt-engine'
Makefile:264: recipe for target 'tmp.built' failed
make[1]: *** [tmp.built] Error 2
make[1]: Leaving directory '/home/0day/git/ovirt-engine'
Makefile:501: recipe for target 'all-dev' failed
make: *** [all-dev] Error 2
then I try instruction given at this link to setup maven
http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
It successfully run without any error but again make install fail.
I have even tried
$mvn initilize -X -e
that also run successfully but didn't help in removing the error.
So,any solution.
Thanks,
Shubham
On Mon, Mar 6, 2017 at 12:32 PM, Maor Lipchuk <mlipchuk(a)redhat.com> wrote:
> Hi shubham,
>
> You should not worry about the java language, the feature is focused
> mainly on logic and the development should be pretty simple.
> Also, your patches will be reviewed through gerrit, so I will help and
> it should be very convenient to develop.
> and of course, you will have the change to get to know a new language,
> which should be fun :)
>
> Regards,
> Maor
>
> On Mon, Mar 6, 2017 at 2:06 AM, shubham dubey <sdubey504(a)gmail.com> wrote:
> > I was expecting it to be in python since major parts of ovirt are in
> > python:).
> > I am little less familiar with java then I am with python.So,did even
> > moderate control on java language will work for project or did I try for
> > some other ideas?
> > Although I have enough experience in Storage technology.
> >
> > Thanks,
> > Shubham
> >
> > On Mon, Mar 6, 2017 at 5:24 AM, Maor Lipchuk <mlipchuk(a)redhat.com>
> wrote:
> >>
> >> Hi Shubham,
> >>
> >> I just saw your message in the oVirt channel, sorry for the late respond
> >> The new backup storage should be developed mainly in Java, it will be
> >> part of the ovirt-engine
> >>
> >> On Sun, Mar 5, 2017 at 10:47 AM, Maor Lipchuk <mlipchuk(a)redhat.com>
> wrote:
> >> > Hi shubham,
> >> >
> >> > Thank you for your interest in the Configuring the backup storage as
> >> > part of the GOSC project.
> >> > The idea is to have an alternative for backing up entities like VMs,
> >> > Templates and disks from a storage domain.
> >> >
> >> > oVirt has many methods of backup entities such as DB backup, import
> >> > storage domain and use of export storage domain.
> >> > With this new capability to define a storage domain as a backup it
> >> > might help oVirt to achieve an easier alternative solution for the
> >> > users to backup entities
> >> >
> >> > I suggest that for start, try to build a working oVirt development
> >> > setup with hosts and a few storage domains.
> >> >
> >> > You can follow this wiki:
> >> >
> >> > http://www.ovirt.org/develop/developer-guide/engine/engine-
> development-environment/
> >> >
> >> > Once you have a working oVirt env we can go forward to the next step.
> >> > Please don't hesitate to ask if you have any questions.
> >> > You can use the #ovirt channel also.
> >> >
> >> > Regards,
> >> > Maor
> >> >
> >> > On Fri, Mar 3, 2017 at 11:27 AM, shubham dubey <sdubey504(a)gmail.com>
> >> > wrote:
> >> >> Hello,
> >> >> I am interested in being part of ovirt for gsoc 2017.
> >> >> I have looked into ovirt project ideas and the project that I find
> most
> >> >> suitable is Configuring the backup storage in ovirt.
> >> >>
> >> >> Since the ovirt online docs have sufficient info for getting started
> >> >> for
> >> >> development so I don't have a questions about that but I want to
> >> >> clarify one
> >> >> doubt that did the previous year mentioned project on ovirt gsoc page
> >> >> are
> >> >> also available to work on?
> >> >> I will also appreciate any discussion about the project or question
> >> >> from
> >> >> mentor side.Even some guideline for start working is welcome.
> >> >>
> >> >> Thanks,
> >> >> Shubham
> >> >>
> >> >> _______________________________________________
> >> >> Users mailing list
> >> >> Users(a)ovirt.org
> >> >> http://lists.ovirt.org/mailman/listinfo/users
> >> >>
> >
> >
>
1
0
Hi, All. We have an ovirt 4.1 cluster setup using multiple paths to a single iSCSI LUN for the data storage domain. I would now like to migrate to a hosted engine.
I setup the new engine VM, shutdown and backed-up the old VM, and restored to the new VM using engine-backup. After updating DNS to change our engine's FQDN to point to the hosted engine, everything seems to work properly. However, when rebooting the entire cluster, the engine VM doesn't come up automatically.
Is there anything that now needs to be done to tell the cluster that it's now using a hosted engine?
I started with a standard engine setup, as I didn't see a way to specify multiple paths to a single iSCSI LUN when using "hosted-engine --deploy."
Any tips would be greatly appreciated.
Many thanks,
Devin
3
2
13 Mar '17
On Mon, Mar 13, 2017 at 11:31 AM, gflwqs gflwqs <gflwqs(a)gmail.com> wrote:
> Hi Simone,
> Since there is a lot of stuff in the setup log that i don't want to send
> to the internet, are there anything specific that i could pick out from the
> log?
>
> Here is the lsbl and fdisk output from the engine:
>
> [root@engine ~]# lsblk
> NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
> sr0 11:0 1 1024M 0 rom
> vda 253:0 0 90G 0 disk
> └─vda1 253:1 0 10G 0 part /
>
>
OK, so the disk is 90 GB but the partition is just 10 GB.
Could you please check if you have cloud-utils-growpart package inside your
engine VM?
If not, could you please report the version of ovirt-engine-appliance you
used at setup time?
> [root@engine ~]# fdisk -l /dev/vda
>
> Disk /dev/vda: 96.6 GB, 96636764160 bytes, 188743680 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk label type: dos
> Disk identifier: 0x0009e1bc
>
> Device Boot Start End Blocks Id System
> /dev/vda1 * 2048 20971519 10484736 83 Linux
>
> 2017-03-13 10:44 GMT+01:00 Simone Tiraboschi <stirabos(a)redhat.com>:
>
>>
>>
>> On Mon, Mar 13, 2017 at 8:26 AM, gflwqs gflwqs <gflwqs(a)gmail.com> wrote:
>>
>>> Hi, I have recently installed a new ovirt environment from scratch.
>>> The version is 4.1.0.
>>> I made a 90GB disk for the hosted engine.
>>> When i had installed it and log into the engine i see that it has only
>>> configured 10GB of those 90GB.
>>> So there is 80GB of unconfigured disk unused.
>>>
>>
>> Can you please attach your hosted-engine-setup log file?
>>
>> Could you please check if the disk is 90GB but the main disk partition is
>> just 10 GB?
>>
>> cloud-utils-growpart should grow the partition on the first boot, if the
>> issue is there we need to understand why it didn't triggered in your case.
>>
>>
>>> That is fine as long as it had used lvm but it has not meaning i have to
>>> take the engine down to extend the disk?
>>>
>>> My questions is:
>>> 1. Why does it not configure all disk?
>>> 2. Why is it not using lvm?
>>> 3. What should i do to extend the disk?
>>>
>>
>> It depends on where the issue is (is on the disk size or just on the
>> partition size?).
>>
>>
>>>
>>> Regards
>>> Christian
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users(a)ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>
1
0
This is a multi-part message in MIME format.
--------------C04C9E032CD3A8F03847AFCD
Content-Type: text/plain; charset=utf-8; format=flowed
Content-Transfer-Encoding: 7bit
Hello everybody,
for production usage i'm testing ovirt with gluster.
All components seems to be running fine but whenever I'm testing huge
workload, then node freez. Not the main OS, but VDSM mgmt and attached
services, VMs eg.
*mgmt *
oVirt - 4.1.0.4
centos 7.3-1611
*nodes* ( installed from ovirt image
/"ovirt-node-ng-installer-ovirt-4.1-2017030804.iso" )/
OS Version:== RHEL - 7 - 3.1611.el7.centos
OS Description:== oVirt Node 4.1.0
Kernel Version:== 3.10.0 - 514.10.2.el7.x86_64
KVM Version:== 2.6.0 - 28.el7_3.3.1
LIBVIRT Version:== libvirt-2.0.0-10.el7_3.5
VDSM Version:== vdsm-4.19.4-1.el7.centos
SPICE Version:== 0.12.4 - 20.el7_3
GlusterFS Version:== glusterfs-3.8.9-1.el7 ( LVM thinprovisioning in
replica 2 - created from ovirt GUI )
concurently running
- huge import from export domain ( net workload )
- sequential write to VMs local disk ( gluster replica sequential workload )
- VMs database huge select ( random IOps )
- huge old snapshot delete ( random IOps )
In this configuration / workload is runnig one hour eg, with no
exceptions , with 70-80% disk load, but in some point VDSM freez all
jobs for a timeout and VM's are in "uknown" status .
The whole system revitalize then automaticaly in cca 20min time frame (
except the import and snapshot deleting(rollback) )
engine.log - focus 10:39:07 time ( Failed in
'HSMGetAllTasksStatusesVDS' method )
========
n child command id: 'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete
2017-03-10 10:39:01,727+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler2) [759c8e1f] Command
'RemoveSnapshotSingleDiskLive' (id:
'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id:
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage' to complete
2017-03-10 10:39:03,929+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 446d0cd3
2017-03-10 10:39:04,343+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED,
16.0.0.164:CONNECTED], log id: 446d0cd3
2017-03-10 10:39:04,353+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 69ea1fda
2017-03-10 10:39:05,128+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 69ea1fda
2017-03-10 10:39:07,163+01 ERROR
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand]
(DefaultQuartzScheduler2) [759c8e1f] Failed in
'HSMGetAllTasksStatusesVDS' method
2017-03-10 10:39:07,178+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(DefaultQuartzScheduler2) [759c8e1f] EVENT_ID:
VDS_BROKER_COMMAND_FAILURE(10,802), Correlation ID: null, Call Stack:
null, Custom Event ID: -1, Message: VDSM 2kvm2 command
HSMGetAllTasksStatusesVDS failed: Connection timed out
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2)
[759c8e1f] BaseAsyncTask::onTaskEndSuccess: Task
'f594bf69-619b-4d1b-8f6d-a9826997e478' (Parent Command 'ImportVm',
Parameters Type
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') ended
successfully.
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.CommandMultiAsyncTasks]
(DefaultQuartzScheduler2) [759c8e1f] Task with DB Task ID
'a05c7c07-9b98-4ab2-ac7b-9e70a75ba7b7' and VDSM Task ID
'7c60369f-70a3-4a6a-80c9-4753ac9ed372' is in state Polling. End action
for command 8deb3fe3-4a83-4605-816c-ffdc63fd9ac1 will proceed when all
the entity's tasks are completed.
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2)
[759c8e1f] SPMAsyncTask::PollTask: Polling task
'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent Command 'DestroyImage',
Parameters Type
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') returned
status 'finished', result 'cleanSuccess'.
2017-03-10 10:39:07,182+01 ERROR
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2)
[759c8e1f] BaseAsyncTask::logEndTaskFailure: Task
'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent Command 'DestroyImage',
Parameters Type
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') ended
with failure:
-- Result: 'cleanSuccess'
-- Message: 'VDSGenericException: VDSErrorException: Failed to
HSMGetAllTasksStatusesVDS, error = Connection timed out, code = 100',
-- Exception: 'VDSGenericException: VDSErrorException: Failed to
HSMGetAllTasksStatusesVDS, error = Connection timed out, code = 100'
2017-03-10 10:39:07,184+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(DefaultQuartzScheduler2) [759c8e1f]
CommandAsyncTask::endActionIfNecessary: All tasks of command
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e' has ended -> executing 'endAction'
2017-03-10 10:39:07,185+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(DefaultQuartzScheduler2) [759c8e1f] CommandAsyncTask::endAction: Ending
action for '1' tasks (command ID:
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e'): calling endAction '.
2017-03-10 10:39:07,185+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::endCommandAction [within thread] context: Attempting
to endAction 'DestroyImage',
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.storage.disk.image.DestroyImageCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] Command
[id=33df2c1e-6ce3-44fd-a39b-d111883b4c4e]: Updating status to 'FAILED',
The command end method logic will be executed by one of its parent commands.
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::HandleEndActionResult [within thread]: endAction for
action type 'DestroyImage' completed, handling the result.
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::HandleEndActionResult [within thread]: endAction for
action type 'DestroyImage' succeeded, clearing tasks.
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
SPMAsyncTask::ClearAsyncTask: Attempting to clear task
'f351e8f6-6dd7-49aa-bf54-650d84fc6352'
2017-03-10 10:39:07,193+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] START,
SPMClearTaskVDSCommand(
SPMTaskGuidBaseVDSCommandParameters:{runAsync='true',
storagePoolId='00000001-0001-0001-0001-000000000311',
ignoreFailoverLimit='false',
taskId='f351e8f6-6dd7-49aa-bf54-650d84fc6352'}), log id: 2b7080c2
2017-03-10 10:39:07,194+01 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] START,
HSMClearTaskVDSCommand(HostName = 2kvm2,
HSMTaskGuidBaseVDSCommandParameters:{runAsync='true',
hostId='905375e1-6de4-4fdf-b69c-b2d546f869c8',
taskId='f351e8f6-6dd7-49aa-bf54-650d84fc6352'}), log id: 2edff460
2017-03-10 10:39:08,208+01 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] FINISH,
HSMClearTaskVDSCommand, log id: 2edff460
2017-03-10 10:39:08,208+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] FINISH,
SPMClearTaskVDSCommand, log id: 2b7080c2
2017-03-10 10:39:08,213+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
BaseAsyncTask::removeTaskFromDB: Removed task
'f351e8f6-6dd7-49aa-bf54-650d84fc6352' from DataBase
2017-03-10 10:39:08,213+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::HandleEndActionResult [within thread]: Removing
CommandMultiAsyncTasks object for entity
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e'
2017-03-10 10:39:10,142+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler10) [a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 2e7278cb
2017-03-10 10:39:11,513+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler10) [a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c]
FINISH, GlusterServersListVDSCommand, return:
[172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id: 2e7278cb
2017-03-10 10:39:11,523+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler10) [a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 43704ef2
2017-03-10 10:39:11,777+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler9) [67e1d8ed] Command 'RemoveSnapshot' (id:
'13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child command id:
'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete
2017-03-10 10:39:11,789+01 WARN
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
(DefaultQuartzScheduler9) [759c8e1f] Child command 'DESTROY_IMAGE'
failed, proceeding to verify
2017-03-10 10:39:11,789+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
(DefaultQuartzScheduler9) [759c8e1f] Executing Live Merge command step
'DESTROY_IMAGE_CHECK'
2017-03-10 10:39:11,832+01 INFO
[org.ovirt.engine.core.bll.DestroyImageCheckCommand] (pool-5-thread-7)
[4856f570] Running command: DestroyImageCheckCommand internal: true.
2017-03-10 10:39:11,833+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMGetVolumeInfoVDSCommand]
(pool-5-thread-7) [4856f570] START, SPMGetVolumeInfoVDSCommand(
SPMGetVolumeInfoVDSCommandParameters:{expectedEngineErrors='[VolumeDoesNotExist]',
runAsync='true', storagePoolId='00000001-0001-0001-0001-000000000311',
ignoreFailoverLimit='false',
storageDomainId='1603cd90-92ef-4c03-922c-cecb282fd00e',
imageGroupId='7543338a-3ca6-4698-bb50-c14f0bd71428',
imageId='50b592f7-bfba-4398-879c-8d6a19a2c000'}), log id: 2c8031f8
2017-03-10 10:39:11,833+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMGetVolumeInfoVDSCommand]
(pool-5-thread-7) [4856f570] Executing GetVolumeInfo using the current SPM
2017-03-10 10:39:11,834+01 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetVolumeInfoVDSCommand]
(pool-5-thread-7) [4856f570] START, GetVolumeInfoVDSCommand(HostName =
2kvm2,
GetVolumeInfoVDSCommandParameters:{expectedEngineErrors='[VolumeDoesNotExist]',
runAsync='true', hostId='905375e1-6de4-4fdf-b69c-b2d546f869c8',
storagePoolId='00000001-0001-0001-0001-000000000311',
storageDomainId='1603cd90-92ef-4c03-922c-cecb282fd00e',
imageGroupId='7543338a-3ca6-4698-bb50-c14f0bd71428',
imageId='50b592f7-bfba-4398-879c-8d6a19a2c000'}), log id: 79ca86cc
2017-03-10 10:39:11,846+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler10) [a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c]
FINISH, GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 43704ef2
2017-03-10 10:39:16,858+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 6542adcd
2017-03-10 10:39:17,394+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f] FINISH,
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED,
16.0.0.164:CONNECTED], log id: 6542adcd
2017-03-10 10:39:17,406+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 44ec33ed
2017-03-10 10:39:18,598+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 44ec33ed
2017-03-10 10:39:21,865+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler6) [67e1d8ed] Command 'RemoveSnapshot' (id:
'13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child command id:
'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete
2017-03-10 10:39:21,881+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler6) [4856f570] Command
'RemoveSnapshotSingleDiskLive' (id:
'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id:
'b1d63b8e-19d3-4d64-8fa8-4eb3e2d1a8fc' type:'DestroyImageCheck' to complete
2017-03-10 10:39:23,611+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 4c2fc22d
2017-03-10 10:39:24,616+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterTasksListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f] START,
GlusterTasksListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 1f169371
2017-03-10 10:39:24,618+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] FINISH,
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED,
16.0.0.164:CONNECTED], log id: 4c2fc22d
2017-03-10 10:39:24,629+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 2ac55735
2017-03-10 10:39:24,822+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterTasksListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f] FINISH,
GlusterTasksListVDSCommand, return: [], log id: 1f169371
2017-03-10 10:39:26,836+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 2ac55735
2017-03-10 10:39:31,849+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 2e8dbcd1
2017-03-10 10:39:31,932+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler6) [67e1d8ed] Command 'RemoveSnapshot' (id:
'13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child command id:
'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete
2017-03-10 10:39:31,944+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler6) [4856f570] Command
'RemoveSnapshotSingleDiskLive' (id:
'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id:
'b1d63b8e-19d3-4d64-8fa8-4eb3e2d1a8fc' type:'DestroyImageCheck' to complete
2017-03-10 10:39:33,213+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED,
16.0.0.164:CONNECTED], log id: 2e8dbcd1
2017-03-10 10:39:33,226+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 1fb3f9e3
2017-03-10 10:39:34,375+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 1fb3f9e3
2017-03-10 10:39:39,392+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 1e0b8eeb
2017-03-10 10:39:40,753+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6] FINISH,
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED,
16.0.0.164:CONNECTED], log id: 1e0b8eeb
2017-03-10 10:39:40,763+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 35b04b33
2017-03-10 10:39:41,952+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 35b04b33
2017-03-10 10:39:41,991+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler6) [67e1d8ed] Command 'RemoveSnapshot' (id:
'13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child command id:
'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete
gluster ( nothing in logs )
======
## "etc-glusterfs-glusterd.vol.log"
[2017-03-10 10:13:52.599019] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:16:48.639635] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
The message "I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1" repeated 3 times between
[2017-03-10 10:16:48.639635] and [2017-03-10 10:17:55.659379]
[2017-03-10 10:18:56.875516] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:19:57.204689] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:21:56.576879] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:21:57.772857] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:24:00.617931] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:30:04.918080] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:31:06.128638] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:32:07.325672] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:32:12.433586] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:32:13.544909] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:35:10.039213] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:37:19.905314] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:37:20.174209] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:38:12.635460] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
[2017-03-10 10:40:14.169864] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume slow1
## "rhev-data-center-mnt-glusterSD-localhost:_slow1.log"
[2017-03-10 09:43:40.346785] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61849:
dentry not found in 43e6968f-9c2a-40d8-8074-caf1a36f60cf
[2017-03-10 09:43:40.347076] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61879:
dentry not found in 902a6e3d-b7aa-439f-8262-cdc1b7f9f022
[2017-03-10 09:43:40.347145] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61935:
dentry not found in 846bbcfc-f2b3-4ab6-af44-aeaa10b39318
[2017-03-10 09:43:40.347211] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61922:
dentry not found in 66ad3bc5-26c7-4360-b33b-a084e3305cf8
[2017-03-10 09:43:40.351571] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61834:
dentry not found in 3b8278e1-40e5-4363-b21e-7bffcd024c62
[2017-03-10 09:43:40.352449] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61870:
dentry not found in 282f4c05-e09a-48e0-96a3-52e079ff2f73
[2017-03-10 09:50:38.829325] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)
[2017-03-10 09:50:42.221775] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)
[2017-03-10 09:50:45.956432] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/3cef54b4-45b9-4f5b-82c2-fcc8def06a37/85287865-38f0-45df-9e6c-1294913cbb88.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/3cef54b4-45b9-4f5b-82c2-fcc8def06a37/85287865-38f0-45df-9e6c-1294913cbb88.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)
[2017-03-10 09:50:40.349563] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)
[2017-03-10 09:50:44.503866] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)
[2017-03-10 09:59:46.860762] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/6e105aa3-a3fc-4aca-be50-78b7642c4072.6684:
dentry not found in d1e65eea-8758-4407-ac2e-3605dc661364
[2017-03-10 10:02:22.500865] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/6e105aa3-a3fc-4aca-be50-78b7642c4072.8767:
dentry not found in e228bb28-9602-4f8e-8323-7434d77849fc
[2017-03-10 10:04:03.103839] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/6e105aa3-a3fc-4aca-be50-78b7642c4072.9787:
dentry not found in 6be71632-aa36-4975-b673-1357e0355027
[2017-03-10 10:06:02.406385] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/2a9c1c6a-f045-4dce-a47b-95a2267eef72/6f264695-0669-4b49-a2f6-e6c92482f2fb.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/2a9c1c6a-f045-4dce-a47b-95a2267eef72/6f264695-0669-4b49-a2f6-e6c92482f2fb.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)
... no other record
messages
========
several times occured:
Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring unsupported value for cmd.
Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring unsupported value for cmd.
Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring unsupported value for cmd.
Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring unsupported value for cmd.
Mar 10 09:10:01 2kvm2 systemd: Started Session 274 of user root.
Mar 10 09:10:01 2kvm2 systemd: Starting Session 274 of user root.
Mar 10 09:20:02 2kvm2 systemd: Started Session 275 of user root.
Mar 10 09:20:02 2kvm2 systemd: Starting Session 275 of user root.
Mar 10 09:22:59 2kvm2 sanlock[1673]: 2017-03-10 09:22:59+0100 136031
[2576]: s3 delta_renew long write time 11 sec
Mar 10 09:24:03 2kvm2 kernel: kswapd1: page allocation failure: order:2,
mode:0x104020
Mar 10 09:24:03 2kvm2 kernel: CPU: 42 PID: 265 Comm: kswapd1 Tainted:
G I ------------ 3.10.0-514.10.2.el7.x86_64 #1
Mar 10 09:24:03 2kvm2 kernel: Hardware name: Supermicro
X10DRC/X10DRi-LN4+, BIOS 1.0a 08/29/2014
Mar 10 09:24:03 2kvm2 kernel: 0000000000104020 00000000f7228dc9
ffff88301f4839d8 ffffffff816864ef
Mar 10 09:24:03 2kvm2 kernel: ffff88301f483a68 ffffffff81186ba0
000068fc00000000 0000000000000000
Mar 10 09:24:03 2kvm2 kernel: fffffffffffffffc 0010402000000000
ffff88301567ae80 00000000f7228dc9
Mar 10 09:24:03 2kvm2 kernel: Call Trace:
Mar 10 09:24:03 2kvm2 kernel: <IRQ> [<ffffffff816864ef>]
dump_stack+0x19/0x1b
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81186ba0>]
warn_alloc_failed+0x110/0x180
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81682083>]
__alloc_pages_slowpath+0x6b7/0x725
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8118b155>]
__alloc_pages_nodemask+0x405/0x420
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811cf30a>]
alloc_pages_current+0xaa/0x170
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81185a7e>] __get_free_pages+0xe/0x50
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811dabae>]
kmalloc_order_trace+0x2e/0xa0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811dd381>] __kmalloc+0x221/0x240
Mar 10 09:24:03 2kvm2 kernel: [<ffffffffa02f83fa>]
bnx2x_frag_alloc.isra.62+0x2a/0x40 [bnx2x]
Mar 10 09:24:03 2kvm2 kernel: [<ffffffffa02f92f7>]
bnx2x_rx_int+0x227/0x17b0 [bnx2x]
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81033669>] ? sched_clock+0x9/0x10
Mar 10 09:24:03 2kvm2 kernel: [<ffffffffa02fc72d>]
bnx2x_poll+0x1dd/0x260 [bnx2x]
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff815705e0>] net_rx_action+0x170/0x380
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8108f2cf>] __do_softirq+0xef/0x280
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8169859c>] call_softirq+0x1c/0x30
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8102d365>] do_softirq+0x65/0xa0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8108f665>] irq_exit+0x115/0x120
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81699138>] do_IRQ+0x58/0xf0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8168e2ad>]
common_interrupt+0x6d/0x6d
Mar 10 09:24:03 2kvm2 kernel: <EOI> [<ffffffff81189a73>] ?
free_hot_cold_page+0x103/0x160
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81189b16>]
free_hot_cold_page_list+0x46/0xa0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81195193>]
shrink_page_list+0x543/0xb00
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81195dda>]
shrink_inactive_list+0x1fa/0x630
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81196975>] shrink_lruvec+0x385/0x770
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81196dd6>] shrink_zone+0x76/0x1a0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8119807c>] balance_pgdat+0x48c/0x5e0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81198343>] kswapd+0x173/0x450
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b17d0>] ?
wake_up_atomic_t+0x30/0x30
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811981d0>] ?
balance_pgdat+0x5e0/0x5e0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b06ff>] kthread+0xcf/0xe0
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b0630>] ?
kthread_create_on_node+0x140/0x140
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81696a58>] ret_from_fork+0x58/0x90
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b0630>] ?
kthread_create_on_node+0x140/0x140
Mar 10 09:24:03 2kvm2 kernel: kswapd1: page allocation failure: order:2,
mode:0x104020
Mar 10 09:24:03 2kvm2 kernel: CPU: 42 PID: 265 Comm: kswapd1 Tainted:
G I ------------ 3.10.0-514.10.2.el7.x86_64 #1
Mar 10 09:24:03 2kvm2 kernel: Hardware name: Supermicro
X10DRC/X10DRi-LN4+, BIOS 1.0a 08/29/2014
Mar 10 09:24:03 2kvm2 kernel: 0000000000104020 00000000f7228dc9
ffff88301f4839d8 ffffffff816864ef
and in critical time again
Mar 10 10:37:53 2kvm2 sanlock[1673]: 2017-03-10 10:37:53+0100 140524
[1673]: s3 check_our_lease warning 73 last_success 140451
Mar 10 10:37:54 2kvm2 sanlock[1673]: 2017-03-10 10:37:54+0100 140525
[1673]: s3 check_our_lease warning 74 last_success 140451
Mar 10 10:37:54 2kvm2 wdmd[1732]: test warning now 140526 ping 140516
close 0 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1
Mar 10 10:37:54 2kvm2 kernel: watchdog watchdog0: watchdog did not stop!
Mar 10 10:37:54 2kvm2 wdmd[1732]: /dev/watchdog0 closed unclean
Mar 10 10:37:55 2kvm2 sanlock[1673]: 2017-03-10 10:37:55+0100 140526
[1673]: s3 check_our_lease warning 75 last_success 140451
Mar 10 10:37:55 2kvm2 wdmd[1732]: test warning now 140527 ping 140516
close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1
Mar 10 10:37:56 2kvm2 sanlock[1673]: 2017-03-10 10:37:56+0100 140527
[1673]: s3 check_our_lease warning 76 last_success 140451
Mar 10 10:37:56 2kvm2 wdmd[1732]: test warning now 140528 ping 140516
close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1
Mar 10 10:37:57 2kvm2 sanlock[1673]: 2017-03-10 10:37:57+0100 140528
[1673]: s3 check_our_lease warning 77 last_success 140451
Mar 10 10:37:57 2kvm2 wdmd[1732]: test warning now 140529 ping 140516
close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1
Mar 10 10:37:58 2kvm2 sanlock[1673]: 2017-03-10 10:37:58+0100 140529
[1673]: s3 check_our_lease warning 78 last_success 140451
Mar 10 10:37:58 2kvm2 wdmd[1732]: test warning now 140530 ping 140516
close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1
Mar 10 10:37:59 2kvm2 sanlock[1673]: 2017-03-10 10:37:59+0100 140530
[1673]: s3 check_our_lease warning 79 last_success 140451
Mar 10 10:37:59 2kvm2 wdmd[1732]: test failed rem 55 now 140531 ping
140516 close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1
Mar 10 10:38:00 2kvm2 sanlock[1673]: 2017-03-10 10:38:00+0100 140531
[1673]: s3 check_our_lease failed 80
Mar 10 10:38:00 2kvm2 sanlock[1673]: 2017-03-10 10:38:00+0100 140531
[1673]: s3 all pids clear
Mar 10 10:38:01 2kvm2 wdmd[1732]: /dev/watchdog0 reopen
Mar 10 10:38:10 2kvm2 journal: Cannot start job (query, none) for domain
TEST-LBS_EBSAPP; current job is (query, none) owned by (3284
remoteDispatchConnectGetAllDomainStats, 0 <null>) for (62s, 0s)
Mar 10 10:38:10 2kvm2 journal: Timed out during operation: cannot
acquire state change lock (held by remoteDispatchConnectGetAllDomainStats)
Mar 10 10:38:11 2kvm2 journal: vdsm vds.dispatcher ERROR SSL error
receiving from <yajsonrpc.betterAsyncore.Dispatcher connected ('::1',
40590, 0, 0) at 0x3acdd88>: unexpected eof
Mar 10 10:38:40 2kvm2 journal: Cannot start job (query, none) for domain
TEST1-LBS_ATRYA; current job is (query, none) owned by (3288
remoteDispatchConnectGetAllDomainStats, 0 <null>) for (47s, 0s)
Mar 10 10:38:40 2kvm2 journal: Timed out during operation: cannot
acquire state change lock (held by remoteDispatchConnectGetAllDomainStats)
Mar 10 10:38:41 2kvm2 journal: vdsm vds.dispatcher ERROR SSL error
receiving from <yajsonrpc.betterAsyncore.Dispatcher connected ('::1',
40592, 0, 0) at 0x3fd5b90>: unexpected eof
Mar 10 10:39:10 2kvm2 journal: Cannot start job (query, none) for domain
TEST-LBS_EBSAPP; current job is (query, none) owned by (3284
remoteDispatchConnectGetAllDomainStats, 0 <null>) for (122s, 0s)
Mar 10 10:39:10 2kvm2 journal: Timed out during operation: cannot
acquire state change lock (held by remoteDispatchConnectGetAllDomainStats)
Mar 10 10:39:10 2kvm2 journal: Cannot start job (query, none) for domain
TEST1-LBS_ATRYA; current job is (query, none) owned by (3288
remoteDispatchConnectGetAllDomainStats, 0 <null>) for (77s, 0s)
Mar 10 10:39:10 2kvm2 journal: Timed out during operation: cannot
acquire state change lock (held by remoteDispatchConnectGetAllDomainStats)
Mar 10 10:39:11 2kvm2 journal: vdsm vds.dispatcher ERROR SSL error
receiving from <yajsonrpc.betterAsyncore.Dispatcher connected ('::1',
40594, 0, 0) at 0x2447290>: unexpected eof
Mar 10 10:39:23 2kvm2 sanlock[1673]: 2017-03-10 10:39:23+0100 140615
[2576]: s3 delta_renew write time 140 error -202
Mar 10 10:39:23 2kvm2 sanlock[1673]: 2017-03-10 10:39:23+0100 140615
[2576]: s3 renewal error -202 delta_length 144 last_success 140451
Mar 10 10:39:40 2kvm2 journal: Cannot start job (query, none) for domain
TEST-LBS_EBSAPP; current job is (query, none) owned by (3284
remoteDispatchConnectGetAllDomainStats, 0 <null>) for (152s, 0s)
Mar 10 10:39:40 2kvm2 journal: Timed out during operation: cannot
acquire state change lock (held by remoteDispatchConnectGetAllDomainStats)
Mar 10 10:39:40 2kvm2 journal: Cannot start job (query, none) for domain
TEST1-LBS_ATRYA; current job is (query, none) owned by (3288
remoteDispatchConnectGetAllDomainStats, 0 <null>) for (107s, 0s)
Mar 10 10:39:40 2kvm2 journal: Timed out during operation: cannot
acquire state change lock (held by remoteDispatchConnectGetAllDomainStats)
Mar 10 10:39:41 2kvm2 journal: vdsm vds.dispatcher ERROR SSL error
receiving from <yajsonrpc.betterAsyncore.Dispatcher connected ('::1',
40596, 0, 0) at 0x2472ef0>: unexpected eof
Mar 10 10:39:49 2kvm2 kernel: INFO: task qemu-img:42107 blocked for more
than 120 seconds.
Mar 10 10:39:49 2kvm2 kernel: "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs" disables this message.
Mar 10 10:39:49 2kvm2 kernel: qemu-img D ffff88010dad3e30 0
42107 3631 0x00000080
Mar 10 10:39:49 2kvm2 kernel: ffff88010dad3b30 0000000000000082
ffff8814491f4e70 ffff88010dad3fd8
Mar 10 10:39:49 2kvm2 kernel: ffff88010dad3fd8 ffff88010dad3fd8
ffff8814491f4e70 ffff88301f096c40
Mar 10 10:39:49 2kvm2 kernel: 0000000000000000 7fffffffffffffff
ffff88181f186c00 ffff88010dad3e30
Mar 10 10:39:49 2kvm2 kernel: Call Trace:
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8168bbb9>] schedule+0x29/0x70
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81689609>]
schedule_timeout+0x239/0x2d0
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8168b15e>]
io_schedule_timeout+0xae/0x130
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8168b1f8>] io_schedule+0x18/0x20
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8124d9e5>]
wait_on_sync_kiocb+0x35/0x80
Mar 10 10:39:49 2kvm2 kernel: [<ffffffffa0a36091>]
fuse_direct_IO+0x231/0x380 [fuse]
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff812a6ddd>] ?
cap_inode_need_killpriv+0x2d/0x40
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff812a8cb6>] ?
security_inode_need_killpriv+0x16/0x20
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81219e3f>] ?
dentry_needs_remove_privs.part.13+0x1f/0x30
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81182a2d>]
generic_file_direct_write+0xcd/0x190
Mar 10 10:39:49 2kvm2 kernel: [<ffffffffa0a36905>]
fuse_file_aio_write+0x185/0x340 [fuse]
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff811fdabd>] do_sync_write+0x8d/0xd0
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff811fe32d>] vfs_write+0xbd/0x1e0
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff811ff002>] SyS_pwrite64+0x92/0xc0
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81696b09>]
system_call_fastpath+0x16/0x1b
Mar 10 10:39:49 2kvm2 kernel: INFO: task qemu-img:42111 blocked for more
than 120 seconds.
Mar 10 10:39:49 2kvm2 kernel: "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs" disables this message.
Mar 10 10:39:49 2kvm2 kernel: qemu-img D ffff8818a76e7e30 0
42111 3632 0x00000080
Mar 10 10:39:49 2kvm2 kernel: ffff8818a76e7b30 0000000000000082
ffff88188aaeaf10 ffff8818a76e7fd8
Mar 10 10:39:49 2kvm2 kernel: ffff8818a76e7fd8 ffff8818a76e7fd8
ffff88188aaeaf10 ffff88301f156c40
memory
=======
# cat /proc/meminfo
MemTotal: 197983472 kB
MemFree: 834228 kB
MemAvailable: 165541204 kB
Buffers: 45548 kB
Cached: 159596272 kB
SwapCached: 119872 kB
Active: 40803264 kB
Inactive: 148022076 kB
Active(anon): 26594112 kB
Inactive(anon): 2626384 kB
Active(file): 14209152 kB
Inactive(file): 145395692 kB
Unevictable: 50488 kB
Mlocked: 50488 kB
SwapTotal: 4194300 kB
SwapFree: 3612188 kB
Dirty: 624 kB
Writeback: 0 kB
AnonPages: 29185032 kB
Mapped: 85176 kB
Shmem: 25908 kB
Slab: 6203384 kB
SReclaimable: 5857240 kB
SUnreclaim: 346144 kB
KernelStack: 19184 kB
PageTables: 86100 kB
NFS_Unstable: 0 kB
Bounce: 0 kB
WritebackTmp: 0 kB
CommitLimit: 103186036 kB
Committed_AS: 52300288 kB
VmallocTotal: 34359738367 kB
VmallocUsed: 1560580 kB
VmallocChunk: 34257341440 kB
HardwareCorrupted: 0 kB
AnonHugePages: 5566464 kB
HugePages_Total: 0
HugePages_Free: 0
HugePages_Rsvd: 0
HugePages_Surp: 0
Hugepagesize: 2048 kB
DirectMap4k: 431292 kB
DirectMap2M: 19382272 kB
DirectMap1G: 183500800 kB
can anybody help me with this ??
I've got a small tip about swap problem ( in messages), but not shure .....
The similar problem occured in older versions in gluster/ovirt testing (
in huge workload freeez - but not fatal overload )
regards
Paf1
--------------C04C9E032CD3A8F03847AFCD
Content-Type: text/html; charset=utf-8
Content-Transfer-Encoding: 8bit
<html>
<head>
<meta http-equiv="content-type" content="text/html; charset=utf-8">
</head>
<body bgcolor="#FFFFFF" text="#000066">
<font face="Ubuntu">Hello everybody,<br>
<br>
for production usage i'm testing ovirt with gluster.<br>
All components seems to be running fine but whenever I'm testing
huge workload, then node freez. Not the main OS, but VDSM mgmt and
attached services, VMs eg.<br>
<br>
<b>mgmt </b><br>
oVirt - 4.1.0.4<br>
centos 7.3-1611<br>
<br>
<br>
<b>nodes</b> ( installed from ovirt image <i>"ovirt-node-ng-installer-ovirt-4.1-2017030804.iso"
)</i><br>
</font><br>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">OS Version:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row0_value"
class="GOJECEMBACD"> == RHEL - 7 - 3.1611.el7.centos</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">OS Description:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row1_value"
class="GOJECEMBACD">== oVirt Node 4.1.0</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">Kernel Version:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row2_value"
class="GOJECEMBACD">== 3.10.0 - 514.10.2.el7.x86_64</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">KVM Version:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row3_value"
class="GOJECEMBACD">== 2.6.0 - 28.el7_3.3.1</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">LIBVIRT Version:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row4_value"
class="GOJECEMBACD">== libvirt-2.0.0-10.el7_3.5</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">VDSM Version:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row5_value"
class="GOJECEMBACD">== vdsm-4.19.4-1.el7.centos</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">SPICE Version:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row6_value"
class="GOJECEMBACD">== 0.12.4 - 20.el7_3</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="col-md-10">GlusterFS Version:<span
id="SubTabHostGeneralSoftwareView_formPanel_col0_row7_value"
class="GOJECEMBACD">== glusterfs-3.8.9-1.el7 ( LVM
thinprovisioning in replica 2 - created from ovirt GUI )<br>
</span></div>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-md-12">
<div class="row">
<div class="col-md-2">
<div class="GOJECEMBPBD"
id="SubTabHostGeneralSoftwareView_formPanel_col0_row8_label"><br>
</div>
</div>
</div>
</div>
</div>
<font face="Ubuntu">concurently running<br>
- huge import from export domain ( net workload )<br>
- sequential write to VMs local disk ( gluster replica sequential
workload )<br>
- VMs database huge select ( random IOps )<br>
- huge old snapshot delete ( random IOps )<br>
<br>
In this configuration / workload is runnig one hour eg, with no
exceptions , with 70-80% disk load, but in some point VDSM freez
all jobs for a timeout and VM's are in "uknown" status .<br>
The whole system revitalize then automaticaly in cca 20min time
frame ( except the import and snapshot deleting(rollback) )<br>
<br>
engine.log - focus 10:39:07 time ( </font><font face="Ubuntu"><font
face="Ubuntu">Failed in 'HSMGetAllTasksStatusesVDS' method )<br>
</font>========<br>
<br>
<font size="-1">n child command id:
'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete<br>
2017-03-10 10:39:01,727+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler2) [759c8e1f] Command
'RemoveSnapshotSingleDiskLive' (id:
'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command
id: '33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage'
to complete<br>
2017-03-10 10:39:03,929+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4]
START, GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
446d0cd3<br>
2017-03-10 10:39:04,343+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4]
FINISH, GlusterServersListVDSCommand, return:
[172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id:
446d0cd3<br>
2017-03-10 10:39:04,353+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4]
START, GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
69ea1fda<br>
2017-03-10 10:39:05,128+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4]
FINISH, GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 69ea1fda<br>
2017-03-10 10:39:07,163+01 ERROR
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand]
(DefaultQuartzScheduler2) [759c8e1f] Failed in
'HSMGetAllTasksStatusesVDS' method<br>
2017-03-10 10:39:07,178+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(DefaultQuartzScheduler2) [759c8e1f] EVENT_ID:
VDS_BROKER_COMMAND_FAILURE(10,802), Correlation ID: null, Call
Stack: null, Custom Event ID: -1, Message: VDSM 2kvm2 command
HSMGetAllTasksStatusesVDS failed: Connection timed out<br>
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask]
(DefaultQuartzScheduler2) [759c8e1f]
BaseAsyncTask::onTaskEndSuccess: Task
'f594bf69-619b-4d1b-8f6d-a9826997e478' (Parent Command
'ImportVm', Parameters Type
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters')
ended successfully.<br>
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.CommandMultiAsyncTasks]
(DefaultQuartzScheduler2) [759c8e1f] Task with DB Task ID
'a05c7c07-9b98-4ab2-ac7b-9e70a75ba7b7' and VDSM Task ID
'7c60369f-70a3-4a6a-80c9-4753ac9ed372' is in state Polling. End
action for command 8deb3fe3-4a83-4605-816c-ffdc63fd9ac1 will
proceed when all the entity's tasks are completed.<br>
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask]
(DefaultQuartzScheduler2) [759c8e1f] SPMAsyncTask::PollTask:
Polling task 'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent
Command 'DestroyImage', Parameters Type
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters')
returned status 'finished', result 'cleanSuccess'.<br>
2017-03-10 10:39:07,182+01 ERROR
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask]
(DefaultQuartzScheduler2) [759c8e1f]
BaseAsyncTask::logEndTaskFailure: Task
'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent Command
'DestroyImage', Parameters Type
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters')
ended with failure:<br>
-- Result: 'cleanSuccess'<br>
-- Message: 'VDSGenericException: VDSErrorException: Failed to
HSMGetAllTasksStatusesVDS, error = Connection timed out, code =
100',<br>
-- Exception: 'VDSGenericException: VDSErrorException: Failed to
HSMGetAllTasksStatusesVDS, error = Connection timed out, code =
100'<br>
2017-03-10 10:39:07,184+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(DefaultQuartzScheduler2) [759c8e1f]
CommandAsyncTask::endActionIfNecessary: All tasks of command
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e' has ended -> executing
'endAction'<br>
2017-03-10 10:39:07,185+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(DefaultQuartzScheduler2) [759c8e1f]
CommandAsyncTask::endAction: Ending action for '1' tasks
(command ID: '33df2c1e-6ce3-44fd-a39b-d111883b4c4e'): calling
endAction '.<br>
2017-03-10 10:39:07,185+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::endCommandAction [within thread] context:
Attempting to endAction 'DestroyImage',<br>
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.storage.disk.image.DestroyImageCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] Command
[id=33df2c1e-6ce3-44fd-a39b-d111883b4c4e]: Updating status to
'FAILED', The command end method logic will be executed by one
of its parent commands.<br>
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::HandleEndActionResult [within thread]:
endAction for action type 'DestroyImage' completed, handling the
result.<br>
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::HandleEndActionResult [within thread]:
endAction for action type 'DestroyImage' succeeded, clearing
tasks.<br>
2017-03-10 10:39:07,192+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
SPMAsyncTask::ClearAsyncTask: Attempting to clear task
'f351e8f6-6dd7-49aa-bf54-650d84fc6352'<br>
2017-03-10 10:39:07,193+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] START,
SPMClearTaskVDSCommand(
SPMTaskGuidBaseVDSCommandParameters:{runAsync='true',
storagePoolId='00000001-0001-0001-0001-000000000311',
ignoreFailoverLimit='false',
taskId='f351e8f6-6dd7-49aa-bf54-650d84fc6352'}), log id:
2b7080c2<br>
2017-03-10 10:39:07,194+01 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] START,
HSMClearTaskVDSCommand(HostName = 2kvm2,
HSMTaskGuidBaseVDSCommandParameters:{runAsync='true',
hostId='905375e1-6de4-4fdf-b69c-b2d546f869c8',
taskId='f351e8f6-6dd7-49aa-bf54-650d84fc6352'}), log id:
2edff460<br>
2017-03-10 10:39:08,208+01 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] FINISH,
HSMClearTaskVDSCommand, log id: 2edff460<br>
2017-03-10 10:39:08,208+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMClearTaskVDSCommand]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f] FINISH,
SPMClearTaskVDSCommand, log id: 2b7080c2<br>
2017-03-10 10:39:08,213+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
BaseAsyncTask::removeTaskFromDB: Removed task
'f351e8f6-6dd7-49aa-bf54-650d84fc6352' from DataBase<br>
2017-03-10 10:39:08,213+01 INFO
[org.ovirt.engine.core.bll.tasks.CommandAsyncTask]
(org.ovirt.thread.pool-6-thread-31) [759c8e1f]
CommandAsyncTask::HandleEndActionResult [within thread]:
Removing CommandMultiAsyncTasks object for entity
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e'<br>
2017-03-10 10:39:10,142+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler10)
[a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
2e7278cb<br>
2017-03-10 10:39:11,513+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler10)
[a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c] FINISH,
GlusterServersListVDSCommand, return:
[172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id:
2e7278cb<br>
2017-03-10 10:39:11,523+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler10)
[a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
43704ef2<br>
2017-03-10 10:39:11,777+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler9) [67e1d8ed] Command 'RemoveSnapshot'
(id: '13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child
command id: 'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete<br>
2017-03-10 10:39:11,789+01 WARN
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
(DefaultQuartzScheduler9) [759c8e1f] Child command
'DESTROY_IMAGE' failed, proceeding to verify<br>
2017-03-10 10:39:11,789+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
(DefaultQuartzScheduler9) [759c8e1f] Executing Live Merge
command step 'DESTROY_IMAGE_CHECK'<br>
2017-03-10 10:39:11,832+01 INFO
[org.ovirt.engine.core.bll.DestroyImageCheckCommand]
(pool-5-thread-7) [4856f570] Running command:
DestroyImageCheckCommand internal: true.<br>
2017-03-10 10:39:11,833+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMGetVolumeInfoVDSCommand]
(pool-5-thread-7) [4856f570] START, SPMGetVolumeInfoVDSCommand(
SPMGetVolumeInfoVDSCommandParameters:{expectedEngineErrors='[VolumeDoesNotExist]',
runAsync='true',
storagePoolId='00000001-0001-0001-0001-000000000311',
ignoreFailoverLimit='false',
storageDomainId='1603cd90-92ef-4c03-922c-cecb282fd00e',
imageGroupId='7543338a-3ca6-4698-bb50-c14f0bd71428',
imageId='50b592f7-bfba-4398-879c-8d6a19a2c000'}), log id:
2c8031f8<br>
2017-03-10 10:39:11,833+01 INFO
[org.ovirt.engine.core.vdsbroker.irsbroker.SPMGetVolumeInfoVDSCommand]
(pool-5-thread-7) [4856f570] Executing GetVolumeInfo using the
current SPM<br>
2017-03-10 10:39:11,834+01 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetVolumeInfoVDSCommand]
(pool-5-thread-7) [4856f570] START,
GetVolumeInfoVDSCommand(HostName = 2kvm2,
GetVolumeInfoVDSCommandParameters:{expectedEngineErrors='[VolumeDoesNotExist]',
runAsync='true', hostId='905375e1-6de4-4fdf-b69c-b2d546f869c8',
storagePoolId='00000001-0001-0001-0001-000000000311',
storageDomainId='1603cd90-92ef-4c03-922c-cecb282fd00e',
imageGroupId='7543338a-3ca6-4698-bb50-c14f0bd71428',
imageId='50b592f7-bfba-4398-879c-8d6a19a2c000'}), log id:
79ca86cc<br>
2017-03-10 10:39:11,846+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler10)
[a86dc7b5-52dc-40d4-a3b9-49d7eabbb93c] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 43704ef2<br>
2017-03-10 10:39:16,858+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f]
START, GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
6542adcd<br>
2017-03-10 10:39:17,394+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f]
FINISH, GlusterServersListVDSCommand, return:
[172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id:
6542adcd<br>
2017-03-10 10:39:17,406+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f]
START, GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
44ec33ed<br>
2017-03-10 10:39:18,598+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f]
FINISH, GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 44ec33ed<br>
2017-03-10 10:39:21,865+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler6) [67e1d8ed] Command 'RemoveSnapshot'
(id: '13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child
command id: 'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete<br>
2017-03-10 10:39:21,881+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler6) [4856f570] Command
'RemoveSnapshotSingleDiskLive' (id:
'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command
id: 'b1d63b8e-19d3-4d64-8fa8-4eb3e2d1a8fc'
type:'DestroyImageCheck' to complete<br>
2017-03-10 10:39:23,611+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
4c2fc22d<br>
2017-03-10 10:39:24,616+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterTasksListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f]
START, GlusterTasksListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
1f169371<br>
2017-03-10 10:39:24,618+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] FINISH,
GlusterServersListVDSCommand, return:
[172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id:
4c2fc22d<br>
2017-03-10 10:39:24,629+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
2ac55735<br>
2017-03-10 10:39:24,822+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterTasksListVDSCommand]
(DefaultQuartzScheduler7) [d82701d9-9fa3-467d-b273-f5fe5a93062f]
FINISH, GlusterTasksListVDSCommand, return: [], log id: 1f169371<br>
2017-03-10 10:39:26,836+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler6) [4856f570] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 2ac55735<br>
2017-03-10 10:39:31,849+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4]
START, GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
2e8dbcd1<br>
2017-03-10 10:39:31,932+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler6) [67e1d8ed] Command 'RemoveSnapshot'
(id: '13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child
command id: 'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete<br>
2017-03-10 10:39:31,944+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler6) [4856f570] Command
'RemoveSnapshotSingleDiskLive' (id:
'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command
id: 'b1d63b8e-19d3-4d64-8fa8-4eb3e2d1a8fc'
type:'DestroyImageCheck' to complete<br>
2017-03-10 10:39:33,213+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4]
FINISH, GlusterServersListVDSCommand, return:
[172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id:
2e8dbcd1<br>
2017-03-10 10:39:33,226+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4]
START, GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
1fb3f9e3<br>
2017-03-10 10:39:34,375+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler8) [fde51205-3e8b-4b84-a478-352dc444ccc4]
FINISH, GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 1fb3f9e3<br>
2017-03-10 10:39:39,392+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6]
START, GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
1e0b8eeb<br>
2017-03-10 10:39:40,753+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6]
FINISH, GlusterServersListVDSCommand, return:
[172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id:
1e0b8eeb<br>
2017-03-10 10:39:40,763+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6]
START, GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id:
35b04b33<br>
2017-03-10 10:39:41,952+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler9) [12d6d15f-e054-4833-bd87-58f6a51e5fa6]
FINISH, GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 35b04b33<br>
2017-03-10 10:39:41,991+01 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(DefaultQuartzScheduler6) [67e1d8ed] Command 'RemoveSnapshot'
(id: '13c2cb7c-0809-4971-aceb-37ae66105ab7') waiting on child
command id: 'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete<br>
<br>
</font><br>
gluster ( nothing in logs )<br>
======<br>
<br>
<br>
</font><font face="Ubuntu">## "etc-glusterfs-glusterd.vol.log"</font><font
face="Ubuntu"><font size="-1"><br>
[2017-03-10 10:13:52.599019] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:16:48.639635] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
The message "I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1"
repeated 3 times between [2017-03-10 10:16:48.639635] and
[2017-03-10 10:17:55.659379]<br>
[2017-03-10 10:18:56.875516] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:19:57.204689] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:21:56.576879] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:21:57.772857] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:24:00.617931] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:30:04.918080] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:31:06.128638] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:32:07.325672] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:32:12.433586] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:32:13.544909] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:35:10.039213] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:37:19.905314] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:37:20.174209] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:38:12.635460] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1<br>
[2017-03-10 10:40:14.169864] I [MSGID: 106499]
[glusterd-handler.c:4349:__glusterd_handle_status_volume]
0-management: Received status volume req for volume slow1</font><br>
<br>
<br>
## "rhev-data-center-mnt-glusterSD-localhost:_slow1.log"<br>
<font size="-1">[2017-03-10 09:43:40.346785] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61849:
dentry not found in 43e6968f-9c2a-40d8-8074-caf1a36f60cf<br>
[2017-03-10 09:43:40.347076] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61879:
dentry not found in 902a6e3d-b7aa-439f-8262-cdc1b7f9f022<br>
[2017-03-10 09:43:40.347145] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61935:
dentry not found in 846bbcfc-f2b3-4ab6-af44-aeaa10b39318<br>
[2017-03-10 09:43:40.347211] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61922:
dentry not found in 66ad3bc5-26c7-4360-b33b-a084e3305cf8<br>
[2017-03-10 09:43:40.351571] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61834:
dentry not found in 3b8278e1-40e5-4363-b21e-7bffcd024c62<br>
[2017-03-10 09:43:40.352449] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/b6f2d08d-2441-4111-ab62-e14abdfaf602.61870:
dentry not found in 282f4c05-e09a-48e0-96a3-52e079ff2f73<br>
[2017-03-10 09:50:38.829325] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)<br>
[2017-03-10 09:50:42.221775] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)<br>
[2017-03-10 09:50:45.956432] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/3cef54b4-45b9-4f5b-82c2-fcc8def06a37/85287865-38f0-45df-9e6c-1294913cbb88.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/3cef54b4-45b9-4f5b-82c2-fcc8def06a37/85287865-38f0-45df-9e6c-1294913cbb88.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)<br>
[2017-03-10 09:50:40.349563] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/014ca3aa-d5f5-4b88-8f84-be8d4c5dfc1e/f147532a-89fa-49e0-8225-f82343fca8be.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)<br>
[2017-03-10 09:50:44.503866] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/4cf7dd90-9dcc-428c-82bc-fbf08dbee0be/12812d56-1606-4bf8-a391-0a2cacbd020b.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)<br>
[2017-03-10 09:59:46.860762] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/6e105aa3-a3fc-4aca-be50-78b7642c4072.6684:
dentry not found in d1e65eea-8758-4407-ac2e-3605dc661364<br>
[2017-03-10 10:02:22.500865] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/6e105aa3-a3fc-4aca-be50-78b7642c4072.8767:
dentry not found in e228bb28-9602-4f8e-8323-7434d77849fc<br>
[2017-03-10 10:04:03.103839] W [MSGID: 101159]
[inode.c:1214:__inode_unlink] 0-inode:
be318638-e8a0-4c6d-977d-7a937aa84806/6e105aa3-a3fc-4aca-be50-78b7642c4072.9787:
dentry not found in 6be71632-aa36-4975-b673-1357e0355027<br>
[2017-03-10 10:06:02.406385] I [MSGID: 109066]
[dht-rename.c:1569:dht_rename] 0-slow1-dht: renaming
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/2a9c1c6a-f045-4dce-a47b-95a2267eef72/6f264695-0669-4b49-a2f6-e6c92482f2fb.meta.new
(hash=slow1-replicate-0/cache=slow1-replicate-0) =>
/1603cd90-92ef-4c03-922c-cecb282fd00e/images/2a9c1c6a-f045-4dce-a47b-95a2267eef72/6f264695-0669-4b49-a2f6-e6c92482f2fb.meta
(hash=slow1-replicate-0/cache=slow1-replicate-0)</font><br>
<font size="-1">... no other record</font><br>
<br>
<br>
messages<br>
========<br>
<br>
several times occured:<br>
<br>
<font size="-1">Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring
unsupported value for cmd.<br>
Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring unsupported
value for cmd.<br>
Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring unsupported
value for cmd.<br>
Mar 10 09:04:38 2kvm2 lvmetad: WARNING: Ignoring unsupported
value for cmd.<br>
Mar 10 09:10:01 2kvm2 systemd: Started Session 274 of user root.<br>
Mar 10 09:10:01 2kvm2 systemd: Starting Session 274 of user
root.<br>
Mar 10 09:20:02 2kvm2 systemd: Started Session 275 of user root.<br>
Mar 10 09:20:02 2kvm2 systemd: Starting Session 275 of user
root.<br>
Mar 10 09:22:59 2kvm2 sanlock[1673]: 2017-03-10 09:22:59+0100
136031 [2576]: s3 delta_renew long write time 11 sec<br>
Mar 10 09:24:03 2kvm2 kernel: kswapd1: page allocation failure:
order:2, mode:0x104020<br>
Mar 10 09:24:03 2kvm2 kernel: CPU: 42 PID: 265 Comm: kswapd1
Tainted: G I ------------
3.10.0-514.10.2.el7.x86_64 #1<br>
Mar 10 09:24:03 2kvm2 kernel: Hardware name: Supermicro
X10DRC/X10DRi-LN4+, BIOS 1.0a 08/29/2014<br>
Mar 10 09:24:03 2kvm2 kernel: 0000000000104020 00000000f7228dc9
ffff88301f4839d8 ffffffff816864ef<br>
Mar 10 09:24:03 2kvm2 kernel: ffff88301f483a68 ffffffff81186ba0
000068fc00000000 0000000000000000<br>
Mar 10 09:24:03 2kvm2 kernel: fffffffffffffffc 0010402000000000
ffff88301567ae80 00000000f7228dc9<br>
Mar 10 09:24:03 2kvm2 kernel: Call Trace:<br>
Mar 10 09:24:03 2kvm2 kernel: <IRQ>
[<ffffffff816864ef>] dump_stack+0x19/0x1b<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81186ba0>]
warn_alloc_failed+0x110/0x180<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81682083>]
__alloc_pages_slowpath+0x6b7/0x725<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8118b155>]
__alloc_pages_nodemask+0x405/0x420<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811cf30a>]
alloc_pages_current+0xaa/0x170<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81185a7e>]
__get_free_pages+0xe/0x50<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811dabae>]
kmalloc_order_trace+0x2e/0xa0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811dd381>]
__kmalloc+0x221/0x240<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffffa02f83fa>]
bnx2x_frag_alloc.isra.62+0x2a/0x40 [bnx2x]<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffffa02f92f7>]
bnx2x_rx_int+0x227/0x17b0 [bnx2x]<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81033669>] ?
sched_clock+0x9/0x10<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffffa02fc72d>]
bnx2x_poll+0x1dd/0x260 [bnx2x]<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff815705e0>]
net_rx_action+0x170/0x380<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8108f2cf>]
__do_softirq+0xef/0x280<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8169859c>]
call_softirq+0x1c/0x30<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8102d365>]
do_softirq+0x65/0xa0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8108f665>]
irq_exit+0x115/0x120<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81699138>]
do_IRQ+0x58/0xf0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8168e2ad>]
common_interrupt+0x6d/0x6d<br>
Mar 10 09:24:03 2kvm2 kernel: <EOI>
[<ffffffff81189a73>] ? free_hot_cold_page+0x103/0x160<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81189b16>]
free_hot_cold_page_list+0x46/0xa0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81195193>]
shrink_page_list+0x543/0xb00<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81195dda>]
shrink_inactive_list+0x1fa/0x630<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81196975>]
shrink_lruvec+0x385/0x770<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81196dd6>]
shrink_zone+0x76/0x1a0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff8119807c>]
balance_pgdat+0x48c/0x5e0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81198343>]
kswapd+0x173/0x450<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b17d0>] ?
wake_up_atomic_t+0x30/0x30<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff811981d0>] ?
balance_pgdat+0x5e0/0x5e0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b06ff>]
kthread+0xcf/0xe0<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b0630>] ?
kthread_create_on_node+0x140/0x140<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff81696a58>]
ret_from_fork+0x58/0x90<br>
Mar 10 09:24:03 2kvm2 kernel: [<ffffffff810b0630>] ?
kthread_create_on_node+0x140/0x140<br>
Mar 10 09:24:03 2kvm2 kernel: kswapd1: page allocation failure:
order:2, mode:0x104020<br>
Mar 10 09:24:03 2kvm2 kernel: CPU: 42 PID: 265 Comm: kswapd1
Tainted: G I ------------
3.10.0-514.10.2.el7.x86_64 #1<br>
Mar 10 09:24:03 2kvm2 kernel: Hardware name: Supermicro
X10DRC/X10DRi-LN4+, BIOS 1.0a 08/29/2014<br>
Mar 10 09:24:03 2kvm2 kernel: 0000000000104020 00000000f7228dc9
ffff88301f4839d8 ffffffff816864ef<br>
</font><br>
<br>
and in critical time again<br>
<br>
<font size="-1">Mar 10 10:37:53 2kvm2 sanlock[1673]: 2017-03-10
10:37:53+0100 140524 [1673]: s3 check_our_lease warning 73
last_success 140451<br>
Mar 10 10:37:54 2kvm2 sanlock[1673]: 2017-03-10 10:37:54+0100
140525 [1673]: s3 check_our_lease warning 74 last_success 140451<br>
Mar 10 10:37:54 2kvm2 wdmd[1732]: test warning now 140526 ping
140516 close 0 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1<br>
Mar 10 10:37:54 2kvm2 kernel: watchdog watchdog0: watchdog did
not stop!<br>
Mar 10 10:37:54 2kvm2 wdmd[1732]: /dev/watchdog0 closed unclean<br>
Mar 10 10:37:55 2kvm2 sanlock[1673]: 2017-03-10 10:37:55+0100
140526 [1673]: s3 check_our_lease warning 75 last_success 140451<br>
Mar 10 10:37:55 2kvm2 wdmd[1732]: test warning now 140527 ping
140516 close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1<br>
Mar 10 10:37:56 2kvm2 sanlock[1673]: 2017-03-10 10:37:56+0100
140527 [1673]: s3 check_our_lease warning 76 last_success 140451<br>
Mar 10 10:37:56 2kvm2 wdmd[1732]: test warning now 140528 ping
140516 close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1<br>
Mar 10 10:37:57 2kvm2 sanlock[1673]: 2017-03-10 10:37:57+0100
140528 [1673]: s3 check_our_lease warning 77 last_success 140451<br>
Mar 10 10:37:57 2kvm2 wdmd[1732]: test warning now 140529 ping
140516 close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1<br>
Mar 10 10:37:58 2kvm2 sanlock[1673]: 2017-03-10 10:37:58+0100
140529 [1673]: s3 check_our_lease warning 78 last_success 140451<br>
Mar 10 10:37:58 2kvm2 wdmd[1732]: test warning now 140530 ping
140516 close 140526 renewal 140451 expire 140531 client 1673
sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1<br>
Mar 10 10:37:59 2kvm2 sanlock[1673]: 2017-03-10 10:37:59+0100
140530 [1673]: s3 check_our_lease warning 79 last_success 140451<br>
Mar 10 10:37:59 2kvm2 wdmd[1732]: test failed rem 55 now 140531
ping 140516 close 140526 renewal 140451 expire 140531 client
1673 sanlock_1603cd90-92ef-4c03-922c-cecb282fd00e:1<br>
Mar 10 10:38:00 2kvm2 sanlock[1673]: 2017-03-10 10:38:00+0100
140531 [1673]: s3 check_our_lease failed 80<br>
Mar 10 10:38:00 2kvm2 sanlock[1673]: 2017-03-10 10:38:00+0100
140531 [1673]: s3 all pids clear<br>
Mar 10 10:38:01 2kvm2 wdmd[1732]: /dev/watchdog0 reopen<br>
Mar 10 10:38:10 2kvm2 journal: Cannot start job (query, none)
for domain TEST-LBS_EBSAPP; current job is (query, none) owned
by (3284 remoteDispatchConnectGetAllDomainStats, 0 <null>)
for (62s, 0s)<br>
Mar 10 10:38:10 2kvm2 journal: Timed out during operation:
cannot acquire state change lock (held by
remoteDispatchConnectGetAllDomainStats)<br>
Mar 10 10:38:11 2kvm2 journal: vdsm vds.dispatcher ERROR SSL
error receiving from <yajsonrpc.betterAsyncore.Dispatcher
connected ('::1', 40590, 0, 0) at 0x3acdd88>: unexpected eof<br>
Mar 10 10:38:40 2kvm2 journal: Cannot start job (query, none)
for domain TEST1-LBS_ATRYA; current job is (query, none) owned
by (3288 remoteDispatchConnectGetAllDomainStats, 0 <null>)
for (47s, 0s)<br>
Mar 10 10:38:40 2kvm2 journal: Timed out during operation:
cannot acquire state change lock (held by
remoteDispatchConnectGetAllDomainStats)<br>
Mar 10 10:38:41 2kvm2 journal: vdsm vds.dispatcher ERROR SSL
error receiving from <yajsonrpc.betterAsyncore.Dispatcher
connected ('::1', 40592, 0, 0) at 0x3fd5b90>: unexpected eof<br>
Mar 10 10:39:10 2kvm2 journal: Cannot start job (query, none)
for domain TEST-LBS_EBSAPP; current job is (query, none) owned
by (3284 remoteDispatchConnectGetAllDomainStats, 0 <null>)
for (122s, 0s)<br>
Mar 10 10:39:10 2kvm2 journal: Timed out during operation:
cannot acquire state change lock (held by
remoteDispatchConnectGetAllDomainStats)<br>
Mar 10 10:39:10 2kvm2 journal: Cannot start job (query, none)
for domain TEST1-LBS_ATRYA; current job is (query, none) owned
by (3288 remoteDispatchConnectGetAllDomainStats, 0 <null>)
for (77s, 0s)<br>
Mar 10 10:39:10 2kvm2 journal: Timed out during operation:
cannot acquire state change lock (held by
remoteDispatchConnectGetAllDomainStats)<br>
Mar 10 10:39:11 2kvm2 journal: vdsm vds.dispatcher ERROR SSL
error receiving from <yajsonrpc.betterAsyncore.Dispatcher
connected ('::1', 40594, 0, 0) at 0x2447290>: unexpected eof<br>
Mar 10 10:39:23 2kvm2 sanlock[1673]: 2017-03-10 10:39:23+0100
140615 [2576]: s3 delta_renew write time 140 error -202<br>
Mar 10 10:39:23 2kvm2 sanlock[1673]: 2017-03-10 10:39:23+0100
140615 [2576]: s3 renewal error -202 delta_length 144
last_success 140451<br>
Mar 10 10:39:40 2kvm2 journal: Cannot start job (query, none)
for domain TEST-LBS_EBSAPP; current job is (query, none) owned
by (3284 remoteDispatchConnectGetAllDomainStats, 0 <null>)
for (152s, 0s)<br>
Mar 10 10:39:40 2kvm2 journal: Timed out during operation:
cannot acquire state change lock (held by
remoteDispatchConnectGetAllDomainStats)<br>
Mar 10 10:39:40 2kvm2 journal: Cannot start job (query, none)
for domain TEST1-LBS_ATRYA; current job is (query, none) owned
by (3288 remoteDispatchConnectGetAllDomainStats, 0 <null>)
for (107s, 0s)<br>
Mar 10 10:39:40 2kvm2 journal: Timed out during operation:
cannot acquire state change lock (held by
remoteDispatchConnectGetAllDomainStats)<br>
Mar 10 10:39:41 2kvm2 journal: vdsm vds.dispatcher ERROR SSL
error receiving from <yajsonrpc.betterAsyncore.Dispatcher
connected ('::1', 40596, 0, 0) at 0x2472ef0>: unexpected eof<br>
Mar 10 10:39:49 2kvm2 kernel: INFO: task qemu-img:42107 blocked
for more than 120 seconds.<br>
Mar 10 10:39:49 2kvm2 kernel: "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs" disables this message.<br>
Mar 10 10:39:49 2kvm2 kernel: qemu-img D
ffff88010dad3e30 0 42107 3631 0x00000080<br>
Mar 10 10:39:49 2kvm2 kernel: ffff88010dad3b30 0000000000000082
ffff8814491f4e70 ffff88010dad3fd8<br>
Mar 10 10:39:49 2kvm2 kernel: ffff88010dad3fd8 ffff88010dad3fd8
ffff8814491f4e70 ffff88301f096c40<br>
Mar 10 10:39:49 2kvm2 kernel: 0000000000000000 7fffffffffffffff
ffff88181f186c00 ffff88010dad3e30<br>
Mar 10 10:39:49 2kvm2 kernel: Call Trace:<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8168bbb9>]
schedule+0x29/0x70<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81689609>]
schedule_timeout+0x239/0x2d0<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8168b15e>]
io_schedule_timeout+0xae/0x130<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8168b1f8>]
io_schedule+0x18/0x20<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff8124d9e5>]
wait_on_sync_kiocb+0x35/0x80<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffffa0a36091>]
fuse_direct_IO+0x231/0x380 [fuse]<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff812a6ddd>] ?
cap_inode_need_killpriv+0x2d/0x40<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff812a8cb6>] ?
security_inode_need_killpriv+0x16/0x20<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81219e3f>] ?
dentry_needs_remove_privs.part.13+0x1f/0x30<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81182a2d>]
generic_file_direct_write+0xcd/0x190<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffffa0a36905>]
fuse_file_aio_write+0x185/0x340 [fuse]<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff811fdabd>]
do_sync_write+0x8d/0xd0<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff811fe32d>]
vfs_write+0xbd/0x1e0<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff811ff002>]
SyS_pwrite64+0x92/0xc0<br>
Mar 10 10:39:49 2kvm2 kernel: [<ffffffff81696b09>]
system_call_fastpath+0x16/0x1b<br>
Mar 10 10:39:49 2kvm2 kernel: INFO: task qemu-img:42111 blocked
for more than 120 seconds.<br>
Mar 10 10:39:49 2kvm2 kernel: "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs" disables this message.<br>
Mar 10 10:39:49 2kvm2 kernel: qemu-img D
ffff8818a76e7e30 0 42111 3632 0x00000080<br>
Mar 10 10:39:49 2kvm2 kernel: ffff8818a76e7b30 0000000000000082
ffff88188aaeaf10 ffff8818a76e7fd8<br>
Mar 10 10:39:49 2kvm2 kernel: ffff8818a76e7fd8 ffff8818a76e7fd8
ffff88188aaeaf10 ffff88301f156c40<br>
</font><br>
memory<br>
=======<br>
<br>
# cat /proc/meminfo <br>
<font size="-1">MemTotal: 197983472 kB<br>
MemFree: 834228 kB<br>
MemAvailable: 165541204 kB<br>
Buffers: 45548 kB<br>
Cached: 159596272 kB<br>
SwapCached: 119872 kB<br>
Active: 40803264 kB<br>
Inactive: 148022076 kB<br>
Active(anon): 26594112 kB<br>
Inactive(anon): 2626384 kB<br>
Active(file): 14209152 kB<br>
Inactive(file): 145395692 kB<br>
Unevictable: 50488 kB<br>
Mlocked: 50488 kB<br>
SwapTotal: 4194300 kB<br>
SwapFree: 3612188 kB<br>
Dirty: 624 kB<br>
Writeback: 0 kB<br>
AnonPages: 29185032 kB<br>
Mapped: 85176 kB<br>
Shmem: 25908 kB<br>
Slab: 6203384 kB<br>
SReclaimable: 5857240 kB<br>
SUnreclaim: 346144 kB<br>
KernelStack: 19184 kB<br>
PageTables: 86100 kB<br>
NFS_Unstable: 0 kB<br>
Bounce: 0 kB<br>
WritebackTmp: 0 kB<br>
CommitLimit: 103186036 kB<br>
Committed_AS: 52300288 kB<br>
VmallocTotal: 34359738367 kB<br>
VmallocUsed: 1560580 kB<br>
VmallocChunk: 34257341440 kB<br>
HardwareCorrupted: 0 kB<br>
AnonHugePages: 5566464 kB<br>
HugePages_Total: 0<br>
HugePages_Free: 0<br>
HugePages_Rsvd: 0<br>
HugePages_Surp: 0<br>
Hugepagesize: 2048 kB<br>
DirectMap4k: 431292 kB<br>
DirectMap2M: 19382272 kB<br>
DirectMap1G: 183500800 kB</font><br>
<br>
<br>
can anybody help me with this ??<br>
I've got a small tip about swap problem ( in messages), but not
shure .....<br>
The similar problem occured in older versions in gluster/ovirt
testing ( in huge workload freeez - but not fatal overload )<br>
<br>
regards<br>
Paf1<br>
<br>
<br>
</font>
</body>
</html>
--------------C04C9E032CD3A8F03847AFCD--
3
4
Hi, I have recently installed a new ovirt environment from scratch.
The version is 4.1.0.
I made a 90GB disk for the hosted engine.
When i had installed it and log into the engine i see that it has only
configured 10GB of those 90GB.
So there is 80GB of unconfigured disk unused.
That is fine as long as it had used lvm but it has not meaning i have to
take the engine down to extend the disk?
My questions is:
1. Why does it not configure all disk?
2. Why is it not using lvm?
3. What should i do to extend the disk?
Regards
Christian
2
1
12 Mar '17
Hi All
I had a storage issue with my gluster volumes running under ovirt hosted.
I now cannot start the hosted engine manager vm from "hosted-engine
--vm-start".
I've scoured the net to find a way, but can't seem to find anything
concrete.
Running Centos7, ovirt 4.0 and gluster 3.8.9
How do I recover the engine manager. Im at a loss!
Engine Status = score between nodes was 0 for all, now node 1 is reading
3400, but all others are 0
{"reason": "bad vm status", "health": "bad", "vm": "down", "detail": "down"}
Logs from agent.log
==================
INFO::2017-03-09
19:32:52,600::state_decorators::51::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(check)
Global maintenance detected
INFO::2017-03-09
19:32:52,603::hosted_engine::612::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_vdsm)
Initializing VDSM
INFO::2017-03-09
19:32:54,820::hosted_engine::639::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_storage_images)
Connecting the storage
INFO::2017-03-09
19:32:54,821::storage_server::219::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
Connecting storage server
INFO::2017-03-09
19:32:59,194::storage_server::226::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
Connecting storage server
INFO::2017-03-09
19:32:59,211::storage_server::233::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
Refreshing the storage domain
INFO::2017-03-09
19:32:59,328::hosted_engine::666::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_storage_images)
Preparing images
INFO::2017-03-09
19:32:59,328::image::126::ovirt_hosted_engine_ha.lib.image.Image::(prepare_images)
Preparing images
INFO::2017-03-09
19:33:01,748::hosted_engine::669::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_initialize_storage_images)
Reloading vm.conf from the shared storage domain
INFO::2017-03-09
19:33:01,748::config::206::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config::(refresh_local_conf_file)
Trying to get a fresher copy of vm configuration from the OVF_STORE
WARNING::2017-03-09
19:33:04,056::ovf_store::107::ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(scan)
Unable to find OVF_STORE
ERROR::2017-03-09
19:33:04,058::config::235::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config::(refresh_local_conf_file)
Unable to get vm.conf from OVF_STORE, falling back to initial vm.conf
ovirt-ha-agent logs
================
ovirt-ha-agent
ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config ERROR Unable
to get vm.conf from OVF_STORE, falling back to initial vm.conf
vdsm
======
vdsm vds.dispatcher ERROR SSL error during reading data: unexpected eof
ovirt-ha-broker
============
ovirt-ha-broker cpu_load_no_engine.EngineHealth ERROR Failed to getVmStats:
'pid'
--
Ian Neilsen
Mobile: 0424 379 762
Linkedin: http://au.linkedin.com/in/ianneilsen
Twitter : ineilsen
3
4
This is a multi-part message in MIME format.
------=_001_NextPart562105536544_=----
Content-Type: text/plain;
charset="ISO-8859-1"
Content-Transfer-Encoding: base64
SGksIGFsbA0KICAgICAgICBJIGhhdmUgbm90aWNlZCB0aGF0IGt2bSBzdXBwb3J0cyBob3N0LW9u
bHkgbmV0d29yayBtb2RlLiBTbyBJIHdhbnQgdG8ga25vdyBob3cgdG8gY3JlYXRlIGEgaG9zdC1v
bmx5IHZpbmMgZm9yIGEgdmlydHVhbCBtYWNoaW5lIGluIG92aXJ0Lg0KICAgICAgICBBbnlvbmUg
Y2FuIGhlbHA/IFRoYW5rcyENCg==
------=_001_NextPart562105536544_=----
Content-Type: text/html;
charset="ISO-8859-1"
Content-Transfer-Encoding: quoted-printable
<html><head><meta http-equiv=3D"content-type" content=3D"text/html; charse=
t=3DISO-8859-1"><style>body { line-height: 1.5; }body { font-size: 10.5pt;=
font-family: ????; color: rgb(0, 0, 0); line-height: 1.5; }body { font-si=
ze: 10.5pt; color: rgb(0, 0, 0); line-height: 1.5; }</style></head><body>H=
i, all<div> <span style=3D"font-size: 10.5pt; line-heigh=
t: 1.5; background-color: window;"> I have noticed that kvm s=
upports host-only network mode. So I want to know how to create a host-onl=
y vinc for a virtual machine in ovirt.</span></div><div>  =
;<span style=3D"font-size: 10.5pt; line-height: 1.5; background-color: win=
dow;"> Anyone can help? Thanks!</span></div></body></html>
------=_001_NextPart562105536544_=------
2
1
--------------7FAFD472380337AE680E33F3
Content-Type: text/plain; charset="windows-1252"; format=flowed
Content-Transfer-Encoding: 7bit
It looks that any memory eager running and swap fails afterwards
will write here who's response of this
On 03/10/2017 01:56 PM, users-request(a)ovirt.org wrote:
> Send Users mailing list submissions to
> users(a)ovirt.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
> http://lists.ovirt.org/mailman/listinfo/users
> or, via email, send a message with subject or body 'help' to
> users-request(a)ovirt.org
>
> You can reach the person managing the list at
> users-owner(a)ovirt.org
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Users digest..."
>
>
> Today's Topics:
>
> 1. Error on Node upgrade 2 (FERNANDO FREDIANI)
> 2. Re: oVirt VM backup and restore (Gianluca Cecchi)
> 3. Re: VDSM hang (Nir Soffer)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Fri, 10 Mar 2017 09:37:00 -0300
> From: FERNANDO FREDIANI <fernando.frediani(a)upx.com>
> To: "users(a)ovirt.org" <users(a)ovirt.org>
> Subject: [ovirt-users] Error on Node upgrade 2
> Message-ID: <3a306136-25aa-4533-4267-f1b8be58fd87(a)upx.com>
> Content-Type: text/plain; charset=utf-8; format=flowed
>
> I am not sure if another email I sent went through but has anyone got
> problems when upgrading a running oVirt-node-ng from 4.1.0 to 4.1.1.
>
> Is the only solution a complete reinstall of the node ?
>
> Thanks
>
> Fernando
>
>
>
> ------------------------------
>
> Message: 2
> Date: Fri, 10 Mar 2017 13:52:35 +0100
> From: Gianluca Cecchi <gianluca.cecchi(a)gmail.com>
> To: Juan Hern?ndez <jhernand(a)redhat.com>
> Cc: "users(a)ovirt.org" <users(a)ovirt.org>
> Subject: Re: [ovirt-users] oVirt VM backup and restore
> Message-ID:
> <CAG2kNCwA3-jubtjcz=bzwX+V5-a5V85=HKi5pr=447Ad-tRD4w(a)mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> On Thu, Mar 9, 2017 at 11:23 AM, Juan Hern?ndez <jhernand(a)redhat.com> wrote:
>
>>> Very good point Gialuca, you are right, the 'persist_memorystate'
>> flag
>>> is 'true' by default, and that makes the pause longer. Would you be
>> so
>>> kind to add it to the 'vm_backup.py' example that is part of version
>> 4
>>> of the SDK?
>>>
>>>
>>> https://github.com/oVirt/ovirt-engine-sdk/blob/master/
>> sdk/examples/vm_backup.py#L143-L151
>>> <https://github.com/oVirt/ovirt-engine-sdk/blob/master/
>> sdk/examples/vm_backup.py#L143-L151>
>>> (Note that that Gibhub is just a mirror, the change needs to be
>> submited
>>> using gerrit.ovirt.org <http://gerrit.ovirt.org>).
>>>
>>>
>>>
>>> I already verified (on a 4.1 infra) that changing vm_backup.py
>>> downloaded yesterday from master this way (apart connection paramters):
>>>
>>> $ diff vm_backup.py vm_backup.py.orig
>>> 150d149
>>> < persist_memorystate=False,
>>>
>>> I get the backup result and snapshot is correctly without memory saved
>>> (and no pause at VM side)
>>>
>> [snip]
>> I see that you already have a gerrit.ovirt.org, so it shouldn't be
>> difficult. The initial setup should be like this:
>>
>> $ git config --global user.name youruser
>> $ git config --global user.email your@email
>> $ git clone ssh://youruser@gerrit.ovirt.org:29418/ovirt-engine-sdk
>> $ gitdir=$(git rev-parse --git-dir); scp -p -P 29418
>> youruser@gerrit.ovirt.org:hooks/commit-msg ${gitdir}/hooks/
>>
>> Then, to submit the patch:
>>
>> $ cd ovirt-engine-sdk
>> $ Edit the vm_backup.py file, and make your changes.
>> $ git commit -a -s
>> $ git push origin HEAD:refs/for/master
>>
>>
> Ok. I found the time to try and apparently it worked as expected.
> The master (you? jenkins CI? ;-) should see my change...
> Just learnt (a little...) another thing
>
> sys admins often try to put an eye inside devs field but the reverse seldom
> happens ;-)
>
1
0
Hi everyone,
Tonight we experienced a hardware fault on one of our PHX storage servers.
The faulty server was used to provide storage for multiple production VMs.
Since automatic failover did not happen they became unavailable.
The outage lasted between 09.03.2017 20:36 UTC and 10.03.2017 09:15 UTC
Unavailable services included all of oVirt's CI infrastructure, mailing
lists
and package repositories. Services in other datecenters such as
gerrit.ovirt.org and www.ovirt.org were not affected.
We brought storage back up and this allowed for VMs to be restarted.
If you see tests that failed or didn't run during this period please
re-trigger them.
If there are still persisting issues please report them to the tracker
ticket
that has a more detailed root cause analysis:
https://ovirt-jira.atlassian.net/browse/OVIRT-1244
Sorry for the inconvenience caused. We are working on improving reliability
of the environment to avoid similar incidents from happening in the future.
--
Regards,
Evgheni Dereveanchin
1
0
Dears Engineers,
I hope you are doing great,
I wish you can help me with a tutorial that shows how to take a scheduled
backups for virtual machines running on oVirt node 4 and how to restore
them.
Best regards
8
18