Hi Thomas, so you are seeing high load on your storage and you are asking 'why'? an answer with the facts you give would be: you are using your storage, so, you have storage IO.

so, if you want to dive deeper:
-which storage are you using, specs would be nice.
-which host model are you using?
-network specs? card model, etc. switch model, etc.

hows your setup made? iscsi? nfs? gluster?

based on the former, we might get a better idea and after this some tests could be made if needed to find if there's a bottleneck or if the environment is working as expected..

regards, 


2018-06-04 14:29 GMT-03:00 Thomas Fecke <thomas.fecke@eset.de>:

Hey Guys,

 

sorry i need to ask again.

 

We got 2 Hypervisor with about 50 running VM´s and a single Storage with 10 Gig connection.

 

 

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util

sda               3,00   694,00 1627,00  947,00 103812,00 61208,00   128,22     6,78    2,63    2,13    3,49   0,39  99,70

 

avg-cpu:  %user   %nice %system %iowait  %steal   %idle

           0,00    0,00    3,70   31,37    0,00   64,93

 

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util

sda               1,00   805,00  836,00  997,00 43916,00 57900,00   111,09     6,00    3,27    1,87    4,44   0,54  99,30

 

avg-cpu:  %user   %nice %system %iowait  %steal   %idle

           0,00    0,00    3,54   29,96    0,00   66,50

 

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util

sda               2,00   822,00 1160,00 1170,00 46700,00 52176,00    84,87     5,68    2,44    1,57    3,30   0,43  99,50

 

avg-cpu:  %user   %nice %system %iowait  %steal   %idle

           0,00    0,00    5,05   31,46    0,00   63,50

 

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util

sda               3,00  1248,00 2337,00 1502,00 134932,00 48536,00    95,58     6,59    1,72    1,53    2,01   0,26  99,30

 

avg-cpu:  %user   %nice %system %iowait  %steal   %idle

           0,00    0,00    3,95   31,79    0,00   64,26

 

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util

sda               0,00   704,00  556,00 1292,00 19908,00 72600,00   100,12     5,50    2,99    1,83    3,48   0,54  99,50

 

avg-cpu:  %user   %nice %system %iowait  %steal   %idle

           0,00    0,00    3,03   28,90    0,00   68,07

 

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util

sda               0,00   544,00  278,00 1095,00  7848,00 66124,00   107,75     5,31    3,87    1,49    4,47   0,72  99,10

 

avg-cpu:  %user   %nice %system %iowait  %steal   %idle

           0,00    0,00    3,03   29,32    0,00   67,65

 

Device:         rrqm/s   wrqm/s     r/s     w/s    rkB/s    wkB/s avgrq-sz avgqu-sz   await r_await w_await  svctm  %util

sda               0,00   464,00  229,00 1172,00  6588,00 72384,00   112,74     5,44    3,88    1,67    4,31   0,71  99,50

 

 

 

 

and this is our Problem. Anyone know why our Storage recive that much of Precesses?

 

Thanks in advance


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/YENR6R4ESX3JCOS7DYA354EOPNL6WGUN/