On 27 October 2016 at 10:42, <joost(a)familiealbers.nl> wrote:
Hi All, i have two ovirt / fakevdsm questions.
I am using fakevdsm to test my ovirt 4.0.5 engine.
I have a somewhat awkward setup where i have a relatively large amount of
dc's.
I cannot get nfs master domain to work from the api.
The api code is solid and is used on a ovirt-engine 3.5 production
cluster.
When creating the nfs storage domain i have no issues (it is created
as id
(00000-00000 more zeros here......) but when i try to attach it to a dc it
fails.
please share the logs
I am not sure this is an ovirt of fakevdsm issue. When i manually add
the
nfs storage domain it does get created which is somewhat worrying.
Manually?
My real aim would be to run 500 dc's with 2 hosts and around 5
vms.
You'd have to have a host per DC, cause every DC has an SPM. You meant
cluster or I'm missing something.
Not being able to use the nfs storage master domain is one of the
issues.
To work around this i decided to try and create 1000 dcs with a single
host using local storage.
Allthough this works it looks like the hosts and vms within the dc's are
actually not having their own storage pool.
In our prod environment we run at a much lower scale serveral dc's with
shared gluster storage (shared between the two hosts in the cluster)
This is considered per dc its own storage pool.
In my tests when reaching dc,cluster, host 249 things fail, this is likely
because of the value of 'maxnumberofhostsinstoragepool'. I would expect
this to be bound to the dcs and not the overall server.
In short i expect each dc to be able to run 249 hosts as each dc is its
own storage pool?
I'm not sure I understand the question. However, the limitation is soft
and can
be overridden.
Similarly when i look at the ovirt ui some hosts actually show they
are
running the total amount of running vms.
Yes, when running with vdsm fake you must set this config value as well
```bash
sudo -i -u postgreexport ENGINE_DB=dbname
psql $ENGINE_DB -c "UPDATE vdc_options set option_value = 'true' WHERE
option_name = 'UseHostNameIdentifier';"
```
I've sent a patch to gerrit for that
https://gerrit.ovirt.org/65804
Again i am not sure if i am correct in assuming each dc is its own
storage
pool.
Correct. DC = Storage Pool
Finally.
As this setup is somewhat awkward i would like to tweak the diverse flags
in ovirt engine to be more relaxed about obtaining data from the hosts and
storage domains.
I believe ovirt engine now checks every 3 seconds the host but i am
guessing i can live with every 60 seconds for example, the same would apply
for the storage part. what i am not sure about though is if there is a
golden ratio between these config items.
I.e. there is vdsRefreshRate, if i set this to 60 instead of 3 do i need
to also change the vdsTimeout value which is set to 180 by default?
Change the VdsRefreshRate to 20s - but bare in mind that we have a
limitation when
starting up a VM - its status will show as up only after
the next 20s interval
Please report with your finding on the above tweak
sorru for the lenght of this message.
the most critical items are.
Should i worry about not being eable to create an nfs master domain or is
this a fakevdsm thing
Why are my dcs, for someone reason related to the main storage pool or am
i right in thinking a dc + cluster +host with local storage, is its own
storage pool?
Thanks you so much.
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users