On Sun, Apr 6, 2014 at 6:32 PM, Gilad Chaplik <gchaplik@redhat.com> wrote:
----- Original Message -----
> From: "Liran Zelkha" <liran.zelkha@gmail.com>
> To: "Kobi Ianko" <kobi@redhat.com>
> Cc: "Gilad Chaplik" <gchaplik@redhat.com>, devel@linode01.ovirt.org, "engine-devel" <engine-devel@ovirt.org>
> Sent: Sunday, April 6, 2014 3:40:13 PM
> Subject: Re: [Devel] [Engine-devel] vds_dynamic refactor
>
> On Sun, Apr 6, 2014 at 3:37 PM, Kobi Ianko <kobi@redhat.com> wrote:
>
> > Joining in...
> > From my point of view, in real life a user should have that many VDSs on
> > one Engine (from a DB point of view).
> > Modern DB system handles tables with millions of records and many
> > relations, Do we really have a performance issue here?
> > We could prefer a more easy to maintain implantation in this case over DB
> > performance
> >
> > Yes we do. We make many queries on the VDS view, which is a VERY complex
> view.
>

Actually I quite agree with Kobi, what is the plan for VMs? why do we start with VDS...
what is the biggest deploy do you know of?
We start with VDS because in an idle system, with 200 hosts and several thousands VMs, this is what you get as the top queries against the database. Look at how many times getvds is called. 
Inline image 1
BTW - the second query is an example of abusing the dynamic query mechanism. The 4th query (an update command) is a set of useless update_vds_dynamic commands. 

For reference, the explain plan of get VDS is something like this:
                                                                        QUERY PLAN                                                                         
-----------------------------------------------------------------------------------------------------------------------------------------------------------
 Nested Loop  (cost=9.30..46.75 rows=6 width=9060) (actual time=0.063..0.068 rows=1 loops=1)
   Join Filter: (vds_static.vds_id = vds_statistics.vds_id)
   ->  Seq Scan on vds_statistics  (cost=0.00..1.01 rows=1 width=109) (actual time=0.008..0.008 rows=1 loops=1)
   ->  Nested Loop  (cost=9.30..45.64 rows=6 width=8983) (actual time=0.048..0.052 rows=1 loops=1)
         Join Filter: (vds_groups.vds_group_id = vds_static.vds_group_id)
         ->  Nested Loop Left Join  (cost=0.00..9.29 rows=1 width=1389) (actual time=0.013..0.013 rows=1 loops=1)
               ->  Seq Scan on vds_groups  (cost=0.00..1.01 rows=1 width=1271) (actual time=0.003..0.003 rows=1 loops=1)
               ->  Index Scan using pk_storage_pool on storage_pool  (cost=0.00..8.27 rows=1 width=134) (actual time=0.008..0.008 rows=1 loops=1)
                     Index Cond: (vds_groups.storage_pool_id = id)
         ->  Hash Right Join  (cost=9.30..36.28 rows=6 width=7610) (actual time=0.033..0.037 rows=1 loops=1)
               Hash Cond: (vds_spm_id_map.vds_id = vds_static.vds_id)
               ->  Seq Scan on vds_spm_id_map  (cost=0.00..22.30 rows=1230 width=20) (actual time=0.003..0.003 rows=1 loops=1)
               ->  Hash  (cost=9.29..9.29 rows=1 width=7606) (actual time=0.019..0.019 rows=1 loops=1)
                     Buckets: 1024  Batches: 1  Memory Usage: 2kB
                     ->  Nested Loop  (cost=0.00..9.29 rows=1 width=7606) (actual time=0.012..0.013 rows=1 loops=1)
                           ->  Seq Scan on vds_dynamic  (cost=0.00..1.01 rows=1 width=1895) (actual time=0.006..0.006 rows=1 loops=1)
                           ->  Index Scan using pk_vds_static on vds_static  (cost=0.00..8.27 rows=1 width=5711) (actual time=0.005..0.006 rows=1 loops=1)
                                 Index Cond: (vds_id = vds_dynamic.vds_id)
 Total runtime: 0.299 ms
(19 rows)

It's terrible. Adding any additional join will make this worse. Please don't add any more tables... 

> >
> > ----- Original Message -----
> > > From: "Gilad Chaplik" <gchaplik@redhat.com>
> > > To: "Liran Zelkha" <liran.zelkha@gmail.com>
> > > Cc: devel@linode01.ovirt.org, "engine-devel" <engine-devel@ovirt.org>
> > > Sent: Sunday, April 6, 2014 3:32:26 PM
> > > Subject: Re: [Devel] [Engine-devel] vds_dynamic refactor
> > >
> > > ----- Original Message -----
> > > > From: "Liran Zelkha" <liran.zelkha@gmail.com>
> > > > To: "Gilad Chaplik" <gchaplik@redhat.com>
> > > > Cc: "Itamar Heim" <iheim@redhat.com>, devel@linode01.ovirt.org,
> > > > "engine-devel" <engine-devel@ovirt.org>
> > > > Sent: Sunday, April 6, 2014 3:26:24 PM
> > > > Subject: Re: [Engine-devel] vds_dynamic refactor
> > > >
> > > > On Sun, Apr 6, 2014 at 3:18 PM, Gilad Chaplik <gchaplik@redhat.com>
> > wrote:
> > > >
> > > > > ----- Original Message -----
> > > > > > From: "Itamar Heim" <iheim@redhat.com>
> > > > > > To: "Liran Zelkha" <liran.zelkha@gmail.com>
> > > > > > Cc: "Gilad Chaplik" <gchaplik@redhat.com>,
> > devel@linode01.ovirt.org,
> > > > > "engine-devel" <engine-devel@ovirt.org>
> > > > > > Sent: Sunday, April 6, 2014 11:33:12 AM
> > > > > > Subject: Re: [Engine-devel] vds_dynamic refactor
> > > > > >
> > > > > > On 04/06/2014 11:32 AM, Liran Zelkha wrote:
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > On Sun, Apr 6, 2014 at 11:22 AM, Itamar Heim <iheim@redhat.com
> > > > > > > <mailto:iheim@redhat.com>> wrote:
> > > > > > >
> > > > > > >     On 04/03/2014 07:51 PM, Liran Zelkha wrote:
> > > > > > >
> > > > > > >         The problem is with both updates and selects.
> > > > > > >         For selects - to get all the information for the VDS we
> > have
> > > > > > >         multiple
> > > > > > >         joins. Adding another one will hurt performance even
> > more.
> > > > > > >         For updates - we have vds_static thats hardly changed.
> > > > > > >         vds_statistics
> > > > > > >         that changes all the time. vds_dynamic is not changed
> > allot -
> > > > > but
> > > > > > >         is
> > > > > > >         updated all the time because of the status. I think it's
> > best
> > > > > to
> > > > > > >         split
> > > > > > >         it to the two existing tables (BTW - relevant for VM as
> > well)
> > > > > > >
> > > > > > >
> > > > > > >     but we don't update it unless the status has changed, which
> > is a
> > > > > > >     rare occurance?
> > > > > > >
> > > > > > > Actually - no. We can definitely see times we are updating
> > > > > > > vds_dynamic
> > > > > > > with no reason at all. I tried to create patches for that - but
> > it
> > > > > > > happens from many different places in the code.
> > > > > >
> > > > > > what would be updated vds_dyanmic for status not originating in
> > update
> > > > > > run time info?
> > > > >
> > > > > We have separate DB flows for that (updateStatus and
> > > > > updatePartialVdsDynamicCalc and more in VdsDynamicDAODbFacadeImpl).
> > > > > A question: do you know if we update status in updateVdsDynamic? :-)
> > not
> > > > > sure but I found a possible race for pending resources (cpu, mem),
> > LOL
> > > > > :-)
> > > > >
> > > > > I think we do but not sure. Will check.
> > >
> > > Of course it is, that was a rhetorical question :-) (a lot of emoticons
> > and
> > > LOLs ;-))
> > >
> > > >
> > > >
> > > > > Still holds my original thought for having vds_on_boot.
> > > > >
> > > > >
> > > > >
> > > > Let's talk f2f on Tuesday?
> > >
> > > I'd prefer to reach conclusions here, I'd like everyone to be involved
> > in a
> > > root issue like this one.
> >
>
> What is the update frequency of this field?
>

which field?
status? pending resources? on boot fields?
iinm, status is updated mostly by user actions, at least in positive scenarios, and not that often.


> > >
> > > >
> > > _______________________________________________
> > > Devel mailing list
> > > Devel@ovirt.org
> > > http://lists.ovirt.org/mailman/listinfo/devel
> > >
> >
>