[ovirt-devel] [ovirt-users] Where do you run oVirt? Here are the answers!
Roy Golan
rgolan at redhat.com
Wed Mar 4 07:07:44 UTC 2015
On 03/03/2015 09:43 PM, Eyal Edri wrote:
> very nice!
> great to see the distribution of OSs,
> can give us a hint on where to focus testing/ci/etc...
hint hint - urrmmm urrmm
>
> e.
>
> ----- Original Message -----
>> From: "Sandro Bonazzola" <sbonazzo at redhat.com>
>> To: devel at ovirt.org, Users at ovirt.org
>> Sent: Tuesday, March 3, 2015 1:56:30 PM
>> Subject: [ovirt-users] Where do you run oVirt? Here are the answers!
>>
>> Hi,
>> This is a summary of the 85 response we got to the last month poll. Thanks
>> everyone who answered!
>>
>> Which distribution are you using for running ovirt-engine?
>> Fedora 20 8 9%
>> CentOS 6 52 61%
>> CentOS 7 22 26%
>> Other 3 4%
>>
>> Which distribution are you using for your nodes?
>> Fedora 20 6 7%
>> CentOS 6 40 47%
>> CentOS 7 31 36%
>> oVirt Node 6 7%
>> Other 2 2%
>>
>> In Other: RHEL 6.6, 7.0, 7.1 and a mixed environment of CentOS 6 and 7.
>>
>> Do you use Hosted Engine?
>> Yes 42 49%
>> No 42 49%
>>
>>
>> Would you like to share more info on your datacenter, vms,...? Tell us about
>> it
>> -------------------------------------------------------------------------------
>>
>> oVirt is so AWESOME! I luv it.
>> --
>> "We currently run engine on CentOS 6 as CentOS 7 was not yet supported. We
>> plan on migrating it to a CentOS 7 machine.
>> Our nodes are currently CentOS 6 but are planning to migrate to CentOS 7.
>> (For the nodes a checkbox for each distribution would be better than the
>> radiobutton, as you can have multiple clusters with different
>> distributions)."
>>
>> --
>> FC Storage (Dell md3620f and IBM Blade-S internal SAS storage)
>>
>> --
>> Please provide ceph support and built in backup tools
>>
>> --
>> "3 separate virtual RHEV datacenters, test, dev, prod.
>> Use direct attach fibre channel luns for application storage heavily on VMs
>> to take advantage of snapshot/restore features on our array.
>> Hosted Engine in test and dev. Physical manager in prod. "
>>
>> --
>> "2 nodes running centos 6.6 from SSD (i3 35w 32GB)
>> 1x NFS datastore
>> 1x ISO datastore
>> 1 node running NFS (i3 35w 8GB, Dell PERC sas controller)
>> All 3 nodes connected via 10Gb ethernet
>> Between 9 en 15 VM's depending on my tests
>> Always active
>> - Zimbra
>> - ldap/dhcp
>> - web/php
>> - devel/php
>> - pfsense
>> Develop/test
>> - Hadoop
>> - Openstack
>> - Gluster
>> - ms..."
>>
>> --
>> "- 4 nodes for KVM, 4 nodes for GlusterFS.
>> - 1Gigabit for management and 4Gigabit channel bonding for GlusterFS replica.
>> - Master Storage Domain lives on replica-3 GlustrerFS volume."
>>
>> --
>> "30 hvs NFS storage over infiniband, custom portal for task automation and
>> classroom abstraction via API"
>>
>> --
>> "We use atm local storage.
>> Current running vm count is over 100.
>> I'd like to use EL7 platform in the future, but I'm uncertain how to best
>> upgrade everything with a minimal downtime.
>> we currently run ovirt-engine 3.3.3
>> we will stick with EL platform and not switch to fedora based, because
>> we need the improved stability.
>> we also do not upgrade to dot zero releases as these introduced
>> some breakage in the past (regressions).
>> I hope this gets better with future releases.
>> Keep up the good work!
>> Sven"
>>
>> --
>> "Storage GlusterFS (Virt+Gluster on Nodes), and FreeNAS via NFS"
>>
>> --
>> "- iSCSI dedicated network
>> - 2x Poweredge M1000e chassis (so, 2 x 16 blades)"
>>
>> --
>> Yes it's NIELIT a gov agency to provide various trannig on virtual
>> environment
>>
>> --
>> "Running production engine on CentOS6 with CentOS6 nodes.
>> Test/Staging environemtn based on CentOS7 and CentOS7 nodes, Hosted-engine on
>> iSCSI."
>>
>> --
>> "Mix of Dell, HP, UCS for compute
>> Netapp for NAS, VNX for FC"
>>
>> --
>> "Cloud used for CI purpose, made from about ""old"" 50 desktop PCs (and still
>> growing) with Celerons, i3, i5 and few i7. VMs are ""light"" nodes for
>> Jenkins (2GB-6GB/2-4cores). Some resources are utilized for cloud's services
>> like vpn, zabbix, httpd, etc. As storage we use MooseFS!"
>>
>> --
>> "This is a sample config for the few installes I have performed, but ideal
>> for a small office.
>> 2x nodes - CentOS 6 with SSD boot and 2x 2TB drives and 2 gluster volumes
>> spread over the 2 - 1 for vm storage and 1 for file storage
>> 1x engine (planning on changing to hosted)
>> 5x vms - 2x DNS/DHCP/Management, 1x webserver for intranet, 1x mailserver and
>> 1x Asterisk PBX
>> "
>>
>> --
>> "I think that we really need more troubleshooting tools and guides more than
>> anything. There are various logs, but there is no reason why we
>> shouldn't be publishing some of this information to the engine UI and even
>> automating certain self-healing.
>> The absolute most important feature in my mind is getting the ability to auto
>> start (restart) VMs after certain failures and attempting to unlock
>> disks, etc.. VMware does a tremendous amount of that in order to provide
>> better HA. We need this."
>>
>> --
>> "Have FC only. Using SVC. Behind it now DS4700. Going to have other storages
>> too.
>> This is BYOD.
>> "
>>
>> --
>> "One node cluster with local storage for education, POC etc. at home."
>>
>> --
>> No
>>
>> --
>> Combo glusterfs storage and vm hosted nodes. Will be migrating engine to
>> centos 7 at some point. Wish libgfapi was properly supported now that it's
>> feasible.
>>
>> --
>> 3 x Supermicro A1SAi-2750F nodes (16 GiB RAM + 8 TiB storage + 8x1GiB/s
>> Ethernet each) with hyperconverged GlusterFS (doubling as an NFS/CIFS
>> storage
>> cluster)
>>
>> --
>> "running 15 vms, for univ apps (LMS/SIS/etc), using dual freeIPA (vms),
>> glusterfs (replicated-distributed-10G net)"
>>
>> --
>> "Tried to install hosted engine on Centos 7, but ran into issues and went for
>> Fedora20 instead. Fedora20 installation was pretty much problem free.
>> Using NFS on Synology NAS for vm disks."
>>
>> --
>> "3 clusters, 1 gluster storage cluster, 50 TB total disk space, with 4 hosts
>> all volumes replica 3
>> 2 virtualisation clusters, SandyBridge with 5 hosts, Nehelem with 2 hosts
>> running about 70 mostly Linux clients, hosted engine runs on 3 of the
>> SandyBridge nodes."
>>
>> --
>> I use oVirt since version 3.0. The evolution of this project is outstanding.
>> The biggest unsolved problem is that there is no good backup solution for
>> oVirt.
>>
>> --
>> Modest server with all-in-one oVirt installation.
>>
>> --
>> "oVirt is for Testing actually, I should say validate. We use Xen Open Source
>> and plan to migrate to oVirt during this year."
>>
>> --
>> "4 nodes, 30 VM's (10 are HA), ISO and export domain on NAS, local drives
>> shared over NFS between nodes, one FC primary storage domain 1TB. Engine is
>> a KVM VM on a CentOS host."
>>
>> --
>> "5 hosts, FC"
>>
>> --
>> Development system for working with other Open Source projects.
>>
>> --
>> "1. DATA_DOMAIN - HP P2000 G3 FC
>> 2. Nodes - Intel server boards SB2400BB
>> 3. VMS - On CentOS 6.6"
>>
>> --
>> We are developing VDI solution based on oVirt/KVM/SPICE.
>>
>> --
>> So I answered for my largest oVirt cluster. I actually have several
>>
>> --
>> "Create multiple VM using ovirt in order to provide platform as a service in
>> my campus environment.
>> All the VM(s) created will be used to run web based application for the
>> purpose of final year project presentation.
>> I did this only on 1 physical server as we currently have the limitation on
>> the hardware part."
>>
>> --
>> "small lab environment, ~10 nodes, ~20-30 VMs."
>>
>> --
>> Ovirt 3.4.4 with 16 VDSM Nodes and 170 virtual machines. We currently use
>> Direct attached disks from our ISCSI SAN and we use the snapshot and
>> replication features of the SAN (Dell Equallogic).
>>
>> --
>> "Lab @Home
>> SuperMicro mainboard with 1 x Xeon quad core cpu (Sandy Bridge)
>> 32 GB RAM
>> Synology NAS with storage for oVirt through iSCSI"
>>
>> --
>> "IBM Blade Center.
>> 1 Engine
>> 2 Ovirt Nodes
>> 1 NAS for NFS"
>>
>> --
>> "We started looking at OVIRT a while ago and it had come a long way. My only
>> objection to migrating it into our production is the issues we have with
>> the network interfaces (bonded VLAN tags on the mgmt) and their removal on
>> boot/reboot. Other than that we have fully tested multi-cluster glusterized
>> environments successfully.
>> Again outside of the networks the only other feature I would suggest is
>> gluster storage for the hosted engine. "
>>
>> --
>> Storage is ZFS shared over NFS via dual 10 gbit links. Running 15 nodes now.
>>
>> --
>> We will upgrade up CentOS7 soon
>>
>> --
>> "2 FC datacenters, 4 VNX SAN, 2 4To Lun, 25 hosts on 2 physical sites, 180
>> centos server vms. We used to recycle all our old servers (1950, 2950, G6
>> etc..) to get additional vCPU and RAM, but it has no more interest since we
>> got 4 new r630 chassis with 128 Go of RAM and 40 vCPUs per server. So the
>> goal is to reduce the number of hosts when keeping the same capacities."
>>
>> --
>> "2 production setups in different DC, beginning their history from ovirt-3.2,
>> thus centos6 and no hosted-engine.
>> dell m1000e bladesystem + fc storage in first one; hp dl580 g7 + fc storage
>> in second. ~150 vms in each"
>>
>> --
>> Storage is ZFS shared over NFS via dual 10 gbit links. Running 15 nodes now.
>>
>> --
>> "Just started using Ovirt, we're using supermicro microcloud servers with
>> 32GB memory and quad core E3-1241v3 Xeons.
>> We're using the jenkins builds of ovirt-node since there isn't a recent
>> ovirt-node build. "
>>
>> --
>> Hyorvisor used also as replica glusterfs nodes.
>>
>> --
>> "We run about 20 VMs on 3 proliant machines with a msa backed shared storage.
>> We have migrated dec/jan to oVirt and I am a big fan of the virtualization
>> solution. We ran VPS before on Virtuozzo, but it did not provide out of the
>> box HA.
>> The only thing I was missing is some way of automated backups.
>> We have finally bought a proprietary backup solution ( R1soft CDP ) to run
>> within the VMs to have disaster and file backups.
>> Overall a great product!
>> "
>>
>>
>>
>> --
>> Sandro Bonazzola
>> Better technology. Faster innovation. Powered by community collaboration.
>> See how it works at redhat.com
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
More information about the Devel
mailing list