disk resize question
by Dev Ops
Hey all, quick question regarding new VM instances in oVirt. Is there any way to grow the thin disk associated with a template on instantiation? This would allow facilities like cloud-init resizefs to actually be useful. Been searching around a bit and it seems like this is not a thing in oVirt.
Thanks!!
6 years, 3 months
Re: Unable to mount NFS lun
by Budur Nagaraju
Hi
Am stuck ,can someone help on this?
Thanks,
Nagarau
On Mon, Sep 17, 2018 at 7:10 PM Budur Nagaraju <nbudoor(a)gmail.com> wrote:
> Hi
>
> We have manually mounted the lun in the node and not seeing any issues
> ,while mounting from the UI its failing , by the way one thing to update is
> that nfs lun is created from the Dell unity NFS version3.
>
> Any help can we get ?
>
> Thanks,
> Nagaraju
> '
>
>
>
> On Mon, Sep 17, 2018 at 6:23 PM Fred Rolland <frolland(a)redhat.com> wrote:
>
>> Hi,
>>
>> Check this page for troubleshooting:
>>
>>
>> https://www.ovirt.org/documentation/how-to/troubleshooting/troubleshootin...
>>
>> Regards,
>> Freddy
>>
>> On Mon, Sep 17, 2018 at 1:08 PM, Budur Nagaraju <nbudoor(a)gmail.com>
>> wrote:
>>
>>> Hi
>>>
>>> When I do a NFS mount in the oVirt getting error , below are the logs.
>>>
>>>
>>>
>>> https://pastebin.com/T1ASaB4V
>>>
>>> Thanks,
>>> Nagaraju
>>>
>>>
>>> _______________________________________________
>>> Users mailing list -- users(a)ovirt.org
>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct:
>>> https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives:
>>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/EBK5VIEPTED...
>>>
>>>
>>
6 years, 3 months
clone snapshot of running vm
by p.staniforth@leedsbeckett.ac.uk
Hello,
I am running 4.2.6 and I can't seem to clone a snapshot of vm when it's running, when I click the clone button it gives an
Uncaught exception occurred. Please try reloading the page. Details: Exception caught: (TypeError) : Cannot read property 'K' of null
Please have your administrator check the UI logs
in the UI logs I have
2018-09-11 12:09:11,001+01 ERROR [org.ovirt.engine.ui.frontend.server.gwt.OvirtRemoteLoggingService] (default task-4854) [] Permutation name: 5F972F481E93244BA7B05ED4E850EBA2
2018-09-11 12:09:11,002+01 ERROR [org.ovirt.engine.ui.frontend.server.gwt.OvirtRemoteLoggingService] (default task-4854) [] Uncaught exception: com.google.gwt.event.shared.UmbrellaException: Exception caught: (TypeError) : Cannot read property 'K' of null
at java.lang.Throwable.Throwable(Throwable.java:70) [rt.jar:1.8.0_181]
at java.lang.RuntimeException.RuntimeException(RuntimeException.java:32) [rt.jar:1.8.0_181]
at com.google.web.bindery.event.shared.UmbrellaException.UmbrellaException(UmbrellaException.java:64) [gwt-servlet.jar:]
at com.google.gwt.event.shared.UmbrellaException.UmbrellaException(UmbrellaException.java:25) [gwt-servlet.jar:]
at com.google.gwt.event.shared.HandlerManager.$fireEvent(HandlerManager.java:117) [gwt-servlet.jar:]
at com.google.gwt.user.client.ui.Widget.$fireEvent(Widget.java:127) [gwt-servlet.jar:]
at com.google.gwt.user.client.ui.Widget.fireEvent(Widget.java:127) [gwt-servlet.jar:]
at com.google.gwt.event.dom.client.DomEvent.fireNativeEvent(DomEvent.java:110) [gwt-servlet.jar:]
at com.google.gwt.user.client.ui.Widget.$onBrowserEvent(Widget.java:163) [gwt-servlet.jar:]
at com.google.gwt.user.client.ui.Widget.onBrowserEvent(Widget.java:163) [gwt-servlet.jar:]
at com.google.gwt.user.client.DOM.dispatchEvent(DOM.java:1415) [gwt-servlet.jar:]
at com.google.gwt.user.client.impl.DOMImplStandard.dispatchEvent(DOMImplStandard.java:312) [gwt-servlet.jar:]
at com.google.gwt.core.client.impl.Impl.apply(Impl.java:236) [gwt-servlet.jar:]
at com.google.gwt.core.client.impl.Impl.entry0(Impl.java:275) [gwt-servlet.jar:]
at Unknown.eval(webadmin-0.js)
Caused by: com.google.gwt.core.client.JavaScriptException: (TypeError) : Cannot read property 'K' of null
at org.ovirt.engine.ui.uicommonweb.models.vms.ExistingVmModelBehavior.updateHaAvailability(ExistingVmModelBehavior.java:481)
at org.ovirt.engine.ui.uicommonweb.models.vms.UnitVmModel.eventRaised(UnitVmModel.java:1933)
at org.ovirt.engine.ui.uicompat.Event.$raise(Event.java:99)
at org.ovirt.engine.ui.uicommonweb.models.ListModel.$setSelectedItem(ListModel.java:82)
at org.ovirt.engine.ui.uicommonweb.models.ListModel.setSelectedItem(ListModel.java:78)
at org.ovirt.engine.ui.common.editor.UiCommonEditorVisitor.$updateListEditor(UiCommonEditorVisitor.java:193)
at org.ovirt.engine.ui.common.editor.UiCommonEditorVisitor.visit(UiCommonEditorVisitor.java:47)
at com.google.gwt.editor.client.impl.AbstractEditorContext.$traverse(AbstractEditorContext.java:127) [gwt-servlet.jar:]
at org.ovirt.engine.ui.common.widget.uicommon.popup.AbstractVmPopupWidget_UiCommonModelEditorDelegate.accept(AbstractVmPopupWidget_UiCommonModelEditorDelegate.java:502)
at com.google.gwt.editor.client.impl.AbstractEditorContext.$traverse(AbstractEditorContext.java:127) [gwt-servlet.jar:]
at org.ovirt.engine.ui.common.widget.uicommon.popup.AbstractVmPopupWidget_DriverImpl.accept(AbstractVmPopupWidget_DriverImpl.java:4)
at org.ovirt.engine.ui.common.editor.AbstractUiCommonModelEditorDriver.$edit(AbstractUiCommonModelEditorDriver.java:32)
at org.ovirt.engine.ui.common.widget.uicommon.popup.AbstractVmPopupWidget.$edit(AbstractVmPopupWidget.java:1518)
at org.ovirt.engine.ui.common.widget.uicommon.popup.AbstractVmPopupWidget.edit(AbstractVmPopupWidget.java:1518)
at org.ovirt.engine.ui.common.widget.uicommon.popup.AbstractVmPopupWidget.edit(AbstractVmPopupWidget.java:1518)
at org.ovirt.engine.ui.common.widget.uicommon.popup.AbstractModeSwitchingPopupWidget.edit(AbstractModeSwitchingPopupWidget.java:80)
at org.ovirt.engine.ui.common.view.popup.AbstractModelBoundWidgetPopupView.edit(AbstractModelBoundWidgetPopupView.java:37)
at org.ovirt.engine.ui.common.presenter.AbstractModelBoundPopupPresenterWidget.$init(AbstractModelBoundPopupPresenterWidget.java:105)
at org.ovirt.engine.ui.common.widget.popup.AbstractVmBasedPopupPresenterWidget.$init(AbstractVmBasedPopupPresenterWidget.java:63)
at org.ovirt.engine.ui.common.widget.popup.AbstractVmBasedPopupPresenterWidget.init(AbstractVmBasedPopupPresenterWidget.java:63)
at org.ovirt.engine.ui.common.widget.popup.AbstractVmBasedPopupPresenterWidget.init(AbstractVmBasedPopupPresenterWidget.java:63)
at org.ovirt.engine.ui.common.uicommon.model.ModelBoundPopupHandler.$handleWindowModelChange(ModelBoundPopupHandler.java:116)
at org.ovirt.engine.ui.common.uicommon.model.ModelBoundPopupHandler.$lambda$0(ModelBoundPopupHandler.java:80)
at org.ovirt.engine.ui.common.uicommon.model.ModelBoundPopupHandler$lambda$0$Type.eventRaised(ModelBoundPopupHandler.java:80)
at org.ovirt.engine.ui.uicompat.Event.$raise(Event.java:99)
at org.ovirt.engine.ui.uicommonweb.models.Model.$onPropertyChanged(Model.java:470)
at org.ovirt.engine.ui.uicommonweb.models.Model.onPropertyChanged(Model.java:470)
at org.ovirt.engine.ui.uicommonweb.models.Model.$setWindow(Model.java:99)
at org.ovirt.engine.ui.uicommonweb.models.vms.VmSnapshotListModel.$cloneVM(VmSnapshotListModel.java:652)
at org.ovirt.engine.ui.uicommonweb.models.vms.VmSnapshotListModel.executeCommand(VmSnapshotListModel.java:798)
at org.ovirt.engine.ui.uicommonweb.UICommand.$execute(UICommand.java:163)
at org.ovirt.engine.ui.common.widget.action.UiCommandButtonDefinition.onClick(UiCommandButtonDefinition.java:125)
at org.ovirt.engine.ui.common.presenter.ActionPanelPresenterWidget.$lambda$2(ActionPanelPresenterWidget.java:140)
at org.ovirt.engine.ui.common.presenter.ActionPanelPresenterWidget$lambda$2$Type.onClick(ActionPanelPresenterWidget.java:140)
at com.google.gwt.event.dom.client.ClickEvent.dispatch(ClickEvent.java:55) [gwt-servlet.jar:]
at com.google.gwt.event.shared.GwtEvent.dispatch(GwtEvent.java:76) [gwt-servlet.jar:]
at com.google.web.bindery.event.shared.SimpleEventBus.$doFire(SimpleEventBus.java:173) [gwt-servlet.jar:]
... 11 more
Regards,
Paul S.
6 years, 3 months
ansible ovirt-host-deploy error on master
by Eitan Raviv
Hi,
On trying to connect a host to engine I get install failed with ansible
ovirt-host-deploy reporting "Could not find or access
/etc/pki/vdsm/libvirt-spice/ca-cert.pem" (log snippet attached) although
said directory holds all files with what looks like correct permissions
(and no manual intervention has been done there).
-rw-r--r--. 1 root kvm 1368 Sep 13 15:09 ca-cert.pem
-rw-r--r--. 1 root kvm 5074 Sep 16 13:37 server-cert.pem
-r--r-----. 1 vdsm kvm 1704 Sep 16 13:37 server-key.pem
setup:
python2-ovirt-host-deploy-1.8.0-0.0.master.20180624095234.git827d6d1.el7.noarch
vdsm-4.30.0-585.gitd4d043e.el7.x86_64
ovirt-engine: master latest (dev build from today)
Any help much appreciated.
6 years, 3 months
Single Hyperconverged Node Gluster Config
by Jeremy Tourville
Hello,
I am trying to setup a single hyperconverged node. The disks that I will be using for the creation of my engine and all VMs are on a RAID 5. This volume is using hardware raid with an LSI controller (LSI 9361-4i). I am unsure of the appropriate values to use for the Gluster config. Here is the info about my environment.
[root@vmh ~]# mount
sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime,seclabel)
proc on /proc type proc (rw,nosuid,nodev,noexec,relatime)
devtmpfs on /dev type devtmpfs (rw,nosuid,seclabel,size=65713924k,nr_inodes=16428481,mode=755)
securityfs on /sys/kernel/security type securityfs (rw,nosuid,nodev,noexec,relatime)
tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev,seclabel)
devpts on /dev/pts type devpts (rw,nosuid,noexec,relatime,seclabel,gid=5,mode=620,ptmxmode=000)
tmpfs on /run type tmpfs (rw,nosuid,nodev,seclabel,mode=755)
tmpfs on /sys/fs/cgroup type tmpfs (ro,nosuid,nodev,noexec,seclabel,mode=755)
cgroup on /sys/fs/cgroup/systemd type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,xattr,release_agent=/usr/lib/systemd/systemd-cgroups-agent,name=systemd)
pstore on /sys/fs/pstore type pstore (rw,nosuid,nodev,noexec,relatime)
cgroup on /sys/fs/cgroup/freezer type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,freezer)
cgroup on /sys/fs/cgroup/pids type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,pids)
cgroup on /sys/fs/cgroup/cpuset type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,cpuset)
cgroup on /sys/fs/cgroup/cpu,cpuacct type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,cpuacct,cpu)
cgroup on /sys/fs/cgroup/blkio type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,blkio)
cgroup on /sys/fs/cgroup/hugetlb type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,hugetlb)
cgroup on /sys/fs/cgroup/net_cls,net_prio type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,net_prio,net_cls)
cgroup on /sys/fs/cgroup/devices type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,devices)
cgroup on /sys/fs/cgroup/perf_event type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,perf_event)
cgroup on /sys/fs/cgroup/memory type cgroup (rw,nosuid,nodev,noexec,relatime,seclabel,memory)
configfs on /sys/kernel/config type configfs (rw,relatime)
/dev/mapper/onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 on / type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered)
selinuxfs on /sys/fs/selinux type selinuxfs (rw,relatime)
systemd-1 on /proc/sys/fs/binfmt_misc type autofs (rw,relatime,fd=22,pgrp=1,timeout=0,minproto=5,maxproto=5,direct,pipe_ino=43386)
debugfs on /sys/kernel/debug type debugfs (rw,relatime)
mqueue on /dev/mqueue type mqueue (rw,relatime,seclabel)
hugetlbfs on /dev/hugepages type hugetlbfs (rw,relatime,seclabel)
hugetlbfs on /dev/hugepages1G type hugetlbfs (rw,relatime,seclabel,pagesize=1G)
/dev/mapper/onn_vmh-tmp on /tmp type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered)
/dev/mapper/onn_vmh-home on /home type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered)
/dev/sdb1 on /boot type ext4 (rw,relatime,seclabel,data=ordered)
/dev/mapper/onn_vmh-var on /var type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered)
/dev/mapper/onn_vmh-var_log on /var/log type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered)
/dev/mapper/onn_vmh-var_crash on /var/crash type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered)
/dev/mapper/onn_vmh-var_log_audit on /var/log/audit type ext4 (rw,relatime,seclabel,discard,stripe=16,data=ordered)
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime)
/dev/mapper/3600605b00a2faca222fb4da81ac9bdb1p1 on /data type ext4 (rw,relatime,seclabel,discard,stripe=64,data=ordered)
tmpfs on /run/user/0 type tmpfs (rw,nosuid,nodev,relatime,seclabel,size=13149252k,mode=700)
[root@vmh ~]# blkid
/dev/sda1: UUID="ce130131-c457-46a0-b6de-e50cc89a6da3" TYPE="ext4" PARTUUID="5fff73ae-70d4-4697-a307-5f68a4c00f4c"
/dev/sdb1: UUID="47422043-e5d0-4541-86ff-193f61a779b0" TYPE="ext4"
/dev/sdb2: UUID="2f2cf71b-b68e-985a-6433-ed1889595df0" UUID_SUB="a3e32abe-109c-b436-7712-6c9d9f1d57c9" LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member"
/dev/sdc1: UUID="2f2cf71b-b68e-985a-6433-ed1889595df0" UUID_SUB="7132acc1-e210-f645-5df2-e1a1eff7f836" LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member"
/dev/md127: UUID="utB9xu-zva6-j5Ci-3E49-uDya-g2AJ-MQu5Rd" TYPE="LVM2_member"
/dev/mapper/onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1: UUID="8656ff1e-d217-4088-b353-6d2b9f602ce3" TYPE="ext4"
/dev/mapper/onn_vmh-swap: UUID="57a165d1-116e-4e64-a694-2618ffa3a79e" TYPE="swap"
/dev/mapper/3600605b00a2faca222fb4da81ac9bdb1p1: UUID="ce130131-c457-46a0-b6de-e50cc89a6da3" TYPE="ext4" PARTUUID="dac9e1fc-b0d7-43da-b52c-66bb059d8137"
/dev/mapper/onn_vmh-root: UUID="7cc65568-d408-43ab-a793-b6c110d7ba98" TYPE="ext4"
/dev/mapper/onn_vmh-home: UUID="c7e344d0-f401-4504-b52e-9b5c6023c10e" TYPE="ext4"
/dev/mapper/onn_vmh-tmp: UUID="3323e010-0dab-4166-b15d-d739a09b4c03" TYPE="ext4"
/dev/mapper/onn_vmh-var: UUID="c20af647-364d-4a11-8776-8e6aac362425" TYPE="ext4"
/dev/mapper/onn_vmh-var_log: UUID="eb2e159a-4d50-411d-912c-c8e11f297fea" TYPE="ext4"
/dev/mapper/onn_vmh-var_log_audit: UUID="65aa85b9-c053-4244-b3b0-c5f91de8ad59" TYPE="ext4"
/dev/mapper/onn_vmh-var_crash: UUID="579bbef5-bab6-49ce-b8ed-ab92afa85ea8" TYPE="ext4"
/dev/mapper/3600605b00a2faca222fb4da81ac9bdb1: PTTYPE="gpt"
[root@vmh ~]# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 8.2T 0 disk
├─sda1 8:1 0 8.2T 0 part
└─3600605b00a2faca222fb4da81ac9bdb1 253:5 0 8.2T 0 mpath
└─3600605b00a2faca222fb4da81ac9bdb1p1 253:6 0 8.2T 0 part /data
sdb 8:16 0 223.6G 0 disk
├─sdb1 8:17 0 1G 0 part /boot
└─sdb2 8:18 0 222.6G 0 part
└─md127 9:127 0 222.5G 0 raid1
├─onn_vmh-pool00_tmeta 253:0 0 1G 0 lvm
│ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm
│ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm /
│ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm
│ ├─onn_vmh-root 253:8 0 146.6G 0 lvm
│ ├─onn_vmh-home 253:9 0 1G 0 lvm /home
│ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp
│ ├─onn_vmh-var 253:11 0 15G 0 lvm /var
│ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log
│ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit
│ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash
├─onn_vmh-pool00_tdata 253:1 0 173.6G 0 lvm
│ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm
│ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm /
│ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm
│ ├─onn_vmh-root 253:8 0 146.6G 0 lvm
│ ├─onn_vmh-home 253:9 0 1G 0 lvm /home
│ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp
│ ├─onn_vmh-var 253:11 0 15G 0 lvm /var
│ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log
│ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit
│ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash
└─onn_vmh-swap 253:4 0 4G 0 lvm [SWAP]
sdc 8:32 0 223.6G 0 disk
└─sdc1 8:33 0 222.6G 0 part
└─md127 9:127 0 222.5G 0 raid1
├─onn_vmh-pool00_tmeta 253:0 0 1G 0 lvm
│ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm
│ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm /
│ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm
│ ├─onn_vmh-root 253:8 0 146.6G 0 lvm
│ ├─onn_vmh-home 253:9 0 1G 0 lvm /home
│ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp
│ ├─onn_vmh-var 253:11 0 15G 0 lvm /var
│ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log
│ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit
│ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash
├─onn_vmh-pool00_tdata 253:1 0 173.6G 0 lvm
│ └─onn_vmh-pool00-tpool 253:2 0 173.6G 0 lvm
│ ├─onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 253:3 0 146.6G 0 lvm /
│ ├─onn_vmh-pool00 253:7 0 173.6G 0 lvm
│ ├─onn_vmh-root 253:8 0 146.6G 0 lvm
│ ├─onn_vmh-home 253:9 0 1G 0 lvm /home
│ ├─onn_vmh-tmp 253:10 0 1G 0 lvm /tmp
│ ├─onn_vmh-var 253:11 0 15G 0 lvm /var
│ ├─onn_vmh-var_log 253:12 0 8G 0 lvm /var/log
│ ├─onn_vmh-var_log_audit 253:13 0 2G 0 lvm /var/log/audit
│ └─onn_vmh-var_crash 253:14 0 10G 0 lvm /var/crash
└─onn_vmh-swap 253:4 0 4G 0 lvm [SWAP]
[root@vmh ~]# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/onn_vmh-ovirt--node--ng--4.2.5.1--0.20180821.0+1 145G 1.9G 135G 2% /
devtmpfs 63G 0 63G 0% /dev
tmpfs 63G 4.0K 63G 1% /dev/shm
tmpfs 63G 19M 63G 1% /run
tmpfs 63G 0 63G 0% /sys/fs/cgroup
/dev/mapper/onn_vmh-tmp 976M 2.8M 906M 1% /tmp
/dev/mapper/onn_vmh-home 976M 2.6M 907M 1% /home
/dev/sdb1 976M 204M 706M 23% /boot
/dev/mapper/onn_vmh-var 15G 43M 14G 1% /var
/dev/mapper/onn_vmh-var_log 7.8G 58M 7.3G 1% /var/log
/dev/mapper/onn_vmh-var_crash 9.8G 37M 9.2G 1% /var/crash
/dev/mapper/onn_vmh-var_log_audit 2.0G 7.7M 1.8G 1% /var/log/audit
/dev/mapper/3600605b00a2faca222fb4da81ac9bdb1p1 8.2T 90M 7.8T 1% /data
tmpfs 13G 0 13G 0% /run/user/0
[root@vmh ~]# lspci | grep RAID
3b:00.0 RAID bus controller: LSI Logic / Symbios Logic MegaRAID SAS-3 3108 [Invader] (rev 02)
[root@vmh ~]# lsmod | grep mega
megaraid_sas 145373 2
[root@vmh ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: enp96s0f0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000
link/ether 0c:c4:7a:f9:b9:88 brd ff:ff:ff:ff:ff:ff
inet 172.30.50.3/24 brd 172.30.50.255 scope global noprefixroute dynamic enp96s0f0
valid_lft 47994sec preferred_lft 47994sec
inet6 fe80::119b:61e4:f9:1dca/64 scope link noprefixroute
valid_lft forever preferred_lft forever
3: enp96s0f1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP group default qlen 1000
link/ether 0c:c4:7a:f9:b9:89 brd ff:ff:ff:ff:ff:ff
inet 172.30.50.4/24 brd 172.30.50.255 scope global noprefixroute dynamic enp96s0f1
valid_lft 47994sec preferred_lft 47994sec
inet6 fe80::498f:1c33:7de6:1fd/64 scope link noprefixroute
valid_lft forever preferred_lft forever
19: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group default qlen 1000
link/ether 2e:78:2c:51:c8:12 brd ff:ff:ff:ff:ff:ff
In DNS I have defined .3 address as vmh.cyber-range.lan and .4 is defined as ovirtbe.cyber-range.lan. .3 is the frontend IP and .4 is the backednd IP.
I have done the following so far with my Gluster Config file, I'm sure it needs further edits.
[hosts]
vmh.cyber-range.lan
[disktype]
jbod #Possible values raid6, raid10, raid5, jbod
[diskcount]
@NUMBER_OF_DATA_DISKS@ #Ignored in case of jbod
[stripesize]
256 #256 in case of jbod
[script1]
action=execute
ignore_script_errors=no
file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d @DEVICE@
#[vdo] # Note: Uncomment if dedupe & compression needs to be enabled on device. Needs kmod-vdo module
#action=create
#names=@VDO_DEVICE_Name@
#devices=@DEVICE@
#logicalsize=@logical_size@T # Note:logicalsize is 10x physical space on disk
##slabsize=32G # Note: used only when the physical size is few TBs
#blockmapcachesize=128M
#readcache=enabled
#readcachesize=20M
#emulate512=enabled
#writepolicy=auto
[pv]
action=create
devices=@DEVICE@ # Change to @VDO_DEVICE_name@ if using vdo
[vg1]
action=create
vgname=gluster_vg1
pvname=@DEVICE@ # Change to @VDO_DEVICE_name@ if using vdo
[lv1]
action=create
vgname=gluster_vg1
lvname=engine_lv
lvtype=thick
size=100GB
mount=/gluster_bricks/engine
[lv2]
action=create
vgname=gluster_vg1
poolname=lvthinpool
lvtype=thinpool
poolmetadatasize=7.7TB
size=7.7TB #For example: 18000GB, depending on device capacity. Units to be specified.
[lv3]
action=create
lvname=lv_vmdisks
poolname=lvthinpool
vgname=gluster_vg1
lvtype=thinlv
mount=/gluster_bricks/vmstore
virtualsize=@SIZE@ # Units to be specified, for instance 5000GB
[lv4]
action=create
lvname=lv_datadisks
poolname=lvthinpool
vgname=gluster_vg1
lvtype=thinlv
mount=/gluster_bricks/data
virtualsize=@SIZE@ # Units to be specified, for instance 5000GB
#[lv5]
#action=setup-cache
#ssd=@SSD_DEVICE@
#vgname=gluster_vg1
#poolname=lvthinpool
#cache_lv=lvcache
#cache_lvsize=5GB # Provide device size
## cachemode=writeback
[shell2]
action=execute
command=vdsm-tool configure --force
[script3]
action=execute
file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh
ignore_script_errors=no
[selinux]
yes
[service3]
action=restart
service=glusterd
slice_setup=yes
[firewalld]
action=add
ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp
services=glusterfs
[script2]
action=execute
file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh
[shell3]
action=execute
command=usermod -a -G gluster qemu
[volume]
action=create
volname=engine
transport=tcp
key=storage.owner-uid,storage.owner-gid,features.shard,performance.low-prio-threads,performance.strict-o-direct,network.remote-dio,network.ping-timeout,user.cifs,nfs.disable,performance.quick-read,performance.read-ahead,performance.io-cache,cluster.eager-lock
value=36,36,on,32,on,off,30,off,on,off,off,off,enable
brick_dirs=/gluster_bricks/engine/engine
ignore_volume_errors=no
[volume2]
action=create
volname=vmstore
transport=tcp
key=storage.owner-uid,storage.owner-gid,features.shard,performance.low-prio-threads,performance.strict-o-direct,network.remote-dio,network.ping-timeout,user.cifs,nfs.disable,performance.quick-read,performance.read-ahead,performance.io-cache,cluster.eager-lock
value=36,36,on,32,on,off,30,off,on,off,off,off,enable
brick_dirs=/gluster_bricks/vmstore/vmstore
ignore_volume_errors=no
[volume3]
action=create
volname=data
transport=tcp
key=storage.owner-uid,storage.owner-gid,features.shard,performance.low-prio-threads,performance.strict-o-direct,network.remote-dio,network.ping-timeout,user.cifs,nfs.disable,performance.quick-read,performance.read-ahead,performance.io-cache,cluster.eager-lock
value=36,36,on,32,on,off,30,off,on,off,off,off,enable
brick_dirs=/gluster_bricks/data/data
ignore_volume_errors=no
Can anyone kindly provide the correct values and corrections needed? I am not nearly as good with Linux storage as I'd like to be. Please let me know if any further info is required. Many thanks in advance!!!
6 years, 3 months
NFS Multipathing
by Thomas Letherby
Hello all,
I've been looking around but I've not found anything definitive on whether
Ovirt can do NFS Multipathing, and if so how?
Does anyone have any good how tos or configuration guides?
Thanks,
Thomas
6 years, 3 months
Re: NFS Multipathing
by spfma.tech@e.mail.fr
Hi, It should be possible, as oVirt is able to support NFS 4.1 I have a Synology NAS which is also able to support this version of the protocol, but never found time to set this together and test it until now. Reagrds
Le 30-Aug-2018 12:16:32 +0200, xrs444(a)xrs444.net a crit:
Hello all, I've been looking around but I've not found anything definitive on whether Ovirt can do NFS Multipathing, and if so how? Does anyone have any good how tos or configuration guides? Thanks, Thomas
-------------------------------------------------------------------------------------------------
FreeMail powered by mail.fr
-------------------------------------------------------------------------------------------------
FreeMail powered by mail.fr
6 years, 3 months
HowTo trace errors or issues ?
by Sven Achtelik
Hi All,
I would like to understand how to find the point that fails when starting from the Event showing in the GUI. With that event I get an correlation ID and how would I trace all the following tasks, actions or events that are connected to that correlation ID ?
Is it something like CorrelationID -> TaskID -> CommandID or how do things connect ? Is there some documentation I could look at ?
Thank you,
Sven
6 years, 3 months