On Thu, 2018-07-26 at 14:01 +0200, Simone Tiraboschi wrote:
On Thu, Jul 26, 2018 at 12:44 PM Karli Sjöberg <karli(a)inparadise.se>
wrote:
> On Thu, 2018-07-26 at 12:38 +0200, Simone Tiraboschi wrote:
> >
> >
> > On Thu, Jul 26, 2018 at 9:30 AM Karli Sjöberg <karli(a)inparadise.s
> e>
> > wrote:
> > > On Thu, 2018-07-26 at 09:27 +0200, Simone Tiraboschi wrote:
> > > >
> > > >
> > > > On Wed, Jul 25, 2018 at 12:04 PM Karli Sjöberg <karli@inparad
> ise.
> > > se>
> > > > wrote:
> > > > > Hey all!
> > > > >
> > > > > I'm trying to deploy Hosted Engine through the Cockpit UI
> and
> > > it's
> > > > > going well until it's time to start the local VM and it
> kernel
> > > > > panics:
> > > > >
> > > > > [ 2.032053] Call Trace:
> > > > > [ 2.032053] [<ffffffffb687e78c>]
> > > load_elf_binary+0x33c/0xe50
> > > > > [ 2.032053] [<ffffffffb68f3919>] ?
> ima_bprm_check+0x49/0x50
> > > > > [ 2.032053] [<ffffffffb687e450>] ?
> > > load_elf_library+0x220/0x220
> > > > > [ 2.032053] [<ffffffffb682236f>]
> > > > > search_binary_handler+0xef/0x310
> > > > > [ 2.032053] [<ffffffffb6823b4b>]
> > > > > do_execve_common.isra.24+0x5db/0x6e0
> > > > > [ 2.032053] [<ffffffffb6823c68>] do_execve+0x18/0x20
> > > > > [ 2.032053] [<ffffffffb66afc1f>]
> > > > > ____call_usermodehelper+0xff/0x140
> > > > > [ 2.032053] [<ffffffffb66afb20>] ?
> > > > > call_usermodehelper+0x60/0x60
> > > > > [ 2.032053] [<ffffffffb6d20677>]
> > > > > ret_from_fork_nospec_begin+0x21/0x21
> > > > > [ 2.032053] [<ffffffffb66afb20>] ?
> > > > > call_usermodehelper+0x60/0x60
> > > > > [ 2.032053] Code: cf e9 ff 4c 89 f7 e8 7b 32 e7 ff e9 4d
> fa
> > > ff
> > > > > ff 65 8b 05 03 a0 7e 49 a8 01 0f 84 85 fc ff ff 31 d2 b8 01
> 00
> > > 00
> > > > > 00 b9 49 00 00 00 <0f> 30 0f 1f 44 00 00 48 c7 c0 10 00
00
> 00
> > > e8 07
> > > > > 00 00 00 f3 90
> > > > > [ 2.032053] RIP [<ffffffffb6823025>]
> > > flush_old_exec+0x725/0x980
> > > > > [ 2.032053] RSP <ffff8add75ecbd00>
> > > > > [ 2.298131] ---[ end trace 354b4039b6fb0889 ]---
> > > > > [ 2.303914] Kernel panic - not syncing: Fatal exception
> > > > > [ 2.304835] Kernel Offset: 0x35600000 from
> > > 0xffffffff81000000
> > > > > (relocation range: 0xffffffff80000000-0xffffffffbfffffff)
> > > > >
> > > > > I've never had this problem so I'd just want to know if
> it's a
> > > > > known
> > > > > issue right now or if I've done anything special to deserve
> > > this:)
> > > > >
> > > > > The "Hosts" I'm deploying this on are VMs with
nested virt
> > > > > activated,
> > > > > and I've done this before but this time around it's
> bombing, as
> > > > > earlier
> > > > > explained.
> > > > >
> > > >
> > > > Thanks for the report,
> > > > which hypervisor are you using on L0?
> > >
> > > In my case it's Xubuntu 18.04 LTS. Is there anything I can do
> to
> > > help
> > > out with this?
> >
> > And you are running your host VM on KVM, right?
>
> Correct, but wait, woa woa, what am I saying, it's not 18.04, it's
> 16.04! I was looking on the wrong computer:)
>
> Just to be as clear as I possibly can, the issue I am facing is
> with
> Xubuntu 16.04.4 LTS as L0 hypervisor.
We are successfully running our CI tests on a nested env on Centos
7.5.
Maybe the issue is just due to an older KVM version on Xubuntu
16.04.4 LTS.
Honestly I never tried that combination but I think it could be worth
to try with a different L0.
OK, lets compare versions:
CentOS 7.5: 3.10.0
Xubuntu 16.04: 4.4.0
In what way could that be, as you say, older?
/K
>
> >
> > /K
> >
> > >
> > > > /K
> > > >
> > > > >
> > > > > > Thanks in advance!
> > > > > > /K
> > > > > > _______________________________________________
> > > > > > Users mailing list -- users(a)ovirt.org
> > > > > > To unsubscribe send an email to users-leave(a)ovirt.org
> > > > > > Privacy Statement:
https://www.ovirt.org/site/privacy-polic
> > y/
> > > > > > oVirt Code of Conduct:
https://www.ovirt.org/community/abou
> > t/co
> > > > mmun
> > > > > > ity-guidelines/
> > > > > > List Archives:
https://lists.ovirt.org/archives/list/users@
> > ovir
> > > > t.or
> > > > > > g/message/UQ6NH54FRQ4BIURTSE5I5OGQZ6HHDT2B/