On Thu, 21 Nov 2019 at 16:42, Nathanaël Blanchet <blanchet(a)abes.fr> wrote:
Le 21/11/2019 à 13:57, Roy Golan a écrit :
On Thu, 21 Nov 2019 at 08:48, Roy Golan <rgolan(a)redhat.com> wrote:
>
>
> On Wed, 20 Nov 2019 at 09:49, Nathanaël Blanchet <blanchet(a)abes.fr>
> wrote:
>
>>
>> Le 19/11/2019 à 19:23, Nathanaël Blanchet a écrit :
>>
>>
>> Le 19/11/2019 à 13:43, Roy Golan a écrit :
>>
>>
>>
>> On Tue, 19 Nov 2019 at 14:34, Nathanaël Blanchet <blanchet(a)abes.fr>
>> wrote:
>>
>>> Le 19/11/2019 à 08:55, Roy Golan a écrit :
>>>
>>> oc get -o json clusterversion
>>>
>>> This is the output of the previous failed deployment, I'll give a try
>>> to a newer one when I'll have a minute to test
>>>
>> Without changing nothing with template, I gave a new try and... nothing
>> works anymore now, none of provided IPs can be pingued : dial tcp
>> 10.34.212.51:6443: connect: no route to host", so none of masters can
>> be provisonned by bootstrap.
>>
>> I tried with the latest rhcos and latest ovirt 4.3.7, it is the same.
>> Obviously something changed since my first attempt 12 days ago... is your
>> docker image for openshift-installer up to date?
>>
>> Are you still able to your side to deploy a valid cluster ?
>>
>> I investigated looking at bootstrap logs (attached) and it seems that
>> every containers die immediately after been started.
>>
>> Nov 20 07:02:33 localhost podman[2024]: 2019-11-20 07:02:33.60107571
>> +0000 UTC m=+0.794838407 container init
>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>
registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>> Nov 20 07:02:33 localhost podman[2024]: 2019-11-20 07:02:33.623197173
>> +0000 UTC m=+0.816959853 container start
>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>
registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>> Nov 20 07:02:33 localhost podman[2024]: 2019-11-20 07:02:33.623814258
>> +0000 UTC m=+0.817576965 container attach
>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>
registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>> Nov 20 07:02:34 localhost systemd[1]:
>> libpod-446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603.scope:
>> Consumed 814ms CPU time
>> Nov 20 07:02:34 localhost podman[2024]: 2019-11-20 07:02:34.100569998
>> +0000 UTC m=+1.294332779 container died
>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>
registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>> Nov 20 07:02:35 localhost podman[2024]: 2019-11-20 07:02:35.138523102
>> +0000 UTC m=+2.332285844 container remove
>> 446dc9b7a04ff3ff4bbcfa6750e3946c084741b39707eb088c9d7ae648e35603 (image=
>>
registry.svc.ci.openshift.org/origin/release:4.3, name=eager_cannon)
>>
>> and this:
>>
>> Nov 20 07:04:16 localhost hyperkube[1909]: E1120 07:04:16.489527 1909
>> remote_runtime.go:200] CreateContainer in sandbox
>> "58f2062aa7b6a5b2bdd6b9cf7b41a9f94ca2b30ad5a20e4fa4dec8a9b82f05e5"
from
>> runtime service failed: rpc error: code = Unknown desc = container create
>> failed: container_linux.go:345: starting container process caused "exec:
>> \"runtimecfg\": executable file not found in $PATH"
>> Nov 20 07:04:16 localhost hyperkube[1909]: E1120 07:04:16.489714 1909
>> kuberuntime_manager.go:783] init container start failed:
>> CreateContainerError: container create failed: container_linux.go:345:
>> starting container process caused "exec: \"runtimecfg\":
executable file
>> not found in $PATH"
>>
>> What do you think about this?
>>
>
> I'm seeing the same now, checking...
>
>
Because of the move upstream to release OKD the release-image that comes
with the installer I gave you are no longer valid.
I need to prepare an installer version with the preview of OKD, you can
find the details here
https://mobile.twitter.com/smarterclayton/status/1196477646885965824
Seems that okd4
https://github.com/openshift/okd/releases/download/4.3.0-0.okd-2019-11-15...
is a compiled binary without the ovirt installer. Can you tell where are
sources so as to include ovirt extensions into the installer build?
Soon the CI will help us in building okd images from pull requests. I'll
update you when this will become available.
(do I need to use the terraform-workers tag instead of latest?)
>>>
>>> docker pull quay.io/rgolangh/openshift-installer:terraform-workers
>>>
>>>
>>> [root@openshift-installer
>>> openshift-origin-client-tools-v3.11.0-0cbc58b-linux-64bit]# ./oc get -o
>>> json clusterversion
>>> {
>>> "apiVersion": "v1",
>>> "items": [
>>> {
>>> "apiVersion": "config.openshift.io/v1",
>>> "kind": "ClusterVersion",
>>> "metadata": {
>>> "creationTimestamp":
"2019-11-07T12:23:06Z",
>>> "generation": 1,
>>> "name": "version",
>>> "namespace": "",
>>> "resourceVersion": "3770202",
>>> "selfLink": "/apis/
>>> config.openshift.io/v1/clusterversions/version",
>>> "uid":
"77600bba-6e71-4b35-a60b-d8ee6e0f545c"
>>> },
>>> "spec": {
>>> "channel": "stable-4.3",
>>> "clusterID":
"6f87b719-e563-4c0b-ab5a-1144172bc983",
>>> "upstream":
>>> "https://api.openshift.com/api/upgrades_info/v1/graph"
>>> <
https://api.openshift.com/api/upgrades_info/v1/graph>
>>> },
>>> "status": {
>>> "availableUpdates": null,
>>> "conditions": [
>>> {
>>> "lastTransitionTime":
"2019-11-07T12:23:12Z",
>>> "status": "False",
>>> "type": "Available"
>>> },
>>> {script
>>> "lastTransitionTime":
"2019-11-07T12:56:15Z",
>>> "message": "Cluster operator
image-registry is
>>> still updating",
>>> "reason":
"ClusterOperatorNotAvailable",
>>> "status": "True",
>>> "type": "Failing"
>>> },
>>> {
>>> "lastTransitionTime":
"2019-11-07T12:23:12Z",
>>> "message": "Unable to apply
>>> 4.3.0-0.okd-2019-10-29-180250: the cluster operator image-registry has not
>>> yet successfully rolled out",
>>> "reason":
"ClusterOperatorNotAvailable",
>>> "status": "True",
>>> "type": "Progressing"
>>> },
>>> {
>>> "lastTransitionTime":
"2019-11-07T12:23:12Z",
>>> "message": "Unable to retrieve
available
>>> updates: currently installed version 4.3.0-0.okd-2019-10-29-180250 not
>>> found in the \"stable-4.3\" channel",
>>> "reason": "RemoteFailed",
>>> "status": "False",
>>> "type": "RetrievedUpdates"
>>> }
>>> ],
>>> "desired": {
>>> "force": false,
>>> "image": "
>>>
registry.svc.ci.openshift.org/origin/release@sha256:68286e07f7d68ebc8a067...
>>> ",
>>> "version":
"4.3.0-0.okd-2019-10-29-180250"
>>>
>>
>> Indeed this version is not the latest and is missing the aforementioned
>> fix for the registry.
>>
>> },
>>> "history": [
>>> {
>>> "completionTime": null,
>>> "image": "
>>>
registry.svc.ci.openshift.org/origin/release@sha256:68286e07f7d68ebc8a067...
>>> ",
>>> "startedTime":
"2019-11-07T12:23:12Z",
>>> "state": "Partial",
>>> "verified": false,
>>> "version":
"4.3.0-0.okd-2019-10-29-180250"
>>> }
>>> ],
>>> "observedGeneration": 1,
>>> "versionHash": "-3onP9QpPTg="
>>> }
>>> }
>>> ],
>>> "kind": "List",
>>> "metadata": {
>>> "resourceVersion": "",
>>> "selfLink": ""
>>> }
>>>
>>> }
>>>
>>>
>>> Can you answer to these few questions please?
>>>
>>> - The latest stable OKD version is 4.2.4. Is it possible to chose
>>> the version of okd when deploying (seems to use 4.3) or does the
installer
>>> always download the latest OKD?
>>>
>>>
>>
>>>
>>> - Can we use FCOS instead of RHCOS?
>>>
>>>
>>
>>> - About the pull secret, do we absolutely need a redhat login to
>>> get this file to deploy an upstream OKD cluster and not downstream
>>> openshift?
>>>
>>>
>>> To answer the 3 of those, this specific build is not really OKD, and
>> will use 4.3 and Red Hat artifact and must use RHCOs, hence the pull secret
>> thing.
>> I frankly don't know when OKD 4.3 is going to be released, I guess it
>> will be on top FCOS.
>> I'll update the list once we have the oVirt installer for OKD ready for
>> testing (on FCOS)
>>
>>
>>
>> --
>>> Nathanaël Blanchet
>>>
>>> Supervision réseau
>>> Pôle Infrastrutures Informatiques
>>> 227 avenue Professeur-Jean-Louis-Viala
>>> 34193 MONTPELLIER CEDEX 5
>>> Tél. 33 (0)4 67 54 84 55
>>> Fax 33 (0)4 67 54 84 14blanchet(a)abes.fr
>>>
>>> --
>> Nathanaël Blanchet
>>
>> Supervision réseau
>> Pôle Infrastrutures Informatiques
>> 227 avenue Professeur-Jean-Louis-Viala
>> 34193 MONTPELLIER CEDEX 5
>> Tél. 33 (0)4 67 54 84 55
>> Fax 33 (0)4 67 54 84 14blanchet(a)abes.fr
>>
>>
>> _______________________________________________
>> Users mailing list -- users(a)ovirt.org
>> To unsubscribe send an email to users-leave(a)ovirt.org
>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/MLO4NW7NWR2...
>>
>> --
>> Nathanaël Blanchet
>>
>> Supervision réseau
>> Pôle Infrastrutures Informatiques
>> 227 avenue Professeur-Jean-Louis-Viala
>> 34193 MONTPELLIER CEDEX 5
>> Tél. 33 (0)4 67 54 84 55
>> Fax 33 (0)4 67 54 84 14blanchet(a)abes.fr
>>
>> --
Nathanaël Blanchet
Supervision réseau
Pôle Infrastrutures Informatiques
227 avenue Professeur-Jean-Louis-Viala
34193 MONTPELLIER CEDEX 5
Tél. 33 (0)4 67 54 84 55
Fax 33 (0)4 67 54 84 14blanchet(a)abes.fr