this is the result of execution of following commands:

[root@openshift-ied ~]# oc project kube-service-catalog
Now using project "kube-service-catalog" on server "https://openshift-ied.install.etux:8443".
[root@openshift-ied ~]# oc get pods
NAME                       READY     STATUS    RESTARTS   AGE
apiserver-h6rg4            1/1       Running   0          1h
controller-manager-d8vkq   1/1       Running   5          3h
[root@openshift-ied ~]# oc describe pod apiserver-h6rg4 
Name:           apiserver-h6rg4
Namespace:      kube-service-catalog
Node:           openshift-ied.install.etux/10.10.4.248
Start Time:     Fri, 15 Jun 2018 14:16:50 +0100
Labels:         app=apiserver
                controller-revision-hash=296530938
                pod-template-generation=2
Annotations:    ca_hash=63a4fb42b7ebe23182afbfd0146c520b4d0bb4cd
                openshift.io/scc=hostmount-anyuid
Status:         Running
IP:             10.128.0.20
Controlled By:  DaemonSet/apiserver
Containers:
  apiserver:
    Container ID:  docker://1eb20cb1b8235c72cb8016a3587a087ae2f954834ec4d259d71bf10e5b25d034
    Image:         docker.io/openshift/origin-service-catalog:v3.9.0
    Image ID:      docker-pullable://docker.io/openshift/origin-service-catalog@sha256:4c8fa186fce466c8b35afbbd715207d473333369cb92b6710faa4a707fb038a5
    Port:          6443/TCP
    Command:
      /usr/bin/service-catalog
    Args:
      apiserver
      --storage-type
      etcd
      --secure-port
      6443
      --etcd-servers
      https://openshift-ied.install.etux:2379
      --etcd-cafile
      /etc/origin/master/master.etcd-ca.crt
      --etcd-certfile
      /etc/origin/master/master.etcd-client.crt
      --etcd-keyfile
      /etc/origin/master/master.etcd-client.key
      -v
      3
      --cors-allowed-origins
      localhost
      --admission-control
      KubernetesNamespaceLifecycle,DefaultServicePlan,ServiceBindingsLifecycle,ServicePlanChangeValidator,BrokerAuthSarCheck
      --feature-gates
      OriginatingIdentity=true
    State:          Running
      Started:      Fri, 15 Jun 2018 14:16:58 +0100
    Ready:          True
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /etc/origin/master from etcd-host-cert (ro)
      /var/run/kubernetes-service-catalog from apiserver-ssl (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from service-catalog-apiserver-token-ptkqz (ro)
Conditions:
  Type           Status
  Initialized    True 
  Ready          True 
  PodScheduled   True 
Volumes:
  apiserver-ssl:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  apiserver-ssl
    Optional:    false
  etcd-host-cert:
    Type:          HostPath (bare host directory volume)
    Path:          /etc/origin/master
    HostPathType:  
  data-dir:
    Type:    EmptyDir (a temporary directory that shares a pod's lifetime)
    Medium:  
  service-catalog-apiserver-token-ptkqz:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  service-catalog-apiserver-token-ptkqz
    Optional:    false
QoS Class:       BestEffort
Node-Selectors:  node-role.kubernetes.io/master=true
Tolerations:     node.kubernetes.io/disk-pressure:NoSchedule
                 node.kubernetes.io/memory-pressure:NoSchedule
                 node.kubernetes.io/not-ready:NoExecute
                 node.kubernetes.io/unreachable:NoExecute
Events:
  Type     Reason                 Age               From                                 Message
  ----     ------                 ----              ----                                 -------
  Normal   SuccessfulMountVolume  1h                kubelet, openshift-ied.install.etux  MountVolume.SetUp succeeded for volume "etcd-host-cert"
  Normal   SuccessfulMountVolume  1h                kubelet, openshift-ied.install.etux  MountVolume.SetUp succeeded for volume "data-dir"
  Normal   SuccessfulMountVolume  1h                kubelet, openshift-ied.install.etux  MountVolume.SetUp succeeded for volume "service-catalog-apiserver-token-ptkqz"
  Normal   SuccessfulMountVolume  1h                kubelet, openshift-ied.install.etux  MountVolume.SetUp succeeded for volume "apiserver-ssl"
  Normal   Pulled                 1h                kubelet, openshift-ied.install.etux  Container image "docker.io/openshift/origin-service-catalog:v3.9.0" already present on machine
  Normal   Created                1h                kubelet, openshift-ied.install.etux  Created container
  Normal   Started                1h                kubelet, openshift-ied.install.etux  Started container
  Warning  DNSConfigForming       4m (x62 over 1h)  kubelet, openshift-ied.install.etux  Search Line limits were exceeded, some search paths have been omitted, the applied search line is: kube-service-catalog.svc.cluster.local svc.cluster.local cluster.local install.etux bofh.etux dmz.etux
[root@openshift-ied ~]# oc describe apiserver
the server doesn't have a resource type "apiserver"

Regards,
Carlos Rodrigues

On Fri, 2018-06-15 at 07:58 -0600, Rich Megginson wrote:
on the metrics machine, log in as root

oc project kube-service-catalog
oc get pods

you should have a pod named apiserver-xxxxx

oc logs apiserver-xxxxx

oc describe pod apiserver-xxxxx

oc describe ds apiserver

On 06/15/2018 07:02 AM, Carlos Rodrigues wrote:
I'm able to go forward in installation but i get another error as you
can see in attachment.

Could you help me?

Regards,
Carlos Rodrigues

On Fri, 2018-06-15 at 05:57 -0600, Rich Megginson wrote:
Are you able to install and run ovirt metrics successfully?

On 06/15/2018 02:27 AM, Carlos Rodrigues wrote:
Hi,

I'm behind a proxy. After configure docker to use proxy, the
following
command run well:

sudo docker pull docker.io/openshift/origin-pod:v3.9.0

Thank you.

Regards,
Carlos Rodrigues

On Thu, 2018-06-14 at 13:43 -0600, Rich Megginson wrote:
This is a different error than the one described in the links
below:

           "5m          1h           128
webconsole-84466b9d97-s4x28.153776d1bf88b3a4 Pod
Warning   FailedCreatePodSandBox   kubelet, openshift-
ied.install.etux
Failed create pod sandbox: rpc error: code = Unknown desc =
failed
pulling image \"docker.io/openshift/origin-pod:v3.9.0\": Get
https://registry-1.docker.io/v2/: net/http: request canceled
while
waiting for connection (Client.Timeout exceeded while awaiting
headers)"

Is it possible that there is some sort of networking issue that
you

from the machine - can you do


can you do

sudo docker pull docker.io/openshift/origin-pod:v3.9.0

?

On 06/14/2018 03:18 AM, Carlos Rodrigues wrote:
Hi, i still get the same error as you can see in attachment.

I also send in attachment the ansible log result from run the
following
command:


ANSIBLE_LOG_PATH=/tmp/ansible.log ansible-playbook -vvv -e
@/root/vars.yaml -i /root/ansible-inventory-origin-39-aio
playbooks/deploy_cluster.yml


Regards,
Carlos Rodrigues


On Tue, 2018-06-12 at 15:11 +0100, Carlos Rodrigues wrote:
Thank you Rich, w'll try this workaround and tell you later.

Regards,
Carlos Rodrigues

On Tue, 2018-06-12 at 07:54 -0600, Rich Megginson wrote:
Sorry, did not mean to send an internal link to an external
list/address.

On 06/12/2018 07:52 AM, Rich Megginson wrote:
http://post-office.corp.redhat.com/archives/aos-devel/201
8-Ju
ne/m
sg
00195.html


"
It smells like https://access.redhat.com/solutions/348092
1 /
me.
"

I think the workaround is to add

oreg_url=registry.access.redhat.com/openshift3/ose-
${component}:${version}


to your inventory for OCP external and

oreg_url=brew-pulp-
docker01.web.prod.ext.phx2.redhat.com:8888/openshift3/ose
-
${component}:${version}


for OCP internal and

oreg_url=docker.io/openshift/origin-
${component}:${version}

for Origin

On 06/12/2018 02:47 AM, Shirly Radco wrote:
-- 

SHIRLY RADCO

BI SeNIOR SOFTWARE ENGINEER

Red Hat Israel <https://www.redhat.com/>

TRIED. TESTED. TRUSTED. <https://redhat.com/trusted>


On Tue, Jun 12, 2018 at 11:43 AM, Shirly Radco <sradco@
redh
at.c
om
    
<mailto:sradco@redhat.com>> wrote:

       Hi Rich,

       Are you families with this OpenShift installation
issue?

familiar*


       Best,

       --

       SHIRLY RADCO

       BI SeNIOR SOFTWARE ENGINEER

       Red Hat Israel <https://www.redhat.com/>

       <https://red.ht/sig>
       TRIED. TESTED. TRUSTED. <https://redhat.com/trust
ed>


       On Mon, Jun 11, 2018 at 4:20 PM, Carlos Rodrigues
       <cmar@eurotux.com <mailto:cmar@eurotux.com>>
wrote:

           Hi,

           i'm trying to install oVirt Metrics following
the
installation
           guide (h
           ttps://www.ovirt.org/develop/release-
           management/features/metrics/metrics-store-
installation/)
but i
           have
           some issues running deploy cluster from here
           https://www.ovirt.org/deve
lop/release-management/features/metrics/setting-up-
viaq-
logging/
<https://www.ovirt.org/develop/release-management/featu
res/
metr
ic
s/setting-up-viaq-logging/>

           cd /usr/share/ansible/openshift-ansible
           # (or wherever you cloned the git repo if
using
git)
           ANSIBLE_LOG_PATH=/tmp/ansible.log ansible-
playbook
-vvv
-e
           @/root/vars.yaml -i /root/ansible-inventory-
origin-39-
aio
           playbooks/deploy_cluster.yml

           I fails on Web console installation:

           2018-06-06 19:48:24,020 p=17586 u=root |
[DEPRECATION
           WARNING]: Using
           tests as filters is deprecated. Instead of
using
           `result|version_compare` instead use `result
is
           version_compare`. This
           feature
            will be removed in version 2.9.
           Deprecation warnings can be disabled by
setting
           deprecation_warnings=False in ansible.cfg.
           2018-06-06 19:48:24,135 p=17586 u=root
 Using

module
file
/usr/lib/python2.7/site-
packages/ansible/modules/commands/command.py
           2018-06-06 19:48:27,093 p=17586 u=root
 fatal:

[localhost]:
           FAILED!
           => {
               "changed": true,
               "cmd": [
                   "oc",
                   "logs",
                   "deployment/webconsole",
                   "--tail=50",
"--config=/tmp/console-ansible-
C8XDuW/admin.kubeconfig",
                   "-n",
                   "openshift-web-console"
               ],
               "delta": "0:00:01.567694",
               "end": "2018-06-06 19:48:26.706407",
               "invocation": {
                   "module_args": {
                       "_raw_params": "oc logs
deployment/webconsole
           --tail=50 --
           config=/tmp/console-ansible-
C8XDuW/admin.kubeconfig -n
           openshift-web-
           console",
                       "_uses_shell": false,
                       "chdir": null,
                       "creates": null,
                       "executable": null,
                       "removes": null,
                       "stdin": null,
                       "warn": true
                   }

               },
               "msg": "non-zero return code",
               "rc": 1,
               "start": "2018-06-06 19:48:25.138713",
               "stderr": "Error from server
(BadRequest):
container
           \"webconsole\"
           in pod \"webconsole-84466b9d97-s4x28\" is
waiting
to
start:
           ContainerCreating",
               "stderr_lines": [
                   "Error from server (BadRequest):
container
           \"webconsole\" in
           pod \"webconsole-84466b9d97-s4x28\" is
waiting to
start:
           ContainerCreating"
               ],
               "stdout": "",
               "stdout_lines": []
           }
           2018-06-06 19:48:27,097 p=17586 u=root |
...ignoring

           Regards,
           Carlos Rodrigues

           On Mon, 2018-06-11 at 13:30 +0300, Shirly
Radco
wrote:
           > Dear users,
           >
           > I would love to get some feedback if
someone has
tried
to
           install and
           > use the oVirt metrics store, released in
4.2,
for
collecting
           metrics
           > and logs, based on Elasticsearch, Kibana,
Collectd
and
           Fluentd on top
           > of OpenShift.
           >
https://www.ovirt.org/develop/release-management/featur
es/m
etri
cs
/met
<https://www.ovirt.org/develop/release-management/featu
res/
metr
ic
s/met>
           > rics-store/
           >
           > How did the installation go? Are you
actively
using
it?
           > And any other feedback would be much
appreciated.
           >
           > Best regards,
           > --
           > SHIRLY RADCO
           > BI SENIOR SOFTWARE ENGINEER
           > Red Hat Israel
           >       TRIED. TESTED. TRUSTED.
           >
_______________________________________________
           > Users mailing list -- users@ovirt.org
<mailto:us
vi
rt.org>
           > To unsubscribe send an email to users-leave
@ovir
t.org
           <mailto:users-leave@ovirt.org>
           > Privacy Statement:
           https://www.ovirt.org/site/privacy-policy/
           <https://www.ovirt.org/site/privacy-policy/>
           > oVirt Code of Conduct:
           https://www.ovirt.org/community/about/communi
t
           <https://www.ovirt.org/community/about/commun
it>
           > y-guidelines/
           > List Archives:
           https://lists.ovirt.org/archives/list/users@o
virt.
org/
<https://lists.ovirt.org/archives/list/users@ovirt.org/
           > message/3G2M3Q35UQZLOHDRAEBMX2INPDAQCOHO/
           --         Carlos Rodrigues

           Engenheiro de Software Sénior

           Eurotux Informática, S.A. | www.eurotux.com
           <http://www.eurotux.com>
           (t) +351 253 680 300 (m) +351 911 926 110




-- 
Carlos Rodrigues

Engenheiro de Software Sénior

Eurotux Informática, S.A. | www.eurotux.com
(t) +351 253 680 300 (m) +351 911 926 110
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/
comm
unit
y-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ov
irt.
org/
message/CUAHIV5AVLIDPL2YQVWRHM7RDA56QJ5T/


-- 
Carlos Rodrigues

Engenheiro de Software Sénior

Eurotux Informática, S.A. | www.eurotux.com
(t) +351 253 680 300 (m) +351 911 926 110