<html>
  <head>
    <meta content="text/html; charset=utf-8" http-equiv="Content-Type">
  </head>
  <body text="#000000" bgcolor="#FFFFFF">
    <p>Ah, spoke to soon.  30 seconds later the network went down with
      IPv6 disabled.  So it does appear to be a host forwarding problem,
      not a VM problem.  I have an oVirt 4.0 cluster on the same network
      that doesn't have these issues, so it must be a configuration
      issue somewhere.  Here is a dump of my ip config on the host:</p>
    <p>[07:57:26]root@ovirt730-01 ~ # ip addr show<br>
      1: lo: &lt;LOOPBACK,UP,LOWER_UP&gt; mtu 65536 qdisc noqueue state
      UNKNOWN qlen 1<br>
          link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00<br>
          inet 127.0.0.1/8 scope host lo<br>
             valid_lft forever preferred_lft forever<br>
          inet6 ::1/128 scope host <br>
             valid_lft forever preferred_lft forever<br>
      2: em1: &lt;BROADCAST,MULTICAST,UP,LOWER_UP&gt; mtu 1500 qdisc mq
      master vmNet state UP qlen 1000<br>
          link/ether 18:66:da:eb:8f:c0 brd ff:ff:ff:ff:ff:ff<br>
      3: em2: &lt;BROADCAST,MULTICAST&gt; mtu 1500 qdisc noop state DOWN
      qlen 1000<br>
          link/ether 18:66:da:eb:8f:c1 brd ff:ff:ff:ff:ff:ff<br>
      4: em3: &lt;BROADCAST,MULTICAST&gt; mtu 1500 qdisc noop state DOWN
      qlen 1000<br>
          link/ether 18:66:da:eb:8f:c2 brd ff:ff:ff:ff:ff:ff<br>
      5: em4: &lt;BROADCAST,MULTICAST&gt; mtu 1500 qdisc noop state DOWN
      qlen 1000<br>
          link/ether 18:66:da:eb:8f:c3 brd ff:ff:ff:ff:ff:ff<br>
      6: p5p1: &lt;BROADCAST,MULTICAST,UP,LOWER_UP&gt; mtu 1500 qdisc mq
      master ovirtmgmt state UP qlen 1000<br>
          link/ether f4:e9:d4:a9:7a:f0 brd ff:ff:ff:ff:ff:ff<br>
      7: p5p2: &lt;BROADCAST,MULTICAST&gt; mtu 1500 qdisc noop state
      DOWN qlen 1000<br>
          link/ether f4:e9:d4:a9:7a:f2 brd ff:ff:ff:ff:ff:ff<br>
      8: vmNet: &lt;BROADCAST,MULTICAST,UP,LOWER_UP&gt; mtu 1500 qdisc
      noqueue state UP qlen 1000<br>
          link/ether 18:66:da:eb:8f:c0 brd ff:ff:ff:ff:ff:ff<br>
          inet 192.168.1.180/24 brd 192.168.1.255 scope global vmNet<br>
             valid_lft forever preferred_lft forever<br>
      10: ovirtmgmt: &lt;BROADCAST,MULTICAST,UP,LOWER_UP&gt; mtu 1500
      qdisc noqueue state UP qlen 1000<br>
          link/ether f4:e9:d4:a9:7a:f0 brd ff:ff:ff:ff:ff:ff<br>
          inet 192.168.130.180/24 brd 192.168.130.255 scope global
      ovirtmgmt<br>
             valid_lft forever preferred_lft forever<br>
      11: ;vdsmdummy;: &lt;BROADCAST,MULTICAST&gt; mtu 1500 qdisc noop
      state DOWN qlen 1000<br>
          link/ether fa:f3:48:35:76:8d brd ff:ff:ff:ff:ff:ff<br>
      14: vnet0: &lt;BROADCAST,MULTICAST,UP,LOWER_UP&gt; mtu 1500 qdisc
      pfifo_fast master ovirtmgmt state UNKNOWN qlen 1000<br>
          link/ether fe:16:3e:3f:fb:ec brd ff:ff:ff:ff:ff:ff<br>
          inet6 fe80::fc16:3eff:fe3f:fbec/64 scope link <br>
             valid_lft forever preferred_lft forever<br>
      15: vnet1: &lt;BROADCAST,MULTICAST,UP,LOWER_UP&gt; mtu 1500 qdisc
      pfifo_fast master vmNet state UNKNOWN qlen 1000<br>
          link/ether fe:1a:4a:16:01:51 brd ff:ff:ff:ff:ff:ff<br>
          inet6 fe80::fc1a:4aff:fe16:151/64 scope link <br>
             valid_lft forever preferred_lft forever<br>
      16: vnet2: &lt;BROADCAST,MULTICAST,UP,LOWER_UP&gt; mtu 1500 qdisc
      pfifo_fast master vmNet state UNKNOWN qlen 1000<br>
          link/ether fe:1a:4a:16:01:52 brd ff:ff:ff:ff:ff:ff<br>
          inet6 fe80::fc1a:4aff:fe16:152/64 scope link <br>
             valid_lft forever preferred_lft forever<br>
    </p>
    <p>[07:57:50]root@ovirt730-01 ~ # ip route show<br>
      default via 192.168.1.254 dev vmNet <br>
      192.168.1.0/24 dev vmNet  proto kernel  scope link  src
      192.168.1.180 <br>
      169.254.0.0/16 dev vmNet  scope link  metric 1008 <br>
      169.254.0.0/16 dev ovirtmgmt  scope link  metric 1010 <br>
      192.168.130.0/24 dev ovirtmgmt  proto kernel  scope link  src
      192.168.130.180 <br>
    </p>
    <p>[07:57:53]root@ovirt730-01 ~ # ip rule show<br>
      0:    from all lookup local <br>
      32760:    from all to 192.168.130.0/24 iif ovirtmgmt lookup
      3232268980 <br>
      32761:    from 192.168.130.0/24 lookup 3232268980 <br>
      32762:    from all to 192.168.1.0/24 iif vmNet lookup 2308294836 <br>
      32763:    from 192.168.1.0/24 lookup 2308294836 <br>
      32766:    from all lookup main <br>
      32767:    from all lookup default <br>
      [07:57:58]root@ovirt730-01 ~ # <br>
      <br>
    </p>
    <br>
    <div class="moz-cite-prefix">On 2017-04-10 07:54 AM, Charles Tassell
      wrote:<br>
    </div>
    <blockquote
      cite="mid:ba6a9823-8f79-11ce-3b5b-bda2bbba456d@islandadmin.ca"
      type="cite">
      <meta content="text/html; charset=utf-8" http-equiv="Content-Type">
      <p>Hi Everyone,</p>
      <p>  Just an update, I installed a new Ubuntu guest VM and it was
        doing the same thing regarding the network going down, then I
        disabled IPv6 and it's been fine for the past 10-15 minutes.  So
        the issue seems to be IPv6 related, and I don't need IPv6 so I
        can just turn it off.  The eth1 NIC disappearing is still
        worrisome though.<br>
      </p>
      <br>
      <div class="moz-cite-prefix">On 2017-04-10 07:13 AM, Charles
        Tassell wrote:<br>
      </div>
      <blockquote
        cite="mid:208a3a45-ad2a-4cea-82c1-474553cff236@gmail.com"
        type="cite">
        <meta content="text/html; charset=utf-8"
          http-equiv="Content-Type">
        <div class="moz-cite-prefix">Hi Everyone,<br>
          <br>
            Thanks for the help, answers below.<br>
          <br>
          On 2017-04-10 05:27 AM, Sandro Bonazzola wrote:<br>
        </div>
        <blockquote
cite="mid:CAPQRNT=w+P46u+Qter3-6yKPVqYogoErLWB66keLBQt7Up0KqA@mail.gmail.com"
          type="cite">
          <div dir="ltr">Adding Simone and Martin, replying inline.<br>
            <div class="gmail_extra"><br>
              <div class="gmail_quote">On Mon, Apr 10, 2017 at 10:16 AM,
                Ondrej Svoboda <span dir="ltr">&lt;<a
                    moz-do-not-send="true"
                    href="mailto:osvoboda@redhat.com" target="_blank">osvoboda@redhat.com</a>&gt;</span>
                wrote:<br>
                <blockquote class="gmail_quote" style="margin:0 0 0
                  .8ex;border-left:1px #ccc solid;padding-left:1ex">
                  <div dir="ltr">
                    <div>Hello Charles,<br>
                      <br>
                    </div>
                    <div>First, can you give us more information
                      regarding the duplicated IPv6 addresses? Since you
                      are going to reinstall the hosted engine, could
                      you make sure that NetworkManager is disabled
                      before adding the second vNIC (and perhaps even
                      disable IPv6 and reboot as well, so we have a
                      solid base and see what makes the difference)?<br>
                    </div>
                  </div>
                </blockquote>
              </div>
            </div>
          </div>
        </blockquote>
        I disabled NetworkManager on the hosts (systemctl disable
        NetworkManager ; service NetworkManager stop) before doing the
        oVirt setup and rebooted to make sure that it didn't come back
        up.  Or are you referring to on the hosted engine VM?  I just
        removed and re-added the eth1 NIC in the hosted engine, and this
        is what showed up in dmesg:<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: [1af4:1000] type 00
        class 0x020000<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: reg 0x10: [io 
        0x0000-0x001f]<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: reg 0x14: [mem
        0x00000000-0x00000fff]<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: reg 0x20: [mem
        0x00000000-0x00003fff 64bit pref]<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: reg 0x30: [mem
        0x00000000-0x0003ffff pref]<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: BAR 6: assigned
        [mem 0xc0000000-0xc003ffff pref]<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: BAR 4: assigned
        [mem 0xc0040000-0xc0043fff 64bit pref]<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: BAR 1: assigned
        [mem 0xc0044000-0xc0044fff]<br>
        [Mon Apr 10 06:46:43 2017] pci 0000:00:08.0: BAR 0: assigned
        [io  0x1000-0x101f]<br>
        [Mon Apr 10 06:46:43 2017] virtio-pci 0000:00:08.0: enabling
        device (0000 -&gt; 0003)<br>
        [Mon Apr 10 06:46:43 2017] virtio-pci 0000:00:08.0: irq 35 for
        MSI/MSI-X<br>
        [Mon Apr 10 06:46:43 2017] virtio-pci 0000:00:08.0: irq 36 for
        MSI/MSI-X<br>
        [Mon Apr 10 06:46:43 2017] virtio-pci 0000:00:08.0: irq 37 for
        MSI/MSI-X<br>
        [Mon Apr 10 06:46:43 2017] IPv6: ADDRCONF(NETDEV_UP): eth1: link
        is not ready<br>
        [Mon Apr 10 06:46:43 2017] IPv6: eth1: IPv6 duplicate address
        fe80::21a:4aff:fe16:151 detected!<br>
        <br>
        Then when the network dropped I started getting these:<br>
        <br>
        [Mon Apr 10 06:48:00 2017] IPv6: eth1: IPv6 duplicate address
        2001:410:e000:902:21a:4aff:fe16:151 detected!<br>
        [Mon Apr 10 06:48:00 2017] IPv6: eth1: IPv6 duplicate address
        2001:410:e000:902:21a:4aff:fe16:151 detected!<br>
        [Mon Apr 10 06:49:51 2017] IPv6: eth1: IPv6 duplicate address
        2001:410:e000:902:21a:4aff:fe16:151 detected!<br>
        [Mon Apr 10 06:51:40 2017] IPv6: eth1: IPv6 duplicate address
        2001:410:e000:902:21a:4aff:fe16:151 detected!<br>
        <br>
        The network on eth1 would go down for a few seconds then come
        back up, but networking stays solid on eth0.  I disabled
        NetworkManager on the HE VM as well to see if that makes a
        difference.  I also disabled IPv6 with sysctl to see if that
        helps.  I'll install a Ubuntu VM on the cluster later today and
        see if it has a similar issue.<br>
        <br>
        <br>
        <blockquote
cite="mid:CAPQRNT=w+P46u+Qter3-6yKPVqYogoErLWB66keLBQt7Up0KqA@mail.gmail.com"
          type="cite">
          <div dir="ltr">
            <div class="gmail_extra">
              <div class="gmail_quote">
                <blockquote class="gmail_quote" style="margin:0 0 0
                  .8ex;border-left:1px #ccc solid;padding-left:1ex">
                  <div dir="ltr">
                    <div><br>
                      What kind of documentation did you follow to
                      install the hosted engine? Was it this page? <a
                        moz-do-not-send="true"
                        href="https://www.ovirt.org/documentation/how-to/hosted-engine/"
                        target="_blank">https://www.ovirt.org/<wbr>documentation/how-to/hosted-<wbr>engine/</a>
                      If so, could you file a bug against VDSM
                      networking and attach /var/log/vdsm/vdsm.log and
                      supervdsm.log, and make sure they include the time
                      period from adding the second vNIC to rebooting?<br>
                      <br>
                    </div>
                    <div>Second, even the vNIC going missing after
                      reboot looks like a bug to me. Even though eth1
                      does not exist in the VM, can you see it defined
                      for the VM in the engine web GUI?<br>
                    </div>
                  </div>
                </blockquote>
                <div><br>
                </div>
                <div>If the HE vm configuration wasn't flushed to the
                  OVF_STORE yet, it make sense it disappeared on
                  restart.</div>
                <div><br>
                </div>
                <div> </div>
              </div>
            </div>
          </div>
        </blockquote>
        The docs I used were <a moz-do-not-send="true"
          class="moz-txt-link-freetext"
href="https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.1/html/self-hosted_engine_guide/chap-deploying_self-hosted_engine#Deploying_Self-Hosted_Engine_on_RHEL">https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.1/html/self-hosted_engine_guide/chap-deploying_self-hosted_engine#Deploying_Self-Hosted_Engine_on_RHEL</a>
        which someone on the list pointed me to last week as being more
        up-to-date than what was on the website (the docs on the website
        don't seem to mention that you need to put the HE on it's own
        datastore and look to be more geared towards bare-metal engine
        rather than the VM self hosted option.)<br>
        <br>
        When I went back into the GUI and looked at the hosted engine
        config the second NIC was listed there, but it wasn't showing up
        in lspci on the VM.  I removed the NIC in the GUI and re-added
        it, and the device appeared again on the VM.  What is the proper
        way to "save" the state of the VM so that the OVF_STORE gets
        updated?  When I do anything on the HE VM that I want to test I
        just type "reboot", but that powers down the VM.  I then login
        to my host and run "hosted-engine --vm-start" which restarts it,
        but of course the last time I did that it restarted without the
        second NIC.<br>
        <br>
        <blockquote
cite="mid:CAPQRNT=w+P46u+Qter3-6yKPVqYogoErLWB66keLBQt7Up0KqA@mail.gmail.com"
          type="cite">
          <div dir="ltr">
            <div class="gmail_extra">
              <div class="gmail_quote">
                <blockquote class="gmail_quote" style="margin:0 0 0
                  .8ex;border-left:1px #ccc solid;padding-left:1ex">
                  <div dir="ltr">
                    <div><br>
                    </div>
                    <div>The steps you took to install the hosted engine
                      with regards to networking look good to me, but I
                      believe Sandro (CC'ed) would be able to give more
                      advice.<br>
                      <br>
                      Sandro, since we want to configure bonding, would
                      you recommend to install the engine physically
                      first, move it to a VM, according to the following
                      method, and only then reconfigure networking? <a
                        moz-do-not-send="true"
href="https://www.ovirt.org/documentation/self-hosted/chap-Migrating_from_Bare_Metal_to_an_EL-Based_Self-Hosted_Environment/"
                        target="_blank">https://www.ovirt.org/<wbr>documentation/self-hosted/<wbr>chap-Migrating_from_Bare_<wbr>Metal_to_an_EL-Based_Self-<wbr>Hosted_Environment/</a></div>
                  </div>
                </blockquote>
                <div><br>
                </div>
                <div><br>
                </div>
                <div>I don't see why a diret HE deployment couldn't be
                  done. Simone, Martin can you help here?</div>
                <div> </div>
                <blockquote class="gmail_quote" style="margin:0 0 0
                  .8ex;border-left:1px #ccc solid;padding-left:1ex">
                  <div dir="ltr">
                    <div><br>
                      <br>
                    </div>
                    <div>Thank you,<br>
                    </div>
                    <div>Ondra<br>
                    </div>
                  </div>
                  <div class="HOEnZb">
                    <div class="h5">
                      <div class="gmail_extra"><br>
                        <div class="gmail_quote">On Mon, Apr 10, 2017 at
                          8:51 AM, Charles Tassell <span dir="ltr">&lt;<a
                              moz-do-not-send="true"
                              href="mailto:ctassell@gmail.com"
                              target="_blank">ctassell@gmail.com</a>&gt;</span>
                          wrote:<br>
                          <blockquote class="gmail_quote"
                            style="margin:0 0 0 .8ex;border-left:1px
                            #ccc solid;padding-left:1ex">Hi Everyone,<br>
                            <br>
                              Okay, I'm again having problems with
                            getting basic networking setup with oVirt
                            4.1  Here is my situation.  I have two
                            servers I want to use to create an oVirt
                            cluster, with two different networks.  My
                            "public" network is a 1G link on device em1
                            connected to my Internet feed, and my
                            "storage" network is a 10G link connected on
                            device p5p1 to my file server.  Since I need
                            to connect to my storage network in order to
                            do the install, I selected p5p1 has the
                            ovirtmgmt interface when installing the
                            hosted engine.  That worked fine, I got
                            everything installed, so I used some
                            ssh-proxy magic to connect to the web
                            console and completed the install (setup a
                            Storage domain and create a new network
                            vmNet for VM networking and added em1 to
                            it.)<br>
                            <br>
                              The problem was that when I added a second
                            network device to the HostedEngine VM (so
                            that I can connect to it from my public
                            network) it would intermittently go down.  I
                            did some digging and found some IPV6 errors
                            in the dmesg (IPv6: eth1: IPv6 duplicate
                            address 2001:410:e000:902:21a:4aff:fe1<wbr>6:151
                            detected!) so I disabled IPv6 on both eth0
                            and eth1 in the HostedEngine and rebooted
                            it.  The problem is that when I restarted
                            the VM, the eth1 device was missing.<br>
                            <br>
                              So, my question is: Can I add a second NIC
                            to the HostedEngine VM and make it stick, or
                            will it be deleted whenever the engine VM is
                            restarted?  </blockquote>
                        </div>
                      </div>
                    </div>
                  </div>
                </blockquote>
                <div><br>
                </div>
                <div>When you change something in the HE Vm using the
                  web ui, it has to be saved also on the OVF_STORE to
                  make it permanent for further reboot.</div>
                <div>Martin can you please elaborate here?</div>
                <div><br>
                </div>
                <div><br>
                </div>
                <div> </div>
                <blockquote class="gmail_quote" style="margin:0 0 0
                  .8ex;border-left:1px #ccc solid;padding-left:1ex">
                  <div class="HOEnZb">
                    <div class="h5">
                      <div class="gmail_extra">
                        <div class="gmail_quote">
                          <blockquote class="gmail_quote"
                            style="margin:0 0 0 .8ex;border-left:1px
                            #ccc solid;padding-left:1ex">Is there a
                            better way to do what I'm trying to do, ie,
                            should I setup ovirtmgmt on the public em1
                            interface, and then create the "storage"
                            network after the fact for connecting to the
                            datastores and such.  Is that even possible,
                            or required?  I was thinking that it would
                            be better for migrations and other
                            management functions to happen on the faster
                            10G network, but if the HostedEngine doesn't
                            need to be able to connect to the storage
                            network maybe it's not worth the effort?<br>
                            <br>
                              Eventually I want to setup LACP on the
                            storage network, but I had to wipe the
                            servers and reinstall from scratch the last
                            time I tried to set that up.  I was thinking
                            that it was because I setup the bonding
                            before installing oVirt, so I didn't do that
                            this time.<br>
                            <br>
                              Here are my /etc/sysconfig/network-scripts<wbr>/ifcfg-*
                            files in case I did something wrong there
                            (I'm more familiar with Debian/Ubuntu
                            network setup than CentOS)<br>
                            <br>
                            ifcfg-eth0: (ovirtmgmt aka storage)<br>
                            ----------------<br>
                            BROADCAST=192.168.130.255<br>
                            NETMASK=255.255.255.0<br>
                            BOOTPROTO=static<br>
                            DEVICE=eth0<br>
                            IPADDR=192.168.130.179<br>
                            ONBOOT=yes<br>
                            DOMAIN=<a moz-do-not-send="true"
                              href="http://public.net" rel="noreferrer"
                              target="_blank">public.net</a><br>
                            ZONE=public<br>
                            IPV6INIT=no<br>
                            <br>
                            <br>
                            ifcfg-eth1: (vmNet aka Internet)<br>
                            ----------------<br>
                            BROADCAST=192.168.1.255<br>
                            NETMASK=255.255.255.0<br>
                            BOOTPROTO=static<br>
                            DEVICE=eth1<br>
                            IPADDR=192.168.1.179<br>
                            GATEWAY=192.168.1.254<br>
                            ONBOOT=yes<br>
                            DNS1=192.168.1.1<br>
                            DNS2=192.168.1.2<br>
                            DOMAIN=<a moz-do-not-send="true"
                              href="http://public.net" rel="noreferrer"
                              target="_blank">public.net</a><br>
                            ZONE=public<br>
                            IPV6INIT=no<br>
                            <br>
                            ______________________________<wbr>_________________<br>
                            Users mailing list<br>
                            <a moz-do-not-send="true"
                              href="mailto:Users@ovirt.org"
                              target="_blank">Users@ovirt.org</a><br>
                            <a moz-do-not-send="true"
                              href="http://lists.ovirt.org/mailman/listinfo/users"
                              rel="noreferrer" target="_blank">http://lists.ovirt.org/mailman<wbr>/listinfo/users</a><br>
                          </blockquote>
                        </div>
                        <br>
                      </div>
                    </div>
                  </div>
                </blockquote>
              </div>
              <br>
              <br clear="all">
              <div><br>
              </div>
              -- <br>
              <div class="gmail_signature"
                data-smartmail="gmail_signature">
                <div dir="ltr">
                  <div>
                    <div dir="ltr">
                      <div dir="ltr">
                        <div dir="ltr">
                          <div dir="ltr">
                            <p
style="color:rgb(0,0,0);font-family:overpass,sans-serif;font-weight:bold;margin:0px;padding:0px;font-size:14px;text-transform:uppercase"><span>SANDRO</span> <span>BONAZZOLA</span></p>
                            <p
style="color:rgb(0,0,0);font-family:overpass,sans-serif;font-size:10px;margin:0px
                              0px 4px;text-transform:uppercase"><span>ASSOCIATE
                                MANAGER, SOFTWARE ENGINEERING, EMEA ENG
                                VIRTUALIZATION R&amp;D</span></p>
                            <p
style="font-family:overpass,sans-serif;margin:0px;font-size:10px;color:rgb(153,153,153)"><a
                                moz-do-not-send="true"
                                href="https://www.redhat.com/"
                                style="color:rgb(0,136,206);margin:0px"
                                target="_blank">Red Hat <span>EMEA</span></a></p>
                            <table
style="color:rgb(0,0,0);font-family:overpass,sans-serif;font-size:medium"
                              border="0">
                              <tbody>
                                <tr>
                                  <td width="100px"><a
                                      moz-do-not-send="true"
                                      href="https://red.ht/sig"
                                      target="_blank"><img
                                        moz-do-not-send="true"
src="https://www.redhat.com/profiles/rh/themes/redhatdotcom/img/logo-red-hat-black.png"
                                        height="auto" width="90"></a></td>
                                  <td style="font-size:10px">
                                    <div><a moz-do-not-send="true"
                                        href="https://redhat.com/trusted"
style="color:rgb(204,0,0);font-weight:bold" target="_blank">TRIED.
                                        TESTED. TRUSTED.</a></div>
                                  </td>
                                </tr>
                              </tbody>
                            </table>
                          </div>
                        </div>
                      </div>
                    </div>
                  </div>
                </div>
              </div>
            </div>
          </div>
        </blockquote>
        <p><br>
        </p>
      </blockquote>
      <br>
    </blockquote>
    <br>
  </body>
</html>