[Users] Adding new disk to Ovort node

Gabi C gabicr at gmail.com
Wed Dec 18 12:14:39 UTC 2013


Still, now I cannot start none of the 2 machines! I get
ID 119 VM proxy2 is down. Exit message: Child quit during startup
handshake: Input/output error.""


Something similar to bug https://bugzilla.redhat.com/show_bug.cgi?id=1033064,
except that in my case selinux is permissive!


On Wed, Dec 18, 2013 at 2:10 PM, Gabi C <gabicr at gmail.com> wrote:

> in my case $brick_path =/data
>
>
> getfattr -d /data return NOTHING on both nodes!!!
>
>
>
> On Wed, Dec 18, 2013 at 1:46 PM, Fabian Deutsch <fabiand at redhat.com>wrote:
>
>> Am Mittwoch, den 18.12.2013, 13:26 +0200 schrieb Gabi C:
>> > Update on Glusterfs issue
>> >
>> >
>> > I manage to recover lost volume after recretaing the same volume name
>> > with same bricks, whisch raised an error message, resolved by, on both
>> > nodes:
>> >
>> > setfattr -x trusted.glusterfs.volume-id $brick_path
>> > setfattr -x trusted.gfid $brick_path
>>
>> Hey,
>>
>> good that you could recover them.
>>
>> Could you please provide $brick_path and getfattr -d $brick_path
>>
>> The question is if and/or why the fattrs are not stored.
>>
>> - fabian
>>
>> >
>> >
>> >
>> > On Wed, Dec 18, 2013 at 12:12 PM, Gabi C <gabicr at gmail.com> wrote:
>> >         node 1:
>> >
>> >         [root at virtual5 admin]# cat /config/files
>> >         /etc/fstab
>> >         /etc/shadow
>> >         /etc/default/ovirt
>> >         /etc/ssh/ssh_host_key
>> >         /etc/ssh/ssh_host_key.pub
>> >         /etc/ssh/ssh_host_dsa_key
>> >         /etc/ssh/ssh_host_dsa_key.pub
>> >         /etc/ssh/ssh_host_rsa_key
>> >         /etc/ssh/ssh_host_rsa_key.pub
>> >         /etc/rsyslog.conf
>> >         /etc/libvirt/libvirtd.conf
>> >         /etc/libvirt/passwd.db
>> >         /etc/passwd
>> >         /etc/sysconfig/network
>> >         /etc/collectd.conf
>> >         /etc/libvirt/qemu/networks
>> >         /etc/ssh/sshd_config
>> >         /etc/pki
>> >         /etc/logrotate.d/ovirt-node
>> >         /var/lib/random-seed
>> >         /etc/iscsi/initiatorname.iscsi
>> >         /etc/libvirt/qemu.conf
>> >         /etc/sysconfig/libvirtd
>> >         /etc/logrotate.d/libvirtd
>> >         /etc/multipath.conf
>> >         /etc/hosts
>> >         /etc/sysconfig/network-scripts/ifcfg-enp3s0
>> >         /etc/sysconfig/network-scripts/ifcfg-lo
>> >         /etc/ntp.conf
>> >         /etc/shadow
>> >         /etc/vdsm-reg/vdsm-reg.conf
>> >         /etc/shadow
>> >         /etc/shadow
>> >         /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt
>> >         /etc/sysconfig/network-scripts/route-ovirtmgmt
>> >         /etc/sysconfig/network-scripts/rule-ovirtmgmt
>> >         /root/.ssh/authorized_keys
>> >         /etc/vdsm/vdsm.id
>> >         /etc/udev/rules.d/12-ovirt-iosched.rules
>> >         /etc/vdsm/vdsm.conf
>> >         /etc/sysconfig/iptables
>> >         /etc/resolv.conf
>> >         /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY
>> >         /etc/sysconfig/network-scripts/ifcfg-enp6s0
>> >         /etc/sysconfig/network-scripts/ifcfg-enp6s0.50
>> >         /etc/glusterfs/glusterd.vol
>> >         /etc/selinux/config
>> >
>> >
>> >
>> >
>> >
>> >
>> >
>> >
>> >         node 2:
>> >
>> >
>> >         [root at virtual4 ~]# cat /config/files
>> >         /etc/fstab
>> >         /etc/shadow
>> >         /etc/default/ovirt
>> >         /etc/ssh/ssh_host_key
>> >         /etc/ssh/ssh_host_key.pub
>> >         /etc/ssh/ssh_host_dsa_key
>> >         /etc/ssh/ssh_host_dsa_key.pub
>> >         /etc/ssh/ssh_host_rsa_key
>> >         /etc/ssh/ssh_host_rsa_key.pub
>> >         /etc/rsyslog.conf
>> >         /etc/libvirt/libvirtd.conf
>> >         /etc/libvirt/passwd.db
>> >         /etc/passwd
>> >         /etc/sysconfig/network
>> >         /etc/collectd.conf
>> >         /etc/libvirt/qemu/networks
>> >         /etc/ssh/sshd_config
>> >         /etc/pki
>> >         /etc/logrotate.d/ovirt-node
>> >         /var/lib/random-seed
>> >         /etc/iscsi/initiatorname.iscsi
>> >         /etc/libvirt/qemu.conf
>> >         /etc/sysconfig/libvirtd
>> >         /etc/logrotate.d/libvirtd
>> >         /etc/multipath.conf
>> >         /etc/hosts
>> >         /etc/sysconfig/network-scripts/ifcfg-enp3s0
>> >         /etc/sysconfig/network-scripts/ifcfg-lo
>> >         /etc/shadow
>> >         /etc/shadow
>> >         /etc/vdsm-reg/vdsm-reg.conf
>> >         /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt
>> >         /etc/sysconfig/network-scripts/route-ovirtmgmt
>> >         /etc/sysconfig/network-scripts/rule-ovirtmgmt
>> >         /root/.ssh/authorized_keys
>> >         /etc/shadow
>> >         /etc/shadow
>> >         /etc/vdsm/vdsm.id
>> >         /etc/udev/rules.d/12-ovirt-iosched.rules
>> >         /etc/sysconfig/iptables
>> >         /etc/vdsm/vdsm.conf
>> >         /etc/shadow
>> >         /etc/resolv.conf
>> >         /etc/ntp.conf
>> >         /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY
>> >         /etc/sysconfig/network-scripts/ifcfg-enp6s0
>> >         /etc/sysconfig/network-scripts/ifcfg-enp6s0.50
>> >         /etc/glusterfs/glusterd.vol
>> >         /etc/selinux/config
>> >
>> >
>> >
>> >
>> >         On Wed, Dec 18, 2013 at 12:07 PM, Fabian Deutsch
>> >         <fabiand at redhat.com> wrote:
>> >                 Am Mittwoch, den 18.12.2013, 12:03 +0200 schrieb Gabi
>> >                 C:
>> >                 > So here it is:
>> >                 >
>> >                 >
>> >                 > in tab volumes add new volume - Replicated, then
>> >                 added storage -
>> >                 > data/glusterfs. Then I impoerted Vm, ran them and at
>> >                 some point,
>> >                 > needing some space for a Redhat Satellite  instance
>> >                 I decided to put
>> >                 > both node in maintenace stop them add new disk
>> >                 devices and restart,
>> >                 > but after restart the gluster volume defined under
>> >                 Volumes Tab
>> >                 > vanished!
>> >
>> >
>> >                 Antoni,
>> >
>> >                 can you tell what log files to look at to find out why
>> >                 that storage
>> >                 domain vanished - from a Engine side?
>> >
>> >                 And do you know what files related to gluster are
>> >                 changed on the Node
>> >                 side?
>> >
>> >                 Gabi,
>> >
>> >                 could you please provide the contents of /config/files
>> >                 on the Node.
>> >
>> >                 > Glusterfs data goes under /data directory which was
>> >                 automatically
>> >                 > configured when I installed the node.
>> >
>> >
>> >                 Yep, /data is on the Data LV - that should be good.
>> >
>> >                 - fabian
>> >
>> >                 >
>> >                 >
>> >                 > On Wed, Dec 18, 2013 at 11:45 AM, Fabian Deutsch
>> >                 <fabiand at redhat.com>
>> >                 > wrote:
>> >                 >         Am Mittwoch, den 18.12.2013, 11:42 +0200
>> >                 schrieb Gabi C:
>> >                 >         > Yes, it is the VM part..I just run into an
>> >                 issue.  My setup
>> >                 >         consist in
>> >                 >         > 2 nodes with glusterfs and after adding
>> >                 supplemental hard
>> >                 >         disk, after
>> >                 >         > reboot I've lost glusterfs volumes!
>> >                 >
>> >                 >
>> >                 >         Could you exactly explain what you
>> >                 configured?
>> >                 >
>> >                 >         >
>> >                 >         > How can I persist any configuration on
>> >                 node and I refer here
>> >                 >         to
>> >                 >         > ''setenforce 0'' - for ssh login to work-
>> >                 and further
>> >                 >
>> >                 >
>> >                 >         How changes can be persisted on Node can be
>> >                 found here:
>> >                 >
>> >
>> http://www.ovirt.org/Node_Troubleshooting#Making_changes_on_the_host
>> >                 >
>> >                 >         Do you know into what path the glusterfs
>> >                 data goes? Or is it
>> >                 >         written
>> >                 >         directly onto a disk/LV?
>> >                 >
>> >                 >         - fabian
>> >                 >
>> >                 >         > ""
>> >                 http://www.ovirt.org/Features/GlusterFS_Storage_Domain
>> >                 >         >       * option rpc-auth-allow-insecure on
>> >                 ==> in
>> >                 >         glusterd.vol (ensure
>> >                 >         >         u restart glusterd service... for
>> >                 this to take
>> >                 >         effect)
>> >                 >
>> >                 >         >       * volume set <volname>
>> >                 server.allow-insecure on ==>
>> >                 >         (ensure u
>> >                 >         >         stop and start the volume.. for
>> >                 this to take
>> >                 >         effect)''
>> >                 >         >
>> >                 >         >
>> >                 >         > Thanks!
>> >                 >         >
>> >                 >         >
>> >                 >         >
>> >                 >         >
>> >                 >         >
>> >                 >         > On Wed, Dec 18, 2013 at 11:35 AM, Fabian
>> >                 Deutsch
>> >                 >         <fabiand at redhat.com>
>> >                 >         > wrote:
>> >                 >         >         Am Mittwoch, den 18.12.2013, 08:34
>> >                 +0200 schrieb
>> >                 >         Gabi C:
>> >                 >         >         > Hello!
>> >                 >         >         >
>> >                 >         >         >
>> >                 >         >         > In order to increase disk space
>> >                 I want to add a
>> >                 >         new disk
>> >                 >         >         drive to
>> >                 >         >         > ovirt node. After adding this
>> >                 should I proceed as
>> >                 >         "normal" -
>> >                 >         >         pvcreate,
>> >                 >         >         > vgcreate, lvcreate and so on -
>> >                 or these
>> >                 >         configuration will
>> >                 >         >         not
>> >                 >         >         > persist?
>> >                 >         >
>> >                 >         >
>> >                 >         >         Hey Gabi,
>> >                 >         >
>> >                 >         >         basically plain LVM is used in
>> >                 Node - so yes
>> >                 >         pvcreate and
>> >                 >         >         lvextend can
>> >                 >         >         be used.
>> >                 >         >         What storage part do you want to
>> >                 extend? The part
>> >                 >         where the
>> >                 >         >         VMs reside?
>> >                 >         >         You will also need to take care to
>> >                 extend the
>> >                 >         filesystem.
>> >                 >         >
>> >                 >         >         - fabian
>> >                 >         >
>> >                 >         >
>> >                 >         >
>> >                 >
>> >                 >
>> >                 >
>> >                 >
>> >
>> >
>> >
>> >
>> >
>> >
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20131218/6cbea6ba/attachment-0001.html>


More information about the Users mailing list