Am Mittwoch, den 18.12.2013, 14:14 +0200 schrieb Gabi C:
> Still, now I cannot start none of the 2 machines! I getCould you try ot find out in what context this IO error appears?
>
> ID 119 VM proxy2 is down. Exit message: Child quit during startup
> handshake: Input/output error.""
- fabian
>
> Something similar to bug
> https://bugzilla.redhat.com/show_bug.cgi?id=1033064, except that in my
> case selinux is permissive!
>
>
>
> On Wed, Dec 18, 2013 at 2:10 PM, Gabi C <gabicr@gmail.com> wrote:
> in my case $brick_path =/data
>
>
> getfattr -d /data return NOTHING on both nodes!!!
>
>
>
>
> On Wed, Dec 18, 2013 at 1:46 PM, Fabian Deutsch
> <fabiand@redhat.com> wrote:
> Am Mittwoch, den 18.12.2013, 13:26 +0200 schrieb Gabi
> C:
> > Update on Glusterfs issue
> >
> >
> > I manage to recover lost volume after recretaing the
> same volume name
> > with same bricks, whisch raised an error message,
> resolved by, on both
> > nodes:
> >
> > setfattr -x trusted.glusterfs.volume-id $brick_path
> > setfattr -x trusted.gfid $brick_path
>
>
> Hey,
>
> good that you could recover them.
>
> Could you please provide $brick_path and getfattr -d
> $brick_path
>
> The question is if and/or why the fattrs are not
> stored.
>
> - fabian
>
> >
> >
> >
> > On Wed, Dec 18, 2013 at 12:12 PM, Gabi C
> <gabicr@gmail.com> wrote:
> > node 1:
> >
> > [root@virtual5 admin]# cat /config/files
> > /etc/fstab
> > /etc/shadow
> > /etc/default/ovirt
> > /etc/ssh/ssh_host_key
> > /etc/ssh/ssh_host_key.pub
> > /etc/ssh/ssh_host_dsa_key
> > /etc/ssh/ssh_host_dsa_key.pub
> > /etc/ssh/ssh_host_rsa_key
> > /etc/ssh/ssh_host_rsa_key.pub
> > /etc/rsyslog.conf
> > /etc/libvirt/libvirtd.conf
> > /etc/libvirt/passwd.db
> > /etc/passwd
> > /etc/sysconfig/network
> > /etc/collectd.conf
> > /etc/libvirt/qemu/networks
> > /etc/ssh/sshd_config
> > /etc/pki
> > /etc/logrotate.d/ovirt-node
> > /var/lib/random-seed
> > /etc/iscsi/initiatorname.iscsi
> > /etc/libvirt/qemu.conf
> > /etc/sysconfig/libvirtd
> > /etc/logrotate.d/libvirtd
> > /etc/multipath.conf
> > /etc/hosts
> > /etc/sysconfig/network-scripts/ifcfg-enp3s0
> > /etc/sysconfig/network-scripts/ifcfg-lo
> > /etc/ntp.conf
> > /etc/shadow
> > /etc/vdsm-reg/vdsm-reg.conf
> > /etc/shadow
> > /etc/shadow
> >
> /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt
> >
> /etc/sysconfig/network-scripts/route-ovirtmgmt
> >
> /etc/sysconfig/network-scripts/rule-ovirtmgmt
> > /root/.ssh/authorized_keys
> > /etc/vdsm/vdsm.id
> > /etc/udev/rules.d/12-ovirt-iosched.rules
> > /etc/vdsm/vdsm.conf
> > /etc/sysconfig/iptables
> > /etc/resolv.conf
> >
> /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY
> > /etc/sysconfig/network-scripts/ifcfg-enp6s0
> >
> /etc/sysconfig/network-scripts/ifcfg-enp6s0.50
> > /etc/glusterfs/glusterd.vol
> > /etc/selinux/config
> >
> >
> >
> >
> >
> >
> >
> >
> > node 2:
> >
> >
> > [root@virtual4 ~]# cat /config/files
> > /etc/fstab
> > /etc/shadow
> > /etc/default/ovirt
> > /etc/ssh/ssh_host_key
> > /etc/ssh/ssh_host_key.pub
> > /etc/ssh/ssh_host_dsa_key
> > /etc/ssh/ssh_host_dsa_key.pub
> > /etc/ssh/ssh_host_rsa_key
> > /etc/ssh/ssh_host_rsa_key.pub
> > /etc/rsyslog.conf
> > /etc/libvirt/libvirtd.conf
> > /etc/libvirt/passwd.db
> > /etc/passwd
> > /etc/sysconfig/network
> > /etc/collectd.conf
> > /etc/libvirt/qemu/networks
> > /etc/ssh/sshd_config
> > /etc/pki
> > /etc/logrotate.d/ovirt-node
> > /var/lib/random-seed
> > /etc/iscsi/initiatorname.iscsi
> > /etc/libvirt/qemu.conf
> > /etc/sysconfig/libvirtd
> > /etc/logrotate.d/libvirtd
> > /etc/multipath.conf
> > /etc/hosts
> > /etc/sysconfig/network-scripts/ifcfg-enp3s0
> > /etc/sysconfig/network-scripts/ifcfg-lo
> > /etc/shadow
> > /etc/shadow
> > /etc/vdsm-reg/vdsm-reg.conf
> >
> /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt
> >
> /etc/sysconfig/network-scripts/route-ovirtmgmt
> >
> /etc/sysconfig/network-scripts/rule-ovirtmgmt
> > /root/.ssh/authorized_keys
> > /etc/shadow
> > /etc/shadow
> > /etc/vdsm/vdsm.id
> > /etc/udev/rules.d/12-ovirt-iosched.rules
> > /etc/sysconfig/iptables
> > /etc/vdsm/vdsm.conf
> > /etc/shadow
> > /etc/resolv.conf
> > /etc/ntp.conf
> >
> /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY
> > /etc/sysconfig/network-scripts/ifcfg-enp6s0
> >
> /etc/sysconfig/network-scripts/ifcfg-enp6s0.50
> > /etc/glusterfs/glusterd.vol
> > /etc/selinux/config
> >
> >
> >
> >
> > On Wed, Dec 18, 2013 at 12:07 PM, Fabian
> Deutsch
> > <fabiand@redhat.com> wrote:
> > Am Mittwoch, den 18.12.2013, 12:03
> +0200 schrieb Gabi
> > C:
> > > So here it is:
> > >
> > >
> > > in tab volumes add new volume -
> Replicated, then
> > added storage -
> > > data/glusterfs. Then I impoerted
> Vm, ran them and at
> > some point,
> > > needing some space for a Redhat
> Satellite instance
> > I decided to put
> > > both node in maintenace stop them
> add new disk
> > devices and restart,
> > > but after restart the gluster
> volume defined under
> > Volumes Tab
> > > vanished!
> >
> >
> > Antoni,
> >
> > can you tell what log files to look
> at to find out why
> > that storage
> > domain vanished - from a Engine
> side?
> >
> > And do you know what files related
> to gluster are
> > changed on the Node
> > side?
> >
> > Gabi,
> >
> > could you please provide the
> contents of /config/files
> > on the Node.
> >
> > > Glusterfs data goes under /data
> directory which was
> > automatically
> > > configured when I installed the
> node.
> >
> >
> > Yep, /data is on the Data LV - that
> should be good.
> >
> > - fabian
> >
> > >
> > >
> > > On Wed, Dec 18, 2013 at 11:45 AM,
> Fabian Deutsch
> > <fabiand@redhat.com>
> > > wrote:
> > > Am Mittwoch, den
> 18.12.2013, 11:42 +0200
> > schrieb Gabi C:
> > > > Yes, it is the VM
> part..I just run into an
> > issue. My setup
> > > consist in
> > > > 2 nodes with glusterfs
> and after adding
> > supplemental hard
> > > disk, after
> > > > reboot I've lost
> glusterfs volumes!
> > >
> > >
> > > Could you exactly explain
> what you
> > configured?
> > >
> > > >
> > > > How can I persist any
> configuration on
> > node and I refer here
> > > to
> > > > ''setenforce 0'' - for
> ssh login to work-
> > and further
> > >
> > >
> > > How changes can be
> persisted on Node can be
> > found here:
> > >
> >
> http://www.ovirt.org/Node_Troubleshooting#Making_changes_on_the_host
> > >
> > > Do you know into what path
> the glusterfs
> > data goes? Or is it
> > > written
> > > directly onto a disk/LV?
> > >
> > > - fabian
> > >
> > > > ""
> >
> http://www.ovirt.org/Features/GlusterFS_Storage_Domain
> > > > * option
> rpc-auth-allow-insecure on
> > ==> in
> > > glusterd.vol (ensure
> > > > u restart
> glusterd service... for
> > this to take
> > > effect)
> > >
> > > > * volume set
> <volname>
> > server.allow-insecure on ==>
> > > (ensure u
> > > > stop and start
> the volume.. for
> > this to take
> > > effect)''
> > > >
> > > >
> > > > Thanks!
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > On Wed, Dec 18, 2013 at
> 11:35 AM, Fabian
> > Deutsch
> > > <fabiand@redhat.com>
> > > > wrote:
> > > > Am Mittwoch, den
> 18.12.2013, 08:34
> > +0200 schrieb
> > > Gabi C:
> > > > > Hello!
> > > > >
> > > > >
> > > > > In order to
> increase disk space
> > I want to add a
> > > new disk
> > > > drive to
> > > > > ovirt node.
> After adding this
> > should I proceed as
> > > "normal" -
> > > > pvcreate,
> > > > > vgcreate,
> lvcreate and so on -
> > or these
> > > configuration will
> > > > not
> > > > > persist?
> > > >
> > > >
> > > > Hey Gabi,
> > > >
> > > > basically plain
> LVM is used in
> > Node - so yes
> > > pvcreate and
> > > > lvextend can
> > > > be used.
> > > > What storage
> part do you want to
> > extend? The part
> > > where the
> > > > VMs reside?
> > > > You will also
> need to take care to
> > extend the
> > > filesystem.
> > > >
> > > > - fabian
> > > >
> > > >
> > > >
> > >
> > >
> > >
> > >
> >
> >
> >
> >
> >
> >
>
>
>
>
>
>