
Hello again! After persisting selinux config, at reboot I get "Curerent mode: enforced"" although ""Mode from config file: permissive"" ! Due to this, i think I get an denied for glusterfsd: type=AVC msg=audit(1387365750.532:5873): avc: denied { relabelfrom } for pid=30249 comm="glusterfsd" name="23fe702e-be59-4f65-8c55-58b1b1e1b023" dev="dm-10" ino=1835015 scontext=system_u:system_r:glusterd_t:s0 tcontext=system_u:object_r:file_t:s0 tclass=file On Wed, Dec 18, 2013 at 3:30 PM, Fabian Deutsch <fabiand@redhat.com> wrote:
Am Mittwoch, den 18.12.2013, 14:14 +0200 schrieb Gabi C:
Still, now I cannot start none of the 2 machines! I get
ID 119 VM proxy2 is down. Exit message: Child quit during startup handshake: Input/output error.""
Could you try ot find out in what context this IO error appears?
- fabian
Something similar to bug https://bugzilla.redhat.com/show_bug.cgi?id=1033064, except that in my case selinux is permissive!
On Wed, Dec 18, 2013 at 2:10 PM, Gabi C <gabicr@gmail.com> wrote: in my case $brick_path =/data
getfattr -d /data return NOTHING on both nodes!!!
On Wed, Dec 18, 2013 at 1:46 PM, Fabian Deutsch <fabiand@redhat.com> wrote: Am Mittwoch, den 18.12.2013, 13:26 +0200 schrieb Gabi C: > Update on Glusterfs issue > > > I manage to recover lost volume after recretaing the same volume name > with same bricks, whisch raised an error message, resolved by, on both > nodes: > > setfattr -x trusted.glusterfs.volume-id $brick_path > setfattr -x trusted.gfid $brick_path
Hey,
good that you could recover them.
Could you please provide $brick_path and getfattr -d $brick_path
The question is if and/or why the fattrs are not stored.
- fabian
> > > > On Wed, Dec 18, 2013 at 12:12 PM, Gabi C <gabicr@gmail.com> wrote: > node 1: > > [root@virtual5 admin]# cat /config/files > /etc/fstab > /etc/shadow > /etc/default/ovirt > /etc/ssh/ssh_host_key > /etc/ssh/ssh_host_key.pub > /etc/ssh/ssh_host_dsa_key > /etc/ssh/ssh_host_dsa_key.pub > /etc/ssh/ssh_host_rsa_key > /etc/ssh/ssh_host_rsa_key.pub > /etc/rsyslog.conf > /etc/libvirt/libvirtd.conf > /etc/libvirt/passwd.db > /etc/passwd > /etc/sysconfig/network > /etc/collectd.conf > /etc/libvirt/qemu/networks > /etc/ssh/sshd_config > /etc/pki > /etc/logrotate.d/ovirt-node > /var/lib/random-seed > /etc/iscsi/initiatorname.iscsi > /etc/libvirt/qemu.conf > /etc/sysconfig/libvirtd > /etc/logrotate.d/libvirtd > /etc/multipath.conf > /etc/hosts > /etc/sysconfig/network-scripts/ifcfg-enp3s0 > /etc/sysconfig/network-scripts/ifcfg-lo > /etc/ntp.conf > /etc/shadow > /etc/vdsm-reg/vdsm-reg.conf > /etc/shadow > /etc/shadow > /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt > /etc/sysconfig/network-scripts/route-ovirtmgmt > /etc/sysconfig/network-scripts/rule-ovirtmgmt > /root/.ssh/authorized_keys > /etc/vdsm/vdsm.id > /etc/udev/rules.d/12-ovirt-iosched.rules > /etc/vdsm/vdsm.conf > /etc/sysconfig/iptables > /etc/resolv.conf > /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY > /etc/sysconfig/network-scripts/ifcfg-enp6s0 > /etc/sysconfig/network-scripts/ifcfg-enp6s0.50 > /etc/glusterfs/glusterd.vol > /etc/selinux/config > > > > > > > > > node 2: > > > [root@virtual4 ~]# cat /config/files > /etc/fstab > /etc/shadow > /etc/default/ovirt > /etc/ssh/ssh_host_key > /etc/ssh/ssh_host_key.pub > /etc/ssh/ssh_host_dsa_key > /etc/ssh/ssh_host_dsa_key.pub > /etc/ssh/ssh_host_rsa_key > /etc/ssh/ssh_host_rsa_key.pub > /etc/rsyslog.conf > /etc/libvirt/libvirtd.conf > /etc/libvirt/passwd.db > /etc/passwd > /etc/sysconfig/network > /etc/collectd.conf > /etc/libvirt/qemu/networks > /etc/ssh/sshd_config > /etc/pki > /etc/logrotate.d/ovirt-node > /var/lib/random-seed > /etc/iscsi/initiatorname.iscsi > /etc/libvirt/qemu.conf > /etc/sysconfig/libvirtd > /etc/logrotate.d/libvirtd > /etc/multipath.conf > /etc/hosts > /etc/sysconfig/network-scripts/ifcfg-enp3s0 > /etc/sysconfig/network-scripts/ifcfg-lo > /etc/shadow > /etc/shadow > /etc/vdsm-reg/vdsm-reg.conf > /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt > /etc/sysconfig/network-scripts/route-ovirtmgmt > /etc/sysconfig/network-scripts/rule-ovirtmgmt > /root/.ssh/authorized_keys > /etc/shadow > /etc/shadow > /etc/vdsm/vdsm.id > /etc/udev/rules.d/12-ovirt-iosched.rules > /etc/sysconfig/iptables > /etc/vdsm/vdsm.conf > /etc/shadow > /etc/resolv.conf > /etc/ntp.conf > /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY > /etc/sysconfig/network-scripts/ifcfg-enp6s0 > /etc/sysconfig/network-scripts/ifcfg-enp6s0.50 > /etc/glusterfs/glusterd.vol > /etc/selinux/config > > > > > On Wed, Dec 18, 2013 at 12:07 PM, Fabian Deutsch > <fabiand@redhat.com> wrote: > Am Mittwoch, den 18.12.2013, 12:03 +0200 schrieb Gabi > C: > > So here it is: > > > > > > in tab volumes add new volume - Replicated, then > added storage - > > data/glusterfs. Then I impoerted Vm, ran them and at > some point, > > needing some space for a Redhat Satellite instance > I decided to put > > both node in maintenace stop them add new disk > devices and restart, > > but after restart the gluster volume defined under > Volumes Tab > > vanished! > > > Antoni, > > can you tell what log files to look at to find out why > that storage > domain vanished - from a Engine side? > > And do you know what files related to gluster are > changed on the Node > side? > > Gabi, > > could you please provide the contents of /config/files > on the Node. > > > Glusterfs data goes under /data directory which was > automatically > > configured when I installed the node. > > > Yep, /data is on the Data LV - that should be good. > > - fabian > > > > > > > On Wed, Dec 18, 2013 at 11:45 AM, Fabian Deutsch > <fabiand@redhat.com> > > wrote: > > Am Mittwoch, den 18.12.2013, 11:42 +0200 > schrieb Gabi C: > > > Yes, it is the VM part..I just run into an > issue. My setup > > consist in > > > 2 nodes with glusterfs and after adding > supplemental hard > > disk, after > > > reboot I've lost glusterfs volumes! > > > > > > Could you exactly explain what you > configured? > > > > > > > > How can I persist any configuration on > node and I refer here > > to > > > ''setenforce 0'' - for ssh login to work- > and further > > > > > > How changes can be persisted on Node can be > found here: > > >
http://www.ovirt.org/Node_Troubleshooting#Making_changes_on_the_host
> > > > Do you know into what path the glusterfs > data goes? Or is it > > written > > directly onto a disk/LV? > > > > - fabian > > > > > "" > http://www.ovirt.org/Features/GlusterFS_Storage_Domain > > > * option rpc-auth-allow-insecure on > ==> in > > glusterd.vol (ensure > > > u restart glusterd service... for > this to take > > effect) > > > > > * volume set <volname> > server.allow-insecure on ==> > > (ensure u > > > stop and start the volume.. for > this to take > > effect)'' > > > > > > > > > Thanks! > > > > > > > > > > > > > > > > > > On Wed, Dec 18, 2013 at 11:35 AM, Fabian > Deutsch > > <fabiand@redhat.com> > > > wrote: > > > Am Mittwoch, den 18.12.2013, 08:34 > +0200 schrieb > > Gabi C: > > > > Hello! > > > > > > > > > > > > In order to increase disk space > I want to add a > > new disk > > > drive to > > > > ovirt node. After adding this > should I proceed as > > "normal" - > > > pvcreate, > > > > vgcreate, lvcreate and so on - > or these > > configuration will > > > not > > > > persist? > > > > > > > > > Hey Gabi, > > > > > > basically plain LVM is used in > Node - so yes > > pvcreate and > > > lvextend can > > > be used. > > > What storage part do you want to > extend? The part > > where the > > > VMs reside? > > > You will also need to take care to > extend the > > filesystem. > > > > > > - fabian > > > > > > > > > > > > > > > > > > > > > > >