[Users] Adding new disk to Ovort node

Fabian Deutsch fabiand at redhat.com
Fri Dec 20 14:50:35 UTC 2013


Am Donnerstag, den 19.12.2013, 12:25 +0200 schrieb Gabi C:
> Hello again!
> 
> 
> After persisting selinux config, at reboot I get  "Curerent mode:
> enforced"" although ""Mode from config file: permissive"" !
> 
> Due to this, i think I get an denied for glusterfsd:
> 
> type=AVC msg=audit(1387365750.532:5873): avc:  denied  { relabelfrom }
> for  pid=30249 comm="glusterfsd"
> name="23fe702e-be59-4f65-8c55-58b1b1e1b023" dev="dm-10" ino=1835015
> scontext=system_u:system_r:glusterd_t:s0
> tcontext=system_u:object_r:file_t:s0 tclass=file
> 

Hey Gabi,

just a small update here.
The problems are all related to some mislabeling during build - ths
builds problems. You'll need a new image to get rid of all selinux
relatde bugs.

- fabian


> 
> 
> 
> 
> On Wed, Dec 18, 2013 at 3:30 PM, Fabian Deutsch <fabiand at redhat.com>
> wrote:
>         Am Mittwoch, den 18.12.2013, 14:14 +0200 schrieb Gabi C:
>         > Still, now I cannot start none of the 2 machines! I get
>         >
>         > ID 119 VM proxy2 is down. Exit message: Child quit during
>         startup
>         > handshake: Input/output error.""
>         
>         
>         Could you try ot find out in what context this IO error
>         appears?
>         
>         - fabian
>         
>         >
>         > Something similar to bug
>         > https://bugzilla.redhat.com/show_bug.cgi?id=1033064, except
>         that in my
>         > case selinux is permissive!
>         >
>         >
>         >
>         > On Wed, Dec 18, 2013 at 2:10 PM, Gabi C <gabicr at gmail.com>
>         wrote:
>         >         in my case $brick_path =/data
>         >
>         >
>         >         getfattr -d /data return NOTHING on both nodes!!!
>         >
>         >
>         >
>         >
>         >         On Wed, Dec 18, 2013 at 1:46 PM, Fabian Deutsch
>         >         <fabiand at redhat.com> wrote:
>         >                 Am Mittwoch, den 18.12.2013, 13:26 +0200
>         schrieb Gabi
>         >                 C:
>         >                 > Update on Glusterfs issue
>         >                 >
>         >                 >
>         >                 > I manage to recover lost volume after
>         recretaing the
>         >                 same volume name
>         >                 > with same bricks, whisch raised an error
>         message,
>         >                 resolved by, on both
>         >                 > nodes:
>         >                 >
>         >                 > setfattr -x trusted.glusterfs.volume-id
>         $brick_path
>         >                 > setfattr -x trusted.gfid $brick_path
>         >
>         >
>         >                 Hey,
>         >
>         >                 good that you could recover them.
>         >
>         >                 Could you please provide $brick_path and
>         getfattr -d
>         >                 $brick_path
>         >
>         >                 The question is if and/or why the fattrs are
>         not
>         >                 stored.
>         >
>         >                 - fabian
>         >
>         >                 >
>         >                 >
>         >                 >
>         >                 > On Wed, Dec 18, 2013 at 12:12 PM, Gabi C
>         >                 <gabicr at gmail.com> wrote:
>         >                 >         node 1:
>         >                 >
>         >                 >         [root at virtual5 admin]#
>         cat /config/files
>         >                 >         /etc/fstab
>         >                 >         /etc/shadow
>         >                 >         /etc/default/ovirt
>         >                 >         /etc/ssh/ssh_host_key
>         >                 >         /etc/ssh/ssh_host_key.pub
>         >                 >         /etc/ssh/ssh_host_dsa_key
>         >                 >         /etc/ssh/ssh_host_dsa_key.pub
>         >                 >         /etc/ssh/ssh_host_rsa_key
>         >                 >         /etc/ssh/ssh_host_rsa_key.pub
>         >                 >         /etc/rsyslog.conf
>         >                 >         /etc/libvirt/libvirtd.conf
>         >                 >         /etc/libvirt/passwd.db
>         >                 >         /etc/passwd
>         >                 >         /etc/sysconfig/network
>         >                 >         /etc/collectd.conf
>         >                 >         /etc/libvirt/qemu/networks
>         >                 >         /etc/ssh/sshd_config
>         >                 >         /etc/pki
>         >                 >         /etc/logrotate.d/ovirt-node
>         >                 >         /var/lib/random-seed
>         >                 >         /etc/iscsi/initiatorname.iscsi
>         >                 >         /etc/libvirt/qemu.conf
>         >                 >         /etc/sysconfig/libvirtd
>         >                 >         /etc/logrotate.d/libvirtd
>         >                 >         /etc/multipath.conf
>         >                 >         /etc/hosts
>         >                 >
>           /etc/sysconfig/network-scripts/ifcfg-enp3s0
>         >                 >
>           /etc/sysconfig/network-scripts/ifcfg-lo
>         >                 >         /etc/ntp.conf
>         >                 >         /etc/shadow
>         >                 >         /etc/vdsm-reg/vdsm-reg.conf
>         >                 >         /etc/shadow
>         >                 >         /etc/shadow
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/route-ovirtmgmt
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/rule-ovirtmgmt
>         >                 >         /root/.ssh/authorized_keys
>         >                 >         /etc/vdsm/vdsm.id
>         >                 >
>           /etc/udev/rules.d/12-ovirt-iosched.rules
>         >                 >         /etc/vdsm/vdsm.conf
>         >                 >         /etc/sysconfig/iptables
>         >                 >         /etc/resolv.conf
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY
>         >                 >
>           /etc/sysconfig/network-scripts/ifcfg-enp6s0
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/ifcfg-enp6s0.50
>         >                 >         /etc/glusterfs/glusterd.vol
>         >                 >         /etc/selinux/config
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >         node 2:
>         >                 >
>         >                 >
>         >                 >         [root at virtual4 ~]#
>         cat /config/files
>         >                 >         /etc/fstab
>         >                 >         /etc/shadow
>         >                 >         /etc/default/ovirt
>         >                 >         /etc/ssh/ssh_host_key
>         >                 >         /etc/ssh/ssh_host_key.pub
>         >                 >         /etc/ssh/ssh_host_dsa_key
>         >                 >         /etc/ssh/ssh_host_dsa_key.pub
>         >                 >         /etc/ssh/ssh_host_rsa_key
>         >                 >         /etc/ssh/ssh_host_rsa_key.pub
>         >                 >         /etc/rsyslog.conf
>         >                 >         /etc/libvirt/libvirtd.conf
>         >                 >         /etc/libvirt/passwd.db
>         >                 >         /etc/passwd
>         >                 >         /etc/sysconfig/network
>         >                 >         /etc/collectd.conf
>         >                 >         /etc/libvirt/qemu/networks
>         >                 >         /etc/ssh/sshd_config
>         >                 >         /etc/pki
>         >                 >         /etc/logrotate.d/ovirt-node
>         >                 >         /var/lib/random-seed
>         >                 >         /etc/iscsi/initiatorname.iscsi
>         >                 >         /etc/libvirt/qemu.conf
>         >                 >         /etc/sysconfig/libvirtd
>         >                 >         /etc/logrotate.d/libvirtd
>         >                 >         /etc/multipath.conf
>         >                 >         /etc/hosts
>         >                 >
>           /etc/sysconfig/network-scripts/ifcfg-enp3s0
>         >                 >
>           /etc/sysconfig/network-scripts/ifcfg-lo
>         >                 >         /etc/shadow
>         >                 >         /etc/shadow
>         >                 >         /etc/vdsm-reg/vdsm-reg.conf
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/route-ovirtmgmt
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/rule-ovirtmgmt
>         >                 >         /root/.ssh/authorized_keys
>         >                 >         /etc/shadow
>         >                 >         /etc/shadow
>         >                 >         /etc/vdsm/vdsm.id
>         >                 >
>           /etc/udev/rules.d/12-ovirt-iosched.rules
>         >                 >         /etc/sysconfig/iptables
>         >                 >         /etc/vdsm/vdsm.conf
>         >                 >         /etc/shadow
>         >                 >         /etc/resolv.conf
>         >                 >         /etc/ntp.conf
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY
>         >                 >
>           /etc/sysconfig/network-scripts/ifcfg-enp6s0
>         >                 >
>         >
>           /etc/sysconfig/network-scripts/ifcfg-enp6s0.50
>         >                 >         /etc/glusterfs/glusterd.vol
>         >                 >         /etc/selinux/config
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >         On Wed, Dec 18, 2013 at 12:07 PM,
>         Fabian
>         >                 Deutsch
>         >                 >         <fabiand at redhat.com> wrote:
>         >                 >                 Am Mittwoch, den
>         18.12.2013, 12:03
>         >                 +0200 schrieb Gabi
>         >                 >                 C:
>         >                 >                 > So here it is:
>         >                 >                 >
>         >                 >                 >
>         >                 >                 > in tab volumes add new
>         volume -
>         >                 Replicated, then
>         >                 >                 added storage -
>         >                 >                 > data/glusterfs. Then I
>         impoerted
>         >                 Vm, ran them and at
>         >                 >                 some point,
>         >                 >                 > needing some space for a
>         Redhat
>         >                 Satellite  instance
>         >                 >                 I decided to put
>         >                 >                 > both node in maintenace
>         stop them
>         >                 add new disk
>         >                 >                 devices and restart,
>         >                 >                 > but after restart the
>         gluster
>         >                 volume defined under
>         >                 >                 Volumes Tab
>         >                 >                 > vanished!
>         >                 >
>         >                 >
>         >                 >                 Antoni,
>         >                 >
>         >                 >                 can you tell what log
>         files to look
>         >                 at to find out why
>         >                 >                 that storage
>         >                 >                 domain vanished - from a
>         Engine
>         >                 side?
>         >                 >
>         >                 >                 And do you know what files
>         related
>         >                 to gluster are
>         >                 >                 changed on the Node
>         >                 >                 side?
>         >                 >
>         >                 >                 Gabi,
>         >                 >
>         >                 >                 could you please provide
>         the
>         >                 contents of /config/files
>         >                 >                 on the Node.
>         >                 >
>         >                 >                 > Glusterfs data goes
>         under /data
>         >                 directory which was
>         >                 >                 automatically
>         >                 >                 > configured when I
>         installed the
>         >                 node.
>         >                 >
>         >                 >
>         >                 >                 Yep, /data is on the Data
>         LV - that
>         >                 should be good.
>         >                 >
>         >                 >                 - fabian
>         >                 >
>         >                 >                 >
>         >                 >                 >
>         >                 >                 > On Wed, Dec 18, 2013 at
>         11:45 AM,
>         >                 Fabian Deutsch
>         >                 >                 <fabiand at redhat.com>
>         >                 >                 > wrote:
>         >                 >                 >         Am Mittwoch, den
>         >                 18.12.2013, 11:42 +0200
>         >                 >                 schrieb Gabi C:
>         >                 >                 >         > Yes, it is the
>         VM
>         >                 part..I just run into an
>         >                 >                 issue.  My setup
>         >                 >                 >         consist in
>         >                 >                 >         > 2 nodes with
>         glusterfs
>         >                 and after adding
>         >                 >                 supplemental hard
>         >                 >                 >         disk, after
>         >                 >                 >         > reboot I've
>         lost
>         >                 glusterfs volumes!
>         >                 >                 >
>         >                 >                 >
>         >                 >                 >         Could you
>         exactly explain
>         >                 what you
>         >                 >                 configured?
>         >                 >                 >
>         >                 >                 >         >
>         >                 >                 >         > How can I
>         persist any
>         >                 configuration on
>         >                 >                 node and I refer here
>         >                 >                 >         to
>         >                 >                 >         > ''setenforce
>         0'' - for
>         >                 ssh login to work-
>         >                 >                 and further
>         >                 >                 >
>         >                 >                 >
>         >                 >                 >         How changes can
>         be
>         >                 persisted on Node can be
>         >                 >                 found here:
>         >                 >                 >
>         >                 >
>         >
>         http://www.ovirt.org/Node_Troubleshooting#Making_changes_on_the_host
>         >                 >                 >
>         >                 >                 >         Do you know into
>         what path
>         >                 the glusterfs
>         >                 >                 data goes? Or is it
>         >                 >                 >         written
>         >                 >                 >         directly onto a
>         disk/LV?
>         >                 >                 >
>         >                 >                 >         - fabian
>         >                 >                 >
>         >                 >                 >         > ""
>         >                 >
>         >
>         http://www.ovirt.org/Features/GlusterFS_Storage_Domain
>         >                 >                 >         >       * option
>         >                 rpc-auth-allow-insecure on
>         >                 >                 ==> in
>         >                 >                 >         glusterd.vol
>         (ensure
>         >                 >                 >         >         u
>         restart
>         >                 glusterd service... for
>         >                 >                 this to take
>         >                 >                 >         effect)
>         >                 >                 >
>         >                 >                 >         >       * volume
>         set
>         >                 <volname>
>         >                 >                 server.allow-insecure on
>         ==>
>         >                 >                 >         (ensure u
>         >                 >                 >         >         stop
>         and start
>         >                 the volume.. for
>         >                 >                 this to take
>         >                 >                 >         effect)''
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >         > Thanks!
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >         > On Wed, Dec
>         18, 2013 at
>         >                 11:35 AM, Fabian
>         >                 >                 Deutsch
>         >                 >                 >
>         <fabiand at redhat.com>
>         >                 >                 >         > wrote:
>         >                 >                 >         >         Am
>         Mittwoch, den
>         >                 18.12.2013, 08:34
>         >                 >                 +0200 schrieb
>         >                 >                 >         Gabi C:
>         >                 >                 >         >         >
>         Hello!
>         >                 >                 >         >         >
>         >                 >                 >         >         >
>         >                 >                 >         >         > In
>         order to
>         >                 increase disk space
>         >                 >                 I want to add a
>         >                 >                 >         new disk
>         >                 >                 >         >         drive
>         to
>         >                 >                 >         >         >
>         ovirt node.
>         >                 After adding this
>         >                 >                 should I proceed as
>         >                 >                 >         "normal" -
>         >                 >                 >         >
>         pvcreate,
>         >                 >                 >         >         >
>         vgcreate,
>         >                 lvcreate and so on -
>         >                 >                 or these
>         >                 >                 >         configuration
>         will
>         >                 >                 >         >         not
>         >                 >                 >         >         >
>         persist?
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >         >         Hey
>         Gabi,
>         >                 >                 >         >
>         >                 >                 >         >
>         basically plain
>         >                 LVM is used in
>         >                 >                 Node - so yes
>         >                 >                 >         pvcreate and
>         >                 >                 >         >
>         lvextend can
>         >                 >                 >         >         be
>         used.
>         >                 >                 >         >         What
>         storage
>         >                 part do you want to
>         >                 >                 extend? The part
>         >                 >                 >         where the
>         >                 >                 >         >         VMs
>         reside?
>         >                 >                 >         >         You
>         will also
>         >                 need to take care to
>         >                 >                 extend the
>         >                 >                 >         filesystem.
>         >                 >                 >         >
>         >                 >                 >         >         -
>         fabian
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >         >
>         >                 >                 >
>         >                 >                 >
>         >                 >                 >
>         >                 >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >                 >
>         >
>         >
>         >
>         >
>         >
>         >
>         
>         
> 
> 

-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 836 bytes
Desc: This is a digitally signed message part
URL: <http://lists.ovirt.org/pipermail/users/attachments/20131220/880b14a4/attachment-0001.sig>


More information about the Users mailing list