Am Donnerstag, den 19.12.2013, 12:25 +0200 schrieb Gabi C: > Hello again! > > > After persisting selinux config, at reboot I get "Curerent mode: > enforced"" although ""Mode from config file: permissive"" ! > > Due to this, i think I get an denied for glusterfsd: > > type=AVC msg=audit(1387365750.532:5873): avc: denied { relabelfrom } > for pid=30249 comm="glusterfsd" > name="23fe702e-be59-4f65-8c55-58b1b1e1b023" dev="dm-10" ino=1835015 > scontext=system_u:system_r:glusterd_t:s0 > tcontext=system_u:object_r:file_t:s0 tclass=file >
Hey Gabi, just a small update here. The problems are all related to some mislabeling during build - ths builds problems. You'll need a new image to get rid of all selinux relatde bugs. - fabian > > > > > On Wed, Dec 18, 2013 at 3:30 PM, Fabian Deutsch <fabi...@redhat.com> > wrote: > Am Mittwoch, den 18.12.2013, 14:14 +0200 schrieb Gabi C: > > Still, now I cannot start none of the 2 machines! I get > > > > ID 119 VM proxy2 is down. Exit message: Child quit during > startup > > handshake: Input/output error."" > > > Could you try ot find out in what context this IO error > appears? > > - fabian > > > > > Something similar to bug > > https://bugzilla.redhat.com/show_bug.cgi?id=1033064, except > that in my > > case selinux is permissive! > > > > > > > > On Wed, Dec 18, 2013 at 2:10 PM, Gabi C <gab...@gmail.com> > wrote: > > in my case $brick_path =/data > > > > > > getfattr -d /data return NOTHING on both nodes!!! > > > > > > > > > > On Wed, Dec 18, 2013 at 1:46 PM, Fabian Deutsch > > <fabi...@redhat.com> wrote: > > Am Mittwoch, den 18.12.2013, 13:26 +0200 > schrieb Gabi > > C: > > > Update on Glusterfs issue > > > > > > > > > I manage to recover lost volume after > recretaing the > > same volume name > > > with same bricks, whisch raised an error > message, > > resolved by, on both > > > nodes: > > > > > > setfattr -x trusted.glusterfs.volume-id > $brick_path > > > setfattr -x trusted.gfid $brick_path > > > > > > Hey, > > > > good that you could recover them. > > > > Could you please provide $brick_path and > getfattr -d > > $brick_path > > > > The question is if and/or why the fattrs are > not > > stored. > > > > - fabian > > > > > > > > > > > > > > On Wed, Dec 18, 2013 at 12:12 PM, Gabi C > > <gab...@gmail.com> wrote: > > > node 1: > > > > > > [root@virtual5 admin]# > cat /config/files > > > /etc/fstab > > > /etc/shadow > > > /etc/default/ovirt > > > /etc/ssh/ssh_host_key > > > /etc/ssh/ssh_host_key.pub > > > /etc/ssh/ssh_host_dsa_key > > > /etc/ssh/ssh_host_dsa_key.pub > > > /etc/ssh/ssh_host_rsa_key > > > /etc/ssh/ssh_host_rsa_key.pub > > > /etc/rsyslog.conf > > > /etc/libvirt/libvirtd.conf > > > /etc/libvirt/passwd.db > > > /etc/passwd > > > /etc/sysconfig/network > > > /etc/collectd.conf > > > /etc/libvirt/qemu/networks > > > /etc/ssh/sshd_config > > > /etc/pki > > > /etc/logrotate.d/ovirt-node > > > /var/lib/random-seed > > > /etc/iscsi/initiatorname.iscsi > > > /etc/libvirt/qemu.conf > > > /etc/sysconfig/libvirtd > > > /etc/logrotate.d/libvirtd > > > /etc/multipath.conf > > > /etc/hosts > > > > /etc/sysconfig/network-scripts/ifcfg-enp3s0 > > > > /etc/sysconfig/network-scripts/ifcfg-lo > > > /etc/ntp.conf > > > /etc/shadow > > > /etc/vdsm-reg/vdsm-reg.conf > > > /etc/shadow > > > /etc/shadow > > > > > > /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt > > > > > > /etc/sysconfig/network-scripts/route-ovirtmgmt > > > > > > /etc/sysconfig/network-scripts/rule-ovirtmgmt > > > /root/.ssh/authorized_keys > > > /etc/vdsm/vdsm.id > > > > /etc/udev/rules.d/12-ovirt-iosched.rules > > > /etc/vdsm/vdsm.conf > > > /etc/sysconfig/iptables > > > /etc/resolv.conf > > > > > > /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY > > > > /etc/sysconfig/network-scripts/ifcfg-enp6s0 > > > > > > /etc/sysconfig/network-scripts/ifcfg-enp6s0.50 > > > /etc/glusterfs/glusterd.vol > > > /etc/selinux/config > > > > > > > > > > > > > > > > > > > > > > > > > > > node 2: > > > > > > > > > [root@virtual4 ~]# > cat /config/files > > > /etc/fstab > > > /etc/shadow > > > /etc/default/ovirt > > > /etc/ssh/ssh_host_key > > > /etc/ssh/ssh_host_key.pub > > > /etc/ssh/ssh_host_dsa_key > > > /etc/ssh/ssh_host_dsa_key.pub > > > /etc/ssh/ssh_host_rsa_key > > > /etc/ssh/ssh_host_rsa_key.pub > > > /etc/rsyslog.conf > > > /etc/libvirt/libvirtd.conf > > > /etc/libvirt/passwd.db > > > /etc/passwd > > > /etc/sysconfig/network > > > /etc/collectd.conf > > > /etc/libvirt/qemu/networks > > > /etc/ssh/sshd_config > > > /etc/pki > > > /etc/logrotate.d/ovirt-node > > > /var/lib/random-seed > > > /etc/iscsi/initiatorname.iscsi > > > /etc/libvirt/qemu.conf > > > /etc/sysconfig/libvirtd > > > /etc/logrotate.d/libvirtd > > > /etc/multipath.conf > > > /etc/hosts > > > > /etc/sysconfig/network-scripts/ifcfg-enp3s0 > > > > /etc/sysconfig/network-scripts/ifcfg-lo > > > /etc/shadow > > > /etc/shadow > > > /etc/vdsm-reg/vdsm-reg.conf > > > > > > /etc/sysconfig/network-scripts/ifcfg-ovirtmgmt > > > > > > /etc/sysconfig/network-scripts/route-ovirtmgmt > > > > > > /etc/sysconfig/network-scripts/rule-ovirtmgmt > > > /root/.ssh/authorized_keys > > > /etc/shadow > > > /etc/shadow > > > /etc/vdsm/vdsm.id > > > > /etc/udev/rules.d/12-ovirt-iosched.rules > > > /etc/sysconfig/iptables > > > /etc/vdsm/vdsm.conf > > > /etc/shadow > > > /etc/resolv.conf > > > /etc/ntp.conf > > > > > > /etc/sysconfig/network-scripts/ifcfg-VPO_IPPROXY > > > > /etc/sysconfig/network-scripts/ifcfg-enp6s0 > > > > > > /etc/sysconfig/network-scripts/ifcfg-enp6s0.50 > > > /etc/glusterfs/glusterd.vol > > > /etc/selinux/config > > > > > > > > > > > > > > > On Wed, Dec 18, 2013 at 12:07 PM, > Fabian > > Deutsch > > > <fabi...@redhat.com> wrote: > > > Am Mittwoch, den > 18.12.2013, 12:03 > > +0200 schrieb Gabi > > > C: > > > > So here it is: > > > > > > > > > > > > in tab volumes add new > volume - > > Replicated, then > > > added storage - > > > > data/glusterfs. Then I > impoerted > > Vm, ran them and at > > > some point, > > > > needing some space for a > Redhat > > Satellite instance > > > I decided to put > > > > both node in maintenace > stop them > > add new disk > > > devices and restart, > > > > but after restart the > gluster > > volume defined under > > > Volumes Tab > > > > vanished! > > > > > > > > > Antoni, > > > > > > can you tell what log > files to look > > at to find out why > > > that storage > > > domain vanished - from a > Engine > > side? > > > > > > And do you know what files > related > > to gluster are > > > changed on the Node > > > side? > > > > > > Gabi, > > > > > > could you please provide > the > > contents of /config/files > > > on the Node. > > > > > > > Glusterfs data goes > under /data > > directory which was > > > automatically > > > > configured when I > installed the > > node. > > > > > > > > > Yep, /data is on the Data > LV - that > > should be good. > > > > > > - fabian > > > > > > > > > > > > > > > On Wed, Dec 18, 2013 at > 11:45 AM, > > Fabian Deutsch > > > <fabi...@redhat.com> > > > > wrote: > > > > Am Mittwoch, den > > 18.12.2013, 11:42 +0200 > > > schrieb Gabi C: > > > > > Yes, it is the > VM > > part..I just run into an > > > issue. My setup > > > > consist in > > > > > 2 nodes with > glusterfs > > and after adding > > > supplemental hard > > > > disk, after > > > > > reboot I've > lost > > glusterfs volumes! > > > > > > > > > > > > Could you > exactly explain > > what you > > > configured? > > > > > > > > > > > > > > How can I > persist any > > configuration on > > > node and I refer here > > > > to > > > > > ''setenforce > 0'' - for > > ssh login to work- > > > and further > > > > > > > > > > > > How changes can > be > > persisted on Node can be > > > found here: > > > > > > > > > > http://www.ovirt.org/Node_Troubleshooting#Making_changes_on_the_host > > > > > > > > Do you know into > what path > > the glusterfs > > > data goes? Or is it > > > > written > > > > directly onto a > disk/LV? > > > > > > > > - fabian > > > > > > > > > "" > > > > > > http://www.ovirt.org/Features/GlusterFS_Storage_Domain > > > > > * option > > rpc-auth-allow-insecure on > > > ==> in > > > > glusterd.vol > (ensure > > > > > u > restart > > glusterd service... for > > > this to take > > > > effect) > > > > > > > > > * volume > set > > <volname> > > > server.allow-insecure on > ==> > > > > (ensure u > > > > > stop > and start > > the volume.. for > > > this to take > > > > effect)'' > > > > > > > > > > > > > > > Thanks! > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Wed, Dec > 18, 2013 at > > 11:35 AM, Fabian > > > Deutsch > > > > > <fabi...@redhat.com> > > > > > wrote: > > > > > Am > Mittwoch, den > > 18.12.2013, 08:34 > > > +0200 schrieb > > > > Gabi C: > > > > > > > Hello! > > > > > > > > > > > > > > > > > > In > order to > > increase disk space > > > I want to add a > > > > new disk > > > > > drive > to > > > > > > > ovirt node. > > After adding this > > > should I proceed as > > > > "normal" - > > > > > > pvcreate, > > > > > > > vgcreate, > > lvcreate and so on - > > > or these > > > > configuration > will > > > > > not > > > > > > > persist? > > > > > > > > > > > > > > > Hey > Gabi, > > > > > > > > > > > basically plain > > LVM is used in > > > Node - so yes > > > > pvcreate and > > > > > > lvextend can > > > > > be > used. > > > > > What > storage > > part do you want to > > > extend? The part > > > > where the > > > > > VMs > reside? > > > > > You > will also > > need to take care to > > > extend the > > > > filesystem. > > > > > > > > > > - > fabian > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > >
signature.asc
Description: This is a digitally signed message part
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users