Hello Bob, and thanks for the reply. I am using RHEL5, 3 node cluster - node2, node3 and node4. Node2 is also a lucy box. Now that I look at the versions of the GFS, I do see some differences. We are not running Xen kernel.
The command gfs_grow was launched from the Cluster node4 and here is what I have found issuing rpm -qa | grep gfs: 2.6.18-92.1.10.el5 #1 SMP Wed Jul 23 03:55:54 EDT 2008 i686 i686 i386 GNU/Linux gfs2-utils-0.1.44-1.el5 gfs-utils-0.1.17-1.el5 kmod-gfs-0.1.23-5.el5 kmod-gfs2-1.92-1.1.el5 kmod-gfs2-xen-1.92-1.1.el5 kmod-gfs-xen-0.1.23-5.e Node 3: 2.6.18-92.1.10.el5 #1 SMP Wed Jul 23 03:55:54 EDT 2008 i686 i686 i386 GNU/Linux gfs2-utils-0.1.44-1.el5_2.1 gfs-utils-0.1.17-1.el5 kmod-gfs-0.1.23-5.el5 *kmod-gfs-0.1.23-5.el5_2.2* kmod-gfs2-1.92-1.1.el5 *kmod-gfs2-1.92-1.1.el5_2.2* kmod-gfs2-xen-1.92-1.1.el5 kmod-gfs2-xen-1.92-1.1.el5_2.2 kmod-gfs-xen-0.1.23-5.el5 kmod-gfs-xen-0.1.23-5.el5_2.2 Node2 - lucy node: 2.6.18-92.1.10.el5 #1 SMP Wed Jul 23 03:55:54 EDT 2008 i686 i686 i386 GNU/Linux gfs2-utils-0.1.44-1.el5_2.1 kmod-gfs-0.1.23-5.el5_2.2 gfs-utils-0.1.17-1.el5 kmod-gfs-0.1.23-5.el5 Let me know if you need any additional information. What would be suggested path to recovery. I tried gfs_fsck but I get: Initializing fsck Unable to open device: /lvm_test2 On Thu, Sep 25, 2008 at 2:46 PM, Bob Peterson <[EMAIL PROTECTED]> wrote: > ----- "Alan A" <[EMAIL PROTECTED]> wrote: > | Hi all! > | > | I have 3 node test cluster utilizing SCSI fencing and GFS. I have made > | 2 GFS > | Logical Volumes - lvm1 and lvm2, both utilizing 5GB on 10GB disks. > | Testing > | the command line tools I did lvextend -L +1G /devicename to bring lvm2 > | to > | 6GB. This went fine without any problems. Then I issued command > | gfs_grow > | /mountpoint and the volume became inaccessible. Any command trying to > | access > | the volume hangs, and umount returns: /sbin/umount.gfs: /lvm2: device > | is > | busy. > | > | Few questions - Since I have two volumes on this cluster and the lvm1 > | works > | just fine, would there be any suggestions to unmounting lvm2 in order > | to try > | and fix it? > | Is gfs_grow - bug free or not (use/do not use)? > | Is there any other way besides restarting the cluster/ nodes to get > | lvm2 > | back in operational state? > | -- > | Alan A. > > Hi Alan, > > Did you check in dmesg for kernel messages relating to the hang? > > I have seen some bugs in gfs_grow, and there are some fixes that > haven't made it out to all users yet, but you did not tell us which > version of the software you're using. You didn't even say whether > this is RHEL4/CentOS4 or RHEL5/Centos5 or another distro. > > I'm not aware of any bugs in the most recent gfs_grow that appears > in the cluster git repository. These gfs_grow fixes will trickle > out to various releases if you're not compiling from the source code, > so you may or may not have the fixed code. > > If your software is not recent, it's likely that an interrupted or > hung gfs_grow will end up corrupting the GFS file system. There is > a new, improved version of gfs_fsck that can repair the damage, but > again, you need a recent version of the software. > > Regards, > > Bob Peterson > Red Hat Clustering & GFS > > -- > Linux-cluster mailing list > Linux-cluster@redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- Alan A.
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster