From: Kushnir, Michael (NIH/NLM/LHC) [C] [mailto:michael.kush...@nih.gov]
Sent: Wednesday, April 25, 2012 12:52 PM
To: Nathan Patwardhan; ocfs2-users@oss.oracle.com
Subject: RE: RHEL 5.8, ocfs2 v1.4.4, stability issues


Ø  Hi Nathan,


Ø  I've no particular insight on splunk...


Ø  We use OCFS 1.4  with RHEL 5.8 VMs on ESX 4.1 and ESXi 5 as storage for a 
Hadoop cluster as well as an index and document store for application servers 
running a search engine application (java, tomcat, apache). We also export the 
same volumes from a few cluster nodes over NFS for developer access. We've had 
no stability issues or corruption.

I think I've identified the issue.

http://communities.vmware.com/thread/164967

In short, we had been allocated a shared vmdk through shared bus rather than as 
an RDM.  As a result, and after some digging into the logs, I found a bunch of 
SCSI reservation conflicts.  I'll get setup with RDM within the next week or so 
and will update with results as I have them.


From: Nathan Patwardhan 
[mailto:npatward...@llbean.com]<mailto:[mailto:npatward...@llbean.com]>
Sent: Tuesday, April 24, 2012 9:44 AM
To: ocfs2-users@oss.oracle.com<mailto:ocfs2-users@oss.oracle.com>
Subject: [Ocfs2-users] RHEL 5.8, ocfs2 v1.4.4, stability issues

Hi everyone,

We're running a two-node cluster w/ocfs2 v.1.4.4 under RHEL 5.8 and we're 
having major stability issues:

[root@splwww02 ~]# lsb_release -a ; uname -a ; rpm -qa |grep -i ocfs
LSB Version:    
:core-4.0-amd64:core-4.0-ia32:core-4.0-noarch:graphics-4.0-amd64:graphics-4.0-ia32:graphics-4.0-noarch:printing-4.0-amd64:printing-4.0-ia32:printing-4.0-noarch
Distributor ID: RedHatEnterpriseServer
Description:    Red Hat Enterprise Linux Server release 5.8 (Tikanga)
Release:        5.8
Codename:       Tikanga

Linux splwww02.llbean.com 2.6.18-308.1.1.el5 #1 SMP Fri Feb 17 16:51:01 EST 
2012 x86_64 x86_64 x86_64 GNU/Linux

ocfs2-2.6.18-308.1.1.el5-1.4.7-1.el5
ocfs2-tools-1.4.4-1.el5
ocfs2console-1.4.4-1.el5

These rpms were downloaded and installed from oracle.com.

The node are ESX VMs and the storage is shared to both VMs through vSphere.

We're using ocfs2 because we needed shared back end storage for a splunk 
implementation (splunk web search heads can used pooled storage and we didn't 
want to just use NFS for the obvious reasons - where shared storage seemed to 
be a better solution all the way around).

/etc/ocfs2/cluster.conf contains the following on each node:

cluster:
        node_count = 2
        name = splwww

node:
        ip_port = 7777
        ip_address = 10.130.245.40
        number = 0
        name = splwww01
        cluster = splwww

node:
        ip_port = 7777
        ip_address = 10.130.245.41
        number = 1
        name = splwww02
        cluster = splwww

Note that each node uses eth0 for clustering.  We do NOT currently have a 
private network setup in ESX for this purpose but CAN if this is recommended by 
people here.

o2cb status on both nodes:

Driver for "configfs": Loaded
Filesystem "configfs": Mounted
Driver for "ocfs2_dlmfs": Loaded
Filesystem "ocfs2_dlmfs": Mounted
Checking O2CB cluster splwww: Online
Heartbeat dead threshold = 3
  Network idle timeout: 90000
  Network keepalive delay: 5000
  Network reconnect delay: 5000
Checking O2CB heartbeat: Active

/etc/sysconfig/o2cb:

O2CB_ENABLED=true
O2CB_STACK=o2cb
O2CB_BOOTCLUSTER=splwww
O2CB_HEARTBEAT_THRESHOLD=3
O2CB_IDLE_TIMEOUT_MS=90000
O2CB_KEEPALIVE_DELAY_MS=5000
O2CB_RECONNECT_DELAY_MS=5000

Each host mounts the (ocfs2) file system then turns around and exports the  
file system where it mounts it on localhost as shown below (yes I know this is 
awful but splunk does not support ocfs2 and as such is unable to lock files 
properly no matter what options I give to mount an ocfs2 file system):

# From /etc/fstab
/dev/mapper/splwwwshared-splwwwshared--lv /opt/splunk/pooled ocfs2 
_netdev,datavolume,errors=remount-ro,rw,noatime 0 0

# From /etc/exports
/opt/splunk/pooled 127.0.0.1(rw,no_root_squash)

# From /bin/mount
/dev/mapper/splwwwshared-splwwwshared--lv on /opt/splunk/pooled type ocfs2 
(rw,_netdev,noatime,datavolume,errors=remount-ro,heartbeat=local)
localhost:/opt/splunk/pooled on /opt/splunk/mnt type nfs 
(rw,nordirplus,addr=127.0.0.1)

The above should supposedly allow us to enjoy the benefits of an ocfs2 
shared/clustered file system while functionally supporting what splunk needs.  
Unfortunately we have experienced a number of problems with stability, 
summarized here:


1.       Cluster timeouts, node(s) leaving cluster.



Apr 21 21:03:50 splwww01 kernel: 
(o2hb-DA9199FC9F,3090,0):o2hb_do_disk_heartbeat:768 ERROR: status = -52

Apr 21 21:03:52 splwww01 kernel: 
(o2hb-DA9199FC9F,3090,0):o2hb_do_disk_heartbeat:777 ERROR: Device "dm-5": 
another node is heartbeating in our slot!



2.       Ocfs2 crashes, nfs mount becoming unavailable, /opt/splunk/pooled 
becoming inaccessible, splunk crashing.

Apr 24 06:53:43 splwww01 kernel: 
(o2hb-DA9199FC9F,3089,0):ocfs2_dlm_eviction_cb:98 device (253,5): dlm has 
evicted node 1


3.       inode irregularities on either/both node(s).



Apr 24 06:40:47 splwww01 kernel: (nfsd,3330,1):ocfs2_read_locked_inode:472 
ERROR: status = -5



Apr 21 10:09:31 splwww01 kernel: 
(ocfs2rec,7676,0):ocfs2_clear_journal_error:690 ERROR: File system error -5 
recorded in journal 0.

Apr 21 10:09:31 splwww01 kernel: 
(ocfs2rec,7676,0):ocfs2_clear_journal_error:692 ERROR: File system on device 
dm-5 needs checking.



4.       Full system crashes on both node(s).

Ideally, I'd love to continue using ocfs2 because it solves an age-old problem 
for us, but since we can't keep either system stable for more than an hour or 
so, I'm looking for some more insight as to what we're seeing and how we might 
resolve some of these issues.

Note also that I'm using monit to check a .state file I created under 
/opt/splunk/pool and re-mount all the fs, restart splunk, etc, but this only 
takes us so far once the system(s) crash and reboot.

--
Nathan Patwardhan, Sr. System Engineer
npatward...@llbean.com<mailto:npatward...@llbean.com>
x26662

_______________________________________________
Ocfs2-users mailing list
Ocfs2-users@oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users

Reply via email to