> Has every instance of this bug involved ssh in the stack trace?

All the ones we've had to date. I initially thought it was apparmor
related because it mentions memory allocation etc but having removed
the packages and rebuilt initramfs I guess not.

If it happens again from now on, I'll make sure we always get a photo.

> Can you describe the software raid stack in detail?

It's really simple. 4x 2TB drives, each with a 2 partitons, the first
being 2GB and the remainder. sda1 and sdb1 are RAID1 as /boot, sdc1
and sdd1 are RAID1 for swap and the sda2,b2,c2,d2 are RAID5 as /
Currently the / filesystem is about 70% full.

> My first guess would be that OCFS is to blame.  Would it be possible to
> run this server for awhile without it, or is that impossible?  Can you
> either do without the SAN, or mount it as another fs type?

It's not possible really. The machine only mounts a snapshot of our
live SAN via iSCSI once a week to copy over virtual machine images. It
TENDS to lock up at those times, however it's also by far the busiest
time on the box as it'll be running rsnapshot backups at the same time
too. It's not unsual for the box to have 180% IO WAIT across the two
CPUs at those times.

We've put more RAM in the box again this week to see if that solves
it. I'm wondering if it's starting to swap to the point that the box
is getting overloaded. Time will tell if that has any effect. It's
hard because munin gives some idea of things like load/memory usage as
the box loads up, but isn't frequent enough to catch a spike!

Cheers

Alex

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/726461

Title:
  sshd on lucid causes kernel panic

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

Reply via email to