My OpenSolaris snv134 on my production system does no more boot. I get kernel
panic on booting with the following message:
panic[cpu0]/thread=ffffff000f639c60: zfs: allocating allocated
segment(offset=422941467136 size=8192)
I attached the full stack trace per file attachment.
I need urgently a solution how to repair the filesystem to reboot the server
with all web services on it!
I tried to import my zpool with zfs-fuse 0.7.0 with Linux:
r...@rescue ~/tmp/official/src/cmd/zpool # ./zpool import
pool: rpool
id: 7742244823439206706
state: ONLINE
status: The pool was last accessed by another system.
action: The pool can be imported using its name or numeric identifier and
the '-f' flag.
see: http://www.sun.com/msg/ZFS-8000-EY
config:
rpool ONLINE
mirror-0 ONLINE
disk/by-id/ata-SAMSUNG_HD753LJ_S13UJDWZ301210-part5 ONLINE
disk/by-id/ata-SAMSUNG_HD753LJ_S13UJDWZ301202-part5 ONLINE
r...@rescue ~/tmp/official/src/cmd/zpool # ./zpool
r...@rescue ~/tmp/official/src/cmd/zpool # mkdir /mnt2
r...@rescue ~/tmp/official/src/cmd/zpool # ./zpool import -R /mnt2
7742244823439206706
cannot import 'rpool': pool may be in use from other system, it was last
accessed by (hostid: 0xc4db24) on Mon Sep 13 04:11:47 2010
use '-f' to import anyway
r...@rescue ~/tmp/official/src/cmd/zpool # ./zpool import -f -R /mnt2
7742244823439206706
zfsfuse_ioctl_read_loop(): file descriptor closed
internal error: out of memory
I need urgently a patch (by preference today). I'm ready to pay for it if
needed!
If I would be at least be able to repair my filesystem with zfs-fuse I would be
happy to recover my data.
I posted a bug about this on:
https://defect.opensolaris.org/bz/show_bug.cgi?id=17026
--
This message posted from opensolaris.org
_______________________________________________
zfs-code mailing list
[email protected]
http://mail.opensolaris.org/mailman/listinfo/zfs-code