Re: [zfs-discuss] ZFS, Kernel Panic on import
So I tried a few more things.. I think the combination of the following in /etc/system made a difference: set pcplusmp:apic_use_acpi=0 set sata:sata_max_queue_depth = 0x1 set zfs:zfs_recover=1 <<< I had this before set aok=1 <<< I had this before too I crossed my fingers, and it actually imported this time.. Somehow .. solaria ~ # zpool status pool: itank state: ONLINE scrub: scrub in progress for 0h7m, 2.76% done, 4h33m to go config: NAME STATE READ WRITE CKSUM itankONLINE 0 0 0 raidz1 ONLINE 0 0 0 c12t1d0 ONLINE 0 0 0 c13t0d0 ONLINE 0 0 0 c11t0d0 ONLINE 0 0 0 c13t1d0 ONLINE 0 0 0 c11t1d0 ONLINE 0 0 0 Running some scrubs on it now, and I HOPE everything is okay... Anything else you suggest I try before it's considered stable? Thanks -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS, Kernel Panic on import
Not too sure if it's much help. I enabled kernel pages and curproc.. Let me know if I need to enable "all" then. solaria crash # echo "::status" | mdb -k debugging live kernel (64-bit) on solaria operating system: 5.11 snv_98 (i86pc) solaria crash # echo "::stack" | mdb -k solaria crash # echo "::msgbuf -v" | mdb -k TIMESTAMP LOGCTL MESSAGE 2008 Nov 7 18:53:55 ff01c901dcf0 capacity = 1953525168 sectors 2008 Nov 7 18:53:55 ff01c901db70 /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED] : 2008 Nov 7 18:53:55 ff01c901d9f0 SATA disk device at port 0 2008 Nov 7 18:53:55 ff01c901d870 model ST31000340AS 2008 Nov 7 18:53:55 ff01c901d6f0 firmware SD15 2008 Nov 7 18:53:55 ff01c901d570 serial number 2008 Nov 7 18:53:55 ff01c901d3f0 supported features: 2008 Nov 7 18:53:55 ff01c901d270 48-bit LBA, DMA, Native Command Queueing, SMART self-test 2008 Nov 7 18:53:55 ff01c901d0f0 SATA Gen1 signaling speed (1.5Gbps) 2008 Nov 7 18:53:55 ff01c901adf0 Supported queue depth 32, limited to 31 2008 Nov 7 18:53:55 ff01c901ac70 capacity = 1953525168 sectors 2008 Nov 7 18:53:55 ff01c901aaf0 /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED] : 2008 Nov 7 18:53:55 ff01c901a970 SATA disk device at port 0 2008 Nov 7 18:53:55 ff01c901a7f0 model Maxtor 6L250S0 2008 Nov 7 18:53:55 ff01c901a670 firmware BANC1G10 2008 Nov 7 18:53:55 ff01c901a4f0 serial number 2008 Nov 7 18:53:55 ff01c901a370 supported features: 2008 Nov 7 18:53:55 ff01c901a2b0 48-bit LBA, DMA, Native Command Queueing, SMART self-test 2008 Nov 7 18:53:55 ff01c901a130 SATA Gen1 signaling speed (1.5Gbps) 2008 Nov 7 18:53:55 ff01c901a070 Supported queue depth 32, limited to 31 2008 Nov 7 18:53:55 ff01c9017ef0 capacity = 490234752 sectors 2008 Nov 7 18:53:55 ff01c9017d70 pseudo-device: ramdisk1024 2008 Nov 7 18:53:55 ff01c9017bf0 ramdisk1024 is /pseudo/[EMAIL PROTECTED] 2008 Nov 7 18:53:55 ff01c9017a70 NOTICE: e1000g0 registered 2008 Nov 7 18:53:55 ff01c90179b0 pcplusmp: pci8086,100e (e1000g) instance 0 vector 0x14 ioapic 0x2 intin 0x14 is bound to cpu 0 2008 Nov 7 18:53:55 ff01c90178f0 Intel(R) PRO/1000 Network Connection, Driver Ver. 5.2.12 2008 Nov 7 18:53:56 ff01c9017830 pseudo-device: lockstat0 2008 Nov 7 18:53:56 ff01c9017770 lockstat0 is /pseudo/[EMAIL PROTECTED] 2008 Nov 7 18:53:56 ff01c90176b0 sd6 at si31240: target 0 lun 0 2008 Nov 7 18:53:56 ff01c90175f0 sd6 is /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL PROTECTED],0 2008 Nov 7 18:53:56 ff01c9017530 sd5 at si31242: target 0 lun 0 2008 Nov 7 18:53:56 ff01c9017470 sd5 is /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL PROTECTED],0 2008 Nov 7 18:53:56 ff01c90173b0 sd4 at si31241: target 0 lun 0 2008 Nov 7 18:53:56 ff01c90172f0 sd4 is /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL PROTECTED],0 2008 Nov 7 18:53:56 ff01c9017230 /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd4) online 2008 Nov 7 18:53:56 ff01c9017170 /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED] : 2008 Nov 7 18:53:56 ff01c90170b0 SATA disk device at port 1 2008 Nov 7 18:53:56 ff01c9087f30 model ST31000340AS 2008 Nov 7 18:53:56 ff01c9087e70 firmware SD15 2008 Nov 7 18:53:56 ff01c9087db0 serial number 2008 Nov 7 18:53:56 ff01c9087cf0 supported features: 2008 Nov 7 18:53:56 ff01c9087c30 48-bit LBA, DMA, Native Command Queueing, SMART self-test 2008 Nov 7 18:53:56 ff01c9087b70 SATA Gen1 signaling speed (1.5Gbps) 2008 Nov 7 18:53:56 ff01c9087ab0 Supported queue depth 32, limited to 31 2008 Nov 7 18:53:56 ff01c90879f0 capacity = 1953525168 sectors 2008 Nov 7 18:53:56 ff01c9087930 /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd6) online 2008 Nov 7 18:53:56 ff01c9087870 /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5) online 2008 Nov 7 18:53:56 ff01c90877b0 /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED] : 2008 Nov 7 18:53:56 ff01c90876f0 SATA disk device at port 1 2008 Nov 7 18:53:56 ff01c9087630 model ST31000340AS 2008 Nov 7 18:53:56 ff01c9087570 firmware SD15 2008 Nov 7 18:53:56 ff01c90874b0 serial number 2008 Nov 7 18:53:56 ff01c90873f0 supported features: 2008 Nov 7 18:53:56 ff01c9087330 48-bit LBA, DMA, Native Command Queueing, SMART self-test 2008 Nov 7 18:53:56 ff01c9087270 SATA Gen1 signaling speed (1.5Gbps) 2008 Nov 7 18:53:56 ff01c90871b0 Supported queu
Re: [zfs-discuss] ZFS, Kernel Panic on import
Andrew пишет: > hey Victor, > > Where would i find that? I'm still somewhat getting used to the > Solaris environment. /var/adm/messages doesn't seem to show any Panic > info.. I only have remote access via SSH, so I hope I can do > something with dtrace to pull it. Do you have anything in /var/crash/ ? If yes, then do something like this and provide output: cd /var/crash/ echo "::status" | mdb -k echo "::stack" | mdb -k echo "::msgbuf -v" | mdb -k victor ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS, Kernel Panic on import
hey Victor, Where would i find that? I'm still somewhat getting used to the Solaris environment. /var/adm/messages doesn't seem to show any Panic info.. I only have remote access via SSH, so I hope I can do something with dtrace to pull it. Thanks, Andrew -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS, Kernel Panic on import
Andrew, Andrew wrote: > Thanks a lot! Google didn't seem to cooperate as well as I had hoped. > > > Still no dice on the import. I only have shell access on my > Blackberry Pearl from where I am, so it's kind of hard, but I'm > managing.. I've tried the OP's exact commands, and even trying to > import array as ro, yet the system still wants to panic.. I really > hope I don't have to redo my array, and lose everything as I still > have faith in ZFS... could you please post a little bit more details - at least panic string and stack backtrace during panic. That would help to get an idea about what might went wrong. regards, victor ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS, Kernel Panic on import
Thanks a lot! Google didn't seem to cooperate as well as I had hoped. Still no dice on the import. I only have shell access on my Blackberry Pearl from where I am, so it's kind of hard, but I'm managing.. I've tried the OP's exact commands, and even trying to import array as ro, yet the system still wants to panic.. I really hope I don't have to redo my array, and lose everything as I still have faith in ZFS... -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS, Kernel Panic on import
Andrew, > I woke up yesterday morning, only to discover my system kept > rebooting.. > > It's been running fine for the last while. I upgraded to snv 98 a > couple weeks back (from 95), and had upgraded my RaidZ Zpool from > version 11 to 13 for improved scrub performance. > > After some research it turned out that, on bootup, importing my 4tb > raidZ array was causing the system to panic (similar to this OP's > error). I got that bypassed, and can now at least boot the system.. > > However, when I try anything (like mdb -kw), it advises me that > there is no command line editing because: "mdb: no terminal data > available for TERM=vt320. term init failed: command-line editing and > prompt will not be available". This means I can't really try what > aldredmr had done in mdb, and I really don't have any experience in > it. I upgraded to snv_100 (November), but experiencing the exact > same issues > > If anyone has some insight, it would be greatly appreciated. Thanks I have the same problem SSH'ing in from my Mac OS X, which sets the TERM type to 'xterm-color', also not supported. Do the following, depending on your default shell. and you should be all set. TERM=vt100; export TERM or setenv TERM vt100 > > -- > This message posted from opensolaris.org > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss Jim ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS, Kernel Panic on import
I woke up yesterday morning, only to discover my system kept rebooting.. It's been running fine for the last while. I upgraded to snv 98 a couple weeks back (from 95), and had upgraded my RaidZ Zpool from version 11 to 13 for improved scrub performance. After some research it turned out that, on bootup, importing my 4tb raidZ array was causing the system to panic (similar to this OP's error). I got that bypassed, and can now at least boot the system.. However, when I try anything (like mdb -kw), it advises me that there is no command line editing because: "mdb: no terminal data available for TERM=vt320. term init failed: command-line editing and prompt will not be available". This means I can't really try what aldredmr had done in mdb, and I really don't have any experience in it. I upgraded to snv_100 (November), but experiencing the exact same issues. If anyone has some insight, it would be greatly appreciated. Thanks -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS, Kernel Panic on import
Ok, I've managed to get around the kernel panic. [EMAIL PROTECTED]:~/Download$ pfexec mdb -kw Loading modules: [ unix genunix specfs dtrace cpu.generic uppc pcplusmp scsi_vhci zfs sd ip hook neti sctp arp usba uhci s1394 fctl md lofs random sppp ipc ptm fcip fcp cpc crypto logindmux ii nsctl sdbc ufs rdc nsmb sv ] > vdev_uberblock_compare+0x49/W 1 vdev_uberblock_compare+0x49:0x = 0x1 > vdev_uberblock_compare+0x3b/W 1 vdev_uberblock_compare+0x3b:0x = 0x1 > zfsvfs_setup+0x60/v 0xeb zfsvfs_setup+0x60: 0x74= 0xeb > This has let me import the pool, without the kernel panicing, and I'm doing a scrub on the pool now. The thing is, I don't know what those commands do, could anyone enlighten me? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS, Kernel Panic on import
G'day, I've got a OpenSolaris server n95, that I use for media, serving. It's uses a DQ35JOE motherboard, dual core, and I have my rpool mirrored on two IDE 40GB drives, and my media mirrored on 2 x 500GB SATA drives. I've got a few CIFS shares on the media drive, and I'm using MediaTomb to stream to my PS3. No problems at all, until today. I was at work (obviously not working too hard :) ), when I thought that I really should scrub my pools, since I hasn't done it for awhile. So I SSHed into the box, and did a scrub on both pools. A few minutes later, I lost my SSH connection... uh oh, but not too worried, I thought that the ADSL must've gone down or something. Came home, and the server is in a reboot loop, kernel panic. Nuts... Booted into the LiveDVD of snv_95, no problem, set about scrubbing my rpool, everything is good, until I decide to import and start scrubbing my storage pool... kernel panic... Nuts... Removed the storage pool drives from the machine, no problem, boots up fine and starts scrubbing the rpool again. No problems. Decided to more the storage drives over to my desktop machine, try to import kernel panic... So, the trick is, how do I fix it? I've read a few posts, and I've seen other people with similar problems, but I have to admit I'm simply not smart enough to solve the problem, so, anyone got any ideas? Here's some info that I hope prove useful. [EMAIL PROTECTED]:~/Desktop$ pfexec zpool import pool: storage id: 6933883927787501942 state: ONLINE status: The pool is formatted using an older on-disk version. action: The pool can be imported using its name or numeric identifier, though some features will not be available without an explicit 'zpool upgrade'. config: storage ONLINE mirrorONLINE c3t3d0 ONLINE c3t2d0 ONLINE [EMAIL PROTECTED]:~/Desktop$ zdb -uuu -e storage Uberblock magic = 00bab10c version = 10 txg = 3818020 guid_sum = 6700303293925244073 timestamp = 1220003402 UTC = Fri Aug 29 17:50:02 2008 rootbp = [L0 DMU objset] 400L/200P DVA[0]=<0:6a00058e00:200> DVA[1]=<0:2a8600:200> DVA[2]=<0:3800050600:200> fletcher4 lzjb LE contiguous birth=3818020 fill=170 cksum=8b56cdef9:38379d3cd95:b809c1c9bb15:197649b024bfd1 [EMAIL PROTECTED]:~/Desktop$ zdb -e -bb storage Traversing all blocks to verify nothing leaked ... No leaks (block sum matches space maps exactly) bp count: 3736040 bp logical:484538716672 avg: 129693 bp physical: 484064542720 avg: 129566compression: 1.00 bp allocated: 484259193344 avg: 129618compression: 1.00 SPA allocated: 484259193344 used: 97.20% Blocks LSIZE PSIZE ASIZE avgcomp %Total Type 105 1.11M339K 1017K9.7K3.35 0.00 deferred free 232K 4K 12.0K 6.00K8.00 0.00 object directory 2 1K 1K 3.00K 1.50K1.00 0.00 object array 116K 1.50K 4.50K 4.50K 10.67 0.00 packed nvlist - - - - - -- packed nvlist size 116K 3.00K 9.00K 9.00K5.33 0.00 bplist - - - - - -- bplist header - - - - - -- SPA space map header 373 2.14M801K 2.35M 6.44K2.73 0.00 SPA space map 3 40.0K 40.0K 40.0K 13.3K1.00 0.00 ZIL intent log 552 8.62M 2.40M 4.82M 8.94K3.60 0.00 DMU dnode 8 8K 4K 8.50K 1.06K2.00 0.00 DMU objset - - - - - -- DSL directory 8 4K 4K 12.0K 1.50K1.00 0.00 DSL directory child map 7 3.50K 3.50K 10.5K 1.50K1.00 0.00 DSL dataset snap map 15 225K 25.0K 75.0K 5.00K8.98 0.00 DSL props - - - - - -- DSL dataset - - - - - -- ZFS znode - - - - - -- ZFS V0 ACL 3.56M 451G451G451G127K1.00 100.00 ZFS plain file 1.55K 9.9M 1.51M 3.03M 1.95K6.55 0.00 ZFS directory 7 3.50K 3.50K 7.00K 1K1.00 0.00 ZFS master node 40 550K 87.0K174K 4.35K6.32 0.00 ZFS delete queue - - - - - -- zvol object - - - - - -- zvol prop - - - - - -- other uint8[] - - - - - -- other uint64[] 1512 512 1.50K 1.50K1.00 0.00 other ZAP - - - - - -- persistent error log 1 128K 10.0K 30.0K 30.0K 12.80 0.00 SPA history - - - - -