Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-08 Thread Andrew
So I tried a few more things..
I think the combination of the following in /etc/system made a difference:
set pcplusmp:apic_use_acpi=0
set sata:sata_max_queue_depth = 0x1
set zfs:zfs_recover=1 <<< I had this before
set aok=1   <<< I had this before too

I crossed my fingers, and it actually imported this time.. Somehow ..

solaria ~ # zpool status
  pool: itank
 state: ONLINE
 scrub: scrub in progress for 0h7m, 2.76% done, 4h33m to go
config:

NAME STATE READ WRITE CKSUM
itankONLINE   0 0 0
  raidz1 ONLINE   0 0 0
c12t1d0  ONLINE   0 0 0
c13t0d0  ONLINE   0 0 0
c11t0d0  ONLINE   0 0 0
c13t1d0  ONLINE   0 0 0
c11t1d0  ONLINE   0 0 0

Running some scrubs on it now, and I HOPE everything is okay...

Anything else you suggest I try before it's considered stable?
Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-07 Thread Andrew
Not too sure if it's much help. I enabled kernel pages and curproc.. Let me 
know if I need to enable "all" then.

solaria crash # echo "::status" | mdb -k
debugging live kernel (64-bit) on solaria
operating system: 5.11 snv_98 (i86pc)
solaria crash # echo "::stack" | mdb -k
solaria crash # echo "::msgbuf -v" | mdb -k
   TIMESTAMP   LOGCTL MESSAGE
2008 Nov  7 18:53:55 ff01c901dcf0   capacity = 1953525168 sectors
2008 Nov  7 18:53:55 ff01c901db70 /[EMAIL PROTECTED],0/pci1022,[EMAIL 
PROTECTED]/pci1095,[EMAIL PROTECTED] :
2008 Nov  7 18:53:55 ff01c901d9f0   SATA disk device at port 0
2008 Nov  7 18:53:55 ff01c901d870
model ST31000340AS
2008 Nov  7 18:53:55 ff01c901d6f0   firmware SD15
2008 Nov  7 18:53:55 ff01c901d570   serial number 
2008 Nov  7 18:53:55 ff01c901d3f0   supported features:
2008 Nov  7 18:53:55 ff01c901d270
 48-bit LBA, DMA, Native Command Queueing, SMART self-test
2008 Nov  7 18:53:55 ff01c901d0f0   SATA Gen1 signaling speed (1.5Gbps)
2008 Nov  7 18:53:55 ff01c901adf0   Supported queue depth 32, limited to 31
2008 Nov  7 18:53:55 ff01c901ac70   capacity = 1953525168 sectors
2008 Nov  7 18:53:55 ff01c901aaf0 /[EMAIL PROTECTED],0/pci1022,[EMAIL 
PROTECTED]/pci1095,[EMAIL PROTECTED] :
2008 Nov  7 18:53:55 ff01c901a970   SATA disk device at port 0
2008 Nov  7 18:53:55 ff01c901a7f0
model Maxtor 6L250S0
2008 Nov  7 18:53:55 ff01c901a670   firmware BANC1G10
2008 Nov  7 18:53:55 ff01c901a4f0   serial number
2008 Nov  7 18:53:55 ff01c901a370   supported features:
2008 Nov  7 18:53:55 ff01c901a2b0
 48-bit LBA, DMA, Native Command Queueing, SMART self-test
2008 Nov  7 18:53:55 ff01c901a130   SATA Gen1 signaling speed (1.5Gbps)
2008 Nov  7 18:53:55 ff01c901a070   Supported queue depth 32, limited to 31
2008 Nov  7 18:53:55 ff01c9017ef0   capacity = 490234752 sectors
2008 Nov  7 18:53:55 ff01c9017d70 pseudo-device: ramdisk1024
2008 Nov  7 18:53:55 ff01c9017bf0 ramdisk1024 is /pseudo/[EMAIL PROTECTED]
2008 Nov  7 18:53:55 ff01c9017a70 NOTICE: e1000g0 registered
2008 Nov  7 18:53:55 ff01c90179b0
pcplusmp: pci8086,100e (e1000g) instance 0 vector 0x14 ioapic 0x2 intin 0x14 is
bound to cpu 0
2008 Nov  7 18:53:55 ff01c90178f0
Intel(R) PRO/1000 Network Connection, Driver Ver. 5.2.12
2008 Nov  7 18:53:56 ff01c9017830 pseudo-device: lockstat0
2008 Nov  7 18:53:56 ff01c9017770 lockstat0 is /pseudo/[EMAIL PROTECTED]
2008 Nov  7 18:53:56 ff01c90176b0 sd6 at si31240: target 0 lun 0
2008 Nov  7 18:53:56 ff01c90175f0
sd6 is /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL 
PROTECTED]/[EMAIL PROTECTED],0
2008 Nov  7 18:53:56 ff01c9017530 sd5 at si31242: target 0 lun 0
2008 Nov  7 18:53:56 ff01c9017470
sd5 is /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL 
PROTECTED]/[EMAIL PROTECTED],0
2008 Nov  7 18:53:56 ff01c90173b0 sd4 at si31241: target 0 lun 0
2008 Nov  7 18:53:56 ff01c90172f0
sd4 is /[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL 
PROTECTED]/[EMAIL PROTECTED],0
2008 Nov  7 18:53:56 ff01c9017230
/[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 (sd4) online
2008 Nov  7 18:53:56 ff01c9017170 /[EMAIL PROTECTED],0/pci1022,[EMAIL 
PROTECTED]/pci1095,[EMAIL PROTECTED] :
2008 Nov  7 18:53:56 ff01c90170b0   SATA disk device at port 1
2008 Nov  7 18:53:56 ff01c9087f30
model ST31000340AS
2008 Nov  7 18:53:56 ff01c9087e70   firmware SD15
2008 Nov  7 18:53:56 ff01c9087db0   serial number
2008 Nov  7 18:53:56 ff01c9087cf0   supported features:
2008 Nov  7 18:53:56 ff01c9087c30
 48-bit LBA, DMA, Native Command Queueing, SMART self-test
2008 Nov  7 18:53:56 ff01c9087b70   SATA Gen1 signaling speed (1.5Gbps)
2008 Nov  7 18:53:56 ff01c9087ab0   Supported queue depth 32, limited to 31
2008 Nov  7 18:53:56 ff01c90879f0   capacity = 1953525168 sectors
2008 Nov  7 18:53:56 ff01c9087930
/[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 (sd6) online
2008 Nov  7 18:53:56 ff01c9087870
/[EMAIL PROTECTED],0/pci1022,[EMAIL PROTECTED]/pci1095,[EMAIL PROTECTED]/[EMAIL 
PROTECTED],0 (sd5) online
2008 Nov  7 18:53:56 ff01c90877b0 /[EMAIL PROTECTED],0/pci1022,[EMAIL 
PROTECTED]/pci1095,[EMAIL PROTECTED] :
2008 Nov  7 18:53:56 ff01c90876f0   SATA disk device at port 1
2008 Nov  7 18:53:56 ff01c9087630
model ST31000340AS
2008 Nov  7 18:53:56 ff01c9087570   firmware SD15
2008 Nov  7 18:53:56 ff01c90874b0   serial number 
2008 Nov  7 18:53:56 ff01c90873f0   supported features:
2008 Nov  7 18:53:56 ff01c9087330
 48-bit LBA, DMA, Native Command Queueing, SMART self-test
2008 Nov  7 18:53:56 ff01c9087270   SATA Gen1 signaling speed (1.5Gbps)
2008 Nov  7 18:53:56 ff01c90871b0   Supported queu

Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-07 Thread Victor Latushkin
Andrew пишет:
> hey Victor,
> 
> Where would i find that? I'm still somewhat getting used to the
> Solaris environment. /var/adm/messages doesn't seem to show any Panic
> info.. I only have remote access via SSH, so I hope I can do
> something with dtrace to pull it.

Do you have anything in /var/crash/ ?

If yes, then do something like this and provide output:

cd /var/crash/
echo "::status" | mdb -k 
echo "::stack" | mdb -k 
echo "::msgbuf -v" | mdb -k 

victor

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-07 Thread Andrew
hey Victor,

Where would i find that? I'm still somewhat getting used to the Solaris 
environment. /var/adm/messages doesn't seem to show any Panic info.. I only 
have remote access via SSH, so I hope I can do something with dtrace to pull it.

Thanks,
Andrew
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-07 Thread Victor Latushkin
Andrew,

Andrew wrote:
> Thanks a lot! Google didn't seem to cooperate as well as I had hoped.
> 
> 
> Still no dice on the import. I only have shell access on my
> Blackberry Pearl from where I am, so it's kind of hard, but I'm
> managing.. I've tried the OP's exact commands, and even trying to
> import array as ro, yet the system still wants to panic.. I really
> hope I don't have to redo my array, and lose everything as I still
> have faith in ZFS...

could you please post a little bit more details - at least panic string 
and stack backtrace during panic. That would help to get an idea about 
what might went wrong.

regards,
victor
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-07 Thread Andrew
Thanks a lot! Google didn't seem to cooperate as well as I had hoped.

Still no dice on the import. I only have shell access on my Blackberry Pearl 
from where I am, so it's kind of hard, but I'm managing.. I've tried the OP's 
exact commands, and even trying to import array as ro, yet the system still 
wants to panic.. I really hope I don't have to redo my array, and lose 
everything as I still have faith in ZFS...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-07 Thread Jim Dunham
Andrew,

> I woke up yesterday morning, only to discover my system kept  
> rebooting..
>
> It's been running fine for the last while. I upgraded to snv 98 a  
> couple weeks back (from 95), and had upgraded my RaidZ Zpool from  
> version 11 to 13 for improved scrub performance.
>
> After some research it turned out that, on bootup, importing my 4tb  
> raidZ array was causing the system to panic (similar to this OP's  
> error). I got that bypassed, and can now at least boot the system..
>
> However, when I try anything (like mdb -kw), it advises me that  
> there is no command line editing because: "mdb: no terminal data  
> available for TERM=vt320. term init failed: command-line editing and  
> prompt will not be available". This means I can't really try what  
> aldredmr had done in mdb, and I really don't have any experience in  
> it. I upgraded to snv_100 (November), but experiencing the exact  
> same issues
>
> If anyone has some insight, it would be greatly appreciated. Thanks

I have the same problem SSH'ing in from my Mac OS X, which sets the  
TERM type to 'xterm-color', also not supported.

Do the following, depending on your default shell. and you should be  
all set.

TERM=vt100; export TERM
or
setenv TERM vt100

>
> -- 
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Jim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-11-07 Thread Andrew
I woke up yesterday morning, only to discover my system kept rebooting..

It's been running fine for the last while. I upgraded to snv 98 a couple weeks 
back (from 95), and had upgraded my RaidZ Zpool from version 11 to 13 for 
improved scrub performance.

After some research it turned out that, on bootup, importing my 4tb raidZ array 
was causing the system to panic (similar to this OP's error). I got that 
bypassed, and can now at least boot the system..

However, when I try anything (like mdb -kw), it advises me that there is no 
command line editing because: "mdb: no terminal data available for TERM=vt320. 
term init failed: command-line editing and prompt will not be available". This 
means I can't really try what aldredmr had done in mdb, and I really don't have 
any experience in it. I upgraded to snv_100 (November), but experiencing the 
exact same issues. 

If anyone has some insight, it would be greatly appreciated. Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS, Kernel Panic on import

2008-08-29 Thread Mike Aldred
Ok, I've managed to get around the kernel panic.

[EMAIL PROTECTED]:~/Download$ pfexec mdb -kw
Loading modules: [ unix genunix specfs dtrace cpu.generic uppc pcplusmp 
scsi_vhci zfs sd ip hook neti sctp arp usba uhci s1394 fctl md lofs random sppp 
ipc ptm fcip fcp cpc crypto logindmux ii nsctl sdbc ufs rdc nsmb sv ]
> vdev_uberblock_compare+0x49/W 1
vdev_uberblock_compare+0x49:0x  =   0x1
> vdev_uberblock_compare+0x3b/W 1
vdev_uberblock_compare+0x3b:0x  =   0x1
> zfsvfs_setup+0x60/v 0xeb
zfsvfs_setup+0x60:  0x74=   0xeb
> 

This has let me import the pool, without the kernel panicing, and I'm doing a 
scrub on the pool now.
The thing is, I don't know what those commands do, could anyone enlighten me?
--
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS, Kernel Panic on import

2008-08-29 Thread Mike Aldred
G'day,

I've got a OpenSolaris server n95, that I use for media, serving.  It's uses a 
DQ35JOE motherboard, dual core, and I have my rpool mirrored on two IDE 40GB 
drives, and my media mirrored on 2 x 500GB SATA drives.

I've got a few CIFS shares on the media drive, and I'm using MediaTomb to 
stream to my PS3. No problems at all, until today.  I was at work (obviously 
not working too hard :) ), when I thought that I really should scrub my pools, 
since I hasn't done it for awhile.  So I SSHed into the box, and did a scrub on 
both pools.

A few minutes later, I lost my SSH connection... uh oh, but not too worried, I 
thought that the ADSL must've gone down or something.

Came home, and the server is in a reboot loop, kernel panic.  Nuts...

Booted into the LiveDVD of snv_95, no problem, set about scrubbing my rpool, 
everything is good, until I decide to import and start scrubbing my storage 
pool... kernel panic... Nuts...

Removed the storage pool drives from the machine, no problem, boots up fine and 
starts scrubbing the rpool again.  No problems.  Decided to more the storage 
drives over to my desktop machine, try to import kernel panic...

So, the trick is, how do I fix it?

I've read a few posts, and I've seen other people with similar problems, but I 
have to admit I'm simply not smart enough to solve the problem, so, anyone got 
any ideas?

Here's some info that I hope prove useful.

[EMAIL PROTECTED]:~/Desktop$ pfexec zpool import
  pool: storage
id: 6933883927787501942
 state: ONLINE
status: The pool is formatted using an older on-disk version.
action: The pool can be imported using its name or numeric identifier, though
some features will not be available without an explicit 'zpool upgrade'.
config:

storage ONLINE
  mirrorONLINE
c3t3d0  ONLINE
c3t2d0  ONLINE

[EMAIL PROTECTED]:~/Desktop$ zdb -uuu -e storage
Uberblock

magic = 00bab10c
version = 10
txg = 3818020
guid_sum = 6700303293925244073
timestamp = 1220003402 UTC = Fri Aug 29 17:50:02 2008
rootbp = [L0 DMU objset] 400L/200P DVA[0]=<0:6a00058e00:200> 
DVA[1]=<0:2a8600:200> DVA[2]=<0:3800050600:200> fletcher4 lzjb LE 
contiguous birth=3818020 fill=170 
cksum=8b56cdef9:38379d3cd95:b809c1c9bb15:197649b024bfd1

[EMAIL PROTECTED]:~/Desktop$ zdb -e -bb storage

Traversing all blocks to verify nothing leaked ...

No leaks (block sum matches space maps exactly)

bp count: 3736040
bp logical:484538716672  avg: 129693
bp physical:   484064542720  avg: 129566compression:   1.00
bp allocated:  484259193344  avg: 129618compression:   1.00
SPA allocated: 484259193344 used: 97.20%

Blocks  LSIZE   PSIZE   ASIZE avgcomp   %Total  Type
   105  1.11M339K   1017K9.7K3.35 0.00  deferred free
 232K  4K   12.0K   6.00K8.00 0.00  object directory
 2 1K  1K   3.00K   1.50K1.00 0.00  object array
 116K   1.50K   4.50K   4.50K   10.67 0.00  packed nvlist
 -  -   -   -   -   --  packed nvlist size
 116K   3.00K   9.00K   9.00K5.33 0.00  bplist
 -  -   -   -   -   --  bplist header
 -  -   -   -   -   --  SPA space map header
   373  2.14M801K   2.35M   6.44K2.73 0.00  SPA space map
 3  40.0K   40.0K   40.0K   13.3K1.00 0.00  ZIL intent log
   552  8.62M   2.40M   4.82M   8.94K3.60 0.00  DMU dnode
 8 8K  4K   8.50K   1.06K2.00 0.00  DMU objset
 -  -   -   -   -   --  DSL directory
 8 4K  4K   12.0K   1.50K1.00 0.00  DSL directory child map
 7  3.50K   3.50K   10.5K   1.50K1.00 0.00  DSL dataset snap map
15   225K   25.0K   75.0K   5.00K8.98 0.00  DSL props
 -  -   -   -   -   --  DSL dataset
 -  -   -   -   -   --  ZFS znode
 -  -   -   -   -   --  ZFS V0 ACL
 3.56M   451G451G451G127K1.00   100.00  ZFS plain file
 1.55K   9.9M   1.51M   3.03M   1.95K6.55 0.00  ZFS directory
 7  3.50K   3.50K   7.00K  1K1.00 0.00  ZFS master node
40   550K   87.0K174K   4.35K6.32 0.00  ZFS delete queue
 -  -   -   -   -   --  zvol object
 -  -   -   -   -   --  zvol prop
 -  -   -   -   -   --  other uint8[]
 -  -   -   -   -   --  other uint64[]
 1512 512   1.50K   1.50K1.00 0.00  other ZAP
 -  -   -   -   -   --  persistent error log
 1   128K   10.0K   30.0K   30.0K   12.80 0.00  SPA history
 -  -   -   -   -