Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-18 Thread Bob Friesenhahn

On Fri, 19 Aug 2011, Edho Arief wrote:


Asking Oracle for help without support contract would be like shouting
in vacuum space...


It seems that obtaining an Oracle support contract or a contract 
renewal is equally frustrating.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-18 Thread Edho Arief
On Fri, Aug 19, 2011 at 12:19 AM, Stu Whitefish  wrote:
>> From: Thomas Gouverneur 
>
>> To: zfs-discuss@opensolaris.org
>> Cc:
>> Sent: Thursday, August 18, 2011 5:11:16 PM
>> Subject: Re: [zfs-discuss] Kernel panic on zpool import. 200G of data 
>> inaccessible!
>>
>> Have you already extracted the core file of the kernel crash ?
>
> Nope, not a clue how to do that and I have installed Windows on this box 
> instead of Solaris since I can't get my data back from ZFS.
> I have my two drives the pool is on disconnected so if this ever gets 
> resolved I can reinstall Solaris and start learning again.
>
>> (and btw activated dump device for such dumping happen at next reboot...)
>
> This was a development box for me to see how I get along with Solaris. I'm 
> afraid I don't have any experience in Solaris to understand your question.
>
>> Have you also tried applying the latest kernel/zfs patches and try importing 
>> the pool afterwards ?
>
> Wish I had them and knew what to do with them if I had them. Somebody on OTN 
> noted this is supposed to be fixed by 142910 but
> I didn't hear back yet whether it fixes an pool ZFS won't import, or it only 
> stops it from happening in the first place. Don't have a service
> contract as I say this box was my first try with Solaris and it is a homebrew 
> system not on Oracle's support list.
>
> I am sure if there is a patch for this or a way to get my 200G of data back 
> some kind soul at Oracle will certainly help me since I lost
> my data and getting it back isn't a matter of convenience. What an 
> opportunity to generate some old fashioned goodwill!  :-)
>

lots of replies and no suggestion to try on FreeBSD. How about trying
on one? I believe if it crashed on FreeBSD, the developers would be
interested in helping to solve it. Try using the 9.0-beta1 since
8.2-release has some problems importing certain zpools.

Asking Oracle for help without support contract would be like shouting
in vacuum space...
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS performance question over NFS

2011-08-18 Thread Bob Friesenhahn

On Thu, 18 Aug 2011, Thomas Nau wrote:


Tim
the client is identical as the server but no SAS drives attached.
Also right now only one 1gbit Intel NIC Is available


I don't know what the request pattern from filebench looks like but it 
seems like your ZEUS RAM devices are not keeping up or else many 
requests are bypassing the ZEUS RAM devices.


Note that very large synchronous writes will bypass your ZEUS RAM 
device and go directly to a log in the main store.  Small (<= 128K) 
writes should directly benefit from the dedicated zil device.


Find a copy of zilstat.ksh and run it while filebench is running in 
order to understand more about what is going on.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS performance question over NFS

2011-08-18 Thread Thomas Nau
Tim
the client is identical as the server but no SAS drives attached.
Also right now only one 1gbit Intel NIC Is available

Thomas


Am 18.08.2011 um 17:49 schrieb Tim Cook :

> What are the specs on the client?
> 
> On Aug 18, 2011 10:28 AM, "Thomas Nau"  wrote:
> > Dear all.
> > We finally got all the parts for our new fileserver following several
> > recommendations we got over this list. We use
> > 
> > Dell R715, 96GB RAM, dual 8-core Opterons
> > 1 10GE Intel dual-port NIC
> > 2 LSI 9205-8e SAS controllers
> > 2 DataON DNS-1600 JBOD chassis
> > 46 Seagate constellation SAS drives
> > 2 STEC ZEUS RAM
> > 
> > 
> > The base zpool config utilizes 42 drives plus the STECs as mirrored
> > log devices. The Seagates are setup as a stripe of 7 times 6-drive-RAIDZ2
> > junks plus as said a dedicated ZIL made of the mirrored STECs.
> > 
> > As a quick'n dirty check we ran "filebench" with the "fileserver"
> > workload. Running locally we get
> > 
> > statfile1 5476ops/s 0.0mb/s 0.6ms/op 179us/op-cpu
> > deletefile1 5476ops/s 0.0mb/s 1.0ms/op 454us/op-cpu
> > closefile3 5476ops/s 0.0mb/s 0.0ms/op 5us/op-cpu
> > readfile1 5476ops/s 729.5mb/s 0.2ms/op 128us/op-cpu
> > openfile2 5477ops/s 0.0mb/s 0.8ms/op 204us/op-cpu
> > closefile2 5477ops/s 0.0mb/s 0.0ms/op 5us/op-cpu
> > appendfilerand1 5477ops/s 42.8mb/s 0.3ms/op 184us/op-cpu
> > openfile1 5477ops/s 0.0mb/s 0.9ms/op 209us/op-cpu
> > closefile1 5477ops/s 0.0mb/s 0.0ms/op 6us/op-cpu
> > wrtfile1 5477ops/s 688.4mb/s 0.4ms/op 220us/op-cpu
> > createfile1 5477ops/s 0.0mb/s 2.7ms/op 1068us/op-cpu
> > 
> > 
> > 
> > with a single remote client (similar Dell System) using NFS
> > 
> > statfile1 90ops/s 0.0mb/s 27.6ms/op 145us/op-cpu
> > deletefile1 90ops/s 0.0mb/s 64.5ms/op 401us/op-cpu
> > closefile3 90ops/s 0.0mb/s 25.8ms/op 40us/op-cpu
> > readfile1 90ops/s 11.4mb/s 3.1ms/op 363us/op-cpu
> > openfile2 90ops/s 0.0mb/s 66.0ms/op 263us/op-cpu
> > closefile2 90ops/s 0.0mb/s 22.6ms/op 124us/op-cpu
> > appendfilerand1 90ops/s 0.7mb/s 0.5ms/op 101us/op-cpu
> > openfile1 90ops/s 0.0mb/s 72.6ms/op 269us/op-cpu
> > closefile1 90ops/s 0.0mb/s 43.6ms/op 189us/op-cpu
> > wrtfile1 90ops/s 11.2mb/s 0.2ms/op 211us/op-cpu
> > createfile1 90ops/s 0.0mb/s 226.5ms/op 709us/op-cpu
> > 
> > 
> > 
> > the same remote client with zpool sync disabled on the server
> > 
> > statfile1 479ops/s 0.0mb/s 6.2ms/op 130us/op-cpu
> > deletefile1 479ops/s 0.0mb/s 13.0ms/op 351us/op-cpu
> > closefile3 480ops/s 0.0mb/s 3.0ms/op 37us/op-cpu
> > readfile1 480ops/s 62.7mb/s 0.8ms/op 174us/op-cpu
> > openfile2 480ops/s 0.0mb/s 14.1ms/op 235us/op-cpu
> > closefile2 480ops/s 0.0mb/s 6.0ms/op 123us/op-cpu
> > appendfilerand1 480ops/s 3.7mb/s 0.2ms/op 53us/op-cpu
> > openfile1 480ops/s 0.0mb/s 13.7ms/op 235us/op-cpu
> > closefile1 480ops/s 0.0mb/s 11.1ms/op 190us/op-cpu
> > wrtfile1 480ops/s 60.3mb/s 0.2ms/op 233us/op-cpu
> > createfile1 480ops/s 0.0mb/s 35.6ms/op 683us/op-cpu
> > 
> > 
> > Disabling ZIL is no option but I expected a much better performance
> > especially the ZEUS RAM only gets us a speed-up of about 1.8x
> > 
> > Is this test realistic for a typical fileserver scenario or does it require 
> > many
> > more clients to push the limits?
> > 
> > Thanks
> > Thomas
> > ___
> > zfs-discuss mailing list
> > zfs-discuss@opensolaris.org
> > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-18 Thread Stu Whitefish
> From: Thomas Gouverneur 

> To: zfs-discuss@opensolaris.org
> Cc: 
> Sent: Thursday, August 18, 2011 5:11:16 PM
> Subject: Re: [zfs-discuss] Kernel panic on zpool import. 200G of data 
> inaccessible!
> 
> Have you already extracted the core file of the kernel crash ?

Nope, not a clue how to do that and I have installed Windows on this box 
instead of Solaris since I can't get my data back from ZFS.
I have my two drives the pool is on disconnected so if this ever gets resolved 
I can reinstall Solaris and start learning again.

> (and btw activated dump device for such dumping happen at next reboot...)

This was a development box for me to see how I get along with Solaris. I'm 
afraid I don't have any experience in Solaris to understand your question.

> Have you also tried applying the latest kernel/zfs patches and try importing 
> the pool afterwards ?

Wish I had them and knew what to do with them if I had them. Somebody on OTN 
noted this is supposed to be fixed by 142910 but
I didn't hear back yet whether it fixes an pool ZFS won't import, or it only 
stops it from happening in the first place. Don't have a service
contract as I say this box was my first try with Solaris and it is a homebrew 
system not on Oracle's support list.

I am sure if there is a patch for this or a way to get my 200G of data back 
some kind soul at Oracle will certainly help me since I lost
my data and getting it back isn't a matter of convenience. What an opportunity 
to generate some old fashioned goodwill!  :-)

Jim

> 
> 
> Thomas
> 
> On 08/18/2011 06:40 PM, Stu Whitefish wrote:
>>  Hi Thomas,
>> 
>>  Thanks for that link. That's very similar but not identical. 
> There's a different line number in zfs_ioctl.c, mine and Preston's fail 
> on line 1815. It could be because of a difference in levels in that module of 
> course, but the traceback is not identical either. Ours show brand_sysenter 
> and 
> the one you linked to shows brand_sys_syscall. I don't know what all that 
> means but it is different. Anyway at least two of us have identical failures.
>> 
>>  I was not using crypto, just a plain jane mirror on 2 drives. Possibly I 
> had compression on a few file systems but everything else was allowed to 
> default.
>> 
>>  Here are our screenshots in case anybody doesn't want to go through the 
> thread.
>> 
>> 
>>  http://imageshack.us/photo/my-images/13/zfsimportfail.jpg/
>> 
>>  http://prestonconnors.com/zvol_get_stats.jpg
>> 
>> 
>>  I hope somebody can help with this. It's not a good feeling having so 
> much data gone.
>> 
>>  Thanks for your help. Oracle, are you listening?
>> 
>>  Jim
>> 
>> 
>> 
>>  - Original Message -
>>     
>>>  From: Thomas Gouverneur
>>>  To: zfs-discuss@opensolaris.org
>>>  Cc: Stu Whitefish
>>>  Sent: Thursday, August 18, 2011 1:57:29 PM
>>>  Subject: Re: [zfs-discuss] Kernel panic on zpool import. 200G of data 
> inaccessible!
>>> 
>>>  You're probably hitting bug 7056738 ->
>>>  http://wesunsolve.net/bugid/id/7056738
>>>  Looks like it's not fixed yet @ oracle anyway...
>>> 
>>>  Were you using crypto on your datasets ?
>>> 
>>> 
>>>  Regards,
>>> 
>>>  Thomas
>>>       
>>  ___
>>  zfs-discuss mailing list
>>  zfs-discuss@opensolaris.org
>>  http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>     
> 
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-18 Thread Thomas Gouverneur

Have you already extracted the core file of the kernel crash ?
(and btw activated dump device for such dumping happen at next reboot...)

Have you also tried applying the latest kernel/zfs patches and try 
importing the pool afterwards ?



Thomas

On 08/18/2011 06:40 PM, Stu Whitefish wrote:

Hi Thomas,

Thanks for that link. That's very similar but not identical. There's a 
different line number in zfs_ioctl.c, mine and Preston's fail on line 1815. It 
could be because of a difference in levels in that module of course, but the 
traceback is not identical either. Ours show brand_sysenter and the one you 
linked to shows brand_sys_syscall. I don't know what all that means but it is 
different. Anyway at least two of us have identical failures.

I was not using crypto, just a plain jane mirror on 2 drives. Possibly I had 
compression on a few file systems but everything else was allowed to default.

Here are our screenshots in case anybody doesn't want to go through the thread.


http://imageshack.us/photo/my-images/13/zfsimportfail.jpg/

http://prestonconnors.com/zvol_get_stats.jpg


I hope somebody can help with this. It's not a good feeling having so much data 
gone.

Thanks for your help. Oracle, are you listening?

Jim



- Original Message -
   

From: Thomas Gouverneur
To: zfs-discuss@opensolaris.org
Cc: Stu Whitefish
Sent: Thursday, August 18, 2011 1:57:29 PM
Subject: Re: [zfs-discuss] Kernel panic on zpool import. 200G of data 
inaccessible!

You're probably hitting bug 7056738 ->
http://wesunsolve.net/bugid/id/7056738
Looks like it's not fixed yet @ oracle anyway...

Were you using crypto on your datasets ?


Regards,

Thomas
 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-18 Thread Stu Whitefish
Hi Thomas,

Thanks for that link. That's very similar but not identical. There's a 
different line number in zfs_ioctl.c, mine and Preston's fail on line 1815. It 
could be because of a difference in levels in that module of course, but the 
traceback is not identical either. Ours show brand_sysenter and the one you 
linked to shows brand_sys_syscall. I don't know what all that means but it is 
different. Anyway at least two of us have identical failures.

I was not using crypto, just a plain jane mirror on 2 drives. Possibly I had 
compression on a few file systems but everything else was allowed to default.

Here are our screenshots in case anybody doesn't want to go through the thread.


http://imageshack.us/photo/my-images/13/zfsimportfail.jpg/

http://prestonconnors.com/zvol_get_stats.jpg


I hope somebody can help with this. It's not a good feeling having so much data 
gone.

Thanks for your help. Oracle, are you listening?

Jim



- Original Message -
> From: Thomas Gouverneur 
> To: zfs-discuss@opensolaris.org
> Cc: Stu Whitefish 
> Sent: Thursday, August 18, 2011 1:57:29 PM
> Subject: Re: [zfs-discuss] Kernel panic on zpool import. 200G of data 
> inaccessible!
> 
> You're probably hitting bug 7056738 -> 
> http://wesunsolve.net/bugid/id/7056738
> Looks like it's not fixed yet @ oracle anyway...
> 
> Were you using crypto on your datasets ?
> 
> 
> Regards,
> 
> Thomas
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS performance question over NFS

2011-08-18 Thread Tim Cook
What are the specs on the client?
On Aug 18, 2011 10:28 AM, "Thomas Nau"  wrote:
> Dear all.
> We finally got all the parts for our new fileserver following several
> recommendations we got over this list. We use
>
> Dell R715, 96GB RAM, dual 8-core Opterons
> 1 10GE Intel dual-port NIC
> 2 LSI 9205-8e SAS controllers
> 2 DataON DNS-1600 JBOD chassis
> 46 Seagate constellation SAS drives
> 2 STEC ZEUS RAM
>
>
> The base zpool config utilizes 42 drives plus the STECs as mirrored
> log devices. The Seagates are setup as a stripe of 7 times 6-drive-RAIDZ2
> junks plus as said a dedicated ZIL made of the mirrored STECs.
>
> As a quick'n dirty check we ran "filebench" with the "fileserver"
> workload. Running locally we get
>
> statfile1 5476ops/s 0.0mb/s 0.6ms/op 179us/op-cpu
> deletefile1 5476ops/s 0.0mb/s 1.0ms/op 454us/op-cpu
> closefile3 5476ops/s 0.0mb/s 0.0ms/op 5us/op-cpu
> readfile1 5476ops/s 729.5mb/s 0.2ms/op 128us/op-cpu
> openfile2 5477ops/s 0.0mb/s 0.8ms/op 204us/op-cpu
> closefile2 5477ops/s 0.0mb/s 0.0ms/op 5us/op-cpu
> appendfilerand1 5477ops/s 42.8mb/s 0.3ms/op 184us/op-cpu
> openfile1 5477ops/s 0.0mb/s 0.9ms/op 209us/op-cpu
> closefile1 5477ops/s 0.0mb/s 0.0ms/op 6us/op-cpu
> wrtfile1 5477ops/s 688.4mb/s 0.4ms/op 220us/op-cpu
> createfile1 5477ops/s 0.0mb/s 2.7ms/op 1068us/op-cpu
>
>
>
> with a single remote client (similar Dell System) using NFS
>
> statfile1 90ops/s 0.0mb/s 27.6ms/op 145us/op-cpu
> deletefile1 90ops/s 0.0mb/s 64.5ms/op 401us/op-cpu
> closefile3 90ops/s 0.0mb/s 25.8ms/op 40us/op-cpu
> readfile1 90ops/s 11.4mb/s 3.1ms/op 363us/op-cpu
> openfile2 90ops/s 0.0mb/s 66.0ms/op 263us/op-cpu
> closefile2 90ops/s 0.0mb/s 22.6ms/op 124us/op-cpu
> appendfilerand1 90ops/s 0.7mb/s 0.5ms/op 101us/op-cpu
> openfile1 90ops/s 0.0mb/s 72.6ms/op 269us/op-cpu
> closefile1 90ops/s 0.0mb/s 43.6ms/op 189us/op-cpu
> wrtfile1 90ops/s 11.2mb/s 0.2ms/op 211us/op-cpu
> createfile1 90ops/s 0.0mb/s 226.5ms/op 709us/op-cpu
>
>
>
> the same remote client with zpool sync disabled on the server
>
> statfile1 479ops/s 0.0mb/s 6.2ms/op 130us/op-cpu
> deletefile1 479ops/s 0.0mb/s 13.0ms/op 351us/op-cpu
> closefile3 480ops/s 0.0mb/s 3.0ms/op 37us/op-cpu
> readfile1 480ops/s 62.7mb/s 0.8ms/op 174us/op-cpu
> openfile2 480ops/s 0.0mb/s 14.1ms/op 235us/op-cpu
> closefile2 480ops/s 0.0mb/s 6.0ms/op 123us/op-cpu
> appendfilerand1 480ops/s 3.7mb/s 0.2ms/op 53us/op-cpu
> openfile1 480ops/s 0.0mb/s 13.7ms/op 235us/op-cpu
> closefile1 480ops/s 0.0mb/s 11.1ms/op 190us/op-cpu
> wrtfile1 480ops/s 60.3mb/s 0.2ms/op 233us/op-cpu
> createfile1 480ops/s 0.0mb/s 35.6ms/op 683us/op-cpu
>
>
> Disabling ZIL is no option but I expected a much better performance
> especially the ZEUS RAM only gets us a speed-up of about 1.8x
>
> Is this test realistic for a typical fileserver scenario or does it
require many
> more clients to push the limits?
>
> Thanks
> Thomas
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS performance question over NFS

2011-08-18 Thread Thomas Nau
Dear all.
We finally got all the parts for our new fileserver following several
recommendations we got over this list. We use

Dell R715, 96GB RAM, dual 8-core Opterons
1 10GE Intel dual-port NIC
2 LSI 9205-8e SAS controllers
2 DataON DNS-1600 JBOD chassis
46 Seagate constellation SAS drives
2 STEC ZEUS RAM


The base zpool config utilizes 42 drives plus the STECs as mirrored
log devices. The Seagates are setup as a stripe of 7 times 6-drive-RAIDZ2
junks plus as said a dedicated ZIL made of the mirrored STECs.

As a quick'n dirty check we ran "filebench" with the "fileserver"
workload. Running locally we get

statfile15476ops/s   0.0mb/s  0.6ms/op  179us/op-cpu
deletefile1  5476ops/s   0.0mb/s  1.0ms/op  454us/op-cpu
closefile3   5476ops/s   0.0mb/s  0.0ms/op5us/op-cpu
readfile15476ops/s 729.5mb/s  0.2ms/op  128us/op-cpu
openfile25477ops/s   0.0mb/s  0.8ms/op  204us/op-cpu
closefile2   5477ops/s   0.0mb/s  0.0ms/op5us/op-cpu
appendfilerand1  5477ops/s  42.8mb/s  0.3ms/op  184us/op-cpu
openfile15477ops/s   0.0mb/s  0.9ms/op  209us/op-cpu
closefile1   5477ops/s   0.0mb/s  0.0ms/op6us/op-cpu
wrtfile1 5477ops/s 688.4mb/s  0.4ms/op  220us/op-cpu
createfile1  5477ops/s   0.0mb/s  2.7ms/op 1068us/op-cpu



with a single remote client (similar Dell System) using NFS

statfile1  90ops/s   0.0mb/s 27.6ms/op  145us/op-cpu
deletefile190ops/s   0.0mb/s 64.5ms/op  401us/op-cpu
closefile3 90ops/s   0.0mb/s 25.8ms/op   40us/op-cpu
readfile1  90ops/s  11.4mb/s  3.1ms/op  363us/op-cpu
openfile2  90ops/s   0.0mb/s 66.0ms/op  263us/op-cpu
closefile2 90ops/s   0.0mb/s 22.6ms/op  124us/op-cpu
appendfilerand190ops/s   0.7mb/s  0.5ms/op  101us/op-cpu
openfile1  90ops/s   0.0mb/s 72.6ms/op  269us/op-cpu
closefile1 90ops/s   0.0mb/s 43.6ms/op  189us/op-cpu
wrtfile1   90ops/s  11.2mb/s  0.2ms/op  211us/op-cpu
createfile190ops/s   0.0mb/s226.5ms/op  709us/op-cpu



the same remote client with zpool sync disabled on the server

statfile1 479ops/s   0.0mb/s  6.2ms/op  130us/op-cpu
deletefile1   479ops/s   0.0mb/s 13.0ms/op  351us/op-cpu
closefile3480ops/s   0.0mb/s  3.0ms/op   37us/op-cpu
readfile1 480ops/s  62.7mb/s  0.8ms/op  174us/op-cpu
openfile2 480ops/s   0.0mb/s 14.1ms/op  235us/op-cpu
closefile2480ops/s   0.0mb/s  6.0ms/op  123us/op-cpu
appendfilerand1   480ops/s   3.7mb/s  0.2ms/op   53us/op-cpu
openfile1 480ops/s   0.0mb/s 13.7ms/op  235us/op-cpu
closefile1480ops/s   0.0mb/s 11.1ms/op  190us/op-cpu
wrtfile1  480ops/s  60.3mb/s  0.2ms/op  233us/op-cpu
createfile1   480ops/s   0.0mb/s 35.6ms/op  683us/op-cpu


Disabling ZIL is no option but I expected a much better performance
especially the ZEUS RAM only gets us a speed-up of about 1.8x

Is this test realistic for a typical fileserver scenario or does it require many
more clients to push the limits?

Thanks
Thomas
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Kernel panic on zpool import. 200G of data inaccessible!

2011-08-18 Thread Thomas Gouverneur
You're probably hitting bug 7056738 -> http://wesunsolve.net/bugid/id/7056738
Looks like it's not fixed yet @ oracle anyway...

Were you using crypto on your datasets ?


Regards,

Thomas

On Tue, 16 Aug 2011 09:33:34 -0700 (PDT)
Stu Whitefish  wrote:

> - Original Message -
> 
> > From: Alexander Lesle 
> > To: zfs-discuss@opensolaris.org
> > Cc: 
> > Sent: Monday, August 15, 2011 8:37:42 PM
> > Subject: Re: [zfs-discuss] Kernel panic on zpool import. 200G of data 
> > inaccessible!
> > 
> > Hello Stu Whitefish and List,
> > 
> > On August, 15 2011, 21:17  wrote in [1]:
> > 
> >>>  7. cannot import old rpool (c0t2d0s0 c0t3d0s0), any attempt causes a
> >>>  kernel panic, even when booted from different OS versions
> > 
> >>  Right. I have tried OpenIndiana 151 and Solaris 11 Express (latest
> >>  from Oracle) several times each as well as 2 new installs of Update 8.
> > 
> > When I understand you right is your primary interest to recover your
> > data on tank pool.
> > 
> > Have you check the way to boot from a Live-DVD, mount your "safe 
> > place"
> > and copy the data on a other machine?
> 
> Hi Alexander,
> 
> Yes of course...the problem is no version of Solaris can import the pool. 
> Please refer to the first message in the thread.
> 
> Thanks,
> 
> Jim
> 
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


-- 
Gouverneur Thomas 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs-discuss Digest, Vol 70, Issue 37

2011-08-18 Thread Gowrisankr Periyasamy
Please check whether you have latest ' MPT ' patch installed on your server
? If not , please install MPT patch . It will fix the issue .

Regards,
Gowrisankar .

On Thu, Aug 18, 2011 at 5:30 PM, wrote:

> Send zfs-discuss mailing list submissions to
>zfs-discuss@opensolaris.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
>http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> or, via email, send a message with subject or body 'help' to
>zfs-discuss-requ...@opensolaris.org
>
> You can reach the person managing the list at
>zfs-discuss-ow...@opensolaris.org
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of zfs-discuss digest..."
>
>
> Today's Topics:
>
>   1. Re: solaris 10u8 hangs with message Disconnected  command
>  timeout for Target 0 (Richard Elling)
>
>
> --
>
> Message: 1
> Date: Wed, 17 Aug 2011 17:20:23 -0700
> From: Richard Elling 
> To: Ding Honghui 
> Cc: zfs-discuss@opensolaris.org
> Subject: Re: [zfs-discuss] solaris 10u8 hangs with message
>Disconnectedcommand timeout for Target 0
> Message-ID: 
> Content-Type: text/plain; charset=iso-8859-1
>
> On Aug 15, 2011, at 11:17 PM, Ding Honghui wrote:
> > My solaris storage hangs. I login to the console and there is messages[1]
> display on the console.
> > I can't login into the console and seems the IO is totally blocked.
> >
> > The system is solaris 10u8 on Dell R710 with disk array Dell MD3000. 2
> HBA cable connect the server and MD3000.
> > The symptom is random.
>
> This symptom is consistent with a broken SATA disk behind a SAS expander.
>
> Unfortunately, the mpt driver is closed source, so we can only infer what
> the
> code does by using the open source mpt_sas driver as (hopefully) a
> derivative.
>
> >
> > It is very appreciated if any one can help me out.
> >
> > Regards,
> > Ding
> >
> > [1]
> > Aug 16 13:14:16 nas-hz-02 scsi: WARNING: /pci@0,0/pci8086,3410@9
> /pci8086,32c@0/pci1028,1f04@8 (mpt1):
> > Aug 16 13:14:16 nas-hz-02   Disconnected command timeout for Target 0
>
> A command did not complete and the mpt driver reset the target.
> If that target is an expander, then everything behind the expander can
> reset, resulting in the aborts of any in-flight commands, as follows...
>
> > Aug 16 13:14:16 nas-hz-02 scsi: WARNING:
> /scsi_vhci/disk@g60026b900053aa1802a44b8f0ded (sd47):
> > Aug 16 13:14:16 nas-hz-02   Error for Command: write(10)
>   Error Level: Retryable
> > Aug 16 13:14:16 nas-hz-02 scsi: Requested Block: 1380679073
>  Error Block: 1380679073
> > Aug 16 13:14:16 nas-hz-02 scsi: Vendor: DELL
>   Serial Number:
> > Aug 16 13:14:16 nas-hz-02 scsi: Sense Key: Unit Attention
> > Aug 16 13:14:16 nas-hz-02 scsi: ASC: 0x29 (device internal
> reset), ASCQ: 0x4, FRU: 0x0
> > Aug 16 13:14:16 nas-hz-02 scsi: WARNING:
> /scsi_vhci/disk@g60026b900053aa18029e4b8f0d61 (sd41):
> > Aug 16 13:14:16 nas-hz-02   Error for Command: write(10)
>   Error Level: Retryable
> > Aug 16 13:14:16 nas-hz-02 scsi: Requested Block: 1380679072
>  Error Block: 1380679072
> > Aug 16 13:14:16 nas-hz-02 scsi: Vendor: DELL
>   Serial Number:
> > Aug 16 13:14:16 nas-hz-02 scsi: Sense Key: Unit Attention
> > Aug 16 13:14:16 nas-hz-02 scsi: ASC: 0x29 (device internal
> reset), ASCQ: 0x4, FRU: 0x0
> > Aug 16 13:14:16 nas-hz-02 scsi: WARNING:
> /scsi_vhci/disk@g60026b900053aa1802a24b8f0dc5 (sd45):
> > Aug 16 13:14:16 nas-hz-02   Error for Command: write(10)
>   Error Level: Retryable
> > Aug 16 13:14:16 nas-hz-02 scsi: Requested Block: 1380679073
>  Error Block: 1380679073
> > Aug 16 13:14:16 nas-hz-02 scsi: Vendor: DELL
>   Serial Number:
> > Aug 16 13:14:16 nas-hz-02 scsi: Sense Key: Unit Attention
> > Aug 16 13:14:16 nas-hz-02 scsi: ASC: 0x29 (device internal
> reset), ASCQ: 0x4, FRU: 0x0
> > Aug 16 13:14:16 nas-hz-02 scsi: WARNING:
> /scsi_vhci/disk@g60026b900053aa18029c4b8f0d35 (sd39):
> > Aug 16 13:14:16 nas-hz-02   Error for Command: write(10)
>   Error Level: Retryable
> > Aug 16 13:14:16 nas-hz-02 scsi: Requested Block: 1380679072
>  Error Block: 1380679072
> > Aug 16 13:14:16 nas-hz-02 scsi: Vendor: DELL
>   Serial Number:
> > Aug 16 13:14:16 nas-hz-02 scsi: Sense Key: Unit Attention
> > Aug 16 13:14:16 nas-hz-02 scsi: ASC: 0x29 (device internal
> reset), ASCQ: 0x4, FRU: 0x0
> > Aug 16 13:14:16 nas-hz-02 scsi: WARNING:
> /scsi_vhci/disk@g60026b900053aa1802984b8f0cd2 (sd35):
> > Aug 16 13:14:16 nas-hz-02   Error for Command: write(10)
>   Error Level: Retryable
> > Aug 16 13:14:16 nas-hz-02 scsi: Requested Block: 1380679072
>  Error Block: 1380679072
> > Aug 16 13:14:16 nas-hz-02 scsi: Vendor: DELL
>   Serial