Re: [zfs-discuss] snapshot and space reservation

2010-02-23 Thread Richard Elling
On Feb 23, 2010, at 6:24 PM, v wrote:

> Hi,
> Another question regarding snapshot.
> If there is no space in zfs pool, will a write to zfs fail ?

Of course.

> Is there a way to reserve space in zfs pool to be used by snapshot or clone?

Snapshots are read only, so no need for a reservation.
Clones can be given reservations.

NB, there are two kinds of reservations:
reservations  work on the dataset size
refreservations work on the dataset and snapshots sizes
the two differ in the accounting of snapshots.

For more information, see the "Setting ZFS Quotas and Reservations" in the
ZFS Administration Guide.
 -- richard

ZFS storage and performance consulting at http://www.RichardElling.com
ZFS training on deduplication, NexentaStor, and NAS performance
http://nexenta-atlanta.eventbrite.com (March 16-18, 2010)




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Import zpool from FreeBSD in OpenSolaris

2010-02-23 Thread Ethan
On Tue, Feb 23, 2010 at 21:22, Bob Friesenhahn  wrote:

> On Wed, 24 Feb 2010, Fajar A. Nugraha wrote:
>
>  On Wed, Feb 24, 2010 at 9:11 AM, patrik  wrote:
>>
>>> This is "zpool import" from my machine with OpenSolaris 2009.06 (all
>>> zpool's are fine in FreeBSD). Notice that the zpool named "temp" can be
>>> imported. Why not "secure" then? Is it because it is raidz1?
>>>
>>
>>  status: One or more devices contains corrupted data.
>>>
>>
>> c8t3d0s8  UNAVAIL  corrupted data
>>>c8t4d0s8  UNAVAIL  corrupted data
>>>
>>
>> I'd suggest you try reimporting them in FreeBSD. It's possible that
>> those disks are really corrupted.
>> Another option is to try latest opensolaris livecd from genunix.org,
>> and try to import it there.
>>
>
> Just a couple of days ago there was discussion of importing disks from
> Linux FUSE zfs.  The import was successful.  The same methods used
> (directory containing symbolic links to desired device partitions) might be
> used.
>
> Bob
> --
>

Yes, I ran into this very recently, moving from zfs-fuse on linux to
OpenSolaris. My import looked almost exactly like yours. I did something
along the lines of:

mkdir dskp0s # create a temp directory to point to the p0 partitions of the
relevant disks
cd dskp0s
ln -s /dev/dsk/c8t1d0p0
ln -s /dev/dsk/c8t2d0p0
ln -s /dev/dsk/c8t3d0p0
ln -s /dev/dsk/c8t4d0p0
zpool import -d . secure

(substituting in info for your pool) and it imported, no problem.

-Ethan
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Import zpool from FreeBSD in OpenSolaris

2010-02-23 Thread Marc Nicholas
send and receive?!

-marc

On Tue, Feb 23, 2010 at 9:25 PM, Thomas Burgess  wrote:

> When i needed to do this, the only way i could get it to work was to do
> this:
>
> Take some disks, use a Opensolaris Live CD and label them EFI
> Create a ZPOOL in FreeBSD with these disks
> copy my data from freebsd to the new zpool
> export the pool
> import the pool
>
>
>
>
> On Tue, Feb 23, 2010 at 9:11 PM, patrik  wrote:
>
>> I want to import my zpool's from FreeBSD 8.0 in OpenSolaris 2009.06.
>>
>> After reading the few posts (links below) I was able to find on the
>> subject, it seems like it there is a differences between FreeBSD and
>> Solaris. FreeBSD operates on directly on the disk and Solaris creates a
>> partion and uses that... is that right? Is it impossible for OpenSolaris to
>> use zpool's from FreeBSD?
>>
>> * http://opensolaris.org/jive/thread.jspa?messageID=445766
>> * http://opensolaris.org/jive/thread.jspa?messageID=450755&;
>> * http://mail.opensolaris.org/pipermail/ug-nzosug/2009-June/27.html
>>
>> This is "zpool import" from my machine with OpenSolaris 2009.06 (all
>> zpool's are fine in FreeBSD). Notice that the zpool named "temp" can be
>> imported. Why not "secure" then? Is it because it is raidz1?
>>
>>  pool: secure
>>id: 15384175022505637073
>>  state: UNAVAIL
>> status: One or more devices contains corrupted data.
>> action: The pool cannot be imported due to damaged devices or data.
>>   see: http://www.sun.com/msg/ZFS-8000-5E
>> config:
>>
>>secureUNAVAIL  insufficient replicas
>>  raidz1  UNAVAIL  insufficient replicas
>>c8t1d0p0  ONLINE
>>c8t2d0s2  ONLINE
>>c8t3d0s8  UNAVAIL  corrupted data
>>c8t4d0s8  UNAVAIL  corrupted data
>>
>>
>>  pool: temp
>>id: 10889808377251842082
>>  state: ONLINE
>> status: The pool is formatted using an older on-disk version.
>> action: The pool can be imported using its name or numeric identifier,
>> though
>>some features will not be available without an explicit 'zpool
>> upgrade'.
>> config:
>>
>>tempONLINE
>>  c8t0d0p0  ONLINE
>> --
>> This message posted from opensolaris.org
>> ___
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>
>
>
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Import zpool from FreeBSD in OpenSolaris

2010-02-23 Thread Thomas Burgess
When i needed to do this, the only way i could get it to work was to do
this:

Take some disks, use a Opensolaris Live CD and label them EFI
Create a ZPOOL in FreeBSD with these disks
copy my data from freebsd to the new zpool
export the pool
import the pool



On Tue, Feb 23, 2010 at 9:11 PM, patrik  wrote:

> I want to import my zpool's from FreeBSD 8.0 in OpenSolaris 2009.06.
>
> After reading the few posts (links below) I was able to find on the
> subject, it seems like it there is a differences between FreeBSD and
> Solaris. FreeBSD operates on directly on the disk and Solaris creates a
> partion and uses that... is that right? Is it impossible for OpenSolaris to
> use zpool's from FreeBSD?
>
> * http://opensolaris.org/jive/thread.jspa?messageID=445766
> * http://opensolaris.org/jive/thread.jspa?messageID=450755&;
> * http://mail.opensolaris.org/pipermail/ug-nzosug/2009-June/27.html
>
> This is "zpool import" from my machine with OpenSolaris 2009.06 (all
> zpool's are fine in FreeBSD). Notice that the zpool named "temp" can be
> imported. Why not "secure" then? Is it because it is raidz1?
>
>  pool: secure
>id: 15384175022505637073
>  state: UNAVAIL
> status: One or more devices contains corrupted data.
> action: The pool cannot be imported due to damaged devices or data.
>   see: http://www.sun.com/msg/ZFS-8000-5E
> config:
>
>secureUNAVAIL  insufficient replicas
>  raidz1  UNAVAIL  insufficient replicas
>c8t1d0p0  ONLINE
>c8t2d0s2  ONLINE
>c8t3d0s8  UNAVAIL  corrupted data
>c8t4d0s8  UNAVAIL  corrupted data
>
>
>  pool: temp
>id: 10889808377251842082
>  state: ONLINE
> status: The pool is formatted using an older on-disk version.
> action: The pool can be imported using its name or numeric identifier,
> though
>some features will not be available without an explicit 'zpool
> upgrade'.
> config:
>
>tempONLINE
>  c8t0d0p0  ONLINE
> --
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] snapshot and space reservation

2010-02-23 Thread v
Hi,
Another question regarding snapshot.
If there is no space in zfs pool, will a write to zfs fail ? Is there a way to 
reserve space in zfs pool to be used by snapshot or clone?

Regards
Victor
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Import zpool from FreeBSD in OpenSolaris

2010-02-23 Thread Bob Friesenhahn

On Wed, 24 Feb 2010, Fajar A. Nugraha wrote:


On Wed, Feb 24, 2010 at 9:11 AM, patrik  wrote:

This is "zpool import" from my machine with OpenSolaris 2009.06 (all zpool's are fine in FreeBSD). 
Notice that the zpool named "temp" can be imported. Why not "secure" then? Is it because 
it is raidz1?



status: One or more devices contains corrupted data.



           c8t3d0s8  UNAVAIL  corrupted data
           c8t4d0s8  UNAVAIL  corrupted data


I'd suggest you try reimporting them in FreeBSD. It's possible that
those disks are really corrupted.
Another option is to try latest opensolaris livecd from genunix.org,
and try to import it there.


Just a couple of days ago there was discussion of importing disks from 
Linux FUSE zfs.  The import was successful.  The same methods used 
(directory containing symbolic links to desired device partitions) 
might be used.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Import zpool from FreeBSD in OpenSolaris

2010-02-23 Thread patrik
On 2010-02-24 03.16, Fajar A. Nugraha wrote:
> On Wed, Feb 24, 2010 at 9:11 AM, patrik  wrote:
>> This is "zpool import" from my machine with OpenSolaris 2009.06 (all zpool's 
>> are fine in FreeBSD). Notice that the zpool named "temp" can be imported. 
>> Why not "secure" then? Is it because it is raidz1?
> 
>> status: One or more devices contains corrupted data.
> 
>>c8t3d0s8  UNAVAIL  corrupted data
>>c8t4d0s8  UNAVAIL  corrupted data
> 
> I'd suggest you try reimporting them in FreeBSD. It's possible that
> those disks are really corrupted.

I have. They are not corrupted. In fact, I'm using them in FreeBSD now. :-)

> Another option is to try latest opensolaris livecd from genunix.org,
> and try to import it there.

I'm gonna give it a try. Thanks.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs sequential read performance

2010-02-23 Thread v
Hi,
Thanks for the reply.
So the problem of sequential read after random write problem exist in zfs.
I wonder if it is a real problem, ie, for example cause longer backup time, 
will it be addressed in future?

So I should ask anther question: is zfs suitable for an environment that has 
lots of data changes? I think for random I/O, there will be no such performance 
penalty, but if you backup a zfs dataset, must the backup utility sequentially 
read blocks of a dataset? Will zfs dataset suitable for database temporary 
tablespace or online redo logs?

Will a defrag utility be implemented?

Regards
Victor
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Import zpool from FreeBSD in OpenSolaris

2010-02-23 Thread Fajar A. Nugraha
On Wed, Feb 24, 2010 at 9:11 AM, patrik  wrote:
> This is "zpool import" from my machine with OpenSolaris 2009.06 (all zpool's 
> are fine in FreeBSD). Notice that the zpool named "temp" can be imported. Why 
> not "secure" then? Is it because it is raidz1?

> status: One or more devices contains corrupted data.

>            c8t3d0s8  UNAVAIL  corrupted data
>            c8t4d0s8  UNAVAIL  corrupted data

I'd suggest you try reimporting them in FreeBSD. It's possible that
those disks are really corrupted.
Another option is to try latest opensolaris livecd from genunix.org,
and try to import it there.

-- 
Fajar
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Import zpool from FreeBSD in OpenSolaris

2010-02-23 Thread patrik
I want to import my zpool's from FreeBSD 8.0 in OpenSolaris 2009.06.

After reading the few posts (links below) I was able to find on the subject, it 
seems like it there is a differences between FreeBSD and Solaris. FreeBSD 
operates on directly on the disk and Solaris creates a partion and uses that... 
is that right? Is it impossible for OpenSolaris to use zpool's from FreeBSD?

* http://opensolaris.org/jive/thread.jspa?messageID=445766
* http://opensolaris.org/jive/thread.jspa?messageID=450755&;
* http://mail.opensolaris.org/pipermail/ug-nzosug/2009-June/27.html

This is "zpool import" from my machine with OpenSolaris 2009.06 (all zpool's 
are fine in FreeBSD). Notice that the zpool named "temp" can be imported. Why 
not "secure" then? Is it because it is raidz1?

  pool: secure
id: 15384175022505637073
 state: UNAVAIL
status: One or more devices contains corrupted data.
action: The pool cannot be imported due to damaged devices or data.
   see: http://www.sun.com/msg/ZFS-8000-5E
config:

secureUNAVAIL  insufficient replicas
  raidz1  UNAVAIL  insufficient replicas
c8t1d0p0  ONLINE
c8t2d0s2  ONLINE
c8t3d0s8  UNAVAIL  corrupted data
c8t4d0s8  UNAVAIL  corrupted data


  pool: temp
id: 10889808377251842082
 state: ONLINE
status: The pool is formatted using an older on-disk version.
action: The pool can be imported using its name or numeric identifier, though
some features will not be available without an explicit 'zpool upgrade'.
config:

tempONLINE
  c8t0d0p0  ONLINE
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SSDs with a SCSI SCA interface? (WHEE! MAYBE!)

2010-02-23 Thread Erik Trimble

Erik Trimble wrote:

Miles Nordin wrote:

"et" == Erik Trimble  writes:



et> I'd kill for a parallel SCSI -> SATA adapter thingy

http://www.google.com/products?q=scsi+to+sata+converter
  



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
  



OK, I think I might have found what I'm looking for:

Acard *ARS-2320H*

http://www.acard.com/english/fb01-product.jsp?idno_no=239&prod_no=ARS-2320&type1_title=SCSIDE%20II%20Bridge&type1_idno=11 




Best price I can find is $125 or so, which, while not cheap, is OK.  I 
might have to get one and try it.




OK, wrong link. That one is for the 68-pin version. Here's the SCA 
hot-swap box:


http://www.acard.com/english/fb01-product.jsp?idno_no=241&prod_no=ARS-2320H&type1_title=SCSIDE%20II%20Bridge&type1_idno=11




--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SSDs with a SCSI SCA interface? (WHEE! MAYBE!)

2010-02-23 Thread Erik Trimble

Miles Nordin wrote:

"et" == Erik Trimble  writes:



et> I'd kill for a parallel SCSI -> SATA adapter thingy

http://www.google.com/products?q=scsi+to+sata+converter
  



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
  



OK, I think I might have found what I'm looking for:

Acard *ARS-2320H*

http://www.acard.com/english/fb01-product.jsp?idno_no=239&prod_no=ARS-2320&type1_title=SCSIDE%20II%20Bridge&type1_idno=11


Best price I can find is $125 or so, which, while not cheap, is OK.  I 
might have to get one and try it.





--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SSDs with a SCSI SCA interface?

2010-02-23 Thread Erik Trimble

Miles Nordin wrote:

"et" == Erik Trimble  writes:



et> I'd kill for a parallel SCSI -> SATA adapter thingy

http://www.google.com/products?q=scsi+to+sata+converter
  


Not quite. All of those convert from SATA-150 to 68-pin LVD Ultra160 (at 
best, some do just Ultra2).



I need something that will fit in a hot-swap bay, so it has to have the 
SCA connector at the standard location.  What I'm really gonna need is a 
2.5" SATA drive -> SCA Ultra320 adapter. Not quite around yet, I see.


--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Adding a zfs mirror drive to rpool - new drive formats to one cylinder less

2010-02-23 Thread tomwaters
Thanks for that.

It seems strange though that the two disks, which are from the same 
manufacturer, same model, same firmware and similar batch/serial's behave 
differently.

I am also puzzled that the rpool disk appears to start at cylinder 0 and not 1.

I did find this quote after googling for the CR6844090 at the best practices 
page http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide

"The size of the replacements vdev, measured by usable sectors, must be the 
same or greater than the vdev being replaced. This can be confusing when whole 
disks are used because different models of disks may provide a different number 
of usable sectors. For example, if a pool was created with a "500 GByte" drive 
and you need to replace it with another "500 GByte" drive, then you may not be 
able to do so if the drives are not of the same make, model, and firmware 
revision. Consider planning ahead and reserving some space by creating a slice 
which is smaller than the whole disk instead of the whole disk. In Nevada, 
build 117, it might be possible to replace or attach a disk that is slight 
smaller than the other disks in a pool. This is CR 6844090. "
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Chris Ridd

On 23 Feb 2010, at 19:53, Bruno Sousa wrote:

> The system becames really slow during the data copy using network, but i copy 
> data between 2 pools of the box i don't notice that issue, so probably i may 
> be hitting some sort of interrupt conflit in the network cards...This system 
> is configured with alot of interfaces, being :
> 
> 4 internal broadcom gigabit
> 1 PCIe 4x, Intel Dual Pro gigabit
> 1 PCIe 4x, Intel 10gbE card
> 2 PCIe 8x Sun non-raid HBA
> 
> 
> With all of this, is there any way to check if there is indeed an interrupt 
> conflit or some other type of conflit that leads this high load? I also 
> noticed some messages about acpi..can this acpi also affect the performance 
> of the system?

To see what interrupts are being shared:

# echo "::interrupts -d" | mdb -k

Running intrstat might also be interesting.

Cheers,

Chris
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [storage-discuss] Horribly bad luck with Unified Storage 7210 - hardware or software?

2010-02-23 Thread Bruno Sousa











Hi,

Just some comments on your situation , please take a look the following
things :

* Sometimes the hw looks the same, i'm talking specifically to the
  SSD's, but they can be somehow different and that may lead to some
  problems in the future . Bottom line is , don't go too cheap for
  production :)
* If you bought 3 7210 full of disks, they put pressure in your Sun
  reseller to get some very cheap (i would say for free!) logzillas
  . Their margin is more than okay to give you at least 1 logzilla
  to your primary box .
* With ZFS always make sure you use a UPS attached to system  and
  make sure that you make a gracefull shutdown of the system in the
  event of power failure and drained batteries on the UPS. If
  possible use some sort of no-break system, like an auxiliary
  generator or so.
* If possible don't use any box whatsoever in production before 2 to
  4 weeks of testing/pre-production . It's quite often to see
  problems when we stop to use benchmarks and start to use the
  systems in production. Benchmarks i often called them "controlled
  chaos", but based on my experiences in production there's only
  "chaos", without the "controlled" part ;) . We never now when some
  application that *should* do some sequencial reads is actually
  doing random reads...
* Don't give up "easily"  on the Sun support guys . Again based on
  my experiences, quite often you need to put pressure in your Sun
  partner/reseller to get the necessary amount of attention to your
  problem..please remember that money is what drives business .
* On your XEN, do you really LVM-over-iscsi? Can you live with NFS
  storage repositorys in the XEN? In my current job, i never
  achieved better performance with iscsi than what i have with NFS,
  and NFS gives less issues..just take a look to the all snapshot
  issues in Citrix Xen forums for people using ISCSI rather than NFS
  . Also in a near-by future (i hope) pNFS will be there so,
  therefore more and more performance for the NFS stack ;)
* Still on the XEN, i found that in my current environment enabling
  jumbo frames in the xen physical servers leads to massive
  problems..so if you can take a look to the jumbo frames support
  for your specific network cards
* Getting out from the OS installed in your 7210, towards anything
  else probably means that you will lose support from Sun..so i
  don't know if that's such a good idea. Again, put pressure in the
  Sun reseller to get attention to your problem and don't take
  shortcuts and don't get mad ;) I been there, i know the feeling of
  having to put something , expensive, to work and it just doesn't
  work as it should.
* Keep this list informed, they are great minds around here, so
  maybe someone can give you some extra help


Good luck,
Bruno

On 23-2-2010 23:45, Nate Carlson wrote:
> Hey all,
>
> I've put up a blog post on issues we're having with 3 brand new Unified 
> Storage 7210 arrays, available here:
>
> http://tinyurl.com/yeo9pft
>
> I'm curious if the symptoms (spurious halt, spurious vlan ipmp group 
> dropping, spurious reboots) that we are seeing are common for the X4540 
> machine, or if you think it's likely to be a problem with the way Sun rolls 
> the Unified Storage software package.
>
> If it is a software issue, and we are unable to return these machines, I'm 
> considering running either OpenSolaris or NexentaStor on the hardware.. 
> hoping that it's a software issue and not hardware.  ;)  (Going OpenSol or 
> Nexenta would also let us add our own SSD's without worrying about Sun 
> complaining about it.. $400 for 32gb x25-e's sounds a lot better than $6k for 
> an 18gb Logzilla!)
>
> Appreciate any thoughts!
>   



smime.p7s
Description: S/MIME Cryptographic Signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] shrinking a zpool - roadmap

2010-02-23 Thread Robert Milkowski

On 23/02/2010 17:20, Richard Elling wrote:

On Feb 23, 2010, at 5:10 AM, Robert Milkowski wrote:
   

On 23/02/2010 02:52, Richard Elling wrote:
 

On Feb 22, 2010, at 6:42 PM, Charles Hedrick wrote:

   

I talked with our enterprise systems people recently. I don't believe they'd 
consider ZFS until it's more flexible. Shrink is a big one, as is removing an 
slog. We also need to be able to expand a raidz, possibly by striping it with a 
second one and then rebalancing the sizes.

 

So what file system do they use that has all of these features? :-P


   

VxVM + VxFS?
 

I did know they still cost $$$, but I didn't know they implemented a slog :-P

   


you got me! :)
I missed the reference to a slog.

--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SSDs with a SCSI SCA interface?

2010-02-23 Thread Miles Nordin
> "et" == Erik Trimble  writes:

et> I'd kill for a parallel SCSI -> SATA adapter thingy

http://www.google.com/products?q=scsi+to+sata+converter


pgpFP7Wm0x6aw.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs sequential read performance

2010-02-23 Thread Robert Milkowski

On 23/02/2010 17:18, Richard Elling wrote:

On Feb 23, 2010, at 1:00 AM, v wrote:

   

Hi experts,
I am new to zfs and ask a question regarding zfs sequential peroformance: I read some 
blogs saying that netapp's WAFL can suffer "sequential read after random 
write(SRARW)" performance penalty, since zfs is also doing no update in place, can 
zfs has such problem?
 

I know of only one study of this effect, by Allan Packer and Neel using MySQL.
http://www.youtube.com/watch?v=a31NhwzlAxs
slides at
http://blogs.sun.com/realneel/resource/MySQL_Conference_2009_ZFS_MySQL.pdf

   


from my own experience with mysql on top of zfs on solaris 10 - after a 
fresh db restore backup times X amount of time. After couple of weeks of 
the database in live the backup takes 2-3x the amount of X.


--
Robert Milkowski
http://milek.blogspot.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Help with corrupted pool

2010-02-23 Thread Ethan
On Sun, Feb 21, 2010 at 12:41, Ethan  wrote:
>
> Update: I'm stuck. Again.
>
> To answer "For curiosity's sake, what happens when you remove (rename) your
> dir with the symlinks?": it finds the devices on p0 with no problem, with
> the symlinks directory deleted.
>
> After clearing the errors and scrubbing again, no errors were encountered
> in the second scrub. Then I offlined the disk which had errors in the first
> scrub.
>
> I followed the suggestion to thoroughly test the disk (and remap any bad
> sectors), filling it with random-looking data by encrypting /dev/zero.
> Reading back and decrypting the drive, it all read back as zeros - all
> good.
>
> I then checked the SMART status of the drive, which had 0 error rates for
> everything. I ran the several-hour "extended self-test", whatever that does,
> after which I had two write errors on one drive which weren't there before.
> I believe it's the same drive that had the zfs errors, but I did the SMART
> stuff in linux, not being able to find SMART tools in solaris, and I haven't
> been able to figure out which drive is which. Is there a way to get a
> drive's serial number in solaris? I could identify it by that.
>
> I scrubbed again with the pool degraded. No errors.
>
> NAME STATE READ WRITE CKSUM
> qONLINE   0 0 0
>   raidz1 ONLINE   0 0 0
> c9t4d0p0 ONLINE   0 0 0
> c9t5d0p0 ONLINE   0 0 0
> c9t2d0p0 ONLINE   0 0 0
> 3763020893739678459  UNAVAIL  0 0 0  was
> /dev/dsk/c9t1d0p0
> c9t0d0p0 ONLINE   0 0 0
>
> errors: No known data errors
>
> I tried zpool replace on the drive.
> # zpool replace q 3763020893739678459 c9t1d0
> cannot replace 3763020893739678459 with c9t1d0: device is too small
>
> Victor was right. I went into 'format' and fought with it for a while.
> Moving the beginning of slice 0 from block 256 down to block 34 was simple
> enough, but I can not figure out how to tell it I don't want 8MB in slice 8.
> Is it even possible? I haven't got 8MB to spare (as silly as that sounds for
> a 1.5TB drive) - if I can't get rid of slice 8, I may have to stick with
> using p0's. I haven't encountered a problem using them so far (who needs
> partition tables anyway?) but I figured I'd ask if anybody had ideas about
> getting back that space.
> What's the 8MB for, anyway? Some stuff seems to indicate that it has to do
> with booting the drive, but this will never be a boot drive. That seems to
> be for VTOC stuff, not EFI, though. I did look at switching to VTOC labels,
> but it seems they don't support disks as large as I am using, so I think
> that's out.
> I also see "Information that was stored in the alternate cylinders area,
> the last two cylinders of the disk, is now stored in slice 8." (
> http://docsun.cites.uiuc.edu/sun_docs/C/solaris_9/SUNWaadm/SYSADV1/p117.html)
> Not sure what an "alternate cylinders area" is - it sounds sort of like
> remapping bad sectors, but that's something that the disk does on its own.
>
> So, can I get the 8MB back? Should I use p0? Is there another option I'm
> not thinking of? (I could always try diving into the EFI label with a hex
> editor and set it the way I please with no silly slice 8)
>
> -Ethan
>
>
I did a replace onto p0 of the drive I'd randomized, and did a scrub. No
errors. (Then I did another scrub, just for the hell of it; no errors
again.)

I feel fairly content staying with p0's, unless there's a good reason not
to. There are a few things I'm not entirely certain about:

- Is there any significant advantage to having a partition table?
- If there is, is it possible to drop the 8MB slice 8 so that I can actually
have enough space to put my raid on slice 0?
- Should I replace the disk that had errors on the initial scrub, or is it
probably sufficient to just be wary of it, scrub frequently, and replace it
if it encounters any more errors?

-Ethan
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SSDs with a SCSI SCA interface?

2010-02-23 Thread Daniel Carosone
On Tue, Feb 23, 2010 at 12:09:20PM -0800, Erik Trimble wrote:
> I've got stacks of both v20z/v40z hardware, plus a whole raft of IBM  
> xSeries (/not/ System X) machines which really, really, really need an  
> SSD for improved I/O.   At this point, I'd kill for a parallel SCSI ->  
> SATA adapter thingy; something that would plug into a SCA connector on  
> one side, and a SATA port on the other. I could at least hack together a  
> mounting bracket for something like that...

I've hoped for target-mode drivers for some of the older parallel-scsi
hba's.  With those, and comstar, one could connect such boxes to a
complete zfs storage system, booting and all.  In my case, the idea is
to get rid of the older, noisy powerhungry disks entirely - but it's
just an idle hope.

--
Dan.

pgp3jLhXv2ZED.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] SSDs with a SCSI SCA interface?

2010-02-23 Thread Erik Trimble

Al Hopper wrote:



On Fri, Dec 4, 2009 at 1:06 AM, Erik Trimble > wrote:


Hey folks.

I've looked around quite a bit, and I can't find something like this:

I have a bunch of older systems which use Ultra320 SCA hot-swap
connectors for their internal drives. (e.g. v20z and similar)

I'd love to be able to use modern flash SSDs with these systems,
but I have yet to find someone who makes anything that would fit
the bill.

I need either:

(a) a SSD with an Ultra160/320 parallel interface (I can always
find an interface adapter, so I'm not particular about whether
it's a 68-pin or SCA)

(b)  a  SAS or SATA to UltraSCSI adapter (preferably with a SCA
interface)


Hi Erik,

One of the less well known facts about SCSI is that all SCSI commands 
are sent in legacy 8-bit mode.  And it takes multiple SCSI commands to 
make a SCSI drive do something useful!   Translation -  it's s-l-o-w.  
Since one of the big upsides of an SSD is I/O Ops/Sec - get ready for 
a disappointment if you use SCSI based connection.   Sure - after the 
drive has received the necessary commands it can move data blocks 
reasonably quickly - but the limit, in terms of an SSD will 
*definitely* be the rate at which commands can be received by the 
drive.  This (8-bit command) design decision was responsible for 
SCSIs' long lasting upward compatibility - but it also turned into its 
achilles heel; that ultimately doomed SCSI to extinction.
Really?  I hadn't realized this was a problem with SSDs and SCSI.   
Exactly how does this impact SSDs with a SAS connection, since that's 
still using the SCSI command set, just over a serial link rather than a 
parallel one.  Or, am I missing something, and is SAS  considerably 
different (protocol wise) from traditional parallel SCSI?



Given the enormous amount of legacy hardware out there that has parallel 
SCSI drive bays (I mean, SAS is really only 2-3 years old in terms of 
server hardware adoption), I am just flabbergasted that there's no 
parallel-SCSI SSD around.


I understand exactly the problem you're solving - and you're not alone 
(got 4 V20Zs in a CoLo in Menlo Park CA that I maintain for 
Genunix.Org and I visit them less than once a year at great expense - 
both in terms of time and dollars)!   IMHO any kind of a hardware 
"hack job" and a couple of 1.8" or 2.5" SATA SSDs, combined with an 
OpenSolaris plugin SATA controller, would be a better solution. But I 
don't like this solution any more than I'm sure you do!


Please contact me offlist if you have any ideas and please let us know 
(on the list) how this works out for you.


Regards,

--
Al Hopper  Logical Approach Inc,Plano,TX a...@logical-approach.com 


  Voice: 972.379.2133 Timezone: US CDT
OpenSolaris Governing Board (OGB) Member - Apr 2005 to Mar 2007
http://www.opensolaris.org/os/community/ogb/ogb_2005-2007/


I've got stacks of both v20z/v40z hardware, plus a whole raft of IBM 
xSeries (/not/ System X) machines which really, really, really need an 
SSD for improved I/O.   At this point, I'd kill for a parallel SCSI -> 
SATA adapter thingy; something that would plug into a SCA connector on 
one side, and a SATA port on the other. I could at least hack together a 
mounting bracket for something like that...






--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Bruno Sousa
The system becames really slow during the data copy using network, but i
copy data between 2 pools of the box i don't notice that issue, so
probably i may be hitting some sort of interrupt conflit in the network
cards...This system is configured with *alot *of interfaces, being :

4 internal broadcom gigabit
1 PCIe 4x, Intel Dual Pro gigabit
1 PCIe 4x, Intel 10gbE card
2 PCIe 8x Sun non-raid HBA


With all of this, is there any way to check if there is indeed an
interrupt conflit or some other type of conflit that leads this high
load? I also noticed some messages about acpi..can this acpi also affect
the performance of the system?

Regards,
Bruno

On 23-2-2010 20:47, Bob Friesenhahn wrote:
> On Tue, 23 Feb 2010, Bruno Sousa wrote:
>
>> I don't have compression and deduplication enabled, but checksums are.
>> However disabling checksums gives a 0.5 load reduction only...
>
> Since high CPU consumption is unusual, I would suspect a device driver
> issue.  Perhaps there is an interrupt conflict such that two devices
> are using the same interrupt.
>
> On my own system (12 disks), I can run a throughput benchmark and the
> system remains completely usable as an interactive desktop system,
> without any large use of CPU or high load factor.  The bandwidth
> bottleneck in my case is the PCIe (4 lane) fiber channel card and its
> duplex connection to the storage array.
>
> Bob
> -- 
> Bob Friesenhahn
> bfrie...@simple.dallas.tx.us,
> http://www.simplesystems.org/users/bfriesen/
> GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
>



smime.p7s
Description: S/MIME Cryptographic Signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Bob Friesenhahn

On Tue, 23 Feb 2010, Bruno Sousa wrote:


I don't have compression and deduplication enabled, but checksums are.
However disabling checksums gives a 0.5 load reduction only...


Since high CPU consumption is unusual, I would suspect a device driver 
issue.  Perhaps there is an interrupt conflict such that two devices 
are using the same interrupt.


On my own system (12 disks), I can run a throughput benchmark and the 
system remains completely usable as an interactive desktop system, 
without any large use of CPU or high load factor.  The bandwidth 
bottleneck in my case is the PCIe (4 lane) fiber channel card and its 
duplex connection to the storage array.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Help with itadm commands

2010-02-23 Thread Nigel Smith
The iSCSI COMSTAR Port Provider is not installed by default.
What release of OpenSolaris are you running?
If pre snv_133 then:

  $ pfexec pkg install  SUNWiscsit

For snv_133, I think it will be:

  $ pfexec pkg install  network/iscsi/target

Regards
Nigel Smith
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Bob Friesenhahn

On Tue, 23 Feb 2010, Eugen Leitl wrote:


In terms of scaling, does zfs on OpenSolaris play well on multiple
cores? How much disks (assuming 100 MByte/s throughput for each)
would be considered pushing it for a current single-socket quadcore?


In any large storage system, most disks are relatively unused.  It is 
not normal for all disks to be pumping through their rated throughput 
at one time.  PCIe interfaces are only capable of a certain amount of 
bandwidth and this will place a hard limit on maximum throughput. 
There are also limits based on the raw memory bandwidth of the 
machine.


OpenSolaris is the king of multi-threading and excels on multiple 
cores.  Without this fine level of threading, SPARC CMT hardware would 
be rendered useless.


With this in mind, some older versions of OpenSolaris did experience a 
thread priority problem when compression was used.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Bruno Sousa
Hi Bob,

I have neither deduplication or compression enabled. The checksum are
enabled, but if try to disable it i gain aroud 0.5 less load on the box,
so it still seems to be to much.

Bruno

On 23-2-2010 20:03, Bob Friesenhahn wrote:
> On Tue, 23 Feb 2010, Bruno Sousa wrote:
>> Could the fact of having a RAIDZ2 configuration be the cause for such a
>> big load on the zfs box, or maybe am i missing something ?
>
> Zfs can consume appreciable CPU if compression, sha256 checksums,
> and/or deduplication is enabled.  Otherwise, substantial CPU
> consumption is unexpected.
>
> Are compression, sha256 checksums, or deduplication enabled for the
> filesystem you are using?
>
> Bob
> -- 
> Bob Friesenhahn
> bfrie...@simple.dallas.tx.us,
> http://www.simplesystems.org/users/bfriesen/
> GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
>




smime.p7s
Description: S/MIME Cryptographic Signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Bruno Sousa
Hi,

I don't have compression and deduplication enabled, but checksums are.
However disabling checksums gives a 0.5 load reduction only...


Bruno
On 23-2-2010 20:27, Eugen Leitl wrote:
> On Tue, Feb 23, 2010 at 01:03:04PM -0600, Bob Friesenhahn wrote:
>
>   
>> Zfs can consume appreciable CPU if compression, sha256 checksums, 
>> and/or deduplication is enabled.  Otherwise, substantial CPU 
>> consumption is unexpected.
>> 
> In terms of scaling, does zfs on OpenSolaris play well on multiple
> cores? How much disks (assuming 100 MByte/s throughput for each)
> would be considered pushing it for a current single-socket quadcore?
>  
>   
>> Are compression, sha256 checksums, or deduplication enabled for the 
>> filesystem you are using?
>> 
>   




smime.p7s
Description: S/MIME Cryptographic Signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Eugen Leitl
On Tue, Feb 23, 2010 at 01:03:04PM -0600, Bob Friesenhahn wrote:

> Zfs can consume appreciable CPU if compression, sha256 checksums, 
> and/or deduplication is enabled.  Otherwise, substantial CPU 
> consumption is unexpected.

In terms of scaling, does zfs on OpenSolaris play well on multiple
cores? How much disks (assuming 100 MByte/s throughput for each)
would be considered pushing it for a current single-socket quadcore?
 
> Are compression, sha256 checksums, or deduplication enabled for the 
> filesystem you are using?

-- 
Eugen* Leitl http://leitl.org";>leitl http://leitl.org
__
ICBM: 48.07100, 11.36820 http://www.ativel.com http://postbiota.org
8B29F6BE: 099D 78BA 2FD3 B014 B08A  7779 75B0 2443 8B29 F6BE
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Help with itadm commands

2010-02-23 Thread Jeff Freeman
Hi - I'm trying to create an iscsi targe and go thru the motions of making the 
following LUN's available - I am not able to run the command: itadm 
create-target as I get the following error: bash: itadm: command not found

I need to get the following dirve seen by vmware 

0. c7t0d0 
  /p...@0,0/pci8086,2...@1/pci8086,3...@0/pci17d3,1...@e/d...@0,0
1. c7t0d1 
  /p...@0,0/pci8086,2...@1/pci8086,3...@0/pci17d3,1...@e/d...@0,1
2. c7t0d2 
  /p...@0,0/pci8086,2...@1/pci8086,3...@0/pci17d3,1...@e/d...@0,2


any help would be greatly appreciated
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Bob Friesenhahn

On Tue, 23 Feb 2010, Bruno Sousa wrote:

Could the fact of having a RAIDZ2 configuration be the cause for such a
big load on the zfs box, or maybe am i missing something ?


Zfs can consume appreciable CPU if compression, sha256 checksums, 
and/or deduplication is enabled.  Otherwise, substantial CPU 
consumption is unexpected.


Are compression, sha256 checksums, or deduplication enabled for the 
filesystem you are using?


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] verging OT: how to buy J4500 w/o overpriced drives

2010-02-23 Thread Bob Friesenhahn

On Tue, 23 Feb 2010, matthew patton wrote:


Of the "we force you to buy our overinflated drives" camp, Dell is 
the cheapest but also the most inefficient by far on power/space. 
The HP puts 70 disks in 4U. NexSan 42, and Sun 48. The clear winner 
here is HP.


What is the performance like with HP?  Is there a loss of bandwidth 
or reliability due to their approach?


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] snv_133 - high cpu

2010-02-23 Thread Bruno Sousa
Hi all,

I'm currently evaluating the possibility of migrating a NFS server
(Linux Centos 5.4 / RHEL 5.4 x64-32) based to a opensolaris box and i'm
seeing some huge cpu usage in the opensolaris box.

The zfs box is a Dell R710 with 2 Quad-Cores (Intel E5506  @ 2.13GHz),
16Gb ram , 2 Sun non-Raid HBA's connected to two J4400 jbods, while the
Linux box is a 2Xeon 3.0Ghz with 8Gb ram, a areca HBA with 512 mb cache,
and both of the servers have a Intel 10gbE card with jumbo frames enabled.

This zfs box has one pool in a raidz2 with multipath enable (to make use
of the 2hbas and 2 j4400), with 20 disks (sata 7.200 rpm seagate
enterprise as supplied by Sun). The raidz2 has 5 vdevs with 4 disks each.
The test is made by mounting in the linux box one nfs share from the zfs
box, and copy around 1.1TB of data , and this data is users's home
directories, so thousands of small files.
During the copy procedure from the linux box to the zfs box the load in
the zfs box is between 8 and 10 while on the linux box it never goes
over 1 .
Could the fact of having a RAIDZ2 configuration be the cause for such a
big load on the zfs box, or maybe am i missing something ?

Thanks for all your time,
Bruno


Here are some more specs from the ZFS box :

r...@zfsbox01:/var/adm# zpool status -v RAIDZ2
  pool: RAIDZ2
 state: ONLINE
 scrub: none requested
config:

NAME   STATE READ WRITE CKSUM
RAIDZ2 ONLINE   0 0 0
  raidz2-0 ONLINE   0 0 0
c0t5000C5001A101764d0  ONLINE   0 0 0
c0t5000C5001A315D0Ad0  ONLINE   0 0 0
c0t5000C5001A10EC6Bd0  ONLINE   0 0 0
c0t5000C5001A0FFF4Bd0  ONLINE   0 0 0
  raidz2-1 ONLINE   0 0 0
c0t5000C50019C0A04Ed0  ONLINE   0 0 0
c0t5000C5001A0FA028d0  ONLINE   0 0 0
c0t5000C50019FCF180d0  ONLINE   0 0 0
c0t5000C5001A11E657d0  ONLINE   0 0 0
  raidz2-2 ONLINE   0 0 0
c0t5000C5001A104A30d0  ONLINE   0 0 0
c0t5000C5001A316841d0  ONLINE   0 0 0
c0t5000C5001A0FF92Ed0  ONLINE   0 0 0
c0t5000C50019EB02FDd0  ONLINE   0 0 0
  raidz2-3 ONLINE   0 0 0
c0t5000C5001A0FDBDCd0  ONLINE   0 0 0
c0t5000C5001A0F2197d0  ONLINE   0 0 0
c0t5000C50019BDBB8Dd0  ONLINE   0 0 0
c0t5000C5001A3152A0d0  ONLINE   0 0 0
  raidz2-4 ONLINE   0 0 0
c0t5000C5001A100DA0d0  ONLINE   0 0 0
c0t5000C5001A31544Cd0  ONLINE   0 0 0
c0t5000C50019F03AF6d0  ONLINE   0 0 0
c0t5000C50019FC3055d0  ONLINE   0 0 0

###

r...@zfsbox01:~# zpool iostat RAIDZ2 5
   capacity operationsbandwidth
poolalloc   free   read  write   read  write
--  -  -  -  -  -  -
RAIDZ2  2.29T  15.8T 43305  3.03M  14.6M
RAIDZ2  2.29T  15.8T114663  12.7M  18.6M
RAIDZ2  2.29T  15.8T129595  14.0M  11.2M
RAIDZ2  2.29T  15.8T881623  13.0M  10.4M
RAIDZ2  2.29T  15.8T227449  8.48M  17.5M
RAIDZ2  2.29T  15.8T 39498  4.55M  29.1M

###

r...@zfsbox01:~# top -b | grep CPU | head -n1
CPU states: 35.2% idle,  2.2% user, 62.6% kernel,  0.0% iowait,  0.0% swap

r...@zfsbox01:~# mpstat
CPU minf mjf xcal  intr ithr  csw icsw migr smtx  srw syscl  usr sys  wt idl
  0   55   0 16969 18180  102  785   55  127 17794   2421  69  
0  30
  1   70   0 18005 168204  926   44  142 18896   1591  65  
0  35
  2   42   0 16659 18091  262  555   53  113 1757   11   2502  68  
0  31
  3   48   0 18221 17380  246  667   40  122 1929   12   1321  66  
0  33
  4   38   0 16547 19965 1766  517   48  107 1775   10   2642  70  
0  29
  5   42   0 18596 19113 1527  595   35  115 19876   1561  69  
0  31
  6   23   0 16284 17921   10 2066   54  109 17634   1151  70  
0  29
  7   32   0 17576 166653 2233   39  134 18475900  64  
0  35

top -b| grep Memory
Memory: 16G phys mem, 2181M free mem, 8187M total swap, 8187M free swap


Feb 18 11:42:36 zfsbox01 unix: [ID 378719 kern.info] NOTICE: cpu_acpi:
_PSS package evaluation failed for with status 5 for CPU 2.
Feb 18 11:42:36 zfsbox01 unix: [ID 388705 kern.info] NOTICE: cpu_acpi:
error parsing _PSS for CPU 2
 

Feb 18 11:43:12 zfsbox01 ixgbe: [ID 611667 kern.info] NOTICE: ixgbe0:
identify 82598 adapter
Feb 18 11:43:12 zfsbox01 ixgbe: [ID 611667 kern.info] NOTICE: ixgbe0:
Request 16 handles, 2 available
Feb 18 11:43:12 zfsbox0

Re: [zfs-discuss] controller cache instead of dedicated ZIL device

2010-02-23 Thread Richard Elling
On Feb 23, 2010, at 9:16 AM, Felix Buenemann wrote:
> Hi,
> 
> as it turns out to be pretty difficult (or expensive), to find high 
> performance dedicated ZIL devices, I had another thought:
> 
> If using a RAID controller with a large cache, eg. 4GB and battery backup in 
> JBOD mode and using on disk ZIL – wouldn't the controller cache work as a 
> great ZIL accelerator, requiring no dedicated ZIL?

Yes. ZIL is a performance problem for HDD JBODs, not so much on devices with
fast, nonvolatile write caches.

> If my understanding is correct, a battery backed RAID controller will ignore 
> cache flush commands and thus the controller cache would be a very low 
> latency intermediate cache wich is preserved over power failure.

Cache flush latency is orthogonal to slog latency.
 -- richard

ZFS storage and performance consulting at http://www.RichardElling.com
ZFS training on deduplication, NexentaStor, and NAS performance
http://nexenta-atlanta.eventbrite.com (March 16-18, 2010)




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] shrinking a zpool - roadmap

2010-02-23 Thread Richard Elling
On Feb 23, 2010, at 5:10 AM, Robert Milkowski wrote:
> On 23/02/2010 02:52, Richard Elling wrote:
>> On Feb 22, 2010, at 6:42 PM, Charles Hedrick wrote:
>>   
>>> I talked with our enterprise systems people recently. I don't believe 
>>> they'd consider ZFS until it's more flexible. Shrink is a big one, as is 
>>> removing an slog. We also need to be able to expand a raidz, possibly by 
>>> striping it with a second one and then rebalancing the sizes.
>>> 
>> So what file system do they use that has all of these features? :-P
>> 
>>   
> VxVM + VxFS?

I did know they still cost $$$, but I didn't know they implemented a slog :-P
 -- richard

ZFS storage and performance consulting at http://www.RichardElling.com
ZFS training on deduplication, NexentaStor, and NAS performance
http://nexenta-atlanta.eventbrite.com (March 16-18, 2010)




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs sequential read performance

2010-02-23 Thread Richard Elling
On Feb 23, 2010, at 1:00 AM, v wrote:

> Hi experts,
> I am new to zfs and ask a question regarding zfs sequential peroformance: I 
> read some blogs saying that netapp's WAFL can suffer "sequential read after 
> random write(SRARW)" performance penalty, since zfs is also doing no update 
> in place, can zfs has such problem?

I know of only one study of this effect, by Allan Packer and Neel using MySQL.
http://www.youtube.com/watch?v=a31NhwzlAxs
slides at
http://blogs.sun.com/realneel/resource/MySQL_Conference_2009_ZFS_MySQL.pdf

However, there is much speculation in the ZFS archives...
 -- richard

ZFS storage and performance consulting at http://www.RichardElling.com
ZFS training on deduplication, NexentaStor, and NAS performance
http://nexenta-atlanta.eventbrite.com (March 16-18, 2010)




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] controller cache instead of dedicated ZIL device

2010-02-23 Thread Felix Buenemann

Hi,

as it turns out to be pretty difficult (or expensive), to find high 
performance dedicated ZIL devices, I had another thought:


If using a RAID controller with a large cache, eg. 4GB and battery 
backup in JBOD mode and using on disk ZIL – wouldn't the controller 
cache work as a great ZIL accelerator, requiring no dedicated ZIL?


If my understanding is correct, a battery backed RAID controller will 
ignore cache flush commands and thus the controller cache would be a 
very low latency intermediate cache wich is preserved over power failure.


Best Regards,
Felix Buenemann

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] future of OpenSolaris

2010-02-23 Thread Bob Friesenhahn

On Tue, 23 Feb 2010, Joerg Schilling wrote:


and what uname -s reports.


It will surely report "OrkOS".

Ork:

 2. (Mythology) A mythical monster of varying descriptions; an
ogre.
[PJC]

  Goblins, hobgoblins, and orcs of the worst
  description.  --J. J.
Tolkien (The
Hobbit)

I am glad to be able to contribute positively and constructively to 
this discussion.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] panic: assertion failed: 0 == dmu_buf_hold_array(os, object, offset, size, FALSE, FTAG, &numbufs, &dbp), file: ../../common/fs/zfs/dmu.c, line: 591

2010-02-23 Thread Brian Kolaci


I recently upgraded a box to Solaris 10 U8.
I've been getting more timeouts and I guess the Adaptec card is suspect, possibly not 
able to keep up, so it issues bus resets at times.  It has apparently corrupted some 
files on the pool, and zpool status -v showed 2 files and one dataset corrupt.  I 
initially was able to bring the pool up and salvage some of the files.  It would not let 
me remove the files listed giving a "Bad exchange descriptor error".  So I 
figured I'd salvage and remove those two datasets and try again.

So while trying to salvage what I could, it apparently stressed the card too 
much (constantly at 98-100% busy), eventually the service times increased high 
enough and then failed with timeouts and another bus reset.

Then it crashed with the following:

panic[cpu2]/thread=c603adc0: assertion failed: 0 == dmu_buf_hold_array(os, object, 
offset, size, FALSE, FTAG, &numbufs, &dbp), file:
../../common/fs/zfs/dmu.c, line: 591

c603abec genunix:assfail+51 (edf9094c, edf90930,)
c603ac34 zfs:dmu_write+150 (c5aa3a20, 86, 0, b5)
c603ac9c zfs:space_map_sync+2ed (c6fde4cc, 1, c6fde3)
c603acec zfs:metaslab_sync+245 (c6fde340, 904f005, )
c603ad14 zfs:vdev_sync+a8 (c0bad040, 904f005, )
c603ad5c zfs:spa_sync+38e (c23196c0, 904f005, )
c603ada8 zfs:txg_sync_thread+22c (c1016600, 0)
c603adb8 unix:thread_start+8 ()

syncing file systems... [1] [1] [1] [1] [1] [1] [1] [1] [1] [1] [1] [1] [1] [1] 
[1] [1] [1] [1] [1] [1] [1] done (not all i/o completed)
dumping to /dev/dsk/c0t0d0s1, offset 215547904, content: kernel
WARNING: This system contains a SCSI HBA card/driver that doesn't support 
software reset. This means that memory being used by the HBA for DMA based 
reads could have been updated after we panic'd.


And then it would not boot anymore.  It just went into a panic loop.  I hopped in the car 
and went to the data center.  I managed to boot off CD, mounted the root file system and 
moved /etc/zfs/zpool.cache out of the way, so now I can boot the OS again.  If I try to 
import the pool, I get the panic as above.  If you just enter "zpool import", I 
get the following:

 state: ONLINE
status: The pool is formatted using an older on-disk version.
action: The pool can be imported using its name or numeric identifier, though
some features will not be available without an explicit 'zpool upgrade'.
config:

pool0   ONLINE
  c2t4d0ONLINE
  c2t4d2ONLINE

So it appears to still be there, but I can't import it.
The two devices are actually hardware RAID devices of 750G each, so I don't 
have redundancy on the system level, only the hardware RAIDs.

I'm not too sure what to do with zdb to see anything.
Any ideas as to what I can do to recover the rest of the data?
There's still some database files on there I need.

Thanks,

Brian




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] verging OT: how to buy J4500 w/o overpriced drives

2010-02-23 Thread matthew patton
For those who are interested in some of the options out there.

DIY DAS:
Supermicro 36 bay case - $1800
Promise 16 bay JBOD VTrak J610sD - $3700
Promise VTE610sD - $7500 (SAS attached head unit with onboard raid controllers, 
takes JBOD expansion)

The following apply to 1TB SATA drive configurations, dual controllers under 
federal pricing:

HP MD600 - $722/TB
DELL MD1000 - $680/TB (780/TB w/ nearline-SAS)
Sun J4500 - $1014/TB (federal discount is a mere 13%, not 40)
NexSan SATAbeast - $1400/TB (FC attached dual-head)


Of the "we force you to buy our overinflated drives" camp, Dell is the cheapest 
but also the most inefficient by far on power/space. The HP puts 70 disks in 
4U. NexSan 42, and Sun 48. The clear winner here is HP.


  
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Opensolaris 2010.03 / snv releases

2010-02-23 Thread Marc Nicholas
Isn't the dedupe bug fixed in svn133?

-marc

On Tue, Feb 23, 2010 at 9:21 AM, Jeffry Molanus wrote:

> There is no clustering package for it and available source seems very old
> also the de-dup bug is there iirc. So if you don't need HA cluster and
> dedup..
>
> BR, Jeffry
>
> > -Original Message-
> > From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> > boun...@opensolaris.org] On Behalf Of Bruno Sousa
> > Sent: dinsdag 23 februari 2010 8:37
> > To: zfs-discuss@opensolaris.org
> > Subject: [zfs-discuss] Opensolaris 2010.03 / snv releases
> >
> > Hi all,
> >
> > According to what i have been reading the opensolaris 2010.03 should be
> > released around March this year, but with all the process of the
> Oracle/Sun
> > deal i was wondering if anyone knows if this schedule still makes sense,
> > and if not does snv_132/133 look very similar to future 2010.03.
> > In other words, without waiting for the opensolaris 2010.03 would anyone
> > risk to put in production any box with snv_132/133 ?
> >
> > Thanks,
> > Bruno
>
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] future of OpenSolaris

2010-02-23 Thread Joerg Schilling
Erik Trimble  wrote:

> But, yes, the proper way to talk all about this now is to say "Oracle" 
> when you mean the company, and "Sun" when you are talking about specific 
> brand-name products.  The latter will almost certainly be restricted to 
> hardware - all software that I know of is having a 's/Sun/Oracle/'  
> regex done on it.

If this is done really everywhere, then I suspect that a lot of software
will no longer compile correctly. 

It would be interesting to know whether

sun, __sun, __SunOS*, __SUNPRO*, __SUN* and similar sre still defined

and what uname -s reports.

Jörg

-- 
 EMail:jo...@schily.isdn.cs.tu-berlin.de (home) Jörg Schilling D-13353 Berlin
   j...@cs.tu-berlin.de(uni)  
   joerg.schill...@fokus.fraunhofer.de (work) Blog: 
http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Opensolaris 2010.03 / snv releases

2010-02-23 Thread Jeffry Molanus
There is no clustering package for it and available source seems very old also 
the de-dup bug is there iirc. So if you don't need HA cluster and dedup.. 

BR, Jeffry

> -Original Message-
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Bruno Sousa
> Sent: dinsdag 23 februari 2010 8:37
> To: zfs-discuss@opensolaris.org
> Subject: [zfs-discuss] Opensolaris 2010.03 / snv releases
> 
> Hi all,
> 
> According to what i have been reading the opensolaris 2010.03 should be
> released around March this year, but with all the process of the Oracle/Sun
> deal i was wondering if anyone knows if this schedule still makes sense,
> and if not does snv_132/133 look very similar to future 2010.03.
> In other words, without waiting for the opensolaris 2010.03 would anyone
> risk to put in production any box with snv_132/133 ?
> 
> Thanks,
> Bruno

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Adding a zfs mirror drive to rpool - new drive formats to one cylinder less

2010-02-23 Thread Mark J Musante

On Mon, 22 Feb 2010, tomwaters wrote:

I have just installed open solaris 2009.6 on my server using a 250G 
laptop drive (using the entire drive).


So, 2009.06 was based on 111b.  There was a fix that went into build 117 
that allows you to mirror to smaller disks if the metaslabs in zfs are 
still the same size even if the disk is one block smaller.  CR 6844090.


If you can upgrade to a later build, this will very probably fix your 
issue.



Regards,
markm
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] More performance questions [on zfs over nfs]

2010-02-23 Thread Harry Putnam
Harry Putnam  writes:

> [Note: This is a repost of question posted about 1.5 days ago that
> has never appeared on the group.. at least not on my server (gmane).
> Sorry if it ends up being a double whammy]

Apparently I missed two informative answers by:
Henrik J.  Bob F.

Thanks for the input... and technical summation, I still don't see my
original post but no matter, my second post threaded with the two
authors answers and let me see them.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] shrinking a zpool - roadmap

2010-02-23 Thread Robert Milkowski

On 23/02/2010 02:52, Richard Elling wrote:

On Feb 22, 2010, at 6:42 PM, Charles Hedrick wrote:
   

I talked with our enterprise systems people recently. I don't believe they'd 
consider ZFS until it's more flexible. Shrink is a big one, as is removing an 
slog. We also need to be able to expand a raidz, possibly by striping it with a 
second one and then rebalancing the sizes.
 

So what file system do they use that has all of these features? :-P

   

VxVM + VxFS?

--
Robert Milkowski
http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs sequential read performance

2010-02-23 Thread Robert Milkowski

On 23/02/2010 09:00, v wrote:

Hi experts,
I am new to zfs and ask a question regarding zfs sequential peroformance: I read some 
blogs saying that netapp's WAFL can suffer "sequential read after random 
write(SRARW)" performance penalty, since zfs is also doing no update in place, can 
zfs has such problem?

Thanks
Victor
   

yes

but on the other hand it usually means much faster random writes.

--
Robert Milkowski
http://milek.blogspot.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] future of OpenSolaris

2010-02-23 Thread Anil Gulecha
Hi,

> Without saying anything negative about Nexenta I would strongly recommend
> you go try to send a single patch to their equivalent of onnv-gate before
> recommending it as any sort of replacement for OpenSolaris.

Not sure what the above is intended to mean.

To clear things, Nexenta project is open to patches. It always makes
sense to send patches to the upstream project (for any project), but
if there's an instance where upstream does not integrate a patch, the
Nexenta project will definitely consider adding it. This discussion is
a bug/rfe report away.

Code/Tracker: http://www.nexenta.org/projects/nexenta-gate

Thanks,
Anil
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [zones-discuss] Zones on shared storage - a warning

2010-02-23 Thread Frank Batschulat (Home)
update on this one: 

a workaround if you so will, or the more appropriate way to do this is 
apparently
to use lofiadm(1M) to create a pseudo block device comprising the file hosted 
on NFS
and use the created lofi device (eg. /dev/lofi/1) as the device for zpool create
and all subsequent I/O (this was not producing the strange CKSUM errors), eg.:

osoldev.root./export/home/batschul.=> mount -F nfs opteron:/pool/zones /nfszone
osoldev.root./export/home/batschul.=> mount -v| grep nfs
opteron:/pool/zones on /nfszone type nfs 
remote/read/write/setuid/devices/xattr/dev=9080001 on Tue Feb  9 10:37:00 2010
osoldev.root./export/home/batschul.=> nfsstat -m
/nfszone from opteron:/pool/zones
 Flags: 
vers=4,proto=tcp,sec=sys,hard,intr,link,symlink,acl,rsize=1048576,wsize=1048576,retrans=5,timeo=600
 Attr cache:acregmin=3,acregmax=60,acdirmin=30,acdirmax=60

osoldev.root./export/home/batschul.=>  mkfile -n 7G /nfszone/remote.file
osoldev.root./export/home/batschul.=>  ls -la /nfszone
total 28243534
drwxrwxrwx   2 nobody   nobody 6 Feb  9 09:36 .
drwxr-xr-x  30 batschul other 32 Feb  8 22:24 ..
-rw---   1 nobody   nobody   7516192768 Feb  9 09:36 remote.file

osoldev.root./export/home/batschul.=> lofiadm -a /nfszone/remote.file
/dev/lofi/1

osoldev.root./export/home/batschul.=> lofiadm
Block Device File   Options
/dev/lofi/1  /nfszone/remote.file   -

osoldev.root./export/home/batschul.=> zpool create -m /tank/zones/nfszone 
nfszone /dev/lofi/1

Feb  9 10:50:35 osoldev zfs: [ID 249136 kern.info] created version 22 pool 
nfszone using 22

osoldev.root./export/home/batschul.=> zpool status -v nfszone
  pool: nfszone
 state: ONLINE
 scrub: none requested
config:

NAME   STATE READ WRITE CKSUM
nfszoneONLINE   0 0 0
  /dev/lofi/1  ONLINE   0 0 0

---
frankB
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Abysmal ISCSI / ZFS Performance

2010-02-23 Thread Kjetil Torgrim Homme
Miles Nordin  writes:

>> "kth" == Kjetil Torgrim Homme  writes:
>
>kth> the SCSI layer handles the replaying of operations after a
>kth> reboot or connection failure.
>
> how?
>
> I do not think it is handled by SCSI layers, not for SAS nor iSCSI.

sorry, I was inaccurate.  error reporting is done by the SCSI layer, and
the filesystem handles it by retrying whatever outstanding operations it
has.

> Also, remember a write command that goes into the write cache is a
> SCSI command that's succeeded, even though it's not actually on disk
> for sure unless you can complete a sync cache command successfully and
> do so with no errors nor ``protocol events'' in the gap between the
> successful write and the successful sync.  A facility to replay failed
> commands won't help because when a drive with write cache on reboots,
> successful writes are rolled back.

this is true, sorry about my lack of precision.  the SCSI layer can't do
this on its own.

-- 
Kjetil T. Homme
Redpill Linpro AS - Changing the game

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Adding a zfs mirror drive to rpool - new drive formats to one cylinder less

2010-02-23 Thread tomwaters
Looks like an issue with the start /length of the partition table...

These are the disks from "fomrat"...
   8. c8t0d0 
  /p...@0,0/pci8086,3...@1f,2/d...@0,0
   9. c8t1d0 
  /p...@0,0/pci8086,3...@1f,2/d...@1,0

Loking at the partitions, the existing rpool disk is formatted like this..

 Total disk size is 30401 cylinders
 Cylinder size is 16065 (512 byte) blocks

   Cylinders
  Partition   StatusType  Start   End   Length%
  =   ==  =   ===   ==   ===
  1   ActiveSolaris2  0  3040130402100


The "new" disk is formatted like this...
 Total disk size is 30401 cylinders
 Cylinder size is 16065 (512 byte) blocks

   Cylinders
  Partition   StatusType  Start   End   Length%
  =   ==  =   ===   ==   ===
  1   ActiveSolaris2  1  3040030400100


I tried to manually create a partition by selecting "part" from the menu but I 
can not tell it to start at cyl. 0 or go to 30399...

Help!

Why is this so hard!
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Poor ZIL SLC SSD performance

2010-02-23 Thread Al Hopper
On Fri, Feb 19, 2010 at 2:17 PM, Felix Buenemann
 wrote:
>
> Am 19.02.10 20:50, schrieb Bob Friesenhahn:
>>
>> On Fri, 19 Feb 2010, David Dyer-Bennet wrote:
>>
 Too bad, I'm getting ~1000 IOPS with an Intel X25-M G2 MLC and around
 300 with a regular USB stick, so 50 IOPS is really poor for an SLC SSD.
>>>
>>> Well, but the Intel X25-M is the drive that really first cracked the
>>> problem (earlier high-performance drives were hideously expensive and
>>> rather brute force). Which was relatively recently. The industry is
>>> still evolving rapidly.
>>
>> What is the problem is it that the X25-M cracked? The X25-M is
>> demonstrated to ignore cache sync and toss transactions. As such, it is
>> useless for a ZIL.
>
> Yes, I see no difference with the X25-M with both zfs_nocacheflush=0 and 
> zfs_nocacheflush=1. After setting zfs_nocacheflush=1, the Mtron SSD also 
> performed at around 1000 IOPS, which is still useless, because the array 
> performs the same IOPS without dedicated ZIL.
> Looking at the X25-E (SLC) benchmarks it should be able to do about 3000 
> IOPS, which would improve array performance.
>
> I think I'll try one of thise inexpensive battery-backed PCI RAM drives from 
> Gigabyte and see how much IOPS they can pull.
>

I've given up on the Gigabyte card - it's basically unstable.  I've
tested it as a disk drive under ZFS and "another" operating system.
Again - it glitches out - sometimes after only a couple of minutes if
you tar up /usr (as a relative tar file), gzip it, and then untar it
onto the Gigabyte "drive".  I've got two of them - both using the
recommended Kingston RAM.  Both are unstable/flaky.  I've tried
removing some of the RAM to see if that makes a difference - it does
not.  Conclusion: Run Away from this product.

Regards,

--
Al Hopper  Logical Approach Inc,Plano,TX a...@logical-approach.com
                  Voice: 972.379.2133 Timezone: US CDT
OpenSolaris Governing Board (OGB) Member - Apr 2005 to Mar 2007
http://www.opensolaris.org/os/community/ogb/ogb_2005-2007/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs sequential read performance

2010-02-23 Thread v
Hi experts,
I am new to zfs and ask a question regarding zfs sequential peroformance: I 
read some blogs saying that netapp's WAFL can suffer "sequential read after 
random write(SRARW)" performance penalty, since zfs is also doing no update in 
place, can zfs has such problem?

Thanks
Victor
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss