Re: [zfs-discuss] How does resilver/scrub work?

2012-05-25 Thread zfs user
On 5/23/12 11:28 PM, Richard Elling wrote: The man page is clear on this topic, IMHO Indeed, even in snv_117 the zpool man page says that. But the console/dmesg message was also quite clear, so go figure whom to trust (or fear) more ;) The FMA message is consistent with the man page. The

Re: [zfs-discuss] webserver zfs root lock contention under heavy load

2012-03-25 Thread zfs user
On 3/25/12 10:25 AM, Aubrey Li wrote: On Mon, Mar 26, 2012 at 12:48 AM, Richard Elling richard.ell...@richardelling.com wrote: This is the wrong forum for general purpose performance tuning. So I won't continue this much farther. Notice the huge number of icsw, that is a bigger symptom than

Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-27 Thread zfs user
On 4/27/11 4:00 AM, Markus Kovero wrote: Sync was disabled on the main pool and then let to inherrit to everything else. The reason for disabled this in the first place was to fix bad NFS write performance (even with Zil on an X25e SSD it was under 1MB/s). I've also tried setting the

Re: [zfs-discuss] show hdd life time ?

2011-01-23 Thread zfs user
On 1/23/11 10:30 AM, Roy Sigurd Karlsbakk wrote: If you're looking for stats to give an indication of likely wear, and thus increasing probably of failure, POH is probably not very useful by itself (or even at all). Things like Head Flying Hours and Load Cycle Count are probably more indicative,

Re: [zfs-discuss] ZFS Crypto in Oracle Solaris 11 Express

2010-11-17 Thread zfs user
On 11/17/10 12:04 PM, Miles Nordin wrote: djm == Darren J Moffatdarr...@opensolaris.org writes: djm http://blogs.sun.com/darren/entry/introducing_zfs_crypto_in_oracle djm http://blogs.sun.com/darren/entry/assued_delete_with_zfs_dataset djm

Re: [zfs-discuss] Performance issues with iSCSI under Linux

2010-10-30 Thread zfs user
So maybe a next step is to run zilstat, arcstat, iostat -xe?? (I forget what people like to use for these params), zpool iostat -v in 4 term windows while running the same test and try to see what is spiking when that high load period occurs. Not sure if there is a better version than this:

Re: [zfs-discuss] Performance issues with iSCSI under Linux

2010-10-30 Thread zfs user
Here is a total guess - but what if it has to do with zfs processing running on one CPU having to talk to the memory owned by a different CPU? I don't know if many people are running fully populated boxes like you are, so maybe it is something people are not seeing due to not having huge

Re: [zfs-discuss] Performance issues with iSCSI under Linux

2010-10-30 Thread zfs user
. I am seeing some spotty performance with my new Mangy Cours CPU... It is like they are asking for it. I think they be better off doing something like Intel core arch names using city names Santa Rosa, etc. On 10/30/10 3:49 PM, Eugen Leitl wrote: On Sat, Oct 30, 2010 at 02:10:49PM -0700, zfs

Re: [zfs-discuss] Performance issues with iSCSI under Linux

2010-10-30 Thread zfs user
:07 PM, zfs user wrote: I did it deliberately - how dumb are these product managers that they name products with weird names and not expect them to be abused? On the other hand, if you do a search for mangy cours you'll find a bunch of hits where it is clearly a misspelling on serious tech articles

[zfs-discuss] AMD/Supermicro machine - AS-2022G-URF

2010-10-14 Thread zfs user
Sorry for the long post but I know trying to decide on hardware often want to see details about what people are using. I have the following AS-2022G-URF machine running OpenGaryIndiana[1] that I am starting to use. I successfully transferred a deduped zpool with 1.x TB of files and 60 or so

Re: [zfs-discuss] Dedup relationship between pool and filesystem

2010-09-23 Thread zfs user
I believe it goes a something like this - ZPS filesystems with dedupe turned on can be thought of as hippie/socialist filesystems, wanting to share, etc. Filesystems with dedupe turned off are a grey Randian landscape where sharing blocks between files is seen as a weakness/defect. They all

Re: [zfs-discuss] Raidz1 p

2009-01-20 Thread zfs user
I would get a new 1.5 TB and make sure it has the new firmware and replace c6t3d0 right away - even if someone here comes up with a magic solution, you don't want to wait for another drive to fail. http://hardware.slashdot.org/article.pl?sid=09/01/17/0115207

Re: [zfs-discuss] OpenSolaris better Than Solaris10u6 with requards to ARECA Raid Card

2009-01-13 Thread zfs user
Charles Wright wrote: Under Solaris 10 u6 , No matter how I configured my ARECA 1261ML Raid card I got errors on all drives that result from SCSI timeout errors. [snip litany of errors] I had similar problems on a 1120 card with 2008.05 I upgraded to 2008.11 and the something*.16 sun areca

Re: [zfs-discuss] ZFS and SAN Arrays

2008-12-17 Thread zfs user
Bob Friesenhahn wrote: On Tue, 16 Dec 2008, Reed Gregory wrote: 8 Hardware RAID-5 Groups ( 5 drives each) and 2 SAN hot spares. zraid of these 8 Raid Groups. ~ 14TB usable. I did read in a FAQ that doing double redundancy is not recommended since parity would have to be calculated twice.

Re: [zfs-discuss] RE : rsync using 100% of a cpu

2008-12-02 Thread zfs user
Francois Dion wrote: Francois Dion wrote: Source is local to rsync, copying from a zfs file system, destination is remote over a dsl connection. Takes forever to just go through the unchanged files. Going the other way is not a problem, it takes a fraction of the time. Anybody