[zfs-discuss] current status of SAM-QFS?

2012-05-01 Thread Fred Liu
The subject says it all.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [developer] Setting default user/group quotas[usage accounting]?

2012-05-01 Thread Fred Liu

On Apr 26, 2012, at 12:27 AM, Fred Liu wrote:


zfs 'userused@' properties and 'zfs userspace' command are good enough to 
gather usage statistics.
I think I mix that with NetApp. If my memory is correct, we have to set quotas 
to get usage statistics under DataOnTAP.
Further, if we can add an ILM-like feature to poll the time-related 
info(atime,mtime,ctime,etc) with that statistic from ZFS, that will be
really cool.

In general, file-based ILM has limitations that cause all sorts of issues for 
things like
operating systems, where files might only be needed infrequently, but when 
they are
needed, they are needed right now

Have you looked at zfs diff  for changed files?

   Here ILM-like feature, I mean we know how the data is 
distributed by time per pool/filesystem like how many data are modified/accessed
   before mm/dd/. And we don't need to do the actual storage-tying 
operations immediately(moving the infrequently-used data to tie-2
   storage). The time-related usage statistics are very useful reference 
for us.
   zfs diff will show the delta but not come with the time info.


Since no one is focusing on enabling default user/group quota now, the 
temporarily remedy could be a script which traverse all the users/groups
in the directory tree. Tough it is not so decent.

The largest market for user/group quotas is .edu. But they represent only a 
small market
when measured by $.

There are also many corner cases in this problem space. One might pine for the 
days of
VMS and its file resource management features, those features don't scale well 
to company-
wide LDAP and thousands of file systems.

My understanding is the quota management is needed as long as 
zfs storage is used in NAS way(shared by multi-users).


So, for now, the fastest method to solve the problem might be to script some 
walkers.
Yes. That is ture.

Thanks.


Fred

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] IOzone benchmarking

2012-05-01 Thread Ray Van Dolson
I'm trying to run some IOzone benchmarking on a new system to get a
feel for baseline performance.

Unfortunately, the system has a lot of memory (144GB), but I have some
time so am approaching my runs as follows:

Throughput:
iozone -m -t 8 -T -r 128k -o -s 36G -R -b bigfile.xls

IOPS:
iozone -O -i 0 -i 1 -i 2 -e -+n -r 128K -s 288G  iops.txt

Not sure what I gain/lose by using threads or not.

Am I off on this?

System is a 240x2TB (7200RPM) system in 20 Dell MD1200 JBODs.  16 vdevs of 15
disks each -- RAIDZ3.  NexentaStor 3.1.2.

Ray
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] current status of SAM-QFS?

2012-05-01 Thread Darren J Moffat

On 04/30/12 04:00, Fred Liu wrote:

The subject says it all.


Still a fully supported product from Oracle:

http://www.oracle.com/us/products/servers-storage/storage/storage-software/qfs-software/overview/index.html

--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [developer] Setting default user/group quotas[usage accounting]?

2012-05-01 Thread Richard Elling

On Apr 29, 2012, at 7:59 PM, Fred Liu wrote:

  
 On Apr 26, 2012, at 12:27 AM, Fred Liu wrote:
 
 “zfs 'userused@' properties” and “'zfs userspace' command” are good enough 
 to gather usage statistics.
 I think I mix that with NetApp. If my memory is correct, we have to set 
 quotas to get usage statistics under DataOnTAP.
 Further, if we can add an ILM-like feature to poll the time-related 
 info(atime,mtime,ctime,etc) with that statistic from ZFS, that will be
 really cool.
  
 In general, file-based ILM has limitations that cause all sorts of issues 
 for things like
 operating systems, where files might only be needed infrequently, but when 
 they are
 needed, they are needed right now
  
 Have you looked at zfs diff  for changed files?
  
Here “ILM-like feature”, I mean we know how the data is 
 distributed by time per pool/filesystem like how many data are 
 modified/accessed
before mm/dd/. And we don’t need to do the actual storage-tying 
 operations immediately(moving the infrequently-used data to tie-2
storage). The time-related usage statistics are very useful reference 
 for us.
“zfs diff” will show the delta but not come with the time info.

The time is the creation time of the snapshots.
 -- richard

--
ZFS Performance and Training
richard.ell...@richardelling.com
+1-760-896-4422







___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] cluster vs nfs

2012-05-01 Thread Maurice R Volaski
Instead we've switched to Linux and DRBD.  And if that doesn't get me
sympathy I don't know what will.

SvSAN does something similar and it does it rather well, I think.
http://www.stormagic.com/SvSAN.php
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Gary Driggs
On May 1, 2012, at 1:41 AM, Ray Van Dolson wrote:

 Throughput:
iozone -m -t 8 -T -r 128k -o -s 36G -R -b bigfile.xls

 IOPS:
iozone -O -i 0 -i 1 -i 2 -e -+n -r 128K -s 288G  iops.txt

Do you expect to be reading or writing 36 or 288Gb files very often on
this array? The largest file size I've used in my still lengthy
benchmarks was 16Gb. If you use the sizes you've proposed, it could
take several days or weeks to complete. Try a web search for iozone
examples if you want more details on the command switches.

-Gary
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [developer] Setting default user/group quotas[usage accounting]?

2012-05-01 Thread Matthew Ahrens
2012/4/25 Richard Elling richard.ell...@gmail.com:
 On Apr 25, 2012, at 8:14 AM, Eric Schrock wrote:

 ZFS will always track per-user usage information even in the absence of
 quotas. See the the zfs 'userused@' properties and 'zfs userspace' command.


 tip: zfs get -H -o value -p userused@username filesystem

 Yes, and this is the logical size, not physical size. Some ZFS features
 increase logical size
 (copies) while others decrease physical size (compression, dedup)

The size accounted for by the userused@ and groupused@ properties is
the referenced space, which is used as the basis for many other
space accounting values in ZFS (e.g. du / ls -s / stat(2), and the
zfs accounting properties referenced, refquota, refreservation,
refcompressratio, written).  It includes changes local to the
dataset (compression, the copies property, file-specific metadata
such as indirect blocks), but ignores pool-wide or cross-dataset
changes (space shared between a clone and its origin, mirroring,
raid-z, dedup[*]).

--matt

[*] Although dedup can be turned on and off per-dataset, the data is
deduplicated against all dedup-enabled data in the pool.  Ie,
identical data in different datasets will be stored only once, if
dedup is enabled for both datasets.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Paul Kraus
On Mon, Apr 30, 2012 at 4:15 PM, Ray Van Dolson rvandol...@esri.com wrote:

 I'm trying to run some IOzone benchmarking on a new system to get a
 feel for baseline performance.

If you have compression turned on (and I highly recommend turning
it on if you have the CPU power to handle it), the IOzone data will be
flawed. I did not look deeper into it, but the data that IOzone uses
compresses very, very well. Much more so than any real data out there.
I used a combination of Filebench and Oracle's Orion to test ZFS
performance. Recently I started writing my own utilities for testing,
as _none_ of the existing offerings tested what I needed (lots and
lots of small, less than 64KB, files). My tool is only OK for relative
measures.

 Unfortunately, the system has a lot of memory (144GB), but I have some
 time so am approaching my runs as follows:

 When I was testing systems with more RAM than I wanted (when does
that ever happen :-), I called the ARC to something rational (2GB, 4GB
etc) and ran the tests with file sizes four times the ARC limit.
Unfortunately, the siwiki site appears to be down (gone ???).

On Solaris 10, the following in /etc/system (and a reboot) will cap
the zfa arc to the amount of RAM specified (in bytes). Not sure on
Nextena (and I have not had to cap the arc on my Nexenta Core system
at home).

set zfs:zfs_arc_max = 4294967296

 Throughput:
    iozone -m -t 8 -T -r 128k -o -s 36G -R -b bigfile.xls

 IOPS:
    iozone -O -i 0 -i 1 -i 2 -e -+n -r 128K -s 288G  iops.txt

 Not sure what I gain/lose by using threads or not.

IOzone without threads is single threaded and will demonstrate the
performance a single user or application will achieve. When you use
threads in IOzone you see performance for N simultaneous users (or
applications). In my experience, the knee in the performance vs. # of
threads curve happens somewhere between one and two times the number
of CPUs in the system. In other words, with a 16 CPU system,
performance scales linearly as the number of threads increases until
you get to somewhere between 16 and 32. At that point the performance
will start flattening out and eventually _decreases_ as you add more
threads.

 Using multiple threads (or processes or clients or etc.) is a
good way to measure how many simultaneous users your system can handle
(at a certain performance level).

 Am I off on this?

 System is a 240x2TB (7200RPM) system in 20 Dell MD1200 JBODs.  16 vdevs of 15
 disks each -- RAIDZ3.  NexentaStor 3.1.2.

-- 
{1-2-3-4-5-6-7-}
Paul Kraus
- Senior Systems Architect, Garnet River ( http://www.garnetriver.com/ )
- Assistant Technical Director, LoneStarCon 3 (http://lonestarcon3.org/)
- Sound Coordinator, Schenectady Light Opera Company (
http://www.sloctheater.org/ )
- Technical Advisor, Troy Civic Theatre Company
- Technical Advisor, RPI Players
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Bob Friesenhahn

On Mon, 30 Apr 2012, Ray Van Dolson wrote:


I'm trying to run some IOzone benchmarking on a new system to get a
feel for baseline performance.


Unfortunately, benchmarking with IOzone is a very poor indicator of 
what performance will be like during normal use.  Forcing the system 
to behave like it is short on memory only tests how the system will 
behave when it is short on memory.


Testing multi-threaded synchronous writes with IOzone might actually 
mean something if it is representative of your work-load.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [developer] current status of SAM-QFS?

2012-05-01 Thread Deepak Honnalli

5.3 is released, not much of an idea about the 5.4 schedule.

On Monday 30 April 2012 08:30 AM, Fred Liu wrote:

The subject says it all.


---
illumos-developer
Archives: https://www.listbox.com/member/archive/182179/=now
RSS Feed: https://www.listbox.com/member/archive/rss/182179/21175038-3914b2ca
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21175038id_secret=21175038-6fd90acc
Powered by Listbox: http://www.listbox.com



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Subscribe

2012-05-01 Thread james cypcar

Subscribe

--
ORACLE
James Cypcar | Solaris and Network Domain, Global Systems Support
Oracle Global Customer Services

Log, update, and monitor your Service Request online 
usinghttps://support.oracle.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Ray Van Dolson
On Tue, May 01, 2012 at 03:21:05AM -0700, Gary Driggs wrote:
 On May 1, 2012, at 1:41 AM, Ray Van Dolson wrote:
 
  Throughput:
 iozone -m -t 8 -T -r 128k -o -s 36G -R -b bigfile.xls
 
  IOPS:
 iozone -O -i 0 -i 1 -i 2 -e -+n -r 128K -s 288G  iops.txt
 
 Do you expect to be reading or writing 36 or 288Gb files very often on
 this array? The largest file size I've used in my still lengthy
 benchmarks was 16Gb. If you use the sizes you've proposed, it could
 take several days or weeks to complete. Try a web search for iozone
 examples if you want more details on the command switches.
 
 -Gary

The problem is this box has 144GB of memory.  If I go with a 16GB file
size (which I did), then memory and caching influences the results
pretty severely (I get around 3GB/sec for writes!).

Obviously, I could yank RAM for purposes of benchmarking. :)

Thanks,
Ray
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Ray Van Dolson
On Tue, May 01, 2012 at 07:18:18AM -0700, Bob Friesenhahn wrote:
 On Mon, 30 Apr 2012, Ray Van Dolson wrote:
 
  I'm trying to run some IOzone benchmarking on a new system to get a
  feel for baseline performance.
 
 Unfortunately, benchmarking with IOzone is a very poor indicator of 
 what performance will be like during normal use.  Forcing the system 
 to behave like it is short on memory only tests how the system will 
 behave when it is short on memory.
 
 Testing multi-threaded synchronous writes with IOzone might actually 
 mean something if it is representative of your work-load.
 
 Bob

Sounds like IOzone may not be my best option here (though it does
produce pretty graphs).

bonnie++ actually gave me more realistic sounding numbers, and I've
been reading good thigns about fio.

Ray
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Gary
On 5/1/12, Ray Van Dolson wrote:

 The problem is this box has 144GB of memory.  If I go with a 16GB file
 size (which I did), then memory and caching influences the results
 pretty severely (I get around 3GB/sec for writes!).

The idea of benchmarking -- IMHO -- is to vaguely attempt to reproduce
real world loads. Obviously, this is an imperfect science but if
you're going to be writing a lot of small files (e.g. NNTP or email
servers used to be a good real world example) then you're going to
want to benchmark for that. If you're going to want to write a bunch
of huge files (are you writing a lot of 16GB files?) then you'll want
to test for that. Caching anywhere in the pipeline is important for
benchmarks because you aren't going to turn off a cache or remove RAM
in production are you?

-Gary
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Paul Kraus
On Tue, May 1, 2012 at 1:45 PM, Gary gdri...@gmail.com wrote:

 The idea of benchmarking -- IMHO -- is to vaguely attempt to reproduce
 real world loads. Obviously, this is an imperfect science but if
 you're going to be writing a lot of small files (e.g. NNTP or email
 servers used to be a good real world example) then you're going to
 want to benchmark for that. If you're going to want to write a bunch
 of huge files (are you writing a lot of 16GB files?) then you'll want
 to test for that. Caching anywhere in the pipeline is important for
 benchmarks because you aren't going to turn off a cache or remove RAM
 in production are you?

It also depends on what you are going to be tuning. When I needed
to decided on a zpool configuration (# of vdev's, type of vdev, etc.)
I did not want the effect of the cache hiding the underlying
performance limitations of the physical drive configuration. In that
case I either needed to use a very large test data set or reduce the
size (effect) of the RAM. By limiting the ARC to 2 GB for my test, I
was able to relatively easily quantify the performance differences
between the various configurations. Once we picked a configuration, we
let the ARC take as much RAM as it wanted and re-ran the benchmark to
see what kind of real world performance we would get. Unfortunately,
we could not easily simulate 400 real world people sitting at desktops
accessing the data. So our ARC limited benchmark was effectively a
worst case number and the full ARC the best case. The real world,
as usual, fell somewhere in between.

   Finding a benchmark tool that matches _my_ work load is why I have
started kludging together my own.

-- 
{1-2-3-4-5-6-7-}
Paul Kraus
- Senior Systems Architect, Garnet River ( http://www.garnetriver.com/ )
- Sound Coordinator, Schenectady Light Opera Company (
http://www.sloctheater.org/ )
- Technical Advisor, RPI Players
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Bob Friesenhahn

On Tue, 1 May 2012, Ray Van Dolson wrote:


Testing multi-threaded synchronous writes with IOzone might actually
mean something if it is representative of your work-load.


Sounds like IOzone may not be my best option here (though it does
produce pretty graphs).

bonnie++ actually gave me more realistic sounding numbers, and I've
been reading good thigns about fio.


None of these benchmarks is really useful other than to stress-test 
your hardware.  Assuming that the hardware is working properly, when 
you intentionally break the cache, IOzone should produce numbers 
similar to what you could have estimated from hardware specification 
sheets and an understanding of the algorithms.


Sun engineers used 'filebench' to do most of their performance testing 
because it allowed configuring the behavior to emulate various usage 
models.  You can get it from 
https://sourceforge.net/projects/filebench/;.


Zfs is all about caching so the cache really does need to be included 
(and not intentionally broken) in any realistic measurement of how the 
system will behave.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] IOzone benchmarking

2012-05-01 Thread Richard Elling
more comments...

On May 1, 2012, at 10:41 AM, Ray Van Dolson wrote:

 On Tue, May 01, 2012 at 07:18:18AM -0700, Bob Friesenhahn wrote:
 On Mon, 30 Apr 2012, Ray Van Dolson wrote:
 
 I'm trying to run some IOzone benchmarking on a new system to get a
 feel for baseline performance.
 
 Unfortunately, benchmarking with IOzone is a very poor indicator of 
 what performance will be like during normal use.  Forcing the system 
 to behave like it is short on memory only tests how the system will 
 behave when it is short on memory.
 
 Testing multi-threaded synchronous writes with IOzone might actually 
 mean something if it is representative of your work-load.
 
 Bob
 
 Sounds like IOzone may not be my best option here (though it does
 produce pretty graphs).

For performance analysis of ZFS systems, you need to consider the advantages
of the hybrid storage pool. I wrote a white paper last summer describing a model
that you can use with your performance measurements or data from vendor 
datasheets. 

http://info.nexenta.com/rs/nexenta/images/tech_brief_nexenta_performance.pdf

And in presentation form, 
http://www.slideshare.net/relling/nexentastor-performance-tuning-openstorage-summit-2011

Recently, this model has been expanded and enhanced. Contact me offline, if you
are interested.

I have used IOzone, filebench, and vdbench for a lot of performance 
characterization
lately. Each has their own strength, but all can build a full characterization 
profile of a
system.

For IOzone, I like to run a full characterization run, which precludes 
multithreaded runs,
for a spectrum of I/O sizes and WSS. Such info can be useful to explore the 
boundaries
of your system's performance and compare to other systems. 

Also, for systems with  50GB of RAM, there are some tunables needed for good 
scaling 
under heavy write load workloads. Alas, there is no perfect answer and no 
single tunable
setting works optimally for all cases. WIP. YMMV.

A single, summary metric is not very useful...

 bonnie++ actually gave me more realistic sounding numbers, and I've
 been reading good thigns about fio.

IMNSHO, bonnie++ is a totally useless benchmark. Roch disected it rather nicely 
at
https://bigip-blogs-cms-adc.oracle.com/roch/entry/decoding_bonnie

[gag me!  Does Oracle have butugly URLs or what? ;-)]
 -- richard

--
ZFS Performance and Training
richard.ell...@richardelling.com
+1-760-896-4422







___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss