[zfs-discuss] Tunable parameter to zfs memory use

2006-08-16 Thread homerun
Hi

been using now zfs since 06/06 u2 release has been out.
one thing have notised.
zfs eats a lot of memory.
right after boot mem usage is about 280M
but after accessing zfs disks usage rises fast to be 900M.
and it seems to stay in level of 90% of tot mem.
also noted it frees used mem but running heavy apps you see performace impact 
while zfs frees memory to other use.

now idea:
could there be /etc/system tunable parameter for zfs so we could manually
set % of total memory zfs can use.

like :
zfs_max_phys_mem 30%
would tell zfs that it can use 30% of available memory.
and via versa
zfs_min_phys_mem 10%
would tell zfs that it can allways use 10% of available memory.
-- these maybe need to global values for all zfs filesystems 

or just simple zfs option that can be set as other zfs options , like 
compression, etc... then this setting could be easily per zfs filesystem

Thanks
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Tunable parameter to zfs memory use

2006-08-16 Thread James C. McPherson

homerun wrote:

been using now zfs since 06/06 u2 release has been out. one thing
have notised. zfs eats a lot of memory. right after boot mem usage is
about 280M but after accessing zfs disks usage rises fast to be 900M.
 and it seems to stay in level of 90% of tot mem. also noted it frees
used mem but running heavy apps you see performace impact while zfs
frees memory to other use.


Yes, that is the case. As I understand things it's an artefact of
how zfs is implemented. There are changes in the Solaris 10 patch
pipeline which should make the kmem usage somewhat better.


now idea: could there be /etc/system tunable parameter for zfs so we
could manually set % of total memory zfs can use.
like : zfs_max_phys_mem 30% would tell zfs that it can use 30% of
available memory. and via versa zfs_min_phys_mem 10% would tell zfs
that it can allways use 10% of available memory. -- these maybe need
to global values for all zfs filesystems
or just simple zfs option that can be set as other zfs options , like
compression, etc... then this setting could be easily per zfs
filesystem


I understand the concern, but I'm not sure that this is the
best way of achieving your goal.

What would you want to observe if your system hit the upper
limit in zfs_max_phys_mem?


cheers,
James C. McPherson



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Query on ZFS

2006-08-16 Thread James C. McPherson

Jaganraj Janarthanan wrote:

Customer had come back saying.
I've tried this and it doesn't work. It still displays the size as 1 gig after 
saying it's bringing the lun back online.
Eric Schrock wrote On 08/15/06 20:49,:

[ For ZFS discussions, try 'zfs-discuss@opensolaris.org' or
  '[EMAIL PROTECTED]' if there is proprietary
  information ]

ZFS will detect the expanded LUN when it re-opens the device.  There's
currently no way to automatically do this.  However, if you do a 'zpool
online pool device', this will have the side effect of reopening the
device and should pick up the expanded size.  If it doesn't, then please
let us know.

- Eric


Hi Jaganraj,
As I recall, one of the projects that's still on the table
within the relevant team is lun expansion. I don't recall
any email (prior to being RIFfed) saying that the project
has reached integration yet. Until it does, you'll probably
have to reboot :(


cheers,
James C. McPherson


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Apple Time Machine

2006-08-16 Thread Steve Hoelzer

On 8/7/06, Adam Leventhal [EMAIL PROTECTED] wrote:

Needless to say, this was a pretty interesting piece of the keynote from a
technical point of view that had quite a few of us scratching our heads.
After talking to some Apple engineers, it seems like what they're doing is
more or less this:

When a file is modified, the kernel fires off an event which a user-land
daemon listens for. Every so often, the user-land daemon does something
like a snapshot of the affected portions of the filesystem with hard links
(including hard links to directories -- I'm not making this up). That
might be a bit off, but it's the impression I was left with.

Anyhow, very slick UI, sort of dubious back end, interesting possibility
for integration with ZFS.


You were spot on with your description and conclusion. More details here:
http://arstechnica.com/staff/fatbits.ars/2006/8/15/4995

Steve
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS Boot Disk

2006-08-16 Thread Joerg Schilling
Dick Davies [EMAIL PROTECTED] wrote:

 As an aside, is there a general method to generate bootable
 opensolaris DVDs? The only way I know of getting opensolaris on
 is installing sxcr and then BFUing on top.

A year ago, I did publish a toolkit to create bootable SchilliX CDs/DVDs.
Would this help?

Jörg

-- 
 EMail:[EMAIL PROTECTED] (home) Jörg Schilling D-13353 Berlin
   [EMAIL PROTECTED](uni)  
   [EMAIL PROTECTED] (work) Blog: http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/old/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import - cannot mount [...] directory is not empty

2006-08-16 Thread Mark Maybee

Robert,

Are you sure that nfs-s5-p0/d5110 and nfs-s5-p0/d5111 are mounted
following the import?  These messages imply that the d5110 and d5111
directories in the top-level filesystem of pool nfs-s5-p0 are not
empty.  Could you verify that 'df /nfs-s5-p0/d5110' displays
nfs-s5-p0/d5110 as the Filesystem (and not just nfs-s5-p0)?

-Mark

Robert Milkowski wrote:

All pools were exported than I tried to import one-by-one and got this with 
only a first pool.

bash-3.00# zpool export nfs-s5-p4 nfs-s5-s5 nfs-s5-s6 nfs-s5-s7 nfs-s5-s8
bash-3.00# zpool import nfs-s5-p4
cannot mount '/nfs-s5-p4/d5139': directory is not empty
cannot mount '/nfs-s5-p4/d5141': directory is not empty
cannot mount '/nfs-s5-p4/d5138': directory is not empty
cannot mount '/nfs-s5-p4/d5142': directory is not empty
bash-3.00# df -h /nfs-s5-p4/d5139
Filesystem size   used  avail capacity  Mounted on
nfs-s5-p4/d5139600G   556G44G93%/nfs-s5-p4/d5139
bash-3.00# zpool export nfs-s5-p4
bash-3.00# ls -l /nfs-s5-p4/d5139
/nfs-s5-p4/d5139: No such file or directory
bash-3.00# ls -l /nfs-s5-p4/
total 0
bash-3.00# zpool import nfs-s5-p4
bash-3.00# uname -a
SunOS XXX 5.11 snv_43 sun4u sparc SUNW,Sun-Fire-V240
bash-3.00#

No problem with other pools - all other pools imported without any warnings.

The same on another server (all pools were exported first):

bash-3.00# zpool import nfs-s5-p0
cannot mount '/nfs-s5-p0/d5110': directory is not empty
use legacy mountpoint to allow this behavior, or use the -O flag
cannot mount 'nfs-s5-p0/d5112': mountpoint or dataset is busy
cannot mount '/nfs-s5-p0/d5111': directory is not empty
use legacy mountpoint to allow this behavior, or use the -O flag
bash-3.00# zpool export nfs-s5-p0
bash-3.00# zpool import nfs-s5-p0
cannot mount '/nfs-s5-p0/d5110': directory is not empty
use legacy mountpoint to allow this behavior, or use the -O flag
cannot mount '/nfs-s5-p0/d5111': directory is not empty
use legacy mountpoint to allow this behavior, or use the -O flag
bash-3.00# zpool export nfs-s5-p0
bash-3.00# ls -la /nfs-s5-p0/
total 4
drwxr-xr-x   2 root other512 Jun 14 14:37 .
drwxr-xr-x  40 root root1024 Aug  8 11:00 ..
bash-3.00# zpool import nfs-s5-p0
cannot mount '/nfs-s5-p0/d5110': directory is not empty
use legacy mountpoint to allow this behavior, or use the -O flag
cannot mount 'nfs-s5-p0/d5112': mountpoint or dataset is busy
cannot mount '/nfs-s5-p0/d5111': directory is not empty
use legacy mountpoint to allow this behavior, or use the -O flag
bash-3.00#
bash-3.00# uname -a
SunOS X 5.11 snv_39 sun4v sparc SUNW,Sun-Fire-T200
bash-3.00#

All filesystems from that pool were however mounted.

No problem with other pools - all other pools imported without any warnings.


All filesystems in a pool have set sharenfs (actually sharenfs is set on a pool 
and then inherited by filesystems). Additionally nfs/server was disabled just 
before I exported pools and automatically started when first pool was imported.



I belive there's already open bug for this.
 
 
This message posted from opensolaris.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re[2]: [zfs-discuss] zpool import - cannot mount [...] directory is not empty

2006-08-16 Thread Robert Milkowski
Hello Mark,

Wednesday, August 16, 2006, 3:23:43 PM, you wrote:

MM Robert,

MM Are you sure that nfs-s5-p0/d5110 and nfs-s5-p0/d5111 are mounted
MM following the import?  These messages imply that the d5110 and d5111
MM directories in the top-level filesystem of pool nfs-s5-p0 are not
MM empty.  Could you verify that 'df /nfs-s5-p0/d5110' displays
MM nfs-s5-p0/d5110 as the Filesystem (and not just nfs-s5-p0)?

They are definitely mounted and not empty. And df reports them as
mounted file systems.

-- 
Best regards,
 Robertmailto:[EMAIL PROTECTED]
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import - cannot mount [...] directory is not empty

2006-08-16 Thread Eric Schrock
What does 'zfs list -o name,mountpoint' and 'zfs mount' show after the
import?  My only guess is that you have some explicit mountpoint set
that's confusing the DSl-orderered mounting code.  If this is the case,
this was fixed in build 46 (likely to be in S10u4) to always mount
datasets in mountpoint order, regardless of their hierarchical
relationship.  However, you may be hitting some other bug that we
haven't seen before.

- Eric

On Wed, Aug 16, 2006 at 03:32:08PM +0200, Robert Milkowski wrote:
 Hello Mark,
 
 Wednesday, August 16, 2006, 3:23:43 PM, you wrote:
 
 MM Robert,
 
 MM Are you sure that nfs-s5-p0/d5110 and nfs-s5-p0/d5111 are mounted
 MM following the import?  These messages imply that the d5110 and d5111
 MM directories in the top-level filesystem of pool nfs-s5-p0 are not
 MM empty.  Could you verify that 'df /nfs-s5-p0/d5110' displays
 MM nfs-s5-p0/d5110 as the Filesystem (and not just nfs-s5-p0)?
 
 They are definitely mounted and not empty. And df reports them as
 mounted file systems.
 
 -- 
 Best regards,
  Robertmailto:[EMAIL PROTECTED]
http://milek.blogspot.com
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool import/export

2006-08-16 Thread Eric Schrock
This seems like a reasonable RFE.  Feel free to file it at
bugs.opensolaris.org.

- Eric

On Wed, Aug 16, 2006 at 06:44:44AM +0200, Robert Milkowski wrote:
 Hello zfs-discuss,
 
   I do have several pools in a SAN shared environment where some pools
   are mounted by one server and some by another.
 
   Now I can do 'zpool export A B C D'
   But I can't do 'zpool import A B C D'
 
   import -a isn't an option since I want import only those pools I
   just exported.
 
   It's not a big deal but it's annoying I can't import these pool with
   one command line like I can export them. Ideally I should just go to
   shell history and change export to import and voila!
 
 -- 
 Best regards,
  Robert  mailto:[EMAIL PROTECTED]
  http://milek.blogspot.com
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

--
Eric Schrock, Solaris Kernel Development   http://blogs.sun.com/eschrock
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Information on ZFS API?

2006-08-16 Thread William Fretts-Saxton
I'm having trouble finding information on any hooks into ZFS.  Is there 
information on a ZFS API so I can access ZFS information directly as opposed to 
having to constantly parse 'zpool' and 'zfs' command output?
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Information on ZFS API?

2006-08-16 Thread Bill Sommerfeld
On Wed, 2006-08-16 at 11:49 -0400, Eric Enright wrote:
 On 8/16/06, William Fretts-Saxton [EMAIL PROTECTED] wrote:
  I'm having trouble finding information on any hooks into ZFS.  Is 
  there information on a ZFS API so I can access ZFS information 
  directly as opposed to having to constantly parse 'zpool' and 'zfs' 
  command output?
 
 libzfs: http://cvs.opensolaris.org/source/xref/on/usr/src/lib/libzfs/common/

Careful -- this isn't an API -- it's part of the implementation of ZFS,
and as a result it is subject to incompatible change at any time without
advanced warning.

In terms of our Interface Taxonomy, this is considered Project
Private.  

If you have specific requirements for a programmatic interface to ZFS,
let's hear them...  it's not uncommon for Private interfaces to get put
through some renovation and housecleaning and then promoted to a public
interface, but it's best for the conversation to happen first...

(And I can't speak for the plans/intentions of the core zfs team on
this..)

- Bill






___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Information on ZFS API?

2006-08-16 Thread William Fretts-Saxton
This is my first post on the opensolaris.org forums.  I'm still trying 
to figure out whether I'm actually on the zfs-discuss alias and, if not, 
how to add myself to it!


Anyway, I'm looking for a way to get information like zpool list or 
zfs get all filesystem through an API instead of parsing the 
information from the CLI.


In the ideal world, I could just get a zpool object from the OS, which 
contained the zpool information and various zfs objects and do whatever 
I wanted with them.


Bill Sommerfeld wrote:

On Wed, 2006-08-16 at 11:49 -0400, Eric Enright wrote:

On 8/16/06, William Fretts-Saxton [EMAIL PROTECTED] wrote:
I'm having trouble finding information on any hooks into ZFS.  Is 
there information on a ZFS API so I can access ZFS information 
directly as opposed to having to constantly parse 'zpool' and 'zfs' 
command output?

libzfs: http://cvs.opensolaris.org/source/xref/on/usr/src/lib/libzfs/common/


Careful -- this isn't an API -- it's part of the implementation of ZFS,
and as a result it is subject to incompatible change at any time without
advanced warning.

In terms of our Interface Taxonomy, this is considered Project
Private.  


If you have specific requirements for a programmatic interface to ZFS,
let's hear them...  it's not uncommon for Private interfaces to get put
through some renovation and housecleaning and then promoted to a public
interface, but it's best for the conversation to happen first...

(And I can't speak for the plans/intentions of the core zfs team on
this..)

- Bill







___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: Re: [zfs-discuss] multi-layer ZFS filesystems and exporting: my stupid question for the day

2006-08-16 Thread Joe Little

On 8/16/06, Frank Cusack [EMAIL PROTECTED] wrote:

On August 16, 2006 10:25:18 AM -0700 Joe Little [EMAIL PROTECTED] wrote:
 Is there a way to allow simple export commands the traverse multiple
 ZFS filesystems for exporting? I'd hate to have to have hundreds of
 mounts required for every point in a given tree (we have users,
 projects, src, etc)

Set the sharenfs property on the filesystems and use the automounter
on the client.



Damn. We are hoping to move away from automounters and maintenance of
such (we use NeoPath, for example, for virtual aggregation of the
paths). In the NAS world, sometimes automounts are not available. So,
if this is true, you'll likely need to count me as one of those people
who says we can't make a filesystem per user, and give us user quotas
now please :)



-frank




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] neopath vs automount

2006-08-16 Thread Frank Cusack

Sorry, I'm an email deleter, not a hoarder, so this is a new thread.
Usually I save a thread I'm interested in for awhile before killing it,
but I jumped the gun this time.  Anyway ...

I looked up neopath, cool product!

But ISTM that to use it would give up some zfs features like snapshots.
I would think that for home dirs, you can easily get away with using
the automounter and you wouldn't give up a lot vs neopath functionality.

-frank
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: Re: Re: [zfs-discuss] multi-layer ZFS filesystems and exporting: my stupid question for the day

2006-08-16 Thread Joe Little

On 8/16/06, Frank Cusack [EMAIL PROTECTED] wrote:

On August 16, 2006 10:34:31 AM -0700 Joe Little [EMAIL PROTECTED] wrote:
 On 8/16/06, Frank Cusack [EMAIL PROTECTED] wrote:
 On August 16, 2006 10:25:18 AM -0700 Joe Little [EMAIL PROTECTED] wrote:
  Is there a way to allow simple export commands the traverse multiple
  ZFS filesystems for exporting? I'd hate to have to have hundreds of
  mounts required for every point in a given tree (we have users,
  projects, src, etc)

 Set the sharenfs property on the filesystems and use the automounter
 on the client.


 Damn. We are hoping to move away from automounters and maintenance of
 such (we use NeoPath, for example, for virtual aggregation of the
 paths).

I don't know what NeoPath is but automounts are trivial or at least
easy to maintain even for very large sites.  You are using path wildcards
and DNS aliases, yes?


used to.. running away quickly. For different *nix flavors, pick one
of autofs, automountd, etc.. Nohide support, no nohide support,
netgroups, LDAP, it all gets ugly pretty quickly. We want to move to
statically definied mount trees similar to AFS managed centrally for
all using NFS as the common protocol (aka the NeoPath)



 In the NAS world, sometimes automounts are not available. So,
 if this is true, you'll likely need to count me as one of those people
 who says we can't make a filesystem per user, and give us user quotas
 now please :)

I don't understand.  If an automount is not available, how is that
different than the nfs server itself not being available.  Or do you
mean some clients do not have an automounter.


Some clients don't have automounters available. Also some servers:
Example is gateways/proxys (think SMB proxy w/o automounter support).

Other clients in heavy use require lots of changes to get what you
propose, such as OSX's automounter (storing the results in NetInfo)



-frank


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] ZFS write performance problem with compression set to ON

2006-08-16 Thread Anantha N. Srirama
Test setup:
  - E2900 with 12 US-IV+ 1.5GHz processor, 96GB memory, 2x2Gbps FC HBAs, MPxIO 
in round-robbin config.
  - 50x64GB EMC disks presented on both 2 FCs.
  - ZFS pool defined using all 50 disks
  - Multiple ZFS filesystems built on the above pool.

I'm observing the following:
  - When the filesystems have compress=OFF and I do bulk reads/writes (8 
parallel 'cp's running between ZFS filesystems) I observe approximately 
200-250MB/S consolidated I/O; writes in the 100MB/S range. I get these numbers 
running 'zpool iostat 5'. I see the same read/write ratio for the duration of 
the test.
  - When the filesystems have compress=ON I see the following: reads from 
compressed filesystems come in waves; zpool will report for long durations (60+ 
seconds) no read activity while the write activity is consistently reported at 
20MB/S (no variation in the write rate throughtout the test.)
  - The machine is mostly idling during the entire test; both cases.
  - ZFS reports 4:1 compresson ratio for my filesystem.

I'm puzzled by the following:
  - Why do reads comes in waves with compression=ON; it almost feels like ZFS 
reads a bunch of data and then proceeds to compress it before writing it out. 
This tells me there is not a read bottleneck; meaning there is no starvation of 
the compress routine due to the following facts: CPU/Machine/IO is not 
saturated in any shape or form.
  - Why then does ZFS generate substantially lower write throughput (magical 
20MB/S spread evenly across the 50 disks, 0.4MB/S each)?

Can anybody shed any light on this anomoloy (?). Mr. Bonwick I hope you're 
reading this post.

BTW, we love the ZFS and are looking forward to rolling out aggresively in our 
new project. I'd like to take advantage of the compression since we're mostly 
I/O bound and we've plenty of CPU/Memory.

Thanks.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: system unresponsive after issuing a zpool attach

2006-08-16 Thread Rustam
I have similar behaviour on S10 U2 but in a different situation.

I had working mirror with one of mirrors failed:

mirror DEGRADED 0 0 0
c0d0 ONLINE 0 0 0
c0d1 UNAVAILABLE 0 0 0

After replacing corrupted hard disk i've run:
# zpool replace tank c0d1

And it started replacement/resilvering... after few minutes system became 
unavailbale. Reboot only gives me a few minutes, then resilvering make system 
unresponsible.

Is there any workaroud or patch for this problem???
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss