Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0

2012-01-05 Thread Peter Radig
It's supposed to be

7111576: arc shrinks in the absence of memory pressure

currently in status "accepted" and an RPE escalation pending.

-Original Message-
From: zfs-discuss-boun...@opensolaris.org 
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Tomas Forsman
Sent: Donnerstag, 5. Januar 2012 10:35
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0

On 04 January, 2012 - Steve Gonczi sent me these 2,5K bytes:

> The interesting bit is what happens inside arc_reclaim_needed(), that 
> is, how it arrives at the conclusion that there is memory pressure.
> 
> Maybe we could trace arg0, which gives the location where we have left 
> the function. This would finger which return path
> arc_reclaim_needed() took. 

It's new code, basically comparing the inuse/total/free from kstat -n 
zfs_file_data, which seems buggered.

> Steve
> 
> 
> - Original Message -
> 
> 
> 
> Well it looks like the only place this get's changed is in the 
> arc_reclaim_thread for opensolaris. I suppose you could dtrace it to see what 
> is going on and investigate what is happening to the return code of the 
> arc_reclaim_needed is. 
> 
> 
> 
> http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/comm
> on/fs/zfs/arc.c#2089
> 
> 
> maybe
> 
> 
> dtrace -n 'fbt:zfs:arc_reclaim_needed:return { trace(arg1) }' 
> 
> 
> Dave
> 
> 
> 
> 
> 


/Tomas
--
Tomas Forsman, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0

2012-01-04 Thread Peter Radig
Thanks. The guys from Oracle are currently looking at some new code that was 
introduced  in arc_reclaim_thread() between b151a and b175.

Peter Radig, Ahornstrasse 34, 85774 Unterföhring, Germany
tel: +49 89 99536751 - fax: +49 89 99536754 - mobile: +49 171 2652977
email: pe...@radig.de<mailto:pe...@radig.de>

From: David Blasingame [mailto:dbla...@yahoo.com]
Sent: Mittwoch, 4. Januar 2012 17:35
To: Peter Radig; st...@acc.umu.se
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0


Well it looks like the only place this get's changed is in the 
arc_reclaim_thread for opensolaris.  I suppose you could dtrace it to see what 
is going on and investigate what is happening to the return code of the 
arc_reclaim_needed is.

http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/fs/zfs/arc.c#2089

maybe

dtrace -n 'fbt:zfs:arc_reclaim_needed:return { trace(arg1) }'

Dave


 Original Message 
Subject:

Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0

Date:

Tue, 3 Jan 2012 23:50:04 +0000

From:

Peter Radig <mailto:pe...@radig.de>

To:

Tomas Forsman <mailto:st...@acc.umu.se>

CC:

zfs-discuss@opensolaris.org<mailto:zfs-discuss@opensolaris.org> 
<mailto:zfs-discuss@opensolaris.org>



Tomas,



Yup, same here.



free0

mem_inuse   16162750464

mem_total   17171480576



Page SummaryPagesMB  %Tot

     

Kernel 842305  3290   20%

ZFS File Data2930110%

Anon44038   1721%

Exec and libs3731140%

Page cache   8580330%

Free (cachelist) 5504210%

Free (freelist)   3284924 12831   78%



Total 4192012 16375

Physical  4192011 16375





I will create an SR with Oracle.



Thanks,

Peter



-Original Message-

From: Tomas Forsman [mailto:st...@acc.umu.se]

Sent: Mittwoch, 4. Januar 2012 00:39

To: Peter Radig

Cc: zfs-discuss@opensolaris.org<mailto:zfs-discuss@opensolaris.org>

Subject: Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0



On 03 January, 2012 - Peter Radig sent me these 3,5K bytes:



> Hello.

>

> I have a Solaris 11/11 x86 box (which I migrated from SolEx 11/10 a couple of 
> weeks ago).

>

> Without no obvious reason (at least for me), after an uptime of 1 to 2 days 
> (observed 3 times now) Solaris sets arc_no_grow to 1 and then never sets it 
> back to 0. ARC is being shrunk to less than 1 GB -- needless to say that 
> performance is terrible. There is not much load on this system.

>

> Memory seems to be not an issue (see below).

>

> I looked at the old Nevada code base of onnv_147 and can't find a reason for 
> this happening.

>

> How can I find out what's causing this?



New code that seems to be counting wrong.. I was planning on filing a bug, but 
am currently struggling to convince oracle that we bought support..



Try this:

kstat -n zfs_file_data



In my case, I get:

free15322

mem_inuse   24324866048

mem_total   25753026560

.. where ::memstat says:

Kernel2638984 10308   42%

ZFS File Data   39260   1531%

Anon   873549  3412   14%

Exec and libs5199200%

Page cache  20019780%

Free (cachelist) 6608250%

Free (freelist)   2703509 10560   43%



On another reboot, it refused to go over 130MB on a 24GB system..





> BTW: I was not seeing this on SolEx 11/10.



Dito.



> Thanks,

> Peter

>

>

>

> *** ::memstat ***

> Page SummaryPagesMB  %Tot

>      

> Kernel 860254  3360   21%

> ZFS File Data3047110%

> Anon38246   1491%

> Exec and libs3765140%

> Page cache   8517330%

> Free (cachelist) 5866220%

> Free (freelist)   3272317 12782   78%

> Total 4192012 16375

> Physical  4192011 16375

>

> mem_inuse   

Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0

2012-01-03 Thread Peter Radig
Tomas,

Yup, same here.

free0
mem_inuse   16162750464
mem_total   17171480576

Page SummaryPagesMB  %Tot
     
Kernel 842305  3290   20%
ZFS File Data2930110%
Anon44038   1721%
Exec and libs3731140%
Page cache   8580330%
Free (cachelist) 5504210%
Free (freelist)   3284924 12831   78%

Total 4192012 16375
Physical  4192011 16375


I will create an SR with Oracle.

Thanks,
Peter

-Original Message-
From: Tomas Forsman [mailto:st...@acc.umu.se] 
Sent: Mittwoch, 4. Januar 2012 00:39
To: Peter Radig
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] arc_no_grow is set to 1 and never set back to 0

On 03 January, 2012 - Peter Radig sent me these 3,5K bytes:

> Hello.
> 
> I have a Solaris 11/11 x86 box (which I migrated from SolEx 11/10 a couple of 
> weeks ago).
> 
> Without no obvious reason (at least for me), after an uptime of 1 to 2 days 
> (observed 3 times now) Solaris sets arc_no_grow to 1 and then never sets it 
> back to 0. ARC is being shrunk to less than 1 GB -- needless to say that 
> performance is terrible. There is not much load on this system.
> 
> Memory seems to be not an issue (see below).
> 
> I looked at the old Nevada code base of onnv_147 and can't find a reason for 
> this happening.
> 
> How can I find out what's causing this?

New code that seems to be counting wrong.. I was planning on filing a bug, but 
am currently struggling to convince oracle that we bought support..

Try this:
kstat -n zfs_file_data

In my case, I get:
free15322
mem_inuse   24324866048
mem_total   25753026560
.. where ::memstat says:
Kernel2638984 10308   42%
ZFS File Data   39260   1531%
Anon   873549  3412   14%
Exec and libs5199200%
Page cache  20019780%
Free (cachelist) 6608250%
Free (freelist)   2703509 10560   43%

On another reboot, it refused to go over 130MB on a 24GB system..


> BTW: I was not seeing this on SolEx 11/10.

Dito.

> Thanks,
> Peter
> 
> 
> 
> *** ::memstat ***
> Page SummaryPagesMB  %Tot
>      
> Kernel 860254  3360   21%
> ZFS File Data3047110%
> Anon38246   1491%
> Exec and libs3765140%
> Page cache   8517330%
> Free (cachelist) 5866220%
> Free (freelist)   3272317 12782   78%
> Total 4192012 16375
> Physical  4192011 16375
> 
> mem_inuse   4145901568
> mem_total   1077466365952
> 
> *** ::arc ***
> hits  = 186279921
> misses=  14366462
> demand_data_hits  =   4648464
> demand_data_misses=   8605873
> demand_metadata_hits  = 171803126
> demand_metadata_misses=   3805675
> prefetch_data_hits=772678
> prefetch_data_misses  =   1464457
> prefetch_metadata_hits=   9055653
> prefetch_metadata_misses  =490457
> mru_hits  =  12295087
> mru_ghost_hits= 0
> mfu_hits  = 175281066
> mfu_ghost_hits= 0
> deleted   =  14462192
> mutex_miss=30
> hash_elements =   3752768
> hash_elements_max =   3752770
> hash_collisions   =  11409790
> hash_chains   =  8256
> hash_chain_max=20
> p =48 MB
> c =   781 MB
> c_min =64 MB
> c_max = 15351 MB
> size  =   788 MB
> buf_size  =   185 MB
> data_size =   289 MB
> other_size=   313 MB
> l2_hits   = 0
> l2_misses =  14366462
> l2_feeds

[zfs-discuss] arc_no_grow is set to 1 and never set back to 0

2012-01-03 Thread Peter Radig
Hello.

I have a Solaris 11/11 x86 box (which I migrated from SolEx 11/10 a couple of 
weeks ago).

Without no obvious reason (at least for me), after an uptime of 1 to 2 days 
(observed 3 times now) Solaris sets arc_no_grow to 1 and then never sets it 
back to 0. ARC is being shrunk to less than 1 GB -- needless to say that 
performance is terrible. There is not much load on this system.

Memory seems to be not an issue (see below).

I looked at the old Nevada code base of onnv_147 and can't find a reason for 
this happening.

How can I find out what's causing this?

BTW: I was not seeing this on SolEx 11/10.

Thanks,
Peter



*** ::memstat ***
Page SummaryPagesMB  %Tot
     
Kernel 860254  3360   21%
ZFS File Data3047110%
Anon38246   1491%
Exec and libs3765140%
Page cache   8517330%
Free (cachelist) 5866220%
Free (freelist)   3272317 12782   78%
Total 4192012 16375
Physical  4192011 16375

mem_inuse   4145901568
mem_total   1077466365952

*** ::arc ***
hits  = 186279921
misses=  14366462
demand_data_hits  =   4648464
demand_data_misses=   8605873
demand_metadata_hits  = 171803126
demand_metadata_misses=   3805675
prefetch_data_hits=772678
prefetch_data_misses  =   1464457
prefetch_metadata_hits=   9055653
prefetch_metadata_misses  =490457
mru_hits  =  12295087
mru_ghost_hits= 0
mfu_hits  = 175281066
mfu_ghost_hits= 0
deleted   =  14462192
mutex_miss=30
hash_elements =   3752768
hash_elements_max =   3752770
hash_collisions   =  11409790
hash_chains   =  8256
hash_chain_max=20
p =48 MB
c =   781 MB
c_min =64 MB
c_max = 15351 MB
size  =   788 MB
buf_size  =   185 MB
data_size =   289 MB
other_size=   313 MB
l2_hits   = 0
l2_misses =  14366462
l2_feeds  = 0
l2_rw_clash   = 0
l2_read_bytes = 0 MB
l2_write_bytes= 0 MB
l2_writes_sent= 0
l2_writes_done= 0
l2_writes_error   = 0
l2_writes_hdr_miss= 0
l2_evict_lock_retry   = 0
l2_evict_reading  = 0
l2_abort_lowmem   =   490
l2_cksum_bad  = 0
l2_io_error   = 0
l2_hdr_size   = 0 MB
memory_throttle_count = 0
meta_used =   499 MB
meta_max  =  1154 MB
meta_limit= 0 MB
arc_no_grow   = 1
arc_tempreserve   = 0 MB
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Snapshot size as reported by the USED property

2010-08-30 Thread Peter Radig
I create snapshots on my datasets quite frequently. My understanding of the 
USED property of a snapshot is that it indicates the amount of data that was 
written to the dataset after the snapshot was taken. But now I'm seeing a 
snapshot with USED == 0 where there was definitely write activity after it was 
taken.

Is my understanding wrong or do I see a thing that is not supposed to happen? I 
am on NCP3.

Thanks,
Peter
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Sharing Issues

2010-02-22 Thread Peter Radig
As I explained earlier, this is not possible with CIFS. This is the RFE entry: 
http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6582165

And the explanation is here: 
http://mail.opensolaris.org/pipermail/cifs-discuss/2009-March/001397.html

Peter

-Original Message-
From: zfs-discuss-boun...@opensolaris.org 
[mailto:zfs-discuss-boun...@opensolaris.org] On Behalf Of Tau
Sent: 22. February 2010 22:40
To: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] Sharing Issues

It will not let me set sharesmb=none.

Also I dont see how ACL's would do what im after here...  I want to create a 
nested dataset inside another one so that I can create snapshots, and export 
that dataset when needed.  Though I do not want the nested dataset to have its 
own share... (if that makes sence)

Scenario;

tank1/websites/website1
  /website2
  /website3

Now when i create website1-3 datasets inside of the websites dataset it shares 
them as well.  So from my windows/nix boxes i see;

websites
websites_website1
websites_website2
websites_website3

now i JUST want to see "websites"  then be able to access the website1-3 
datasets from navigating through the websites share.

The way it is sharing right now im going to end up with 100's of shares were i 
only really need to have 6 or 8, and the rest accessible through the root 
datasets.

I hope this is a better explanation.
--
This message posted from opensolaris.org 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Sharing Issues

2010-02-21 Thread Peter Radig
It doesn't work with CIFS. There is an open RFE on that for quite some  
time now.

Peter

On 22.02.2010, at 08:09, "Tau"  wrote:

> I am having a bit of an issue I have an opensolaris box setup as a  
> fileserver.  Running through CIFS to provide shares to some windows  
> machines.
>
> Now lets call my zpool /tank1, when i create a zfs filesystem  
> called /test it gets shared as /test and i can see it as "test" on  
> my windows machines...  Now when i create a child system inside the  
> test system (lets call this /tank1/test/child) the child system gets  
> shared as well on its own as test_child as seen on the windows system.
>
> I want to be able to create  nested filesystems, and not have the  
> nested systems shared through cifs  i want to access them  
> through the root system, and only have the root systems shared to  
> the windows machines...
>
> I have been trolling through the manuals, and forums, but cant seem  
> to find the ansear.
>
> I'm sure im missing something simple, can someone shed some light  
> onto this issue?
> -- 
> This message posted from opensolaris.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] File created with CIFS is not immediately deletable on local file system

2010-02-21 Thread Peter Radig
Box running osol_133 with smb/server enabled. I create a file on a Windows box 
that has a remote ZFS fs mounted. I go to the Solaris box and try to remove the 
file and get "permission denied" for up 30 sec. Than it works. A "sync" 
immediately before the rm seems to speed things up and rm is successful 
immediately afterwards, too. The permissions of the file don't change in 
between and look as follow:
--+  1 petersysprog0 Feb 21 11:37 NEWFILE3.txt
 user:peter:rwxpdDaARWcCos:---:allow
   group:2147483648:rwxpdDaARWcCos:---:allow

Is this a bug or a feature?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Impact of an enterprise class SSD on ZIL performance

2010-02-04 Thread Peter Radig
I was interested in the impact the type of an SSD has on the performance of the 
ZIL. So I did some benchmarking and just want to share the results.

My test case is simply untarring the latest ON source (528 MB, 53k files) on an 
Linux system that has a ZFS file system mounted via NFS over gigabit ethernet.

I got the following results:
- locally on the Solaris box: 30 sec
- remotely with no dedicated ZIL device: 36 min 37 sec (factor 73 compared to 
local)
- remotely with ZIL disabled: 1 min 54 sec (factor 3.8 compared to local)
- remotely with a OCZ VERTEX SATA II 120 GB as ZIL device: 14 min 40 sec 
(factor 29.3 compared to local)
- remotely with an Intel X25-E 32 GB as ZIL device: 3 min 11 sec (factor 6.4 
compared to local)

So it really makes a difference what type of SSD you use for your ZIL device. I 
was expecting a good performance from the X25-E, but was really suprised that 
it is that good (only 1.7 times slower than it takes with ZIL completely 
disabled). So I will use the X25-E as ZIL device on my box and will not 
consider disabling ZIL at all to improve NFS performance.

-- Peter
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss