[zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-18 Thread Rainer Heilke
  This problem was fixed in snv_48 last September
  and will be
  in S10_U4.

U4 doesn't help us any. We need the fix now. :-(  By the time U4 is out, we may 
even be finished (certainly well on our way) our RAC/ASM migration and this 
whole issue will be moot.

Rainer
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-18 Thread Rainer Heilke
Thanks for the detailed explanation of the bug. This makes it clearer to us as 
to what's happening, and why (which is something I _always_ appreciate!). 
Unfortunately, U4 doesn't buy us anything for our current problem.

Rainer
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-18 Thread Bev Crair

Rainer,
Have you considered looking for a patch?  If you have the supported 
version(s) of Solaris (which it sound like you do), this may already be 
available in a patch.

Bev.

Rainer Heilke wrote:

Thanks for the detailed explanation of the bug. This makes it clearer to us as 
to what's happening, and why (which is something I _always_ appreciate!). 
Unfortunately, U4 doesn't buy us anything for our current problem.

Rainer
 
 
This message posted from opensolaris.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
  

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-18 Thread Rainer Heilke
Sorry, I should have qualified that effective better. I was specifically 
speaking in terms of Solaris and price. For companies without a SAN (especially 
using Linux), something like a NetApp Filer using UFS is the way to go, I 
realize. If you're running Solaris, the cost of QFS becomes a major factor. If 
you have a SAN, then getting a NetApp Filer seems silly. And so on.

Oracle has suggested RAW disk for some time, I think. (Some?) DBA's don't seem 
to like it largely because they cannot see the files, and so on. ASM still has 
some of these limitations, but it's getting better, and DBA's are starting to 
get used to the new paradigms. If I remember a conversation last year 
correctly, OEM will become the window into some of these ideas. Once ASM has 
industry acceptance on a large scale, then yes, making file systems perform 
well especially for Oracle databases will be chasing the wind. But, that may be 
a while down the road. I don't know, my crystal ball got cracked during the 
last comet transition.  ;-)

Rainer
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-18 Thread Rainer Heilke
Rats, didn't proof accurately. For UFS, I meant NFS.

Rainer
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-17 Thread Rainer Heilke
 The limit is documented as 1 million inodes per TB.
  So something
 ust not have gone right.  But many people have
 complained and
 you could take the newfs source and fix the
 limitation.

Patching the source ourselves would not fly very far, but thanks for the 
clarification. I guess I have to assume, then, that somewhere around this 
million mark we also ran out of inodes. With the wide range in file sizes for 
the files, this doesn't surprise me. There was no way to tune the file system 
for anything.

 The discontinuity when going from 1TB to over 1TB is
 appaling.
 (1TB allows for 137million inodes; = 1TB allows for
 1million per).

Either way, we were stuck. Our test/devl environment goes way beyond 1 million 
files (read: inodes). I think we hit the ceiling half-way into our data copy, 
if memory serves.

I think the argument I saw for this inode disparity was that a 1TB FS was 
only for database files and not the binaries, or something to that effect.

 The rationale is fsck time (but logging is forced
 anyway)

I can't remember for sure, but this might have been mentioned in one of the 
notes I found.

 The 1 million limit is arbitrary and too low...
 
 Casper

Thank you very much for the clarification, and for the candor. It is greatly 
appreciated.

Rainer
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-17 Thread Anton B. Rang
 Yes, Anantha is correct that is the bug id, which could be responsible
 for more disk writes than expected.

I believe, though, that this would explain at most a factor of 2 of write 
expansion (user data getting pushed to disk once in the intent log, then again 
in its final location). If the writes are relatively large, there'd be even 
less expansion, because the ZIL will write a large enough block of data (would 
this be 128K?) into a block which can be used as its final location. (If I'm 
understanding some earlier conversations right; haven't looked at the code 
lately.)

Anton
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Re: Re: Heavy writes freezing system

2007-01-17 Thread Neil Perrin



Anton B. Rang wrote On 01/17/07 20:31,:

Yes, Anantha is correct that is the bug id, which could be responsible
for more disk writes than expected.



I believe, though, that this would explain at most a factor of 2

 of write expansion (user data getting pushed to disk once in the
 intent log, then again in its final location).

Agreed.


If the writes are

 relatively large, there'd be even less expansion, because the ZIL
 will write a large enough block of data (would this be 128K?)

Anything over zfs_immediate_write_sz (currently 32KB) is written
in this way.

 into a  block which can be used as its final location. (If I'm
 understanding some earlier conversations right; haven't looked at
 the code lately.)


Anton
 
 
This message posted from opensolaris.org

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss