Based upon the information below, zfs is under consideration for our
disk farm server system. At one point, we had to run lustre to meet an
external funding "recommendation" -- but we do not have that aegis at
present. However, one important question:
Porting a file system to an OS environment is not always trivial, and
can result in actual performance (and in some cases, reliability)
reduction/degradation. Is the port of zfs to ELNx x86-64 (N currently
6) professionally supported, and if so, by which entity? Do understand
that I regard SL as being professionally supported because there are
(paid) professional staff working on SL via Fermilab/CERN -- and TUV EL
definitely is so supported.
I found: Native ZFS on Linux
Produced at Lawrence Livermore National Laboratory
from:
http://zfsonlinux.org/
that references:
http://zfsonlinux.org/zfs-disclaimer.html
Is LLNL actually supporting zfs?
Yasha Karant
On 07/25/2013 10:57 AM, Brown, Chris (GE Healthcare) wrote:
Overview:
http://www.nexenta.com/corp/zfs-education/203-nexentastor-an-introduction-to-zfss-hybrid-storage-pool-
The ZIL:
See:
https://blogs.oracle.com/realneel/entry/the_zfs_intent_log
https://blogs.oracle.com/perrin/entry/the_lumberjack
http://nex7.blogspot.com/2013/04/zfs-intent-log.html
Accordingly it is actually quite ok to use cheap SSD.
Two things to do if doing so however:
1) low latency is key keep this in mind when selecting the prospective SSD to
use
2) Mirror and stripe the vdev EG: RAID10 ZIL 4x SSD to safe
The L2ARC:
https://blogs.oracle.com/brendan/entry/test
http://www.zfsbuild.com/2010/04/15/explanation-of-arc-and-l2arc/
Accordingly with the L2ARC it is also ok to use cheap SSD same above to two
rules apply. However due to the nature of the cache data a striped vdev of 2
SSD is fine as well.
Foregoing details but one can also achieve the same sort of general idea to a
point as the above with an external journal with ext4.
Also with BTRFS mkfs.btrfs -m raid10 SSD SSD SSD SDD -d raid10 <disk> <disk> <disk>
<disk>
- Chris
-----Original Message-----
From: owner-scientific-linux-us...@listserv.fnal.gov
[mailto:owner-scientific-linux-us...@listserv.fnal.gov] On Behalf Of Graham
Allan
Sent: Thursday, July 25, 2013 12:54 PM
To: Yasha Karant
Cc: scientific-linux-users
Subject: Re: Large filesystem recommendation
I'm not sure if anyone really knows what the reliability will be, but the hope
is obviously that these SLC-type drives should be longer-lasting (and they are
in a mirror).
Losing the ZIL used to be a fairly fatal event, but that was a long time ago
(ZFS v19 or something). I think with current ZFS versions you just lose the
performance boost if the dedicated ZIL device fails or goes away.
There's a good explanation here:
http://www.nexentastor.org/boards/2/topics/6890
Graham
On Thu, Jul 25, 2013 at 10:41:50AM -0700, Yasha Karant wrote:
How reliable are the SSDs, including actual non-corrected BER, and
what is the failure rate / interval ?
If a ZFS log on a SSD fails, what happens? Is the log automagically
recreated on a secondary SSD? Are the drives (spinning and/or SSD)
mirrored? Are primary (non-log) data lost?
Yasha Karant