Re: [zfs-discuss] # devices in raidz.

2006-11-15 Thread Richard Elling - PAE
Torrey McMahon wrote: Richard Elling - PAE wrote: Torrey McMahon wrote: Robert Milkowski wrote: Hello Torrey, Friday, November 10, 2006, 11:31:31 PM, you wrote: [SNIP] Tunable in a form of pool property, with default 100%. On the other hand maybe simple algorithm Veritas has used is good

Re: [zfs-discuss] # devices in raidz.

2006-11-13 Thread Richard Elling - PAE
Torrey McMahon wrote: Robert Milkowski wrote: Hello Torrey, Friday, November 10, 2006, 11:31:31 PM, you wrote: TM> Robert Milkowski wrote: Also scrub can consume all CPU power on smaller and older machines and that's not always what I would like. REP> The big question, though, is "10% of

[zfs-discuss] CR 6483250 closed: will not fix

2006-11-09 Thread Richard Elling - PAE
ZFS fans, Recalling our conversation about hot-plug and hot-swap terminology and use, I afraid to say that CR 6483250 has been closed as will-not-fix. No explaination was given. If you feel strongly about this, please open another CR and pile on. *Change Request ID*: 6483250 *Synopsis*: X2100 r

Re: [zfs-discuss] Best Practices recommendation on x4200

2006-11-08 Thread Richard Elling - PAE
Mike Gerdts wrote: On 11/7/06, Richard Elling - PAE <[EMAIL PROTECTED]> wrote: > d10 mirror of c0t2d0s0 and c0t3d0s0swap (2+2GB, to match above) Also a waste, use a swap file. Add a dumpdev if you care about kernel dumps, no need to mirror a dumpdev. How do you fi

Re: [zfs-discuss] Best Practices recommendation on x4200

2006-11-07 Thread Richard Elling - PAE
The best thing about best practices is that there are so many of them :-) Robert Milkowski wrote: Hello John, Tuesday, November 7, 2006, 7:45:46 PM, you wrote: JT> Greetings all- JT> I have a new X4200 that I'm getting ready to deploy. It has JT> four 146 GB SAS drives. I'd like to setup

Re: [zfs-discuss] # devices in raidz.

2006-11-07 Thread Richard Elling - PAE
Daniel Rock wrote: Richard Elling - PAE schrieb: The big question, though, is "10% of what?" User CPU? iops? Maybe something like the "slow" parameter of VxVM? slow[=iodelay] Reduces toe system performance impact of copy

Re: [zfs-discuss] # devices in raidz.

2006-11-07 Thread Richard Elling - PAE
Robert Milkowski wrote: Saturday, November 4, 2006, 12:46:05 AM, you wrote: REP> Incidentally, since ZFS schedules the resync iops itself, then it can REP> really move along on a mostly idle system. You should be able to resync REP> at near the media speed for an idle system. By contrast, a har

Re: [zfs-discuss] # devices in raidz.

2006-11-03 Thread Richard Elling - PAE
Al Hopper wrote: [1] Using MTTDL = MTBF^2 / (N * (N-1) * MTTR) But ... I'm not sure I buy into your numbers given the probability that more than one disk will fail inside the service window - given that the disks are identical? Or ... a disk failure occurs at 5:01 PM (quitting time) on a Frida

Re: [zfs-discuss] # devices in raidz.

2006-11-03 Thread Richard Elling - PAE
ozan s. yigit wrote: for s10u2, documentation recommends 3 to 9 devices in raidz. what is the basis for this recommendation? i assume it is performance and not failure resilience, but i am just guessing... [i know, recommendation was intended for people who know their raid cold, so it needed no f

Re: [zfs-discuss] Default zpool on Thumpers

2006-11-02 Thread Richard Elling - PAE
Robert Milkowski wrote: REP> P.S. did you upgrade the OS? I'd consider the need for 'zpool upgrade' to be REP> a bug. on one thumper I reinstalled OS to S10U3 beta and imported default pool. On another I put snv_49 and imported pool. Then I destroyed pools and I'm experimenting with different

Re: [zfs-discuss] raid-z random read performance

2006-11-02 Thread Richard Elling - PAE
Wow. Thanks for the data. This is somewhat consistent with what I predict in RAIDoptimizer. Robert Milkowski wrote: Hello zfs-discuss, Server: x4500, 2x Opetron 285 (dual-core), 16GB RAM, 48x500GB filebench/randomread script, filesize=256GB Your performance numbers are better than I predi

Re: [zfs-discuss] Default zpool on Thumpers

2006-11-02 Thread Richard Elling - PAE
Robert Milkowski wrote: Thumpers come with Solaris pre-installed and already configured one pool. It's a collection of raid-z1 groups but some groups are smaller than the others. I'll reconfigure it anyway but I'm just curious what side-effects can there be with such a config? Any performanc

Re: [storage-discuss] Re: [zfs-discuss] ZFS/iSCSI target integration

2006-11-02 Thread Richard Elling - PAE
comment below... Robert Milkowski wrote: Hello Richard, Wednesday, November 1, 2006, 11:36:14 PM, you wrote: REP> Adam Leventhal wrote: On Wed, Nov 01, 2006 at 04:00:43PM -0500, Torrey McMahon wrote: Lets say server A has the pool with NFS shared, or iSCSI shared, volumes. Server A exports t

Re: [zfs-discuss] ZFS/iSCSI target integration

2006-11-01 Thread Richard Elling - PAE
Rick McNeal wrote: Looking at the code it doesn't seem like the backing store being zeroed. In case of regular file a single sector (512 byte) of uninitialized data from stack (bad practice ?) is written to the very end of the file. And in case I hang my head in shame. I've fixed the code.

Re: [zfs-discuss] ZFS/iSCSI target integration

2006-11-01 Thread Richard Elling - PAE
Adam Leventhal wrote: On Wed, Nov 01, 2006 at 04:00:43PM -0500, Torrey McMahon wrote: Lets say server A has the pool with NFS shared, or iSCSI shared, volumes. Server A exports the pool or goes down. Server B imports the pool. Which clients would still be active on the filesystem(s)? The ones

Re: [zfs-discuss] ZFS Automatic Device Error Notification?

2006-10-31 Thread Richard Elling - PAE
There are several ways to do this. Two of the most popular are syslog and SNMP. syslog works, just like it always did (or didn't). For more details on FMA and how it works with SNMP traps, see the conversations on the OpenSolaris fault management community, http://www.opensolaris.org/os

Re: [zfs-discuss] Re: ZFS Performance Question

2006-10-31 Thread Richard Elling - PAE
Jay Grogan wrote: To answer your question "Yes I did expect the same or better performance than standard UFS" based on all the hype and to quote Sun "Blazing performance ZFS is based on a transactional object model that removes most of the traditional constraints on the order of issuing I/Os, w

Re: [zfs-discuss] Re: Recommended Minimum Hardware for ZFS Fileserver?

2006-10-30 Thread Richard Elling - PAE
Wes Williams wrote: Thanks gents for your replies. I've used to a very large config W2100z and ZFS for awhile but didn't know "how low can you go" for ZFS to shine, though a 64-bit CPU seems to be the minimum performance threshold. Now that Sun's store is [sort of] working again, I can see s

Re: [zfs-discuss] Recommended Minimum Hardware for ZFS Fileserver?

2006-10-30 Thread Richard Elling - PAE
Wes Williams wrote: I could use the list's help. My goal: Build a cheap ZFS file server with OpenSolairs on UFS boot (for now) 10,000 rpm U320 SCSI drive while having a ZFS pool in the same machine. The ZFS pool will either be a mirror or raidz setup consisting of either two or three 500Gb

Re: [zfs-discuss] Current status of a ZFS root

2006-10-30 Thread Richard Elling - PAE
[Richard removes his Sun hat...] Ceri Davies wrote: On Sun, Oct 29, 2006 at 12:01:45PM -0800, Richard Elling - PAE wrote: Chris Adams wrote: We're looking at replacing a current Linux server with a T1000 + a fiber channel enclosure to take advantage of ZFS. Unfortunately, the T1000 only

CLARIFICATION: Re: [zfs-discuss] Current status of a ZFS root

2006-10-29 Thread Richard Elling - PAE
CLARIFICATION below. Richard Elling - PAE wrote: Chris Adams wrote: We're looking at replacing a current Linux server with a T1000 + a fiber channel enclosure to take advantage of ZFS. Unfortunately, the T1000 only has a single drive bay (!) which makes it impossible to follow our n

Re: [zfs-discuss] Current status of a ZFS root

2006-10-29 Thread Richard Elling - PAE
Chris Adams wrote: We're looking at replacing a current Linux server with a T1000 + a fiber channel enclosure to take advantage of ZFS. Unfortunately, the T1000 only has a single drive bay (!) which makes it impossible to follow our normal practice of mirroring the root file system; naturally t

Re: [zfs-discuss] zfs on removable drive

2006-10-26 Thread Richard Elling - PAE
It is supposed to work, though I haven't tried it. Gary Gendel wrote: Here is the problem I'm trying to solve... Ive been using a sparc machine as my primary home server for years. A few years back the motherboard died. I did a nightly backup on an external USB drive formatted in ufs format.

Re: [zfs-discuss] Re: Mirrored Raidz

2006-10-24 Thread Richard Elling - PAE
Pedantic question, what would this gain us other than better data retention? Space and (especially?) performance would be worse with RAID-Z+1 than 2-way mirrors. -- richard Frank Cusack wrote: On October 24, 2006 9:19:07 AM -0700 "Anton B. Rang" <[EMAIL PROTECTED]> wrote: Our thinking is that

Re: [zfs-discuss] Migrating vdevs to new pools

2006-10-23 Thread Richard Elling - PAE
Matty wrote: We use VxVM quite a bit at my place of employment, and are extremely interested in moving to ZFS to reduce complexity and costs. One useful feature that is in VxVM that doesn't seem to be in ZFS is the ability to migrate vdevs between pools. Could you be more specific? Are you

Re: [zfs-discuss] ZFS RAID-10

2006-10-23 Thread Richard Elling - PAE
Dennis Clarke wrote: Dennis Clarke wrote: While ZFS may do a similar thing *I don't know* if there is a published document yet that shows conclusively that ZFS will survive multiple disk failures. ?? why not? Perhaps this is just too simple and therefore doesn't get explained well. That is

Re: [zfs-discuss] ZFS RAID-10

2006-10-22 Thread Richard Elling - PAE
Dennis Clarke wrote: While ZFS may do a similar thing *I don't know* if there is a published document yet that shows conclusively that ZFS will survive multiple disk failures. ?? why not? Perhaps this is just too simple and therefore doesn't get explained well. Note that SVM (nee Solstice Di

Re: [zfs-discuss] Re: [osol-discuss] Cloning a disk w/ ZFS in it

2006-10-20 Thread Richard Elling - PAE
Victor Latushkin wrote: It seems that if you clone disks bit for bit you'll end up with the same hostid for all boxes and this may be confusing. Geez, just spoof them to something different, if you must. -- richard ___ zfs-discuss mailing list zfs-di

Re: [zfs-discuss] Re: [osol-discuss] Cloning a disk w/ ZFS in it

2006-10-20 Thread Richard Elling - PAE
minor adjustments below... Darren J Moffat wrote: Asif Iqbal wrote: Hi I have a X2100 with two 74G disks. I build the OS on the first disk with slice0 root 10G ufs, slice1 2.5G swap, slice6 25MB ufs and slice7 62G zfs. What is the fastest way to clone it to the second disk. I have to build 10

Re: [zfs-discuss] Mirrored Raidz

2006-10-20 Thread Richard Elling - PAE
Frank Cusack wrote: On October 20, 2006 8:43:03 AM -0700 Ed Gould <[EMAIL PROTECTED]> wrote: On Oct 20, 2006, at 0:48, Torrey McMahon wrote: Anthony Miller wrote: I want to create create a raidz on one array and have it mirrored to the other array. Do you think this will get you more availab

Re: [zfs-discuss] ZFS and dual-pathed A5200

2006-10-20 Thread Richard Elling - PAE
Victor Latushkin wrote: The next natural question is Richard Elling - PAE wrote: Isn't this in a FAQ somewhere? IIRC, if ZFS finds a disk via two paths, then it will pick one. Will it (try to) failover to another one if picked one fails? No, not automatically. MPXIO provides auto

Re: [zfs-discuss] Mirrored Raidz

2006-10-20 Thread Richard Elling - PAE
Anthony Miller wrote: Hi, I've search the forums and not found any answer to the following. I have 2 JBOD arrays each with 4 disks. I want to create create a raidz on one array and have it mirrored to the other array. Today, the top level raid sets are assembled using dynamic striping. The

Re: [zfs-discuss] ZFS and dual-pathed A5200

2006-10-20 Thread Richard Elling - PAE
Isn't this in a FAQ somewhere? IIRC, if ZFS finds a disk via two paths, then it will pick one. -- richard Torrey McMahon wrote: Robert Milkowski wrote: This of course does work. I guess the real question was what will happen if you now export your pool, then disable mpxio so you will see the

Re: [zfs-discuss] Re: ZFS Inexpensive SATA Whitebox

2006-10-17 Thread Richard Elling - PAE
still more below... Frank Cusack wrote: On October 17, 2006 1:45:45 PM -0700 Richard Elling - PAE <[EMAIL PROTECTED]> wrote: Ah, more terminology below... Daniel Rock wrote: I still haven't found the document which states that hot-plugging of disks is not supported by Sol

CAVEAT: Re: [zfs-discuss] Re: ZFS Inexpensive SATA Whitebox

2006-10-17 Thread Richard Elling - PAE
Richard Elling - PAE wrote: All SATA drives are hot-pluggable. The caveat here is that some enclosures will cause a shutdown when opened to access the drives. The drives themselves are hot-pluggable, but access may not possible without a shutdown. -- richard

Re: [zfs-discuss] Re: ZFS Inexpensive SATA Whitebox

2006-10-17 Thread Richard Elling - PAE
Ah, more terminology below... Daniel Rock wrote: Richard Elling - PAE schrieb: Frank Cusack wrote: I'm sorry, but that's ridiculous. Sun sells a hardware product which their software does not support. The worst part is it is advertised as working. <http://www.sun.com/server

Re: [zfs-discuss] Re: ZFS Inexpensive SATA Whitebox

2006-10-17 Thread Richard Elling - PAE
Frank Cusack wrote: On October 17, 2006 10:59:51 AM -0700 Richard Elling - PAE <[EMAIL PROTECTED]> wrote: The realities of the hardware world strike again. Sun does use the Siig SATA chips in some products, Marvell in others, and NVidia MCPs in others. The difference is in who writ

Re: [zfs-discuss] Re: ZFS Inexpensive SATA Whitebox

2006-10-17 Thread Richard Elling - PAE
Dale Ghent wrote: On Oct 12, 2006, at 12:23 AM, Frank Cusack wrote: On October 11, 2006 11:14:59 PM -0400 Dale Ghent <[EMAIL PROTECTED]> wrote: Today, in 2006 - much different story. I even had Linux AND Solaris problems with my machine's MCP51 chipset when it first came out. Both forcedeth and

Re: [zfs-discuss] Re: Configuring a 3510 for ZFS

2006-10-17 Thread Richard Elling - PAE
[editorial comment below :-)] Matthew Ahrens wrote: Torrey McMahon wrote: Richard Elling - PAE wrote: Anantha N. Srirama wrote: I'm glad you asked this question. We are currently expecting 3511 storage sub-systems for our servers. We were wondering about their configuration as well.

Re: [zfs-discuss] Self-tuning recordsize

2006-10-16 Thread Richard Elling - PAE
Roch wrote: Oracle will typically create it's files with 128K writes not recordsize ones. Blast from the past... http://www.sun.com/blueprints/0400/ram-vxfs.pdf -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.

Re: [zfs-discuss] Re: Configuring a 3510 for ZFS

2006-10-16 Thread Richard Elling - PAE
Anantha N. Srirama wrote: I'm glad you asked this question. We are currently expecting 3511 storage sub-systems for our servers. We were wondering about their configuration as well. This ZFS thing throws a wrench in the old line think ;-) Seriously, we now have to put on a new hat to figure out

Re: [zfs-discuss] Thumper and ZFS

2006-10-16 Thread Richard Elling - PAE
Robert Milkowski wrote: Hello Richard, Friday, October 13, 2006, 8:05:18 AM, you wrote: REP> Do you want data availability, data retention, space, or performance? data availability, space, performance Thumper is not designed for high data availability, more traditional RAID arrays are fully

Re: [zfs-discuss] Thumper and ZFS

2006-10-12 Thread Richard Elling - PAE
Do you want data availability, data retention, space, or performance? -- richard Robert Milkowski wrote: Hello zfs-discuss, While waiting for Thumpers to come I'm thinking how to configure them. I would like to use raid-z. As thumper has 6 SATA controllers each 8-port then maybe it would

Re: [zfs-discuss] A versioning FS

2006-10-06 Thread Richard Elling - PAE
Erik Trimble wrote: The problem is we are comparing apples to oranges in user bases here. TOPS-20 systems had a couple of dozen users (or, at most, a few hundred). VMS only slightly more. UNIX/POSIX systems have 10s of thousands. IIRC, I had about a dozen files under VMS, not counting ver

Re: [zfs-discuss] A versioning FS

2006-10-05 Thread Richard Elling - PAE
Brian Hechinger wrote: On Thu, Oct 05, 2006 at 11:19:19AM -0700, David Dyer-Bennet wrote: On 10/5/06, Jeremy Teo <[EMAIL PROTECTED]> wrote: What would a version FS buy us that cron+ zfs snapshots doesn't? Finer granularity; no chance of missing a change. TOPS-20 did this, and it was *tremendo

Re: [zfs-discuss] zfs mirror resurrection

2006-10-05 Thread Richard Elling - PAE
Dick Davies wrote: Need a bit of help salvaging a perfectly working ZFS mirror that I've managed to render unbootable. I've had a ZFS root (x86, mirored zpool, SXCR b46 ) working fine for months. I very foolishly decided to mirror /grub using SVM (so I could boot easily if a disk died). Shran

Re: [zfs-discuss] Re: please remove my ignorance of raiding and mirroring

2006-09-30 Thread Richard Elling - PAE
Anton B. Rang wrote: This is a difference from RAID-5. In RAID-5, small reads are more efficient than mirroring because there are more disk spindles available, and a small read usually uses only one. Small writes are less efficient than mirroring in RAID-5 because they require a pre-read phase

Re: [zfs-discuss] jbod questions

2006-09-28 Thread Richard Elling - PAE
Keith Clay wrote: We are in the process of purchasing new san/s that our mail server runs on (JES3). We have moved our mailstores to zfs and continue to have checksum errors -- they are corrected but this improves on the ufs inode errors that require system shutdown and fsck. So, I am recomm

Re: [zfs-discuss] Metaslab alignment on RAID-Z

2006-09-27 Thread Richard Elling - PAE
observations below... Bill Moore wrote: Thanks, Chris, for digging into this and sharing your results. These seemingly stranded sectors are actually properly accounted for in terms of space utilization, since they are actually unusable while maintaining integrity in the face of a single drive f

Re: [zfs-discuss] Metaslab alignment on RAID-Z

2006-09-26 Thread Richard Elling - PAE
Chris Csanady wrote: I believe I have tracked down the problem discussed in the "low disk performance thread." It seems that an alignment issue will cause small file/block performance to be abysmal on a RAID-Z. metaslab_ff_alloc() seems to naturally align all allocations, and so all blocks will

Re: [zfs-discuss] Re: slow reads question...

2006-09-25 Thread Richard Elling - PAE
Harley Gorrell wrote: I do wonder what accounts for the improvement -- seek time, transfer rate, disk cache, or something else? Does anywone have a dtrace script to measure this which they would share? You might also be seeing the effects of defect management. As drives get older, they ten

Re: [zfs-discuss] Building large home file server with SATA

2006-09-21 Thread Richard Elling - PAE
Alexei Rodriguez wrote: I currently have a linux system at home with a pair of 3ware RAID (pci) controllers (4 port each) with a total of 8x250GB drives attached. I would like to move this existing setup to zfs but the problem I keep running into is finding suitable SATA controllers to replace

Re: [zfs-discuss] how do I find out if I am on a zfs filesystem

2006-09-21 Thread Richard Elling - PAE
mount -v -- richard Jan Hendrik Mangold wrote: This may be a dumb question, but is there a way to find out if an arbitrary filesystem is actually a zfs filesystem? Like if I were to write a script that needs to do different steps based on the underlying filesystem. Any pointers welcome.

Re: [zfs-discuss] Proposal: multiple copies of user data

2006-09-19 Thread Richard Elling - PAE
Torrey McMahon wrote: Richard Elling - PAE wrote: Non-recoverable reads may not represent permanent failures. In the case of a RAID array, the data should be reconstructed and a rewrite + verify attempted with the possibility of sparing the sector. ZFS can reconstruct the data and relocate

Re: [zfs-discuss] ZFS and HDS ShadowImage

2006-09-19 Thread Richard Elling - PAE
still more below... Torrey McMahon wrote: Darren Dunham wrote: In my experience, we would not normally try to mount two different copies of the same data at the same time on a single host. To avoid confusion, we would especially not want to do this if the data represents two different points

Re: [zfs-discuss] Proposal: multiple copies of user data

2006-09-19 Thread Richard Elling - PAE
reply below... Torrey McMahon wrote: Richard Elling - PAE wrote: This question was asked many times in this thread. IMHO, it is the single biggest reason we should implement ditto blocks for data. We did a study of disk failures in an enterprise RAID array a few years ago. One failure mode

Re: [zfs-discuss] Disk Layout for New Storage Server

2006-09-19 Thread Richard Elling - PAE
more below... Eric Schrock wrote: On Tue, Sep 19, 2006 at 02:40:14PM -0700, Eric Hill wrote: We are implementing a ZFS storage server (NAS) to replace a NetApp box. I have a Sun server with two dual Ultra320 PCIX cards connected to 4 shelves of 12 500GB disks each, yielding a total of 24TB of

Re: [zfs-discuss] ZFS layout on hardware RAID-5?

2006-09-19 Thread Richard Elling - PAE
In general, you will have to trade-off data availability, reliability, performance, and space. You haven't really given us requirements which would point us in one direction or another. I'm developing a tool to help you make these trade-offs. But it isn't yet ready for public consumption. RSN

Re: [zfs-discuss] Proposal: multiple copies of user data

2006-09-19 Thread Richard Elling - PAE
[pardon the digression] David Dyer-Bennet wrote: On 9/18/06, Richard Elling - PAE <[EMAIL PROTECTED]> wrote: Interestingly, the operation may succeed and yet we will get an error which recommends replacing the drive. For example, if the failure prediction threshold is exceeded. You

Re: [zfs-discuss] Possible file corruption on a ZFS mirror

2006-09-19 Thread Richard Elling - PAE
ditto data blocks anyone? :-) -- richard Ian Collins wrote: Dick Davies wrote: What does zpool status say? Knew I'd forgotten to check something: # zpool status -v pool: tank state: ONLINE status: One or more devices has experienced an error resulting in data corruption. Applic

Re: [zfs-discuss] Proposal: multiple copies of user data

2006-09-18 Thread Richard Elling - PAE
more below... David Dyer-Bennet wrote: On 9/18/06, Richard Elling - PAE <[EMAIL PROTECTED]> wrote: [appologies for being away from my data last week] David Dyer-Bennet wrote: > The more I look at it the more I think that a second copy on the same > disk doesn't protect again

Re: [zfs-discuss] Proposal: multiple copies of user data

2006-09-18 Thread Richard Elling - PAE
[appologies for being away from my data last week] David Dyer-Bennet wrote: The more I look at it the more I think that a second copy on the same disk doesn't protect against very much real-world risk. Am I wrong here? Are partial(small) disk corruptions more common than I think? I don't have

Re: [zfs-discuss] ZFS and HDS ShadowImage

2006-09-18 Thread Richard Elling - PAE
Joerg Haederli wrote: I'm really not an expert on ZFS, but at least from my point to handle such cases ZFS has to handle at least the following points - GUID a new/different GUID has to be assigned - LUNs ZFS has to be aware that device trees are different, if these are part of some k

Re: [zfs-discuss] Comments on a ZFS multiple use of a pool, RFE.

2006-09-18 Thread Richard Elling - PAE
Robert Milkowski wrote: Hello James, I belive that storing hostid, etc. in a label and checking if it matches on auto-import is the right solution. Before it's implemented you can use -R right now with home-clusters and don't worry about auto-import. hostid isn't sufficient (got a scar), so pe

Re: [zfs-discuss] Hotswap not working

2006-09-10 Thread Richard Elling - PAE
David Dyer-Bennet wrote: Any idea how I could approach finding drivers to let me run in a more real SATA mode? Is it worth trying Express or CE just to see? Today, use the Marvell controllers. See marvell88sx(7d) SuperMicro has several products which use these controllers. -- richard ___

Re: [zfs-discuss] Re: Re: Recommendation ZFS on StorEdge 3320 - offtopic

2006-09-08 Thread Richard Elling - PAE
Anton B. Rang wrote: JBOD probably isn't dead, simply because motherboard manufacturers are unlikely to pay the extra $10 it might cost to use a RAID-enabled chip rather than a plain chip (and the cost is more if you add cache RAM); but basic RAID is at least cheap. NVidia MCPs (later NForce

Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320 - offtopic

2006-09-08 Thread Richard Elling - PAE
Ed Gould wrote: On Sep 8, 2006, at 11:35, Torrey McMahon wrote: If I read between the lines here I think you're saying that the raid functionality is in the chipset but the management can only be done by software running on the outside. (Right?) No. All that's in the chipset is enough to rea

Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320 - offtopic

2006-09-08 Thread Richard Elling - PAE
[EMAIL PROTECTED] wrote: I don't quite see this in my crystal ball. Rather, I see all of the SAS/SATA chipset vendors putting RAID in the chipset. Basically, you can't get a "dumb" interface anymore, except for fibre channel :-). In other words, if we were to design a system in a chassis with

Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320 - offtopic

2006-09-07 Thread Richard Elling - PAE
[EMAIL PROTECTED] wrote: This is the case where I don't understand Sun's politics at all: Sun doesn't offer really cheap JBOD which can be bought just for ZFS. And don't even tell me about 3310/3320 JBODs - they are horrible expansive :-( Yep, multipacks are EOL for some time now -- killed by b

Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320

2006-09-07 Thread Richard Elling - PAE
Torrey McMahon wrote: Raid calculations take CPU time but I haven't seen numbers on ZFS usage. SVM is known for using a fair bit of CPU when performing R5 calculations and I'm sure other OS have the same issue. EMC used to go around saying that offloading raid calculations to their storage arra

Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Richard Elling - PAE
Darren Dunham wrote: Let's say the devices are named thus (and I'm making this up): /devices/../../SUNW,[EMAIL PROTECTED]/[EMAIL PROTECTED],0/WWN:sliceno [EMAIL PROTECTED] denotes the FLX380 frame, [0-6] [EMAIL PROTECTED],n denotes the virtual disk,LUN, [0-19],[0-3] How do I know that my strip

Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Richard Elling - PAE
There is another option. I'll call it "grow into your storage." Pre-ZFS, for most systems you would need to allocate the storage well in advance of its use. For the 7xFLX380 case using SVM and UFS, you would typically setup the FLX380 LUNs, merge them together using SVM, and newfs. Growing is s

Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-06 Thread Richard Elling - PAE
Oatway, Ted wrote: Thanks for the response Richard. Forgive my ignorance but the following questions come to mind as I read your response. I would then have to create 80 RAIDz(6+1) Volumes and the process of creating these Volumes can be scripted. But - 1) I would then have to create 80 mount p

Re: [zfs-discuss] Need input on implementing a ZFS layout

2006-09-05 Thread Richard Elling - PAE
Oatway, Ted wrote: IHAC that has 560+ LUNs that will be assigned to ZFS Pools and some level of protection. The LUNs are provided by seven Sun StorageTek FLX380s. Each FLX380 is configured with 20 Virtual Disks. Each Virtual Disk presents four Volumes/LUNs. (4 Volumes x 20 Virtual Disks x 7 Di

Re: [zfs-discuss] Re: Recommendation ZFS on StorEdge 3320

2006-09-05 Thread Richard Elling - PAE
Jonathan Edwards wrote: Here's 10 options I can think of to summarize combinations of zfs with hw redundancy: # ZFS ARRAY HWCAPACITYCOMMENTS -- --- 1 R0 R1 N/2 hw mirror - no zfs healing (XXX) 2 R0 R5

Re: [zfs-discuss] does anybody port the zfs webadmin to webmin?

2006-08-25 Thread Richard Elling - PAE
Hawk Tsai wrote: Webmin is faster and light weight compared to SMC. ... and most people don't know it ships with Solaris. See webmin(1m) and webminsetup(1m). -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris

Re: [zfs-discuss] Niagara and ZFS compression?

2006-08-21 Thread Richard Elling - PAE
Mike Gerdts wrote: not an expert, but most if not all compression is integer based, and I don't think floating point is supported inside the kernel anyway so it has to be integer based. Not too long ago Roch said "compression runs in the context of a single thread per pool", which makes me wor

Re: [zfs-discuss] ZFS Filesytem Corrpution

2006-08-18 Thread Richard Elling - PAE
Srivastava, Sanjaya wrote: I have been seeing data corruption on the ZFS filesystem. Here are some details. The machine is running s10 on X86 platform with a single 160Gb SATA disk. (root on s0 and zfs on s7) I'd wager that it is a hardware problem. Personally, I've had less than satisfa

Re: [zfs-discuss] Home Server with ZFS

2006-08-18 Thread Richard Elling - PAE
Ben Short wrote: Hi, I'm plan to build home server that will host my svn repository, fileserver, mailserver and webserver. This is my plan.. I have an old dell precision 420 dual 933Mhz pIII cpus. Inside this i have one scsi 9.1G hdd and 2 80G ide hdds. I am going to install solaris 10 on th

Re: [zfs-discuss] Why is ZFS raidz slower than simple ufs mount?

2006-08-17 Thread Richard Elling - PAE
Bob Evans wrote: Hi, this is a follow up to "Significant pauses to zfs writes". I'm getting about 15% slower performance using ZFS raidz than if I just mount the same type of drive using ufs. What is your expectation? -- richard ___ zfs-discuss mai

Re: [zfs-discuss] Re: user quotas vs filesystem quotas?

2006-08-15 Thread Richard Elling - PAE
Peter Bortas wrote: On 8/15/06, David Dyer-Bennet <[EMAIL PROTECTED]> wrote: On 8/15/06, Richard Elling - PAE <[EMAIL PROTECTED]> wrote: > This can be configured with the local mail delivery agent. You could even > put incoming mail in someone's $HOME, however that is

Re: [zfs-discuss] Re: user quotas vs filesystem quotas?

2006-08-15 Thread Richard Elling - PAE
Anton B. Rang wrote: One problem with this approach is that software expects /var/mail to be full of files, not directories, for each user. I don't think you can get the right semantics out of ZFS for this yet (loopback mounting a file comes to mind, but breaks down if something tries to delet

Re: [zfs-discuss] Looking for motherboard/chipset experience, again

2006-08-11 Thread Richard Elling - PAE
This is a great question for the Solaris forum at NVidia. http://www.nvnews.net/vbulletin/forumdisplay.php?f=45 My experience has been that NVidia does a pretty good job keeping the NForce software compatible with the hardware going forward. For Solaris, pre-NForce4 is a little spotty, but that

Re: [zfs-discuss] Describing ZFS RAID configs

2006-08-09 Thread Richard Elling - PAE
n parentheses. But it makes more sense to take dynamic striping as a fact of life, and not differentiate a single mirror from two mirrors, etc. - Eric On Wed, Aug 09, 2006 at 02:16:40PM -0700, Richard Elling - PAE wrote: I'd like to get a concensus of how to describe ZFS RAID configs in a

[zfs-discuss] Describing ZFS RAID configs

2006-08-09 Thread Richard Elling - PAE
I'd like to get a concensus of how to describe ZFS RAID configs in a short-hand method. For example, single-level no RAID (1 disk) RAID-0 (dynamic stripe, > 1 disk) RAID-1 RAID-Z RAID-Z2 mutliple l