Re: [zfs-discuss] # disks per vdev

2011-06-18 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Marty Scholes
> 
> On a busy array it is hard even to use the leds as indicators.

Offline the disk.  Light stays off.
Use dd to read the disk.  Light stays on.
That should make it easy enough.

Also, depending on your HBA, lots of times you can blink an Amber LED
instead of the standard green one.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Jim Klimov

2011-06-18 0:24, marty scholes wrote:
>> It makes me wonder how large shops with thousands of spindles 
handle this.


> We pay for the brand-name disk enclosures or servers where the 
fault-management stuff is supported by Solaris.

> Including the blinky lights.
>  


Funny you say that.

My Sun v40z connected a pair of Sun A5200 arrays running OSol 128a 
can't see the enclosures. The luxadm command comes up blank.


Except for that annoyance (and similar other issues) the Sun gear 
works well with a Sun operating system.




For the sake of weekend sarcasm:

Why would you wonder? That's a wrong brand name, it is too old.
Does it say "Oracle" anywhere on the label? Really, "v40z", pff!
When was it made? Like, in two-thousand-zeros, back when
dinosaurs roamed the earth and Sun was high above horizon?
Is it still supported at all, let alone Solaris (not OSol, may I add)?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread marty scholes
Funny you say that. 

My Sun v40z connected a pair of Sun A5200 arrays running OSol 128a can't see 
the enclosures. The luxadm command comes up blank. 

Except for that annoyance (and similar other issues) the Sun gear works well 
with a Sun operating system. 

Sent from Yahoo! Mail on Android

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Erik Trimble

On 6/17/2011 6:52 AM, Marty Scholes wrote:

Lights. Good.

Agreed. In a fit of desperation and stupidity I once enumerated disks by 
pulling them one by one from the array to see which zfs device faulted.

On a busy array it is hard even to use the leds as indicators.

It makes me wonder how large shops with thousands of spindles handle this.


We pay for the brand-name disk enclosures or servers where the 
fault-management stuff is supported by Solaris.


Including the blinky lights.



--
Erik Trimble
Java Platform Group Infrastructure
Mailstop:  usca22-317
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (UTC-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Marty Scholes
> Lights. Good.

Agreed. In a fit of desperation and stupidity I once enumerated disks by 
pulling them one by one from the array to see which zfs device faulted.

On a busy array it is hard even to use the leds as indicators.

It makes me wonder how large shops with thousands of spindles handle this.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Marty Scholes
> Lights.  Good.

Agreed.  In a fit of desperation and stupidity I once enumerated disks by 
pulling them one by one from the array to see which zfs device faulted.

On a busy array it is hard even to use the leds as indicators.

It makes me wonder how large shops with thousands of spindles handle this.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Lanky Doodle
>  
> or is it completely random leaving me with some trial and error to work
out
> what disk is on what port?

It's highly desirable to have drives with lights on them.  So you can
manually make the light blink (or stay on) just by reading the drive with
dd.

Even if you dig down and quantify precisely how the drives are numbered in
which order ... You would have to find labels printed on the system board or
other sata controllers, and trace the spaghetti of the sata cables, and if
you make any mistake along the way, you destroy your pool.  (Being dramatic,
but not necessarily unrealistic.)

Lights.  Good.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Erik Trimble

On 6/17/2011 12:55 AM, Lanky Doodle wrote:

Thanks Richard.

How does ZFS enumerate the disks? In terms of listing them does it do them 
logically, i.e;

controller #1 (motherboard)
 |
 |--- disk1
 |--- disk2
controller #3
 |--- disk3
 |--- disk4
 |--- disk5
 |--- disk6
 |--- disk7
 |--- disk8
 |--- disk9
 |--- disk10
controller #4
 |--- disk11
 |--- disk12
 |--- disk13
 |--- disk14
 |--- disk15
 |--- disk16
 |--- disk17
 |--- disk18

or is it completely random leaving me with some trial and error to work out 
what disk is on what port?


This is not a ZFS issue, this is the Solaris device driver issue.

Solaris uses a location-based disk naming scheme, NOT the 
BSD/Linux-style of simply incrementing the disk numbers. I.e. drives are 
usually named something like ctd


In most cases, the on-board controllers receive a lower controller 
number than any add-in adapters, and add-in adapters are enumerated in 
PCI ID order. However, there is no good explanation of exactly *what* 
number a given controller may be assigned.


After receiving a controller number, disks are enumerated in ascending 
order by ATA ID, SCSI ID, SAS WWN, or FC WWN.


The naming rules can get a bit complex.

--
Erik Trimble
Java Platform Group Infrastructure
Mailstop:  usca22-317
Phone:  x17195
Santa Clara, CA
Timezone: US/Pacific (UTC-0800)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Lanky Doodle
> 4 - the 16th port
> 
> Can you find somewhere inside the case for an SSD as
> L2ARC on your
> last port?

Although saying that, if we are saying hot spares may be bad in my scenario, I 
could ditch it and use an 3.5" SSD in the 15th drive's place?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Lanky Doodle
>I was planning on using one of
> these
> http://www.scan.co.uk/products/icy-dock-mb994sp-4s-4in
> 1-sas-sata-hot-swap-backplane-525-raid-cage

Imagine if 2.5" 2TB disks were price neutral compared to 3.5" equivalents.

I could have 40 of the buggers in my system giving 80TB raw storage! I'd 
happily use mirrors all the way in that scenario
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Lanky Doodle
Thanks Richard.

How does ZFS enumerate the disks? In terms of listing them does it do them 
logically, i.e;

controller #1 (motherboard)
|
|--- disk1
|--- disk2
controller #3
|--- disk3
|--- disk4
|--- disk5
|--- disk6
|--- disk7
|--- disk8
|--- disk9
|--- disk10
controller #4
|--- disk11
|--- disk12
|--- disk13
|--- disk14
|--- disk15
|--- disk16
|--- disk17
|--- disk18

or is it completely random leaving me with some trial and error to work out 
what disk is on what port?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-17 Thread Lanky Doodle
> 1 - are the 2 vdevs in the same pool, or two separate
> pools?
> 
I was planning on having the 2 z2 vdevs in one pool. Although having 2 pools 
and having them sync'd sounds really good, I fear it may be overkill for the 
intended purpose.

> 
> 
> 3 - spare temperature
> 
> for levels raidz2 and better, you might be happier
> with a warm spare
> and manual replacement, compared to overly-aggressive
> automated
> replacement if there is a cascade of errors.  See
> recent threads.
> 
> You may also consider a cold spare, leaving a drive
> bay free for
> disks-as-backup-tapes swapping.  If you replace the
> 1Tb's now,
> repurpose them for this rather than reselling.  
> 
I have considered this. The fact I am using cheap disks inevitably means they 
will fail sooner and more often than enterprise equivalents so the hot spare 
may be need to be over-used.

Could I have different sized vdevs and still have them both in one pool - i.e. 
an 8 disk z2 vdev and a 7 disk z2 vdev.

> 
> 4 - the 16th port
> 
> Can you find somewhere inside the case for an SSD as
> L2ARC on your
> last port?  Could be very worthwhile for some of your
> other data and
> metadata (less so the movies).

Yes! I have 10 5.1/4" drive bays in my case. 9 of them are occupied by the 
5-in-3 hot swop caddies leaving 1 bay left. I was planning on using one of 
these 
http://www.scan.co.uk/products/icy-dock-mb994sp-4s-4in1-sas-sata-hot-swap-backplane-525-raid-cage
 in the drive bay and having 2x 2.5" SATA drives mirrored for the root pool, 
leaving 2 drive bays spare.

For the mirrored root pool I was going to use 2 of the 6 motherboard SATA II 
ports so they are entirely seperate to the 'data' controllers. So I could 
either use the 16th port on the Supermicro controllers for an SSD or one of the 
remaining motherboard ports.

What size would you recommend for the L2ARC disk. I ask as I have a 72GB SAS 
10k disk spare so could use this for now (being faster than SATA), but it would 
have to be on the Supermicro card as this also supports SAS drives. SSD's are a 
bit out of range price wise at the moment so i'd wait to use one. Also ZFS 
doesn't support TRIM yet does it?

Thank you for you excellent post! :)
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-16 Thread Daniel Carosone
On Thu, Jun 16, 2011 at 07:06:48PM +0200, Roy Sigurd Karlsbakk wrote:
> > I have decided to bite the bullet and change to 2TB disks now rather
> > than go through all the effort using 1TB disks and then maybe changing
> > in 6-12 months time or whatever. The price difference between 1TB and
> > 2TB disks is marginal and I can always re-sell my 6x 1TB disks.
> > 
> > I think I have also narrowed down the raid config to these 4;
> > 
> > 2x 7 disk raid-z2 with 1 hot spare - 20TB usable
> > 3x 5 disk raid-z2 with 0 hot spare - 18TB usable
> > 2x 6 disk raid-z2 with 2 hot spares - 16TB usable
> > 
> > with option 1 probably being preferred at the moment.
> 
> I would choose option 1. I have similar configurations in
> production. A hot spare can be very good when a drive dies while
> you're not watching. 

I would probably also go for option 1, with some additional
considerations:

1 - are the 2 vdevs in the same pool, or two separate pools?

If the majority of your bulk data can be balanced manually or by
application software across 2 filesystems/pools, this offers you the
opportunity to replicate smaller more critical data between pools (and
controllers).  This offers better protection against whole-pool
problems (bugs, fat fingers).  With careful arrangement, you could
even have one pool spun down most of the time. 

You mentioned something early on that implied this kind of thinking,
but it seems to have gone by the wayside since.

If you can, I would recommend 2 pools if you go for 2
vdevs. Conversely, in one pool, you might as well go for 15xZ3 since
even this will likely cover performance needs (and see #4).

2 - disk purchase schedule

With 2 vdevs, regardless of 1 or 2 pools, you could defer purchase of
half the 2Tb drives.  With 2 pools, you can use the 6x1Tb and change
that later to 7x with the next purchase, with some juggling of
data. You might be best to buy 1 more 1Tb to get the shape right at 
the start for in-place upgrades, and in a single pool this is
essentially mandatory.

By the time you need more space to buy the second tranche of drives,
3+Tb drives may be the better option.

3 - spare temperature

for levels raidz2 and better, you might be happier with a warm spare
and manual replacement, compared to overly-aggressive automated
replacement if there is a cascade of errors.  See recent threads.

You may also consider a cold spare, leaving a drive bay free for
disks-as-backup-tapes swapping.  If you replace the 1Tb's now,
repurpose them for this rather than reselling.  

Whatever happens, if you have a mix of drive sizes, your spare should
be of the larger size. Sorry for stating the obvious! :-)

4 - the 16th port

Can you find somewhere inside the case for an SSD as L2ARC on your
last port?  Could be very worthwhile for some of your other data and
metadata (less so the movies).

--
Dan.

pgpFH3G2Esfc9.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-16 Thread Richard Elling
On Jun 16, 2011, at 2:07 AM, Lanky Doodle wrote:

> Thanks guys.
> 
> I have decided to bite the bullet and change to 2TB disks now rather than go 
> through all the effort using 1TB disks and then maybe changing in 6-12 months 
> time or whatever. The price difference between 1TB and 2TB disks is marginal 
> and I can always re-sell my 6x 1TB disks.
> 
> I think I have also narrowed down the raid config to these 4;
> 
> 2x 7 disk raid-z2 with 1 hot spare - 20TB usable
> 3x 5 disk raid-z2 with 0 hot spare - 18TB usable
> 2x 6 disk raid-z2 with 2 hot spares - 16TB usable
> 
> with option 1 probably being preferred at the moment.

Sounds good to me.

> 
> I am aware that bad batches of disks do exist so I tend to either a) buy them 
> in sets from different suppliers or b) use different manufacturers. How 
> sensitive to different disks is ZFS, in terms of disk features (NCQ, RPM 
> speed, firmware/software versions, cache etc).

Actually, ZFS has no idea it is talking to a disk. ZFS uses block devices. So 
there is nothing
in ZFS that knows about NCQ, speed, or any of those sorts of attributes. For 
the current disk
drive market, you really don't have much choice... most vendors offer very 
similar disks.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-16 Thread Roy Sigurd Karlsbakk
> I have decided to bite the bullet and change to 2TB disks now rather
> than go through all the effort using 1TB disks and then maybe changing
> in 6-12 months time or whatever. The price difference between 1TB and
> 2TB disks is marginal and I can always re-sell my 6x 1TB disks.
> 
> I think I have also narrowed down the raid config to these 4;
> 
> 2x 7 disk raid-z2 with 1 hot spare - 20TB usable
> 3x 5 disk raid-z2 with 0 hot spare - 18TB usable
> 2x 6 disk raid-z2 with 2 hot spares - 16TB usable
> 
> with option 1 probably being preferred at the moment.

I would choose option 1. I have similar configurations in production. A hot 
spare can be very good when a drive dies while you're not watching.

> I am aware that bad batches of disks do exist so I tend to either a)
> buy them in sets from different suppliers or b) use different
> manufacturers. How sensitive to different disks is ZFS, in terms of
> disk features (NCQ, RPM speed, firmware/software versions, cache etc).

For a home server, it shouldn't make much difference - the network is likely to 
be the bottleneck anyway. If you choose drives with different spin rate in a 
pool/vdev, the lower ones will probably pull down performance, so if you're 
considering "green" drives, you should use that for all the drives. Mixing 
Seagate, Samsung and Western drives should work well for this.

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-16 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Lanky Doodle
> 
> can you have one vdev that is a duplicate of another
> vdev? By that I mean say you had 2x 7 disk raid-z2 vdevs, instead of them
> both being used in one large pool could you have one that is a backup of
the
> other, allowing you to destroy one of them and re-build without data loss?

Well, you can't make a vdev from other vdev's, so you can't make a mirror of
raidz, if that's what you were hoping.

As Cindy mentioned, you can split mirrors...

Or you could use zfs send | zfs receive, to sync one pool to another pool.
This would not care if the architecture of the two pools are the same (the
2nd pool could have different or nonexistent redundancy.)  But this will be
based on snapshots.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-16 Thread Lanky Doodle
Thanks guys.

I have decided to bite the bullet and change to 2TB disks now rather than go 
through all the effort using 1TB disks and then maybe changing in 6-12 months 
time or whatever. The price difference between 1TB and 2TB disks is marginal 
and I can always re-sell my 6x 1TB disks.

I think I have also narrowed down the raid config to these 4;

2x 7 disk raid-z2 with 1 hot spare - 20TB usable
3x 5 disk raid-z2 with 0 hot spare - 18TB usable
2x 6 disk raid-z2 with 2 hot spares - 16TB usable

with option 1 probably being preferred at the moment.

I am aware that bad batches of disks do exist so I tend to either a) buy them 
in sets from different suppliers or b) use different manufacturers. How 
sensitive to different disks is ZFS, in terms of disk features (NCQ, RPM speed, 
firmware/software versions, cache etc).

Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Roy Sigurd Karlsbakk
> 3x 5 disk raid-z. 3 disk failures in the right scenario, 12TB storage
> 2x 7 disk raid-z + hot spare. 2 disk failures in the right scenario,
> 12TB storage
> 1x 15 disk raid-z2. 2 disk failures, 13TB storage
> 2x 7 disk raid-z2 + hot spare. 4 disk failures in the right scenario,
> 10TB storage

If paranoid, use two RAIDz2 VDEVs and a spare. If not, use a single RAIDz2 or 
RAIDz3 VDEV with 14-15 drives and 1-2 spares. If you choose two VDEVs, 
replacing the drives in one of them with bigger ones as the pool grows will be 
more flexible, but may lead to badly balanced pools (although I just saw that 
fixed in Illumos/openindiana - dunno about s11ex, fbsd or other platforms). 
Personally, I'm a bit paranoid, and prefer to use smaller VDEVs. With 7 drives 
per VDEV in RAIDz2, and a spare, you may still have sufficient space for some 
time. If this isn't backed up somewhere else, I'd be a wee bit paranoid indeed 
:)

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er 
et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av 
idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og 
relevante synonymer på norsk.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Cindy Swearingen

Hi Lanky,

If you created a mirrored pool instead of a RAIDZ pool, you could use
the zpool split feature to split your mirrored pool into two identical
pools.

For example, If you had 3-way mirrored pool, your primary pool will
remain redundant with 2-way mirrors after the split. Then, you would
have a non-redundant pool as a backup. You could also attach more disks
to the backup pool to make it redundant.

At the end of the week or so, destroy the non-redundant pool and
re-attach the disks to your primary pool and repeat.

This is what I would do with daily snapshots and a monthly backup.

Make sure you develop a backup strategy for any pool you build.

Thanks,

Cindy



On 06/15/11 06:20, Lanky Doodle wrote:

That's how I understood autoexpand, about not doing so until all disks have 
been done.

I do indeed rip from disc rather than grab torrents - to VIDEO_TS folders and 
not ISO - on my laptop then copy the whole folder up to WHS in one go. So while 
they're not one large single file, they are lots of small .vob files, but being 
written in one hit.

This is a bit OT, but can you have one vdev that is a duplicate of another 
vdev? By that I mean say you had 2x 7 disk raid-z2 vdevs, instead of them both 
being used in one large pool could you have one that is a backup of the other, 
allowing you to destroy one of them and re-build without data loss?

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Marty Scholes
It sounds like you are getting a good plan together.

> The only thing though I seem to remember reading that adding vdevs to
> pools way after the creation of the pool and data had been written to it,
> that things aren't spread evenly - is that right? So it might actually make
> sense to buy all the disks now and start fresh with the final build.

In this scenario, balancing would not impact your performance.  You would start 
with the performance of a single vdev.  Adding the second vdev later will only 
increase performance, even if horribly imbalanced.  Over time it will start to 
balance itself.  If you want it balanced, you can force zfs to start balancing 
by copying files then deleting the originals.

> Starting with only 6 disks would leave growth for another 6 disk
> raid-z2 (to keep matching geometry) leaving 3 disks spare which is
> not ideal. 

Maintaining identical geometry only matters if all of the disks are identical.  
If you later add 2TB disks, then pick whatever geometry works for you.  The 
most important thing is to maintain consistent vdev types, e.g. all RAIDZ2.

> I do like the idea of having a hot spare

I'm not sure I agree.  In my anecdotal experience, sometimes my array would 
offline (for whatever reason) and zfs would try to replace as many disks as it 
could with the hot spares.  If there weren't enough hot spares for the whole 
array, then the pool was left irreversibly damaged, having several disks in the 
middle of being replaced.  This has only happened once or twice and in the 
panic I might have handled it incorrectly, but it has spooked me from having 
hot spares.

> This is a bit OT, but can you have one vdev that is a duplicate of
> another vdev? By that I mean say you had 2x 7 disk raid-z2 vdevs, 
> instead of them both being used in one large pool could you have one
> that is a backup of the other, allowing you to destroy one of them
> and re-build without data loss? 

Absolutely.  I do this very thing with large, slow disks holding a backup for 
the main disks.  My home server has an SMF service which regularly synchronizes 
the time-slider snapshots from each main pool to the backup pool.  This has 
saved me when a whole pool disappeared (see above) and has allowed me to make 
changes to the layout of the main pools.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Krunal Desai
On Wed, Jun 15, 2011 at 8:20 AM, Lanky Doodle  wrote:
> That's how I understood autoexpand, about not doing so until all disks have 
> been done.
>
> I do indeed rip from disc rather than grab torrents - to VIDEO_TS folders and 
> not ISO - on my laptop then copy the whole folder up to WHS in one go. So 
> while they're not one large single file, they are lots of small .vob files, 
> but being written in one hit.

I decided on 3x 6-drive RAID-Z2s for my home media server, made up of
2TB drives (mix of Barracuda LP 5900rpm and 5K3000), it's been quite
solid so far. Performance is entirely limited by GigE.

--khd
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Jim Klimov
> This is a bit OT, but can you have one vdev that is a duplicate 
> of another vdev? By that I mean say you had 2x 7 disk raid-z2 
> vdevs, instead of them both being used in one large pool could 
> you have one that is a backup of the other, allowing you to 
> destroy one of them and re-build without data loss?

At least two ways I can think of: maybe you can make a mirror 
of raidz top-level vdevs, or simply use regular zfs send/recv syncs.
This may possibly solve some of fragmentation troubles by 
regrouping blocks during send/recv - but I asked about this 
recently on the list and did not get a definite answer.
 
//Jim
 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Lanky Doodle
That's how I understood autoexpand, about not doing so until all disks have 
been done.

I do indeed rip from disc rather than grab torrents - to VIDEO_TS folders and 
not ISO - on my laptop then copy the whole folder up to WHS in one go. So while 
they're not one large single file, they are lots of small .vob files, but being 
written in one hit.

This is a bit OT, but can you have one vdev that is a duplicate of another 
vdev? By that I mean say you had 2x 7 disk raid-z2 vdevs, instead of them both 
being used in one large pool could you have one that is a backup of the other, 
allowing you to destroy one of them and re-build without data loss?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Lanky Doodle
> 
> In that case what 'option' would you choose - smaller raid-z vdevs or
larger
> raid-z2 vdevs.

The more redundant disks you have, the more protection you get, and the
smaller available disk space.  So that's entirely up to you.


> When 4-5TB
> drives come to market 2-3TB drives will drop in price so I could always
> upgrade them - can you do this with raid-z vdevs, in terms of autoexand?

Yup.  But you won't see any increase until you replace all the drives in the
vdev.


> There might be the odd deletion here and there if a movie is truly turd,
but
> as you say 99% of the time it will be written and left.

That won't matter.  The thing that matters is ...  File fragmentation.

For example, if you run bit torrent directly onto the file server, then
you're going to get terrible performance for everything because bittorrent
grabs tiny little fragments all over the place in essentially random order.
But if you rip directly from disc to a file, then it'll be fine because it's
all serialized.  Or use bittorrent onto your laptop and then copy the file
all at once to the server.

The thing that's bad for performance, especially on raidz, is when you're
performing lots of small random operations.  And when you come back to a
large file making small random modifications after it has already been
written and snapshotted...

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-15 Thread Lanky Doodle
Thanks Edward.

In that case what 'option' would you choose - smaller raid-z vdevs or larger 
raid-z2 vdevs.

I do like the idea of having a hot spare so 2x 7 disk raid-z2 may be the better 
option rather than 3x 5 disk raid-z with no hot spare. 2TB loss in the former 
could be acceptable I suppose for the sake of better protection. When 4-5TB 
drives come to market 2-3TB drives will drop in price so I could always upgrade 
them - can you do this with raid-z vdevs, in terms of autoexand?

There might be the odd deletion here and there if a movie is truly turd, but as 
you say 99% of the time it will be written and left.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-14 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Lanky Doodle
> 
> But as it for a home media server, which is mainly WORM access and will be
> storing (legal!) DVD/Bluray rips i'm not so sure I can sacrify the space.

For your purposes, raidzN will work very well.  And since you're going to
sequentially write your data once initially and leave it in place, even the
resilver should perform pretty well.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-14 Thread Lanky Doodle
Thanks martysch.

That is what I meant about adding disks to vdevs - not adding disks to vdevs 
but adding vdevs to pools.

If the geometry of the vdevs should ideally be the same, it would make sense to 
buy one more disk now and have a 7 disk raid-z2 to start with, then buy disks 
as and when and create a further 7 disk raid-z2 leaving the 15th disk as a hot 
spare. Would 'only' give 10TB usable though.

The only thing though I seem to remember reading that adding vdevs to pools way 
after the creation of the pool and data had been written to it, that things 
aren't spread evenly - is that right? So it might actually make sense to buy 
all the disks now and start fresh with the final build.

Starting with only 6 disks would leave growth for another 6 disk raid-z2 (to 
keep matching geometry) leaving 3 disks spare which is not ideal.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-14 Thread Marty Scholes
I am asssuming you will put all of the vdevs into a single pool, which is a 
good idea unless you have a specific reason for keeping them separate, e.g. you 
want to be able to destroy / rebuild a particular vdev while leaving the others 
intact.

Fewer disks per vdev implies more vdevs, providing better random performance, 
lower scrub and resilver times and the ability to expand a vdev by replacing 
only the few disks in it.

The downside of more vdevs is that you dedicate your parity to each vdev, e.g. 
a RAIDZ2 would need two parity disks per vdev.

> I'm in two minds with mirrors. I know they provide
> the best performance and protection, and if this was
> a business critical machine I wouldn't hesitate.
> 
> But as it for a home media server, which is mainly
> WORM access and will be storing (legal!) DVD/Bluray
> rips i'm not so sure I can sacrify the space.

For a home media server, all accesses are essentially sequential, so random 
performance should not be a deciding factor.

> 7x 2 way mirrors would give me 7TB usable with 1 hot
> spare, using 1TB disks, which is a big drop from
> 12TB! I could always jump to 2TB disks giving me 14TB
> usable but I already have 6x 1TB disks in my WHS
> build which i'd like to re-use.

I would be tempted to start with a 4+2 (six disk RAIDZ2) vdev using your 
current disks and plan from there.  There is no reason you should feel 
compelled to buy more 1TB disks just because you already have some.

> Am I right in saying that single disks cannot be
> added to a raid-z* vdev so a minimum of 3 would be
> required each time. However a mirror is just 2 disks
> so if adding disks over a period of time mirrors
> would be cheaper each time.

That is not correct.  You cannot ever add disks to a vdev.  Well, you can add 
additional disks to a mirror vdev, but otherwise, once you set the geometry, a 
vdev is stuck for life.

However, you can add any vdev you want to an existing pool.  You can take a 
pool with a single vdev set up as a 6x RAIDZ2 and add a single disk to that 
pool.  The previous example is a horrible idea because it makes the entire pool 
dependent upon a single disk.  The example also illustrates that you can add 
any type of vdev to a pool.

Most agree it is best to make the pool from vdevs of identical geometry, but 
that is not enforced by zfs.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-14 Thread Lanky Doodle
Thanks Edward.

I'm in two minds with mirrors. I know they provide the best performance and 
protection, and if this was a business critical machine I wouldn't hesitate.

But as it for a home media server, which is mainly WORM access and will be 
storing (legal!) DVD/Bluray rips i'm not so sure I can sacrify the space.

7x 2 way mirrors would give me 7TB usable with 1 hot spare, using 1TB disks, 
which is a big drop from 12TB! I could always jump to 2TB disks giving me 14TB 
usable but I already have 6x 1TB disks in my WHS build which i'd like to re-use.

Hmmm!
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] # disks per vdev

2011-06-14 Thread Edward Ned Harvey
> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
> boun...@opensolaris.org] On Behalf Of Lanky Doodle
> 
> The ZFS install will be mirrored, but I am not sure how to configure the
15
> data disks from a performance (inc. resilvering) vs protection vs usable
space
> perspective;
> 
> 3x 5 disk raid-z. 3 disk failures in the right scenario, 12TB storage
> 2x 7 disk raid-z + hot spare. 2 disk failures in the right scenario, 12TB
storage
> 1x 15 disk raid-z2. 2 disk failures, 13TB storage
> 2x 7 disk raid-z2 + hot spare. 4 disk failures in the right scenario, 10TB
storage

The above all provide highest usable space (lowest hardware cost.)
But if you want performance, go for mirrors.  (highest hardware cost.)

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] # disks per vdev

2011-06-14 Thread Lanky Doodle
Hiya,

I am just in the planning stages for my ZFS Home Media Server build at the 
moment (to replace WHS v1).

I plan to use 2x motherboard ports and 2x Supermicro AOC-SASLP-MV8 8 port SATA 
cards to give 17* drive connections; 2 disks (120GB SATA 2.5") will be used for 
the ZFS install using the motherboard ports and the remaing 15 disks (1TB SATA) 
will be used for data using the 2x 8 port cards.

* = the total number of ports is 18 but I only have enough space in the chassis 
for 17 drives (2x 2.5" in 1x 3.5" bay and 15x 3.5" by using 5-in-3 hotswop 
caddies in 9x 5.1/4" bays).

All disks are 5400RPM to keep power requirements down.

The ZFS install will be mirrored, but I am not sure how to configure the 15 
data disks from a performance (inc. resilvering) vs protection vs usable space 
perspective;

3x 5 disk raid-z. 3 disk failures in the right scenario, 12TB storage
2x 7 disk raid-z + hot spare. 2 disk failures in the right scenario, 12TB 
storage
1x 15 disk raid-z2. 2 disk failures, 13TB storage
2x 7 disk raid-z2 + hot spare. 4 disk failures in the right scenario, 10TB 
storage

Without having a mash of different raid-z* levels I can't think of any other 
options.

I am leaning towards the first option as it gives seperation between all the 
disks; I would have seperate Movie folders on each of them while having 
critical data (pictures, home videos, documents etc) stored on each set of 
raid-z.

Suggestions welcomed.

Thanks
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss