[zfs-discuss] Custom Jumpstart and RAID-10 ZFS rpool

2008-10-29 Thread Stephen Le
Is it possible to create a custom Jumpstart profile to install Nevada
on a RAID-10 rpool? From the ZFS Boot FAQ [1], you can create a
profile to install Nevada with a RAID-1 rpool using the following
line:

pool newpool auto auto auto mirror c0t0d0s0 c0t1d0s0

Is there an equivalent line for RAID-10? I've tried using the
following line, but the Jumpstart check script complains about it
being invalid:

pool newpool auto auto auto mirror c0t0d0s0 c0t1d0s0 mirror c0t2d0s0 c0t3d0s0

Thanks,
Stephen Le

[1] http://opensolaris.org/os/community/zfs/boot/zfsbootFAQ/#jumpinstall
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Custom Jumpstart and RAID-10 ZFS rpool

2008-10-29 Thread Ian Collins
Stephen Le wrote:
 Is it possible to create a custom Jumpstart profile to install Nevada
 on a RAID-10 rpool? 

No, simple mirrors only.

-- 
Ian.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [osol-bugs] what's the story wtih bug #6592835?

2008-10-29 Thread Nils Goroll
Hi Graham,

(this message was posed on opensolaris-bugs initially, I am CC'ing and 
reply-to'ing zfs-discuss as it seems to be a more appropriate place to discuss 
this.)

 I'm surprised to see that the status of bug 6592835 hasn't moved beyond yes 
 that's a problem.

My understanding is that the resilver speed is tied to fact that the currenct 
resilver implementation follows the ZFS on disk structures, which needs 
random-like I/O operations while a traditional RAID rebuild issues sequential 
I/O only. Simply put, the former is very slow while the latter is very fast 
with 
respect to the amounts of data having to be touched.

IIRC, this issue has been discussed on zfs-discuss several times already and my 
understanding it that it would be very difficult to implement some kind of 
sequential resilver (walking disk blocks sequentially rather than ZFS on disk 
structures), so I doubt if an improvement can be expected anytime soon.

That said, the core developers on zfs-discuss will know more than me and might 
be willing to give more background on this.

Nils
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-29 Thread Matt Harrison
On Tue, Oct 28, 2008 at 05:30:55PM -0700, Nigel Smith wrote:
 Hi Matt.
 Ok, got the capture and successfully 'unzipped' it.
 (Sorry, I guess I'm using old software to do this!)
 
 I see 12840 packets. The capture is a TCP conversation 
 between two hosts using the SMB aka CIFS protocol.
 
 10.194.217.10 is the client - Presumably Windows?
 10.194.217.3 is the server - Presumably OpenSolaris - CIFS server?

All correct so far

 Using WireShark,
 Menu: 'Statistics  Endpoints' show:
 
 The Client has transmitted 4849 packets, and
 the Server has transmitted 7991 packets.
 
 Menu: 'Analyze  Expert info Composite':
 The 'Errors' tab shows:
 4849 packets with a 'Bad TCP checksum' error - These are all transmitted by 
 the Client.
 
 (Apply a filter of 'ip.src_host == 10.194.217.10' to confirm this.)
 
 The 'Notes' tab shows:
 ..numerous 'Duplicate Ack's'
 For example, for 60 different ACK packets, the exact same packet was 
 re-transmitted 7 times!
 Packet #3718 was duplicated 17 times.
 Packet #8215 was duplicated 16 times.
 packet #6421 was duplicated 15 times, etc.
 These bursts of duplicate ACK packets are all coming from the client side.
 
 This certainly looks strange to me - I've not seen anything like this before.
 It's not going to help the speed to unnecessarily duplicate packets like
 that, and these burst are often closely followed by a short delay, ~0.2 
 seconds.
 And as far as I can see, it looks to point towards the client as the source
 of the problem.
 If you are seeing the same problem with other client PC, then I guess we need 
 to 
 suspect the 'switch' that connects them.

I have another switch on the way to move to. I will see if this helps.

Thanks for your input

Matt


pgpPpxSVqiW79.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] HELP! SNV_97, 98, 99 zfs with iscsitadm and VMWare!

2008-10-29 Thread Nigel Smith
Hi Tano
Great to hear that you've now got this working!!

I understand you are using a Broadcom network card,
from your previous posts I can see you are using the 'bnx' driver.

I will raise this as a bug, but first please would you run 
'/usr/X11/bin/scanpci'
to indentify the exact 'vendor id' and 'device id' for the Broadcom network 
chipset,
and report that back here.

I must admit that this is the first I have heard of 'I/OAT DMA',
so I did some Googling on it, and found this links:

http://opensolaris.org/os/community/arc/caselog/2008/257/onepager/

To quote from that ARC case:

  All new Sun Intel based platforms have Intel I/OAT (I/O Acceleration
   Technology) hardware.

   The first such hardware is an on-systemboard asynchronous DMA engine
   code named Crystal Beach.

   Through a set of RFEs Solaris will use this hardware to implement
   TCP receive side zero CPU copy via a socket.

Ok, so I think that makes some sense, in the context of
the problem we were seeing. It's referring to how the network
adaptor transfers the data it has received, out of the buffer
and onto the rest of the operating system.

I've just looked to see if I can find the source code for 
the BNX driver, but I cannot find it.

Digging deeper we find on this page:
http://www.opensolaris.org/os/about/no_source/
..on the 'ON' tab, that:

Components for which there are currently no plans to release source
bnx driver (B)  Broadcom NetXtreme II Gigabit Ethernet driver

So the bnx driver is closed source :-(
Regards
Nigel Smith
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-29 Thread Matt Harrison
On Tue, Oct 28, 2008 at 05:45:48PM -0700, Richard Elling wrote:
 I replied to Matt directly, but didn't hear back.  It may be a driver issue
 with checksum offloading.  Certainly the symptoms are consistent.
 To test with a workaround see
 http://bugs.opensolaris.org/view_bug.do?bug_id=6686415

Hi, Sorry for not replying, we had some problems with our email provider
yesterday and I was up all night restoring backups.

I did try the workaround, but it didn't have any effect, presumbably because
it's not using the rge driver as you stated before.

I'll try swapping the switch out and post back my results.

Many Thanks

Matt


pgphN0KxBmHEz.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Cannot remove slog device from zpool

2008-10-29 Thread Robert Milkowski
Hello Neil,

Tuesday, October 28, 2008, 10:34:28 PM, you wrote:
NP However, we wanted to make sure it fit into the framework for
NP the removal of any device. This a much harder problem which we
NP have made progress, but it's not there yet...

I think a lot of people here would be interested in more details and any
ETA (no commitments) on this - could you write couple of sentences on
what you guys are actually doing re disk eviction? FOr example - would
it be possible to change raidz2 - raidz1 or raid10 on a fly?


-- 
Best regards,
 Robert Milkowskimailto:[EMAIL PROTECTED]
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [osol-bugs] what's the story wtih bug #6592835?

2008-10-29 Thread Graham McArdle
Hi Nils,

thanks for the detailed info. I've tried searching the zfs-discuss archive for 
both the bug id and 'resilver', but in both cases the only result I can find 
from the whole history is this thread:
http://www.opensolaris.org/jive/thread.jspa?messageID=276358#276358
Maybe the discussions you recall aren't fully indexed for searching on these 
keywords or they were in another forum, but thanks for giving me the gist of 
it. It is potentially quite an Achilles heel for ZFS though. I've argued 
locally to migrate our main online data archive (currently 3.5TB) to ZFS, but 
if the recovery time for disk failures keeps getting slower as the archive 
grows and accumulates snapshots etc., some questions might be asked about this 
policy. I've suggested we can do continuous data replication to a secondary 
server by sending incremental snapshot streams, but if the CDR had to be 
suspended for a significant time (days) to allow a resilver, this would be a 
real problem, at least until the 6343667 bugfix from snv_94 finds its way into 
a Solaris 10 patch (will this happen?).
Does the severity of the problem depend on access / write patterns used, i.e. 
would a simple archiving system where data is only ever added in a sequential 
fashion be less susceptible to slow rebuilds than a system where data is 
written, snapshotted, moved, modified, deleted, etc.
Does the time taken to scrub the pool give some indication of the likely 
resilvering time, or does that process walk a different kind of tree?

Graham
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Lost Disk Space

2008-10-29 Thread Turanga Leela
 No takers? :)
 
 benr.

I'm quite curious about finding out about this too, to be honest :)

And its not just ZFS on Solaris because I've filled up and imported pools into 
ZFS Fuse 0.5.0 (which is based on the latest ZFS code) in Linux, and on FreeBSD 
too.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-29 Thread Martti Kuparinen
Bob Friesenhahn wrote:
 AMD Athelon/Opteron dual core likely matches or exceeds 
 Intel quad core for ZFS use due to a less bottlenecked memory channel.

How big is the difference? Does anyone have benchmarking results (maybe even 
when using ZFS on Solaris 10)?

Martti
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Strange result when syncing between SPARC and x86

2008-10-29 Thread Turanga Leela
 Example is:
 
 [EMAIL PROTECTED] ls -la
 /data/zones/testfs/root/etc/services
 lrwxrwxrwx   1 root root  15 Oct 13 14:35
 /data/zones/testfs/root/etc/services -
 ./inet/services
 
 [EMAIL PROTECTED] ls -la /data/zones/testfs/root/etc/services
 lrwxrwxrwx   1 root root  15 Oct 13 14:35
 /data/zones/testfs/root/etc/services -
 s/teni/.ervices

Ouch, thats a bad one.

I downloaded and burnt b101 to dvd for x86 and solaris, i'm gonna install them 
tomorrow at work and try moving a pool between them to see what happens...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-29 Thread gm_sjo
Out of interest, and reasonably on-topic, can anyone predict
performance comparison (CIFS) between these two setups?

1) Dedicated Windows 2003 Server, Intel hardware SATA RAID controller
(single raid 5 array, 8 disks)

2) OpenSolaris+ZFS+CIFS, 8 drives with a SuperMicro controller
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [osol-bugs] what's the story wtih bug #6592835?

2008-10-29 Thread Bob Friesenhahn
On Wed, 29 Oct 2008, Nils Goroll wrote:
 My understanding is that the resilver speed is tied to fact that the currenct
 resilver implementation follows the ZFS on disk structures, which needs
 random-like I/O operations while a traditional RAID rebuild issues sequential
 I/O only. Simply put, the former is very slow while the latter is very fast 
 with
 respect to the amounts of data having to be touched.

If this is indeed an issue, then the pool itself is severely 
fragmented and will be slow in normal use.  This could happen if the 
pool is allowed to become overly full.

Bob
==
Bob Friesenhahn
[EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [osol-bugs] what's the story wtih bug #6592835?

2008-10-29 Thread Bob Friesenhahn
On Wed, 29 Oct 2008, Graham McArdle wrote:

 Maybe the discussions you recall aren't fully indexed for searching 
 on these keywords or they were in another forum, but thanks for 
 giving me the gist of it. It is potentially quite an Achilles heel 
 for ZFS though. I've argued locally to migrate our main online data 
 archive (currently 3.5TB) to ZFS, but if the recovery time for disk 
 failures keeps getting slower as the archive grows and accumulates 
 snapshots etc., some questions might be asked about this policy.

The simple solution is to use a reasonable pool design.  Limit the 
maximum LUN size to a size which can be resilved in reasonable time. 
Don't do something silly like building a mirror across two 10TB LUNs.

Huge disks will take longer to resilver, and it is more likely for a 
secondary failure to occur during resilvering.  Manage your potential 
losses by managing the size of a loss, and therefore the time to 
recover.

Another rules is to control how full a pool is allowed to become.  If 
you fill it to 99% you can be assured that the pool will become 
fragmented and resilvers will take much longer.  The pool will be 
slower in normal use as well.

Bob
==
Bob Friesenhahn
[EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Custom Jumpstart and RAID-10 ZFS rpool

2008-10-29 Thread Kyle McDonald
Ian Collins wrote:
 Stephen Le wrote:
   
 Is it possible to create a custom Jumpstart profile to install Nevada
 on a RAID-10 rpool? 
 

 No, simple mirrors only.
   
Though a finish sscript could add additional simple mirrors to create 
the config his example would have created.
Pretty sure that's still not RAID10 though.

And any files laid down by the installer would be constrained to the 
first mirrored pair, only new files would have a chance at be 
distributed over the addtional pairs.

 -Kyle

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-29 Thread Bob Friesenhahn
On Wed, 29 Oct 2008, Martti Kuparinen wrote:

 Bob Friesenhahn wrote:
 AMD Athelon/Opteron dual core likely matches or exceeds
 Intel quad core for ZFS use due to a less bottlenecked memory channel.

 How big is the difference? Does anyone have benchmarking results (maybe even
 when using ZFS on Solaris 10)?

The big question would be what should be benchmarked.  ZFS is like a 
big RAM cache.  The more RAM the better.  You would be surprised how 
little disk activity there can really be on systems with a lot of RAM 
as long as synchronous writes are avoided.  As a result, some common 
scenarios mostly exercise RAM rather than the disk channel.

Unless you need a higher power CPU for other purposes, a ZFS-based 
server should focus on maximizing installed RAM.

Bob
==
Bob Friesenhahn
[EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Strange result when syncing between SPARC and x86

2008-10-29 Thread Mike Futerko
Hi

 [EMAIL PROTECTED] ls -la
 /data/zones/testfs/root/etc/services
 lrwxrwxrwx   1 root root  15 Oct 13 14:35
 /data/zones/testfs/root/etc/services -
 ./inet/services

 [EMAIL PROTECTED] ls -la /data/zones/testfs/root/etc/services
 lrwxrwxrwx   1 root root  15 Oct 13 14:35
 /data/zones/testfs/root/etc/services -
 s/teni/.ervices
 
 Ouch, thats a bad one.
 
 I downloaded and burnt b101 to dvd for x86 and solaris, i'm gonna install 
 them tomorrow at work and try moving a pool between them to see what 
 happens...

Would be interesting to know how it'll work if move whole zpool not just
sync with send/recv. But I think all will be fine there as is seems the
problem is in send/recv part on the file system itself on different
architectures.


Thanks
Mike
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Hotplug issues on USB removable media.

2008-10-29 Thread Ross
I don't remember anyone saying they couldn't be stored, just that if they are 
stored it's not ZFS' fault if they go corrupt as it's outside of its control.

I'm actually planning to store the zfs send dump on external USB devices 
myself, but since the USB device will be running on ZFS I expect to be able to 
find out if there are any corruption problems, at which point I'll just run the 
send / receive again.

However, I don't think that's what they're talking about here.  I think they're 
talking about a ZFS pool that consists of an external USB device, and doing a 
send / receive directly to that pool.  That way the USB device is a true backup 
copy of your ZFS pool, and I think the idea is that you can then delete 
snapshots from your main system, confident that they are still present on the 
USB backup.

If it works it's a nice idea, especially with an integrated restore interface.

And yeah, auto mounting isn't something you're going to want to enable on 
production servers.  But for desktop / development / small scale use, it's a 
great idea.

And I agree 100% about hotplug stuff being untrusted input.  You shouldn't be 
able to crash anything with a USB stick, especially not a bunch of unrelated 
ZFS filesystems.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-29 Thread Richard Elling
Matt Harrison wrote:
 On Tue, Oct 28, 2008 at 05:45:48PM -0700, Richard Elling wrote:
   
 I replied to Matt directly, but didn't hear back.  It may be a driver issue
 with checksum offloading.  Certainly the symptoms are consistent.
 To test with a workaround see
 http://bugs.opensolaris.org/view_bug.do?bug_id=6686415
 

 Hi, Sorry for not replying, we had some problems with our email provider
 yesterday and I was up all night restoring backups.

 I did try the workaround, but it didn't have any effect, presumbably because
 it's not using the rge driver as you stated before.
   

The dohwcksum is not an rge option, it is an ip option (hence it is
named ip:dohwcksum) and it transcends NIC drivers.  But if that
didn't fix anything, then you should be able to safely ignore it.

 I'll try swapping the switch out and post back my results.
   

Yeah, I've seen this sort of thing before, too.  I once had a switch
that lost its mind and wouldn't let big packets through unscathed.
We could telnet, ping, ftp, and do all sorts of things through the switch,
but we couldn't push NFS through.  These sorts of failures can be
difficult to isolate.
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Hotplug issues on USB removable media.

2008-10-29 Thread Niall Power
 
 However, I don't think that's what they're talking
 about here.  I think they're talking about a ZFS pool
 that consists of an external USB device, and doing a
 send / receive directly to that pool.  That way the
 USB device is a true backup copy of your ZFS pool,
 and I think the idea is that you can then delete
 snapshots from your main system, confident that they
 are still present on the USB backup. 

That's exactly it. While it's great to have snapshots stored on the same device
as the filesystem for convenience, this is limited in that it's not a full 
recovery
solution since it doesn't offer any protection from physical hardware failure, 
especially
on laptops and workstations where there will most likely be just one hard disk 
on the
system. What we have right now is protection from user and software errors 
which is
probably the source of most cases of data loss or corruption, but we need to 
extend this
to cover hardware failure, and we need to be able to backup to secondary 
devices to
acheive that. For most single users that means an external USB disk or thumb 
drive.

 
 If it works it's a nice idea, especially with an
 integrated restore interface.

That would also be part of our plan.

Cheers,
Niall

 
 And yeah, auto mounting isn't something you're going
 to want to enable on production servers.  But for
 desktop / development / small scale use, it's a great
 idea.
 
 And I agree 100% about hotplug stuff being untrusted
 input.  You shouldn't be able to crash anything with
 a USB stick, especially not a bunch of unrelated ZFS
 filesystems.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance with tens of thousands of zfs filesystems

2008-10-29 Thread Al Hopper
On Tue, Oct 28, 2008 at 9:27 AM, Morten-Christian Bernson [EMAIL PROTECTED] 
wrote:
 I have been reading this forum for a little while, and am interested in more 
 information about the performance of ZFS when creating large amounts of 
 filesystems.   We are considering using ZFS for the user's home folders, and 
 this could potentially be 30'000 filesystems, and if using snapshots that 
 number would be multiplied by x snapshots as well.

 I am a bit nervous after reading this forum, that the performance when 
 getting huge numbers of filesystems is not very good.  Using hours to boot 
 the server, and possibly weeks to make the filesystems seems not right.

You'll have to solve this issue by using multiple ZFS servers so that
the number of filesystems per server is a reasonably good fit with
ZFS' current capabilities.  OTOH  ZFS is improving in this arena over
time and, even with multiple servers, ZFS will still provide a cost
effective solution.   In any case, with 5 ZFS servers, would it not be
better to have only 20% of your user community affected if a server
were to have a catastrophic failure?  (rhetorical question)

 Any official input on how this will be in the upcoming release of Solaris 10?

You won't get any official input on Solaris (the commercial product)
related topics on this list - which is dedicated to OpenSolaris.  But,
to give you something that resembles an answer, Sol10 Update 6 is
based on Build 88 AFAIR.  So looking at the features/performance of
that release will give you a good feel for Sol 10U6 .. the usual
disclaimers YMMV etc.

Regards,

-- 
Al Hopper  Logical Approach Inc,Plano,TX [EMAIL PROTECTED]
   Voice: 972.379.2133 Timezone: US CDT
OpenSolaris Governing Board (OGB) Member - Apr 2005 to Mar 2007
http://www.opensolaris.org/os/community/ogb/ogb_2005-2007/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-29 Thread Al Hopper
On Wed, Oct 29, 2008 at 8:43 AM, Bob Friesenhahn
[EMAIL PROTECTED] wrote:
 On Wed, 29 Oct 2008, Martti Kuparinen wrote:

 Bob Friesenhahn wrote:
 AMD Athelon/Opteron dual core likely matches or exceeds
 Intel quad core for ZFS use due to a less bottlenecked memory channel.

 How big is the difference? Does anyone have benchmarking results (maybe even
 when using ZFS on Solaris 10)?

 The big question would be what should be benchmarked.  ZFS is like a
 big RAM cache.  The more RAM the better.  You would be surprised how
 little disk activity there can really be on systems with a lot of RAM
 as long as synchronous writes are avoided.  As a result, some common
 scenarios mostly exercise RAM rather than the disk channel.

 Unless you need a higher power CPU for other purposes, a ZFS-based
 server should focus on maximizing installed RAM.

Agreed 100%

It's easy to find DDR2 RAM at around $20/gigabyte (based on 1Gb or 2Gb
DIMMs) and I've seen some deals as low a $8/Gb for Kingston RAM.

Regards,

-- 
Al Hopper  Logical Approach Inc,Plano,TX [EMAIL PROTECTED]
   Voice: 972.379.2133 Timezone: US CDT
OpenSolaris Governing Board (OGB) Member - Apr 2005 to Mar 2007
http://www.opensolaris.org/os/community/ogb/ogb_2005-2007/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Peter Baer Galvin
This seems like a n00b question but I'm stuck.

Nevada build 101. Doing fresh install (in vmware fusion). I don't see any way 
to select zfs as the root file system. Looks to me like UFS is the default, but 
I don't see any option box to allow that to be changed to zfs. What am I 
missing?! Thanks.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Rich Teer
On Wed, 29 Oct 2008, Peter Baer Galvin wrote:

 Nevada build 101. Doing fresh install (in vmware fusion). I don't see
 any way to select zfs as the root file system. Looks to me like UFS is
 the default, but I don't see any option box to allow that to be changed
 to zfs. What am I missing?! Thanks.

You have to use the text installer (rather than the default GUI one).
Part way through the process you will be offered a choice of UFS or
ZFS root file system (UFS is the default still).  Select ZFS and away
you go!

-- 
Rich Teer, SCSA, SCNA, SCSECA

CEO,
My Online Home Inventory

URLs: http://www.rite-group.com/rich
  http://www.linkedin.com/in/richteer
  http://www.myonlinehomeinventory.com
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Daryl Doami

Hi Peter,

It's there, you just can't use the GUI installer.  You have to choose 
the text interactive installer.  It'll give you the choice there.


Regards.

 Original Message 
Subject: [zfs-discuss] zfs boot / root in Nevada build 101
From: Peter Baer Galvin [EMAIL PROTECTED]
To: zfs-discuss@opensolaris.org
Date: Wed Oct 29 09:28:46 2008

This seems like a n00b question but I'm stuck.

Nevada build 101. Doing fresh install (in vmware fusion). I don't see any way 
to select zfs as the root file system. Looks to me like UFS is the default, but 
I don't see any option box to allow that to be changed to zfs. What am I 
missing?! Thanks.
  


begin:vcard
fn:Daryl Doami
n:Doami;Daryl
org:Sun Microsystems Federal, Inc.;DoD, Intel,  NASA West Regions
adr;dom:;;222 N. Sepulveda Blvd. 10th Floor;El Segundo;CA;90245
email;internet:[EMAIL PROTECTED]
title:Systems Engineer
tel;work:310-242-6463
tel;fax:310-242-6463
x-mozilla-html:FALSE
url:http://www.sun.com/government/
version:2.1
end:vcard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Cindy . Swearingen
Hi Peter,

You need to select the text-mode install option to select a ZFS root
file system.

Other ZFS root installation tips are described here:

http://docs.sun.com/app/docs/doc/817-2271/zfsboot-1?a=view

I'll be attending Richard Elling's ZFS workshop at LISA08.

Hope to see you. :-)

Cindy

Peter Baer Galvin wrote:
 This seems like a n00b question but I'm stuck.
 
 Nevada build 101. Doing fresh install (in vmware fusion). I don't see any way 
 to select zfs as the root file system. Looks to me like UFS is the default, 
 but I don't see any option box to allow that to be changed to zfs. What am I 
 missing?! Thanks.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Glenn Lagasse
* Peter Baer Galvin ([EMAIL PROTECTED]) wrote:
 This seems like a n00b question but I'm stuck.
 
 Nevada build 101. Doing fresh install (in vmware fusion). I don't see
 any way to select zfs as the root file system. Looks to me like UFS is
 the default, but I don't see any option box to allow that to be
 changed to zfs. What am I missing?! Thanks.

You need to use the text mode installer, that's the only installer that
has support for ZFS root installations in SXCE.

Cheers,

Glenn
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] [osol-bugs] what's the story wtih bug #6592835?

2008-10-29 Thread Graham McArdle
We have a 24-disk server, so the current design is 2-disk root mirror and 2x 
11-disk RAIDZ2 vdevs. I suppose another solution could have been to have 3x 
7-disk vdevs plus a hot spare, but the capacity starts to get compromised. 
Using 1TB disks in our current config will give us growth capacity to 16TiB. 
Obviously 3.5TiB is a small starting point, but we are facing an exponential 
growth curve.
It seems like the recommendation is to keep expanding out in disk quantity 
rather than upgrading disk size to meet growth requirements. Perhaps 1TB SATA 
disk is already too big a lump for ZFS resilver.
Looks like there is also some hope that if bp rewrite becomes a reality it 
will also help by allowing online defragmentation:
http://www.opensolaris.org/jive/thread.jspa?messageID=186582

NEWS
Actually I've just noticed that Matt Ahrens has updated the status of this bug 
to need more information and added a comment that the bugfix for 6343667 
involved a major rewrite, which might have removed the problem. Has anyone out 
there got a large zpool running under snv_94 or higher, and if so have you had 
to rebuild a disk yet? If this has fixed both bugs then I'm definitely hoping 
for an early backport to Solaris 10.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Peter Baer Galvin
Ah, thanks much all!

Cindy I'll try to hit the Elling talk too but my time in San Diego will be 
short, unfortunately. I'll keep an eye out for you...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Peter Baer Galvin
Just occurred to me that S10U6 won't support ZFS root install via the GUI 
either?  That will be confusing...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-29 Thread Nigel Smith
Hi Matt
Can you just confirm if that Ethernet capture file, that you made available,
was done on the client, or on the server. I'm beginning to suspect you
did it on the client.

You can get a capture file on the server (OpenSolaris) using the 'snoop'
command, as per one of my previous emails.  You can still view the
capture file with WireShark as it supports the 'snoop' file format.

Normally it would not be too important where the capture was obtained,
but here, where something strange is happening, it could be critical to 
understanding what is going wrong and where.

It would be interesting to do two separate captures - one on the client
and the one on the server, at the same time, as this would show if the
switch was causing disruption.  Try to have the clocks on the client 
server synchronised as close as possible.
Thanks
Nigel Smith
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Peter Baer Galvin
Rats, text install hangs after printing out that very early set of ...  
More debugging...or I could wait for S10U6...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Cindy . Swearingen
Good point and we've tried to document this issue all over the place
and will continue to publicize this fact.

With the new ZFS boot and install features, it is a good idea to read
the docs first. Tell you friends.

I will send out a set of s10 10/08 doc pointers as soon as they are
available. Thanks for reminding me.

Cindy


Peter Baer Galvin wrote:
 Just occurred to me that S10U6 won't support ZFS root install via the GUI 
 either?  That will be confusing...
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-29 Thread Richard Elling
Al Hopper wrote:
 On Wed, Oct 29, 2008 at 8:43 AM, Bob Friesenhahn
 [EMAIL PROTECTED] wrote:
   
 On Wed, 29 Oct 2008, Martti Kuparinen wrote:

 
 Bob Friesenhahn wrote:
   
 AMD Athelon/Opteron dual core likely matches or exceeds
 Intel quad core for ZFS use due to a less bottlenecked memory channel.
 
 How big is the difference? Does anyone have benchmarking results (maybe even
 when using ZFS on Solaris 10)?
   
 The big question would be what should be benchmarked.  ZFS is like a
 big RAM cache.  The more RAM the better.  You would be surprised how
 little disk activity there can really be on systems with a lot of RAM
 as long as synchronous writes are avoided.  As a result, some common
 scenarios mostly exercise RAM rather than the disk channel.

 Unless you need a higher power CPU for other purposes, a ZFS-based
 server should focus on maximizing installed RAM.
 

 Agreed 100%

 It's easy to find DDR2 RAM at around $20/gigabyte (based on 1Gb or 2Gb
 DIMMs) and I've seen some deals as low a $8/Gb for Kingston RAM.
   

ECC?
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Vincent Fox
Dunno about the text installer mentioned in other replies as I never use it.

JumpStart installs working fine though with ZFS root.

I am also in finish script pre-creating some additional filesystems etc.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-29 Thread Rob Logan

  ECC?

$60 unbuffered 4GB 800MHz DDR2 ECC CL5 DIMM (Kit Of 2)
http://www.provantage.com/kingston-technology-kvr800d2e5k2-4g~7KIN90H4.htm

for Intel 32x0 north bridge like
http://www.provantage.com/supermicro-x7sbe~7SUPM11K.htm
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Hotplug issues on USB removable media.

2008-10-29 Thread Miles Nordin
 r == Ross  [EMAIL PROTECTED] writes:

 r I don't remember anyone saying they couldn't be stored, just
 r that if they are stored it's not ZFS' fault if they go corrupt
 r as it's outside of its control.  

they can't be stored because they ``go corrupt'' in a transactional
all-or-nothing way that other checksummed storable things like zpools,
tarballs, zipfiles, do not.  The software's written with the
assumption it'll be used in a pipe, and I think fails appropriately
for that use.  But it's utterly without grace if you accidentally
separate a stored 'zfs send' from your opportunity to re-send it.

 r I'm actually planning to store the zfs send dump on external
 r USB devices myself, but since the USB device will be running
 r on ZFS I expect to be able to find out if there are any
 r corruption problems,

yeah I still think it's not a good idea.  Our best practices are based
on a feel for the statistics of bit-flips and their consequences

If you flip one bit:

 in tar: 
 you get the file back, with the flipped bit, and a warning from tar.

 in UFS: 
 you get the file back with a bit flipped.  

 in ZFS with redundancy:
 you get the file back with the bit unflipped.

 in ZFS without redundancy:
 you are warned about the flipped bit by losing the whole file

 in 'zfs send':
 you lose the ENTIRE FILESYSTEM because of one bit.

'zfs recv' will do a great job of warning you about corruption
problems, whether it has ZFS underneath it or not.  I don't think
anyone could mistake the signal it sends for affection or
negotiability.  That's not the issue.  It's the probability you'll be
able to successfully restore your backup n years from now.

We have a feel, not as good as we should but some, for this
probability given 'tar' on tape, DVDR, disk, stick.  I think 'zfs
send' reduces this probability enough compared to other kinds of
backup that you'd be crazy to use it.  

I also think this observation isn't adequately communicated to the
average sysadmin by the disclaimery-sounding IT warning ``*BRRK*
*BRRK* 'zfs send' is not a Backup Soloooshun!!''  It sounds like
someone wants you to spend more money on CYAware and not, as it
should, like they are begging you to use tar instead.

There are other problems besides fragileness with storing 'zfs send'
that are solved by both 'tar' and piping immediately to 'zfs recv':

 * stream format has not been as forward-portable across zfs versions
   and endiness as the pool structure itself.  You could easily find
   yourself with a stream that won't restore, with no user-readable
   marker as to what it requires to restore itself.  

   The choices are unstable Nevada builds you may not even be able to
   obtain two years from now much less get to boot on the hardware for
   sale at the time, and machine endynesses(!)---much harder to
   iteratively test than versions of GNU tar, if tar had this same
   problem which it doesn't.  It may even demand to be restored ONTO a
   specific version of zpool, so your desperate restore attempt
   strategy is to reinstall an older Nevada, destroy and recreate a
   blank zpool, 'zfs recv', repeat.

   It's a bad situation.  Store a zpool instead and you won't be in it
   because there's a commitment to forward-portability and
   endyness-independence.

 * no equivalent of 'zpool scrub' or 'tar t'.  The only way to verify
   a 'zfs send' dump is to restore it.  

   I think it's a best-practice when making backups to read back what
   you wrote.  This catches medium errors, broken network pipes,
   filesystems you didn't quite sync before ejecting, whatever.  You
   don't need to do silly Windows things like compare the backup to
   the files still on the disk and then scream OMG OMG something
   Changed!, but you ought to at least read it back now if you expect
   to restore it later.

Is something missing from the toolbox?  hell yes!  Even beyond these
inadequacies I'd like a way to restore a ZFS snapshot tree onto an
LVM2 or Netapp snapshot tree.  so I rather think it'll be some
stream-storing rsync or deduplicating GNUtar instead of something
coming from the ZFS project.  (AIUI GNUtar incrementals include whole
changed files, which is great for portability but useless for VM disk
images)

There are interesting questions like, would it drastically improve the
efficiency of incremental backups if ZFS exposed some interface to its
snapshot tree besides two copies of the file?  Maybe the 'zfs send'
stream itself is sufficient exposure, or maybe it needs to be some
file-level rather than filesystem-level map.  But this should probably
come after we have userspace tools that do the job correctly but
inefficiently given two copies of the file.  

And how do we design a storable incremental format that will take
bit-flips gracefully, without seg-faulting and without killing the
entire snapshot tree because one block is lost forever?

Thirdly think it'd be good to write backups so one ``tape'' can be
lost.  ex:


Re: [zfs-discuss] Custom Jumpstart and RAID-10 ZFS rpool

2008-10-29 Thread kristof
I don't think this is possible.

I already tried to add extra vdevs after install, but I got an error message 
telling me that multiple vdevs for rpool are not allowed.

K
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Custom Jumpstart and RAID-10 ZFS rpool

2008-10-29 Thread Kyle McDonald
kristof wrote:
 I don't think this is possible.

 I already tried to add extra vdevs after install, but I got an error message 
 telling me that multiple vdevs for rpool are not allowed.

 K
   
Oh. Ok. Good to know.

I always put all my 'data' diskspace in a separate pool anyway to make 
migration to another host easier, so I haven't actually tried it.

  -Kyle

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] HELP! SNV_97, 98, 99 zfs with iscsitadm and VMWare!

2008-10-29 Thread Miles Nordin
 ns == Nigel Smith [EMAIL PROTECTED] writes:

ns the bnx driver is closed source :-(

The GPL'd Linux driver is contributed by Broadcom: 

 
http://git.kernel.org/?p=linux/kernel/git/stable/linux-2.6.27.y.git;a=blob;f=drivers/net/bnx2.c;h=2486a656f12d9f47ff27ead587e084a3c337a1a3;hb=HEAD

and I believe the chip itself is newer than the Solaris 10 ``all new
bits will be open-source'' pitch.


pgpnZzPhLnb3Y.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-29 Thread Matt Harrison
On Wed, Oct 29, 2008 at 10:01:09AM -0700, Nigel Smith wrote:
 Hi Matt
 Can you just confirm if that Ethernet capture file, that you made available,
 was done on the client, or on the server. I'm beginning to suspect you
 did it on the client.

That capture was done from the client

 You can get a capture file on the server (OpenSolaris) using the 'snoop'
 command, as per one of my previous emails.  You can still view the
 capture file with WireShark as it supports the 'snoop' file format.

I am uploading a snoop from the server to 

http://distfiles.genestate.com/snoop.zip

Please note this snoop will include traffic to ssh as I can't work out how
to filter that out :P

 Normally it would not be too important where the capture was obtained,
 but here, where something strange is happening, it could be critical to 
 understanding what is going wrong and where.
 
 It would be interesting to do two separate captures - one on the client
 and the one on the server, at the same time, as this would show if the
 switch was causing disruption.  Try to have the clocks on the client 
 server synchronised as close as possible.

Clocks are synced via ntp as we're using Active Directory with CIFS.

On another note, I've just moved the offending network to another switch and
it's even worse I think. I've noticed that under high load, the link light
for the server's connection blinks on and off, not quite steadily but about
every 2 seconds.

This appears in /var/adm/messages:

Oct 29 18:24:22 exodus mac: [ID 435574 kern.info] NOTICE: rtls0 link up, 100
Mbps, full duplex
Oct 29 18:24:24 exodus mac: [ID 486395 kern.info] NOTICE: rtls0 link down
Oct 29 18:24:25 exodus mac: [ID 435574 kern.info] NOTICE: rtls0 link up, 100
Mbps, full duplex
Oct 29 18:24:27 exodus mac: [ID 486395 kern.info] NOTICE: rtls0 link down
Oct 29 18:24:28 exodus mac: [ID 435574 kern.info] NOTICE: rtls0 link up, 100
Mbps, full duplex
Oct 29 18:24:30 exodus mac: [ID 486395 kern.info] NOTICE: rtls0 link down
Oct 29 18:24:31 exodus mac: [ID 435574 kern.info] NOTICE: rtls0 link up, 100
Mbps, full duplex

I think it's got to be the NIC, the network runs full duplex quite happily
so I don't think its an auto-neg problem.

Thanks for sticking with this :)

Matt


pgpgEL3gweH2e.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Peter Baer Galvin
Hi Cindy, I googled quite a lot before posting my question. This issue isn't 
mentioned in the ZFS boot FAQ for example or anywhere (that I saw) on the 
Opensolaris ZFS pages. Of course I could have read the ZFS Admin book at docs. 
sun.com...
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs zpool recommendation

2008-10-29 Thread Mike
Hi all,

I have been asked to build a new server and would like to get some opinions on 
how to setup a zfs pool for the application running on the server.  The server 
will be exclusively for running netbackup application.

Now which would be better? setting up a raidz pool with 6x146gig drives or 
setting up 3 mirrors of 2x146gig into a pool?

Currently I have it setup with the mirrors.. but am wondering if the raidz 
possibility would be a better choice.

Thanks in advance.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs zpool recommendation

2008-10-29 Thread Tomas Ögren
On 29 October, 2008 - Mike sent me these 0,7K bytes:

 Hi all,
 
 I have been asked to build a new server and would like to get some
 opinions on how to setup a zfs pool for the application running on the
 server.  The server will be exclusively for running netbackup
 application.
 
 Now which would be better? setting up a raidz pool with 6x146gig
 drives or setting up 3 mirrors of 2x146gig into a pool?
 
 Currently I have it setup with the mirrors.. but am wondering if the
 raidz possibility would be a better choice.

Define better?

The raidz option will give you more storage at less performance.. The
mirror thing has the possibility of achieving higher reliability.. 1 to
3 disks can fail without interruptions, depending on how Murphy picks
them.. The raidz1 one can handle 1 disk only..

/Tomas
-- 
Tomas Ögren, [EMAIL PROTECTED], http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs boot / root in Nevada build 101

2008-10-29 Thread Daryl Doami
Hi Peter,

It's mentioned here under Annoucements:
http://opensolaris.org/os/community/zfs/boot/

It's just not very obvious.

 Original Message 
Subject: Re: [zfs-discuss] zfs boot / root in Nevada build 101
From: Peter Baer Galvin [EMAIL PROTECTED]
To: zfs-discuss@opensolaris.org
Date: Wed Oct 29 11:25:20 2008
 Hi Cindy, I googled quite a lot before posting my question. This issue isn't 
 mentioned in the ZFS boot FAQ for example or anywhere (that I saw) on the 
 Opensolaris ZFS pages. Of course I could have read the ZFS Admin book at 
 docs. sun.com...
   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Custom Jumpstart and RAID-10 ZFS rpool

2008-10-29 Thread Lori Alt

Kyle McDonald wrote:

Ian Collins wrote:
  

Stephen Le wrote:
  


Is it possible to create a custom Jumpstart profile to install Nevada
on a RAID-10 rpool? 

  

No, simple mirrors only.
  

Though a finish sscript could add additional simple mirrors to create 
the config his example would have created.

Pretty sure that's still not RAID10 though.
  

No, this isn't possible.  The attempt to add additional
top-level vdevs will fail because it's not supported for
root pools.

One top-level vdev only.  We hope to relax this
restriction and support both multiple top-level
vdevs and RAID-Z for root pools in the future.
But that work hasn't started yet, so I have no
estimates at all for when it might be available.

lori

And any files laid down by the installer would be constrained to the 
first mirrored pair, only new files would have a chance at be 
distributed over the addtional pairs.


 -Kyle

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
  


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Resilvering after drive failure - keeps restarting - any guesses why/how to fix?

2008-10-29 Thread Michael Stalnaker
All;

I have a large zfs tank with four raidz2 groups in it. Each of these groups
is 11 disks, and I have four hot spare disks in the system.  The system is
running Open Solaris build snv_90.   One of these groups has had a disk
failure, which the OS correctly detected, and replaced with one of the hot
spares, and began rebuilding.

Now it gets interesting. The resilver runs for about 1 hour, then stops. If
I put zpool status ­v in a while loop with a 10 minute sleep, I see the
repair proceed, then with no messages of ANY kind, it¹ll silently quit and
start over. I¹m attaching the output of zpool status ­v from an hour ago and
then from just now below. Has anyone seen this, or have any ideas as to the
cause? Is there a timeout or priorty I need to change in a tuneable or
something?

--Mike



One Hour Ago:

[EMAIL PROTECTED]:~$ zpool status -v
  pool: tank1
 state: DEGRADED
status: One or more devices are faulted in response to persistent errors.
Sufficient replicas exist for the pool to continue functioning in a
degraded state.
action: Replace the faulted device, or use 'zpool clear' to mark the device
repaired.
 scrub: resilver in progress for 0h46m, 3.96% done, 18h39m to go
config:

NAME   STATE READ WRITE CKSUM
tank1  DEGRADED 0 0 0
  raidz2   DEGRADED 0 0 0
c1t13d0ONLINE   0 0 0
spare  DEGRADED 0 0 0
  c1t14d0  FAULTED  0 0 0  too many errors
  c1t11d0  ONLINE   0 0 0
c1t15d0ONLINE   0 0 0
c1t16d0ONLINE   0 0 0
c1t17d0ONLINE   0 0 0
c1t18d0ONLINE   0 0 0
c1t19d0ONLINE   0 0 0
c1t20d0ONLINE   0 0 0
c1t21d0ONLINE   0 0 0
c1t22d0ONLINE   0 0 0
c1t23d0ONLINE   0 0 0
  raidz2   ONLINE   0 0 0
c1t0d0 ONLINE   0 0 0
c1t1d0 ONLINE   0 0 0
c1t2d0 ONLINE   0 0 0
c1t3d0 ONLINE   0 0 0
c1t4d0 ONLINE   0 0 0
c1t5d0 ONLINE   0 0 0
c1t6d0 ONLINE   0 0 0
c1t7d0 ONLINE   0 0 0
c1t8d0 ONLINE   0 0 0
c1t9d0 ONLINE   0 0 0
c1t10d0ONLINE   0 0 0
  raidz2   ONLINE   0 0 0
c2t13d0ONLINE   0 0 0
c2t14d0ONLINE   0 0 0
c2t15d0ONLINE   0 0 0
c2t16d0ONLINE   0 0 0
c2t17d0ONLINE   0 0 0
c2t18d0ONLINE   0 0 0
c2t19d0ONLINE   0 0 0
c2t20d0ONLINE   0 0 0
c2t21d0ONLINE   0 0 0
c2t22d0ONLINE   0 0 0
c2t23d0ONLINE   0 0 0
  raidz2   ONLINE   0 0 0
c2t0d0 ONLINE   0 0 0
c2t1d0 ONLINE   0 0 0
c2t2d0 ONLINE   0 0 0
c2t3d0 ONLINE   0 0 0
c1t24d0ONLINE   0 0 0
c2t5d0 ONLINE   0 0 0
c2t6d0 ONLINE   0 0 0
c2t7d0 ONLINE   0 0 0
c2t8d0 ONLINE   0 0 0
c2t9d0 ONLINE   0 0 0
c2t10d0ONLINE   0 0 0
spares
  c1t11d0  INUSE currently in use
  c2t24d0  AVAIL
  c2t11d0  AVAIL
  c2t4d0   AVAIL

errors: No known data errors

Just Now:

[EMAIL PROTECTED]:~$ zpool status -v
  pool: tank1
 state: DEGRADED
status: One or more devices are faulted in response to persistent errors.
Sufficient replicas exist for the pool to continue functioning in a
degraded state.
action: Replace the faulted device, or use 'zpool clear' to mark the device
repaired.
 scrub: resilver in progress for 0h24m, 2.23% done, 17h51m to go
config:

NAME   STATE READ WRITE CKSUM
tank1  DEGRADED 0 0 0
  raidz2   DEGRADED 0 0 0
c1t13d0ONLINE   0 0 0
spare  DEGRADED 0 0 0
  c1t14d0  FAULTED  0 0 0  too many errors
  c1t11d0  ONLINE   0 0 0
c1t15d0ONLINE   0 0 0
c1t16d0ONLINE   0 0 0
c1t17d0ONLINE   0 0 0
c1t18d0ONLINE   

Re: [zfs-discuss] Hotplug issues on USB removable media.

2008-10-29 Thread Ross
Hi Miles,

I probably should have explained that storing the zfs send on a USB device is 
just one part of the strategy, and in fact that's just our way of getting the 
backup off-site.

Once off-site, we do zfs receive that into another pool, and in fact we plan to 
have two offsite zfs pools, plus standard tar backups on tape (just in case the 
zpools all get corrupted somehow).

We're also considering streaming zfs send to file, and then ftp'ing that file 
to the remote servers.  I don't fancy restarting an entire zfs send because a 
packet got lost 4/5 of the way through.  I'd much rather use restartable FTP, 
and do the zfs receive later on.

Ross
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs zpool recommendation

2008-10-29 Thread Bob Friesenhahn

On Wed, 29 Oct 2008, Tomas Ögren wrote:


The raidz option will give you more storage at less performance.. The
mirror thing has the possibility of achieving higher reliability.. 1 to
3 disks can fail without interruptions, depending on how Murphy picks
them.. The raidz1 one can handle 1 disk only..


Mirrors also offer ease of administration.  Half the drives can be 
administratively removed and the pool will still work.  The sizes of 
drives in individual pairs can be updated one by one in order to offer 
more space rather than requiring all of the drives in a raidz to be 
replaced before the space is seen.  Resilvering will be faster when 
using mirrors since only one disk has to be read to reconstruct the 
data.  If you feel uneasy about reliability, then you can easily use 
tripple mirroring.  However raidz2 offers more reliability than 
dual-mirrors and raidz1.


Bob
==
Bob Friesenhahn
[EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs zpool recommendation

2008-10-29 Thread Bob Friesenhahn
On Wed, 29 Oct 2008, Mike wrote:

 I am not seeing how using raidz would be a performance hit. Usually 
 stripes perform faster than mirrors.

The mirrors load-share and offer a lot more disk seeking capacity 
(more IOPS).

Bob
==
Bob Friesenhahn
[EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs zpool recommendation

2008-10-29 Thread Scott Laird
On Wed, Oct 29, 2008 at 3:42 PM, Mike [EMAIL PROTECTED] wrote:
 By Better I meant the best practice for a server running the Netbackup 
 application.

 I am not seeing how using raidz would be a performance hit. Usually stripes 
 perform faster than mirrors.

raidz performs reads from all devices in parallel, so you get 1
drive's worth of I/O operations, not 6 drives' worth.  With 3 mirrors,
you'd get 6 drives' worth of reads and 3 drives' worth of writes.
Using raidz might get you slightly better read and write bandwidth,
though.


Scott
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] HELP! SNV_97, 98, 99 zfs with iscsitadm and VMWare!

2008-10-29 Thread Tano
ns I will raise this as a bug, but first please would you run 
'/usr/X11/bin/scanpci'
to indentify the exact 'vendor id' and 'device id' for the Broadcom network 
chipset,
and report that back here

Primary network interface Embedded NIC:  
pci bus 0x0005 cardnum 0x00 function 0x00: vendor 0x14e4 device 0x164c
 Broadcom Corporation NetXtreme II BCM5708 Gigabit Ethernet


Plus the two external add on Broadcom cards: (CURRENTLY NOT IN USE)
pci bus 0x000b cardnum 0x00 function 0x00: vendor 0x1166 device 0x0103
 Broadcom EPB PCI-Express to PCI-X Bridge

pci bus 0x000c cardnum 0x00 function 0x00: vendor 0x14e4 device 0x164c
 Broadcom Corporation NetXtreme II BCM5708 Gigabit Ethernet

pci bus 0x000d cardnum 0x00 function 0x00: vendor 0x1166 device 0x0103
 Broadcom EPB PCI-Express to PCI-X Bridge

pci bus 0x000e cardnum 0x00 function 0x00: vendor 0x14e4 device 0x164c
 Broadcom Corporation NetXtreme II BCM5708 Gigabit Ethernet

I will submit the information that you had asked in email very soon.

Tano
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS + OpenSolaris for home NAS?

2008-10-29 Thread Mario Goebbels
Rob Logan wrote:
   ECC?
 
 $60 unbuffered 4GB 800MHz DDR2 ECC CL5 DIMM (Kit Of 2)
 http://www.provantage.com/kingston-technology-kvr800d2e5k2-4g~7KIN90H4.htm

Geez, I have to move to the US for cheap hardware. I've paid 120€ for
exactly that 4GB ECC kit (well, I bought two of these, so 240€) in Germany.

-mg




signature.asc
Description: OpenPGP digital signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs shrinking swap

2008-10-29 Thread Richard Elling
Karl Rossing wrote:
 $zfs list
 NAME   USED  AVAIL  REFER  MOUNTPOINT
 rpool 48.4G  10.6G31K  /rpool
 rpool/ROOT36.4G  10.6G18K  /rpool/ROOT
 rpool/ROOT/snv_90_zfs 29.6G  10.6G  29.3G  /.alt.tmp.b-Ugf.mnt/
 rpool/ROOT/[EMAIL PROTECTED]   319M  -  29.6G  -
 rpool/ROOT/snv_93 6.82G  10.6G  16.3G  /
 rpool/dump3.95G  10.6G  3.95G  -
 rpool/swap8.00G  17.6G  1.05G  -

 I'm on build 93. I would like to shrink swap.

 Is it as simple as zfs set volsize=4G rpool/swap on a live system? Is 
 there anything else I need to do?
   

Yes.  You need to remove it from use first.
# swap -l
swapfile devswaplo   blocks free
/dev/zvol/dsk/rpool/swap 182,2 8  4194296  4194296
# zfs get volsize rpool/swap
NAMEPROPERTY  VALUE  SOURCE
rpool/swap  volsize   2G -
# swap -d /dev/zvol/dsk/rpool/swap
# swap -l
No swap devices configured
# zfs set volsize=1G rpool/swap
# /sbin/swapadd
# swap -l
swapfile devswaplo   blocks free
/dev/zvol/dsk/rpool/swap 182,2 8  2097144  2097144


Note: in the first swap -l above, blocks == free, so the swap device
was not actually being used.  If your swap device is in use, then you
might not be able to delete it, so you'll have to make another, then
shuffle the data around.
 -- richard
 Thanks
 Karl













 CONFIDENTIALITY NOTICE:  This communication (including all attachments) is
 confidential and is intended for the use of the named addressee(s) only and
 may contain information that is private, confidential, privileged, and
 exempt from disclosure under law.  All rights to privilege are expressly
 claimed and reserved and are not waived.  Any use, dissemination,
 distribution, copying or disclosure of this message and any attachments, in
 whole or in part, by anyone other than the intended recipient(s) is strictly
 prohibited.  If you have received this communication in error, please notify
 the sender immediately, delete this communication from all data storage
 devices and destroy all hard copies.
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
   

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-29 Thread Nigel Smith
Hi Matt
In your previous capture, (which you have now confirmed was done
on the Windows client), all those 'Bad TCP checksum' packets sent by the 
client, 
are explained, because you must be doing hardware TCP checksum offloading
on the client network adaptor.  WireShark will capture the packets before
that hardware calculation is done, so the checksum all appear to be wrong,
as they have not yet been calculated!

  http://wiki.wireshark.org/TCP_checksum_offload
  http://www.wireshark.org/docs/wsug_html_chunked/ChAdvChecksums.html

Ok, so lets look at the new capture, 'snoop'ed on the OpenSolaris box.

I was surprised how small that snoop capture file was
 - only 753400 bytes after unzipping.
I soon realized why...

The strange thing is that I'm only seeing half of the conversation!
I see packets sent from client to server.
That is from source: 10.194.217.10 to destination: 10.194.217.3

I can also see some packets from
source: 10.194.217.5 (Your AD domain controller) to destination  10.194.217.3

But you've not capture anything transmitted from your
OpenSolaris server - source: 10.194.217.3

(I checked, and I did not have any filters applied in WireShark
that would cause the missing half!)
Strange! I'm not sure how you did that.

The half of the conversation that I can see looks fine - there
does not seem to be any problem.  I'm not seeing any duplication
of ACK's from the client in this capture.  
(So again somewhat strange, unless you've fixed the problem!)

I'm assuming your using a single network card in the Solaris server, 
but maybe you had better just confirm that.

Regarding not capturing SSH traffic and only capturing traffic from
( hopefully to) the client, try this:

 # snoop -o test.cap -d rtls0 host 10.194.217.10 and not port 22

Regarding those 'link down', 'link up' messages, '/var/adm/messages'.
I can tie up some of those events with your snoop capture file,
but it just shows that no packets are being received while the link is down,
which is exactly what you would expect.
But dropping the link for a second will surely disrupt your video playback!

If the switch is ok, and the cable from the switch is ok, then it does
now point towards the network card in the OpenSolaris box.  
Maybe as simple as a bad mechanical connection on the cable socket

BTW, just run '/usr/X11/bin/scanpci'  and identify the 'vendor id' and
'device id' for the network card, just in case it turns out to be a driver bug.
Regards
Nigel Smith
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] diagnosing read performance problem

2008-10-29 Thread Matt Harrison
On Wed, Oct 29, 2008 at 05:32:39PM -0700, Nigel Smith wrote:
 Hi Matt
 In your previous capture, (which you have now confirmed was done
 on the Windows client), all those 'Bad TCP checksum' packets sent by the 
 client, 
 are explained, because you must be doing hardware TCP checksum offloading
 on the client network adaptor.  WireShark will capture the packets before
 that hardware calculation is done, so the checksum all appear to be wrong,
 as they have not yet been calculated!

I know that the client I was using has an nForce board with nVidia network
controllers. There is an option to offload to hardware but I believe that
was disabled.

 The strange thing is that I'm only seeing half of the conversation!
 I see packets sent from client to server.
 That is from source: 10.194.217.10 to destination: 10.194.217.3
 
 I can also see some packets from
 source: 10.194.217.5 (Your AD domain controller) to destination  10.194.217.3
 
 But you've not capture anything transmitted from your
 OpenSolaris server - source: 10.194.217.3
 
 (I checked, and I did not have any filters applied in WireShark
 that would cause the missing half!)
 Strange! I'm not sure how you did that.

I believe i was using the wrong filter expression...my bad :(

 The half of the conversation that I can see looks fine - there
 does not seem to be any problem.  I'm not seeing any duplication
 of ACK's from the client in this capture.  
 (So again somewhat strange, unless you've fixed the problem!)

 I'm assuming your using a single network card in the Solaris server, 
 but maybe you had better just confirm that.

Confirmed, there is a single PCI NIC that i'm using (there is the dual
onboard but they don't work for me anymore).
 
 Regarding not capturing SSH traffic and only capturing traffic from
 ( hopefully to) the client, try this:
 
  # snoop -o test.cap -d rtls0 host 10.194.217.10 and not port 22

Much better thanks. I am attaching a second snoop from the server with the
full conversation.

http://distfiles.genestate.com/snoop2.zip

Incidentally, this is talking to a different client, which although doesn't
show checksum errors, does still have a load of duplicate ACKs. If this
confuses the issue, I can do it from the old client as soon as it becomes
free.

 Regarding those 'link down', 'link up' messages, '/var/adm/messages'.
 I can tie up some of those events with your snoop capture file,
 but it just shows that no packets are being received while the link is down,
 which is exactly what you would expect.
 But dropping the link for a second will surely disrupt your video playback!
 
 If the switch is ok, and the cable from the switch is ok, then it does
 now point towards the network card in the OpenSolaris box.  
 Maybe as simple as a bad mechanical connection on the cable socket

Very possible. I have an Intell Pro 1000 and a new GB switch on the way.

 BTW, just run '/usr/X11/bin/scanpci'  and identify the 'vendor id' and
 'device id' for the network card, just in case it turns out to be a driver 
 bug.

pci bus 0x0001 cardnum 0x06 function 0x00: vendor 0x10ec device 0x8139
 Realtek Semiconductor Co., Ltd. RTL-8139/8139C/8139C+

and the two onboards that no longer function:

pci bus 0x cardnum 0x08 function 0x00: vendor 0x10de device 0x0373
 nVidia Corporation MCP55 Ethernet

pci bus 0x cardnum 0x09 function 0x00: vendor 0x10de device 0x0373
 nVidia Corporation MCP55 Ethernet

Thanks

Matt


pgpG9O5VYjefY.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Fwd: zpool import problem

2008-10-29 Thread Terry Heatlie
oops, meant to reply-all...

-- Forwarded message --
From: Terry Heatlie [EMAIL PROTECTED]
Date: Wed, Oct 29, 2008 at 8:14 PM
Subject: Re: [zfs-discuss] zpool import problem
To: Eric Schrock [EMAIL PROTECTED]


well, this does seem to be the case:

bash-3.2# dtrace -s raidz_open2.d
run 'zpool import' to generate trace

1145357764648 BEGIN RAIDZ OPEN
1145357764648 config asize = 1600340623360
1145357764648 config ashift = 9
1145358131986 child[0]: asize = 320071851520, ashift = 9
1145358861331 child[1]: asize = 400088457216, ashift = 9
1145396437606 child[2]: asize = 400088457216, ashift = 9
1145396891657 child[3]: asize = 320072933376, ashift = 9
1145397584944 child[4]: asize = 400087375360, ashift = 9
1145397920504 child[5]: asize = 400087375360, ashift = 9
1145398947963 asize = 1600335380480
1145398947963 ashift = 9
1145398947963 END RAIDZ OPEN

But I still don't see a difference between the partition maps of the drive
with only 2 labels and a good one... c2 is bad, c4 is good...

# prtvtoc /dev/dsk/c2d0p0  /tmp/vtoc_c2
# prtvtoc /dev/dsk/c4d0p0  /tmp/vtoc_c4
# diff /tmp/vtoc_c2 /tmp/vtoc_c4
1c1
 * /dev/dsk/c2d0p0 partition map
---
 * /dev/dsk/c4d0p0 partition map
#


On Tue, Oct 28, 2008 at 3:53 AM, Eric Schrock [EMAIL PROTECTED] wrote:

 These are the symptoms of a shrinking device in a RAID-Z pool.  You can
 try to run the attached script during the import to see if this the
 case.  There's a bug filed on this, but I don't have it handy.

 [...]
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss