Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-05 Thread Blake
How I do recursive, selective snapshot destroys:

http://blog.clockworm.com/2008/03/remove-old-zfs-snapshots.html




>
> Saturday, February 28, 2009, 10:14:20 PM, you wrote:
>
> TW> I would really add : make insane zfs destroy <-r|> poolname  as
> TW> harmless as zpool destroy poolname (recoverable)
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-04 Thread Robert Milkowski
Hello Thomas,

Saturday, February 28, 2009, 10:14:20 PM, you wrote:

TW> I would really add : make insane zfs destroy <-r|> poolname  as
TW> harmless as zpool destroy poolname (recoverable)

TW>   zfs destroy <-r|> poolname<|/filesystem>

TW>   this should behave like that:

TW>   o snapshot the filesystem to be deleted (each, name it 
@deletedby_operatorname_date)

TW>   o hide the snapshot as long as the pool has enough space and
TW>  property snapshotbeforedelete=on (default off) is set 'on'

[...]

I can't disagree more. You can always write a small script/wrapper
doing something like this for you.

-- 
Best regards,
 Robert Milkowski
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-04 Thread Robert Milkowski
Hello George,

Tuesday, March 3, 2009, 3:01:43 PM, you wrote:

GW> Matthew Ahrens wrote:
>>>
>>> pool-shrinking (and an option to shrink disk A when i want disk B to
>>> become a mirror, but A is a few blocks bigger)
>>
>> I'm working on it.
>>


>>>
>>> automated installgrub when mirroring an rpool
GW> I'm working on it.


 Wouldn't it be useful for everyone if the ZFS team could set-up and
 maintain a publicly available web page with a list of features
 currently being worked on or scheduled for a near future with a one
 or two sentence description if needed? That would probably reduce the
 amount of questions being asked hundreds of times - like - can I
 shrink a pool?

 I know that you probably can't share all of it but still some of it
 would be useful for the community.

-- 
Best regards,
 Robert Milkowski
   http://milek.blogspot.com

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-03 Thread Matthew Ahrens

Greg Mason wrote:

Just my $0.02, but would pool shrinking be the same as vdev evacuation?


Yes.


basically, what I'm thinking is:

zpool remove mypool 

Allow time for ZFS to vacate the vdev(s), and then light up the "OK to 
remove" light on each evacuated disk.


That's the goal.

--matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-03 Thread Greg Mason

Just my $0.02, but would pool shrinking be the same as vdev evacuation?

I'm quite interested in vdev evacuation as an upgrade path for 
multi-disk pools. This would be yet another reason to for folks to use 
ZFS at home (you only have to buy cheap disks), but it would also be a 
good to have that ability from an enterprise perspective, as I'm sure 
we've all engineered ourselves into a corner one time or another...


It's a much cleaner, safer, and possibly much faster alternative to 
systematically pulling drives and letting zfs resilver onto a larger 
disk, in order to upgrade a pool in-place, and in production.


basically, what I'm thinking is:

zpool remove mypool 

Allow time for ZFS to vacate the vdev(s), and then light up the "OK to 
remove" light on each evacuated disk.


-Greg

Blake Irvin wrote:

Shrinking pools would also solve the right-sizing dilemma.

Sent from my iPhone

On Feb 28, 2009, at 3:37 AM, Joe Esposito  wrote:


I'm using opensolaris and zfs at my house for my photography storage
as well as for an offsite backup location for my employer and several
side web projects.

I have an 80g drive as my root drive.  I recently took posesion of 2
74g 10k drives which I'd love to add as a mirror to replace the 80 g
drive.

From what I gather it is only possible if I zfs export my storage
array and reinstall solaris on the new disks.

So I guess I'm hoping zfs shrink and grow commands show up sooner or 
later.


Just a data point.

Joe Esposito
www.j-espo.com

On 2/28/09, "C. Bergström"  wrote:

Blake wrote:

Gnome GUI for desktop ZFS administration



On Fri, Feb 27, 2009 at 9:13 PM, Blake  wrote:


zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B to
become a mirror, but A is a few blocks bigger)


This may be interesting... I'm not sure how often you need to shrink a
pool though?  Could this be classified more as a Home or SME level 
feature?

install to mirror from the liveCD gui


I'm not working on OpenSolaris at all, but for when my projects
installer is more ready /we/ can certainly do this..

zfs recovery tools (sometimes bad things happen)


Agreed.. part of what I think keeps zfs so stable though is the complete
lack of dependence on any recovery tools..  It forces customers to bring
up the issue instead of dirty hack and nobody knows.

automated installgrub when mirroring an rpool


This goes back to an installer option?

./C

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-03 Thread Richard Elling

George Wilson wrote:
Along these lines you can envision a restore tool that is capable of 
reading multiple 'zfs send' streams to construct the various versions 
of files which are available. In addition, it would be nice if the 
tool could read in the streams and then make it easy to traverse and 
construct a single file from all available streams. For example, if I 
have 5 send streams then the tool would be able to ingest all the data 
and provide a directory structure similar to .zfs which would allow 
you to restore any file which is completely intact.


In essence, this is how HSM works (qv ADM project).  You get a view of the
file system for which the data in the files may reside on media elsewhere.
Good stuff.
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-03 Thread George Wilson

Richard Elling wrote:

David Magda wrote:


On Feb 27, 2009, at 18:23, C. Bergström wrote:


Blake wrote:

Care to share any of those in advance?  It might be cool to see input
from listees and generally get some wheels turning...


raidz boot support in grub 2 is pretty high on my list to be honest..

Which brings up another question of where is the raidz stuff mostly?

usr/src/uts/common/fs/zfs/vdev_raidz.c ?

Any high level summary, docs or blog entries of what the process 
would look like for a raidz boot support is also appreciated.


Given the threads that have appeared on this list lately, how about 
codifying / standardizing the output of "zfs send" so that it can be 
backed up to tape? :)


It wouldn't help.  zfs send is a data stream which contains parts of 
files,

not files (in the usual sense), so there is no real way to take a send
stream and extract a file, other than by doing a receive.

At the risk of repeating the Best Practices Guide (again):
The zfs send and receive commands do not provide an enterprise-level 
backup solution.

-- richard
Along these lines you can envision a restore tool that is capable of 
reading multiple 'zfs send' streams to construct the various versions of 
files which are available. In addition, it would be nice if the tool 
could read in the streams and then make it easy to traverse and 
construct a single file from all available streams. For example, if I 
have 5 send streams then the tool would be able to ingest all the data 
and provide a directory structure similar to .zfs which would allow you 
to restore any file which is completely intact.


- George
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-03 Thread George Wilson

Matthew Ahrens wrote:

Blake wrote:

zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B to
become a mirror, but A is a few blocks bigger)


I'm working on it.


install to mirror from the liveCD gui

zfs recovery tools (sometimes bad things happen)


We've actually discussed this at length and there will be some work 
started soon.


automated installgrub when mirroring an rpool

I'm working on it.

- George


--matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread David Magda

On Mar 2, 2009, at 18:37, Miles Nordin wrote:

And I'm getting frustrated pointing out these issues for the 10th  
time [...]


http://www.xkcd.com/386/

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread David Magda

On Mar 2, 2009, at 19:31, David wrote:

So nobody is interested in Raidz grow support? i.e. you have 4 disks  
is a
raidz and you only have room for a 5th disk(physically), so you add  
the 5th
disk to the raidz. It would be a great feature for a home server and  
its the

only thing stopping solaris going on my home file server.


The block pointer rewrite / relocate code has already been written, so  
once that code is pushed back it:


will allow us to traverse all the blocks in the pool and move them  
around. This will be used to move all the used space off a disk so  
that it can be removed. But we realized that since bp relocate has  
to visit all the blocks in the pool, it can also be used to scrub or  
resilver the pool.



http://blogs.sun.com/ahrens/entry/new_scrub_code

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread David
So nobody is interested in Raidz grow support? i.e. you have 4 disks is a
raidz and you only have room for a 5th disk(physically), so you add the 5th
disk to the raidz. It would be a great feature for a home server and its the
only thing stopping solaris going on my home file server.



On Tue, Mar 3, 2009 at 12:06 AM, Miles Nordin  wrote:

> > "cb" == C Bergström  writes:
>
>cb> ideas for good zfs GSoC projects, but wanted to stir some
>cb> interest.
>
> Read-only vdev support.
>
> 1. possibility to import a zpool on DVD.  All filesystems within would
>   be read-only.  DVD should be scrubbable: result would be a list of
>   files with defects.
>
> 2. possible to import a zpool of writeable devices as ``read-mostly''.
>   All filesystems within it would be read-only, but you could still
>   scrub it, and future ``vdev shrink'' features would work on it, and
>   the existing silent-fsck features like label rewriting or ZIL
>   rolling or whatever would occur.  This would be used for creating
>   the DVD master images above.
>
> 3. possible to add a read-write vdev to the read-only zpool, and make
>   the zpool become read-write.
>
> 4. (maybe) possible to import zpool as non-persistent.  All writes
>   would be stored in ARC/L2ARC, and the max size of writes could be
>   capped.
>
>
> use cases:
>
>  0. mounting other machines' pools without rewriting labels
>
>  1. live CD's (obvious),
>
>  2. SOX compliant backups, when backups must be done to WORM media.
>
>you import several read-only vdev's and keep attaching one
>read-write vdev to the ``end'' of the pool when you want to add a
>blob of data.  the first and each successive vdev becomes full,
>incremental, incremental, incremental---thus replication and
>backup converge.
>
>Through (2) ``read-mostly'' import with bp-rewrite it's possible
>to condense vdev's, for example to coalesce fifteen WORM daily
>incrementals whose write fuse has already blown, into a single new
>big vdev, and then detach the fifteen tiny ones.
>
>A variety of apparent pools are importable depending on what
>combination of vdev's you would like to use.  'zpool import' would
>have to become a bit more complicated, to list the workable
>combinations and mask unwanted devices as you order it.
>
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Miles Nordin
> "cb" == C Bergström  writes:

cb> ideas for good zfs GSoC projects, but wanted to stir some
cb> interest.

Read-only vdev support.

1. possibility to import a zpool on DVD.  All filesystems within would
   be read-only.  DVD should be scrubbable: result would be a list of
   files with defects.

2. possible to import a zpool of writeable devices as ``read-mostly''.
   All filesystems within it would be read-only, but you could still
   scrub it, and future ``vdev shrink'' features would work on it, and
   the existing silent-fsck features like label rewriting or ZIL
   rolling or whatever would occur.  This would be used for creating
   the DVD master images above.

3. possible to add a read-write vdev to the read-only zpool, and make
   the zpool become read-write.

4. (maybe) possible to import zpool as non-persistent.  All writes
   would be stored in ARC/L2ARC, and the max size of writes could be
   capped.


use cases: 

 0. mounting other machines' pools without rewriting labels

 1. live CD's (obvious), 

 2. SOX compliant backups, when backups must be done to WORM media.  

you import several read-only vdev's and keep attaching one
read-write vdev to the ``end'' of the pool when you want to add a
blob of data.  the first and each successive vdev becomes full,
incremental, incremental, incremental---thus replication and
backup converge.

Through (2) ``read-mostly'' import with bp-rewrite it's possible
to condense vdev's, for example to coalesce fifteen WORM daily
incrementals whose write fuse has already blown, into a single new
big vdev, and then detach the fifteen tiny ones.

A variety of apparent pools are importable depending on what
combination of vdev's you would like to use.  'zpool import' would
have to become a bit more complicated, to list the workable
combinations and mask unwanted devices as you order it.


pgpsLZiG3CGUm.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Miles Nordin
> "ma" == Matthew Ahrens  writes:

ma> We will soon be changing the manpage to indicate that the zfs
ma> send stream will be receivable on all future versions of ZFS.

still not strong enough statement for this case:

   old system   new system

1.  zfs send   --backup--->  zfs recv

2.  zfs recv  <--restore---  zfs send   ***FAIL***


...as I have said a handful of times already.  Is this message short
enough for everyone to read it?  I'm almost certain it isn't.


pgpa95lRZ0eXz.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Miles Nordin
> "dm" == David Magda  writes:

dm> Yes, in its current state; hopefully that will change some
dm> point in the future

I don't think it will or should.  A replication tool and a backup tool
seem similar, but they're not similar enough.

With replication, you want an exact copy, and if for some reason the
copy is not exact then you need something more: you want atomicity of
operations so the overwatcher scheduler:

 * can safely retry the send|recv until it works,

 * can always tell its minder with certainty how much has safely been
   replicated so far,

 * can attempt further replication without risking existing data.  

These things are a bit hard to achieve.  And zfs send|recv does them:

 * If a single bit is flipped the whole stream should be discarded

 * If, on a 'send' timeline of , , ,
   , one of the preceeding blobs did not make it, or
   became bad on the replication target (because somebody wrote to it,
   for example---a FAQ), it should become impossible to restore
   further incremental backups.  The error should not be best-effort
   worked around, or simply silently concealed, the way it is and
   should be with restoring incremental backups.

 * reslicing the unit of replication after writing the stream is
   irrelevant, because you can just reslice on the replication-source
   if you need to do this.  The possibility of reslicing interferes
   with the atomicity I spoke of which makes the replication scheduler
   so much easier to get correct.

 * there's no need to test a stream's validity without restoring it.
   The replication target will always be available and have enough
   free space to test-by-receiving.

 * there's no need to restore the stream on an unimagined future
   filesystem.  It's more important that all fancyfeatures, ACL's,
   gizmos, properties, compression modes, record sizes, whatever, make
   it to the replication target exactly to avoid surprises.  No one is
   worried about data being locked in to a certain filesystem because
   it's all there for you to get with rsync on both replication source
   and target.

Try to use such a tool for backup, and you court disaster.  Your pile
of backups becomes an increasingly large time-lapse gamma ray
detector, which signals a ray's presence by destroying ALL the data
not just the bit, not even just the file, that the ray hit.  The
impossibility of reslicing (extracting a single file from the backup)
means that you can find yourself needing 48TB of empty disk on a
development system somewhere to get out a 100kB file locked inside the
atomic blob, which is an unacceptable burden in time and expense.  The
other points have obvious problems for backups, too---I'll leave
inventing imaginary restore scenarios as an exercise for the reader.
All these harmful points are things that replication wants/needs.  The
goals are incompatible.

If there's going to be a snapshot-aware incremental backup tool for
ZFS, I do not think zfs send|recv is a good starting point.  And I'm
getting frustrated pointing out these issues for the 10th time---it
seems like, I mention five relatively unsolveable problems, and people
seize onto the easiest one, misinterpret it, and then forget the other
four.

the versioning issue (NOT mentioned above) is a problem for
replication among different Solaris releases, not just backup.  It
means you could potentially have to upgrade a whole mess of machines
at once.  At the very least you ought to be able to upgrade the target
before you upgrade the source, so you don't have to break replication
while doing the upgrade---coincidentally that's the right direction
for upgrade-test-downgrade, too, since it's on the target that you'd
possibly have to destroy the zpools if you decide you need to
downgrade.  We should want this and don't have it yet.

It makes having a single backup pool for a lab full of different-aged
systems impossible (even with backward-only compatibility, how do you
restore?), so it is worth solving for that use case too.  I think the
'zfs send' format of a given filesystem should be bit-for-bit
identical given a certain ZFS version, irrespective of zpool version
or kernel release on the sending system.  That's enough to solve the
single-lab-backup-pool problem, and it's also
regression-testable---keep some old streams around, recv them into the
pool under test, send them back out, and make sure they come out
identical.  And the 'zfs recv' panics need fixing.  those would both
be great things, but they would *NOT* make zfs send|recv into a backup
system!  They would make it into a better replication system.

zfs send|recv will not become backup tools if you manage to check off
a few other list-of-things-to-fix, either.  They can't be both a good
replication system and a good backup system at the same time.

no, I don't think the si wiki explains the full size of the issue
adequately.  It makes it sound like the tools are just a little too
small, and maybe someday they will get bigg

Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Nicolas Williams
On Sat, Feb 28, 2009 at 09:45:12PM -0600, Mike Gerdts wrote:
> On Sat, Feb 28, 2009 at 8:34 PM, Nicolas Williams
> > Right, but normally each head in a cluster will have only one pool
> > imported.
> 
> Not necessarily.  Suppose I have a group of servers with a bunch of
> zones.  Each zone represents a service group that needs to
> independently fail over between servers.  In that case, I may have a
> zpool per zone.  It seems this is how it is done in the real world.[1]

That model forces you to allocate storage.  If you're willing to take
the pain of managine storage at a higher level of granularity then
you're welcome to it.  As has just been posted (and as anyone who
read Matt Ahrens' blog entry on "block pointer rewrite" could have read
between the lines), pool shrinking is coming.  But I still don't
recommend it.

> > The Sun Storage 7xxx do this.  One pool per-head, two pools altogether
> > in a cluster.
> 
> Makes sense for your use case.  If you are looking at a zpool per
> zone, it is likely a zpool created on a LUN provided by a Sun Storage
> 7xxx that is presented to multiple hosts.  That is, ZFS on top of ZFS.

Running ZFS on an LUN backed by a ZFS zvol is fine.  That does not force
you to manage storage at the physical partition/cylinder/chunk-of-
sectors level.

> > This gets you back into managing physical space allocation.  Do you
> > really want that?  If you're using zvols you can do "array based copies"
> > of you zvols.  If you're using filesystems then you should just use
> > normal backup tools.
> 
> There are times when you have no real choice.  If a regulation or a
> lawyer's interpretation of a regulation says that you need to have
> physically separate components, you need to have physically separate
> components.  If your disaster recovery requirements mean that you need
> to have a copy of data at a different site and array based copies have
> historically been used - it is unlikely that "while true ; do zfs send
> | ssh | zfs receive" will be adapted in the first round of
> implementation.  Given this, zvols don't do it today.

Physically separate components accessed by a single node are not likely
to meet such a regulation -- either that or the lawyers need technology
training.

Chinese wall requirements might need to be met by physically separate
storage heads, with physically separate networks.  Except that no one
builds networks like that any more, at least not in the corporat world
-- it's all switches and VLANs, so there will be some degree of logical
separation, and the question really is: to what degree?

Nico
-- 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Matthew Ahrens

Blake wrote:

zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B to
become a mirror, but A is a few blocks bigger)


I'm working on it.


install to mirror from the liveCD gui

zfs recovery tools (sometimes bad things happen)

automated installgrub when mirroring an rpool


--matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Matthew Ahrens

David Magda wrote:
Given the threads that have appeared on this list lately, how about 
codifying / standardizing the output of "zfs send" so that it can be 
backed up to tape? :)


We will soon be changing the manpage to indicate that the zfs send stream 
will be receivable on all future versions of ZFS.  However, as has been 
mentioned, that does not necessarily make it an enterprise backup solution :-)


--matt
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-02 Thread Blake
excellent!  i wasn't sure if that was the case, though i had heard rumors.


On Mon, Mar 2, 2009 at 12:36 PM, Matthew Ahrens  wrote:
> Blake wrote:
>>
>> zfs send is great for moving a filesystem with lots of tiny files,
>> since it just handles the blocks :)
>>
>>
>>
>> I'd like to see:
>>
>> pool-shrinking (and an option to shrink disk A when i want disk B to
>> become a mirror, but A is a few blocks bigger)
>
> I'm working on it.
>
>> install to mirror from the liveCD gui
>>
>> zfs recovery tools (sometimes bad things happen)
>>
>> automated installgrub when mirroring an rpool
>
> --matt
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-03-01 Thread Kees Nuyt
On Sat, 28 Feb 2009 21:45:12 -0600, Mike Gerdts
 wrote:

>On Sat, Feb 28, 2009 at 8:34 PM, Nicolas Williams
> wrote:

[snip]

>> Right, but normally each head in a cluster will have only one pool
>> imported.
>
>Not necessarily.  Suppose I have a group of servers with a bunch of
>zones.  Each zone represents a service group that needs to
>independently fail over between servers.  In that case, I may have a
>zpool per zone.  It seems this is how it is done in the real world.[1]
>1. Upton, Tom. "A  Conversation with Jason Hoffman."  ACM Queue.
>January/February 2008. 9.

Exactly. Or even a zpool per application, if there is more
than one application in a zone. In that sense, I would call
a zpool a "unit of maintenance". Or a "unit of failure".
-- 
  (  Kees Nuyt
  )
c[_]
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Mike Gerdts
On Sat, Feb 28, 2009 at 8:34 PM, Nicolas Williams
 wrote:
> On Sat, Feb 28, 2009 at 05:19:26PM -0600, Mike Gerdts wrote:
>> On Sat, Feb 28, 2009 at 4:33 PM, Nicolas Williams
>>  wrote:
>> > On Sat, Feb 28, 2009 at 10:44:59PM +0100, Thomas Wagner wrote:
>> >> > >> pool-shrinking (and an option to shrink disk A when i want disk B to
>> >> > >> become a mirror, but A is a few blocks bigger)
>> >> >  This may be interesting... I'm not sure how often you need to shrink a 
>> >> > pool
>> >> >  though?  Could this be classified more as a Home or SME level feature?
>> >>
>> >> Enterprise level especially in SAN environments need this.
>> >>
>> >> Projects own theyr own pools and constantly grow and *shrink* space.
>> >> And they have no downtime available for that.
>> >
>> > Multiple pools on one server only makes sense if you are going to have
>> > different RAS for each pool for business reasons.  It's a lot easier to
>> > have a single pool though.  I recommend it.
>>
>> Other scenarios for multiple pools include:
>>
>> - Need independent portability of data between servers.  For example,
>> in a HA cluster environment, various workloads will be mapped to
>> various pools.  Since ZFS does not do active-active clustering, a
>> single pool for anything other than a simple active-standby cluster is
>> not useful.
>
> Right, but normally each head in a cluster will have only one pool
> imported.

Not necessarily.  Suppose I have a group of servers with a bunch of
zones.  Each zone represents a service group that needs to
independently fail over between servers.  In that case, I may have a
zpool per zone.  It seems this is how it is done in the real world.[1]

1. Upton, Tom. "A  Conversation with Jason Hoffman."  ACM Queue.
January/February 2008. 9.

> The Sun Storage 7xxx do this.  One pool per-head, two pools altogether
> in a cluster.

Makes sense for your use case.  If you are looking at a zpool per
zone, it is likely a zpool created on a LUN provided by a Sun Storage
7xxx that is presented to multiple hosts.  That is, ZFS on top of ZFS.

>> - Array based copies are needed.  There are times when copies of data
>> are performed at a storage array level to allow testing and support
>> operations to happen "on different spindles".  For example, in a
>> consolidated database environment, each database may be constrained to
>> a set of spindles so that each database can be replicated or copied
>> independent of the various others.
>
> This gets you back into managing physical space allocation.  Do you
> really want that?  If you're using zvols you can do "array based copies"
> of you zvols.  If you're using filesystems then you should just use
> normal backup tools.

There are times when you have no real choice.  If a regulation or a
lawyer's interpretation of a regulation says that you need to have
physically separate components, you need to have physically separate
components.  If your disaster recovery requirements mean that you need
to have a copy of data at a different site and array based copies have
historically been used - it is unlikely that "while true ; do zfs send
| ssh | zfs receive" will be adapted in the first round of
implementation.  Given this, zvols don't do it today.

When you have a smoking hole, the gap in transactions left by normal
backup tools is not always good enough - especially if some of that
smoke is coming from the tape library.  Array based replication tends
to allow you to keep much tighter tolerances on just how many
committed transactions you are willing to lose.

-- 
Mike Gerdts
http://mgerdts.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Nicolas Williams
On Sat, Feb 28, 2009 at 05:19:26PM -0600, Mike Gerdts wrote:
> On Sat, Feb 28, 2009 at 4:33 PM, Nicolas Williams
>  wrote:
> > On Sat, Feb 28, 2009 at 10:44:59PM +0100, Thomas Wagner wrote:
> >> > >> pool-shrinking (and an option to shrink disk A when i want disk B to
> >> > >> become a mirror, but A is a few blocks bigger)
> >> >  This may be interesting... I'm not sure how often you need to shrink a 
> >> > pool
> >> >  though?  Could this be classified more as a Home or SME level feature?
> >>
> >> Enterprise level especially in SAN environments need this.
> >>
> >> Projects own theyr own pools and constantly grow and *shrink* space.
> >> And they have no downtime available for that.
> >
> > Multiple pools on one server only makes sense if you are going to have
> > different RAS for each pool for business reasons.  It's a lot easier to
> > have a single pool though.  I recommend it.
> 
> Other scenarios for multiple pools include:
> 
> - Need independent portability of data between servers.  For example,
> in a HA cluster environment, various workloads will be mapped to
> various pools.  Since ZFS does not do active-active clustering, a
> single pool for anything other than a simple active-standby cluster is
> not useful.

Right, but normally each head in a cluster will have only one pool
imported.

The Sun Storage 7xxx do this.  One pool per-head, two pools altogether
in a cluster.

> - Array based copies are needed.  There are times when copies of data
> are performed at a storage array level to allow testing and support
> operations to happen "on different spindles".  For example, in a
> consolidated database environment, each database may be constrained to
> a set of spindles so that each database can be replicated or copied
> independent of the various others.

This gets you back into managing physical space allocation.  Do you
really want that?  If you're using zvols you can do "array based copies"
of you zvols.  If you're using filesystems then you should just use
normal backup tools.

Nico
-- 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Bob Netherton

> Multiple pools on one server only makes sense if you are going to have
> different RAS for each pool for business reasons.  It's a lot easier to
> have a single pool though.  I recommend it.

A couple of other things to consider to go with that recommendation.

- never build a pool larger than you are willing to restore.   Bad
things can still happen that would require you to restore the entire
pool.  Convenience and SLAs aren't always in agreement :-)   The
advances in ZFS availability might make me look at my worst case
restore scenario a little different though - but there will still
be a restore case that worries me.

- as I look at the recent lifecycle improvements with zones (in the
Solaris 10 context of zones), I really like upgrade on attach.   That
means I will be slinging zones more freely.   So I need to design my
pools to match that philosophy.

- if you are using clustering technologies, pools will go hand in
hand with failover boundaries.   So if I have multiple failover
zones, I will have multiple pools.


Bob


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Aaron Blew
Absolutely agree. I'l love to be able to free up some LUNs that I
don't need in the pool any more.

Also, concatenation of devices in a zpool would be great for devices
that have LUN limits.  It also seems like it may be an easy thing to
implement.

-Aaron

On 2/28/09, Thomas Wagner  wrote:
>> >> pool-shrinking (and an option to shrink disk A when i want disk B to
>> >> become a mirror, but A is a few blocks bigger)
>>  This may be interesting... I'm not sure how often you need to shrink a
>> pool
>>  though?  Could this be classified more as a Home or SME level feature?
>
> Enterprise level especially in SAN environments need this.
>
> Projects own theyr own pools and constantly grow and *shrink* space.
> And they have no downtime available for that.
>
> give a +1 if you agree
>
> Thomas
>
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>

-- 
Sent from my mobile device
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Mike Gerdts
On Sat, Feb 28, 2009 at 4:33 PM, Nicolas Williams
 wrote:
> On Sat, Feb 28, 2009 at 10:44:59PM +0100, Thomas Wagner wrote:
>> > >> pool-shrinking (and an option to shrink disk A when i want disk B to
>> > >> become a mirror, but A is a few blocks bigger)
>> >  This may be interesting... I'm not sure how often you need to shrink a 
>> > pool
>> >  though?  Could this be classified more as a Home or SME level feature?
>>
>> Enterprise level especially in SAN environments need this.
>>
>> Projects own theyr own pools and constantly grow and *shrink* space.
>> And they have no downtime available for that.
>
> Multiple pools on one server only makes sense if you are going to have
> different RAS for each pool for business reasons.  It's a lot easier to
> have a single pool though.  I recommend it.

Other scenarios for multiple pools include:

- Need independent portability of data between servers.  For example,
in a HA cluster environment, various workloads will be mapped to
various pools.  Since ZFS does not do active-active clustering, a
single pool for anything other than a simple active-standby cluster is
not useful.

- Array based copies are needed.  There are times when copies of data
are performed at a storage array level to allow testing and support
operations to happen "on different spindles".  For example, in a
consolidated database environment, each database may be constrained to
a set of spindles so that each database can be replicated or copied
independent of the various others.

-- 
Mike Gerdts
http://mgerdts.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Nicolas Williams
On Sat, Feb 28, 2009 at 10:44:59PM +0100, Thomas Wagner wrote:
> > >> pool-shrinking (and an option to shrink disk A when i want disk B to
> > >> become a mirror, but A is a few blocks bigger)
> >  This may be interesting... I'm not sure how often you need to shrink a 
> > pool 
> >  though?  Could this be classified more as a Home or SME level feature?
> 
> Enterprise level especially in SAN environments need this.
> 
> Projects own theyr own pools and constantly grow and *shrink* space.
> And they have no downtime available for that.

Multiple pools on one server only makes sense if you are going to have
different RAS for each pool for business reasons.  It's a lot easier to
have a single pool though.  I recommend it.

Nico
-- 
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Thomas Wagner
I would really add : make insane zfs destroy <-r|> poolname  as harmless as 
zpool destroy poolname (recoverable)

  zfs destroy <-r|> poolname<|/filesystem>

  this should behave like that:

  o snapshot the filesystem to be deleted (each, name it 
@deletedby_operatorname_date)

  o hide the snapshot as long as the pool has enough space and
 property snapshotbeforedelete=on (default off) is set 'on'

  o free space by removing those snapshots no earlier then configured
in a inheritable pool/filesystem property snapshotbeforedeleteremoval=3days
(=0 preserve forever, 30min preserve for 30 minutes, ...)


  o prevent deletion of a pool or filesystem if at least one 
snapshot from the above save actions exists down the tree

  o purging of snapshots would be done by 


To be honest, I don't want a discussion like the "rm -rf" is one.
In front of the keyboard or inside scripts we are all humans with
all theyr mistakes. In opposite to the rm -rf, the ZFS Design 
should take this extension without major changes. It should be 
a generic rule of dump to implement safety if it is possible 
at resonable low cost.

I think the full range of users, Enterprise to Home will appreciate
that theyr multi-million-$$-business/home_data does not go down 
accidentially with the interactive=on (Bryan) or the the idea 
written here. This in case someone makes an error and all the 
data could still be there (!)...ZFS should protect the user as well
and not only look at the hardware redundancy.

Thomas

PS: think of the day where simple operator $NAME makes a typo
zfs destroy -r poolname and all the data still sits on the
disk. But no one is able to bring that valueable data back,
except restoration from tape with hours of downtime.
Sorry for repeating that, it hurts so much to not having
this feature.

On Sat, Feb 28, 2009 at 04:35:05AM -0500, Bryan Allen wrote:
> I for one would like an "interactive" attribute for zpools and
> filesystems, specifically for destroy.
> 
> The existing behavior (no prompt) could be the default, but all
> filesystems would inherit from the zpool's attrib. so I'd only
> need to set interactive=on for the pool itself, not for each
> filesystem.
> 
> I have yet (in almost two years of using ZFS) to bone myself by
> accidentally destroying tank/worthmorethanyourjob, but it's only
> a matter of time, regardless of how careful I am.
> 
> The argument rm vs zfs destroy doesn't hold much water to me. I
> don't use rm -i, but destroying a single file or a hierarchy of
> directories is somewhat different than destroying a filesytem or
> entire pool. At least to my mind.
> 
> As such, consider it a piece of mind feature.
> -- 
> bda
> Cyberpunk is dead.  Long live cyberpunk.
> http://mirrorshades.org
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> 

-- 
Thomas Wagner
+49-171-6135989  http://www.wagner-net.net
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Thomas Wagner
> >> pool-shrinking (and an option to shrink disk A when i want disk B to
> >> become a mirror, but A is a few blocks bigger)
>  This may be interesting... I'm not sure how often you need to shrink a pool 
>  though?  Could this be classified more as a Home or SME level feature?

Enterprise level especially in SAN environments need this.

Projects own theyr own pools and constantly grow and *shrink* space.
And they have no downtime available for that.

give a +1 if you agree

Thomas

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Bob Friesenhahn

On Sat, 28 Feb 2009, Tim wrote:


That's not entirely true.  Maybe it will put it to shame at streaming
sequential I/O.  The 10k drives will still wipe the floor with any modern
7200rpm drive for random IO and seek times.


Or perhaps streaming sequential I/O will have similar performance, 
with much better performance for random IO and seek times.  It is 
always best to consult the vendor spec sheet.


Regardless, it is much easier to update with the same size, or a 
larger size drive.


Bob
--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Tim
On Sat, Feb 28, 2009 at 8:28 AM, Joe Esposito  wrote:

>
>
> On Sat, Feb 28, 2009 at 8:31 AM,  wrote:
>
>>
>> >I'm using opensolaris and zfs at my house for my photography storage
>> >as well as for an offsite backup location for my employer and several
>> >side web projects.
>> >
>> >I have an 80g drive as my root drive.  I recently took posesion of 2
>> >74g 10k drives which I'd love to add as a mirror to replace the 80 g
>> >drive.
>>
>> Why do you want to use a small 10K rpm disk?
>>
>> A modern 1TB disk at 5400/7200 rpm (at $100) will put it to shame.
>>
>> Casper
>>
>
> fair enough.  I just have a pair of these sitting here from a pull at work.
>  The data array is currently 4x1TB with another hot swap bay ready for 4x???
> when the need arises.
>

That's not entirely true.  Maybe it will put it to shame at streaming
sequential I/O.  The 10k drives will still wipe the floor with any modern
7200rpm drive for random IO and seek times.

--Tim
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Joe Esposito
On Sat, Feb 28, 2009 at 8:31 AM,  wrote:

>
> >I'm using opensolaris and zfs at my house for my photography storage
> >as well as for an offsite backup location for my employer and several
> >side web projects.
> >
> >I have an 80g drive as my root drive.  I recently took posesion of 2
> >74g 10k drives which I'd love to add as a mirror to replace the 80 g
> >drive.
>
> Why do you want to use a small 10K rpm disk?
>
> A modern 1TB disk at 5400/7200 rpm (at $100) will put it to shame.
>
> Casper
>

fair enough.  I just have a pair of these sitting here from a pull at work.
 The data array is currently 4x1TB with another hot swap bay ready for 4x???
when the need arises.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Casper . Dik

>I'm using opensolaris and zfs at my house for my photography storage
>as well as for an offsite backup location for my employer and several
>side web projects.
>
>I have an 80g drive as my root drive.  I recently took posesion of 2
>74g 10k drives which I'd love to add as a mirror to replace the 80 g
>drive.

Why do you want to use a small 10K rpm disk?

A modern 1TB disk at 5400/7200 rpm (at $100) will put it to shame.

Casper

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Mike Gerdts
On Sat, Feb 28, 2009 at 1:20 AM, Richard Elling
 wrote:
> David Magda wrote:
>> On Feb 27, 2009, at 20:02, Richard Elling wrote:
>>> At the risk of repeating the Best Practices Guide (again):
>>> The zfs send and receive commands do not provide an enterprise-level
>>> backup solution.
>>
>> Yes, in its current state; hopefully that will change some point in the
>> future (which is what we're talking about with GSoC--the potential to change
>> the status quo).
>
> I suppose, but considering that enterprise backup solutions exist,
> and some are open source, why reinvent the wheel?
> -- richard

The default mode of operation for every enterprise backup tool that I
have used is file level backups.  The determination of which files
need to be backed up seems to be to crawl the file system looking for
files that have an mtime after the previous backup.

Areas of strength for such tools include:

- Works with any file system that provides a POSIX interface
- Restore of a full backup is an accurate representation of the data backed up
- Restore can happen to a different file system type
- Restoring an individual file is possible

Areas of weakness include:

- Extremely inefficient for file systems with lots of files and little change.
- Restore of full + incremental tends to have extra files because of
spotty support or performance overhead of tool that would prevent it.
- Large files that have blocks rewritten get backed up in full each time
- Restores of file systems with lots of small files (especially in one
directory) are extremely slow

There exist features (sometimes expensive add-ons) that deal with some
of these shortcomings via:

- Keeping track of deleted files so that a restore is more
representative of what is on disk during the incremental backup.
Administration manuals typically warn that this has a big performance
and/or size overhead on the database used by the backup software.
- Including add-ons that hook into other components (e.g. VxFS storage
checkpoints, Oracle RMAN) that provide something similar to
block-level incremental backups

Why re-invent the wheel?

- People are more likely to have snapshots available for file-level
restores, and as such a "zfs send" data stream would only be used in
the event of a complete pool loss.
- It is possible to provide a general block-level backup solution so
that every product doesn't have to invent it.  This gives ZFS another
feature benefit to put it higher in the procurement priority.
- File creation slowness can likely be avoided allowing restore to
happen at tape speed
- To be competitive with NetApp "snapmirror to tape"
- Even having a zfs(1M) option that could list the files that change
between snapshots could be very helpful to prevent file system crawls
and to avoid being fooled by bogus mtimes.


-- 
Mike Gerdts
http://mgerdts.blogspot.com/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Blake Irvin

Shrinking pools would also solve the right-sizing dilemma.

Sent from my iPhone

On Feb 28, 2009, at 3:37 AM, Joe Esposito  wrote:


I'm using opensolaris and zfs at my house for my photography storage
as well as for an offsite backup location for my employer and several
side web projects.

I have an 80g drive as my root drive.  I recently took posesion of 2
74g 10k drives which I'd love to add as a mirror to replace the 80 g
drive.

From what I gather it is only possible if I zfs export my storage
array and reinstall solaris on the new disks.

So I guess I'm hoping zfs shrink and grow commands show up sooner or  
later.


Just a data point.

Joe Esposito
www.j-espo.com

On 2/28/09, "C. Bergström"  wrote:

Blake wrote:

Gnome GUI for desktop ZFS administration



On Fri, Feb 27, 2009 at 9:13 PM, Blake   
wrote:



zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B  
to

become a mirror, but A is a few blocks bigger)

This may be interesting... I'm not sure how often you need to  
shrink a
pool though?  Could this be classified more as a Home or SME level  
feature?

install to mirror from the liveCD gui


I'm not working on OpenSolaris at all, but for when my projects
installer is more ready /we/ can certainly do this..

zfs recovery tools (sometimes bad things happen)

Agreed.. part of what I think keeps zfs so stable though is the  
complete
lack of dependence on any recovery tools..  It forces customers to  
bring

up the issue instead of dirty hack and nobody knows.

automated installgrub when mirroring an rpool


This goes back to an installer option?

./C

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Joe Esposito
I'm using opensolaris and zfs at my house for my photography storage
as well as for an offsite backup location for my employer and several
side web projects.

I have an 80g drive as my root drive.  I recently took posesion of 2
74g 10k drives which I'd love to add as a mirror to replace the 80 g
drive.

>From what I gather it is only possible if I zfs export my storage
array and reinstall solaris on the new disks.

So I guess I'm hoping zfs shrink and grow commands show up sooner or later.

Just a data point.

Joe Esposito
www.j-espo.com

On 2/28/09, "C. Bergström"  wrote:
> Blake wrote:
>> Gnome GUI for desktop ZFS administration
>>
>>
>>
>> On Fri, Feb 27, 2009 at 9:13 PM, Blake  wrote:
>>
>>> zfs send is great for moving a filesystem with lots of tiny files,
>>> since it just handles the blocks :)
>>>
>>>
>>>
>>> I'd like to see:
>>>
>>> pool-shrinking (and an option to shrink disk A when i want disk B to
>>> become a mirror, but A is a few blocks bigger)
>>>
> This may be interesting... I'm not sure how often you need to shrink a
> pool though?  Could this be classified more as a Home or SME level feature?
>>> install to mirror from the liveCD gui
>>>
> I'm not working on OpenSolaris at all, but for when my projects
> installer is more ready /we/ can certainly do this..
>>> zfs recovery tools (sometimes bad things happen)
>>>
> Agreed.. part of what I think keeps zfs so stable though is the complete
> lack of dependence on any recovery tools..  It forces customers to bring
> up the issue instead of dirty hack and nobody knows.
>>> automated installgrub when mirroring an rpool
>>>
> This goes back to an installer option?
>
> ./C
>
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Tomas Ögren
On 28 February, 2009 - Bryan Allen sent me these 1,0K bytes:

> I for one would like an "interactive" attribute for zpools and
> filesystems, specifically for destroy.
> 
> The existing behavior (no prompt) could be the default, but all
> filesystems would inherit from the zpool's attrib. so I'd only
> need to set interactive=on for the pool itself, not for each
> filesystem.
> 
> I have yet (in almost two years of using ZFS) to bone myself by
> accidentally destroying tank/worthmorethanyourjob, but it's only
> a matter of time, regardless of how careful I am.
> 
> The argument rm vs zfs destroy doesn't hold much water to me. I
> don't use rm -i, but destroying a single file or a hierarchy of
> directories is somewhat different than destroying a filesytem or
> entire pool. At least to my mind.
> 
> As such, consider it a piece of mind feature.

Or a spinoff; -t argument to destroy, so you can pretty safely script
snapshot destruction without risking the entire pool/fs being zapped..

That is: zfs destroy -t snapshot b...@foo

And possibly: zfs destroy -r -t snapshot mypool/myfsto kill all
snapshots below myfs.

/Tomas
-- 
Tomas Ögren, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Bryan Allen
I for one would like an "interactive" attribute for zpools and
filesystems, specifically for destroy.

The existing behavior (no prompt) could be the default, but all
filesystems would inherit from the zpool's attrib. so I'd only
need to set interactive=on for the pool itself, not for each
filesystem.

I have yet (in almost two years of using ZFS) to bone myself by
accidentally destroying tank/worthmorethanyourjob, but it's only
a matter of time, regardless of how careful I am.

The argument rm vs zfs destroy doesn't hold much water to me. I
don't use rm -i, but destroying a single file or a hierarchy of
directories is somewhat different than destroying a filesytem or
entire pool. At least to my mind.

As such, consider it a piece of mind feature.
-- 
bda
Cyberpunk is dead.  Long live cyberpunk.
http://mirrorshades.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-28 Thread Chris Ridd


On 28 Feb 2009, at 07:26, C. Bergström wrote:


Blake wrote:

Gnome GUI for desktop ZFS administration

With the libzfs java bindings I am plotting a web based interface..  
I'm not sure if that would meet this gnome requirement though..   
Knowing specifically what you'd want to do in that interface would  
be good.. I planned to compare it to fishworks and the nexenta  
appliance as a base..


Recent builds of OpenSolaris come with SWT from the Eclipse project,  
which makes it possible for Java apps to use real GNOME/GTK native  
UIs. So your libzfs bindings may well be useful with that.


Cheers,

Chris
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread C. Bergström

Blake wrote:

Gnome GUI for desktop ZFS administration
  
With the libzfs java bindings I am plotting a web based interface.. I'm 
not sure if that would meet this gnome requirement though..  Knowing 
specifically what you'd want to do in that interface would be good.. I 
planned to compare it to fishworks and the nexenta appliance as a base..

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread C. Bergström

Blake wrote:

Gnome GUI for desktop ZFS administration



On Fri, Feb 27, 2009 at 9:13 PM, Blake  wrote:
  

zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B to
become a mirror, but A is a few blocks bigger)

This may be interesting... I'm not sure how often you need to shrink a 
pool though?  Could this be classified more as a Home or SME level feature?

install to mirror from the liveCD gui

I'm not working on OpenSolaris at all, but for when my projects 
installer is more ready /we/ can certainly do this..

zfs recovery tools (sometimes bad things happen)

Agreed.. part of what I think keeps zfs so stable though is the complete 
lack of dependence on any recovery tools..  It forces customers to bring 
up the issue instead of dirty hack and nobody knows.

automated installgrub when mirroring an rpool


This goes back to an installer option?

./C

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread Richard Elling

David Magda wrote:

On Feb 27, 2009, at 20:02, Richard Elling wrote:

It wouldn't help.  zfs send is a data stream which contains parts of 
files,

not files (in the usual sense), so there is no real way to take a send
stream and extract a file, other than by doing a receive.


If you create a non-incremental stream of a snapshot, doesn't it 
contain all the information necessary to recreate the file system up 
to the point the snapshot was created? Now take that stream and drop 
it onto a tape (real or VTL) and you would potentially have a back up 
(a la ufsdump).


Who uses only one snapshot?  Not many people.


At the risk of repeating the Best Practices Guide (again):
The zfs send and receive commands do not provide an enterprise-level 
backup solution.


Yes, in its current state; hopefully that will change some point in 
the future (which is what we're talking about with GSoC--the potential 
to change the status quo).


I suppose, but considering that enterprise backup solutions exist,
and some are open source, why reinvent the wheel?
-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread David Magda

On Feb 27, 2009, at 20:02, Richard Elling wrote:

It wouldn't help.  zfs send is a data stream which contains parts of  
files,

not files (in the usual sense), so there is no real way to take a send
stream and extract a file, other than by doing a receive.


If you create a non-incremental stream of a snapshot, doesn't it  
contain all the information necessary to recreate the file system up  
to the point the snapshot was created? Now take that stream and drop  
it onto a tape (real or VTL) and you would potentially have a back up  
(a la ufsdump).



At the risk of repeating the Best Practices Guide (again):
The zfs send and receive commands do not provide an enterprise-level  
backup solution.


Yes, in its current state; hopefully that will change some point in  
the future (which is what we're talking about with GSoC--the potential  
to change the status quo).


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread Blake
Gnome GUI for desktop ZFS administration



On Fri, Feb 27, 2009 at 9:13 PM, Blake  wrote:
> zfs send is great for moving a filesystem with lots of tiny files,
> since it just handles the blocks :)
>
>
>
> I'd like to see:
>
> pool-shrinking (and an option to shrink disk A when i want disk B to
> become a mirror, but A is a few blocks bigger)
>
> install to mirror from the liveCD gui
>
> zfs recovery tools (sometimes bad things happen)
>
> automated installgrub when mirroring an rpool
>
>
>
>
> On Fri, Feb 27, 2009 at 8:02 PM, Richard Elling
>  wrote:
>> David Magda wrote:
>>>
>>> On Feb 27, 2009, at 18:23, C. Bergström wrote:
>>>
 Blake wrote:
>
> Care to share any of those in advance?  It might be cool to see input
> from listees and generally get some wheels turning...
>
 raidz boot support in grub 2 is pretty high on my list to be honest..

 Which brings up another question of where is the raidz stuff mostly?

 usr/src/uts/common/fs/zfs/vdev_raidz.c ?

 Any high level summary, docs or blog entries of what the process would
 look like for a raidz boot support is also appreciated.
>>>
>>> Given the threads that have appeared on this list lately, how about
>>> codifying / standardizing the output of "zfs send" so that it can be backed
>>> up to tape? :)
>>
>> It wouldn't help.  zfs send is a data stream which contains parts of files,
>> not files (in the usual sense), so there is no real way to take a send
>> stream and extract a file, other than by doing a receive.
>>
>> At the risk of repeating the Best Practices Guide (again):
>> The zfs send and receive commands do not provide an enterprise-level backup
>> solution.
>> -- richard
>>
>> ___
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread Blake
zfs send is great for moving a filesystem with lots of tiny files,
since it just handles the blocks :)



I'd like to see:

pool-shrinking (and an option to shrink disk A when i want disk B to
become a mirror, but A is a few blocks bigger)

install to mirror from the liveCD gui

zfs recovery tools (sometimes bad things happen)

automated installgrub when mirroring an rpool




On Fri, Feb 27, 2009 at 8:02 PM, Richard Elling
 wrote:
> David Magda wrote:
>>
>> On Feb 27, 2009, at 18:23, C. Bergström wrote:
>>
>>> Blake wrote:

 Care to share any of those in advance?  It might be cool to see input
 from listees and generally get some wheels turning...

>>> raidz boot support in grub 2 is pretty high on my list to be honest..
>>>
>>> Which brings up another question of where is the raidz stuff mostly?
>>>
>>> usr/src/uts/common/fs/zfs/vdev_raidz.c ?
>>>
>>> Any high level summary, docs or blog entries of what the process would
>>> look like for a raidz boot support is also appreciated.
>>
>> Given the threads that have appeared on this list lately, how about
>> codifying / standardizing the output of "zfs send" so that it can be backed
>> up to tape? :)
>
> It wouldn't help.  zfs send is a data stream which contains parts of files,
> not files (in the usual sense), so there is no real way to take a send
> stream and extract a file, other than by doing a receive.
>
> At the risk of repeating the Best Practices Guide (again):
> The zfs send and receive commands do not provide an enterprise-level backup
> solution.
> -- richard
>
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread Richard Elling

David Magda wrote:


On Feb 27, 2009, at 18:23, C. Bergström wrote:


Blake wrote:

Care to share any of those in advance?  It might be cool to see input
from listees and generally get some wheels turning...


raidz boot support in grub 2 is pretty high on my list to be honest..

Which brings up another question of where is the raidz stuff mostly?

usr/src/uts/common/fs/zfs/vdev_raidz.c ?

Any high level summary, docs or blog entries of what the process 
would look like for a raidz boot support is also appreciated.


Given the threads that have appeared on this list lately, how about 
codifying / standardizing the output of "zfs send" so that it can be 
backed up to tape? :)


It wouldn't help.  zfs send is a data stream which contains parts of files,
not files (in the usual sense), so there is no real way to take a send
stream and extract a file, other than by doing a receive.

At the risk of repeating the Best Practices Guide (again):
The zfs send and receive commands do not provide an enterprise-level 
backup solution.

-- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread David Magda


On Feb 27, 2009, at 18:23, C. Bergström wrote:


Blake wrote:

Care to share any of those in advance?  It might be cool to see input
from listees and generally get some wheels turning...


raidz boot support in grub 2 is pretty high on my list to be honest..

Which brings up another question of where is the raidz stuff mostly?

usr/src/uts/common/fs/zfs/vdev_raidz.c ?

Any high level summary, docs or blog entries of what the process  
would look like for a raidz boot support is also appreciated.


Given the threads that have appeared on this list lately, how about  
codifying / standardizing the output of "zfs send" so that it can be  
backed up to tape? :)


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread David
RaidZ grow support



On Fri, Feb 27, 2009 at 11:23 PM, "C. Bergström"
wrote:

> Blake wrote:
>
>> Care to share any of those in advance?  It might be cool to see input
>> from listees and generally get some wheels turning...
>>
>>
> raidz boot support in grub 2 is pretty high on my list to be honest..
>
> Which brings up another question of where is the raidz stuff mostly?
>
> usr/src/uts/common/fs/zfs/vdev_raidz.c ?
>
> Any high level summary, docs or blog entries of what the process would look
> like for a raidz boot support is also appreciated.
>
>
> Cheers,
>
>
> ./Christopher
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-27 Thread C. Bergström

Blake wrote:

Care to share any of those in advance?  It might be cool to see input
from listees and generally get some wheels turning...
  

raidz boot support in grub 2 is pretty high on my list to be honest..

Which brings up another question of where is the raidz stuff mostly?

usr/src/uts/common/fs/zfs/vdev_raidz.c ?

Any high level summary, docs or blog entries of what the process would 
look like for a raidz boot support is also appreciated.



Cheers,

./Christopher
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-26 Thread Blake
Care to share any of those in advance?  It might be cool to see input
from listees and generally get some wheels turning...

On Wed, Feb 25, 2009 at 4:39 AM, "C. Bergström"
 wrote:
>
> Hi everyone.
>
> I've got a couple ideas for good zfs GSoC projects, but wanted to stir some
> interest.  Anyone interested to help mentor?  The deadline is around the
> corner so if planning hasn't happened yet it should start soon.  If there is
> interest who would the org administrator be?
>
> Thanks
>
> ./Christopher
> ___
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] GSoC 09 zfs ideas?

2009-02-25 Thread Darren J Moffat

C. Bergström wrote:


Hi everyone.

I've got a couple ideas for good zfs GSoC projects, but wanted to stir 
some interest.  Anyone interested to help mentor?  The deadline is 
around the corner so if planning hasn't happened yet it should start 
soon.  If there is interest who would the org administrator be?


I might be interested in mentoring.  I've done GSoC mentoring in the past.

--
Darren J Moffat
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] GSoC 09 zfs ideas?

2009-02-25 Thread C. Bergström


Hi everyone.

I've got a couple ideas for good zfs GSoC projects, but wanted to stir 
some interest.  Anyone interested to help mentor?  The deadline is 
around the corner so if planning hasn't happened yet it should start 
soon.  If there is interest who would the org administrator be?


Thanks

./Christopher
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss