Re: bug? ZFS crypto vs. scrub

2011-05-11 Thread Darren J Moffat

On 11/05/2011 01:07, Daniel Carosone wrote:

Sorry for abusing the mailing list, but I don't know how to report
bugs anymore and have no visibility of whether this is a
known/resolved issue.  So, just in case it is not...


Log a support call with Oracle if you have a support contract.


With Solaris 11 Express, scrubbing a pool with encrypted datasets for
which no key is currently loaded, unrecoverable read errors are
reported. The error count applies to the pool, and not to any specific
device, which is also somewhat at odds with the helpful message text
for diagnostic status and suggested action:


Known issue:

6989185 scrubbing a pool with encrypted filesystems and snapshots can 
report false positive errors.


If you have a support contract you may be able to request that fix be 
back ported into an SRU (note I'm not guaranteeing it will be just 
saying that it is technically possible)



When this has happened previously (on this and other pools) mounting
the dataset by supplying the key, and rerunning the scrub, removes the
errors.

For some reason, I can't in this case (keeps complaining that
the key is wrong). That may be a different issue that has also
happened before, and I will post about separately, once I'm sure I
didn't just made a typo (twice) when first setting the key.


Since you are saying typo I'm assuming you have 
keysource=passphrase,prompt (ie the default).  Have you ever done a 
send|recv of the encrypted datasets ? and if so where there multiple 
snapshots recv'd ?


--
Darren J Moffat
___
zfs-crypto-discuss mailing list
zfs-crypto-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-crypto-discuss


RE: ZFS crypto source

2011-05-11 Thread Rob O'Leary
I guessed you wouldn't be able to say, even if...

The only shortfall in capability that I'm aware of is the secure boot/FDE,
which we discussed previously.

I am mostly interested in the source to see how features have been
implemented and to understand the system structure. I certainly wouldn't
presume to make changes!

On the slightly more general topic of source on opensolaris, are the designs
for subsystems/features available? I've found PSARC cases for some things
but I expect that more detailed design, system interaction and use cases are
documented as part of the development process. Are any of these types of
document made public to assist in understanding at a higher level than the
source code? Again, this is really to help me understand the system, rather
than to attempt any modification.

Regards
Rob

-Original Message-
From: Darren J Moffat [mailto:darr...@opensolaris.org]
Sent: 10 May 2011 11:17
To: Rob O'Leary
Cc: zfs-crypto-discuss@opensolaris.org
Subject: Re: ZFS crypto source


On 07/05/2011 10:57, Rob O'Leary wrote:
 Is the source for ZFS crypto likely to be released on opensolaris.org?

Older versions of the source area available from the zfs-crypto
project gates:  /zfs-crypto/gate/  However in some important areas these
differ quite a bit from what was finally integrated and are not on disk
compatible.

 I searched in /onnv/onnv-gate/usr/src/uts/common/fs/zfs, which may have
been
 the wrong place, for aes and crypt and got no results so I assume that the
 zfs encryption has not been released to date.

Correct the source has not been released.

I do not know anything about future plans nor would I be able to comment
here at this time even if I did.  Please bring this up with your Oracle
account/support team representative if it is important to your business.

Is there something in particular you want to do with the source if you
had it available to you ?  Are there changes you want to make ?

--
Darren J Moffat

___
zfs-crypto-discuss mailing list
zfs-crypto-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-crypto-discuss


Re: [zfs-discuss] Summary: Dedup and L2ARC memory requirements

2011-05-11 Thread Frank Van Damme
Op 10-05-11 06:56, Edward Ned Harvey schreef:
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Edward Ned Harvey

 BTW, here's how to tune it:

 echo arc_meta_limit/Z 0x3000 | sudo mdb -kw

 echo ::arc | sudo mdb -k | grep meta_limit
 arc_meta_limit=   768 MB
 
 Well ... I don't know what to think yet.  I've been reading these numbers
 for like an hour, finding interesting things here and there, but nothing to
 really solidly point my finger at.
 
 The one thing I know for sure...  The free mem drops at an unnatural rate.
 Initially the free mem disappears at a rate approx 2x faster than the sum of
 file size and metadata combined.  Meaning the system could be caching the
 entire file and all the metadata, and that would only explain half of the
 memory disappearance.

I'm seeing similar things. Yesterday I first rebooted with set
zfs:zfs_arc_meta_limit=0x1 (that's 4 GiB) set in /etc/system and
monitored while the box was doing its regular job (taking backups).
zfs_arc_min is also set to 4 GiB. What I noticed is that shortly after
the reboot, the arc started filling up rapidly, mostly with metadata. It
shot up to:

arc_meta_max  =  3130 MB

afterwards, the number for arc_meta_used steadily dropped. Some 12 hours
ago, I started deleting files, it has deleted about 600 files since
then. Now at the moment the arc size stays right at the minimum of 2
GiB, of which metadata fluctuates around 1650 MB.

This is the output of the getmemstats.sh script you posted.

Memory: 6135M phys mem, 539M free mem, 6144M total swap, 6144M free swap
zfs:0:arcstats:c2147483648  = 2 GiB target size
zfs:0:arcstats:c_max5350862848  = 5 GiB
zfs:0:arcstats:c_min2147483648  = 2 GiB
zfs:0:arcstats:data_size829660160   = 791 MiB
zfs:0:arcstats:hdr_size 93396336= 89 MiB
zfs:0:arcstats:other_size   411215168   = 392 MiB
zfs:0:arcstats:size 1741492896  = 1661 Mi
arc_meta_used =  1626 MB
arc_meta_limit=  4096 MB
arc_meta_max  =  3130 MB

I get way more cache misses then I'd like:

Time  read  miss  miss%  dmis  dm%  pmis  pm%  mmis  mm%  arcszc
10:01:133K   380 10   1667   214   15   2597 1G   2G
10:02:132K   340 16372   302   46   323   16 1G   2G
10:03:132K   368 18473   321   46   347   17 1G   2G
10:04:131K   348 25444   303   63   335   24 1G   2G
10:05:132K   420 15874   332   36   383   14 1G   2G
10:06:133K   489 16   1326   357   35   427   14 1G   2G
10:07:132K   405 15492   355   39   401   15 1G   2G
10:08:132K   366 13402   326   37   366   13 1G   2G
10:09:131K   364 20181   345   58   363   20 1G   2G
10:10:134K   370  8592   311   21   3698 1G   2G
10:11:134K   351  8572   294   21   3508 1G   2G
10:12:133K   378 10592   319   26   372   10 1G   2G
10:13:133K   393 11532   339   28   393   11 1G   2G
10:14:132K   403 13402   363   35   402   13 1G   2G
10:15:133K   365 11482   317   30   365   11 1G   2G
10:16:132K   374 15402   334   40   374   15 1G   2G
10:17:133K   385 12432   341   28   383   12 1G   2G
10:18:134K   343  8642   279   19   3438 1G   2G
10:19:133K   391 10592   332   23   391   10 1G   2G


So, one explanation I can think of is that the rest of the memory are
l2arc pointers, supposing they are not actually counted in the arc
memory usage totals (AFAIK l2arc pointers are considered to be part of
arc). Then again my l2arc is still growing (slowly) and I'm only caching
metadata at the moment, so you'd think it'd shrink if there's no more
room for l2arc pointers. Besides I'm getting very little reads from ssd:

 capacity operationsbandwidth
pool  alloc   free   read  write   read  write
  -  -  -  -  -  -
backups   5.49T  1.57T415121  3.13M  1.58M
  raidz1  5.49T  1.57T415121  3.13M  1.58M
c0t0d0s1  -  -170 16  2.47M   551K
c0t1d0s1  -  -171 16  2.46M   550K
c0t2d0s1  -  -170 16  2.53M   552K
c0t3d0s1  -  -170 16  2.44M   550K
cache -  -  -  -  -  -
  c1t5d0  63.4G  48.4G 20  0  2.45M  0
  -  -  -  -  -  -

(typical statistic over 1 minute)


I might try the windows solution and reboot the machine to free up
memory and let it fill the cache all over again and see if I get more
cache hits... hmmm...

 I set the 

Re: [zfs-discuss] bootfs ID on zfs root

2011-05-11 Thread Jim Klimov
Technically bootfs ID is a string which names the root dataset, typically 
rpool/ROOT/solarisReleaseNameCode. This string can be passed to Solaris 
kernel as a parameter manually or by bootloader, otherwise a default current 
bootfs is read from the root pool's attributes (not dataset attributes! - see 
zpool get/set bootfs).

In your case it seems that the attribute points to an invalid name, and your 
root dataset may be named somehing else - just set the pool attribute.

I don't know of bootfs ID numbers, but maybe that's a concept in your 
company's scripting and patching environment.

It is also possible that device names changed (i.e. on x86 - when SATA HDD 
access mode in BIOS changed from IDE to AHCI) and the boot device name saved in 
eeprom or its GRUB emulator is no longer valid. But this has different error 
strings ;)

Good luck,
//Jim
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Modify stmf_sbd_lu properties

2011-05-11 Thread Jim Klimov
You can try to workaround - no idea if this would really work -
0) Disable stmf and iscsi/* services
1) Create your volume's clone
2) Rename the original live volume dataset to some other name
3) Rename the clone to original dataset's name
4) Promote the clone
- now for the system it SHOULD seem like this clone was always here with this 
original naming, and your current newer dataset is a cloned deviation. 
Hopefully this will fool STMF into using this data instead of new data, with 
existing GUID...
5) Enable stmf and iscsi/* services

*) Tell us if it works ;)

HTH,
//Jim
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance problem suggestions?

2011-05-11 Thread Jim Klimov
  Disks that have been in use for a longer time may have very fragmented free
  space on one hand, and not so much of it on another, but ZFS is still 
  trying to push
  bits around evenly. And while it's waiting on some disks, others may be 
  blocked as
  well. Something like that...
 This could explain why performance would go up after a large delete but I've 
 not 
 seen large wait times for any of my disks. The service time, percent busy, 
 and 
 every other metric continues to show nearly idle disks.

I believe, in this situation the older fuller disks would show some activity 
and others can show zero or few IOs - because ZFS has no tasks for them. It 
sent a series of blocks to write from the queue, newer disks wrote them and 
stay dormant, while older disks seek around to fit that piece of data... When 
old disks complete the writes, ZFS batches them a new set of tasks.

 If this is the problem- it would be nice if there were a simple zfs or dtrace 
 query 
 that would show it to you.

Well, it seems that the bridge between email and web interfaces to OpenSolaris 
forums has been fixed, for new posts at least, and hopefully Richard Elling or 
some other experts would come up with an idea of a dtrace for your situation.

I have little non-zero hope that the experts would also come to the web-forums 
and review the past month's posts and give their comments to my, your and 
others' questions and findings ;)

//Jim Klimov
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS backup and restore

2011-05-11 Thread Jim Klimov
Sorry, I did not hit this type of error...

AFAIK the pool writes during zfs receive are done by current code (i.e. ZFSv22 
for you) based on data read from the backup stream. So unless there are 
corruptions on the pool which happened to be at the same time as you did your 
restore, this procedure should not have broken your pool. It was just a 
(special) write by your OS version's code, nearly same as any other.

Did you have any hardware problems at that time, like brown-outs, overheating, 
occasional push of a reset button?

Maybe some experts here would be able to comment better or guide you through 
troubleshooting and collecting data, if you post your OS version details, error 
messages especially, etc.

For example, if for some reason your hostid changed (i.e. you used the safe 
mode miniroot) and the pool tank was not exported afterwards, the error may be 
like Pool is busy or used by another system with a solution as simple as that 
you'd have to do a forced import (zpool import -f tank) - if it is indeed a 
local non-networked pool and no other machine really uses it.

HTH,
//Jim
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS backup and restore

2011-05-11 Thread Naveen surisetty
Hi,

Thanks for the response, Here is my problem.

I have a zfs stream back up took on zfs version 15, currently i have upgraded 
my OS, so new zfs version is 22. Restore process went well from old stream 
backup to new zfs pool. but on reboot i got error unable to mount pool tank. 


So there is incompatibilities between zfs versions, especially send/receive.

So looking for alternate backup solutions.

Thanks
kumar
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance problem suggestions?

2011-05-11 Thread Jim Litchfield
Keep in mind zfs_vdev_max_pending. In the latest version of S10, this is set
to 10. ZFS will not issue more than the value of this variable requests at
a time for a LUN. Your disks may look relatively idle while ZFS
has a lot of data piled up inside just waiting to be read or written.
I have tweaked this on the fly. 

One key indicator is if your disk queues hover around 10.

Jim
---

- Original Message -
From: jimkli...@cos.ru
To: zfs-discuss@opensolaris.org
Sent: Wednesday, May 11, 2011 3:22:19 AM GMT -08:00 US/Canada Pacific
Subject: Re: [zfs-discuss] Performance problem suggestions?

  Disks that have been in use for a longer time may have very fragmented free
  space on one hand, and not so much of it on another, but ZFS is still 
  trying to push
  bits around evenly. And while it's waiting on some disks, others may be 
  blocked as
  well. Something like that...
 This could explain why performance would go up after a large delete but I've 
 not 
 seen large wait times for any of my disks. The service time, percent busy, 
 and 
 every other metric continues to show nearly idle disks.

I believe, in this situation the older fuller disks would show some activity 
and others can show zero or few IOs - because ZFS has no tasks for them. It 
sent a series of blocks to write from the queue, newer disks wrote them and 
stay dormant, while older disks seek around to fit that piece of data... When 
old disks complete the writes, ZFS batches them a new set of tasks.

 If this is the problem- it would be nice if there were a simple zfs or dtrace 
 query 
 that would show it to you.

Well, it seems that the bridge between email and web interfaces to OpenSolaris 
forums has been fixed, for new posts at least, and hopefully Richard Elling or 
some other experts would come up with an idea of a dtrace for your situation.

I have little non-zero hope that the experts would also come to the web-forums 
and review the past month's posts and give their comments to my, your and 
others' questions and findings ;)

//Jim Klimov
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Modify stmf_sbd_lu properties

2011-05-11 Thread Don
I can't actually disable the STMF framework to do this but I can try renaming 
things and dumping the properties from one device to another and see if it 
works- it might actually do it. I will let you know.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Backup complete rpool structure and data to tape

2011-05-11 Thread Arjun YK
Hello,

Trying to understand how to backup mirrored zfs boot pool 'rpool' to tape,
and restore it back if in case the disks are lost.
Backup would be done with an enterprise tool like tsm, legato etc.

As an example, here is the layout:

# zfs list
NAME  USED  AVAIL  REFER  MOUNTPOINT
rpool64.5G   209G97K  legacy
rpool/ROOT   24.0G   209G21K  legacy
rpool/ROOT/s10s_u8wos_08a  24G  14.8G  5.15G  /
rpool/ROOT/s10s_u8wos_08a/var   4G  3.93G  74.6M  /var
rpool/dump   2.50G   209G  2.50G  -
rpool/swap 16G   225G   136M  -
#

Could you answer these queries:

1. Is it possible to backup 'rpool' as a single entity, or do we need to
backup each filesystems, volumes etc. within rpool seperately ?

2. How do we backup the whole structure of zfs (pool, filesystem, volume,
snapshot etc.) along with all its property settings. Not just the actual
data stored within.

3. If in case the whole structure cannot be backed up using enterprise
backup, how do we save and restore zfs sctructure if in case the disks are
lost. I have read about 'zfs send receive ...'. Is this the only recommended
way ?

4. I have never tried to restore a whole boot disk from tape. Could you
share some details on how to rebuild the boot disks by restoring from tape.

5. I have set 'rpool/ROOT' mountpoint to 'legacy' as I don't see any reason
to mount it. Not sure if it's a right thing to do. Any suggestions ?


Thanks
Arjun
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Modify stmf_sbd_lu properties

2011-05-11 Thread Evaldas Auryla

 On 05/10/11 09:45 PM, Don wrote:

Is it possible to modify the GUID associated with a ZFS volume imported
into STMF?

To clarify- I have a ZFS volume I have imported into STMF and export via
iscsi. I have a number of snapshots of this volume. I need to temporarily
go back to an older snapshot without removing all the more recent ones. I
can delete the current sbd LU, clone the snapshot I want to test, and then
bring that back in to sbd.

The problem is that you need to use sbdadm create-lu and that creates a
new GUID. (sbdadm import-lu on a clone will give you a metafile error).


Hi, maybe this could do:

stmfadm create-lu -p guid=XXX..

Regards,

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Performance problem suggestions?

2011-05-11 Thread Don
 It sent a series of blocks to write from the queue, newer disks wrote them 
 and stay
 dormant, while older disks seek around to fit that piece of data... When old 
 disks
 complete the writes, ZFS batches them a new set of tasks.
The thing is- as far as I know the OS doesn't ask the disk to find a place to 
fit the data. Instead the OS tracks what space on the disk is free and then 
tells the disk where to write the data.

Even if ZFS was waiting for the IO to complete I would expect to see that delay 
reflected in the disk service times. In our case we see no high service times, 
no busy disks, nothing. It seems like ZFS is just sitting there quietly and 
thinking to itself. If the processor were busy that might make sense but even 
there- our processor seems largely idle.

At the same time- even a scrub on this system is a joke right now and that's a 
read intensive operation. I'm seeing a scrub speed of 400K/s but almost no IO's 
to my disks.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Intel Z68 chipset and SSD 311 (Larson Creek)

2011-05-11 Thread David Magda
The press embargo on Intel Z68 chipset has been lifted and so there's a bunch 
of press on it. One feature called Smart Response Technology (SRT) will sound 
familiar to users of ZFS:

 Intel's SRT functions like an actual cache. Rather than caching individual 
 files, Intel focuses on frequently accessed LBAs (logical block addresses). 
 Read a block enough times or write to it enough times and those accesses will 
 get pulled into the SSD cache until it's full. When full, the least recently 
 used data gets evicted making room for new data.


http://www.anandtech.com/show/4329/2

In addition to the chipset (already shipping in the 2011 Apple iMacs), there's 
also an SLC SSD:

 The big difference here is the SSD 311 comes with 20GB of 34nm SLC NAND. If 
 you remember back to the SSD Anthology, SLC NAND is architecturally identical 
 to MLC NAND. With half the number of data stored per NAND cell SLC NAND not 
 only lasts longer than MLC NAND but it also is much faster, particularly for 
 writes.
[...]
 The SSD 311 basically offers the performance of a 160GB X25-M G2 but with 
 fewer NAND channels and a much lower capacity.
 
 Remember this is SLC NAND so despite only being a 20GB drive, it's priced 
 more like a 40GB MLC drive: Intel expects the SSD 311 to retail for $110. 
 Thankfully you aren't locked in to only using Intel drives as Smart Response 
 Technology will work with any SSD.


http://www.anandtech.com/show/4329/3

Some good pros and cons from Tom's Hardware of this system:

http://tinyurl.com/42rzkyz
http://www.tomshardware.com/reviews/intel-z68-express-smart-response-technology-ssd-caching,2938-3.html

Plenty of other articles on the subject popping up. I think that the fact that 
the SSD is SLC and not too expensive would make it useful for ZFS users as 
well. Given that the ZIL can never be more than half of RAM, it would mean that 
you could put one in a server that has 40 GB of RAM. Or in a home machine that 
has, say, only 8 GB you could partition 4 GB of the SSD for a 'slog' device, 
and the rest for L2ARC cache.

In a perfect world it'd have a supercap, but as it stands, I think it's a good 
addition to the product choices we have. Hopefully more manufacturers will get 
on the bandwagon of small SLC SSDs now that Intel is helping mainstream this 
idea by putting it right in their chipsets.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS backup and restore

2011-05-11 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Naveen surisetty
 
 I have a zfs stream back up took on zfs version 15, currently i have
upgraded
 my OS, so new zfs version is 22. Restore process went well from old stream
 backup to new zfs pool. but on reboot i got error unable to mount pool
tank.

Let me make sure I get this straight ...

You did zfs send on zpool version 15, save to a file.
Then you upgraded system and have zpool version 22.  You do zfs receive
and there are no errors.

Simply having successfully completed the zfs receive should guarantee it
was successfully received.  But since you're having a problem, I suggest you
zpool export and zpool import immediately after doing your zfs receive.
If you're able to do this, it guarantees the pool is in fact importable.
Then if you reboot and encounter a problem, you know you're running into
something else that's weird.  Like perhaps unsupported or buggy hardware
that does something bad to disk during the reboot process.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backup complete rpool structure and data to tape

2011-05-11 Thread Edward Ned Harvey
 From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
 boun...@opensolaris.org] On Behalf Of Arjun YK
 
 Trying to understand how to backup mirrored zfs boot pool 'rpool' to tape,
 and restore it back if in case the disks are lost.
 Backup would be done with an enterprise tool like tsm, legato etc.

Backup/restore of bootable rpool to tape with a 3rd party application like
legato etc is kind of difficult.  Because if you need to do a bare metal
restore, how are you going to do it?  The root of the problem is the fact
that you need an OS with legato in order to restore the OS.  It's a
catch-22.  It is much easier if you can restore the rpool from some storage
that doesn't require the 3rd party tool to access it ... 

I might suggest:  If you use zfs send to backup rpool to a file in the
data pool...  And then use legato etc to backup the data pool...  If you
need to do a bare metal restore some day, you would just install a new OS,
install legato or whatever, and restore your data pool.  Then you could boot
to a command prompt of the installation disc, and restore (obliterate) the
rpool using the rpool backup file.

But I hope you can completely abandon the whole 3rd party backup software
and tapes.  Some people can, and others cannot.  By far, the fastest best
way to backup ZFS is to use zfs send | zfs receive on another system or a
set of removable disks.  ZFS send has the major advantage that it doesn't
need to crawl the whole filesystem scanning for changes.  It just knows what
incremental blocks have changed, and it instantly fetches only the necessary
blocks.


 1. Is it possible to backup 'rpool' as a single entity, or do we need to
backup
 each filesystems, volumes etc. within rpool seperately ?

You can do it either way you like.  Specify a single filesystem, or
recursively do all of its children.  man zfs send


 2. How do we backup the whole structure of zfs (pool, filesystem, volume,
 snapshot etc.) along with all its property settings. Not just the actual
data
 stored within.

Regarding pool  filesystem properties, I believe this changed at some
point.  There was a time in history when I decided to zpool get all mypool
and zfs get all mypool and store those text files alongside the backup.
But if you check the man page for zfs send, I think this is automatic now.

No matter what, you'll have to create a pool before you can restore.  So
you'll just have to take it upon yourself to remember your pool architecture
... striping, mirroring, raidz, cache  log devices etc.

Incidentally, when you do incremental zfs send, you have to specify the
from and to snapshots.  So there must be at least one identical snapshot
in the sending and receiving system (or else your only option is to do a
complete full send.)  Point is:  You can take a lot of baby steps if you
wish, keeping all the snapshots if you wish.  Or you can jump straight from
the oldest matching snapshot to the latest snap.  You'll complete somewhat
faster but lose granularity in the backups if you do that.


 3. If in case the whole structure cannot be backed up using enterprise
 backup, how do we save and restore zfs sctructure if in case the disks are
 lost. I have read about 'zfs send receive ...'. Is this the only
recommended
 way ?

For anything other than rpool, you can use any normal backup tool you like.
Netbackup, legato, tar, cpio.  Whatever.  (For rpool, I wouldn't really
trust those - I recommend zfs send for rpool.)  You can also use zfs send 
receive for data pools.  You gain performance (potentially many orders of
magnitude shorter backup window) if zfs send  receive are acceptable in
your environment.  But it's not suitable for everyone for many reasons...
You can't exclude anything from zfs send...  And you can't do a selective
zfs receive.  It's the whole filesystem or nothing.  And a single bit
corruption will render the whole backup unusable, so it's not recommended to
store a zfs send data stream for later use.  It's recommended to pipe a
zfs send directly into a zfs receive.  Which implies disk-to-disk, no tape.


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS backup and restore

2011-05-11 Thread Richard Elling
On May 10, 2011, at 11:21 PM, Naveen surisetty wrote:

 Hi,
 
 Thanks for the response, Here is my problem.
 
 I have a zfs stream back up took on zfs version 15, currently i have upgraded 
 my OS, so new zfs version is 22. Restore process went well from old stream 
 backup to new zfs pool. but on reboot i got error unable to mount pool tank. 

There is no such thing as mounting a pool  Can you post the exact error 
message?
 -- richard

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backup complete rpool structure and data to tape

2011-05-11 Thread Glenn Lagasse
* Edward Ned Harvey (opensolarisisdeadlongliveopensola...@nedharvey.com) wrote:
  From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
  boun...@opensolaris.org] On Behalf Of Arjun YK
  
  Trying to understand how to backup mirrored zfs boot pool 'rpool' to tape,
  and restore it back if in case the disks are lost.
  Backup would be done with an enterprise tool like tsm, legato etc.
 
 Backup/restore of bootable rpool to tape with a 3rd party application like
 legato etc is kind of difficult.  Because if you need to do a bare metal
 restore, how are you going to do it?  The root of the problem is the fact
 that you need an OS with legato in order to restore the OS.  It's a

If you're talking about Solaris 11 Express, you could create your own
liveCD using the Distribution Constructor[1] and include the backup
software on the cd image.  You'll have to customize the Distribution
Constructor to install the backup software (presumably via an SVR4
package[2]) but that's not too difficult.  Once you've created the image,
you're good to go forevermore (unless you need to update the backup
software on the image, in which case if you keep your Distribution
Constructor manifests around should be a simple edit to just point at
the newer backup software package)..

If you're talking about S10, then that's a tougher nut to crack.

Cheers,

-- 
Glenn

[1] - http://download.oracle.com/docs/cd/E19963-01/html/820-6564/
[2] - http://download.oracle.com/docs/cd/E19963-01/html/820-6564/addpkg.html
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] bootfs ID on zfs root

2011-05-11 Thread Ketan
So y my system is not coming up .. i jumpstarted the system again ...  but it 
panics like earlier .. so how should i recover it .. and get it up ? 

System was booted from network into single user mode and then rpool imported 
and following is the listing 


# zpool list
NAMESIZE  ALLOC   FREECAP  HEALTH  ALTROOT
rpool68G  4.08G  63.9G 5%  ONLINE  -
# zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
rpool 9.15G  57.8G98K  /rpool
rpool/ROOT4.08G  57.8G21K  /rpool/ROOT
rpool/ROOT/zfsBE_patched  4.08G  57.8G  4.08G  /
rpool/dump3.01G  60.8G16K  -
rpool/swap2.06G  59.9G16K  -
#



Dataset mos [META], ID 0, cr_txg 4, 137K, 62 objects
Dataset rpool/ROOT/zfsBE_patched [ZPL], ID 47, cr_txg 40, 4.08G, 110376 objects
Dataset rpool/ROOT [ZPL], ID 39, cr_txg 32, 21.0K, 4 objects
Dataset rpool/dump [ZVOL], ID 71, cr_txg 74, 16K, 2 objects
Dataset rpool/swap [ZVOL], ID 65, cr_txg 71, 16K, 2 objects
Dataset rpool [ZPL], ID 16, cr_txg 1, 98.0K, 10 objects


But when system is rebooted it again panics .. Is there any way to recover it ? 
I tried all the things which i know 


SunOS Release 5.10 Version Generic_142900-13 64-bit
Copyright 1983-2010 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
NOTICE: zfs_parse_bootfs: error 48
Cannot mount root on rpool/47 fstype zfs

panic[cpu0]/thread=180e000: vfs_mountroot: cannot mount root
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] bootfs ID on zfs root

2011-05-11 Thread Cindy Swearingen

Hi Ketan,

What steps lead up to this problem?

I believe the boot failure messages below are related to a mismatch
between the pool version and the installed OS version.

If you're using the JumpStart installation method, then the root pool is 
re-created each time, I believe. Does it also install a patch that

upgrades the pool version?

Thanks,

Cindy


On 05/11/11 13:27, Ketan wrote:
So y my system is not coming up .. i jumpstarted the system again ...  but it panics like earlier .. so how should i recover it .. and get it up ? 

System was booted from network into single user mode and then rpool imported and following is the listing 



# zpool list
NAMESIZE  ALLOC   FREECAP  HEALTH  ALTROOT
rpool68G  4.08G  63.9G 5%  ONLINE  -
# zfs list
NAME   USED  AVAIL  REFER  MOUNTPOINT
rpool 9.15G  57.8G98K  /rpool
rpool/ROOT4.08G  57.8G21K  /rpool/ROOT
rpool/ROOT/zfsBE_patched  4.08G  57.8G  4.08G  /
rpool/dump3.01G  60.8G16K  -
rpool/swap2.06G  59.9G16K  -
#



Dataset mos [META], ID 0, cr_txg 4, 137K, 62 objects
Dataset rpool/ROOT/zfsBE_patched [ZPL], ID 47, cr_txg 40, 4.08G, 110376 objects
Dataset rpool/ROOT [ZPL], ID 39, cr_txg 32, 21.0K, 4 objects
Dataset rpool/dump [ZVOL], ID 71, cr_txg 74, 16K, 2 objects
Dataset rpool/swap [ZVOL], ID 65, cr_txg 71, 16K, 2 objects
Dataset rpool [ZPL], ID 16, cr_txg 1, 98.0K, 10 objects


But when system is rebooted it again panics .. Is there any way to recover it ? I tried all the things which i know 



SunOS Release 5.10 Version Generic_142900-13 64-bit
Copyright 1983-2010 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
NOTICE: zfs_parse_bootfs: error 48
Cannot mount root on rpool/47 fstype zfs

panic[cpu0]/thread=180e000: vfs_mountroot: cannot mount root

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Modify stmf_sbd_lu properties

2011-05-11 Thread Don
It turns out this was actually as simple as:
stmfadm create-lu -p guid=XXX..

I kept looking at modify-lu to change this and never thought to check the 
create-lu options.

Thanks to Evaldas for the suggestion.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Backup complete rpool structure and data to tape

2011-05-11 Thread Peter Jeremy
On 2011-May-12 00:20:28 +0800, Edward Ned Harvey 
opensolarisisdeadlongliveopensola...@nedharvey.com wrote:
Backup/restore of bootable rpool to tape with a 3rd party application like
legato etc is kind of difficult.  Because if you need to do a bare metal
restore, how are you going to do it?

This is a generic problem, not limited to ZFS.  The generic solutions
are either:
a) Customised boot disk that includes the 3rd party restore client
b) Separate backup of root+client in a format that's restorable using
   tools only on the generic boot disk (eg tar or ufsdump).
(Where boot disk could be network boot instead of a physical CD/DVD).

I might suggest:  If you use zfs send to backup rpool to a file in the
data pool...  And then use legato etc to backup the data pool...

As Edward pointed out later, this might be OK as a disaster-recovery
approach but isn't suitable for the situation where you want to
restore a subset of the files (eg you need to recover a file someone
accidently deleted) and a zfs send stream isn't intended for storage.

Another potential downside is that the only way to read the stream is
using zfs recv into ZFS - this could present a problem if you wanted
to migrate the data into a different filesystem.  (All other restore
utilities I'm aware of use normal open/write/chmod/... interfaces so
you can restore your backup into any filesystem).

Finally, the send/recv protocol is not guaranteed to be compatible
between ZFS versions.  I'm not aware of any specific issues (though
someone reports that a zfs.v15 send | zfs.v22 recv caused pool
corruption in another recent thread) and would hope that zfs recv
would always maintain full compatibility with older zfs send.

But I hope you can completely abandon the whole 3rd party backup software
and tapes.  Some people can, and others cannot.  By far, the fastest best
way to backup ZFS is to use zfs send | zfs receive on another system or a
set of removable disks.

Unfortunately, this doesn't fit cleanly into the traditional
enterprise backup solution where Legato/NetBackup/TSM/... backs up
into a SILO with automatic tape replication and off-site rotation.

Incidentally, when you do incremental zfs send, you have to specify the
from and to snapshots.  So there must be at least one identical snapshot
in the sending and receiving system (or else your only option is to do a
complete full send.)

And (at least on v15) if you are using an incremental replication
stream and you create (or clone) a new descendent filesystem, you will
need to manually manage the initial replication of that filesystem.

BTW, if you do elect to build a bootable, removable drive for backups,
you should be aware that gzip compression isn't supported - at least
in v15, trying to make a gzip compressed filesystem bootable or trying
to set compression=gzip on a bootable filesystem gives a very
uninformative error message and it took a fair amount of trawling
through the source code to find the real cause.

-- 
Peter Jeremy


pgpnNCrRwuYrc.pgp
Description: PGP signature
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] bootfs ID on zfs root

2011-05-11 Thread Ketan
Hello Jim,

Thanks for the reply following is my o/p before setting bootfs parameter 


# zpool get all rpool
NAME   PROPERTY   VALUE SOURCE
rpool  size   68G   -
rpool  capacity   5%-
rpool  altroot- default
rpool  health ONLINE-
rpool  guid   8812174757237060985   default
rpool  version22default
rpool  bootfs rpool/ROOT/zfsBE_patched  local
rpool  delegation ondefault
rpool  autoreplaceoff   default
rpool  cachefile  - default
rpool  failmode   continue  local
rpool  listsnapshots  ondefault
rpool  autoexpand off   default
rpool  free   63.9G -
rpool  allocated  4.08G -

But i still ran the command .. but it didn't help me and system still 
panics 

# zpool set bootfs=rpool/ROOT/zfsBE_patched rpool

# zpool get all rpool
NAME   PROPERTY   VALUE SOURCE
rpool  size   68G   -
rpool  capacity   5%-
rpool  altroot- default
rpool  health ONLINE-
rpool  guid   8812174757237060985   default
rpool  version22default
rpool  bootfs rpool/ROOT/zfsBE_patched  local
rpool  delegation ondefault
rpool  autoreplaceoff   default
rpool  cachefile  - default
rpool  failmode   continue  local
rpool  listsnapshots  ondefault
rpool  autoexpand off   default
rpool  free   63.9G -
rpool  allocated  4.08G -
# init 6
#
The system is being restarted.
syncing file systems... done
rebooting...
Resetting...
POST Sequence 01 CPU Check
POST Sequence 02 Banner
LSB#00 (XSB#00-0): POST 2.14.0 (2010/05/13 13:27)
POST Sequence 03 Fatal Check
POST Sequence 04 CPU Register
POST Sequence 05 STICK
POST Sequence 06 MMU
POST Sequence 07 Memory Initialize
POST Sequence 08 Memory
POST Sequence 09 Raw UE In Cache
POST Sequence 0A Floating Point Unit
POST Sequence 0B SC
POST Sequence 0C Cacheable Instruction
POST Sequence 0D Softint
POST Sequence 0E CPU Cross Call
POST Sequence 0F CMU-CH
POST Sequence 10 PCI-CH
POST Sequence 11 Master Device
POST Sequence 12 DSCP
POST Sequence 13 SC Check Before STICK Diag
POST Sequence 14 STICK Stop
POST Sequence 15 STICK Start
POST Sequence 16 Error CPU Check
POST Sequence 17 System Configuration
POST Sequence 18 System Status Check
POST Sequence 19 System Status Check After Sync
POST Sequence 1A OpenBoot Start...
POST Sequence Complete.
ChassisSerialNumber BCF080207K

Sun SPARC Enterprise M5000 Server, using Domain console
Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
Copyright 2010 Sun Microsystems, Inc. and Fujitsu Limited. All rights reserved.
OpenBoot 4.24.14, 65536 MB memory installed, Serial #75515882.
Ethernet address 0:14:4f:80:47:ea, Host ID: 848047ea.



Rebooting with command: boot
Boot device: rootmirror  File and args:
SunOS Release 5.10 Version Generic_142900-13 64-bit
Copyright 1983-2010 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
NOTICE: zfs_parse_bootfs: error 48
Cannot mount root on rpool/47 fstype zfs

panic[cpu0]/thread=180e000: vfs_mountroot: cannot mount root

0180b950 genunix:vfs_mountroot+358 (800, 200, 0, 18ef800, 1918000, 
194dc00)
  %l0-3: 010c2400 010c22ec 018f5178 011f5400
  %l4-7: 011f5400 01950400 0600 0200
0180ba10 genunix:main+9c (0, 180c000, 1892260, 1833358, 1839738, 
1940800)
  %l0-3: 0180c000 0180c000 70002000 
  %l4-7: 0189c800  0180c000 0001

And the bootfs ID thing i read it from the following opensolaris link where one 
user was getting was the same error as mine. 

http://opensolaris.org/jive/thread.jspa?messageID=315743

Can you plz tell me some other way to get it rt ?
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss