The whole pool. Although you can choose to exclude individual datasets
from the flar when creating it.
lori
On 09/25/09 12:03, Peter Pickford wrote:
Hi Lori,
Is the u8 flash support for the whole root pool or an individual BE
using live upgrade?
Thanks
Peter
2009/9/24 Lori Alt <lori....@sun.com>:
On 09/24/09 15:54, Peter Pickford wrote:
Hi Cindy,
Wouldn't
touch /reconfigure
mv /etc/path_to_inst* /var/tmp/
regenerate all device information?
It might, but it's hard to say whether that would accomplish everything
needed to move a root file system from one system to another.
I just got done modifying flash archive support to work with zfs root on
Solaris 10 Update 8. For those not familiar with it, "flash archives" are a
way to clone full boot environments across multiple machines. The S10
Solaris installer knows how to install one of these flash archives on a
system and then do all the customizations to adapt it to the local hardware
and local network environment. I'm pretty sure there's more to the
customization than just a device reconfiguration.
So feel free to hack together your own solution. It might work for you, but
don't assume that you've come up with a completely general way to clone root
pools.
lori
AFIK zfs doesn't care about the device names it scans for them
it would only affect things like vfstab.
I did a restore from a E2900 to V890 and is seemed to work
Created the pool and zfs recieve.
I would like to be able to have a zfs send of a minimal build and
install it in an abe and activate it.
I tried that is test and it seems to work.
It seems to work but IM just wondering what I may have missed.
I saw someone else has done this on the list and was going to write a blog.
It seems like a good way to get a minimal install on a server with
reduced downtime.
Now if I just knew how to run the installer in and abe without there
being an OS there already that would be cool too.
Thanks
Peter
2009/9/24 Cindy Swearingen <cindy.swearin...@sun.com>:
Hi Peter,
I can't provide it because I don't know what it is.
Even if we could provide a list of items, tweaking
the device informaton if the systems are not identical
would be too difficult.
cs
On 09/24/09 12:04, Peter Pickford wrote:
Hi Cindy,
Could you provide a list of system specific info stored in the root pool?
Thanks
Peter
2009/9/24 Cindy Swearingen <cindy.swearin...@sun.com>:
Hi Karl,
Manually cloning the root pool is difficult. We have a root pool recovery
procedure that you might be able to apply as long as the
systems are identical. I would not attempt this with LiveUpgrade
and manually tweaking.
http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide#Complete_Solaris_ZFS_Root_Pool_Recovery
The problem is that the amount system-specific info stored in the root
pool and any kind of device differences might be insurmountable.
Solaris 10 ZFS/flash archive support is available with patches but not
for the Nevada release.
The ZFS team is working on a split-mirrored-pool feature and that might
be an option for future root pool cloning.
If you're still interested in a manual process, see the steps below
attempted by another community member who moved his root pool to a
larger disk on the same system.
This is probably more than you wanted to know...
Cindy
# zpool create -f altrpool c1t1d0s0
# zpool set listsnapshots=on rpool
# SNAPNAME=`date +%Y%m%d`
# zfs snapshot -r rpool/r...@$snapname
# zfs list -t snapshot
# zfs send -R rp...@$snapname | zfs recv -vFd altrpool
# installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk
/dev/rdsk/c1t1d0s0
for x86 do
# installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c1t1d0s0
Set the bootfs property on the root pool BE.
# zpool set bootfs=altrpool/ROOT/zfsBE altrpool
# zpool export altrpool
# init 5
remove source disk (c1t0d0s0) and move target disk (c1t1d0s0) to slot0
-insert solaris10 dvd
ok boot cdrom -s
# zpool import altrpool rpool
# init 0
ok boot disk1
On 09/24/09 10:06, Karl Rossing wrote:
I would like to clone the configuration on a v210 with snv_115.
The current pool looks like this:
-bash-3.2$ /usr/sbin/zpool status pool: rpool
state: ONLINE
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
rpool ONLINE 0 0 0
mirror ONLINE 0 0 0
c1t0d0s0 ONLINE 0 0 0
c1t1d0s0 ONLINE 0 0 0
errors: No known data errors
After I run zpool detach rpool c1t1d0s0, how can I remount c1t1d0s0 to
/tmp/a so that I can make the changes I need prior to removing the drive
and
putting it into the new v210.
I supose I could lucreate -n new_v210, lumount new_v210, edit what I
need
to, luumount new_v210, luactivate new_v210, zpool detach rpool c1t1d0s0
and
then luactivate the original boot environment.
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss