Hi Lori, Is the u8 flash support for the whole root pool or an individual BE using live upgrade?
Thanks Peter 2009/9/24 Lori Alt <lori....@sun.com>: > On 09/24/09 15:54, Peter Pickford wrote: > > Hi Cindy, > > Wouldn't > > touch /reconfigure > mv /etc/path_to_inst* /var/tmp/ > > regenerate all device information? > > > It might, but it's hard to say whether that would accomplish everything > needed to move a root file system from one system to another. > > I just got done modifying flash archive support to work with zfs root on > Solaris 10 Update 8. For those not familiar with it, "flash archives" are a > way to clone full boot environments across multiple machines. The S10 > Solaris installer knows how to install one of these flash archives on a > system and then do all the customizations to adapt it to the local hardware > and local network environment. I'm pretty sure there's more to the > customization than just a device reconfiguration. > > So feel free to hack together your own solution. It might work for you, but > don't assume that you've come up with a completely general way to clone root > pools. > > lori > > AFIK zfs doesn't care about the device names it scans for them > it would only affect things like vfstab. > > I did a restore from a E2900 to V890 and is seemed to work > > Created the pool and zfs recieve. > > I would like to be able to have a zfs send of a minimal build and > install it in an abe and activate it. > I tried that is test and it seems to work. > > It seems to work but IM just wondering what I may have missed. > > I saw someone else has done this on the list and was going to write a blog. > > It seems like a good way to get a minimal install on a server with > reduced downtime. > > Now if I just knew how to run the installer in and abe without there > being an OS there already that would be cool too. > > Thanks > > Peter > > 2009/9/24 Cindy Swearingen <cindy.swearin...@sun.com>: > > > Hi Peter, > > I can't provide it because I don't know what it is. > > Even if we could provide a list of items, tweaking > the device informaton if the systems are not identical > would be too difficult. > > cs > > On 09/24/09 12:04, Peter Pickford wrote: > > > Hi Cindy, > > Could you provide a list of system specific info stored in the root pool? > > Thanks > > Peter > > 2009/9/24 Cindy Swearingen <cindy.swearin...@sun.com>: > > > Hi Karl, > > Manually cloning the root pool is difficult. We have a root pool recovery > procedure that you might be able to apply as long as the > systems are identical. I would not attempt this with LiveUpgrade > and manually tweaking. > > > http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide#Complete_Solaris_ZFS_Root_Pool_Recovery > > The problem is that the amount system-specific info stored in the root > pool and any kind of device differences might be insurmountable. > > Solaris 10 ZFS/flash archive support is available with patches but not > for the Nevada release. > > The ZFS team is working on a split-mirrored-pool feature and that might > be an option for future root pool cloning. > > If you're still interested in a manual process, see the steps below > attempted by another community member who moved his root pool to a > larger disk on the same system. > > This is probably more than you wanted to know... > > Cindy > > > > # zpool create -f altrpool c1t1d0s0 > # zpool set listsnapshots=on rpool > # SNAPNAME=`date +%Y%m%d` > # zfs snapshot -r rpool/r...@$snapname > # zfs list -t snapshot > # zfs send -R rp...@$snapname | zfs recv -vFd altrpool > # installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk > /dev/rdsk/c1t1d0s0 > for x86 do > # installgrub /boot/grub/stage1 /boot/grub/stage2 /dev/rdsk/c1t1d0s0 > Set the bootfs property on the root pool BE. > # zpool set bootfs=altrpool/ROOT/zfsBE altrpool > # zpool export altrpool > # init 5 > remove source disk (c1t0d0s0) and move target disk (c1t1d0s0) to slot0 > -insert solaris10 dvd > ok boot cdrom -s > # zpool import altrpool rpool > # init 0 > ok boot disk1 > > On 09/24/09 10:06, Karl Rossing wrote: > > > I would like to clone the configuration on a v210 with snv_115. > > The current pool looks like this: > > -bash-3.2$ /usr/sbin/zpool status pool: rpool > state: ONLINE > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > rpool ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c1t0d0s0 ONLINE 0 0 0 > c1t1d0s0 ONLINE 0 0 0 > > errors: No known data errors > > After I run zpool detach rpool c1t1d0s0, how can I remount c1t1d0s0 to > /tmp/a so that I can make the changes I need prior to removing the drive > and > putting it into the new v210. > > I supose I could lucreate -n new_v210, lumount new_v210, edit what I > need > to, luumount new_v210, luactivate new_v210, zpool detach rpool c1t1d0s0 > and > then luactivate the original boot environment. > > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss