Re: [zfs-discuss] separate home partition?
On Fri, Jan 9, 2009 at 11:51 AM, Johan Hartzenberg jhart...@gmail.comwrote: I have this situation working and use my shared pool between Linux and Solaris. Note: The shared pool needs to reside on a whole physical disk or on a primary fdisk partition, Unless something changed since I last checked, Solaris' support for Logical Partitions are... not quite there yet. I just chanced apon the following in the SNV Build105 Change logs: PSARC case 2006/379 : Solaris on Extended partition BUG/RFE:6644364http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6644364Extended partitions need to be supported on Solaris BUG/RFE:6713308http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6713308Macro UNUSED in fdisk.h needs to be changed since id 100 is Novell Netware 286's partition ID BUG/RFE:6713318http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6713318Need to differentiate between solaris old partition and Linux swap BUG/RFE:6745175http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6745175Partitions can be created using fdisk table with invalid partition line by fdisk -F BUG/RFE:6745740http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6745740Multiple extended partition can be created by fdisk -A Files Changed: update:usr/src/Makefile.linthttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/Makefile.lint update:usr/src/Targetdirshttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/Targetdirs update:usr/src/cmd/boot/installgrub/Makefilehttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/boot/installgrub/Makefile update:usr/src/cmd/boot/installgrub/installgrub.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/boot/installgrub/installgrub.c update:usr/src/cmd/devfsadm/disk_link.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/devfsadm/disk_link.c update:usr/src/cmd/fdisk/Makefilehttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/fdisk/Makefile update:usr/src/cmd/fdisk/fdisk.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/fdisk/fdisk.c update:usr/src/cmd/format/Makefilehttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/format/Makefile update:usr/src/cmd/format/menu_fdisk.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/format/menu_fdisk.c update:usr/src/lib/Makefilehttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/lib/Makefile update:usr/src/pkgdefs/SUNWarc/prototype_i386http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/pkgdefs/SUNWarc/prototype_i386 update:usr/src/pkgdefs/SUNWarcr/prototype_i386http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/pkgdefs/SUNWarcr/prototype_i386 update:usr/src/pkgdefs/SUNWcsl/prototype_i386http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/pkgdefs/SUNWcsl/prototype_i386 update:usr/src/pkgdefs/SUNWcslr/prototype_i386http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/pkgdefs/SUNWcslr/prototype_i386 update:usr/src/pkgdefs/SUNWhea/prototype_i386http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/pkgdefs/SUNWhea/prototype_i386 update:usr/src/uts/common/io/cmlb.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/io/cmlb.c update:usr/src/uts/common/io/scsi/targets/sd.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/io/scsi/targets/sd.c update:usr/src/uts/common/sys/cmlb_impl.hhttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/sys/cmlb_impl.h update:usr/src/uts/common/sys/dkio.hhttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/sys/dkio.h update:usr/src/uts/common/sys/dktp/fdisk.hhttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/sys/dktp/fdisk.h update:usr/src/uts/common/sys/scsi/targets/sddef.hhttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/sys/scsi/targets/sddef.h update:usr/src/uts/common/xen/io/xdf.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/xen/io/xdf.c update:usr/src/uts/intel/io/dktp/disk/cmdk.chttp://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/intel/io/dktp/disk/cmdk.c create:usr/src/lib/libfdisk/Makefile create:usr/src/lib/libfdisk/i386/Makefile create:usr/src/lib/libfdisk/i386/libfdisk.c create:usr/src/lib/libfdisk/i386/libfdisk.h create:usr/src/lib/libfdisk/i386/llib-lfdisk create:usr/src/lib/libfdisk/i386/mapfile-vers -- Any sufficiently advanced technology is indistinguishable from magic. Arthur C. Clarke My blog: http://initialprogramload.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
On Tue, Jan 13, 2009 at 2:42 PM, Johan Hartzenberg jhart...@gmail.com wrote: On Fri, Jan 9, 2009 at 11:51 AM, Johan Hartzenberg jhart...@gmail.com wrote: I have this situation working and use my shared pool between Linux and Solaris. Note: The shared pool needs to reside on a whole physical disk or on a primary fdisk partition, Unless something changed since I last checked, Solaris' support for Logical Partitions are... not quite there yet. I just chanced apon the following in the SNV Build105 Change logs: This work was removed later - see http://hg.genunix.org/onnv-gate.hg/rev/de8038a7796e PSARC case 2006/379 : Solaris on Extended partition BUG/RFE:6644364Extended partitions need to be supported on Solaris BUG/RFE:6713308Macro UNUSED in fdisk.h needs to be changed since id 100 is Novell Netware 286's partition ID BUG/RFE:6713318Need to differentiate between solaris old partition and Linux swap BUG/RFE:6745175Partitions can be created using fdisk table with invalid partition line by fdisk -F BUG/RFE:6745740Multiple extended partition can be created by fdisk -A Files Changed: update:usr/src/Makefile.lint update:usr/src/Targetdirs update:usr/src/cmd/boot/installgrub/Makefile update:usr/src/cmd/boot/installgrub/installgrub.c update:usr/src/cmd/devfsadm/disk_link.c update:usr/src/cmd/fdisk/Makefile update:usr/src/cmd/fdisk/fdisk.c update:usr/src/cmd/format/Makefile update:usr/src/cmd/format/menu_fdisk.c update:usr/src/lib/Makefile update:usr/src/pkgdefs/SUNWarc/prototype_i386 update:usr/src/pkgdefs/SUNWarcr/prototype_i386 update:usr/src/pkgdefs/SUNWcsl/prototype_i386 update:usr/src/pkgdefs/SUNWcslr/prototype_i386 update:usr/src/pkgdefs/SUNWhea/prototype_i386 update:usr/src/uts/common/io/cmlb.c update:usr/src/uts/common/io/scsi/targets/sd.c update:usr/src/uts/common/sys/cmlb_impl.h update:usr/src/uts/common/sys/dkio.h update:usr/src/uts/common/sys/dktp/fdisk.h update:usr/src/uts/common/sys/scsi/targets/sddef.h update:usr/src/uts/common/xen/io/xdf.c update:usr/src/uts/intel/io/dktp/disk/cmdk.c create:usr/src/lib/libfdisk/Makefile create:usr/src/lib/libfdisk/i386/Makefile create:usr/src/lib/libfdisk/i386/libfdisk.c create:usr/src/lib/libfdisk/i386/libfdisk.h create:usr/src/lib/libfdisk/i386/llib-lfdisk create:usr/src/lib/libfdisk/i386/mapfile-vers -- Any sufficiently advanced technology is indistinguishable from magic. Arthur C. Clarke My blog: http://initialprogramload.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- Regards, Cyril ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
On Fri, Jan 9, 2009 at 4:10 AM, noz sf2...@gmail.com wrote: Here's my solution: (1) n...@holodeck:~# zpool create epool mirror c4t1d0 c4t2d0 c4t3d0 n...@holodeck:~# zfs list NAME USED AVAIL REFER MOUNTPOINT epool 69K 15.6G18K /epool rpool 3.68G 11.9G72K /rpool rpool/ROOT 2.81G 11.9G18K legacy rpool/ROOT/opensolaris 2.81G 11.9G 2.68G / rpool/dump 383M 11.9G 383M - rpool/export 632K 11.9G19K /export rpool/export/home612K 11.9G19K /export/home rpool/export/home/noz594K 11.9G 594K /export/home/noz rpool/swap 512M 12.4G 21.1M - n...@holodeck:~# (2) n...@holodeck:~# zfs snapshot -r rpool/exp...@now (3) n...@holodeck:~# zfs send -R rpool/exp...@now /tmp/export_now (4) n...@holodeck:~# zfs destroy -r -f rpool/export (5) n...@holodeck:~# zfs recv -d epool /tmp/export_now The above is very dangerous, if it will even work. The output of the zfs send is redirected to /tmp, which is a ramdisk. If you have enough space (RAM + Swap), it will work, but if there is a reboot or crash before the zfs receive completes then everything is gone. In stead, do the following: (2) n...@holodeck:~# zfs snapshot -r rpool/exp...@now (3) n...@holodeck:~# zfs send -R rpool/exp...@now | zfs recv -d epool (4) Check that all the data looks OK in epool (5) n...@holodeck:~# zfs destroy -r -f rpool/export -- Any sufficiently advanced technology is indistinguishable from magic. Arthur C. Clarke My blog: http://initialprogramload.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
On Fri, Jan 9, 2009 at 9:55 AM, hardware technician figh...@yahoo.comwrote: I want to create a separate home, shared, read/write zfs partition on a tri-boot OpenSolaris, Ubuntu, and CentOS system. I have successfully created and exported the zpools that I would like to use, in Ubuntu using zfs-fuse. However, I boot into OpenSolaris, and I type zpool import with no options. The only pool I see to import is on the primary partition, and I haven't been able to see or import the pool that is on the extended partition. I have tried importing using the name, and ID. In OpenSolaris /dev/dsk/c3d0 shows 15 slices, so I think the slices are there, but then I type format, select the disk, and the partition option, but it doesn't show (zfs) partitions from linux. In format, the fdisk option recognizes the (zfs) linux partitions. The partition that I was able to import is on the first partition, and is named c3d0p1, and is not a slice. Are there any ideas how I could import the other pool? I have this situation working and use my shared pool between Linux and Solaris. Note: The shared pool needs to reside on a whole physical disk or on a primary fdisk partition, Unless something changed since I last checked, Solaris' support for Logical Partitions are... not quite there yet. P.S. I blogged about my setup (Linux + Solaris with a Shared ZFS pool) here http://initialprogramload.blogspot.com/search?q=zfs-fuse+linux ... However this was a long time ago and I don't know whether the statement about Grub ZFS support in point 3 is still true. Aparently some bugs pertaining to time stomping between ubuntu and solaris has been fixed, so you may not need to do step 4. An Alternative to step 4 is to run this in Solaris: pfexec /usr/sbin/rtc -z UTC In addition, at point nr 7, use bootadm list-menu to find out where Solaris has decided to save the grub menu.lst file. -- Any sufficiently advanced technology is indistinguishable from magic. Arthur C. Clarke My blog: http://initialprogramload.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
The above is very dangerous, if it will even work. The output of the zfs send is redirected to /tmp, which is a ramdisk. If you have enough space (RAM + Swap), it will work, but if there is a reboot or crash before the zfs receive completes then everything is gone. In stead, do the following: (2) n...@holodeck:~# zfs snapshot -r rpool/exp...@now (3) n...@holodeck:~# zfs send -R rpool/exp...@now | zfs recv -d epool (4) Check that all the data looks OK in epool (5) n...@holodeck:~# zfs destroy -r -f rpool/export Thanks for the tip. Is there an easy way to do your revised step 4? Can I use a diff or something similar? e.g. diff rpool/export epool/export -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
On Fri, Jan 9, 2009 at 6:25 PM, noz sf2...@gmail.com wrote: The above is very dangerous, if it will even work. The output of the zfs send is redirected to /tmp, which is a ramdisk. If you have enough space (RAM + Swap), it will work, but if there is a reboot or crash before the zfs receive completes then everything is gone. In stead, do the following: (2) n...@holodeck:~# zfs snapshot -r rpool/exp...@now (3) n...@holodeck:~# zfs send -R rpool/exp...@now | zfs recv -d epool (4) Check that all the data looks OK in epool (5) n...@holodeck:~# zfs destroy -r -f rpool/export Thanks for the tip. Is there an easy way to do your revised step 4? Can I use a diff or something similar? e.g. diff rpool/export epool/export Personally I would just browse around the structure, open a few files at random, and consider it done. But that is me, and my data, of which I _DO_ make backups. You could use find to create an index of all the files and save these in files, and compare those. Depending on exactly how you do the find, you might be able to just diff the files. Of course if you want to be realy pedantic, you would do cd /rpool/export; find . | xargs cksum /rpool_checksums cd /epool/export; find . | xargs cksum /epool_checksums diff /?pool_checksums But be prepared to wait a very very very long time for the two checksum processes to run. Unless you have very little data. Cheers, _J -- Any sufficiently advanced technology is indistinguishable from magic. Arthur C. Clarke My blog: http://initialprogramload.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
Kyle wrote: So if preserving the home filesystem through re-installs are really important, putting the home filesystem in a separate pool may be in order. My problem similar to the original thread author, and this scenario is exactly the one I had in mind. I figured out a workable solution from the zfs admin guide, but I've only tested this in virtualbox. I have no idea how well this would work if I actually had hundreds of gigabytes of data. I also don't know if my solution is the recommended way to do this, so please let me know if anyone has a better method. Here's my solution: (1) n...@holodeck:~# zpool create epool mirror c4t1d0 c4t2d0 c4t3d0 n...@holodeck:~# zfs list NAME USED AVAIL REFER MOUNTPOINT epool 69K 15.6G18K /epool rpool 3.68G 11.9G72K /rpool rpool/ROOT 2.81G 11.9G18K legacy rpool/ROOT/opensolaris 2.81G 11.9G 2.68G / rpool/dump 383M 11.9G 383M - rpool/export 632K 11.9G19K /export rpool/export/home612K 11.9G19K /export/home rpool/export/home/noz594K 11.9G 594K /export/home/noz rpool/swap 512M 12.4G 21.1M - n...@holodeck:~# (2) n...@holodeck:~# zfs snapshot -r rpool/exp...@now (3) n...@holodeck:~# zfs send -R rpool/exp...@now #62; /tmp/export_now (4) n...@holodeck:~# zfs destroy -r -f rpool/export (5) n...@holodeck:~# zfs recv -d epool #60; /tmp/export_now n...@holodeck:~# zfs list NAME USED AVAIL REFER MOUNTPOINT epool756K 15.6G18K /epool epool/export 630K 15.6G19K /export epool/export/home612K 15.6G19K /export/home epool/export/home/noz592K 15.6G 592K /export/home/noz rpool 3.68G 11.9G72K /rpool rpool/ROOT 2.81G 11.9G18K legacy rpool/ROOT/opensolaris 2.81G 11.9G 2.68G / rpool/dump 383M 11.9G 383M - rpool/swap 512M 12.4G 21.1M - n...@holodeck:~# (6) n...@holodeck:~# zfs mount -a or (6) reboot The only part I'm uncomfortable with is when I have to destroy rpool's export filesystem (step 4), because trying to destroy without the -f switch results in a filesystem is active error. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
To do step no 4, you need to login as root, or create new user which home dir not at export. Sent from my iPhone On Jan 9, 2009, at 10:10 AM, noz sf2...@gmail.com wrote: Kyle wrote: So if preserving the home filesystem through re-installs are really important, putting the home filesystem in a separate pool may be in order. My problem similar to the original thread author, and this scenario is exactly the one I had in mind. I figured out a workable solution from the zfs admin guide, but I've only tested this in virtualbox. I have no idea how well this would work if I actually had hundreds of gigabytes of data. I also don't know if my solution is the recommended way to do this, so please let me know if anyone has a better method. Here's my solution: (1) n...@holodeck:~# zpool create epool mirror c4t1d0 c4t2d0 c4t3d0 n...@holodeck:~# zfs list NAME USED AVAIL REFER MOUNTPOINT epool 69K 15.6G18K /epool rpool 3.68G 11.9G72K /rpool rpool/ROOT 2.81G 11.9G18K legacy rpool/ROOT/opensolaris 2.81G 11.9G 2.68G / rpool/dump 383M 11.9G 383M - rpool/export 632K 11.9G19K /export rpool/export/home612K 11.9G19K /export/home rpool/export/home/noz594K 11.9G 594K /export/home/noz rpool/swap 512M 12.4G 21.1M - n...@holodeck:~# (2) n...@holodeck:~# zfs snapshot -r rpool/exp...@now (3) n...@holodeck:~# zfs send -R rpool/exp...@now #62; /tmp/export_now (4) n...@holodeck:~# zfs destroy -r -f rpool/export (5) n...@holodeck:~# zfs recv -d epool #60; /tmp/export_now n...@holodeck:~# zfs list NAME USED AVAIL REFER MOUNTPOINT epool756K 15.6G18K /epool epool/export 630K 15.6G19K /export epool/export/home612K 15.6G19K /export/home epool/export/home/noz592K 15.6G 592K /export/home/noz rpool 3.68G 11.9G72K /rpool rpool/ROOT 2.81G 11.9G18K legacy rpool/ROOT/opensolaris 2.81G 11.9G 2.68G / rpool/dump 383M 11.9G 383M - rpool/swap 512M 12.4G 21.1M - n...@holodeck:~# (6) n...@holodeck:~# zfs mount -a or (6) reboot The only part I'm uncomfortable with is when I have to destroy rpool's export filesystem (step 4), because trying to destroy without the -f switch results in a filesystem is active error. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
To do step no 4, you need to login as root, or create new user which home dir not at export. Sent from my iPhone I tried to login as root at the login screen but it wouldn't let me, some error about roles. Is there another way to login as root? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
You can edit /etc/user_attr file. Sent from my iPhone On Jan 9, 2009, at 11:13 AM, noz sf2...@gmail.com wrote: To do step no 4, you need to login as root, or create new user which home dir not at export. Sent from my iPhone I tried to login as root at the login screen but it wouldn't let me, some error about roles. Is there another way to login as root? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
I want to create a separate home, shared, read/write zfs partition on a tri-boot OpenSolaris, Ubuntu, and CentOS system. I have successfully created and exported the zpools that I would like to use, in Ubuntu using zfs-fuse. However, I boot into OpenSolaris, and I type zpool import with no options. The only pool I see to import is on the primary partition, and I haven't been able to see or import the pool that is on the extended partition. I have tried importing using the name, and ID. In OpenSolaris /dev/dsk/c3d0 shows 15 slices, so I think the slices are there, but then I type format, select the disk, and the partition option, but it doesn't show (zfs) partitions from linux. In format, the fdisk option recognizes the (zfs) linux partitions. The partition that I was able to import is on the first partition, and is named c3d0p1, and is not a slice. Are there any ideas how I could import the other pool? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
at the risk of venturing off topic: that looks like a good revisioning scheme. opensolaris creates a new (default) boot environment during the update process, which seems like a very cool feature. seems like. when i update my 2008.11 install, nwam breaks, apparently a known bug (the workaround didn't work for me, probably due to inexperience). no problem me thinks, i just boot into the old environment. nwam still broken. i had naively assumed that the new BE was a delta, again, i know little. anyway, back to zfs, you didn't voice any alarm at my virtual home folder scheme. regarding root partition, i'll think it over, but given my luck with updates, i don't imagine doing any. thank you once again for all of your valuable input. scott -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
thanks for the input. since i have no interest in multibooting (virtualbox will suit my needs), i created a 10gb partition on my 500gb drive for opensolaris and reserved the rest for files (130gb worth). after installing the os and fdisking the rest of the space to solaris2, i created a zpool called DOCUMENTS (good tips with the upper case), which i then mounted to Documents in my home folder. the logic is, if i have to reinstall, i just export DOCUMENTS and re-import it into the reinstalled os (or import -f in a worst-case scenario). after having done all the setup, i partitioned drive 2 using identical cylinder locs and mirrored each into their respective pools (rpool and DOCUMENTS). replacing drive 1 with 2 and starting back up, everything boots fine and i see all my data, so it worked. obviously i'm a noob, and yet even i find my own method a little suspicious. i look at the disk usage analyzer and see that / is 100% used. while i'm sure that this is in some kind of virtual sense, it leaves me with a feeling that i've done a goofy thing. comments about this last concern are greatly appreciated! -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
On Sat, Dec 27, 2008 at 7:33 AM, scott new.mail.ba...@gmail.com wrote: do you mean a pool on a SEPARATE partition? -- That is what I do. In particular, I have: fdisk partition 1 = Solaris partition type 0xbf = rpool = 40 GB fdisk partition 2 = MSDOS partition type = SHARED zpool = 190 GB fdisk partition 3 = 30 GB Extended partition. Logical partition 5 used for Ubuntu Root, Logical Partition 6 = Ubuntu Swap. This leaves me with the option of creating an fdisk partition 4 for another operating system. Disadvantages: 1. Partitioning means ZFS does not turn on write-caching. 2. Also there is wasted space. (Partitioning implies pre-allocating space, which means you have to dedicate space that you may not use) Advantages: 1. I can import the SHARED zpool under Ubuntu and thus I have the perfect shared space solution between the two operating systems, without having to worry about clashing mount points which would be present if I tried to import the root pool. 2. If I needed to re-install, I would only wipe/destroy/touch the OS, not my user data. I have not yet made the move from Solaris Express to OpenSolaris, so I am still using Live Upgrade. I generally upgrade to every new release, sometimes to my sorrow. But it does not touch my SHARED data zpool. One other thing: I started a convention of using all-capital names for my ZFS pool names. It makes them stand out nicely in the output of df and mount, but in particular ir distinguishes nicely between the pool name and the mountpoint because I then mount the SHARED pool on /shared. -- Any sufficiently advanced technology is indistinguishable from magic. Arthur C. Clarke My blog: http://initialprogramload.blogspot.com ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] separate home partition?
(i use the term loosely because i know that zfs likes whole volumes better) when installing ubuntu, i got in the habit of using a separate partition for my home directory so that my data and gnome settings would all remain intact when i reinstalled or upgraded. i'm running osol 2008.11 on an ultra 20, which has only two drives. i've got all my data located in my home directory, and the two drives are cloned (mirrored and bootable). i want to have a similar option as my linux setup, but want to play by the zfs (best practice) rules as much as possible. thanks in advance -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] separate home partition?
do you mean a pool on a SEPARATE partition? -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss