Re: [zfs-discuss] ZFS ARC cache issue
When I looked for references on ARC freeing algo, I did find some lines of codes talking about freeing ARC when memory is under pressure. Nice...but what could be memory under pressure in the kernel syntax ? Jumping from C lines to blogs to docsI went back to basics : - lotsfree - fastscan IMHO the lotsfree (The greater of 1/64th of physical memory or 512 Kbytes) is stupid when you're using ZFS. Nicolas ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] [zones-discuss] Solaris 8/9 branded zones on ZFS root?
Le 26 févr. 09 à 15:47, Timothy Kennedy a écrit : Timothy Kennedy wrote: Nicolas Dorfsman wrote: Which ACL model is then used ? From: System Administration Guide: Solaris 8 Containers ( http://docs.sun.com/app/docs/doc/820-2914/gfjbk?a=view ) Using ZFS Although the zone cannot use a delegated ZFS dataset, the zone can reside on a ZFS file system. You can add a ZFS file system to share with the global zone through the zonecfg fs resource. See Step 7 in How to Configure a solaris8 Branded Zone. Note that the setfacl and getfacl commands cannot be used with ZFS. When a cpio or a tar archive with ACLs set on the files is unpacked, the archive will receive warnings about not being able to set the ACLs, although the files will be unpacked successfully. These commands can be used with UFS. Thanks Timothy ! Nico ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] [zones-discuss] Solaris 8/9 branded zones on ZFS root?
Le 25 févr. 09 à 23:12, Timothy Kennedy a écrit : Rich Teer wrote: I have a situation where I need to consolidate a few servers running Solaris 9 and 8. If the application doesn't run natively on Solaris 10 or Nevada, I was thinking of using Solars 9 or 8 branded zones. My intent would be for the global zone to use ZFS boot/root; would I be correct in thinking that this will be OK for the branded zones? That's correct. I have some solaris 8 zones running under cluster control, where zonepath is zfs, and they're doing just fine. Nothing special had to be done. Which ACL model is then used ? Nico ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Computer usable output for zpool commands
Hi, I wrote an hobbit script around lunmap/hbamap commands to monitor SAN health. I'd like to add detail on what is being hosted by those luns. With svm metastat -p is helpful. With zfs, zpool status output is awful for script. Is there somewhere an utility to show zpool informations in a scriptable format ? Nico This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] SAN arrays with NVRAM cache : ZIL and zfs_nocacheflush
Hi, I read some articles on solarisinternals.com like ZFS_Evil_Tuning_Guide on http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide . They clearly suggest to disable cache flush http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide#FLUSH . It seems to be the only serious article on the net about this subject. Could someone here state on this tuning suggestion ? My cu is running a HDS SAN array with Oracle on ZFS, I'd like to be clear in my brain. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Why root zone can't be on ZFS for upgrade ?
Hi, Something is unclear in Solaris containers and Solaris ZFS docs Two extracts : http://docs.sun.com/app/docs/doc/819-5461/6n7ht6qsm?q=zonea=view Consider the following interactions when working with ZFS on a system with Solaris zones installed: A ZFS file system that is added to a non-global zone must have its mountpoint property set to legacy. A ZFS file system cannot serve as zone root because of issues with the Solaris upgrade process. Do not include any system-related software that is accessed by the patch or upgrade process in a ZFS file system that is delegated to a non-global zone. http://docs.sun.com/app/docs/doc/817-1592/6mhahuop2?a=view 4. Set the zone path, /export/home/my-zone in this procedure. zonecfg:my-zone set zonepath=/export/home/my-zone Do not place the zonepath on ZFS for this release. I can't understand why the upgrade process need to have non-global root zone on anything else than zfs. Does the boot cdrom can't mount ZFS volumes ? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Why root zone can't be on ZFS for upgrade ?
Jeff wrote : The installation software does not yet understand ZFS, and is not able to upgrade a Solaris 10 system with a ZFS root file system. Further, it is not able to upgrade a Solaris 10 system with a non-global zone that has a ZFS file system as its zonepath. Thanks Jeff. Any idea on when Install Software will be able to see zfs vol ? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Veritas NetBackup Support for ZFS
I am using Netbackup 6.0 MP3 on several ZFS systems just fine. I think that NBU won't back up some exotic ACLs of ZFS, but if you are using ZFS like other filesystems (UFS, etc) then there aren't any issues. Hum. ACLs are not so exotic. This IS a really BIG issue. If you are using ACLs, even POSIX, moving production to ZFS filesystems means loosing any ACLs in backups. In other words, if you're using 30 years old UNIX rights, no problem. If I'd have to give a list of complaint on ZFS, that would be the first on my list ! Sun SHOULD make pressure on backup software editor (or send them some engineer) to support ZFS. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Snapshots and backing store
Hi, There's something really bizarre in ZFS snaphot specs : Uses no separate backing store. . Hum...if I want to mutualize one physical volume somewhere in my SAN as THE snaphots backing-store...it becomes impossible to do ! Really bad. Is there any chance to have a backing-store-file option in a future release ? In the same idea, it would be great to have some sort of propertie to add a disk/LUN/physical_space to a pool, only reserved to backing-store. At now, the only thing I see to disallow users to use my backing-store space for their usage is to put quota. Nico This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Snapshots and backing store
Well. ZFS isn't copy-on-write in the same way that things like ufssnap are. ufssnap is copy-on-write in that when you write something, it copies out the old data and writes it somewhere else (the backing store). ZFS doesn't need to do this - it simply writes the new data to a new location, and leaves the old data where it is. If that old data is needed for a snapshot then it's left unchanged, if it's not then it's freed. We need to think ZFS as ZFS, and not as a new filesystem ! I mean, the whole concept is different. So. What could be the best architecture ? With UFS, I used to have separate metadevices/LUNs for each application. With ZFS, I thought it would be nice to use a separate pool for each application. But, it means multiply snapshot backing-store OR dynamically remove/add this space/LUN to pool where we need to do backups. Knowing that I can't serialize backups, my only option is to multiply reservation for backing-stores. Uh ! Another option would be to create a single pool and put all apllications in it...don't think this as a solution. Any suggestion ? This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Snapshots and backing store
If you want to copy your filesystems (or snapshots) to other disks, you can use 'zfs send' to send them to a different pool (which may even be on a different machine!). Oh no ! It means copy the whole filesystem. The target here is definitively to snapshot the filesystem and them backup the snapshot. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: zfs share=.foo-internal.bar.edu on multipleinterfaces?
I have a Sun x4200 with 4x gigabit ethernet NICs. I have several of them configured with distinct IP addresses on an internal (10.0.0.0) network. [off topic] Why are you using distinct IP addresses instead of IPMP ? [/off] This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] ZFS on production servers with SLA
Hi, I'm currently doing some tests on a SF15K domain with Solaris 10 installed. The target is to convince my cu to use Solaris 10 for this domain AND establish a list of recommendations. The ZFS perimeter is really an issue for me. For now, I'm waiting for fresh informations from the backup software vendor about ZFS support. No ZFS-acl support could be annoying. Regarding system partitions (/var, /opt, all mirrored + alternate disk), what would be YOUR recommendations ? ZFS or not ? TIA Nicolas This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Changes during zpool import
extract from man zpool : [I]zpool import [-d dir] [-D] [-f] [-o opts] [-R root] pool | id [newpool] Imports a specific pool. A pool can be identified by its name or the numeric identifier. If newpool is specified, the pool is imported using the name newpool. Otherwise, it is imported with the same name as its exported name. If a device is removed from a system without running zpool export first, the device appears as potentially active. It cannot be determined if this was a failed export, or whether the device is really in use from another host. To import a pool in this state, the -f option is required. -d dir Searches for devices or files in dir. The -d option can be specified multiple times. -D Imports destroyed pool. The -f option is also required. -f Forces import, even if the pool appears to be potentially active. -o opts Comma-separated list of mount options to use when mounting datasets within the pool. See zfs(1M) for a description of dataset properties and mount options. -R root Imports pool(s) with an alternate root. See the Alternate Root Pools section.[/I] So, you could import it with '-R', and then modify its properties[i][/i][i][/i][b][/b][i][/i] This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss