[zfs-discuss] How to recover from "can't open objset", "cannot iterate filesystems"?
Recently upgraded a system from b98 to b114. Also replaced two 400G Seagate Barracudea 7200.8 SATA disks with two WD 750G RE3 SATA disks from a 6-device raidz1 pool. Replacing the first 750G went ok. While replacing the second 750G disk, I noticed CKSUM errors on the first disk. Once the second disk was replaced, I halted the system, upgraded to b114, and rebooted. Both b98 and b114 gave the errors: WARNING: can't open objset for tww/opt/dists/cd-8.1 cannot iterate filesystems: I/O error How do I recover from this? # zpool status tww pool: tww state: ONLINE status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: http://www.sun.com/msg/ZFS-8000-8A scrub: none requested config: NAMESTATE READ WRITE CKSUM tww ONLINE 0 0 3 raidz1ONLINE 0 012 c4t0d0 ONLINE 0 0 0 c4t1d0 ONLINE 0 0 0 c4t4d0 ONLINE 0 0 0 c4t5d0 ONLINE 0 0 0 c4t6d0 ONLINE 0 0 0 c4t7d0 ONLINE 0 0 0 errors: 855 data errors, use '-v' for a list -- albert chin (ch...@thewrittenword.com) ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re-import RAID-Z2 with faulted disk
I'm running vanilla 2009.06 since its release. I'll definitely give it a shot with the Live CD. Also I tried importing with only the five good disks physically attached and get the same message. - Kyle On Mon, Sep 21, 2009 at 3:50 AM, Chris Murray wrote: > That really sounds like a scenario that ZFS would be able to cope with. > > What operating system are you using now? I've had good results > 'fixing' pools with problems which were preventing import in the past by > using the Opensolaris 2009.06 Live CD. It would depend what build you're on > now as to whether that will yield any results, of course ... > > Also, is the import any more successful if the new, empty disk just isn't > present at all, and c9t2d0 is connected to nothing? > > Chris > > 2009/9/21 Kyle J. Aleshire > >> Hi all, I have a RAID-Z2 setup with 6x 500Gb SATA disks. I exported the >> array to use under a different system but during or after the export one of >> the disks failed: >> >> k...@localhost:~$ pfexec zpool import >> pool: chronicle >> id: 11592382930413748377 >> state: DEGRADED >> status: One or more devices are missing from the system. >> action: The pool can be imported despite missing or damaged devices. The >> fault tolerance of the pool may be compromised if imported. >>see: http://www.sun.com/msg/ZFS-8000-2Q >> config: >> >> chronicle DEGRADED >> raidz2DEGRADED >> c9t2d0 UNAVAIL cannot open >> c9t1d0 ONLINE >> c9t0d0 ONLINE >> c9t4d0 ONLINE >> c9t5d0 ONLINE >> c9t3d0 ONLINE >> >> I have no success trying to reimport the pool: >> k...@localhost:~$ pfexec zpool import -f chronicle >> cannot import 'chronicle': one or more devices is currently unavailable >> >> The disk has since been replaced, so now: >> k...@localhost:~$ pfexec zpool import >> pool: chronicle >> id: 11592382930413748377 >> state: DEGRADED >> status: One or more devices contains corrupted data. >> action: The pool can be imported despite missing or damaged devices. The >> fault tolerance of the pool may be compromised if imported. >>see: http://www.sun.com/msg/ZFS-8000-4J >> config: >> >> chronicle DEGRADED >> raidz2DEGRADED >> c9t2d0 FAULTED corrupted data >> c9t1d0 ONLINE >> c9t0d0 ONLINE >> c9t4d0 ONLINE >> c9t5d0 ONLINE >> c9t3d0 ONLINE >> >> but "pfexec zpool import -f chronicle" still fails with the same message. >> >> I've Google'd this several times for a fix but to no avail. Any assistance >> is appreciated. >> >> ___ >> zfs-discuss mailing list >> zfs-discuss@opensolaris.org >> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss >> >> > ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re-import RAID-Z2 with faulted disk
On Mon, Sep 21, 2009 at 3:37 AM, wrote: > > > > >The disk has since been replaced, so now: > >k...@localhost:~$ pfexec zpool import > > pool: chronicle > >id: 11592382930413748377 > > state: DEGRADED > >status: One or more devices contains corrupted data. > >action: The pool can be imported despite missing or damaged devices. The > >fault tolerance of the pool may be compromised if imported. > > see: http://www.sun.com/msg/ZFS-8000-4J > >config: > > > >chronicle DEGRADED > > raidz2DEGRADED > >c9t2d0 FAULTED corrupted data > >c9t1d0 ONLINE > >c9t0d0 ONLINE > >c9t4d0 ONLINE > >c9t5d0 ONLINE > >c9t3d0 ONLINE > > > >but "pfexec zpool import -f chronicle" still fails with the same message. > > > > That sounds like a bug; if ZFS can recover then zpool import should > also work. > > So what was wrong with the broken disk? Not just badly plugged in? > > Cas[er > > The disk physically failed. When I powered the system on, it made frighteningly loud clicking sounds and would not be acknowledged by any system I was running, like it wasn't even there. - Kyle ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] possibilities of AFP ever making it into ZFS like NFS an
I was able to get Netatalk built on OpenSolaris for my ZFS NAS at home. Everything is running great so far, and I'm planning on using it on the 96TB NAS I'm building for my office. It would be nice to have this supported out of the box, but there are probably licensing issues involved. -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS file disk usage
On Sep 21, 2009, at 2:43 PM, Andrew Deason wrote: On Mon, 21 Sep 2009 17:13:26 -0400 Richard Elling wrote: OK, so the problem you are trying to solve is "how much stuff can I place in the remaining free space?" I don't think this is knowable for a dynamic file system like ZFS where metadata is dynamically allocated. Yes. And I acknowledge that we can't know that precisely; I'm trying for an estimate on the bound. You don't know the max overhead for the file before it is allocated. You could guess at a max of 3x size + at least three blocks. Since you can't control this, it seems like the worst case is when copies=3. Is that max with copies=3? Assume copies=1; what is it then? 1x size + 1 block. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS file disk usage
On Mon, 21 Sep 2009 17:13:26 -0400 Richard Elling wrote: > OK, so the problem you are trying to solve is "how much stuff can I > place in the remaining free space?" I don't think this is knowable > for a dynamic file system like ZFS where metadata is dynamically > allocated. Yes. And I acknowledge that we can't know that precisely; I'm trying for an estimate on the bound. > You don't know the max overhead for the file before it is allocated. > You could guess at a max of 3x size + at least three blocks. Since > you can't control this, it seems like the worst case is when copies=3. Is that max with copies=3? Assume copies=1; what is it then? -- Andrew Deason adea...@sinenomine.net ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] If you have ZFS in production, willing to share some details (with me)?
On Fri, Sep 18, 2009 at 01:51:52PM -0400, Steffen Weiberle wrote: > I am trying to compile some deployment scenarios of ZFS. > > # of systems One, our e-mail server for the entire campus. > amount of storage 2 TB that's 58% used. > application profile(s) This is our Cyrus IMAP spool. In addition to user's e-mail folders (directories) and messages (files), it contains global, per-folder, and per-user databases. The latter two types are quite small. > type of workload (low, high; random, sequential; read-only, read-write, > write-only) It's quite active. Message files arrive randomly and are deleted randomly. As a result, files in a directory are not located in proximity on the storage. Individual users often read all of their folders and messages in one IMAP session. Databases are quite active. Each incoming message adds a file to a directory and reads or updates several databases. Most IMAP I/O is done with mmap() rather than with read()/write(). So far, IMAP peformance is adequate. The backup, done by EMC Networker, is very slow because it must read thousands of small files in directory order. > storage type(s) We are using an Iscsi SAN with storage on a Netapp filer. It exports four 500-gb LUNs that are striped into one ZFS pool. All disk mangement is done on the Netapp. We have had several disk failures and replacements on the Netapp, with no effect on the e-mail server. > industry A University with 35,000 enabled e-mail accounts. > whether it is private or I can share in a summary > anything else that might be of interest You are welcome to share this information. -- -Gary Mills--Unix Group--Computer and Network Services- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS file disk usage
On Sep 21, 2009, at 7:11 AM, Andrew Deason wrote: On Sun, 20 Sep 2009 20:31:57 -0400 Richard Elling wrote: If you are just building a cache, why not just make a file system and put a reservation on it? Turn off auto snapshots and set other features as per best practices for your workload? In other words, treat it like we treat dump space. I think that we are getting caught up in trying to answer the question you ask rather than solving the problem you have... perhaps because we don't understand the problem. Yes, possibly... some of these suggestions dont quite make a lot of sense to me. We can't just make a filesystem and put a reservation on it; we are just an application the administrator puts on a machine for it to access AFS. So I'm not sure when you are imagining we do that; when the client starts up? Or part of the installation procedure? Requiring a separate filesystem seems unnecessarily restrictive. And I still don't see how that helps. Making an fs with a reservation would definitely limit us to the specified space, but we still can't get an accurate picture of the current disk usage. I already mentioned why using statvfs is not usable with that commit delay. OK, so the problem you are trying to solve is "how much stuff can I place in the remaining free space?" I don't think this is knowable for a dynamic file system like ZFS where metadata is dynamically allocated. But solving the general problem for me isn't necessary. If I could just get a ballpark estimate of the max overhead for a file, I would be fine. I haven't payed attention to it before, so I don't even have an intuitive feel for what it is. You don't know the max overhead for the file before it is allocated. You could guess at a max of 3x size + at least three blocks. Since you can't control this, it seems like the worst case is when copies=3. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Directory size value
On 21 September, 2009 - Chris Banal sent me these 4,4K bytes: > It appears as though zfs reports the size of a directory to be one byte per > file. Traditional file systems such as ufs or ext3 report the actual size of > the data needed to store the directory. Or rather, "the size needed at some random point in time" .. because it seldom shrinks when you remove files.. At least not on UFS.. So basically it's an uninteresting and stale piece of information. The number of entries in the directory is much more useful. > This causes some trouble with the default behavior of some nfs clients > (linux) to decide to to use a readdirplus call when directory contents are > small vs a readdir call when the contents are large. We've found this > particularly troublesome with Maildir style mail folders. The speedup by not > using readdirplus is a factor of 100 in this particular situation. While we > have a work around it would seem that this non-standard behavior might cause > trouble for others and in other areas. Are there any suggestions for dealing > with this difference and/or why zfs does not represent its directory sizes > in a more traditional manner? This has been discussed before regarding software that totally broke (segfault) with ZFS's reporting.. The POSIX spec says nothing about what size on directories should be, iirc.. Other (Novell I think?) filesystems has used the same size reporting as well.. > >From the linux kernel source. > > ./fs/nfs/inode.c:#define NFS_LIMIT_READDIRPLUS (8*PAGE_SIZE) Basically, it's an assumption given informativen given in an unspecified format.. Since there is nothing else to go on, it's not a totally crappy assumption, but it doesn't work all the time either.. Hrm. > ie. ZFS > > zfshost:~/testdir> ls -1 | wc -l > 330 > zfshost:~/testdir> stat . > File: `.' > Size: 332 Blocks: 486IO Block: 32768 directory > Device: 29h/41d Inode: 540058 Links: 2 > Access: (0775/drwxrwxr-x) Uid: ( 2891/ banal) Gid: ( 101/film) > Access: 2008-11-05 18:40:16.0 -0800 > Modify: 2009-09-01 16:09:52.782674099 -0700 > Change: 2009-09-01 16:09:52.782674099 -0700 > > ie. ext3 > > ext3host:~/testdir> ls -1 | wc -l > 330 > ext3host:~/testdir> stat . > File: `.' > Size: 36864 Blocks: 72 IO Block: 4096 directory > Device: 807h/2055d Inode: 23887981Links: 2 > Access: (0775/drwxrwxr-x) Uid: ( 2891/ banal) Gid: ( 101/film) > Access: 2009-09-21 13:44:00.0 -0700 > Modify: 2009-09-21 13:44:31.0 -0700 > Change: 2009-09-21 13:44:31.0 -0700 > > Thanks, > Chris > ___ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss /Tomas -- Tomas Ögren, st...@acc.umu.se, http://www.acc.umu.se/~stric/ |- Student at Computing Science, University of Umeå `- Sysadmin at {cs,acc}.umu.se ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Directory size value
It appears as though zfs reports the size of a directory to be one byte per file. Traditional file systems such as ufs or ext3 report the actual size of the data needed to store the directory. This causes some trouble with the default behavior of some nfs clients (linux) to decide to to use a readdirplus call when directory contents are small vs a readdir call when the contents are large. We've found this particularly troublesome with Maildir style mail folders. The speedup by not using readdirplus is a factor of 100 in this particular situation. While we have a work around it would seem that this non-standard behavior might cause trouble for others and in other areas. Are there any suggestions for dealing with this difference and/or why zfs does not represent its directory sizes in a more traditional manner? >From the linux kernel source. ./fs/nfs/inode.c:#define NFS_LIMIT_READDIRPLUS (8*PAGE_SIZE) ie. ZFS zfshost:~/testdir> ls -1 | wc -l 330 zfshost:~/testdir> stat . File: `.' Size: 332 Blocks: 486IO Block: 32768 directory Device: 29h/41d Inode: 540058 Links: 2 Access: (0775/drwxrwxr-x) Uid: ( 2891/ banal) Gid: ( 101/film) Access: 2008-11-05 18:40:16.0 -0800 Modify: 2009-09-01 16:09:52.782674099 -0700 Change: 2009-09-01 16:09:52.782674099 -0700 ie. ext3 ext3host:~/testdir> ls -1 | wc -l 330 ext3host:~/testdir> stat . File: `.' Size: 36864 Blocks: 72 IO Block: 4096 directory Device: 807h/2055d Inode: 23887981Links: 2 Access: (0775/drwxrwxr-x) Uid: ( 2891/ banal) Gid: ( 101/film) Access: 2009-09-21 13:44:00.0 -0700 Modify: 2009-09-21 13:44:31.0 -0700 Change: 2009-09-21 13:44:31.0 -0700 Thanks, Chris ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] zfsdle eating all resources..
Hi, I've got some strange problems with my serer today. When I boot b123, it stops at "reading zfs config". I've tried several times to get past this point, but it seems to freeze there. Then I tried single user mode, from GRUB, and it seems to get me a little further. After a few minutes however, the OS becomes unresponsive, and soon its impossible to do anything. I get messages like this "File system full, swap space limit exceeded", and some messages about shell being unable to fork. If I'm really fast after booting single user, I can start "top", and it looks like I have tons of "zfsdle" processes, using all available resources.. The rpool is 2 disks in mirror. The other pool is 16 disks in 4x4 raidz. I've recently lost "disk 12", and was trying to replace it just before this happened. Best Regards, Vidar Nilsen ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS Recv slow with high CPU
Tristan Ball wrote: Hi Everyone, I have a couple of systems running opensolaris b118, one of which sends hourly snapshots to the other. This has been working well, however as of today, the receiving zfs process has started running extremely slowly, and is running at 100% CPU on one core, completely in kernel mode. A little bit of exploration with lockstat and dtrace seems to imply that the issue is around the "dbuf_free_range" function - or at least, that's what it looks like to my inexperienced eye! This is probably RFE 6812603 "zfs send can aggregate free records", which is currently being worked on. --matt ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] lots of zil_clean threads
Thinking more about this I'm confused about what you are seeing. The function dsl_pool_zil_clean() will serialise separate calls to zil_clean() within a pool. I don't expect you have >1037 pools on your laptop! So I don't know what's going on. What is the typical call stack for those zil_clean() threads? Neil. On 09/21/09 08:53, Neil Perrin wrote: Nils, A zil_clean() is started for each dataset after every txg. this includes snapshots (which is perhaps a bit inefficient). Still, zil_clean() is fairly lightweight if there's nothing to do (grab a non contended lock; find nothing on a list; drop the lock & exit). Neil. On 09/21/09 08:08, Nils Goroll wrote: Hi All, out of curiosity: Can anyone come up with a good idea about why my snv_111 laptop computer should run more than 1000 zil_clean threads? ff0009a9dc60 fbc2c0300 tq:zil_clean ff0009aa3c60 fbc2c0300 tq:zil_clean ff0009aa9c60 fbc2c0300 tq:zil_clean ff0009aafc60 fbc2c0300 tq:zil_clean ff0009ab5c60 fbc2c0300 tq:zil_clean ff0009abbc60 fbc2c0300 tq:zil_clean ff0009ac1c60 fbc2c0300 tq:zil_clean > ::threadlist!grep zil_clean| wc -l 1037 Thanks, Nils P.S.: Please don't spend too much time on this, for me, this question is really academic - but I'd be grateful for any good answers. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Incremental backup via zfs send / zfs receive
On 09/21/09 12:04 PM, David Pacheco wrote: This is 6839260 want zfs send with properties Could it be amended to ask for an option on incrementals /not/ to send properties? IMO non-recursive streams should be able to be consistent. I wonder what the thinking was to not send properties on a new file system, but to over- ride them on an increment. Seems odd :-) Thanks -- Frank ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] resizing zpools by growing LUN
Hej Richard. think I'll update all our servers to the same version of zfs... That will hopefully make sure that this doesn't happen again :-) Darren and Richard: Thank you very much for your help ! Sascha -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Real help
On Sep 21, 2009, at 4:34 AM, David Magda wrote: On Sep 21, 2009, at 06:52, Chris Ridd wrote: Does zpool destroy prompt "are you sure" in any way? Some admin tools do (beadm destroy for example) but there's not a lot of consistency. No it doesn't, which I always found strange. Personally I always thought you should be queried for a "zfs destroy", but add a "-f" option for things like scripts. Not sure if things can be changed now. This was one of the first, great debates on ZFS-discuss http://www.opensolaris.org/jive/thread.jspa?messageID=15363㰃 I believe all possible observations are made in that thread. -- richard ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Incremental backup via zfs send / zfs receive
Frank Middleton wrote: The problem with the regular stream is that most of the file system properties (such as mountpoint) are not copied as they are with a recursive stream. This may seem an advantage to some, (e.g., if the remote mountpoint is already in use, the mountpoint seems to default to legacy). However, did I miss anything in the documentation, or would it be worth submitting an RFE for an option to send/recv properties in a non-recursive stream? This is 6839260 want zfs send with properties -- Dave -- David Pacheco, Sun Microsystems Fishworks. http://blogs.sun.com/dap/ ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] resizing zpools by growing LUN
On Sep 21, 2009, at 8:59 AM, Sascha wrote: Hi Darren, sorry that it took so long before I could answer. The good thing: I found out what went wrong. What I did: After resizing a Disk on the Storage, solaris recognizes it immediately. Everytime you resize a disk, the EVA storage updates the discription which contains the size. So typing "echo |format" shows promptly the new discription. So that seemed to work fine. But It turned out that "format" doesn't automagically update the amount of cylinders/sectors. So i tried to "auto configure" the disk. When I did that, the first sector of partition 0 changed from sector # 34 to # 256. (which makes label 0 and 1 of zpool unaccessible). This was a change made long ago, but it finally caught up with you. You must have created the original EFI label with an older version of [Open]Solaris. If you then relabel with automatic settings, the starting sector will change to the 256 value, which is a much better starting point. -- richard The last sector changed to the new end of the disk (which makes label 3 and 4 of the zpool unaccessible). Partition 8 changed accordingly and correctly. If I then relabel the disk the zpool is destroyed. Label 0 and 1 of the zpool are gone and, obviously label 2 and 3 too. What I did to solve it: 1. Changed the size of a disk/LUN on the storage 2. Verified the position of the starting sector of partition 0 3. Did an auto configure 4. Changed starting sector of partition 0 to the formerly starting sector from step #2 5. labeled the disk I then was able to impoort the zpool Do you know why the "auto configure" changes the starting cylinder ? What I also found out is that it did not happen every time. Sometimes the first sector of partition 0 changes, sometimes not. Up to now I can't find any correlation between when and why. Sascha -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] resizing zpools by growing LUN
Hi Darren, sorry that it took so long before I could answer. The good thing: I found out what went wrong. What I did: After resizing a Disk on the Storage, solaris recognizes it immediately. Everytime you resize a disk, the EVA storage updates the discription which contains the size. So typing "echo |format" shows promptly the new discription. So that seemed to work fine. But It turned out that "format" doesn't automagically update the amount of cylinders/sectors. So i tried to "auto configure" the disk. When I did that, the first sector of partition 0 changed from sector # 34 to # 256. (which makes label 0 and 1 of zpool unaccessible). The last sector changed to the new end of the disk (which makes label 3 and 4 of the zpool unaccessible). Partition 8 changed accordingly and correctly. If I then relabel the disk the zpool is destroyed. Label 0 and 1 of the zpool are gone and, obviously label 2 and 3 too. What I did to solve it: 1. Changed the size of a disk/LUN on the storage 2. Verified the position of the starting sector of partition 0 3. Did an auto configure 4. Changed starting sector of partition 0 to the formerly starting sector from step #2 5. labeled the disk I then was able to impoort the zpool Do you know why the "auto configure" changes the starting cylinder ? What I also found out is that it did not happen every time. Sometimes the first sector of partition 0 changes, sometimes not. Up to now I can't find any correlation between when and why. Sascha -- This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] lots of zil_clean threads
Nils, A zil_clean() is started for each dataset after every txg. this includes snapshots (which is perhaps a bit inefficient). Still, zil_clean() is fairly lightweight if there's nothing to do (grab a non contended lock; find nothing on a list; drop the lock & exit). Neil. On 09/21/09 08:08, Nils Goroll wrote: Hi All, out of curiosity: Can anyone come up with a good idea about why my snv_111 laptop computer should run more than 1000 zil_clean threads? ff0009a9dc60 fbc2c0300 tq:zil_clean ff0009aa3c60 fbc2c0300 tq:zil_clean ff0009aa9c60 fbc2c0300 tq:zil_clean ff0009aafc60 fbc2c0300 tq:zil_clean ff0009ab5c60 fbc2c0300 tq:zil_clean ff0009abbc60 fbc2c0300 tq:zil_clean ff0009ac1c60 fbc2c0300 tq:zil_clean > ::threadlist!grep zil_clean| wc -l 1037 Thanks, Nils P.S.: Please don't spend too much time on this, for me, this question is really academic - but I'd be grateful for any good answers. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] ZFS file disk usage
On Sun, 20 Sep 2009 20:31:57 -0400 Richard Elling wrote: > If you are just building a cache, why not just make a file system and > put a reservation on it? Turn off auto snapshots and set other > features as per best practices for your workload? In other words, > treat it like we > treat dump space. > > I think that we are getting caught up in trying to answer the question > you ask rather than solving the problem you have... perhaps because > we don't understand the problem. Yes, possibly... some of these suggestions dont quite make a lot of sense to me. We can't just make a filesystem and put a reservation on it; we are just an application the administrator puts on a machine for it to access AFS. So I'm not sure when you are imagining we do that; when the client starts up? Or part of the installation procedure? Requiring a separate filesystem seems unnecessarily restrictive. And I still don't see how that helps. Making an fs with a reservation would definitely limit us to the specified space, but we still can't get an accurate picture of the current disk usage. I already mentioned why using statvfs is not usable with that commit delay. But solving the general problem for me isn't necessary. If I could just get a ballpark estimate of the max overhead for a file, I would be fine. I haven't payed attention to it before, so I don't even have an intuitive feel for what it is. -- Andrew Deason adea...@sinenomine.net ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] lots of zil_clean threads
Hi All, out of curiosity: Can anyone come up with a good idea about why my snv_111 laptop computer should run more than 1000 zil_clean threads? ff0009a9dc60 fbc2c0300 tq:zil_clean ff0009aa3c60 fbc2c0300 tq:zil_clean ff0009aa9c60 fbc2c0300 tq:zil_clean ff0009aafc60 fbc2c0300 tq:zil_clean ff0009ab5c60 fbc2c0300 tq:zil_clean ff0009abbc60 fbc2c0300 tq:zil_clean ff0009ac1c60 fbc2c0300 tq:zil_clean > ::threadlist!grep zil_clean| wc -l 1037 Thanks, Nils P.S.: Please don't spend too much time on this, for me, this question is really academic - but I'd be grateful for any good answers. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Real help
On Mon, Sep 21, 2009 at 13:34, David Magda wrote: > On Sep 21, 2009, at 06:52, Chris Ridd wrote: > >> Does zpool destroy prompt "are you sure" in any way? Some admin tools do >> (beadm destroy for example) but there's not a lot of consistency. > > No it doesn't, which I always found strange. > > Personally I always thought you should be queried for a "zfs destroy", but > add a "-f" option for things like scripts. Not sure if things can be changed > now. You can import a destroyed pool, you can find them wiht zpool import -D. But the problem in this case was not zpool destroy, it was zpool create. zpool create will overwrite whatever was on the partition. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re-import RAID-Z2 with faulted disk
> >The disk has since been replaced, so now: >k...@localhost:~$ pfexec zpool import > pool: chronicle >id: 11592382930413748377 > state: DEGRADED >status: One or more devices contains corrupted data. >action: The pool can be imported despite missing or damaged devices. The >fault tolerance of the pool may be compromised if imported. > see: http://www.sun.com/msg/ZFS-8000-4J >config: > >chronicle DEGRADED > raidz2DEGRADED >c9t2d0 FAULTED corrupted data >c9t1d0 ONLINE >c9t0d0 ONLINE >c9t4d0 ONLINE >c9t5d0 ONLINE >c9t3d0 ONLINE > >but "pfexec zpool import -f chronicle" still fails with the same message. > That sounds like a bug; if ZFS can recover then zpool import should also work. So what was wrong with the broken disk? Not just badly plugged in? Cas[er ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Real help
On Sep 21, 2009, at 06:52, Chris Ridd wrote: Does zpool destroy prompt "are you sure" in any way? Some admin tools do (beadm destroy for example) but there's not a lot of consistency. No it doesn't, which I always found strange. Personally I always thought you should be queried for a "zfs destroy", but add a "-f" option for things like scripts. Not sure if things can be changed now. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re-import RAID-Z2 with faulted disk
Hi all, I have a RAID-Z2 setup with 6x 500Gb SATA disks. I exported the array to use under a different system but during or after the export one of the disks failed: k...@localhost:~$ pfexec zpool import pool: chronicle id: 11592382930413748377 state: DEGRADED status: One or more devices are missing from the system. action: The pool can be imported despite missing or damaged devices. The fault tolerance of the pool may be compromised if imported. see: http://www.sun.com/msg/ZFS-8000-2Q config: chronicle DEGRADED raidz2DEGRADED c9t2d0 UNAVAIL cannot open c9t1d0 ONLINE c9t0d0 ONLINE c9t4d0 ONLINE c9t5d0 ONLINE c9t3d0 ONLINE I have no success trying to reimport the pool: k...@localhost:~$ pfexec zpool import -f chronicle cannot import 'chronicle': one or more devices is currently unavailable The disk has since been replaced, so now: k...@localhost:~$ pfexec zpool import pool: chronicle id: 11592382930413748377 state: DEGRADED status: One or more devices contains corrupted data. action: The pool can be imported despite missing or damaged devices. The fault tolerance of the pool may be compromised if imported. see: http://www.sun.com/msg/ZFS-8000-4J config: chronicle DEGRADED raidz2DEGRADED c9t2d0 FAULTED corrupted data c9t1d0 ONLINE c9t0d0 ONLINE c9t4d0 ONLINE c9t5d0 ONLINE c9t3d0 ONLINE but "pfexec zpool import -f chronicle" still fails with the same message. I've Google'd this several times for a fix but to no avail. Any assistance is appreciated. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Real help
On 20 Sep 2009, at 19:46, dick hoogendijk wrote: On Sun, 2009-09-20 at 11:41 -0700, vattini giacomo wrote: Hi there,i'm in a bad situation,under Ubuntu i was tring to import a solaris zpool that is in /dev/sda1,while the Ubuntu is in /dev/ sda5;not being able to mount the solaris pool i decide to destroy the pool created like that sudo zfs-fuse sudo zpool create hazz0 /dev/sda1 sudo zpool destroy hazz0 sudo reboot Now opensolaris is not booting everything is vanished Is there anyhow to restore everything? Any idea about the meaning of the verb DESTROY ? Does zpool destroy prompt "are you sure" in any way? Some admin tools do (beadm destroy for example) but there's not a lot of consistency. Cheers, Chris ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Migrate from iscsitgt to comstar?
Is it possible to migrate data from iscsitgt for comstar iscsi target? I guess comstar wants metadata at beginning of volume and this makes things difficult? Yours Markus Kovero ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss