Re: [zfs-discuss] replacing a drive in a raidz vdev
On 12/5/06, Bill Sommerfeld [EMAIL PROTECTED] wrote: On Mon, 2006-12-04 at 13:56 -0500, Krzys wrote: mypool2/[EMAIL PROTECTED] 34.4M - 151G - mypool2/[EMAIL PROTECTED] 141K - 189G - mypool2/d3 492G 254G 11.5G legacy I am so confused with all of this... Why its taking so long to replace that one bad disk? To workaround a bug where a pool traverse gets lost when the snapshot configuration of a pool changes, both scrubs and resilvers will start over again any time you create or delete a snapshot. Unfortunately, this workaround has problems of its own -- If your inter-snapshot interval is less than the time required to complete a scrub, the resilver will never complete. The open bug is: 6343667 scrub/resilver has to start over when a snapshot is taken if it's not going to be fixed any time soon, perhaps we need a better workaround: Anyone internal working on this? -- Regards, Jeremy ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] need Clarification on ZFS
dudekula mastan wrote: 5) Like fsck command on Linux, is there any command to check the consistency of the ZFS file system ? As others have mentioned, ZFS doesn't require off line consistency checking. You can run 'zpool scrub' on a live system and check the result with 'zpool status': zpool status pool: tank state: ONLINE scrub: scrub completed with 0 errors on Tue Dec 5 21:07:08 2006 config: NAMESTATE READ WRITE CKSUM tankONLINE 0 0 0 mirrorONLINE 0 0 0 c3d0ONLINE 0 0 0 c4d0ONLINE 0 0 0 errors: No known data errors Ian ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] need Clarification on ZFS
dudekula mastan [EMAIL PROTECTED] wrote: 1) On Linux to know the presence of ext2/ext3 file systems on a device we use tune2fs command. Similar to tune2fs command is there any command to know the presence of ZFS file system on a device ? 2) When a device is shared between two machines , What our project does is, - Create ext2 file system on device a) Mount the device on machine 1 b) Write data on the device c) unmount the device from machine 1 d)mount the device on machine 2 e) read the data on the device f) compare the current read data with previous write data and report the result If you like to compare directora trees, I rcommend to use star -diff -vv ftp://ftp.berlios.de/pub/star/alpha/ Jörg -- EMail:[EMAIL PROTECTED] (home) Jörg Schilling D-13353 Berlin [EMAIL PROTECTED](uni) [EMAIL PROTECTED] (work) Blog: http://schily.blogspot.com/ URL: http://cdrecord.berlios.de/old/private/ ftp://ftp.berlios.de/pub/schily ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] need Clarification on ZFS
Le 04/12/2006 à 23:34:39-0800, Jason A. Hoffman a écrit Hi Mastan, Like this , Can We share zfs file system between two machines. If so please explain it. It's always going from machine 1 to machine 2? zfs send [EMAIL PROTECTED] | ssh [EMAIL PROTECTED] | zfs recv filesystem-one-machine2 will stream a snapshot from the first machine to a filesystem/device/ snapshot on machine2 That's impressive. Whath the size of the file you send throught ssh ? Is that size is exactly same of the FS or the occupation of FS ? Can I send just the diff ? For example At t=0 I send a big file using your command at t=t+1I just send the diff not a big file Regards. -- Albert SHIH Universite de Paris 7 (Denis DIDEROT) U.F.R. de Mathematiques. 7 ième étage, plateau D, bureau 10 Heure local/Local time: Tue Dec 5 14:53:13 CET 2006 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: ZFS on multi-volume
Hi Albert, On Tue, 2006-12-05 at 14:16 +0100, Albert Shih wrote: It's possible to configure the server, the high level raid array, and the pool of my old array raid to do : 1/ When the server read/write he do from high level raid 2/ The server make a copie of all data from high level raid to the pool of my old array «when he have the time». But I want this automatics. I don't want this by using something like rsync. Using zfs send/recv, you can have one system running ZFS send copies to other systems running ZFS. It's analogous to using rsync, but should be a bit quicker. I don't know of an existing automated way to do this send/recv only when the sending zpool isn't busy, for some given definition of busy.. (you're the 2nd person I've heard from in recent days that's asked for this - Theo asked a similar question at http://blogs.sun.com/timf/entry/zfs_automatic_snapshot_service_logging#comments I wonder is it a useful RFE for the ZFS automatic snapshot service ? ) Here's what I'm thinking: if you know what times the system is likely to be idle, you can use a cron job to send/receive the data between systems -- would this be sufficient ? Remember that you can send/recv incremental snapshots as well, so every 10 minutes, you could take a snapshot of your data, and decide whether to send/recv that (which would reduce the amount of IO you need to do) If the system is busy, you just remember which incremental snapshot you last sent, and record that somewhere. As soon as the system is idle, take another snapshot, and do an incremental send of the difference between that and your recorded snapshot. This probably isn't elegant, but it would do the job I think. What I want to do is make a NFS server with the new high level raid array with primary data. But I want also using my old-low-level raid array to make backup (in case I'm lost my high-level raid array) and only backup. Sounds like you really want i/o throttling of send/recv operations as against normal pool operations - I don't know enough to suggest how this could be implemented (except via brutal pstop/prun hacks on the zfs send process whenever your pool exceeds some given IO threshold) cheers, tim -- Tim Foster, Sun Microsystems Inc, Solaris Engineering Ops http://blogs.sun.com/timf ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] need Clarification on ZFS
On Tue, 2006-12-05 at 14:56 +0100, Albert Shih wrote: That's impressive. Whath the size of the file you send throught ssh ? Is that size is exactly same of the FS or the occupation of FS ? Can I send just the diff ? For example At t=0 I send a big file using your command at t=t+1I just send the diff not a big file Yes. The zfs send contents are the amount of occupied space of the filesystem at a given point in time (a snapshot). You can use zfs send -i to send incremental snapshots, containing only the data that differs between [EMAIL PROTECTED] and [EMAIL PROTECTED] More info in the ZFS man pages, the (excellent) ZFS Administration guide, at http://docs.sun.com/app/docs/doc/819-5461/6n7ht6qt0?a=view (you want Chapter 6) and a blog post from Matt at: http://blogs.sun.com/ahrens/entry/is_it_magic Matt has mentioned some additional features to zfs send/recv coming soon, including ways to send all incremental snapshots, send nested filesystem, and ways to preserve filesystem properties while sending. cheers, tim -- Tim Foster, Sun Microsystems Inc, Solaris Engineering Ops http://blogs.sun.com/timf ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: ZFS on multi-volume
How about attaching the slow storage and kick off a scrub during the nights ? Then detach in the morning ? Downside: you are running an unreplicated pool during the day. Storage side errors won't be recoverable. -r Albert Shih writes: Le 04/12/2006 à 21:24:26-0800, Anton B. Rang a écrit It is possible to configure ZFS in the way you describe, but your performance will be limited by the older array. All mirror writes have to be stored on both arrays before they are considered complete, so writes will be as slow as the slowest disk or array involved. OK. It's possible to configure the server, the high level raid array, and the pool of my old array raid to do : 1/ When the server read/write he do from high level raid 2/ The server make a copie of all data from high level raid to the pool of my old array «when he have the time». But I want this automatics. I don't want this by using something like rsync. What I want to do is make a NFS server with the new high level raid array with primary data. But I want also using my old-low-level raid array to make backup (in case I'm lost my high-level raid array) and only backup. Do you think ZFS can help me ? Best regards -- Albert SHIH Universite de Paris 7 (Denis DIDEROT) U.F.R. de Mathematiques. 7 ième étage, plateau D, bureau 10 Tel : 01 44 27 86 88 FAX : 01 44 27 69 35 GSM(UFR) : 06 85 05 58 43 Heure local/Local time: Tue Dec 5 14:16:01 CET 2006 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: need Clarification on ZFS
is there any command to know the presence of ZFS file system on a device ? fstyp is the Solaris command to determine what type of file system may be present on a disk: # fstyp /dev/dsk/c0t1d0s6 zfs When a device is shared between two machines [ ... ] You can use the same mount/unmount mechanism with ZFS, except that you will both need to unmount the file systems and export the pool. It's critical that you export the pool from machine A before trying to access it from machine B. Can we create ZFS pools (or ZFS file system ) on VxVm volumes ? if so, how ? Yes. Just give the name of the VxVM volume to the 'zpool create' command, where the name of a disk would go. [Disclaimer: I haven't tested this.] Can we share ZFS pools ( ZFS file system ) between two machines ? If you mean, can you have a pool simultaneously mounted on two machines, the answer is no. You must unmount/export the pool from one machine before allowing another machine to access it. Like fsck command on Linux, is there any command to check the consistency of the ZFS file system ? No. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: ZFS related kernel panic
But it's still not the application's problem to handle the underlying device failure. But it is the application's problem to handle an error writing to the file system -- that's why the file system is allowed to return errors. ;-) Some applications might not check them, some applications might not have anything reasonable to do (though they can usually at least output a useful message to stderr), but other applications may be more robust. It's not particularly uncommon for an application to encounter an error writing to volume X and then choose to write to volume Y instead; or to report the error back to another component or the end user. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] weird thing with zfs
ok, two weeks ago I did notice one of my disk in zpool got problems. I was getting Corrupt label; wrong magic number messages, then when I looked in format it did not see that disk... (last disk) I had that setup running for few months now and all of the sudden last disk failed. So I ordered another disk, had it replaced like a week ago, I did issue replace command after disk replacement, it was resilvering disks since forever, then I got hints from this group that snaps could be causing it so yesterday I did disable snaps and this morning I di dnotice the same disk that I replaced is gone... Does it seem weird that this disk would fail? Its new disk... I have Solaris 10 U2, 4 internal drives and then 7 external drives which are in single enclousures connected via scsi chain to each other... So it seems like last disk is failing. Those nipacks from sun have self termination so there is no terminator at the end... Any ideas what should I do? Do I need to order another drive and replace that one too? Or will it happen again? What do you think could be the problem? Ah, when I look at that enclosure I do see green light on it so it seems like it did not fail... format Searching for disks... efi_alloc_and_init failed. done AVAILABLE DISK SELECTIONS: 0. c1t0d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809 /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 1. c1t1d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809 /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 2. c1t2d0 SEAGATE-ST337LC-D703-279.40GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 3. c1t3d0 SEAGATE-ST337LC-D703-279.40GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 4. c3t0d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 5. c3t1d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 6. c3t2d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 7. c3t3d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 8. c3t4d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 9. c3t5d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 10. c3t6d0 drive type unknown /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 zpool status -v pool: mypool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM mypool ONLINE 0 0 0 mirrorONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 errors: No known data errors pool: mypool2 state: DEGRADED scrub: resilver completed with 0 errors on Mon Dec 4 22:34:57 2006 config: NAME STATE READ WRITE CKSUM mypool2 DEGRADED 0 0 0 raidz DEGRADED 0 0 0 c3t0d0ONLINE 0 0 0 c3t1d0ONLINE 0 0 0 c3t2d0ONLINE 0 0 0 c3t3d0ONLINE 0 0 0 c3t4d0ONLINE 0 0 0 c3t5d0ONLINE 0 0 0 replacing UNAVAIL 0 775 0 insufficient replicas c3t6d0s0/o UNAVAIL 0 0 0 cannot open c3t6d0 UNAVAIL 0 940 0 cannot open errors: No known data errors ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( format Searching for disks... efi_alloc_and_init failed. done AVAILABLE DISK SELECTIONS: 0. c1t0d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809 /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 1. c1t1d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809 /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 2. c1t2d0 SEAGATE-ST337LC-D703-279.40GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 3. c1t3d0 SEAGATE-ST337LC-D703-279.40GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 4. c3t0d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 5. c3t1d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 6. c3t2d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 7. c3t3d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 8. c3t4d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 9. c3t5d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 10. c3t6d0 drive type unknown /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 Specify disk (enter its number): 10 Segmentation Fault (core dumped) :( Cant even get to format menu on that drive... Chris On Tue, 5 Dec 2006, Nicholas Senedzuk wrote: The only time that I have seen a format return drive type unknown is when the drive has failed. You may just have another bad drive and want to try replacing it again. If that does not work you, you may have another problem such as a bad backplane or a bad SCSI cable assuming the drive is an external drive. Hope that helps. On 12/5/06, Krzys [EMAIL PROTECTED] wrote: ok, two weeks ago I did notice one of my disk in zpool got problems. I was getting Corrupt label; wrong magic number messages, then when I looked in format it did not see that disk... (last disk) I had that setup running for few months now and all of the sudden last disk failed. So I ordered another disk, had it replaced like a week ago, I did issue replace command after disk replacement, it was resilvering disks since forever, then I got hints from this group that snaps could be causing it so yesterday I did disable snaps and this morning I di dnotice the same disk that I replaced is gone... Does it seem weird that this disk would fail? Its new disk... I have Solaris 10 U2, 4 internal drives and then 7 external drives which are in single enclousures connected via scsi chain to each other... So it seems like last disk is failing. Those nipacks from sun have self termination so there is no terminator at the end... Any ideas what should I do? Do I need to order another drive and replace that one too? Or will it happen again? What do you think could be the problem? Ah, when I look at that enclosure I do see green light on it so it seems like it did not fail... format Searching for disks... efi_alloc_and_init failed. done AVAILABLE DISK SELECTIONS: 0. c1t0d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809 /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 1. c1t1d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809 /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 2. c1t2d0 SEAGATE-ST337LC-D703-279.40GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 3. c1t3d0 SEAGATE-ST337LC-D703-279.40GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 4. c3t0d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 5. c3t1d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 6. c3t2d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 7. c3t3d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 8. c3t4d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 9. c3t5d0 SEAGATE-ST3146807LC-0007-136.73GB /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 10. c3t6d0 drive type unknown /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 zpool status -v pool: mypool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM mypool ONLINE
Re: [zfs-discuss] weird thing with zfs
Krzys wrote: Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( Run pstack /path/to/core and send the output. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
[12:00:40] [EMAIL PROTECTED]: /d/d3/nb1 pstack core core 'core' of 29506: format -e - lwp# 1 / thread# 1 000239b8 c_disk (51800, 52000, 4bde4, 525f4, 54e78, 0) + 4e0 00020fb4 main (2, 0, ffbff8e8, 0, 52000, 29000) + 46c 000141a8 _start (0, 0, 0, 0, 0, 0) + 108 - lwp# 2 / thread# 2 ff241818 _door_return (0, 0, 0, 0, fef92400, ff26cbc0) + 10 ff0c0c30 door_create_func (0, feefc000, 0, 0, ff0c0c10, 0) + 20 ff2400b0 _lwp_start (0, 0, 0, 0, 0, 0) - lwp# 3 / thread# 3 ff240154 __lwp_park (75e78, 75e88, 0, 0, 0, 0) + 14 ff23a1e4 cond_wait_queue (75e78, 75e88, 0, 0, 0, 0) + 28 ff23a764 cond_wait (75e78, 75e88, 1, 0, 0, ff26cbc0) + 10 ff142a60 subscriber_event_handler (551d8, fedfc000, 0, 0, ff142a2c, 0) + 34 ff2400b0 _lwp_start (0, 0, 0, 0, 0, 0) On Tue, 5 Dec 2006, Torrey McMahon wrote: Krzys wrote: Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( Run pstack /path/to/core and send the output. !DSPAM:122,45759fd826586021468! ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
On Tue, 5 Dec 2006, Krzys wrote: Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( ... snip Try zeroing out the disk label with something like: dd if=/dev/zero of=/dev/rdsk/c?t?d?p0 bs=1024k count=1024 Regards, Al Hopper Logical Approach Inc, Plano, TX. [EMAIL PROTECTED] Voice: 972.379.2133 Fax: 972.379.2134 Timezone: US CDT OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005 OpenSolaris Governing Board (OGB) Member - Feb 2006 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
Al Hopper [EMAIL PROTECTED] wrote: On Tue, 5 Dec 2006, Krzys wrote: Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( ... snip Try zeroing out the disk label with something like: dd if=/dev/zero of=/dev/rdsk/c?t?d?p0 bs=1024k count=1024 Do you expect a 1 GB disk label? Jörg -- EMail:[EMAIL PROTECTED] (home) Jörg Schilling D-13353 Berlin [EMAIL PROTECTED](uni) [EMAIL PROTECTED] (work) Blog: http://schily.blogspot.com/ URL: http://cdrecord.berlios.de/old/private/ ftp://ftp.berlios.de/pub/schily ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
On Tue, 5 Dec 2006, Joerg Schilling wrote: Al Hopper [EMAIL PROTECTED] wrote: On Tue, 5 Dec 2006, Krzys wrote: Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( ... snip Try zeroing out the disk label with something like: dd if=/dev/zero of=/dev/rdsk/c?t?d?p0 bs=1024k count=1024 Do you expect a 1 GB disk label? No. :) Al Hopper Logical Approach Inc, Plano, TX. [EMAIL PROTECTED] Voice: 972.379.2133 Fax: 972.379.2134 Timezone: US CDT OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005 OpenSolaris Governing Board (OGB) Member - Feb 2006 ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
Does not work :( dd if=/dev/zero of=/dev/rdsk/c3t6d0s0 bs=1024k count=1024 dd: opening `/dev/rdsk/c3t6d0s0': I/O error That is so strange... it seems like I lost another disk... I will try to reboot and see what I get, but I guess I need to order another disk then and give it a try... Chris On Tue, 5 Dec 2006, Al Hopper wrote: On Tue, 5 Dec 2006, Krzys wrote: Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( ... snip Try zeroing out the disk label with something like: dd if=/dev/zero of=/dev/rdsk/c?t?d?p0 bs=1024k count=1024 Regards, Al Hopper Logical Approach Inc, Plano, TX. [EMAIL PROTECTED] Voice: 972.379.2133 Fax: 972.379.2134 Timezone: US CDT OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005 OpenSolaris Governing Board (OGB) Member - Feb 2006 !DSPAM:122,4575a7731650371292! ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Managed to corrupt my pool
So the questions are: - is this fixable? I don't see an inum I could run find on to remove, and I can't even do a zfs volinit anyway: nextest-01# zfs volinit cannot iterate filesystems: I/O error - would not enabling zil_disable have prevented this? - Should I have been doing a 3-way mirror? - Is there a more optimum configuration to help prevent this kind of corruption? Anyone have any thoughts on this? I'd really like to be able to build a nice ZFS box for file service but if a hardware failure can corrupt a disk pool I'll have to try to find another solution, I'm afraid. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
Ok, so here is an update I did restart my sysyte, I power it off and power it on. Here is screen capture of my boot. I certainly do have some hard drive issues and will need to take a look at them... But I got my disk back visible to the system and zfs is doing resilvering again Rebooting with command: boot Boot device: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0:a File and args: SunOS Release 5.10 Version Generic_118833-24 64-bit Copyright 1983-2006 Sun Microsystems, Inc. All rights reserved. Use is subject to license terms. Hardware watchdog enabled Hostname: chrysek WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 6 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5): Error for Command: read(10)Error Level: Retryable Requested Block: 286732066 Error Block: 286732066 Vendor: SEAGATESerial Number: 3HY14PVS Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 3 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd23): Error for Command: read(10)Error Level: Retryable Requested Block: 283623842 Error Block: 283623842 Vendor: SEAGATESerial Number: 3HY8HS7L Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 5 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd25): Error for Command: read(10)Error Level: Retryable Requested Block: 283623458 Error Block: 283623458 Vendor: SEAGATESerial Number: 3HY0LF18 Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 /kernel/drv/sparcv9/zpool symbol avl_add multiply defined /kernel/drv/sparcv9/zpool symbol assfail3 multiply defined WARNING: kstat_create('unix', 0, 'dmu_buf_impl_t'): namespace collision mypool2/d3 uncorrectable error checking ufs filesystems /dev/rdsk/c1t0d0s7: is logging. chrysek console login: VERITAS SCSA Generic Revision: 3.5c Dec 5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing Dec 5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing Dec 5 13:01:46 chrysek VERITAS: No proxy found. Dec 5 13:01:52 chrysek vmd[546]: ready for connections Dec 5 13:01:53 chrysek VERITAS: No proxy found. Dec 5 13:01:54 chrysek VERITAS: No proxy found. Dec 5 13:02:00 chrysek VERITAS: No proxy found. Dec 5 13:02:01 chrysek VERITAS: No proxy found. Dec 5 13:02:03 chrysek VERITAS: No proxy found. starting NetWorker daemons: nsrexecd lgtolmd Dec 5 13:02:20 chrysek CNS Transport[841]: cctransport started Dec 5 13:02:48 chrysek webmin[1353]: Webmin starting Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Dec 5 13:19:07 chrysek Target 6 disabled wide SCSI mode Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Dec 5 13:19:07 chrysek Target 6 reverting to async. mode Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5): Dec 5 13:19:07 chrysek Error for Command: write(10) Error Level: Retryable Dec 5 13:19:07 chrysek scsi: Requested Block: 137163259 Error Block: 137163259 Dec 5 13:19:07 chrysek scsi: Vendor: SEAGATE Serial Number: 3HY14PVS Dec 5 13:19:07 chrysek scsi: Sense Key: Aborted Command Dec 5 13:19:07 chrysek scsi: ASC: 0x47 (scsi parity error), ASCQ: 0x0, FRU: 0x3 but now when I do zpool status -v pool: mypool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM mypool ONLINE 0 0 0 mirrorONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 errors: No known data errors pool: mypool2 state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scrub: resilver in progress, 4.40% done, 11h40m to go
Re: [zfs-discuss] replacing a drive in a raidz vdev
Jeremy Teo wrote: On 12/5/06, Bill Sommerfeld [EMAIL PROTECTED] wrote: On Mon, 2006-12-04 at 13:56 -0500, Krzys wrote: mypool2/[EMAIL PROTECTED] 34.4M - 151G - mypool2/[EMAIL PROTECTED] 141K - 189G - mypool2/d3 492G 254G 11.5G legacy I am so confused with all of this... Why its taking so long to replace that one bad disk? To workaround a bug where a pool traverse gets lost when the snapshot configuration of a pool changes, both scrubs and resilvers will start over again any time you create or delete a snapshot. Unfortunately, this workaround has problems of its own -- If your inter-snapshot interval is less than the time required to complete a scrub, the resilver will never complete. The open bug is: 6343667 scrub/resilver has to start over when a snapshot is taken if it's not going to be fixed any time soon, perhaps we need a better workaround: Anyone internal working on this? Yes. But its going to be a few months. -Mark ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Managed to corrupt my pool
Anyone have any thoughts on this? I'd really like to be able to build a nice ZFS box for file service but if a hardware failure can corrupt a disk pool I'll have to try to find another solution, I'm afraid. Sorry, I worded this poorly -- if the loss of a disk in a mirror can corrupt the pool it's going to give me pause in implementing a ZFS solution. Jim This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: weird thing with zfs
What os is this? What is the hardware? can you try running format with efi_debug set. You have to run format using a debugger and patch the variable. Here is how using mdb (set a break point in main so that the dynamic linker has done it's stuff, then update the value of efi_debug to be 1, then continue): # mdb /usr/sbin/format main:b :r mdb: stop at main mdb: target stopped at: main: pushl %ebp efi_debug/x libefi.so.1`efi_debug: libefi.so.1`efi_debug: 0 efi_debug/w 1 libefi.so.1`efi_debug: 0 = 0x1 :c Searching for disks...done --chris This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: ZFS related kernel panic
So ZFS should be more resilient against write errors, and the SCSI disk or FC drivers should be more resilient against LIPs (the most likely cause of your problem) or other transient errors. (Alternatively, the ifp driver should be updated to support the maximum number of targets on a loop, which might also solve your second problem.) NB. LIPs are a normal part of everyday life for fibre channel, they are not an error. Right. I don't think it's the LIP's that's the problem but rather (a guess, not verified) the fact that the HBA loses light on it's fiber interface when the switch reboots... I think I also saw the same ZFS-induced panic when I (stupid, I know, but...) moved a fiber cable from one GBIC in the switch to another on the run. I also saw this with the ifp driver btw. And as someone wrote - the ifp driver will never be updated since it's for EOL'd hardware :-) This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: ZFS related kernel panic
Hmm... I just noticed this qla2100.conf option: # During link down conditions enable/disable the reporting of # errors. #0 = disabled, 1 = enable hba0-link-down-error=1; hba1-link-down-error=1; I _wonder_ what might possibly happen if I change that 1 to a 0 (zero)... :-) This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
This looks more like a cabling or connector problem. When that happens you should see parity errors and transfer rate negotiations. -- richard Krzys wrote: Ok, so here is an update I did restart my sysyte, I power it off and power it on. Here is screen capture of my boot. I certainly do have some hard drive issues and will need to take a look at them... But I got my disk back visible to the system and zfs is doing resilvering again Rebooting with command: boot Boot device: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0:a File and args: SunOS Release 5.10 Version Generic_118833-24 64-bit Copyright 1983-2006 Sun Microsystems, Inc. All rights reserved. Use is subject to license terms. Hardware watchdog enabled Hostname: chrysek WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 6 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5): Error for Command: read(10)Error Level: Retryable Requested Block: 286732066 Error Block: 286732066 Vendor: SEAGATESerial Number: 3HY14PVS Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 3 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd23): Error for Command: read(10)Error Level: Retryable Requested Block: 283623842 Error Block: 283623842 Vendor: SEAGATESerial Number: 3HY8HS7L Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 5 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd25): Error for Command: read(10)Error Level: Retryable Requested Block: 283623458 Error Block: 283623458 Vendor: SEAGATESerial Number: 3HY0LF18 Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 /kernel/drv/sparcv9/zpool symbol avl_add multiply defined /kernel/drv/sparcv9/zpool symbol assfail3 multiply defined WARNING: kstat_create('unix', 0, 'dmu_buf_impl_t'): namespace collision mypool2/d3 uncorrectable error checking ufs filesystems /dev/rdsk/c1t0d0s7: is logging. chrysek console login: VERITAS SCSA Generic Revision: 3.5c Dec 5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing Dec 5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing Dec 5 13:01:46 chrysek VERITAS: No proxy found. Dec 5 13:01:52 chrysek vmd[546]: ready for connections Dec 5 13:01:53 chrysek VERITAS: No proxy found. Dec 5 13:01:54 chrysek VERITAS: No proxy found. Dec 5 13:02:00 chrysek VERITAS: No proxy found. Dec 5 13:02:01 chrysek VERITAS: No proxy found. Dec 5 13:02:03 chrysek VERITAS: No proxy found. starting NetWorker daemons: nsrexecd lgtolmd Dec 5 13:02:20 chrysek CNS Transport[841]: cctransport started Dec 5 13:02:48 chrysek webmin[1353]: Webmin starting Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Dec 5 13:19:07 chrysek Target 6 disabled wide SCSI mode Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Dec 5 13:19:07 chrysek Target 6 reverting to async. mode Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5): Dec 5 13:19:07 chrysek Error for Command: write(10) Error Level: Retryable Dec 5 13:19:07 chrysek scsi: Requested Block: 137163259 Error Block: 137163259 Dec 5 13:19:07 chrysek scsi: Vendor: SEAGATE Serial Number: 3HY14PVS Dec 5 13:19:07 chrysek scsi: Sense Key: Aborted Command Dec 5 13:19:07 chrysek scsi: ASC: 0x47 (scsi parity error), ASCQ: 0x0, FRU: 0x3 but now when I do zpool status -v pool: mypool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM mypool ONLINE 0 0 0 mirrorONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 errors: No known data errors pool: mypool2 state: DEGRADED status: One or more devices is currently being resilvered. The pool will
Re: [zfs-discuss] weird thing with zfs
BTW, there is a way to check what the SCSI negotiations resolved to. I wrote about it once in a BluePrint http://www.sun.com/blueprints/0500/sysperfnc.pdf See page 11 -- richard Richard Elling wrote: This looks more like a cabling or connector problem. When that happens you should see parity errors and transfer rate negotiations. -- richard Krzys wrote: Ok, so here is an update I did restart my sysyte, I power it off and power it on. Here is screen capture of my boot. I certainly do have some hard drive issues and will need to take a look at them... But I got my disk back visible to the system and zfs is doing resilvering again Rebooting with command: boot Boot device: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0:a File and args: SunOS Release 5.10 Version Generic_118833-24 64-bit Copyright 1983-2006 Sun Microsystems, Inc. All rights reserved. Use is subject to license terms. Hardware watchdog enabled Hostname: chrysek WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 6 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5): Error for Command: read(10)Error Level: Retryable Requested Block: 286732066 Error Block: 286732066 Vendor: SEAGATESerial Number: 3HY14PVS Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 3 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd23): Error for Command: read(10)Error Level: Retryable Requested Block: 283623842 Error Block: 283623842 Vendor: SEAGATESerial Number: 3HY8HS7L Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): SCSI bus DATA IN phase parity error WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Target 5 reducing sync. transfer rate WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd25): Error for Command: read(10)Error Level: Retryable Requested Block: 283623458 Error Block: 283623458 Vendor: SEAGATESerial Number: 3HY0LF18 Sense Key: Aborted Command ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 0x2 /kernel/drv/sparcv9/zpool symbol avl_add multiply defined /kernel/drv/sparcv9/zpool symbol assfail3 multiply defined WARNING: kstat_create('unix', 0, 'dmu_buf_impl_t'): namespace collision mypool2/d3 uncorrectable error checking ufs filesystems /dev/rdsk/c1t0d0s7: is logging. chrysek console login: VERITAS SCSA Generic Revision: 3.5c Dec 5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing Dec 5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing Dec 5 13:01:46 chrysek VERITAS: No proxy found. Dec 5 13:01:52 chrysek vmd[546]: ready for connections Dec 5 13:01:53 chrysek VERITAS: No proxy found. Dec 5 13:01:54 chrysek VERITAS: No proxy found. Dec 5 13:02:00 chrysek VERITAS: No proxy found. Dec 5 13:02:01 chrysek VERITAS: No proxy found. Dec 5 13:02:03 chrysek VERITAS: No proxy found. starting NetWorker daemons: nsrexecd lgtolmd Dec 5 13:02:20 chrysek CNS Transport[841]: cctransport started Dec 5 13:02:48 chrysek webmin[1353]: Webmin starting Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Dec 5 13:19:07 chrysek Target 6 disabled wide SCSI mode Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2): Dec 5 13:19:07 chrysek Target 6 reverting to async. mode Dec 5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5): Dec 5 13:19:07 chrysek Error for Command: write(10) Error Level: Retryable Dec 5 13:19:07 chrysek scsi: Requested Block: 137163259 Error Block: 137163259 Dec 5 13:19:07 chrysek scsi: Vendor: SEAGATE Serial Number: 3HY14PVS Dec 5 13:19:07 chrysek scsi: Sense Key: Aborted Command Dec 5 13:19:07 chrysek scsi: ASC: 0x47 (scsi parity error), ASCQ: 0x0, FRU: 0x3 but now when I do zpool status -v pool: mypool state: ONLINE scrub: none requested config: NAMESTATE READ WRITE CKSUM mypool ONLINE 0 0 0 mirrorONLINE 0 0 0
Re: [zfs-discuss] Re: Re: ZFS related kernel panic
On 12/5/06, Peter Eriksson [EMAIL PROTECTED] wrote: Hmm... I just noticed this qla2100.conf option: # During link down conditions enable/disable the reporting of # errors. #0 = disabled, 1 = enable hba0-link-down-error=1; hba1-link-down-error=1; This is the driver the we are using in this configuration. Excellent insight. This will be added to the testing. Although, we are moving away from the Qlogic cards, back to the Sun branded Qlogic cards to use MPxIO. Which works flawlessly with UFS and raw drives. I wonder if using MPxIO, and dual connected / path HBAs would also reduce these errors. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] weird thing with zfs
Hm. If the disk has no label, why would it have an s0? Or, did you mean p0? Nathan. On Wed, 2006-12-06 at 04:45, Krzys wrote: Does not work :( dd if=/dev/zero of=/dev/rdsk/c3t6d0s0 bs=1024k count=1024 dd: opening `/dev/rdsk/c3t6d0s0': I/O error That is so strange... it seems like I lost another disk... I will try to reboot and see what I get, but I guess I need to order another disk then and give it a try... Chris On Tue, 5 Dec 2006, Al Hopper wrote: On Tue, 5 Dec 2006, Krzys wrote: Thanks, ah another wird thing is that when I run format on that frive I get a coredump :( ... snip Try zeroing out the disk label with something like: dd if=/dev/zero of=/dev/rdsk/c?t?d?p0 bs=1024k count=1024 Regards, Al Hopper Logical Approach Inc, Plano, TX. [EMAIL PROTECTED] Voice: 972.379.2133 Fax: 972.379.2134 Timezone: US CDT OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005 OpenSolaris Governing Board (OGB) Member - Feb 2006 !DSPAM:122,4575a7731650371292! ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] raidz DEGRADED state
Quoth Thomas Garner on Thu, Nov 30, 2006 at 06:41:15PM -0500: I currently have a 400GB disk that is full of data on a linux system. If I buy 2 more disks and put them into a raid-z'ed zfs under solaris, is there a generally accepted way to build an degraded array with the 2 disks, copy the data to the new filesystem, and then move the original disk to complete the array? No, because we currently can't add disks to a raidz array. You could create a mirror instead and then add in the other disk to make a three-way mirror, though. Even doing that would be dicey if you only have a single machine, though, since Solaris can't natively read the popular Linux filesystems. I believe there is freeware to do it, but nothing supported. David ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] raidz DEGRADED state
So there is no current way to specify the creation of a 3 disk raid-z array with a known missing disk? On 12/5/06, David Bustos [EMAIL PROTECTED] wrote: Quoth Thomas Garner on Thu, Nov 30, 2006 at 06:41:15PM -0500: I currently have a 400GB disk that is full of data on a linux system. If I buy 2 more disks and put them into a raid-z'ed zfs under solaris, is there a generally accepted way to build an degraded array with the 2 disks, copy the data to the new filesystem, and then move the original disk to complete the array? No, because we currently can't add disks to a raidz array. You could create a mirror instead and then add in the other disk to make a three-way mirror, though. Even doing that would be dicey if you only have a single machine, though, since Solaris can't natively read the popular Linux filesystems. I believe there is freeware to do it, but nothing supported. David ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Re: [zfs-discuss] Re: Managed to corrupt my pool
Jim, I'm not at all sure what happened to your pool. However, I can answer some of your questions. Jim Hranicky wrote On 12/05/06 11:32,: So the questions are: - is this fixable? I don't see an inum I could run find on to remove, I think the pool is busted. Even the message printed in your previous email is bad: DATASET OBJECT RANGE 15 0 lvl=4294967295 blkid=0 as level is way out of range. and I can't even do a zfs volinit anyway: nextest-01# zfs volinit cannot iterate filesystems: I/O error I'm not sure why you're using zfs volinit which I believe creates the zvol links, but this further shows problems. - would not enabling zil_disable have prevented this? No the intent log is not needed for pool integrity. It ensures the synchronous semantics of O_DSYNC/fsync are obeyed. ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Shared ZFS pools
You specify the mirroring configuration. The top-level vdevs are implicitly striped. So if you, for instance, request something like zpool create mirror AA BA mirror AB BB then you will have a pool consisting of a stripe of two mirrors. Each mirror will have one copy of its data at each location. Hence if site B goes down, you'll lose BA and BB, and each mirror will transition to a degraded state. So the answer to your first question is yes, with this configuration you can be sure that no data will reside solely on LUNs at site B. On your second question, I don't believe there is work currently being done on shared pools, but I'll let the ZFS team clarify the road map. This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: Re: Managed to corrupt my pool
I think the pool is busted. Even the message printed in your previous email is bad: DATASET OBJECT RANGE 15 0 lvl=4294967295 blkid=0 as level is way out of range. I think this could be from dmu_objset_open_impl(). It sets object to 0 and level to -1 (= 4294967295). [Hmmm, this also seems to indicate a truncation from 64 to 32 bits somewhere.] Would zdb would show any more detail? (Actually, it looks like the ZIL also sets object to 0 and level to -1 when accessing its blocks, but since the ZIL was disabled, I'd guess this isn't the issue here.) This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
[zfs-discuss] Re: raidz DEGRADED state
Creating an array configuration with one element being a sparse file, then removing that file, comes to mind, but I wouldn't want to be the first to attempt it. ;-) This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss