Re: [zfs-discuss] replacing a drive in a raidz vdev

2006-12-05 Thread Jeremy Teo

On 12/5/06, Bill Sommerfeld [EMAIL PROTECTED] wrote:

On Mon, 2006-12-04 at 13:56 -0500, Krzys wrote:
 mypool2/[EMAIL PROTECTED]  34.4M  -   151G  -
 mypool2/[EMAIL PROTECTED]  141K  -   189G  -
 mypool2/d3 492G   254G  11.5G  legacy

 I am so confused with all of this... Why its taking so long to replace that 
one
 bad disk?

To workaround a bug where a pool traverse gets lost when the snapshot
configuration of a pool changes, both scrubs and resilvers will start
over again any time you create or delete a snapshot.

Unfortunately, this workaround has problems of its own -- If your
inter-snapshot interval is less than the time required to complete a
scrub, the resilver will never complete.

The open bug is:

6343667 scrub/resilver has to start over when a snapshot is taken

if it's not going to be fixed any time soon, perhaps we need a better
workaround:


Anyone internal working on this?
--
Regards,
Jeremy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] need Clarification on ZFS

2006-12-05 Thread Ian Collins
dudekula mastan wrote:

  
 5)  Like fsck command on Linux, is there any command  to check the
 consistency of the ZFS file system ?
  

As others have mentioned, ZFS doesn't require off line consistency
checking.  You can run 'zpool scrub' on a live system and check the
result with 'zpool status':

zpool status
  pool: tank
 state: ONLINE
 scrub: scrub completed with 0 errors on Tue Dec  5 21:07:08 2006
config:

NAMESTATE READ WRITE CKSUM
tankONLINE   0 0 0
  mirrorONLINE   0 0 0
c3d0ONLINE   0 0 0
c4d0ONLINE   0 0 0

errors: No known data errors

Ian

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] need Clarification on ZFS

2006-12-05 Thread Joerg Schilling
dudekula mastan [EMAIL PROTECTED] wrote:

   1) On Linux to know the presence of ext2/ext3 file systems on a device we 
 use tune2fs command. Similar to tune2fs command is there any command to know 
 the presence of ZFS file system on a device ?

   2) When a device is shared between two machines , What our project does is,

   - Create ext2 file system on device 
   a) Mount the device on machine 1
b) Write data on the device 
   c) unmount the device from machine 1
   d)mount the device on machine 2
   e) read the data on the device
   f) compare the current read data with previous write data  and report the 
 result

If you like to compare directora trees, I rcommend to use star -diff -vv
ftp://ftp.berlios.de/pub/star/alpha/

Jörg

-- 
 EMail:[EMAIL PROTECTED] (home) Jörg Schilling D-13353 Berlin
   [EMAIL PROTECTED](uni)  
   [EMAIL PROTECTED] (work) Blog: http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/old/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] need Clarification on ZFS

2006-12-05 Thread Albert Shih
 Le 04/12/2006 à 23:34:39-0800, Jason A. Hoffman a écrit
 Hi Mastan,
 
 Like this , Can We share zfs file system between two machines. If  
 so please explain it.
 
 It's always going from machine 1 to machine 2?
 
 zfs send [EMAIL PROTECTED] | ssh [EMAIL PROTECTED] | zfs  
 recv filesystem-one-machine2
 
 will stream a snapshot from the first machine to a filesystem/device/ 
 snapshot on machine2

That's impressive. Whath the size of the file you send throught ssh ? Is
that size is exactly same of the FS or the occupation of FS ? Can I send
just the diff ? For example

At t=0  I send a big file using your command
at t=t+1I just send the diff not a big file 

Regards.

 
--
Albert SHIH
Universite de Paris 7 (Denis DIDEROT)
U.F.R. de Mathematiques.
7 ième étage, plateau D, bureau 10
Heure local/Local time:
Tue Dec 5 14:53:13 CET 2006
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: ZFS on multi-volume

2006-12-05 Thread Tim Foster
Hi Albert,

On Tue, 2006-12-05 at 14:16 +0100, Albert Shih wrote:
 It's possible to configure the server, the high level raid array, and the
 pool of my old array raid to do :
 
   1/ When the server read/write he do from high level raid
   2/ The server make a copie of all data from high level raid to the
   pool of my old array «when he have the time». But I want this
   automatics. I don't want this by using something like rsync.

Using zfs send/recv, you can have one system running ZFS send copies to
other systems running ZFS. It's analogous to using rsync, but should be
a bit quicker.

I don't know of an existing automated way to do this send/recv only when
the sending zpool isn't busy, for some given definition of busy..

(you're the 2nd person I've heard from in recent days that's asked for
this - Theo asked a similar question at
http://blogs.sun.com/timf/entry/zfs_automatic_snapshot_service_logging#comments 

I wonder is it a useful RFE for the ZFS automatic snapshot service ?
)


Here's what I'm thinking: if you know what times the system is likely to
be idle, you can use a cron job to send/receive the data between systems
-- would this be sufficient ?

Remember that you can send/recv incremental snapshots as well, so every
10 minutes, you could take a snapshot of your data, and decide whether
to send/recv that (which would reduce the amount of IO you need to do)

If the system is busy, you just remember which incremental snapshot
you last sent, and record that somewhere. As soon as the system is idle,
take another snapshot, and do an incremental send of the difference
between that and your recorded snapshot.

This probably isn't elegant, but it would do the job I think.

 What I want to do is make a NFS server with the new high level raid array
 with primary data. But I want also using my old-low-level raid array to
 make backup (in case I'm lost my high-level raid array) and only backup.

Sounds like you really want i/o throttling of send/recv operations as
against normal pool operations - I don't know enough to suggest how
this could be implemented (except via brutal pstop/prun hacks on the
zfs send process whenever your pool exceeds some given IO threshold)

cheers,
tim


-- 
Tim Foster, Sun Microsystems Inc, Solaris Engineering Ops
http://blogs.sun.com/timf

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] need Clarification on ZFS

2006-12-05 Thread Tim Foster
On Tue, 2006-12-05 at 14:56 +0100, Albert Shih wrote:
 That's impressive. Whath the size of the file you send throught ssh ? Is
 that size is exactly same of the FS or the occupation of FS ? Can I send
 just the diff ? For example
 
   At t=0  I send a big file using your command
   at t=t+1I just send the diff not a big file 

Yes. The zfs send contents are the amount of occupied space of the
filesystem at a given point in time (a snapshot). You can use zfs send
-i to send incremental snapshots, containing only the data that differs
between [EMAIL PROTECTED] and [EMAIL PROTECTED]

More info in the ZFS man pages, the (excellent) ZFS Administration
guide, at
http://docs.sun.com/app/docs/doc/819-5461/6n7ht6qt0?a=view (you want
Chapter 6)

and a blog post from Matt at:
http://blogs.sun.com/ahrens/entry/is_it_magic

Matt has mentioned some additional features to zfs send/recv coming
soon, including ways to send all incremental snapshots, send nested
filesystem, and ways to preserve filesystem properties while sending.

cheers,
tim
-- 
Tim Foster, Sun Microsystems Inc, Solaris Engineering Ops
http://blogs.sun.com/timf

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: ZFS on multi-volume

2006-12-05 Thread Roch - PAE


How about attaching the slow storage and kick off a 
scrub during the nights ? Then detach in the morning ?

Downside: you are running an unreplicated pool during the
day. Storage side errors won't be recoverable.

-r

Albert Shih writes:
   Le 04/12/2006 à 21:24:26-0800, Anton B. Rang a écrit
   It is possible to configure ZFS in the way you describe, but your 
   performance will be limited by the older array.
   
   All mirror writes have to be stored on both arrays before they are 
   considered complete, so writes will be as slow as the slowest disk or 
   array involved.
   
  
  OK. 
  
  It's possible to configure the server, the high level raid array, and the
  pool of my old array raid to do :
  
   1/ When the server read/write he do from high level raid
   2/ The server make a copie of all data from high level raid to the
   pool of my old array «when he have the time». But I want this
   automatics. I don't want this by using something like rsync.
  
  What I want to do is make a NFS server with the new high level raid array
  with primary data. But I want also using my old-low-level raid array to
  make backup (in case I'm lost my high-level raid array) and only backup.
  
  Do you think ZFS can help me ?
  
  Best regards
  
  --
  Albert SHIH
  Universite de Paris 7 (Denis DIDEROT)
  U.F.R. de Mathematiques.
  7 ième étage, plateau D, bureau 10
  Tel  : 01 44 27 86 88
  FAX   : 01 44 27 69 35
  GSM(UFR) : 06 85 05 58 43
  Heure local/Local time:
  Tue Dec 5 14:16:01 CET 2006
  ___
  zfs-discuss mailing list
  zfs-discuss@opensolaris.org
  http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: need Clarification on ZFS

2006-12-05 Thread Anton B. Rang
 is there any command to know the presence of ZFS file system on a device ?

fstyp is the Solaris command to determine what type of file system may be 
present on a disk:

  # fstyp /dev/dsk/c0t1d0s6
  zfs

 When a device is shared between two machines [ ... ]

You can use the same mount/unmount mechanism with ZFS, except that you will 
both need to unmount the file systems and export the pool.  It's critical that 
you export the pool from machine A before trying to access it from machine B.

 Can we create ZFS pools (or ZFS file system ) on VxVm volumes ? if so, how ?

Yes.  Just give the name of the VxVM volume to the 'zpool create' command, 
where the name of a disk would go.  [Disclaimer: I haven't tested this.]

 Can we share ZFS pools ( ZFS file system ) between two machines ?

If you mean, can you have a pool simultaneously mounted on two machines, the 
answer is no.  You must unmount/export the pool from one machine before 
allowing another machine to access it.

 Like fsck command on Linux, is there any command  to check the consistency of
 the ZFS file system ?

No.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: ZFS related kernel panic

2006-12-05 Thread Anton B. Rang
 But it's still not the application's problem to handle the underlying
 device failure.

But it is the application's problem to handle an error writing to the file 
system -- that's why the file system is allowed to return errors.  ;-)

Some applications might not check them, some applications might not have 
anything reasonable to do (though they can usually at least output a useful 
message to stderr), but other applications may be more robust.  It's not 
particularly uncommon for an application to encounter an error writing to 
volume X and then choose to write to volume Y instead; or to report the error 
back to another component or the end user.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] weird thing with zfs

2006-12-05 Thread Krzys


ok, two weeks ago I did notice one of my disk in zpool got problems.
I was getting Corrupt label; wrong magic number messages, then when I looked 
in format it did not see that disk... (last disk) I had that setup running for 
few months now and all of the sudden last disk failed. So I ordered another 
disk, had it replaced like a week ago, I did issue replace command after disk 
replacement, it was resilvering disks since forever, then I got hints from this 
group that snaps could be causing it so yesterday I did disable snaps and this 
morning I di dnotice the same disk that I replaced is gone... Does it seem weird 
that this disk would fail? Its new disk... I have Solaris 10 U2, 4 internal 
drives and then 7 external drives which are in single enclousures connected via 
scsi chain to each other... So it seems like last disk is failing. Those nipacks 
from sun have self termination so there is no terminator at the end... Any ideas 
what should I do? Do I need to order another drive and replace that one too? Or 
will it happen again? What do you think could be the problem? Ah, when I look at 
that enclosure I do see green light on it so it seems like it did not fail...


format
Searching for disks...
efi_alloc_and_init failed.
done


AVAILABLE DISK SELECTIONS:
   0. c1t0d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   1. c1t1d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   2. c1t2d0 SEAGATE-ST337LC-D703-279.40GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   3. c1t3d0 SEAGATE-ST337LC-D703-279.40GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   4. c3t0d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   5. c3t1d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   6. c3t2d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   7. c3t3d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   8. c3t4d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   9. c3t5d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
  10. c3t6d0 drive type unknown
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0



zpool status -v
  pool: mypool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
mypool  ONLINE   0 0 0
  mirrorONLINE   0 0 0
c1t2d0  ONLINE   0 0 0
c1t3d0  ONLINE   0 0 0

errors: No known data errors

  pool: mypool2
 state: DEGRADED
 scrub: resilver completed with 0 errors on Mon Dec  4 22:34:57 2006
config:

NAME  STATE READ WRITE CKSUM
mypool2   DEGRADED 0 0 0
  raidz   DEGRADED 0 0 0
c3t0d0ONLINE   0 0 0
c3t1d0ONLINE   0 0 0
c3t2d0ONLINE   0 0 0
c3t3d0ONLINE   0 0 0
c3t4d0ONLINE   0 0 0
c3t5d0ONLINE   0 0 0
replacing UNAVAIL  0   775 0  insufficient replicas
  c3t6d0s0/o  UNAVAIL  0 0 0  cannot open
  c3t6d0  UNAVAIL  0   940 0  cannot open

errors: No known data errors

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Krzys
Thanks, ah another wird thing is that when I run format on that frive I get 
a coredump :(


format
Searching for disks...
efi_alloc_and_init failed.
done


AVAILABLE DISK SELECTIONS:
   0. c1t0d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   1. c1t1d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   2. c1t2d0 SEAGATE-ST337LC-D703-279.40GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   3. c1t3d0 SEAGATE-ST337LC-D703-279.40GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   4. c3t0d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   5. c3t1d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   6. c3t2d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   7. c3t3d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   8. c3t4d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   9. c3t5d0 SEAGATE-ST3146807LC-0007-136.73GB
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
  10. c3t6d0 drive type unknown
  /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
Specify disk (enter its number): 10

Segmentation Fault (core dumped)

:( Cant even get to format menu on that drive...

Chris



On Tue, 5 Dec 2006, Nicholas Senedzuk wrote:


The only time that I have seen a format return drive type unknown is when
the drive has failed. You may just have another bad drive and want to try
replacing it again. If that does not work you, you may have another problem
such as a bad backplane or a bad SCSI cable assuming the drive is an
external drive. Hope that helps.




On 12/5/06, Krzys [EMAIL PROTECTED] wrote:



ok, two weeks ago I did notice one of my disk in zpool got problems.
I was getting Corrupt label; wrong magic number messages, then when I
looked
in format it did not see that disk... (last disk) I had that setup running
for
few months now and all of the sudden last disk failed. So I ordered
another
disk, had it replaced like a week ago, I did issue replace command after
disk
replacement, it was resilvering disks since forever, then I got hints from
this
group that snaps could be causing it so yesterday I did disable snaps and
this
morning I di dnotice the same disk that I replaced is gone... Does it seem
weird
that this disk would fail? Its new disk... I have Solaris 10 U2, 4
internal
drives and then 7 external drives which are in single enclousures
connected via
scsi chain to each other... So it seems like last disk is failing. Those
nipacks
from sun have self termination so there is no terminator at the end... Any
ideas
what should I do? Do I need to order another drive and replace that one
too? Or
will it happen again? What do you think could be the problem? Ah, when I
look at
that enclosure I do see green light on it so it seems like it did not
fail...

format
Searching for disks...
efi_alloc_and_init failed.
done


AVAILABLE DISK SELECTIONS:
0. c1t0d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
1. c1t1d0 SEAGATE-ST337LC-D703 cyl 45265 alt 2 hd 16 sec 809
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
2. c1t2d0 SEAGATE-ST337LC-D703-279.40GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
3. c1t3d0 SEAGATE-ST337LC-D703-279.40GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
4. c3t0d0 SEAGATE-ST3146807LC-0007-136.73GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
5. c3t1d0 SEAGATE-ST3146807LC-0007-136.73GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
6. c3t2d0 SEAGATE-ST3146807LC-0007-136.73GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
7. c3t3d0 SEAGATE-ST3146807LC-0007-136.73GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
8. c3t4d0 SEAGATE-ST3146807LC-0007-136.73GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
9. c3t5d0 SEAGATE-ST3146807LC-0007-136.73GB
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0
   10. c3t6d0 drive type unknown
   /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0



zpool status -v
   pool: mypool
  state: ONLINE
  scrub: none requested
config:

 NAMESTATE READ WRITE CKSUM
 mypool  ONLINE 

Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Torrey McMahon

Krzys wrote:
Thanks, ah another wird thing is that when I run format on that 
frive I get a coredump :(


Run pstack /path/to/core and send the output.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Krzys

[12:00:40] [EMAIL PROTECTED]: /d/d3/nb1  pstack core
core 'core' of 29506:   format -e
-  lwp# 1 / thread# 1  
 000239b8 c_disk   (51800, 52000, 4bde4, 525f4, 54e78, 0) + 4e0
 00020fb4 main (2, 0, ffbff8e8, 0, 52000, 29000) + 46c
 000141a8 _start   (0, 0, 0, 0, 0, 0) + 108
-  lwp# 2 / thread# 2  
 ff241818 _door_return (0, 0, 0, 0, fef92400, ff26cbc0) + 10
 ff0c0c30 door_create_func (0, feefc000, 0, 0, ff0c0c10, 0) + 20
 ff2400b0 _lwp_start (0, 0, 0, 0, 0, 0)
-  lwp# 3 / thread# 3  
 ff240154 __lwp_park (75e78, 75e88, 0, 0, 0, 0) + 14
 ff23a1e4 cond_wait_queue (75e78, 75e88, 0, 0, 0, 0) + 28
 ff23a764 cond_wait (75e78, 75e88, 1, 0, 0, ff26cbc0) + 10
 ff142a60 subscriber_event_handler (551d8, fedfc000, 0, 0, ff142a2c, 0) + 34
 ff2400b0 _lwp_start (0, 0, 0, 0, 0, 0)



On Tue, 5 Dec 2006, Torrey McMahon wrote:


Krzys wrote:
Thanks, ah another wird thing is that when I run format on that frive I 
get a coredump :(


Run pstack /path/to/core and send the output.


!DSPAM:122,45759fd826586021468!


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Al Hopper
On Tue, 5 Dec 2006, Krzys wrote:

 Thanks, ah another wird thing is that when I run format on that frive I 
 get
 a coredump :(
... snip 

Try zeroing out the disk label with something like:

dd if=/dev/zero of=/dev/rdsk/c?t?d?p0  bs=1024k count=1024

Regards,

Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
   Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
 OpenSolaris Governing Board (OGB) Member - Feb 2006
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Joerg Schilling
Al Hopper [EMAIL PROTECTED] wrote:

 On Tue, 5 Dec 2006, Krzys wrote:

  Thanks, ah another wird thing is that when I run format on that frive I 
  get
  a coredump :(
 ... snip 

 Try zeroing out the disk label with something like:

 dd if=/dev/zero of=/dev/rdsk/c?t?d?p0  bs=1024k count=1024

Do you expect a 1 GB disk label?

Jörg

-- 
 EMail:[EMAIL PROTECTED] (home) Jörg Schilling D-13353 Berlin
   [EMAIL PROTECTED](uni)  
   [EMAIL PROTECTED] (work) Blog: http://schily.blogspot.com/
 URL:  http://cdrecord.berlios.de/old/private/ ftp://ftp.berlios.de/pub/schily
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Al Hopper
On Tue, 5 Dec 2006, Joerg Schilling wrote:

 Al Hopper [EMAIL PROTECTED] wrote:

  On Tue, 5 Dec 2006, Krzys wrote:
 
   Thanks, ah another wird thing is that when I run format on that frive 
   I get
   a coredump :(
  ... snip 
 
  Try zeroing out the disk label with something like:
 
  dd if=/dev/zero of=/dev/rdsk/c?t?d?p0  bs=1024k count=1024

 Do you expect a 1 GB disk label?

No.  :)

Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
   Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
 OpenSolaris Governing Board (OGB) Member - Feb 2006
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Krzys

Does not work :(

dd if=/dev/zero of=/dev/rdsk/c3t6d0s0 bs=1024k count=1024
dd: opening `/dev/rdsk/c3t6d0s0': I/O error

That is so strange... it seems like I lost another disk... I will try to reboot 
and see what I get, but I guess I need to order another disk then and give it a 
try...


Chris





On Tue, 5 Dec 2006, Al Hopper wrote:


On Tue, 5 Dec 2006, Krzys wrote:


Thanks, ah another wird thing is that when I run format on that frive I get
a coredump :(

... snip 

Try zeroing out the disk label with something like:

dd if=/dev/zero of=/dev/rdsk/c?t?d?p0  bs=1024k count=1024

Regards,

Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
  Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
OpenSolaris Governing Board (OGB) Member - Feb 2006


!DSPAM:122,4575a7731650371292!


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Managed to corrupt my pool

2006-12-05 Thread Jim Hranicky
 So the questions are:
 
 - is this fixable? I don't see an inum I could run
  find on to remove, 
and I can't even do a zfs volinit anyway:
nextest-01# zfs volinit
  cannot iterate filesystems: I/O error
 
 - would not enabling zil_disable have prevented
  this?
 
- Should I have been doing a 3-way mirror?
 - Is there a more optimum configuration to help
  prevent this  kind of corruption?

Anyone have any thoughts on this? I'd really like to be 
able to build a nice ZFS box for file service but if a 
hardware failure can corrupt a disk pool I'll have to 
try to find another solution, I'm afraid.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Krzys

Ok, so here is an update

I did restart my sysyte, I power it off and power it on. Here is screen capture 
of my boot. I certainly do have some hard drive issues and will need to take a 
look at them... But I got my disk back visible to the system and zfs is doing 
resilvering again


Rebooting with command: boot
Boot device: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0:a  
File and args:
SunOS Release 5.10 Version Generic_118833-24 64-bit
Copyright 1983-2006 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
Hardware watchdog enabled
Hostname: chrysek
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 6 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5):
Error for Command: read(10)Error Level: Retryable
Requested Block: 286732066 Error Block: 286732066
Vendor: SEAGATESerial Number: 3HY14PVS
Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 
0x2

WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 3 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd23):
Error for Command: read(10)Error Level: Retryable
Requested Block: 283623842 Error Block: 283623842
Vendor: SEAGATESerial Number: 3HY8HS7L
Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 
0x2

WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 5 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd25):
Error for Command: read(10)Error Level: Retryable
Requested Block: 283623458 Error Block: 283623458
Vendor: SEAGATESerial Number: 3HY0LF18
Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 0x0, FRU: 
0x2

/kernel/drv/sparcv9/zpool symbol avl_add multiply defined
/kernel/drv/sparcv9/zpool symbol assfail3 multiply defined
WARNING: kstat_create('unix', 0, 'dmu_buf_impl_t'): namespace collision
mypool2/d3 uncorrectable error
checking ufs filesystems
/dev/rdsk/c1t0d0s7: is logging.

chrysek console login: VERITAS SCSA Generic Revision: 3.5c
Dec  5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing
Dec  5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp missing
Dec  5 13:01:46 chrysek VERITAS: No proxy found.
Dec  5 13:01:52 chrysek vmd[546]: ready for connections
Dec  5 13:01:53 chrysek VERITAS: No proxy found.
Dec  5 13:01:54 chrysek VERITAS: No proxy found.
Dec  5 13:02:00 chrysek VERITAS: No proxy found.
Dec  5 13:02:01 chrysek VERITAS: No proxy found.
Dec  5 13:02:03 chrysek VERITAS: No proxy found.
starting NetWorker daemons:
 nsrexecd
 lgtolmd
Dec  5 13:02:20 chrysek CNS Transport[841]: cctransport started
Dec  5 13:02:48 chrysek webmin[1353]: Webmin starting
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED] (glm2):
Dec  5 13:19:07 chrysek Target 6 disabled wide SCSI mode
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED] (glm2):
Dec  5 13:19:07 chrysek Target 6 reverting to async. mode
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED]/[EMAIL PROTECTED],0 (sd5):
Dec  5 13:19:07 chrysek Error for Command: write(10)   Error 
Level: Retryable
Dec  5 13:19:07 chrysek scsi:   Requested Block: 137163259 Error 
Block: 137163259
Dec  5 13:19:07 chrysek scsi:   Vendor: SEAGATE 
Serial Number: 3HY14PVS

Dec  5 13:19:07 chrysek scsi:   Sense Key: Aborted Command
Dec  5 13:19:07 chrysek scsi:   ASC: 0x47 (scsi parity error), ASCQ: 0x0, FRU: 
0x3



but now when I do zpool status -v
  pool: mypool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
mypool  ONLINE   0 0 0
  mirrorONLINE   0 0 0
c1t2d0  ONLINE   0 0 0
c1t3d0  ONLINE   0 0 0

errors: No known data errors

  pool: mypool2
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
 scrub: resilver in progress, 4.40% done, 11h40m to go

Re: [zfs-discuss] replacing a drive in a raidz vdev

2006-12-05 Thread Mark Maybee

Jeremy Teo wrote:

On 12/5/06, Bill Sommerfeld [EMAIL PROTECTED] wrote:


On Mon, 2006-12-04 at 13:56 -0500, Krzys wrote:
 mypool2/[EMAIL PROTECTED]  34.4M  -   151G  -
 mypool2/[EMAIL PROTECTED]  141K  -   189G  -
 mypool2/d3 492G   254G  11.5G  legacy

 I am so confused with all of this... Why its taking so long to 
replace that one

 bad disk?

To workaround a bug where a pool traverse gets lost when the snapshot
configuration of a pool changes, both scrubs and resilvers will start
over again any time you create or delete a snapshot.

Unfortunately, this workaround has problems of its own -- If your
inter-snapshot interval is less than the time required to complete a
scrub, the resilver will never complete.

The open bug is:

6343667 scrub/resilver has to start over when a snapshot is taken

if it's not going to be fixed any time soon, perhaps we need a better
workaround:



Anyone internal working on this?


Yes.  But its going to be a few months.

-Mark
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Managed to corrupt my pool

2006-12-05 Thread Jim Hranicky
 Anyone have any thoughts on this? I'd really like to
 be able to build a nice ZFS box for file service but if
 a  hardware failure can corrupt a disk pool I'll have to
  try to find another solution, I'm afraid.

Sorry, I worded this poorly -- if the loss of a disk in a mirror
can corrupt the pool it's going to give me pause in implementing
a ZFS solution. 

Jim
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: weird thing with zfs

2006-12-05 Thread Chris Gerhard
What os is this?

What is the hardware?

can you try running format with efi_debug set. You have to run format using a 
debugger and patch the variable. Here is how using mdb (set a break point in 
main so that the dynamic linker has done it's stuff, then update the value of 
efi_debug to be 1, then continue):

# mdb /usr/sbin/format
 main:b
 :r
mdb: stop at main
mdb: target stopped at:
main:   pushl  %ebp
 efi_debug/x
libefi.so.1`efi_debug:
libefi.so.1`efi_debug:  0   
 efi_debug/w 1
libefi.so.1`efi_debug:  0   =   0x1
 :c
Searching for disks...done

--chris
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: ZFS related kernel panic

2006-12-05 Thread Peter Eriksson
 So ZFS should be more resilient against write errors, and the SCSI disk or 
 FC drivers
 should be more resilient against LIPs (the most likely cause of your 
 problem) or other
 transient errors. (Alternatively, the ifp driver should be updated to 
 support the
 maximum number of targets on a loop, which might also solve your second 
 problem.)

 NB. LIPs are a normal part of everyday life for fibre channel, they are not 
 an error.

Right. I don't think it's the LIP's that's the problem but rather (a guess, not 
verified) the fact that the HBA loses light on it's fiber interface when the 
switch reboots... I think I also saw the same ZFS-induced panic when I (stupid, 
I know, but...) moved a fiber cable from one GBIC in the switch to another on 
the run. I also saw this with the ifp driver btw. And as someone wrote - the 
ifp driver will never be updated since it's for EOL'd hardware :-)
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: ZFS related kernel panic

2006-12-05 Thread Peter Eriksson
Hmm... I just noticed this qla2100.conf option:

# During link down conditions enable/disable the reporting of
# errors.
#0 = disabled, 1 = enable
hba0-link-down-error=1;
hba1-link-down-error=1;

I _wonder_ what might possibly happen if I change that 1 to a 0 (zero)... :-)
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Richard Elling

This looks more like a cabling or connector problem.  When that happens
you should see parity errors and transfer rate negotiations.
 -- richard

Krzys wrote:

Ok, so here is an update

I did restart my sysyte, I power it off and power it on. Here is screen 
capture of my boot. I certainly do have some hard drive issues and will 
need to take a look at them... But I got my disk back visible to the 
system and zfs is doing resilvering again


Rebooting with command: boot
Boot device: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0:a  
File and args:
SunOS Release 5.10 Version Generic_118833-24 64-bit
Copyright 1983-2006 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
Hardware watchdog enabled
Hostname: chrysek
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 6 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5):
Error for Command: read(10)Error Level: Retryable
Requested Block: 286732066 Error Block: 286732066
Vendor: SEAGATESerial Number: 3HY14PVS
Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 
0x0, FRU: 0x2

WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 3 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd23):
Error for Command: read(10)Error Level: Retryable
Requested Block: 283623842 Error Block: 283623842
Vendor: SEAGATESerial Number: 3HY8HS7L
Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 
0x0, FRU: 0x2

WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 5 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd25):
Error for Command: read(10)Error Level: Retryable
Requested Block: 283623458 Error Block: 283623458
Vendor: SEAGATESerial Number: 3HY0LF18
Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 
0x0, FRU: 0x2

/kernel/drv/sparcv9/zpool symbol avl_add multiply defined
/kernel/drv/sparcv9/zpool symbol assfail3 multiply defined
WARNING: kstat_create('unix', 0, 'dmu_buf_impl_t'): namespace collision
mypool2/d3 uncorrectable error
checking ufs filesystems
/dev/rdsk/c1t0d0s7: is logging.

chrysek console login: VERITAS SCSA Generic Revision: 3.5c
Dec  5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp 
missing
Dec  5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp 
missing

Dec  5 13:01:46 chrysek VERITAS: No proxy found.
Dec  5 13:01:52 chrysek vmd[546]: ready for connections
Dec  5 13:01:53 chrysek VERITAS: No proxy found.
Dec  5 13:01:54 chrysek VERITAS: No proxy found.
Dec  5 13:02:00 chrysek VERITAS: No proxy found.
Dec  5 13:02:01 chrysek VERITAS: No proxy found.
Dec  5 13:02:03 chrysek VERITAS: No proxy found.
starting NetWorker daemons:
 nsrexecd
 lgtolmd
Dec  5 13:02:20 chrysek CNS Transport[841]: cctransport started
Dec  5 13:02:48 chrysek webmin[1353]: Webmin starting
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED] (glm2):
Dec  5 13:19:07 chrysek Target 6 disabled wide SCSI mode
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED] (glm2):
Dec  5 13:19:07 chrysek Target 6 reverting to async. mode
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED]/[EMAIL PROTECTED],0 (sd5):
Dec  5 13:19:07 chrysek Error for Command: 
write(10)   Error Level: Retryable
Dec  5 13:19:07 chrysek scsi:   Requested Block: 
137163259 Error Block: 137163259

Dec  5 13:19:07 chrysek scsi:   Vendor: SEAGATE Serial Number: 3HY14PVS
Dec  5 13:19:07 chrysek scsi:   Sense Key: Aborted Command
Dec  5 13:19:07 chrysek scsi:   ASC: 0x47 (scsi parity error), ASCQ: 
0x0, FRU: 0x3



but now when I do zpool status -v
  pool: mypool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
mypool  ONLINE   0 0 0
  mirrorONLINE   0 0 0
c1t2d0  ONLINE   0 0 0
c1t3d0  ONLINE   0 0 0

errors: No known data errors

  pool: mypool2
 state: DEGRADED
status: One or more devices is currently being resilvered.  The pool will
 

Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Richard Elling

BTW, there is a way to check what the SCSI negotiations resolved to.
I wrote about it once in a BluePrint
http://www.sun.com/blueprints/0500/sysperfnc.pdf
See page 11
 -- richard

Richard Elling wrote:

This looks more like a cabling or connector problem.  When that happens
you should see parity errors and transfer rate negotiations.
 -- richard

Krzys wrote:

Ok, so here is an update

I did restart my sysyte, I power it off and power it on. Here is 
screen capture of my boot. I certainly do have some hard drive issues 
and will need to take a look at them... But I got my disk back visible 
to the system and zfs is doing resilvering again


Rebooting with command: boot
Boot device: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0:a  
File and args:
SunOS Release 5.10 Version Generic_118833-24 64-bit
Copyright 1983-2006 Sun Microsystems, Inc.  All rights reserved.
Use is subject to license terms.
Hardware watchdog enabled
Hostname: chrysek
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 6 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd5):
Error for Command: read(10)Error Level: Retryable
Requested Block: 286732066 Error Block: 286732066
Vendor: SEAGATESerial Number: 
3HY14PVS

Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 
0x0, FRU: 0x2

WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 3 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd23):
Error for Command: read(10)Error Level: Retryable
Requested Block: 283623842 Error Block: 283623842
Vendor: SEAGATESerial Number: 
3HY8HS7L

Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 
0x0, FRU: 0x2

WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
SCSI bus DATA IN phase parity error
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED] (glm2):
Target 5 reducing sync. transfer rate
WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 (sd25):
Error for Command: read(10)Error Level: Retryable
Requested Block: 283623458 Error Block: 283623458
Vendor: SEAGATESerial Number: 
3HY0LF18

Sense Key: Aborted Command
ASC: 0x48 (initiator detected error message received), ASCQ: 
0x0, FRU: 0x2

/kernel/drv/sparcv9/zpool symbol avl_add multiply defined
/kernel/drv/sparcv9/zpool symbol assfail3 multiply defined
WARNING: kstat_create('unix', 0, 'dmu_buf_impl_t'): namespace collision
mypool2/d3 uncorrectable error
checking ufs filesystems
/dev/rdsk/c1t0d0s7: is logging.

chrysek console login: VERITAS SCSA Generic Revision: 3.5c
Dec  5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp 
missing
Dec  5 13:01:38 chrysek root: CAPTURE_UPTIME ERROR: /var/opt/SUNWsrsrp 
missing

Dec  5 13:01:46 chrysek VERITAS: No proxy found.
Dec  5 13:01:52 chrysek vmd[546]: ready for connections
Dec  5 13:01:53 chrysek VERITAS: No proxy found.
Dec  5 13:01:54 chrysek VERITAS: No proxy found.
Dec  5 13:02:00 chrysek VERITAS: No proxy found.
Dec  5 13:02:01 chrysek VERITAS: No proxy found.
Dec  5 13:02:03 chrysek VERITAS: No proxy found.
starting NetWorker daemons:
 nsrexecd
 lgtolmd
Dec  5 13:02:20 chrysek CNS Transport[841]: cctransport started
Dec  5 13:02:48 chrysek webmin[1353]: Webmin starting
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED] (glm2):
Dec  5 13:19:07 chrysek Target 6 disabled wide SCSI mode
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL 
PROTECTED] (glm2):
Dec  5 13:19:07 chrysek Target 6 reverting to async. mode
Dec  5 13:19:07 chrysek scsi: WARNING: /[EMAIL PROTECTED],60/[EMAIL PROTECTED]/[EMAIL PROTECTED],0 
(sd5):
Dec  5 13:19:07 chrysek Error for Command: 
write(10)   Error Level: Retryable
Dec  5 13:19:07 chrysek scsi:   Requested Block: 
137163259 Error Block: 137163259

Dec  5 13:19:07 chrysek scsi:   Vendor: SEAGATE Serial Number: 3HY14PVS
Dec  5 13:19:07 chrysek scsi:   Sense Key: Aborted Command
Dec  5 13:19:07 chrysek scsi:   ASC: 0x47 (scsi parity error), ASCQ: 
0x0, FRU: 0x3



but now when I do zpool status -v
  pool: mypool
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
mypool  ONLINE   0 0 0
  mirrorONLINE   0 0 0

Re: [zfs-discuss] Re: Re: ZFS related kernel panic

2006-12-05 Thread Douglas Denny

On 12/5/06, Peter Eriksson [EMAIL PROTECTED] wrote:

Hmm... I just noticed this qla2100.conf option:

# During link down conditions enable/disable the reporting of
# errors.
#0 = disabled, 1 = enable
hba0-link-down-error=1;
hba1-link-down-error=1;


This is the driver the we are using in this configuration. Excellent
insight. This will be added to the testing. Although, we are moving
away from the Qlogic cards, back to the Sun branded Qlogic cards to
use MPxIO. Which works flawlessly with UFS and raw drives.

I wonder if using MPxIO, and dual connected / path HBAs would also
reduce these errors.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] weird thing with zfs

2006-12-05 Thread Nathan Kroenert
Hm. If the disk has no label, why would it have an s0?

Or, did you mean p0?

Nathan.


On Wed, 2006-12-06 at 04:45, Krzys wrote:
 Does not work :(
 
 dd if=/dev/zero of=/dev/rdsk/c3t6d0s0 bs=1024k count=1024
 dd: opening `/dev/rdsk/c3t6d0s0': I/O error
 
 That is so strange... it seems like I lost another disk... I will try to 
 reboot 
 and see what I get, but I guess I need to order another disk then and give it 
 a 
 try...
 
 Chris
 
 
 
 
 
 On Tue, 5 Dec 2006, Al Hopper wrote:
 
  On Tue, 5 Dec 2006, Krzys wrote:
 
  Thanks, ah another wird thing is that when I run format on that frive 
  I get
  a coredump :(
  ... snip 
 
  Try zeroing out the disk label with something like:
 
  dd if=/dev/zero of=/dev/rdsk/c?t?d?p0  bs=1024k count=1024
 
  Regards,
 
  Al Hopper  Logical Approach Inc, Plano, TX.  [EMAIL PROTECTED]
Voice: 972.379.2133 Fax: 972.379.2134  Timezone: US CDT
  OpenSolaris.Org Community Advisory Board (CAB) Member - Apr 2005
  OpenSolaris Governing Board (OGB) Member - Feb 2006
 
 
  !DSPAM:122,4575a7731650371292!
 
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
-- 

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] raidz DEGRADED state

2006-12-05 Thread David Bustos
Quoth Thomas Garner on Thu, Nov 30, 2006 at 06:41:15PM -0500:
 I currently have a 400GB disk that is full of data on a linux system.
 If I buy 2 more disks and put them into a raid-z'ed zfs under solaris,
 is there a generally accepted way to build an degraded array with the
 2 disks, copy the data to the new filesystem, and then move the
 original disk to complete the array?

No, because we currently can't add disks to a raidz array.  You could
create a mirror instead and then add in the other disk to make
a three-way mirror, though.

Even doing that would be dicey if you only have a single machine,
though, since Solaris can't natively read the popular Linux filesystems.
I believe there is freeware to do it, but nothing supported.


David
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] raidz DEGRADED state

2006-12-05 Thread Thomas Garner

So there is no current way to specify the creation of a 3 disk raid-z
array with a known missing disk?

On 12/5/06, David Bustos [EMAIL PROTECTED] wrote:

Quoth Thomas Garner on Thu, Nov 30, 2006 at 06:41:15PM -0500:
 I currently have a 400GB disk that is full of data on a linux system.
 If I buy 2 more disks and put them into a raid-z'ed zfs under solaris,
 is there a generally accepted way to build an degraded array with the
 2 disks, copy the data to the new filesystem, and then move the
 original disk to complete the array?

No, because we currently can't add disks to a raidz array.  You could
create a mirror instead and then add in the other disk to make
a three-way mirror, though.

Even doing that would be dicey if you only have a single machine,
though, since Solaris can't natively read the popular Linux filesystems.
I believe there is freeware to do it, but nothing supported.


David


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Re: Managed to corrupt my pool

2006-12-05 Thread Neil Perrin

Jim,

I'm not at all sure what happened to your pool.
However, I can answer some of your questions.

Jim Hranicky wrote On 12/05/06 11:32,:

So the questions are:

- is this fixable? I don't see an inum I could run
find on to remove, 


I think the pool is busted. Even the message printed in your
previous email is bad:

  DATASET  OBJECT  RANGE
  15   0   lvl=4294967295 blkid=0

as level is way out of range.


  and I can't even do a zfs volinit anyway:
  nextest-01# zfs volinit
cannot iterate filesystems: I/O error


I'm not sure why you're using zfs volinit which I believe creates
the zvol links, but this further shows problems.



- would not enabling zil_disable have prevented this?


No the intent log is not needed for pool integrity.
It ensures the synchronous semantics of O_DSYNC/fsync are obeyed.

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Shared ZFS pools

2006-12-05 Thread Anton B. Rang
You specify the mirroring configuration. The top-level vdevs are implicitly 
striped. So if you, for instance, request something like

  zpool create mirror AA BA mirror AB BB

then you will have a pool consisting of a stripe of two mirrors. Each mirror 
will have one copy of its data at each location. Hence if site B goes down, 
you'll lose BA and BB, and each mirror will transition to a degraded state.

So the answer to your first question is yes, with this configuration you can be 
sure that no data will reside solely on LUNs at site B.

On your second question, I don't believe there is work currently being done on 
shared pools, but I'll let the ZFS team clarify the road map.
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Managed to corrupt my pool

2006-12-05 Thread Anton B. Rang
 I think the pool is busted. Even the message printed in your
 previous email is bad:
 
DATASET  OBJECT  RANGE
 15   0   lvl=4294967295 blkid=0
 
 as level is way out of range.

I think this could be from dmu_objset_open_impl().

It sets object to 0 and level to -1 (= 4294967295).  [Hmmm, this also seems to 
indicate a truncation from 64 to 32 bits somewhere.]

Would zdb would show any more detail?

(Actually, it looks like the ZIL also sets object to 0 and level to -1 when 
accessing its blocks, but since the ZIL was disabled, I'd guess this isn't the 
issue here.)
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: raidz DEGRADED state

2006-12-05 Thread Anton B. Rang
Creating an array configuration with one element being a sparse file, then 
removing that file, comes to mind, but I wouldn't want to be the first to 
attempt it.  ;-)
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss