Re: [zfs-discuss] How to recover -- LUNs go offline, now permanent errors?

2011-07-18 Thread David Smith
Cindy,

I gave your suggestion a try.  I did the zpool clear and then did another zpool 
scrub  and all is happy now.  Thank you for your help.

David
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] How to recover -- LUNs go offline, now permanent errors?

2011-07-15 Thread David Smith
Cindy,

Thanks for the reply.  I'll get that a try and then send an update.

Thanks,

David
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How to recover -- LUNs go offline, now permanent errors?

2011-07-13 Thread David Smith
I recently had an issue with my LUNs from our storage unit going offline.  This 
caused the zpool to get numerous errors on the luns.  The pool is on-line, and 
I did a scrub, but one of the raid sets is
degraded:

   raidz2-3 DEGRADED 0 0 0
c7t60001FF011C6F3103B00011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F3023900011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F2F53700011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F2E43500011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F2D23300011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F2A93100011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F29A2F00011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F2682D00011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F24C2B00011D1BF1d0  DEGRADED 0 0 0  
too many errors
c7t60001FF011C6F2192900011D1BF1d0  DEGRADED 0 0 0  
too many errors

Also I have the following:
errors: Permanent errors have been detected in the following files:

<0x3a>:<0x3b04>

Originally, there was a file, and then a directory listed, but I removed them.  
Now I'm stuck with
the hex codes above.  How do I interpret them?  Can this pool be recovered, or 
basically how do
I proceed?

The system is Solaris 10 U9 with all recent patches.

Thanks,

David
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Zpool metadata corruption from S10U9 to S11 express

2011-06-22 Thread David Smith
An update:

I had mirrored my boot drive when I installed Solaris 10U9 originally, so I 
went ahead and rebooted the system to this disk instead of my Solaris 11 
install.  After getting the system up, I imported the zpool, and everything 
worked normally.  

So I guess there is some sort of incompatibility between Solaris 10 and Solaris 
11.  I would have thought that Solaris 11 could import an older pool level.

Any other insight on importing pools between these two versions of Solaris 
would be helpful.

Thanks,

David
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Zpool metadata corruption from S10U9 to S11 express

2011-06-22 Thread David Smith
I was recently running Solaris 10 U9 and I decided that I would like to go  

to Solaris 11 Express so I exported my zpool, hoping that I would just do   

an import once I had the new system installed with Solaris 11.  Now when I  

try to do an import I'm getting the following:  



# /home/dws# zpool import   

  pool: tank

id: 13155614069147461689

 state: FAULTED 

status: The pool metadata is corrupted. 

action: The pool cannot be imported due to damaged devices or data. 

   see: http://www.sun.com/msg/ZFS-8000-72  

config: 



tank FAULTED  corrupted data

logs

  mirror-6   ONLINE 

c9t57d0  ONLINE 

c9t58d0  ONLINE 

  mirror-7   ONLINE 

c9t59d0  ONLINE 

c9t60d0  ONLINE 



Is there something else I can do to see what is wrong.  



Original attempt when specifying the name resulted in:  



# /home/dws# zpool import tank  

cannot import 'tank': I/O error 

Destroy and re-create the pool from 

a backup source.



I verified that I have all 60 of my luns.  The controller numbers have  

changed, but I don't believe that should matter. 

Any suggestions about getting additional information about what is happening

would be greatly appreciated.   



Thanks, 



David
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How to map solaris disk devices to physical location for ZFS pool setup

2009-09-15 Thread David Smith
Hi, I'm setting up a ZFS environment running on a Sun x4440 + J4400 arrays 
(similar to 7410 environment) and I was trying to figure out the best way to 
map a disk drive physical location (tray and slot) to the Solaris device 
c#t#d#.   Do I need to install the CAM software to do this, or is there another 
way?  I would like to understand the solaris device to physical drive location 
so that I can setup my ZFS pool mirrors/raid properly.

I'm currently running Solaris Express build 119.

Thanks,

David
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs destroy is taking a long time...

2009-01-08 Thread David Smith
A few more details:

The system is a Sun x4600 running Solaris 10 Update 4.
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs destroy is taking a long time...

2009-01-08 Thread David Smith
I was wondering if anyone has any experience with how long a "zfs destroy" of 
about 40 TB should take?  So far, it has been about an hour...  Is there any 
good way to tell if it is working or if it is hung?

Doing a "zfs list" just hangs.  If you do a more specific zfs list, then it is 
okay... zfs list pool/another-fs

Thanks,

David
-- 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Replacing a Lun (Raid 0) (Santricity)

2008-03-14 Thread David Smith
Yes! That worked to get the spare back to an available state.  Thanks!


So that leaves me with the trying to put together a recommended procedure to 
replace a failed lun/disk from our Flexline 380.   Does anyone have 
configuration in
which they are using a RAID 0 lun, which they need to replace?

Thanks,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Replacing a Lun (Raid 0) (Santricity)

2008-03-14 Thread David Smith
Addtional information:

It looks like perhaps the original drive is in use, and the hot spare is 
assigned but not in use see below about zpool iostat:

  raidz22.76T  4.49T  0  0  29.0K  18.4K
c10t600A0B80001139967CF945E80E95d0  -  -  0  0  2.46K  
1.33K
c10t600A0B8000115EA2FEEB45E8145Ed0  -  -  0  0  2.46K  
1.33K
spare   -  -  0  0  1.81K  1.33K
  c10t600A0B80001139967D2345E81075d0  -  -  0  0  2.47K 
 1.33K
  c10t600A0B80001139967CE145E80D4Dd0  -  -  0  0  0 
 1.33K
c10t600A0B80001139967D3F45E81157d0  -  -  0  0  2.47K  
1.33K
c10t600A0B8000115EA2FF1145E817BEd0  -  -  0  0  2.46K  
1.33K
c10t600A0B80001139967D5D45E813EBd0  -  -  0  0  2.46K  
1.33K
c10t600A0B8000115EA2FE7D45E80CDEd0  -  -  0  0  2.47K  
1.33K
c10t600A0B80001139967C6145E808C7d0  -  -  0  0  2.47K  
1.33K
c10t600A0B8000115EA2FE9945E80E6Ad0  -  -  0  0  2.46K  
1.33K
c10t600A0B80001139967C8B45E80A59d0  -  -  0  0  2.46K  
1.33K
c10t600A0B80001139967CA745E80B21d0  -  -  0  0  2.47K  
1.33K
c10t600A0B8000115EA2FEB545E810D4d0  -  -  0  0  2.47K  
1.33K
c10t600A0B80001139967CD145E80CD7d0  -  -  0  0  2.47K  
1.33K
c10t600A0B8000115EA2FED145E8129Cd0  -  -  0  0  2.47K  
1.33K
c10t600A0B80001139967CFB45E80EA5d0  -  -  0  0  2.47K  
1.33K
c10t600A0B8000115EA2FEED45E8146Ed0  -  -  0  0  2.46K  
1.33K

spares
  c10t600A0B80001139967CE145E80D4Dd0  INUSE currently in use

So how do I get the hot spare out of the inuse state?
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Replacing a Lun (Raid 0) (Santricity)

2008-03-14 Thread David Smith
I would like advise about how to replace a raid 0 lun.  The lun is basically a 
raid 0 lun which is from a single disk volume group / volume from our Flexline 
380 unit.  So every disk in the unit is a volume group/volume/lun mapped to the 
host.  We then let ZFS do the raid.  

We have a lun now which has been getting read errors and basically the 
underlying drive needs to be replaced.  I've done this in the past but as I 
remember it was cumbersome and didn't go that smoothly so I would like some 
advise about how to go about it.  As I recall if you just fail the disk via 
Santricity, the lun really doesn't go offline from a host point of via and 
hence ZFS still tries to write to the lun/disk.  I believe I either unmapped 
the lun to the host at which point ZFS kicked in a hot spare, or I offlined the 
disk via ZFS first.  Then at that point the drive can be replace on the storage 
and the volume re-initialized then remapped back to the host.  So part of my 
question... does the above sound reasonable, or should I be doing this 
differently?  Also I'm a little unsure about how to get the original lun back 
in operation and the spare back to a spare.

For example I have the following situation now:

  raidz2  ONLINE   0 0 0
c10t600A0B80001139967CF945E80E95d0ONLINE   0 0 0
c10t600A0B8000115EA2FEEB45E8145Ed0ONLINE   0 0 0
spare ONLINE   0 0 0
  c10t600A0B80001139967D2345E81075d0  ONLINE   0 0 0
  c10t600A0B80001139967CE145E80D4Dd0  ONLINE   0 0 0
c10t600A0B80001139967D3F45E81157d0ONLINE   0 0 0
c10t600A0B8000115EA2FF1145E817BEd0ONLINE   0 0 0
c10t600A0B80001139967D5D45E813EBd0ONLINE   0 0 0
c10t600A0B8000115EA2FE7D45E80CDEd0ONLINE   0 0 0
c10t600A0B80001139967C6145E808C7d0ONLINE   0 0 0
c10t600A0B8000115EA2FE9945E80E6Ad0ONLINE   0 0 0
c10t600A0B80001139967C8B45E80A59d0ONLINE   0 0 0
c10t600A0B80001139967CA745E80B21d0ONLINE   0 0 0
c10t600A0B8000115EA2FEB545E810D4d0ONLINE   0 0 0
c10t600A0B80001139967CD145E80CD7d0ONLINE   0 0 0
c10t600A0B8000115EA2FED145E8129Cd0ONLINE   0 0 0
c10t600A0B80001139967CFB45E80EA5d0ONLINE   0 0 0
c10t600A0B8000115EA2FEED45E8146Ed0ONLINE   0 0 0

spares
  c10t600A0B80001139967CE145E80D4Dd0  INUSE currently in use
  c10t600A0B8000115EA2FEE145E81328d0  AVAIL   
  c10t600A0B80001139967D0B45E80F21d0  AVAIL   
  c10t600A0B8000115EA2FEFD45E81506d0  AVAIL   
  c10t600A0B80001139967D3545E81107d0  AVAIL   
  c10t600A0B80001139967D5345E81289d0  AVAIL   
  c10t600A0B8000115EA2FF2345E81864d0  AVAIL   
  c10t600A0B80001139967D6F45E8149Bd0  AVAIL   

I thought if I replace the original device with the same (but new) lun that 
should work, but I get
the following:

zpool replace tank c10t600A0B80001139967D2345E81075d0 
c10t600A0B80001139967D2345E81075d0
invalid vdev specification
use '-f' to override the following errors:
/dev/dsk/c10t600A0B80001139967D2345E81075d0s0 is part of active ZFS pool 
tank. Please see zpool(1M).

The above lun is the same as before, but the underlying disk on the storage was 
replaced.  Do I need to do something to this lun to make ZFS think it is a new 
disk?  Or should I be doing something different?

I now have another disk which has gone back so I need to fix my above situation 
with the hot
spare first, and then go through the process again for my second failure.

The environment is S10U4, running on a x4600 with Flexline 380 storage units.

Tia,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool status (advanced listing)?

2007-09-26 Thread David Smith
Under the GUI, there is an "advanced" option which shows vdev capacity, etc.  
I'm drawing a blank about how to get with the commands... 

Thanks,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] remove snapshots

2007-08-17 Thread David Smith
To list your snapshots:

/usr/sbin/zfs list -H -t snapshot -o name

Then you could use that in a for loop:

for i in `/usr/sbin/zfs list -H -t snapshot -o name` ;
do
  echo "Destroying snapshot: $i"
  /usr/sbin/zfs destroy $i
done

The above would destroy all your snapshots.  You could put a grep on the end to 
select 
certain ones by a specific string...

For example:

/usr/sbin/zfs list -H -t snapshot -o name | grep "aug-16"

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] General recommendations on raidz groups of different sizes

2007-07-18 Thread David Smith
What are your thoughts or recommendations on having a zpool made up of
raidz groups of different sizes?  Are there going to be performance issues?

For example:

  pool: testpool1
 state: ONLINE
 scrub: none requested
config:

NAMESTATE READ WRITE CKSUM
testpool1   ONLINE   0 0 0

  raidz1ONLINE   0 0 0
c12t600A0B800029E5EA07234685122Ad0  ONLINE   0 0 0
c12t600A0B800029E5EA07254685123Cd0  ONLINE   0 0 0
c12t600A0B800029E5EA072F46851256d0  ONLINE   0 0 0
c12t600A0B800029E5EA073146851266d0  ONLINE   0 0 0
c12t600A0B800029E5EA073746851278d0  ONLINE   0 0 0
c12t600A0B800029E5EA074146851292d0  ONLINE   0 0 0
c12t600A0B800029E5EA0747468512B6d0  ONLINE   0 0 0
c12t600A0B800029E5EA0749468512C2d0  ONLINE   0 0 0
  raidz1ONLINE   0 0 0
c12t600A0B800029E5EA074F468512E0d0  ONLINE   0 0 0
c12t600A0B800029E5EA0751468512E8d0  ONLINE   0 0 0
c12t600A0B800029E5EA07574685130Cd0  ONLINE   0 0 0
c12t600A0B800029E5EA075946851318d0  ONLINE   0 0 0
c12t600A0B800029E5EA075F4685132Ed0  ONLINE   0 0 0
c12t600A0B800029E5EA076546851342d0  ONLINE   0 0 0


Thanks,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zfs list hangs if zfs send is killed (leaving zfs receive process)

2007-07-13 Thread David Smith
Well, the zfs receive process finally died, and now my zfs list works just fine.

If there is a better way to capture what is going on, please let me know and I 
can duplicate the hang.

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Again ZFS with expanding LUNs!

2007-07-13 Thread David Smith
I don't believe LUN expansion is quite yet possible under Solaris 10 (11/06).  
I believe this might make it into the next update but I'm not sure on that.  
Someone from Sun would need to comment on when this will make it into the 
production release of Solaris.

I know this because I was working with a person from Sun testing a utility 
which will detect the larger LUN and update the disk label.  Then at that point 
if you export the pool and import it again the added space will show up.  The 
utility was just for testing however.  I'm waiting as well for the automatic 
LUN expansion stuff.  

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zfs list hangs if zfs send is killed (leaving zfs receive process)

2007-07-13 Thread David Smith
I was in the process of doing a large zfs send | zfs receive when I decided 
that I wanted to terminate the the zfs send process. I killed it, but the zfs 
receive doesn't want to die... In the meantime my zfs list command just hangs.

Here is the tail end of the truss output from a "truss zfs list":

ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484)  = 0
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C)  = 0
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) Err#3 ESRCH
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484)  = 0
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0
brk(0x082CA000) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C)  = 0
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) Err#3 ESRCH
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484)  = 0
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C)  = 0
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) Err#3 ESRCH
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484)  = 0
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08045788) Err#3 ESRCH
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) Err#3 ESRCH
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08046950) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C)  = 0
ioctl(3, ZFS_IOC_SNAPSHOT_LIST_NEXT, 0x08046950) Err#3 ESRCH
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08045814)  = 0
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08046950) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x0804464C)  = 0
ioctl(3, ZFS_IOC_DATASET_LIST_NEXT, 0x08045788) = 0
ioctl(3, ZFS_IOC_OBJSET_STATS, 0x08043484)  = 0

Let me know if something else is needed to figure out what is going on.

The system is an x4600 running Solaris 10, with a fairly recent patch set.

bash-3.00# uname -a
SunOS gdo-node-1 5.10 Generic_125101-07 i86pc i386 i86pc

bash-3.00# ps -ef | grep zfs
root 17593 17438   0   Jul 10 pts/3 605:57 zfs receive 
gdo-node-2/pcmdi/ftp/pub-backup
root 25972 18474   0 17:04:03 pts/4   0:00 truss zfs list
root 25968 18474   0 17:03:12 pts/4   0:08 zfs list
root 26086 17438   0 17:31:55 pts/3   0:00 grep zfs
root 25926 23059   0 16:55:44 pts/5   0:12 zfs list
root 25973 25972   0 17:04:03 pts/4   0:08 zfs list


David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool status -v: machine readable format?

2007-07-03 Thread David Smith
I was wondering if anyone had a script to parse the "zpool status -v" output 
into a more machine readable format?

Thanks,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: How to get new ZFS Solaris 10 U3 features going from Solaris 10 U2

2006-12-18 Thread David Smith
Thank you to everyone that has replied.  It sounds like I have a few options 
with regards to upgrading or just waiting and patching the current environment. 
 

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How to get new ZFS Solaris 10 U3 features going from Solaris 10 U2

2006-12-15 Thread David Smith
We currently have a couple of servers at Solaris 10 U2, and we would like to 
get to Solaris 10 U3 for the new zfs features.  Can this be accomplished via 
patching, or do you have to do an upgrade from S10U2 to S10U3?  Also what about 
a system with Zones?  What is the best practice for upgrading a system with 
zones?

Thanks,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Re: Corrupted LUN in RAIDZ group -- How to repair?

2006-09-14 Thread David Smith
I have run zpool scrub again, and I now see checksum errors again.  Wouldn't 
the checksum errors gotten fixed  with the first zpool scrub?

Can anyone recommend actions I should do at this point?  

Thanks,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] How to NOT mount a ZFS storage pool/ZFS file system?

2006-09-12 Thread David Smith
I currently have a system which has two ZFS storage pools.  One of the pools is 
coming from a faulty piece of hardware.  I would like to bring up our server 
mounting the storage pool which is okay and NOT mounting the one with from the 
hardware with problems.   Is there a simple way to NOT mount one of my ZFS 
storage pools?

The system is currently down due to the disk issues from one of the above 
pools.  

Thanks,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Re: Corrupted LUN in RAIDZ group -- How to repair?

2006-09-10 Thread David Smith
James,

Thanks for the reply.

It looks like now the scrub has completed.  Should I now clear these warnings?

bash-3.00# zpool status -x
  pool: mypool
 state: ONLINE
status: One or more devices has experienced an unrecoverable error.  An
attempt was made to correct the error.  Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
using 'zpool clear' or replace the device with 'zpool replace'.
   see: http://www.sun.com/msg/ZFS-8000-9P
 scrub: scrub completed with 0 errors on Sun Sep 10 07:44:36 2006
config:

NAMESTATE READ WRITE CKSUM
mypoolONLINE   0 0 0
  raidz ONLINE   0 0 0
c10t600A0B800011730E66C544C5EBB8d0  ONLINE   0 0 0
c10t600A0B800011730E66CA44C5EBEAd0  ONLINE   0 0 0
c10t600A0B800011730E66CF44C5EC1Cd0  ONLINE   0 0 0
c10t600A0B800011730E66D444C5EC5Cd0  ONLINE   0 0 0
c10t600A0B800011730E66D944C5ECA0d0  ONLINE   0 0 0
c10t600A0B800011652EE5C144C5ECDFd0  ONLINE   0 0 0
c10t600A0B800011730E66E244C5ED2Cd0  ONLINE   0 0 0
c10t600A0B800011652EE5C644C5ED87d0  ONLINE   0 0 0
c10t600A0B800011730E66EB44C5EDD8d0  ONLINE   0 0 0
c10t600A0B800011652EE5CB44C5EE29d0  ONLINE   0 0 0
c10t600A0B800011730E66F444C5EE7Ed0  ONLINE   0 015
c10t600A0B800011652EE5D044C5EEC9d0  ONLINE   0 0 0
c10t600A0B800011730E66FD44C5EF1Ad0  ONLINE  50 0 0
c10t600A0B800011652EE5D544C5EF63d0  ONLINE   0 0 0
  raidz ONLINE   0 0 0
c10t600A0B800011652EE5B844C5EBCBd0  ONLINE   0 0 0
c10t600A0B800011652EE5BA44C5EBF5d0  ONLINE   0 0 0
c10t600A0B800011652EE5BC44C5EC2Dd0  ONLINE   0 0 0
c10t600A0B800011652EE5BE44C5EC6Bd0  ONLINE   0 0 0
c10t600A0B800011730E66DB44C5ECB4d0  ONLINE   0 0 0
c10t600A0B800011652EE5C344C5ECF9d0  ONLINE   0 0 0
c10t600A0B800011730E66E444C5ED5Ad0  ONLINE   0 0 0
c10t600A0B800011652EE5C844C5EDA1d0  ONLINE   0 0 0
c10t600A0B800011730E66ED44C5EDFAd0  ONLINE   0 0 0
c10t600A0B800011652EE5CD44C5EE47d0  ONLINE   0 0 0
c10t600A0B800011730E66F644C5EE96d0  ONLINE   0 014
c10t600A0B800011652EE5D244C5EEE7d0  ONLINE   0 0 0
c10t600A0B800011730E66FF44C5EF32d0  ONLINE  70 0 0
c10t600A0B800011652EE5D744C5EF7Fd0  ONLINE   0 0 0


David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Corrupted LUN in RAIDZ group -- How to repair?

2006-09-09 Thread David Smith
Background:  We have a ZFS pool setup from LUNS which are from a SAN connected 
StorageTek/Engenio Flexline 380 storage system.   Just this past Friday the 
storage environment went down causing the system to go down.

After looking at the storage environment, we had several volume groups which 
needed to be carefully put back together to prevent corruption.  Well, one of 
the volume groups and the volumes/LUNs coming from it got corrupted.  Since our 
ZFS pools is setup to only have a LUN from each volume group we basically ended 
up with a single disk loss in our RAIDZ group.  So I believe we should be able 
to recover from this.

My question is how to replace this disk (LUN).  Basically the LUN is again 
okay, but the data on the LUN is not.  

I have tried to do a zpool replace, but ZFS seems to know that the disk/lun is 
the same device.  Using a -f (force) didn't work either.  How does one replace 
a LUN with ZFS?

I'm currently doing a "scrub", but don't know if that will help.

I first just had read errors on a lun in the raidz group, but just tonight 
noticed that I now have a checksum error on another lun as well. (see zpool 
status command output below)

Below is a zpool status -x output.   Can anyone advise how to recover from 
this?  

# zpool status -x
  pool: mypool
 state: ONLINE
status: One or more devices has experienced an unrecoverable error.  An
attempt was made to correct the error.  Applications are unaffected.
action: Determine if the device needs to be replaced, and clear the errors
using 'zpool clear' or replace the device with 'zpool replace'.
   see: http://www.sun.com/msg/ZFS-8000-9P
 scrub: scrub in progress, 66.00% done, 10h45m to go
config:

NAMESTATE READ WRITE CKSUM
mypoolONLINE   0 0 0
  raidz ONLINE   0 0 0
c10t600A0B800011730E66C544C5EBB8d0  ONLINE   0 0 0
c10t600A0B800011730E66CA44C5EBEAd0  ONLINE   0 0 0
c10t600A0B800011730E66CF44C5EC1Cd0  ONLINE   0 0 0
c10t600A0B800011730E66D444C5EC5Cd0  ONLINE   0 0 0
c10t600A0B800011730E66D944C5ECA0d0  ONLINE   0 0 0
c10t600A0B800011652EE5C144C5ECDFd0  ONLINE   0 0 0
c10t600A0B800011730E66E244C5ED2Cd0  ONLINE   0 0 0
c10t600A0B800011652EE5C644C5ED87d0  ONLINE   0 0 0
c10t600A0B800011730E66EB44C5EDD8d0  ONLINE   0 0 0
c10t600A0B800011652EE5CB44C5EE29d0  ONLINE   0 0 0
c10t600A0B800011730E66F444C5EE7Ed0  ONLINE   0 0 9
c10t600A0B800011652EE5D044C5EEC9d0  ONLINE   0 0 0
c10t600A0B800011730E66FD44C5EF1Ad0  ONLINE  50 0 0
c10t600A0B800011652EE5D544C5EF63d0  ONLINE   0 0 0
  raidz ONLINE   0 0 0
c10t600A0B800011652EE5B844C5EBCBd0  ONLINE   0 0 0
c10t600A0B800011652EE5BA44C5EBF5d0  ONLINE   0 0 0
c10t600A0B800011652EE5BC44C5EC2Dd0  ONLINE   0 0 0
c10t600A0B800011652EE5BE44C5EC6Bd0  ONLINE   0 0 0
c10t600A0B800011730E66DB44C5ECB4d0  ONLINE   0 0 0
c10t600A0B800011652EE5C344C5ECF9d0  ONLINE   0 0 0
c10t600A0B800011730E66E444C5ED5Ad0  ONLINE   0 0 0
c10t600A0B800011652EE5C844C5EDA1d0  ONLINE   0 0 0
c10t600A0B800011730E66ED44C5EDFAd0  ONLINE   0 0 0
c10t600A0B800011652EE5CD44C5EE47d0  ONLINE   0 0 0
c10t600A0B800011730E66F644C5EE96d0  ONLINE   0 0 6
c10t600A0B800011652EE5D244C5EEE7d0  ONLINE   0 0 0
c10t600A0B800011730E66FF44C5EF32d0  ONLINE  70 0 0
c10t600A0B800011652EE5D744C5EF7Fd0  ONLINE   0 0 0

This system is at Solaris 10, U2.

Thank you,

David
 
 
This message posted from opensolaris.org
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss