[zfs-discuss] Convert from rz2 to rz1

2010-03-11 Thread Lars-Gunnar Persson
Is it possible to convert a rz2 array to rz1 array?

I have a pool with to rz2 arrays. I would like to convert them to rz1. Would 
that be possible?

If not, is it ok to remove one disk from a rz2 array and just let the array 
keep running with one disk missing?

Regards,

Lars-Gunnar Persson
System Administrator

Nansen Environmental and Remote Sensing Center
Address: Thormøhlensgate 47, 5006 Bergen
Direct: +47 55 20 58 31, switchboard: +47 55 20 58 00, fax: +47 55 20 58 01
Internet: http://www.nersc.no, e-mail: lars-gunnar.pers...@nersc.no

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Nexsan SATABeast and ZFS

2009-03-11 Thread Lars-Gunnar Persson

I would like to go back to my question for a second:

I checked with my Nexsan supplier and they confirmed that access to  
every single disk in SATABeast is not possible. The smallest entities  
I can create on the SATABeast are RAID 0 or 1 arrays. With RAID 1 I'll  
loose too much disk space and I believe that leaves me with RAID 0 as  
the only reasonable option. But with this unsecure RAID format I'll  
need higher redundancy in the ZFS configuration. I think I'll go with  
the following configuration:


On the Nexsan SATABeast:
* 14 disks configured in 7 RAID arrays with RAID level 0 (each disk is  
1 TB which gives me a total of 14 TB raw disk space).

* Each RAID 0 array configured as one volume.

On the Sun Fire X4100 M2 with Solaris 10:
* Add all 7 volumes to one zpool configured in on raidz2 (gives me  
approx. 8,8 TB available disk space)


Any comments or suggestions?

Best regards, Lars-Gunnar Persson

On 11. mars. 2009, at 02.39, Bob Friesenhahn wrote:


On Tue, 10 Mar 2009, A Darren Dunham wrote:


What part isn't true?  ZFS has a independent checksum for the data
block.  But if the data block is spread over multiple disks, then  
each

of the disks have to be read to verify the checksum.


I interpreted what you said to imply that RAID6 type algorithms were  
being used to validate the data, rather than to correct wrong data.   
I agree that it is necessary to read a full ZFS block in order to  
use the ZFS block checksum.  I also agree that a raidz2 vdev has  
IOPS behavior which is similar to a single disk.



From what I understand, a raidz2 with a very large number of disks
won't use all of the disks to store one ZFS block.  There is a  
maximum number of disks in a stripe which can be supported by the  
ZFS block size.


--
Bob Friesenhahn
bfrie...@simple.dallas.tx.us, http://www.simplesystems.org/users/bfriesen/
GraphicsMagick Maintainer,http://www.GraphicsMagick.org/
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Nexsan SATABeast and ZFS

2009-03-10 Thread Lars-Gunnar Persson

How about this configuration?

On the Nexsan SATABeast, add all disk to one RAID 5 or 6 group. Then  
on the Nexsan define several smaller volumes and then add those  
volumes to a raidz2/raidz zpool?


Could that be an useful configuration? Maybe I'll loose too much space  
with double raid 5 or 6 configuration? What about performance?


Regards,

Lars-Gunnar Persson

On 10. mars. 2009, at 00.26, Kees Nuyt wrote:


On Mon, 9 Mar 2009 12:06:40 +0100, Lars-Gunnar Persson
lars-gunnar.pers...@nersc.no wrote:


1. On the external disk array, I not able to configure JBOD or RAID 0
or 1 with just one disk.


In some arrays it seems to be possible to configure separate
disks by offering the array just one disk in one slot at a
time, and, very important, leaving all other slots empty(!).

Repeat for as many disks as you have, seating each disk in
its own slot, and all other slots empty.

(ok, it's just hear-say, but it might be worth a try with
the first 4 disks or so).
--
 (  Kees Nuyt
 )
c[_]
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] Nexsan SATABeast and ZFS

2009-03-10 Thread Lars-Gunnar Persson
I realized that I'll loose too much disk space with the double raid  
configuration suggested below. Agree?


I've done some performance testing with raidz/raidz1 vs raidz2:

bash-3.00# zpool status -v raid5
  pool: raid5
 state: ONLINE
 scrub: none requested
config:

NAME   STATE READ  
WRITE CKSUM
raid5  ONLINE   0  
0 0
  raidz1   ONLINE   0  
0 0
c7t6000402001FC442C609DC5A3d0  ONLINE   0  
0 0
c7t6000402001FC442C609DCA4Ad0  ONLINE   0  
0 0
c7t6000402001FC442C609DCA22d0  ONLINE   0  
0 0
c7t6000402001FC442C609DCABFd0  ONLINE   0  
0 0
c7t6000402001FC442C609DCADBd0  ONLINE   0  
0 0
c7t6000402001FC442C609DCAF8d0  ONLINE   0  
0 0
c7t6000402001FC442C609F0291d0  ONLINE   0  
0 0


errors: No known data errors

bash-3.00# zpool list
NAMESIZEUSED   AVAILCAP  HEALTH ALTROOT
raid5  12.6T141K   12.6T 0%  ONLINE -

bash-3.00# df -h /raid5
Filesystem size   used  avail capacity  Mounted on
raid5   11T41K11T 1%/raid5

bash-3.00# echo zfs_nocacheflush/D | mdb -k
zfs_nocacheflush:
zfs_nocacheflush:   0
bash-3.00# ./filesync-1 /raid5 1
Time in seconds to create and unlink 1 files with O_DSYNC: 9.871197

bash-3.00# echo zfs_nocacheflush/W1 | mdb -kw
zfs_nocacheflush:   0   =   0x1
bash-3.00# ./filesync-1 /raid5 1
Time in seconds to create and unlink 1 files with O_DSYNC: 7.363303

Then I destroyed the raid5 pool and created a raid6 pool:

bash-3.00# zpool status -v raid6  pool: raid6 state: ONLINE scrub:  
none requested

config:

NAME   STATE READ  
WRITE CKSUM
raid6  ONLINE   0  
0 0
  raidz2   ONLINE   0  
0 0
c7t6000402001FC442C609DC5A3d0  ONLINE   0  
0 0
c7t6000402001FC442C609DCA4Ad0  ONLINE   0  
0 0
c7t6000402001FC442C609DCA22d0  ONLINE   0  
0 0
c7t6000402001FC442C609DCABFd0  ONLINE   0  
0 0
c7t6000402001FC442C609DCADBd0  ONLINE   0  
0 0
c7t6000402001FC442C609DCAF8d0  ONLINE   0  
0 0
c7t6000402001FC442C609F0291d0  ONLINE   0  
0 0


errors: No known data errors

bash-3.00# zpool list
NAMESIZEUSED   AVAILCAP  HEALTH ALTROOT
raid6  12.6T195K   12.6T 0%  ONLINE -

bash-3.00# df -h /raid6
Filesystem size   used  avail capacity  Mounted on
raid6  8.8T52K   8.8T 1%/raid6

bash-3.00# echo zfs_nocacheflush/D | mdb -k
zfs_nocacheflush:
zfs_nocacheflush:   0
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC: 9.879219

bash-3.00# echo zfs_nocacheflush/W1 | mdb -kw
zfs_nocacheflush:   0   =   0x1
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC: 7.560435

My conclusion on raidz1 vs raidz2 would be no difference in  
performance and big difference in disk space available.



On 10. mars. 2009, at 09.13, Lars-Gunnar Persson wrote:


How about this configuration?

On the Nexsan SATABeast, add all disk to one RAID 5 or 6 group. Then  
on the Nexsan define several smaller volumes and then add those  
volumes to a raidz2/raidz zpool?


Could that be an useful configuration? Maybe I'll loose too much  
space with double raid 5 or 6 configuration? What about performance?


Regards,

Lars-Gunnar Persson

On 10. mars. 2009, at 00.26, Kees Nuyt wrote:


On Mon, 9 Mar 2009 12:06:40 +0100, Lars-Gunnar Persson
lars-gunnar.pers...@nersc.no wrote:

1. On the external disk array, I not able to configure JBOD or  
RAID 0

or 1 with just one disk.


In some arrays it seems to be possible to configure separate
disks by offering the array just one disk in one slot at a
time, and, very important, leaving all other slots empty(!).

Repeat for as many disks as you have, seating each disk in
its own slot, and all other slots empty.

(ok, it's just hear-say, but it might be worth a try with
the first 4 disks or so).
--
(  Kees Nuyt
)
c[_]
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http

Re: [zfs-discuss] Nexsan SATABeast and ZFS

2009-03-10 Thread Lars-Gunnar Persson

Test 1:
bash-3.00# echo zfs_nocacheflush/D | mdb -k
zfs_nocacheflush:
zfs_nocacheflush:   0
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC:  
292.223081


bash-3.00# echo zfs_nocacheflush/W1 | mdb -kw
zfs_nocacheflush:   0   =   0x1
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC:  
288.099066


Test 2:
ash-3.00# echo zfs_nocacheflush/D | mdb -k
zfs_nocacheflush:
zfs_nocacheflush:   0
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC: 13.092332

bash-3.00# echo zfs_nocacheflush/W1 | mdb -kw
zfs_nocacheflush:   0   =   0x1
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC: 9.591622


Test 3:
bash-3.00# echo zfs_nocacheflush/D | mdb -k
zfs_nocacheflush:
zfs_nocacheflush:   0
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC: 9.879219

bash-3.00# echo zfs_nocacheflush/W1 | mdb -kw
zfs_nocacheflush:   0   =   0x1
bash-3.00# ./filesync-1 /raid6 1
Time in seconds to create and unlink 1 files with O_DSYNC: 7.560435


Thank you for your reply. If I make a raidz or a raidz2 on the Solaris  
box, then I get enough redundancy?


The Nexsan can't do block-level snapshots.

On 9. mars. 2009, at 18.27, Miles Nordin wrote:

lp == Lars-Gunnar Persson lars-gunnar.pers...@nersc.no  
writes:


   lp Ignore force unit access (FUA) bit: [X]

   lp Any thoughts about this?

run three tests

(1) write cache disabled

(2) write cache enabled, ignore FUA off

(3) write cache enabled, ignore FUA [X]

if all three are the same, either the test is broken or something is  
wrong.


If (2) and (3) are the same, then ZFS is working as expected.  run
with either (2) or (3).

If (1) and (2) are the same, then ZFS doesn't seem to have the
cache-flush changes implemented, or else they aren't sufficient for
your array.  You could look into it more, file a bug, something like
that.

If you're not interested in investigating more and filing a bug in the
last case, then you could just set (3), and do no testing at all.

It might be good to do some of the testing just to catch wtf cases
like, ``oh sorry, we didn't sell you a cache.'' or ``the battery is
dead?  how'd that happen?'' but there are so many places for wtf cases
maybe this isn't the one to worry about.

but I'm not sure why you are switching to one big single-disk vdev
with FC if you are having ZFS corruption problems.  I think everyone's
been saying you are more likely to have problems by using FC instead
of direct attached, and also by using FC without vdev redundancy (the
redundancy seems to work around some bugs).  At least the people
reporting lots of lost pools are the ones using FC and iSCSI, who lose
pools during target reboots or SAN outages.

I suppose if the NexSAN can do block-level snapshots, you could
snapshot exported copies of your pool from time to time, and roll back
to a snapshot if ZFS refuses to import your ``corrupt'' pool.  In that
way it could help you?



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] Nexsan SATABeast and ZFS

2009-03-09 Thread Lars-Gunnar Persson
I'm trying to implement a Nexsan SATABeast (an external disk array,  
read more: http://www.nexsan.com/satabeast.php, 14 disks available)  
with a Sun Fire X4100 M2 server running Solaris 10 u6 (connected via  
fiber) and have a couple of questions:


(My motivation for this is the corrupted ZFS volume discussion I had  
earlier with no result, and this time I'm trying to make a more robust  
implementation)


1. On the external disk array, I not able to configure JBOD or RAID 0  
or 1 with just one disk. I can't find any options for my Solaris  
server to access the disk directly so I have to configure some raids  
on the SATABeast. I was thinking of striping two disks in each raid  
and then add all 7 raids to one zpool as a zraid. The problem with  
this is if one disk breaks down, I'll loose one RAID 0 disk but maybe  
ZFS can handle this? Should I rather implement RAID5 disks one the  
SATABeast and then export them to the Solaris machine? 14 disks would  
give me 4 RAID5 volumes and 2 spare disks? I'll loose a lot of disk  
space. What about create larger RAID volumes on the SATABeast? Like 3  
RAID volumes with 5 disks in 2 RAIDS and 4 disks in one RAID? I'm  
really not sure what to choose ... At the moment I've striped two  
disks in one RAID volume.


2. After reading from the ZFS Evil Tuning Guide (http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide#Cache_Flushes 
) about cache flushes I checked the cache configuration on the  
SATABaest and I can change these settings:


System Admin
Configure Cache

Cache Configuration
Current write cache state: Enabled, FUA ignored - 495 MB
Manually override current write cache status: [ ] Force write cache to  
Disabled

Desired write cache state: [X] Enabled [ ] Disabled
Allow attached host to override write cache configuration: [ ]
Ignore force unit access (FUA) bit: [X]
Write cache streaming mode: [ ]
Cache optimization setting:
 [ ] Random access
 [X] Mixed sequential/random
 [ ] Sequential access

And from the help section:

Write cache will normally speed up host writes,  data is buffered in  
the RAID controllers memory when the installed disk drives are not  
ready to accept the write data. The RAID controller write cache  
memory is battery backed, this allows any unwritten array data to be  
kept intact during a power failure situation. When power is restored  
this battery backed data will be flushed out to the RAID array.


Current write cache state - This is the current state of the write  
cache that the RAID system is using.


Manually override current write cache status - This allows the write  
caching to be forced on or off by the user, this change will take  
effect immediately.


Desired write cache state - This is the state of the write cache the  
user wishes to have after boot up.


Allow attached host to override write cache configuration - This  
allows the host system software to issue commands to the RAID system  
via the host interface that will either turn off or on the write  
caching.


Ignore force unit access (FUA) bit - When the force unit access  
(FUA) bit is set by a host system on a per command basis data is  
written / read directly to / from the disks without using the  
onboard cache. This will incur a time overhead, but guarantees the  
data is on the media. Set this option to force the controller to  
ignore the FUA bit such that command execution times are more  
consistent.


Write cache streaming mode - When the write cache is configured in  
streaming mode (check box ticked), the system continuously flushes  
the cache (it runs empty). This provides maximum cache buffering to  
protect against raid system delays adversely affecting command  
response times to the host.
When the write cache operates in non-streaming mode (check box not  
ticked) the system runs with a full write cache to maximise cache  
hits and maximise random IO performance.


Cache optimization setting - The cache optimization setting adjusts  
the cache behaviour to maximize performance for the expected host I/ 
O pattern.


Note that the write cache will be flushed 5 seconds after the last  
host write. It is recommended that all host activity is stopped 30  
seconds before powering the system off.


Any thoughts about this?

Regards,

Lars-Gunnar Persson
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


[zfs-discuss] zpool vs df

2009-03-09 Thread Lars-Gunnar Persson
I've a interesting situation. I've created two pool now and one pool  
named Data and another named raid5. Check the details here:


bash-3.00# zpool list
NAMESIZEUSED   AVAILCAP  HEALTH ALTROOT
Data   10.7T   9.82T892G91%  ONLINE -
raid5  10.9T145K   10.9T 0%  ONLINE -

As you see, the sizes are approximately the same. If I run the df  
command, it reports:


bash-3.00# df -h /Data
Filesystem size   used  avail capacity  Mounted on
Data11T   108M   154G 1%/Data
bash-3.00# df -h /raid5
Filesystem size   used  avail capacity  Mounted on
raid5  8.9T40K   8.9T 1%/raid5

You see that the Data has 11 TB when zpool reported 10.7 TB and the  
raid5 has 10.9TB in zpool but only 8.9 TB when using df. Thats a  
difference of 2 TB. Where did they go?


Any explanation would be find.

Regards,

Lars-Gunnar Persson
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] zpool vs df

2009-03-09 Thread Lars-Gunnar Persson

Here is what zpool status reports:

bash-3.00# zpool status
  pool: Data
 state: ONLINE
 scrub: none requested
config:

NAME STATE READ WRITE CKSUM
Data ONLINE   0 0 0
  c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors

  pool: raid5
 state: ONLINE
 scrub: none requested
config:

NAME   STATE READ  
WRITE CKSUM
raid5  ONLINE   0  
0 0
  raidz1   ONLINE   0  
0 0
c7t6000402001FC442C609DCA22d0  ONLINE   0  
0 0
c7t6000402001FC442C609DCA4Ad0  ONLINE   0  
0 0
c7t6000402001FC442C609DCAA2d0  ONLINE   0  
0 0
c7t6000402001FC442C609DCABFd0  ONLINE   0  
0 0
c7t6000402001FC442C609DCADBd0  ONLINE   0  
0 0
c7t6000402001FC442C609DCAF8d0  ONLINE   0  
0 0


errors: No known data errors


On 9. mars. 2009, at 14.29, Tomas Ögren wrote:


On 09 March, 2009 - Lars-Gunnar Persson sent me these 1,1K bytes:


I've a interesting situation. I've created two pool now and one pool
named Data and another named raid5. Check the details here:

bash-3.00# zpool list
NAMESIZEUSED   AVAILCAP  HEALTH  
ALTROOT

Data   10.7T   9.82T892G91%  ONLINE -
raid5  10.9T145K   10.9T 0%  ONLINE -

As you see, the sizes are approximately the same. If I run the df
command, it reports:

bash-3.00# df -h /Data
Filesystem size   used  avail capacity  Mounted on
Data11T   108M   154G 1%/Data
bash-3.00# df -h /raid5
Filesystem size   used  avail capacity  Mounted on
raid5  8.9T40K   8.9T 1%/raid5

You see that the Data has 11 TB when zpool reported 10.7 TB and the
raid5 has 10.9TB in zpool but only 8.9 TB when using df. Thats a
difference of 2 TB. Where did they go?


To your raid5 (raidz) parity.

Check 'zpool status' to see how your two pools differ.. zpool list  
shows

the disk space you have.. zfs/df shows how much you can store there..

/Tomas
--
Tomas Ögren, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss



.--.
|Lars-Gunnar  
Persson   |
|IT- 
sjef   |
| 
  |
|Nansen senteret for miljø og  
fjernmåling  |
|Adresse  : Thormøhlensgate 47, 5006  
Bergen|
|Direkte  : 55 20 58 31, sentralbord: 55 20 58 00, fax: 55 20 58  
01|
|Internett: http://www.nersc.no, e-post: lars- 
gunnar.pers...@nersc.no  |

'--'

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS volume corrupted?

2009-03-03 Thread Lars-Gunnar Persson
I thought a ZFS file system wouldn't destroy a ZFS volume? Hmm, I'm  
not sure what to do now ...


First of all, this zfs volume Data/subversion1 has been working for a  
year and suddenly after a reboot of the Solaris server, running of the  
zpool export and zpool import command, I get  problems with this ZFS  
volume?


Today I checked some more, after reading this guide: 
http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide

My main question is: Is my ZFS volume which is part of a zpool lost or  
can I recover it?


If I upgrade the Solaris server to the latest and do a zpool export  
and zpool import help?


All advices appreciated :-)

Here is some more information:

-bash-3.00$ zfs list -o  
name,type,used,avail,ratio,compression,reserv,volsize Data/subversion1

NAMETYPE   USED  AVAIL  RATIO  COMPRESS  RESERV  VOLSIZE
Data/subversion1  volume  22.5K   511G  1.00x   off250G 250G

I've also learned the the AVAIL column reports what's available in the  
zpool and NOT what's available in the ZFS volume.


-bash-3.00$ sudo zpool status -v
Password:
  pool: Data
 state: ONLINE
 scrub: scrub in progress, 5.86% done, 12h46m to go
config:

NAME STATE READ WRITE CKSUM
Data ONLINE   0 0 0
  c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors

Interesting thing here is that the scrub process should be finished  
today but the progress is much slower than reported here. And will the  
scrub process help anything in my case?



-bash-3.00$ sudo fmdump
TIME UUID SUNW-MSG-ID
Nov 15 2007 10:16:38 8aa789d2-7f3a-45d5-9f5c-c101d73b795e ZFS-8000-CS
Oct 14 09:31:40.8179 8c7d9847-94b7-ec09-8da7-c352de405b78 FMD-8000-2K

bash-3.00$ sudo fmdump -ev
TIME CLASS ENA
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e688d11500401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68926e600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68a3d3900401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68bc6741
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68d8bb600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68e98191
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e690a11000401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e69038551
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e69038551
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e690a11000401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e692a4ca1
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68bc6741
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e690a11000401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68bc6741
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e69038551
Nov 15 2007 09:33:52 ereport.fs.zfs.data
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68a3d3900401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68a3d3900401
Nov 15 2007 10:16:12 ereport.fs.zfs.vdev.open_failed
0x0533bb1b56400401
Nov 15 2007 10:16:12 ereport.fs.zfs.zpool   
0x0533bb1b56400401
Oct 14 09:31:31.6092 ereport.fm.fmd.log_append  
0x02eb96a8b6502801
Oct 14 09:31:31.8643 ereport.fm.fmd.mod_init
0x02ec89eadd100401



On 3. mars. 2009, at 08.10, Lars-Gunnar Persson wrote:


I've turned off iSCSI sharing at the moment.

My first question is: how can zfs report available is larger than  
reservation on a zfs volume? I also know that used mshould be larger  
than 22.5 K. Isn't

Re: [zfs-discuss] ZFS volume corrupted?

2009-03-03 Thread Lars-Gunnar Persson
Thank you for your long reply. I don't believe that will help me get  
my ZFS volume back though,


From my last reply to this list I confirm that I do understand what  
the AVAIL column is reporting when running the zfs list command.


hmm, still confused ...

Regards,

Lars-Gunnar Persson

On 3. mars. 2009, at 11.26, O'Shea, Damien wrote:



Hi,

The reason zfs is saying that the available is larger is because in  
Zfs the
size of the pool is always available to the all the zfs filesystems  
that
reside in the pool. Setting a reservation will gaurntee that the  
reservation
size is reserved for the filesystem/volume but you can change that  
on the

fly.

You can see that if you create another filsystem within the pool  
that the
reservation in use by your volume will have be deducted from the  
available

size.

Like below:



r...@testfs create -V 10g testpool/test
r...@testfs get all testpool
NAME  PROPERTY VALUE  SOURCE
testpool  type filesystem -
testpool  creation Wed Feb 11 13:17 2009  -
testpool  used 10.1G  -
testpool  available124G   -
testpool  referenced   100M   -
testpool  compressratio1.00x  -
testpool  mounted  yes-

Here the available is 124g as the volume has been set to 10g from a  
pool of

134g. If we set a reservation like this

r...@test1 set reservation=10g testpool/test
r...@test1 zfs get all testpool/test
NAME   PROPERTY VALUE  SOURCE
testpool/test  type volume -
testpool/test  creation Tue Mar  3 10:13 2009  -
testpool/test  used 10G-
testpool/test  available134G   -
testpool/test  referenced   16K-
testpool/test  compressratio1.00x  -

We can see that the available is now 134g, which is the avilable  
size of the
rest of the pool + the 10g reservation that we have set. So in  
theory this

volume can grow to the complete size of the pool.

So if we have a look at the availble space now in the pool we see

r...@test1# zfs get all testpool
NAME  PROPERTY VALUE  SOURCE
testpool  type filesystem -
testpool  creation Wed Feb 11 13:17 2009  -
testpool  used 10.1G  -
testpool  available124G   -
testpool  referenced   100M   -
testpool  compressratio1.00x  -
testpool  mounted  yes-

124g with 10g used to account for the size of the volume !

So if we now create another filesystem like this

r...@test1# zfs create testpool/test3
r...@test1# zfs get all testpool/test3
NAMEPROPERTY VALUE  SOURCE
testpool/test3  type filesystem -
testpool/test3  creation Tue Mar  3 10:19 2009  -
testpool/test3  used 18K-
testpool/test3  available124G   -
testpool/test3  referenced   18K-
testpool/test3  compressratio1.00x  -
testpool/test3  mounted  yes-

We see that the total amount available to the filesystem is the  
amount of the
space in the pool minus the 10g reservation. Lets set the  
reservation to

something bigger.

r...@test1# zfs set volsize=100g testpool/test
r...@test1# zfs set reservation=100g testpool/test
r...@test1# zfs get all testpool/test
NAME   PROPERTY VALUE  SOURCE
testpool/test  type volume -
testpool/test  creation Tue Mar  3 10:13 2009  -
testpool/test  used 100G   -
testpool/test  available134G   -
testpool/test  referenced   16K-

So the available is still 134G, which is the rest of the pool + the
reservation set.

r...@test1# zfs get all testpool
NAME  PROPERTY VALUE  SOURCE
testpool  type filesystem -
testpool  creation Wed Feb 11 13:17 2009  -
testpool  used 100G   -
testpool  available33.8G  -
testpool  referenced   100M   -
testpool  compressratio1.00x  -
testpool  mounted  yes-

The pool however now only has 33.8G left, which should be the same  
for all

the other filesystems in the pool.

Hope that helps.




-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org]on Behalf Of Lars-Gunnar
Persson
Sent: 03 March 2009 07:11
To: Richard Elling
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] ZFS volume corrupted?


*

This e-mail has been

Re: [zfs-discuss] ZFS volume corrupted?

2009-03-03 Thread Lars-Gunnar Persson

I run a new command now zdb. Here is the current output:

-bash-3.00$ sudo zdb Data
version=4
name='Data'
state=0
txg=9806565
pool_guid=6808539022472427249
vdev_tree
type='root'
id=0
guid=6808539022472427249
children[0]
type='disk'
id=0
guid=2167768931511572294
path='/dev/dsk/c4t5000402001FC442Cd0s0'
devid='id1,s...@n6000402001fc442c6e1a0e97/a'
whole_disk=1
metaslab_array=14
metaslab_shift=36
ashift=9
asize=11801587875840
Uberblock

magic = 00bab10c
version = 4
txg = 9842225
guid_sum = 8976307953983999543
timestamp = 1236084668 UTC = Tue Mar  3 13:51:08 2009

Dataset mos [META], ID 0, cr_txg 4, 392M, 1213 objects
... [snip]

Dataset Data/subversion1 [ZVOL], ID 3527, cr_txg 2514080, 22.5K, 3  
objects


... [snip]
Dataset Data [ZPL], ID 5, cr_txg 4, 108M, 2898 objects

Traversing all blocks to verify checksums and verify nothing leaked ...

and I'm still waiting for this process to finish.


On 3. mars. 2009, at 11.18, Lars-Gunnar Persson wrote:

I thought a ZFS file system wouldn't destroy a ZFS volume? Hmm, I'm  
not sure what to do now ...


First of all, this zfs volume Data/subversion1 has been working for  
a year and suddenly after a reboot of the Solaris server, running of  
the zpool export and zpool import command, I get  problems with this  
ZFS volume?


Today I checked some more, after reading this guide: 
http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide

My main question is: Is my ZFS volume which is part of a zpool lost  
or can I recover it?


If I upgrade the Solaris server to the latest and do a zpool export  
and zpool import help?


All advices appreciated :-)

Here is some more information:

-bash-3.00$ zfs list -o  
name,type,used,avail,ratio,compression,reserv,volsize Data/subversion1
NAMETYPE   USED  AVAIL  RATIO  COMPRESS  RESERV   
VOLSIZE
Data/subversion1  volume  22.5K   511G  1.00x   off250G  
250G


I've also learned the the AVAIL column reports what's available in  
the zpool and NOT what's available in the ZFS volume.


-bash-3.00$ sudo zpool status -v
Password:
 pool: Data
state: ONLINE
scrub: scrub in progress, 5.86% done, 12h46m to go
config:

   NAME STATE READ WRITE CKSUM
   Data ONLINE   0 0 0
 c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors

Interesting thing here is that the scrub process should be finished  
today but the progress is much slower than reported here. And will  
the scrub process help anything in my case?



-bash-3.00$ sudo fmdump
TIME UUID SUNW-MSG-ID
Nov 15 2007 10:16:38 8aa789d2-7f3a-45d5-9f5c-c101d73b795e ZFS-8000-CS
Oct 14 09:31:40.8179 8c7d9847-94b7-ec09-8da7-c352de405b78 FMD-8000-2K

bash-3.00$ sudo fmdump -ev
TIME CLASS ENA
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e688d11500401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68926e600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68a3d3900401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68bc6741
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68d8bb600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68e98191
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e690a11000401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e69038551
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e69038551
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e690a11000401
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e692a4ca1
Nov 15 2007 09:33:52 ereport.fs.zfs.io  
0x915e68bc6741

Re: [zfs-discuss] ZFS volume corrupted?

2009-03-03 Thread Lars-Gunnar Persson

And then the zdb process ends with:

Traversing all blocks to verify checksums and verify nothing leaked ...
out of memory -- generating core dump
Abort (core dumped)

hmm, what does that mean??


I also ran these commands:

-bash-3.00$ sudo fmstat
module ev_recv ev_acpt wait  svc_t  %w  %b  open solve   
memsz  bufsz
cpumem-retire0   0  0.00.1   0   0 0  
0  0  0
disk-transport   0   0  0.04.1   0   0 0 0 
32b  0
eft  0   0  0.05.7   0   0 0 0
1.4M  0
fmd-self-diagnosis   0   0  0.00.2   0   0 0  
0  0  0
io-retire0   0  0.00.2   0   0 0  
0  0  0
snmp-trapgen 0   0  0.00.1   0   0 0 0 
32b  0
sysevent-transport   0   0  0.0 1520.8   0   0 0  
0  0  0
syslog-msgs  0   0  0.00.1   0   0 0  
0  0  0
zfs-diagnosis  301   0  0.00.0   0   0 2 0
120b80b
zfs-retire   0   0  0.00.3   0   0 0  
0  0  0

-bash-3.00$ sudo fmadm config
MODULE   VERSION STATUS  DESCRIPTION
cpumem-retire1.1 active  CPU/Memory Retire Agent
disk-transport   1.0 active  Disk Transport Agent
eft  1.16active  eft diagnosis engine
fmd-self-diagnosis   1.0 active  Fault Manager Self-Diagnosis
io-retire1.0 active  I/O Retire Agent
snmp-trapgen 1.0 active  SNMP Trap Generation Agent
sysevent-transport   1.0 active  SysEvent Transport Agent
syslog-msgs  1.0 active  Syslog Messaging Agent
zfs-diagnosis1.0 active  ZFS Diagnosis Engine
zfs-retire   1.0 active  ZFS Retire Agent
-bash-3.00$ sudo zpool upgrade -v
This system is currently running ZFS version 4.

The following versions are supported:

VER  DESCRIPTION
---  
 1   Initial ZFS version
 2   Ditto blocks (replicated metadata)
 3   Hot spares and double parity RAID-Z
 4   zpool history

For more information on a particular version, including supported  
releases, see:


http://www.opensolaris.org/os/community/zfs/version/N

Where 'N' is the version number.


I hope I've provided enough information for all you ZFS experts out  
there.


Any tips or solutions in sight? Or is this ZFS gone completely?

Lars-Gunnar Persson


On 3. mars. 2009, at 13.58, Lars-Gunnar Persson wrote:


I run a new command now zdb. Here is the current output:

-bash-3.00$ sudo zdb Data
   version=4
   name='Data'
   state=0
   txg=9806565
   pool_guid=6808539022472427249
   vdev_tree
   type='root'
   id=0
   guid=6808539022472427249
   children[0]
   type='disk'
   id=0
   guid=2167768931511572294
   path='/dev/dsk/c4t5000402001FC442Cd0s0'
   devid='id1,s...@n6000402001fc442c6e1a0e97/a'
   whole_disk=1
   metaslab_array=14
   metaslab_shift=36
   ashift=9
   asize=11801587875840
Uberblock

   magic = 00bab10c
   version = 4
   txg = 9842225
   guid_sum = 8976307953983999543
   timestamp = 1236084668 UTC = Tue Mar  3 13:51:08 2009

Dataset mos [META], ID 0, cr_txg 4, 392M, 1213 objects
... [snip]

Dataset Data/subversion1 [ZVOL], ID 3527, cr_txg 2514080, 22.5K, 3  
objects


... [snip]
Dataset Data [ZPL], ID 5, cr_txg 4, 108M, 2898 objects

Traversing all blocks to verify checksums and verify nothing  
leaked ...


and I'm still waiting for this process to finish.


On 3. mars. 2009, at 11.18, Lars-Gunnar Persson wrote:

I thought a ZFS file system wouldn't destroy a ZFS volume? Hmm, I'm  
not sure what to do now ...


First of all, this zfs volume Data/subversion1 has been working for  
a year and suddenly after a reboot of the Solaris server, running  
of the zpool export and zpool import command, I get  problems with  
this ZFS volume?


Today I checked some more, after reading this guide: 
http://www.solarisinternals.com/wiki/index.php/ZFS_Troubleshooting_Guide

My main question is: Is my ZFS volume which is part of a zpool lost  
or can I recover it?


If I upgrade the Solaris server to the latest and do a zpool export  
and zpool import help?


All advices appreciated :-)

Here is some more information:

-bash-3.00$ zfs list -o  
name,type,used,avail,ratio,compression,reserv,volsize Data/ 
subversion1
NAMETYPE   USED  AVAIL  RATIO  COMPRESS  RESERV   
VOLSIZE
Data/subversion1  volume  22.5K   511G  1.00x   off250G  
250G


I've also learned the the AVAIL column reports what's available in  
the zpool and NOT what's available in the ZFS volume.


-bash-3.00$ sudo zpool status -v
Password:
pool: Data
state: ONLINE
scrub: scrub in progress, 5.86

Re: [zfs-discuss] ZFS volume corrupted?

2009-03-03 Thread Lars-Gunnar Persson


On 3. mars. 2009, at 14.51, Sanjeev wrote:

Thank you for your reply.


Lars-Gunnar,


On Tue, Mar 03, 2009 at 11:18:27AM +0100, Lars-Gunnar Persson wrote:

-bash-3.00$ zfs list -o
name,type,used,avail,ratio,compression,reserv,volsize Data/ 
subversion1
NAMETYPE   USED  AVAIL  RATIO  COMPRESS  RESERV   
VOLSIZE
Data/subversion1  volume  22.5K   511G  1.00x   off250G  
250G


This shows that the volume still exists.
Correct me if I am wrong here :
- Did you mean that the contents of the volume subversion1 are  
corrupted ?
I'm not 100% sure if it's the content of this volume or if it's the  
zpool that is corrupted. It was iSCSI exported to a Linux host where  
it was formatted as an ext3 file system.




What does that volume have on it ? Does it contain a filesystem  
which can
can be mounted on Solaris ? If so, we could try mounting it locally  
on the

Solaris box. This is to rule out any iSCSI issues.

I don't think that Solaris supports mounting of ext3 file systems or ?


Also, do you have any snapshots of the volume ? If so, you could  
rollback
to the latest snapshot. But, that would mean we lose some amount of  
data.
Nope, No snapshots - since this is a subversion repository with  
versioning built in. I didn't think I'll end up in this situation.




Also, you mentioned that the volume was in use for a year. But, I  
see in the
above output that it has only about 22.5K used. Is that correct ? I  
would

have expected it to be higher.
You're absolutely right, the 22.5K is wrong. That is why I suspect zfs  
is doing something wrong ...




You should also check what 'zpool history -i ' says.


it says:

-bash-3.00$ sudo zpool history Data | grep subversion
2008-04-02.09:08:53 zfs create -V 250GB Data/subversion1
2008-04-02.09:08:53 zfs set shareiscsi=on Data/subversion1
2008-08-14.14:13:58 zfs set shareiscsi=off Data/subversion1
2008-08-29.15:08:50 zfs set shareiscsi=on Data/subversion1
2009-03-02.10:37:36 zfs set shareiscsi=off Data/subversion1
2009-03-02.10:37:55 zfs set shareiscsi=on Data/subversion1
2009-03-02.11:37:22 zfs set shareiscsi=off Data/subversion1
2009-03-03.09:37:34 zfs set shareiscsi=on Data/subversion1

and:

2009-03-01.11:26:22 zpool export -f Data
2009-03-01.13:21:58 zpool import Data

2009-03-01.14:32:04 zpool scrub Data




Thanks and regards,
Sanjeev


More info:
I just rebooted the SOlaris server and no change in status:

-bash-3.00$ zpool status -v
  pool: Data
 state: ONLINE
 scrub: none requested
config:

NAME STATE READ WRITE CKSUM
Data ONLINE   0 0 0
  c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors


The scrubing has stopped and the zdb command crashed the server.





I've also learned the the AVAIL column reports what's available in  
the

zpool and NOT what's available in the ZFS volume.

-bash-3.00$ sudo zpool status -v
Password:
 pool: Data
state: ONLINE
scrub: scrub in progress, 5.86% done, 12h46m to go
config:

   NAME STATE READ WRITE CKSUM
   Data ONLINE   0 0 0
 c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors

Interesting thing here is that the scrub process should be finished
today but the progress is much slower than reported here. And will  
the

scrub process help anything in my case?


-bash-3.00$ sudo fmdump
TIME UUID SUNW-MSG-ID
Nov 15 2007 10:16:38 8aa789d2-7f3a-45d5-9f5c-c101d73b795e ZFS-8000-CS
Oct 14 09:31:40.8179 8c7d9847-94b7-ec09-8da7-c352de405b78 FMD-8000-2K

bash-3.00$ sudo fmdump -ev
TIME CLASS ENA
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e6850ff400401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e688d11500401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68926e600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68a3d3900401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68bc6741
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68d8bb600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68da5b51
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e6897db600401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68e98191
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e690a11000401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e68f0c9800401
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e69038551
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e69038551
Nov 15 2007 09:33:52 ereport.fs.zfs.io
0x915e690a11000401
Nov 15 2007

[zfs-discuss] ZFS volume corrupted?

2009-03-02 Thread Lars-Gunnar Persson

Hey to everyone on this mailing list (since this is my first post)!

We've a Sun Fire X4100 M2 server running Solaris 10 u6 and after some  
system work this weekend we have a problem with only one ZFS volume.


We have a pool called /Data with many file systems and two volumes.  
The status of my zpool is:


-bash-3.00$ zpool status
  pool: Data
 state: ONLINE
 scrub: scrub in progress, 5.99% done, 13h38m to go
config:

NAME STATE READ WRITE CKSUM
Data ONLINE   0 0 0
  c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors


Yesterday I started the scrub process because I read that was a smart  
thing to do after a zpool export and zpool import procedure. I did  
this because I wanted to move the zpool to another OS installation but  
changed my mind and did a zpool import on the same OS as I did an  
export.


After checking as much information as I could find on the web, I was  
advised to to run the zpool scrub after an import.


Well, the problem now is that one volume in this zpool is not working.  
I've shared it via iscsi to a Linux host (all of this was working on  
Friday). The Linux host reports that it can't find a partition table.  
Here is the log from the Linux host:


Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr  
sectors (268435 MB)

Mar  2 11:09:36 eva kernel: SCSI device sdb: drive cache: write through
Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr  
sectors (268435 MB)

Mar  2 11:09:37 eva kernel: SCSI device sdb: drive cache: write through
Mar  2 11:09:37 eva kernel:  sdb: unknown partition table
Mar  2 11:09:37 eva kernel: Attached scsi disk sdb at scsi28, channel  
0, id 0, lun 0



So I checked the status on my Solaris server and I found this  
information a bit strange;:


-bash-3.00$ zfs list Data/subversion1
NAME   USED  AVAIL  REFER  MOUNTPOINT
Data/subversion1  22.5K   519G  22.5K  -

How  can it bed 519GB available on a volume that is 250GB in size?  
Here are more details:


-bash-3.00$ zfs get all Data/subversion1
NAME  PROPERTY   VALUE  SOURCE
Data/subversion1  type   volume -
Data/subversion1  creation   Wed Apr  2  9:06 2008  -
Data/subversion1  used   22.5K  -
Data/subversion1  available  519G   -
Data/subversion1  referenced 22.5K  -
Data/subversion1  compressratio  1.00x  -
Data/subversion1  reservation250G   local
Data/subversion1  volsize250G   -
Data/subversion1  volblocksize   8K -
Data/subversion1  checksum   on default
Data/subversion1  compressionoffdefault
Data/subversion1  readonly   offdefault
Data/subversion1  shareiscsi offlocal


Will this be fixed after the scrub process is finished tomorrow or is  
this volume lost forever?


Hoping for some quick answers as the data is quite important for us.

Regards,

Lars-Gunnar Persson

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS volume corrupted?

2009-03-02 Thread Lars-Gunnar Persson
That is correct. It's a raid 6 disk shelf with one volume connected  
via fibre.


Lars-Gunnar Persson

Den 2. mars. 2009 kl. 16.57 skrev Blake blake.ir...@gmail.com:

It looks like you only have one physical device in this pool.  Is  
that correct?




On Mon, Mar 2, 2009 at 9:01 AM, Lars-Gunnar Persson
lars-gunnar.pers...@nersc.no wrote:

Hey to everyone on this mailing list (since this is my first post)!

We've a Sun Fire X4100 M2 server running Solaris 10 u6 and after  
some system

work this weekend we have a problem with only one ZFS volume.

We have a pool called /Data with many file systems and two volumes.  
The

status of my zpool is:

-bash-3.00$ zpool status
 pool: Data
 state: ONLINE
 scrub: scrub in progress, 5.99% done, 13h38m to go
config:

   NAME STATE READ WRITE CKSUM
   Data ONLINE   0 0 0
 c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors


Yesterday I started the scrub process because I read that was a  
smart thing
to do after a zpool export and zpool import procedure. I did this  
because I
wanted to move the zpool to another OS installation but changed my  
mind and

did a zpool import on the same OS as I did an export.

After checking as much information as I could find on the web, I  
was advised

to to run the zpool scrub after an import.

Well, the problem now is that one volume in this zpool is not  
working. I've
shared it via iscsi to a Linux host (all of this was working on  
Friday). The
Linux host reports that it can't find a partition table. Here is  
the log

from the Linux host:

Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte  
hdwr sectors

(268435 MB)
Mar  2 11:09:36 eva kernel: SCSI device sdb: drive cache: write  
through
Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte  
hdwr sectors

(268435 MB)
Mar  2 11:09:37 eva kernel: SCSI device sdb: drive cache: write  
through

Mar  2 11:09:37 eva kernel:  sdb: unknown partition table
Mar  2 11:09:37 eva kernel: Attached scsi disk sdb at scsi28,  
channel 0, id

0, lun 0


So I checked the status on my Solaris server and I found this  
information a

bit strange;:

-bash-3.00$ zfs list Data/subversion1
NAME   USED  AVAIL  REFER  MOUNTPOINT
Data/subversion1  22.5K   519G  22.5K  -

How  can it bed 519GB available on a volume that is 250GB in size?  
Here are

more details:

-bash-3.00$ zfs get all Data/subversion1
NAME  PROPERTY   VALUE  SOURCE
Data/subversion1  type   volume -
Data/subversion1  creation   Wed Apr  2  9:06 2008  -
Data/subversion1  used   22.5K  -
Data/subversion1  available  519G   -
Data/subversion1  referenced 22.5K  -
Data/subversion1  compressratio  1.00x  -
Data/subversion1  reservation250G   local
Data/subversion1  volsize250G   -
Data/subversion1  volblocksize   8K -
Data/subversion1  checksum   on default
Data/subversion1  compressionoffdefault
Data/subversion1  readonly   offdefault
Data/subversion1  shareiscsi offlocal


Will this be fixed after the scrub process is finished tomorrow or  
is this

volume lost forever?

Hoping for some quick answers as the data is quite important for us.

Regards,

Lars-Gunnar Persson

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss




___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] ZFS volume corrupted?

2009-03-02 Thread Lars-Gunnar Persson
The Linux host can still see the device. I showed you the log from the  
Linux host.


I tried the fdisk -l and it listed the iSCSI disks.

Lars-Gunnar Persson

Den 2. mars. 2009 kl. 17.02 skrev O'Shea, Damien daos...@revenue.ie:



I could be wrong but this looks like an issue on the Linux side

A zpool status is returning the healthy pool

What does format/fdisk show you on the Linux side ? Can it still see  
the

iSCSI device that is being shared from the Solaris server ?



Regards,
Damien O'Shea
Strategy  Unix Systems
Revenue Backup Site
VPN: 35603
daos...@revenue.ie mailto:daos...@revenue.ie


-Original Message-
From: zfs-discuss-boun...@opensolaris.org
[mailto:zfs-discuss-boun...@opensolaris.org]on Behalf Of Blake
Sent: 02 March 2009 15:57
To: Lars-Gunnar Persson
Cc: zfs-discuss@opensolaris.org
Subject: Re: [zfs-discuss] ZFS volume corrupted?


*

This e-mail has been received by the Revenue Internet e-mail  
service. (IP)


*

It looks like you only have one physical device in this pool.  Is that
correct?



On Mon, Mar 2, 2009 at 9:01 AM, Lars-Gunnar Persson
lars-gunnar.pers...@nersc.no wrote:

Hey to everyone on this mailing list (since this is my first post)!

We've a Sun Fire X4100 M2 server running Solaris 10 u6 and after some

system

work this weekend we have a problem with only one ZFS volume.

We have a pool called /Data with many file systems and two volumes.  
The

status of my zpool is:

-bash-3.00$ zpool status
 pool: Data
 state: ONLINE
 scrub: scrub in progress, 5.99% done, 13h38m to go
config:

   NAME STATE READ WRITE CKSUM
   Data ONLINE   0 0 0
 c4t5000402001FC442Cd0  ONLINE   0 0 0

errors: No known data errors


Yesterday I started the scrub process because I read that was a  
smart thing
to do after a zpool export and zpool import procedure. I did this  
because I
wanted to move the zpool to another OS installation but changed my  
mind and

did a zpool import on the same OS as I did an export.

After checking as much information as I could find on the web, I was

advised

to to run the zpool scrub after an import.

Well, the problem now is that one volume in this zpool is not  
working. I've
shared it via iscsi to a Linux host (all of this was working on  
Friday).

The
Linux host reports that it can't find a partition table. Here is  
the log

from the Linux host:

Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr

sectors

(268435 MB)
Mar  2 11:09:36 eva kernel: SCSI device sdb: drive cache: write  
through

Mar  2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr

sectors

(268435 MB)
Mar  2 11:09:37 eva kernel: SCSI device sdb: drive cache: write  
through

Mar  2 11:09:37 eva kernel:  sdb: unknown partition table
Mar  2 11:09:37 eva kernel: Attached scsi disk sdb at scsi28,  
channel 0, id

0, lun 0


So I checked the status on my Solaris server and I found this  
information a

bit strange;:

-bash-3.00$ zfs list Data/subversion1
NAME   USED  AVAIL  REFER  MOUNTPOINT
Data/subversion1  22.5K   519G  22.5K  -

How  can it bed 519GB available on a volume that is 250GB in size?  
Here are

more details:

-bash-3.00$ zfs get all Data/subversion1
NAME  PROPERTY   VALUE  SOURCE
Data/subversion1  type   volume -
Data/subversion1  creation   Wed Apr  2  9:06 2008  -
Data/subversion1  used   22.5K  -
Data/subversion1  available  519G   -
Data/subversion1  referenced 22.5K  -
Data/subversion1  compressratio  1.00x  -
Data/subversion1  reservation250G   local
Data/subversion1  volsize250G   -
Data/subversion1  volblocksize   8K -
Data/subversion1  checksum   on default
Data/subversion1  compressionoffdefault
Data/subversion1  readonly   offdefault
Data/subversion1  shareiscsi offlocal


Will this be fixed after the scrub process is finished tomorrow or  
is this

volume lost forever?

Hoping for some quick answers as the data is quite important for us.

Regards,

Lars-Gunnar Persson

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss




This message has been delivered to the Internet by the Revenue  
Internet e-mail service (OP)


*

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman

Re: [zfs-discuss] ZFS volume corrupted?

2009-03-02 Thread Lars-Gunnar Persson

I've turned off iSCSI sharing at the moment.

My first question is: how can zfs report available is larger than  
reservation on a zfs volume? I also know that used mshould be larger  
than 22.5 K. Isn't this strange?


Lars-Gunnar Persson

Den 3. mars. 2009 kl. 00.38 skrev Richard Elling richard.ell...@gmail.com 
:



Lars-Gunnar Persson wrote:

Hey to everyone on this mailing list (since this is my first post)!


Welcome!



We've a Sun Fire X4100 M2 server running Solaris 10 u6 and after  
some system work this weekend we have a problem with only one ZFS  
volume.


We have a pool called /Data with many file systems and two volumes.  
The status of my zpool is:


-bash-3.00$ zpool status
pool: Data
state: ONLINE
scrub: scrub in progress, 5.99% done, 13h38m to go
config:

NAME STATE READ WRITE CKSUM
Data ONLINE 0 0 0
c4t5000402001FC442Cd0 ONLINE 0 0 0

errors: No known data errors


Yesterday I started the scrub process because I read that was a  
smart thing to do after a zpool export and zpool import procedure.  
I did this because I wanted to move the zpool to another OS  
installation but changed my mind and did a zpool import on the same  
OS as I did an export.


After checking as much information as I could find on the web, I  
was advised to to run the zpool scrub after an import.


Well, the problem now is that one volume in this zpool is not  
working. I've shared it via iscsi to a Linux host (all of this was  
working on Friday). The Linux host reports that it can't find a  
partition table. Here is the log from the Linux host:


Mar 2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr  
sectors (268435 MB)
Mar 2 11:09:36 eva kernel: SCSI device sdb: drive cache: write  
through
Mar 2 11:09:36 eva kernel: SCSI device sdb: 524288000 512-byte hdwr  
sectors (268435 MB)
Mar 2 11:09:37 eva kernel: SCSI device sdb: drive cache: write  
through

Mar 2 11:09:37 eva kernel: sdb: unknown partition table
Mar 2 11:09:37 eva kernel: Attached scsi disk sdb at scsi28,  
channel 0, id 0, lun 0



So I checked the status on my Solaris server and I found this  
information a bit strange;:


-bash-3.00$ zfs list Data/subversion1
NAME USED AVAIL REFER MOUNTPOINT
Data/subversion1 22.5K 519G 22.5K -

How can it bed 519GB available on a volume that is 250GB in size?  
Here are more details:


-bash-3.00$ zfs get all Data/subversion1
NAME PROPERTY VALUE SOURCE
Data/subversion1 type volume -
Data/subversion1 creation Wed Apr 2 9:06 2008 -
Data/subversion1 used 22.5K -
Data/subversion1 available 519G -
Data/subversion1 referenced 22.5K -
Data/subversion1 compressratio 1.00x -
Data/subversion1 reservation 250G local
Data/subversion1 volsize 250G -
Data/subversion1 volblocksize 8K -
Data/subversion1 checksum on default
Data/subversion1 compression off default
Data/subversion1 readonly off default
Data/subversion1 shareiscsi off local


It does not appear that Data/subversion1 is being shared via iscsi?
-- richard



___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss