Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-28 Thread Markus Kovero
 failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo, 
 spa-spa_meta_objset) == 0, file ../zdb.c, line 571, function dump_metaslab

 Is this something I should worry about?

 uname -a
 SunOS E55000 5.11 oi_148 i86pc i386 i86pc Solaris

I thought we were talking about solaris 11 express, not oi?
Anyway, no idea about how openindiana should work or not.

Yours
Markus Kovero
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-28 Thread Stephan Budach



Sync was disabled on the main pool and then let to inherrit to everything else. 
The  reason for disabled this in the first place was to fix bad NFS write 
performance (even with  Zil on an X25e SSD it was under 1MB/s).
I've also tried setting the logbias to throughput and latency but they both 
perform  around the same level.
Thanks
-Matt

I believe you're hitting bug 7000208: Space map trashing affects NFS write 
throughput. We also did, and it did impact iscsi as well.

If you have enough ram you can try enabling metaslab debug (which makes problem 
vanish);

# echo metaslab_debug/W1 | mdb -kw

And calculating amount of ram needed:


/usr/sbin/amd64/zdb -mmpoolname/tmp/zdb-mm.out

awk '/segments/ {s+=$2}END {printf(sum=%d\n,s)}' zdb_mm.out

93373117 sum of segments
16 VDEVs
116 metaslabs
1856 metaslabs in total

93373117/1856 = 50308 average number of segments per metaslab

50308*1856*64
5975785472

5975785472/1024/1024/1024
5.56

= 5.56 GB

Yours
Markus Kovero
Out of curiosity, I just tried the command on one of my zpools, that has 
an number of ZFS volumes and was presented the following:


root@solaris11c:/obelixData/9_Testkunde/01_Etat01# 
/usr/sbin/amd64/zdb -mm obelixData  /tmp/zdb-mm.out

WARNING: can't open objset for obelixData/15035_RWE
zdb: can't open 'obelixData': I/O error

So, actually one of the volumes seems to have a problem, of which I 
wasn't aware - but this volume seems to behave just normal and it 
doesn't seem to show any erratic behaviour.


Can anybody maybe shed some light on what might be wrong with that 
particular volume?


Thanks,
budy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-28 Thread Stephan Budach

Am 28.04.11 11:51, schrieb Markus Kovero:

failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo,
spa-spa_meta_objset) == 0, file ../zdb.c, line 571, function dump_metaslab
Is this something I should worry about?
uname -a
SunOS E55000 5.11 oi_148 i86pc i386 i86pc Solaris

I thought we were talking about solaris 11 express, not oi?
Anyway, no idea about how openindiana should work or not.

Yours
Markus Kovero
___
Maybe this hasn't anything to do with Sol11Expr vs. oi. I do get the 
same error on one of my Sol11Expr hosts, when I run it against a 50 TB 
zpool:


root@solaris11a:~# uname -a
SunOS solaris11a 5.11 snv_151a i86pc i386 i86pc

root@solaris11a:~# /usr/sbin/amd64/zdb -mm backupPool_01  /tmp/zdb-mm.out
Assertion failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo, 
spa-spa_meta_objset) == 0, file ../zdb.c, line 593, function dump_metaslab



Cheers,
budy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-28 Thread Victor Latushkin

On Apr 28, 2011, at 5:04 PM, Stephan Budach wrote:

 Am 28.04.11 11:51, schrieb Markus Kovero:
 failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo,
 spa-spa_meta_objset) == 0, file ../zdb.c, line 571, function dump_metaslab
 Is this something I should worry about?
 uname -a
 SunOS E55000 5.11 oi_148 i86pc i386 i86pc Solaris
 I thought we were talking about solaris 11 express, not oi?
 Anyway, no idea about how openindiana should work or not.
 
 Yours
 Markus Kovero
 ___
 Maybe this hasn't anything to do with Sol11Expr vs. oi. I do get the same 
 error on one of my Sol11Expr hosts, when I run it against a 50 TB zpool:
 
 root@solaris11a:~# uname -a
 SunOS solaris11a 5.11 snv_151a i86pc i386 i86pc
 
 root@solaris11a:~# /usr/sbin/amd64/zdb -mm backupPool_01  /tmp/zdb-mm.out
 Assertion failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo, 
 spa-spa_meta_objset) == 0, file ../zdb.c, line 593, function dump_metaslab

zdb is not intended to be run against changing pools/datasets.
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-28 Thread Stephan Budach

Am 28.04.11 15:16, schrieb Victor Latushkin:

On Apr 28, 2011, at 5:04 PM, Stephan Budach wrote:


Am 28.04.11 11:51, schrieb Markus Kovero:

failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo,
spa-spa_meta_objset) == 0, file ../zdb.c, line 571, function dump_metaslab
Is this something I should worry about?
uname -a
SunOS E55000 5.11 oi_148 i86pc i386 i86pc Solaris

I thought we were talking about solaris 11 express, not oi?
Anyway, no idea about how openindiana should work or not.

Yours
Markus Kovero
___

Maybe this hasn't anything to do with Sol11Expr vs. oi. I do get the same error 
on one of my Sol11Expr hosts, when I run it against a 50 TB zpool:

root@solaris11a:~# uname -a
SunOS solaris11a 5.11 snv_151a i86pc i386 i86pc

root@solaris11a:~# /usr/sbin/amd64/zdb -mm backupPool_01  /tmp/zdb-mm.out
Assertion failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo, 
spa-spa_meta_objset) == 0, file ../zdb.c, line 593, function dump_metaslab

zdb is not intended to be run against changing pools/datasets.
Well, that explains it then (and probably the I/O erorr I got from my 
other zpool as well). Thanks.


Cheers,
budy
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-27 Thread Andrew Gabriel

Matthew Anderson wrote:

Hi All,

I've run into a massive performance problem after upgrading to Solaris 11 
Express from oSol 134.

Previously the server was performing a batch write every 10-15 seconds and the 
client servers (connected via NFS and iSCSI) had very low wait times. Now I'm 
seeing constant writes to the array with a very low throughput and high wait 
times on the client servers. Zil is currently disabled.


How/Why?


 There is currently one failed disk that is being replaced shortly.

Is there any ZFS tunable to revert Solaris 11 back to the behaviour of oSol 134?
  


What does zfs get sync report?

--
Andrew Gabriel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-27 Thread Matthew Anderson
NAMEPROPERTY  VALUE SOURCE
MirrorPool  sync  disabled  local
MirrorPool/CCIT sync  disabled  local
MirrorPool/EX01 sync  disabled  inherited from MirrorPool
MirrorPool/EX02 sync  disabled  inherited from MirrorPool
MirrorPool/FileStore1   sync  disabled  inherited from MirrorPool


Sync was disabled on the main pool and then let to inherrit to everything else. 
The reason for disabled this in the first place was to fix bad NFS write 
performance (even with Zil on an X25e SSD it was under 1MB/s).
I've also tried setting the logbias to throughput and latency but they both 
perform around the same level.

Thanks
-Matt


-Original Message-
From: Andrew Gabriel [mailto:andrew.gabr...@oracle.com] 
Sent: Wednesday, 27 April 2011 3:41 PM
To: Matthew Anderson
Cc: 'zfs-discuss@opensolaris.org'
Subject: Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 
Express

Matthew Anderson wrote:
 Hi All,

 I've run into a massive performance problem after upgrading to Solaris 11 
 Express from oSol 134.

 Previously the server was performing a batch write every 10-15 seconds and 
 the client servers (connected via NFS and iSCSI) had very low wait times. Now 
 I'm seeing constant writes to the array with a very low throughput and high 
 wait times on the client servers. Zil is currently disabled.

How/Why?

  There is currently one failed disk that is being replaced shortly.

 Is there any ZFS tunable to revert Solaris 11 back to the behaviour of oSol 
 134?
   

What does zfs get sync report?

-- 
Andrew Gabriel
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-27 Thread Markus Kovero


 Sync was disabled on the main pool and then let to inherrit to everything 
 else. The  reason for disabled this in the first place was to fix bad NFS 
 write performance (even with  Zil on an X25e SSD it was under 1MB/s).
 I've also tried setting the logbias to throughput and latency but they both 
 perform  around the same level.

 Thanks
 -Matt

I believe you're hitting bug 7000208: Space map trashing affects NFS write 
throughput. We also did, and it did impact iscsi as well.

If you have enough ram you can try enabling metaslab debug (which makes problem 
vanish);

# echo metaslab_debug/W1 | mdb -kw

And calculating amount of ram needed:


/usr/sbin/amd64/zdb -mm poolname  /tmp/zdb-mm.out

awk '/segments/ {s+=$2}END {printf(sum=%d\n,s)}' zdb_mm.out

93373117 sum of segments
16 VDEVs
116 metaslabs
1856 metaslabs in total

93373117/1856 = 50308 average number of segments per metaslab

50308*1856*64
5975785472

5975785472/1024/1024/1024
5.56

= 5.56 GB

Yours
Markus Kovero
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-27 Thread Tomas Ögren
On 27 April, 2011 - Matthew Anderson sent me these 3,2K bytes:

 Hi All,
 
 I've run into a massive performance problem after upgrading to Solaris 11 
 Express from oSol 134.
 
 Previously the server was performing a batch write every 10-15 seconds and 
 the client servers (connected via NFS and iSCSI) had very low wait times. Now 
 I'm seeing constant writes to the array with a very low throughput and high 
 wait times on the client servers. Zil is currently disabled. There is 
 currently one failed disk that is being replaced shortly.
 
 Is there any ZFS tunable to revert Solaris 11 back to the behaviour of oSol 
 134?
 I attempted to remove Sol 11 and reinstall 134 but it keeps freezing during 
 install which is probably another issue entirely...
 
 IOstat output is below. When running iostat -v 2 that level is writes OP's 
 and throughput is very constant.
 
capacity operationsbandwidth
 poolalloc   free   read  write   read  write
 --  -  -  -  -  -  -
 MirrorPool  12.2T  4.11T153  4.63K  6.06M  33.6M
   mirror1.04T   325G 11416   400K  2.80M
 c7t0d0  -  -  5114   163K  2.80M
 c7t1d0  -  -  6114   237K  2.80M
   mirror1.04T   324G 10374   426K  2.79M
 c7t2d0  -  -  5108   190K  2.79M
 c7t3d0  -  -  5107   236K  2.79M
   mirror1.04T   324G 15425   537K  3.15M
 c7t4d0  -  -  7115   290K  3.15M
 c7t5d0  -  -  8116   247K  3.15M
   mirror1.04T   325G 13412   572K  3.00M
 c7t6d0  -  -  7115   313K  3.00M
 c7t7d0  -  -  6116   259K  3.00M
   mirror1.04T   324G 13381   580K  2.85M
 c7t8d0  -  -  7111   362K  2.85M
 c7t9d0  -  -  5111   219K  2.85M
   mirror1.04T   325G 15408   654K  3.10M
 c7t10d0  -  -  7122   336K  3.10M
 c7t11d0  -  -  7123   318K  3.10M
   mirror1.04T   325G 14461   681K  3.22M
 c7t12d0  -  -  8130   403K  3.22M
 c7t13d0  -  -  6132   278K  3.22M
   mirror 749G   643G  1279   140K  1.07M
 c4t14d0  -  -  0  0  0  0
 c7t15d0  -  -  1 83   140K  1.07M
   mirror1.05T   319G 18333   672K  2.74M
 c7t16d0  -  - 11 96   406K  2.74M
 c7t17d0  -  -  7 96   266K  2.74M
   mirror1.04T   323G 13353   540K  2.85M
 c7t18d0  -  -  7 98   279K  2.85M
 c7t19d0  -  -  6100   261K  2.85M
   mirror1.04T   324G 12459   543K  2.99M
 c7t20d0  -  -  7118   285K  2.99M
 c7t21d0  -  -  4119   258K  2.99M
   mirror1.04T   324G 11431   465K  3.04M
 c7t22d0  -  -  5116   195K  3.04M
 c7t23d0  -  -  6117   272K  3.04M
   c8t2d00  29.5G  0  0  0  0

Btw, this disk seems alone, unmirrored and a bit small..?

 cache   -  -  -  -  -  -
   c8t3d059.4G  3.88M113 64  6.51M  7.31M
   c8t1d059.5G48K 95 69  5.69M  8.08M
 
 
 Thanks
 -Matt
 ___
 zfs-discuss mailing list
 zfs-discuss@opensolaris.org
 http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


/Tomas
-- 
Tomas Ögren, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


Re: [zfs-discuss] No write coalescing after upgrade to Solaris 11 Express

2011-04-27 Thread zfs user

On 4/27/11 4:00 AM, Markus Kovero wrote:




Sync was disabled on the main pool and then let to inherrit to everything else. 
The  reason for disabled this in the first place was to fix bad NFS write 
performance (even with  Zil on an X25e SSD it was under 1MB/s).
I've also tried setting the logbias to throughput and latency but they both 
perform  around the same level.



Thanks
-Matt


I believe you're hitting bug 7000208: Space map trashing affects NFS write 
throughput. We also did, and it did impact iscsi as well.

If you have enough ram you can try enabling metaslab debug (which makes problem 
vanish);

# echo metaslab_debug/W1 | mdb -kw

And calculating amount of ram needed:


/usr/sbin/amd64/zdb -mmpoolname/tmp/zdb-mm.out


metaslab 65   offset  410   spacemap258   free   Assertion 
failed: space_map_load(sm, zfs_metaslab_ops, SM_FREE, smo, 
spa-spa_meta_objset) == 0, file ../zdb.c, line 571, function dump_metaslab


Is this something I should worry about?

uname -a
SunOS E55000 5.11 oi_148 i86pc i386 i86pc Solaris



awk '/segments/ {s+=$2}END {printf(sum=%d\n,s)}' zdb_mm.out

93373117 sum of segments
16 VDEVs
116 metaslabs
1856 metaslabs in total

93373117/1856 = 50308 average number of segments per metaslab

50308*1856*64
5975785472

5975785472/1024/1024/1024
5.56

= 5.56 GB

Yours
Markus Kovero
___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

___
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss