> Update ...
> 
> iostat output during "zpool scrub"
> 
> extended device statistics
>                   
>   w/s   Mr/s   Mw/s wait actv  svc_t  %w  %b 
> 34         2.0  395.2    0.1    0.6  0.0 34.8   87.7
>   0 100 
> 35        21.0  312.2    1.2    2.9  0.0 26.0   78.0
>   0  79 
> 36        20.0    1.0    1.2    0.0  0.0  0.7   31.4
>   0  13 
> 37        20.0    1.0    1.0    0.0  0.0  0.7   35.1
>   0  21 
> sd34 is always at 100% ...


  pool: zpool1
 state: ONLINE
 scrub: scrub in progress, 0.13% done, 72h39m to go
config:

        NAME                                       STATE     READ WRITE CKSUM
        zpool1                                    ONLINE       0     0     0
          raidz                                    ONLINE       0     0     0
            c4t60001FE100118DB000091190724700C7d0  ONLINE       0     0     0
            c4t60001FE100118DB000091190724700C9d0  ONLINE       0     0     0
            c4t60001FE100118DB000091190724700CBd0  ONLINE       0     0     0
            c4t60001FE100118DB000091190724700CCd0  ONLINE       0     0     0

72hours?? isn't too much for 370GB of data?
 
 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to