Good Morning Michelle. Am 20.01.11 08:05, schrieb Michelle Knight: > hi Folks, > > OI 148. > > Three 1.5tb drives were replaced with three 2tb drives. They are here on > internal SATA channels c2t2d0, c2t3d0 and c2t4d0. One is a Seagate Barracuda > and the other two are Western Digital Greens. > > mich@jaguar:~# cfgadm -lv > Ap_Id Receptacle Occupant Condition > Information > When Type Busy Phys_Id > Slot8 connected configured ok Location: > Slot8 > Jan 1 1970 unknown n /devices/pci@0,0/pci8086,3b4a@1c,4:Slot8 > sata0/0::dsk/c2t0d0 connected configured ok Mod: > INTEL > SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PH040GGN > unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:0 > sata0/1::dsk/c2t1d0 connected configured ok Mod: > INTEL > SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PC040GGN > unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:1 > sata0/2::dsk/c2t2d0 connected configured ok Mod: > ST32000542AS FRev: CC34 SN: 5XW17ARW > unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:2 > sata0/3::dsk/c2t3d0 connected configured ok Mod: WDC > WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0555575 > unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:3 > sata0/4::dsk/c2t4d0 connected configured ok Mod: WDC > WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0484508 > > A zpool export and subsequent import, which should have taken the set to 4tb > overall storage in the raidz, appears to have not worked despite the import > taking what must have been about ten to fifteen minutes to do the import. > (during which time the drives were silent and the zpool process was mostly 0% > very occasionally peaking to 25%, and the system being very slow to respond > during that period) > Personally i assume the peaks were triggerd by resilvering the pool. Its not uncommon that you have a high load, if your pool is resilvering.
Best Practice in this case would have been creating a new Zpool, e.g raidz.. zfs send from $oldpool zfs receive $newpool... :-( I have three Questions for you. First.. is this a production Box? Second.. could you provide the output of zpool history? Third, no offence, but do you have proper Literature for ZFS? If not please have a look at Solarisinternals http://www.solarisinternals.com/wiki/index.php/Solaris_Internals_and_Performance_FAQ http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide Or have a look at the Open Solaris Bible. Fourth... what would you like to achieve with OI ? Sorry now i made four Questions out of it. ;) Cheers :-) > Any ideas please? Or is there still some process running in the background > that I can't see? > > mich@jaguar:~# zfs list > NAME USED AVAIL REFER MOUNTPOINT > data 2.27T 401G 2.27T /mirror > rpool 7.69G 28.7G 45K /rpool > rpool/ROOT 3.70G 28.7G 31K legacy > rpool/ROOT/openindiana 3.70G 28.7G 3.59G / > rpool/dump 1.93G 28.7G 1.93G - > rpool/export 5.22M 28.7G 32K /export > rpool/export/home 5.19M 28.7G 32K /export/home > rpool/export/home/mich 5.16M 28.7G 5.16M /export/home/mich > rpool/swap 2.05G 30.7G 126M - > > > mich@jaguar:~# zpool status > pool: data > state: ONLINE > scan: resilvered 1.13T in 12h26m with 0 errors on Wed Jan 19 23:42:23 2011 > config: > > NAME STATE READ WRITE CKSUM > data ONLINE 0 0 0 > raidz1-0 ONLINE 0 0 0 > c2t2d0 ONLINE 0 0 0 > c2t3d0 ONLINE 0 0 0 > c2t4d0 ONLINE 0 0 0 > > errors: No known data errors > That took a very long time, for only 1 TB > > last pid: 1802; load avg: 0.61, 0.56, 0.61; up 0+19:55:20 > 07:06:01 > 74 processes: 73 sleeping, 1 on cpu > CPU states: 99.8% idle, 0.0% user, 0.3% kernel, 0.0% iowait, 0.0% swap > Kernel: 375 ctxsw, 653 intr, 120 syscall > Memory: 3959M phys mem, 401M free mem, 1979M total swap, 1979M free swap > > PID USERNAME NLWP PRI NICE SIZE RES STATE TIME CPU COMMAND > 1197 gdm 1 59 0 95M 28M sleep 1:23 0.04% gdm-simple-gree > 922 root 3 59 0 102M 51M sleep 0:37 0.02% Xorg > 1801 root 1 59 0 4036K 2460K cpu/3 0:00 0.01% top > 1196 gdm 1 59 0 80M 13M sleep 0:00 0.00% metacity > 1190 gdm 1 59 0 7892K 6028K sleep 0:00 0.00% at-spi-registry > 640 root 16 59 0 14M 9072K sleep 0:09 0.00% smbd > 1737 mich 1 59 0 13M 5392K sleep 0:00 0.00% sshd > 148 root 1 59 0 8312K 1608K sleep 0:00 0.00% dhcpagent > 672 root 26 59 0 27M 15M sleep 0:01 0.00% fmd > 1198 gdm 1 59 0 87M 18M sleep 0:02 0.00% gnome-power-man > 1247 root 1 59 0 6080K 2500K sleep 0:00 0.00% sendmail > 11 root 21 59 0 15M 13M sleep 0:06 0.00% svc.configd > 272 root 6 59 0 11M 4784K sleep 0:01 0.00% devfsadm > 1220 root 24 59 0 13M 4404K sleep 0:01 0.00% nscd > 45 netcfg 5 59 0 4716K 3268K sleep 0:00 0.00% netcfgd > 1309 admin 1 59 0 16M 8404K sleep 14:30 0.00% sshd > 1312 admin 1 59 0 16M 5468K sleep 3:47 0.00% sshd > 787 root 1 59 0 12M 5928K sleep 0:05 0.00% intrd > 1237 root 4 59 0 8400K 1936K sleep 0:03 0.00% automountd > 9 root 15 59 0 20M 12M sleep 0:02 0.00% svc.startd > 387 root 5 59 0 7500K 6024K sleep 0:01 0.00% hald > 1308 root 1 59 0 13M 4788K sleep 0:01 0.00% sshd > 252 root 5 60 -20 2544K 1460K sleep 0:00 0.00% zonestatd > 1191 gdm 1 59 0 111M 42M sleep 0:00 0.00% gnome-settings- > 1175 gdm 2 59 0 20M 10M sleep 0:00 0.00% gnome-session > 790 root 4 59 0 15M 6432K sleep 0:00 0.00% rad > 1188 gdm 1 59 0 12M 5944K sleep 0:00 0.00% gconfd-2 > 291 root 1 59 0 16M 5604K sleep 0:00 0.00% cupsd > 1156 mich 1 59 0 13M 5432K sleep 0:00 0.00% sshd > 132 daemon 3 59 0 13M 5104K sleep 0:00 0.00% kcfd > _______________________________________________ > OpenIndiana-discuss mailing list > OpenIndiana-discuss@openindiana.org > http://openindiana.org/mailman/listinfo/openindiana-discuss -- with kind regards _.-|-/\-._ \-' '-. / /\ /\ \/ \/ < . > ./. \/ _ / < > /___\ |. .< \ / < /\ > ( #) |#) | | < /\ -. __\ \ < < V > )./_._(\ .)/\ < < .- / \_'_) )-.. \ < ./ / > > /._./ /\ < '-' > > / '-._ < v > _.-' / '-.______.ยท' \ \/ *********************************************************** *This message has been scanned by DrWeb AV and Spamassassin _______________________________________________ OpenIndiana-discuss mailing list OpenIndiana-discuss@openindiana.org http://openindiana.org/mailman/listinfo/openindiana-discuss