hi Folks, OI 148.
Three 1.5tb drives were replaced with three 2tb drives. They are here on internal SATA channels c2t2d0, c2t3d0 and c2t4d0. One is a Seagate Barracuda and the other two are Western Digital Greens. mich@jaguar:~# cfgadm -lv Ap_Id Receptacle Occupant Condition Information When Type Busy Phys_Id Slot8 connected configured ok Location: Slot8 Jan 1 1970 unknown n /devices/pci@0,0/pci8086,3b4a@1c,4:Slot8 sata0/0::dsk/c2t0d0 connected configured ok Mod: INTEL SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PH040GGN unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:0 sata0/1::dsk/c2t1d0 connected configured ok Mod: INTEL SSDSA2M040G2GC FRev: 2CV102HB SN: CVGB949301PC040GGN unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:1 sata0/2::dsk/c2t2d0 connected configured ok Mod: ST32000542AS FRev: CC34 SN: 5XW17ARW unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:2 sata0/3::dsk/c2t3d0 connected configured ok Mod: WDC WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0555575 unavailable disk n /devices/pci@0,0/pci1458,b005@1f,2:3 sata0/4::dsk/c2t4d0 connected configured ok Mod: WDC WD20EARS-00MVWB0 FRev: 51.0AB51 SN: WD-WMAZA0484508 A zpool export and subsequent import, which should have taken the set to 4tb overall storage in the raidz, appears to have not worked despite the import taking what must have been about ten to fifteen minutes to do the import. (during which time the drives were silent and the zpool process was mostly 0% very occasionally peaking to 25%, and the system being very slow to respond during that period) Any ideas please? Or is there still some process running in the background that I can't see? mich@jaguar:~# zfs list NAME USED AVAIL REFER MOUNTPOINT data 2.27T 401G 2.27T /mirror rpool 7.69G 28.7G 45K /rpool rpool/ROOT 3.70G 28.7G 31K legacy rpool/ROOT/openindiana 3.70G 28.7G 3.59G / rpool/dump 1.93G 28.7G 1.93G - rpool/export 5.22M 28.7G 32K /export rpool/export/home 5.19M 28.7G 32K /export/home rpool/export/home/mich 5.16M 28.7G 5.16M /export/home/mich rpool/swap 2.05G 30.7G 126M - mich@jaguar:~# zpool status pool: data state: ONLINE scan: resilvered 1.13T in 12h26m with 0 errors on Wed Jan 19 23:42:23 2011 config: NAME STATE READ WRITE CKSUM data ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 c2t2d0 ONLINE 0 0 0 c2t3d0 ONLINE 0 0 0 c2t4d0 ONLINE 0 0 0 errors: No known data errors last pid: 1802; load avg: 0.61, 0.56, 0.61; up 0+19:55:20 07:06:01 74 processes: 73 sleeping, 1 on cpu CPU states: 99.8% idle, 0.0% user, 0.3% kernel, 0.0% iowait, 0.0% swap Kernel: 375 ctxsw, 653 intr, 120 syscall Memory: 3959M phys mem, 401M free mem, 1979M total swap, 1979M free swap PID USERNAME NLWP PRI NICE SIZE RES STATE TIME CPU COMMAND 1197 gdm 1 59 0 95M 28M sleep 1:23 0.04% gdm-simple-gree 922 root 3 59 0 102M 51M sleep 0:37 0.02% Xorg 1801 root 1 59 0 4036K 2460K cpu/3 0:00 0.01% top 1196 gdm 1 59 0 80M 13M sleep 0:00 0.00% metacity 1190 gdm 1 59 0 7892K 6028K sleep 0:00 0.00% at-spi-registry 640 root 16 59 0 14M 9072K sleep 0:09 0.00% smbd 1737 mich 1 59 0 13M 5392K sleep 0:00 0.00% sshd 148 root 1 59 0 8312K 1608K sleep 0:00 0.00% dhcpagent 672 root 26 59 0 27M 15M sleep 0:01 0.00% fmd 1198 gdm 1 59 0 87M 18M sleep 0:02 0.00% gnome-power-man 1247 root 1 59 0 6080K 2500K sleep 0:00 0.00% sendmail 11 root 21 59 0 15M 13M sleep 0:06 0.00% svc.configd 272 root 6 59 0 11M 4784K sleep 0:01 0.00% devfsadm 1220 root 24 59 0 13M 4404K sleep 0:01 0.00% nscd 45 netcfg 5 59 0 4716K 3268K sleep 0:00 0.00% netcfgd 1309 admin 1 59 0 16M 8404K sleep 14:30 0.00% sshd 1312 admin 1 59 0 16M 5468K sleep 3:47 0.00% sshd 787 root 1 59 0 12M 5928K sleep 0:05 0.00% intrd 1237 root 4 59 0 8400K 1936K sleep 0:03 0.00% automountd 9 root 15 59 0 20M 12M sleep 0:02 0.00% svc.startd 387 root 5 59 0 7500K 6024K sleep 0:01 0.00% hald 1308 root 1 59 0 13M 4788K sleep 0:01 0.00% sshd 252 root 5 60 -20 2544K 1460K sleep 0:00 0.00% zonestatd 1191 gdm 1 59 0 111M 42M sleep 0:00 0.00% gnome-settings- 1175 gdm 2 59 0 20M 10M sleep 0:00 0.00% gnome-session 790 root 4 59 0 15M 6432K sleep 0:00 0.00% rad 1188 gdm 1 59 0 12M 5944K sleep 0:00 0.00% gconfd-2 291 root 1 59 0 16M 5604K sleep 0:00 0.00% cupsd 1156 mich 1 59 0 13M 5432K sleep 0:00 0.00% sshd 132 daemon 3 59 0 13M 5104K sleep 0:00 0.00% kcfd _______________________________________________ OpenIndiana-discuss mailing list OpenIndiana-discuss@openindiana.org http://openindiana.org/mailman/listinfo/openindiana-discuss