Speak for yourself, I just built a bunch of 36 disk datanodes :) -j
On May 9, 2011, at 2:33 AM, Eric wrote: > Just a small warning: I've seen kernel panics with the XFS kernel module once > you have many disks (in my case: > 20 disks). This is an exotic amount of > disks to put in one server so it shouldn't hold anyone back from using XFS :-) > > 2011/5/7 Rita <rmorgan...@gmail.com> > Sheng, > > How big is your each XFS volume? We noticed if its over 4TB hdfs won't pick > it up. > > > 2011/5/6 Ferdy Galema <ferdy.gal...@kalooga.com> > No unfortunately not, we couldn't because of our kernel versions. > > > On 05/06/2011 04:00 AM, ShengChang Gu wrote: >> >> Many thanks. >> >> We use xfs all the time.Have you try the ext4 filesystem? >> >> 2011/5/6 Ferdy Galema <ferdy.gal...@kalooga.com> >> Hi, >> >> We've performed tests for ext3 and xfs filesystems using different settings. >> The results might be useful for anyone else. >> >> The datanode cluster consists of 15 slave nodes, each equipped with 1Gbit >> ethernet, X3220@2.40GHz quadcores and 4x1TB disks. The disk read speeds vary >> from about 90 to 130MB/s. (Tested using hdparm -t). >> >> Hadoop: Cloudera CDH3u0 (4 concurrent mappers / node) >> OS: Linux version 2.6.18-238.5.1.el5 (mockbu...@builder10.centos.org) (gcc >> version 4.1.2 20080704 (Red Hat 4.1.2-50)) >> >> #our command >> for i in `seq 1 10`; do ./hadoop jar ../hadoop-examples-0.20.2-cdh3u0.jar >> randomwriter -Ddfs.replication=1 /rand$i && ./hadoop fs -rmr /rand$i/_logs >> /rand$i/_SUCCESS && ./hadoop distcp -Ddfs.replication=1 /rand$i >> /rand-copy$i; done >> >> Our benchmark consists of a standard random-writer job followed by a distcp >> of the same data, both using a replication of 1. This is to make sure only >> the disks get hit. Each benchmark is ran several times for every >> configuration. Because of the occasional hickup, I will list both the >> average and the fastest times for each configuration. I read the execution >> times off the jobtracker. >> >> The configurations (with exection times in seconds of Avg-writer / >> Min-writer / Avg-distcp / Min-distcp) >> ext3-default 158 / 136 / 411 / 343 >> ext3-tuned 159 / 132 / 330 / 297 >> ra1024 ext3-tuned 159 / 132 / 292 / 264 >> ra1024 xfs-tuned 128 / 122 / 220 / 202 >> >> To explain, ext3-tuned is with tuned mount options >> [noatime,nodiratime,data=writeback,rw] and ra1024 means a read-ahead buffer >> of 1024 blocks. The xfs disks are created using mkfs options >> [size=128m,lazy-count=1] and mount options [noatime,nodiratime,logbufs=8]. >> >> In conclusion it seems that using tuned xfs filesystems combined with >> increased read-ahead buffers increased our basic hdfs performance with about >> 10% (random-writer) to 40% (distcp). >> >> Hopefully this is useful to anyone. Although I won't be performing more >> tests soon I'd be happy to provide more details. >> Ferdy. >> >> >> >> -- >> 阿昌 > > > > -- > --- Get your facts first, then you can distort them as you please.-- >