Speak for yourself, I just built a bunch of 36 disk datanodes :)

-j

On May 9, 2011, at 2:33 AM, Eric wrote:

> Just a small warning: I've seen kernel panics with the XFS kernel module once 
> you have many disks (in my case: > 20 disks). This is an exotic amount of 
> disks to put in one server so it shouldn't hold anyone back from using XFS :-)
> 
> 2011/5/7 Rita <rmorgan...@gmail.com>
> Sheng,
> 
> How big is your each XFS volume? We noticed if its over 4TB hdfs won't pick 
> it up.
> 
> 
> 2011/5/6 Ferdy Galema <ferdy.gal...@kalooga.com>
> No unfortunately not, we couldn't because of our kernel versions.
> 
> 
> On 05/06/2011 04:00 AM, ShengChang Gu wrote:
>> 
>> Many thanks.
>> 
>> We use xfs all the time.Have you try the ext4 filesystem?
>> 
>> 2011/5/6 Ferdy Galema <ferdy.gal...@kalooga.com>
>> Hi,
>> 
>> We've performed tests for ext3 and xfs filesystems using different settings. 
>> The results might be useful for anyone else.
>> 
>> The datanode cluster consists of 15 slave nodes, each equipped with 1Gbit 
>> ethernet, X3220@2.40GHz quadcores and 4x1TB disks. The disk read speeds vary 
>> from about 90 to 130MB/s. (Tested using hdparm -t).
>> 
>> Hadoop: Cloudera CDH3u0 (4 concurrent mappers / node)
>> OS: Linux version 2.6.18-238.5.1.el5 (mockbu...@builder10.centos.org) (gcc 
>> version 4.1.2 20080704 (Red Hat 4.1.2-50))
>> 
>> #our command
>> for i in `seq 1 10`; do ./hadoop jar ../hadoop-examples-0.20.2-cdh3u0.jar 
>> randomwriter -Ddfs.replication=1 /rand$i && ./hadoop fs -rmr /rand$i/_logs 
>> /rand$i/_SUCCESS && ./hadoop distcp -Ddfs.replication=1 /rand$i 
>> /rand-copy$i; done
>> 
>> Our benchmark consists of a standard random-writer job followed by a distcp 
>> of the same data, both using a replication of 1. This is to make sure only 
>> the disks get hit. Each benchmark is ran several times for every 
>> configuration. Because of the occasional hickup, I will list both the 
>> average and the fastest times for each configuration. I read the execution 
>> times off the jobtracker.
>> 
>> The configurations (with exection times in seconds of Avg-writer / 
>> Min-writer / Avg-distcp / Min-distcp)
>> ext3-default      158 / 136 / 411 / 343
>> ext3-tuned        159 / 132 / 330 / 297
>> ra1024 ext3-tuned 159 / 132 / 292 / 264
>> ra1024 xfs-tuned  128 / 122 / 220 / 202
>> 
>> To explain, ext3-tuned is with tuned mount options 
>> [noatime,nodiratime,data=writeback,rw] and ra1024 means a read-ahead buffer 
>> of 1024 blocks. The xfs disks are created using mkfs options 
>> [size=128m,lazy-count=1] and mount options [noatime,nodiratime,logbufs=8].
>> 
>> In conclusion it seems that using tuned xfs filesystems combined with 
>> increased read-ahead buffers increased our basic hdfs performance with about 
>> 10% (random-writer) to 40% (distcp).
>> 
>> Hopefully this is useful to anyone. Although I won't be performing more 
>> tests soon I'd be happy to provide more details.
>> Ferdy.
>> 
>> 
>> 
>> -- 
>> 阿昌
> 
> 
> 
> -- 
> --- Get your facts first, then you can distort them as you please.--
> 

Reply via email to