Thanks for the help. I checked the quotas, it seems they're used for setting the maximum size on the files inside the hdfs, and not the datanode itself. For example, if I set my dfs.data.dir to /media/newhard (which I've mounted my new hard disk to), I can't use dfsadmin -setSpaceQuota n /media/newhard to set the size of this directory, I can change the sizes of the directories inside hdfs (tmp, user, ...), which don't have any effect on the capacity of the datanode. I can set the my new mounted volume as the datanode directory and it runs without a problem, but the capacity is the default 5 GB.
On Sun, Jan 1, 2012 at 10:41 AM, Rajiv Chittajallu <raj...@yahoo-inc.com>wrote: > Once you updated the configuration is the datanode, restarted? Check if > the datanode log indicated that it was able to setup the new volume. > > > > >________________________________ > > From: Hamed Ghavamnia <ghavamni...@gmail.com> > >To: hdfs-user@hadoop.apache.org > >Sent: Sunday, January 1, 2012 11:33 AM > >Subject: HDFS Datanode Capacity > > > > > >Hi, > >I've been searching on how to configure the maximum capacity of a > datanode. I've added big volumes to one of my datanodes, but the configured > capacity doesn't get bigger than the default 5GB. If I want a datanode with > 100GB of capacity, I have to add 20 directories, each having 5GB so the > maximum capacity reaches 100. Is there anywhere this can be set? Can > different datanodes have different capacities? > > > >Also it seems like the dfs.datanode.du.reserved doesn't work either, > because I've set it to zero, but it still leaves 50% of the free space for > non-dfs usage. > > > >Thanks, > >Hamed > > > >P.S. This is my first message in the mailing list, so if I have to follow > any rules for sending emails, I'll be thankful if you let me know. :) > > > > > > >