You can safely move block files between disks. Follow the instructions here: http://wiki.apache.org/hadoop/FAQ#On_an_individual_data_node.2C_how_do_you_balance_the_blocks_on_the_disk.3F
On Tue, Mar 4, 2014 at 11:47 PM, divye sheth <divs.sh...@gmail.com> wrote: > Thanks Harsh. The jira is fixed in version 2.1.0 whereas I am using Hadoop > 0.20.2 (we are in a process of upgrading) is there a workaround for the > short term to balance the disk utilization? The patch in the Jira, if > applied to the version that I am using, will it break anything? > > Thanks > Divye Sheth > > > On Wed, Mar 5, 2014 at 11:28 AM, Harsh J <ha...@cloudera.com> wrote: >> >> You're probably looking for >> https://issues.apache.org/jira/browse/HDFS-1804 >> >> On Tue, Mar 4, 2014 at 5:54 AM, divye sheth <divs.sh...@gmail.com> wrote: >> > Hi, >> > >> > I am new to the mailing list. >> > >> > I am using Hadoop 0.20.2 with an append r1056497 version. The question I >> > have is related to balancing. I have a 5 datanode cluster and each node >> > has >> > 2 disks attached to it. The second disk was added when the first disk >> > was >> > reaching its capacity. >> > >> > Now the scenario that I am facing is, when the new disk was added hadoop >> > automatically moved over some data to the new disk. But over the time I >> > notice that data is no longer being written to the second disk. I have >> > also >> > faced an issue on the datanode where the first disk had 100% >> > utilization. >> > >> > How can I overcome such scenario, is it not hadoop's job to balance the >> > disk >> > utilization between multiple disks on single datanode? >> > >> > Thanks >> > Divye Sheth >> >> >> >> -- >> Harsh J > > -- Harsh J