Problem with hadoop decommission

2011-10-04 Thread trang van anh
Dear all, I setup hadoop cluster with following struture 1 Namenode named server1; 4 datanodes named server1,server2,server3,server4. Hadoop cluster capacity: 128G DFS used : 49G i want to detach server1 from cluster but time consumed slowly. I don't know why? Any idea for me Thanks in

setInt getInt

2011-10-04 Thread Ratner, Alan S (IS)
I have no problem with Hadoop.mapred using JobConf to setInt integers and pass them to my map(s) for getInt as shown in the first program below. However, when I use Hadoop.mapreduce using Configuration to setInt these values are invisible to my map's getInt's. Please tell me what I am doing

Re: setInt getInt

2011-10-04 Thread Joey Echeverria
The Job class copies the Configuraiton that you pass in. You either need to do your conf.setInt(number, 12345) before you create the Job object or you need call job.getConfiguration().setInt(number, 12345). -Joey On Tue, Oct 4, 2011 at 12:28 PM, Ratner, Alan S (IS) alan.rat...@ngc.com wrote: I

Re: How do I diagnose IO bounded errors using the framework counters?

2011-10-04 Thread W.P. McNeill
Here's an even more basic question. I tried to figure out what the FILE_BYTES_READ means by searching every file in the hadoop 0.20.203.0 installation for the string FILE_BYTES_READ installation by running find . -type f | xargs grep FILE_BYTES_READ I only found this string in source files

Error using hadoop distcp

2011-10-04 Thread praveenesh kumar
I am trying to use distcp to copy a file from one HDFS to another. But while copying I am getting the following exception : hadoop distcp hdfs://ub13:54310/user/hadoop/weblog hdfs://ub16:54310/user/hadoop/weblog 11/10/05 10:41:01 INFO mapred.JobClient: Task Id :

ways to expand hadoop.tmp.dir capacity?

2011-10-04 Thread Meng Mao
Currently, we've got defined: property namehadoop.tmp.dir/name value/hadoop/hadoop-metadata/cache//value /property In our experiments with SOLR, the intermediate files are so large that they tend to blow out disk space and fail (and annoyingly leave behind their huge failed

Re: ways to expand hadoop.tmp.dir capacity?

2011-10-04 Thread Meng Mao
I just read this: MapReduce performance can also be improved by distributing the temporary data generated by MapReduce tasks across multiple disks on each machine: property namemapred.local.dir/name value/d1/mapred/local,/d2/mapred/local,/d3/mapred/local,/d4/mapred/local/value