Does smell like a bug as that number you get is simply Long.MAX_VALUE,
or 8 exbibytes.

Looking at the sources, this turns out to be a rather funny Java issue
(there's a divide by zero happening and [1] suggests Long.MAX_VALUE
return in such a case). I've logged a bug report for this at
https://issues.apache.org/jira/browse/MAPREDUCE-5288 with a
reproducible case.

Does this happen consistently for you?

[1] http://docs.oracle.com/javase/6/docs/api/java/lang/Math.html#round(double)

On Sat, Jun 1, 2013 at 7:27 PM, Lanati, Matteo <matteo.lan...@lrz.de> wrote:
> Hi all,
>
> I stumbled upon this problem as well while trying to run the default 
> wordcount shipped with Hadoop 1.2.0. My testbed is made up of 2 virtual 
> machines: Debian 7, Oracle Java 7, 2 GB RAM, 25 GB hard disk. One node is 
> used as JT+NN, the other as TT+DN. Security is enabled. The input file is 
> about 600 kB and the error is
>
> 2013-06-01 12:22:51,999 WARN org.apache.hadoop.mapred.JobInProgress: No room 
> for map task. Node 10.156.120.49 has 22854692864 bytes free; but we expect 
> map to take 9223372036854775807
>
> The logfile is attached, together with the configuration files. The version 
> I'm using is
>
> Hadoop 1.2.0
> Subversion https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.2 
> -r 1479473
> Compiled by hortonfo on Mon May  6 06:59:37 UTC 2013
> From source with checksum 2e0dac51ede113c1f2ca8e7d82fb3405
> This command was run using 
> /home/lu95jib/hadoop-exmpl/hadoop-1.2.0/hadoop-core-1.2.0.jar
>
> If I run the default configuration (i.e. no securty), then the job succeeds.
>
> Is there something missing in how I set up my nodes? How is it possible that 
> the envisaged value for the needed space is so big?
>
> Thanks in advance.
>
> Matteo
>
>
>
>>Which version of Hadoop are you using. A quick search shows me a bug
>>https://issues.apache.org/jira/browse/HADOOP-5241 that seems to show
>>similar symptoms. However, that was fixed a long while ago.
>>
>>
>>On Sat, Mar 23, 2013 at 4:40 PM, Redwane belmaati cherkaoui <
>>reduno1...@googlemail.com> wrote:
>>
>>> This the content of the jobtracker log file :
>>> 2013-03-23 12:06:48,912 INFO org.apache.hadoop.mapred.JobInProgress: Input
>>> size for job job_201303231139_0001 = 6950001. Number of splits = 7
>>> 2013-03-23 12:06:48,925 INFO org.apache.hadoop.mapred.JobInProgress:
>>> tip:task_201303231139_0001_m_000000 has split on
>>> node:/default-rack/hadoop0.novalocal
>>> 2013-03-23 12:06:48,927 INFO org.apache.hadoop.mapred.JobInProgress:
>>> tip:task_201303231139_0001_m_000001 has split on
>>> node:/default-rack/hadoop0.novalocal
>>> 2013-03-23 12:06:48,930 INFO org.apache.hadoop.mapred.JobInProgress:
>>> tip:task_201303231139_0001_m_000002 has split on
>>> node:/default-rack/hadoop0.novalocal
>>> 2013-03-23 12:06:48,931 INFO org.apache.hadoop.mapred.JobInProgress:
>>> tip:task_201303231139_0001_m_000003 has split on
>>> node:/default-rack/hadoop0.novalocal
>>> 2013-03-23 12:06:48,933 INFO org.apache.hadoop.mapred.JobInProgress:
>>> tip:task_201303231139_0001_m_000004 has split on
>>> node:/default-rack/hadoop0.novalocal
>>> 2013-03-23 12:06:48,934 INFO org.apache.hadoop.mapred.JobInProgress:
>>> tip:task_201303231139_0001_m_000005 has split on
>>> node:/default-rack/hadoop0.novalocal
>>> 2013-03-23 12:06:48,939 INFO org.apache.hadoop.mapred.JobInProgress:
>>> tip:task_201303231139_0001_m_000006 has split on
>>> node:/default-rack/hadoop0.novalocal
>>> 2013-03-23 12:06:48,950 INFO org.apache.hadoop.mapred.JobInProgress:
>>> job_201303231139_0001 LOCALITY_WAIT_FACTOR=0.5
>>> 2013-03-23 12:06:48,978 INFO org.apache.hadoop.mapred.JobInProgress: Job
>>> job_201303231139_0001 initialized successfully with 7 map tasks and 1
>>> reduce tasks.
>>> 2013-03-23 12:06:50,855 INFO org.apache.hadoop.mapred.JobTracker: Adding
>>> task (JOB_SETUP) 'attempt_201303231139_0001_m_000008_0' to tip
>>> task_201303231139_0001_m_000008, for tracker
>>> 'tracker_hadoop0.novalocal:hadoop0.novalocal/127.0.0.1:44879'
>>> 2013-03-23 12:08:00,340 INFO org.apache.hadoop.mapred.JobInProgress: Task
>>> 'attempt_201303231139_0001_m_000008_0' has completed
>>> task_201303231139_0001_m_000008 successfully.
>>> 2013-03-23 12:08:00,538 WARN org.apache.hadoop.mapred.JobInProgress: No
>>> room for map task. Node hadoop0.novalocal has 8791543808 bytes free; but we
>>> expect map to take 1317624576693539401
>>> 2013-03-23 12:08:00,543 WARN org.apache.hadoop.mapred.JobInProgress: No
>>> room for map task. Node hadoop0.novalocal has 8791543808 bytes free; but we
>>> expect map to take 1317624576693539401
>>> 2013-03-23 12:08:00,544 WARN org.apache.hadoop.mapred.JobInProgress: No
>>> room for map task. Node hadoop0.novalocal has 8791543808 bytes free; but we
>>> expect map to take 1317624576693539401
>>> 2013-03-23 12:08:00,544 WARN org.apache.hadoop.mapred.JobInProgress: No
>>> room for map task. Node hadoop0.novalocal has 8791543808 bytes free; but we
>>> expect map to take 1317624576693539401
>>> 2013-03-23 12:08:01,264 WARN org.apache.hadoop.mapred.JobInProgress: No
>>> room for map task. Node hadoop1.novalocal has 8807518208 bytes free; but we
>>> expect map to take 1317624576693539401
>>>
>>>
>>> The value in we excpect map to take is too huge   1317624576693539401
>>> bytes  !!!!!!!
>>>
>>> On Sat, Mar 23, 2013 at 11:37 AM, Redwane belmaati cherkaoui <
>>> reduno1...@googlemail.com> wrote:
>>>
>>>> The estimated value that the hadoop compute is too huge for the simple
>>>> example that i am running .
>>>>
>>>> ---------- Forwarded message ----------
>>>> From: Redwane belmaati cherkaoui <reduno1...@googlemail.com>
>>>>  Date: Sat, Mar 23, 2013 at 11:32 AM
>>>> Subject: Re: About running a simple wordcount mapreduce
>>>> To: Abdelrahman Shettia <ashet...@hortonworks.com>
>>>> Cc: user@hadoop.apache.org, reduno1985 <reduno1...@gmail.com>
>>>>
>>>>
>>>> This the output that I get I am running two machines  as you can see  do
>>>> u see anything suspicious ?
>>>> Configured Capacity: 21145698304 (19.69 GB)
>>>> Present Capacity: 17615499264 (16.41 GB)
>>>> DFS Remaining: 17615441920 (16.41 GB)
>>>> DFS Used: 57344 (56 KB)
>>>> DFS Used%: 0%
>>>> Under replicated blocks: 0
>>>> Blocks with corrupt replicas: 0
>>>> Missing blocks: 0
>>>>
>>>> -------------------------------------------------
>>>> Datanodes available: 2 (2 total, 0 dead)
>>>>
>>>> Name: 11.1.0.6:50010
>>>> Decommission Status : Normal
>>>> Configured Capacity: 10572849152 (9.85 GB)
>>>> DFS Used: 28672 (28 KB)
>>>> Non DFS Used: 1765019648 (1.64 GB)
>>>> DFS Remaining: 8807800832(8.2 GB)
>>>> DFS Used%: 0%
>>>> DFS Remaining%: 83.31%
>>>> Last contact: Sat Mar 23 11:30:10 CET 2013
>>>>
>>>>
>>>> Name: 11.1.0.3:50010
>>>> Decommission Status : Normal
>>>> Configured Capacity: 10572849152 (9.85 GB)
>>>> DFS Used: 28672 (28 KB)
>>>> Non DFS Used: 1765179392 (1.64 GB)
>>>> DFS Remaining: 8807641088(8.2 GB)
>>>> DFS Used%: 0%
>>>> DFS Remaining%: 83.3%
>>>> Last contact: Sat Mar 23 11:30:08 CET 2013
>>>>
>>>>
>>>> On Fri, Mar 22, 2013 at 10:19 PM, Abdelrahman Shettia <
>>>> ashet...@hortonworks.com> wrote:
>>>>
>>>>> Hi Redwane,
>>>>>
>>>>> Please run the following command as hdfs user on any datanode. The
>>>>> output will be something like this. Hope this helps
>>>>>
>>>>> hadoop dfsadmin -report
>>>>> Configured Capacity: 81075068925 (75.51 GB)
>>>>> Present Capacity: 70375292928 (65.54 GB)
>>>>> DFS Remaining: 69895163904 (65.09 GB)
>>>>> DFS Used: 480129024 (457.89 MB)
>>>>> DFS Used%: 0.68%
>>>>> Under replicated blocks: 0
>>>>> Blocks with corrupt replicas: 0
>>>>> Missing blocks: 0
>>>>>
>>>>> Thanks
>>>>> -Abdelrahman
>>>>>
>>>>>
>>>>> On Fri, Mar 22, 2013 at 12:35 PM, reduno1985 
>>>>> <reduno1...@googlemail.com>wrote:
>>>>>
>>>>>>
>>>>>> I have my hosts running on openstack virtual machine instances each
>>>>>> instance has 10gb hard disc . Is there a way too see how much space is in
>>>>>> the hdfs without web ui .
>>>>>>
>>>>>>
>>>>>> Sent from Samsung Mobile
>>>>>>
>>>>>> Serge Blazhievsky <hadoop...@gmail.com> wrote:
>>>>>> Check web ui how much space you have on hdfs???
>>>>>>
>>>>>> Sent from my iPhone
>>>>>>
>>>>>> On Mar 22, 2013, at 11:41 AM, Abdelrahman Shettia <
>>>>>> ashet...@hortonworks.com> wrote:
>>>>>>
>>>>>> Hi Redwane ,
>>>>>>
>>>>>> It is possible that the hosts which are running tasks are do not have
>>>>>> enough space. Those dirs are confiugred in mapred-site.xml
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Mar 22, 2013 at 8:42 AM, Redwane belmaati cherkaoui <
>>>>>> reduno1...@googlemail.com> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> ---------- Forwarded message ----------
>>>>>>> From: Redwane belmaati cherkaoui <reduno1...@googlemail.com>
>>>>>>> Date: Fri, Mar 22, 2013 at 4:39 PM
>>>>>>> Subject: About running a simple wordcount mapreduce
>>>>>>> To: mapreduce-iss...@hadoop.apache.org
>>>>>>>
>>>>>>>
>>>>>>> Hi
>>>>>>> I am trying to run  a wordcount mapreduce job on several files (<20
>>>>>>> mb) using two machines . I get stuck on 0% map 0% reduce.
>>>>>>> The jobtracker log file shows the following warning:
>>>>>>>  WARN org.apache.hadoop.mapred.JobInProgress: No room for map task.
>>>>>>> Node hadoop0.novalocal has 8791384064 bytes free; but we expect map to
>>take
>>>>>>> 1317624576693539401
>>>>>>>
>>>>>>> Please help me ,
>>>>>>> Best Regards,
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>>
>>>
>
>
> Matteo Lanati
> Distributed Resources Group
> Leibniz-Rechenzentrum (LRZ)
> Boltzmannstrasse 1
> 85748 Garching b. München (Germany)
> Phone: +49 89 35831 8724



--
Harsh J

Reply via email to