thanks Chris. I only run nutch, so no to the external command. And I just checked and it happens or has happened on all the nodes at some point. I have to say thou that it doesn't cause the job to fail or anything. the map tasks that fail will finish when they are re-spawn again. it is just annoying and makes me think that some value some where in the config files are either not correct or not optimal.

On 04/22/2013 02:49 PM, Chris Nauroth wrote:
I'm not aware of any Hadoop-specific meaning for exit code 126.
  Typically, this is a standard Unix exit code used to indicate that a
command couldn't be executed.  Some reasons for this might be that the
command is not an executable file, or the command is an executable file
but the user doesn't have execute permissions.  (See below for an
example of each of these.)

Does your job code attempt to exec an external command?  Also, are the
task failures consistently happening on the same set of nodes in your
cluster?  If so, then I recommend checking that the command has been
deployed and has the correct permissions on those nodes.

Even if your code doesn't exec an external command, various parts of the
Hadoop code do this internally, so you still might have a case of a
misconfigured node.

Hope this helps,
--Chris

[chris@Chriss-MacBook-Pro:ttys000] hadoop-common
 > ./BUILDING.txt
-bash: ./BUILDING.txt: Permission denied
[chris@Chriss-MacBook-Pro:ttys000] hadoop-common
 > echo $?
126

[chris@Chriss-MacBook-Pro:ttys000] test
 > ls -lrt exec
-rwx------  1 root  staff     0B Apr 22 14:43 exec*
[chris@Chriss-MacBook-Pro:ttys000] test
 > whoami
chris
[chris@Chriss-MacBook-Pro:ttys000] test
 > ./exec
bash: ./exec: Permission denied
[chris@Chriss-MacBook-Pro:ttys000] test
 > echo $?
126



On Mon, Apr 22, 2013 at 2:09 PM, kaveh minooie <ka...@plutoz.com
<mailto:ka...@plutoz.com>> wrote:

    thanks. that is the issue, there is no other log files. when i go to
    the attempt directory of that failed map task (e.g.
    userlogs/job_201304191712___0015/attempt_201304191712___0015_m_000019_0
    ) it is empty. there is no other log file. thou based on the counter
    value, I can say that it happens right at the beginning of the map
    task (counter is only 1 )




    On 04/22/2013 02:12 AM, 姚吉龙 wrote:

        Hi


        I have the same problem before
        I think this is caused by the lack of memory shortage for map task.
        It is just a suggestion,you can post your log


        BRs
        Geelong
        —
        Sent from Mailbox <https://bit.ly/SZvoJe> for iPhone



        On Mon, Apr 22, 2013 at 4:34 PM, kaveh minooie <ka...@plutoz.com
        <mailto:ka...@plutoz.com>
        <mailto:ka...@plutoz.com <mailto:ka...@plutoz.com>>> wrote:

             HI

             regardless of what job I run, there are always a few map
        tasks that
             fail with the following, very unhelpful, message: ( that is the
             entire error message)

             java.lang.Throwable: Child Error
                 at
        org.apache.hadoop.mapred.__TaskRunner.run(TaskRunner.__java:271)
             Caused by: java.io.IOException: Task process exit with
        nonzero status of 126.
                 at
        org.apache.hadoop.mapred.__TaskRunner.run(TaskRunner.__java:258)


             I would appreciate it if someone could show me how I could
        figure
             out why this error keeps happening.

             thanks,



    --
    Kaveh Minooie



--
Kaveh Minooie

Reply via email to