[
https://issues.apache.org/jira/browse/HADOOP-1857?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12527222
]
Runping Qi commented on HADOOP-1857:
------------------------------------
I think we need to handle the case for C++ pipe and the case for streaming
differently.
In all cases, it will be helpful to log the offending key/value pair, and the
progress stats (how many key/value pairs have been processed, etc)
In C++ pipe case, we know the executable is in C++, thus Raghu's suggestion is
good.
For streaming, the executable can be anything. Not sure how do you get
stacktrace.
It is really up to the executable.
> Ability to run a script when a task fails to capture stack traces
> -----------------------------------------------------------------
>
> Key: HADOOP-1857
> URL: https://issues.apache.org/jira/browse/HADOOP-1857
> Project: Hadoop
> Issue Type: Improvement
> Components: mapred
> Affects Versions: 0.14.0
> Reporter: Amareshwari Sri Ramadasu
> Assignee: Amareshwari Sri Ramadasu
> Fix For: 0.15.0
>
>
> This basically is for providing a better user interface for debugging failed
> jobs. Today we see stack traces for failed tasks on the job ui if the job
> happened to be a Java MR job. For non-Java jobs like Streaming, Pipes, the
> diagnostic info on the job UI is not helpful enough to debug what might have
> gone wrong. They are usually framework traces and not app traces.
> We want to be able to provide a facility, via user-provided scripts, for doing
> post-processing on task logs, input, output, etc. There should be some default
> scripts like running core dumps under gdb for locating illegal instructions,
> the last few lines from stderr, etc. These outputs could be sent to the
> tasktracker and in turn to the jobtracker which would then display it on the
> job UI on demand.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.