[ 
https://issues.apache.org/jira/browse/HIVE-690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12735229#action_12735229
 ] 

Zheng Shao commented on HIVE-690:
---------------------------------

The major goal of status reporting is probably more for the detecting dead 
process/dead node than for infinite loops.

Also, users will probably just output stderr messages using timers in their 
python code.
If that's the case, it does not prevent our users' program from infinite loop.

I think we should check this fix in first (after all it's so simple a fix). 
Thoughts?


> ScriptOperator still times out
> ------------------------------
>
>                 Key: HIVE-690
>                 URL: https://issues.apache.org/jira/browse/HIVE-690
>             Project: Hadoop Hive
>          Issue Type: Bug
>            Reporter: Zheng Shao
>            Assignee: Zheng Shao
>             Fix For: 0.4.0
>
>         Attachments: HIVE-690.1.patch
>
>
> The code is using exp_interval instead of exp_int in the 
> "scheduleAtFxiedRate".
> {code}
>       Integer exp_interval = null;
>       int exp_int;
>       exp_interval = 
> Integer.decode(hconf.get("mapred.tasktracker.expiry.interval"));
>       if (exp_interval != null)
>         exp_int = exp_interval.intValue() / 2;
>       else
>         exp_int = 300000;
>       rpTimer = new Timer(true);
>       rpTimer.scheduleAtFixedRate(new ReporterTask(reporter), 0, 
> exp_interval);
> {code}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to