[
https://issues.apache.org/jira/browse/HIVE-1100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12805365#action_12805365
]
Zheng Shao commented on HIVE-1100:
----------------------------------
What will be the semantics of pre-execution and post-execution hooks?
Shall we re-execute pre-execution hooks every time we resume the job?
Shall we fail the query if the post-execution hooks failed?
My preference is to treat the hooks as a task in the query plan - once it
succeeds then we don't run it again when resuming the query.
> Make it possible for users to retry map-reduce jobs in a single Hive query
> --------------------------------------------------------------------------
>
> Key: HIVE-1100
> URL: https://issues.apache.org/jira/browse/HIVE-1100
> Project: Hadoop Hive
> Issue Type: New Feature
> Affects Versions: 0.6.0
> Reporter: Zheng Shao
> Assignee: Zheng Shao
>
> Sometimes a single hive query get compiled into several map-reduce jobs, and
> one of the jobs failed because of some transient errors.
> Users would need to start from scratch.
> We should allow the user to start from the point of failure to continue the
> query.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.