Hi,

Consider a scenario where we submit a coordinator job. After a few
weeks, we want to change the coordinator file or workflow file or
properties file and hence we have to submit a new coordinator job for
the same job. The previous one is either in SUCCESS state or KILLED
state.


Now, consider that after a while we want to re-run instances of this job
for a given time line such that a few instances are from the first
coordinator and others are from the second coordinator.


There might be issues such as frequency of job has changed, code might
not be backward compatible, hadoop cluster configuration has changed
i.e. say we added LZO compression in between etc. Although these issues
are rare, they are worth considering.

Also we will have to rerun dependent jobs.

Has anyone come across such situation? Is there any good solution for
this?

We are brainstorming on ideas for this. I will post these ideas within a
couple of days, but would like to hear from you first and don't want to
bias your ideas towards our ideas.

Regards,

Reply via email to