Hi, Consider a scenario where we submit a coordinator job. After a few weeks, we want to change the coordinator file or workflow file or properties file and hence we have to submit a new coordinator job for the same job. The previous one is either in SUCCESS state or KILLED state.
Now, consider that after a while we want to re-run instances of this job for a given time line such that a few instances are from the first coordinator and others are from the second coordinator. There might be issues such as frequency of job has changed, code might not be backward compatible, hadoop cluster configuration has changed i.e. say we added LZO compression in between etc. Although these issues are rare, they are worth considering. Also we will have to rerun dependent jobs. Has anyone come across such situation? Is there any good solution for this? We are brainstorming on ideas for this. I will post these ideas within a couple of days, but would like to hear from you first and don't want to bias your ideas towards our ideas. Regards,
