Thanks for the suggestions,
My concerns are that I can't actually copyToLocal from the dfs because the
data is huge.

Say if my hadoop was 0.20 and I am upgrading to 0.20.205 I can do a
namenode upgrade. I don't have to copy data out of dfs.

But here I am having Apache hadoop 0.20.205 and I want to use CDH3 now,
which is based on 0.20
Now it is actually a downgrade as 0.20.205's namenode info has to be used
by 0.20's namenode.

Any idea how I can achieve what I am trying to do?

Thanks.

On Thu, May 3, 2012 at 12:23 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:

> i can think of following options
>
> 1) write a simple get and put code which gets the data from DFS and loads
> it in dfs
> 2) see if the distcp  between both versions are compatible
> 3) this is what I had done (and my data was hardly few hundred GB) .. did a
> dfs -copyToLocal and then in the new grid did a copyFromLocal
>
> On Thu, May 3, 2012 at 11:41 AM, Austin Chungath <austi...@gmail.com>
> wrote:
>
> > Hi,
> > I am migrating from Apache hadoop 0.20.205 to CDH3u3.
> > I don't want to lose the data that is in the HDFS of Apache hadoop
> > 0.20.205.
> > How do I migrate to CDH3u3 but keep the data that I have on 0.20.205.
> > What is the best practice/ techniques to do this?
> >
> > Thanks & Regards,
> > Austin
> >
>
>
>
> --
> Nitin Pawar
>

Reply via email to