Thanks for the suggestions, My concerns are that I can't actually copyToLocal from the dfs because the data is huge.
Say if my hadoop was 0.20 and I am upgrading to 0.20.205 I can do a namenode upgrade. I don't have to copy data out of dfs. But here I am having Apache hadoop 0.20.205 and I want to use CDH3 now, which is based on 0.20 Now it is actually a downgrade as 0.20.205's namenode info has to be used by 0.20's namenode. Any idea how I can achieve what I am trying to do? Thanks. On Thu, May 3, 2012 at 12:23 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote: > i can think of following options > > 1) write a simple get and put code which gets the data from DFS and loads > it in dfs > 2) see if the distcp between both versions are compatible > 3) this is what I had done (and my data was hardly few hundred GB) .. did a > dfs -copyToLocal and then in the new grid did a copyFromLocal > > On Thu, May 3, 2012 at 11:41 AM, Austin Chungath <austi...@gmail.com> > wrote: > > > Hi, > > I am migrating from Apache hadoop 0.20.205 to CDH3u3. > > I don't want to lose the data that is in the HDFS of Apache hadoop > > 0.20.205. > > How do I migrate to CDH3u3 but keep the data that I have on 0.20.205. > > What is the best practice/ techniques to do this? > > > > Thanks & Regards, > > Austin > > > > > > -- > Nitin Pawar >