Good summary, Brian.

This should be added to ref guide.

Cheers

On Tue, Dec 16, 2014 at 4:17 AM, Brian Jeltema <
brian.jelt...@digitalenvoy.net> wrote:
>
> I have been able to export snapshots from 0.94 to 0.98. I’ve pasted the
> instructions that I developed
>
> and published on our internal wiki. I also had to significantly increase
> retry count parameters due to
>
> a high number of timeout failures during the export.
>
>
> Cross-cluster transfers
>
> To export a snapshot to a different cluster:
>
>    hbase org.apache.hadoop.hbase.snapshot.ExportSnapshot -snapshot snappy
> -copy-to proto://remhost/apps/hbase/data -mappers nummaps
>
> where snappy is the local snapshot to export, remhost is the host name the
> default filesystem in the remote cluster which is the target of the export
> (i.e. the value of fs.defaultFS), nummaps is the number of mappers to run
> to perform the export, and proto is the protocol to use, either hftp, hdfs
> or webhdfs. Use hdfs if the clusters are compatible. Run this as the hbase
> user. If you see exceptions being thrown during the transfer related to
> lease expirations, reduce the number of mappers or try using the -bandwidth
> option. You may also see many "File does not exist" warnings in the log
> output. These can be displayed continuously for several minutes for a large
> table, but appear to be noise and can be ignored, so be patient. However,
> it is also very common for this command to fail due to a variety of file
> ownership conflicts, so you may need to fiddle to get everything right. A
> failure of this command often leaves garbage on the target system that must
> be deleted; if that is the case, the command will fail with info on what
> needs to be cleaned up.
>
> If exporting from an HBase 0.94 cluster to an HBase 0.98 cluster, you will
> need to use the webhdfs protocol (or possibly hftp, though I couldn’t get
> that to work). You also need to manually move some files around because
> snapshot layouts have changed. Based on the example above, on the 0.98
> cluster do the following:
>
>    check whether any imports already exist for the table:
>
>   hadoop fs -ls /apps/hbase/data/archive/data/default
>
>   and check whether  snappyTable is listed, where snappyTable is the
> source of the snapshot (e.g. hosts). If the source table is listed, then
> merge the new ssnapshot data into the existing snapshot data:
>
>   hadoop fs -mv /apps/hbase/data/.archive/snappyTable/*
> /apps/hbase/data/archive/data/default/snappyTable
>
>     hadoop fs -rm r /apps/hbase/data/.archive/snappyTable
>
> otherwise, create and populate the snapshot data directory:
>
>   hadoop fs -mv /apps/hbase/data/.archive/snappyTable
> /apps/hbase/data/archive/data/default (snappyTable is the source of snappy)
>
> in either case, update the snapshot metadata files as follows:
>
>      hadoop fs -mkdir /apps/hbase/data/.hbase-snapshot/snappy/.tabledesc
>
>   hadoop fs -mv
> /apps/hbase/data/.hbase-snapshot/snappy/.tableinfo.0000000001
> /apps/hbase/data/.hbase-snapshot/snappy/.tabledesc
>
> at this point, you should be able to do  a restore_snapshot from the HBase
> shell.
>
>
> On Dec 15, 2014, at 8:52 PM, lars hofhansl <la...@apache.org> wrote:
>
> > Nope :(Replication uses RPC and that was changed to protobufs. AFAIK
> snapshots can also not be exported from 0.94 and 0.98. We have a really
> shitty story here.      From: Sean Busbey <bus...@cloudera.com>
> > To: user <user@hbase.apache.org>
> > Sent: Monday, December 15, 2014 5:04 PM
> > Subject: Re: 0.94 going forward
> >
> > Does replication and snapshot export work from 0.94.6+ to a 0.96 or 0.98
> > cluster?
> >
> > Presuming it does, shouldn't a site be able to use a multiple cluster set
> > up to do a cut over of a client application?
> >
> > That doesn't help with needing downtime for to do the eventual upgrade,
> but
> > it mitigates the impact on the downstream app.
> >
> > --
> > Sean
> >
> >
> > On Dec 15, 2014 6:51 PM, "Jeremy Carroll" <phobos...@gmail.com> wrote:
> >
> >> Which is why I feel that a lot of customers are still on 0.94. Pretty
> much
> >> trapped unless you want to take downtime for your site. Any type of
> >> guidance would be helpful. We are currently in the process of designing
> our
> >> own system to deal with this.
> >>
> >> On Mon, Dec 15, 2014 at 4:47 PM, Andrew Purtell <apurt...@apache.org>
> >> wrote:
> >>>
> >>> Zero downtime upgrade from 0.94 won't be possible. See
> >>> http://hbase.apache.org/book.html#d0e5199
> >>>
> >>>
> >>> On Mon, Dec 15, 2014 at 4:44 PM, Jeremy Carroll <phobos...@gmail.com>
> >>> wrote:
> >>>>
> >>>> Looking for guidance on how to do a zero downtime upgrade from 0.94 ->
> >>> 0.98
> >>>> (or 1.0 if it launches soon). As soon as we can figure this out, we
> >> will
> >>>> migrate over.
> >>>>
> >>>> On Mon, Dec 15, 2014 at 1:37 PM, Esteban Gutierrez <
> >> este...@cloudera.com
> >>>>
> >>>> wrote:
> >>>>>
> >>>>> Hi Lars,
> >>>>>
> >>>>> Thanks for bringing this for discussion. From my experience I can
> >> tell
> >>>> that
> >>>>> 0.94 is very stable but that shouldn't be a blocker to consider to
> >>>> EOL'ing.
> >>>>> Are you considering any specific timeframe for that?
> >>>>>
> >>>>> thanks,
> >>>>> esteban.
> >>>>>
> >>>>> --
> >>>>> Cloudera, Inc.
> >>>>>
> >>>>>
> >>>>> On Mon, Dec 15, 2014 at 11:46 AM, Koert Kuipers <ko...@tresata.com>
> >>>> wrote:
> >>>>>>
> >>>>>> given that CDH4 is hbase 0.94 i dont believe nobody is using it.
> >> for
> >>>> our
> >>>>>> clients the majority is on 0.94 (versus 0.96 and up).
> >>>>>>
> >>>>>> so i am going with 1), its very stable!
> >>>>>>
> >>>>>> On Mon, Dec 15, 2014 at 1:53 PM, lars hofhansl <la...@apache.org>
> >>>> wrote:
> >>>>>>>
> >>>>>>> Over the past few months the rate of the change into 0.94 has
> >>> slowed
> >>>>>>> significantly.
> >>>>>>> 0.94.25 was released on Nov 15th, and since then we had only 4
> >>>> changes.
> >>>>>>>
> >>>>>>> This could mean two things: (1) 0.94 is very stable now or (2)
> >>> nobody
> >>>>> is
> >>>>>>> using it (at least nobody is contributing to it anymore).
> >>>>>>>
> >>>>>>>
> >>>>>>> If anybody out there is still using 0.94 and is not planning to
> >>>> upgrade
> >>>>>> to
> >>>>>>> 0.98 or later soon (which will required downtime), please speak
> >> up.
> >>>>>>> Otherwise it might be time to think about EOL'ing 0.94.
> >>>>>>>
> >>>>>>> It's not actually much work to do these releases, especially when
> >>>> they
> >>>>>> are
> >>>>>>> so small, but I'd like to continue only if they are actually
> >> used.
> >>>>>>> In any case, I am going to spin 0.94.26 with the current 4 fixes
> >>>> today
> >>>>> or
> >>>>>>> tomorrow.
> >>>>>>>
> >>>>>>> -- Lars
> >>>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>>
> >>>
> >>> --
> >>> Best regards,
> >>>
> >>>     - Andy
> >>>
> >>> Problems worthy of attack prove their worth by hitting back. - Piet
> Hein
> >>> (via Tom White)
> >>>
> >>
> >
> >
>
>
>

Reply via email to