[13/22] hbase-site git commit: Published site at ac5bb8155b618194fe9cf1131f0e72c99b7b534c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a7f7dba1/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.LogRecoveredEditsOutputSink.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.LogRecoveredEditsOutputSink.html b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.LogRecoveredEditsOutputSink.html index 99f53c4..2d5c946 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.LogRecoveredEditsOutputSink.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALSplitter.LogRecoveredEditsOutputSink.html @@ -123,7 +123,7 @@ 115 public static final boolean SPLIT_SKIP_ERRORS_DEFAULT = false; 116 117 // Parameters for split process -118 protected final Path rootDir; +118 protected final Path walDir; 119 protected final FileSystem fs; 120 protected final Configuration conf; 121 @@ -156,14 +156,14 @@ 148 149 150 @VisibleForTesting -151 WALSplitter(final WALFactory factory, Configuration conf, Path rootDir, +151 WALSplitter(final WALFactory factory, Configuration conf, Path walDir, 152 FileSystem fs, LastSequenceId idChecker, 153 SplitLogWorkerCoordination splitLogWorkerCoordination) { 154this.conf = HBaseConfiguration.create(conf); 155String codecClassName = conf 156 .get(WALCellCodec.WAL_CELL_CODEC_CLASS_KEY, WALCellCodec.class.getName()); 157 this.conf.set(HConstants.RPC_CODEC_CONF_KEY, codecClassName); -158this.rootDir = rootDir; +158this.walDir = walDir; 159this.fs = fs; 160this.sequenceIdChecker = idChecker; 161this.splitLogWorkerCoordination = splitLogWorkerCoordination; @@ -194,11 +194,11 @@ 186 * p 187 * @return false if it is interrupted by the progress-able. 188 */ -189 public static boolean splitLogFile(Path rootDir, FileStatus logfile, FileSystem fs, +189 public static boolean splitLogFile(Path walDir, FileStatus logfile, FileSystem fs, 190 Configuration conf, CancelableProgressable reporter, LastSequenceId idChecker, 191 SplitLogWorkerCoordination splitLogWorkerCoordination, final WALFactory factory) 192 throws IOException { -193WALSplitter s = new WALSplitter(factory, conf, rootDir, fs, idChecker, +193WALSplitter s = new WALSplitter(factory, conf, walDir, fs, idChecker, 194splitLogWorkerCoordination); 195return s.splitLogFile(logfile, reporter); 196 } @@ -330,10 +330,10 @@ 322 LOG.warn("Could not parse, corrupted WAL={}", logPath, e); 323 if (splitLogWorkerCoordination != null) { 324// Some tests pass in a csm of null. -325 splitLogWorkerCoordination.markCorrupted(rootDir, logfile.getPath().getName(), fs); +325 splitLogWorkerCoordination.markCorrupted(walDir, logfile.getPath().getName(), fs); 326 } else { 327// for tests only -328ZKSplitLog.markCorrupted(rootDir, logfile.getPath().getName(), fs); +328ZKSplitLog.markCorrupted(walDir, logfile.getPath().getName(), fs); 329 } 330 isCorrupted = true; 331} catch (IOException e) { @@ -465,1437 +465,1442 @@ 457 * codelogEntry/code: e.g. /hbase/some_table/2323432434/recovered.edits/2332. 458 * This method also ensures existence of RECOVERED_EDITS_DIR under the region 459 * creating it if necessary. -460 * @param fs -461 * @param logEntry -462 * @param rootDir HBase root dir. -463 * @param fileNameBeingSplit the file being split currently. Used to generate tmp file name. -464 * @return Path to file into which to dump split log edits. -465 * @throws IOException -466 */ -467 @SuppressWarnings("deprecation") -468 @VisibleForTesting -469 static Path getRegionSplitEditsPath(final FileSystem fs, -470 final Entry logEntry, final Path rootDir, String fileNameBeingSplit) -471 throws IOException { -472Path tableDir = FSUtils.getTableDir(rootDir, logEntry.getKey().getTableName()); -473String encodedRegionName = Bytes.toString(logEntry.getKey().getEncodedRegionName()); -474Path regiondir = HRegion.getRegionDir(tableDir, encodedRegionName); -475Path dir = getRegionDirRecoveredEditsDir(regiondir); -476 -477if (!fs.exists(regiondir)) { -478 LOG.info("This region's directory does not exist: {}." -479 + "It is very likely that it was already split so it is " -480 + "safe to discard those edits.", regiondir); -481 return null; -482} -483if (fs.exists(dir) fs.isFile(dir)) { -484 Path tmp = new Path("/tmp"); -485 if (!fs.exists(tmp)) { -486fs.mkdirs(tmp); -487 } -488 tmp = new Path(tmp, -489HConstants.RECOVERED_EDITS_DIR + "_" + encodedRegionName); -490 LOG.warn("Found existing old file: {}. It could be some " -491+ "leftover of an old installation. It should be a folder instead. " -492+ "So moving it to {}", dir, tmp); -493 if
[13/22] hbase-site git commit: Published site at 46cb5dfa226892fd2580f26ce9ce77225bd7e67c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html index 23b678a..c2d61b1 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html @@ -137,7 +137,7 @@ var activeTableTab = "activeTableTab"; -public static class ThriftServerRunner.IOErrorWithCause +public static class ThriftServerRunner.IOErrorWithCause extends org.apache.hadoop.hbase.thrift.generated.IOError See Also: @@ -273,7 +273,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError cause -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable cause +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable cause @@ -290,7 +290,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError IOErrorWithCause -publicIOErrorWithCause(https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablecause) +publicIOErrorWithCause(https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablecause) @@ -307,7 +307,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError getCause -publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">ThrowablegetCause() +publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">ThrowablegetCause() Overrides: https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--; title="class or interface in java.lang">getCausein classhttps://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable @@ -320,7 +320,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError equals -publicbooleanequals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectother) +publicbooleanequals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectother) Overrides: equalsin classorg.apache.hadoop.hbase.thrift.generated.IOError @@ -333,7 +333,7 @@ extends org.apache.hadoop.hbase.thrift.generated.IOError hashCode -publicinthashCode() +publicinthashCode() Overrides: hashCodein classorg.apache.hadoop.hbase.thrift.generated.IOError http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8bc57a1a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html index 82d106c..0392c62 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -static enum ThriftServerRunner.ImplType +static enum ThriftServerRunner.ImplType extends https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumThriftServerRunner.ImplType An enum of server implementation selections @@ -281,7 +281,7 @@ the order they are declared. HS_HA -public static finalThriftServerRunner.ImplType HS_HA +public static finalThriftServerRunner.ImplType HS_HA @@ -290,7 +290,7 @@ the order they are declared. NONBLOCKING -public static finalThriftServerRunner.ImplType NONBLOCKING +public static finalThriftServerRunner.ImplType NONBLOCKING @@ -299,7 +299,7 @@ the order they are declared. THREAD_POOL -public static finalThriftServerRunner.ImplType THREAD_POOL +public static finalThriftServerRunner.ImplType THREAD_POOL @@ -308,7 +308,7 @@ the order they are declared. THREADED_SELECTOR -public static finalThriftServerRunner.ImplType THREADED_SELECTOR +public static finalThriftServerRunner.ImplType THREADED_SELECTOR @@ -325,7 +325,7 @@ the order they are declared. DEFAULT -public static finalThriftServerRunner.ImplType DEFAULT +public static finalThriftServerRunner.ImplType DEFAULT @@ -334,7 +334,7
[13/22] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/34a201e5/testdevapidocs/index-all.html -- diff --git a/testdevapidocs/index-all.html b/testdevapidocs/index-all.html index 42630f4..4777def 100644 --- a/testdevapidocs/index-all.html +++ b/testdevapidocs/index-all.html @@ -364,6 +364,25 @@ AccessWriteAction(TableName) - Constructor for class org.apache.hadoop.hbase.client.TestSnapshotWithAcl.AccessWriteAction +AcidGuaranteesTestTool - Class in org.apache.hadoop.hbase + +A test tool that uses multiple threads to read and write multifamily rows into a table, verifying + that reads never see partially-complete writes + +AcidGuaranteesTestTool() - Constructor for class org.apache.hadoop.hbase.AcidGuaranteesTestTool + +AcidGuaranteesTestTool.AtomicGetReader - Class in org.apache.hadoop.hbase + +Thread that does single-row reads in a table, looking for partially completed rows. + +AcidGuaranteesTestTool.AtomicityWriter - Class in org.apache.hadoop.hbase + +Thread that does random full-row writes into a table. + +AcidGuaranteesTestTool.AtomicScanReader - Class in org.apache.hadoop.hbase + +Thread that does full scans of the table looking for any partially completed rows. + acquire() - Method in class org.apache.hadoop.hbase.master.procedure.TestMasterProcedureSchedulerConcurrency.TestTableProcSet acquireBarrier() - Method in class org.apache.hadoop.hbase.procedure.SimpleRSProcedureManager.SimpleSubprocedure @@ -612,6 +631,8 @@ addNumber(SetLong, long) - Static method in class org.apache.hadoop.hbase.util.TestByteBufferUtils +addOptions() - Method in class org.apache.hadoop.hbase.AcidGuaranteesTestTool + addOptions() - Method in class org.apache.hadoop.hbase.chaos.util.ChaosMonkeyRunner addOptions() - Method in class org.apache.hadoop.hbase.IntegrationTestBackupRestore @@ -1853,7 +1874,7 @@ AsyncTest(AsyncConnection, PerformanceEvaluation.TestOptions, PerformanceEvaluation.Status) - Constructor for class org.apache.hadoop.hbase.PerformanceEvaluation.AsyncTest -AtomicGetReader(MultithreadedTestUtil.TestContext, byte[], byte[][], ExecutorService) - Constructor for class org.apache.hadoop.hbase.TestAcidGuarantees.AtomicGetReader +AtomicGetReader(MultithreadedTestUtil.TestContext, byte[], byte[][], ExecutorService) - Constructor for class org.apache.hadoop.hbase.AcidGuaranteesTestTool.AtomicGetReader AtomicHFileLoader(TableName, MultithreadedTestUtil.TestContext, byte[][]) - Constructor for class org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad.AtomicHFileLoader @@ -1861,13 +1882,13 @@ AtomicHFileLoader(TableName, MultithreadedTestUtil.TestContext, byte[][]) - Constructor for class org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoadWithOldSecureEndpoint.AtomicHFileLoader -AtomicityWriter(MultithreadedTestUtil.TestContext, byte[][], byte[][], ExecutorService) - Constructor for class org.apache.hadoop.hbase.TestAcidGuarantees.AtomicityWriter +AtomicityWriter(MultithreadedTestUtil.TestContext, byte[][], byte[][], ExecutorService) - Constructor for class org.apache.hadoop.hbase.AcidGuaranteesTestTool.AtomicityWriter AtomicOperation(HRegion, int, AtomicLong, AtomicInteger) - Constructor for class org.apache.hadoop.hbase.regionserver.TestAtomicOperation.AtomicOperation -AtomicScanReader(TableName, MultithreadedTestUtil.TestContext, byte[][]) - Constructor for class org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad.AtomicScanReader +AtomicScanReader(MultithreadedTestUtil.TestContext, byte[][], ExecutorService) - Constructor for class org.apache.hadoop.hbase.AcidGuaranteesTestTool.AtomicScanReader -AtomicScanReader(MultithreadedTestUtil.TestContext, byte[][], ExecutorService) - Constructor for class org.apache.hadoop.hbase.TestAcidGuarantees.AtomicScanReader +AtomicScanReader(TableName, MultithreadedTestUtil.TestContext, byte[][]) - Constructor for class org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad.AtomicScanReader AUTH_A - Static variable in class org.apache.hadoop.hbase.security.access.TestZKPermissionWatcher @@ -5087,6 +5108,8 @@ compressTags - Variable in class org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers +compType - Variable in class org.apache.hadoop.hbase.TestAcidGuarantees + concat(byte[], int) - Method in class org.apache.hadoop.hbase.client.TestAsyncTable CONCURRENCY - Static variable in class org.apache.hadoop.hbase.client.TestAsyncNonMetaRegionLocatorConcurrenyLimit @@ -5453,8 +5476,6 @@ conf - Variable in class org.apache.hadoop.hbase.test.IntegrationTestLoadAndVerify.LoadMapper -conf - Variable in class org.apache.hadoop.hbase.TestAcidGuarantees - conf - Static variable in class org.apache.hadoop.hbase.TestJMXConnectorServer conf - Static variable in class org.apache.hadoop.hbase.TestStochasticBalancerJmxMetrics @@ -5691,6 +5712,12 @@ conn - Variable in class
[13/22] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6f0c8299/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.CreateCommand.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.CreateCommand.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.CreateCommand.html index df7d03c..59d8ee8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.CreateCommand.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.CreateCommand.html @@ -40,994 +40,1004 @@ 032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; 033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; 034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; -035 -036import java.io.IOException; -037import java.net.URI; -038import java.util.List; -039 -040import org.apache.commons.cli.CommandLine; -041import org.apache.commons.cli.HelpFormatter; -042import org.apache.commons.cli.Options; -043import org.apache.commons.lang.StringUtils; -044import org.apache.hadoop.conf.Configuration; -045import org.apache.hadoop.conf.Configured; -046import org.apache.hadoop.fs.FileSystem; -047import org.apache.hadoop.fs.Path; -048import org.apache.hadoop.hbase.HBaseConfiguration; -049import org.apache.hadoop.hbase.TableName; -050import org.apache.hadoop.hbase.backup.BackupAdmin; -051import org.apache.hadoop.hbase.backup.BackupInfo; -052import org.apache.hadoop.hbase.backup.BackupInfo.BackupState; -053import org.apache.hadoop.hbase.backup.BackupRequest; -054import org.apache.hadoop.hbase.backup.BackupRestoreConstants; -055import org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand; -056import org.apache.hadoop.hbase.backup.BackupType; -057import org.apache.hadoop.hbase.backup.util.BackupSet; -058import org.apache.hadoop.hbase.backup.util.BackupUtils; -059import org.apache.hadoop.hbase.classification.InterfaceAudience; -060import org.apache.hadoop.hbase.client.Connection; -061import org.apache.hadoop.hbase.client.ConnectionFactory; -062import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists; -063import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -064 -065/** -066 * General backup commands, options and usage messages -067 */ -068 -069@InterfaceAudience.Private -070public final class BackupCommands { -071 -072 public final static String INCORRECT_USAGE = "Incorrect usage"; +035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; +036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; +037 +038import java.io.IOException; +039import java.net.URI; +040import java.util.List; +041 +042import org.apache.commons.cli.CommandLine; +043import org.apache.commons.cli.HelpFormatter; +044import org.apache.commons.cli.Options; +045import org.apache.commons.lang.StringUtils; +046import org.apache.hadoop.conf.Configuration; +047import org.apache.hadoop.conf.Configured; +048import org.apache.hadoop.fs.FileSystem; +049import org.apache.hadoop.fs.Path; +050import org.apache.hadoop.hbase.HBaseConfiguration; +051import org.apache.hadoop.hbase.TableName; +052import org.apache.hadoop.hbase.backup.BackupAdmin; +053import org.apache.hadoop.hbase.backup.BackupInfo; +054import org.apache.hadoop.hbase.backup.BackupInfo.BackupState; +055import org.apache.hadoop.hbase.backup.BackupRequest; +056import org.apache.hadoop.hbase.backup.BackupRestoreConstants; +057import org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand; +058import org.apache.hadoop.hbase.backup.BackupType; +059import org.apache.hadoop.hbase.backup.util.BackupSet; +060import org.apache.hadoop.hbase.backup.util.BackupUtils; +061import org.apache.hadoop.hbase.classification.InterfaceAudience; +062import org.apache.hadoop.hbase.client.Connection; +063import org.apache.hadoop.hbase.client.ConnectionFactory; +064import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists; +065import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +066 +067/** +068 * General backup commands, options and usage messages +069 */ +070 +071@InterfaceAudience.Private +072public final class BackupCommands { 073 -074 public static final String USAGE = "Usage: hbase backup COMMAND [command-specific arguments]\n" -075 + "where COMMAND is one of:\n" + " create create a new backup image\n" -076 + " delete delete an existing backup image\n" -077 + " describe show the detailed information of a backup image\n" -078 + " historyshow history of all successful backups\n" -079 + " progress show the progress of the latest backup request\n" -080 + " setbackup set management\n" -081 + "
[13/22] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9830530/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.RegionCloseOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.RegionCloseOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.RegionCloseOperation.html index f0a9b50..13cebd8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.RegionCloseOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.RegionCloseOperation.html @@ -69,7 +69,7 @@ 061 "hbase.regionserver.rpc.startup.waittime"; 062 private static final int DEFAULT_RS_RPC_STARTUP_WAIT_TIME = 6; 063 -064 private static final int RS_VERSION_WITH_EXEC_PROCS = 0x0201000; // 2.1 +064 private static final int RS_VERSION_WITH_EXEC_PROCS = 0x020; // 2.0 065 066 protected final MasterServices master; 067 protected final long rsStartupWaitTime; @@ -267,286 +267,289 @@ 259 final MasterProcedureEnv env = master.getMasterProcedureExecutor().getEnvironment(); 260 261 request = ExecuteProceduresRequest.newBuilder(); -262 splitAndResolveOperation(getServerName(), operations, this); -263 -264 try { -265final ExecuteProceduresResponse response = sendRequest(getServerName(), request.build()); -266remoteCallCompleted(env, response); -267 } catch (IOException e) { -268e = unwrapException(e); -269// TODO: In the future some operation may want to bail out early. -270// TODO: How many times should we retry (use numberOfAttemptsSoFar) -271if (!scheduleForRetry(e)) { -272 remoteCallFailed(env, e); -273} -274 } -275 return null; -276} -277 -278public void dispatchOpenRequests(final MasterProcedureEnv env, -279final ListRegionOpenOperation operations) { -280 request.addOpenRegion(buildOpenRegionRequest(env, getServerName(), operations)); -281} -282 -283public void dispatchCloseRequests(final MasterProcedureEnv env, -284final ListRegionCloseOperation operations) { -285 for (RegionCloseOperation op: operations) { -286 request.addCloseRegion(op.buildCloseRegionRequest(getServerName())); -287 } -288} -289 -290protected ExecuteProceduresResponse sendRequest(final ServerName serverName, -291final ExecuteProceduresRequest request) throws IOException { -292 try { -293return getRsAdmin().executeProcedures(null, request); -294 } catch (ServiceException se) { -295throw ProtobufUtil.getRemoteException(se); -296 } -297} -298 -299 -300private void remoteCallCompleted(final MasterProcedureEnv env, -301final ExecuteProceduresResponse response) { -302 /* -303 for (RemoteProcedure proc: operations) { -304proc.remoteCallCompleted(env, getServerName(), response); -305 }*/ -306} -307 -308private void remoteCallFailed(final MasterProcedureEnv env, final IOException e) { -309 for (RemoteProcedure proc: operations) { -310proc.remoteCallFailed(env, getServerName(), e); -311 } -312} -313 } -314 -315 // == -316 // Compatibility calls -317 // Since we don't have a "batch proc-exec" request on the target RS -318 // we have to chunk the requests by type and dispatch the specific request. -319 // == -320 private static OpenRegionRequest buildOpenRegionRequest(final MasterProcedureEnv env, -321 final ServerName serverName, final ListRegionOpenOperation operations) { -322final OpenRegionRequest.Builder builder = OpenRegionRequest.newBuilder(); -323 builder.setServerStartCode(serverName.getStartcode()); -324 builder.setMasterSystemTime(EnvironmentEdgeManager.currentTime()); -325for (RegionOpenOperation op: operations) { -326 builder.addOpenInfo(op.buildRegionOpenInfoRequest(env)); -327} -328return builder.build(); -329 } -330 -331 private final class OpenRegionRemoteCall extends AbstractRSRemoteCall { -332private final ListRegionOpenOperation operations; +262 if (LOG.isTraceEnabled()) { +263LOG.trace("Building request with operations count=" + operations.size()); +264 } +265 splitAndResolveOperation(getServerName(), operations, this); +266 +267 try { +268final ExecuteProceduresResponse response = sendRequest(getServerName(), request.build()); +269remoteCallCompleted(env, response); +270 } catch (IOException e) { +271e = unwrapException(e); +272// TODO: In the future some operation may want to bail out
[13/22] hbase-site git commit: Published site at da5fb27eabed4a4b4d251be973ee945fb52895bf.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7b1830cf/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.html b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.html index 2a09edb..b6fb3b4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/snapshot/ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.html @@ -37,131 +37,131 @@ 029import java.util.Comparator; 030import java.util.LinkedList; 031import java.util.List; -032import java.util.Random; -033 -034import org.apache.commons.cli.CommandLine; -035import org.apache.commons.cli.Option; -036import org.apache.commons.logging.Log; -037import org.apache.commons.logging.LogFactory; -038import org.apache.hadoop.hbase.classification.InterfaceAudience; -039import org.apache.hadoop.conf.Configuration; -040import org.apache.hadoop.fs.FSDataInputStream; -041import org.apache.hadoop.fs.FSDataOutputStream; -042import org.apache.hadoop.fs.FileChecksum; -043import org.apache.hadoop.fs.FileStatus; -044import org.apache.hadoop.fs.FileSystem; -045import org.apache.hadoop.fs.FileUtil; -046import org.apache.hadoop.fs.Path; -047import org.apache.hadoop.fs.permission.FsPermission; -048import org.apache.hadoop.hbase.TableName; -049import org.apache.hadoop.hbase.HBaseConfiguration; -050import org.apache.hadoop.hbase.HConstants; -051import org.apache.hadoop.hbase.HRegionInfo; -052import org.apache.hadoop.hbase.io.FileLink; -053import org.apache.hadoop.hbase.io.HFileLink; -054import org.apache.hadoop.hbase.io.WALLink; -055import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -056import org.apache.hadoop.hbase.mob.MobUtils; -057import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescription; -058import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo; -059import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; -060import org.apache.hadoop.hbase.util.AbstractHBaseTool; -061import org.apache.hadoop.hbase.util.FSUtils; -062import org.apache.hadoop.hbase.util.HFileArchiveUtil; -063import org.apache.hadoop.hbase.util.Pair; -064import org.apache.hadoop.io.BytesWritable; -065import org.apache.hadoop.io.IOUtils; -066import org.apache.hadoop.io.NullWritable; -067import org.apache.hadoop.io.Writable; -068import org.apache.hadoop.mapreduce.Job; -069import org.apache.hadoop.mapreduce.JobContext; -070import org.apache.hadoop.mapreduce.Mapper; -071import org.apache.hadoop.mapreduce.InputFormat; -072import org.apache.hadoop.mapreduce.InputSplit; -073import org.apache.hadoop.mapreduce.RecordReader; -074import org.apache.hadoop.mapreduce.TaskAttemptContext; -075import org.apache.hadoop.mapreduce.lib.output.NullOutputFormat; -076import org.apache.hadoop.mapreduce.security.TokenCache; -077import org.apache.hadoop.hbase.io.hadoopbackport.ThrottledInputStream; -078import org.apache.hadoop.util.StringUtils; -079import org.apache.hadoop.util.Tool; -080 -081/** -082 * Export the specified snapshot to a given FileSystem. -083 * -084 * The .snapshot/name folder is copied to the destination cluster -085 * and then all the hfiles/wals are copied using a Map-Reduce Job in the .archive/ location. -086 * When everything is done, the second cluster can restore the snapshot. -087 */ -088@InterfaceAudience.Public -089public class ExportSnapshot extends AbstractHBaseTool implements Tool { -090 public static final String NAME = "exportsnapshot"; -091 /** Configuration prefix for overrides for the source filesystem */ -092 public static final String CONF_SOURCE_PREFIX = NAME + ".from."; -093 /** Configuration prefix for overrides for the destination filesystem */ -094 public static final String CONF_DEST_PREFIX = NAME + ".to."; -095 -096 private static final Log LOG = LogFactory.getLog(ExportSnapshot.class); -097 -098 private static final String MR_NUM_MAPS = "mapreduce.job.maps"; -099 private static final String CONF_NUM_SPLITS = "snapshot.export.format.splits"; -100 private static final String CONF_SNAPSHOT_NAME = "snapshot.export.format.snapshot.name"; -101 private static final String CONF_SNAPSHOT_DIR = "snapshot.export.format.snapshot.dir"; -102 private static final String CONF_FILES_USER = "snapshot.export.files.attributes.user"; -103 private static final String CONF_FILES_GROUP = "snapshot.export.files.attributes.group"; -104 private static final String CONF_FILES_MODE = "snapshot.export.files.attributes.mode"; -105 private static final