Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 5C319200CC4 for ; Wed, 28 Jun 2017 17:01:17 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 5AADF160BFF; Wed, 28 Jun 2017 15:01:17 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 146F0160C02 for ; Wed, 28 Jun 2017 17:01:14 +0200 (CEST) Received: (qmail 65510 invoked by uid 500); 28 Jun 2017 15:01:09 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 64342 invoked by uid 99); 28 Jun 2017 15:01:08 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 28 Jun 2017 15:01:08 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id BC603E96DA; Wed, 28 Jun 2017 15:01:06 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: git-site-role@apache.org To: commits@hbase.apache.org Date: Wed, 28 Jun 2017 15:01:35 -0000 Message-Id: In-Reply-To: <4a3527e9985f440e86627a5c1aa5b2df@git.apache.org> References: <4a3527e9985f440e86627a5c1aa5b2df@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [31/38] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd. archived-at: Wed, 28 Jun 2017 15:01:17 -0000 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2848f431/devapidocs/index-all.html ---------------------------------------------------------------------- diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html index 0cf4600..edffa3b 100644 --- a/devapidocs/index-all.html +++ b/devapidocs/index-all.html @@ -6384,8 +6384,10 @@
 
BLOCKSIZE_BYTES - Static variable in class org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder
 
-
blockSizeMap - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
+
blockSizeDetails - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
+
Serialize column family to block size map to configuration.
+
blockSizeWritten() - Method in class org.apache.hadoop.hbase.io.hfile.HFileBlock.Writer
Returns the number of bytes written into the current block so far, or @@ -6567,8 +6569,10 @@
The type of bloom
-
bloomTypeMap - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
+
bloomTypeDetails - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
+
Serialize column family to bloom type map to configuration.
+
bName - Variable in class org.apache.hadoop.hbase.NamespaceDescriptor.Builder
 
body - Variable in class org.apache.hadoop.hbase.rest.client.Response
@@ -12203,10 +12207,6 @@
 
close() - Method in class org.apache.hadoop.hbase.mapreduce.HFileInputFormat.HFileRecordReader
 
-
close(StoreFileWriter) - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
-
close(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
close() - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSerialization.KeyValueDeserializer
 
close() - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSerialization.KeyValueSerializer
@@ -13401,6 +13401,8 @@
 
CombinedCacheStats(CacheStats, CacheStats) - Constructor for class org.apache.hadoop.hbase.io.hfile.CombinedBlockCache.CombinedCacheStats
 
+
combineTableNameSuffix(byte[], byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
 
combineWith(CompactionRequest) - Method in class org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest
Combines the request with other request.
@@ -13796,8 +13798,6 @@
 
CompactionDelPartitionId(byte[], byte[]) - Constructor for class org.apache.hadoop.hbase.mob.compactions.PartitionedMobCompactionRequest.CompactionDelPartitionId
 
-
compactionExclude - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
compactionInputFileCountHisto - Variable in class org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceImpl
 
CompactionInputFormat() - Constructor for class org.apache.hadoop.hbase.regionserver.CompactionTool.CompactionInputFormat
@@ -14650,8 +14650,6 @@
 
compareTo(Import.KeyValueWritableComparable) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueWritableComparable
 
-
compareTo(MultiTableHFileOutputFormat.MultiHFilePartitioner.TableSplitKeyPair) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner.TableSplitKeyPair
-
 
compareTo(TableSplit) - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
Compares this split against the given one.
@@ -15180,8 +15178,10 @@
compressionContext - Variable in class org.apache.hadoop.hbase.wal.WALKey
 
-
compressionMap - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
+
compressionDetails - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
+
Serialize column family to compression algorithm map to configuration.
+
CompressionState - Class in org.apache.hadoop.hbase.io.encoding
Stores the state of data block encoder at the beginning of new key.
@@ -15476,12 +15476,8 @@
 
conf - Variable in class org.apache.hadoop.hbase.mapreduce.HFileInputFormat.HFileRecordReader
 
-
conf - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
conf - Variable in class org.apache.hadoop.hbase.mapreduce.HRegionPartitioner
 
-
conf - Variable in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
 
conf - Variable in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormat
The configuration.
@@ -15834,27 +15830,11 @@
 
configure(JobConf) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormat
 
-
configureBlockSize(HTableDescriptor, Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
-
Serialize column family to block size map to configuration.
-
-
configureBloomType(HTableDescriptor, Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
-
Serialize column family to bloom type map to configuration.
-
configureCfRenaming(Configuration, Map<String, String>) - Static method in class org.apache.hadoop.hbase.mapreduce.Import
Sets a configuration property with key Import.CF_RENAME_PROP in conf that tells the mapper how to rename column families.
-
configureCompression(Configuration, HTableDescriptor) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
-
Serialize column family to compression algorithm map to configuration.
-
-
configureDataBlockEncoding(HTableDescriptor, Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
-
Serialize column family to data block encoding map to configuration.
-
configuredReadTableTimeouts - Variable in class org.apache.hadoop.hbase.tool.Canary
 
configuredReadTableTimeouts - Variable in class org.apache.hadoop.hbase.tool.Canary.RegionMonitor
@@ -15863,11 +15843,6 @@
 
configuredWriteTableTimeout - Variable in class org.apache.hadoop.hbase.tool.Canary.RegionMonitor
 
-
configureForOneTable(Configuration, HTableDescriptor) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
-
Configure for one table, should be used before creating a new HFileRecordWriter, - Set compression algorithms and related configuration based on column families
-
configureForRegion(HRegion) - Method in class org.apache.hadoop.hbase.regionserver.BusyRegionSplitPolicy
 
configureForRegion(HRegion) - Method in class org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy
@@ -15907,24 +15882,15 @@
Configure a MapReduce Job to perform an incremental load into the given table.
-
configureIncrementalLoad(Job, HTableDescriptor, RegionLocator, Class<? extends OutputFormat<?, ?>>) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
 
-
configureIncrementalLoad(Job, List<TableName>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
-
Configure a MapReduce Job to output HFiles for performing an incremental load into - the multiple tables.
-
-
configureIncrementalLoad(Job, List<TableName>, Class<? extends OutputFormat<?, ?>>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
+
configureIncrementalLoad(Job, List<HFileOutputFormat2.TableInfo>, Class<? extends OutputFormat<?, ?>>) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
 
-
configureIncrementalLoad(Job, Map<ImmutableBytesWritable, List<ImmutableBytesWritable>>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
+
configureIncrementalLoad(Job, List<HFileOutputFormat2.TableInfo>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
Same purpose as configureIncrementalLoad(Job job, List tables) - Used when region startKeys of each table is available, input as > - - Caller needs to transfer TableName and byte[] to ImmutableBytesWritable
+
Analogous to + HFileOutputFormat2.configureIncrementalLoad(Job, HTableDescriptor, RegionLocator), + this function will configure the requisite number of reducers to write HFiles for multple + tables simultaneously
-
configureIncrementalLoad(Job, Map<ImmutableBytesWritable, List<ImmutableBytesWritable>>, Class<? extends OutputFormat<?, ?>>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
 
configureIncrementalLoadMap(Job, HTableDescriptor) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
 
configureInfoServer() - Method in class org.apache.hadoop.hbase.master.HMaster
@@ -15939,7 +15905,7 @@
Job configuration.
-
configurePartitioner(Job, List<ImmutableBytesWritable>) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
configurePartitioner(Job, List<ImmutableBytesWritable>, boolean) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
Configure job with a TotalOrderPartitioner, partitioning against splitPoints.
@@ -16635,8 +16601,6 @@
 
Context() - Constructor for class org.apache.hadoop.hbase.ipc.RpcScheduler.Context
 
-
context - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
context - Variable in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
 
context - Variable in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionRecordReader
@@ -18353,6 +18317,20 @@
 
createCompositeImmutableSegment(CellComparator, List<ImmutableSegment>) - Method in class org.apache.hadoop.hbase.regionserver.SegmentFactory
 
+
createCompositeKey(byte[], byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
+
+
Creates a composite key to use as a mapper output key when using + MultiTableHFileOutputFormat.configureIncrementaLoad to set up bulk ingest job
+
+
createCompositeKey(byte[], ImmutableBytesWritable) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
+
+
Alternate api which accepts an ImmutableBytesWritable for the suffix
+
+
createCompositeKey(String, ImmutableBytesWritable) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
+
+
Alternate api which accepts a String for the tableName and ImmutableBytesWritable for the + suffix
+
createCompressionStream(OutputStream, Compressor, int) - Method in enum org.apache.hadoop.hbase.io.compress.Compression.Algorithm
 
createConfigurationSubset() - Method in class org.apache.hadoop.hbase.master.MasterRpcServices
@@ -19029,8 +19007,6 @@
 
createMoveRegionProcedure(RegionPlan) - Method in class org.apache.hadoop.hbase.master.assignment.AssignmentManager
 
-
createMultiHFileRecordWriter(TaskAttemptContext) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
 
createMutableSegment(Configuration, CellComparator) - Method in class org.apache.hadoop.hbase.regionserver.SegmentFactory
 
createMyEphemeralNode() - Method in class org.apache.hadoop.hbase.regionserver.HRegionServer
@@ -20621,8 +20597,10 @@
 
dataBlockEncodingCtx - Variable in class org.apache.hadoop.hbase.io.hfile.HFileBlock.Writer
 
-
datablockEncodingMap - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
+
dataBlockEncodingDetails - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
+
Serialize column family to data block encoding map to configuration.
+
dataBlockIndexReader - Variable in class org.apache.hadoop.hbase.io.hfile.HFileReaderImpl
Data block index reader keeping the root data index in memory
@@ -22212,8 +22190,6 @@
 
DEFAULT_OVERLAPS_TO_SIDELINE - Static variable in class org.apache.hadoop.hbase.util.HBaseFsck
 
-
DEFAULT_PATH - Static variable in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
 
DEFAULT_PERIODIC_ROLL - Static variable in class org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore
 
DEFAULT_PREFETCH_BLOCKS_ON_OPEN - Static variable in class org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder
@@ -22603,8 +22579,6 @@
DefaultCompactor(Configuration, Store) - Constructor for class org.apache.hadoop.hbase.regionserver.compactions.DefaultCompactor
 
-
defaultCompression - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
defaultContexts - Variable in class org.apache.hadoop.hbase.http.HttpServer
 
defaultDecodingCtx - Variable in class org.apache.hadoop.hbase.io.hfile.HFileBlock.FSReaderImpl
@@ -31150,8 +31124,6 @@
 
fs - Variable in class org.apache.hadoop.hbase.io.hfile.HFile.WriterFactory
 
-
fs - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
fs - Variable in class org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles
 
fs - Variable in class org.apache.hadoop.hbase.master.cleaner.CleanerChore
@@ -34695,8 +34667,6 @@
Returns the current configuration.
-
getConf() - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
 
getConf() - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormat
Returns the current configuration.
@@ -37926,6 +37896,8 @@
Return HStore instance.
+
getHTableDescriptor() - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.TableInfo
+
 
getHTableDescriptors(List<TableName>) - Method in class org.apache.hadoop.hbase.util.HBaseFsck
 
getHTableMultiplexerStatus() - Method in class org.apache.hadoop.hbase.client.HTableMultiplexer
@@ -41250,8 +41222,6 @@
getNewRpcRetryingCallerFactory(Configuration) - Method in class org.apache.hadoop.hbase.client.ConnectionImplementation
 
-
getNewWriter(byte[], Configuration, InetSocketAddress[]) - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
getNextBlockOnDiskSize() - Method in class org.apache.hadoop.hbase.io.hfile.HFileBlock
 
getNextCallSeq() - Method in class org.apache.hadoop.hbase.regionserver.RSRpcServices.RegionScannerHolder
@@ -42437,14 +42407,8 @@
getPartition(Import.KeyValueWritableComparable, KeyValue, int) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueWritableComparablePartitioner
 
-
getPartition(ImmutableBytesWritable, Cell, int) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
 
getPartition(ImmutableBytesWritable, VALUE, int) - Method in class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
 
-
getPartitionFile(Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
-
Get the path to the SequenceFile storing the sorted .
-
getPartitionId() - Method in class org.apache.hadoop.hbase.mob.compactions.PartitionedMobCompactionRequest.CompactionPartition
 
getPassword(Configuration, String, String) - Static method in class org.apache.hadoop.hbase.HBaseConfiguration
@@ -43822,8 +43786,6 @@
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
 
-
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
@@ -43848,10 +43810,6 @@
 
getRecoveryMode() - Method in class org.apache.hadoop.hbase.master.SplitLogManager
 
-
getReducerNumber(Map<ImmutableBytesWritable, List<ImmutableBytesWritable>>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
-
Get the number of reducers by tables' split keys
-
getRefCellTags() - Method in class org.apache.hadoop.hbase.regionserver.HMobStore
 
getRefCount(BlockCacheKey) - Method in class org.apache.hadoop.hbase.io.hfile.bucket.BucketCache
@@ -44277,6 +44235,8 @@
 
getRegionLocator() - Method in class org.apache.hadoop.hbase.client.HTable
 
+
getRegionLocator() - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.TableInfo
+
 
getRegionLocator() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
Allows subclasses to get the RegionLocator.
@@ -44612,7 +44572,7 @@
 
getRegionsStateInTransition() - Method in class org.apache.hadoop.hbase.master.assignment.RegionStates
 
-
getRegionStartKeys(RegionLocator) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
getRegionStartKeys(List<RegionLocator>, boolean) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
Return the start keys of all of the regions in this table, as a list of ImmutableBytesWritable.
@@ -47894,6 +47854,8 @@
 
getSuccessfulCoalescings() - Method in interface org.apache.hadoop.hbase.thrift.IncrementCoalescerMBean
 
+
getSuffix(byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
+
 
getSuffix() - Method in class org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.Builder
@@ -48055,10 +48017,6 @@
getTable() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
 
-
getTable(byte[], Connection, Admin) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
-
Check if table exist, should not dependent on HBase instance
-
getTable() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
Allows subclasses to get the Table.
@@ -48401,6 +48359,8 @@
getTableName() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
 
+
getTableName(byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
+
 
getTableName() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
Returns the table name converted to a byte array.
@@ -48531,6 +48491,8 @@
 
getTableNamespaceManager(MasterProcedureEnv) - Method in class org.apache.hadoop.hbase.master.procedure.ModifyNamespaceProcedure
 
+
getTableNameSuffixedWithFamily(byte[], byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
 
getTableNameTag(Cell) - Static method in class org.apache.hadoop.hbase.mob.MobUtils
Gets the table name tag.
@@ -48746,10 +48708,6 @@
 
getTableSplits(TableName) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin
 
-
getTablesRegionStartKeys(Configuration, List<TableName>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
-
Return map of
-
getTableStartRowForMeta(Optional<TableName>, MetaTableAccessor.QueryType) - Static method in class org.apache.hadoop.hbase.AsyncMetaTableAccessor
 
getTableStartRowForMeta(TableName, MetaTableAccessor.QueryType) - Static method in class org.apache.hadoop.hbase.MetaTableAccessor
@@ -54023,7 +53981,7 @@
HFileOutputFormat2() - Constructor for class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
 
-
HFileOutputFormat2.HFileRecordWriter<V extends Cell> - Class in org.apache.hadoop.hbase.mapreduce
+
HFileOutputFormat2.TableInfo - Class in org.apache.hadoop.hbase.mapreduce
 
HFileOutputFormat2.WriterLength - Class in org.apache.hadoop.hbase.mapreduce
 
@@ -54075,10 +54033,6 @@
HFileRecordReader() - Constructor for class org.apache.hadoop.hbase.mapreduce.HFileInputFormat.HFileRecordReader
 
-
HFileRecordWriter(TaskAttemptContext, Path) - Constructor for class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
-
Mapredue job will create a temp path for outputting results.
-
hfileRefsZNode - Variable in class org.apache.hadoop.hbase.replication.ReplicationStateZKBase
The name of the znode that contains queues of hfile references to be replicated
@@ -54729,6 +54683,8 @@
Deprecated.
 
+
hTableDescriptor - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.TableInfo
+
 
hTableDescriptor - Variable in class org.apache.hadoop.hbase.master.procedure.CloneSnapshotProcedure
 
hTableDescriptor - Variable in class org.apache.hadoop.hbase.master.procedure.CreateTableProcedure
@@ -68205,8 +68161,6 @@
maxSize - Variable in class org.apache.hadoop.hbase.io.hfile.LruCachedBlockQueue
 
-
maxsize - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
maxSize - Variable in class org.apache.hadoop.hbase.util.PoolMap.ReusablePool
 
maxSize - Variable in class org.apache.hadoop.hbase.util.PoolMap.RoundRobinPool
@@ -71394,6 +71348,8 @@
MULTI_BYTE_BUFFER_HEAP_SIZE - Static variable in class org.apache.hadoop.hbase.io.hfile.HFileBlock
 
+
MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
+
 
multiAction - Variable in class org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.SingleServerRequestRunnable
 
MultiAction - Class in org.apache.hadoop.hbase.client
@@ -71446,8 +71402,6 @@
Create multiple Connection instances and initialize a thread pool executor
-
MultiHFilePartitioner() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
 
multiMutate(Connection, Table, byte[], Mutation...) - Static method in class org.apache.hadoop.hbase.MetaTableAccessor
 
multiMutate(Connection, Table, byte[], List<Mutation>) - Static method in class org.apache.hadoop.hbase.MetaTableAccessor
@@ -71545,32 +71499,20 @@
 
MultiTableHFileOutputFormat - Class in org.apache.hadoop.hbase.mapreduce
-
Create 3 level tree directory, first level is using table name as parent directory and then use - family name as child directory, and all related HFiles for one family are under child directory +
Create 3 level tree directory, first level is using table name as parent + directory and then use family name as child directory, and all related HFiles + for one family are under child directory -tableName1 - -columnFamilyName1 - -HFile (region1) - -columnFamilyName2 - -HFile1 (region1) - -HFile2 (region2) - -HFile3 (region3) + -columnFamilyName1 + -columnFamilyName2 + -HFiles -tableName2 - -columnFamilyName1 - -HFile (region1) - family directory and its hfiles match the output of HFileOutputFormat2
+ -columnFamilyName1 + -HFiles + -columnFamilyName2
MultiTableHFileOutputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
 
-
MultiTableHFileOutputFormat.MultiHFilePartitioner - Class in org.apache.hadoop.hbase.mapreduce
-
-
MultiTableHFileOutputFormat writes files based on partitions created by MultiHFilePartitioner - The input is partitioned based on table's name and its region boundaries with the table.
-
-
MultiTableHFileOutputFormat.MultiHFilePartitioner.TableSplitKeyPair - Class in org.apache.hadoop.hbase.mapreduce
-
-
A class store pair, has two main usage - 1.
-
MultiTableInputFormat - Class in org.apache.hadoop.hbase.mapreduce
Convert HBase tabular data from multiple scanners into a format that @@ -73923,8 +73865,6 @@
now - Variable in class org.apache.hadoop.hbase.io.hfile.BlockCacheUtil.CachedBlocksByFile
 
-
now - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
now - Variable in class org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher
 
now - Variable in class org.apache.hadoop.hbase.regionserver.StoreScanner
@@ -75986,16 +75926,12 @@ service.
 
outputArrayOffsets - Variable in class org.apache.hadoop.hbase.codec.prefixtree.encode.column.ColumnSectionWriter
 
-
outputDir - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
outputFileSize - Variable in class org.apache.hadoop.hbase.regionserver.HStore.StoreFlusherImpl
 
outputFs - Variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot.ExportMapper
 
outputJSON - Variable in class org.apache.hadoop.hbase.wal.WALPrettyPrinter
 
-
outputPath - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
outputRoot - Variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot.ExportMapper
 
outputRoot - Variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot
@@ -76058,8 +75994,6 @@ service.
Returns true if the first range start1...end1 overlaps with the second range start2...end2, assuming the byte arrays represent row keys
-
overriddenEncoding - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
overwrite - Variable in class org.apache.hadoop.hbase.backup.RestoreRequest
 
OVERWRITE - Static variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot.Options
@@ -76797,12 +76731,8 @@ service.
 
PartitionedMobCompactor.DelPartitionComparator - Class in org.apache.hadoop.hbase.mob.compactions
 
-
PARTITIONER_PATH - Static variable in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
 
partitionId - Variable in class org.apache.hadoop.hbase.mob.compactions.PartitionedMobCompactionRequest.CompactionPartition
 
-
partitionMap - Variable in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
 
partitions - Variable in class org.apache.hadoop.hbase.mapreduce.HashTable.TableHash
 
PARTITIONS_FILE_NAME - Static variable in class org.apache.hadoop.hbase.mapreduce.HashTable
@@ -81011,8 +80941,6 @@ service.
Try to position the scanner in the row before the current row.
-
previousRow - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
previousRowInternal() - Method in class org.apache.hadoop.hbase.codec.prefixtree.decode.PrefixTreeArrayReversibleScanner
 
previousState - Variable in class org.apache.hadoop.hbase.master.procedure.ServerCrashProcedure
@@ -84466,10 +84394,6 @@ service.
readTableDescriptor(FileSystem, FileStatus) - Static method in class org.apache.hadoop.hbase.util.FSTableDescriptors
 
-
readTableSplitKeys(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat.MultiHFilePartitioner
-
-
read partition file into map
-
readThreads - Variable in class org.apache.hadoop.hbase.ipc.SimpleRpcServer
 
readTimestamp(ByteBuffer) - Method in class org.apache.hadoop.hbase.io.encoding.CompressionState
@@ -85742,6 +85666,8 @@ service.
regionLocator - Variable in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
 
+
regionLocator - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.TableInfo
+
 
regionLocator - Variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
The RegionLocator of the table.
@@ -91245,8 +91171,6 @@ service.
 
rollperiod - Variable in class org.apache.hadoop.hbase.regionserver.LogRoller
 
-
rollRequested - Variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
rollRetries - Variable in class org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore
 
rollThreshold - Variable in class org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore
@@ -91289,8 +91213,6 @@ service.
This lock makes sure only one log roll runs at a time.
-
rollWriters() - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.HFileRecordWriter
-
 
rollWriterWithRetries() - Method in class org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore
 
root - Variable in class org.apache.hadoop.hbase.codec.prefixtree.encode.tokenize.Tokenizer
@@ -94736,6 +94658,8 @@ service.
Write trailer data as protobuf
+
serializeColumnFamilyAttribute