Return-Path:
List<String>
Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer>
run(String dirPath,
Map<byte[],List<org.apache.hadoop.fs.Path>> map,
TableName tableName)
List<String>
Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer>
doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Admin admin,
Table table,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f17356a7/apidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
----------------------------------------------------------------------
diff --git a/apidocs/org/apache/hadoop/hbase/client/class-use/Admin.html b/apidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
index db081bc..76328a1 100644
--- a/apidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
+++ b/apidocs/org/apache/hadoop/hbase/client/class-use/Admin.html
@@ -147,7 +147,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods.
List<String>
Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer>
doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Admin admin,
Table table,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f17356a7/apidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
----------------------------------------------------------------------
diff --git a/apidocs/org/apache/hadoop/hbase/client/class-use/Connection.html b/apidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
index 106f0d4..78bf964 100644
--- a/apidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
+++ b/apidocs/org/apache/hadoop/hbase/client/class-use/Connection.html
@@ -206,12 +206,13 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods.
protected void
-LoadIncrementalHFiles.bulkLoadPhase(Table table,
+LoadIncrementalHFiles.bulkLoadPhase(Table table,
Connection conn,
ExecutorService pool,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups,
- boolean copyFile)
+ boolean copyFile,
+ Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> item2RegionMap)
This takes the LQI's grouped by likely regions and attempts to bulk load
them.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f17356a7/apidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
----------------------------------------------------------------------
diff --git a/apidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html b/apidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
index c5f06a1..cd36c53 100644
--- a/apidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
+++ b/apidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html
@@ -165,7 +165,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods.
-List<String>
+Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer>
LoadIncrementalHFiles.doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Admin admin,
Table table,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f17356a7/apidocs/org/apache/hadoop/hbase/client/class-use/Table.html
----------------------------------------------------------------------
diff --git a/apidocs/org/apache/hadoop/hbase/client/class-use/Table.html b/apidocs/org/apache/hadoop/hbase/client/class-use/Table.html
index ecc24f7..f109f4a 100644
--- a/apidocs/org/apache/hadoop/hbase/client/class-use/Table.html
+++ b/apidocs/org/apache/hadoop/hbase/client/class-use/Table.html
@@ -203,12 +203,13 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods.
protected void
-LoadIncrementalHFiles.bulkLoadPhase(Table table,
+LoadIncrementalHFiles.bulkLoadPhase(Table table,
Connection conn,
ExecutorService pool,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups,
- boolean copyFile)
+ boolean copyFile,
+ Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> item2RegionMap)
This takes the LQI's grouped by likely regions and attempts to bulk load
them.
@@ -223,7 +224,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods.
-List<String>
+Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer>
LoadIncrementalHFiles.doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Admin admin,
Table table,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f17356a7/apidocs/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.html
----------------------------------------------------------------------
diff --git a/apidocs/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.html b/apidocs/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.html
index a220262..c3e8882 100644
--- a/apidocs/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.html
+++ b/apidocs/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.html
@@ -196,18 +196,19 @@ implements org.apache.hadoop.util.Tool
protected void
-bulkLoadPhase(Table table,
+bulkLoadPhase(Table table,
Connection conn,
ExecutorService pool,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups,
- boolean copyFile)
+ boolean copyFile,
+ Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> item2RegionMap)
This takes the LQI's grouped by likely regions and attempts to bulk load
them.
-List<String>
+Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer>
doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Admin admin,
Table table,
@@ -315,7 +316,7 @@ implements org.apache.hadoop.util.Tool
run(String[] args)
-List<String>
+Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer>
run(String dirPath,
Map<byte[],List<org.apache.hadoop.fs.Path>> map,
TableName tableName)
@@ -507,14 +508,14 @@ implements org.apache.hadoop.util.Tool
-
doBulkLoad
-public List<String> doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
- Admin admin,
- Table table,
- RegionLocator regionLocator,
- boolean silence,
- boolean copyFile)
- throws TableNotFoundException,
- IOException
+public Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
+ Admin admin,
+ Table table,
+ RegionLocator regionLocator,
+ boolean silence,
+ boolean copyFile)
+ throws TableNotFoundException,
+ IOException
Perform a bulk load of the given directory into the given
pre-existing table. This method is not threadsafe.
@@ -526,7 +527,7 @@ implements org.apache.hadoop.util.Tool
silence
- true to ignore unmatched column families
copyFile
- always copy hfiles if true
- Returns:
-- List of filenames which were not found
+- Map of LoadQueueItem to region
- Throws:
TableNotFoundException
- if table does not yet exist
IOException
@@ -570,7 +571,7 @@ implements org.apache.hadoop.util.Tool
-
prepareHFileQueue
-public void prepareHFileQueue(org.apache.hadoop.fs.Path hfilesDir,
+public void prepareHFileQueue(org.apache.hadoop.fs.Path hfilesDir,
Table table,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
boolean validateHFile)
@@ -595,7 +596,7 @@ implements org.apache.hadoop.util.Tool
-
prepareHFileQueue
-public void prepareHFileQueue(org.apache.hadoop.fs.Path hfilesDir,
+public void prepareHFileQueue(org.apache.hadoop.fs.Path hfilesDir,
Table table,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
boolean validateHFile,
@@ -622,7 +623,7 @@ implements org.apache.hadoop.util.Tool
-
prepareHFileQueue
-public void prepareHFileQueue(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
+public void prepareHFileQueue(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Table table,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
boolean silence)
@@ -647,7 +648,7 @@ implements org.apache.hadoop.util.Tool
-
loadHFileQueue
-public void loadHFileQueue(Table table,
+public void loadHFileQueue(Table table,
Connection conn,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys)
@@ -675,7 +676,7 @@ implements org.apache.hadoop.util.Tool
-
loadHFileQueue
-public void loadHFileQueue(Table table,
+public void loadHFileQueue(Table table,
Connection conn,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys,
@@ -698,18 +699,19 @@ implements org.apache.hadoop.util.Tool
-
+
-
bulkLoadPhase
-protected void bulkLoadPhase(Table table,
+protected void bulkLoadPhase(Table table,
Connection conn,
ExecutorService pool,
Deque<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> queue,
com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups,
- boolean copyFile)
+ boolean copyFile,
+ Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> item2RegionMap)
throws IOException
This takes the LQI's grouped by likely regions and attempts to bulk load
them. Any failures are re-queued for another pass with the
@@ -726,7 +728,7 @@ implements org.apache.hadoop.util.Tool
-
splitStoreFile
-protected List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> splitStoreFile(org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem item,
+protected List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> splitStoreFile(org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem item,
Table table,
byte[] startKey,
byte[] splitKey)
@@ -743,7 +745,7 @@ implements org.apache.hadoop.util.Tool
-
groupOrSplit
-protected Pair<List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem>,String> groupOrSplit(com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups,
+protected Pair<List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem>,String> groupOrSplit(com.google.common.collect.Multimap<ByteBuffer,org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> regionGroups,
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem item,
Table table,
Pair<byte[][],byte[][]> startEndKeys)
@@ -766,7 +768,7 @@ implements org.apache.hadoop.util.Tool
-
tryAtomicRegionLoad
-protected List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> tryAtomicRegionLoad(Connection conn,
+protected List<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> tryAtomicRegionLoad(Connection conn,
TableName tableName,
byte[] first,
Collection<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem> lqis,
@@ -795,7 +797,7 @@ implements org.apache.hadoop.util.Tool
-
inferBoundaries
-public static byte[][] inferBoundaries(TreeMap<byte[],Integer> bdryMap)
+public static byte[][] inferBoundaries(TreeMap<byte[],Integer> bdryMap)
@@ -804,10 +806,10 @@ implements org.apache.hadoop.util.Tool
-
run
-public List<String> run(String dirPath,
- Map<byte[],List<org.apache.hadoop.fs.Path>> map,
- TableName tableName)
- throws Exception
+public Map<org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> run(String dirPath,
+ Map<byte[],List<org.apache.hadoop.fs.Path>> map,
+ TableName tableName)
+ throws Exception
- Throws:
Exception
@@ -820,7 +822,7 @@ implements org.apache.hadoop.util.Tool
-
run
-public int run(String[] args)
+public int run(String[] args)
throws Exception
- Specified by:
@@ -836,7 +838,7 @@ implements org.apache.hadoop.util.Tool
-
main
-public static void main(String[] args)
+public static void main(String[] args)
throws Exception
- Throws:
@@ -850,7 +852,7 @@ implements org.apache.hadoop.util.Tool
-
setBulkToken
-public void setBulkToken(String stagingDir)
+public void setBulkToken(String stagingDir)
Called from replication sink, where it manages bulkToken(staging directory) by itself. This is
used only when SecureBulkLoadEndpoint is configured in hbase.coprocessor.region.classes
property. This directory is used as a temporary directory where all files are initially