Return-Path: X-Original-To: apmail-hbase-commits-archive@www.apache.org Delivered-To: apmail-hbase-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 91FA518CE3 for ; Wed, 11 Nov 2015 02:24:17 +0000 (UTC) Received: (qmail 37053 invoked by uid 500); 11 Nov 2015 02:24:09 -0000 Delivered-To: apmail-hbase-commits-archive@hbase.apache.org Received: (qmail 36975 invoked by uid 500); 11 Nov 2015 02:24:09 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 36089 invoked by uid 99); 11 Nov 2015 02:24:09 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 11 Nov 2015 02:24:09 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 1EC24E0A36; Wed, 11 Nov 2015 02:24:09 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: misty@apache.org To: commits@hbase.apache.org Date: Wed, 11 Nov 2015 02:24:41 -0000 Message-Id: <09e5aa766f0f4b88833c2dee628feda4@git.apache.org> In-Reply-To: <1ce9f925b68d42dd953d92cb12baabd1@git.apache.org> References: <1ce9f925b68d42dd953d92cb12baabd1@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [35/52] [partial] hbase git commit: Published website at 40c7b09ac86bdbf560321fa71867112aea92a324 http://git-wip-us.apache.org/repos/asf/hbase/blob/1619ae4c/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html ---------------------------------------------------------------------- diff --git a/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html b/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html index 5423853..78dda33 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html +++ b/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html @@ -36,7 +36,7 @@ - + MiniHBaseCluster startMiniCluster(int numMasters, int numSlaves, @@ -1541,7 +1527,7 @@ extends Same as startMiniCluster(int, int), but with custom number of datanodes. - + MiniHBaseCluster startMiniCluster(int numMasters, int numSlaves, @@ -1550,7 +1536,7 @@ extends Class<? extends org.apache.hadoop.hbase.master.HMaster> masterClass, Class<? extends MiniHBaseCluster.MiniHBaseClusterRegionServer> regionserverClass)  - + MiniHBaseCluster startMiniCluster(int numMasters, int numSlaves, @@ -1563,7 +1549,7 @@ extends + MiniHBaseCluster startMiniCluster(int numMasters, int numSlaves, @@ -1571,14 +1557,14 @@ extends Start up a minicluster of hbase, optionally dfs, and zookeeper. - + MiniHBaseCluster startMiniCluster(int numMasters, int numSlaves, String[] dataNodeHosts, boolean create)  - + MiniHBaseCluster startMiniCluster(int numMasters, int numSlaves, @@ -1588,41 +1574,41 @@ extends Start up a minicluster of hbase, optionally dfs, and zookeeper. - + org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers)
Start a minidfscluster.
- + org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers, String[] hosts)
Start a minidfscluster.
- + org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers, String[] racks, String[] hosts)  - + org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(String[] hosts)
Start a minidfscluster.
- + org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSClusterForTestWAL(int namenodePort)  - + MiniHBaseCluster startMiniHBaseCluster(int numMasters, int numSlaves)  - + MiniHBaseCluster startMiniHBaseCluster(int numMasters, int numSlaves, @@ -1632,30 +1618,30 @@ extends Starts up mini hbase cluster. - + org.apache.hadoop.mapred.MiniMRCluster startMiniMapReduceCluster()
Starts a MiniMRCluster with a default number of TaskTracker's.
- + private void startMiniMapReduceCluster(int servers)
Starts a MiniMRCluster.
- + org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster startMiniZKCluster()
Call this if you only want a zk cluster.
- + private org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster startMiniZKCluster(File dir)  - + private org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster startMiniZKCluster(File dir, int zooKeeperServerNum, @@ -1663,34 +1649,34 @@ extends Start a mini ZK cluster. - + org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster startMiniZKCluster(int zooKeeperServerNum, int... clientPortList)
Call this if you only want a zk cluster.
- + org.apache.hadoop.hbase.client.HTable truncateTable(org.apache.hadoop.hbase.TableName tableName)
Truncate a table using the admin command.
- + org.apache.hadoop.hbase.client.HTable truncateTable(org.apache.hadoop.hbase.TableName tableName, boolean preserveRegions)
Truncate a table using the admin command.
- + void verifyNumericRows(org.apache.hadoop.hbase.regionserver.HRegion region, byte[] f, int startRow, int endRow)  - + void verifyNumericRows(org.apache.hadoop.hbase.regionserver.HRegion region, byte[] f, @@ -1698,14 +1684,14 @@ extends + void verifyNumericRows(org.apache.hadoop.hbase.regionserver.Region region, byte[] f, int startRow, int endRow)  - + void verifyNumericRows(org.apache.hadoop.hbase.regionserver.Region region, byte[] f, @@ -1713,7 +1699,7 @@ extends + void verifyNumericRows(org.apache.hadoop.hbase.client.Table table, byte[] f, @@ -1721,7 +1707,7 @@ extends + <E extends Exception
long
waitFor(long timeout, long interval, @@ -1730,7 +1716,7 @@ extends Wrapper method for Waiter.waitFor(Configuration, long, long, boolean, Predicate). - + <E extends Exception
long
waitFor(long timeout, long interval, @@ -1738,87 +1724,87 @@ extends Wrapper method for Waiter.waitFor(Configuration, long, long, Predicate). - + <E extends Exception
long
waitFor(long timeout, Waiter.Predicate<E> predicate) - + static void waitForHostPort(String host, int port)  - + void waitLabelAvailable(long timeoutMillis, String... labels)
Wait until labels is ready in VisibilityLabelsCache.
- + void waitTableAvailable(byte[] table, long timeoutMillis)
Wait until all regions in a table have been assigned
- + void waitTableAvailable(org.apache.hadoop.hbase.TableName table)
Wait until all regions in a table have been assigned.
- + void waitTableAvailable(org.apache.hadoop.hbase.TableName table, long timeoutMillis)  - + void waitTableDisabled(byte[] table)
Waits for a table to be 'disabled'.
- + void waitTableDisabled(byte[] table, long timeoutMillis)
Waits for a table to be 'disabled'.
- + void waitTableDisabled(org.apache.hadoop.hbase.TableName table, long millisTimeout)  - + void waitTableEnabled(byte[] table, long timeoutMillis)
Waits for a table to be 'enabled'.
- + void waitTableEnabled(org.apache.hadoop.hbase.TableName table)
Waits for a table to be 'enabled'.
- + void waitTableEnabled(org.apache.hadoop.hbase.TableName table, long timeoutMillis)  - + void waitUntilAllRegionsAssigned(org.apache.hadoop.hbase.TableName tableName)
Wait until all regions for a table in hbase:meta have a non-empty info:server, up to 60 seconds.
- + void waitUntilAllRegionsAssigned(org.apache.hadoop.hbase.TableName tableName, long timeout) @@ -1826,7 +1812,7 @@ extends + void waitUntilNoRegionsInTransition(long timeout)
Wait until no regions in transition.
@@ -2026,23 +2012,13 @@ private static final Filesystem URI used for map-reduce mini-cluster setup - - - - @@ -2052,7 +2028,7 @@ private static final 
  • BOOLEAN_PARAMETERIZED

    -
    public static final List<Object[]> BOOLEAN_PARAMETERIZED
    +
    public static final List<Object[]> BOOLEAN_PARAMETERIZED
    This is for unit tests parameterized with a two booleans.
  • @@ -2062,7 +2038,7 @@ private static final 
  • MEMSTORETS_TAGS_PARAMETRIZED

    -
    public static final List<Object[]> MEMSTORETS_TAGS_PARAMETRIZED
    +
    public static final List<Object[]> MEMSTORETS_TAGS_PARAMETRIZED
    This is for unit tests parameterized with a single boolean.
  • @@ -2072,7 +2048,7 @@ private static final 
  • COMPRESSION_ALGORITHMS

    -
    public static final org.apache.hadoop.hbase.io.compress.Compression.Algorithm[] COMPRESSION_ALGORITHMS
    +
    public static final org.apache.hadoop.hbase.io.compress.Compression.Algorithm[] COMPRESSION_ALGORITHMS
    Compression algorithms to use in testing
  • @@ -2082,7 +2058,7 @@ private static final 
  • BLOOM_AND_COMPRESSION_COMBINATIONS

    -
    public static final Collection<Object[]> BLOOM_AND_COMPRESSION_COMBINATIONS
    +
    public static final Collection<Object[]> BLOOM_AND_COMPRESSION_COMBINATIONS
  • @@ -2091,7 +2067,7 @@ private static final 
  • fam1

    -
    public static final byte[] fam1
    +
    public static final byte[] fam1
  • @@ -2100,7 +2076,7 @@ private static final 
  • fam2

    -
    public static final byte[] fam2
    +
    public static final byte[] fam2
  • @@ -2109,7 +2085,7 @@ private static final 
  • fam3

    -
    public static final byte[] fam3
    +
    public static final byte[] fam3
  • @@ -2118,7 +2094,7 @@ private static final 
  • COLUMNS

    -
    public static final byte[][] COLUMNS
    +
    public static final byte[][] COLUMNS
  • @@ -2127,7 +2103,7 @@ private static final 
  • MAXVERSIONS

    -
    private static final int MAXVERSIONS
    +
    private static final int MAXVERSIONS
    See Also:
    Constant Field Values
  • @@ -2137,7 +2113,7 @@ private static final 
  • FIRST_CHAR

    -
    public static final char FIRST_CHAR
    +
    public static final char FIRST_CHAR
    See Also:
    Constant Field Values
  • @@ -2147,7 +2123,7 @@ private static final 
  • LAST_CHAR

    -
    public static final char LAST_CHAR
    +
    public static final char LAST_CHAR
    See Also:
    Constant Field Values
  • @@ -2157,7 +2133,7 @@ private static final 
  • START_KEY_BYTES

    -
    public static final byte[] START_KEY_BYTES
    +
    public static final byte[] START_KEY_BYTES
  • @@ -2166,7 +2142,7 @@ private static final 
  • START_KEY

    -
    public static final String START_KEY
    +
    public static final String START_KEY
  • @@ -2175,7 +2151,7 @@ private static final 
  • ROWS

    -
    public static final byte[][] ROWS
    +
    public static final byte[][] ROWS
    All the row values for the data loaded by #loadTable(HTable, byte[])
  • @@ -2185,7 +2161,7 @@ private static final 
  • KEYS

    -
    public static final byte[][] KEYS
    +
    public static final byte[][] KEYS
  • @@ -2194,7 +2170,7 @@ private static final 
  • KEYS_FOR_HBA_CREATE_TABLE

    -
    public static final byte[][] KEYS_FOR_HBA_CREATE_TABLE
    +
    public static final byte[][] KEYS_FOR_HBA_CREATE_TABLE
  • @@ -2203,7 +2179,7 @@ private static final 
  • hbaseAdmin

    -
    private org.apache.hadoop.hbase.client.HBaseAdmin hbaseAdmin
    +
    private org.apache.hadoop.hbase.client.HBaseAdmin hbaseAdmin
  • @@ -2212,36 +2188,25 @@ private static final 
  • zooKeeperWatcher

    -
    private org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher zooKeeperWatcher
    -
  • - - - - - - + - + @@ -2258,7 +2223,7 @@ private static final 
  • HBaseTestingUtility

    -
    public HBaseTestingUtility()
    +
    public HBaseTestingUtility()
  • @@ -2267,7 +2232,7 @@ private static final 
  • HBaseTestingUtility

    -
    public HBaseTestingUtility(org.apache.hadoop.conf.Configuration conf)
    +
    public HBaseTestingUtility(org.apache.hadoop.conf.Configuration conf)
  • @@ -2284,7 +2249,7 @@ private static final 
  • available

    -
    public static boolean available(int port)
    +
    public static boolean available(int port)
    Checks to see if a specific port is available.
    Parameters:
    port - the port number to check for availability
    Returns:
    true if the port is available, or false if not
    @@ -2296,7 +2261,7 @@ private static final 
  • bloomAndCompressionCombinations

    -
    private static List<Object[]> bloomAndCompressionCombinations()
    +
    private static List<Object[]> bloomAndCompressionCombinations()
    Create all combinations of Bloom filters and compression algorithms for testing.
  • @@ -2307,7 +2272,7 @@ private static final 
  • memStoreTSAndTagsCombination

    -
    private static List<Object[]> memStoreTSAndTagsCombination()
    +
    private static List<Object[]> memStoreTSAndTagsCombination()
    Create combination of memstoreTS and tags
  • @@ -2317,7 +2282,7 @@ private static final 
  • memStoreTSTagsAndOffheapCombination

    -
    public static List<Object[]> memStoreTSTagsAndOffheapCombination()
    +
    public static List<Object[]> memStoreTSTagsAndOffheapCombination()
  • @@ -2326,7 +2291,7 @@ private static final 
  • createLocalHTU

    -
    public static HBaseTestingUtility createLocalHTU()
    +
    public static HBaseTestingUtility createLocalHTU()
    Create an HBaseTestingUtility where all tmp files are written to the local test data dir. It is needed to properly base FSUtil.getRootDirs so that they drop temp files in the proper test dir. Use this when you aren't using an Mini HDFS cluster.
    @@ -2339,7 +2304,7 @@ private static final 
  • createLocalHTU

    -
    public static HBaseTestingUtility createLocalHTU(org.apache.hadoop.conf.Configuration c)
    +
    public static HBaseTestingUtility createLocalHTU(org.apache.hadoop.conf.Configuration c)
    Create an HBaseTestingUtility where all tmp files are written to the local test data dir. It is needed to properly base FSUtil.getRootDirs so that they drop temp files in the proper test dir. Use this when you aren't using an Mini HDFS cluster.
    @@ -2353,7 +2318,7 @@ private static final 
  • closeRegionAndWAL

    -
    public static void closeRegionAndWAL(org.apache.hadoop.hbase.regionserver.Region r)
    +
    public static void closeRegionAndWAL(org.apache.hadoop.hbase.regionserver.Region r)
                                   throws IOException
    Close both the region r and it's underlying WAL. For use in tests.
    Throws:
    @@ -2366,7 +2331,7 @@ private static final 
  • closeRegionAndWAL

    -
    public static void closeRegionAndWAL(org.apache.hadoop.hbase.regionserver.HRegion r)
    +
    public static void closeRegionAndWAL(org.apache.hadoop.hbase.regionserver.HRegion r)
                                   throws IOException
    Close both the HRegion r and it's underlying WAL. For use in tests.
    Throws:
    @@ -2379,7 +2344,7 @@ private static final 
  • getConfiguration

    -
    public org.apache.hadoop.conf.Configuration getConfiguration()
    +
    public org.apache.hadoop.conf.Configuration getConfiguration()
    Returns this classes's instance of Configuration. Be careful how you use the returned Configuration since HConnection instances can be shared. The Map of HConnections is keyed by the Configuration. If @@ -2400,7 +2365,7 @@ private static final 
  • setHBaseCluster

    -
    public void setHBaseCluster(HBaseCluster hbaseCluster)
    +
    public void setHBaseCluster(HBaseCluster hbaseCluster)
  • @@ -2409,7 +2374,7 @@ private static final 
  • setupDataTestDir

    -
    protected org.apache.hadoop.fs.Path setupDataTestDir()
    +
    protected org.apache.hadoop.fs.Path setupDataTestDir()
    Home our data in a dir under HBaseCommonTestingUtility.DEFAULT_BASE_TEST_DIRECTORY. Give it a random name so can have many concurrent tests running if we need to. It needs to amend the TEST_DIRECTORY_KEY @@ -2434,7 +2399,7 @@ private static final 
  • createSubDirAndSystemProperty

    -
    private void createSubDirAndSystemProperty(String propertyName,
    +
    private void createSubDirAndSystemProperty(String propertyName,
                                      org.apache.hadoop.fs.Path parent,
                                      String subDirName)
  • @@ -2445,7 +2410,7 @@ private static final 
  • getBaseTestDirOnTestFS

    -
    private org.apache.hadoop.fs.Path getBaseTestDirOnTestFS()
    +
    private org.apache.hadoop.fs.Path getBaseTestDirOnTestFS()
                                                       throws IOException
    Returns:
    Where to write test data on the test filesystem; Returns working directory for the test filesystem by default
    @@ -2460,7 +2425,7 @@ private static final 
  • getMetaTableDescriptor

    -
    public org.apache.hadoop.hbase.HTableDescriptor getMetaTableDescriptor()
    +
    public org.apache.hadoop.hbase.HTableDescriptor getMetaTableDescriptor()
    Returns:
    META table descriptor
  • @@ -2470,7 +2435,7 @@ private static final 
  • getClusterTestDir

    -
    org.apache.hadoop.fs.Path getClusterTestDir()
    +
    org.apache.hadoop.fs.Path getClusterTestDir()
    Returns:
    Where the DFS cluster will write data on the local subsystem. Creates it if it does not exist already. A subdir of HBaseCommonTestingUtility.getBaseTestDir()
    See Also:
    getTestFileSystem()
  • @@ -2481,7 +2446,7 @@ private static final 
  • setupClusterTestDir

    -
    private void setupClusterTestDir()
    +
    private void setupClusterTestDir()
    Creates a directory for the DFS cluster, under the test data
  • @@ -2491,7 +2456,7 @@ private static final 
  • getDataTestDirOnTestFS

    -
    public org.apache.hadoop.fs.Path getDataTestDirOnTestFS()
    +
    public org.apache.hadoop.fs.Path getDataTestDirOnTestFS()
                                                      throws IOException
    Returns a Path in the test filesystem, obtained from getTestFileSystem() to write temporary test data. Call this method after setting up the mini dfs cluster @@ -2507,7 +2472,7 @@ private static final 
  • getDataTestDirOnTestFS

    -
    public org.apache.hadoop.fs.Path getDataTestDirOnTestFS(String subdirName)
    +
    public org.apache.hadoop.fs.Path getDataTestDirOnTestFS(String subdirName)
                                                      throws IOException
    Returns a Path in the test filesystem, obtained from getTestFileSystem() to write temporary test data. Call this method after setting up the mini dfs cluster @@ -2524,7 +2489,7 @@ private static final 
  • setupDataTestDirOnTestFS

    -
    private void setupDataTestDirOnTestFS()
    +
    private void setupDataTestDirOnTestFS()
                                    throws IOException
    Sets up a path in test filesystem to be used by tests. Creates a new directory if not already setup.
    @@ -2538,7 +2503,7 @@ private static final 
  • getNewDataTestDirOnTestFS

    -
    private org.apache.hadoop.fs.Path getNewDataTestDirOnTestFS()
    +
    private org.apache.hadoop.fs.Path getNewDataTestDirOnTestFS()
                                                          throws IOException
    Sets up a new path in test filesystem to be used by tests.
    Throws:
    @@ -2551,7 +2516,7 @@ private static final 
  • cleanupDataTestDirOnTestFS

    -
    public boolean cleanupDataTestDirOnTestFS()
    +
    public boolean cleanupDataTestDirOnTestFS()
                                        throws IOException
    Cleans the test data directory on the test filesystem.
    Returns:
    True if we removed the test dirs
    @@ -2565,7 +2530,7 @@ private static final 
  • cleanupDataTestDirOnTestFS

    -
    public boolean cleanupDataTestDirOnTestFS(String subdirName)
    +
    public boolean cleanupDataTestDirOnTestFS(String subdirName)
                                        throws IOException
    Cleans a subdirectory under the test data directory on the test filesystem.
    Returns:
    True if we removed child
    @@ -2579,7 +2544,7 @@ private static final 
  • startMiniDFSCluster

    -
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers)
    +
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers)
                                                               throws Exception
    Start a minidfscluster.
    Parameters:
    servers - How many DNs to start.
    @@ -2594,7 +2559,7 @@ private static final 
  • startMiniDFSCluster

    -
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(String[] hosts)
    +
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(String[] hosts)
                                                               throws Exception
    Start a minidfscluster. This is useful if you want to run datanode on distinct hosts for things @@ -2613,7 +2578,7 @@ private static final 
  • startMiniDFSCluster

    -
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers,
    +
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers,
                                                             String[] hosts)
                                                               throws Exception
    Start a minidfscluster. @@ -2630,7 +2595,7 @@ private static final 
  • setFs

    -
    private void setFs()
    +
    private void setFs()
                 throws IOException
    Throws:
    IOException
    @@ -2642,7 +2607,7 @@ private static final 
  • startMiniDFSCluster

    -
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers,
    +
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSCluster(int servers,
                                                             String[] racks,
                                                             String[] hosts)
                                                               throws Exception
    @@ -2656,7 +2621,7 @@ private static final 
  • startMiniDFSClusterForTestWAL

    -
    public org.apache.hadoop.hdfs.MiniDFSCluster startMiniDFSClusterForTestWAL(int namenodePort)
    +
    public org.apache.hadoop.hdfs.MiniDFSCluster