hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From syuanji...@apache.org
Subject [26/50] [abbrv] hbase git commit: HBASE-17532 Replaced explicit type with diamond operator
Date Fri, 10 Mar 2017 22:10:06 GMT
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java
index 1dab633..01e9ef3 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/mapreduce/IndexBuilder.java
@@ -103,7 +103,7 @@ public class IndexBuilder extends Configured implements Tool {
       String[] fields = configuration.getStrings("index.fields");
       String familyName = configuration.get("index.familyname");
       family = Bytes.toBytes(familyName);
-      indexes = new TreeMap<byte[], ImmutableBytesWritable>(Bytes.BYTES_COMPARATOR);
+      indexes = new TreeMap<>(Bytes.BYTES_COMPARATOR);
       for(String field : fields) {
         // if the table is "people" and the field to index is "email", then the
         // index table will be called "people-email"

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
index b16ef7b..cb0cfbb 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/DemoClient.java
@@ -114,7 +114,7 @@ public class DemoClient {
     private void run() throws Exception {
         TTransport transport = new TSocket(host, port);
         if (secure) {
-          Map<String, String> saslProperties = new HashMap<String, String>();
+          Map<String, String> saslProperties = new HashMap<>();
           saslProperties.put(Sasl.QOP, "auth-conf,auth-int,auth");
           /**
            * The Thrift server the DemoClient is trying to connect to
@@ -154,7 +154,7 @@ public class DemoClient {
         //
         // Create the demo table with two column families, entry: and unused:
         //
-        ArrayList<ColumnDescriptor> columns = new ArrayList<ColumnDescriptor>(2);
+        ArrayList<ColumnDescriptor> columns = new ArrayList<>(2);
         ColumnDescriptor col;
         col = new ColumnDescriptor();
         col.name = ByteBuffer.wrap(bytes("entry:"));
@@ -194,7 +194,7 @@ public class DemoClient {
 
         ArrayList<Mutation> mutations;
         // non-utf8 is fine for data
-        mutations = new ArrayList<Mutation>(1);
+        mutations = new ArrayList<>(1);
         mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("entry:foo")),
             ByteBuffer.wrap(invalid), writeToWal));
         client.mutateRow(ByteBuffer.wrap(t), ByteBuffer.wrap(bytes("foo")),
@@ -202,19 +202,19 @@ public class DemoClient {
 
 
         // this row name is valid utf8
-        mutations = new ArrayList<Mutation>(1);
+        mutations = new ArrayList<>(1);
         mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("entry:foo")), ByteBuffer.wrap(valid), writeToWal));
         client.mutateRow(ByteBuffer.wrap(t), ByteBuffer.wrap(valid), mutations, dummyAttributes);
 
         // non-utf8 is now allowed in row names because HBase stores values as binary
 
-        mutations = new ArrayList<Mutation>(1);
+        mutations = new ArrayList<>(1);
         mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("entry:foo")), ByteBuffer.wrap(invalid), writeToWal));
         client.mutateRow(ByteBuffer.wrap(t), ByteBuffer.wrap(invalid), mutations, dummyAttributes);
 
 
         // Run a scanner on the rows we just created
-        ArrayList<ByteBuffer> columnNames = new ArrayList<ByteBuffer>();
+        ArrayList<ByteBuffer> columnNames = new ArrayList<>();
         columnNames.add(ByteBuffer.wrap(bytes("entry:")));
 
         System.out.println("Starting scanner...");
@@ -238,7 +238,7 @@ public class DemoClient {
             nf.setGroupingUsed(false);
             byte[] row = bytes(nf.format(i));
 
-            mutations = new ArrayList<Mutation>(1);
+            mutations = new ArrayList<>(1);
             mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("unused:")), ByteBuffer.wrap(bytes("DELETE_ME")), writeToWal));
             client.mutateRow(ByteBuffer.wrap(t), ByteBuffer.wrap(row), mutations, dummyAttributes);
             printRow(client.getRow(ByteBuffer.wrap(t), ByteBuffer.wrap(row), dummyAttributes));
@@ -251,14 +251,14 @@ public class DemoClient {
                 // no-op
             }
 
-            mutations = new ArrayList<Mutation>(2);
+            mutations = new ArrayList<>(2);
             mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("entry:num")), ByteBuffer.wrap(bytes("0")), writeToWal));
             mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("entry:foo")), ByteBuffer.wrap(bytes("FOO")), writeToWal));
             client.mutateRow(ByteBuffer.wrap(t), ByteBuffer.wrap(row), mutations, dummyAttributes);
             printRow(client.getRow(ByteBuffer.wrap(t), ByteBuffer.wrap(row), dummyAttributes));
 
             Mutation m;
-            mutations = new ArrayList<Mutation>(2);
+            mutations = new ArrayList<>(2);
             m = new Mutation();
             m.column = ByteBuffer.wrap(bytes("entry:foo"));
             m.isDelete = true;
@@ -270,7 +270,7 @@ public class DemoClient {
             client.mutateRow(ByteBuffer.wrap(t), ByteBuffer.wrap(row), mutations, dummyAttributes);
             printRow(client.getRow(ByteBuffer.wrap(t), ByteBuffer.wrap(row), dummyAttributes));
 
-            mutations = new ArrayList<Mutation>();
+            mutations = new ArrayList<>();
             mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("entry:num")), ByteBuffer.wrap(bytes(Integer.toString(i))), writeToWal));
             mutations.add(new Mutation(false, ByteBuffer.wrap(bytes("entry:sqr")), ByteBuffer.wrap(bytes(Integer.toString(i * i))), writeToWal));
             client.mutateRow(ByteBuffer.wrap(t), ByteBuffer.wrap(row), mutations, dummyAttributes);
@@ -347,7 +347,7 @@ public class DemoClient {
     private void printRow(TRowResult rowResult) {
         // copy values into a TreeMap to get them in sorted order
 
-        TreeMap<String, TCell> sorted = new TreeMap<String, TCell>();
+        TreeMap<String, TCell> sorted = new TreeMap<>();
         for (Map.Entry<ByteBuffer, TCell> column : rowResult.columns.entrySet()) {
             sorted.put(utf8(column.getKey().array()), column.getValue());
         }
@@ -379,7 +379,7 @@ public class DemoClient {
         new Configuration() {
           @Override
           public AppConfigurationEntry[] getAppConfigurationEntry(String name) {
-            Map<String, String> options = new HashMap<String, String>();
+            Map<String, String> options = new HashMap<>();
             options.put("useKeyTab", "false");
             options.put("storeKey", "false");
             options.put("doNotPrompt", "true");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
index 666891c..25fdc4a 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift/HttpDoAsClient.java
@@ -151,7 +151,7 @@ public class HttpDoAsClient {
     //
     // Create the demo table with two column families, entry: and unused:
     //
-    ArrayList<ColumnDescriptor> columns = new ArrayList<ColumnDescriptor>(2);
+    ArrayList<ColumnDescriptor> columns = new ArrayList<>(2);
     ColumnDescriptor col;
     col = new ColumnDescriptor();
     col.name = ByteBuffer.wrap(bytes("entry:"));
@@ -236,7 +236,7 @@ public class HttpDoAsClient {
   private void printRow(TRowResult rowResult) {
     // copy values into a TreeMap to get them in sorted order
 
-    TreeMap<String, TCell> sorted = new TreeMap<String, TCell>();
+    TreeMap<String, TCell> sorted = new TreeMap<>();
     for (Map.Entry<ByteBuffer, TCell> column : rowResult.columns.entrySet()) {
       sorted.put(utf8(column.getKey().array()), column.getValue());
     }
@@ -261,7 +261,7 @@ public class HttpDoAsClient {
         new Configuration() {
           @Override
           public AppConfigurationEntry[] getAppConfigurationEntry(String name) {
-            Map<String, String> options = new HashMap<String, String>();
+            Map<String, String> options = new HashMap<>();
             options.put("useKeyTab", "false");
             options.put("storeKey", "false");
             options.put("doNotPrompt", "true");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java
index 4083792..666997e 100644
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java
+++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java
@@ -102,7 +102,7 @@ public class DemoClient {
        *
        * The HBase cluster must be secure, allow proxy user.
        */
-      Map<String, String> saslProperties = new HashMap<String, String>();
+      Map<String, String> saslProperties = new HashMap<>();
       saslProperties.put(Sasl.QOP, "auth-conf,auth-int,auth");
       transport = new TSaslClientTransport("GSSAPI", null,
         user != null ? user : "hbase",// Thrift server user name, should be an authorized proxy user
@@ -126,7 +126,7 @@ public class DemoClient {
     columnValue.setFamily("family1".getBytes());
     columnValue.setQualifier("qualifier1".getBytes());
     columnValue.setValue("value1".getBytes());
-    List<TColumnValue> columnValues = new ArrayList<TColumnValue>(1);
+    List<TColumnValue> columnValues = new ArrayList<>(1);
     columnValues.add(columnValue);
     put.setColumnValues(columnValues);
 
@@ -159,7 +159,7 @@ public class DemoClient {
       new Configuration() {
         @Override
         public AppConfigurationEntry[] getAppConfigurationEntry(String name) {
-          Map<String, String> options = new HashMap<String, String>();
+          Map<String, String> options = new HashMap<>();
           options.put("useKeyTab", "false");
           options.put("storeKey", "false");
           options.put("doNotPrompt", "true");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-external-blockcache/src/main/java/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.java
----------------------------------------------------------------------
diff --git a/hbase-external-blockcache/src/main/java/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.java b/hbase-external-blockcache/src/main/java/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.java
index 69d8521..e741760 100644
--- a/hbase-external-blockcache/src/main/java/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.java
+++ b/hbase-external-blockcache/src/main/java/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.java
@@ -100,7 +100,7 @@ public class MemcachedBlockCache implements BlockCache {
     // case.
     String serverListString = c.get(MEMCACHED_CONFIG_KEY,"localhost:11211");
     String[] servers = serverListString.split(",");
-    List<InetSocketAddress> serverAddresses = new ArrayList<InetSocketAddress>(servers.length);
+    List<InetSocketAddress> serverAddresses = new ArrayList<>(servers.length);
     for (String s:servers) {
       serverAddresses.add(Addressing.createInetSocketAddressFromHostAndPortStr(s));
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/CompatibilitySingletonFactory.java
----------------------------------------------------------------------
diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/CompatibilitySingletonFactory.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/CompatibilitySingletonFactory.java
index 78442ba..be6d6d1 100644
--- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/CompatibilitySingletonFactory.java
+++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/CompatibilitySingletonFactory.java
@@ -34,7 +34,7 @@ public class CompatibilitySingletonFactory extends CompatibilityFactory {
   public static enum SingletonStorage {
     INSTANCE;
     private final Object lock = new Object();
-    private final Map<Class, Object> instances = new HashMap<Class, Object>();
+    private final Map<Class, Object> instances = new HashMap<>();
   }
   private static final Log LOG = LogFactory.getLog(CompatibilitySingletonFactory.class);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/TestCompatibilitySingletonFactory.java
----------------------------------------------------------------------
diff --git a/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/TestCompatibilitySingletonFactory.java b/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/TestCompatibilitySingletonFactory.java
index f942059..168f6c7 100644
--- a/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/TestCompatibilitySingletonFactory.java
+++ b/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/TestCompatibilitySingletonFactory.java
@@ -55,9 +55,8 @@ public class TestCompatibilitySingletonFactory {
 
   @Test
   public void testGetInstance() throws Exception {
-    List<TestCompatibilitySingletonFactoryCallable> callables =
-        new ArrayList<TestCompatibilitySingletonFactoryCallable>(ITERATIONS);
-    List<String> resultStrings = new ArrayList<String>(ITERATIONS);
+    List<TestCompatibilitySingletonFactoryCallable> callables = new ArrayList<>(ITERATIONS);
+    List<String> resultStrings = new ArrayList<>(ITERATIONS);
 
 
     // Create the callables.

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java
----------------------------------------------------------------------
diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java
index 76bbb09..78893ab 100644
--- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java
+++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java
@@ -28,10 +28,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 public class MetricsHBaseServerSourceFactoryImpl extends MetricsHBaseServerSourceFactory {
   private enum SourceStorage {
     INSTANCE;
-    HashMap<String, MetricsHBaseServerSource>
-        sources =
-        new HashMap<String, MetricsHBaseServerSource>();
-
+    HashMap<String, MetricsHBaseServerSource> sources = new HashMap<>();
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/master/balancer/MetricsStochasticBalancerSourceImpl.java
----------------------------------------------------------------------
diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/master/balancer/MetricsStochasticBalancerSourceImpl.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/master/balancer/MetricsStochasticBalancerSourceImpl.java
index f658a27..c304fb9 100644
--- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/master/balancer/MetricsStochasticBalancerSourceImpl.java
+++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/master/balancer/MetricsStochasticBalancerSourceImpl.java
@@ -46,7 +46,7 @@ public class MetricsStochasticBalancerSourceImpl extends MetricsBalancerSourceIm
           return size() > mruCap;
         }
       };
-  private Map<String, String> costFunctionDescs = new ConcurrentHashMap<String, String>();
+  private Map<String, String> costFunctionDescs = new ConcurrentHashMap<>();
 
   /**
    * Calculates the mru cache capacity from the metrics size
@@ -79,7 +79,7 @@ public class MetricsStochasticBalancerSourceImpl extends MetricsBalancerSourceIm
     synchronized (stochasticCosts) {
       Map<String, Double> costs = stochasticCosts.get(tableName);
       if (costs == null) {
-        costs = new ConcurrentHashMap<String, Double>();
+        costs = new ConcurrentHashMap<>();
       }
 
       costs.put(costFunctionName, cost);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/metrics/Interns.java
----------------------------------------------------------------------
diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/metrics/Interns.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/metrics/Interns.java
index 7905561..565b853 100644
--- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/metrics/Interns.java
+++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/metrics/Interns.java
@@ -45,14 +45,14 @@ public final class Interns {
       CacheBuilder.newBuilder().expireAfterAccess(1, TimeUnit.DAYS)
           .build(new CacheLoader<String, ConcurrentHashMap<String, MetricsInfo>>() {
             public ConcurrentHashMap<String, MetricsInfo> load(String key) {
-              return new ConcurrentHashMap<String, MetricsInfo>();
+              return new ConcurrentHashMap<>();
             }
           });
   private static LoadingCache<MetricsInfo, ConcurrentHashMap<String, MetricsTag>> tagCache =
       CacheBuilder.newBuilder().expireAfterAccess(1, TimeUnit.DAYS)
           .build(new CacheLoader<MetricsInfo, ConcurrentHashMap<String, MetricsTag>>() {
             public ConcurrentHashMap<String, MetricsTag> load(MetricsInfo key) {
-              return new ConcurrentHashMap<String, MetricsTag>();
+              return new ConcurrentHashMap<>();
             }
           });
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/util/MetricSampleQuantiles.java
----------------------------------------------------------------------
diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/util/MetricSampleQuantiles.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/util/MetricSampleQuantiles.java
index a968aca..3e4016d 100644
--- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/util/MetricSampleQuantiles.java
+++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/metrics2/util/MetricSampleQuantiles.java
@@ -73,7 +73,7 @@ public class MetricSampleQuantiles {
 
   public MetricSampleQuantiles(MetricQuantile[] quantiles) {
     this.quantiles = Arrays.copyOf(quantiles, quantiles.length);
-    this.samples = new LinkedList<SampleItem>();
+    this.samples = new LinkedList<>();
   }
 
   /**
@@ -235,7 +235,7 @@ public class MetricSampleQuantiles {
   synchronized public Map<MetricQuantile, Long> snapshot() throws IOException {
     // flush the buffer first for best results
     insertBatch();
-    Map<MetricQuantile, Long> values = new HashMap<MetricQuantile, Long>(quantiles.length);
+    Map<MetricQuantile, Long> values = new HashMap<>(quantiles.length);
     for (int i = 0; i < quantiles.length; i++) {
       values.put(quantiles[i], query(quantiles[i].quantile));
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java
----------------------------------------------------------------------
diff --git a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java
index 4291eb7..19a8ad2 100644
--- a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java
+++ b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java
@@ -37,9 +37,9 @@ import static org.junit.Assert.*;
  *  A helper class that will allow tests to get into hadoop2's metrics2 values.
  */
 public class MetricsAssertHelperImpl implements MetricsAssertHelper {
-  private Map<String, String> tags = new HashMap<String, String>();
-  private Map<String, Number> gauges = new HashMap<String, Number>();
-  private Map<String, Long> counters = new HashMap<String, Long>();
+  private Map<String, String> tags = new HashMap<>();
+  private Map<String, Number> gauges = new HashMap<>();
+  private Map<String, Long> counters = new HashMap<>();
 
   public class MockMetricsBuilder implements MetricsCollector {
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
index d35ef84..431ba42 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
@@ -317,7 +317,7 @@ public class DistributedHBaseCluster extends HBaseCluster {
   }
 
   protected boolean restoreMasters(ClusterStatus initial, ClusterStatus current) {
-    List<IOException> deferred = new ArrayList<IOException>();
+    List<IOException> deferred = new ArrayList<>();
     //check whether current master has changed
     final ServerName initMaster = initial.getMaster();
     if (!ServerName.isSameHostnameAndPort(initMaster, current.getMaster())) {
@@ -371,8 +371,8 @@ public class DistributedHBaseCluster extends HBaseCluster {
       }
     } else {
       //current master has not changed, match up backup masters
-      Set<ServerName> toStart = new TreeSet<ServerName>(new ServerNameIgnoreStartCodeComparator());
-      Set<ServerName> toKill = new TreeSet<ServerName>(new ServerNameIgnoreStartCodeComparator());
+      Set<ServerName> toStart = new TreeSet<>(new ServerNameIgnoreStartCodeComparator());
+      Set<ServerName> toKill = new TreeSet<>(new ServerNameIgnoreStartCodeComparator());
       toStart.addAll(initial.getBackupMasters());
       toKill.addAll(current.getBackupMasters());
 
@@ -429,8 +429,8 @@ public class DistributedHBaseCluster extends HBaseCluster {
   }
 
   protected boolean restoreRegionServers(ClusterStatus initial, ClusterStatus current) {
-    Set<ServerName> toStart = new TreeSet<ServerName>(new ServerNameIgnoreStartCodeComparator());
-    Set<ServerName> toKill = new TreeSet<ServerName>(new ServerNameIgnoreStartCodeComparator());
+    Set<ServerName> toStart = new TreeSet<>(new ServerNameIgnoreStartCodeComparator());
+    Set<ServerName> toKill = new TreeSet<>(new ServerNameIgnoreStartCodeComparator());
     toStart.addAll(initial.getServers());
     toKill.addAll(current.getServers());
 
@@ -443,7 +443,7 @@ public class DistributedHBaseCluster extends HBaseCluster {
       toKill.remove(server);
     }
 
-    List<IOException> deferred = new ArrayList<IOException>();
+    List<IOException> deferred = new ArrayList<>();
 
     for(ServerName sn:toStart) {
       try {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java
index b6f1aeb..07014e5 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java
@@ -312,7 +312,7 @@ public class HBaseClusterManager extends Configured implements ClusterManager {
     LOG.info("Executed remote command, exit code:" + shell.getExitCode()
         + " , output:" + shell.getOutput());
 
-    return new Pair<Integer, String>(shell.getExitCode(), shell.getOutput());
+    return new Pair<>(shell.getExitCode(), shell.getOutput());
   }
 
   private Pair<Integer, String> execWithRetries(String hostname, ServiceType service, String... cmd)

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestDDLMasterFailover.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestDDLMasterFailover.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestDDLMasterFailover.java
index 5d79722..2d3693a 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestDDLMasterFailover.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestDDLMasterFailover.java
@@ -125,17 +125,13 @@ public class IntegrationTestDDLMasterFailover extends IntegrationTestBase {
 
   protected int numThreads, numRegions;
 
-  ConcurrentHashMap<String, NamespaceDescriptor> namespaceMap =
-      new ConcurrentHashMap<String, NamespaceDescriptor>();
+  ConcurrentHashMap<String, NamespaceDescriptor> namespaceMap = new ConcurrentHashMap<>();
 
-  ConcurrentHashMap<TableName, HTableDescriptor> enabledTables =
-      new ConcurrentHashMap<TableName, HTableDescriptor>();
+  ConcurrentHashMap<TableName, HTableDescriptor> enabledTables = new ConcurrentHashMap<>();
 
-  ConcurrentHashMap<TableName, HTableDescriptor> disabledTables =
-      new ConcurrentHashMap<TableName, HTableDescriptor>();
+  ConcurrentHashMap<TableName, HTableDescriptor> disabledTables = new ConcurrentHashMap<>();
 
-  ConcurrentHashMap<TableName, HTableDescriptor> deletedTables =
-      new ConcurrentHashMap<TableName, HTableDescriptor>();
+  ConcurrentHashMap<TableName, HTableDescriptor> deletedTables = new ConcurrentHashMap<>();
 
   @Override
   public void setUpCluster() throws Exception {
@@ -256,7 +252,7 @@ public class IntegrationTestDDLMasterFailover extends IntegrationTestBase {
         if (namespaceMap.isEmpty()) {
           return null;
         }
-        ArrayList<String> namespaceList = new ArrayList<String>(namespaceMap.keySet());
+        ArrayList<String> namespaceList = new ArrayList<>(namespaceMap.keySet());
         String randomKey = namespaceList.get(RandomUtils.nextInt(namespaceList.size()));
         NamespaceDescriptor randomNsd = namespaceMap.get(randomKey);
         // remove from namespaceMap
@@ -396,7 +392,7 @@ public class IntegrationTestDDLMasterFailover extends IntegrationTestBase {
         if (tableMap.isEmpty()) {
           return null;
         }
-        ArrayList<TableName> tableList = new ArrayList<TableName>(tableMap.keySet());
+        ArrayList<TableName> tableList = new ArrayList<>(tableMap.keySet());
         TableName randomKey = tableList.get(RandomUtils.nextInt(tableList.size()));
         HTableDescriptor randomHtd = tableMap.get(randomKey);
         // remove from tableMap
@@ -770,7 +766,7 @@ public class IntegrationTestDDLMasterFailover extends IntegrationTestBase {
       Admin admin = connection.getAdmin();
       TableName tableName = selected.getTableName();
       try (Table table = connection.getTable(tableName)){
-        ArrayList<HRegionInfo> regionInfos = new ArrayList<HRegionInfo>(admin.getTableRegions(
+        ArrayList<HRegionInfo> regionInfos = new ArrayList<>(admin.getTableRegions(
             selected.getTableName()));
         int numRegions = regionInfos.size();
         // average number of rows to be added per action to each region

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngest.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngest.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngest.java
index d4bcacd..7b6635e 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngest.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngest.java
@@ -207,7 +207,7 @@ public class IntegrationTestIngest extends IntegrationTestBase {
   }
 
   protected String[] getArgsForLoadTestToolInitTable() {
-    List<String> args = new ArrayList<String>();
+    List<String> args = new ArrayList<>();
     args.add("-tn");
     args.add(getTablename().getNameAsString());
     // pass all remaining args from conf with keys <test class name>.<load test tool arg>
@@ -225,7 +225,7 @@ public class IntegrationTestIngest extends IntegrationTestBase {
 
   protected String[] getArgsForLoadTestTool(String mode, String modeSpecificArg, long startKey,
       long numKeys) {
-    List<String> args = new ArrayList<String>(11);
+    List<String> args = new ArrayList<>(11);
     args.add("-tn");
     args.add(getTablename().getNameAsString());
     args.add("-families");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java
index 82eef1a..d129279 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithACL.java
@@ -68,7 +68,7 @@ public class IntegrationTestIngestWithACL extends IntegrationTestIngest {
   protected String[] getArgsForLoadTestTool(String mode, String modeSpecificArg, long startKey,
       long numKeys) {
     String[] args = super.getArgsForLoadTestTool(mode, modeSpecificArg, startKey, numKeys);
-    List<String> tmp = new ArrayList<String>(Arrays.asList(args));
+    List<String> tmp = new ArrayList<>(Arrays.asList(args));
     tmp.add(HYPHEN + LoadTestTool.OPT_GENERATOR);
     StringBuilder sb = new StringBuilder(LoadTestDataGeneratorWithACL.class.getName());
     sb.append(COLON);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java
index 13a5936..cd9e355 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithMOB.java
@@ -61,7 +61,7 @@ public class IntegrationTestIngestWithMOB extends IntegrationTestIngest {
 
   @Override
   protected String[] getArgsForLoadTestToolInitTable() {
-    List<String> args = new ArrayList<String>();
+    List<String> args = new ArrayList<>();
     args.add("-tn");
     args.add(getTablename().getNameAsString());
     // pass all remaining args from conf with keys <test class name>.<load test tool arg>
@@ -133,7 +133,7 @@ public class IntegrationTestIngestWithMOB extends IntegrationTestIngest {
   protected String[] getArgsForLoadTestTool(String mode, String modeSpecificArg, long startKey,
       long numKeys) {
     String[] args = super.getArgsForLoadTestTool(mode, modeSpecificArg, startKey, numKeys);
-    List<String> tmp = new ArrayList<String>(Arrays.asList(args));
+    List<String> tmp = new ArrayList<>(Arrays.asList(args));
     // LoadTestDataGeneratorMOB:mobColumnFamily:minMobDataSize:maxMobDataSize
     tmp.add(HIPHEN + LoadTestTool.OPT_GENERATOR);
     StringBuilder sb = new StringBuilder(LoadTestDataGeneratorWithMOB.class.getName());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithTags.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithTags.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithTags.java
index f1b2c68..08bd4e5 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithTags.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithTags.java
@@ -45,7 +45,7 @@ public class IntegrationTestIngestWithTags extends IntegrationTestIngest {
   protected String[] getArgsForLoadTestTool(String mode, String modeSpecificArg, long startKey,
       long numKeys) {
     String[] args = super.getArgsForLoadTestTool(mode, modeSpecificArg, startKey, numKeys);
-    List<String> tmp = new ArrayList<String>(Arrays.asList(args));
+    List<String> tmp = new ArrayList<>(Arrays.asList(args));
     // LoadTestDataGeneratorWithTags:minNumTags:maxNumTags:minTagLength:maxTagLength
     tmp.add(HIPHEN + LoadTestTool.OPT_GENERATOR);
     StringBuilder sb = new StringBuilder(LoadTestDataGeneratorWithTags.class.getName());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithVisibilityLabels.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithVisibilityLabels.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithVisibilityLabels.java
index 133be1a..b7d8dad 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithVisibilityLabels.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestIngestWithVisibilityLabels.java
@@ -42,31 +42,31 @@ public class IntegrationTestIngestWithVisibilityLabels extends IntegrationTestIn
   private static final String[] VISIBILITY_EXPS = { "secret & confidential & !private",
       "topsecret | confidential", "confidential & private", "public", "topsecret & private",
       "!public | private", "(secret | topsecret) & private" };
-  private static final List<List<String>> AUTHS = new ArrayList<List<String>>();
+  private static final List<List<String>> AUTHS = new ArrayList<>();
 
   static {
-    ArrayList<String> tmp = new ArrayList<String>(2);
+    ArrayList<String> tmp = new ArrayList<>(2);
     tmp.add("secret");
     tmp.add("confidential");
     AUTHS.add(tmp);
-    tmp = new ArrayList<String>(1);
+    tmp = new ArrayList<>(1);
     tmp.add("topsecret");
     AUTHS.add(tmp);
-    tmp = new ArrayList<String>(2);
+    tmp = new ArrayList<>(2);
     tmp.add("confidential");
     tmp.add("private");
     AUTHS.add(tmp);
-    tmp = new ArrayList<String>(1);
+    tmp = new ArrayList<>(1);
     tmp.add("public");
     AUTHS.add(tmp);
-    tmp = new ArrayList<String>(2);
+    tmp = new ArrayList<>(2);
     tmp.add("topsecret");
     tmp.add("private");
     AUTHS.add(tmp);
-    tmp = new ArrayList<String>(1);
+    tmp = new ArrayList<>(1);
     tmp.add("confidential");
     AUTHS.add(tmp);
-    tmp = new ArrayList<String>(2);
+    tmp = new ArrayList<>(2);
     tmp.add("topsecret");
     tmp.add("private");
     AUTHS.add(tmp);
@@ -88,7 +88,7 @@ public class IntegrationTestIngestWithVisibilityLabels extends IntegrationTestIn
   protected String[] getArgsForLoadTestTool(String mode, String modeSpecificArg, long startKey,
       long numKeys) {
     String[] args = super.getArgsForLoadTestTool(mode, modeSpecificArg, startKey, numKeys);
-    List<String> tmp = new ArrayList<String>(Arrays.asList(args));
+    List<String> tmp = new ArrayList<>(Arrays.asList(args));
     tmp.add(HIPHEN + LoadTestTool.OPT_GENERATOR);
     StringBuilder sb = new StringBuilder(LoadTestDataGeneratorWithVisibilityLabels.class.getName());
     sb.append(COLON);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
index 548ff53..6efe9d8 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
@@ -92,8 +92,7 @@ public class IntegrationTestLazyCfLoading {
     public static final byte[] VALUE_COLUMN = Bytes.toBytes("val");
     public static final long ACCEPTED_VALUE = 1L;
 
-    private static final Map<byte[], byte[][]> columnMap = new TreeMap<byte[], byte[][]>(
-        Bytes.BYTES_COMPARATOR);
+    private static final Map<byte[], byte[][]> columnMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
 
     private final AtomicLong expectedNumberOfKeys = new AtomicLong(0);
     private final AtomicLong totalNumberOfKeys = new AtomicLong(0);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaPerf.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaPerf.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaPerf.java
index e609f0b..c3c5df3 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaPerf.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaPerf.java
@@ -266,8 +266,8 @@ public class IntegrationTestRegionReplicaPerf extends IntegrationTestBase {
       format("--nomapred --table=%s --latency --sampleRate=0.1 randomRead 4", tableName);
     String replicaReadOpts = format("%s %s", replicas, readOpts);
 
-    ArrayList<TimingResult> resultsWithoutReplicas = new ArrayList<TimingResult>(maxIters);
-    ArrayList<TimingResult> resultsWithReplicas = new ArrayList<TimingResult>(maxIters);
+    ArrayList<TimingResult> resultsWithoutReplicas = new ArrayList<>(maxIters);
+    ArrayList<TimingResult> resultsWithReplicas = new ArrayList<>(maxIters);
 
     // create/populate the table, replicas disabled
     LOG.debug("Populating table.");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaReplication.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaReplication.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaReplication.java
index 98d53e9..b6cfdcd 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaReplication.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRegionReplicaReplication.java
@@ -126,7 +126,7 @@ public class IntegrationTestRegionReplicaReplication extends IntegrationTestInge
     protected BlockingQueue<Long> createWriteKeysQueue(Configuration conf) {
       this.delayMs = conf.getLong(String.format("%s.%s",
         IntegrationTestRegionReplicaReplication.class.getSimpleName(), OPT_READ_DELAY_MS), 5000);
-      return new ConstantDelayQueue<Long>(TimeUnit.MILLISECONDS, delayMs);
+      return new ConstantDelayQueue<>(TimeUnit.MILLISECONDS, delayMs);
     }
   }
 
@@ -145,7 +145,7 @@ public class IntegrationTestRegionReplicaReplication extends IntegrationTestInge
     protected BlockingQueue<Long> createWriteKeysQueue(Configuration conf) {
       this.delayMs = conf.getLong(String.format("%s.%s",
         IntegrationTestRegionReplicaReplication.class.getSimpleName(), OPT_READ_DELAY_MS), 5000);
-      return new ConstantDelayQueue<Long>(TimeUnit.MILLISECONDS, delayMs);
+      return new ConstantDelayQueue<>(TimeUnit.MILLISECONDS, delayMs);
     }
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java
index 04a3b05..03ba460 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java
@@ -336,7 +336,7 @@ public class RESTApiClusterManager extends Configured implements ClusterManager
   // ClusterManager methods take a "ServiceType" object (e.g. "HBASE_MASTER," "HADOOP_NAMENODE").
   // These "service types," which cluster managers call "roles" or "components," need to be mapped
   // to their corresponding service (e.g. "HBase," "HDFS") in order to be controlled.
-  private static Map<ServiceType, Service> roleServiceType = new HashMap<ServiceType, Service>();
+  private static Map<ServiceType, Service> roleServiceType = new HashMap<>();
   static {
     roleServiceType.put(ServiceType.HADOOP_NAMENODE, Service.HDFS);
     roleServiceType.put(ServiceType.HADOOP_DATANODE, Service.HDFS);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java
index d1a32b1..4c7be8c 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java
@@ -123,7 +123,7 @@ public class Action {
     if (count == 1) {
       return new ServerName [] {};
     }
-    ArrayList<ServerName> tmp = new ArrayList<ServerName>(count);
+    ArrayList<ServerName> tmp = new ArrayList<>(count);
     tmp.addAll(regionServers);
     tmp.remove(master);
     return tmp.toArray(new ServerName[count-1]);
@@ -192,11 +192,11 @@ public class Action {
   protected void unbalanceRegions(ClusterStatus clusterStatus,
       List<ServerName> fromServers, List<ServerName> toServers,
       double fractionOfRegions) throws Exception {
-    List<byte[]> victimRegions = new LinkedList<byte[]>();
+    List<byte[]> victimRegions = new LinkedList<>();
     for (ServerName server : fromServers) {
       ServerLoad serverLoad = clusterStatus.getLoad(server);
       // Ugh.
-      List<byte[]> regions = new LinkedList<byte[]>(serverLoad.getRegionsLoad().keySet());
+      List<byte[]> regions = new LinkedList<>(serverLoad.getRegionsLoad().keySet());
       int victimRegionCount = (int)Math.ceil(fractionOfRegions * regions.size());
       LOG.debug("Removing " + victimRegionCount + " regions from " + server.getServerName());
       for (int i = 0; i < victimRegionCount; ++i) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java
index ce66000..75414ae 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java
@@ -43,7 +43,7 @@ public class BatchRestartRsAction extends RestartActionBaseAction {
     List<ServerName> selectedServers = PolicyBasedChaosMonkey.selectRandomItems(getCurrentServers(),
         ratio);
 
-    Set<ServerName> killedServers = new HashSet<ServerName>();
+    Set<ServerName> killedServers = new HashSet<>();
 
     for (ServerName server : selectedServers) {
       // Don't keep killing servers if we're

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java
index 7299e79..f5349dc 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java
@@ -49,7 +49,7 @@ public class RestartRandomDataNodeAction extends RestartActionBaseAction {
     DistributedFileSystem fs = (DistributedFileSystem) FSUtils.getRootDir(getConf())
         .getFileSystem(getConf());
     DFSClient dfsClient = fs.getClient();
-    List<ServerName> hosts = new LinkedList<ServerName>();
+    List<ServerName> hosts = new LinkedList<>();
     for (DatanodeInfo dataNode: dfsClient.datanodeReport(HdfsConstants.DatanodeReportType.LIVE)) {
       hosts.add(ServerName.valueOf(dataNode.getHostName(), -1, -1));
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java
index e79ff5b..ba25198 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java
@@ -59,8 +59,8 @@ public class RollingBatchRestartRsAction extends BatchRestartRsAction {
         (int)(ratio * 100)));
     List<ServerName> selectedServers = selectServers();
 
-    Queue<ServerName> serversToBeKilled = new LinkedList<ServerName>(selectedServers);
-    Queue<ServerName> deadServers = new LinkedList<ServerName>();
+    Queue<ServerName> serversToBeKilled = new LinkedList<>(selectedServers);
+    Queue<ServerName> deadServers = new LinkedList<>();
 
     // loop while there are servers to be killed or dead servers to be restarted
     while ((!serversToBeKilled.isEmpty() || !deadServers.isEmpty())  && !context.isStopping()) {
@@ -123,7 +123,7 @@ public class RollingBatchRestartRsAction extends BatchRestartRsAction {
       @Override
       protected ServerName[] getCurrentServers() throws IOException {
         final int count = 4;
-        List<ServerName> serverNames = new ArrayList<ServerName>(count);
+        List<ServerName> serverNames = new ArrayList<>(count);
         for (int i = 0; i < 4; i++) {
           serverNames.add(ServerName.valueOf(i + ".example.org", i, i));
         }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java
index 1ac1458..a40c8b1 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java
@@ -52,13 +52,13 @@ public class UnbalanceKillAndRebalanceAction extends Action {
   @Override
   public void perform() throws Exception {
     ClusterStatus status = this.cluster.getClusterStatus();
-    List<ServerName> victimServers = new LinkedList<ServerName>(status.getServers());
-    Set<ServerName> killedServers = new HashSet<ServerName>();
+    List<ServerName> victimServers = new LinkedList<>(status.getServers());
+    Set<ServerName> killedServers = new HashSet<>();
 
     int liveCount = (int)Math.ceil(FRC_SERVERS_THAT_HOARD_AND_LIVE * victimServers.size());
     int deadCount = (int)Math.ceil(FRC_SERVERS_THAT_HOARD_AND_DIE * victimServers.size());
     Assert.assertTrue((liveCount + deadCount) < victimServers.size());
-    List<ServerName> targetServers = new ArrayList<ServerName>(liveCount);
+    List<ServerName> targetServers = new ArrayList<>(liveCount);
     for (int i = 0; i < liveCount + deadCount; ++i) {
       int victimIx = RandomUtils.nextInt(victimServers.size());
       targetServers.add(victimServers.remove(victimIx));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java
index 2779bd1..bdffdb1 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java
@@ -48,9 +48,9 @@ public class UnbalanceRegionsAction extends Action {
   public void perform() throws Exception {
     LOG.info("Unbalancing regions");
     ClusterStatus status = this.cluster.getClusterStatus();
-    List<ServerName> victimServers = new LinkedList<ServerName>(status.getServers());
+    List<ServerName> victimServers = new LinkedList<>(status.getServers());
     int targetServerCount = (int)Math.ceil(fractionOfServers * victimServers.size());
-    List<ServerName> targetServers = new ArrayList<ServerName>(targetServerCount);
+    List<ServerName> targetServers = new ArrayList<>(targetServerCount);
     for (int i = 0; i < targetServerCount; ++i) {
       int victimIx = RandomUtils.nextInt(victimServers.size());
       targetServers.add(victimServers.remove(victimIx));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java
index 57f7c83..951f8f8 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java
@@ -92,7 +92,7 @@ public class PolicyBasedChaosMonkey extends ChaosMonkey {
   public static <T> List<T> selectRandomItems(T[] items, float ratio) {
     int remaining = (int)Math.ceil(items.length * ratio);
 
-    List<T> selectedItems = new ArrayList<T>(remaining);
+    List<T> selectedItems = new ArrayList<>(remaining);
 
     for (int i=0; i<items.length && remaining > 0; i++) {
       if (RandomUtils.nextFloat() < ((float)remaining/(items.length-i))) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java
index e03816a..35f06eb 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java
@@ -31,7 +31,7 @@ public class DoActionsOncePolicy extends PeriodicPolicy {
 
   public DoActionsOncePolicy(long periodMs, List<Action> actions) {
     super(periodMs);
-    this.actions = new ArrayList<Action>(actions);
+    this.actions = new ArrayList<>(actions);
   }
 
   public DoActionsOncePolicy(long periodMs, Action... actions) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java
index 8912467..8b76e49 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java
@@ -46,9 +46,9 @@ public class PeriodicRandomActionPolicy extends PeriodicPolicy {
 
   public PeriodicRandomActionPolicy(long periodMs, Action... actions) {
     super(periodMs);
-    this.actions = new ArrayList<Pair<Action, Integer>>(actions.length);
+    this.actions = new ArrayList<>(actions.length);
     for (Action action : actions) {
-      this.actions.add(new Pair<Action, Integer>(action, 1));
+      this.actions.add(new Pair<>(action, 1));
     }
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestBulkLoad.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestBulkLoad.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestBulkLoad.java
index fd062d1..e39d0fe 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestBulkLoad.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestBulkLoad.java
@@ -353,7 +353,7 @@ public class IntegrationTestBulkLoad extends IntegrationTestBase {
     @Override
     public List<InputSplit> getSplits(JobContext context) throws IOException, InterruptedException {
       int numSplits = context.getConfiguration().getInt(NUM_MAPS_KEY, NUM_MAPS);
-      ArrayList<InputSplit> ret = new ArrayList<InputSplit>(numSplits);
+      ArrayList<InputSplit> ret = new ArrayList<>(numSplits);
       for (int i = 0; i < numSplits; ++i) {
         ret.add(new EmptySplit());
       }
@@ -376,7 +376,7 @@ public class IntegrationTestBulkLoad extends IntegrationTestBase {
       chainId = chainId - (chainId % numMapTasks) + taskId; // ensure that chainId is unique per task and across iterations
       LongWritable[] keys = new LongWritable[] {new LongWritable(chainId)};
 
-      return new FixedRecordReader<LongWritable, LongWritable>(keys, keys);
+      return new FixedRecordReader<>(keys, keys);
     }
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
index 42b6ae7..9d04bf9 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/mapreduce/IntegrationTestImportTsv.java
@@ -194,7 +194,7 @@ public class IntegrationTestImportTsv extends Configured implements Tool {
         util.getDataTestDirOnTestFS(table.getNameAsString()), "hfiles");
 
 
-    Map<String, String> args = new HashMap<String, String>();
+    Map<String, String> args = new HashMap<>();
     args.put(ImportTsv.BULK_OUTPUT_CONF_KEY, hfiles.toString());
     args.put(ImportTsv.COLUMNS_CONF_KEY,
         format("HBASE_ROW_KEY,HBASE_TS_KEY,%s:c1,%s:c2", cf, cf));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
index f87cc86..bd14c31 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/mttr/IntegrationTestMTTR.java
@@ -297,9 +297,9 @@ public class IntegrationTestMTTR {
     LOG.info("Starting " + testName + " with " + maxIters + " iterations.");
 
     // Array to keep track of times.
-    ArrayList<TimingResult> resultPuts = new ArrayList<TimingResult>(maxIters);
-    ArrayList<TimingResult> resultScan = new ArrayList<TimingResult>(maxIters);
-    ArrayList<TimingResult> resultAdmin = new ArrayList<TimingResult>(maxIters);
+    ArrayList<TimingResult> resultPuts = new ArrayList<>(maxIters);
+    ArrayList<TimingResult> resultScan = new ArrayList<>(maxIters);
+    ArrayList<TimingResult> resultAdmin = new ArrayList<>(maxIters);
     long start = System.nanoTime();
 
     try {
@@ -357,7 +357,7 @@ public class IntegrationTestMTTR {
    */
   private static class TimingResult {
     DescriptiveStatistics stats = new DescriptiveStatistics();
-    ArrayList<Long> traces = new ArrayList<Long>(10);
+    ArrayList<Long> traces = new ArrayList<>(10);
 
     /**
      * Add a result to this aggregate result.

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
index dff1828..1b23de8 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedList.java
@@ -350,7 +350,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
       public List<InputSplit> getSplits(JobContext job) throws IOException, InterruptedException {
         int numMappers = job.getConfiguration().getInt(GENERATOR_NUM_MAPPERS_KEY, 1);
 
-        ArrayList<InputSplit> splits = new ArrayList<InputSplit>(numMappers);
+        ArrayList<InputSplit> splits = new ArrayList<>(numMappers);
 
         for (int i = 0; i < numMappers; i++) {
           splits.add(new GeneratorInputSplit());
@@ -956,7 +956,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
     throws IOException, InterruptedException {
       Path keysInputDir = new Path(conf.get(SEARCHER_INPUTDIR_KEY));
       FileSystem fs = FileSystem.get(conf);
-      SortedSet<byte []> result = new TreeSet<byte []>(Bytes.BYTES_COMPARATOR);
+      SortedSet<byte []> result = new TreeSet<>(Bytes.BYTES_COMPARATOR);
       if (!fs.exists(keysInputDir)) {
         throw new FileNotFoundException(keysInputDir.toString());
       }
@@ -977,7 +977,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
     private static SortedSet<byte[]> readFileToSearch(final Configuration conf,
         final FileSystem fs, final LocatedFileStatus keyFileStatus) throws IOException,
         InterruptedException {
-      SortedSet<byte []> result = new TreeSet<byte []>(Bytes.BYTES_COMPARATOR);
+      SortedSet<byte []> result = new TreeSet<>(Bytes.BYTES_COMPARATOR);
       // Return entries that are flagged Counts.UNDEFINED in the value. Return the row. This is
       // what is missing.
       TaskAttemptContext context = new TaskAttemptContextImpl(conf, new TaskAttemptID());
@@ -1064,7 +1064,7 @@ public class IntegrationTestBigLinkedList extends IntegrationTestBase {
      */
     public static class VerifyReducer extends
         Reducer<BytesWritable, BytesWritable, BytesWritable, BytesWritable> {
-      private ArrayList<byte[]> refs = new ArrayList<byte[]>();
+      private ArrayList<byte[]> refs = new ArrayList<>();
       private final BytesWritable UNREF = new BytesWritable(addPrefixFlag(
         Counts.UNREFERENCED.ordinal(), new byte[] {}));
       private final BytesWritable LOSTFAM = new BytesWritable(addPrefixFlag(

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
index dec565f..9eacc5a 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestLoadAndVerify.java
@@ -456,7 +456,7 @@ public void cleanUpCluster() throws Exception {
       throws IOException, InterruptedException {
     Path keysInputDir = new Path(conf.get(SEARCHER_INPUTDIR_KEY));
     FileSystem fs = FileSystem.get(conf);
-    SortedSet<byte []> result = new TreeSet<byte []>(Bytes.BYTES_COMPARATOR);
+    SortedSet<byte []> result = new TreeSet<>(Bytes.BYTES_COMPARATOR);
     if (!fs.exists(keysInputDir)) {
       throw new FileNotFoundException(keysInputDir.toString());
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestReplication.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestReplication.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestReplication.java
index 141b24d..bf534f3 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestReplication.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestReplication.java
@@ -234,7 +234,7 @@ public class IntegrationTestReplication extends IntegrationTestBigLinkedList {
 
         // set the test table to be the table to replicate
         HashMap<TableName, ArrayList<String>> toReplicate = new HashMap<>();
-        toReplicate.put(tableName, new ArrayList<String>(0));
+        toReplicate.put(tableName, new ArrayList<>(0));
 
         replicationAdmin.addPeer("TestPeer", peerConfig, toReplicate);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
index b7463bd..327d879 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/trace/IntegrationTestSendTraceRequests.java
@@ -233,7 +233,7 @@ public class IntegrationTestSendTraceRequests extends AbstractHBaseTool {
   }
 
   private LinkedBlockingQueue<Long> insertData() throws IOException, InterruptedException {
-    LinkedBlockingQueue<Long> rowKeys = new LinkedBlockingQueue<Long>(25000);
+    LinkedBlockingQueue<Long> rowKeys = new LinkedBlockingQueue<>(25000);
     BufferedMutator ht = util.getConnection().getBufferedMutator(this.tableName);
     byte[] value = new byte[300];
     for (int x = 0; x < 5000; x++) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/ArraySearcherPool.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/ArraySearcherPool.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/ArraySearcherPool.java
index e6df88a..5aa5d88 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/ArraySearcherPool.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/ArraySearcherPool.java
@@ -42,8 +42,7 @@ public class ArraySearcherPool {
    */
   private static final Integer MAX_POOL_SIZE = 1000;
 
-  protected Queue<PrefixTreeArraySearcher> pool
-    = new LinkedBlockingQueue<PrefixTreeArraySearcher>(MAX_POOL_SIZE);
+  protected Queue<PrefixTreeArraySearcher> pool = new LinkedBlockingQueue<>(MAX_POOL_SIZE);
 
   public PrefixTreeArraySearcher checkOut(ByteBuff buffer, boolean includesMvccVersion) {
     PrefixTreeArraySearcher searcher = pool.poll();//will return null if pool is empty

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeCell.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeCell.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeCell.java
index 3ca4236..255c8a3 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeCell.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeCell.java
@@ -87,7 +87,7 @@ public class PrefixTreeCell extends ByteBufferCell implements SettableSequenceId
   protected int tagsOffset;
   protected int tagsLength;
   // Pair to set the value ByteBuffer and its offset
-  protected ObjectIntPair<ByteBuffer> pair = new ObjectIntPair<ByteBuffer>();
+  protected ObjectIntPair<ByteBuffer> pair = new ObjectIntPair<>();
 
   /********************** Cell methods ******************/
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/EncoderPoolImpl.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/EncoderPoolImpl.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/EncoderPoolImpl.java
index 8a5ffba..a8f0082 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/EncoderPoolImpl.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/EncoderPoolImpl.java
@@ -26,8 +26,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 @InterfaceAudience.Private
 public class EncoderPoolImpl implements EncoderPool {
 
-  private BlockingQueue<PrefixTreeEncoder> unusedEncoders = 
-      new LinkedBlockingQueue<PrefixTreeEncoder>();
+  private BlockingQueue<PrefixTreeEncoder> unusedEncoders = new LinkedBlockingQueue<>();
 
   @Override
   public PrefixTreeEncoder checkOut(OutputStream outputStream, boolean includeMvccVersion) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
index 3291d72..3597fbe 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
@@ -60,7 +60,7 @@ public class LongEncoder {
   /****************** construct ****************************/
 
   public LongEncoder() {
-    this.uniqueValues = new HashSet<Long>();
+    this.uniqueValues = new HashSet<>();
   }
 
   public void reset() {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
index f44017b..e2824b0 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
@@ -161,7 +161,7 @@ public class Tokenizer{
   }
 
   public List<byte[]> getArrays() {
-    List<TokenizerNode> nodes = new ArrayList<TokenizerNode>();
+    List<TokenizerNode> nodes = new ArrayList<>();
     root.appendNodesToExternalList(nodes, true, true);
     List<byte[]> byteArrays = Lists.newArrayListWithCapacity(CollectionUtils.nullSafeSize(nodes));
     for (int i = 0; i < nodes.size(); ++i) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeHashSet.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeHashSet.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeHashSet.java
index 9ce6163..dbaa508 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeHashSet.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeHashSet.java
@@ -39,7 +39,7 @@ public class ByteRangeHashSet extends ByteRangeSet {
   /************************ constructors *****************************/
 
   public ByteRangeHashSet() {
-    this.uniqueIndexByUniqueRange = new HashMap<ByteRange, Integer>();
+    this.uniqueIndexByUniqueRange = new HashMap<>();
   }
 
   public ByteRangeHashSet(List<ByteRange> rawByteArrays) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeTreeSet.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeTreeSet.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeTreeSet.java
index b5c0b1a..4ee7b28 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeTreeSet.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/byterange/impl/ByteRangeTreeSet.java
@@ -36,7 +36,7 @@ public class ByteRangeTreeSet extends ByteRangeSet {
   /************************ constructors *****************************/
 
   public ByteRangeTreeSet() {
-    this.uniqueIndexByUniqueRange = new TreeMap<ByteRange, Integer>();
+    this.uniqueIndexByUniqueRange = new TreeMap<>();
   }
 
   public ByteRangeTreeSet(List<ByteRange> rawByteArrays) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
index 39140a3..1f9b459 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
@@ -43,7 +43,7 @@ public class TestRowDataExerciseFInts extends BaseTestRowData{
 
   static List<ByteRange> rows;
   static{
-    List<String> rowStrings = new ArrayList<String>(16);
+    List<String> rowStrings = new ArrayList<>(16);
     rowStrings.add("com.edsBlog/directoryAa/pageAaa");
     rowStrings.add("com.edsBlog/directoryAa/pageBbb");
     rowStrings.add("com.edsBlog/directoryAa/pageCcc");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivialWithTags.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivialWithTags.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivialWithTags.java
index 2d3901f..a7edfe7 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivialWithTags.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivialWithTags.java
@@ -46,7 +46,7 @@ public class TestRowDataTrivialWithTags extends BaseTestRowData{
 
   static List<KeyValue> d = Lists.newArrayList();
   static {
-    List<Tag> tagList = new ArrayList<Tag>(2);
+    List<Tag> tagList = new ArrayList<>(2);
     Tag t = new ArrayBackedTag((byte) 1, "visisbility");
     tagList.add(t);
     t = new ArrayBackedTag((byte) 2, "ACL");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
index a71daaa..0276617 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
@@ -41,7 +41,7 @@ public class TestRowDataUrls extends BaseTestRowData{
 
   static List<ByteRange> rows;
   static{
-    List<String> rowStrings = new ArrayList<String>(16);
+    List<String> rowStrings = new ArrayList<>(16);
     rowStrings.add("com.edsBlog/directoryAa/pageAaa");
     rowStrings.add("com.edsBlog/directoryAa/pageBbb");
     rowStrings.add("com.edsBlog/directoryAa/pageCcc");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataBasic.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataBasic.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataBasic.java
index bccff6d..d4fbb4d 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataBasic.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataBasic.java
@@ -27,7 +27,7 @@ public class TestTimestampDataBasic implements TestTimestampData {
 
   @Override
   public List<Long> getInputs() {
-    List<Long> d = new ArrayList<Long>(5);
+    List<Long> d = new ArrayList<>(5);
     d.add(5L);
     d.add(3L);
     d.add(0L);
@@ -43,7 +43,7 @@ public class TestTimestampDataBasic implements TestTimestampData {
 
   @Override
   public List<Long> getOutputs() {
-    List<Long> d = new ArrayList<Long>(4);
+    List<Long> d = new ArrayList<>(4);
     d.add(0L);
     d.add(1L);
     d.add(3L);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataNumbers.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataNumbers.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataNumbers.java
index 2a5dcae..d0bc837 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataNumbers.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataNumbers.java
@@ -29,7 +29,7 @@ public class TestTimestampDataNumbers implements TestTimestampData {
 
   @Override
   public List<Long> getInputs() {
-    List<Long> d = new ArrayList<Long>(5);
+    List<Long> d = new ArrayList<>(5);
     d.add(5L << shift);
     d.add(3L << shift);
     d.add(7L << shift);
@@ -45,7 +45,7 @@ public class TestTimestampDataNumbers implements TestTimestampData {
 
   @Override
   public List<Long> getOutputs() {
-    List<Long> d = new ArrayList<Long>(4);
+    List<Long> d = new ArrayList<>(4);
     d.add(1L << shift);
     d.add(3L << shift);
     d.add(5L << shift);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataRepeats.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataRepeats.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataRepeats.java
index 2186528..3320d66 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataRepeats.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/timestamp/data/TestTimestampDataRepeats.java
@@ -29,7 +29,7 @@ public class TestTimestampDataRepeats implements TestTimestampData {
 
   @Override
   public List<Long> getInputs() {
-    List<Long> d = new ArrayList<Long>(5);
+    List<Long> d = new ArrayList<>(5);
     d.add(t);
     d.add(t);
     d.add(t);
@@ -45,7 +45,7 @@ public class TestTimestampDataRepeats implements TestTimestampData {
 
   @Override
   public List<Long> getOutputs() {
-    List<Long> d = new ArrayList<Long>();
+    List<Long> d = new ArrayList<>();
     return d;
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
index b38b96c..0856aa2 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
@@ -209,33 +209,28 @@ public class ProcedureExecutor<TEnvironment> {
    * Once a Root-Procedure completes (success or failure), the result will be added to this map.
    * The user of ProcedureExecutor should call getResult(procId) to get the result.
    */
-  private final ConcurrentHashMap<Long, ProcedureInfo> completed =
-    new ConcurrentHashMap<Long, ProcedureInfo>();
+  private final ConcurrentHashMap<Long, ProcedureInfo> completed = new ConcurrentHashMap<>();
 
   /**
    * Map the the procId returned by submitProcedure(), the Root-ProcID, to the RootProcedureState.
    * The RootProcedureState contains the execution stack of the Root-Procedure,
    * It is added to the map by submitProcedure() and removed on procedure completion.
    */
-  private final ConcurrentHashMap<Long, RootProcedureState> rollbackStack =
-    new ConcurrentHashMap<Long, RootProcedureState>();
+  private final ConcurrentHashMap<Long, RootProcedureState> rollbackStack = new ConcurrentHashMap<>();
 
   /**
    * Helper map to lookup the live procedures by ID.
    * This map contains every procedure. root-procedures and subprocedures.
    */
-  private final ConcurrentHashMap<Long, Procedure> procedures =
-    new ConcurrentHashMap<Long, Procedure>();
+  private final ConcurrentHashMap<Long, Procedure> procedures = new ConcurrentHashMap<>();
 
   /**
    * Helper map to lookup whether the procedure already issued from the same client.
    * This map contains every root procedure.
    */
-  private final ConcurrentHashMap<NonceKey, Long> nonceKeysToProcIdsMap =
-      new ConcurrentHashMap<NonceKey, Long>();
+  private final ConcurrentHashMap<NonceKey, Long> nonceKeysToProcIdsMap = new ConcurrentHashMap<>();
 
-  private final CopyOnWriteArrayList<ProcedureExecutorListener> listeners =
-    new CopyOnWriteArrayList<ProcedureExecutorListener>();
+  private final CopyOnWriteArrayList<ProcedureExecutorListener> listeners = new CopyOnWriteArrayList<>();
 
   private Configuration conf;
   private ThreadGroup threadGroup;
@@ -399,7 +394,7 @@ public class ProcedureExecutor<TEnvironment> {
           break;
         case WAITING_TIMEOUT:
           if (waitingSet == null) {
-            waitingSet = new HashSet<Procedure>();
+            waitingSet = new HashSet<>();
           }
           waitingSet.add(proc);
           break;
@@ -498,7 +493,7 @@ public class ProcedureExecutor<TEnvironment> {
 
     // Create the workers
     workerId.set(0);
-    workerThreads = new CopyOnWriteArrayList<WorkerThread>();
+    workerThreads = new CopyOnWriteArrayList<>();
     for (int i = 0; i < corePoolSize; ++i) {
       workerThreads.add(new WorkerThread(threadGroup));
     }
@@ -979,8 +974,7 @@ public class ProcedureExecutor<TEnvironment> {
    * @return the procedures in a list
    */
   public List<ProcedureInfo> listProcedures() {
-    final List<ProcedureInfo> procedureLists =
-        new ArrayList<ProcedureInfo>(procedures.size() + completed.size());
+    final List<ProcedureInfo> procedureLists = new ArrayList<>(procedures.size() + completed.size());
     for (Map.Entry<Long, Procedure> p: procedures.entrySet()) {
       procedureLists.add(ProcedureUtil.convertToProcedureInfo(p.getValue()));
     }
@@ -1614,7 +1608,7 @@ public class ProcedureExecutor<TEnvironment> {
   //  Timeout Thread
   // ==========================================================================
   private final class TimeoutExecutorThread extends StoppableThread {
-    private final DelayQueue<DelayedWithTimeout> queue = new DelayQueue<DelayedWithTimeout>();
+    private final DelayQueue<DelayedWithTimeout> queue = new DelayQueue<>();
 
     public TimeoutExecutorThread(final ThreadGroup group) {
       super(group, "ProcedureTimeoutExecutor");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RootProcedureState.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RootProcedureState.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RootProcedureState.java
index 2f118b7..4f9b136 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RootProcedureState.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RootProcedureState.java
@@ -147,7 +147,7 @@ class RootProcedureState {
       state = State.FAILED;
     }
     if (subprocStack == null) {
-      subprocStack = new ArrayList<Procedure>();
+      subprocStack = new ArrayList<>();
     }
     proc.addStackIndex(subprocStack.size());
     subprocStack.add(proc);
@@ -156,7 +156,7 @@ class RootProcedureState {
   protected synchronized void addSubProcedure(final Procedure proc) {
     if (!proc.hasParent()) return;
     if (subprocs == null) {
-      subprocs = new HashSet<Procedure>();
+      subprocs = new HashSet<>();
     }
     subprocs.add(proc);
   }
@@ -173,7 +173,7 @@ class RootProcedureState {
     int[] stackIndexes = proc.getStackIndexes();
     if (stackIndexes != null) {
       if (subprocStack == null) {
-        subprocStack = new ArrayList<Procedure>();
+        subprocStack = new ArrayList<>();
       }
       int diff = (1 + stackIndexes[stackIndexes.length - 1]) - subprocStack.size();
       if (diff > 0) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
index 3f9a7b7..5c3a4c7 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
@@ -130,7 +130,7 @@ public abstract class StateMachineProcedure<TEnvironment, TState>
    */
   protected void addChildProcedure(Procedure... subProcedure) {
     if (subProcList == null) {
-      subProcList = new ArrayList<Procedure>(subProcedure.length);
+      subProcList = new ArrayList<>(subProcedure.length);
     }
     for (int i = 0; i < subProcedure.length; ++i) {
       Procedure proc = subProcedure[i];

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreBase.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreBase.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreBase.java
index 0e0e46f..63eff37 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreBase.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/ProcedureStoreBase.java
@@ -25,8 +25,7 @@ import java.util.concurrent.atomic.AtomicBoolean;
  * Base class for {@link ProcedureStore}s.
  */
 public abstract class ProcedureStoreBase implements ProcedureStore {
-  private final CopyOnWriteArrayList<ProcedureStoreListener> listeners =
-      new CopyOnWriteArrayList<ProcedureStoreListener>();
+  private final CopyOnWriteArrayList<ProcedureStoreListener> listeners = new CopyOnWriteArrayList<>();
 
   private final AtomicBoolean running = new AtomicBoolean(false);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
index ec59607..5ad96e1 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/ProcedureWALPrettyPrinter.java
@@ -155,7 +155,7 @@ public class ProcedureWALPrettyPrinter extends Configured implements Tool {
     options.addOption("h", "help", false, "Output help message");
     options.addOption("f", "file", true, "File to print");
 
-    final List<Path> files = new ArrayList<Path>();
+    final List<Path> files = new ArrayList<>();
     try {
       CommandLine cmd = new PosixParser().parse(options, args);
 


Mime
View raw message