hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From syuanji...@apache.org
Subject [28/50] [abbrv] hbase git commit: HBASE-17532 Replaced explicit type with diamond operator
Date Fri, 10 Mar 2017 22:10:08 GMT
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesClientImpl.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesClientImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesClientImpl.java
index dcbed7a..3507547 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesClientImpl.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesClientImpl.java
@@ -72,7 +72,7 @@ public class TableBasedReplicationQueuesClientImpl extends ReplicationTableBase
 
   @Override
   public Set<String> getAllWALs() {
-    Set<String> allWals = new HashSet<String>();
+    Set<String> allWals = new HashSet<>();
     ResultScanner allQueues = null;
     try (Table replicationTable = getOrBlockOnReplicationTable()) {
       allQueues = replicationTable.getScanner(new Scan());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesImpl.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesImpl.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesImpl.java
index 1023e0d..bf55e8c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesImpl.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/TableBasedReplicationQueuesImpl.java
@@ -201,7 +201,7 @@ public class TableBasedReplicationQueuesImpl extends ReplicationTableBase
   public List<String> getLogsInQueue(String queueId) {
     String errMsg = "Failed getting logs in queue queueId=" + queueId;
     byte[] rowKey = queueIdToRowKey(queueId);
-    List<String> logs = new ArrayList<String>();
+    List<String> logs = new ArrayList<>();
     try {
       Get getQueue = new Get(rowKey);
       Result queue = getResultIfOwner(getQueue);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SecurityInfo.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SecurityInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SecurityInfo.java
index e48f81d..0e8a68d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SecurityInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SecurityInfo.java
@@ -33,7 +33,7 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProto
 @InterfaceAudience.Private
 public class SecurityInfo {
   /** Maps RPC service names to authentication information */
-  private static ConcurrentMap<String,SecurityInfo> infos = new ConcurrentHashMap<String,SecurityInfo>();
+  private static ConcurrentMap<String,SecurityInfo> infos = new ConcurrentHashMap<>();
   // populate info for known services
   static {
     infos.put(AdminProtos.AdminService.getDescriptor().getName(),

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
index eeac9c7..1c4a868 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
@@ -266,7 +266,7 @@ public class AccessControlClient {
     HBaseRpcController controller
       = ((ClusterConnection) connection).getRpcControllerFactory().newController();
       */
-    List<UserPermission> permList = new ArrayList<UserPermission>();
+    List<UserPermission> permList = new ArrayList<>();
     try (Table table = connection.getTable(ACL_TABLE_NAME)) {
       try (Admin admin = connection.getAdmin()) {
         CoprocessorRpcChannel service = table.coprocessorService(HConstants.EMPTY_START_ROW);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlUtil.java
index 1d26366..1873ea3 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlUtil.java
@@ -367,7 +367,7 @@ public class AccessControlUtil {
    */
   public static List<Permission.Action> toPermissionActions(
       List<AccessControlProtos.Permission.Action> protoActions) {
-    List<Permission.Action> actions = new ArrayList<Permission.Action>(protoActions.size());
+    List<Permission.Action> actions = new ArrayList<>(protoActions.size());
     for (AccessControlProtos.Permission.Action a : protoActions) {
       actions.add(toPermissionAction(a));
     }
@@ -644,7 +644,7 @@ public class AccessControlUtil {
     AccessControlProtos.GetUserPermissionsRequest request = builder.build();
     AccessControlProtos.GetUserPermissionsResponse response =
         protocol.getUserPermissions(controller, request);
-    List<UserPermission> perms = new ArrayList<UserPermission>(response.getUserPermissionCount());
+    List<UserPermission> perms = new ArrayList<>(response.getUserPermissionCount());
     for (AccessControlProtos.UserPermission perm: response.getUserPermissionList()) {
       perms.add(toUserPermission(perm));
     }
@@ -672,7 +672,7 @@ public class AccessControlUtil {
     AccessControlProtos.GetUserPermissionsRequest request = builder.build();
     AccessControlProtos.GetUserPermissionsResponse response =
         protocol.getUserPermissions(controller, request);
-    List<UserPermission> perms = new ArrayList<UserPermission>(response.getUserPermissionCount());
+    List<UserPermission> perms = new ArrayList<>(response.getUserPermissionCount());
     for (AccessControlProtos.UserPermission perm: response.getUserPermissionList()) {
       perms.add(toUserPermission(perm));
     }
@@ -700,7 +700,7 @@ public class AccessControlUtil {
     AccessControlProtos.GetUserPermissionsRequest request = builder.build();
     AccessControlProtos.GetUserPermissionsResponse response =
         protocol.getUserPermissions(controller, request);
-    List<UserPermission> perms = new ArrayList<UserPermission>(response.getUserPermissionCount());
+    List<UserPermission> perms = new ArrayList<>(response.getUserPermissionCount());
     for (AccessControlProtos.UserPermission perm: response.getUserPermissionList()) {
       perms.add(toUserPermission(perm));
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/Authorizations.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/Authorizations.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/Authorizations.java
index 4b3ed54..5fdeee9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/Authorizations.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/Authorizations.java
@@ -34,7 +34,7 @@ public class Authorizations {
 
   private List<String> labels;
   public Authorizations(String... labels) {
-    this.labels = new ArrayList<String>(labels.length);
+    this.labels = new ArrayList<>(labels.length);
     Collections.addAll(this.labels, labels);
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
index cd153f1..d87bf14 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
@@ -130,7 +130,7 @@ public class VisibilityClient {
           new Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse>() {
             ServerRpcController controller = new ServerRpcController();
             CoprocessorRpcUtils.BlockingRpcCallback<VisibilityLabelsResponse> rpcCallback =
-                new CoprocessorRpcUtils.BlockingRpcCallback<VisibilityLabelsResponse>();
+                new CoprocessorRpcUtils.BlockingRpcCallback<>();
 
             public VisibilityLabelsResponse call(VisibilityLabelsService service)
                 throws IOException {
@@ -215,7 +215,7 @@ public class VisibilityClient {
             new Batch.Call<VisibilityLabelsService, GetAuthsResponse>() {
           ServerRpcController controller = new ServerRpcController();
           CoprocessorRpcUtils.BlockingRpcCallback<GetAuthsResponse> rpcCallback =
-              new CoprocessorRpcUtils.BlockingRpcCallback<GetAuthsResponse>();
+              new CoprocessorRpcUtils.BlockingRpcCallback<>();
 
           public GetAuthsResponse call(VisibilityLabelsService service) throws IOException {
             GetAuthsRequest.Builder getAuthReqBuilder = GetAuthsRequest.newBuilder();
@@ -268,7 +268,7 @@ public class VisibilityClient {
           new Batch.Call<VisibilityLabelsService, ListLabelsResponse>() {
             ServerRpcController controller = new ServerRpcController();
             CoprocessorRpcUtils.BlockingRpcCallback<ListLabelsResponse> rpcCallback =
-                new CoprocessorRpcUtils.BlockingRpcCallback<ListLabelsResponse>();
+                new CoprocessorRpcUtils.BlockingRpcCallback<>();
 
             public ListLabelsResponse call(VisibilityLabelsService service) throws IOException {
               ListLabelsRequest.Builder listAuthLabelsReqBuilder = ListLabelsRequest.newBuilder();
@@ -340,7 +340,7 @@ public class VisibilityClient {
             new Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse>() {
           ServerRpcController controller = new ServerRpcController();
           CoprocessorRpcUtils.BlockingRpcCallback<VisibilityLabelsResponse> rpcCallback =
-              new CoprocessorRpcUtils.BlockingRpcCallback<VisibilityLabelsResponse>();
+              new CoprocessorRpcUtils.BlockingRpcCallback<>();
 
           public VisibilityLabelsResponse call(VisibilityLabelsService service) throws IOException {
             SetAuthsRequest.Builder setAuthReqBuilder = SetAuthsRequest.newBuilder();

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
index 24302be..38ae04a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ProtobufUtil.java
@@ -191,8 +191,7 @@ public final class ProtobufUtil {
   /**
    * Primitive type to class mapping.
    */
-  private final static Map<String, Class<?>>
-    PRIMITIVES = new HashMap<String, Class<?>>();
+  private final static Map<String, Class<?>> PRIMITIVES = new HashMap<>();
 
   /**
    * Many results are simple: no cell, exists true or false. To save on object creations,
@@ -1491,7 +1490,7 @@ public final class ProtobufUtil {
       return proto.getStale() ? EMPTY_RESULT_STALE : EMPTY_RESULT;
     }
 
-    List<Cell> cells = new ArrayList<Cell>(values.size());
+    List<Cell> cells = new ArrayList<>(values.size());
     for (CellProtos.Cell c : values) {
       cells.add(toCell(c));
     }
@@ -1525,7 +1524,7 @@ public final class ProtobufUtil {
     List<Cell> cells = null;
     if (proto.hasAssociatedCellCount()) {
       int count = proto.getAssociatedCellCount();
-      cells = new ArrayList<Cell>(count + values.size());
+      cells = new ArrayList<>(count + values.size());
       for (int i = 0; i < count; i++) {
         if (!scanner.advance()) throw new IOException("Failed get " + i + " of " + count);
         cells.add(scanner.current());
@@ -1533,7 +1532,7 @@ public final class ProtobufUtil {
     }
 
     if (!values.isEmpty()){
-      if (cells == null) cells = new ArrayList<Cell>(values.size());
+      if (cells == null) cells = new ArrayList<>(values.size());
       for (CellProtos.Cell c: values) {
         cells.add(toCell(c));
       }
@@ -1903,7 +1902,7 @@ public final class ProtobufUtil {
    */
   static List<HRegionInfo> getRegionInfos(final GetOnlineRegionResponse proto) {
     if (proto == null) return null;
-    List<HRegionInfo> regionInfos = new ArrayList<HRegionInfo>(proto.getRegionInfoList().size());
+    List<HRegionInfo> regionInfos = new ArrayList<>(proto.getRegionInfoList().size());
     for (RegionInfo regionInfo: proto.getRegionInfoList()) {
       regionInfos.add(HRegionInfo.convert(regionInfo));
     }
@@ -2719,7 +2718,7 @@ public final class ProtobufUtil {
 
   public static List<ReplicationLoadSource> toReplicationLoadSourceList(
       List<ClusterStatusProtos.ReplicationLoadSource> clsList) {
-    ArrayList<ReplicationLoadSource> rlsList = new ArrayList<ReplicationLoadSource>(clsList.size());
+    ArrayList<ReplicationLoadSource> rlsList = new ArrayList<>(clsList.size());
     for (ClusterStatusProtos.ReplicationLoadSource cls : clsList) {
       rlsList.add(toReplicationLoadSource(cls));
     }
@@ -2976,26 +2975,26 @@ public final class ProtobufUtil {
   public static ClusterStatus convert(ClusterStatusProtos.ClusterStatus proto) {
 
     Map<ServerName, ServerLoad> servers = null;
-    servers = new HashMap<ServerName, ServerLoad>(proto.getLiveServersList().size());
+    servers = new HashMap<>(proto.getLiveServersList().size());
     for (LiveServerInfo lsi : proto.getLiveServersList()) {
       servers.put(ProtobufUtil.toServerName(
           lsi.getServer()), new ServerLoad(lsi.getServerLoad()));
     }
 
     Collection<ServerName> deadServers = null;
-    deadServers = new ArrayList<ServerName>(proto.getDeadServersList().size());
+    deadServers = new ArrayList<>(proto.getDeadServersList().size());
     for (HBaseProtos.ServerName sn : proto.getDeadServersList()) {
       deadServers.add(ProtobufUtil.toServerName(sn));
     }
 
     Collection<ServerName> backupMasters = null;
-    backupMasters = new ArrayList<ServerName>(proto.getBackupMastersList().size());
+    backupMasters = new ArrayList<>(proto.getBackupMastersList().size());
     for (HBaseProtos.ServerName sn : proto.getBackupMastersList()) {
       backupMasters.add(ProtobufUtil.toServerName(sn));
     }
 
     Set<RegionState> rit = null;
-    rit = new HashSet<RegionState>(proto.getRegionsInTransitionList().size());
+    rit = new HashSet<>(proto.getRegionsInTransitionList().size());
     for (RegionInTransition region : proto.getRegionsInTransitionList()) {
       RegionState value = RegionState.convert(region.getRegionState());
       rit.add(value);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
index d3ef7b8..998b3c0 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java
@@ -572,8 +572,7 @@ public final class RequestConverter {
             .setService(userToken.getService().toString()).build();
     }
 
-    List<ClientProtos.BulkLoadHFileRequest.FamilyPath> protoFamilyPaths =
-        new ArrayList<ClientProtos.BulkLoadHFileRequest.FamilyPath>(familyPaths.size());
+    List<ClientProtos.BulkLoadHFileRequest.FamilyPath> protoFamilyPaths = new ArrayList<>(familyPaths.size());
     if (!familyPaths.isEmpty()) {
       ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder pathBuilder
         = ClientProtos.BulkLoadHFileRequest.FamilyPath.newBuilder();

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ResponseConverter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ResponseConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ResponseConverter.java
index cbcad80..ecadbbc 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ResponseConverter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/ResponseConverter.java
@@ -233,7 +233,7 @@ public final class ResponseConverter {
   public static List<RegionOpeningState> getRegionOpeningStateList(
       final OpenRegionResponse proto) {
     if (proto == null) return null;
-    List<RegionOpeningState> regionOpeningStates = new ArrayList<RegionOpeningState>(proto.getOpeningStateCount());
+    List<RegionOpeningState> regionOpeningStates = new ArrayList<>(proto.getOpeningStateCount());
     for (int i = 0; i < proto.getOpeningStateCount(); i++) {
       regionOpeningStates.add(RegionOpeningState.valueOf(
           proto.getOpeningState(i).name()));
@@ -394,7 +394,7 @@ public final class ResponseConverter {
         boolean isPartial =
             response.getPartialFlagPerResultCount() > i ?
                 response.getPartialFlagPerResult(i) : false;
-        List<Cell> cells = new ArrayList<Cell>(noOfCells);
+        List<Cell> cells = new ArrayList<>(noOfCells);
         for (int j = 0; j < noOfCells; j++) {
           try {
             if (cellScanner.advance() == false) {
@@ -426,7 +426,7 @@ public final class ResponseConverter {
   }
 
   public static Map<String, Long> getScanMetrics(ScanResponse response) {
-    Map<String, Long> metricMap = new HashMap<String, Long>();
+    Map<String, Long> metricMap = new HashMap<>();
     if (response == null || !response.hasScanMetrics() || response.getScanMetrics() == null) {
       return metricMap;
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java
index b683fcc..2131db3 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java
@@ -57,7 +57,7 @@ public class PoolMap<K, V> implements Map<K, V> {
 
   private int poolMaxSize;
 
-  private Map<K, Pool<V>> pools = new ConcurrentHashMap<K, Pool<V>>();
+  private Map<K, Pool<V>> pools = new ConcurrentHashMap<>();
 
   public PoolMap(PoolType poolType) {
     this.poolType = poolType;
@@ -107,7 +107,7 @@ public class PoolMap<K, V> implements Map<K, V> {
 
   @Override
   public Collection<V> values() {
-    Collection<V> values = new ArrayList<V>();
+    Collection<V> values = new ArrayList<>();
     for (Pool<V> pool : pools.values()) {
       Collection<V> poolValues = pool.values();
       if (poolValues != null) {
@@ -118,7 +118,7 @@ public class PoolMap<K, V> implements Map<K, V> {
   }
 
   public Collection<V> values(K key) {
-    Collection<V> values = new ArrayList<V>();
+    Collection<V> values = new ArrayList<>();
     Pool<V> pool = pools.get(key);
     if (pool != null) {
       Collection<V> poolValues = pool.values();
@@ -185,7 +185,7 @@ public class PoolMap<K, V> implements Map<K, V> {
 
   @Override
   public Set<Map.Entry<K, V>> entrySet() {
-    Set<Map.Entry<K, V>> entries = new HashSet<Entry<K, V>>();
+    Set<Map.Entry<K, V>> entries = new HashSet<>();
     for (Map.Entry<K, Pool<V>> poolEntry : pools.entrySet()) {
       final K poolKey = poolEntry.getKey();
       final Pool<V> pool = poolEntry.getValue();
@@ -271,11 +271,11 @@ public class PoolMap<K, V> implements Map<K, V> {
   protected Pool<V> createPool() {
     switch (poolType) {
     case Reusable:
-      return new ReusablePool<V>(poolMaxSize);
+      return new ReusablePool<>(poolMaxSize);
     case RoundRobin:
-      return new RoundRobinPool<V>(poolMaxSize);
+      return new RoundRobinPool<>(poolMaxSize);
     case ThreadLocal:
-      return new ThreadLocalPool<V>();
+      return new ThreadLocalPool<>();
     }
     return null;
   }
@@ -389,7 +389,7 @@ public class PoolMap<K, V> implements Map<K, V> {
    *          the type of the resource
    */
   static class ThreadLocalPool<R> extends ThreadLocal<R> implements Pool<R> {
-    private static final Map<ThreadLocalPool<?>, AtomicInteger> poolSizes = new HashMap<ThreadLocalPool<?>, AtomicInteger>();
+    private static final Map<ThreadLocalPool<?>, AtomicInteger> poolSizes = new HashMap<>();
 
     public ThreadLocalPool() {
     }
@@ -441,7 +441,7 @@ public class PoolMap<K, V> implements Map<K, V> {
 
     @Override
     public Collection<R> values() {
-      List<R> values = new ArrayList<R>();
+      List<R> values = new ArrayList<>();
       values.add(get());
       return values;
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/util/Writables.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/util/Writables.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/util/Writables.java
index 940d523..abe3079 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/util/Writables.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/util/Writables.java
@@ -68,7 +68,7 @@ public class Writables {
    * @throws IOException e
    */
   public static byte [] getBytes(final Writable... ws) throws IOException {
-    List<byte []> bytes = new ArrayList<byte []>(ws.length);
+    List<byte []> bytes = new ArrayList<>(ws.length);
     int size = 0;
     for (Writable w: ws) {
       byte [] b = getBytes(w);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
index 6127997..0090b6f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java
@@ -106,7 +106,7 @@ public class HQuorumPeer {
         conf.get("hbase.zookeeper.dns.interface","default"),
         conf.get("hbase.zookeeper.dns.nameserver","default")));
 
-    List<String> ips = new ArrayList<String>();
+    List<String> ips = new ArrayList<>();
 
     // Add what could be the best (configured) match
     ips.add(myAddress.contains(".") ?

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/InstancePending.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/InstancePending.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/InstancePending.java
index 7458ac7..e63bfc5 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/InstancePending.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/InstancePending.java
@@ -74,7 +74,7 @@ class InstancePending<T> {
    */
   void prepare(T instance) {
     assert instance != null;
-    instanceHolder = new InstanceHolder<T>(instance);
+    instanceHolder = new InstanceHolder<>(instance);
     pendingLatch.countDown();
   }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java
index e8431a2..afab54a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java
@@ -106,8 +106,8 @@ public class MetaTableLocator {
   public List<Pair<HRegionInfo, ServerName>> getMetaRegionsAndLocations(ZooKeeperWatcher zkw,
       int replicaId) {
     ServerName serverName = getMetaRegionLocation(zkw, replicaId);
-    List<Pair<HRegionInfo, ServerName>> list = new ArrayList<Pair<HRegionInfo, ServerName>>(1);
-    list.add(new Pair<HRegionInfo, ServerName>(RegionReplicaUtil.getRegionInfoForReplica(
+    List<Pair<HRegionInfo, ServerName>> list = new ArrayList<>(1);
+    list.add(new Pair<>(RegionReplicaUtil.getRegionInfoForReplica(
         HRegionInfo.FIRST_META_REGIONINFO, replicaId), serverName));
     return list;
   }
@@ -135,7 +135,7 @@ public class MetaTableLocator {
   private List<HRegionInfo> getListOfHRegionInfos(
       final List<Pair<HRegionInfo, ServerName>> pairs) {
     if (pairs == null || pairs.isEmpty()) return null;
-    List<HRegionInfo> result = new ArrayList<HRegionInfo>(pairs.size());
+    List<HRegionInfo> result = new ArrayList<>(pairs.size());
     for (Pair<HRegionInfo, ServerName> pair: pairs) {
       result.add(pair.getFirst());
     }
@@ -550,7 +550,7 @@ public class MetaTableLocator {
           throws InterruptedException {
     int numReplicasConfigured = 1;
 
-    List<ServerName> servers = new ArrayList<ServerName>();
+    List<ServerName> servers = new ArrayList<>();
     // Make the blocking call first so that we do the wait to know
     // the znodes are all in place or timeout.
     ServerName server = blockUntilAvailable(zkw, timeout);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/PendingWatcher.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/PendingWatcher.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/PendingWatcher.java
index 11d0e5d..da7d176 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/PendingWatcher.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/PendingWatcher.java
@@ -33,7 +33,7 @@ import org.apache.zookeeper.Watcher;
  * and then call the method {@code PendingWatcher.prepare}.
  */
 class PendingWatcher implements Watcher {
-  private final InstancePending<Watcher> pending = new InstancePending<Watcher>();
+  private final InstancePending<Watcher> pending = new InstancePending<>();
 
   @Override
   public void process(WatchedEvent event) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
index 14532cf..43a5ad9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
@@ -637,7 +637,7 @@ public class RecoverableZooKeeper {
   throws UnsupportedOperationException {
     if(ops == null) return null;
 
-    List<Op> preparedOps = new LinkedList<Op>();
+    List<Op> preparedOps = new LinkedList<>();
     for (Op op : ops) {
       if (op.getType() == ZooDefs.OpCode.create) {
         CreateRequest create = (CreateRequest)op.toRequestRecord();
@@ -777,7 +777,7 @@ public class RecoverableZooKeeper {
    */
   private static List<String> filterByPrefix(List<String> nodes,
       String... prefixes) {
-    List<String> lockChildren = new ArrayList<String>();
+    List<String> lockChildren = new ArrayList<>();
     for (String child : nodes){
       for (String prefix : prefixes){
         if (child.startsWith(prefix)){

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
index 3e00e04..c678a7c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
@@ -235,8 +235,7 @@ public class ZKUtil {
     private static final String CLIENT_KEYTAB_KERBEROS_CONFIG_NAME =
       "zookeeper-client-keytab-kerberos";
 
-    private static final Map<String, String> BASIC_JAAS_OPTIONS =
-      new HashMap<String,String>();
+    private static final Map<String, String> BASIC_JAAS_OPTIONS = new HashMap<>();
     static {
       String jaasEnvVar = System.getenv("HBASE_JAAS_DEBUG");
       if (jaasEnvVar != null && "true".equalsIgnoreCase(jaasEnvVar)) {
@@ -244,8 +243,7 @@ public class ZKUtil {
       }
     }
 
-    private static final Map<String,String> KEYTAB_KERBEROS_OPTIONS =
-      new HashMap<String,String>();
+    private static final Map<String,String> KEYTAB_KERBEROS_OPTIONS = new HashMap<>();
     static {
       KEYTAB_KERBEROS_OPTIONS.put("doNotPrompt", "true");
       KEYTAB_KERBEROS_OPTIONS.put("storeKey", "true");
@@ -746,7 +744,7 @@ public class ZKUtil {
     List<String> nodes =
       ZKUtil.listChildrenAndWatchForNewChildren(zkw, baseNode);
     if (nodes != null) {
-      List<NodeAndData> newNodes = new ArrayList<NodeAndData>();
+      List<NodeAndData> newNodes = new ArrayList<>();
       for (String node : nodes) {
         String nodePath = ZKUtil.joinZNode(baseNode, node);
         byte[] data = ZKUtil.getDataAndWatch(zkw, nodePath);
@@ -905,11 +903,11 @@ public class ZKUtil {
       return Ids.OPEN_ACL_UNSAFE;
     }
     if (isSecureZooKeeper) {
-      ArrayList<ACL> acls = new ArrayList<ACL>();
+      ArrayList<ACL> acls = new ArrayList<>();
       // add permission to hbase supper user
       String[] superUsers = zkw.getConfiguration().getStrings(Superusers.SUPERUSER_CONF_KEY);
       if (superUsers != null) {
-        List<String> groups = new ArrayList<String>();
+        List<String> groups = new ArrayList<>();
         for (String user : superUsers) {
           if (AuthUtil.isGroupPrincipal(user)) {
             // TODO: Set node ACL for groups when ZK supports this feature
@@ -1327,7 +1325,7 @@ public class ZKUtil {
       LOG.warn("Given path is not valid!");
       return;
     }
-    List<ZKUtilOp> ops = new ArrayList<ZKUtil.ZKUtilOp>();
+    List<ZKUtilOp> ops = new ArrayList<>();
     for (String eachRoot : pathRoots) {
       List<String> children = listChildrenBFSNoWatch(zkw, eachRoot);
       // Delete the leaves first and eventually get rid of the root
@@ -1377,7 +1375,7 @@ public class ZKUtil {
       LOG.warn("Given path is not valid!");
       return;
     }
-    List<ZKUtilOp> ops = new ArrayList<ZKUtil.ZKUtilOp>();
+    List<ZKUtilOp> ops = new ArrayList<>();
     for (String eachRoot : pathRoots) {
       // ZooKeeper Watches are one time triggers; When children of parent nodes are deleted
       // recursively, must set another watch, get notified of delete node
@@ -1415,8 +1413,8 @@ public class ZKUtil {
    */
   private static List<String> listChildrenBFSNoWatch(ZooKeeperWatcher zkw,
       final String znode) throws KeeperException {
-    Deque<String> queue = new LinkedList<String>();
-    List<String> tree = new ArrayList<String>();
+    Deque<String> queue = new LinkedList<>();
+    List<String> tree = new ArrayList<>();
     queue.add(znode);
     while (true) {
       String node = queue.pollFirst();
@@ -1451,8 +1449,8 @@ public class ZKUtil {
    */
   private static List<String> listChildrenBFSAndWatchThem(ZooKeeperWatcher zkw, final String znode)
       throws KeeperException {
-    Deque<String> queue = new LinkedList<String>();
-    List<String> tree = new ArrayList<String>();
+    Deque<String> queue = new LinkedList<>();
+    List<String> tree = new ArrayList<>();
     queue.add(znode);
     while (true) {
       String node = queue.pollFirst();
@@ -1648,7 +1646,7 @@ public class ZKUtil {
     }
     if (ops == null) return;
 
-    List<Op> zkOps = new LinkedList<Op>();
+    List<Op> zkOps = new LinkedList<>();
     for (ZKUtilOp op : ops) {
       zkOps.add(toZooKeeperOp(zkw, op));
     }
@@ -1816,7 +1814,7 @@ public class ZKUtil {
 
   private static void appendRSZnodes(ZooKeeperWatcher zkw, String znode, StringBuilder sb)
       throws KeeperException {
-    List<String> stack = new LinkedList<String>();
+    List<String> stack = new LinkedList<>();
     stack.add(znode);
     do {
       String znodeToProcess = stack.remove(stack.size() - 1);
@@ -1927,7 +1925,7 @@ public class ZKUtil {
         socket.getInputStream()));
       out.println("stat");
       out.flush();
-      ArrayList<String> res = new ArrayList<String>();
+      ArrayList<String> res = new ArrayList<>();
       while (true) {
         String line = in.readLine();
         if (line != null) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
index c8462fb..f18b8ba 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
@@ -83,8 +83,7 @@ public class ZooKeeperWatcher implements Watcher, Abortable, Closeable {
   public final ZNodePaths znodePaths;
 
   // listeners to be notified
-  private final List<ZooKeeperListener> listeners =
-    new CopyOnWriteArrayList<ZooKeeperListener>();
+  private final List<ZooKeeperListener> listeners = new CopyOnWriteArrayList<>();
 
   // Used by ZKUtil:waitForZKConnectionIfAuthenticating to wait for SASL
   // negotiation to complete
@@ -374,7 +373,7 @@ public class ZooKeeperWatcher implements Watcher, Abortable, Closeable {
    */
   public List<String> getMetaReplicaNodes() throws KeeperException {
     List<String> childrenOfBaseNode = ZKUtil.listChildrenNoWatch(this, znodePaths.baseZNode);
-    List<String> metaReplicaNodes = new ArrayList<String>(2);
+    List<String> metaReplicaNodes = new ArrayList<>(2);
     if (childrenOfBaseNode != null) {
       String pattern = conf.get("zookeeper.znode.metaserver","meta-region-server");
       for (String child : childrenOfBaseNode) {
@@ -416,7 +415,7 @@ public class ZooKeeperWatcher implements Watcher, Abortable, Closeable {
    * Get a copy of current registered listeners
    */
   public List<ZooKeeperListener> getListeners() {
-    return new ArrayList<ZooKeeperListener>(listeners);
+    return new ArrayList<>(listeners);
   }
 
   /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/test/java/org/apache/hadoop/hbase/TestInterfaceAudienceAnnotations.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/TestInterfaceAudienceAnnotations.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/TestInterfaceAudienceAnnotations.java
index d0b6317..9acbb43 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/TestInterfaceAudienceAnnotations.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/TestInterfaceAudienceAnnotations.java
@@ -342,7 +342,7 @@ public class TestInterfaceAudienceAnnotations {
   @Test
   public void testProtosInReturnTypes() throws ClassNotFoundException, IOException, LinkageError {
     Set<Class<?>> classes = findPublicClasses();
-    List<Pair<Class<?>, Method>> protosReturnType = new ArrayList<Pair<Class<?>, Method>>();
+    List<Pair<Class<?>, Method>> protosReturnType = new ArrayList<>();
     for (Class<?> clazz : classes) {
       findProtoInReturnType(clazz, protosReturnType);
     }
@@ -374,8 +374,7 @@ public class TestInterfaceAudienceAnnotations {
   @Test
   public void testProtosInParamTypes() throws ClassNotFoundException, IOException, LinkageError {
     Set<Class<?>> classes = findPublicClasses();
-    List<Triple<Class<?>, Method, Class<?>>> protosParamType =
-        new ArrayList<Triple<Class<?>, Method, Class<?>>>();
+    List<Triple<Class<?>, Method, Class<?>>> protosParamType = new ArrayList<>();
     for (Class<?> clazz : classes) {
       findProtoInParamType(clazz, protosParamType);
     }
@@ -395,7 +394,7 @@ public class TestInterfaceAudienceAnnotations {
   @Test
   public void testProtosInConstructors() throws ClassNotFoundException, IOException, LinkageError {
     Set<Class<?>> classes = findPublicClasses();
-    List<Class<?>> classList = new ArrayList<Class<?>>();
+    List<Class<?>> classList = new ArrayList<>();
     for (Class<?> clazz : classes) {
       Constructor<?>[] constructors = clazz.getConstructors();
       for (Constructor<?> cons : constructors) {
@@ -424,7 +423,7 @@ public class TestInterfaceAudienceAnnotations {
 
   private void findProtoInReturnType(Class<?> clazz,
       List<Pair<Class<?>, Method>> protosReturnType) {
-    Pair<Class<?>, Method> returnTypePair = new Pair<Class<?>, Method>();
+    Pair<Class<?>, Method> returnTypePair = new Pair<>();
     Method[] methods = clazz.getMethods();
     returnTypePair.setFirst(clazz);
     for (Method method : methods) {
@@ -443,7 +442,7 @@ public class TestInterfaceAudienceAnnotations {
 
   private void findProtoInParamType(Class<?> clazz,
       List<Triple<Class<?>, Method, Class<?>>> protosParamType) {
-    Triple<Class<?>, Method, Class<?>> paramType = new Triple<Class<?>, Method, Class<?>>();
+    Triple<Class<?>, Method, Class<?>> paramType = new Triple<>();
     Method[] methods = clazz.getMethods();
     paramType.setFirst(clazz);
     for (Method method : methods) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java
index 75199a6..f2f0467 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestAsyncProcess.java
@@ -147,7 +147,7 @@ public class TestAsyncProcess {
   static class MyAsyncProcess extends AsyncProcess {
     final AtomicInteger nbMultiResponse = new AtomicInteger();
     final AtomicInteger nbActions = new AtomicInteger();
-    public List<AsyncRequestFuture> allReqs = new ArrayList<AsyncRequestFuture>();
+    public List<AsyncRequestFuture> allReqs = new ArrayList<>();
     public AtomicInteger callsCt = new AtomicInteger();
 
     private long previousTimeout = -1;
@@ -162,7 +162,7 @@ public class TestAsyncProcess {
           return DUMMY_TABLE;
         }
       };
-      AsyncRequestFutureImpl<Res> r = new MyAsyncRequestFutureImpl<Res>(
+      AsyncRequestFutureImpl<Res> r = new MyAsyncRequestFutureImpl<>(
           wrap, actions, nonceGroup, this);
       allReqs.add(r);
       return r;
@@ -326,9 +326,9 @@ public class TestAsyncProcess {
     }
   }
   class MyAsyncProcessWithReplicas extends MyAsyncProcess {
-    private Set<byte[]> failures = new TreeSet<byte[]>(new Bytes.ByteArrayComparator());
+    private Set<byte[]> failures = new TreeSet<>(new Bytes.ByteArrayComparator());
     private long primarySleepMs = 0, replicaSleepMs = 0;
-    private Map<ServerName, Long> customPrimarySleepMs = new HashMap<ServerName, Long>();
+    private Map<ServerName, Long> customPrimarySleepMs = new HashMap<>();
     private final AtomicLong replicaCalls = new AtomicLong(0);
 
     public void addFailures(HRegionInfo... hris) {
@@ -683,7 +683,7 @@ public class TestAsyncProcess {
     ClusterConnection hc = createHConnection();
     MyAsyncProcess ap = new MyAsyncProcess(hc, CONF);
 
-    List<Put> puts = new ArrayList<Put>(1);
+    List<Put> puts = new ArrayList<>(1);
     puts.add(createPut(1, true));
 
     ap.submit(null, DUMMY_TABLE, puts, false, null, false);
@@ -702,7 +702,7 @@ public class TestAsyncProcess {
     };
     MyAsyncProcess ap = new MyAsyncProcess(hc, CONF);
 
-    List<Put> puts = new ArrayList<Put>(1);
+    List<Put> puts = new ArrayList<>(1);
     puts.add(createPut(1, true));
 
     final AsyncRequestFuture ars = ap.submit(null, DUMMY_TABLE, puts, false, cb, false);
@@ -719,7 +719,7 @@ public class TestAsyncProcess {
       SimpleRequestController.class.getName());
     MyAsyncProcess ap = new MyAsyncProcess(conn, CONF);
     SimpleRequestController controller = (SimpleRequestController) ap.requestController;
-    List<Put> puts = new ArrayList<Put>(1);
+    List<Put> puts = new ArrayList<>(1);
     puts.add(createPut(1, true));
 
     for (int i = 0; i != controller.maxConcurrentTasksPerRegion; ++i) {
@@ -748,7 +748,7 @@ public class TestAsyncProcess {
     SimpleRequestController controller = (SimpleRequestController) ap.requestController;
     controller.taskCounterPerServer.put(sn2, new AtomicInteger(controller.maxConcurrentTasksPerServer));
 
-    List<Put> puts = new ArrayList<Put>(4);
+    List<Put> puts = new ArrayList<>(4);
     puts.add(createPut(1, true));
     puts.add(createPut(3, true)); // <== this one won't be taken, the rs is busy
     puts.add(createPut(1, true)); // <== this one will make it, the region is already in
@@ -770,7 +770,7 @@ public class TestAsyncProcess {
   public void testFail() throws Exception {
     MyAsyncProcess ap = new MyAsyncProcess(createHConnection(), CONF, false);
 
-    List<Put> puts = new ArrayList<Put>(1);
+    List<Put> puts = new ArrayList<>(1);
     Put p = createPut(1, false);
     puts.add(p);
 
@@ -818,7 +818,7 @@ public class TestAsyncProcess {
       }
     };
 
-    List<Put> puts = new ArrayList<Put>(1);
+    List<Put> puts = new ArrayList<>(1);
     Put p = createPut(1, true);
     puts.add(p);
 
@@ -844,7 +844,7 @@ public class TestAsyncProcess {
   public void testFailAndSuccess() throws Exception {
     MyAsyncProcess ap = new MyAsyncProcess(createHConnection(), CONF, false);
 
-    List<Put> puts = new ArrayList<Put>(3);
+    List<Put> puts = new ArrayList<>(3);
     puts.add(createPut(1, false));
     puts.add(createPut(1, true));
     puts.add(createPut(1, true));
@@ -871,7 +871,7 @@ public class TestAsyncProcess {
   public void testFlush() throws Exception {
     MyAsyncProcess ap = new MyAsyncProcess(createHConnection(), CONF, false);
 
-    List<Put> puts = new ArrayList<Put>(3);
+    List<Put> puts = new ArrayList<>(3);
     puts.add(createPut(1, false));
     puts.add(createPut(1, true));
     puts.add(createPut(1, true));
@@ -956,7 +956,7 @@ public class TestAsyncProcess {
       }
     };
 
-    List<Put> puts = new ArrayList<Put>(1);
+    List<Put> puts = new ArrayList<>(1);
     puts.add(createPut(1, true));
 
     t.start();
@@ -981,7 +981,7 @@ public class TestAsyncProcess {
     t2.start();
 
     long start = System.currentTimeMillis();
-    ap.submit(null, DUMMY_TABLE, new ArrayList<Row>(), false, null, false);
+    ap.submit(null, DUMMY_TABLE, new ArrayList<>(), false, null, false);
     long end = System.currentTimeMillis();
 
     //Adds 100 to secure us against approximate timing.
@@ -1008,7 +1008,7 @@ public class TestAsyncProcess {
     setMockLocation(hc, DUMMY_BYTES_1, hrls1);
     setMockLocation(hc, DUMMY_BYTES_2, hrls2);
     setMockLocation(hc, DUMMY_BYTES_3, hrls3);
-    List<HRegionLocation> locations = new ArrayList<HRegionLocation>();
+    List<HRegionLocation> locations = new ArrayList<>();
     for (HRegionLocation loc : hrls1.getRegionLocations()) {
       locations.add(loc);
     }
@@ -1172,7 +1172,7 @@ public class TestAsyncProcess {
     HTable ht = new HTable(conn, mutator);
     ht.multiAp = new MyAsyncProcess(conn, CONF, false);
 
-    List<Put> puts = new ArrayList<Put>(7);
+    List<Put> puts = new ArrayList<>(7);
     puts.add(createPut(1, true));
     puts.add(createPut(1, true));
     puts.add(createPut(1, true));
@@ -1309,8 +1309,8 @@ public class TestAsyncProcess {
   @Test
   public void testThreadCreation() throws Exception {
     final int NB_REGS = 100;
-    List<HRegionLocation> hrls = new ArrayList<HRegionLocation>(NB_REGS);
-    List<Get> gets = new ArrayList<Get>(NB_REGS);
+    List<HRegionLocation> hrls = new ArrayList<>(NB_REGS);
+    List<Get> gets = new ArrayList<>(NB_REGS);
     for (int i = 0; i < NB_REGS; i++) {
       HRegionInfo hri = new HRegionInfo(
           DUMMY_TABLE, Bytes.toBytes(i * 10L), Bytes.toBytes(i * 10L + 9L), false, i);
@@ -1518,7 +1518,7 @@ public class TestAsyncProcess {
   }
 
   private static List<Get> makeTimelineGets(byte[]... rows) {
-    List<Get> result = new ArrayList<Get>(rows.length);
+    List<Get> result = new ArrayList<>(rows.length);
     for (byte[] row : rows) {
       Get get = new Get(row);
       get.setConsistency(Consistency.TIMELINE);
@@ -1609,10 +1609,10 @@ public class TestAsyncProcess {
     ClusterConnection hc = createHConnection();
     MyThreadPoolExecutor myPool =
         new MyThreadPoolExecutor(1, 20, 60, TimeUnit.SECONDS,
-            new LinkedBlockingQueue<Runnable>(200));
+            new LinkedBlockingQueue<>(200));
     AsyncProcess ap = new AsyncProcessForThrowableCheck(hc, CONF);
 
-    List<Put> puts = new ArrayList<Put>(1);
+    List<Put> puts = new ArrayList<>(1);
     puts.add(createPut(1, true));
     AsyncProcessTask task = AsyncProcessTask.newBuilder()
             .setPool(myPool)

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
index a4be9a2..d20c7c8 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestClientNoCluster.java
@@ -368,8 +368,7 @@ public class TestClientNoCluster extends Configured implements Tool {
     throws IOException {
       super(conf, pool, user);
       int serverCount = conf.getInt("hbase.test.servers", 10);
-      this.serversByClient =
-        new HashMap<ServerName, ClientService.BlockingInterface>(serverCount);
+      this.serversByClient = new HashMap<>(serverCount);
       this.meta = makeMeta(Bytes.toBytes(
         conf.get("hbase.test.tablename", Bytes.toString(BIG_USER_TABLE))),
         conf.getInt("hbase.test.regions", 100),
@@ -694,14 +693,13 @@ public class TestClientNoCluster extends Configured implements Tool {
       final int regionCount, final long namespaceSpan, final int serverCount) {
     // I need a comparator for meta rows so we sort properly.
     SortedMap<byte [], Pair<HRegionInfo, ServerName>> meta =
-      new ConcurrentSkipListMap<byte[], Pair<HRegionInfo,ServerName>>(new MetaRowsComparator());
+      new ConcurrentSkipListMap<>(new MetaRowsComparator());
     HRegionInfo [] hris = makeHRegionInfos(tableName, regionCount, namespaceSpan);
     ServerName [] serverNames = makeServerNames(serverCount);
     int per = regionCount / serverCount;
     int count = 0;
     for (HRegionInfo hri: hris) {
-      Pair<HRegionInfo, ServerName> p =
-        new Pair<HRegionInfo, ServerName>(hri, serverNames[count++ / per]);
+      Pair<HRegionInfo, ServerName> p = new Pair<>(hri, serverNames[count++ / per]);
       meta.put(hri.getRegionName(), p);
     }
     return meta;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/test/java/org/apache/hadoop/hbase/filter/TestKeyOnlyFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/filter/TestKeyOnlyFilter.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/filter/TestKeyOnlyFilter.java
index e93319a..f22e5d4 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/filter/TestKeyOnlyFilter.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/filter/TestKeyOnlyFilter.java
@@ -48,7 +48,7 @@ public class TestKeyOnlyFilter {
 
   @Parameters
   public static Collection<Object[]> parameters() {
-    List<Object[]> paramList = new ArrayList<Object[]>(2);
+    List<Object[]> paramList = new ArrayList<>(2);
     {
       paramList.add(new Object[] { false });
       paramList.add(new Object[] { true });

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/test/java/org/apache/hadoop/hbase/ipc/TestHBaseRpcControllerImpl.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/ipc/TestHBaseRpcControllerImpl.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/ipc/TestHBaseRpcControllerImpl.java
index 0659f30..0ec78ad 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/ipc/TestHBaseRpcControllerImpl.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/ipc/TestHBaseRpcControllerImpl.java
@@ -40,7 +40,7 @@ public class TestHBaseRpcControllerImpl {
   @Test
   public void testListOfCellScannerables() throws IOException {
     final int count = 10;
-    List<CellScannable> cells = new ArrayList<CellScannable>(count);
+    List<CellScannable> cells = new ArrayList<>(count);
 
     for (int i = 0; i < count; i++) {
       cells.add(createCell(i));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/test/java/org/apache/hadoop/hbase/util/BuilderStyleTest.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/util/BuilderStyleTest.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/util/BuilderStyleTest.java
index d2d0a53..771cf52 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/util/BuilderStyleTest.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/util/BuilderStyleTest.java
@@ -80,7 +80,7 @@ public class BuilderStyleTest {
           }
           Set<Method> sigMethods = methodsBySignature.get(sig);
           if (sigMethods == null) {
-            sigMethods = new HashSet<Method>();
+            sigMethods = new HashSet<>();
             methodsBySignature.put(sig, sigMethods);
           }
           sigMethods.add(method);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-client/src/test/java/org/apache/hadoop/hbase/zookeeper/TestInstancePending.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/test/java/org/apache/hadoop/hbase/zookeeper/TestInstancePending.java b/hbase-client/src/test/java/org/apache/hadoop/hbase/zookeeper/TestInstancePending.java
index 667fed8..e67c9fd 100644
--- a/hbase-client/src/test/java/org/apache/hadoop/hbase/zookeeper/TestInstancePending.java
+++ b/hbase-client/src/test/java/org/apache/hadoop/hbase/zookeeper/TestInstancePending.java
@@ -29,8 +29,8 @@ import org.junit.experimental.categories.Category;
 public class TestInstancePending {
   @Test(timeout = 1000)
   public void test() throws Exception {
-    final InstancePending<String> pending = new InstancePending<String>();
-    final AtomicReference<String> getResultRef = new AtomicReference<String>();
+    final InstancePending<String> pending = new InstancePending<>();
+    final AtomicReference<String> getResultRef = new AtomicReference<>();
 
     new Thread() {
       @Override

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
index 28c1d88..5930928 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
@@ -1597,7 +1597,7 @@ public final class CellUtil {
    * @return Tags in the given Cell as a List
    */
   public static List<Tag> getTags(Cell cell) {
-    List<Tag> tags = new ArrayList<Tag>();
+    List<Tag> tags = new ArrayList<>();
     Iterator<Tag> tagsItr = tagsIterator(cell);
     while (tagsItr.hasNext()) {
       tags.add(tagsItr.next());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
index 99dc163..d4ec48e 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
@@ -134,8 +134,8 @@ public class ChoreService implements ChoreServicer {
     }
 
     scheduler.setRemoveOnCancelPolicy(true);
-    scheduledChores = new HashMap<ScheduledChore, ScheduledFuture<?>>();
-    choresMissingStartTime = new HashMap<ScheduledChore, Boolean>();
+    scheduledChores = new HashMap<>();
+    choresMissingStartTime = new HashMap<>();
   }
 
   /**
@@ -348,7 +348,7 @@ public class ChoreService implements ChoreServicer {
   }
 
   private void cancelAllChores(final boolean mayInterruptIfRunning) {
-    ArrayList<ScheduledChore> choresToCancel = new ArrayList<ScheduledChore>(scheduledChores.keySet().size());
+    ArrayList<ScheduledChore> choresToCancel = new ArrayList<>(scheduledChores.keySet().size());
     // Build list of chores to cancel so we can iterate through a set that won't change
     // as chores are cancelled. If we tried to cancel each chore while iterating through
     // keySet the results would be undefined because the keySet would be changing
@@ -365,7 +365,7 @@ public class ChoreService implements ChoreServicer {
    * Prints a summary of important details about the chore. Used for debugging purposes
    */
   private void printChoreDetails(final String header, ScheduledChore chore) {
-    LinkedHashMap<String, String> output = new LinkedHashMap<String, String>();
+    LinkedHashMap<String, String> output = new LinkedHashMap<>();
     output.put(header, "");
     output.put("Chore name: ", chore.getName());
     output.put("Chore period: ", Integer.toString(chore.getPeriod()));
@@ -380,7 +380,7 @@ public class ChoreService implements ChoreServicer {
    * Prints a summary of important details about the service. Used for debugging purposes
    */
   private void printChoreServiceDetails(final String header) {
-    LinkedHashMap<String, String> output = new LinkedHashMap<String, String>();
+    LinkedHashMap<String, String> output = new LinkedHashMap<>();
     output.put(header, "");
     output.put("ChoreService corePoolSize: ", Integer.toString(getCorePoolSize()));
     output.put("ChoreService scheduledChores: ", Integer.toString(getNumberOfScheduledChores()));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/CompoundConfiguration.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CompoundConfiguration.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CompoundConfiguration.java
index 0eda1e5..a7fcba6 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CompoundConfiguration.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CompoundConfiguration.java
@@ -72,8 +72,7 @@ public class CompoundConfiguration extends Configuration {
     int size();
   }
 
-  private final List<ImmutableConfigMap> configs
-    = new ArrayList<ImmutableConfigMap>();
+  private final List<ImmutableConfigMap> configs = new ArrayList<>();
 
   static class ImmutableConfWrapper implements  ImmutableConfigMap {
    private final Configuration c;
@@ -167,7 +166,7 @@ public class CompoundConfiguration extends Configuration {
 
       @Override
       public Iterator<Map.Entry<String,String>> iterator() {
-        Map<String, String> ret = new HashMap<String, String>();
+        Map<String, String> ret = new HashMap<>();
         for (Map.Entry<Bytes, Bytes> entry : map.entrySet()) {
           String key = Bytes.toString(entry.getKey().get());
           String val = entry.getValue() == null ? null : Bytes.toString(entry.getValue().get());
@@ -366,7 +365,7 @@ public class CompoundConfiguration extends Configuration {
 
   @Override
   public Iterator<Map.Entry<String, String>> iterator() {
-    Map<String, String> ret = new HashMap<String, String>();
+    Map<String, String> ret = new HashMap<>();
 
     // add in reverse order so that oldest get overridden.
     if (!configs.isEmpty()) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index 0434820..96fc30b 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -81,7 +81,7 @@ import com.google.common.annotations.VisibleForTesting;
  */
 @InterfaceAudience.Private
 public class KeyValue implements ExtendedCell {
-  private static final ArrayList<Tag> EMPTY_ARRAY_LIST = new ArrayList<Tag>();
+  private static final ArrayList<Tag> EMPTY_ARRAY_LIST = new ArrayList<>();
 
   private static final Log LOG = LogFactory.getLog(KeyValue.class);
 
@@ -1174,7 +1174,7 @@ public class KeyValue implements ExtendedCell {
    * @return the Map&lt;String,?&gt; containing data from this key
    */
   public Map<String, Object> toStringMap() {
-    Map<String, Object> stringMap = new HashMap<String, Object>();
+    Map<String, Object> stringMap = new HashMap<>();
     stringMap.put("row", Bytes.toStringBinary(getRowArray(), getRowOffset(), getRowLength()));
     stringMap.put("family",
       Bytes.toStringBinary(getFamilyArray(), getFamilyOffset(), getFamilyLength()));
@@ -1184,7 +1184,7 @@ public class KeyValue implements ExtendedCell {
     stringMap.put("vlen", getValueLength());
     List<Tag> tags = getTags();
     if (tags != null) {
-      List<String> tagsString = new ArrayList<String>(tags.size());
+      List<String> tagsString = new ArrayList<>(tags.size());
       for (Tag t : tags) {
         tagsString.add(t.toString());
       }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
index ca990cf..807749a 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
@@ -496,7 +496,7 @@ public class KeyValueUtil {
         return KeyValueUtil.ensureKeyValue(arg0);
       }
     });
-    return new ArrayList<KeyValue>(lazyList);
+    return new ArrayList<>(lazyList);
   }
   /**
    * Write out a KeyValue in the manner in which we used to when KeyValue was a

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/NamespaceDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/NamespaceDescriptor.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/NamespaceDescriptor.java
index e1ceace..23876ab 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/NamespaceDescriptor.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/NamespaceDescriptor.java
@@ -57,14 +57,14 @@ public class NamespaceDescriptor {
 
   public final static Set<String> RESERVED_NAMESPACES;
   static {
-    Set<String> set = new HashSet<String>();
+    Set<String> set = new HashSet<>();
     set.add(NamespaceDescriptor.DEFAULT_NAMESPACE_NAME_STR);
     set.add(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR);
     RESERVED_NAMESPACES = Collections.unmodifiableSet(set);
   }
   public final static Set<byte[]> RESERVED_NAMESPACES_BYTES;
   static {
-    Set<byte[]> set = new TreeSet<byte[]>(Bytes.BYTES_RAWCOMPARATOR);
+    Set<byte[]> set = new TreeSet<>(Bytes.BYTES_RAWCOMPARATOR);
     for(String name: RESERVED_NAMESPACES) {
       set.add(Bytes.toBytes(name));
     }
@@ -165,7 +165,7 @@ public class NamespaceDescriptor {
   @InterfaceStability.Evolving
   public static class Builder {
     private String bName;
-    private Map<String, String> bConfiguration = new TreeMap<String, String>();
+    private Map<String, String> bConfiguration = new TreeMap<>();
 
     private Builder(NamespaceDescriptor ns) {
       this.bName = ns.name;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
index 499ffd9..0c0a7ff 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ServerName.java
@@ -99,7 +99,7 @@ public class ServerName implements Comparable<ServerName>, Serializable {
    * @see #getVersionedBytes()
    */
   private byte [] bytes;
-  public static final List<ServerName> EMPTY_SERVER_LIST = new ArrayList<ServerName>(0);
+  public static final List<ServerName> EMPTY_SERVER_LIST = new ArrayList<>(0);
 
   protected ServerName(final String hostname, final int port, final long startcode) {
     this(Address.fromParts(hostname, port), startcode);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
index 63066b3..9b9755b 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/TableName.java
@@ -57,7 +57,7 @@ import org.apache.hadoop.hbase.KeyValue.KVComparator;
 public final class TableName implements Comparable<TableName> {
 
   /** See {@link #createTableNameIfNecessary(ByteBuffer, ByteBuffer)} */
-  private static final Set<TableName> tableCache = new CopyOnWriteArraySet<TableName>();
+  private static final Set<TableName> tableCache = new CopyOnWriteArraySet<>();
 
   /** Namespace delimiter */
   //this should always be only 1 byte long

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
index f9668dd..936d8c2 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
@@ -72,7 +72,7 @@ public final class TagUtil {
    * @return List of tags
    */
   public static List<Tag> asList(byte[] b, int offset, int length) {
-    List<Tag> tags = new ArrayList<Tag>();
+    List<Tag> tags = new ArrayList<>();
     int pos = offset;
     while (pos < offset + length) {
       int tagLen = Bytes.readAsInt(b, pos, TAG_LENGTH_SIZE);
@@ -91,7 +91,7 @@ public final class TagUtil {
    * @return List of tags
    */
   public static List<Tag> asList(ByteBuffer b, int offset, int length) {
-    List<Tag> tags = new ArrayList<Tag>();
+    List<Tag> tags = new ArrayList<>();
     int pos = offset;
     while (pos < offset + length) {
       int tagLen = ByteBufferUtils.readAsInt(b, pos, TAG_LENGTH_SIZE);
@@ -239,7 +239,7 @@ public final class TagUtil {
     }
     List<Tag> tags = tagsOrNull;
     if (tags == null) {
-      tags = new ArrayList<Tag>();
+      tags = new ArrayList<>();
     }
     while (itr.hasNext()) {
       tags.add(itr.next());
@@ -276,7 +276,7 @@ public final class TagUtil {
     // in the array so set its size to '1' (I saw this being done in earlier version of
     // tag-handling).
     if (tags == null) {
-      tags = new ArrayList<Tag>(1);
+      tags = new ArrayList<>(1);
     }
     tags.add(new ArrayBackedTag(TagType.TTL_TAG_TYPE, Bytes.toBytes(ttl)));
     return tags;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/BoundedByteBufferPool.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/BoundedByteBufferPool.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/BoundedByteBufferPool.java
index 939d12d..079a277 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/BoundedByteBufferPool.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/BoundedByteBufferPool.java
@@ -51,7 +51,7 @@ import com.google.common.annotations.VisibleForTesting;
 public class BoundedByteBufferPool {
   private static final Log LOG = LogFactory.getLog(BoundedByteBufferPool.class);
 
-  private final Queue<ByteBuffer> buffers = new ConcurrentLinkedQueue<ByteBuffer>();
+  private final Queue<ByteBuffer> buffers = new ConcurrentLinkedQueue<>();
 
   @VisibleForTesting
   int getQueueSize() {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferListOutputStream.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferListOutputStream.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferListOutputStream.java
index c334a5a..9c6796e 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferListOutputStream.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferListOutputStream.java
@@ -44,8 +44,8 @@ public class ByteBufferListOutputStream extends ByteBufferOutputStream {
   // it is not available will make a new one our own and keep writing to that. We keep track of all
   // the BBs that we got from pool, separately so that on closeAndPutbackBuffers, we can make sure
   // to return back all of them to pool
-  protected List<ByteBuffer> allBufs = new ArrayList<ByteBuffer>();
-  protected List<ByteBuffer> bufsFromPool = new ArrayList<ByteBuffer>();
+  protected List<ByteBuffer> allBufs = new ArrayList<>();
+  protected List<ByteBuffer> bufsFromPool = new ArrayList<>();
 
   private boolean lastBufFlipped = false;// Indicate whether the curBuf/lastBuf is flipped already
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferPool.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferPool.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferPool.java
index 115671d..07ba3db 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferPool.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/ByteBufferPool.java
@@ -54,7 +54,7 @@ public class ByteBufferPool {
   public static final int DEFAULT_BUFFER_SIZE = 64 * 1024;// 64 KB. Making it same as the chunk size
                                                           // what we will write/read to/from the
                                                           // socket channel.
-  private final Queue<ByteBuffer> buffers = new ConcurrentLinkedQueue<ByteBuffer>();
+  private final Queue<ByteBuffer> buffers = new ConcurrentLinkedQueue<>();
 
   private final int bufferSize;
   private final int maxPoolSize;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/Encryption.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/Encryption.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/Encryption.java
index ad89ca0..b6c2e97 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/Encryption.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/Encryption.java
@@ -533,15 +533,14 @@ public final class Encryption {
     }
   }
 
-  static final Map<Pair<String,String>,KeyProvider> keyProviderCache =
-      new ConcurrentHashMap<Pair<String,String>,KeyProvider>();
+  static final Map<Pair<String,String>,KeyProvider> keyProviderCache = new ConcurrentHashMap<>();
 
   public static KeyProvider getKeyProvider(Configuration conf) {
     String providerClassName = conf.get(HConstants.CRYPTO_KEYPROVIDER_CONF_KEY,
       KeyStoreKeyProvider.class.getName());
     String providerParameters = conf.get(HConstants.CRYPTO_KEYPROVIDER_PARAMETERS_KEY, "");
     try {
-      Pair<String,String> providerCacheKey = new Pair<String,String>(providerClassName,
+      Pair<String,String> providerCacheKey = new Pair<>(providerClassName,
         providerParameters);
       KeyProvider provider = keyProviderCache.get(providerCacheKey);
       if (provider != null) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
index cef51d8..22d7e3e 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java
@@ -738,7 +738,7 @@ abstract class BufferedDataBlockEncoder extends AbstractDataBlockEncoder {
     protected  KeyValue.KeyOnlyKeyValue keyOnlyKV = new KeyValue.KeyOnlyKeyValue();
     // A temp pair object which will be reused by ByteBuff#asSubByteBuffer calls. This avoids too
     // many object creations.
-    protected final ObjectIntPair<ByteBuffer> tmpPair = new ObjectIntPair<ByteBuffer>();
+    protected final ObjectIntPair<ByteBuffer> tmpPair = new ObjectIntPair<>();
     protected STATE current, previous;
 
     public BufferedEncodedSeeker(CellComparator comparator,

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexCodecV1.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexCodecV1.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexCodecV1.java
index d81bb4a..0f8ea01 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexCodecV1.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexCodecV1.java
@@ -110,7 +110,7 @@ public class RowIndexCodecV1 extends AbstractDataBlockEncoder {
       RowIndexSeekerV1 seeker = new RowIndexSeekerV1(CellComparator.COMPARATOR,
           decodingCtx);
       seeker.setCurrentBuffer(new SingleByteBuff(sourceAsBuffer));
-      List<Cell> kvs = new ArrayList<Cell>();
+      List<Cell> kvs = new ArrayList<>();
       kvs.add(seeker.getCell());
       while (seeker.next()) {
         kvs.add(seeker.getCell());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexSeekerV1.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexSeekerV1.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexSeekerV1.java
index 4e14acb..6ac5645 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexSeekerV1.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/encoding/RowIndexSeekerV1.java
@@ -40,7 +40,7 @@ public class RowIndexSeekerV1 extends AbstractEncodedSeeker {
 
   // A temp pair object which will be reused by ByteBuff#asSubByteBuffer calls. This avoids too
   // many object creations.
-  protected final ObjectIntPair<ByteBuffer> tmpPair = new ObjectIntPair<ByteBuffer>();
+  protected final ObjectIntPair<ByteBuffer> tmpPair = new ObjectIntPair<>();
 
   private ByteBuff currentBuffer;
   private SeekerState current = new SeekerState(); // always valid

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/LRUDictionary.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/LRUDictionary.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/LRUDictionary.java
index 99780ba..2456961 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/LRUDictionary.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/LRUDictionary.java
@@ -86,7 +86,7 @@ public class LRUDictionary implements Dictionary {
     private Node head;
     private Node tail;
 
-    private HashMap<Node, Short> nodeToIndex = new HashMap<Node, Short>();
+    private HashMap<Node, Short> nodeToIndex = new HashMap<>();
     private Node[] indexToNode;
     private int initSize = 0;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/StreamUtils.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/StreamUtils.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/StreamUtils.java
index 0e1c3ae..9325284 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/StreamUtils.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/util/StreamUtils.java
@@ -135,7 +135,7 @@ public class StreamUtils {
     int newOffset = offset;
     byte tmp = input[newOffset++];
     if (tmp >= 0) {
-      return new Pair<Integer, Integer>((int) tmp, newOffset - offset);
+      return new Pair<>((int) tmp, newOffset - offset);
     }
     int result = tmp & 0x7f;
     tmp = input[newOffset++];
@@ -160,7 +160,7 @@ public class StreamUtils {
             for (int i = 0; i < 5; i++) {
               tmp = input[newOffset++];
               if (tmp >= 0) {
-                return new Pair<Integer, Integer>(result, newOffset - offset);
+                return new Pair<>(result, newOffset - offset);
               }
             }
             throw new IOException("Malformed varint");
@@ -168,7 +168,7 @@ public class StreamUtils {
         }
       }
     }
-    return new Pair<Integer, Integer>(result, newOffset - offset);
+    return new Pair<>(result, newOffset - offset);
   }
 
   public static Pair<Integer, Integer> readRawVarint32(ByteBuffer input, int offset)
@@ -176,7 +176,7 @@ public class StreamUtils {
     int newOffset = offset;
     byte tmp = input.get(newOffset++);
     if (tmp >= 0) {
-      return new Pair<Integer, Integer>((int) tmp, newOffset - offset);
+      return new Pair<>((int) tmp, newOffset - offset);
     }
     int result = tmp & 0x7f;
     tmp = input.get(newOffset++);
@@ -201,7 +201,7 @@ public class StreamUtils {
             for (int i = 0; i < 5; i++) {
               tmp = input.get(newOffset++);
               if (tmp >= 0) {
-                return new Pair<Integer, Integer>(result, newOffset - offset);
+                return new Pair<>(result, newOffset - offset);
               }
             }
             throw new IOException("Malformed varint");
@@ -209,7 +209,7 @@ public class StreamUtils {
         }
       }
     }
-    return new Pair<Integer, Integer>(result, newOffset - offset);
+    return new Pair<>(result, newOffset - offset);
   }
 
   public static short toShort(byte hi, byte lo) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java
index 0e45410..9f6b7b5 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/nio/SingleByteBuff.java
@@ -202,7 +202,7 @@ public class SingleByteBuff extends ByteBuff {
     } else {
       // TODO we can do some optimization here? Call to asSubByteBuffer might
       // create a copy.
-      ObjectIntPair<ByteBuffer> pair = new ObjectIntPair<ByteBuffer>();
+      ObjectIntPair<ByteBuffer> pair = new ObjectIntPair<>();
       src.asSubByteBuffer(srcOffset, length, pair);
       if (pair.getFirst() != null) {
         ByteBufferUtils.copyFromBufferToBuffer(pair.getFirst(), this.buf, pair.getSecond(), offset,

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
index c480dad..be2a0d3 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/User.java
@@ -354,8 +354,7 @@ public abstract class User {
   }
 
   static class TestingGroups extends Groups {
-    private final Map<String, List<String>> userToGroupsMapping =
-        new HashMap<String,List<String>>();
+    private final Map<String, List<String>> userToGroupsMapping = new HashMap<>();
     private Groups underlyingImplementation;
 
     TestingGroups(Groups underlyingImplementation) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
index 43b1c89..955abfc 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/security/UserProvider.java
@@ -90,7 +90,7 @@ public class UserProvider extends BaseConfigurable {
 
           private String[] getGroupStrings(String ugi) {
             try {
-              Set<String> result = new LinkedHashSet<String>(groups.getGroups(ugi));
+              Set<String> result = new LinkedHashSet<>(groups.getGroups(ugi));
               return result.toArray(new String[result.size()]);
             } catch (Exception e) {
               return new String[0];

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/trace/SpanReceiverHost.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/trace/SpanReceiverHost.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/trace/SpanReceiverHost.java
index f632ae0..1b6a67d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/trace/SpanReceiverHost.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/trace/SpanReceiverHost.java
@@ -63,7 +63,7 @@ public class SpanReceiverHost {
   }
 
   SpanReceiverHost(Configuration conf) {
-    receivers = new HashSet<SpanReceiver>();
+    receivers = new HashSet<>();
     this.conf = conf;
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/types/StructBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/types/StructBuilder.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/types/StructBuilder.java
index c9c3b64..d73a17d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/types/StructBuilder.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/types/StructBuilder.java
@@ -30,7 +30,7 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
 @InterfaceStability.Evolving
 public class StructBuilder {
 
-  protected final List<DataType<?>> fields = new ArrayList<DataType<?>>();
+  protected final List<DataType<?>> fields = new ArrayList<>();
 
   /**
    * Create an empty {@code StructBuilder}.

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ArrayUtils.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ArrayUtils.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ArrayUtils.java
index 4e3374e..51b87f7 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ArrayUtils.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ArrayUtils.java
@@ -102,7 +102,7 @@ public class ArrayUtils {
 
   public static ArrayList<Long> toList(long[] array){
     int length = length(array);
-    ArrayList<Long> list = new ArrayList<Long>(length);
+    ArrayList<Long> list = new ArrayList<>(length);
     for(int i=0; i < length; ++i){
       list.add(array[i]);
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AvlUtil.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AvlUtil.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AvlUtil.java
index 260a8b2..58c50a8 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AvlUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/AvlUtil.java
@@ -261,7 +261,7 @@ public final class AvlUtil {
         final AvlNodeVisitor<TNode> visitor) {
       if (root == null) return;
 
-      final AvlTreeIterator<TNode> iterator = new AvlTreeIterator<TNode>(root);
+      final AvlTreeIterator<TNode> iterator = new AvlTreeIterator<>(root);
       boolean visitNext = true;
       while (visitNext && iterator.hasNext()) {
         visitNext = visitor.visitNode(iterator.next());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/BoundedCompletionService.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/BoundedCompletionService.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/BoundedCompletionService.java
index c3fa547..ba38097 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/BoundedCompletionService.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/BoundedCompletionService.java
@@ -58,8 +58,8 @@ public class BoundedCompletionService<V> {
 
   public BoundedCompletionService(Executor executor, int maxTasks) {
     this.executor = executor;
-    this.tasks = new ArrayList<Future<V>>(maxTasks);
-    this.completed = new ArrayBlockingQueue<Future<V>>(maxTasks);
+    this.tasks = new ArrayList<>(maxTasks);
+    this.completed = new ArrayBlockingQueue<>(maxTasks);
   }
 
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRangeUtils.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRangeUtils.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRangeUtils.java
index 7de1b13..9248b41 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRangeUtils.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ByteRangeUtils.java
@@ -49,7 +49,7 @@ public class ByteRangeUtils {
 
   public static ArrayList<byte[]> copyToNewArrays(Collection<ByteRange> ranges) {
     if (ranges == null) {
-      return new ArrayList<byte[]>(0);
+      return new ArrayList<>(0);
     }
     ArrayList<byte[]> arrays = Lists.newArrayListWithCapacity(ranges.size());
     for (ByteRange range : ranges) {
@@ -60,7 +60,7 @@ public class ByteRangeUtils {
 
   public static ArrayList<ByteRange> fromArrays(Collection<byte[]> arrays) {
     if (arrays == null) {
-      return new ArrayList<ByteRange>(0);
+      return new ArrayList<>(0);
     }
     ArrayList<ByteRange> ranges = Lists.newArrayListWithCapacity(arrays.size());
     for (byte[] array : arrays) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
index 8cc71a3..1470d5c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
@@ -34,8 +34,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 @InterfaceAudience.Private
 public class CollectionUtils {
 
-  private static final List<Object> EMPTY_LIST = Collections.unmodifiableList(
-    new ArrayList<Object>(0));
+  private static final List<Object> EMPTY_LIST = Collections.unmodifiableList(new ArrayList<>(0));
 
   
   @SuppressWarnings("unchecked")

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ConcatenatedLists.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ConcatenatedLists.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ConcatenatedLists.java
index 0f00132..ba54f9d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ConcatenatedLists.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ConcatenatedLists.java
@@ -34,7 +34,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
  */
 @InterfaceAudience.Private
 public class ConcatenatedLists<T> implements Collection<T> {
-  protected final ArrayList<List<T>> components = new ArrayList<List<T>>();
+  protected final ArrayList<List<T>> components = new ArrayList<>();
   protected int size = 0;
 
   public void addAllSublists(List<? extends List<T>> items) {


Mime
View raw message