hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From st...@apache.org
Subject [11/22] hbase git commit: HBASE-17532 Replaced explicit type with diamond operator
Date Tue, 07 Mar 2017 19:23:24 GMT
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java
index 5cb7d75..a38e264 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java
@@ -147,14 +147,14 @@ public class ReplicationSourceManager implements ReplicationListener {
       final Path oldLogDir, final UUID clusterId) throws IOException {
     //CopyOnWriteArrayList is thread-safe.
     //Generally, reading is more than modifying.
-    this.sources = new CopyOnWriteArrayList<ReplicationSourceInterface>();
+    this.sources = new CopyOnWriteArrayList<>();
     this.replicationQueues = replicationQueues;
     this.replicationPeers = replicationPeers;
     this.replicationTracker = replicationTracker;
     this.server = server;
-    this.walsById = new HashMap<String, Map<String, SortedSet<String>>>();
-    this.walsByIdRecoveredQueues = new ConcurrentHashMap<String, Map<String, SortedSet<String>>>();
-    this.oldsources = new CopyOnWriteArrayList<ReplicationSourceInterface>();
+    this.walsById = new HashMap<>();
+    this.walsByIdRecoveredQueues = new ConcurrentHashMap<>();
+    this.oldsources = new CopyOnWriteArrayList<>();
     this.conf = conf;
     this.fs = fs;
     this.logDir = logDir;
@@ -170,8 +170,7 @@ public class ReplicationSourceManager implements ReplicationListener {
     // use a short 100ms sleep since this could be done inline with a RS startup
     // even if we fail, other region servers can take care of it
     this.executor = new ThreadPoolExecutor(nbWorkers, nbWorkers,
-        100, TimeUnit.MILLISECONDS,
-        new LinkedBlockingQueue<Runnable>());
+        100, TimeUnit.MILLISECONDS, new LinkedBlockingQueue<>());
     ThreadFactoryBuilder tfb = new ThreadFactoryBuilder();
     tfb.setNameFormat("ReplicationExecutor-%d");
     tfb.setDaemon(true);
@@ -277,7 +276,7 @@ public class ReplicationSourceManager implements ReplicationListener {
           this.replicationPeers, server, id, this.clusterId, peerConfig, peer);
     synchronized (this.walsById) {
       this.sources.add(src);
-      Map<String, SortedSet<String>> walsByGroup = new HashMap<String, SortedSet<String>>();
+      Map<String, SortedSet<String>> walsByGroup = new HashMap<>();
       this.walsById.put(id, walsByGroup);
       // Add the latest wal to that source's queue
       synchronized (latestPaths) {
@@ -285,7 +284,7 @@ public class ReplicationSourceManager implements ReplicationListener {
           for (Path logPath : latestPaths) {
             String name = logPath.getName();
             String walPrefix = AbstractFSWALProvider.getWALPrefixFromWALName(name);
-            SortedSet<String> logs = new TreeSet<String>();
+            SortedSet<String> logs = new TreeSet<>();
             logs.add(name);
             walsByGroup.put(walPrefix, logs);
             try {
@@ -423,7 +422,7 @@ public class ReplicationSourceManager implements ReplicationListener {
         if (!existingPrefix) {
           // The new log belongs to a new group, add it into this peer
           LOG.debug("Start tracking logs for wal group " + logPrefix + " for peer " + peerId);
-          SortedSet<String> wals = new TreeSet<String>();
+          SortedSet<String> wals = new TreeSet<>();
           wals.add(logName);
           walsByPrefix.put(logPrefix, wals);
         }
@@ -570,8 +569,7 @@ public class ReplicationSourceManager implements ReplicationListener {
         + sources.size() + " and another "
         + oldsources.size() + " that were recovered");
     String terminateMessage = "Replication stream was removed by a user";
-    List<ReplicationSourceInterface> oldSourcesToDelete =
-        new ArrayList<ReplicationSourceInterface>();
+    List<ReplicationSourceInterface> oldSourcesToDelete = new ArrayList<>();
     // synchronized on oldsources to avoid adding recovered source for the to-be-removed peer
     // see NodeFailoverWorker.run
     synchronized (oldsources) {
@@ -589,7 +587,7 @@ public class ReplicationSourceManager implements ReplicationListener {
     LOG.info("Number of deleted recovered sources for " + id + ": "
         + oldSourcesToDelete.size());
     // Now look for the one on this cluster
-    List<ReplicationSourceInterface> srcToRemove = new ArrayList<ReplicationSourceInterface>();
+    List<ReplicationSourceInterface> srcToRemove = new ArrayList<>();
     // synchronize on replicationPeers to avoid adding source for the to-be-removed peer
     synchronized (this.replicationPeers) {
       for (ReplicationSourceInterface src : this.sources) {
@@ -735,13 +733,13 @@ public class ReplicationSourceManager implements ReplicationListener {
             continue;
           }
           // track sources in walsByIdRecoveredQueues
-          Map<String, SortedSet<String>> walsByGroup = new HashMap<String, SortedSet<String>>();
+          Map<String, SortedSet<String>> walsByGroup = new HashMap<>();
           walsByIdRecoveredQueues.put(peerId, walsByGroup);
           for (String wal : walsSet) {
             String walPrefix = AbstractFSWALProvider.getWALPrefixFromWALName(wal);
             SortedSet<String> wals = walsByGroup.get(walPrefix);
             if (wals == null) {
-              wals = new TreeSet<String>();
+              wals = new TreeSet<>();
               walsByGroup.put(walPrefix, wals);
             }
             wals.add(wal);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java
index f06330c..c1aad93 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java
@@ -286,7 +286,7 @@ public class AccessControlLists {
             ACL_KEY_DELIMITER, columnName, ACL_KEY_DELIMITER,
             ACL_KEY_DELIMITER, columnName))));
 
-    Set<byte[]> qualifierSet = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);
+    Set<byte[]> qualifierSet = new TreeSet<>(Bytes.BYTES_COMPARATOR);
     ResultScanner scanner = null;
     try {
       scanner = table.getScanner(scan);
@@ -384,8 +384,7 @@ public class AccessControlLists {
       throw new IOException("Can only load permissions from "+ACL_TABLE_NAME);
     }
 
-    Map<byte[], ListMultimap<String, TablePermission>> allPerms =
-        new TreeMap<byte[], ListMultimap<String, TablePermission>>(Bytes.BYTES_RAWCOMPARATOR);
+    Map<byte[], ListMultimap<String, TablePermission>> allPerms = new TreeMap<>(Bytes.BYTES_RAWCOMPARATOR);
 
     // do a full scan of _acl_ table
 
@@ -397,7 +396,7 @@ public class AccessControlLists {
       iScanner = aclRegion.getScanner(scan);
 
       while (true) {
-        List<Cell> row = new ArrayList<Cell>();
+        List<Cell> row = new ArrayList<>();
 
         boolean hasNext = iScanner.next(row);
         ListMultimap<String,TablePermission> perms = ArrayListMultimap.create();
@@ -436,8 +435,7 @@ public class AccessControlLists {
    */
   static Map<byte[], ListMultimap<String,TablePermission>> loadAll(
       Configuration conf) throws IOException {
-    Map<byte[], ListMultimap<String,TablePermission>> allPerms =
-        new TreeMap<byte[], ListMultimap<String,TablePermission>>(Bytes.BYTES_RAWCOMPARATOR);
+    Map<byte[], ListMultimap<String,TablePermission>> allPerms = new TreeMap<>(Bytes.BYTES_RAWCOMPARATOR);
 
     // do a full scan of _acl_, filtering on only first table region rows
 
@@ -530,7 +528,7 @@ public class AccessControlLists {
     ListMultimap<String,TablePermission> allPerms = getPermissions(
         conf, entryName, null);
 
-    List<UserPermission> perms = new ArrayList<UserPermission>();
+    List<UserPermission> perms = new ArrayList<>();
 
     if(isNamespaceEntry(entryName)) {  // Namespace
       for (Map.Entry<String, TablePermission> entry : allPerms.entries()) {
@@ -591,8 +589,7 @@ public class AccessControlLists {
 
     //Handle namespace entry
     if(isNamespaceEntry(entryName)) {
-      return new Pair<String, TablePermission>(username,
-          new TablePermission(Bytes.toString(fromNamespaceEntry(entryName)), value));
+      return new Pair<>(username, new TablePermission(Bytes.toString(fromNamespaceEntry(entryName)), value));
     }
 
     //Handle table and global entry
@@ -612,8 +609,7 @@ public class AccessControlLists {
       }
     }
 
-    return new Pair<String,TablePermission>(username,
-        new TablePermission(TableName.valueOf(entryName), permFamily, permQualifier, value));
+    return new Pair<>(username, new TablePermission(TableName.valueOf(entryName), permFamily, permQualifier, value));
   }
 
   /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
index 814f209..64ac900 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
@@ -259,8 +259,7 @@ public class AccessController implements MasterObserver, RegionObserver, RegionS
    */
   void updateACL(RegionCoprocessorEnvironment e,
       final Map<byte[], List<Cell>> familyMap) {
-    Set<byte[]> entries =
-        new TreeSet<byte[]>(Bytes.BYTES_RAWCOMPARATOR);
+    Set<byte[]> entries = new TreeSet<>(Bytes.BYTES_RAWCOMPARATOR);
     for (Map.Entry<byte[], List<Cell>> f : familyMap.entrySet()) {
       List<Cell> cells = f.getValue();
       for (Cell cell: cells) {
@@ -793,7 +792,7 @@ public class AccessController implements MasterObserver, RegionObserver, RegionS
     // This Map is identical to familyMap. The key is a BR rather than byte[].
     // It will be easy to do gets over this new Map as we can create get keys over the Cell cf by
     // new SimpleByteRange(cell.familyArray, cell.familyOffset, cell.familyLen)
-    Map<ByteRange, List<Cell>> familyMap1 = new HashMap<ByteRange, List<Cell>>();
+    Map<ByteRange, List<Cell>> familyMap1 = new HashMap<>();
     for (Entry<byte[], ? extends Collection<?>> entry : familyMap.entrySet()) {
       if (entry.getValue() instanceof List) {
         familyMap1.put(new SimpleMutableByteRange(entry.getKey()), (List<Cell>) entry.getValue());
@@ -882,7 +881,7 @@ public class AccessController implements MasterObserver, RegionObserver, RegionS
       List<Cell> newCells = Lists.newArrayList();
       for (Cell cell: e.getValue()) {
         // Prepend the supplied perms in a new ACL tag to an update list of tags for the cell
-        List<Tag> tags = new ArrayList<Tag>();
+        List<Tag> tags = new ArrayList<>();
         tags.add(new ArrayBackedTag(AccessControlLists.ACL_TAG_TYPE, perms));
         Iterator<Tag> tagIterator = CellUtil.tagsIterator(cell);
         while (tagIterator.hasNext()) {
@@ -990,7 +989,7 @@ public class AccessController implements MasterObserver, RegionObserver, RegionS
   public void preCreateTable(ObserverContext<MasterCoprocessorEnvironment> c,
       HTableDescriptor desc, HRegionInfo[] regions) throws IOException {
     Set<byte[]> families = desc.getFamiliesKeys();
-    Map<byte[], Set<byte[]>> familyMap = new TreeMap<byte[], Set<byte[]>>(Bytes.BYTES_COMPARATOR);
+    Map<byte[], Set<byte[]>> familyMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
     for (byte[] family: families) {
       familyMap.put(family, null);
     }
@@ -2407,8 +2406,7 @@ public class AccessController implements MasterObserver, RegionObserver, RegionS
                   tperm.getTableName()));
             }
 
-            Map<byte[], Set<byte[]>> familyMap =
-                new TreeMap<byte[], Set<byte[]>>(Bytes.BYTES_COMPARATOR);
+            Map<byte[], Set<byte[]>> familyMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
             if (tperm.getFamily() != null) {
               if (tperm.getQualifier() != null) {
                 Set<byte[]> qualifiers = Sets.newTreeSet(Bytes.BYTES_COMPARATOR);
@@ -2515,7 +2513,7 @@ public class AccessController implements MasterObserver, RegionObserver, RegionS
       return null;
     }
 
-    Map<byte[], Collection<byte[]>> familyMap = new TreeMap<byte[], Collection<byte[]>>(Bytes.BYTES_COMPARATOR);
+    Map<byte[], Collection<byte[]>> familyMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
     familyMap.put(family, qualifier != null ? ImmutableSet.of(qualifier) : null);
     return familyMap;
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/TableAuthManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/TableAuthManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/TableAuthManager.java
index eae9e4e..0d539ce 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/TableAuthManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/TableAuthManager.java
@@ -103,10 +103,10 @@ public class TableAuthManager implements Closeable {
   private volatile PermissionCache<Permission> globalCache;
 
   private ConcurrentSkipListMap<TableName, PermissionCache<TablePermission>> tableCache =
-      new ConcurrentSkipListMap<TableName, PermissionCache<TablePermission>>();
+      new ConcurrentSkipListMap<>();
 
   private ConcurrentSkipListMap<String, PermissionCache<TablePermission>> nsCache =
-    new ConcurrentSkipListMap<String, PermissionCache<TablePermission>>();
+    new ConcurrentSkipListMap<>();
 
   private Configuration conf;
   private ZKPermissionWatcher zkperms;
@@ -143,7 +143,7 @@ public class TableAuthManager implements Closeable {
       throw new IOException("Unable to obtain the current user, " +
           "authorization checks for internal operations will not work correctly!");
     }
-    PermissionCache<Permission> newCache = new PermissionCache<Permission>();
+    PermissionCache<Permission> newCache = new PermissionCache<>();
     String currentUser = user.getShortName();
 
     // the system user is always included
@@ -239,7 +239,7 @@ public class TableAuthManager implements Closeable {
    */
   private void updateTableCache(TableName table,
                                 ListMultimap<String,TablePermission> tablePerms) {
-    PermissionCache<TablePermission> newTablePerms = new PermissionCache<TablePermission>();
+    PermissionCache<TablePermission> newTablePerms = new PermissionCache<>();
 
     for (Map.Entry<String,TablePermission> entry : tablePerms.entries()) {
       if (AuthUtil.isGroupPrincipal(entry.getKey())) {
@@ -263,7 +263,7 @@ public class TableAuthManager implements Closeable {
    */
   private void updateNsCache(String namespace,
                              ListMultimap<String, TablePermission> tablePerms) {
-    PermissionCache<TablePermission> newTablePerms = new PermissionCache<TablePermission>();
+    PermissionCache<TablePermission> newTablePerms = new PermissionCache<>();
 
     for (Map.Entry<String, TablePermission> entry : tablePerms.entries()) {
       if (AuthUtil.isGroupPrincipal(entry.getKey())) {
@@ -734,8 +734,7 @@ public class TableAuthManager implements Closeable {
     return mtime.get();
   }
 
-  private static Map<ZooKeeperWatcher,TableAuthManager> managerMap =
-    new HashMap<ZooKeeperWatcher,TableAuthManager>();
+  private static Map<ZooKeeperWatcher,TableAuthManager> managerMap = new HashMap<>();
 
   private static Map<TableAuthManager, Integer> refCount = new HashMap<>();
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
index f21e877..3324b90 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
@@ -58,8 +58,7 @@ public class ZKPermissionWatcher extends ZooKeeperListener implements Closeable
   TableAuthManager authManager;
   String aclZNode;
   CountDownLatch initialized = new CountDownLatch(1);
-  AtomicReference<List<ZKUtil.NodeAndData>> nodes =
-      new AtomicReference<List<ZKUtil.NodeAndData>>(null);
+  AtomicReference<List<ZKUtil.NodeAndData>> nodes = new AtomicReference<>(null);
   ExecutorService executor;
 
   public ZKPermissionWatcher(ZooKeeperWatcher watcher,

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenSecretManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenSecretManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenSecretManager.java
index 26448b1..a569cf3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenSecretManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenSecretManager.java
@@ -70,8 +70,7 @@ public class AuthenticationTokenSecretManager
   private LeaderElector leaderElector;
   private ZKClusterId clusterId;
 
-  private Map<Integer,AuthenticationKey> allKeys =
-      new ConcurrentHashMap<Integer, AuthenticationKey>();
+  private Map<Integer,AuthenticationKey> allKeys = new ConcurrentHashMap<>();
   private AuthenticationKey currentKey;
 
   private int idSeq;
@@ -181,8 +180,7 @@ public class AuthenticationTokenSecretManager
   public Token<AuthenticationTokenIdentifier> generateToken(String username) {
     AuthenticationTokenIdentifier ident =
         new AuthenticationTokenIdentifier(username);
-    Token<AuthenticationTokenIdentifier> token =
-        new Token<AuthenticationTokenIdentifier>(ident, this);
+    Token<AuthenticationTokenIdentifier> token = new Token<>(ident, this);
     if (clusterId.hasId()) {
       token.setService(new Text(clusterId.getId()));
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
index f767ed3..1d42450 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
@@ -314,7 +314,7 @@ public class TokenUtil {
    * @return the Token instance
    */
   public static Token<AuthenticationTokenIdentifier> toToken(AuthenticationProtos.Token proto) {
-    return new Token<AuthenticationTokenIdentifier>(
+    return new Token<>(
         proto.hasIdentifier() ? proto.getIdentifier().toByteArray() : null,
         proto.hasPassword() ? proto.getPassword().toByteArray() : null,
         AuthenticationTokenIdentifier.AUTH_TOKEN_TYPE,

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefaultVisibilityLabelServiceImpl.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefaultVisibilityLabelServiceImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefaultVisibilityLabelServiceImpl.java
index 9abb3a2..d4a5627 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefaultVisibilityLabelServiceImpl.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefaultVisibilityLabelServiceImpl.java
@@ -151,10 +151,10 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
   protected List<List<Cell>> getExistingLabelsWithAuths() throws IOException {
     Scan scan = new Scan();
     RegionScanner scanner = labelsRegion.getScanner(scan);
-    List<List<Cell>> existingLabels = new ArrayList<List<Cell>>();
+    List<List<Cell>> existingLabels = new ArrayList<>();
     try {
       while (true) {
-        List<Cell> cells = new ArrayList<Cell>();
+        List<Cell> cells = new ArrayList<>();
         scanner.next(cells);
         if (cells.isEmpty()) {
           break;
@@ -169,8 +169,8 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
 
   protected Pair<Map<String, Integer>, Map<String, List<Integer>>> extractLabelsAndAuths(
       List<List<Cell>> labelDetails) {
-    Map<String, Integer> labels = new HashMap<String, Integer>();
-    Map<String, List<Integer>> userAuths = new HashMap<String, List<Integer>>();
+    Map<String, Integer> labels = new HashMap<>();
+    Map<String, List<Integer>> userAuths = new HashMap<>();
     for (List<Cell> cells : labelDetails) {
       for (Cell cell : cells) {
         if (CellUtil.matchingQualifier(cell, LABEL_QUALIFIER)) {
@@ -183,14 +183,14 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
               cell.getQualifierLength());
           List<Integer> auths = userAuths.get(user);
           if (auths == null) {
-            auths = new ArrayList<Integer>();
+            auths = new ArrayList<>();
             userAuths.put(user, auths);
           }
           auths.add(CellUtil.getRowAsInt(cell));
         }
       }
     }
-    return new Pair<Map<String, Integer>, Map<String, List<Integer>>>(labels, userAuths);
+    return new Pair<>(labels, userAuths);
   }
 
   protected void addSystemLabel(Region region, Map<String, Integer> labels,
@@ -207,7 +207,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
   public OperationStatus[] addLabels(List<byte[]> labels) throws IOException {
     assert labelsRegion != null;
     OperationStatus[] finalOpStatus = new OperationStatus[labels.size()];
-    List<Mutation> puts = new ArrayList<Mutation>(labels.size());
+    List<Mutation> puts = new ArrayList<>(labels.size());
     int i = 0;
     for (byte[] label : labels) {
       String labelStr = Bytes.toString(label);
@@ -235,7 +235,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
   public OperationStatus[] setAuths(byte[] user, List<byte[]> authLabels) throws IOException {
     assert labelsRegion != null;
     OperationStatus[] finalOpStatus = new OperationStatus[authLabels.size()];
-    List<Mutation> puts = new ArrayList<Mutation>(authLabels.size());
+    List<Mutation> puts = new ArrayList<>(authLabels.size());
     int i = 0;
     for (byte[] auth : authLabels) {
       String authStr = Bytes.toString(auth);
@@ -269,7 +269,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
     else {
       currentAuths = this.getUserAuths(user, true);
     }
-    List<Mutation> deletes = new ArrayList<Mutation>(authLabels.size());
+    List<Mutation> deletes = new ArrayList<>(authLabels.size());
     int i = 0;
     for (byte[] authLabel : authLabels) {
       String authLabelStr = Bytes.toString(authLabel);
@@ -334,10 +334,10 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
     Filter filter = VisibilityUtils.createVisibilityLabelFilter(this.labelsRegion,
         new Authorizations(SYSTEM_LABEL));
     s.setFilter(filter);
-    ArrayList<String> auths = new ArrayList<String>();
+    ArrayList<String> auths = new ArrayList<>();
     RegionScanner scanner = this.labelsRegion.getScanner(s);
     try {
-      List<Cell> results = new ArrayList<Cell>(1);
+      List<Cell> results = new ArrayList<>(1);
       while (true) {
         scanner.next(results);
         if (results.isEmpty()) break;
@@ -371,10 +371,10 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
     Filter filter = VisibilityUtils.createVisibilityLabelFilter(this.labelsRegion,
         new Authorizations(SYSTEM_LABEL));
     s.setFilter(filter);
-    Set<String> auths = new HashSet<String>();
+    Set<String> auths = new HashSet<>();
     RegionScanner scanner = this.labelsRegion.getScanner(s);
     try {
-      List<Cell> results = new ArrayList<Cell>(1);
+      List<Cell> results = new ArrayList<>(1);
       while (true) {
         scanner.next(results);
         if (results.isEmpty()) break;
@@ -389,7 +389,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
     } finally {
       scanner.close();
     }
-    return new ArrayList<String>(auths);
+    return new ArrayList<>(auths);
   }
 
   @Override
@@ -401,7 +401,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
     labels.remove(SYSTEM_LABEL);
     if (regex != null) {
       Pattern pattern = Pattern.compile(regex);
-      ArrayList<String> matchedLabels = new ArrayList<String>();
+      ArrayList<String> matchedLabels = new ArrayList<>();
       for (String label : labels.keySet()) {
         if (pattern.matcher(label).matches()) {
           matchedLabels.add(label);
@@ -409,13 +409,13 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
       }
       return matchedLabels;
     }
-    return new ArrayList<String>(labels.keySet());
+    return new ArrayList<>(labels.keySet());
   }
 
   @Override
   public List<Tag> createVisibilityExpTags(String visExpression, boolean withSerializationFormat,
       boolean checkAuths) throws IOException {
-    Set<Integer> auths = new HashSet<Integer>();
+    Set<Integer> auths = new HashSet<>();
     if (checkAuths) {
       User user = VisibilityUtils.getActiveUser();
       auths.addAll(this.labelsCache.getUserAuthsAsOrdinals(user.getShortName()));
@@ -461,7 +461,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
       try {
         // null authorizations to be handled inside SLG impl.
         authLabels = scanLabelGenerator.getLabels(VisibilityUtils.getActiveUser(), authorizations);
-        authLabels = (authLabels == null) ? new ArrayList<String>() : authLabels;
+        authLabels = (authLabels == null) ? new ArrayList<>() : authLabels;
         authorizations = new Authorizations(authLabels);
       } catch (Throwable t) {
         LOG.error(t);
@@ -605,7 +605,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
   }
 
   private static List<List<Integer>> sortTagsBasedOnOrdinal(List<Tag> tags) throws IOException {
-    List<List<Integer>> fullTagsList = new ArrayList<List<Integer>>();
+    List<List<Integer>> fullTagsList = new ArrayList<>();
     for (Tag tag : tags) {
       if (tag.getType() == VISIBILITY_TAG_TYPE) {
         getSortedTagOrdinals(fullTagsList, tag);
@@ -616,7 +616,7 @@ public class DefaultVisibilityLabelServiceImpl implements VisibilityLabelService
 
   private static void getSortedTagOrdinals(List<List<Integer>> fullTagsList, Tag tag)
       throws IOException {
-    List<Integer> tagsOrdinalInSortedOrder = new ArrayList<Integer>();
+    List<Integer> tagsOrdinalInSortedOrder = new ArrayList<>();
     int offset = tag.getValueOffset();
     int endOffset = offset + tag.getValueLength();
     while (offset < endOffset) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefinedSetFilterScanLabelGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefinedSetFilterScanLabelGenerator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefinedSetFilterScanLabelGenerator.java
index 2c7d253..2126ee7 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefinedSetFilterScanLabelGenerator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefinedSetFilterScanLabelGenerator.java
@@ -62,18 +62,18 @@ public class DefinedSetFilterScanLabelGenerator implements ScanLabelGenerator {
     if (authorizations != null) {
       List<String> labels = authorizations.getLabels();
       String userName = user.getShortName();
-      Set<String> auths = new HashSet<String>();
+      Set<String> auths = new HashSet<>();
       auths.addAll(this.labelsCache.getUserAuths(userName));
       auths.addAll(this.labelsCache.getGroupAuths(user.getGroupNames()));
-      return dropLabelsNotInUserAuths(labels, new ArrayList<String>(auths), userName);
+      return dropLabelsNotInUserAuths(labels, new ArrayList<>(auths), userName);
     }
     return null;
   }
 
   private List<String> dropLabelsNotInUserAuths(List<String> labels, List<String> auths,
       String userName) {
-    List<String> droppedLabels = new ArrayList<String>();
-    List<String> passedLabels = new ArrayList<String>(labels.size());
+    List<String> droppedLabels = new ArrayList<>();
+    List<String> passedLabels = new ArrayList<>(labels.size());
     for (String label : labels) {
       if (auths.contains(label)) {
         passedLabels.add(label);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/EnforcingScanLabelGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/EnforcingScanLabelGenerator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/EnforcingScanLabelGenerator.java
index dd0497c..177f4d2 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/EnforcingScanLabelGenerator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/EnforcingScanLabelGenerator.java
@@ -62,10 +62,10 @@ public class EnforcingScanLabelGenerator implements ScanLabelGenerator {
     if (authorizations != null) {
       LOG.warn("Dropping authorizations requested by user " + userName + ": " + authorizations);
     }
-    Set<String> auths = new HashSet<String>();
+    Set<String> auths = new HashSet<>();
     auths.addAll(this.labelsCache.getUserAuths(userName));
     auths.addAll(this.labelsCache.getGroupAuths(user.getGroupNames()));
-    return new ArrayList<String>(auths);
+    return new ArrayList<>(auths);
   }
 
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ExpressionParser.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ExpressionParser.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ExpressionParser.java
index db3caff..2b9a56e 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ExpressionParser.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ExpressionParser.java
@@ -40,7 +40,7 @@ public class ExpressionParser {
   private static final char DOUBLE_QUOTES = '"';
   public ExpressionNode parse(String expS) throws ParseException {
     expS = expS.trim();
-    Stack<ExpressionNode> expStack = new Stack<ExpressionNode>();
+    Stack<ExpressionNode> expStack = new Stack<>();
     int index = 0;
     byte[] exp = Bytes.toBytes(expS);
     int endPos = exp.length;
@@ -68,7 +68,7 @@ public class ExpressionParser {
           // We have to rewrite the expression within double quotes as incase of expressions 
           // with escape characters we may have to avoid them as the original expression did
           // not have them
-          List<Byte> list = new ArrayList<Byte>();
+          List<Byte> list = new ArrayList<>();
           while (index < endPos && !endDoubleQuotesFound(exp[index])) {
             if (exp[index] == '\\') {
               index++;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/FeedUserAuthScanLabelGenerator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/FeedUserAuthScanLabelGenerator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/FeedUserAuthScanLabelGenerator.java
index 1f90682..f4cf762 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/FeedUserAuthScanLabelGenerator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/FeedUserAuthScanLabelGenerator.java
@@ -65,10 +65,10 @@ public class FeedUserAuthScanLabelGenerator implements ScanLabelGenerator {
     if (authorizations == null || authorizations.getLabels() == null
         || authorizations.getLabels().isEmpty()) {
       String userName = user.getShortName();
-      Set<String> auths = new HashSet<String>();
+      Set<String> auths = new HashSet<>();
       auths.addAll(this.labelsCache.getUserAuths(userName));
       auths.addAll(this.labelsCache.getGroupAuths(user.getGroupNames()));
-      return new ArrayList<String>(auths);
+      return new ArrayList<>(auths);
     }
     return authorizations.getLabels();
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
index fb685bc..476921b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
@@ -145,7 +145,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
   boolean authorizationEnabled;
 
   // Add to this list if there are any reserved tag types
-  private static ArrayList<Byte> RESERVED_VIS_TAG_TYPES = new ArrayList<Byte>();
+  private static ArrayList<Byte> RESERVED_VIS_TAG_TYPES = new ArrayList<>();
   static {
     RESERVED_VIS_TAG_TYPES.add(TagType.VISIBILITY_TAG_TYPE);
     RESERVED_VIS_TAG_TYPES.add(TagType.VISIBILITY_EXP_SERIALIZATION_FORMAT_TAG_TYPE);
@@ -328,7 +328,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
       return;
     }
     // TODO this can be made as a global LRU cache at HRS level?
-    Map<String, List<Tag>> labelCache = new HashMap<String, List<Tag>>();
+    Map<String, List<Tag>> labelCache = new HashMap<>();
     for (int i = 0; i < miniBatchOp.size(); i++) {
       Mutation m = miniBatchOp.getOperation(i);
       CellVisibility cellVisibility = null;
@@ -341,7 +341,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
       }
       boolean sanityFailure = false;
       boolean modifiedTagFound = false;
-      Pair<Boolean, Tag> pair = new Pair<Boolean, Tag>(false, null);
+      Pair<Boolean, Tag> pair = new Pair<>(false, null);
       for (CellScanner cellScanner = m.cellScanner(); cellScanner.advance();) {
         pair = checkForReservedVisibilityTagPresence(cellScanner.current(), pair);
         if (!pair.getFirst()) {
@@ -381,7 +381,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
             }
           }
           if (visibilityTags != null) {
-            List<Cell> updatedCells = new ArrayList<Cell>();
+            List<Cell> updatedCells = new ArrayList<>();
             for (CellScanner cellScanner = m.cellScanner(); cellScanner.advance();) {
               Cell cell = cellScanner.current();
               List<Tag> tags = CellUtil.getTags(cell);
@@ -427,7 +427,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
     }
     // The check for checkForReservedVisibilityTagPresence happens in preBatchMutate happens.
     // It happens for every mutation and that would be enough.
-    List<Tag> visibilityTags = new ArrayList<Tag>();
+    List<Tag> visibilityTags = new ArrayList<>();
     if (cellVisibility != null) {
       String labelsExp = cellVisibility.getExpression();
       try {
@@ -474,7 +474,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
   private Pair<Boolean, Tag> checkForReservedVisibilityTagPresence(Cell cell,
       Pair<Boolean, Tag> pair) throws IOException {
     if (pair == null) {
-      pair = new Pair<Boolean, Tag>(false, null);
+      pair = new Pair<>(false, null);
     } else {
       pair.setFirst(false);
       pair.setSecond(null);
@@ -782,7 +782,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
         new VisibilityControllerNotReadyException("VisibilityController not yet initialized!"),
         response);
     } else {
-      List<byte[]> labels = new ArrayList<byte[]>(visLabels.size());
+      List<byte[]> labels = new ArrayList<>(visLabels.size());
       try {
         if (authorizationEnabled) {
           checkCallingUserAuth();
@@ -844,7 +844,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
         response);
     } else {
       byte[] user = request.getUser().toByteArray();
-      List<byte[]> labelAuths = new ArrayList<byte[]>(auths.size());
+      List<byte[]> labelAuths = new ArrayList<>(auths.size());
       try {
         if (authorizationEnabled) {
           checkCallingUserAuth();
@@ -959,7 +959,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
           "VisibilityController not yet initialized"), response);
     } else {
       byte[] requestUser = request.getUser().toByteArray();
-      List<byte[]> labelAuths = new ArrayList<byte[]>(auths.size());
+      List<byte[]> labelAuths = new ArrayList<>(auths.size());
       try {
         // When AC is ON, do AC based user auth check
         if (authorizationEnabled && accessControllerAvailable && !isSystemOrSuperUser()) {
@@ -1071,7 +1071,7 @@ public class VisibilityController implements MasterObserver, RegionObserver,
 
     @Override
     public ReturnCode filterKeyValue(Cell cell) throws IOException {
-      List<Tag> putVisTags = new ArrayList<Tag>();
+      List<Tag> putVisTags = new ArrayList<>();
       Byte putCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(cell, putVisTags);
       boolean matchFound = VisibilityLabelServiceManager
           .getInstance().getVisibilityLabelService()

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityLabelsCache.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityLabelsCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityLabelsCache.java
index 0948520..e27a4f8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityLabelsCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityLabelsCache.java
@@ -54,10 +54,10 @@ public class VisibilityLabelsCache implements VisibilityLabelOrdinalProvider {
   private static VisibilityLabelsCache instance;
 
   private ZKVisibilityLabelWatcher zkVisibilityWatcher;
-  private Map<String, Integer> labels = new HashMap<String, Integer>();
-  private Map<Integer, String> ordinalVsLabels = new HashMap<Integer, String>();
-  private Map<String, Set<Integer>> userAuths = new HashMap<String, Set<Integer>>();
-  private Map<String, Set<Integer>> groupAuths = new HashMap<String, Set<Integer>>();
+  private Map<String, Integer> labels = new HashMap<>();
+  private Map<Integer, String> ordinalVsLabels = new HashMap<>();
+  private Map<String, Set<Integer>> userAuths = new HashMap<>();
+  private Map<String, Set<Integer>> groupAuths = new HashMap<>();
 
   /**
    * This covers the members labels, ordinalVsLabels and userAuths
@@ -145,10 +145,9 @@ public class VisibilityLabelsCache implements VisibilityLabelOrdinalProvider {
       for (UserAuthorizations userAuths : multiUserAuths.getUserAuthsList()) {
         String user = Bytes.toString(userAuths.getUser().toByteArray());
         if (AuthUtil.isGroupPrincipal(user)) {
-          this.groupAuths.put(AuthUtil.getGroupName(user),
-            new HashSet<Integer>(userAuths.getAuthList()));
+          this.groupAuths.put(AuthUtil.getGroupName(user), new HashSet<>(userAuths.getAuthList()));
         } else {
-          this.userAuths.put(user, new HashSet<Integer>(userAuths.getAuthList()));
+          this.userAuths.put(user, new HashSet<>(userAuths.getAuthList()));
         }
       }
     } finally {
@@ -210,7 +209,7 @@ public class VisibilityLabelsCache implements VisibilityLabelOrdinalProvider {
       List<String> auths = EMPTY_LIST;
       Set<Integer> authOrdinals = getUserAuthsAsOrdinals(user);
       if (!authOrdinals.equals(EMPTY_SET)) {
-        auths = new ArrayList<String>(authOrdinals.size());
+        auths = new ArrayList<>(authOrdinals.size());
         for (Integer authOrdinal : authOrdinals) {
           auths.add(ordinalVsLabels.get(authOrdinal));
         }
@@ -227,7 +226,7 @@ public class VisibilityLabelsCache implements VisibilityLabelOrdinalProvider {
       List<String> auths = EMPTY_LIST;
       Set<Integer> authOrdinals = getGroupAuthsAsOrdinals(groups);
       if (!authOrdinals.equals(EMPTY_SET)) {
-        auths = new ArrayList<String>(authOrdinals.size());
+        auths = new ArrayList<>(authOrdinals.size());
         for (Integer authOrdinal : authOrdinals) {
           auths.add(ordinalVsLabels.get(authOrdinal));
         }
@@ -263,7 +262,7 @@ public class VisibilityLabelsCache implements VisibilityLabelOrdinalProvider {
   public Set<Integer> getGroupAuthsAsOrdinals(String[] groups) {
     this.lock.readLock().lock();
     try {
-      Set<Integer> authOrdinals = new HashSet<Integer>();
+      Set<Integer> authOrdinals = new HashSet<>();
       if (groups != null && groups.length > 0) {
         Set<Integer> groupAuthOrdinals = null;
         for (String group : groups) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.java
index c1c3852..c77b776 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.java
@@ -69,9 +69,9 @@ public class VisibilityReplicationEndpoint implements ReplicationEndpoint {
       // string based tags. But for intra cluster replication like region
       // replicas it is not needed.
       List<Entry> entries = replicateContext.getEntries();
-      List<Tag> visTags = new ArrayList<Tag>();
-      List<Tag> nonVisTags = new ArrayList<Tag>();
-      List<Entry> newEntries = new ArrayList<Entry>(entries.size());
+      List<Tag> visTags = new ArrayList<>();
+      List<Tag> nonVisTags = new ArrayList<>();
+      List<Entry> newEntries = new ArrayList<>(entries.size());
       for (Entry entry : entries) {
         WALEdit newEdit = new WALEdit();
         ArrayList<Cell> cells = entry.getEdit().getCells();

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.java
index 2595fe0..67181e1 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.java
@@ -53,17 +53,14 @@ public class VisibilityScanDeleteTracker extends ScanDeleteTracker {
   // type would solve this problem and also ensure that the combination of different type
   // of deletes with diff ts would also work fine
   // Track per TS
-  private List<Triple<List<Tag>, Byte, Long>> visibilityTagsDeleteFamily =
-      new ArrayList<Triple<List<Tag>, Byte, Long>>();
+  private List<Triple<List<Tag>, Byte, Long>> visibilityTagsDeleteFamily = new ArrayList<>();
   // Delete family version with different ts and different visibility expression could come.
   // Need to track it per ts.
-  private List<Triple<List<Tag>, Byte, Long>> visibilityTagsDeleteFamilyVersion =
-      new ArrayList<Triple<List<Tag>, Byte, Long>>();
+  private List<Triple<List<Tag>, Byte, Long>> visibilityTagsDeleteFamilyVersion = new ArrayList<>();
   private List<Pair<List<Tag>, Byte>> visibilityTagsDeleteColumns;
   // Tracking as List<List> is to handle same ts cell but different visibility tag. 
   // TODO : Need to handle puts with same ts but different vis tags.
-  private List<Pair<List<Tag>, Byte>> visiblityTagsDeleteColumnVersion =
-      new ArrayList<Pair<List<Tag>, Byte>>();
+  private List<Pair<List<Tag>, Byte>> visiblityTagsDeleteColumnVersion = new ArrayList<>();
 
   public VisibilityScanDeleteTracker() {
     super();
@@ -117,50 +114,46 @@ public class VisibilityScanDeleteTracker extends ScanDeleteTracker {
       Byte deleteCellVisTagsFormat = null;
       switch (type) {
       case DeleteFamily:
-        List<Tag> delTags = new ArrayList<Tag>();
+        List<Tag> delTags = new ArrayList<>();
         if (visibilityTagsDeleteFamily == null) {
-          visibilityTagsDeleteFamily = new ArrayList<Triple<List<Tag>, Byte, Long>>();
+          visibilityTagsDeleteFamily = new ArrayList<>();
         }
         deleteCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(delCell, delTags);
         if (!delTags.isEmpty()) {
-          visibilityTagsDeleteFamily.add(new Triple<List<Tag>, Byte, Long>(delTags,
-              deleteCellVisTagsFormat, delCell.getTimestamp()));
+          visibilityTagsDeleteFamily.add(new Triple<>(delTags, deleteCellVisTagsFormat, delCell.getTimestamp()));
           hasVisTag = true;
         }
         break;
       case DeleteFamilyVersion:
         if(visibilityTagsDeleteFamilyVersion == null) {
-          visibilityTagsDeleteFamilyVersion = new ArrayList<Triple<List<Tag>, Byte, Long>>();
+          visibilityTagsDeleteFamilyVersion = new ArrayList<>();
         }
-        delTags = new ArrayList<Tag>();
+        delTags = new ArrayList<>();
         deleteCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(delCell, delTags);
         if (!delTags.isEmpty()) {
-          visibilityTagsDeleteFamilyVersion.add(new Triple<List<Tag>, Byte, Long>(delTags,
-              deleteCellVisTagsFormat, delCell.getTimestamp()));
+          visibilityTagsDeleteFamilyVersion.add(new Triple<>(delTags, deleteCellVisTagsFormat, delCell.getTimestamp()));
           hasVisTag = true;
         }
         break;
       case DeleteColumn:
         if (visibilityTagsDeleteColumns == null) {
-          visibilityTagsDeleteColumns = new ArrayList<Pair<List<Tag>, Byte>>();
+          visibilityTagsDeleteColumns = new ArrayList<>();
         }
-        delTags = new ArrayList<Tag>();
+        delTags = new ArrayList<>();
         deleteCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(delCell, delTags);
         if (!delTags.isEmpty()) {
-          visibilityTagsDeleteColumns.add(new Pair<List<Tag>, Byte>(delTags,
-              deleteCellVisTagsFormat));
+          visibilityTagsDeleteColumns.add(new Pair<>(delTags, deleteCellVisTagsFormat));
           hasVisTag = true;
         }
         break;
       case Delete:
         if (visiblityTagsDeleteColumnVersion == null) {
-          visiblityTagsDeleteColumnVersion = new ArrayList<Pair<List<Tag>, Byte>>();
+          visiblityTagsDeleteColumnVersion = new ArrayList<>();
         }
-        delTags = new ArrayList<Tag>();
+        delTags = new ArrayList<>();
         deleteCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(delCell, delTags);
         if (!delTags.isEmpty()) {
-          visiblityTagsDeleteColumnVersion.add(new Pair<List<Tag>, Byte>(delTags,
-              deleteCellVisTagsFormat));
+          visiblityTagsDeleteColumnVersion.add(new Pair<>(delTags, deleteCellVisTagsFormat));
           hasVisTag = true;
         }
         break;
@@ -182,7 +175,7 @@ public class VisibilityScanDeleteTracker extends ScanDeleteTracker {
               // visibilityTagsDeleteFamily is ArrayList
               Triple<List<Tag>, Byte, Long> triple = visibilityTagsDeleteFamily.get(i);
               if (timestamp <= triple.getThird()) {
-                List<Tag> putVisTags = new ArrayList<Tag>();
+                List<Tag> putVisTags = new ArrayList<>();
                 Byte putCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(cell, putVisTags);
                 boolean matchFound = VisibilityLabelServiceManager.getInstance()
                     .getVisibilityLabelService().matchVisibility(putVisTags, putCellVisTagsFormat,
@@ -218,7 +211,7 @@ public class VisibilityScanDeleteTracker extends ScanDeleteTracker {
               // visibilityTagsDeleteFamilyVersion is ArrayList
               Triple<List<Tag>, Byte, Long> triple = visibilityTagsDeleteFamilyVersion.get(i);
               if (timestamp == triple.getThird()) {
-                List<Tag> putVisTags = new ArrayList<Tag>();
+                List<Tag> putVisTags = new ArrayList<>();
                 Byte putCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(cell, putVisTags);
                 boolean matchFound = VisibilityLabelServiceManager.getInstance()
                     .getVisibilityLabelService().matchVisibility(putVisTags, putCellVisTagsFormat,
@@ -248,7 +241,7 @@ public class VisibilityScanDeleteTracker extends ScanDeleteTracker {
             if (visibilityTagsDeleteColumns != null) {
               if (!visibilityTagsDeleteColumns.isEmpty()) {
                 for (Pair<List<Tag>, Byte> tags : visibilityTagsDeleteColumns) {
-                  List<Tag> putVisTags = new ArrayList<Tag>();
+                  List<Tag> putVisTags = new ArrayList<>();
                   Byte putCellVisTagsFormat =
                       VisibilityUtils.extractVisibilityTags(cell, putVisTags);
                   boolean matchFound = VisibilityLabelServiceManager.getInstance()
@@ -277,7 +270,7 @@ public class VisibilityScanDeleteTracker extends ScanDeleteTracker {
             if (visiblityTagsDeleteColumnVersion != null) {
               if (!visiblityTagsDeleteColumnVersion.isEmpty()) {
                 for (Pair<List<Tag>, Byte> tags : visiblityTagsDeleteColumnVersion) {
-                  List<Tag> putVisTags = new ArrayList<Tag>();
+                  List<Tag> putVisTags = new ArrayList<>();
                   Byte putCellVisTagsFormat =
                       VisibilityUtils.extractVisibilityTags(cell, putVisTags);
                   boolean matchFound = VisibilityLabelServiceManager.getInstance()

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java
index 1db506d..4441c08 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java
@@ -175,7 +175,7 @@ public class VisibilityUtils {
     String slgClassesCommaSeparated = conf.get(VISIBILITY_LABEL_GENERATOR_CLASS);
     // We have only System level SLGs now. The order of execution will be same as the order in the
     // comma separated config value
-    List<ScanLabelGenerator> slgs = new ArrayList<ScanLabelGenerator>();
+    List<ScanLabelGenerator> slgs = new ArrayList<>();
     if (StringUtils.isNotEmpty(slgClassesCommaSeparated)) {
       String[] slgClasses = slgClassesCommaSeparated.split(COMMA);
       for (String slgClass : slgClasses) {
@@ -266,7 +266,7 @@ public class VisibilityUtils {
 
   public static Filter createVisibilityLabelFilter(Region region, Authorizations authorizations)
       throws IOException {
-    Map<ByteRange, Integer> cfVsMaxVersions = new HashMap<ByteRange, Integer>();
+    Map<ByteRange, Integer> cfVsMaxVersions = new HashMap<>();
     for (HColumnDescriptor hcd : region.getTableDesc().getFamilies()) {
       cfVsMaxVersions.put(new SimpleMutableByteRange(hcd.getName()), hcd.getMaxVersions());
     }
@@ -302,10 +302,10 @@ public class VisibilityUtils {
       throw new IOException(e);
     }
     node = EXP_EXPANDER.expand(node);
-    List<Tag> tags = new ArrayList<Tag>();
+    List<Tag> tags = new ArrayList<>();
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
-    List<Integer> labelOrdinals = new ArrayList<Integer>();
+    List<Integer> labelOrdinals = new ArrayList<>();
     // We will be adding this tag before the visibility tags and the presence of this
     // tag indicates we are supporting deletes with cell visibility
     if (withSerializationFormat) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/expression/NonLeafExpressionNode.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/expression/NonLeafExpressionNode.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/expression/NonLeafExpressionNode.java
index 4399ecc..9903b9b 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/expression/NonLeafExpressionNode.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/expression/NonLeafExpressionNode.java
@@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 @InterfaceAudience.Private
 public class NonLeafExpressionNode implements ExpressionNode {
   private Operator op;
-  private List<ExpressionNode> childExps = new ArrayList<ExpressionNode>(2);
+  private List<ExpressionNode> childExps = new ArrayList<>(2);
 
   public NonLeafExpressionNode() {
 
@@ -46,7 +46,7 @@ public class NonLeafExpressionNode implements ExpressionNode {
 
   public NonLeafExpressionNode(Operator op, ExpressionNode... exps) {
     this.op = op;
-    List<ExpressionNode> expLst = new ArrayList<ExpressionNode>();
+    List<ExpressionNode> expLst = new ArrayList<>();
     Collections.addAll(expLst, exps);
     this.childExps = expLst;
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
index a5507fc..efae7e4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
@@ -564,7 +564,7 @@ public class ExportSnapshot extends AbstractHBaseTool implements Tool {
       final FileSystem fs, final Path snapshotDir) throws IOException {
     SnapshotDescription snapshotDesc = SnapshotDescriptionUtils.readSnapshotInfo(fs, snapshotDir);
 
-    final List<Pair<SnapshotFileInfo, Long>> files = new ArrayList<Pair<SnapshotFileInfo, Long>>();
+    final List<Pair<SnapshotFileInfo, Long>> files = new ArrayList<>();
     final TableName table = TableName.valueOf(snapshotDesc.getTable());
 
     // Get snapshot files
@@ -591,7 +591,7 @@ public class ExportSnapshot extends AbstractHBaseTool implements Tool {
             } else {
               size = HFileLink.buildFromHFileLinkPattern(conf, path).getFileStatus(fs).getLen();
             }
-            files.add(new Pair<SnapshotFileInfo, Long>(fileInfo, size));
+            files.add(new Pair<>(fileInfo, size));
           }
         }
     });
@@ -618,8 +618,7 @@ public class ExportSnapshot extends AbstractHBaseTool implements Tool {
     });
 
     // create balanced groups
-    List<List<Pair<SnapshotFileInfo, Long>>> fileGroups =
-      new LinkedList<List<Pair<SnapshotFileInfo, Long>>>();
+    List<List<Pair<SnapshotFileInfo, Long>>> fileGroups = new LinkedList<>();
     long[] sizeGroups = new long[ngroups];
     int hi = files.size() - 1;
     int lo = 0;
@@ -630,7 +629,7 @@ public class ExportSnapshot extends AbstractHBaseTool implements Tool {
 
     while (hi >= lo) {
       if (g == fileGroups.size()) {
-        group = new LinkedList<Pair<SnapshotFileInfo, Long>>();
+        group = new LinkedList<>();
         fileGroups.add(group);
       } else {
         group = fileGroups.get(g);
@@ -703,7 +702,7 @@ public class ExportSnapshot extends AbstractHBaseTool implements Tool {
       public ExportSnapshotInputSplit(final List<Pair<SnapshotFileInfo, Long>> snapshotFiles) {
         this.files = new ArrayList(snapshotFiles.size());
         for (Pair<SnapshotFileInfo, Long> fileInfo: snapshotFiles) {
-          this.files.add(new Pair<BytesWritable, Long>(
+          this.files.add(new Pair<>(
             new BytesWritable(fileInfo.getFirst().toByteArray()), fileInfo.getSecond()));
           this.length += fileInfo.getSecond();
         }
@@ -726,13 +725,13 @@ public class ExportSnapshot extends AbstractHBaseTool implements Tool {
       @Override
       public void readFields(DataInput in) throws IOException {
         int count = in.readInt();
-        files = new ArrayList<Pair<BytesWritable, Long>>(count);
+        files = new ArrayList<>(count);
         length = 0;
         for (int i = 0; i < count; ++i) {
           BytesWritable fileInfo = new BytesWritable();
           fileInfo.readFields(in);
           long size = in.readLong();
-          files.add(new Pair<BytesWritable, Long>(fileInfo, size));
+          files.add(new Pair<>(fileInfo, size));
           length += size;
         }
       }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
index 98afe8b..63839c4 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
@@ -108,11 +108,9 @@ import org.apache.hadoop.io.IOUtils;
 public class RestoreSnapshotHelper {
   private static final Log LOG = LogFactory.getLog(RestoreSnapshotHelper.class);
 
-  private final Map<byte[], byte[]> regionsMap =
-        new TreeMap<byte[], byte[]>(Bytes.BYTES_COMPARATOR);
+  private final Map<byte[], byte[]> regionsMap = new TreeMap<>(Bytes.BYTES_COMPARATOR);
 
-  private final Map<String, Pair<String, String> > parentsMap =
-      new HashMap<String, Pair<String, String> >();
+  private final Map<String, Pair<String, String> > parentsMap = new HashMap<>();
 
   private final ForeignExceptionDispatcher monitor;
   private final MonitoredTask status;
@@ -187,7 +185,7 @@ public class RestoreSnapshotHelper {
 
     // Take a copy of the manifest.keySet() since we are going to modify
     // this instance, by removing the regions already present in the restore dir.
-    Set<String> regionNames = new HashSet<String>(regionManifests.keySet());
+    Set<String> regionNames = new HashSet<>(regionManifests.keySet());
 
     HRegionInfo mobRegion = MobUtils.getMobRegionInfo(snapshotManifest.getTableDescriptor()
         .getTableName());
@@ -213,7 +211,7 @@ public class RestoreSnapshotHelper {
       status.setStatus("Restoring table regions...");
       if (regionNames.contains(mobRegion.getEncodedName())) {
         // restore the mob region in case
-        List<HRegionInfo> mobRegions = new ArrayList<HRegionInfo>(1);
+        List<HRegionInfo> mobRegions = new ArrayList<>(1);
         mobRegions.add(mobRegion);
         restoreHdfsMobRegions(exec, regionManifests, mobRegions);
         regionNames.remove(mobRegion.getEncodedName());
@@ -230,7 +228,7 @@ public class RestoreSnapshotHelper {
 
     // Regions to Add: present in the snapshot but not in the current table
     if (regionNames.size() > 0) {
-      List<HRegionInfo> regionsToAdd = new ArrayList<HRegionInfo>(regionNames.size());
+      List<HRegionInfo> regionsToAdd = new ArrayList<>(regionNames.size());
 
       monitor.rethrowException();
       // add the mob region
@@ -344,14 +342,14 @@ public class RestoreSnapshotHelper {
 
     void addRegionToRemove(final HRegionInfo hri) {
       if (regionsToRemove == null) {
-        regionsToRemove = new LinkedList<HRegionInfo>();
+        regionsToRemove = new LinkedList<>();
       }
       regionsToRemove.add(hri);
     }
 
     void addRegionToRestore(final HRegionInfo hri) {
       if (regionsToRestore == null) {
-        regionsToRestore = new LinkedList<HRegionInfo>();
+        regionsToRestore = new LinkedList<>();
       }
       regionsToRestore.add(hri);
     }
@@ -361,7 +359,7 @@ public class RestoreSnapshotHelper {
       if (regionInfos == null || parentsMap.isEmpty()) return;
 
       // Extract region names and offlined regions
-      Map<String, HRegionInfo> regionsByName = new HashMap<String, HRegionInfo>(regionInfos.size());
+      Map<String, HRegionInfo> regionsByName = new HashMap<>(regionInfos.size());
       List<HRegionInfo> parentRegions = new LinkedList<>();
       for (HRegionInfo regionInfo: regionInfos) {
         if (regionInfo.isSplitParent()) {
@@ -441,10 +439,10 @@ public class RestoreSnapshotHelper {
   private Map<String, List<SnapshotRegionManifest.StoreFile>> getRegionHFileReferences(
       final SnapshotRegionManifest manifest) {
     Map<String, List<SnapshotRegionManifest.StoreFile>> familyMap =
-      new HashMap<String, List<SnapshotRegionManifest.StoreFile>>(manifest.getFamilyFilesCount());
+      new HashMap<>(manifest.getFamilyFilesCount());
     for (SnapshotRegionManifest.FamilyFiles familyFiles: manifest.getFamilyFilesList()) {
       familyMap.put(familyFiles.getFamilyName().toStringUtf8(),
-        new ArrayList<SnapshotRegionManifest.StoreFile>(familyFiles.getStoreFilesList()));
+        new ArrayList<>(familyFiles.getStoreFilesList()));
     }
     return familyMap;
   }
@@ -489,8 +487,7 @@ public class RestoreSnapshotHelper {
       List<SnapshotRegionManifest.StoreFile> snapshotFamilyFiles =
           snapshotFiles.remove(familyDir.getName());
       if (snapshotFamilyFiles != null) {
-        List<SnapshotRegionManifest.StoreFile> hfilesToAdd =
-            new ArrayList<SnapshotRegionManifest.StoreFile>();
+        List<SnapshotRegionManifest.StoreFile> hfilesToAdd = new ArrayList<>();
         for (SnapshotRegionManifest.StoreFile storeFile: snapshotFamilyFiles) {
           if (familyFiles.contains(storeFile.getName())) {
             // HFile already present
@@ -546,7 +543,7 @@ public class RestoreSnapshotHelper {
     FileStatus[] hfiles = FSUtils.listStatus(fs, familyDir);
     if (hfiles == null) return Collections.emptySet();
 
-    Set<String> familyFiles = new HashSet<String>(hfiles.length);
+    Set<String> familyFiles = new HashSet<>(hfiles.length);
     for (int i = 0; i < hfiles.length; ++i) {
       String hfileName = hfiles[i].getPath().getName();
       familyFiles.add(hfileName);
@@ -564,8 +561,7 @@ public class RestoreSnapshotHelper {
       final List<HRegionInfo> regions) throws IOException {
     if (regions == null || regions.isEmpty()) return null;
 
-    final Map<String, HRegionInfo> snapshotRegions =
-      new HashMap<String, HRegionInfo>(regions.size());
+    final Map<String, HRegionInfo> snapshotRegions = new HashMap<>(regions.size());
 
     // clone region info (change embedded tableName with the new one)
     HRegionInfo[] clonedRegionsInfo = new HRegionInfo[regions.size()];
@@ -742,7 +738,7 @@ public class RestoreSnapshotHelper {
     synchronized (parentsMap) {
       Pair<String, String> daughters = parentsMap.get(clonedRegionName);
       if (daughters == null) {
-        daughters = new Pair<String, String>(regionName, null);
+        daughters = new Pair<>(regionName, null);
         parentsMap.put(clonedRegionName, daughters);
       } else if (!regionName.equals(daughters.getFirst())) {
         daughters.setSecond(regionName);
@@ -778,7 +774,7 @@ public class RestoreSnapshotHelper {
     FileStatus[] regionDirs = FSUtils.listStatus(fs, tableDir, new FSUtils.RegionDirFilter(fs));
     if (regionDirs == null) return null;
 
-    List<HRegionInfo> regions = new ArrayList<HRegionInfo>(regionDirs.length);
+    List<HRegionInfo> regions = new ArrayList<>(regionDirs.length);
     for (int i = 0; i < regionDirs.length; ++i) {
       HRegionInfo hri = HRegionFileSystem.loadRegionInfoFileContent(fs, regionDirs[i].getPath());
       regions.add(hri);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
index 85d3af3..6dbd3f0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
@@ -598,8 +598,7 @@ public final class SnapshotInfo extends AbstractHBaseTool {
     Path snapshotDir = SnapshotDescriptionUtils.getSnapshotsDir(rootDir);
     FileStatus[] snapshots = fs.listStatus(snapshotDir,
         new SnapshotDescriptionUtils.CompletedSnaphotDirectoriesFilter(fs));
-    List<SnapshotDescription> snapshotLists =
-      new ArrayList<SnapshotDescription>(snapshots.length);
+    List<SnapshotDescription> snapshotLists = new ArrayList<>(snapshots.length);
     for (FileStatus snapshotDirStat: snapshots) {
       HBaseProtos.SnapshotDescription snapshotDesc =
           SnapshotDescriptionUtils.readSnapshotInfo(fs, snapshotDirStat.getPath());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
index 47e3073..4e838ad 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
@@ -220,7 +220,7 @@ public final class SnapshotManifest {
       Object familyData = visitor.familyOpen(regionData, store.getFamily().getName());
       monitor.rethrowException();
 
-      List<StoreFile> storeFiles = new ArrayList<StoreFile>(store.getStorefiles());
+      List<StoreFile> storeFiles = new ArrayList<>(store.getStorefiles());
       if (LOG.isDebugEnabled()) {
         LOG.debug("Adding snapshot references for " + storeFiles  + " hfiles");
       }
@@ -305,7 +305,7 @@ public final class SnapshotManifest {
     FileStatus[] stats = FSUtils.listStatus(fs, storeDir);
     if (stats == null) return null;
 
-    ArrayList<StoreFileInfo> storeFiles = new ArrayList<StoreFileInfo>(stats.length);
+    ArrayList<StoreFileInfo> storeFiles = new ArrayList<>(stats.length);
     for (int i = 0; i < stats.length; ++i) {
       storeFiles.add(new StoreFileInfo(conf, fs, stats[i]));
     }
@@ -374,8 +374,7 @@ public final class SnapshotManifest {
             tpool.shutdown();
           }
           if (v1Regions != null && v2Regions != null) {
-            regionManifests =
-              new ArrayList<SnapshotRegionManifest>(v1Regions.size() + v2Regions.size());
+            regionManifests = new ArrayList<>(v1Regions.size() + v2Regions.size());
             regionManifests.addAll(v1Regions);
             regionManifests.addAll(v2Regions);
           } else if (v1Regions != null) {
@@ -427,8 +426,7 @@ public final class SnapshotManifest {
   public Map<String, SnapshotRegionManifest> getRegionManifestsMap() {
     if (regionManifests == null || regionManifests.isEmpty()) return null;
 
-    HashMap<String, SnapshotRegionManifest> regionsMap =
-        new HashMap<String, SnapshotRegionManifest>(regionManifests.size());
+    HashMap<String, SnapshotRegionManifest> regionsMap = new HashMap<>(regionManifests.size());
     for (SnapshotRegionManifest manifest: regionManifests) {
       String regionName = getRegionNameFromManifest(manifest);
       regionsMap.put(regionName, manifest);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
index cceeebc..46893f9 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
@@ -121,7 +121,7 @@ public final class SnapshotManifestV1 {
     }
 
     final ExecutorCompletionService<SnapshotRegionManifest> completionService =
-      new ExecutorCompletionService<SnapshotRegionManifest>(executor);
+      new ExecutorCompletionService<>(executor);
     for (final FileStatus region: regions) {
       completionService.submit(new Callable<SnapshotRegionManifest>() {
         @Override
@@ -132,8 +132,7 @@ public final class SnapshotManifestV1 {
       });
     }
 
-    ArrayList<SnapshotRegionManifest> regionsManifest =
-        new ArrayList<SnapshotRegionManifest>(regions.length);
+    ArrayList<SnapshotRegionManifest> regionsManifest = new ArrayList<>(regions.length);
     try {
       for (int i = 0; i < regions.length; ++i) {
         regionsManifest.add(completionService.take().get());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
index a1341fb..567f42d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
@@ -139,7 +139,7 @@ public final class SnapshotManifestV2 {
     if (manifestFiles == null || manifestFiles.length == 0) return null;
 
     final ExecutorCompletionService<SnapshotRegionManifest> completionService =
-      new ExecutorCompletionService<SnapshotRegionManifest>(executor);
+      new ExecutorCompletionService<>(executor);
     for (final FileStatus st: manifestFiles) {
       completionService.submit(new Callable<SnapshotRegionManifest>() {
         @Override
@@ -157,8 +157,7 @@ public final class SnapshotManifestV2 {
       });
     }
 
-    ArrayList<SnapshotRegionManifest> regionsManifest =
-        new ArrayList<SnapshotRegionManifest>(manifestFiles.length);
+    ArrayList<SnapshotRegionManifest> regionsManifest = new ArrayList<>(manifestFiles.length);
     try {
       for (int i = 0; i < manifestFiles.length; ++i) {
         regionsManifest.add(completionService.take().get());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
index 8cd438e..7a2bfe6 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
@@ -210,8 +210,7 @@ public final class SnapshotReferenceUtil {
       return;
     }
 
-    final ExecutorCompletionService<Void> completionService =
-      new ExecutorCompletionService<Void>(exec);
+    final ExecutorCompletionService<Void> completionService = new ExecutorCompletionService<>(exec);
 
     for (final SnapshotRegionManifest regionManifest : regionManifests) {
       completionService.submit(new Callable<Void>() {
@@ -345,7 +344,7 @@ public final class SnapshotReferenceUtil {
   private static Set<String> getHFileNames(final Configuration conf, final FileSystem fs,
       final Path snapshotDir, final SnapshotDescription snapshotDesc)
       throws IOException {
-    final Set<String> names = new HashSet<String>();
+    final Set<String> names = new HashSet<>();
     visitTableStoreFiles(conf, fs, snapshotDir, snapshotDesc, new StoreFileVisitor() {
       @Override
       public void storeFile(final HRegionInfo regionInfo, final String family,

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
index 73160bc..ee93cdb 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
@@ -140,8 +140,8 @@ public final class Canary implements Tool {
     private AtomicLong readFailureCount = new AtomicLong(0),
         writeFailureCount = new AtomicLong(0);
 
-    private Map<String, String> readFailures = new ConcurrentHashMap<String, String>();
-    private Map<String, String> writeFailures = new ConcurrentHashMap<String, String>();
+    private Map<String, String> readFailures = new ConcurrentHashMap<>();
+    private Map<String, String> writeFailures = new ConcurrentHashMap<>();
 
     @Override
     public long getReadFailureCount() {
@@ -949,7 +949,7 @@ public final class Canary implements Tool {
     public void run() {
       if (this.initAdmin()) {
         try {
-          List<Future<Void>> taskFutures = new LinkedList<Future<Void>>();
+          List<Future<Void>> taskFutures = new LinkedList<>();
           if (this.targets != null && this.targets.length > 0) {
             String[] tables = generateMonitorTables(this.targets);
             this.initialized = true;
@@ -996,7 +996,7 @@ public final class Canary implements Tool {
       if (this.useRegExp) {
         Pattern pattern = null;
         HTableDescriptor[] tds = null;
-        Set<String> tmpTables = new TreeSet<String>();
+        Set<String> tmpTables = new TreeSet<>();
         try {
           if (LOG.isDebugEnabled()) {
             LOG.debug(String.format("reading list of tables"));
@@ -1040,7 +1040,7 @@ public final class Canary implements Tool {
       if (LOG.isDebugEnabled()) {
         LOG.debug(String.format("reading list of tables"));
       }
-      List<Future<Void>> taskFutures = new LinkedList<Future<Void>>();
+      List<Future<Void>> taskFutures = new LinkedList<>();
       for (HTableDescriptor table : admin.listTables()) {
         if (admin.isTableEnabled(table.getTableName())
             && (!table.getTableName().equals(writeTableName))) {
@@ -1078,7 +1078,7 @@ public final class Canary implements Tool {
         admin.deleteTable(writeTableName);
         createWriteTable(numberOfServers);
       }
-      HashSet<ServerName> serverSet = new HashSet<ServerName>();
+      HashSet<ServerName> serverSet = new HashSet<>();
       for (Pair<HRegionInfo, ServerName> pair : pairs) {
         serverSet.add(pair.getSecond());
       }
@@ -1165,7 +1165,7 @@ public final class Canary implements Tool {
     } else {
       LOG.warn(String.format("Table %s is not enabled", tableName));
     }
-    return new LinkedList<Future<Void>>();
+    return new LinkedList<>();
   }
 
   /*
@@ -1183,7 +1183,7 @@ public final class Canary implements Tool {
     try {
       table = admin.getConnection().getTable(tableDesc.getTableName());
     } catch (TableNotFoundException e) {
-      return new ArrayList<Future<Void>>();
+      return new ArrayList<>();
     }
     finally {
       if (table !=null) {
@@ -1191,7 +1191,7 @@ public final class Canary implements Tool {
       }
     }
 
-    List<RegionTask> tasks = new ArrayList<RegionTask>();
+    List<RegionTask> tasks = new ArrayList<>();
     RegionLocator regionLocator = null;
     try {
       regionLocator = admin.getConnection().getRegionLocator(tableDesc.getTableName());
@@ -1290,7 +1290,7 @@ public final class Canary implements Tool {
     }
 
     private boolean checkNoTableNames() {
-      List<String> foundTableNames = new ArrayList<String>();
+      List<String> foundTableNames = new ArrayList<>();
       TableName[] tableNames = null;
 
       if (LOG.isDebugEnabled()) {
@@ -1323,8 +1323,8 @@ public final class Canary implements Tool {
     }
 
     private void monitorRegionServers(Map<String, List<HRegionInfo>> rsAndRMap) {
-      List<RegionServerTask> tasks = new ArrayList<RegionServerTask>();
-      Map<String, AtomicLong> successMap = new HashMap<String, AtomicLong>();
+      List<RegionServerTask> tasks = new ArrayList<>();
+      Map<String, AtomicLong> successMap = new HashMap<>();
       Random rand = new Random();
       for (Map.Entry<String, List<HRegionInfo>> entry : rsAndRMap.entrySet()) {
         String serverName = entry.getKey();
@@ -1379,7 +1379,7 @@ public final class Canary implements Tool {
     }
 
     private Map<String, List<HRegionInfo>> getAllRegionServerByName() {
-      Map<String, List<HRegionInfo>> rsAndRMap = new HashMap<String, List<HRegionInfo>>();
+      Map<String, List<HRegionInfo>> rsAndRMap = new HashMap<>();
       Table table = null;
       RegionLocator regionLocator = null;
       try {
@@ -1400,7 +1400,7 @@ public final class Canary implements Tool {
             if (rsAndRMap.containsKey(rsName)) {
               regions = rsAndRMap.get(rsName);
             } else {
-              regions = new ArrayList<HRegionInfo>();
+              regions = new ArrayList<>();
               rsAndRMap.put(rsName, regions);
             }
             regions.add(r);
@@ -1438,7 +1438,7 @@ public final class Canary implements Tool {
       Map<String, List<HRegionInfo>> filteredRsAndRMap = null;
 
       if (this.targets != null && this.targets.length > 0) {
-        filteredRsAndRMap = new HashMap<String, List<HRegionInfo>>();
+        filteredRsAndRMap = new HashMap<>();
         Pattern pattern = null;
         Matcher matcher = null;
         boolean regExpFound = false;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/util/BoundedPriorityBlockingQueue.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/BoundedPriorityBlockingQueue.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/BoundedPriorityBlockingQueue.java
index 4a93151..354382c 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/BoundedPriorityBlockingQueue.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/BoundedPriorityBlockingQueue.java
@@ -152,7 +152,7 @@ public class BoundedPriorityBlockingQueue<E> extends AbstractQueue<E> implements
    */
   public BoundedPriorityBlockingQueue(int capacity,
       Comparator<? super E> comparator) {
-    this.queue = new PriorityQueue<E>(capacity, comparator);
+    this.queue = new PriorityQueue<>(capacity, comparator);
   }
 
   public boolean offer(E e) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java
index 3f05969..9e36290 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java
@@ -66,7 +66,7 @@ public class CollectionBackedScanner extends NonReversedNonLazyKeyValueScanner {
       Cell... array) {
     this.comparator = comparator;
 
-    List<Cell> tmp = new ArrayList<Cell>(array.length);
+    List<Cell> tmp = new ArrayList<>(array.length);
     Collections.addAll(tmp, array);
     Collections.sort(tmp, comparator);
     data = tmp;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ConnectionCache.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ConnectionCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ConnectionCache.java
index 0659a0d..87e867f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ConnectionCache.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ConnectionCache.java
@@ -50,9 +50,8 @@ import org.apache.commons.logging.LogFactory;
 public class ConnectionCache {
   private static final Log LOG = LogFactory.getLog(ConnectionCache.class);
 
-  private final Map<String, ConnectionInfo>
-   connections = new ConcurrentHashMap<String, ConnectionInfo>();
-  private final KeyLocker<String> locker = new KeyLocker<String>();
+  private final Map<String, ConnectionInfo> connections = new ConcurrentHashMap<>();
+  private final KeyLocker<String> locker = new KeyLocker<>();
   private final String realUserName;
   private final UserGroupInformation realUser;
   private final UserProvider userProvider;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/util/EncryptionTest.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/EncryptionTest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/EncryptionTest.java
index f45ecff..6692ee8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/EncryptionTest.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/EncryptionTest.java
@@ -40,10 +40,9 @@ import org.apache.hadoop.hbase.security.EncryptionUtil;
 public class EncryptionTest {
   private static final Log LOG = LogFactory.getLog(EncryptionTest.class);
 
-  static final Map<String, Boolean> keyProviderResults = new ConcurrentHashMap<String, Boolean>();
-  static final Map<String, Boolean> cipherProviderResults =
-    new ConcurrentHashMap<String, Boolean>();
-  static final Map<String, Boolean> cipherResults = new ConcurrentHashMap<String, Boolean>();
+  static final Map<String, Boolean> keyProviderResults = new ConcurrentHashMap<>();
+  static final Map<String, Boolean> cipherProviderResults = new ConcurrentHashMap<>();
+  static final Map<String, Boolean> cipherResults = new ConcurrentHashMap<>();
 
   private EncryptionTest() {
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java
index 0d880d0..de49d38 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java
@@ -58,7 +58,7 @@ public class FSHDFSUtils extends FSUtils {
    */
   private static Set<InetSocketAddress> getNNAddresses(DistributedFileSystem fs,
                                                       Configuration conf) {
-    Set<InetSocketAddress> addresses = new HashSet<InetSocketAddress>();
+    Set<InetSocketAddress> addresses = new HashSet<>();
     String serviceName = fs.getCanonicalServiceName();
 
     if (serviceName.startsWith("ha-hdfs")) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSRegionScanner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSRegionScanner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSRegionScanner.java
index b0af52b..0bc8783 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSRegionScanner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSRegionScanner.java
@@ -72,7 +72,7 @@ class FSRegionScanner implements Runnable {
   public void run() {
     try {
       // empty the map for each region
-      Map<String, AtomicInteger> blockCountMap = new HashMap<String, AtomicInteger>();
+      Map<String, AtomicInteger> blockCountMap = new HashMap<>();
 
       //get table name
       String tableName = regionPath.getParent().getName();
@@ -145,7 +145,7 @@ class FSRegionScanner implements Runnable {
       }
 
       if (regionDegreeLocalityMapping != null && totalBlkCount > 0) {
-        Map<String, Float> hostLocalityMap = new HashMap<String, Float>();
+        Map<String, Float> hostLocalityMap = new HashMap<>();
         for (Map.Entry<String, AtomicInteger> entry : blockCountMap.entrySet()) {
           String host = entry.getKey();
           if (host.endsWith(".")) {


Mime
View raw message