Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id D510A200C42 for ; Fri, 10 Mar 2017 23:09:49 +0100 (CET) Received: by cust-asf.ponee.io (Postfix) id D3990160B85; Fri, 10 Mar 2017 22:09:49 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 87ED0160B67 for ; Fri, 10 Mar 2017 23:09:47 +0100 (CET) Received: (qmail 53964 invoked by uid 500); 10 Mar 2017 22:09:42 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 52969 invoked by uid 99); 10 Mar 2017 22:09:41 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 10 Mar 2017 22:09:41 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 9EE67DFF47; Fri, 10 Mar 2017 22:09:41 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: syuanjiang@apache.org To: commits@hbase.apache.org Date: Fri, 10 Mar 2017 22:10:02 -0000 Message-Id: In-Reply-To: References: X-Mailer: ASF-Git Admin Mailer Subject: [22/50] [abbrv] hbase git commit: HBASE-17532 Replaced explicit type with diamond operator archived-at: Fri, 10 Mar 2017 22:09:50 -0000 http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.java index f6ae9af..b0e088c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/BaseLoadBalancer.java @@ -73,7 +73,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { protected static final int MIN_SERVER_BALANCE = 2; private volatile boolean stopped = false; - private static final List EMPTY_REGION_LIST = new ArrayList(0); + private static final List EMPTY_REGION_LIST = new ArrayList<>(0); static final Predicate IDLE_SERVER_PREDICATOR = load -> load.getNumberOfRegions() == 0; @@ -187,19 +187,19 @@ public abstract class BaseLoadBalancer implements LoadBalancer { unassignedRegions = EMPTY_REGION_LIST; } - serversToIndex = new HashMap(); - hostsToIndex = new HashMap(); - racksToIndex = new HashMap(); - tablesToIndex = new HashMap(); + serversToIndex = new HashMap<>(); + hostsToIndex = new HashMap<>(); + racksToIndex = new HashMap<>(); + tablesToIndex = new HashMap<>(); //TODO: We should get the list of tables from master - tables = new ArrayList(); + tables = new ArrayList<>(); this.rackManager = rackManager != null ? rackManager : new DefaultRackManager(); numRegions = 0; - List> serversPerHostList = new ArrayList>(); - List> serversPerRackList = new ArrayList>(); + List> serversPerHostList = new ArrayList<>(); + List> serversPerRackList = new ArrayList<>(); this.clusterState = clusterState; this.regionFinder = regionFinder; @@ -211,7 +211,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { } if (!hostsToIndex.containsKey(sn.getHostname())) { hostsToIndex.put(sn.getHostname(), numHosts++); - serversPerHostList.add(new ArrayList(1)); + serversPerHostList.add(new ArrayList<>(1)); } int serverIndex = serversToIndex.get(sn.getHostAndPort()); @@ -221,7 +221,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { String rack = this.rackManager.getRack(sn); if (!racksToIndex.containsKey(rack)) { racksToIndex.put(rack, numRacks++); - serversPerRackList.add(new ArrayList()); + serversPerRackList.add(new ArrayList<>()); } int rackIndex = racksToIndex.get(rack); serversPerRackList.get(rackIndex).add(serverIndex); @@ -233,7 +233,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { } numRegions += unassignedRegions.size(); - regionsToIndex = new HashMap(numRegions); + regionsToIndex = new HashMap<>(numRegions); servers = new ServerName[numServers]; serversPerHost = new int[numHosts][]; serversPerRack = new int[numRacks][]; @@ -1064,7 +1064,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { public static final String TABLES_ON_MASTER = "hbase.balancer.tablesOnMaster"; - protected final Set tablesOnMaster = new HashSet(); + protected final Set tablesOnMaster = new HashSet<>(); protected MetricsBalancer metricsBalancer = null; protected ClusterStatus clusterStatus = null; protected ServerName masterServerName; @@ -1170,7 +1170,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { // Move this region away from the master regionserver RegionPlan plan = new RegionPlan(region, masterServerName, dest); if (plans == null) { - plans = new ArrayList(); + plans = new ArrayList<>(); } plans.add(plan); } @@ -1183,7 +1183,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { // Move this region to the master regionserver RegionPlan plan = new RegionPlan(region, server.getKey(), masterServerName); if (plans == null) { - plans = new ArrayList(); + plans = new ArrayList<>(); } plans.add(plan); } @@ -1199,10 +1199,9 @@ public abstract class BaseLoadBalancer implements LoadBalancer { if (servers == null || regions == null || regions.isEmpty()) { return null; } - Map> assignments - = new TreeMap>(); + Map> assignments = new TreeMap<>(); if (masterServerName != null && servers.contains(masterServerName)) { - assignments.put(masterServerName, new ArrayList()); + assignments.put(masterServerName, new ArrayList<>()); for (HRegionInfo region: regions) { if (shouldBeOnMaster(region)) { assignments.get(masterServerName).add(region); @@ -1303,12 +1302,12 @@ public abstract class BaseLoadBalancer implements LoadBalancer { metricsBalancer.incrMiscInvocations(); Map> assignments = assignMasterRegions(regions, servers); if (assignments != null && !assignments.isEmpty()) { - servers = new ArrayList(servers); + servers = new ArrayList<>(servers); // Guarantee not to put other regions on master servers.remove(masterServerName); List masterRegions = assignments.get(masterServerName); if (!masterRegions.isEmpty()) { - regions = new ArrayList(regions); + regions = new ArrayList<>(regions); for (HRegionInfo region: masterRegions) { regions.remove(region); } @@ -1331,17 +1330,17 @@ public abstract class BaseLoadBalancer implements LoadBalancer { if (numServers == 1) { // Only one server, nothing fancy we can do here ServerName server = servers.get(0); - assignments.put(server, new ArrayList(regions)); + assignments.put(server, new ArrayList<>(regions)); return assignments; } Cluster cluster = createCluster(servers, regions, false); - List unassignedRegions = new ArrayList(); + List unassignedRegions = new ArrayList<>(); roundRobinAssignment(cluster, regions, unassignedRegions, servers, assignments); - List lastFewRegions = new ArrayList(); + List lastFewRegions = new ArrayList<>(); // assign the remaining by going through the list and try to assign to servers one-by-one int serverIdx = RANDOM.nextInt(numServers); for (HRegionInfo region : unassignedRegions) { @@ -1351,7 +1350,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { if (!cluster.wouldLowerAvailability(region, serverName)) { List serverRegions = assignments.get(serverName); if (serverRegions == null) { - serverRegions = new ArrayList(); + serverRegions = new ArrayList<>(); assignments.put(serverName, serverRegions); } serverRegions.add(region); @@ -1372,7 +1371,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { ServerName server = servers.get(i); List serverRegions = assignments.get(server); if (serverRegions == null) { - serverRegions = new ArrayList(); + serverRegions = new ArrayList<>(); assignments.put(server, serverRegions); } serverRegions.add(region); @@ -1416,7 +1415,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { if (shouldBeOnMaster(regionInfo)) { return masterServerName; } - servers = new ArrayList(servers); + servers = new ArrayList<>(servers); // Guarantee not to put other regions on master servers.remove(masterServerName); } @@ -1465,12 +1464,12 @@ public abstract class BaseLoadBalancer implements LoadBalancer { Map> assignments = assignMasterRegions(regions.keySet(), servers); if (assignments != null && !assignments.isEmpty()) { - servers = new ArrayList(servers); + servers = new ArrayList<>(servers); // Guarantee not to put other regions on master servers.remove(masterServerName); List masterRegions = assignments.get(masterServerName); if (!masterRegions.isEmpty()) { - regions = new HashMap(regions); + regions = new HashMap<>(regions); for (HRegionInfo region: masterRegions) { regions.remove(region); } @@ -1487,7 +1486,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { } if (numServers == 1) { // Only one server, nothing fancy we can do here ServerName server = servers.get(0); - assignments.put(server, new ArrayList(regions.keySet())); + assignments.put(server, new ArrayList<>(regions.keySet())); return assignments; } @@ -1499,7 +1498,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { // servers on the same host on different ports. ArrayListMultimap serversByHostname = ArrayListMultimap.create(); for (ServerName server : servers) { - assignments.put(server, new ArrayList()); + assignments.put(server, new ArrayList<>()); serversByHostname.put(server.getHostname(), server); } @@ -1516,7 +1515,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { for (Map.Entry entry : regions.entrySet()) { HRegionInfo region = entry.getKey(); ServerName oldServerName = entry.getValue(); - List localServers = new ArrayList(); + List localServers = new ArrayList<>(); if (oldServerName != null) { localServers = serversByHostname.get(oldServerName.getHostname()); } @@ -1629,7 +1628,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { for (int j = 0; j < numServers; j++) { ServerName server = servers.get((j + serverIdx) % numServers); - List serverRegions = new ArrayList(max); + List serverRegions = new ArrayList<>(max); for (int i = regionIdx; i < numRegions; i += numServers) { HRegionInfo region = regions.get(i % numRegions); if (cluster.wouldLowerAvailability(region, server)) { @@ -1649,7 +1648,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer { if (this.services != null && this.services.getAssignmentManager() != null) { return this.services.getAssignmentManager().getSnapShotOfAssignment(regions); } else { - return new HashMap>(); + return new HashMap<>(); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/ClusterLoadState.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/ClusterLoadState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/ClusterLoadState.java index e5f0e3b..8c337bd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/ClusterLoadState.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/ClusterLoadState.java @@ -39,7 +39,7 @@ public class ClusterLoadState { this.numRegions = 0; this.numServers = clusterState.size(); this.clusterState = clusterState; - serversByLoad = new TreeMap>(); + serversByLoad = new TreeMap<>(); // Iterate so we can count regions as we build the map for (Map.Entry> server : clusterState.entrySet()) { List regions = server.getValue(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java index d5edfab..f7e166d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java @@ -179,7 +179,7 @@ class RegionLocationFinder { */ protected List getTopBlockLocations(HRegionInfo region, String currentHost) { HDFSBlocksDistribution blocksDistribution = getBlockDistribution(region); - List topHosts = new ArrayList(); + List topHosts = new ArrayList<>(); for (String host : blocksDistribution.getTopHosts()) { if (host.equals(currentHost)) { break; @@ -250,15 +250,15 @@ class RegionLocationFinder { return Lists.newArrayList(); } - List topServerNames = new ArrayList(); + List topServerNames = new ArrayList<>(); Collection regionServers = status.getServers(); // create a mapping from hostname to ServerName for fast lookup - HashMap> hostToServerName = new HashMap>(); + HashMap> hostToServerName = new HashMap<>(); for (ServerName sn : regionServers) { String host = sn.getHostname(); if (!hostToServerName.containsKey(host)) { - hostToServerName.put(host, new ArrayList()); + hostToServerName.put(host, new ArrayList<>()); } hostToServerName.get(host).add(sn); } @@ -309,8 +309,7 @@ class RegionLocationFinder { } public void refreshAndWait(Collection hris) { - ArrayList> regionLocationFutures = - new ArrayList>(hris.size()); + ArrayList> regionLocationFutures = new ArrayList<>(hris.size()); for (HRegionInfo hregionInfo : hris) { regionLocationFutures.add(asyncGetBlockDistribution(hregionInfo)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/SimpleLoadBalancer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/SimpleLoadBalancer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/SimpleLoadBalancer.java index a6a0774..7e8d696 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/SimpleLoadBalancer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/SimpleLoadBalancer.java @@ -255,7 +255,7 @@ public class SimpleLoadBalancer extends BaseLoadBalancer { if (clusterMap.size() <= 2) { return null; } - clusterMap = new HashMap>(clusterMap); + clusterMap = new HashMap<>(clusterMap); clusterMap.remove(masterServerName); } @@ -285,14 +285,13 @@ public class SimpleLoadBalancer extends BaseLoadBalancer { // TODO: Look at data block locality or a more complex load to do this MinMaxPriorityQueue regionsToMove = MinMaxPriorityQueue.orderedBy(rpComparator).create(); - regionsToReturn = new ArrayList(); + regionsToReturn = new ArrayList<>(); // Walk down most loaded, pruning each to the max int serversOverloaded = 0; // flag used to fetch regions from head and tail of list, alternately boolean fetchFromTail = false; - Map serverBalanceInfo = - new TreeMap(); + Map serverBalanceInfo = new TreeMap<>(); for (Map.Entry> server: serversByLoad.descendingMap().entrySet()) { ServerAndLoad sal = server.getKey(); @@ -330,7 +329,7 @@ public class SimpleLoadBalancer extends BaseLoadBalancer { int neededRegions = 0; // number of regions needed to bring all up to min fetchFromTail = false; - Map underloadedServers = new HashMap(); + Map underloadedServers = new HashMap<>(); int maxToTake = numRegions - min; for (Map.Entry> server: serversByLoad.entrySet()) { @@ -524,8 +523,7 @@ public class SimpleLoadBalancer extends BaseLoadBalancer { // A structure help to map ServerName to it's load and index in ServerLoadList Map> SnLoadMap = new HashMap<>(); for (int i = 0; i < serverLoadList.size(); i++) { - SnLoadMap.put(serverLoadList.get(i).getServerName(), - new Pair(serverLoadList.get(i), i)); + SnLoadMap.put(serverLoadList.get(i).getServerName(), new Pair<>(serverLoadList.get(i), i)); } Pair shredLoad; // A List to help mark the plan in regionsToMove that should be removed http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.java index f2329bb..f68afb6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.java @@ -121,7 +121,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer { private static final Random RANDOM = new Random(System.currentTimeMillis()); private static final Log LOG = LogFactory.getLog(StochasticLoadBalancer.class); - Map> loads = new HashMap>(); + Map> loads = new HashMap<>(); // values are defaults private int maxSteps = 1000000; @@ -332,7 +332,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer { if (clusterState.size() <= 2) { return null; } - clusterState = new HashMap>(clusterState); + clusterState = new HashMap<>(clusterState); clusterState.remove(masterServerName); } @@ -482,7 +482,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer { * @return List of RegionPlan's that represent the moves needed to get to desired final state. */ private List createRegionPlans(Cluster cluster) { - List plans = new LinkedList(); + List plans = new LinkedList<>(); for (int regionIndex = 0; regionIndex < cluster.regionIndexToServerIndex.length; regionIndex++) { int initialServerIndex = cluster.initialRegionIndexToServerIndex[regionIndex]; @@ -511,7 +511,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer { // We create a new hashmap so that regions that are no longer there are removed. // However we temporarily need the old loads so we can use them to keep the rolling average. Map> oldLoads = loads; - loads = new HashMap>(); + loads = new HashMap<>(); for (ServerName sn : clusterStatus.getServers()) { ServerLoad sl = clusterStatus.getLoad(sn); @@ -522,7 +522,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer { Deque rLoads = oldLoads.get(Bytes.toString(entry.getKey())); if (rLoads == null) { // There was nothing there - rLoads = new ArrayDeque(); + rLoads = new ArrayDeque<>(); } else if (rLoads.size() >= numRegionLoadsToRemember) { rLoads.remove(); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java index c6b6f62..dddad36 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java @@ -92,7 +92,7 @@ public abstract class CleanerChore extends Schedu * @param confKey key to get the file cleaner classes from the configuration */ private void initCleanerChain(String confKey) { - this.cleanersChain = new LinkedList(); + this.cleanersChain = new LinkedList<>(); String[] logCleaners = conf.getStrings(confKey); if (logCleaners != null) { for (String className : logCleaners) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationMetaCleaner.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationMetaCleaner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationMetaCleaner.java index 5c56271..45b2401 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationMetaCleaner.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationMetaCleaner.java @@ -74,7 +74,7 @@ public class ReplicationMetaCleaner extends ScheduledChore { } } if (hasSerialScope) { - serialTables.put(entry.getValue().getTableName().getNameAsString(), new HashSet()); + serialTables.put(entry.getValue().getTableName().getNameAsString(), new HashSet<>()); } } if (serialTables.isEmpty()){ http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationZKNodeCleaner.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationZKNodeCleaner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationZKNodeCleaner.java index c0a1b75..dafc4f8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationZKNodeCleaner.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/ReplicationZKNodeCleaner.java @@ -82,7 +82,7 @@ public class ReplicationZKNodeCleaner { for (String queueId : queueIds) { ReplicationQueueInfo queueInfo = new ReplicationQueueInfo(queueId); if (!peerIds.contains(queueInfo.getPeerId())) { - undeletedQueues.computeIfAbsent(replicator, (key) -> new ArrayList()).add( + undeletedQueues.computeIfAbsent(replicator, (key) -> new ArrayList<>()).add( queueId); if (LOG.isDebugEnabled()) { LOG.debug("Undeleted replication queue for removed peer found: " http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/locking/LockProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/locking/LockProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/locking/LockProcedure.java index c4b49f0..8e490eb 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/locking/LockProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/locking/LockProcedure.java @@ -477,4 +477,4 @@ public final class LockProcedure extends Procedure env.getProcedureScheduler().wakeRegions(LockProcedure.this, tableName, regionInfos); } } -} \ No newline at end of file +} http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java index 5c67258..9c8358b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/normalizer/SimpleRegionNormalizer.java @@ -118,7 +118,7 @@ public class SimpleRegionNormalizer implements RegionNormalizer { return null; } - List plans = new ArrayList(); + List plans = new ArrayList<>(); List tableRegions = masterServices.getAssignmentManager().getRegionStates(). getRegionsOfTable(table); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java index aefd14c..347d01d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CloneSnapshotProcedure.java @@ -70,8 +70,7 @@ public class CloneSnapshotProcedure private HTableDescriptor hTableDescriptor; private SnapshotDescription snapshot; private List newRegions = null; - private Map > parentsToChildrenPairMap = - new HashMap>(); + private Map > parentsToChildrenPairMap = new HashMap<>(); // Monitor private MonitoredTask monitorStatus = null; @@ -264,18 +263,18 @@ public class CloneSnapshotProcedure if (cloneSnapshotMsg.getRegionInfoCount() == 0) { newRegions = null; } else { - newRegions = new ArrayList(cloneSnapshotMsg.getRegionInfoCount()); + newRegions = new ArrayList<>(cloneSnapshotMsg.getRegionInfoCount()); for (HBaseProtos.RegionInfo hri: cloneSnapshotMsg.getRegionInfoList()) { newRegions.add(HRegionInfo.convert(hri)); } } if (cloneSnapshotMsg.getParentToChildRegionsPairListCount() > 0) { - parentsToChildrenPairMap = new HashMap>(); + parentsToChildrenPairMap = new HashMap<>(); for (MasterProcedureProtos.RestoreParentToChildRegionsPair parentToChildrenPair: cloneSnapshotMsg.getParentToChildRegionsPairListList()) { parentsToChildrenPairMap.put( parentToChildrenPair.getParentRegionName(), - new Pair( + new Pair<>( parentToChildrenPair.getChild1RegionName(), parentToChildrenPair.getChild2RegionName())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CreateTableProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CreateTableProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CreateTableProcedure.java index 2421dfc..ced7abc 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CreateTableProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/CreateTableProcedure.java @@ -208,7 +208,7 @@ public class CreateTableProcedure if (state.getRegionInfoCount() == 0) { newRegions = null; } else { - newRegions = new ArrayList(state.getRegionInfoCount()); + newRegions = new ArrayList<>(state.getRegionInfoCount()); for (HBaseProtos.RegionInfo hri: state.getRegionInfoList()) { newRegions.add(HRegionInfo.convert(hri)); } @@ -364,8 +364,7 @@ public class CreateTableProcedure if (numRegionReplicas <= 0) { return regions; } - List hRegionInfos = - new ArrayList((numRegionReplicas+1)*regions.size()); + List hRegionInfos = new ArrayList<>((numRegionReplicas+1)*regions.size()); for (int i = 0; i < regions.size(); i++) { for (int j = 1; j <= numRegionReplicas; j++) { hRegionInfos.add(RegionReplicaUtil.getRegionInfoForReplica(regions.get(i), j)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteTableProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteTableProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteTableProcedure.java index 06b666b..9d0a283 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteTableProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/DeleteTableProcedure.java @@ -221,7 +221,7 @@ public class DeleteTableProcedure if (state.getRegionInfoCount() == 0) { regions = null; } else { - regions = new ArrayList(state.getRegionInfoCount()); + regions = new ArrayList<>(state.getRegionInfoCount()); for (HBaseProtos.RegionInfo hri: state.getRegionInfoList()) { regions.add(HRegionInfo.convert(hri)); } @@ -343,7 +343,7 @@ public class DeleteTableProcedure Scan tableScan = MetaTableAccessor.getScanForTableName(connection, tableName); try (Table metaTable = connection.getTable(TableName.META_TABLE_NAME)) { - List deletes = new ArrayList(); + List deletes = new ArrayList<>(); try (ResultScanner resScanner = metaTable.getScanner(tableScan)) { for (Result result : resScanner) { deletes.add(new Delete(result.getRow())); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java index f4ecf15..4d67edd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/EnableTableProcedure.java @@ -357,7 +357,7 @@ public class EnableTableProcedure // need to potentially create some regions for the replicas List unrecordedReplicas = - AssignmentManager.replicaRegionsNotRecordedInMeta(new HashSet( + AssignmentManager.replicaRegionsNotRecordedInMeta(new HashSet<>( regionsToAssign.keySet()), masterServices); Map> srvToUnassignedRegs = assignmentManager.getBalancer().roundRobinAssignment(unrecordedReplicas, @@ -464,8 +464,7 @@ public class EnableTableProcedure private static Map regionsToAssignWithServerName( final MasterProcedureEnv env, final List> regionsInMeta) throws IOException { - Map regionsToAssign = - new HashMap(regionsInMeta.size()); + Map regionsToAssign = new HashMap<>(regionsInMeta.size()); RegionStates regionStates = env.getMasterServices().getAssignmentManager().getRegionStates(); for (Pair regionLocation : regionsInMeta) { HRegionInfo hri = regionLocation.getFirst(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MasterDDLOperationHelper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MasterDDLOperationHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MasterDDLOperationHelper.java index 980bf94..4b9a7ab 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MasterDDLOperationHelper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MasterDDLOperationHelper.java @@ -97,12 +97,12 @@ public final class MasterDDLOperationHelper { regionLocations = locator.getAllRegionLocations(); } // Convert List to Map. - NavigableMap hri2Sn = new TreeMap(); + NavigableMap hri2Sn = new TreeMap<>(); for (HRegionLocation location : regionLocations) { hri2Sn.put(location.getRegionInfo(), location.getServerName()); } TreeMap> serverToRegions = Maps.newTreeMap(); - List reRegions = new ArrayList(); + List reRegions = new ArrayList<>(); for (HRegionInfo hri : regionInfoList) { ServerName sn = hri2Sn.get(hri); // Skip the offlined split parent region http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MergeTableRegionsProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MergeTableRegionsProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MergeTableRegionsProcedure.java index d7fe5f6..366378a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MergeTableRegionsProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/MergeTableRegionsProcedure.java @@ -678,7 +678,7 @@ public class MergeTableRegionsProcedure final MasterCoprocessorHost cpHost = env.getMasterCoprocessorHost(); if (cpHost != null) { @MetaMutationAnnotation - final List metaEntries = new ArrayList(); + final List metaEntries = new ArrayList<>(); boolean ret = cpHost.preMergeRegionsCommit(regionsToMerge, metaEntries, getUser()); if (ret) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.java index f1b411a..6a70f62 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.java @@ -337,7 +337,7 @@ public class ModifyTableProcedure final int newReplicaCount = newHTableDescriptor.getRegionReplication(); if (newReplicaCount < oldReplicaCount) { - Set tableRows = new HashSet(); + Set tableRows = new HashSet<>(); Connection connection = env.getMasterServices().getConnection(); Scan scan = MetaTableAccessor.getScanForTableName(connection, getTableName()); scan.addColumn(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java index d99bd6b..f8c9d8f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.java @@ -66,8 +66,7 @@ public class RestoreSnapshotProcedure private List regionsToRestore = null; private List regionsToRemove = null; private List regionsToAdd = null; - private Map> parentsToChildrenPairMap = - new HashMap>(); + private Map> parentsToChildrenPairMap = new HashMap<>(); private SnapshotDescription snapshot; @@ -275,8 +274,7 @@ public class RestoreSnapshotProcedure if (restoreSnapshotMsg.getRegionInfoForRestoreCount() == 0) { regionsToRestore = null; } else { - regionsToRestore = - new ArrayList(restoreSnapshotMsg.getRegionInfoForRestoreCount()); + regionsToRestore = new ArrayList<>(restoreSnapshotMsg.getRegionInfoForRestoreCount()); for (HBaseProtos.RegionInfo hri: restoreSnapshotMsg.getRegionInfoForRestoreList()) { regionsToRestore.add(HRegionInfo.convert(hri)); } @@ -284,8 +282,7 @@ public class RestoreSnapshotProcedure if (restoreSnapshotMsg.getRegionInfoForRemoveCount() == 0) { regionsToRemove = null; } else { - regionsToRemove = - new ArrayList(restoreSnapshotMsg.getRegionInfoForRemoveCount()); + regionsToRemove = new ArrayList<>(restoreSnapshotMsg.getRegionInfoForRemoveCount()); for (HBaseProtos.RegionInfo hri: restoreSnapshotMsg.getRegionInfoForRemoveList()) { regionsToRemove.add(HRegionInfo.convert(hri)); } @@ -293,7 +290,7 @@ public class RestoreSnapshotProcedure if (restoreSnapshotMsg.getRegionInfoForAddCount() == 0) { regionsToAdd = null; } else { - regionsToAdd = new ArrayList(restoreSnapshotMsg.getRegionInfoForAddCount()); + regionsToAdd = new ArrayList<>(restoreSnapshotMsg.getRegionInfoForAddCount()); for (HBaseProtos.RegionInfo hri: restoreSnapshotMsg.getRegionInfoForAddList()) { regionsToAdd.add(HRegionInfo.convert(hri)); } @@ -303,7 +300,7 @@ public class RestoreSnapshotProcedure restoreSnapshotMsg.getParentToChildRegionsPairListList()) { parentsToChildrenPairMap.put( parentToChildrenPair.getParentRegionName(), - new Pair( + new Pair<>( parentToChildrenPair.getChild1RegionName(), parentToChildrenPair.getChild2RegionName())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java index 7b4eb6e..484decc 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.java @@ -99,7 +99,7 @@ implements ServerProcedureInterface { public static final int DEFAULT_WAIT_ON_RIT = 30000; - private static final Set META_REGION_SET = new HashSet(); + private static final Set META_REGION_SET = new HashSet<>(); static { META_REGION_SET.add(HRegionInfo.FIRST_META_REGIONINFO); } @@ -424,7 +424,7 @@ implements ServerProcedureInterface { private List calcRegionsToAssign(final MasterProcedureEnv env) throws IOException { AssignmentManager am = env.getMasterServices().getAssignmentManager(); - List regionsToAssignAggregator = new ArrayList(); + List regionsToAssignAggregator = new ArrayList<>(); int replicaCount = env.getMasterConfiguration().getInt(HConstants.META_REPLICAS_NUM, HConstants.DEFAULT_META_REPLICA_NUM); for (int i = 1; i < replicaCount; i++) { @@ -625,14 +625,14 @@ implements ServerProcedureInterface { this.shouldSplitWal = state.getShouldSplitWal(); int size = state.getRegionsOnCrashedServerCount(); if (size > 0) { - this.regionsOnCrashedServer = new HashSet(size); + this.regionsOnCrashedServer = new HashSet<>(size); for (RegionInfo ri: state.getRegionsOnCrashedServerList()) { this.regionsOnCrashedServer.add(HRegionInfo.convert(ri)); } } size = state.getRegionsAssignedCount(); if (size > 0) { - this.regionsAssigned = new ArrayList(size); + this.regionsAssigned = new ArrayList<>(size); for (RegionInfo ri: state.getRegionsOnCrashedServerList()) { this.regionsAssigned.add(HRegionInfo.convert(ri)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/SplitTableRegionProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/SplitTableRegionProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/SplitTableRegionProcedure.java index 69b89be..3cd6c66 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/SplitTableRegionProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/SplitTableRegionProcedure.java @@ -547,7 +547,7 @@ public class SplitTableRegionProcedure } if (nbFiles == 0) { // no file needs to be splitted. - return new Pair(0,0); + return new Pair<>(0,0); } // Default max #threads to use is the smaller of table's configured number of blocking store // files or the available number of logical cores. @@ -561,7 +561,7 @@ public class SplitTableRegionProcedure " using " + maxThreads + " threads"); ThreadPoolExecutor threadPool = (ThreadPoolExecutor) Executors.newFixedThreadPool( maxThreads, Threads.getNamedThreadFactory("StoreFileSplitter-%1$d")); - List>> futures = new ArrayList>> (nbFiles); + List>> futures = new ArrayList<>(nbFiles); // Split each store file. final HTableDescriptor htd = env.getMasterServices().getTableDescriptors().get(getTableName()); @@ -617,7 +617,7 @@ public class SplitTableRegionProcedure LOG.debug("Split storefiles for region " + parentHRI + " Daughter A: " + daughterA + " storefiles, Daughter B: " + daughterB + " storefiles."); } - return new Pair(daughterA, daughterB); + return new Pair<>(daughterA, daughterB); } private void assertReferenceFileCount( @@ -646,7 +646,7 @@ public class SplitTableRegionProcedure if (LOG.isDebugEnabled()) { LOG.debug("Splitting complete for store file: " + sf.getPath() + " for region: " + parentHRI); } - return new Pair(path_first, path_second); + return new Pair<>(path_first, path_second); } /** @@ -684,7 +684,7 @@ public class SplitTableRegionProcedure **/ private void preSplitRegionBeforePONR(final MasterProcedureEnv env) throws IOException, InterruptedException { - final List metaEntries = new ArrayList(); + final List metaEntries = new ArrayList<>(); final MasterCoprocessorHost cpHost = env.getMasterCoprocessorHost(); if (cpHost != null) { if (cpHost.preSplitBeforePONRAction(getSplitRow(), metaEntries, getUser())) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.java index 7482831..2ab142a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.java @@ -250,7 +250,7 @@ public class TruncateTableProcedure if (state.getRegionInfoCount() == 0) { regions = null; } else { - regions = new ArrayList(state.getRegionInfoCount()); + regions = new ArrayList<>(state.getRegionInfoCount()); for (HBaseProtos.RegionInfo hri: state.getRegionInfoList()) { regions.add(HRegionInfo.convert(hri)); } @@ -258,7 +258,7 @@ public class TruncateTableProcedure } private static List recreateRegionInfo(final List regions) { - ArrayList newRegions = new ArrayList(regions.size()); + ArrayList newRegions = new ArrayList<>(regions.size()); for (HRegionInfo hri: regions) { newRegions.add(new HRegionInfo(hri.getTable(), hri.getStartKey(), hri.getEndKey())); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java index 113ce58..5f86e08 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/DisabledTableSnapshotHandler.java @@ -74,7 +74,7 @@ public class DisabledTableSnapshotHandler extends TakeSnapshotHandler { // 1. get all the regions hosting this table. // extract each pair to separate lists - Set regions = new HashSet(); + Set regions = new HashSet<>(); for (Pair p : regionsAndLocations) { // Don't include non-default regions HRegionInfo hri = p.getFirst(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java index e63727a..73cd4d7 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/EnabledTableSnapshotHandler.java @@ -71,7 +71,7 @@ public class EnabledTableSnapshotHandler extends TakeSnapshotHandler { @Override protected void snapshotRegions(List> regions) throws HBaseSnapshotException, IOException { - Set regionServers = new HashSet(regions.size()); + Set regionServers = new HashSet<>(regions.size()); for (Pair region : regions) { if (region != null && region.getFirst() != null && region.getSecond() != null) { HRegionInfo hri = region.getFirst(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java index f03344c..b6641de 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java @@ -90,13 +90,12 @@ public class SnapshotFileCache implements Stoppable { private final FileSystem fs; private final SnapshotFileInspector fileInspector; private final Path snapshotDir; - private final Set cache = new HashSet(); + private final Set cache = new HashSet<>(); /** * This is a helper map of information about the snapshot directories so we don't need to rescan * them if they haven't changed since the last time we looked. */ - private final Map snapshots = - new HashMap(); + private final Map snapshots = new HashMap<>(); private final Timer refreshTimer; private long lastModifiedTime = Long.MIN_VALUE; @@ -229,7 +228,7 @@ public class SnapshotFileCache implements Stoppable { // 2.clear the cache this.cache.clear(); - Map known = new HashMap(); + Map known = new HashMap<>(); // 3. check each of the snapshot directories FileStatus[] snapshots = FSUtils.listStatus(fs, snapshotDir); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java index b950079..54b68d3 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java @@ -147,8 +147,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable // The map is always accessed and modified under the object lock using synchronized. // snapshotTable() will insert an Handler in the table. // isSnapshotDone() will remove the handler requested if the operation is finished. - private Map snapshotHandlers = - new HashMap(); + private Map snapshotHandlers = new HashMap<>(); // Restore map, with table name as key, procedure ID as value. // The map is always accessed and modified under the object lock using synchronized. @@ -157,7 +156,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable // TODO: just as the Apache HBase 1.x implementation, this map would not survive master // restart/failover. This is just a stopgap implementation until implementation of taking // snapshot using Procedure-V2. - private Map restoreTableToProcIdMap = new HashMap(); + private Map restoreTableToProcIdMap = new HashMap<>(); private Path rootDir; private ExecutorService executorService; @@ -168,7 +167,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable * - create snapshot * - SnapshotCleaner * */ - private KeyLocker locks = new KeyLocker(); + private KeyLocker locks = new KeyLocker<>(); @@ -209,7 +208,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable * @throws IOException File system exception */ private List getCompletedSnapshots(Path snapshotDir) throws IOException { - List snapshotDescs = new ArrayList(); + List snapshotDescs = new ArrayList<>(); // first create the snapshot root path and check to see if it exists FileSystem fs = master.getMasterFileSystem().getFileSystem(); if (snapshotDir == null) snapshotDir = SnapshotDescriptionUtils.getSnapshotsDir(rootDir); @@ -1032,11 +1031,11 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable boolean userDisabled = (enabled != null && enabled.trim().length() > 0 && !snapshotEnabled); // Extract cleaners from conf - Set hfileCleaners = new HashSet(); + Set hfileCleaners = new HashSet<>(); String[] cleaners = conf.getStrings(HFileCleaner.MASTER_HFILE_CLEANER_PLUGINS); if (cleaners != null) Collections.addAll(hfileCleaners, cleaners); - Set logCleaners = new HashSet(); + Set logCleaners = new HashSet<>(); cleaners = conf.getStrings(HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS); if (cleaners != null) Collections.addAll(logCleaners, cleaners); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java index 992f28e..123758f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java @@ -183,7 +183,7 @@ public abstract class TakeSnapshotHandler extends EventHandler implements Snapsh monitor.rethrowException(); // extract each pair to separate lists - Set serverNames = new HashSet(); + Set serverNames = new HashSet<>(); for (Pair p : regionsAndLocations) { if (p != null && p.getFirst() != null && p.getSecond() != null) { HRegionInfo hri = p.getFirst(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java index 697286c..56c0242 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java @@ -174,7 +174,7 @@ public class DefaultMobStoreCompactor extends DefaultCompactor { long bytesWrittenProgressForShippedCall = 0; // Since scanner.next() can return 'false' but still be delivering data, // we have to use a do/while loop. - List cells = new ArrayList(); + List cells = new ArrayList<>(); // Limit to "hbase.hstore.compaction.kv.max" (default 10) to avoid OOME int closeCheckSizeLimit = HStore.getCloseCheckInterval(); long lastMillis = 0; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java index 3c6a071..2456a41 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreFlusher.java @@ -98,7 +98,7 @@ public class DefaultMobStoreFlusher extends DefaultStoreFlusher { @Override public List flushSnapshot(MemStoreSnapshot snapshot, long cacheFlushId, MonitoredTask status, ThroughputController throughputController) throws IOException { - ArrayList result = new ArrayList(); + ArrayList result = new ArrayList<>(); long cellsCount = snapshot.getCellsCount(); if (cellsCount == 0) return result; // don't flush if there are no entries @@ -179,7 +179,7 @@ public class DefaultMobStoreFlusher extends DefaultStoreFlusher { byte[] fileName = Bytes.toBytes(mobFileWriter.getPath().getName()); ScannerContext scannerContext = ScannerContext.newBuilder().setBatchLimit(compactionKVMax).build(); - List cells = new ArrayList(); + List cells = new ArrayList<>(); boolean hasMore; String flushName = ThroughputControlUtil.getNameForThrottling(store, "flush"); boolean control = throughputController != null && !store.getRegionInfo().isSystemTable(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFile.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFile.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFile.java index 8c760e6..cd4c079 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFile.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFile.java @@ -55,7 +55,7 @@ public class MobFile { * @throws IOException */ public StoreFileScanner getScanner() throws IOException { - List sfs = new ArrayList(); + List sfs = new ArrayList<>(); sfs.add(sf); List sfScanners = StoreFileScanner.getScannersForStoreFiles(sfs, false, true, false, false, sf.getMaxMemstoreTS()); @@ -85,7 +85,7 @@ public class MobFile { public Cell readCell(Cell search, boolean cacheMobBlocks, long readPt) throws IOException { Cell result = null; StoreFileScanner scanner = null; - List sfs = new ArrayList(); + List sfs = new ArrayList<>(); sfs.add(sf); try { List sfScanners = StoreFileScanner.getScannersForStoreFiles(sfs, http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFileCache.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFileCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFileCache.java index fd62340..25328b1 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFileCache.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobFileCache.java @@ -102,7 +102,7 @@ public class MobFileCache { this.mobFileMaxCacheSize = conf.getInt(MobConstants.MOB_FILE_CACHE_SIZE_KEY, MobConstants.DEFAULT_MOB_FILE_CACHE_SIZE); isCacheEnabled = (mobFileMaxCacheSize > 0); - map = new ConcurrentHashMap(mobFileMaxCacheSize); + map = new ConcurrentHashMap<>(mobFileMaxCacheSize); if (isCacheEnabled) { long period = conf.getLong(MobConstants.MOB_CACHE_EVICT_PERIOD, MobConstants.DEFAULT_MOB_CACHE_EVICT_PERIOD); // in seconds @@ -136,12 +136,12 @@ public class MobFileCache { return; } printStatistics(); - List evictedFiles = new ArrayList(); + List evictedFiles = new ArrayList<>(); try { if (map.size() <= mobFileMaxCacheSize) { return; } - List files = new ArrayList(map.values()); + List files = new ArrayList<>(map.values()); Collections.sort(files); int start = (int) (mobFileMaxCacheSize * evictRemainRatio); if (start >= 0) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java index 8191828..eb75120 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/MobUtils.java @@ -314,7 +314,7 @@ public final class MobUtils { // no file found return; } - List filesToClean = new ArrayList(); + List filesToClean = new ArrayList<>(); int deletedFileCount = 0; for (FileStatus file : stats) { String fileName = file.getPath().getName(); @@ -482,7 +482,7 @@ public final class MobUtils { public static Cell createMobRefCell(Cell cell, byte[] fileName, Tag tableNameTag) { // Append the tags to the KeyValue. // The key is same, the value is the filename of the mob file - List tags = new ArrayList(); + List tags = new ArrayList<>(); // Add the ref tag as the 1st one. tags.add(MobConstants.MOB_REF_TAG); // Add the tag of the source table name, this table is where this mob file is flushed @@ -832,7 +832,7 @@ public final class MobUtils { if (maxThreads == 0) { maxThreads = 1; } - final SynchronousQueue queue = new SynchronousQueue(); + final SynchronousQueue queue = new SynchronousQueue<>(); ThreadPoolExecutor pool = new ThreadPoolExecutor(1, maxThreads, 60, TimeUnit.SECONDS, queue, Threads.newDaemonThreadFactory("MobCompactor"), new RejectedExecutionHandler() { @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactionRequest.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactionRequest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactionRequest.java index b6cf814..f1dcaee 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactionRequest.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/compactions/PartitionedMobCompactionRequest.java @@ -71,7 +71,7 @@ public class PartitionedMobCompactionRequest extends MobCompactionRequest { * the same partition. */ protected static class CompactionPartition { - private List files = new ArrayList(); + private List files = new ArrayList<>(); private CompactionPartitionId partitionId; // The startKey and endKey of this partition, both are inclusive. http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java index 7ff7db6..b49df28 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java @@ -224,7 +224,7 @@ public class MonitoredRPCHandlerImpl extends MonitoredTaskImpl if (getState() != State.RUNNING) { return map; } - Map rpcJSON = new HashMap(); + Map rpcJSON = new HashMap<>(); ArrayList paramList = new ArrayList(); map.put("rpcCall", rpcJSON); rpcJSON.put("queuetimems", getRPCQueueTime()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java index 27aaceb..dda77ac 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java @@ -148,7 +148,7 @@ class MonitoredTaskImpl implements MonitoredTask { @Override public Map toMap() { - Map map = new HashMap(); + Map map = new HashMap<>(); map.put("description", getDescription()); map.put("status", getStatus()); map.put("state", getState()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java index 949b036..ff92704 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java @@ -72,7 +72,7 @@ public class TaskMonitor { MonitoredTask proxy = (MonitoredTask) Proxy.newProxyInstance( stat.getClass().getClassLoader(), new Class[] { MonitoredTask.class }, - new PassthroughInvocationHandler(stat)); + new PassthroughInvocationHandler<>(stat)); TaskAndWeakRefPair pair = new TaskAndWeakRefPair(stat, proxy); if (tasks.isFull()) { purgeExpiredTasks(); @@ -87,7 +87,7 @@ public class TaskMonitor { MonitoredRPCHandler proxy = (MonitoredRPCHandler) Proxy.newProxyInstance( stat.getClass().getClassLoader(), new Class[] { MonitoredRPCHandler.class }, - new PassthroughInvocationHandler(stat)); + new PassthroughInvocationHandler<>(stat)); TaskAndWeakRefPair pair = new TaskAndWeakRefPair(stat, proxy); rpcTasks.add(pair); return proxy; @@ -189,7 +189,7 @@ public class TaskMonitor { public TaskAndWeakRefPair(MonitoredTask stat, MonitoredTask proxy) { this.impl = stat; - this.weakProxy = new WeakReference(proxy); + this.weakProxy = new WeakReference<>(proxy); } public MonitoredTask get() { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceStateManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceStateManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceStateManager.java index 523b056..604f211 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceStateManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceStateManager.java @@ -47,7 +47,7 @@ class NamespaceStateManager { private volatile boolean initialized = false; public NamespaceStateManager(MasterServices masterServices) { - nsStateCache = new ConcurrentHashMap(); + nsStateCache = new ConcurrentHashMap<>(); master = masterServices; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceTableAndRegionInfo.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceTableAndRegionInfo.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceTableAndRegionInfo.java index 86651e4..d30de6e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceTableAndRegionInfo.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/namespace/NamespaceTableAndRegionInfo.java @@ -40,7 +40,7 @@ class NamespaceTableAndRegionInfo { public NamespaceTableAndRegionInfo(String namespace) { this.name = namespace; - this.tableAndRegionInfo = new HashMap(); + this.tableAndRegionInfo = new HashMap<>(); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/MasterProcedureManagerHost.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/MasterProcedureManagerHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/MasterProcedureManagerHost.java index 8161ffe..222c933 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/MasterProcedureManagerHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/MasterProcedureManagerHost.java @@ -33,8 +33,7 @@ import org.apache.zookeeper.KeeperException; public class MasterProcedureManagerHost extends ProcedureManagerHost { - private Hashtable procedureMgrMap - = new Hashtable(); + private Hashtable procedureMgrMap = new Hashtable<>(); @Override public void loadProcedures(Configuration conf) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java index 0279a60..1d20ba5 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Procedure.java @@ -125,9 +125,9 @@ public class Procedure implements Callable, ForeignExceptionListener { public Procedure(ProcedureCoordinator coord, ForeignExceptionDispatcher monitor, long wakeFreq, long timeout, String procName, byte[] args, List expectedMembers) { this.coord = coord; - this.acquiringMembers = new ArrayList(expectedMembers); - this.inBarrierMembers = new ArrayList(acquiringMembers.size()); - this.dataFromFinishedMembers = new HashMap(); + this.acquiringMembers = new ArrayList<>(expectedMembers); + this.inBarrierMembers = new ArrayList<>(acquiringMembers.size()); + this.dataFromFinishedMembers = new HashMap<>(); this.procName = procName; this.args = args; this.monitor = monitor; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java index b7e0c04..8a64cc8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureCoordinator.java @@ -113,7 +113,7 @@ public class ProcedureCoordinator { public static ThreadPoolExecutor defaultPool(String coordName, int opThreads, long keepAliveMillis) { return new ThreadPoolExecutor(1, opThreads, keepAliveMillis, TimeUnit.MILLISECONDS, - new SynchronousQueue(), + new SynchronousQueue<>(), new DaemonThreadFactory("(" + coordName + ")-proc-coordinator-pool")); } @@ -325,6 +325,6 @@ public class ProcedureCoordinator { * @return Return set of all procedure names. */ public Set getProcedureNames() { - return new HashSet(procedures.keySet()); + return new HashSet<>(procedures.keySet()); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureManagerHost.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureManagerHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureManagerHost.java index 3ab4ac5..f61ce14 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureManagerHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureManagerHost.java @@ -46,7 +46,7 @@ public abstract class ProcedureManagerHost { private static final Log LOG = LogFactory.getLog(ProcedureManagerHost.class); - protected Set procedures = new HashSet(); + protected Set procedures = new HashSet<>(); /** * Load system procedures. Read the class names from configuration. @@ -60,7 +60,7 @@ public abstract class ProcedureManagerHost { if (defaultProcClasses == null || defaultProcClasses.length == 0) return; - List configured = new ArrayList(); + List configured = new ArrayList<>(); for (String className : defaultProcClasses) { className = className.trim(); ClassLoader cl = this.getClass().getClassLoader(); @@ -105,7 +105,7 @@ public abstract class ProcedureManagerHost { } public Set getProcedureManagers() { - Set returnValue = new HashSet(); + Set returnValue = new HashSet<>(); for (E e: procedures) { returnValue.add(e); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMember.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMember.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMember.java index 485821e..baed1f3 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMember.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureMember.java @@ -86,7 +86,7 @@ public class ProcedureMember implements Closeable { public static ThreadPoolExecutor defaultPool(String memberName, int procThreads, long keepAliveMillis) { return new ThreadPoolExecutor(1, procThreads, keepAliveMillis, TimeUnit.MILLISECONDS, - new SynchronousQueue(), + new SynchronousQueue<>(), new DaemonThreadFactory("member: '" + memberName + "' subprocedure-pool")); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/MasterFlushTableProcedureManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/MasterFlushTableProcedureManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/MasterFlushTableProcedureManager.java index 802a5ab..7b624a5 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/MasterFlushTableProcedureManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/MasterFlushTableProcedureManager.java @@ -68,7 +68,7 @@ public class MasterFlushTableProcedureManager extends MasterProcedureManager { private MasterServices master; private ProcedureCoordinator coordinator; - private Map procMap = new HashMap(); + private Map procMap = new HashMap<>(); private boolean stopped; public MasterFlushTableProcedureManager() {}; @@ -135,7 +135,7 @@ public class MasterFlushTableProcedureManager extends MasterProcedureManager { master.getConnection(), tableName, false); } - Set regionServers = new HashSet(regionsAndLocations.size()); + Set regionServers = new HashSet<>(regionsAndLocations.size()); for (Pair region : regionsAndLocations) { if (region != null && region.getFirst() != null && region.getSecond() != null) { HRegionInfo hri = region.getFirst(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/RegionServerFlushTableProcedureManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/RegionServerFlushTableProcedureManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/RegionServerFlushTableProcedureManager.java index 1aa959c..147c013 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/RegionServerFlushTableProcedureManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/flush/RegionServerFlushTableProcedureManager.java @@ -201,7 +201,7 @@ public class RegionServerFlushTableProcedureManager extends RegionServerProcedur private final ExecutorCompletionService taskPool; private final ThreadPoolExecutor executor; private volatile boolean stopped; - private final List> futures = new ArrayList>(); + private final List> futures = new ArrayList<>(); private final String name; FlushTableSubprocedurePool(String name, Configuration conf, Abortable abortable) { @@ -213,10 +213,10 @@ public class RegionServerFlushTableProcedureManager extends RegionServerProcedur int threads = conf.getInt(CONCURENT_FLUSH_TASKS_KEY, DEFAULT_CONCURRENT_FLUSH_TASKS); this.name = name; executor = new ThreadPoolExecutor(threads, threads, keepAlive, TimeUnit.MILLISECONDS, - new LinkedBlockingQueue(), new DaemonThreadFactory("rs(" + new LinkedBlockingQueue<>(), new DaemonThreadFactory("rs(" + name + ")-flush-proc-pool")); executor.allowCoreThreadTimeOut(true); - taskPool = new ExecutorCompletionService(executor); + taskPool = new ExecutorCompletionService<>(executor); } boolean hasTasks() { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java index c301759..8f681f0 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/protobuf/ReplicationProtbufUtil.java @@ -99,7 +99,7 @@ public class ReplicationProtbufUtil { buildReplicateWALEntryRequest(final Entry[] entries, byte[] encodedRegionName, String replicationClusterId, Path sourceBaseNamespaceDir, Path sourceHFileArchiveDir) { // Accumulate all the Cells seen in here. - List> allCells = new ArrayList>(entries.length); + List> allCells = new ArrayList<>(entries.length); int size = 0; WALProtos.FamilyScope.Builder scopeBuilder = WALProtos.FamilyScope.newBuilder(); AdminProtos.WALEntry.Builder entryBuilder = AdminProtos.WALEntry.newBuilder(); @@ -165,7 +165,7 @@ public class ReplicationProtbufUtil { builder.setSourceHFileArchiveDirPath(sourceHFileArchiveDir.toString()); } - return new Pair(builder.build(), + return new Pair<>(builder.build(), getCellScanner(allCells, size)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java index 647a770..5dab2e3 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/MasterQuotaManager.java @@ -80,9 +80,9 @@ public class MasterQuotaManager implements RegionStateListener { } LOG.info("Initializing quota support"); - namespaceLocks = new NamedLock(); - tableLocks = new NamedLock(); - userLocks = new NamedLock(); + namespaceLocks = new NamedLock<>(); + tableLocks = new NamedLock<>(); + userLocks = new NamedLock<>(); namespaceQuotaManager = new NamespaceAuditor(masterServices); namespaceQuotaManager.start(); @@ -460,7 +460,7 @@ public class MasterQuotaManager implements RegionStateListener { } private static class NamedLock { - private HashSet locks = new HashSet(); + private HashSet locks = new HashSet<>(); public void lock(final T name) throws InterruptedException { synchronized (locks) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaCache.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaCache.java index 1451052..ad91617 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaCache.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaCache.java @@ -65,12 +65,9 @@ public class QuotaCache implements Stoppable { // for testing purpose only, enforce the cache to be always refreshed static boolean TEST_FORCE_REFRESH = false; - private final ConcurrentHashMap namespaceQuotaCache = - new ConcurrentHashMap(); - private final ConcurrentHashMap tableQuotaCache = - new ConcurrentHashMap(); - private final ConcurrentHashMap userQuotaCache = - new ConcurrentHashMap(); + private final ConcurrentHashMap namespaceQuotaCache = new ConcurrentHashMap<>(); + private final ConcurrentHashMap tableQuotaCache = new ConcurrentHashMap<>(); + private final ConcurrentHashMap userQuotaCache = new ConcurrentHashMap<>(); private final RegionServerServices rsServices; private QuotaRefresherChore refreshChore; @@ -262,8 +259,8 @@ public class QuotaCache implements Stoppable { long evictPeriod = refreshPeriod * EVICT_PERIOD_FACTOR; // Find the quota entries to update - List gets = new ArrayList(); - List toRemove = new ArrayList(); + List gets = new ArrayList<>(); + List toRemove = new ArrayList<>(); for (Map.Entry entry: quotasMap.entrySet()) { long lastUpdate = entry.getValue().getLastUpdate(); long lastQuery = entry.getValue().getLastQuery(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java index ab646b9..fd12fa1 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/QuotaUtil.java @@ -167,7 +167,7 @@ public class QuotaUtil extends QuotaTableUtil { long nowTs = EnvironmentEdgeManager.currentTime(); Result[] results = doGet(connection, gets); - Map userQuotas = new HashMap(results.length); + Map userQuotas = new HashMap<>(results.length); for (int i = 0; i < results.length; ++i) { byte[] key = gets.get(i).getRow(); assert isUserRowKey(key); @@ -232,7 +232,7 @@ public class QuotaUtil extends QuotaTableUtil { long nowTs = EnvironmentEdgeManager.currentTime(); Result[] results = doGet(connection, gets); - Map globalQuotas = new HashMap(results.length); + Map globalQuotas = new HashMap<>(results.length); for (int i = 0; i < results.length; ++i) { byte[] row = gets.get(i).getRow(); K key = kfr.getKeyFromRow(row); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java index cb00c34..21b4768 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/quotas/UserQuotaState.java @@ -123,7 +123,7 @@ public class UserQuotaState extends QuotaState { private Map setLimiter(Map limiters, final K key, final Quotas quotas) { if (limiters == null) { - limiters = new HashMap(); + limiters = new HashMap<>(); } QuotaLimiter limiter = quotas.hasThrottle() ? @@ -164,7 +164,7 @@ public class UserQuotaState extends QuotaState { if (otherMap != null) { // To Remove - Set toRemove = new HashSet(map.keySet()); + Set toRemove = new HashSet<>(map.keySet()); toRemove.removeAll(otherMap.keySet()); map.keySet().removeAll(toRemove);