http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java index 98f39da..69b486d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java @@ -451,7 +451,7 @@ public class TableMapReduceUtil { job.setMapperClass(mapper); Configuration conf = job.getConfiguration(); HBaseConfiguration.merge(conf, HBaseConfiguration.create(conf)); - List<String> scanStrings = new ArrayList<String>(); + List<String> scanStrings = new ArrayList<>(); for (Scan scan : scans) { scanStrings.add(convertScanToString(scan)); @@ -807,7 +807,7 @@ public class TableMapReduceUtil { if (conf == null) { throw new IllegalArgumentException("Must provide a configuration object."); } - Set<String> paths = new HashSet<String>(conf.getStringCollection("tmpjars")); + Set<String> paths = new HashSet<>(conf.getStringCollection("tmpjars")); if (paths.isEmpty()) { throw new IllegalArgumentException("Configuration contains no tmpjars."); } @@ -879,13 +879,13 @@ public class TableMapReduceUtil { Class<?>... classes) throws IOException { FileSystem localFs = FileSystem.getLocal(conf); - Set<String> jars = new HashSet<String>(); + Set<String> jars = new HashSet<>(); // Add jars that are already in the tmpjars variable jars.addAll(conf.getStringCollection("tmpjars")); // add jars as we find them to a map of contents jar name so that we can avoid // creating new jars for classes that have already been packaged. - Map<String, String> packagedClasses = new HashMap<String, String>(); + Map<String, String> packagedClasses = new HashMap<>(); // Add jars containing the specified classes for (Class<?> clazz : classes) {
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.java index c40396f..b2db319 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormat.java @@ -183,7 +183,7 @@ public class TableSnapshotInputFormat extends InputFormat<ImmutableBytesWritable @Override public List<InputSplit> getSplits(JobContext job) throws IOException, InterruptedException { - List<InputSplit> results = new ArrayList<InputSplit>(); + List<InputSplit> results = new ArrayList<>(); for (TableSnapshotInputFormatImpl.InputSplit split : TableSnapshotInputFormatImpl.getSplits(job.getConfiguration())) { results.add(new TableSnapshotRegionSplit(split)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java index d52703a..69beef8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSnapshotInputFormatImpl.java @@ -311,7 +311,7 @@ public class TableSnapshotInputFormatImpl { Path tableDir = FSUtils.getTableDir(restoreDir, htd.getTableName()); - List<InputSplit> splits = new ArrayList<InputSplit>(); + List<InputSplit> splits = new ArrayList<>(); for (HRegionInfo hri : regionManifests) { // load region descriptor @@ -346,7 +346,7 @@ public class TableSnapshotInputFormatImpl { */ public static List<String> getBestLocations( Configuration conf, HDFSBlocksDistribution blockDistribution) { - List<String> locations = new ArrayList<String>(3); + List<String> locations = new ArrayList<>(3); HostAndWeight[] hostAndWeights = blockDistribution.getTopHostsWithWeights(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TextSortReducer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TextSortReducer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TextSortReducer.java index 1e09f03..05a4820 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TextSortReducer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TextSortReducer.java @@ -146,7 +146,7 @@ public class TextSortReducer extends "reducer.row.threshold", 1L * (1<<30)); Iterator<Text> iter = lines.iterator(); while (iter.hasNext()) { - Set<KeyValue> kvs = new TreeSet<KeyValue>(CellComparator.COMPARATOR); + Set<KeyValue> kvs = new TreeSet<>(CellComparator.COMPARATOR); long curSize = 0; // stop at the end or the RAM threshold while (iter.hasNext() && curSize < threshold) { @@ -160,7 +160,7 @@ public class TextSortReducer extends ttl = parsed.getCellTTL(); // create tags for the parsed line - List<Tag> tags = new ArrayList<Tag>(); + List<Tag> tags = new ArrayList<>(); if (cellVisibilityExpr != null) { tags.addAll(kvCreator.getVisibilityExpressionResolver().createVisibilityExpTags( cellVisibilityExpr)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java index 94bcb43..08b5aab 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java @@ -113,7 +113,7 @@ extends Mapper<LongWritable, Text, ImmutableBytesWritable, Put> throw new RuntimeException("No row key column specified"); } this.kvCreator = new CellCreator(conf); - tags = new ArrayList<Tag>(); + tags = new ArrayList<>(); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALInputFormat.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALInputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALInputFormat.java index 02fcbba..8514ace 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALInputFormat.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALInputFormat.java @@ -239,7 +239,7 @@ public class WALInputFormat extends InputFormat<WALKey, WALEdit> { FileSystem fs = inputDir.getFileSystem(conf); List<FileStatus> files = getFiles(fs, inputDir, startTime, endTime); - List<InputSplit> splits = new ArrayList<InputSplit>(files.size()); + List<InputSplit> splits = new ArrayList<>(files.size()); for (FileStatus file : files) { splits.add(new WALSplit(file.getPath().toString(), file.getLen(), startTime, endTime)); } @@ -248,7 +248,7 @@ public class WALInputFormat extends InputFormat<WALKey, WALEdit> { private List<FileStatus> getFiles(FileSystem fs, Path dir, long startTime, long endTime) throws IOException { - List<FileStatus> result = new ArrayList<FileStatus>(); + List<FileStatus> result = new ArrayList<>(); LOG.debug("Scanning " + dir.toString() + " for WAL files"); FileStatus[] files = fs.listStatus(dir); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java index 2c67baf..cca2041 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/WALPlayer.java @@ -132,7 +132,7 @@ public class WALPlayer extends Configured implements Tool { */ protected static class WALMapper extends Mapper<WALKey, WALEdit, ImmutableBytesWritable, Mutation> { - private Map<TableName, TableName> tables = new TreeMap<TableName, TableName>(); + private Map<TableName, TableName> tables = new TreeMap<>(); @Override public void map(WALKey key, WALEdit value, Context context) http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java index 60ad545..69ebd97 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java @@ -115,7 +115,7 @@ public class AssignmentManager { private AtomicInteger numRegionsOpened = new AtomicInteger(0); - final private KeyLocker<String> locker = new KeyLocker<String>(); + final private KeyLocker<String> locker = new KeyLocker<>(); Set<HRegionInfo> replicasToClose = Collections.synchronizedSet(new HashSet<HRegionInfo>()); @@ -141,8 +141,7 @@ public class AssignmentManager { // TODO: When do plans get cleaned out? Ever? In server open and in server // shutdown processing -- St.Ack // All access to this Map must be synchronized. - final NavigableMap<String, RegionPlan> regionPlans = - new TreeMap<String, RegionPlan>(); + final NavigableMap<String, RegionPlan> regionPlans = new TreeMap<>(); private final TableStateManager tableStateManager; @@ -183,8 +182,7 @@ public class AssignmentManager { * because we don't expect this to happen frequently; we don't * want to copy this information over during each state transition either. */ - private final ConcurrentHashMap<String, AtomicInteger> - failedOpenTracker = new ConcurrentHashMap<String, AtomicInteger>(); + private final ConcurrentHashMap<String, AtomicInteger> failedOpenTracker = new ConcurrentHashMap<>(); // In case not using ZK for region assignment, region states // are persisted in meta with a state store @@ -197,7 +195,7 @@ public class AssignmentManager { public static boolean TEST_SKIP_SPLIT_HANDLING = false; /** Listeners that are called on assignment events. */ - private List<AssignmentListener> listeners = new CopyOnWriteArrayList<AssignmentListener>(); + private List<AssignmentListener> listeners = new CopyOnWriteArrayList<>(); private RegionStateListener regionStateListener; @@ -382,7 +380,7 @@ public class AssignmentManager { pending++; } } - return new Pair<Integer, Integer>(pending, hris.size()); + return new Pair<>(pending, hris.size()); } /** @@ -748,16 +746,16 @@ public class AssignmentManager { return true; } LOG.info("Assigning " + regionCount + " region(s) to " + destination.toString()); - Set<String> encodedNames = new HashSet<String>(regionCount); + Set<String> encodedNames = new HashSet<>(regionCount); for (HRegionInfo region : regions) { encodedNames.add(region.getEncodedName()); } - List<HRegionInfo> failedToOpenRegions = new ArrayList<HRegionInfo>(); + List<HRegionInfo> failedToOpenRegions = new ArrayList<>(); Map<String, Lock> locks = locker.acquireLocks(encodedNames); try { - Map<String, RegionPlan> plans = new HashMap<String, RegionPlan>(regionCount); - List<RegionState> states = new ArrayList<RegionState>(regionCount); + Map<String, RegionPlan> plans = new HashMap<>(regionCount); + List<RegionState> states = new ArrayList<>(regionCount); for (HRegionInfo region : regions) { String encodedName = region.getEncodedName(); if (!isDisabledorDisablingRegionInRIT(region)) { @@ -797,8 +795,7 @@ public class AssignmentManager { // that unnecessary timeout on RIT is reduced. this.addPlans(plans); - List<Pair<HRegionInfo, List<ServerName>>> regionOpenInfos = - new ArrayList<Pair<HRegionInfo, List<ServerName>>>(states.size()); + List<Pair<HRegionInfo, List<ServerName>>> regionOpenInfos = new ArrayList<>(states.size()); for (RegionState state: states) { HRegionInfo region = state.getRegion(); regionStates.updateRegionState( @@ -807,8 +804,7 @@ public class AssignmentManager { if (shouldAssignFavoredNodes(region)) { favoredNodes = server.getFavoredNodesManager().getFavoredNodesWithDNPort(region); } - regionOpenInfos.add(new Pair<HRegionInfo, List<ServerName>>( - region, favoredNodes)); + regionOpenInfos.add(new Pair<>(region, favoredNodes)); } // Move on to open regions. @@ -908,7 +904,7 @@ public class AssignmentManager { } // wait for assignment completion - ArrayList<HRegionInfo> userRegionSet = new ArrayList<HRegionInfo>(regions.size()); + ArrayList<HRegionInfo> userRegionSet = new ArrayList<>(regions.size()); for (HRegionInfo region: regions) { if (!region.getTable().isSystemTable()) { userRegionSet.add(region); @@ -1443,7 +1439,7 @@ public class AssignmentManager { */ public boolean waitForAssignment(HRegionInfo regionInfo) throws InterruptedException { - ArrayList<HRegionInfo> regionSet = new ArrayList<HRegionInfo>(1); + ArrayList<HRegionInfo> regionSet = new ArrayList<>(1); regionSet.add(regionInfo); return waitForAssignment(regionSet, true, Long.MAX_VALUE); } @@ -1588,7 +1584,7 @@ public class AssignmentManager { } // invoke assignment (async) - ArrayList<HRegionInfo> userRegionSet = new ArrayList<HRegionInfo>(regions); + ArrayList<HRegionInfo> userRegionSet = new ArrayList<>(regions); for (Map.Entry<ServerName, List<HRegionInfo>> plan: bulkPlan.entrySet()) { if (!assign(plan.getKey(), plan.getValue()) && !server.isStopped()) { for (HRegionInfo region: plan.getValue()) { @@ -1640,7 +1636,7 @@ public class AssignmentManager { if (retainAssignment) { assign(allRegions); } else { - List<HRegionInfo> regions = new ArrayList<HRegionInfo>(regionsFromMetaScan); + List<HRegionInfo> regions = new ArrayList<>(regionsFromMetaScan); assign(regions); } @@ -1687,7 +1683,7 @@ public class AssignmentManager { */ public static List<HRegionInfo> replicaRegionsNotRecordedInMeta( Set<HRegionInfo> regionsRecordedInMeta, MasterServices master)throws IOException { - List<HRegionInfo> regionsNotRecordedInMeta = new ArrayList<HRegionInfo>(); + List<HRegionInfo> regionsNotRecordedInMeta = new ArrayList<>(); for (HRegionInfo hri : regionsRecordedInMeta) { TableName table = hri.getTable(); if(master.getTableDescriptors().get(table) == null) @@ -1723,7 +1719,7 @@ public class AssignmentManager { // Get any new but slow to checkin region server that joined the cluster Set<ServerName> onlineServers = serverManager.getOnlineServers().keySet(); // Set of offline servers to be returned - Set<ServerName> offlineServers = new HashSet<ServerName>(); + Set<ServerName> offlineServers = new HashSet<>(); // Iterate regions in META for (Result result : results) { if (result == null && LOG.isDebugEnabled()){ @@ -2446,7 +2442,7 @@ public class AssignmentManager { threadPoolExecutorService.submit(splitReplicasCallable); // wait for assignment completion - ArrayList<HRegionInfo> regionAssignSet = new ArrayList<HRegionInfo>(2); + ArrayList<HRegionInfo> regionAssignSet = new ArrayList<>(2); regionAssignSet.add(daughterAHRI); regionAssignSet.add(daughterBHRI); while (!waitForAssignment(regionAssignSet, true, regionAssignSet.size(), @@ -2558,7 +2554,7 @@ public class AssignmentManager { final HRegionInfo a = HRegionInfo.convert(transition.getRegionInfo(1)); final HRegionInfo b = HRegionInfo.convert(transition.getRegionInfo(2)); - Set<String> encodedNames = new HashSet<String>(2); + Set<String> encodedNames = new HashSet<>(2); encodedNames.add(a.getEncodedName()); encodedNames.add(b.getEncodedName()); Map<String, Lock> locks = locker.acquireLocks(encodedNames); @@ -2645,7 +2641,7 @@ public class AssignmentManager { threadPoolExecutorService.submit(mergeReplicasCallable); // wait for assignment completion - ArrayList<HRegionInfo> regionAssignSet = new ArrayList<HRegionInfo>(1); + ArrayList<HRegionInfo> regionAssignSet = new ArrayList<>(1); regionAssignSet.add(mergedRegion); while (!waitForAssignment(regionAssignSet, true, regionAssignSet.size(), Long.MAX_VALUE)) { LOG.debug("The merged region " + mergedRegion + " is still in transition. "); @@ -2754,7 +2750,7 @@ public class AssignmentManager { final HRegionInfo hri_b) { // Close replicas for the original unmerged regions. create/assign new replicas // for the merged parent. - List<HRegionInfo> unmergedRegions = new ArrayList<HRegionInfo>(); + List<HRegionInfo> unmergedRegions = new ArrayList<>(); unmergedRegions.add(hri_a); unmergedRegions.add(hri_b); Map<ServerName, List<HRegionInfo>> map = regionStates.getRegionAssignments(unmergedRegions); @@ -2768,7 +2764,7 @@ public class AssignmentManager { } } int numReplicas = getNumReplicas(server, mergedHri.getTable()); - List<HRegionInfo> regions = new ArrayList<HRegionInfo>(); + List<HRegionInfo> regions = new ArrayList<>(); for (int i = 1; i < numReplicas; i++) { regions.add(RegionReplicaUtil.getRegionInfoForReplica(mergedHri, i)); } @@ -2790,7 +2786,7 @@ public class AssignmentManager { // the replica1s of daughters will be on the same machine int numReplicas = getNumReplicas(server, parentHri.getTable()); // unassign the old replicas - List<HRegionInfo> parentRegion = new ArrayList<HRegionInfo>(); + List<HRegionInfo> parentRegion = new ArrayList<>(); parentRegion.add(parentHri); Map<ServerName, List<HRegionInfo>> currentAssign = regionStates.getRegionAssignments(parentRegion); @@ -2804,7 +2800,7 @@ public class AssignmentManager { } } // assign daughter replicas - Map<HRegionInfo, ServerName> map = new HashMap<HRegionInfo, ServerName>(); + Map<HRegionInfo, ServerName> map = new HashMap<>(); for (int i = 1; i < numReplicas; i++) { prepareDaughterReplicaForAssignment(hri_a, parentHri, i, map); prepareDaughterReplicaForAssignment(hri_b, parentHri, i, map); @@ -2856,7 +2852,7 @@ public class AssignmentManager { sendRegionClosedNotification(regionInfo); // also note that all the replicas of the primary should be closed if (state != null && state.equals(State.SPLIT)) { - Collection<HRegionInfo> c = new ArrayList<HRegionInfo>(1); + Collection<HRegionInfo> c = new ArrayList<>(1); c.add(regionInfo); Map<ServerName, List<HRegionInfo>> map = regionStates.getRegionAssignments(c); Collection<List<HRegionInfo>> allReplicas = map.values(); @@ -2865,7 +2861,7 @@ public class AssignmentManager { } } else if (state != null && state.equals(State.MERGED)) { - Collection<HRegionInfo> c = new ArrayList<HRegionInfo>(1); + Collection<HRegionInfo> c = new ArrayList<>(1); c.add(regionInfo); Map<ServerName, List<HRegionInfo>> map = regionStates.getRegionAssignments(c); Collection<List<HRegionInfo>> allReplicas = map.values(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentVerificationReport.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentVerificationReport.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentVerificationReport.java index d290f26..e1922af 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentVerificationReport.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentVerificationReport.java @@ -51,16 +51,13 @@ public class AssignmentVerificationReport { private int totalRegions = 0; private int totalRegionServers = 0; // for unassigned regions - private List<HRegionInfo> unAssignedRegionsList = - new ArrayList<HRegionInfo>(); + private List<HRegionInfo> unAssignedRegionsList = new ArrayList<>(); // For regions without valid favored nodes - private List<HRegionInfo> regionsWithoutValidFavoredNodes = - new ArrayList<HRegionInfo>(); + private List<HRegionInfo> regionsWithoutValidFavoredNodes = new ArrayList<>(); // For regions not running on the favored nodes - private List<HRegionInfo> nonFavoredAssignedRegionList = - new ArrayList<HRegionInfo>(); + private List<HRegionInfo> nonFavoredAssignedRegionList = new ArrayList<>(); // For regions running on the favored nodes private int totalFavoredAssignments = 0; @@ -73,26 +70,20 @@ public class AssignmentVerificationReport { private float avgRegionsOnRS = 0; private int maxRegionsOnRS = 0; private int minRegionsOnRS = Integer.MAX_VALUE; - private Set<ServerName> mostLoadedRSSet = - new HashSet<ServerName>(); - private Set<ServerName> leastLoadedRSSet = - new HashSet<ServerName>(); + private Set<ServerName> mostLoadedRSSet = new HashSet<>(); + private Set<ServerName> leastLoadedRSSet = new HashSet<>(); private float avgDispersionScore = 0; private float maxDispersionScore = 0; - private Set<ServerName> maxDispersionScoreServerSet = - new HashSet<ServerName>(); + private Set<ServerName> maxDispersionScoreServerSet = new HashSet<>(); private float minDispersionScore = Float.MAX_VALUE; - private Set<ServerName> minDispersionScoreServerSet = - new HashSet<ServerName>(); + private Set<ServerName> minDispersionScoreServerSet = new HashSet<>(); private float avgDispersionNum = 0; private float maxDispersionNum = 0; - private Set<ServerName> maxDispersionNumServerSet = - new HashSet<ServerName>(); + private Set<ServerName> maxDispersionNumServerSet = new HashSet<>(); private float minDispersionNum = Float.MAX_VALUE; - private Set<ServerName> minDispersionNumServerSet = - new HashSet<ServerName>(); + private Set<ServerName> minDispersionNumServerSet = new HashSet<>(); public void fillUp(TableName tableName, SnapshotOfRegionAssignmentFromMeta snapshot, Map<String, Map<String, Float>> regionLocalityMap) { @@ -111,13 +102,10 @@ public class AssignmentVerificationReport { Map<HRegionInfo, ServerName> currentAssignment = snapshot.getRegionToRegionServerMap(); // Initialize the server to its hosing region counter map - Map<ServerName, Integer> serverToHostingRegionCounterMap = - new HashMap<ServerName, Integer>(); + Map<ServerName, Integer> serverToHostingRegionCounterMap = new HashMap<>(); - Map<ServerName, Integer> primaryRSToRegionCounterMap = - new HashMap<ServerName, Integer>(); - Map<ServerName, Set<ServerName>> primaryToSecTerRSMap = - new HashMap<ServerName, Set<ServerName>>(); + Map<ServerName, Integer> primaryRSToRegionCounterMap = new HashMap<>(); + Map<ServerName, Set<ServerName>> primaryToSecTerRSMap = new HashMap<>(); // Check the favored nodes and its locality information // Also keep tracker of the most loaded and least loaded region servers @@ -164,7 +152,7 @@ public class AssignmentVerificationReport { // Update the primary rs to secondary and tertiary rs map Set<ServerName> secAndTerSet = primaryToSecTerRSMap.get(primaryRS); if (secAndTerSet == null) { - secAndTerSet = new HashSet<ServerName>(); + secAndTerSet = new HashSet<>(); } secAndTerSet.add(secondaryRS); secAndTerSet.add(tertiaryRS); @@ -340,10 +328,8 @@ public class AssignmentVerificationReport { plan = newPlan; } // Get the region to region server mapping - Map<ServerName, Integer> primaryRSToRegionCounterMap = - new HashMap<ServerName, Integer>(); - Map<ServerName, Set<ServerName>> primaryToSecTerRSMap = - new HashMap<ServerName, Set<ServerName>>(); + Map<ServerName, Integer> primaryRSToRegionCounterMap = new HashMap<>(); + Map<ServerName, Set<ServerName>> primaryToSecTerRSMap = new HashMap<>(); // Check the favored nodes and its locality information // Also keep tracker of the most loaded and least loaded region servers @@ -375,7 +361,7 @@ public class AssignmentVerificationReport { // Update the primary rs to secondary and tertiary rs map Set<ServerName> secAndTerSet = primaryToSecTerRSMap.get(primaryRS); if (secAndTerSet == null) { - secAndTerSet = new HashSet<ServerName>(); + secAndTerSet = new HashSet<>(); } secAndTerSet.add(secondaryRS); secAndTerSet.add(tertiaryRS); @@ -451,7 +437,7 @@ public class AssignmentVerificationReport { * */ public List<Float> getDispersionInformation() { - List<Float> dispersion = new ArrayList<Float>(); + List<Float> dispersion = new ArrayList<>(); dispersion.add(avgDispersionScore); dispersion.add(maxDispersionScore); dispersion.add(minDispersionScore); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java index 606dce4..d8c511e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java @@ -59,7 +59,7 @@ public class BulkReOpen extends BulkAssigner { .entrySet()) { final List<HRegionInfo> hris = e.getValue(); // add plans for the regions that need to be reopened - Map<String, RegionPlan> plans = new HashMap<String, RegionPlan>(); + Map<String, RegionPlan> plans = new HashMap<>(); for (HRegionInfo hri : hris) { RegionPlan reOpenPlan = assignmentManager.getRegionReopenPlan(hri); plans.put(hri.getEncodedName(), reOpenPlan); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java index ef042af..affd44c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java @@ -151,9 +151,8 @@ public class CatalogJanitor extends ScheduledChore { final AtomicInteger count = new AtomicInteger(0); // Keep Map of found split parents. There are candidates for cleanup. // Use a comparator that has split parents come before its daughters. - final Map<HRegionInfo, Result> splitParents = - new TreeMap<HRegionInfo, Result>(new SplitParentFirstComparator()); - final Map<HRegionInfo, Result> mergedRegions = new TreeMap<HRegionInfo, Result>(); + final Map<HRegionInfo, Result> splitParents = new TreeMap<>(new SplitParentFirstComparator()); + final Map<HRegionInfo, Result> mergedRegions = new TreeMap<>(); // This visitor collects split parents and counts rows in the hbase:meta table MetaTableAccessor.Visitor visitor = new MetaTableAccessor.Visitor() { @@ -181,8 +180,7 @@ public class CatalogJanitor extends ScheduledChore { // the start row MetaTableAccessor.scanMetaForTableRegions(this.connection, visitor, tableName); - return new Triple<Integer, Map<HRegionInfo, Result>, Map<HRegionInfo, Result>>( - count.get(), mergedRegions, splitParents); + return new Triple<>(count.get(), mergedRegions, splitParents); } /** @@ -275,7 +273,7 @@ public class CatalogJanitor extends ScheduledChore { // Now work on our list of found parents. See if any we can clean up. int splitCleaned = 0; // regions whose parents are still around - HashSet<String> parentNotCleaned = new HashSet<String>(); + HashSet<String> parentNotCleaned = new HashSet<>(); for (Map.Entry<HRegionInfo, Result> e : splitParents.entrySet()) { if (this.services.isInMaintenanceMode()) { // Stop cleaning if the master is in maintenance mode @@ -398,7 +396,7 @@ public class CatalogJanitor extends ScheduledChore { Pair<Boolean, Boolean> checkDaughterInFs(final HRegionInfo parent, final HRegionInfo daughter) throws IOException { if (daughter == null) { - return new Pair<Boolean, Boolean>(Boolean.FALSE, Boolean.FALSE); + return new Pair<>(Boolean.FALSE, Boolean.FALSE); } FileSystem fs = this.services.getMasterFileSystem().getFileSystem(); @@ -411,12 +409,12 @@ public class CatalogJanitor extends ScheduledChore { try { if (!FSUtils.isExists(fs, daughterRegionDir)) { - return new Pair<Boolean, Boolean>(Boolean.FALSE, Boolean.FALSE); + return new Pair<>(Boolean.FALSE, Boolean.FALSE); } } catch (IOException ioe) { LOG.error("Error trying to determine if daughter region exists, " + "assuming exists and has references", ioe); - return new Pair<Boolean, Boolean>(Boolean.TRUE, Boolean.TRUE); + return new Pair<>(Boolean.TRUE, Boolean.TRUE); } boolean references = false; @@ -433,9 +431,9 @@ public class CatalogJanitor extends ScheduledChore { } catch (IOException e) { LOG.error("Error trying to determine referenced files from : " + daughter.getEncodedName() + ", to: " + parent.getEncodedName() + " assuming has references", e); - return new Pair<Boolean, Boolean>(Boolean.TRUE, Boolean.TRUE); + return new Pair<>(Boolean.TRUE, Boolean.TRUE); } - return new Pair<Boolean, Boolean>(Boolean.TRUE, Boolean.valueOf(references)); + return new Pair<>(Boolean.TRUE, Boolean.valueOf(references)); } private HTableDescriptor getTableDescriptor(final TableName tableName) http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterSchemaServiceImpl.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterSchemaServiceImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterSchemaServiceImpl.java index 52af89e..bf3ae7e 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterSchemaServiceImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterSchemaServiceImpl.java @@ -122,7 +122,7 @@ class ClusterSchemaServiceImpl implements ClusterSchemaService { checkIsRunning(); Set<NamespaceDescriptor> set = getTableNamespaceManager().list(); if (set == null || set.isEmpty()) return EMPTY_NAMESPACE_LIST; - List<NamespaceDescriptor> list = new ArrayList<NamespaceDescriptor>(set.size()); + List<NamespaceDescriptor> list = new ArrayList<>(set.size()); list.addAll(set); return Collections.unmodifiableList(list); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java index 3b19ada..ea5516d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ClusterStatusPublisher.java @@ -97,8 +97,7 @@ public class ClusterStatusPublisher extends ScheduledChore { private long lastMessageTime = 0; private final HMaster master; private final int messagePeriod; // time between two message - private final ConcurrentMap<ServerName, Integer> lastSent = - new ConcurrentHashMap<ServerName, Integer>(); + private final ConcurrentMap<ServerName, Integer> lastSent = new ConcurrentHashMap<>(); private Publisher publisher; private boolean connected = false; @@ -194,7 +193,7 @@ public class ClusterStatusPublisher extends ScheduledChore { } // We're sending the new deads first. - List<Map.Entry<ServerName, Integer>> entries = new ArrayList<Map.Entry<ServerName, Integer>>(); + List<Map.Entry<ServerName, Integer>> entries = new ArrayList<>(); entries.addAll(lastSent.entrySet()); Collections.sort(entries, new Comparator<Map.Entry<ServerName, Integer>>() { @Override @@ -205,7 +204,7 @@ public class ClusterStatusPublisher extends ScheduledChore { // With a limit of MAX_SERVER_PER_MESSAGE int max = entries.size() > MAX_SERVER_PER_MESSAGE ? MAX_SERVER_PER_MESSAGE : entries.size(); - List<ServerName> res = new ArrayList<ServerName>(max); + List<ServerName> res = new ArrayList<>(max); for (int i = 0; i < max; i++) { Map.Entry<ServerName, Integer> toSend = entries.get(i); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java index c33cdcc..faceba2 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java @@ -51,7 +51,7 @@ public class DeadServer { * and it's server logs are recovered, it will be told to call server startup * because by then, its regions have probably been reassigned. */ - private final Map<ServerName, Long> deadServers = new HashMap<ServerName, Long>(); + private final Map<ServerName, Long> deadServers = new HashMap<>(); /** * Number of dead servers currently being processed @@ -102,7 +102,7 @@ public class DeadServer { public synchronized boolean areDeadServersInProgress() { return processing; } public synchronized Set<ServerName> copyServerNames() { - Set<ServerName> clone = new HashSet<ServerName>(deadServers.size()); + Set<ServerName> clone = new HashSet<>(deadServers.size()); clone.addAll(deadServers.keySet()); return clone; } @@ -177,11 +177,11 @@ public class DeadServer { * @return a sorted array list, by death time, lowest values first. */ public synchronized List<Pair<ServerName, Long>> copyDeadServersSince(long ts){ - List<Pair<ServerName, Long>> res = new ArrayList<Pair<ServerName, Long>>(size()); + List<Pair<ServerName, Long>> res = new ArrayList<>(size()); for (Map.Entry<ServerName, Long> entry:deadServers.entrySet()){ if (entry.getValue() >= ts){ - res.add(new Pair<ServerName, Long>(entry.getKey(), entry.getValue())); + res.add(new Pair<>(entry.getKey(), entry.getValue())); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/GeneralBulkAssigner.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/GeneralBulkAssigner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/GeneralBulkAssigner.java index 43ea523..fc3607f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/GeneralBulkAssigner.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/GeneralBulkAssigner.java @@ -43,8 +43,7 @@ import org.apache.hadoop.hbase.ServerName; public class GeneralBulkAssigner extends BulkAssigner { private static final Log LOG = LogFactory.getLog(GeneralBulkAssigner.class); - private Map<ServerName, List<HRegionInfo>> failedPlans - = new ConcurrentHashMap<ServerName, List<HRegionInfo>>(); + private Map<ServerName, List<HRegionInfo>> failedPlans = new ConcurrentHashMap<>(); private ExecutorService pool; final Map<ServerName, List<HRegionInfo>> bulkPlan; @@ -82,7 +81,7 @@ public class GeneralBulkAssigner extends BulkAssigner { @Override protected boolean waitUntilDone(final long timeout) throws InterruptedException { - Set<HRegionInfo> regionSet = new HashSet<HRegionInfo>(); + Set<HRegionInfo> regionSet = new HashSet<>(); for (List<HRegionInfo> regionList : bulkPlan.values()) { regionSet.addAll(regionList); } @@ -164,7 +163,7 @@ public class GeneralBulkAssigner extends BulkAssigner { } private int reassignFailedPlans() { - List<HRegionInfo> reassigningRegions = new ArrayList<HRegionInfo>(); + List<HRegionInfo> reassigningRegions = new ArrayList<>(); for (Map.Entry<ServerName, List<HRegionInfo>> e : failedPlans.entrySet()) { LOG.info("Failed assigning " + e.getValue().size() + " regions to server " + e.getKey() + ", reassigning them"); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java index 78f1783..501d3bd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java @@ -1063,7 +1063,7 @@ public class HMaster extends HRegionServer implements MasterServices { //start the hfile archive cleaner thread Path archiveDir = HFileArchiveUtil.getArchivePath(conf); - Map<String, Object> params = new HashMap<String, Object>(); + Map<String, Object> params = new HashMap<>(); params.put(MASTER, this); this.hfileCleaner = new HFileCleaner(cleanerInterval, this, conf, getMasterFileSystem() .getFileSystem(), archiveDir, params); @@ -1327,7 +1327,7 @@ public class HMaster extends HRegionServer implements MasterServices { Map<TableName, Map<ServerName, List<HRegionInfo>>> assignmentsByTable = this.assignmentManager.getRegionStates().getAssignmentsByTable(); - List<RegionPlan> plans = new ArrayList<RegionPlan>(); + List<RegionPlan> plans = new ArrayList<>(); //Give the balancer the current cluster state. this.balancer.setClusterStatus(getClusterStatus()); @@ -2235,8 +2235,7 @@ public class HMaster extends HRegionServer implements MasterServices { Pair<HRegionInfo, ServerName> getTableRegionForRow( final TableName tableName, final byte [] rowKey) throws IOException { - final AtomicReference<Pair<HRegionInfo, ServerName>> result = - new AtomicReference<Pair<HRegionInfo, ServerName>>(null); + final AtomicReference<Pair<HRegionInfo, ServerName>> result = new AtomicReference<>(null); MetaTableAccessor.Visitor visitor = new MetaTableAccessor.Visitor() { @Override @@ -2345,7 +2344,7 @@ public class HMaster extends HRegionServer implements MasterServices { List<ServerName> backupMasters = null; if (backupMasterStrings != null && !backupMasterStrings.isEmpty()) { - backupMasters = new ArrayList<ServerName>(backupMasterStrings.size()); + backupMasters = new ArrayList<>(backupMasterStrings.size()); for (String s: backupMasterStrings) { try { byte [] bytes; @@ -2852,7 +2851,7 @@ public class HMaster extends HRegionServer implements MasterServices { */ List<NamespaceDescriptor> getNamespaces() throws IOException { checkInitialized(); - final List<NamespaceDescriptor> nsds = new ArrayList<NamespaceDescriptor>(); + final List<NamespaceDescriptor> nsds = new ArrayList<>(); boolean bypass = false; if (cpHost != null) { bypass = cpHost.preListNamespaceDescriptors(nsds); @@ -2918,7 +2917,7 @@ public class HMaster extends HRegionServer implements MasterServices { public List<HTableDescriptor> listTableDescriptors(final String namespace, final String regex, final List<TableName> tableNameList, final boolean includeSysTables) throws IOException { - List<HTableDescriptor> htds = new ArrayList<HTableDescriptor>(); + List<HTableDescriptor> htds = new ArrayList<>(); boolean bypass = cpHost != null? cpHost.preGetTableDescriptors(tableNameList, htds, regex): false; if (!bypass) { @@ -2939,13 +2938,13 @@ public class HMaster extends HRegionServer implements MasterServices { */ public List<TableName> listTableNames(final String namespace, final String regex, final boolean includeSysTables) throws IOException { - List<HTableDescriptor> htds = new ArrayList<HTableDescriptor>(); + List<HTableDescriptor> htds = new ArrayList<>(); boolean bypass = cpHost != null? cpHost.preGetTableNames(htds, regex): false; if (!bypass) { htds = getTableDescriptors(htds, namespace, regex, null, includeSysTables); if (cpHost != null) cpHost.postGetTableNames(htds, regex); } - List<TableName> result = new ArrayList<TableName>(htds.size()); + List<TableName> result = new ArrayList<>(htds.size()); for (HTableDescriptor htd: htds) result.add(htd.getTableName()); return result; } @@ -3262,7 +3261,7 @@ public class HMaster extends HRegionServer implements MasterServices { @Override public List<ServerName> listDrainingRegionServers() { String parentZnode = getZooKeeper().znodePaths.drainingZNode; - List<ServerName> serverNames = new ArrayList<ServerName>(); + List<ServerName> serverNames = new ArrayList<>(); List<String> serverStrs = null; try { serverStrs = ZKUtil.listChildrenNoWatch(getZooKeeper(), parentZnode); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMetaBootstrap.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMetaBootstrap.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMetaBootstrap.java index 5e1917b..1988e2d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMetaBootstrap.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMetaBootstrap.java @@ -100,7 +100,7 @@ public class MasterMetaBootstrap { throws IOException, InterruptedException, KeeperException { int numReplicas = master.getConfiguration().getInt(HConstants.META_REPLICAS_NUM, HConstants.DEFAULT_META_REPLICA_NUM); - final Set<ServerName> EMPTY_SET = new HashSet<ServerName>(); + final Set<ServerName> EMPTY_SET = new HashSet<>(); for (int i = 1; i < numReplicas; i++) { assignMeta(EMPTY_SET, i); } @@ -241,7 +241,7 @@ public class MasterMetaBootstrap { */ private Set<ServerName> getPreviouselyFailedMetaServersFromZK() throws KeeperException { final ZooKeeperWatcher zooKeeper = master.getZooKeeper(); - Set<ServerName> result = new HashSet<ServerName>(); + Set<ServerName> result = new HashSet<>(); String metaRecoveringZNode = ZKUtil.joinZNode(zooKeeper.znodePaths.recoveringRegionsZNode, HRegionInfo.FIRST_META_REGIONINFO.getEncodedName()); List<String> regionFailedServers = ZKUtil.listChildrenNoWatch(zooKeeper, metaRecoveringZNode); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMobCompactionThread.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMobCompactionThread.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMobCompactionThread.java index fc0ecfb..2b1232a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMobCompactionThread.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterMobCompactionThread.java @@ -56,7 +56,7 @@ public class MasterMobCompactionThread { final String n = Thread.currentThread().getName(); // this pool is used to run the mob compaction this.masterMobPool = new ThreadPoolExecutor(1, 2, 60, TimeUnit.SECONDS, - new SynchronousQueue<Runnable>(), new ThreadFactory() { + new SynchronousQueue<>(), new ThreadFactory() { @Override public Thread newThread(Runnable r) { String name = n + "-MasterMobCompaction-" + EnvironmentEdgeManager.currentTime(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java index 3beda05..177ee32 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java @@ -234,7 +234,7 @@ public class MasterRpcServices extends RSRpcServices * @return list of blocking services and their security info classes that this server supports */ protected List<BlockingServiceAndInterface> getServices() { - List<BlockingServiceAndInterface> bssi = new ArrayList<BlockingServiceAndInterface>(5); + List<BlockingServiceAndInterface> bssi = new ArrayList<>(5); bssi.add(new BlockingServiceAndInterface( MasterService.newReflectiveBlockingService(this), MasterService.BlockingInterface.class)); @@ -1333,7 +1333,7 @@ public class MasterRpcServices extends RSRpcServices Pair<HRegionInfo, ServerName> pair = MetaTableAccessor.getRegion(master.getConnection(), regionName); if (Bytes.equals(HRegionInfo.FIRST_META_REGIONINFO.getRegionName(),regionName)) { - pair = new Pair<HRegionInfo, ServerName>(HRegionInfo.FIRST_META_REGIONINFO, + pair = new Pair<>(HRegionInfo.FIRST_META_REGIONINFO, master.getMetaTableLocator().getMetaRegionLocation(master.getZooKeeper())); } if (pair == null) { @@ -1491,7 +1491,7 @@ public class MasterRpcServices extends RSRpcServices throw new DoNotRetryIOException("Table " + tableName + " is not enabled"); } boolean allFiles = false; - List<HColumnDescriptor> compactedColumns = new ArrayList<HColumnDescriptor>(); + List<HColumnDescriptor> compactedColumns = new ArrayList<>(); HColumnDescriptor[] hcds = master.getTableDescriptors().get(tableName).getColumnFamilies(); byte[] family = null; if (request.hasFamily()) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java index 1f9729c..27aca94 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterWalManager.java @@ -155,7 +155,7 @@ public class MasterWalManager { boolean retrySplitting = !conf.getBoolean("hbase.hlog.split.skip.errors", WALSplitter.SPLIT_SKIP_ERRORS_DEFAULT); - Set<ServerName> serverNames = new HashSet<ServerName>(); + Set<ServerName> serverNames = new HashSet<>(); Path logsDirPath = new Path(this.rootDir, HConstants.HREGION_LOGDIR_NAME); do { @@ -218,7 +218,7 @@ public class MasterWalManager { } public void splitLog(final ServerName serverName) throws IOException { - Set<ServerName> serverNames = new HashSet<ServerName>(); + Set<ServerName> serverNames = new HashSet<>(); serverNames.add(serverName); splitLog(serverNames); } @@ -228,7 +228,7 @@ public class MasterWalManager { * @param serverName logs belonging to this server will be split */ public void splitMetaLog(final ServerName serverName) throws IOException { - Set<ServerName> serverNames = new HashSet<ServerName>(); + Set<ServerName> serverNames = new HashSet<>(); serverNames.add(serverName); splitMetaLog(serverNames); } @@ -245,7 +245,7 @@ public class MasterWalManager { "We only release this lock when we set it. Updates to code that uses it should verify use " + "of the guard boolean.") private List<Path> getLogDirs(final Set<ServerName> serverNames) throws IOException { - List<Path> logDirs = new ArrayList<Path>(); + List<Path> logDirs = new ArrayList<>(); boolean needReleaseLock = false; if (!this.services.isInitialized()) { // during master initialization, we could have multiple places splitting a same wal http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RackManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RackManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RackManager.java index 2b1fb50..5c06857 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RackManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RackManager.java @@ -80,7 +80,7 @@ public class RackManager { public List<String> getRack(List<ServerName> servers) { // just a note - switchMapping caches results (at least the implementation should unless the // resolution is really a lightweight process) - List<String> serversAsString = new ArrayList<String>(servers.size()); + List<String> serversAsString = new ArrayList<>(servers.size()); for (ServerName server : servers) { serversAsString.add(server.getHostname()); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java index 7acf9df..ffc3e5b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java @@ -106,7 +106,7 @@ public class RegionPlacementMaintainer { this.conf = conf; this.enforceLocality = enforceLocality; this.enforceMinAssignmentMove = enforceMinAssignmentMove; - this.targetTableSet = new HashSet<TableName>(); + this.targetTableSet = new HashSet<>(); this.rackManager = new RackManager(conf); try { this.connection = ConnectionFactory.createConnection(this.conf); @@ -163,7 +163,7 @@ public class RegionPlacementMaintainer { if (this.enforceLocality == true) { regionLocalityMap = FSUtils.getRegionDegreeLocalityMappingFromFS(conf); } - List<AssignmentVerificationReport> reports = new ArrayList<AssignmentVerificationReport>(); + List<AssignmentVerificationReport> reports = new ArrayList<>(); // Iterate all the tables to fill up the verification report for (TableName table : tables) { if (!this.targetTableSet.isEmpty() && @@ -204,7 +204,7 @@ public class RegionPlacementMaintainer { assignmentSnapshot.getRegionToRegionServerMap(); // Get the all the region servers - List<ServerName> servers = new ArrayList<ServerName>(); + List<ServerName> servers = new ArrayList<>(); try (Admin admin = this.connection.getAdmin()) { servers.addAll(admin.getClusterStatus().getServers()); } @@ -255,15 +255,14 @@ public class RegionPlacementMaintainer { // Compute the total rack locality for each region in each rack. The total // rack locality is the sum of the localities of a region on all servers in // a rack. - Map<String, Map<HRegionInfo, Float>> rackRegionLocality = - new HashMap<String, Map<HRegionInfo, Float>>(); + Map<String, Map<HRegionInfo, Float>> rackRegionLocality = new HashMap<>(); for (int i = 0; i < numRegions; i++) { HRegionInfo region = regions.get(i); for (int j = 0; j < regionSlots; j += slotsPerServer) { String rack = rackManager.getRack(servers.get(j / slotsPerServer)); Map<HRegionInfo, Float> rackLocality = rackRegionLocality.get(rack); if (rackLocality == null) { - rackLocality = new HashMap<HRegionInfo, Float>(); + rackLocality = new HashMap<>(); rackRegionLocality.put(rack, rackLocality); } Float localityObj = rackLocality.get(region); @@ -395,8 +394,7 @@ public class RegionPlacementMaintainer { tertiaryAssignment = randomizedMatrix.invertIndices(tertiaryAssignment); for (int i = 0; i < numRegions; i++) { - List<ServerName> favoredServers = - new ArrayList<ServerName>(FavoredNodeAssignmentHelper.FAVORED_NODES_NUM); + List<ServerName> favoredServers = new ArrayList<>(FavoredNodeAssignmentHelper.FAVORED_NODES_NUM); ServerName s = servers.get(primaryAssignment[i] / slotsPerServer); favoredServers.add(ServerName.valueOf(s.getHostname(), s.getPort(), ServerName.NON_STARTCODE)); @@ -417,7 +415,7 @@ public class RegionPlacementMaintainer { LOG.info("Assignment plan for secondary and tertiary generated " + "using MunkresAssignment"); } else { - Map<HRegionInfo, ServerName> primaryRSMap = new HashMap<HRegionInfo, ServerName>(); + Map<HRegionInfo, ServerName> primaryRSMap = new HashMap<>(); for (int i = 0; i < numRegions; i++) { primaryRSMap.put(regions.get(i), servers.get(primaryAssignment[i] / slotsPerServer)); } @@ -427,8 +425,7 @@ public class RegionPlacementMaintainer { Map<HRegionInfo, ServerName[]> secondaryAndTertiaryMap = favoredNodeHelper.placeSecondaryAndTertiaryWithRestrictions(primaryRSMap); for (int i = 0; i < numRegions; i++) { - List<ServerName> favoredServers = - new ArrayList<ServerName>(FavoredNodeAssignmentHelper.FAVORED_NODES_NUM); + List<ServerName> favoredServers = new ArrayList<>(FavoredNodeAssignmentHelper.FAVORED_NODES_NUM); HRegionInfo currentRegion = regions.get(i); ServerName s = primaryRSMap.get(currentRegion); favoredServers.add(ServerName.valueOf(s.getHostname(), s.getPort(), @@ -614,8 +611,7 @@ public class RegionPlacementMaintainer { if (plan == null) return; LOG.info("========== Start to print the assignment plan ================"); // sort the map based on region info - Map<String, List<ServerName>> assignmentMap = - new TreeMap<String, List<ServerName>>(plan.getAssignmentMap()); + Map<String, List<ServerName>> assignmentMap = new TreeMap<>(plan.getAssignmentMap()); for (Map.Entry<String, List<ServerName>> entry : assignmentMap.entrySet()) { @@ -666,13 +662,11 @@ public class RegionPlacementMaintainer { // track of the failed and succeeded updates int succeededNum = 0; - Map<ServerName, Exception> failedUpdateMap = - new HashMap<ServerName, Exception>(); + Map<ServerName, Exception> failedUpdateMap = new HashMap<>(); for (Map.Entry<ServerName, List<HRegionInfo>> entry : currentAssignment.entrySet()) { - List<Pair<HRegionInfo, List<ServerName>>> regionUpdateInfos = - new ArrayList<Pair<HRegionInfo, List<ServerName>>>(); + List<Pair<HRegionInfo, List<ServerName>>> regionUpdateInfos = new ArrayList<>(); try { // Keep track of the favored updates for the current region server FavoredNodesPlan singleServerPlan = null; @@ -687,8 +681,7 @@ public class RegionPlacementMaintainer { } // Update the single server update singleServerPlan.updateFavoredNodesMap(region, favoredServerList); - regionUpdateInfos.add( - new Pair<HRegionInfo, List<ServerName>>(region, favoredServerList)); + regionUpdateInfos.add(new Pair<>(region, favoredServerList)); } } if (singleServerPlan != null) { @@ -749,7 +742,7 @@ public class RegionPlacementMaintainer { */ public Map<TableName, Integer> getRegionsMovement(FavoredNodesPlan newPlan) throws IOException { - Map<TableName, Integer> movesPerTable = new HashMap<TableName, Integer>(); + Map<TableName, Integer> movesPerTable = new HashMap<>(); SnapshotOfRegionAssignmentFromMeta snapshot = this.getRegionAssignmentSnapshot(); Map<TableName, List<HRegionInfo>> tableToRegions = snapshot .getTableToRegionMap(); @@ -944,7 +937,7 @@ public class RegionPlacementMaintainer { if (favoredNodesArray == null) return null; - List<ServerName> serverList = new ArrayList<ServerName>(); + List<ServerName> serverList = new ArrayList<>(); for (String hostNameAndPort : favoredNodesArray) { serverList.add(ServerName.valueOf(hostNameAndPort, ServerName.NON_STARTCODE)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionStates.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionStates.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionStates.java index 4125eea..a1e24f2 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionStates.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionStates.java @@ -80,41 +80,35 @@ public class RegionStates { /** * Regions currently in transition. */ - final HashMap<String, RegionState> regionsInTransition = - new HashMap<String, RegionState>(); + final HashMap<String, RegionState> regionsInTransition = new HashMap<>(); /** * Region encoded name to state map. * All the regions should be in this map. */ - private final Map<String, RegionState> regionStates = - new HashMap<String, RegionState>(); + private final Map<String, RegionState> regionStates = new HashMap<>(); /** * Holds mapping of table -> region state */ - private final Map<TableName, Map<String, RegionState>> regionStatesTableIndex = - new HashMap<TableName, Map<String, RegionState>>(); + private final Map<TableName, Map<String, RegionState>> regionStatesTableIndex = new HashMap<>(); /** * Server to regions assignment map. * Contains the set of regions currently assigned to a given server. */ - private final Map<ServerName, Set<HRegionInfo>> serverHoldings = - new HashMap<ServerName, Set<HRegionInfo>>(); + private final Map<ServerName, Set<HRegionInfo>> serverHoldings = new HashMap<>(); /** * Maintains the mapping from the default region to the replica regions. */ - private final Map<HRegionInfo, Set<HRegionInfo>> defaultReplicaToOtherReplicas = - new HashMap<HRegionInfo, Set<HRegionInfo>>(); + private final Map<HRegionInfo, Set<HRegionInfo>> defaultReplicaToOtherReplicas = new HashMap<>(); /** * Region to server assignment map. * Contains the server a given region is currently assigned to. */ - private final TreeMap<HRegionInfo, ServerName> regionAssignments = - new TreeMap<HRegionInfo, ServerName>(); + private final TreeMap<HRegionInfo, ServerName> regionAssignments = new TreeMap<>(); /** * Encoded region name to server assignment map for re-assignment @@ -126,8 +120,7 @@ public class RegionStates { * is offline while the info in lastAssignments is cleared when * the region is closed or the server is dead and processed. */ - private final HashMap<String, ServerName> lastAssignments = - new HashMap<String, ServerName>(); + private final HashMap<String, ServerName> lastAssignments = new HashMap<>(); /** * Encoded region name to server assignment map for the @@ -138,16 +131,14 @@ public class RegionStates { * to match the meta. We need this map to find out the old server * whose serverHoldings needs cleanup, given a moved region. */ - private final HashMap<String, ServerName> oldAssignments = - new HashMap<String, ServerName>(); + private final HashMap<String, ServerName> oldAssignments = new HashMap<>(); /** * Map a host port pair string to the latest start code * of a region server which is known to be dead. It is dead * to us, but server manager may not know it yet. */ - private final HashMap<String, Long> deadServers = - new HashMap<String, Long>(); + private final HashMap<String, Long> deadServers = new HashMap<>(); /** * Map a dead servers to the time when log split is done. @@ -156,8 +147,7 @@ public class RegionStates { * on a configured time. By default, we assume a dead * server should be done with log splitting in two hours. */ - private final HashMap<ServerName, Long> processedServers = - new HashMap<ServerName, Long>(); + private final HashMap<ServerName, Long> processedServers = new HashMap<>(); private long lastProcessedServerCleanTime; private final TableStateManager tableStateManager; @@ -181,7 +171,7 @@ public class RegionStates { * @return a copy of the region assignment map */ public synchronized Map<HRegionInfo, ServerName> getRegionAssignments() { - return new TreeMap<HRegionInfo, ServerName>(regionAssignments); + return new TreeMap<>(regionAssignments); } /** @@ -191,7 +181,7 @@ public class RegionStates { */ synchronized Map<ServerName, List<HRegionInfo>> getRegionAssignments( Collection<HRegionInfo> regions) { - Map<ServerName, List<HRegionInfo>> map = new HashMap<ServerName, List<HRegionInfo>>(); + Map<ServerName, List<HRegionInfo>> map = new HashMap<>(); for (HRegionInfo region : regions) { HRegionInfo defaultReplica = RegionReplicaUtil.getRegionInfoForDefaultReplica(region); Set<HRegionInfo> allReplicas = defaultReplicaToOtherReplicas.get(defaultReplica); @@ -201,7 +191,7 @@ public class RegionStates { if (server != null) { List<HRegionInfo> regionsOnServer = map.get(server); if (regionsOnServer == null) { - regionsOnServer = new ArrayList<HRegionInfo>(1); + regionsOnServer = new ArrayList<>(1); map.put(server, regionsOnServer); } regionsOnServer.add(hri); @@ -220,11 +210,11 @@ public class RegionStates { * Get regions in transition and their states */ public synchronized Set<RegionState> getRegionsInTransition() { - return new HashSet<RegionState>(regionsInTransition.values()); + return new HashSet<>(regionsInTransition.values()); } public synchronized SortedSet<RegionState> getRegionsInTransitionOrderedByTimestamp() { - final TreeSet<RegionState> rit = new TreeSet<RegionState>(REGION_STATE_COMPARATOR); + final TreeSet<RegionState> rit = new TreeSet<>(REGION_STATE_COMPARATOR); for (RegionState rs: regionsInTransition.values()) { rit.add(rs); } @@ -404,7 +394,7 @@ public class RegionStates { RegionState oldState = regionStates.put(encodedName, regionState); Map<String, RegionState> map = regionStatesTableIndex.get(table); if (map == null) { - map = new HashMap<String, RegionState>(); + map = new HashMap<>(); regionStatesTableIndex.put(table, map); } map.put(encodedName, regionState); @@ -483,7 +473,7 @@ public class RegionStates { private void addToServerHoldings(ServerName serverName, HRegionInfo hri) { Set<HRegionInfo> regions = serverHoldings.get(serverName); if (regions == null) { - regions = new HashSet<HRegionInfo>(); + regions = new HashSet<>(); serverHoldings.put(serverName, regions); } regions.add(hri); @@ -494,7 +484,7 @@ public class RegionStates { Set<HRegionInfo> replicas = defaultReplicaToOtherReplicas.get(defaultReplica); if (replicas == null) { - replicas = new HashSet<HRegionInfo>(); + replicas = new HashSet<>(); defaultReplicaToOtherReplicas.put(defaultReplica, replicas); } replicas.add(hri); @@ -618,16 +608,16 @@ public class RegionStates { */ public List<HRegionInfo> serverOffline(final ServerName sn) { // Offline all regions on this server not already in transition. - List<HRegionInfo> rits = new ArrayList<HRegionInfo>(); - Set<HRegionInfo> regionsToCleanIfNoMetaEntry = new HashSet<HRegionInfo>(); + List<HRegionInfo> rits = new ArrayList<>(); + Set<HRegionInfo> regionsToCleanIfNoMetaEntry = new HashSet<>(); // Offline regions outside the loop and synchronized block to avoid // ConcurrentModificationException and deadlock in case of meta anassigned, // but RegionState a blocked. - Set<HRegionInfo> regionsToOffline = new HashSet<HRegionInfo>(); + Set<HRegionInfo> regionsToOffline = new HashSet<>(); synchronized (this) { Set<HRegionInfo> assignedRegions = serverHoldings.get(sn); if (assignedRegions == null) { - assignedRegions = new HashSet<HRegionInfo>(); + assignedRegions = new HashSet<>(); } for (HRegionInfo region : assignedRegions) { @@ -711,7 +701,7 @@ public class RegionStates { * @return Online regions from <code>tableName</code> */ public synchronized List<HRegionInfo> getRegionsOfTable(TableName tableName) { - List<HRegionInfo> tableRegions = new ArrayList<HRegionInfo>(); + List<HRegionInfo> tableRegions = new ArrayList<>(); // boundary needs to have table's name but regionID 0 so that it is sorted // before all table's regions. HRegionInfo boundary = new HRegionInfo(tableName, null, null, false, 0L); @@ -733,10 +723,9 @@ public class RegionStates { */ public synchronized Map<RegionState.State, List<HRegionInfo>> getRegionByStateOfTable(TableName tableName) { - Map<RegionState.State, List<HRegionInfo>> tableRegions = - new HashMap<State, List<HRegionInfo>>(); + Map<RegionState.State, List<HRegionInfo>> tableRegions = new HashMap<>(); for (State state : State.values()) { - tableRegions.put(state, new ArrayList<HRegionInfo>()); + tableRegions.put(state, new ArrayList<>()); } Map<String, RegionState> indexMap = regionStatesTableIndex.get(tableName); if (indexMap == null) @@ -774,7 +763,7 @@ public class RegionStates { * We loop through all regions assuming we don't delete tables too much. */ public void tableDeleted(final TableName tableName) { - Set<HRegionInfo> regionsToDelete = new HashSet<HRegionInfo>(); + Set<HRegionInfo> regionsToDelete = new HashSet<>(); synchronized (this) { for (RegionState state: regionStates.values()) { HRegionInfo region = state.getRegion(); @@ -794,7 +783,7 @@ public class RegionStates { public synchronized Set<HRegionInfo> getServerRegions(ServerName serverName) { Set<HRegionInfo> regions = serverHoldings.get(serverName); if (regions == null) return null; - return new HashSet<HRegionInfo>(regions); + return new HashSet<>(regions); } /** @@ -954,7 +943,7 @@ public class RegionStates { */ synchronized Map<HRegionInfo, ServerName> closeAllUserRegions(Set<TableName> excludedTables) { boolean noExcludeTables = excludedTables == null || excludedTables.isEmpty(); - Set<HRegionInfo> toBeClosed = new HashSet<HRegionInfo>(regionStates.size()); + Set<HRegionInfo> toBeClosed = new HashSet<>(regionStates.size()); for(RegionState state: regionStates.values()) { HRegionInfo hri = state.getRegion(); if (state.isSplit() || hri.isSplit()) { @@ -966,8 +955,7 @@ public class RegionStates { toBeClosed.add(hri); } } - Map<HRegionInfo, ServerName> allUserRegions = - new HashMap<HRegionInfo, ServerName>(toBeClosed.size()); + Map<HRegionInfo, ServerName> allUserRegions = new HashMap<>(toBeClosed.size()); for (HRegionInfo hri: toBeClosed) { RegionState regionState = updateRegionState(hri, State.CLOSED); allUserRegions.put(hri, regionState.getServerName()); @@ -1032,7 +1020,7 @@ public class RegionStates { for (Map<ServerName, List<HRegionInfo>> map: result.values()) { for (ServerName svr: onlineSvrs.keySet()) { if (!map.containsKey(svr)) { - map.put(svr, new ArrayList<HRegionInfo>()); + map.put(svr, new ArrayList<>()); } } map.keySet().removeAll(drainingServers); @@ -1041,20 +1029,19 @@ public class RegionStates { } private Map<TableName, Map<ServerName, List<HRegionInfo>>> getTableRSRegionMap(Boolean bytable){ - Map<TableName, Map<ServerName, List<HRegionInfo>>> result = - new HashMap<TableName, Map<ServerName,List<HRegionInfo>>>(); + Map<TableName, Map<ServerName, List<HRegionInfo>>> result = new HashMap<>(); for (Map.Entry<ServerName, Set<HRegionInfo>> e: serverHoldings.entrySet()) { for (HRegionInfo hri: e.getValue()) { if (hri.isMetaRegion()) continue; TableName tablename = bytable ? hri.getTable() : TableName.valueOf(HConstants.ENSEMBLE_TABLE_NAME); Map<ServerName, List<HRegionInfo>> svrToRegions = result.get(tablename); if (svrToRegions == null) { - svrToRegions = new HashMap<ServerName, List<HRegionInfo>>(serverHoldings.size()); + svrToRegions = new HashMap<>(serverHoldings.size()); result.put(tablename, svrToRegions); } List<HRegionInfo> regions = svrToRegions.get(e.getKey()); if (regions == null) { - regions = new ArrayList<HRegionInfo>(); + regions = new ArrayList<>(); svrToRegions.put(e.getKey(), regions); } regions.add(hri); @@ -1072,10 +1059,9 @@ public class RegionStates { * @return a Map of ServerName to a List of HRegionInfo's */ protected synchronized Map<ServerName, List<HRegionInfo>> getRegionAssignmentsByServer() { - Map<ServerName, List<HRegionInfo>> regionsByServer = - new HashMap<ServerName, List<HRegionInfo>>(serverHoldings.size()); + Map<ServerName, List<HRegionInfo>> regionsByServer = new HashMap<>(serverHoldings.size()); for (Map.Entry<ServerName, Set<HRegionInfo>> e: serverHoldings.entrySet()) { - regionsByServer.put(e.getKey(), new ArrayList<HRegionInfo>(e.getValue())); + regionsByServer.put(e.getKey(), new ArrayList<>(e.getValue())); } return regionsByServer; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java index eb96f97..e6b60d8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java @@ -131,29 +131,25 @@ public class ServerManager { * The last flushed sequence id for a region. */ private final ConcurrentNavigableMap<byte[], Long> flushedSequenceIdByRegion = - new ConcurrentSkipListMap<byte[], Long>(Bytes.BYTES_COMPARATOR); + new ConcurrentSkipListMap<>(Bytes.BYTES_COMPARATOR); /** * The last flushed sequence id for a store in a region. */ private final ConcurrentNavigableMap<byte[], ConcurrentNavigableMap<byte[], Long>> - storeFlushedSequenceIdsByRegion = - new ConcurrentSkipListMap<byte[], ConcurrentNavigableMap<byte[], Long>>(Bytes.BYTES_COMPARATOR); + storeFlushedSequenceIdsByRegion = new ConcurrentSkipListMap<>(Bytes.BYTES_COMPARATOR); /** Map of registered servers to their current load */ - private final ConcurrentNavigableMap<ServerName, ServerLoad> onlineServers = - new ConcurrentSkipListMap<ServerName, ServerLoad>(); + private final ConcurrentNavigableMap<ServerName, ServerLoad> onlineServers = new ConcurrentSkipListMap<>(); /** * Map of admin interfaces per registered regionserver; these interfaces we use to control * regionservers out on the cluster */ - private final Map<ServerName, AdminService.BlockingInterface> rsAdmins = - new HashMap<ServerName, AdminService.BlockingInterface>(); + private final Map<ServerName, AdminService.BlockingInterface> rsAdmins = new HashMap<>(); /** List of region servers that should not get any more new regions. */ - private final ArrayList<ServerName> drainingServers = - new ArrayList<ServerName>(); + private final ArrayList<ServerName> drainingServers = new ArrayList<>(); private final MasterServices master; private final ClusterConnection connection; @@ -182,7 +178,7 @@ public class ServerManager { * So this is a set of region servers known to be dead but not submitted to * ServerShutdownHandler for processing yet. */ - private Set<ServerName> queuedDeadServers = new HashSet<ServerName>(); + private Set<ServerName> queuedDeadServers = new HashSet<>(); /** * Set of region servers which are dead and submitted to ServerShutdownHandler to process but not @@ -199,11 +195,10 @@ public class ServerManager { * is currently in startup mode. In this case, the dead server will be parked in this set * temporarily. */ - private Map<ServerName, Boolean> requeuedDeadServers - = new ConcurrentHashMap<ServerName, Boolean>(); + private Map<ServerName, Boolean> requeuedDeadServers = new ConcurrentHashMap<>(); /** Listeners that are called on server events. */ - private List<ServerListener> listeners = new CopyOnWriteArrayList<ServerListener>(); + private List<ServerListener> listeners = new CopyOnWriteArrayList<>(); /** * Constructor. @@ -1111,7 +1106,7 @@ public class ServerManager { public List<ServerName> getOnlineServersList() { // TODO: optimize the load balancer call so we don't need to make a new list // TODO: FIX. THIS IS POPULAR CALL. - return new ArrayList<ServerName>(this.onlineServers.keySet()); + return new ArrayList<>(this.onlineServers.keySet()); } /** @@ -1139,14 +1134,14 @@ public class ServerManager { * @return A copy of the internal list of draining servers. */ public List<ServerName> getDrainingServersList() { - return new ArrayList<ServerName>(this.drainingServers); + return new ArrayList<>(this.drainingServers); } /** * @return A copy of the internal set of deadNotExpired servers. */ Set<ServerName> getDeadNotExpiredServers() { - return new HashSet<ServerName>(this.queuedDeadServers); + return new HashSet<>(this.queuedDeadServers); } /** @@ -1287,11 +1282,9 @@ public class ServerManager { LOG.warn("Attempting to send favored nodes update rpc to server " + server.toString() + " failed because no RPC connection found to this server"); } else { - List<Pair<HRegionInfo, List<ServerName>>> regionUpdateInfos = - new ArrayList<Pair<HRegionInfo, List<ServerName>>>(); + List<Pair<HRegionInfo, List<ServerName>>> regionUpdateInfos = new ArrayList<>(); for (Entry<HRegionInfo, List<ServerName>> entry : favoredNodes.entrySet()) { - regionUpdateInfos.add(new Pair<HRegionInfo, List<ServerName>>(entry.getKey(), - entry.getValue())); + regionUpdateInfos.add(new Pair<>(entry.getKey(), entry.getValue())); } UpdateFavoredNodesRequest request = RequestConverter.buildUpdateFavoredNodesRequest(regionUpdateInfos); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotOfRegionAssignmentFromMeta.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotOfRegionAssignmentFromMeta.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotOfRegionAssignmentFromMeta.java index 8fedb40..6e477bc 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotOfRegionAssignmentFromMeta.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SnapshotOfRegionAssignmentFromMeta.java @@ -80,19 +80,19 @@ public class SnapshotOfRegionAssignmentFromMeta { private final boolean excludeOfflinedSplitParents; public SnapshotOfRegionAssignmentFromMeta(Connection connection) { - this(connection, new HashSet<TableName>(), false); + this(connection, new HashSet<>(), false); } public SnapshotOfRegionAssignmentFromMeta(Connection connection, Set<TableName> disabledTables, boolean excludeOfflinedSplitParents) { this.connection = connection; - tableToRegionMap = new HashMap<TableName, List<HRegionInfo>>(); - regionToRegionServerMap = new HashMap<HRegionInfo, ServerName>(); - currentRSToRegionMap = new HashMap<ServerName, List<HRegionInfo>>(); - primaryRSToRegionMap = new HashMap<ServerName, List<HRegionInfo>>(); - secondaryRSToRegionMap = new HashMap<ServerName, List<HRegionInfo>>(); - teritiaryRSToRegionMap = new HashMap<ServerName, List<HRegionInfo>>(); - regionNameToRegionInfoMap = new TreeMap<String, HRegionInfo>(); + tableToRegionMap = new HashMap<>(); + regionToRegionServerMap = new HashMap<>(); + currentRSToRegionMap = new HashMap<>(); + primaryRSToRegionMap = new HashMap<>(); + secondaryRSToRegionMap = new HashMap<>(); + teritiaryRSToRegionMap = new HashMap<>(); + regionNameToRegionInfoMap = new TreeMap<>(); existingAssignmentPlan = new FavoredNodesPlan(); this.disabledTables = disabledTables; this.excludeOfflinedSplitParents = excludeOfflinedSplitParents; @@ -180,7 +180,7 @@ public class SnapshotOfRegionAssignmentFromMeta { TableName tableName = regionInfo.getTable(); List<HRegionInfo> regionList = tableToRegionMap.get(tableName); if (regionList == null) { - regionList = new ArrayList<HRegionInfo>(); + regionList = new ArrayList<>(); } // Add the current region info into the tableToRegionMap regionList.add(regionInfo); @@ -196,7 +196,7 @@ public class SnapshotOfRegionAssignmentFromMeta { // Process the region server to region map List<HRegionInfo> regionList = currentRSToRegionMap.get(server); if (regionList == null) { - regionList = new ArrayList<HRegionInfo>(); + regionList = new ArrayList<>(); } regionList.add(regionInfo); currentRSToRegionMap.put(server, regionList); @@ -206,7 +206,7 @@ public class SnapshotOfRegionAssignmentFromMeta { // Process the region server to region map List<HRegionInfo> regionList = primaryRSToRegionMap.get(server); if (regionList == null) { - regionList = new ArrayList<HRegionInfo>(); + regionList = new ArrayList<>(); } regionList.add(regionInfo); primaryRSToRegionMap.put(server, regionList); @@ -216,7 +216,7 @@ public class SnapshotOfRegionAssignmentFromMeta { // Process the region server to region map List<HRegionInfo> regionList = secondaryRSToRegionMap.get(server); if (regionList == null) { - regionList = new ArrayList<HRegionInfo>(); + regionList = new ArrayList<>(); } regionList.add(regionInfo); secondaryRSToRegionMap.put(server, regionList); @@ -226,7 +226,7 @@ public class SnapshotOfRegionAssignmentFromMeta { // Process the region server to region map List<HRegionInfo> regionList = teritiaryRSToRegionMap.get(server); if (regionList == null) { - regionList = new ArrayList<HRegionInfo>(); + regionList = new ArrayList<>(); } regionList.add(regionInfo); teritiaryRSToRegionMap.put(server, regionList); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java index 20fef35..7017d29 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java @@ -118,7 +118,7 @@ public class SplitLogManager { protected final ReentrantLock recoveringRegionLock = new ReentrantLock(); @VisibleForTesting - final ConcurrentMap<String, Task> tasks = new ConcurrentHashMap<String, Task>(); + final ConcurrentMap<String, Task> tasks = new ConcurrentHashMap<>(); private TimeoutMonitor timeoutMonitor; private volatile Set<ServerName> deadWorkers = null; @@ -176,7 +176,7 @@ public class SplitLogManager { public static FileStatus[] getFileList(final Configuration conf, final List<Path> logDirs, final PathFilter filter) throws IOException { - List<FileStatus> fileStatus = new ArrayList<FileStatus>(); + List<FileStatus> fileStatus = new ArrayList<>(); for (Path logDir : logDirs) { final FileSystem fs = logDir.getFileSystem(conf); if (!fs.exists(logDir)) { @@ -201,7 +201,7 @@ public class SplitLogManager { * @throws IOException */ public long splitLogDistributed(final Path logDir) throws IOException { - List<Path> logDirs = new ArrayList<Path>(); + List<Path> logDirs = new ArrayList<>(); logDirs.add(logDir); return splitLogDistributed(logDirs); } @@ -218,7 +218,7 @@ public class SplitLogManager { if (logDirs.isEmpty()) { return 0; } - Set<ServerName> serverNames = new HashSet<ServerName>(); + Set<ServerName> serverNames = new HashSet<>(); for (Path logDir : logDirs) { try { ServerName serverName = AbstractFSWALProvider.getServerNameFromWALDirectoryName(logDir); @@ -398,7 +398,7 @@ public class SplitLogManager { } if (serverNames == null || serverNames.isEmpty()) return; - Set<String> recoveredServerNameSet = new HashSet<String>(); + Set<String> recoveredServerNameSet = new HashSet<>(); for (ServerName tmpServerName : serverNames) { recoveredServerNameSet.add(tmpServerName.getServerName()); } @@ -410,8 +410,7 @@ public class SplitLogManager { } catch (IOException e) { LOG.warn("removeRecoveringRegions got exception. Will retry", e); if (serverNames != null && !serverNames.isEmpty()) { - this.failedRecoveringRegionDeletions.add(new Pair<Set<ServerName>, Boolean>(serverNames, - isMetaRecovery)); + this.failedRecoveringRegionDeletions.add(new Pair<>(serverNames, isMetaRecovery)); } } finally { this.recoveringRegionLock.unlock(); @@ -426,7 +425,7 @@ public class SplitLogManager { */ void removeStaleRecoveringRegions(final Set<ServerName> failedServers) throws IOException, InterruptedIOException { - Set<String> knownFailedServers = new HashSet<String>(); + Set<String> knownFailedServers = new HashSet<>(); if (failedServers != null) { for (ServerName tmpServerName : failedServers) { knownFailedServers.add(tmpServerName.getServerName()); @@ -519,7 +518,7 @@ public class SplitLogManager { // to reason about concurrency. Makes it easier to retry. synchronized (deadWorkersLock) { if (deadWorkers == null) { - deadWorkers = new HashSet<ServerName>(100); + deadWorkers = new HashSet<>(100); } deadWorkers.add(workerName); } @@ -529,7 +528,7 @@ public class SplitLogManager { void handleDeadWorkers(Set<ServerName> serverNames) { synchronized (deadWorkersLock) { if (deadWorkers == null) { - deadWorkers = new HashSet<ServerName>(100); + deadWorkers = new HashSet<>(100); } deadWorkers.addAll(serverNames); } @@ -749,7 +748,7 @@ public class SplitLogManager { getSplitLogManagerCoordination().getDetails().getFailedDeletions(); // Retry previously failed deletes if (failedDeletions.size() > 0) { - List<String> tmpPaths = new ArrayList<String>(failedDeletions); + List<String> tmpPaths = new ArrayList<>(failedDeletions); for (String tmpPath : tmpPaths) { // deleteNode is an async call getSplitLogManagerCoordination().deleteTask(tmpPath); @@ -766,7 +765,7 @@ public class SplitLogManager { // inside the function there have more checks before GC anything if (!failedRecoveringRegionDeletions.isEmpty()) { List<Pair<Set<ServerName>, Boolean>> previouslyFailedDeletions = - new ArrayList<Pair<Set<ServerName>, Boolean>>(failedRecoveringRegionDeletions); + new ArrayList<>(failedRecoveringRegionDeletions); failedRecoveringRegionDeletions.removeAll(previouslyFailedDeletions); for (Pair<Set<ServerName>, Boolean> failedDeletion : previouslyFailedDeletions) { removeRecoveringRegions(failedDeletion.getFirst(), failedDeletion.getSecond());
