http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java index c3635cb..bc51440 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CoprocessorClassLoader.java @@ -130,13 +130,13 @@ public class CoprocessorClassLoader extends ClassLoaderBase { /** * A locker used to synchronize class loader initialization per coprocessor jar file */ - private static final KeyLocker<String> locker = new KeyLocker<String>(); + private static final KeyLocker<String> locker = new KeyLocker<>(); /** * A set used to synchronized parent path clean up. Generally, there * should be only one parent path, but using a set so that we can support more. */ - static final HashSet<String> parentDirLockSet = new HashSet<String>(); + static final HashSet<String> parentDirLockSet = new HashSet<>(); /** * Creates a JarClassLoader that loads classes from the given paths.
http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Counter.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Counter.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Counter.java index 0d3a5c6..36ca7ad 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Counter.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Counter.java @@ -94,7 +94,7 @@ public class Counter { } private Counter(Cell initCell) { - containerRef = new AtomicReference<Container>(new Container(initCell)); + containerRef = new AtomicReference<>(new Container(initCell)); } private static int hash() { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/DynamicClassLoader.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/DynamicClassLoader.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/DynamicClassLoader.java index 595cc5b..1a73069 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/DynamicClassLoader.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/DynamicClassLoader.java @@ -102,7 +102,7 @@ public class DynamicClassLoader extends ClassLoaderBase { // FindBugs: Making synchronized to avoid IS2_INCONSISTENT_SYNC complaints about // remoteDirFs and jarModifiedTime being part synchronized protected. private synchronized void initTempDir(final Configuration conf) { - jarModifiedTime = new HashMap<String, Long>(); + jarModifiedTime = new HashMap<>(); String localDirPath = conf.get( LOCAL_DIR_KEY, DEFAULT_LOCAL_DIR) + DYNAMIC_JARS_DIR; localDir = new File(localDirPath); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/IterableUtils.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/IterableUtils.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/IterableUtils.java index 41e837d..862da43 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/IterableUtils.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/IterableUtils.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience; public class IterableUtils { private static final List<Object> EMPTY_LIST = Collections - .unmodifiableList(new ArrayList<Object>(0)); + .unmodifiableList(new ArrayList<>(0)); @SuppressWarnings("unchecked") public static <T> Iterable<T> nullSafe(Iterable<T> in) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/KeyLocker.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/KeyLocker.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/KeyLocker.java index 05bd66d..6acf584 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/KeyLocker.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/KeyLocker.java @@ -49,7 +49,7 @@ public class KeyLocker<K> { private static final int NB_CONCURRENT_LOCKS = 1000; private final WeakObjectPool<K, ReentrantLock> lockPool = - new WeakObjectPool<K, ReentrantLock>( + new WeakObjectPool<>( new WeakObjectPool.ObjectFactory<K, ReentrantLock>() { @Override public ReentrantLock createObject(K key) { @@ -85,7 +85,7 @@ public class KeyLocker<K> { Arrays.sort(keyArray); lockPool.purge(); - Map<K, Lock> locks = new LinkedHashMap<K, Lock>(keyArray.length); + Map<K, Lock> locks = new LinkedHashMap<>(keyArray.length); for (Object o : keyArray) { @SuppressWarnings("unchecked") K key = (K)o; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Pair.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Pair.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Pair.java index 159924f..719d1ee 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Pair.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Pair.java @@ -64,7 +64,7 @@ public class Pair<T1, T2> implements Serializable * @return a new pair containing the passed arguments */ public static <T1,T2> Pair<T1,T2> newPair(T1 a, T2 b) { - return new Pair<T1,T2>(a, b); + return new Pair<>(a, b); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Threads.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Threads.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Threads.java index 21b376c..279ce95 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Threads.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Threads.java @@ -191,7 +191,7 @@ public class Threads { ThreadFactory threadFactory) { ThreadPoolExecutor boundedCachedThreadPool = new ThreadPoolExecutor(maxCachedThread, maxCachedThread, timeout, - unit, new LinkedBlockingQueue<Runnable>(), threadFactory); + unit, new LinkedBlockingQueue<>(), threadFactory); // allow the core pool threads timeout and terminate boundedCachedThreadPool.allowCoreThreadTimeOut(true); return boundedCachedThreadPool; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Triple.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Triple.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Triple.java index 1de6bee..9ee0ab5 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Triple.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/Triple.java @@ -41,7 +41,7 @@ public class Triple<A, B, C> { // ctor cannot infer types w/o warning but a method can. public static <A, B, C> Triple<A, B, C> create(A first, B second, C third) { - return new Triple<A, B, C>(first, second, third); + return new Triple<>(first, second, third); } public int hashCode() { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WeakObjectPool.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WeakObjectPool.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WeakObjectPool.java index 7757c6c..478864b 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WeakObjectPool.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/WeakObjectPool.java @@ -50,7 +50,7 @@ public class WeakObjectPool<K, V> { V createObject(K key); } - private final ReferenceQueue<V> staleRefQueue = new ReferenceQueue<V>(); + private final ReferenceQueue<V> staleRefQueue = new ReferenceQueue<>(); private class ObjectReference extends WeakReference<V> { final K key; @@ -126,8 +126,7 @@ public class WeakObjectPool<K, V> { } this.objectFactory = objectFactory; - this.referenceCache = new ConcurrentHashMap<K, ObjectReference>( - initialCapacity, 0.75f, concurrencyLevel); + this.referenceCache = new ConcurrentHashMap<>(initialCapacity, 0.75f, concurrencyLevel); // 0.75f is the default load factor threshold of ConcurrentHashMap. } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/main/java/org/apache/hadoop/hbase/util/test/RedundantKVGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/test/RedundantKVGenerator.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/test/RedundantKVGenerator.java index c73705a..87d56a9 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/test/RedundantKVGenerator.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/test/RedundantKVGenerator.java @@ -171,7 +171,7 @@ public class RedundantKVGenerator { private List<byte[]> generateRows() { // generate prefixes - List<byte[]> prefixes = new ArrayList<byte[]>(); + List<byte[]> prefixes = new ArrayList<>(); prefixes.add(new byte[0]); for (int i = 1; i < numberOfRowPrefixes; ++i) { int prefixLength = averagePrefixLength; @@ -184,7 +184,7 @@ public class RedundantKVGenerator { } // generate rest of the row - List<byte[]> rows = new ArrayList<byte[]>(); + List<byte[]> rows = new ArrayList<>(); for (int i = 0; i < numberOfRows; ++i) { int suffixLength = averageSuffixLength; suffixLength += randomizer.nextInt(2 * suffixLengthVariance + 1) - @@ -213,10 +213,10 @@ public class RedundantKVGenerator { * @return sorted list of key values */ public List<KeyValue> generateTestKeyValues(int howMany, boolean useTags) { - List<KeyValue> result = new ArrayList<KeyValue>(); + List<KeyValue> result = new ArrayList<>(); List<byte[]> rows = generateRows(); - Map<Integer, List<byte[]>> rowsToQualifier = new HashMap<Integer, List<byte[]>>(); + Map<Integer, List<byte[]>> rowsToQualifier = new HashMap<>(); if(family==null){ family = new byte[columnFamilyLength]; @@ -249,7 +249,7 @@ public class RedundantKVGenerator { // add it to map if (!rowsToQualifier.containsKey(rowId)) { - rowsToQualifier.put(rowId, new ArrayList<byte[]>()); + rowsToQualifier.put(rowId, new ArrayList<>()); } rowsToQualifier.get(rowId).add(qualifier); } else if (qualifierChance > chanceForSameQualifier) { @@ -299,9 +299,9 @@ public class RedundantKVGenerator { * @return sorted list of key values */ public List<Cell> generateTestExtendedOffheapKeyValues(int howMany, boolean useTags) { - List<Cell> result = new ArrayList<Cell>(); + List<Cell> result = new ArrayList<>(); List<byte[]> rows = generateRows(); - Map<Integer, List<byte[]>> rowsToQualifier = new HashMap<Integer, List<byte[]>>(); + Map<Integer, List<byte[]>> rowsToQualifier = new HashMap<>(); if (family == null) { family = new byte[columnFamilyLength]; @@ -334,7 +334,7 @@ public class RedundantKVGenerator { // add it to map if (!rowsToQualifier.containsKey(rowId)) { - rowsToQualifier.put(rowId, new ArrayList<byte[]>()); + rowsToQualifier.put(rowId, new ArrayList<>()); } rowsToQualifier.get(rowId).add(qualifier); } else if (qualifierChance > chanceForSameQualifier) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java index 01d387c..0aa30ee 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java @@ -148,8 +148,8 @@ public class ClassFinder { final Pattern jarResourceRe = Pattern.compile("^file:(.+\\.jar)!/" + path + "$"); Enumeration<URL> resources = ClassLoader.getSystemClassLoader().getResources(path); - List<File> dirs = new ArrayList<File>(); - List<String> jars = new ArrayList<String>(); + List<File> dirs = new ArrayList<>(); + List<String> jars = new ArrayList<>(); while (resources.hasMoreElements()) { URL resource = resources.nextElement(); @@ -168,7 +168,7 @@ public class ClassFinder { } } - Set<Class<?>> classes = new HashSet<Class<?>>(); + Set<Class<?>> classes = new HashSet<>(); for (File directory : dirs) { classes.addAll(findClassesFromFiles(directory, packageName, proceedOnExceptions)); } @@ -189,7 +189,7 @@ public class ClassFinder { throw ioEx; } - Set<Class<?>> classes = new HashSet<Class<?>>(); + Set<Class<?>> classes = new HashSet<>(); JarEntry entry = null; try { while (true) { @@ -236,7 +236,7 @@ public class ClassFinder { private Set<Class<?>> findClassesFromFiles(File baseDirectory, String packageName, boolean proceedOnExceptions) throws ClassNotFoundException, LinkageError { - Set<Class<?>> classes = new HashSet<Class<?>>(); + Set<Class<?>> classes = new HashSet<>(); if (!baseDirectory.exists()) { LOG.warn(baseDirectory.getAbsolutePath() + " does not exist"); return classes; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceChecker.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceChecker.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceChecker.java index 310a2fb..c0b9836 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceChecker.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceChecker.java @@ -94,7 +94,7 @@ public class ResourceChecker { public List<String> getStringsToLog() { return null; } } - private List<ResourceAnalyzer> ras = new ArrayList<ResourceAnalyzer>(); + private List<ResourceAnalyzer> ras = new ArrayList<>(); private int[] initialValues; private int[] endingValues; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceCheckerJUnitListener.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceCheckerJUnitListener.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceCheckerJUnitListener.java index 6264a5e..751b9e3 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceCheckerJUnitListener.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/ResourceCheckerJUnitListener.java @@ -41,10 +41,10 @@ import org.junit.runner.notification.RunListener; * When surefire forkMode=once/always/perthread, this code is executed on the forked process. */ public class ResourceCheckerJUnitListener extends RunListener { - private Map<String, ResourceChecker> rcs = new ConcurrentHashMap<String, ResourceChecker>(); + private Map<String, ResourceChecker> rcs = new ConcurrentHashMap<>(); static class ThreadResourceAnalyzer extends ResourceChecker.ResourceAnalyzer { - private static Set<String> initialThreadNames = new HashSet<String>(); + private static Set<String> initialThreadNames = new HashSet<>(); private static List<String> stringsToLog = null; @Override @@ -57,7 +57,7 @@ public class ResourceCheckerJUnitListener extends RunListener { } } else if (phase == Phase.END) { if (stackTraces.size() > initialThreadNames.size()) { - stringsToLog = new ArrayList<String>(); + stringsToLog = new ArrayList<>(); for (Thread t : stackTraces.keySet()) { if (!initialThreadNames.contains(t.getName())) { stringsToLog.add("\nPotentially hanging thread: " + t.getName() + "\n"); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCellUtil.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCellUtil.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCellUtil.java index 41a011d..441d1b5 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCellUtil.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCellUtil.java @@ -215,13 +215,13 @@ public class TestCellUtil { consume(doCreateCellArray(1), 1); consume(doCreateCellArray(0), 0); consume(doCreateCellArray(3), 3); - List<CellScannable> cells = new ArrayList<CellScannable>(hundredK); + List<CellScannable> cells = new ArrayList<>(hundredK); for (int i = 0; i < hundredK; i++) { cells.add(new TestCellScannable(1)); } consume(CellUtil.createCellScanner(cells), hundredK * 1); - NavigableMap<byte [], List<Cell>> m = new TreeMap<byte [], List<Cell>>(Bytes.BYTES_COMPARATOR); - List<Cell> cellArray = new ArrayList<Cell>(hundredK); + NavigableMap<byte [], List<Cell>> m = new TreeMap<>(Bytes.BYTES_COMPARATOR); + List<Cell> cellArray = new ArrayList<>(hundredK); for (int i = 0; i < hundredK; i++) cellArray.add(new TestCell(i)); m.put(new byte [] {'f'}, cellArray); consume(CellUtil.createCellScanner(m), hundredK * 1); @@ -237,7 +237,7 @@ public class TestCellUtil { private CellScanner doCreateCellScanner(final int listsCount, final int itemsPerList) throws IOException { - List<CellScannable> cells = new ArrayList<CellScannable>(listsCount); + List<CellScannable> cells = new ArrayList<>(listsCount); for (int i = 0; i < listsCount; i++) { CellScannable cs = new CellScannable() { @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/TestClassFinder.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestClassFinder.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestClassFinder.java index 5154810..244c267 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestClassFinder.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestClassFinder.java @@ -381,7 +381,7 @@ public class TestClassFinder { // Directory entries for all packages have to be added explicitly for // resources to be findable via ClassLoader. Directory entries must end // with "/"; the initial one is expected to, also. - Set<String> pathsInJar = new HashSet<String>(); + Set<String> pathsInJar = new HashSet<>(); for (FileAndPath fileAndPath : filesInJar) { String pathToAdd = fileAndPath.path; while (pathsInJar.add(pathToAdd)) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCompoundConfiguration.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCompoundConfiguration.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCompoundConfiguration.java index 57409b6..0a0a1d2 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCompoundConfiguration.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestCompoundConfiguration.java @@ -121,8 +121,7 @@ public class TestCompoundConfiguration extends TestCase { @Test public void testWithIbwMap() { - Map<Bytes, Bytes> map = - new HashMap<Bytes, Bytes>(); + Map<Bytes, Bytes> map = new HashMap<>(); map.put(strToIb("B"), strToIb("2b")); map.put(strToIb("C"), strToIb("33")); map.put(strToIb("D"), strToIb("4")); @@ -162,7 +161,7 @@ public class TestCompoundConfiguration extends TestCase { @Test public void testWithStringMap() { - Map<String, String> map = new HashMap<String, String>(); + Map<String, String> map = new HashMap<>(); map.put("B", "2b"); map.put("C", "33"); map.put("D", "4"); @@ -199,10 +198,10 @@ public class TestCompoundConfiguration extends TestCase { @Test public void testLaterConfigsOverrideEarlier() { - Map<String, String> map1 = new HashMap<String, String>(); + Map<String, String> map1 = new HashMap<>(); map1.put("A", "2"); map1.put("D", "5"); - Map<String, String> map2 = new HashMap<String, String>(); + Map<String, String> map2 = new HashMap<>(); String newValueForA = "3", newValueForB = "4"; map2.put("A", newValueForA); map2.put("B", newValueForB); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/TestKeyValue.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestKeyValue.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestKeyValue.java index 4e0090d..562c008 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestKeyValue.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestKeyValue.java @@ -222,7 +222,7 @@ public class TestKeyValue extends TestCase { } public void testBinaryKeys() throws Exception { - Set<KeyValue> set = new TreeSet<KeyValue>(CellComparator.COMPARATOR); + Set<KeyValue> set = new TreeSet<>(CellComparator.COMPARATOR); final byte [] fam = Bytes.toBytes("col"); final byte [] qf = Bytes.toBytes("umn"); final byte [] nb = new byte[0]; @@ -248,7 +248,7 @@ public class TestKeyValue extends TestCase { } assertTrue(assertion); // Make set with good comparator - set = new TreeSet<KeyValue>(CellComparator.META_COMPARATOR); + set = new TreeSet<>(CellComparator.META_COMPARATOR); Collections.addAll(set, keys); count = 0; for (KeyValue k: set) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestBoundedByteBufferPool.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestBoundedByteBufferPool.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestBoundedByteBufferPool.java index 8775d7f..eca7712 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestBoundedByteBufferPool.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestBoundedByteBufferPool.java @@ -93,7 +93,7 @@ public class TestBoundedByteBufferPool { @Test public void testBufferSizeGrowWithMultiThread() throws Exception { - final ConcurrentLinkedDeque<ByteBuffer> bufferQueue = new ConcurrentLinkedDeque<ByteBuffer>(); + final ConcurrentLinkedDeque<ByteBuffer> bufferQueue = new ConcurrentLinkedDeque<>(); int takeBufferThreadsCount = 30; int putBufferThreadsCount = 1; Thread takeBufferThreads[] = new Thread[takeBufferThreadsCount]; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestTagCompressionContext.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestTagCompressionContext.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestTagCompressionContext.java index 5e609ad..dddd9e7 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestTagCompressionContext.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/io/TestTagCompressionContext.java @@ -150,7 +150,7 @@ public class TestTagCompressionContext { } private KeyValue createKVWithTags(int noOfTags) { - List<Tag> tags = new ArrayList<Tag>(); + List<Tag> tags = new ArrayList<>(); for (int i = 0; i < noOfTags; i++) { tags.add(new ArrayBackedTag((byte) i, "tagValue" + i)); } @@ -159,7 +159,7 @@ public class TestTagCompressionContext { } private Cell createOffheapKVWithTags(int noOfTags) { - List<Tag> tags = new ArrayList<Tag>(); + List<Tag> tags = new ArrayList<>(); for (int i = 0; i < noOfTags; i++) { tags.add(new ArrayBackedTag((byte) i, "tagValue" + i)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/nio/TestMultiByteBuff.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/nio/TestMultiByteBuff.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/nio/TestMultiByteBuff.java index af4c464..48922d9 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/nio/TestMultiByteBuff.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/nio/TestMultiByteBuff.java @@ -243,7 +243,7 @@ public class TestMultiByteBuff { assertFalse(bb2 == sub); assertEquals(l2, ByteBufferUtils.toLong(sub, sub.position())); multi.rewind(); - ObjectIntPair<ByteBuffer> p = new ObjectIntPair<ByteBuffer>(); + ObjectIntPair<ByteBuffer> p = new ObjectIntPair<>(); multi.asSubByteBuffer(8, Bytes.SIZEOF_LONG, p); assertFalse(bb1 == p.getFirst()); assertFalse(bb2 == p.getFirst()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestFixedLengthWrapper.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestFixedLengthWrapper.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestFixedLengthWrapper.java index b259429..c2c5a6d 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestFixedLengthWrapper.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestFixedLengthWrapper.java @@ -51,7 +51,7 @@ public class TestFixedLengthWrapper { for (Order ord : new Order[] { Order.ASCENDING, Order.DESCENDING }) { for (byte[] val : VALUES) { buff.setPosition(0); - DataType<byte[]> type = new FixedLengthWrapper<byte[]>(new RawBytes(ord), limit); + DataType<byte[]> type = new FixedLengthWrapper<>(new RawBytes(ord), limit); assertEquals(limit, type.encode(buff, val)); buff.setPosition(0); byte[] actual = type.decode(buff); @@ -67,21 +67,21 @@ public class TestFixedLengthWrapper { @Test(expected = IllegalArgumentException.class) public void testInsufficientRemainingRead() { PositionedByteRange buff = new SimplePositionedMutableByteRange(0); - DataType<byte[]> type = new FixedLengthWrapper<byte[]>(new RawBytes(), 3); + DataType<byte[]> type = new FixedLengthWrapper<>(new RawBytes(), 3); type.decode(buff); } @Test(expected = IllegalArgumentException.class) public void testInsufficientRemainingWrite() { PositionedByteRange buff = new SimplePositionedMutableByteRange(0); - DataType<byte[]> type = new FixedLengthWrapper<byte[]>(new RawBytes(), 3); + DataType<byte[]> type = new FixedLengthWrapper<>(new RawBytes(), 3); type.encode(buff, Bytes.toBytes("")); } @Test(expected = IllegalArgumentException.class) public void testOverflowPassthrough() { PositionedByteRange buff = new SimplePositionedMutableByteRange(3); - DataType<byte[]> type = new FixedLengthWrapper<byte[]>(new RawBytes(), 0); + DataType<byte[]> type = new FixedLengthWrapper<>(new RawBytes(), 0); type.encode(buff, Bytes.toBytes("foo")); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestStructNullExtension.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestStructNullExtension.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestStructNullExtension.java index e87438d..2b2efe6 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestStructNullExtension.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestStructNullExtension.java @@ -58,7 +58,7 @@ public class TestStructNullExtension { Struct shorter = builder.toStruct(); Struct longer = builder // intentionally include a wrapped instance to test wrapper behavior. - .add(new TerminatedWrapper<String>(OrderedString.ASCENDING, "/")) + .add(new TerminatedWrapper<>(OrderedString.ASCENDING, "/")) .add(OrderedNumeric.ASCENDING) .toStruct(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestTerminatedWrapper.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestTerminatedWrapper.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestTerminatedWrapper.java index e36a141..310067b 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestTerminatedWrapper.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/types/TestTerminatedWrapper.java @@ -47,7 +47,7 @@ public class TestTerminatedWrapper { @Test(expected = IllegalArgumentException.class) public void testEmptyDelimiter() { - new TerminatedWrapper<byte[]>(new RawBytes(), ""); + new TerminatedWrapper<>(new RawBytes(), ""); } @Test(expected = IllegalArgumentException.class) @@ -58,7 +58,7 @@ public class TestTerminatedWrapper { @Test(expected = IllegalArgumentException.class) public void testEncodedValueContainsTerm() { - DataType<byte[]> type = new TerminatedWrapper<byte[]>(new RawBytes(), "foo"); + DataType<byte[]> type = new TerminatedWrapper<>(new RawBytes(), "foo"); PositionedByteRange buff = new SimplePositionedMutableByteRange(16); type.encode(buff, Bytes.toBytes("hello foobar!")); } @@ -72,7 +72,7 @@ public class TestTerminatedWrapper { for (byte[] term : TERMINATORS) { for (String val : VALUES_STRINGS) { buff.setPosition(0); - DataType<String> type = new TerminatedWrapper<String>(t, term); + DataType<String> type = new TerminatedWrapper<>(t, term); assertEquals(val.length() + 2 + term.length, type.encode(buff, val)); buff.setPosition(0); assertEquals(val, type.decode(buff)); @@ -89,7 +89,7 @@ public class TestTerminatedWrapper { for (byte[] term : TERMINATORS) { for (byte[] val : VALUES_BYTES) { buff.setPosition(0); - DataType<byte[]> type = new TerminatedWrapper<byte[]>(new RawBytes(ord), term); + DataType<byte[]> type = new TerminatedWrapper<>(new RawBytes(ord), term); assertEquals(val.length + term.length, type.encode(buff, val)); buff.setPosition(0); assertArrayEquals(val, type.decode(buff)); @@ -108,7 +108,7 @@ public class TestTerminatedWrapper { for (byte[] term : TERMINATORS) { for (String val : VALUES_STRINGS) { buff.setPosition(0); - DataType<String> type = new TerminatedWrapper<String>(t, term); + DataType<String> type = new TerminatedWrapper<>(t, term); int expected = val.length() + 2 + term.length; assertEquals(expected, type.encode(buff, val)); buff.setPosition(0); @@ -126,7 +126,7 @@ public class TestTerminatedWrapper { for (byte[] term : TERMINATORS) { for (byte[] val : VALUES_BYTES) { buff.setPosition(0); - DataType<byte[]> type = new TerminatedWrapper<byte[]>(new RawBytes(ord), term); + DataType<byte[]> type = new TerminatedWrapper<>(new RawBytes(ord), term); int expected = type.encode(buff, val); buff.setPosition(0); assertEquals(expected, type.skip(buff)); @@ -139,7 +139,7 @@ public class TestTerminatedWrapper { @Test(expected = IllegalArgumentException.class) public void testInvalidSkip() { PositionedByteRange buff = new SimplePositionedMutableByteRange(Bytes.toBytes("foo")); - DataType<byte[]> type = new TerminatedWrapper<byte[]>(new RawBytes(), new byte[] { 0x00 }); + DataType<byte[]> type = new TerminatedWrapper<>(new RawBytes(), new byte[] { 0x00 }); type.skip(buff); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/ClassLoaderTestHelper.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/ClassLoaderTestHelper.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/ClassLoaderTestHelper.java index 30e33d9..ba6cea0 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/ClassLoaderTestHelper.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/ClassLoaderTestHelper.java @@ -133,13 +133,13 @@ public class ClassLoaderTestHelper { // compile it by JavaCompiler JavaCompiler compiler = ToolProvider.getSystemJavaCompiler(); - ArrayList<String> srcFileNames = new ArrayList<String>(1); + ArrayList<String> srcFileNames = new ArrayList<>(1); srcFileNames.add(sourceCodeFile.toString()); StandardJavaFileManager fm = compiler.getStandardFileManager(null, null, null); Iterable<? extends JavaFileObject> cu = fm.getJavaFileObjects(sourceCodeFile); - List<String> options = new ArrayList<String>(2); + List<String> options = new ArrayList<>(2); options.add("-classpath"); // only add hbase classes to classpath. This is a little bit tricky: assume // the classpath is {hbaseSrc}/target/classes. http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestAvlUtil.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestAvlUtil.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestAvlUtil.java index 3c7b680..554e108 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestAvlUtil.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestAvlUtil.java @@ -48,7 +48,7 @@ public class TestAvlUtil { final int MAX_KEY = 99999999; final int NELEM = 10000; - final TreeMap<Integer, Object> treeMap = new TreeMap<Integer, Object>(); + final TreeMap<Integer, Object> treeMap = new TreeMap<>(); TestAvlNode root = null; final Random rand = new Random(); @@ -117,7 +117,7 @@ public class TestAvlUtil { root = AvlTree.insert(root, new TestAvlNode(i)); } - AvlTreeIterator<TestAvlNode> iter = new AvlTreeIterator<TestAvlNode>(root); + AvlTreeIterator<TestAvlNode> iter = new AvlTreeIterator<>(root); assertTrue(iter.hasNext()); long prevKey = 0; while (iter.hasNext()) { @@ -139,7 +139,7 @@ public class TestAvlUtil { } for (int i = MIN_KEY - 1; i < MAX_KEY + 1; ++i) { - AvlTreeIterator<TestAvlNode> iter = new AvlTreeIterator<TestAvlNode>(root, i, KEY_COMPARATOR); + AvlTreeIterator<TestAvlNode> iter = new AvlTreeIterator<>(root, i, KEY_COMPARATOR); if (i < MAX_KEY) { assertTrue(iter.hasNext()); } else { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBase64.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBase64.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBase64.java index 09ef707..7c74bca 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBase64.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBase64.java @@ -55,7 +55,7 @@ public class TestBase64 extends TestCase { * @throws UnsupportedEncodingException */ public void testBase64() throws UnsupportedEncodingException { - TreeMap<String, String> sorted = new TreeMap<String, String>(); + TreeMap<String, String> sorted = new TreeMap<>(); for (int i = 0; i < uris.length; i++) { byte[] bytes = uris[i].getBytes("UTF-8"); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBoundedArrayQueue.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBoundedArrayQueue.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBoundedArrayQueue.java index 2cc3751..6d9c496 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBoundedArrayQueue.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBoundedArrayQueue.java @@ -31,7 +31,7 @@ import org.junit.experimental.categories.Category; public class TestBoundedArrayQueue { private int qMaxElements = 5; - private BoundedArrayQueue<Integer> queue = new BoundedArrayQueue<Integer>(qMaxElements); + private BoundedArrayQueue<Integer> queue = new BoundedArrayQueue<>(qMaxElements); @Test public void testBoundedArrayQueueOperations() throws Exception { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteBufferUtils.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteBufferUtils.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteBufferUtils.java index e94293c..b78574a 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteBufferUtils.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteBufferUtils.java @@ -77,7 +77,7 @@ public class TestByteBufferUtils { } static { - SortedSet<Long> a = new TreeSet<Long>(); + SortedSet<Long> a = new TreeSet<>(); for (int i = 0; i <= 63; ++i) { long v = (-1L) << i; assertTrue(v < 0); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteRangeWithKVSerialization.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteRangeWithKVSerialization.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteRangeWithKVSerialization.java index 717e24c..8ae2a29 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteRangeWithKVSerialization.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestByteRangeWithKVSerialization.java @@ -64,7 +64,7 @@ public class TestByteRangeWithKVSerialization { final byte[] QUALIFIER = Bytes.toBytes("q1"); final byte[] VALUE = Bytes.toBytes("v"); int kvCount = 1000000; - List<KeyValue> kvs = new ArrayList<KeyValue>(kvCount); + List<KeyValue> kvs = new ArrayList<>(kvCount); int totalSize = 0; Tag[] tags = new Tag[] { new ArrayBackedTag((byte) 1, "tag1") }; for (int i = 0; i < kvCount; i++) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBytes.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBytes.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBytes.java index e145642..38b01b8 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBytes.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestBytes.java @@ -498,7 +498,7 @@ public class TestBytes extends TestCase { } public void testToFromHex() { - List<String> testStrings = new ArrayList<String>(8); + List<String> testStrings = new ArrayList<>(8); testStrings.addAll(Arrays.asList(new String[] { "", "00", @@ -517,7 +517,7 @@ public class TestBytes extends TestCase { Assert.assertTrue(testString.equalsIgnoreCase(result)); } - List<byte[]> testByteData = new ArrayList<byte[]>(5); + List<byte[]> testByteData = new ArrayList<>(5); testByteData.addAll(Arrays.asList(new byte[][] { new byte[0], new byte[1], http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestConcatenatedLists.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestConcatenatedLists.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestConcatenatedLists.java index fd4baf5..cfd288d 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestConcatenatedLists.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestConcatenatedLists.java @@ -39,7 +39,7 @@ public class TestConcatenatedLists { @Test public void testUnsupportedOps() { // If adding support, add tests. - ConcatenatedLists<Long> c = new ConcatenatedLists<Long>(); + ConcatenatedLists<Long> c = new ConcatenatedLists<>(); c.addSublist(Arrays.asList(0L, 1L)); try { c.add(2L); @@ -82,19 +82,19 @@ public class TestConcatenatedLists { @Test public void testEmpty() { - verify(new ConcatenatedLists<Long>(), -1); + verify(new ConcatenatedLists<>(), -1); } @Test public void testOneOne() { - ConcatenatedLists<Long> c = new ConcatenatedLists<Long>(); + ConcatenatedLists<Long> c = new ConcatenatedLists<>(); c.addSublist(Arrays.asList(0L)); verify(c, 0); } @Test public void testOneMany() { - ConcatenatedLists<Long> c = new ConcatenatedLists<Long>(); + ConcatenatedLists<Long> c = new ConcatenatedLists<>(); c.addSublist(Arrays.asList(0L, 1L, 2L)); verify(c, 2); } @@ -102,7 +102,7 @@ public class TestConcatenatedLists { @Test @SuppressWarnings("unchecked") public void testManyOne() { - ConcatenatedLists<Long> c = new ConcatenatedLists<Long>(); + ConcatenatedLists<Long> c = new ConcatenatedLists<>(); c.addSublist(Arrays.asList(0L)); c.addAllSublists(Arrays.asList(Arrays.asList(1L), Arrays.asList(2L))); verify(c, 2); @@ -111,7 +111,7 @@ public class TestConcatenatedLists { @Test @SuppressWarnings("unchecked") public void testManyMany() { - ConcatenatedLists<Long> c = new ConcatenatedLists<Long>(); + ConcatenatedLists<Long> c = new ConcatenatedLists<>(); c.addAllSublists(Arrays.asList(Arrays.asList(0L, 1L))); c.addSublist(Arrays.asList(2L, 3L, 4L)); c.addAllSublists(Arrays.asList(Arrays.asList(5L), Arrays.asList(6L, 7L))); @@ -123,7 +123,7 @@ public class TestConcatenatedLists { assertEquals(last + 1, c.size()); assertTrue(c.containsAll(c)); Long[] array = c.toArray(new Long[c.size()]); - List<Long> all = new ArrayList<Long>(); + List<Long> all = new ArrayList<>(); Iterator<Long> iter = c.iterator(); for (Long i = 0L; i <= last; ++i) { assertTrue(iter.hasNext()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestKeyLocker.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestKeyLocker.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestKeyLocker.java index 40b918c..edf2f78 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestKeyLocker.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestKeyLocker.java @@ -30,7 +30,7 @@ import org.junit.experimental.categories.Category; public class TestKeyLocker { @Test public void testLocker(){ - KeyLocker<String> locker = new KeyLocker<String>(); + KeyLocker<String> locker = new KeyLocker<>(); ReentrantLock lock1 = locker.acquireLock("l1"); Assert.assertTrue(lock1.isHeldByCurrentThread()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestLoadTestKVGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestLoadTestKVGenerator.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestLoadTestKVGenerator.java index 120f2b6..cf74a3e 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestLoadTestKVGenerator.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestLoadTestKVGenerator.java @@ -64,7 +64,7 @@ public class TestLoadTestKVGenerator { @Test public void testCorrectAndUniqueKeys() { - Set<String> keys = new HashSet<String>(); + Set<String> keys = new HashSet<>(); for (int i = 0; i < 1000; ++i) { String k = LoadTestKVGenerator.md5PrefixedKey(i); assertFalse(keys.contains(k)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestWeakObjectPool.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestWeakObjectPool.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestWeakObjectPool.java index bf1b4eb..d9fefa2 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestWeakObjectPool.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/util/TestWeakObjectPool.java @@ -35,7 +35,7 @@ public class TestWeakObjectPool { @Before public void setUp() { - pool = new WeakObjectPool<String, Object>( + pool = new WeakObjectPool<>( new WeakObjectPool.ObjectFactory<String, Object>() { @Override public Object createObject(String key) { @@ -94,7 +94,7 @@ public class TestWeakObjectPool { final int THREAD_COUNT = 100; final AtomicBoolean assertionFailed = new AtomicBoolean(); - final AtomicReference<Object> expectedObjRef = new AtomicReference<Object>(); + final AtomicReference<Object> expectedObjRef = new AtomicReference<>(); final CountDownLatch prepareLatch = new CountDownLatch(THREAD_COUNT); final CountDownLatch startLatch = new CountDownLatch(1); final CountDownLatch endLatch = new CountDownLatch(THREAD_COUNT); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java index 304722e..810778b 100644 --- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java +++ b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java @@ -208,7 +208,7 @@ public class AggregationClient implements Closeable { public R call(AggregateService instance) throws IOException { RpcController controller = new AggregationClientRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.getMax(controller, requestArg, rpcCallback); AggregateResponse response = rpcCallback.get(); if (controller.failed()) { @@ -280,7 +280,7 @@ public class AggregationClient implements Closeable { public R call(AggregateService instance) throws IOException { RpcController controller = new AggregationClientRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.getMin(controller, requestArg, rpcCallback); AggregateResponse response = rpcCallback.get(); if (controller.failed()) { @@ -355,7 +355,7 @@ public class AggregationClient implements Closeable { public Long call(AggregateService instance) throws IOException { RpcController controller = new AggregationClientRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.getRowNum(controller, requestArg, rpcCallback); AggregateResponse response = rpcCallback.get(); if (controller.failed()) { @@ -421,7 +421,7 @@ public class AggregationClient implements Closeable { RpcController controller = new AggregationClientRpcController(); // Not sure what is going on here why I have to do these casts. TODO. CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.getSum(controller, requestArg, rpcCallback); AggregateResponse response = rpcCallback.get(); if (controller.failed()) { @@ -472,7 +472,7 @@ public class AggregationClient implements Closeable { Long rowCount = 0l; public synchronized Pair<S, Long> getAvgArgs() { - return new Pair<S, Long>(sum, rowCount); + return new Pair<>(sum, rowCount); } @Override @@ -488,13 +488,13 @@ public class AggregationClient implements Closeable { public Pair<S, Long> call(AggregateService instance) throws IOException { RpcController controller = new AggregationClientRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.getAvg(controller, requestArg, rpcCallback); AggregateResponse response = rpcCallback.get(); if (controller.failed()) { throw new IOException(controller.errorText()); } - Pair<S, Long> pair = new Pair<S, Long>(null, 0L); + Pair<S, Long> pair = new Pair<>(null, 0L); if (response.getFirstPartCount() == 0) { return pair; } @@ -569,10 +569,10 @@ public class AggregationClient implements Closeable { S sumVal = null, sumSqVal = null; public synchronized Pair<List<S>, Long> getStdParams() { - List<S> l = new ArrayList<S>(2); + List<S> l = new ArrayList<>(2); l.add(sumVal); l.add(sumSqVal); - Pair<List<S>, Long> p = new Pair<List<S>, Long>(l, rowCountVal); + Pair<List<S>, Long> p = new Pair<>(l, rowCountVal); return p; } @@ -592,17 +592,17 @@ public class AggregationClient implements Closeable { public Pair<List<S>, Long> call(AggregateService instance) throws IOException { RpcController controller = new AggregationClientRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.getStd(controller, requestArg, rpcCallback); AggregateResponse response = rpcCallback.get(); if (controller.failed()) { throw new IOException(controller.errorText()); } - Pair<List<S>, Long> pair = new Pair<List<S>, Long>(new ArrayList<S>(), 0L); + Pair<List<S>, Long> pair = new Pair<>(new ArrayList<>(), 0L); if (response.getFirstPartCount() == 0) { return pair; } - List<S> list = new ArrayList<S>(); + List<S> list = new ArrayList<>(); for (int i = 0; i < response.getFirstPartCount(); i++) { ByteString b = response.getFirstPart(i); T t = getParsedGenericInstance(ci.getClass(), 4, b); @@ -680,17 +680,15 @@ public class AggregationClient implements Closeable { getMedianArgs(final Table table, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable { final AggregateRequest requestArg = validateArgAndGetPB(scan, ci, false); - final NavigableMap<byte[], List<S>> map = - new TreeMap<byte[], List<S>>(Bytes.BYTES_COMPARATOR); + final NavigableMap<byte[], List<S>> map = new TreeMap<>(Bytes.BYTES_COMPARATOR); class StdCallback implements Batch.Callback<List<S>> { S sumVal = null, sumWeights = null; public synchronized Pair<NavigableMap<byte[], List<S>>, List<S>> getMedianParams() { - List<S> l = new ArrayList<S>(2); + List<S> l = new ArrayList<>(2); l.add(sumVal); l.add(sumWeights); - Pair<NavigableMap<byte[], List<S>>, List<S>> p = - new Pair<NavigableMap<byte[], List<S>>, List<S>>(map, l); + Pair<NavigableMap<byte[], List<S>>, List<S>> p = new Pair<>(map, l); return p; } @@ -708,14 +706,14 @@ public class AggregationClient implements Closeable { public List<S> call(AggregateService instance) throws IOException { RpcController controller = new AggregationClientRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<AggregateResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.getMedian(controller, requestArg, rpcCallback); AggregateResponse response = rpcCallback.get(); if (controller.failed()) { throw new IOException(controller.errorText()); } - List<S> list = new ArrayList<S>(); + List<S> list = new ArrayList<>(); for (int i = 0; i < response.getFirstPartCount(); i++) { ByteString b = response.getFirstPart(i); T t = getParsedGenericInstance(ci.getClass(), 4, b); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java index bccb76a..3fbbd52 100644 --- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java +++ b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java @@ -83,7 +83,7 @@ extends AggregateService implements CoprocessorService, Coprocessor { T temp; Scan scan = ProtobufUtil.toScan(request.getScan()); scanner = env.getRegion().getScanner(scan); - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); byte[] colFamily = scan.getFamilies()[0]; NavigableSet<byte[]> qualifiers = scan.getFamilyMap().get(colFamily); byte[] qualifier = null; @@ -138,7 +138,7 @@ extends AggregateService implements CoprocessorService, Coprocessor { T temp; Scan scan = ProtobufUtil.toScan(request.getScan()); scanner = env.getRegion().getScanner(scan); - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); byte[] colFamily = scan.getFamilies()[0]; NavigableSet<byte[]> qualifiers = scan.getFamilyMap().get(colFamily); byte[] qualifier = null; @@ -198,7 +198,7 @@ extends AggregateService implements CoprocessorService, Coprocessor { if (qualifiers != null && !qualifiers.isEmpty()) { qualifier = qualifiers.pollFirst(); } - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); boolean hasMoreRows = false; do { hasMoreRows = scanner.next(results); @@ -237,7 +237,7 @@ extends AggregateService implements CoprocessorService, Coprocessor { RpcCallback<AggregateResponse> done) { AggregateResponse response = null; long counter = 0l; - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); InternalScanner scanner = null; try { Scan scan = ProtobufUtil.toScan(request.getScan()); @@ -308,7 +308,7 @@ extends AggregateService implements CoprocessorService, Coprocessor { if (qualifiers != null && !qualifiers.isEmpty()) { qualifier = qualifiers.pollFirst(); } - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); boolean hasMoreRows = false; do { @@ -368,7 +368,7 @@ extends AggregateService implements CoprocessorService, Coprocessor { if (qualifiers != null && !qualifiers.isEmpty()) { qualifier = qualifiers.pollFirst(); } - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); boolean hasMoreRows = false; @@ -434,7 +434,7 @@ extends AggregateService implements CoprocessorService, Coprocessor { // if weighted median is requested, get qualifier for the weight column weightQualifier = qualifiers.pollLast(); } - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); boolean hasMoreRows = false; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java index a9d10e8..b52e5f9 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpoint.java @@ -84,7 +84,7 @@ implements Coprocessor, CoprocessorService { InternalScanner scanner = null; try { scanner = this.env.getRegion().getScanner(scan); - List<Cell> curVals = new ArrayList<Cell>(); + List<Cell> curVals = new ArrayList<>(); boolean hasMore = false; do { curVals.clear(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java index 22dac6d..54e3358 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointNullResponse.java @@ -94,7 +94,7 @@ implements Coprocessor, CoprocessorService { return; } scanner = region.getScanner(scan); - List<Cell> curVals = new ArrayList<Cell>(); + List<Cell> curVals = new ArrayList<>(); boolean hasMore = false; do { curVals.clear(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java index c75fb31..6e8c571 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/ColumnAggregationEndpointWithErrors.java @@ -94,7 +94,7 @@ implements Coprocessor, CoprocessorService { throw new DoNotRetryIOException("An expected exception"); } scanner = region.getScanner(scan); - List<Cell> curVals = new ArrayList<Cell>(); + List<Cell> curVals = new ArrayList<>(); boolean hasMore = false; do { curVals.clear(); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java index 3ed8a56..56fdca6 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java @@ -168,8 +168,7 @@ public class TestClassLoading { // verify that the coprocessors were loaded boolean foundTableRegion=false; boolean found1 = true, found2 = true, found2_k1 = true, found2_k2 = true, found2_k3 = true; - Map<Region, Set<ClassLoader>> regionsActiveClassLoaders = - new HashMap<Region, Set<ClassLoader>>(); + Map<Region, Set<ClassLoader>> regionsActiveClassLoaders = new HashMap<>(); MiniHBaseCluster hbase = TEST_UTIL.getHBaseCluster(); for (Region region: hbase.getRegionServer(0).getOnlineRegionsLocalContext()) { @@ -209,7 +208,7 @@ public class TestClassLoading { " of external jar files", 2, CoprocessorClassLoader.getAllCached().size()); //check if region active classloaders are shared across all RS regions - Set<ClassLoader> externalClassLoaders = new HashSet<ClassLoader>( + Set<ClassLoader> externalClassLoaders = new HashSet<>( CoprocessorClassLoader.getAllCached()); for (Map.Entry<Region, Set<ClassLoader>> regionCP : regionsActiveClassLoaders.entrySet()) { assertTrue("Some CP classloaders for region " + regionCP.getKey() + " are not cached." @@ -312,7 +311,7 @@ public class TestClassLoading { // add 2 coprocessor by using new htd.addCoprocessor() api htd.addCoprocessor(cpName5, new Path(getLocalPath(jarFile5)), Coprocessor.PRIORITY_USER, null); - Map<String, String> kvs = new HashMap<String, String>(); + Map<String, String> kvs = new HashMap<>(); kvs.put("k1", "v1"); kvs.put("k2", "v2"); kvs.put("k3", "v3"); @@ -466,8 +465,7 @@ public class TestClassLoading { * @return subset of all servers. */ Map<ServerName, ServerLoad> serversForTable(String tableName) { - Map<ServerName, ServerLoad> serverLoadHashMap = - new HashMap<ServerName, ServerLoad>(); + Map<ServerName, ServerLoad> serverLoadHashMap = new HashMap<>(); for(Map.Entry<ServerName,ServerLoad> server: TEST_UTIL.getMiniHBaseCluster().getMaster().getServerManager(). getOnlineServers().entrySet()) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java index 547b7e9..adfd8d5 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorEndpoint.java @@ -124,7 +124,7 @@ public class TestCoprocessorEndpoint { public Long call(ColumnAggregationProtos.ColumnAggregationService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<ColumnAggregationProtos.SumResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<ColumnAggregationProtos.SumResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); ColumnAggregationProtos.SumRequest.Builder builder = ColumnAggregationProtos.SumRequest.newBuilder(); builder.setFamily(ByteStringer.wrap(family)); @@ -193,7 +193,7 @@ public class TestCoprocessorEndpoint { throws IOException { LOG.debug("Default response is " + TestProtos.EchoRequestProto.getDefaultInstance()); CoprocessorRpcUtils.BlockingRpcCallback<TestProtos.EchoResponseProto> callback = - new CoprocessorRpcUtils.BlockingRpcCallback<TestProtos.EchoResponseProto>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.echo(controller, request, callback); TestProtos.EchoResponseProto response = callback.get(); LOG.debug("Batch.Call returning result " + response); @@ -226,7 +226,7 @@ public class TestCoprocessorEndpoint { throws IOException { LOG.debug("Default response is " + TestProtos.EchoRequestProto.getDefaultInstance()); CoprocessorRpcUtils.BlockingRpcCallback<TestProtos.EchoResponseProto> callback = - new CoprocessorRpcUtils.BlockingRpcCallback<TestProtos.EchoResponseProto>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.echo(controller, request, callback); TestProtos.EchoResponseProto response = callback.get(); LOG.debug("Batch.Call returning result " + response); @@ -271,7 +271,7 @@ public class TestCoprocessorEndpoint { public String call(TestRpcServiceProtos.TestProtobufRpcProto instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<TestProtos.EchoResponseProto> callback = - new CoprocessorRpcUtils.BlockingRpcCallback<TestProtos.EchoResponseProto>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.echo(controller, request, callback); TestProtos.EchoResponseProto response = callback.get(); LOG.debug("Batch.Call got result " + response); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java index 0af655a..0783131 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestCoprocessorTableEndpoint.java @@ -117,7 +117,7 @@ public class TestCoprocessorTableEndpoint { public Long call(ColumnAggregationProtos.ColumnAggregationService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<ColumnAggregationProtos.SumResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<ColumnAggregationProtos.SumResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); ColumnAggregationProtos.SumRequest.Builder builder = ColumnAggregationProtos.SumRequest.newBuilder(); builder.setFamily(ByteString.copyFrom(family)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java index 69742a6..9dc4822 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionServerCoprocessorEndpoint.java @@ -73,8 +73,7 @@ public class TestRegionServerCoprocessorEndpoint { final ServerName serverName = TEST_UTIL.getHBaseCluster().getRegionServer(0).getServerName(); final ServerRpcController controller = new ServerRpcController(); final CoprocessorRpcUtils.BlockingRpcCallback<DummyRegionServerEndpointProtos.DummyResponse> - rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<DummyRegionServerEndpointProtos.DummyResponse>(); + rpcCallback = new CoprocessorRpcUtils.BlockingRpcCallback<>(); DummyRegionServerEndpointProtos.DummyService service = ProtobufUtil.newServiceStub(DummyRegionServerEndpointProtos.DummyService.class, TEST_UTIL.getAdmin().coprocessorService(serverName)); @@ -91,8 +90,7 @@ public class TestRegionServerCoprocessorEndpoint { final ServerName serverName = TEST_UTIL.getHBaseCluster().getRegionServer(0).getServerName(); final ServerRpcController controller = new ServerRpcController(); final CoprocessorRpcUtils.BlockingRpcCallback<DummyRegionServerEndpointProtos.DummyResponse> - rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<DummyRegionServerEndpointProtos.DummyResponse>(); + rpcCallback = new CoprocessorRpcUtils.BlockingRpcCallback<>(); DummyRegionServerEndpointProtos.DummyService service = ProtobufUtil.newServiceStub(DummyRegionServerEndpointProtos.DummyService.class, TEST_UTIL.getAdmin().coprocessorService(serverName)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java index 07d2042..ed53027 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java @@ -166,10 +166,9 @@ public class TestRowProcessorEndpoint { ProcessResponse protoResult = service.process(null, request); FriendsOfFriendsProcessorResponse response = FriendsOfFriendsProcessorResponse.parseFrom(protoResult.getRowProcessorResult()); - Set<String> result = new HashSet<String>(); + Set<String> result = new HashSet<>(); result.addAll(response.getResultList()); - Set<String> expected = - new HashSet<String>(Arrays.asList(new String[]{"d", "e", "f", "g"})); + Set<String> expected = new HashSet<>(Arrays.asList(new String[]{"d", "e", "f", "g"})); Get get = new Get(ROW); LOG.debug("row keyvalues:" + stringifyKvs(table.get(get).listCells())); assertEquals(expected, result); @@ -349,7 +348,7 @@ public class TestRowProcessorEndpoint { public void process(long now, HRegion region, List<Mutation> mutations, WALEdit walEdit) throws IOException { // Scan current counter - List<Cell> kvs = new ArrayList<Cell>(); + List<Cell> kvs = new ArrayList<>(); Scan scan = new Scan(row, row); scan.addColumn(FAM, COUNTER); doScan(region, scan, kvs); @@ -398,7 +397,7 @@ public class TestRowProcessorEndpoint { BaseRowProcessor<FriendsOfFriendsProcessorRequest, FriendsOfFriendsProcessorResponse> { byte[] row = null; byte[] person = null; - final Set<String> result = new HashSet<String>(); + final Set<String> result = new HashSet<>(); /** * Empty constructor for Writable @@ -432,7 +431,7 @@ public class TestRowProcessorEndpoint { @Override public void process(long now, HRegion region, List<Mutation> mutations, WALEdit walEdit) throws IOException { - List<Cell> kvs = new ArrayList<Cell>(); + List<Cell> kvs = new ArrayList<>(); { // First scan to get friends of the person Scan scan = new Scan(row, row); scan.addColumn(FAM, person); @@ -497,7 +496,7 @@ public class TestRowProcessorEndpoint { @Override public Collection<byte[]> getRowsToLock() { - List<byte[]> rows = new ArrayList<byte[]>(2); + List<byte[]> rows = new ArrayList<>(2); rows.add(row1); rows.add(row2); return rows; @@ -522,8 +521,8 @@ public class TestRowProcessorEndpoint { now = myTimer.getAndIncrement(); // Scan both rows - List<Cell> kvs1 = new ArrayList<Cell>(); - List<Cell> kvs2 = new ArrayList<Cell>(); + List<Cell> kvs1 = new ArrayList<>(); + List<Cell> kvs2 = new ArrayList<>(); doScan(region, new Scan(row1, row1), kvs1); doScan(region, new Scan(row2, row2), kvs2); @@ -538,7 +537,7 @@ public class TestRowProcessorEndpoint { swapped = !swapped; // Add and delete keyvalues - List<List<Cell>> kvs = new ArrayList<List<Cell>>(2); + List<List<Cell>> kvs = new ArrayList<>(2); kvs.add(kvs1); kvs.add(kvs2); byte[][] rows = new byte[][]{row1, row2}; http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/SecureBulkLoadEndpointClient.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/SecureBulkLoadEndpointClient.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/SecureBulkLoadEndpointClient.java index 06e45eb..323999d 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/SecureBulkLoadEndpointClient.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/SecureBulkLoadEndpointClient.java @@ -64,7 +64,7 @@ public class SecureBulkLoadEndpointClient { ServerRpcController controller = new ServerRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<PrepareBulkLoadResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PrepareBulkLoadResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); PrepareBulkLoadRequest request = PrepareBulkLoadRequest.newBuilder() @@ -92,7 +92,7 @@ public class SecureBulkLoadEndpointClient { ServerRpcController controller = new ServerRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<CleanupBulkLoadResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<CleanupBulkLoadResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); CleanupBulkLoadRequest request = CleanupBulkLoadRequest.newBuilder() @@ -133,7 +133,7 @@ public class SecureBulkLoadEndpointClient { } List<ClientProtos.BulkLoadHFileRequest.FamilyPath> protoFamilyPaths = - new ArrayList<ClientProtos.BulkLoadHFileRequest.FamilyPath>(familyPaths.size()); + new ArrayList<>(familyPaths.size()); for(Pair<byte[], String> el: familyPaths) { protoFamilyPaths.add(ClientProtos.BulkLoadHFileRequest.FamilyPath.newBuilder() .setFamily(ByteStringer.wrap(el.getFirst())) @@ -148,8 +148,7 @@ public class SecureBulkLoadEndpointClient { ServerRpcController controller = new ServerRpcController(); CoprocessorRpcUtils.BlockingRpcCallback<SecureBulkLoadProtos.SecureBulkLoadHFilesResponse> - rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<SecureBulkLoadProtos.SecureBulkLoadHFilesResponse>(); + rpcCallback = new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.secureBulkLoadHFiles(controller, request, rpcCallback); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoadWithOldSecureEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoadWithOldSecureEndpoint.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoadWithOldSecureEndpoint.java index c0d2719..10a4d19 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoadWithOldSecureEndpoint.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegionServerBulkLoadWithOldSecureEndpoint.java @@ -92,8 +92,7 @@ public class TestHRegionServerBulkLoadWithOldSecureEndpoint extends TestHRegionS // create HFiles for different column families FileSystem fs = UTIL.getTestFileSystem(); byte[] val = Bytes.toBytes(String.format("%010d", iteration)); - final List<Pair<byte[], String>> famPaths = new ArrayList<Pair<byte[], String>>( - NUM_CFS); + final List<Pair<byte[], String>> famPaths = new ArrayList<>(NUM_CFS); for (int i = 0; i < NUM_CFS; i++) { Path hfile = new Path(dir, family(i)); byte[] fam = Bytes.toBytes(family(i)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java index 9bff701..83c7dbf 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/regionserver/TestServerCustomProtocol.java @@ -198,7 +198,7 @@ public class TestServerCustomProtocol { @Override public Integer call(PingProtos.PingService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.CountResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.CountResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.count(null, PingProtos.CountRequest.newBuilder().build(), rpcCallback); return rpcCallback.get().getCount(); } @@ -215,7 +215,7 @@ public class TestServerCustomProtocol { @Override public Integer call(PingProtos.PingService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.IncrementCountResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.IncrementCountResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.increment(null, PingProtos.IncrementCountRequest.newBuilder().setDiff(diff).build(), rpcCallback); @@ -253,7 +253,7 @@ public class TestServerCustomProtocol { @Override public String call(PingProtos.PingService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.HelloResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.HelloResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); PingProtos.HelloRequest.Builder builder = PingProtos.HelloRequest.newBuilder(); if (send != null) builder.setName(send); instance.hello(null, builder.build(), rpcCallback); @@ -272,7 +272,7 @@ public class TestServerCustomProtocol { @Override public String call(PingProtos.PingService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.HelloResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.HelloResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); PingProtos.HelloRequest.Builder builder = PingProtos.HelloRequest.newBuilder(); // Call ping on same instance. Use result calling hello on same instance. builder.setName(doPing(instance)); @@ -291,7 +291,7 @@ public class TestServerCustomProtocol { @Override public String call(PingProtos.PingService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.NoopResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.NoopResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); PingProtos.NoopRequest.Builder builder = PingProtos.NoopRequest.newBuilder(); instance.noop(null, builder.build(), rpcCallback); rpcCallback.get(); @@ -311,7 +311,7 @@ public class TestServerCustomProtocol { @Override public String call(PingProtos.PingService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.PingResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.PingResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.ping(null, PingProtos.PingRequest.newBuilder().build(), rpcCallback); return rpcCallback.get().getPong(); } @@ -406,7 +406,7 @@ public class TestServerCustomProtocol { private static String doPing(PingProtos.PingService instance) throws IOException { CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.PingResponse> rpcCallback = - new CoprocessorRpcUtils.BlockingRpcCallback<PingProtos.PingResponse>(); + new CoprocessorRpcUtils.BlockingRpcCallback<>(); instance.ping(null, PingProtos.PingRequest.newBuilder().build(), rpcCallback); return rpcCallback.get().getPong(); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpToolWithBulkLoadedData.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpToolWithBulkLoadedData.java b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpToolWithBulkLoadedData.java index f54c632..75f8ee2 100644 --- a/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpToolWithBulkLoadedData.java +++ b/hbase-endpoint/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpToolWithBulkLoadedData.java @@ -71,7 +71,7 @@ public class TestReplicationSyncUpToolWithBulkLoadedData extends TestReplication * Prepare 16 random hfile ranges required for creating hfiles */ Iterator<String> randomHFileRangeListIterator = null; - Set<String> randomHFileRanges = new HashSet<String>(16); + Set<String> randomHFileRanges = new HashSet<>(16); for (int i = 0; i < 16; i++) { randomHFileRanges.add(UUID.randomUUID().toString()); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java index 7e6c290..79ff25b 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/BulkDeleteEndpoint.java @@ -133,9 +133,9 @@ public class BulkDeleteEndpoint extends BulkDeleteService implements Coprocessor // filter and having necessary column(s). scanner = region.getScanner(scan); while (hasMore) { - List<List<Cell>> deleteRows = new ArrayList<List<Cell>>(rowBatchSize); + List<List<Cell>> deleteRows = new ArrayList<>(rowBatchSize); for (int i = 0; i < rowBatchSize; i++) { - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); hasMore = scanner.next(results); if (results.size() > 0) { deleteRows.add(results); @@ -202,14 +202,14 @@ public class BulkDeleteEndpoint extends BulkDeleteService implements Coprocessor byte[] row = CellUtil.cloneRow(deleteRow.get(0)); Delete delete = new Delete(row, ts); if (deleteType == DeleteType.FAMILY) { - Set<byte[]> families = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR); + Set<byte[]> families = new TreeSet<>(Bytes.BYTES_COMPARATOR); for (Cell kv : deleteRow) { if (families.add(CellUtil.cloneFamily(kv))) { delete.addFamily(CellUtil.cloneFamily(kv), ts); } } } else if (deleteType == DeleteType.COLUMN) { - Set<Column> columns = new HashSet<Column>(); + Set<Column> columns = new HashSet<>(); for (Cell kv : deleteRow) { Column column = new Column(CellUtil.cloneFamily(kv), CellUtil.cloneQualifier(kv)); if (columns.add(column)) { @@ -231,7 +231,7 @@ public class BulkDeleteEndpoint extends BulkDeleteService implements Coprocessor noOfVersionsToDelete++; } } else { - Set<Column> columns = new HashSet<Column>(); + Set<Column> columns = new HashSet<>(); for (Cell kv : deleteRow) { Column column = new Column(CellUtil.cloneFamily(kv), CellUtil.cloneQualifier(kv)); // Only one version of particular column getting deleted. http://git-wip-us.apache.org/repos/asf/hbase/blob/b53f3547/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java ---------------------------------------------------------------------- diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java index c2387c5..36d8488 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.java @@ -75,7 +75,7 @@ public class RowCountEndpoint extends ExampleProtos.RowCountService InternalScanner scanner = null; try { scanner = env.getRegion().getScanner(scan); - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); boolean hasMore = false; byte[] lastRow = null; long count = 0; @@ -115,7 +115,7 @@ public class RowCountEndpoint extends ExampleProtos.RowCountService InternalScanner scanner = null; try { scanner = env.getRegion().getScanner(new Scan()); - List<Cell> results = new ArrayList<Cell>(); + List<Cell> results = new ArrayList<>(); boolean hasMore = false; long count = 0; do {
