http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/jobhistory/JobHistoryEventHandler.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/jobhistory/JobHistoryEventHandler.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/jobhistory/JobHistoryEventHandler.java index ca204a6..aad63d3 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/jobhistory/JobHistoryEventHandler.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/jobhistory/JobHistoryEventHandler.java @@ -23,7 +23,6 @@ import java.util.Collections; import java.util.EnumSet; import java.util.HashMap; import java.util.Iterator; -import java.util.Locale; import java.util.Map; import java.util.Timer; import java.util.TimerTask; @@ -712,8 +711,7 @@ public class JobHistoryEventHandler extends AbstractService private void processEventForTimelineServer(HistoryEvent event, JobId jobId, long timestamp) { TimelineEvent tEvent = new TimelineEvent(); - tEvent.setEventType( - event.getEventType().name().toUpperCase(Locale.ENGLISH)); + tEvent.setEventType(event.getEventType().name().toUpperCase()); tEvent.setTimestamp(timestamp); TimelineEntity tEntity = new TimelineEntity();
http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/webapp/AppController.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/webapp/AppController.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/webapp/AppController.java index 440ff49..53f21db 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/webapp/AppController.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/webapp/AppController.java @@ -227,7 +227,7 @@ public class AppController extends Controller implements AMParams { try { String tt = $(TASK_TYPE); tt = tt.isEmpty() ? "All" : StringUtils.capitalize(MRApps.taskType(tt). - toString().toLowerCase(Locale.ENGLISH)); + toString().toLowerCase(Locale.US)); setTitle(join(tt, " Tasks for ", $(JOB_ID))); } catch (Exception e) { LOG.error("Failed to render tasks page with task type : " http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/TypeConverter.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/TypeConverter.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/TypeConverter.java index e0c4773..553ba70 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/TypeConverter.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/TypeConverter.java @@ -21,7 +21,6 @@ package org.apache.hadoop.mapreduce; import java.util.ArrayList; import java.util.HashMap; import java.util.List; -import java.util.Locale; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.mapred.JobPriority; @@ -315,7 +314,7 @@ public class TypeConverter { QueueState state) { org.apache.hadoop.mapreduce.QueueState qState = org.apache.hadoop.mapreduce.QueueState.getState( - state.toString().toLowerCase(Locale.ENGLISH)); + state.toString().toLowerCase()); return qState; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java index 37cfb7a..08b44f8 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java @@ -30,7 +30,6 @@ import java.util.ArrayList; import java.util.Arrays; import java.util.HashMap; import java.util.List; -import java.util.Locale; import java.util.Map; import com.google.common.annotations.VisibleForTesting; @@ -304,7 +303,7 @@ public class MRApps extends Apps { remoteFS.getWorkingDirectory())); String name = (null == u.getFragment()) ? p.getName() : u.getFragment(); - if (!name.toLowerCase(Locale.ENGLISH).endsWith(".jar")) { + if (!name.toLowerCase().endsWith(".jar")) { linkLookup.put(p, name); } } @@ -318,7 +317,7 @@ public class MRApps extends Apps { if (name == null) { name = p.getName(); } - if(!name.toLowerCase(Locale.ENGLISH).endsWith(".jar")) { + if(!name.toLowerCase().endsWith(".jar")) { MRApps.addToEnvironment( environment, classpathEnvVar, http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/TestTypeConverter.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/TestTypeConverter.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/TestTypeConverter.java index d360811..cc42b9c 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/TestTypeConverter.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/test/java/org/apache/hadoop/mapreduce/TestTypeConverter.java @@ -22,7 +22,6 @@ import static org.mockito.Mockito.when; import java.util.ArrayList; import java.util.List; -import java.util.Locale; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.mapreduce.JobStatus.State; @@ -154,8 +153,7 @@ public class TestTypeConverter { org.apache.hadoop.mapreduce.QueueInfo returned = TypeConverter.fromYarn(queueInfo, new Configuration()); Assert.assertEquals("queueInfo translation didn't work.", - returned.getState().toString(), - queueInfo.getQueueState().toString().toLowerCase(Locale.ENGLISH)); + returned.getState().toString(), queueInfo.getQueueState().toString().toLowerCase()); } /** http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Task.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Task.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Task.java index 73dab4f..5274438 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Task.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Task.java @@ -28,7 +28,6 @@ import java.util.ArrayList; import java.util.HashMap; import java.util.Iterator; import java.util.List; -import java.util.Locale; import java.util.Map; import java.util.NoSuchElementException; import java.util.concurrent.atomic.AtomicBoolean; @@ -117,7 +116,7 @@ abstract public class Task implements Writable, Configurable { * BYTES_READ counter and second one is of the BYTES_WRITTEN counter. */ protected static String[] getFileSystemCounterNames(String uriScheme) { - String scheme = uriScheme.toUpperCase(Locale.ENGLISH); + String scheme = uriScheme.toUpperCase(); return new String[]{scheme+"_BYTES_READ", scheme+"_BYTES_WRITTEN"}; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/counters/FileSystemCounterGroup.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/counters/FileSystemCounterGroup.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/counters/FileSystemCounterGroup.java index e8d5d1d..a53b76a 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/counters/FileSystemCounterGroup.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/counters/FileSystemCounterGroup.java @@ -227,7 +227,7 @@ public abstract class FileSystemCounterGroup<C extends Counter> } private String checkScheme(String scheme) { - String fixed = scheme.toUpperCase(Locale.ENGLISH); + String fixed = scheme.toUpperCase(Locale.US); String interned = schemes.putIfAbsent(fixed, fixed); if (schemes.size() > MAX_NUM_SCHEMES) { // mistakes or abuses http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java index ef9b0a4..eaa5af8 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java @@ -473,7 +473,7 @@ public class DistributedCache { if (fragment == null) { return false; } - String lowerCaseFragment = fragment.toLowerCase(Locale.ENGLISH); + String lowerCaseFragment = fragment.toLowerCase(); if (fragments.contains(lowerCaseFragment)) { return false; } @@ -488,7 +488,7 @@ public class DistributedCache { if (fragment == null) { return false; } - String lowerCaseFragment = fragment.toLowerCase(Locale.ENGLISH); + String lowerCaseFragment = fragment.toLowerCase(); if (fragments.contains(lowerCaseFragment)) { return false; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java index 7faa736..00fbeda 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java @@ -29,7 +29,6 @@ import java.sql.SQLException; import java.sql.Statement; import java.util.ArrayList; import java.util.List; -import java.util.Locale; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -163,8 +162,7 @@ public class DBInputFormat<T extends DBWritable> this.connection = createConnection(); DatabaseMetaData dbMeta = connection.getMetaData(); - this.dbProductName = - dbMeta.getDatabaseProductName().toUpperCase(Locale.ENGLISH); + this.dbProductName = dbMeta.getDatabaseProductName().toUpperCase(); } catch (Exception ex) { throw new RuntimeException(ex); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/tools/CLI.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/tools/CLI.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/tools/CLI.java index 6af815b..37ba5b7 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/tools/CLI.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/tools/CLI.java @@ -22,7 +22,6 @@ import java.io.OutputStreamWriter; import java.io.PrintWriter; import java.util.ArrayList; import java.util.List; -import java.util.Locale; import java.util.Set; import java.util.HashSet; import java.util.Arrays; @@ -223,12 +222,12 @@ public class CLI extends Configured implements Tool { taskType = argv[2]; taskState = argv[3]; displayTasks = true; - if (!taskTypes.contains(taskType.toUpperCase(Locale.ENGLISH))) { + if (!taskTypes.contains(taskType.toUpperCase())) { System.out.println("Error: Invalid task-type: " + taskType); displayUsage(cmd); return exitCode; } - if (!taskStates.contains(taskState.toLowerCase(Locale.ENGLISH))) { + if (!taskStates.contains(taskState.toLowerCase())) { System.out.println("Error: Invalid task-state: " + taskState); displayUsage(cmd); return exitCode; @@ -594,8 +593,7 @@ public class CLI extends Configured implements Tool { throws IOException, InterruptedException { TaskReport[] reports=null; - reports = job.getTaskReports( - TaskType.valueOf(type.toUpperCase(Locale.ENGLISH))); + reports = job.getTaskReports(TaskType.valueOf(type.toUpperCase())); for (TaskReport report : reports) { TIPStatus status = report.getCurrentStatus(); if ((state.equalsIgnoreCase("pending") && status ==TIPStatus.PENDING) || http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestDFSIO.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestDFSIO.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestDFSIO.java index 9420497..d9cd07b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestDFSIO.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestDFSIO.java @@ -29,7 +29,6 @@ import java.io.InputStreamReader; import java.io.OutputStream; import java.io.PrintStream; import java.util.Date; -import java.util.Locale; import java.util.Random; import java.util.StringTokenizer; import org.apache.commons.logging.Log; @@ -155,16 +154,16 @@ public class TestDFSIO implements Tool { static ByteMultiple parseString(String sMultiple) { if(sMultiple == null || sMultiple.isEmpty()) // MB by default return MB; - String sMU = sMultiple.toUpperCase(Locale.ENGLISH); - if(B.name().toUpperCase(Locale.ENGLISH).endsWith(sMU)) + String sMU = sMultiple.toUpperCase(); + if(B.name().toUpperCase().endsWith(sMU)) return B; - if(KB.name().toUpperCase(Locale.ENGLISH).endsWith(sMU)) + if(KB.name().toUpperCase().endsWith(sMU)) return KB; - if(MB.name().toUpperCase(Locale.ENGLISH).endsWith(sMU)) + if(MB.name().toUpperCase().endsWith(sMU)) return MB; - if(GB.name().toUpperCase(Locale.ENGLISH).endsWith(sMU)) + if(GB.name().toUpperCase().endsWith(sMU)) return GB; - if(TB.name().toUpperCase(Locale.ENGLISH).endsWith(sMU)) + if(TB.name().toUpperCase().endsWith(sMU)) return TB; throw new IllegalArgumentException("Unsupported ByteMultiple "+sMultiple); } @@ -737,7 +736,7 @@ public class TestDFSIO implements Tool { } for (int i = 0; i < args.length; i++) { // parse command line - if (args[i].toLowerCase(Locale.ENGLISH).startsWith("-read")) { + if (args[i].toLowerCase().startsWith("-read")) { testType = TestType.TEST_TYPE_READ; } else if (args[i].equalsIgnoreCase("-write")) { testType = TestType.TEST_TYPE_WRITE; @@ -756,10 +755,9 @@ public class TestDFSIO implements Tool { testType = TestType.TEST_TYPE_TRUNCATE; } else if (args[i].equalsIgnoreCase("-clean")) { testType = TestType.TEST_TYPE_CLEANUP; - } else if (args[i].toLowerCase(Locale.ENGLISH).startsWith("-seq")) { + } else if (args[i].toLowerCase().startsWith("-seq")) { isSequential = true; - } else if ( - args[i].toLowerCase(Locale.ENGLISH).startsWith("-compression")) { + } else if (args[i].toLowerCase().startsWith("-compression")) { compressionClass = args[++i]; } else if (args[i].equalsIgnoreCase("-nrfiles")) { nrFiles = Integer.parseInt(args[++i]); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestFileSystem.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestFileSystem.java index e2b8985..13e27cd 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestFileSystem.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/TestFileSystem.java @@ -24,7 +24,6 @@ import java.io.OutputStream; import java.security.PrivilegedExceptionAction; import java.util.Arrays; import java.util.Collections; -import java.util.Locale; import java.util.Random; import java.util.List; import java.util.ArrayList; @@ -557,10 +556,7 @@ public class TestFileSystem extends TestCase { static void checkPath(MiniDFSCluster cluster, FileSystem fileSys) throws IOException { InetSocketAddress add = cluster.getNameNode().getNameNodeAddress(); // Test upper/lower case - fileSys.checkPath( - new Path("hdfs://" - + add.getHostName().toUpperCase(Locale.ENGLISH) - + ":" + add.getPort())); + fileSys.checkPath(new Path("hdfs://" + add.getHostName().toUpperCase() + ":" + add.getPort())); } public void testFsClose() throws Exception { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/Constants.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/Constants.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/Constants.java index be2648e..0642052 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/Constants.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/Constants.java @@ -18,8 +18,6 @@ package org.apache.hadoop.fs.slive; -import java.util.Locale; - /** * Constants used in various places in slive */ @@ -37,7 +35,7 @@ class Constants { enum Distribution { BEG, END, UNIFORM, MID; String lowerName() { - return this.name().toLowerCase(Locale.ENGLISH); + return this.name().toLowerCase(); } } @@ -47,7 +45,7 @@ class Constants { enum OperationType { READ, APPEND, RENAME, LS, MKDIR, DELETE, CREATE, TRUNCATE; String lowerName() { - return this.name().toLowerCase(Locale.ENGLISH); + return this.name().toLowerCase(); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationData.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationData.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationData.java index 462db86..b4c98f7 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationData.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationData.java @@ -18,7 +18,6 @@ package org.apache.hadoop.fs.slive; -import java.util.Locale; import org.apache.hadoop.fs.slive.Constants.Distribution; /** @@ -53,8 +52,7 @@ class OperationData { percent = (Double.parseDouble(pieces[0]) / 100.0d); } else if (pieces.length >= 2) { percent = (Double.parseDouble(pieces[0]) / 100.0d); - distribution = - Distribution.valueOf(pieces[1].toUpperCase(Locale.ENGLISH)); + distribution = Distribution.valueOf(pieces[1].toUpperCase()); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationOutput.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationOutput.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationOutput.java index d60a607..57ef017 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationOutput.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/OperationOutput.java @@ -18,7 +18,6 @@ package org.apache.hadoop.fs.slive; -import java.util.Locale; import org.apache.hadoop.io.Text; /** @@ -68,8 +67,7 @@ class OperationOutput { "Invalid key format - no type seperator - " + TYPE_SEP); } try { - dataType = OutputType.valueOf( - key.substring(0, place).toUpperCase(Locale.ENGLISH)); + dataType = OutputType.valueOf(key.substring(0, place).toUpperCase()); } catch (Exception e) { throw new IllegalArgumentException( "Invalid key format - invalid output type", e); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/SliveTest.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/SliveTest.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/SliveTest.java index 443089d..ce1837f 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/SliveTest.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/slive/SliveTest.java @@ -27,7 +27,6 @@ import java.io.InputStreamReader; import java.io.PrintWriter; import java.util.ArrayList; import java.util.List; -import java.util.Locale; import java.util.Map; import java.util.TreeMap; @@ -158,7 +157,7 @@ public class SliveTest implements Tool { if (val == null) { return false; } - String cleanupOpt = val.toLowerCase(Locale.ENGLISH).trim(); + String cleanupOpt = val.toLowerCase().trim(); if (cleanupOpt.equals("true") || cleanupOpt.equals("1")) { return true; } else { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/io/FileBench.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/io/FileBench.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/io/FileBench.java index 0830f37..f155dae 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/io/FileBench.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/io/FileBench.java @@ -23,7 +23,6 @@ import java.util.ArrayList; import java.util.Date; import java.util.EnumSet; import java.util.HashMap; -import java.util.Locale; import java.util.Map; import java.util.Random; @@ -215,25 +214,23 @@ public class FileBench extends Configured implements Tool { if (!(fmt == Format.txt || cod == CCodec.pln)) { for (CType typ : ct) { String fn = - fmt.name().toUpperCase(Locale.ENGLISH) + "_" + - cod.name().toUpperCase(Locale.ENGLISH) + "_" + - typ.name().toUpperCase(Locale.ENGLISH); + fmt.name().toUpperCase() + "_" + + cod.name().toUpperCase() + "_" + + typ.name().toUpperCase(); typ.configure(job); - System.out.print( - rwop.name().toUpperCase(Locale.ENGLISH) + " " + fn + ": "); + System.out.print(rwop.name().toUpperCase() + " " + fn + ": "); System.out.println(rwop.exec(fn, job) / 1000 + " seconds"); } } else { String fn = - fmt.name().toUpperCase(Locale.ENGLISH) + "_" + - cod.name().toUpperCase(Locale.ENGLISH); + fmt.name().toUpperCase() + "_" + + cod.name().toUpperCase(); Path p = new Path(root, fn); if (rwop == RW.r && !fs.exists(p)) { fn += cod.getExt(); } - System.out.print( - rwop.name().toUpperCase(Locale.ENGLISH) + " " + fn + ": "); + System.out.print(rwop.name().toUpperCase() + " " + fn + ": "); System.out.println(rwop.exec(fn, job) / 1000 + " seconds"); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMapRed.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMapRed.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMapRed.java index b9bbd60..02a083b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMapRed.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMapRed.java @@ -28,7 +28,6 @@ import java.io.File; import java.util.EnumSet; import java.util.HashSet; import java.util.Iterator; -import java.util.Locale; import java.util.Random; import org.apache.hadoop.conf.Configuration; @@ -281,7 +280,7 @@ public class TestMapRed extends Configured implements Tool { public void map(WritableComparable key, Text value, OutputCollector<Text, Text> output, Reporter reporter) throws IOException { - String str = value.toString().toLowerCase(Locale.ENGLISH); + String str = value.toString().toLowerCase(); output.collect(new Text(str), value); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java b/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java index 4379cd5..270ddc9 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java @@ -27,7 +27,6 @@ import java.sql.PreparedStatement; import java.sql.ResultSet; import java.sql.SQLException; import java.sql.Statement; -import java.util.Locale; import java.util.Random; import org.apache.commons.logging.Log; @@ -103,7 +102,7 @@ public class DBCountPageView extends Configured implements Tool { private void createConnection(String driverClassName , String url) throws Exception { - if(driverClassName.toLowerCase(Locale.ENGLISH).contains("oracle")) { + if(driverClassName.toLowerCase().contains("oracle")) { isOracle = true; } Class.forName(driverClassName); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-maven-plugins/src/main/java/org/apache/hadoop/maven/plugin/versioninfo/VersionInfoMojo.java ---------------------------------------------------------------------- diff --git a/hadoop-maven-plugins/src/main/java/org/apache/hadoop/maven/plugin/versioninfo/VersionInfoMojo.java b/hadoop-maven-plugins/src/main/java/org/apache/hadoop/maven/plugin/versioninfo/VersionInfoMojo.java index f0ec59c..f342463 100644 --- a/hadoop-maven-plugins/src/main/java/org/apache/hadoop/maven/plugin/versioninfo/VersionInfoMojo.java +++ b/hadoop-maven-plugins/src/main/java/org/apache/hadoop/maven/plugin/versioninfo/VersionInfoMojo.java @@ -35,7 +35,6 @@ import java.util.Collections; import java.util.Comparator; import java.util.Date; import java.util.List; -import java.util.Locale; import java.util.TimeZone; /** @@ -330,8 +329,7 @@ public class VersionInfoMojo extends AbstractMojo { } private String normalizePath(File file) { - return file.getPath().toUpperCase(Locale.ENGLISH) - .replaceAll("\\\\", "/"); + return file.getPath().toUpperCase().replaceAll("\\\\", "/"); } }); byte[] md5 = computeMD5(files); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java index bd8ca6e..2412698 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java @@ -979,8 +979,8 @@ public class AzureNativeFileSystemStore implements NativeFileSystemStore { private String verifyAndConvertToStandardFormat(String rawDir) throws URISyntaxException { URI asUri = new URI(rawDir); if (asUri.getAuthority() == null - || asUri.getAuthority().toLowerCase(Locale.ENGLISH).equalsIgnoreCase( - sessionUri.getAuthority().toLowerCase(Locale.ENGLISH))) { + || asUri.getAuthority().toLowerCase(Locale.US).equalsIgnoreCase( + sessionUri.getAuthority().toLowerCase(Locale.US))) { // Applies to me. return trim(asUri.getPath(), "/"); } else { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java index 36873c7..71e84a1 100644 --- a/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java +++ b/hadoop-tools/hadoop-distcp/src/main/java/org/apache/hadoop/tools/util/DistCpUtils.java @@ -121,9 +121,8 @@ public class DistCpUtils { */ public static Class<? extends InputFormat> getStrategy(Configuration conf, DistCpOptions options) { - String confLabel = "distcp." - + options.getCopyStrategy().toLowerCase(Locale.ENGLISH) - + ".strategy.impl"; + String confLabel = "distcp." + + options.getCopyStrategy().toLowerCase(Locale.getDefault()) + ".strategy.impl"; return conf.getClass(confLabel, UniformSizeInputFormat.class, InputFormat.class); } @@ -222,8 +221,7 @@ public class DistCpUtils { final boolean preserveXAttrs = attributes.contains(FileAttribute.XATTR); if (preserveXAttrs || preserveRawXattrs) { - final String rawNS = - XAttr.NameSpace.RAW.name().toLowerCase(Locale.ENGLISH); + final String rawNS = XAttr.NameSpace.RAW.name().toLowerCase(); Map<String, byte[]> srcXAttrs = srcFileStatus.getXAttrs(); Map<String, byte[]> targetXAttrs = getXAttrs(targetFS, path); if (srcXAttrs != null && !srcXAttrs.equals(targetXAttrs)) { @@ -323,8 +321,7 @@ public class DistCpUtils { copyListingFileStatus.setXAttrs(srcXAttrs); } else { Map<String, byte[]> trgXAttrs = Maps.newHashMap(); - final String rawNS = - XAttr.NameSpace.RAW.name().toLowerCase(Locale.ENGLISH); + final String rawNS = XAttr.NameSpace.RAW.name().toLowerCase(); for (Map.Entry<String, byte[]> ent : srcXAttrs.entrySet()) { final String xattrName = ent.getKey(); if (xattrName.startsWith(rawNS)) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-extras/src/main/java/org/apache/hadoop/tools/DistCpV1.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-extras/src/main/java/org/apache/hadoop/tools/DistCpV1.java b/hadoop-tools/hadoop-extras/src/main/java/org/apache/hadoop/tools/DistCpV1.java index 1db17f7..f46c421 100644 --- a/hadoop-tools/hadoop-extras/src/main/java/org/apache/hadoop/tools/DistCpV1.java +++ b/hadoop-tools/hadoop-extras/src/main/java/org/apache/hadoop/tools/DistCpV1.java @@ -31,7 +31,6 @@ import java.util.HashSet; import java.util.Iterator; import java.util.LinkedList; import java.util.List; -import java.util.Locale; import java.util.Random; import java.util.Stack; import java.util.StringTokenizer; @@ -170,9 +169,7 @@ public class DistCpV1 implements Tool { final char symbol; - private FileAttribute() { - symbol = toString().toLowerCase(Locale.ENGLISH).charAt(0); - } + private FileAttribute() {symbol = toString().toLowerCase().charAt(0);} static EnumSet<FileAttribute> parse(String s) { if (s == null || s.length() == 0) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/GridmixJobSubmissionPolicy.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/GridmixJobSubmissionPolicy.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/GridmixJobSubmissionPolicy.java index b010725..83eb947 100644 --- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/GridmixJobSubmissionPolicy.java +++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/GridmixJobSubmissionPolicy.java @@ -24,7 +24,6 @@ import org.apache.hadoop.mapred.gridmix.Statistics.JobStats; import org.apache.hadoop.mapred.gridmix.Statistics.ClusterStats; import java.util.concurrent.CountDownLatch; -import java.util.Locale; import java.io.IOException; enum GridmixJobSubmissionPolicy { @@ -85,6 +84,6 @@ enum GridmixJobSubmissionPolicy { public static GridmixJobSubmissionPolicy getPolicy( Configuration conf, GridmixJobSubmissionPolicy defaultPolicy) { String policy = conf.get(JOB_SUBMISSION_POLICY, defaultPolicy.name()); - return valueOf(policy.toUpperCase(Locale.ENGLISH)); + return valueOf(policy.toUpperCase()); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/HadoopLogsAnalyzer.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/HadoopLogsAnalyzer.java b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/HadoopLogsAnalyzer.java index 776391d..47fdb1a 100644 --- a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/HadoopLogsAnalyzer.java +++ b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/HadoopLogsAnalyzer.java @@ -28,7 +28,6 @@ import java.util.Arrays; import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; -import java.util.Locale; import java.util.Map; import java.util.StringTokenizer; import java.util.ArrayList; @@ -320,43 +319,42 @@ public class HadoopLogsAnalyzer extends Configured implements Tool { } for (int i = 0; i < args.length - (inputFilename == null ? 0 : 1); ++i) { - if ("-h".equals(args[i].toLowerCase(Locale.ENGLISH)) - || "-help".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-h".equals(args[i].toLowerCase()) + || "-help".equals(args[i].toLowerCase())) { usage(); return 0; } - if ("-c".equals(args[i].toLowerCase(Locale.ENGLISH)) - || "-collect-prefixes".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-c".equals(args[i].toLowerCase()) + || "-collect-prefixes".equals(args[i].toLowerCase())) { collecting = true; continue; } // these control the job digest - if ("-write-job-trace".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-write-job-trace".equals(args[i].toLowerCase())) { ++i; jobTraceFilename = new Path(args[i]); continue; } - if ("-single-line-job-traces".equals( - args[i].toLowerCase(Locale.ENGLISH))) { + if ("-single-line-job-traces".equals(args[i].toLowerCase())) { prettyprintTrace = false; continue; } - if ("-omit-task-details".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-omit-task-details".equals(args[i].toLowerCase())) { omitTaskDetails = true; continue; } - if ("-write-topology".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-write-topology".equals(args[i].toLowerCase())) { ++i; topologyFilename = new Path(args[i]); continue; } - if ("-job-digest-spectra".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-job-digest-spectra".equals(args[i].toLowerCase())) { ArrayList<Integer> values = new ArrayList<Integer>(); ++i; @@ -386,13 +384,13 @@ public class HadoopLogsAnalyzer extends Configured implements Tool { continue; } - if ("-d".equals(args[i].toLowerCase(Locale.ENGLISH)) - || "-debug".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-d".equals(args[i].toLowerCase()) + || "-debug".equals(args[i].toLowerCase())) { debug = true; continue; } - if ("-spreads".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-spreads".equals(args[i].toLowerCase())) { int min = Integer.parseInt(args[i + 1]); int max = Integer.parseInt(args[i + 2]); @@ -406,22 +404,22 @@ public class HadoopLogsAnalyzer extends Configured implements Tool { } // These control log-wide CDF outputs - if ("-delays".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-delays".equals(args[i].toLowerCase())) { delays = true; continue; } - if ("-runtimes".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-runtimes".equals(args[i].toLowerCase())) { runtimes = true; continue; } - if ("-tasktimes".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-tasktimes".equals(args[i].toLowerCase())) { collectTaskTimes = true; continue; } - if ("-v1".equals(args[i].toLowerCase(Locale.ENGLISH))) { + if ("-v1".equals(args[i].toLowerCase())) { version = 1; continue; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/JobBuilder.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/JobBuilder.java b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/JobBuilder.java index 4b512d2..eaa9547 100644 --- a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/JobBuilder.java +++ b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/JobBuilder.java @@ -20,7 +20,6 @@ package org.apache.hadoop.tools.rumen; import java.util.ArrayList; import java.util.HashMap; import java.util.List; -import java.util.Locale; import java.util.Map; import java.util.Properties; import java.util.StringTokenizer; @@ -434,7 +433,7 @@ public class JobBuilder { return Values.SUCCESS; } - return Values.valueOf(name.toUpperCase(Locale.ENGLISH)); + return Values.valueOf(name.toUpperCase()); } private void processTaskUpdatedEvent(TaskUpdatedEvent event) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTask.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTask.java b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTask.java index 068ac5b..903d5fb 100644 --- a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTask.java +++ b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTask.java @@ -20,7 +20,6 @@ package org.apache.hadoop.tools.rumen; import java.util.ArrayList; import java.util.Collections; import java.util.List; -import java.util.Locale; import java.util.Set; import java.util.TreeSet; @@ -244,7 +243,7 @@ public class LoggedTask implements DeepCompare { } private static String canonicalizeCounterName(String nonCanonicalName) { - String result = nonCanonicalName.toLowerCase(Locale.ENGLISH); + String result = nonCanonicalName.toLowerCase(); result = result.replace(' ', '|'); result = result.replace('-', '|'); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTaskAttempt.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTaskAttempt.java b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTaskAttempt.java index c7823c4..d1b365e 100644 --- a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTaskAttempt.java +++ b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/LoggedTaskAttempt.java @@ -20,7 +20,6 @@ package org.apache.hadoop.tools.rumen; import java.util.ArrayList; import java.util.List; -import java.util.Locale; import java.util.Set; import java.util.TreeSet; @@ -612,7 +611,7 @@ public class LoggedTaskAttempt implements DeepCompare { } private static String canonicalizeCounterName(String nonCanonicalName) { - String result = nonCanonicalName.toLowerCase(Locale.ENGLISH); + String result = nonCanonicalName.toLowerCase(); result = result.replace(' ', '|'); result = result.replace('-', '|'); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/Environment.java ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/Environment.java b/hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/Environment.java index 72ff3b0..98d8aa03 100644 --- a/hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/Environment.java +++ b/hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/Environment.java @@ -43,7 +43,7 @@ public class Environment extends Properties { // http://lopica.sourceforge.net/os.html String command = null; String OS = System.getProperty("os.name"); - String lowerOs = OS.toLowerCase(Locale.ENGLISH); + String lowerOs = OS.toLowerCase(); if (OS.indexOf("Windows") > -1) { command = "cmd /C set"; } else if (lowerOs.indexOf("ix") > -1 || lowerOs.indexOf("linux") > -1 http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/ApplicationCLI.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/ApplicationCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/ApplicationCLI.java index 1cb3e58..de8f740 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/ApplicationCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/ApplicationCLI.java @@ -26,7 +26,6 @@ import java.text.DecimalFormat; import java.util.EnumSet; import java.util.HashSet; import java.util.List; -import java.util.Locale; import java.util.Set; import org.apache.commons.cli.CommandLine; @@ -174,7 +173,7 @@ public class ApplicationCLI extends YarnCLI { if (types != null) { for (String type : types) { if (!type.trim().isEmpty()) { - appTypes.add(type.toUpperCase(Locale.ENGLISH).trim()); + appTypes.add(type.toUpperCase().trim()); } } } @@ -193,7 +192,7 @@ public class ApplicationCLI extends YarnCLI { } try { appStates.add(YarnApplicationState.valueOf(state - .toUpperCase(Locale.ENGLISH).trim())); + .toUpperCase().trim())); } catch (IllegalArgumentException ex) { sysout.println("The application state " + state + " is invalid."); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/NodeCLI.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/NodeCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/NodeCLI.java index ce1c0fd..d603626 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/NodeCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/NodeCLI.java @@ -27,7 +27,6 @@ import java.util.Collections; import java.util.Date; import java.util.HashSet; import java.util.List; -import java.util.Locale; import java.util.Set; import org.apache.commons.cli.CommandLine; @@ -112,8 +111,7 @@ public class NodeCLI extends YarnCLI { if (types != null) { for (String type : types) { if (!type.trim().isEmpty()) { - nodeStates.add( - NodeState.valueOf(type.trim().toUpperCase(Locale.ENGLISH))); + nodeStates.add(NodeState.valueOf(type.trim().toUpperCase())); } } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/protocolrecords/impl/pb/GetApplicationsRequestPBImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/protocolrecords/impl/pb/GetApplicationsRequestPBImpl.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/protocolrecords/impl/pb/GetApplicationsRequestPBImpl.java index 0b39dfe..a8996f0 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/protocolrecords/impl/pb/GetApplicationsRequestPBImpl.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/protocolrecords/impl/pb/GetApplicationsRequestPBImpl.java @@ -21,7 +21,6 @@ package org.apache.hadoop.yarn.api.protocolrecords.impl.pb; import java.util.EnumSet; import java.util.HashSet; import java.util.List; -import java.util.Locale; import java.util.Set; import org.apache.commons.lang.math.LongRange; @@ -214,7 +213,7 @@ public class GetApplicationsRequestPBImpl extends GetApplicationsRequest { // Convert applicationTags to lower case and add this.applicationTags = new HashSet<String>(); for (String tag : tags) { - this.applicationTags.add(tag.toLowerCase(Locale.ENGLISH)); + this.applicationTags.add(tag.toLowerCase()); } } @@ -259,8 +258,7 @@ public class GetApplicationsRequestPBImpl extends GetApplicationsRequest { public void setApplicationStates(Set<String> applicationStates) { EnumSet<YarnApplicationState> appStates = null; for (YarnApplicationState state : YarnApplicationState.values()) { - if (applicationStates.contains( - state.name().toLowerCase(Locale.ENGLISH))) { + if (applicationStates.contains(state.name().toLowerCase())) { if (appStates == null) { appStates = EnumSet.of(state); } else { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/records/impl/pb/ApplicationSubmissionContextPBImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/records/impl/pb/ApplicationSubmissionContextPBImpl.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/records/impl/pb/ApplicationSubmissionContextPBImpl.java index fe89f81..303b437 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/records/impl/pb/ApplicationSubmissionContextPBImpl.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/records/impl/pb/ApplicationSubmissionContextPBImpl.java @@ -19,7 +19,6 @@ package org.apache.hadoop.yarn.api.records.impl.pb; import java.util.HashSet; -import java.util.Locale; import java.util.Set; import org.apache.hadoop.classification.InterfaceAudience.Private; @@ -292,7 +291,7 @@ extends ApplicationSubmissionContext { // Convert applicationTags to lower case and add this.applicationTags = new HashSet<String>(); for (String tag : tags) { - this.applicationTags.add(tag.toLowerCase(Locale.ENGLISH)); + this.applicationTags.add(tag.toLowerCase()); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/HamletGen.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/HamletGen.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/HamletGen.java index 45d9b61..c848828 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/HamletGen.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/hamlet/HamletGen.java @@ -241,8 +241,7 @@ public class HamletGen { puts(indent, "\n", "private <T extends _> ", retName, "<T> ", methodName, "_(T e, boolean inline) {\n", - " return new ", retName, "<T>(\"", - retName.toLowerCase(Locale.ENGLISH), + " return new ", retName, "<T>(\"", retName.toLowerCase(Locale.US), "\", e, opt(", !endTagOptional.contains(retName), ", inline, ", retName.equals("PRE"), ")); }"); } @@ -259,7 +258,7 @@ public class HamletGen { puts(0, ") {"); puts(indent, topMode ? "" : " closeAttrs();\n", - " return ", retName.toLowerCase(Locale.ENGLISH), "_(this, ", + " return ", retName.toLowerCase(Locale.US), "_(this, ", isInline(className, retName), ");\n", "}"); } else if (params.length == 1) { puts(0, "String selector) {"); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java index 4140df4..2040f57 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/webapp/AHSWebServices.java @@ -19,7 +19,6 @@ package org.apache.hadoop.yarn.server.applicationhistoryservice.webapp; import java.util.Collections; -import java.util.Locale; import java.util.Set; import javax.servlet.http.HttpServletRequest; @@ -148,8 +147,7 @@ public class AHSWebServices extends WebServices { } Set<String> appStates = parseQueries(statesQuery, true); for (String appState : appStates) { - switch (YarnApplicationState.valueOf( - appState.toUpperCase(Locale.ENGLISH))) { + switch (YarnApplicationState.valueOf(appState.toUpperCase())) { case FINISHED: case FAILED: case KILLED: http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java index 44ba2b9..0907f2c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/timeline/webapp/TimelineWebServices.java @@ -24,7 +24,6 @@ import java.util.Collection; import java.util.EnumSet; import java.util.HashSet; import java.util.List; -import java.util.Locale; import java.util.Set; import java.util.SortedSet; import java.util.TreeSet; @@ -418,7 +417,7 @@ public class TimelineWebServices { String[] strs = str.split(delimiter); List<Field> fieldList = new ArrayList<Field>(); for (String s : strs) { - s = s.trim().toUpperCase(Locale.ENGLISH); + s = s.trim().toUpperCase(); if (s.equals("EVENTS")) { fieldList.add(Field.EVENTS); } else if (s.equals("LASTEVENTONLY")) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/java/org/apache/hadoop/yarn/server/webapp/WebServices.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/java/org/apache/hadoop/yarn/server/webapp/WebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/java/org/apache/hadoop/yarn/server/webapp/WebServices.java index cc9000f..385d10a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/java/org/apache/hadoop/yarn/server/webapp/WebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/java/org/apache/hadoop/yarn/server/webapp/WebServices.java @@ -23,7 +23,6 @@ import java.security.PrivilegedExceptionAction; import java.util.Arrays; import java.util.Collection; import java.util.HashSet; -import java.util.Locale; import java.util.Set; import javax.servlet.http.HttpServletRequest; @@ -165,7 +164,7 @@ public class WebServices { if (checkAppStates && !appStates.contains(appReport.getYarnApplicationState().toString() - .toLowerCase(Locale.ENGLISH))) { + .toLowerCase())) { continue; } if (finalStatusQuery != null && !finalStatusQuery.isEmpty()) { @@ -187,7 +186,7 @@ public class WebServices { } if (checkAppTypes && !appTypes.contains(appReport.getApplicationType().trim() - .toLowerCase(Locale.ENGLISH))) { + .toLowerCase())) { continue; } @@ -369,8 +368,7 @@ public class WebServices { if (isState) { try { // enum string is in the uppercase - YarnApplicationState.valueOf( - paramStr.trim().toUpperCase(Locale.ENGLISH)); + YarnApplicationState.valueOf(paramStr.trim().toUpperCase()); } catch (RuntimeException e) { YarnApplicationState[] stateArray = YarnApplicationState.values(); @@ -380,7 +378,7 @@ public class WebServices { + allAppStates); } } - params.add(paramStr.trim().toLowerCase(Locale.ENGLISH)); + params.add(paramStr.trim().toLowerCase()); } } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java index 8541766..46b45f8 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java @@ -29,7 +29,6 @@ import java.util.Collections; import java.util.EnumSet; import java.util.Iterator; import java.util.List; -import java.util.Locale; import java.util.Map; import java.util.Set; import java.util.concurrent.atomic.AtomicInteger; @@ -757,7 +756,7 @@ public class ClientRMService extends AbstractService implements if (applicationTypes != null && !applicationTypes.isEmpty()) { String appTypeToMatch = caseSensitive ? application.getApplicationType() - : application.getApplicationType().toLowerCase(Locale.ENGLISH); + : application.getApplicationType().toLowerCase(); if (!applicationTypes.contains(appTypeToMatch)) { continue; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/resource/ResourceWeights.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/resource/ResourceWeights.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/resource/ResourceWeights.java index e95b725..230f9a9 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/resource/ResourceWeights.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/resource/ResourceWeights.java @@ -18,7 +18,6 @@ package org.apache.hadoop.yarn.server.resourcemanager.resource; -import java.util.Locale; import org.apache.hadoop.classification.InterfaceAudience.Private; import org.apache.hadoop.classification.InterfaceStability.Evolving; @@ -62,7 +61,7 @@ public class ResourceWeights { sb.append(", "); } ResourceType resourceType = ResourceType.values()[i]; - sb.append(resourceType.name().toLowerCase(Locale.ENGLISH)); + sb.append(resourceType.name().toLowerCase()); sb.append(String.format(" weight=%.1f", getWeight(resourceType))); } sb.append(">"); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacitySchedulerConfiguration.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacitySchedulerConfiguration.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacitySchedulerConfiguration.java index 5e6d3eb..3528c2d 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacitySchedulerConfiguration.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacitySchedulerConfiguration.java @@ -24,7 +24,6 @@ import java.util.Collections; import java.util.HashMap; import java.util.HashSet; import java.util.List; -import java.util.Locale; import java.util.Map; import java.util.Set; import java.util.StringTokenizer; @@ -395,8 +394,7 @@ public class CapacitySchedulerConfiguration extends ReservationSchedulerConfigur public QueueState getState(String queue) { String state = get(getQueuePrefix(queue) + STATE); return (state != null) ? - QueueState.valueOf(state.toUpperCase(Locale.ENGLISH)) : - QueueState.RUNNING; + QueueState.valueOf(state.toUpperCase()) : QueueState.RUNNING; } public void setAccessibleNodeLabels(String queue, Set<String> labels) { @@ -492,7 +490,7 @@ public class CapacitySchedulerConfiguration extends ReservationSchedulerConfigur } private static String getAclKey(QueueACL acl) { - return "acl_" + acl.toString().toLowerCase(Locale.ENGLISH); + return "acl_" + acl.toString().toLowerCase(); } public AccessControlList getAcl(String queue, QueueACL acl) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairSchedulerConfiguration.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairSchedulerConfiguration.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairSchedulerConfiguration.java index 0922092..32ef906 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairSchedulerConfiguration.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairSchedulerConfiguration.java @@ -20,7 +20,6 @@ package org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair; import java.io.File; import java.util.ArrayList; import java.util.List; -import java.util.Locale; import java.util.regex.Matcher; import java.util.regex.Pattern; @@ -242,7 +241,7 @@ public class FairSchedulerConfiguration extends Configuration { public static Resource parseResourceConfigValue(String val) throws AllocationConfigurationException { try { - val = val.toLowerCase(Locale.ENGLISH); + val = val.toLowerCase(); int memory = findResource(val, "mb"); int vcores = findResource(val, "vcores"); return BuilderUtils.newResource(memory, vcores); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/SchedulingPolicy.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/SchedulingPolicy.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/SchedulingPolicy.java index 23e7b81..cc28afc 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/SchedulingPolicy.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/SchedulingPolicy.java @@ -17,7 +17,6 @@ */ package org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair; -import java.util.Locale; import org.apache.hadoop.classification.InterfaceAudience.Public; import org.apache.hadoop.classification.InterfaceStability.Evolving; import org.apache.hadoop.util.ReflectionUtils; @@ -73,7 +72,7 @@ public abstract class SchedulingPolicy { throws AllocationConfigurationException { @SuppressWarnings("rawtypes") Class clazz; - String text = policy.toLowerCase(Locale.ENGLISH); + String text = policy.toLowerCase(); if (text.equalsIgnoreCase(FairSharePolicy.NAME)) { clazz = FairSharePolicy.class; } else if (text.equalsIgnoreCase(FifoPolicy.NAME)) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/NodesPage.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/NodesPage.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/NodesPage.java index 881ecb7..f28a9a8 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/NodesPage.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/NodesPage.java @@ -18,7 +18,6 @@ package org.apache.hadoop.yarn.server.resourcemanager.webapp; -import java.util.Locale; import static org.apache.hadoop.yarn.webapp.YarnWebParams.NODE_STATE; import static org.apache.hadoop.yarn.webapp.YarnWebParams.NODE_LABEL; import static org.apache.hadoop.yarn.webapp.view.JQueryUI.DATATABLES; @@ -78,7 +77,7 @@ class NodesPage extends RmView { .th(".nodeManagerVersion", "Version")._()._().tbody(); NodeState stateFilter = null; if (type != null && !type.isEmpty()) { - stateFilter = NodeState.valueOf(type.toUpperCase(Locale.ENGLISH)); + stateFilter = NodeState.valueOf(type.toUpperCase()); } Collection<RMNode> rmNodes = this.rm.getRMContext().getRMNodes().values(); boolean isInactive = false; http://git-wip-us.apache.org/repos/asf/hadoop/blob/9cedad11/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/RMWebServices.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/RMWebServices.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/RMWebServices.java index fb40054..1834b6a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/RMWebServices.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/RMWebServices.java @@ -30,7 +30,6 @@ import java.util.EnumSet; import java.util.HashMap; import java.util.HashSet; import java.util.List; -import java.util.Locale; import java.util.Map; import java.util.Set; import java.util.concurrent.ConcurrentMap; @@ -258,8 +257,7 @@ public class RMWebServices { } else { acceptedStates = EnumSet.noneOf(NodeState.class); for (String stateStr : states.split(",")) { - acceptedStates.add(NodeState.valueOf( - stateStr.toUpperCase(Locale.ENGLISH))); + acceptedStates.add(NodeState.valueOf(stateStr.toUpperCase())); } } @@ -508,7 +506,7 @@ public class RMWebServices { // if no states, returns the counts of all RMAppStates if (states.size() == 0) { for (YarnApplicationState state : YarnApplicationState.values()) { - states.add(state.toString().toLowerCase(Locale.ENGLISH)); + states.add(state.toString().toLowerCase()); } } // in case we extend to multiple applicationTypes in the future @@ -520,9 +518,8 @@ public class RMWebServices { ConcurrentMap<ApplicationId, RMApp> apps = rm.getRMContext().getRMApps(); for (RMApp rmapp : apps.values()) { YarnApplicationState state = rmapp.createApplicationState(); - String type = - rmapp.getApplicationType().trim().toLowerCase(Locale.ENGLISH); - if (states.contains(state.toString().toLowerCase(Locale.ENGLISH))) { + String type = rmapp.getApplicationType().trim().toLowerCase(); + if (states.contains(state.toString().toLowerCase())) { if (types.contains(ANY)) { countApp(scoreboard, state, ANY); } else if (types.contains(type)) { @@ -557,8 +554,7 @@ public class RMWebServices { if (isState) { try { // enum string is in the uppercase - YarnApplicationState.valueOf( - paramStr.trim().toUpperCase(Locale.ENGLISH)); + YarnApplicationState.valueOf(paramStr.trim().toUpperCase()); } catch (RuntimeException e) { YarnApplicationState[] stateArray = YarnApplicationState.values(); @@ -568,7 +564,7 @@ public class RMWebServices { + " specified. It should be one of " + allAppStates); } } - params.add(paramStr.trim().toLowerCase(Locale.ENGLISH)); + params.add(paramStr.trim().toLowerCase()); } } } @@ -586,8 +582,7 @@ public class RMWebServices { for (String state : states) { Map<String, Long> partScoreboard = new HashMap<String, Long>(); scoreboard.put( - YarnApplicationState.valueOf( - state.toUpperCase(Locale.ENGLISH)), partScoreboard); + YarnApplicationState.valueOf(state.toUpperCase()), partScoreboard); // types is verified no to be empty for (String type : types) { partScoreboard.put(type, 0L);