HADOOP-14539. Move commons logging APIs over to slf4j in hadoop-common. Contributed by Wenxin He.
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/ccaf0366 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/ccaf0366 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/ccaf0366 Branch: refs/heads/HDFS-7240 Commit: ccaf036662e22da14583942054898c99fa51dae5 Parents: 5b00792 Author: Akira Ajisaka <aajis...@apache.org> Authored: Tue Jul 18 13:32:37 2017 +0900 Committer: Akira Ajisaka <aajis...@apache.org> Committed: Tue Jul 18 13:32:37 2017 +0900 ---------------------------------------------------------------------- .../org/apache/hadoop/conf/Configuration.java | 17 +++++++++-------- .../apache/hadoop/conf/ReconfigurableBase.java | 7 ++++--- .../hadoop/conf/ReconfigurationServlet.java | 8 ++++---- .../hadoop/crypto/JceAesCtrCryptoCodec.java | 8 ++++---- .../hadoop/crypto/OpensslAesCtrCryptoCodec.java | 8 ++++---- .../org/apache/hadoop/crypto/OpensslCipher.java | 8 ++++---- .../crypto/random/OpensslSecureRandom.java | 8 ++++---- .../hadoop/crypto/random/OsSecureRandom.java | 9 +++++---- .../apache/hadoop/fs/AbstractFileSystem.java | 6 +++--- .../java/org/apache/hadoop/fs/ChecksumFs.java | 8 ++++---- .../hadoop/fs/DelegationTokenRenewer.java | 10 +++++----- .../org/apache/hadoop/fs/FSInputChecker.java | 9 +++++---- .../java/org/apache/hadoop/fs/FileContext.java | 10 +++++----- .../java/org/apache/hadoop/fs/FileUtil.java | 10 +++++----- .../main/java/org/apache/hadoop/fs/FsShell.java | 6 +++--- .../apache/hadoop/fs/FsShellPermissions.java | 4 ++-- .../main/java/org/apache/hadoop/fs/Globber.java | 7 ++++--- .../org/apache/hadoop/fs/HarFileSystem.java | 11 ++++++----- .../org/apache/hadoop/fs/LocalDirAllocator.java | 9 +++++---- .../main/java/org/apache/hadoop/fs/Trash.java | 7 ++++--- .../apache/hadoop/fs/TrashPolicyDefault.java | 8 ++++---- .../org/apache/hadoop/fs/ftp/FTPFileSystem.java | 8 ++++---- .../hadoop/fs/permission/FsPermission.java | 6 +++--- .../hadoop/fs/sftp/SFTPConnectionPool.java | 7 ++++--- .../apache/hadoop/fs/sftp/SFTPFileSystem.java | 7 ++++--- .../org/apache/hadoop/fs/shell/Command.java | 6 +++--- .../apache/hadoop/ha/ActiveStandbyElector.java | 15 ++++++++------- .../apache/hadoop/ha/FailoverController.java | 10 +++++----- .../main/java/org/apache/hadoop/ha/HAAdmin.java | 8 ++++---- .../org/apache/hadoop/ha/HealthMonitor.java | 8 ++++---- .../java/org/apache/hadoop/ha/NodeFencer.java | 6 +++--- .../org/apache/hadoop/ha/SshFenceByTcpPort.java | 11 +++-------- .../apache/hadoop/ha/ZKFailoverController.java | 20 ++++++++++---------- ...HAServiceProtocolServerSideTranslatorPB.java | 6 +++--- .../org/apache/hadoop/http/HttpServer2.java | 6 +++--- .../hadoop/http/lib/StaticUserWebFilter.java | 7 ++++--- .../java/org/apache/hadoop/io/BloomMapFile.java | 6 +++--- .../apache/hadoop/io/FastByteComparisons.java | 7 +++---- .../main/java/org/apache/hadoop/io/IOUtils.java | 5 ++--- .../main/java/org/apache/hadoop/io/MapFile.java | 8 ++++---- .../org/apache/hadoop/io/ReadaheadPool.java | 6 +++--- .../java/org/apache/hadoop/io/SequenceFile.java | 7 ++++--- .../main/java/org/apache/hadoop/io/UTF8.java | 5 +++-- .../apache/hadoop/io/compress/CodecPool.java | 6 +++--- .../io/compress/CompressionCodecFactory.java | 8 ++++---- .../apache/hadoop/io/compress/DefaultCodec.java | 6 +++--- .../io/compress/bzip2/Bzip2Compressor.java | 8 ++++---- .../io/compress/bzip2/Bzip2Decompressor.java | 8 ++++---- .../hadoop/io/compress/bzip2/Bzip2Factory.java | 6 +++--- .../hadoop/io/compress/lz4/Lz4Compressor.java | 8 ++++---- .../hadoop/io/compress/lz4/Lz4Decompressor.java | 8 ++++---- .../io/compress/snappy/SnappyCompressor.java | 8 ++++---- .../io/compress/snappy/SnappyDecompressor.java | 8 ++++---- .../io/compress/zlib/BuiltInZlibDeflater.java | 8 ++++---- .../hadoop/io/compress/zlib/ZlibCompressor.java | 8 ++++---- .../hadoop/io/compress/zlib/ZlibFactory.java | 8 ++++---- .../apache/hadoop/io/erasurecode/CodecUtil.java | 6 +++--- .../io/erasurecode/ErasureCodeNative.java | 8 ++++---- .../org/apache/hadoop/io/file/tfile/BCFile.java | 6 +++--- .../hadoop/io/file/tfile/Compression.java | 6 +++--- .../org/apache/hadoop/io/file/tfile/TFile.java | 8 ++++---- .../hadoop/io/file/tfile/TFileDumper.java | 8 ++++---- .../org/apache/hadoop/io/nativeio/NativeIO.java | 16 ++++++++-------- .../nativeio/SharedFileDescriptorFactory.java | 7 ++++--- .../apache/hadoop/io/retry/RetryPolicies.java | 6 +++--- .../org/apache/hadoop/io/retry/RetryUtils.java | 6 +++--- .../io/serializer/SerializationFactory.java | 8 ++++---- .../org/apache/hadoop/ipc/CallQueueManager.java | 7 ++++--- .../main/java/org/apache/hadoop/ipc/Client.java | 6 +++--- .../org/apache/hadoop/ipc/FairCallQueue.java | 6 +++--- .../apache/hadoop/ipc/ProtobufRpcEngine.java | 7 ++++--- .../main/java/org/apache/hadoop/ipc/RPC.java | 6 +++--- .../org/apache/hadoop/ipc/RefreshRegistry.java | 7 ++++--- .../java/org/apache/hadoop/ipc/RetryCache.java | 6 +++--- .../main/java/org/apache/hadoop/ipc/Server.java | 16 ++++++++-------- .../ipc/WeightedRoundRobinMultiplexer.java | 8 ++++---- .../apache/hadoop/ipc/WritableRpcEngine.java | 6 +++--- .../hadoop/ipc/metrics/RetryCacheMetrics.java | 6 +++--- .../hadoop/ipc/metrics/RpcDetailedMetrics.java | 8 ++++---- .../apache/hadoop/ipc/metrics/RpcMetrics.java | 6 +++--- .../org/apache/hadoop/jmx/JMXJsonServlet.java | 7 ++++--- .../hadoop/metrics2/impl/MBeanInfoBuilder.java | 2 +- .../hadoop/metrics2/impl/MetricsConfig.java | 8 ++++---- .../metrics2/impl/MetricsSinkAdapter.java | 9 +++++---- .../metrics2/impl/MetricsSourceAdapter.java | 7 ++++--- .../hadoop/metrics2/impl/MetricsSystemImpl.java | 6 +++--- .../hadoop/metrics2/lib/MethodMetric.java | 7 ++++--- .../metrics2/lib/MetricsSourceBuilder.java | 7 ++++--- .../metrics2/lib/MutableMetricsFactory.java | 7 ++++--- .../hadoop/metrics2/lib/MutableRates.java | 7 +++---- .../lib/MutableRatesWithAggregation.java | 7 ++++--- .../hadoop/metrics2/sink/GraphiteSink.java | 7 ++++--- .../sink/ganglia/AbstractGangliaSink.java | 10 +++++----- .../metrics2/sink/ganglia/GangliaSink30.java | 6 +++--- .../metrics2/sink/ganglia/GangliaSink31.java | 7 ++++--- .../org/apache/hadoop/metrics2/util/MBeans.java | 6 +++--- .../hadoop/metrics2/util/MetricsCache.java | 6 +++--- .../main/java/org/apache/hadoop/net/DNS.java | 6 +++--- .../java/org/apache/hadoop/net/NetUtils.java | 6 +++--- .../apache/hadoop/net/ScriptBasedMapping.java | 8 ++++---- .../apache/hadoop/net/SocketIOWithTimeout.java | 6 +++--- .../org/apache/hadoop/net/TableMapping.java | 6 +++--- .../apache/hadoop/net/unix/DomainSocket.java | 10 +++++----- .../hadoop/net/unix/DomainSocketWatcher.java | 12 ++++++------ .../AuthenticationWithProxyUserFilter.java | 8 ++++---- .../hadoop/security/CompositeGroupsMapping.java | 7 ++++--- .../org/apache/hadoop/security/Credentials.java | 10 +++++----- .../java/org/apache/hadoop/security/Groups.java | 7 +++---- .../HttpCrossOriginFilterInitializer.java | 8 ++++---- .../security/JniBasedUnixGroupsMapping.java | 8 ++++---- .../JniBasedUnixGroupsMappingWithFallback.java | 8 ++++---- .../JniBasedUnixGroupsNetgroupMapping.java | 6 +++--- ...edUnixGroupsNetgroupMappingWithFallback.java | 8 ++++---- .../hadoop/security/LdapGroupsMapping.java | 7 ++++--- .../apache/hadoop/security/ProviderUtils.java | 7 ++++--- .../apache/hadoop/security/SaslInputStream.java | 7 ++++--- .../apache/hadoop/security/SaslRpcClient.java | 7 ++++--- .../apache/hadoop/security/SaslRpcServer.java | 6 +++--- .../apache/hadoop/security/SecurityUtil.java | 9 ++++----- .../hadoop/security/ShellBasedIdMapping.java | 8 ++++---- .../ShellBasedUnixGroupsNetgroupMapping.java | 8 ++++---- .../hadoop/security/WhitelistBasedResolver.java | 7 ++++--- .../alias/AbstractJavaKeyStoreProvider.java | 6 +++--- .../authorize/ServiceAuthorizationManager.java | 9 +++++---- .../hadoop/security/http/CrossOriginFilter.java | 7 ++++--- .../security/ssl/FileBasedKeyStoresFactory.java | 8 ++++---- .../security/ssl/ReloadingX509TrustManager.java | 7 ++++--- .../hadoop/security/token/DtFileOperations.java | 7 ++++--- .../hadoop/security/token/DtUtilShell.java | 6 +++--- .../org/apache/hadoop/security/token/Token.java | 6 +++--- .../AbstractDelegationTokenSecretManager.java | 8 ++++---- .../apache/hadoop/service/AbstractService.java | 9 +++++---- .../apache/hadoop/service/CompositeService.java | 7 ++++--- .../service/LoggingStateChangeListener.java | 11 ++++++----- .../hadoop/service/ServiceOperations.java | 5 +++-- .../tracing/TracerConfigurationManager.java | 8 ++++---- .../hadoop/util/ApplicationClassLoader.java | 10 +++++----- .../apache/hadoop/util/AsyncDiskService.java | 7 ++++--- .../apache/hadoop/util/CombinedIPWhiteList.java | 7 ++++--- .../org/apache/hadoop/util/FileBasedIPList.java | 11 ++++++----- .../main/java/org/apache/hadoop/util/GSet.java | 6 +++--- .../hadoop/util/GenericOptionsParser.java | 7 ++++--- .../org/apache/hadoop/util/HostsFileReader.java | 9 +++++---- .../apache/hadoop/util/IntrusiveCollection.java | 7 ++++--- .../org/apache/hadoop/util/JvmPauseMonitor.java | 6 +++--- .../org/apache/hadoop/util/MachineList.java | 6 +++--- .../apache/hadoop/util/NativeCodeLoader.java | 8 ++++---- .../hadoop/util/NodeHealthScriptRunner.java | 7 ++++--- .../java/org/apache/hadoop/util/Progress.java | 6 +++--- .../apache/hadoop/util/ShutdownHookManager.java | 7 ++++--- .../hadoop/util/ShutdownThreadsHelper.java | 7 ++++--- .../org/apache/hadoop/util/SysInfoLinux.java | 8 ++++---- .../org/apache/hadoop/util/SysInfoWindows.java | 7 ++++--- .../java/org/apache/hadoop/util/ThreadUtil.java | 7 +++---- .../org/apache/hadoop/util/VersionInfo.java | 8 ++++---- .../hadoop/util/concurrent/AsyncGetFuture.java | 7 ++++--- .../hadoop/util/concurrent/ExecutorHelper.java | 8 ++++---- .../HadoopScheduledThreadPoolExecutor.java | 8 ++++---- .../concurrent/HadoopThreadPoolExecutor.java | 8 ++++---- .../hadoop/crypto/CryptoStreamsTestBase.java | 6 +++--- .../apache/hadoop/crypto/TestCryptoCodec.java | 7 ++++--- .../apache/hadoop/fs/FCStatisticsBaseTest.java | 7 ++++--- .../org/apache/hadoop/fs/TestFileContext.java | 7 ++++--- .../org/apache/hadoop/fs/TestFileStatus.java | 8 ++++---- .../java/org/apache/hadoop/fs/TestFileUtil.java | 6 +++--- .../org/apache/hadoop/fs/TestFsShellCopy.java | 6 +++--- .../apache/hadoop/fs/TestFsShellReturnCode.java | 8 ++++---- .../org/apache/hadoop/fs/TestFsShellTouch.java | 6 +++--- .../org/apache/hadoop/fs/TestHarFileSystem.java | 7 ++++--- .../fs/contract/AbstractBondedFSContract.java | 8 ++++---- .../hadoop/fs/loadGenerator/LoadGenerator.java | 10 +++++----- .../hadoop/ha/ActiveStandbyElectorTestUtil.java | 6 +++--- .../org/apache/hadoop/ha/DummyHAService.java | 7 ++++--- .../org/apache/hadoop/ha/MiniZKFCCluster.java | 7 ++++--- .../java/org/apache/hadoop/ha/TestHAAdmin.java | 6 +++--- .../org/apache/hadoop/ha/TestHealthMonitor.java | 6 +++--- .../apache/hadoop/http/TestGlobalFilter.java | 6 +++--- .../org/apache/hadoop/http/TestHttpServer.java | 6 +++--- .../apache/hadoop/http/TestHttpServerLogs.java | 6 +++--- .../hadoop/http/TestHttpServerWebapps.java | 9 +++++---- .../hadoop/http/TestHttpServerWithSpengo.java | 7 ++++--- .../org/apache/hadoop/http/TestPathFilter.java | 6 +++--- .../apache/hadoop/http/TestSSLHttpServer.java | 7 ++++--- .../apache/hadoop/http/TestServletFilter.java | 6 +++--- .../hadoop/http/resource/JerseyResource.java | 6 +++--- .../org/apache/hadoop/io/TestArrayFile.java | 7 +++++-- .../hadoop/io/TestDefaultStringifier.java | 8 +++++--- .../org/apache/hadoop/io/TestSequenceFile.java | 7 ++++--- .../java/org/apache/hadoop/io/TestSetFile.java | 7 +++---- .../org/apache/hadoop/io/TestWritableUtils.java | 7 ++++--- .../apache/hadoop/io/compress/TestCodec.java | 8 ++++---- .../io/compress/TestCompressionStreamReuse.java | 9 +++++---- .../apache/hadoop/io/nativeio/TestNativeIO.java | 10 +++++----- .../TestSharedFileDescriptorFactory.java | 7 ++++--- .../org/apache/hadoop/ipc/TestAsyncIPC.java | 10 +++++----- .../java/org/apache/hadoop/ipc/TestIPC.java | 17 ++++++++--------- .../hadoop/ipc/TestIPCServerResponder.java | 10 +++++----- .../ipc/TestProtoBufRpcServerHandoff.java | 12 ++++++------ .../java/org/apache/hadoop/ipc/TestRPC.java | 6 +++--- .../apache/hadoop/ipc/TestRPCCompatibility.java | 8 ++++---- .../hadoop/ipc/TestRPCServerShutdown.java | 7 ++++--- .../apache/hadoop/ipc/TestRpcServerHandoff.java | 8 ++++---- .../java/org/apache/hadoop/ipc/TestSaslRPC.java | 7 +++---- .../java/org/apache/hadoop/ipc/TestServer.java | 4 ++-- .../ipc/TestWeightedRoundRobinMultiplexer.java | 7 ++++--- .../metrics2/impl/TestGangliaMetrics.java | 7 ++++--- .../hadoop/metrics2/impl/TestMetricsConfig.java | 7 ++++--- .../metrics2/impl/TestMetricsSystemImpl.java | 13 +++++++------ .../hadoop/metrics2/impl/TestSinkQueue.java | 10 ++++++---- .../hadoop/metrics2/lib/TestMutableMetrics.java | 7 ++++--- .../hadoop/metrics2/util/TestMetricsCache.java | 8 +++++--- .../org/apache/hadoop/net/ServerSocketUtil.java | 9 +++++---- .../java/org/apache/hadoop/net/TestDNS.java | 6 +++--- .../org/apache/hadoop/net/TestNetUtils.java | 6 +++--- .../hadoop/net/TestSocketIOWithTimeout.java | 8 +++++--- .../apache/hadoop/net/TestStaticMapping.java | 7 ++++--- .../hadoop/net/unix/TestDomainSocket.java | 2 +- .../net/unix/TestDomainSocketWatcher.java | 15 ++++++++------- .../security/TestCompositeGroupMapping.java | 7 ++++--- .../hadoop/security/TestDoAsEffectiveUser.java | 8 ++++---- .../hadoop/security/TestGroupFallback.java | 7 ++++--- .../hadoop/security/TestGroupsCaching.java | 8 ++++---- .../TestShellBasedUnixGroupsMapping.java | 8 ++++---- .../alias/TestCredentialProviderFactory.java | 7 ++++--- .../authorize/TestAccessControlList.java | 8 ++++---- .../security/authorize/TestProxyUsers.java | 8 ++++---- .../token/delegation/TestDelegationToken.java | 7 ++++--- .../hadoop/service/TestCompositeService.java | 7 ++++--- .../hadoop/service/TestServiceLifecycle.java | 7 ++++--- .../org/apache/hadoop/test/MetricsAsserts.java | 6 +++--- .../hadoop/test/MultithreadedTestUtil.java | 8 ++++---- .../hadoop/test/TestGenericTestUtils.java | 5 +---- .../org/apache/hadoop/test/TestJUnitSetup.java | 7 ++++--- .../hadoop/util/TestAsyncDiskService.java | 7 ++++--- .../org/apache/hadoop/util/TestClasspath.java | 9 +++++---- .../org/apache/hadoop/util/TestFindClass.java | 7 ++++--- .../hadoop/util/TestIdentityHashStore.java | 7 ++++--- .../apache/hadoop/util/TestLightWeightGSet.java | 7 ++++--- .../util/TestLightWeightResizableGSet.java | 7 ++++--- .../hadoop/util/TestNativeCodeLoader.java | 6 +++--- .../apache/hadoop/util/TestSignalLogger.java | 11 ++++++----- .../org/apache/hadoop/util/TestWinUtils.java | 6 +++--- .../org/apache/hadoop/mount/MountdBase.java | 12 ++++++------ .../java/org/apache/hadoop/nfs/NfsExports.java | 6 +++--- .../org/apache/hadoop/nfs/nfs3/FileHandle.java | 6 +++--- .../org/apache/hadoop/nfs/nfs3/Nfs3Base.java | 10 +++++----- .../hadoop/oncrpc/RegistrationClient.java | 7 ++++--- .../java/org/apache/hadoop/oncrpc/RpcCall.java | 10 +++++----- .../org/apache/hadoop/oncrpc/RpcProgram.java | 6 +++--- .../java/org/apache/hadoop/oncrpc/RpcUtil.java | 11 ++++++----- .../hadoop/oncrpc/SimpleTcpClientHandler.java | 7 ++++--- .../apache/hadoop/oncrpc/SimpleTcpServer.java | 7 ++++--- .../apache/hadoop/oncrpc/SimpleUdpServer.java | 7 ++++--- .../hadoop/oncrpc/security/Credentials.java | 6 +++--- .../hadoop/oncrpc/security/SecurityHandler.java | 7 ++++--- .../java/org/apache/hadoop/portmap/Portmap.java | 8 ++++---- .../hadoop/portmap/RpcProgramPortmap.java | 7 ++++--- 257 files changed, 1021 insertions(+), 932 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java index 550aee7..de52fbb 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java @@ -79,8 +79,6 @@ import javax.xml.transform.stream.StreamResult; import com.google.common.base.Charsets; import org.apache.commons.collections.map.UnmodifiableMap; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.CommonConfigurationKeysPublic; @@ -98,6 +96,8 @@ import org.apache.hadoop.util.StringInterner; import org.apache.hadoop.util.StringUtils; import org.codehaus.stax2.XMLInputFactory2; import org.codehaus.stax2.XMLStreamReader2; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.w3c.dom.Document; import org.w3c.dom.Element; @@ -192,11 +192,12 @@ import com.google.common.base.Strings; @InterfaceStability.Stable public class Configuration implements Iterable<Map.Entry<String,String>>, Writable { - private static final Log LOG = - LogFactory.getLog(Configuration.class); + private static final Logger LOG = + LoggerFactory.getLogger(Configuration.class); - private static final Log LOG_DEPRECATION = - LogFactory.getLog("org.apache.hadoop.conf.Configuration.deprecation"); + private static final Logger LOG_DEPRECATION = + LoggerFactory.getLogger( + "org.apache.hadoop.conf.Configuration.deprecation"); private boolean quietmode = true; @@ -2885,10 +2886,10 @@ public class Configuration implements Iterable<Map.Entry<String,String>>, } return null; } catch (IOException e) { - LOG.fatal("error parsing conf " + name, e); + LOG.error("error parsing conf " + name, e); throw new RuntimeException(e); } catch (XMLStreamException e) { - LOG.fatal("error parsing conf " + name, e); + LOG.error("error parsing conf " + name, e); throw new RuntimeException(e); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurableBase.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurableBase.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurableBase.java index bdd006d..146c6d8 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurableBase.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurableBase.java @@ -22,9 +22,10 @@ import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Optional; import com.google.common.base.Preconditions; import com.google.common.collect.Maps; -import org.apache.commons.logging.*; import org.apache.hadoop.util.Time; import org.apache.hadoop.conf.ReconfigurationUtil.PropertyChange; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import java.io.IOException; import java.util.Collection; @@ -41,8 +42,8 @@ import java.util.Map; public abstract class ReconfigurableBase extends Configured implements Reconfigurable { - private static final Log LOG = - LogFactory.getLog(ReconfigurableBase.class); + private static final Logger LOG = + LoggerFactory.getLogger(ReconfigurableBase.class); // Use for testing purpose. private ReconfigurationUtil reconfigurationUtil = new ReconfigurationUtil(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java index bb221ee..5a616f7 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/ReconfigurationServlet.java @@ -18,8 +18,6 @@ package org.apache.hadoop.conf; -import org.apache.commons.logging.*; - import org.apache.commons.lang.StringEscapeUtils; import java.util.Collection; @@ -33,6 +31,8 @@ import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletResponse; import org.apache.hadoop.util.StringUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * A servlet for changing a node's configuration. @@ -45,8 +45,8 @@ public class ReconfigurationServlet extends HttpServlet { private static final long serialVersionUID = 1L; - private static final Log LOG = - LogFactory.getLog(ReconfigurationServlet.class); + private static final Logger LOG = + LoggerFactory.getLogger(ReconfigurationServlet.class); // the prefix used to fing the attribute holding the reconfigurable // for a given request http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/JceAesCtrCryptoCodec.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/JceAesCtrCryptoCodec.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/JceAesCtrCryptoCodec.java index 61ee743..de0e5dd 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/JceAesCtrCryptoCodec.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/JceAesCtrCryptoCodec.java @@ -26,12 +26,12 @@ import javax.crypto.Cipher; import javax.crypto.spec.IvParameterSpec; import javax.crypto.spec.SecretKeySpec; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import com.google.common.base.Preconditions; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_CRYPTO_JCE_PROVIDER_KEY; import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_JAVA_SECURE_RANDOM_ALGORITHM_KEY; @@ -42,8 +42,8 @@ import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY */ @InterfaceAudience.Private public class JceAesCtrCryptoCodec extends AesCtrCryptoCodec { - private static final Log LOG = - LogFactory.getLog(JceAesCtrCryptoCodec.class.getName()); + private static final Logger LOG = + LoggerFactory.getLogger(JceAesCtrCryptoCodec.class.getName()); private Configuration conf; private String provider; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslAesCtrCryptoCodec.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslAesCtrCryptoCodec.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslAesCtrCryptoCodec.java index d08e588..8d01f42 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslAesCtrCryptoCodec.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslAesCtrCryptoCodec.java @@ -26,22 +26,22 @@ import java.security.GeneralSecurityException; import java.security.SecureRandom; import java.util.Random; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import com.google.common.base.Preconditions; import org.apache.hadoop.crypto.random.OsSecureRandom; import org.apache.hadoop.util.ReflectionUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Implement the AES-CTR crypto codec using JNI into OpenSSL. */ @InterfaceAudience.Private public class OpensslAesCtrCryptoCodec extends AesCtrCryptoCodec { - private static final Log LOG = - LogFactory.getLog(OpensslAesCtrCryptoCodec.class.getName()); + private static final Logger LOG = + LoggerFactory.getLogger(OpensslAesCtrCryptoCodec.class.getName()); private Configuration conf; private Random random; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslCipher.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslCipher.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslCipher.java index 6a03bb6..133a9f9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslCipher.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/OpensslCipher.java @@ -26,13 +26,13 @@ import javax.crypto.IllegalBlockSizeException; import javax.crypto.NoSuchPaddingException; import javax.crypto.ShortBufferException; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.util.NativeCodeLoader; import com.google.common.base.Preconditions; import org.apache.hadoop.util.PerformanceAdvisory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * OpenSSL cipher using JNI. @@ -41,8 +41,8 @@ import org.apache.hadoop.util.PerformanceAdvisory; */ @InterfaceAudience.Private public final class OpensslCipher { - private static final Log LOG = - LogFactory.getLog(OpensslCipher.class.getName()); + private static final Logger LOG = + LoggerFactory.getLogger(OpensslCipher.class.getName()); public static final int ENCRYPT_MODE = 1; public static final int DECRYPT_MODE = 0; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OpensslSecureRandom.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OpensslSecureRandom.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OpensslSecureRandom.java index 6c53a0a..1219bf9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OpensslSecureRandom.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OpensslSecureRandom.java @@ -19,13 +19,13 @@ package org.apache.hadoop.crypto.random; import java.util.Random; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.util.NativeCodeLoader; import com.google.common.base.Preconditions; import org.apache.hadoop.util.PerformanceAdvisory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * OpenSSL secure random using JNI. @@ -44,8 +44,8 @@ import org.apache.hadoop.util.PerformanceAdvisory; @InterfaceAudience.Private public class OpensslSecureRandom extends Random { private static final long serialVersionUID = -7828193502768789584L; - private static final Log LOG = - LogFactory.getLog(OpensslSecureRandom.class.getName()); + private static final Logger LOG = + LoggerFactory.getLogger(OpensslSecureRandom.class.getName()); /** If native SecureRandom unavailable, use java SecureRandom */ private java.security.SecureRandom fallback = null; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OsSecureRandom.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OsSecureRandom.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OsSecureRandom.java index 9428b98..6671591 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OsSecureRandom.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/random/OsSecureRandom.java @@ -23,12 +23,12 @@ import java.io.FileInputStream; import java.io.IOException; import java.util.Random; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.io.IOUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_SECURE_RANDOM_DEVICE_FILE_PATH_KEY; import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY_SECURE_RANDOM_DEVICE_FILE_PATH_DEFAULT; @@ -39,7 +39,8 @@ import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.HADOOP_SECURITY */ @InterfaceAudience.Private public class OsSecureRandom extends Random implements Closeable, Configurable { - public static final Log LOG = LogFactory.getLog(OsSecureRandom.class); + public static final Logger LOG = + LoggerFactory.getLogger(OsSecureRandom.class); private static final long serialVersionUID = 6391500337172057900L; @@ -112,7 +113,7 @@ public class OsSecureRandom extends Random implements Closeable, Configurable { @Override synchronized public void close() { if (stream != null) { - IOUtils.cleanup(LOG, stream); + IOUtils.cleanupWithLogger(LOG, stream); stream = null; } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java index ef68437..9bea8f9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java @@ -32,8 +32,6 @@ import java.util.NoSuchElementException; import java.util.StringTokenizer; import java.util.concurrent.ConcurrentHashMap; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.HadoopIllegalArgumentException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; @@ -52,6 +50,8 @@ import org.apache.hadoop.security.token.Token; import org.apache.hadoop.util.Progressable; import com.google.common.annotations.VisibleForTesting; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * This class provides an interface for implementors of a Hadoop file system @@ -66,7 +66,7 @@ import com.google.common.annotations.VisibleForTesting; @InterfaceAudience.Public @InterfaceStability.Stable public abstract class AbstractFileSystem { - static final Log LOG = LogFactory.getLog(AbstractFileSystem.class); + static final Logger LOG = LoggerFactory.getLogger(AbstractFileSystem.class); /** Recording statistics per a file system class. */ private static final Map<URI, Statistics> http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFs.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFs.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFs.java index 0a8cc73..75622ad 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFs.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFs.java @@ -27,14 +27,14 @@ import java.util.ArrayList; import java.util.Arrays; import java.util.EnumSet; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.Options.ChecksumOpt; import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.util.DataChecksum; import org.apache.hadoop.util.Progressable; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Abstract Checksumed Fs. @@ -110,8 +110,8 @@ public abstract class ChecksumFs extends FilterFs { * It verifies that data matches checksums. *******************************************************/ private static class ChecksumFSInputChecker extends FSInputChecker { - public static final Log LOG - = LogFactory.getLog(FSInputChecker.class); + public static final Logger LOG = + LoggerFactory.getLogger(FSInputChecker.class); private static final int HEADER_LENGTH = 8; private ChecksumFs fs; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/DelegationTokenRenewer.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/DelegationTokenRenewer.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/DelegationTokenRenewer.java index 3542a9b..09c3a8a 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/DelegationTokenRenewer.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/DelegationTokenRenewer.java @@ -26,12 +26,12 @@ import java.util.concurrent.DelayQueue; import java.util.concurrent.Delayed; import java.util.concurrent.TimeUnit; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.security.token.Token; import org.apache.hadoop.security.token.TokenIdentifier; import org.apache.hadoop.util.Time; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * A daemon thread that waits for the next file system to renew. @@ -39,8 +39,8 @@ import org.apache.hadoop.util.Time; @InterfaceAudience.Private public class DelegationTokenRenewer extends Thread { - private static final Log LOG = LogFactory - .getLog(DelegationTokenRenewer.class); + private static final Logger LOG = LoggerFactory + .getLogger(DelegationTokenRenewer.class); /** The renewable interface used by the renewer. */ public interface Renewable { @@ -243,7 +243,7 @@ public class DelegationTokenRenewer LOG.error("Interrupted while canceling token for " + fs.getUri() + "filesystem"); if (LOG.isDebugEnabled()) { - LOG.debug(ie.getStackTrace()); + LOG.debug("Exception in removeRenewAction: ", ie); } } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FSInputChecker.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FSInputChecker.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FSInputChecker.java index 9b66c95..4f06e26 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FSInputChecker.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FSInputChecker.java @@ -22,11 +22,12 @@ import java.io.IOException; import java.io.InputStream; import java.util.zip.Checksum; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.util.StringUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + import java.nio.ByteBuffer; import java.nio.IntBuffer; @@ -37,8 +38,8 @@ import java.nio.IntBuffer; @InterfaceAudience.LimitedPrivate({"HDFS"}) @InterfaceStability.Unstable abstract public class FSInputChecker extends FSInputStream { - public static final Log LOG - = LogFactory.getLog(FSInputChecker.class); + public static final Logger LOG = + LoggerFactory.getLogger(FSInputChecker.class); /** The file name from which data is read from */ protected Path file; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java index 160a63d..fef968b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java @@ -35,8 +35,6 @@ import java.util.Stack; import java.util.TreeSet; import java.util.Map.Entry; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.HadoopIllegalArgumentException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; @@ -63,6 +61,8 @@ import org.apache.hadoop.util.ShutdownHookManager; import com.google.common.base.Preconditions; import org.apache.htrace.core.Tracer; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * The FileContext class provides an interface for users of the Hadoop @@ -169,7 +169,7 @@ import org.apache.htrace.core.Tracer; @InterfaceStability.Stable public class FileContext { - public static final Log LOG = LogFactory.getLog(FileContext.class); + public static final Logger LOG = LoggerFactory.getLogger(FileContext.class); /** * Default permission for directory and symlink * In previous versions, this default permission was also used to @@ -332,7 +332,7 @@ public class FileContext { } }); } catch (InterruptedException ex) { - LOG.error(ex); + LOG.error(ex.toString()); throw new IOException("Failed to get the AbstractFileSystem for path: " + uri, ex); } @@ -446,7 +446,7 @@ public class FileContext { } catch (UnsupportedFileSystemException ex) { throw ex; } catch (IOException ex) { - LOG.error(ex); + LOG.error(ex.toString()); throw new RuntimeException(ex); } return getFileContext(defaultAfs, aConf); http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileUtil.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileUtil.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileUtil.java index b656a87..eb8a5c3 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileUtil.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileUtil.java @@ -45,8 +45,6 @@ import java.util.zip.ZipFile; import org.apache.commons.collections.map.CaseInsensitiveMap; import org.apache.commons.compress.archivers.tar.TarArchiveEntry; import org.apache.commons.compress.archivers.tar.TarArchiveInputStream; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -57,6 +55,8 @@ import org.apache.hadoop.io.nativeio.NativeIO; import org.apache.hadoop.util.Shell; import org.apache.hadoop.util.Shell.ShellCommandExecutor; import org.apache.hadoop.util.StringUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * A collection of file-processing util methods @@ -65,7 +65,7 @@ import org.apache.hadoop.util.StringUtils; @InterfaceStability.Evolving public class FileUtil { - private static final Log LOG = LogFactory.getLog(FileUtil.class); + private static final Logger LOG = LoggerFactory.getLogger(FileUtil.class); /* The error code is defined in winutils to indicate insufficient * privilege to create symbolic links. This value need to keep in @@ -697,7 +697,7 @@ public class FileUtil { entry = tis.getNextTarEntry(); } } finally { - IOUtils.cleanup(LOG, tis, inputStream); + IOUtils.cleanupWithLogger(LOG, tis, inputStream); } } @@ -1287,7 +1287,7 @@ public class FileUtil { bos = new BufferedOutputStream(fos); jos = new JarOutputStream(bos, jarManifest); } finally { - IOUtils.cleanup(LOG, jos, bos, fos); + IOUtils.cleanupWithLogger(LOG, jos, bos, fos); } String[] jarCp = {classPathJar.getCanonicalPath(), unexpandedWildcardClasspath.toString()}; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShell.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShell.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShell.java index 59d15c2..721f4df 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShell.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShell.java @@ -24,8 +24,6 @@ import java.util.Arrays; import java.util.LinkedList; import org.apache.commons.lang.WordUtils; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; @@ -39,12 +37,14 @@ import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; import org.apache.htrace.core.TraceScope; import org.apache.htrace.core.Tracer; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** Provide command line access to a FileSystem. */ @InterfaceAudience.Private public class FsShell extends Configured implements Tool { - static final Log LOG = LogFactory.getLog(FsShell.class); + static final Logger LOG = LoggerFactory.getLogger(FsShell.class); private static final int MAX_LINE_WIDTH = 80; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShellPermissions.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShellPermissions.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShellPermissions.java index 0a82929..76e379c 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShellPermissions.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FsShellPermissions.java @@ -22,7 +22,6 @@ import java.util.LinkedList; import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.logging.Log; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.permission.ChmodParser; @@ -32,6 +31,7 @@ import org.apache.hadoop.fs.shell.CommandFormat; import org.apache.hadoop.fs.shell.FsCommand; import org.apache.hadoop.fs.shell.PathData; import org.apache.hadoop.util.Shell; +import org.slf4j.Logger; /** * This class is the home for file permissions related commands. @@ -41,7 +41,7 @@ import org.apache.hadoop.util.Shell; @InterfaceStability.Unstable public class FsShellPermissions extends FsCommand { - static Log LOG = FsShell.LOG; + static final Logger LOG = FsShell.LOG; /** * Register the permission related commands with the factory http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java index 7c69167..ca3db1d 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java @@ -23,18 +23,19 @@ import java.util.ArrayList; import java.util.Arrays; import java.util.List; -import org.apache.commons.logging.LogFactory; -import org.apache.commons.logging.Log; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.htrace.core.TraceScope; import org.apache.htrace.core.Tracer; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; @InterfaceAudience.Private @InterfaceStability.Unstable class Globber { - public static final Log LOG = LogFactory.getLog(Globber.class.getName()); + public static final Logger LOG = + LoggerFactory.getLogger(Globber.class.getName()); private final FileSystem fs; private final FileContext fc; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HarFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HarFileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HarFileSystem.java index 6a1e8bd..4c2fd1b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HarFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HarFileSystem.java @@ -17,14 +17,14 @@ */ package org.apache.hadoop.fs; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.io.Text; import org.apache.hadoop.util.LineReader; import org.apache.hadoop.util.Progressable; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import java.io.EOFException; import java.io.FileNotFoundException; @@ -50,7 +50,8 @@ import java.util.*; public class HarFileSystem extends FileSystem { - private static final Log LOG = LogFactory.getLog(HarFileSystem.class); + private static final Logger LOG = + LoggerFactory.getLogger(HarFileSystem.class); public static final String METADATA_CACHE_ENTRIES_KEY = "fs.har.metadatacache.entries"; public static final int METADATA_CACHE_ENTRIES_DEFAULT = 10; @@ -1173,7 +1174,7 @@ public class HarFileSystem extends FileSystem { LOG.warn("Encountered exception ", ioe); throw ioe; } finally { - IOUtils.cleanup(LOG, lin, in); + IOUtils.cleanupWithLogger(LOG, lin, in); } FSDataInputStream aIn = fs.open(archiveIndexPath); @@ -1198,7 +1199,7 @@ public class HarFileSystem extends FileSystem { } } } finally { - IOUtils.cleanup(LOG, aIn); + IOUtils.cleanupWithLogger(LOG, aIn); } } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/LocalDirAllocator.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/LocalDirAllocator.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/LocalDirAllocator.java index 1ed01ea..c1e9d21 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/LocalDirAllocator.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/LocalDirAllocator.java @@ -23,14 +23,15 @@ import java.util.*; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicReference; -import org.apache.commons.logging.*; import org.apache.hadoop.util.*; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.util.DiskChecker.DiskErrorException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; -import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.conf.Configuration; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** An implementation of a round-robin scheme for disk allocation for creating * files. The way it works is that it is kept track what disk was last @@ -245,8 +246,8 @@ public class LocalDirAllocator { private static class AllocatorPerContext { - private final Log LOG = - LogFactory.getLog(AllocatorPerContext.class); + private static final Logger LOG = + LoggerFactory.getLogger(AllocatorPerContext.class); private Random dirIndexRandomizer = new Random(); private String contextCfgItemName; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java index b771812..49cd600 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Trash.java @@ -19,11 +19,12 @@ package org.apache.hadoop.fs; import java.io.IOException; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Provides a trash facility which supports pluggable Trash policies. @@ -34,8 +35,8 @@ import org.apache.hadoop.conf.Configured; @InterfaceAudience.Public @InterfaceStability.Stable public class Trash extends Configured { - private static final org.apache.commons.logging.Log LOG = - LogFactory.getLog(Trash.class); + private static final Logger LOG = + LoggerFactory.getLogger(Trash.class); private TrashPolicy trashPolicy; // configured trash policy instance http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java index c65e16a..265e967 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java @@ -30,8 +30,6 @@ import java.text.SimpleDateFormat; import java.util.Collection; import java.util.Date; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -41,6 +39,8 @@ import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.util.Time; import com.google.common.annotations.VisibleForTesting; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** Provides a <i>trash</i> feature. Files are moved to a user's trash * directory, a subdirectory of their home directory named ".Trash". Files are @@ -54,8 +54,8 @@ import com.google.common.annotations.VisibleForTesting; @InterfaceAudience.Private @InterfaceStability.Evolving public class TrashPolicyDefault extends TrashPolicy { - private static final Log LOG = - LogFactory.getLog(TrashPolicyDefault.class); + private static final Logger LOG = + LoggerFactory.getLogger(TrashPolicyDefault.class); private static final Path CURRENT = new Path("Current"); http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPFileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPFileSystem.java index 5f4c8552..4c1236b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPFileSystem.java @@ -25,8 +25,6 @@ import java.net.URI; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.commons.net.ftp.FTP; import org.apache.commons.net.ftp.FTPClient; import org.apache.commons.net.ftp.FTPFile; @@ -45,6 +43,8 @@ import org.apache.hadoop.fs.permission.FsAction; import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.net.NetUtils; import org.apache.hadoop.util.Progressable; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * <p> @@ -56,8 +56,8 @@ import org.apache.hadoop.util.Progressable; @InterfaceStability.Stable public class FTPFileSystem extends FileSystem { - public static final Log LOG = LogFactory - .getLog(FTPFileSystem.class); + public static final Logger LOG = LoggerFactory + .getLogger(FTPFileSystem.class); public static final int DEFAULT_BUFFER_SIZE = 1024 * 1024; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/FsPermission.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/FsPermission.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/FsPermission.java index ddb2724..73ab5f6 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/FsPermission.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/permission/FsPermission.java @@ -24,8 +24,6 @@ import java.io.InvalidObjectException; import java.io.ObjectInputValidation; import java.io.Serializable; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -33,6 +31,8 @@ import org.apache.hadoop.fs.CommonConfigurationKeys; import org.apache.hadoop.io.Writable; import org.apache.hadoop.io.WritableFactories; import org.apache.hadoop.io.WritableFactory; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * A class for file/directory permissions. @@ -41,7 +41,7 @@ import org.apache.hadoop.io.WritableFactory; @InterfaceStability.Stable public class FsPermission implements Writable, Serializable, ObjectInputValidation { - private static final Log LOG = LogFactory.getLog(FsPermission.class); + private static final Logger LOG = LoggerFactory.getLogger(FsPermission.class); private static final long serialVersionUID = 0x2fe08564; static final WritableFactory FACTORY = new WritableFactory() { http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPConnectionPool.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPConnectionPool.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPConnectionPool.java index c7fae7b..de86bab 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPConnectionPool.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPConnectionPool.java @@ -23,19 +23,20 @@ import java.util.HashSet; import java.util.Iterator; import java.util.Set; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.util.StringUtils; import com.jcraft.jsch.ChannelSftp; import com.jcraft.jsch.JSch; import com.jcraft.jsch.JSchException; import com.jcraft.jsch.Session; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** Concurrent/Multiple Connections. */ class SFTPConnectionPool { - public static final Log LOG = LogFactory.getLog(SFTPFileSystem.class); + public static final Logger LOG = + LoggerFactory.getLogger(SFTPFileSystem.class); // Maximum number of allowed live connections. This doesn't mean we cannot // have more live connections. It means that when we have more // live connections than this threshold, any unused connection will be http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPFileSystem.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPFileSystem.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPFileSystem.java index 6de69fa..421769d 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPFileSystem.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/sftp/SFTPFileSystem.java @@ -26,8 +26,6 @@ import java.net.URLDecoder; import java.util.ArrayList; import java.util.Vector; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; @@ -41,11 +39,14 @@ import com.jcraft.jsch.ChannelSftp; import com.jcraft.jsch.ChannelSftp.LsEntry; import com.jcraft.jsch.SftpATTRS; import com.jcraft.jsch.SftpException; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** SFTP FileSystem. */ public class SFTPFileSystem extends FileSystem { - public static final Log LOG = LogFactory.getLog(SFTPFileSystem.class); + public static final Logger LOG = + LoggerFactory.getLogger(SFTPFileSystem.class); private SFTPConnectionPool connectionPool; private URI uri; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Command.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Command.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Command.java index 4c5cbad..c292cf6 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Command.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Command.java @@ -27,8 +27,6 @@ import java.util.Arrays; import java.util.LinkedList; import java.util.List; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -36,6 +34,8 @@ import org.apache.hadoop.conf.Configured; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.PathNotFoundException; import org.apache.hadoop.util.StringUtils; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * An abstract class for the execution of a file system command @@ -59,7 +59,7 @@ abstract public class Command extends Configured { private int depth = 0; protected ArrayList<Exception> exceptions = new ArrayList<Exception>(); - private static final Log LOG = LogFactory.getLog(Command.class); + private static final Logger LOG = LoggerFactory.getLogger(Command.class); /** allows stdout to be captured if necessary */ public PrintStream out = System.out; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java index cf95a49..93fd2cf 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ActiveStandbyElector.java @@ -26,8 +26,6 @@ import java.util.concurrent.TimeUnit; import java.util.concurrent.locks.Lock; import java.util.concurrent.locks.ReentrantLock; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.HadoopIllegalArgumentException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; @@ -47,6 +45,8 @@ import org.apache.zookeeper.KeeperException.Code; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * @@ -141,7 +141,8 @@ public class ActiveStandbyElector implements StatCallback, StringCallback { @VisibleForTesting protected static final String BREADCRUMB_FILENAME = "ActiveBreadCrumb"; - public static final Log LOG = LogFactory.getLog(ActiveStandbyElector.class); + public static final Logger LOG = + LoggerFactory.getLogger(ActiveStandbyElector.class); private static final int SLEEP_AFTER_FAILURE_TO_BECOME_ACTIVE = 1000; @@ -712,7 +713,7 @@ public class ActiveStandbyElector implements StatCallback, StringCallback { } private void fatalError(String errorMessage) { - LOG.fatal(errorMessage); + LOG.error(errorMessage); reset(); appClient.notifyFatalError(errorMessage); } @@ -824,10 +825,10 @@ public class ActiveStandbyElector implements StatCallback, StringCallback { createConnection(); success = true; } catch(IOException e) { - LOG.warn(e); + LOG.warn(e.toString()); sleepFor(5000); } catch(KeeperException e) { - LOG.warn(e); + LOG.warn(e.toString()); sleepFor(5000); } ++connectionRetryCount; @@ -866,7 +867,7 @@ public class ActiveStandbyElector implements StatCallback, StringCallback { try { tempZk.close(); } catch(InterruptedException e) { - LOG.warn(e); + LOG.warn(e.toString()); } zkConnectionState = ConnectionState.TERMINATED; wantToBeInElection = false; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/FailoverController.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/FailoverController.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/FailoverController.java index d952e29..3c05a25 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/FailoverController.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/FailoverController.java @@ -19,9 +19,6 @@ package org.apache.hadoop.ha; import java.io.IOException; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; - import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -32,6 +29,8 @@ import org.apache.hadoop.ha.HAServiceProtocol.RequestSource; import org.apache.hadoop.ipc.RPC; import com.google.common.base.Preconditions; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * The FailOverController is responsible for electing an active service @@ -43,7 +42,8 @@ import com.google.common.base.Preconditions; @InterfaceStability.Evolving public class FailoverController { - private static final Log LOG = LogFactory.getLog(FailoverController.class); + private static final Logger LOG = + LoggerFactory.getLogger(FailoverController.class); private final int gracefulFenceTimeout; private final int rpcTimeoutToNewActive; @@ -252,7 +252,7 @@ public class FailoverController { } catch (FailoverFailedException ffe) { msg += ". Failback to " + fromSvc + " failed (" + ffe.getMessage() + ")"; - LOG.fatal(msg); + LOG.error(msg); } } throw new FailoverFailedException(msg, cause); http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HAAdmin.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HAAdmin.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HAAdmin.java index 5eff14c..9b7d7ba 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HAAdmin.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HAAdmin.java @@ -28,8 +28,6 @@ import org.apache.commons.cli.Options; import org.apache.commons.cli.CommandLine; import org.apache.commons.cli.GnuParser; import org.apache.commons.cli.ParseException; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; @@ -43,6 +41,8 @@ import org.apache.hadoop.util.ToolRunner; import com.google.common.base.Preconditions; import com.google.common.collect.ImmutableMap; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * A command-line tool for making calls in the HAServiceProtocol. @@ -62,7 +62,7 @@ public abstract class HAAdmin extends Configured implements Tool { * operation, which is why it is not documented in the usage below. */ private static final String FORCEMANUAL = "forcemanual"; - private static final Log LOG = LogFactory.getLog(HAAdmin.class); + private static final Logger LOG = LoggerFactory.getLogger(HAAdmin.class); private int rpcTimeoutForChecks = -1; @@ -449,7 +449,7 @@ public abstract class HAAdmin extends Configured implements Tool { if (cmdLine.hasOption(FORCEMANUAL)) { if (!confirmForceManual()) { - LOG.fatal("Aborted"); + LOG.error("Aborted"); return -1; } // Instruct the NNs to honor this request even if they're http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HealthMonitor.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HealthMonitor.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HealthMonitor.java index 24c149c..a93df75 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HealthMonitor.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/HealthMonitor.java @@ -23,8 +23,6 @@ import java.util.LinkedList; import java.util.List; import org.apache.hadoop.classification.InterfaceAudience; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import static org.apache.hadoop.fs.CommonConfigurationKeys.*; import org.apache.hadoop.ha.HAServiceProtocol; @@ -35,6 +33,8 @@ import org.apache.hadoop.ipc.RPC; import org.apache.hadoop.util.Daemon; import com.google.common.base.Preconditions; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * This class is a daemon which runs in a loop, periodically heartbeating @@ -47,7 +47,7 @@ import com.google.common.base.Preconditions; */ @InterfaceAudience.Private public class HealthMonitor { - private static final Log LOG = LogFactory.getLog( + private static final Logger LOG = LoggerFactory.getLogger( HealthMonitor.class); private Daemon daemon; @@ -283,7 +283,7 @@ public class HealthMonitor { setUncaughtExceptionHandler(new UncaughtExceptionHandler() { @Override public void uncaughtException(Thread t, Throwable e) { - LOG.fatal("Health monitor failed", e); + LOG.error("Health monitor failed", e); enterState(HealthMonitor.State.HEALTH_MONITOR_FAILED); } }); http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/NodeFencer.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/NodeFencer.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/NodeFencer.java index 1afd937..2247a34 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/NodeFencer.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/NodeFencer.java @@ -22,8 +22,6 @@ import java.util.Map; import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -31,6 +29,8 @@ import org.apache.hadoop.util.ReflectionUtils; import com.google.common.collect.ImmutableMap; import com.google.common.collect.Lists; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * This class parses the configured list of fencing methods, and @@ -61,7 +61,7 @@ public class NodeFencer { private static final Pattern HASH_COMMENT_RE = Pattern.compile("#.*$"); - private static final Log LOG = LogFactory.getLog(NodeFencer.class); + private static final Logger LOG = LoggerFactory.getLogger(NodeFencer.class); /** * Standard fencing methods included with Hadoop. http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java index 64cd5a8..9ae113b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java @@ -23,8 +23,6 @@ import java.util.Collection; import java.util.regex.Matcher; import java.util.regex.Pattern; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configured; import com.google.common.annotations.VisibleForTesting; @@ -272,7 +270,7 @@ public class SshFenceByTcpPort extends Configured * Adapter from JSch's logger interface to our log4j */ private static class LogAdapter implements com.jcraft.jsch.Logger { - static final Log LOG = LogFactory.getLog( + static final Logger LOG = LoggerFactory.getLogger( SshFenceByTcpPort.class.getName() + ".jsch"); @Override @@ -285,9 +283,8 @@ public class SshFenceByTcpPort extends Configured case com.jcraft.jsch.Logger.WARN: return LOG.isWarnEnabled(); case com.jcraft.jsch.Logger.ERROR: - return LOG.isErrorEnabled(); case com.jcraft.jsch.Logger.FATAL: - return LOG.isFatalEnabled(); + return LOG.isErrorEnabled(); default: return false; } @@ -306,10 +303,8 @@ public class SshFenceByTcpPort extends Configured LOG.warn(message); break; case com.jcraft.jsch.Logger.ERROR: - LOG.error(message); - break; case com.jcraft.jsch.Logger.FATAL: - LOG.fatal(message); + LOG.error(message); break; default: break; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ZKFailoverController.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ZKFailoverController.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ZKFailoverController.java index 055bcaa..20a4681 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ZKFailoverController.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/ZKFailoverController.java @@ -28,8 +28,6 @@ import java.util.concurrent.Executors; import java.util.concurrent.ScheduledExecutorService; import java.util.concurrent.TimeUnit; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.HadoopIllegalArgumentException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; @@ -56,11 +54,13 @@ import org.apache.zookeeper.data.ACL; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Preconditions; import com.google.common.util.concurrent.ThreadFactoryBuilder; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; @InterfaceAudience.LimitedPrivate("HDFS") public abstract class ZKFailoverController { - static final Log LOG = LogFactory.getLog(ZKFailoverController.class); + static final Logger LOG = LoggerFactory.getLogger(ZKFailoverController.class); public static final String ZK_QUORUM_KEY = "ha.zookeeper.quorum"; private static final String ZK_SESSION_TIMEOUT_KEY = "ha.zookeeper.session-timeout.ms"; @@ -162,7 +162,7 @@ public abstract class ZKFailoverController { public int run(final String[] args) throws Exception { if (!localTarget.isAutoFailoverEnabled()) { - LOG.fatal("Automatic failover is not enabled for " + localTarget + "." + + LOG.error("Automatic failover is not enabled for " + localTarget + "." + " Please ensure that automatic failover is enabled in the " + "configuration before running the ZK failover controller."); return ERR_CODE_AUTO_FAILOVER_NOT_ENABLED; @@ -184,7 +184,7 @@ public abstract class ZKFailoverController { } }); } catch (RuntimeException rte) { - LOG.fatal("The failover controller encounters runtime error: " + rte); + LOG.error("The failover controller encounters runtime error: " + rte); throw (Exception)rte.getCause(); } } @@ -195,7 +195,7 @@ public abstract class ZKFailoverController { try { initZK(); } catch (KeeperException ke) { - LOG.fatal("Unable to start failover controller. Unable to connect " + LOG.error("Unable to start failover controller. Unable to connect " + "to ZooKeeper quorum at " + zkQuorum + ". Please check the " + "configured value for " + ZK_QUORUM_KEY + " and ensure that " + "ZooKeeper is running."); @@ -221,7 +221,7 @@ public abstract class ZKFailoverController { } if (!elector.parentZNodeExists()) { - LOG.fatal("Unable to start failover controller. " + LOG.error("Unable to start failover controller. " + "Parent znode does not exist.\n" + "Run with -formatZK flag to initialize ZooKeeper."); return ERR_CODE_NO_PARENT_ZNODE; @@ -230,7 +230,7 @@ public abstract class ZKFailoverController { try { localTarget.checkFencingConfigured(); } catch (BadFencingConfigurationException e) { - LOG.fatal("Fencing is not configured for " + localTarget + ".\n" + + LOG.error("Fencing is not configured for " + localTarget + ".\n" + "You must configure a fencing method before using automatic " + "failover.", e); return ERR_CODE_NO_FENCER; @@ -376,7 +376,7 @@ public abstract class ZKFailoverController { } private synchronized void fatalError(String err) { - LOG.fatal("Fatal error occurred:" + err); + LOG.error("Fatal error occurred:" + err); fatalError = err; notifyAll(); } @@ -395,7 +395,7 @@ public abstract class ZKFailoverController { } catch (Throwable t) { String msg = "Couldn't make " + localTarget + " active"; - LOG.fatal(msg, t); + LOG.error(msg, t); recordActiveAttempt(new ActiveAttemptRecord(false, msg + "\n" + StringUtils.stringifyException(t))); http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/protocolPB/HAServiceProtocolServerSideTranslatorPB.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/protocolPB/HAServiceProtocolServerSideTranslatorPB.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/protocolPB/HAServiceProtocolServerSideTranslatorPB.java index 63bfbca..7f75582 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/protocolPB/HAServiceProtocolServerSideTranslatorPB.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/protocolPB/HAServiceProtocolServerSideTranslatorPB.java @@ -19,8 +19,6 @@ package org.apache.hadoop.ha.protocolPB; import java.io.IOException; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.ha.HAServiceProtocol; @@ -42,6 +40,8 @@ import org.apache.hadoop.ipc.RPC; import com.google.protobuf.RpcController; import com.google.protobuf.ServiceException; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * This class is used on the server side. Calls come across the wire for the @@ -61,7 +61,7 @@ public class HAServiceProtocolServerSideTranslatorPB implements TransitionToActiveResponseProto.newBuilder().build(); private static final TransitionToStandbyResponseProto TRANSITION_TO_STANDBY_RESP = TransitionToStandbyResponseProto.newBuilder().build(); - private static final Log LOG = LogFactory.getLog( + private static final Logger LOG = LoggerFactory.getLogger( HAServiceProtocolServerSideTranslatorPB.class); public HAServiceProtocolServerSideTranslatorPB(HAServiceProtocol server) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java index d7436b2..28b9bb0 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java @@ -53,8 +53,6 @@ import com.google.common.base.Preconditions; import com.google.common.collect.ImmutableMap; import com.google.common.collect.Lists; import com.sun.jersey.spi.container.servlet.ServletContainer; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.HadoopIllegalArgumentException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; @@ -103,6 +101,8 @@ import org.eclipse.jetty.util.MultiException; import org.eclipse.jetty.util.ssl.SslContextFactory; import org.eclipse.jetty.util.thread.QueuedThreadPool; import org.eclipse.jetty.webapp.WebAppContext; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Create a Jetty embedded server to answer http requests. The primary goal is @@ -117,7 +117,7 @@ import org.eclipse.jetty.webapp.WebAppContext; @InterfaceAudience.Private @InterfaceStability.Evolving public final class HttpServer2 implements FilterContainer { - public static final Log LOG = LogFactory.getLog(HttpServer2.class); + public static final Logger LOG = LoggerFactory.getLogger(HttpServer2.class); public static final String HTTP_SCHEME = "http"; public static final String HTTPS_SCHEME = "https"; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/lib/StaticUserWebFilter.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/lib/StaticUserWebFilter.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/lib/StaticUserWebFilter.java index 9ca5b92..fc64697 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/lib/StaticUserWebFilter.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/lib/StaticUserWebFilter.java @@ -29,11 +29,11 @@ import javax.servlet.ServletResponse; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletRequestWrapper; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.http.FilterContainer; import org.apache.hadoop.http.FilterInitializer; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import javax.servlet.Filter; @@ -47,7 +47,8 @@ import static org.apache.hadoop.fs.CommonConfigurationKeys.DEFAULT_HADOOP_HTTP_S public class StaticUserWebFilter extends FilterInitializer { static final String DEPRECATED_UGI_KEY = "dfs.web.ugi"; - private static final Log LOG = LogFactory.getLog(StaticUserWebFilter.class); + private static final Logger LOG = + LoggerFactory.getLogger(StaticUserWebFilter.class); static class User implements Principal { private final String name; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/BloomMapFile.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/BloomMapFile.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/BloomMapFile.java index d4514c6..519fcd7 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/BloomMapFile.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/BloomMapFile.java @@ -22,8 +22,6 @@ import java.io.DataInputStream; import java.io.DataOutputStream; import java.io.IOException; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -36,6 +34,8 @@ import org.apache.hadoop.util.bloom.DynamicBloomFilter; import org.apache.hadoop.util.bloom.Filter; import org.apache.hadoop.util.bloom.Key; import org.apache.hadoop.util.hash.Hash; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_MAPFILE_BLOOM_ERROR_RATE_DEFAULT; import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_MAPFILE_BLOOM_ERROR_RATE_KEY; @@ -52,7 +52,7 @@ import static org.apache.hadoop.fs.CommonConfigurationKeysPublic.IO_MAPFILE_BLOO @InterfaceAudience.Public @InterfaceStability.Stable public class BloomMapFile { - private static final Log LOG = LogFactory.getLog(BloomMapFile.class); + private static final Logger LOG = LoggerFactory.getLogger(BloomMapFile.class); public static final String BLOOM_FILE_NAME = "bloom"; public static final int HASH_COUNT = 5; http://git-wip-us.apache.org/repos/asf/hadoop/blob/ccaf0366/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java index 705678e..a2903f8 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java @@ -22,11 +22,10 @@ import java.nio.ByteOrder; import java.security.AccessController; import java.security.PrivilegedAction; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import sun.misc.Unsafe; -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; - import com.google.common.primitives.Longs; import com.google.common.primitives.UnsignedBytes; @@ -36,7 +35,7 @@ import com.google.common.primitives.UnsignedBytes; * class to be able to compare arrays that start at non-zero offsets. */ abstract class FastByteComparisons { - static final Log LOG = LogFactory.getLog(FastByteComparisons.class); + static final Logger LOG = LoggerFactory.getLogger(FastByteComparisons.class); /** * Lexicographically compare two byte arrays. --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org