HBASE-20770 WAL cleaner logs way too much; gets clogged when lots of work to do
General log cleanup; setting stuff that can flood the log to TRACE. Amending-Author: Andrew Purtell <apurt...@apache.org> Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/7bfec2a8 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/7bfec2a8 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/7bfec2a8 Branch: refs/heads/branch-1.3 Commit: 7bfec2a8989083c70c7fcc1b8c7a1b7af4d65eb8 Parents: 03885b4 Author: Michael Stack <st...@apache.org> Authored: Sat Jun 23 23:29:11 2018 -0700 Committer: Andrew Purtell <apurt...@apache.org> Committed: Wed Dec 12 18:08:19 2018 -0800 ---------------------------------------------------------------------- .../hadoop/hbase/security/HBaseSaslRpcClient.java | 16 ++++++++-------- .../hadoop/hbase/security/SaslClientHandler.java | 12 ++++++------ .../java/org/apache/hadoop/hbase/ipc/RpcServer.java | 4 ++-- .../hbase/master/balancer/RegionLocationFinder.java | 3 +-- .../hadoop/hbase/master/cleaner/CleanerChore.java | 6 ++---- 5 files changed, 19 insertions(+), 22 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/7bfec2a8/hbase-client/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcClient.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcClient.java index bb6763f..f18b489 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcClient.java @@ -200,8 +200,8 @@ public class HBaseSaslRpcClient { return false; } saslToken = new byte[len]; - if (LOG.isDebugEnabled()) - LOG.debug("Will read input token of size " + saslToken.length + if (LOG.isTraceEnabled()) + LOG.trace("Will read input token of size " + saslToken.length + " for processing by initSASLContext"); inStream.readFully(saslToken); } @@ -209,8 +209,8 @@ public class HBaseSaslRpcClient { while (!saslClient.isComplete()) { saslToken = saslClient.evaluateChallenge(saslToken); if (saslToken != null) { - if (LOG.isDebugEnabled()) - LOG.debug("Will send token of size " + saslToken.length + if (LOG.isTraceEnabled()) + LOG.trace("Will send token of size " + saslToken.length + " from initSASLContext."); outStream.writeInt(saslToken.length); outStream.write(saslToken, 0, saslToken.length); @@ -219,14 +219,14 @@ public class HBaseSaslRpcClient { if (!saslClient.isComplete()) { readStatus(inStream); saslToken = new byte[inStream.readInt()]; - if (LOG.isDebugEnabled()) - LOG.debug("Will read input token of size " + saslToken.length + if (LOG.isTraceEnabled()) + LOG.trace("Will read input token of size " + saslToken.length + " for processing by initSASLContext"); inStream.readFully(saslToken); } } - if (LOG.isDebugEnabled()) { - LOG.debug("SASL client context established. Negotiated QoP: " + if (LOG.isTraceEnabled()) { + LOG.trace("SASL client context established. Negotiated QoP: " + saslClient.getNegotiatedProperty(Sasl.QOP)); } return true; http://git-wip-us.apache.org/repos/asf/hbase/blob/7bfec2a8/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslClientHandler.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslClientHandler.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslClientHandler.java index c79cde7..f43cb24 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslClientHandler.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslClientHandler.java @@ -212,16 +212,16 @@ public class SaslClientHandler extends ChannelDuplexHandler { } } saslToken = new byte[len]; - if (LOG.isDebugEnabled()) { - LOG.debug("Will read input token of size " + saslToken.length + if (LOG.isTraceEnabled()) { + LOG.trace("Will read input token of size " + saslToken.length + " for processing by initSASLContext"); } in.readBytes(saslToken); saslToken = evaluateChallenge(saslToken); if (saslToken != null) { - if (LOG.isDebugEnabled()) { - LOG.debug("Will send token of size " + saslToken.length + " from initSASLContext."); + if (LOG.isTraceEnabled()) { + LOG.trace("Will send token of size " + saslToken.length + " from initSASLContext."); } writeSaslToken(ctx, saslToken); } @@ -230,8 +230,8 @@ public class SaslClientHandler extends ChannelDuplexHandler { if (saslClient.isComplete()) { String qop = (String) saslClient.getNegotiatedProperty(Sasl.QOP); - if (LOG.isDebugEnabled()) { - LOG.debug("SASL client context established. Negotiated QoP: " + qop); + if (LOG.isTraceEnabled()) { + LOG.trace("SASL client context established. Negotiated QoP: " + qop); } boolean useWrap = qop != null && !"auth".equalsIgnoreCase(qop); http://git-wip-us.apache.org/repos/asf/hbase/blob/7bfec2a8/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java index afae42d..975309c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java @@ -1514,8 +1514,8 @@ public class RpcServer implements RpcServerInterface, ConfigurationObserver { throw e; } if (replyToken != null) { - if (LOG.isDebugEnabled()) { - LOG.debug("Will send token of size " + replyToken.length + if (LOG.isTraceEnabled()) { + LOG.trace("Will send token of size " + replyToken.length + " from saslServer."); } doRawSaslReply(SaslStatus.SUCCESS, new BytesWritable(replyToken), null, http://git-wip-us.apache.org/repos/asf/hbase/blob/7bfec2a8/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java index 6c5cb19..1ccd55f2a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/RegionLocationFinder.java @@ -284,8 +284,7 @@ class RegionLocationFinder { blockDistbn = cache.get(hri); return blockDistbn; } else { - LOG.debug("HDFSBlocksDistribution not found in cache for region " - + hri.getRegionNameAsString()); + LOG.trace("HDFSBlocksDistribution not found in cache for region " + hri.getRegionNameAsString()); blockDistbn = internalGetTopBlockLocation(hri); cache.put(hri, blockDistbn); return blockDistbn; http://git-wip-us.apache.org/repos/asf/hbase/blob/7bfec2a8/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java index 28023c6..b3f1f0a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java @@ -292,9 +292,7 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu POOL.updatePool((long) (0.8 * getTimeUnit().toMillis(getPeriod()))); } } else { - if (LOG.isTraceEnabled()) { - LOG.trace("Cleaner chore disabled! Not cleaning."); - } + LOG.trace("Cleaner chore disabled! Not cleaning."); } } @@ -432,7 +430,7 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu @Override protected Boolean compute() { - LOG.debug("Cleaning under " + dir); + LOG.trace("Cleaning under " + dir); List<FileStatus> subDirs; final List<FileStatus> files; try {