This is an automated email from the ASF dual-hosted git repository.

zhangduo pushed a commit to branch branch-3
in repository https://gitbox.apache.org/repos/asf/hbase.git


The following commit(s) were added to refs/heads/branch-3 by this push:
     new 2ad5b1a2bc3 HBASE-28255 Correcting spelling errors or annotations with 
non-standard spelling (#5577)
2ad5b1a2bc3 is described below

commit 2ad5b1a2bc3efa9456753f62ec3f52ecdaf38cba
Author: Sigma <[email protected]>
AuthorDate: Thu Jan 25 17:07:30 2024 +0800

    HBASE-28255 Correcting spelling errors or annotations with non-standard 
spelling (#5577)
    
    Co-authored-by: mazhengxuan <[email protected]>
    Co-authored-by: Duo Zhang <[email protected]>
    Signed-off-by: Duo Zhang <[email protected]>
    (cherry picked from commit caa6aed287c92abd1b630f56e02a2953172373a8)
---
 .../java/org/apache/hadoop/hbase/Abortable.java    |  2 +-
 .../apache/hadoop/hbase/ipc/AbstractRpcClient.java |  2 +-
 .../example/ZooKeeperScanPolicyObserver.java       |  2 +-
 .../procedure2/AbstractProcedureScheduler.java     |  2 +-
 .../master/replication/AbstractPeerProcedure.java  |  2 +-
 .../hbase/regionserver/AbstractMemStore.java       |  2 +-
 .../regionserver/AbstractMultiFileWriter.java      |  2 +-
 .../hbase/regionserver/wal/AbstractFSWAL.java      | 54 +++++++++++-----------
 .../wal/AbstractProtobufWALReader.java             | 12 ++---
 9 files changed, 40 insertions(+), 40 deletions(-)

diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/Abortable.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/Abortable.java
index b9736d57345..b0a5a86d50b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/Abortable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/Abortable.java
@@ -37,7 +37,7 @@ public interface Abortable {
   void abort(String why, Throwable e);
 
   /**
-   * It just call another abort method and the Throwable parameter is null.
+   * It just calls another abort method and the Throwable parameter is null.
    * @param why Why we're aborting.
    * @see Abortable#abort(String, Throwable)
    */
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java
index fcded9f5b69..5926539d067 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AbstractRpcClient.java
@@ -209,7 +209,7 @@ public abstract class AbstractRpcClient<T extends 
RpcConnection> implements RpcC
       for (T conn : connections.values()) {
         // Remove connection if it has not been chosen by anyone for more than 
maxIdleTime, and the
         // connection itself has already shutdown. The latter check is because 
we may still
-        // have some pending calls on connection so we should not shutdown the 
connection outside.
+        // have some pending calls on connection, so we should not shut down 
the connection outside.
         // The connection itself will disconnect if there is no pending call 
for maxIdleTime.
         if (conn.getLastTouched() < closeBeforeTime && !conn.isActive()) {
           if (LOG.isTraceEnabled()) {
diff --git 
a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java
 
b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java
index 53288be872e..fa7ccf73736 100644
--- 
a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java
+++ 
b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.java
@@ -41,7 +41,7 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
 import org.apache.yetus.audience.InterfaceAudience;
 
 /**
- * This is an example showing how a RegionObserver could configured via 
ZooKeeper in order to
+ * This is an example showing how a RegionObserver could be configured via 
ZooKeeper in order to
  * control a Region compaction, flush, and scan policy. This also demonstrated 
the use of shared
  * {@link org.apache.hadoop.hbase.coprocessor.RegionObserver} state. See
  * {@link RegionCoprocessorEnvironment#getSharedData()}.
diff --git 
a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.java
 
b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.java
index 4d1d5c1ccd9..61f73544b1b 100644
--- 
a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.java
+++ 
b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.java
@@ -236,7 +236,7 @@ public abstract class AbstractProcedureScheduler implements 
ProcedureScheduler {
   // ==========================================================================
 
   /**
-   * Wake up all of the given events. Note that we first take scheduler lock 
and then wakeInternal()
+   * Wake up all the given events. Note that we first take scheduler lock and 
then wakeInternal()
    * synchronizes on the event. Access should remain package-private. Use 
ProcedureEvent class to
    * wake/suspend events.
    * @param events the list of events to wake
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/replication/AbstractPeerProcedure.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/replication/AbstractPeerProcedure.java
index cd4cca0f918..b3b01f675c7 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/replication/AbstractPeerProcedure.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/replication/AbstractPeerProcedure.java
@@ -52,7 +52,7 @@ public abstract class AbstractPeerProcedure<TState> extends 
AbstractPeerNoLockPr
   // The sleep interval when waiting table to be enabled or disabled.
   protected static final int SLEEP_INTERVAL_MS = 1000;
 
-  // used to keep compatible with old client where we can only returns after 
updateStorage.
+  // used to keep compatible with old client where we can only return after 
updateStorage.
   protected ProcedurePrepareLatch latch;
 
   protected AbstractPeerProcedure() {
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
index 5cd3a92e5b6..62ff6f9a92f 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMemStore.java
@@ -157,7 +157,7 @@ public abstract class AbstractMemStore implements MemStore {
     Cell toAdd = maybeCloneWithAllocator(currentActive, cell, false);
     boolean mslabUsed = (toAdd != cell);
     // This cell data is backed by the same byte[] where we read request in 
RPC(See
-    // HBASE-15180). By default MSLAB is ON and we might have copied cell to 
MSLAB area. If
+    // HBASE-15180). By default, MSLAB is ON and we might have copied cell to 
MSLAB area. If
     // not we must do below deep copy. Or else we will keep referring to the 
bigger chunk of
     // memory and prevent it from getting GCed.
     // Copy to MSLAB would not have happened if
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
index a02b05f66ba..6370d6a79cc 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/AbstractMultiFileWriter.java
@@ -64,7 +64,7 @@ public abstract class AbstractMultiFileWriter implements 
CellSink, ShipperListen
    * Commit all writers.
    * <p>
    * Notice that here we use the same <code>maxSeqId</code> for all output 
files since we haven't
-   * find an easy to find enough sequence ids for different output files in 
some corner cases. See
+   * found an easy to find enough sequence ids for different output files in 
some corner cases. See
    * comments in HBASE-15400 for more details.
    */
   public List<Path> commitWriters(long maxSeqId, boolean majorCompaction) 
throws IOException {
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.java
index 1a5b5384b01..ef25068512f 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractFSWAL.java
@@ -259,8 +259,8 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
   protected final long blocksize;
 
   /*
-   * If more than this many logs, force flush of oldest region to oldest edit 
goes to disk. If too
-   * many and we crash, then will take forever replaying. Keep the number of 
logs tidy.
+   * If more than this many logs, force flush of oldest region to the oldest 
edit goes to disk. If
+   * too many and we crash, then will take forever replaying. Keep the number 
of logs tidy.
    */
   protected final int maxLogs;
 
@@ -336,7 +336,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
 
     /**
      * The log file size. Notice that the size may not be accurate if we do 
asynchronous close in
-     * sub classes.
+     * subclasses.
      */
     private final long logSize;
 
@@ -346,7 +346,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
     private final long rollTimeNs;
 
     /**
-     * If we do asynchronous close in sub classes, it is possible that when 
adding WALProps to the
+     * If we do asynchronous close in subclasses, it is possible that when 
adding WALProps to the
      * rolled map, the file is not closed yet, so in cleanOldLogs we should 
not archive this file,
      * for safety.
      */
@@ -404,9 +404,9 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
   protected Supplier<Boolean> hasConsumerTask;
 
   private static final int MAX_EPOCH = 0x3FFFFFFF;
-  // the lowest bit is waitingRoll, which means new writer is created and we 
are waiting for old
+  // the lowest bit is waitingRoll, which means new writer is created, and we 
are waiting for old
   // writer to be closed.
-  // the second lowest bit is writerBroken which means the current writer is 
broken and rollWriter
+  // the second-lowest bit is writerBroken which means the current writer is 
broken and rollWriter
   // is needed.
   // all other bits are the epoch number of the current writer, this is used 
to detect whether the
   // writer is still the one when you issue the sync.
@@ -807,7 +807,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
    * If the number of un-archived WAL files ('live' WALs) is greater than 
maximum allowed, check the
    * first (oldest) WAL, and return those regions which should be flushed so 
that it can be
    * let-go/'archived'.
-   * @return stores of regions (encodedRegionNames) to flush in order to 
archive oldest WAL file.
+   * @return stores of regions (encodedRegionNames) to flush in order to 
archive the oldest WAL file
    */
   Map<byte[], List<byte[]>> findRegionsToForceFlush() throws IOException {
     Map<byte[], List<byte[]>> regions = null;
@@ -861,7 +861,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
     long now = System.nanoTime();
     boolean mayLogTooOld = nextLogTooOldNs <= now;
     ArrayList<byte[]> regionsBlockingWal = null;
-    // For each log file, look at its Map of regions to highest sequence id; 
if all sequence ids
+    // For each log file, look at its Map of regions to the highest sequence 
id; if all sequence ids
     // are older than what is currently in memory, the WAL can be GC'd.
     for (Map.Entry<Path, WALProps> e : this.walFile2Props.entrySet()) {
       if (!e.getValue().closed) {
@@ -1084,7 +1084,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
       try {
         Path oldPath = getOldPath();
         Path newPath = getNewPath();
-        // Any exception from here on is catastrophic, non-recoverable so we 
currently abort.
+        // Any exception from here on is catastrophic, non-recoverable, so we 
currently abort.
         W nextWriter = this.createWriterInstance(fs, newPath);
         if (remoteFs != null) {
           // create a remote wal if necessary
@@ -1107,7 +1107,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
           regionsToFlush = findRegionsToForceFlush();
         }
       } catch (CommonFSUtils.StreamLacksCapabilityException exception) {
-        // If the underlying FileSystem can't do what we ask, treat as IO 
failure so
+        // If the underlying FileSystem can't do what we ask, treat as IO 
failure, so
         // we'll abort.
         throw new IOException(
           "Underlying FileSystem can't meet stream requirements. See RS log " 
+ "for details.",
@@ -1195,9 +1195,9 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
         throw new IOException(e.getCause());
       }
     } finally {
-      // in shutdown we may call cleanOldLogs so shutdown this executor in the 
end.
-      // In sync replication implementation, we may shutdown a WAL without 
shutting down the whole
-      // region server, if we shutdown this executor earlier we may get reject 
execution exception
+      // in shutdown, we may call cleanOldLogs so shutdown this executor in 
the end.
+      // In sync replication implementation, we may shut down a WAL without 
shutting down the whole
+      // region server, if we shut down this executor earlier we may get 
reject execution exception
       // and abort the region server
       logArchiveExecutor.shutdown();
     }
@@ -1467,8 +1467,8 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
   // return whether we have successfully set readyForRolling to true.
   private boolean trySetReadyForRolling() {
     // Check without holding lock first. Usually we will just return here.
-    // waitingRoll is volatile and unacedEntries is only accessed inside event 
loop so it is safe to
-    // check them outside the consumeLock.
+    // waitingRoll is volatile and unacedEntries is only accessed inside event 
loop, so it is safe
+    // to check them outside the consumeLock.
     if (!waitingRoll(epochAndState) || !unackedAppends.isEmpty()) {
       return false;
     }
@@ -1532,13 +1532,13 @@ public abstract class AbstractFSWAL<W extends 
WriterBase> implements WAL {
     // changed, i.e, we have already rolled the writer, or the writer is 
already broken, we should
     // just skip here, to avoid mess up the state or accidentally release some 
WAL entries and
     // cause data corruption.
-    // The syncCompleted call is on the critical write path so we should try 
our best to make it
+    // The syncCompleted call is on the critical write path, so we should try 
our best to make it
     // fast. So here we do not hold consumeLock, for increasing performance. 
It is safe because
     // there are only 3 possible situations:
     // 1. For normal case, the only place where we change epochAndState is 
when rolling the writer.
     // Before rolling actually happen, we will only change the state to 
waitingRoll which is another
     // bit than writerBroken, and when we actually change the epoch, we can 
make sure that there is
-    // no out going sync request. So we will always pass the check here and 
there is no problem.
+    // no outgoing sync request. So we will always pass the check here and 
there is no problem.
     // 2. The writer is broken, but we have not called syncFailed yet. In this 
case, since
     // syncFailed and syncCompleted are executed in the same thread, we will 
just face the same
     // situation with #1.
@@ -1706,7 +1706,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
     for (Iterator<FSWALEntry> iter = toWriteAppends.iterator(); 
iter.hasNext();) {
       FSWALEntry entry = iter.next();
       /**
-       * For {@link FSHog},here may throws IOException,but for {@link 
AsyncFSWAL}, here would not
+       * For {@link FSHog},here may throw IOException,but for {@link 
AsyncFSWAL}, here would not
        * throw any IOException.
        */
       boolean appended = appendEntry(writer, entry);
@@ -1753,7 +1753,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
     }
     if (writer.getLength() == fileLengthAtLastSync) {
       // we haven't written anything out, just advance the 
highestSyncedSequence since we may only
-      // stamped some region sequence id.
+      // stamp some region sequence id.
       if (unackedAppends.isEmpty()) {
         highestSyncedTxid.set(highestProcessedAppendTxid);
         finishSync();
@@ -1761,7 +1761,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
       }
       return;
     }
-    // reach here means that we have some unsynced data but haven't reached 
the batch size yet
+    // reach here means that we have some unsynced data but haven't reached 
the batch size yet,
     // but we will not issue a sync directly here even if there are sync 
requests because we may
     // have some new data in the ringbuffer, so let's just return here and 
delay the decision of
     // whether to issue a sync in the caller method.
@@ -1876,12 +1876,12 @@ public abstract class AbstractFSWAL<W extends 
WriterBase> implements WAL {
    * have its region edit/sequence id assigned else it messes up our 
unification of mvcc and
    * sequenceid. On return <code>key</code> will have the region edit/sequence 
id filled in.
    * <p/>
-   * NOTE: This append, at a time that is usually after this call returns, 
starts an mvcc
+   * NOTE: This appends, at a time that is usually after this call returns, 
starts a mvcc
    * transaction by calling 'begin' wherein which we assign this update a 
sequenceid. At assignment
    * time, we stamp all the passed in Cells inside WALEdit with their 
sequenceId. You must
    * 'complete' the transaction this mvcc transaction by calling
    * MultiVersionConcurrencyControl#complete(...) or a variant otherwise mvcc 
will get stuck. Do it
-   * in the finally of a try/finally block within which this append lives and 
any subsequent
+   * in the finally of a try/finally block within which this appends lives and 
any subsequent
    * operations like sync or update of memstore, etc. Get the WriteEntry to 
pass mvcc out of the
    * passed in WALKey <code>walKey</code> parameter. Be warned that the 
WriteEntry is not
    * immediately available on return from this method. It WILL be available 
subsequent to a sync of
@@ -2034,14 +2034,14 @@ public abstract class AbstractFSWAL<W extends 
WriterBase> implements WAL {
    * Notice that you need to clear the {@link #rollRequested} flag in this 
method, as the new writer
    * will begin to work before returning from this method. If we clear the 
flag after returning from
    * this call, we may miss a roll request. The implementation class should 
choose a proper place to
-   * clear the {@link #rollRequested} flag so we do not miss a roll request, 
typically before you
+   * clear the {@link #rollRequested} flag, so we do not miss a roll request, 
typically before you
    * start writing to the new writer.
    */
   protected void doReplaceWriter(Path oldPath, Path newPath, W nextWriter) 
throws IOException {
     Preconditions.checkNotNull(nextWriter);
     waitForSafePoint();
     /**
-     * For {@link FSHLog},here would shutdown {@link FSHLog.SyncRunner}.
+     * For {@link FSHLog},here would shut down {@link FSHLog.SyncRunner}.
      */
     doCleanUpResources();
     // we will call rollWriter in init method, where we want to create the 
first writer and
@@ -2084,7 +2084,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
   protected void doShutdown() throws IOException {
     waitForSafePoint();
     /**
-     * For {@link FSHLog},here would shutdown {@link FSHLog.SyncRunner}.
+     * For {@link FSHLog},here would shut down {@link FSHLog.SyncRunner}.
      */
     doCleanUpResources();
     if (this.writer != null) {
@@ -2214,7 +2214,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
   // So here we need to skip the creation of remote writer and make it 
possible to write the region
   // close marker.
   // Setting markerEdit only to true is for transiting from A to S, where we 
need to give up writing
-  // any pending wal entries as they will be discarded. The remote cluster 
will replicated the
+  // any pending wal entries as they will be discarded. The remote cluster 
will replicate the
   // correct data back later. We still need to allow writing marker edits such 
as close region event
   // to allow closing a region.
   @Override
@@ -2261,7 +2261,7 @@ public abstract class AbstractFSWAL<W extends WriterBase> 
implements WAL {
   }
 
   /**
-   * Pass one or more log file names and it will either dump out a text 
version on
+   * Pass one or more log file names, and it will either dump out a text 
version on
    * <code>stdout</code> or split the specified log files.
    */
   public static void main(String[] args) throws IOException {
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufWALReader.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufWALReader.java
index f5e65e08c84..5d51750ba5f 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufWALReader.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/AbstractProtobufWALReader.java
@@ -121,7 +121,7 @@ public abstract class AbstractProtobufWALReader
    * Get or create the input stream used by cell decoder.
    * <p/>
    * For implementing replication, we may need to limit the bytes we can read, 
so here we provide a
-   * method so sub classes can wrap the original input stream.
+   * method so subclasses can wrap the original input stream.
    */
   protected abstract InputStream getCellCodecInputStream(FSDataInputStream 
stream);
 
@@ -366,7 +366,7 @@ public abstract class AbstractProtobufWALReader
     this.fileLength = stat.getLen();
     this.walEditsStopOffset = this.fileLength;
     long currentPos = stream.getPos();
-    // we will reset walEditsStopOffset if trailer if available
+    // we will reset walEditsStopOffset if trailer is available
     trailerPresent = setTrailerIfPresent(stream);
     if (currentPos != stream.getPos()) {
       // seek back
@@ -509,18 +509,18 @@ public abstract class AbstractProtobufWALReader
    * This is used to determine whether we have already reached the WALTrailer. 
As the size and magic
    * are at the end of the WAL file, it is possible that these two options are 
missing while
    * writing, so we will consider there is no trailer. And when we actually 
reach the WALTrailer, we
-   * will try to decode it as WALKey and we will fail but the error could be 
vary as it is parsing
+   * will try to decode it as WALKey and we will fail but the error could be 
varied as it is parsing
    * WALTrailer actually.
    * @return whether this is a WALTrailer and we should throw EOF to upper 
layer the file is done
    */
   protected final boolean isWALTrailer(long startPosition) throws IOException {
-    // We have nothing in the WALTrailer PB message now so its size is just a 
int length size and a
+    // We have nothing in the WALTrailer PB message now so its size is just an 
int length size and a
     // magic at the end
     int trailerSize = PB_WAL_COMPLETE_MAGIC.length + Bytes.SIZEOF_INT;
     if (fileLength - startPosition >= trailerSize) {
       // We still have more than trailerSize bytes before reaching the EOF so 
this is not a trailer.
       // We also test for == here because if this is a valid trailer, we can 
read it while opening
-      // the reader so we should not reach here
+      // the reader, so we should not reach here
       return false;
     }
     inputStream.seek(startPosition);
@@ -548,7 +548,7 @@ public abstract class AbstractProtobufWALReader
         return false;
       }
     }
-    // in fact we should not reach here, as this means the trailer bytes are 
all matched and
+    // in fact, we should not reach here, as this means the trailer bytes are 
all matched and
     // complete, then we should not call this method...
     return true;
   }

Reply via email to