This is an automated email from the ASF dual-hosted git repository.
srowen pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new 920fb673b26 [MINOR] Fix typos
920fb673b26 is described below
commit 920fb673b264c0bdcad0426020dedf57d8b11cc7
Author: shuoer86 <[email protected]>
AuthorDate: Sat Oct 21 16:37:27 2023 -0500
[MINOR] Fix typos
Closes #43434 from shuoer86/master.
Authored-by: shuoer86 <[email protected]>
Signed-off-by: Sean Owen <[email protected]>
---
binder/postBuild | 4 ++--
.../scala/org/apache/spark/sql/connect/service/SessionHolder.scala | 2 +-
.../spark/sql/connect/plugin/SparkConnectPluginRegistrySuite.scala | 2 +-
core/src/main/scala/org/apache/spark/storage/BlockInfoManager.scala | 2 +-
core/src/main/scala/org/apache/spark/storage/DiskBlockManager.scala | 6 +++---
.../main/scala/org/apache/spark/ui/jobs/TaskThreadDumpPage.scala | 2 +-
.../scala/org/apache/spark/status/AutoCleanupLiveUIDirSuite.scala | 2 +-
docs/sql-ref-syntax-ddl-declare-variable.md | 2 +-
8 files changed, 11 insertions(+), 11 deletions(-)
diff --git a/binder/postBuild b/binder/postBuild
index 70ae23b3937..b6bdf72324c 100644
--- a/binder/postBuild
+++ b/binder/postBuild
@@ -38,7 +38,7 @@ else
pip install plotly "pandas<2.0.0"
"pyspark[sql,ml,mllib,pandas_on_spark]$SPECIFIER$VERSION"
fi
-# Set 'PYARROW_IGNORE_TIMEZONE' to surpress warnings from PyArrow.
+# Set 'PYARROW_IGNORE_TIMEZONE' to suppress warnings from PyArrow.
echo "export PYARROW_IGNORE_TIMEZONE=1" >> ~/.profile
# Add sbin to PATH to run `start-connect-server.sh`.
@@ -50,7 +50,7 @@ echo "export SPARK_HOME=${SPARK_HOME}" >> ~/.profile
SPARK_VERSION=$(python -c "import pyspark; print(pyspark.__version__)")
echo "export SPARK_VERSION=${SPARK_VERSION}" >> ~/.profile
-# Surpress warnings from Spark jobs, and UI progress bar.
+# Suppress warnings from Spark jobs, and UI progress bar.
mkdir -p ~/.ipython/profile_default/startup
echo """from pyspark.sql import SparkSession
SparkSession.builder.config('spark.ui.showConsoleProgress',
'false').getOrCreate().sparkContext.setLogLevel('FATAL')
diff --git
a/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/service/SessionHolder.scala
b/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/service/SessionHolder.scala
index 27f471233f1..dcced21f371 100644
---
a/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/service/SessionHolder.scala
+++
b/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/service/SessionHolder.scala
@@ -77,7 +77,7 @@ case class SessionHolder(userId: String, sessionId: String,
session: SparkSessio
private[service] def addExecuteHolder(executeHolder: ExecuteHolder): Unit = {
val oldExecute = executions.putIfAbsent(executeHolder.operationId,
executeHolder)
if (oldExecute != null) {
- // the existance of this should alrady be checked by
SparkConnectExecutionManager
+ // the existence of this should alrady be checked by
SparkConnectExecutionManager
throw new IllegalStateException(
s"ExecuteHolder with opId=${executeHolder.operationId} already
exists!")
}
diff --git
a/connector/connect/server/src/test/scala/org/apache/spark/sql/connect/plugin/SparkConnectPluginRegistrySuite.scala
b/connector/connect/server/src/test/scala/org/apache/spark/sql/connect/plugin/SparkConnectPluginRegistrySuite.scala
index ea9ae3ed9d9..e1de6b04d21 100644
---
a/connector/connect/server/src/test/scala/org/apache/spark/sql/connect/plugin/SparkConnectPluginRegistrySuite.scala
+++
b/connector/connect/server/src/test/scala/org/apache/spark/sql/connect/plugin/SparkConnectPluginRegistrySuite.scala
@@ -226,7 +226,7 @@ class SparkConnectPluginRegistrySuite extends
SharedSparkSession with SparkConne
}
}
- test("Emtpy registries are really empty and work") {
+ test("Empty registries are really empty and work") {
assert(SparkConnectPluginRegistry.loadRelationPlugins().isEmpty)
assert(SparkConnectPluginRegistry.loadExpressionPlugins().isEmpty)
assert(SparkConnectPluginRegistry.loadCommandPlugins().isEmpty)
diff --git
a/core/src/main/scala/org/apache/spark/storage/BlockInfoManager.scala
b/core/src/main/scala/org/apache/spark/storage/BlockInfoManager.scala
index f80190c96e8..73e72b7f1df 100644
--- a/core/src/main/scala/org/apache/spark/storage/BlockInfoManager.scala
+++ b/core/src/main/scala/org/apache/spark/storage/BlockInfoManager.scala
@@ -259,7 +259,7 @@ private[storage] class
BlockInfoManager(trackingCacheVisibility: Boolean = false
}
/**
- * Apply function `f` on the [[BlockInfo]] object and the aquisition
[[Condition]] for `blockId`.
+ * Apply function `f` on the [[BlockInfo]] object and the acquisition
[[Condition]] for `blockId`.
* Function `f` will be executed while holding the lock for the
[[BlockInfo]] object. If `blockId`
* was not registered, an error will be thrown.
*/
diff --git
a/core/src/main/scala/org/apache/spark/storage/DiskBlockManager.scala
b/core/src/main/scala/org/apache/spark/storage/DiskBlockManager.scala
index 1e8289287cb..6118631e549 100644
--- a/core/src/main/scala/org/apache/spark/storage/DiskBlockManager.scala
+++ b/core/src/main/scala/org/apache/spark/storage/DiskBlockManager.scala
@@ -79,9 +79,9 @@ private[spark] class DiskBlockManager(
private val shutdownHook = addShutdownHook()
// If either of these features are enabled, we must change permissions on
block manager
- // directories and files to accomodate the shuffle service deleting files in
a secure environment.
- // Parent directories are assumed to be restrictive to prevent unauthorized
users from accessing
- // or modifying world readable files.
+ // directories and files to accommodate the shuffle service deleting files
in a secure
+ // environment. Parent directories are assumed to be restrictive to prevent
unauthorized users
+ // from accessing or modifying world readable files.
private val permissionChangingRequired =
conf.get(config.SHUFFLE_SERVICE_ENABLED) && (
conf.get(config.SHUFFLE_SERVICE_REMOVE_SHUFFLE_ENABLED) ||
conf.get(config.SHUFFLE_SERVICE_FETCH_RDD_ENABLED)
diff --git
a/core/src/main/scala/org/apache/spark/ui/jobs/TaskThreadDumpPage.scala
b/core/src/main/scala/org/apache/spark/ui/jobs/TaskThreadDumpPage.scala
index 5349361e014..7b6da1147c5 100644
--- a/core/src/main/scala/org/apache/spark/ui/jobs/TaskThreadDumpPage.scala
+++ b/core/src/main/scala/org/apache/spark/ui/jobs/TaskThreadDumpPage.scala
@@ -102,7 +102,7 @@ private[spark] class TaskThreadDumpPage(
</div>
</div>
}.getOrElse{
- Text(s"Task $taskId finished or some error occured during dumping
thread")
+ Text(s"Task $taskId finished or some error occurred during dumping
thread")
}
UIUtils.headerSparkPage(request, s"Thread dump for task $taskId", content,
parent)
}
diff --git
a/core/src/test/scala/org/apache/spark/status/AutoCleanupLiveUIDirSuite.scala
b/core/src/test/scala/org/apache/spark/status/AutoCleanupLiveUIDirSuite.scala
index f717299a1ed..4ada18e8d93 100644
---
a/core/src/test/scala/org/apache/spark/status/AutoCleanupLiveUIDirSuite.scala
+++
b/core/src/test/scala/org/apache/spark/status/AutoCleanupLiveUIDirSuite.scala
@@ -36,7 +36,7 @@ class AutoCleanupLiveUIDirSuite extends SparkFunSuite {
_ + _
}
.collect()
- // `baseUIDir` should exists and not emtpy before SparkContext stop.
+ // `baseUIDir` should exists and not empty before SparkContext stop.
assert(baseUIDir.exists())
val subDirs = baseUIDir.listFiles()
assert(subDirs.nonEmpty)
diff --git a/docs/sql-ref-syntax-ddl-declare-variable.md
b/docs/sql-ref-syntax-ddl-declare-variable.md
index eea6222646f..f4daeb25579 100644
--- a/docs/sql-ref-syntax-ddl-declare-variable.md
+++ b/docs/sql-ref-syntax-ddl-declare-variable.md
@@ -69,7 +69,7 @@ DECLARE [ OR REPLACE ] [ VARIABLE ]
### Examples
```sql
--- The dense form of declaring a variabel with default
+-- The dense form of declaring a variable with default
DECLARE five = 5;
-- STRING variable initialialized to `NULL`
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]