venkata91 commented on a change in pull request #28874:
URL: https://github.com/apache/spark/pull/28874#discussion_r443089270



##########
File path: core/src/test/scala/org/apache/spark/ui/UISeleniumSuite.scala
##########
@@ -48,24 +48,24 @@ import org.apache.spark.util.CallSite
 
 private[spark] class SparkUICssErrorHandler extends DefaultCssErrorHandler {
 
-  private val cssWhiteList = List("bootstrap.min.css", 
"vis-timeline-graph2d.min.css")
+  private val cssExcludeList = List("bootstrap.min.css", 
"vis-timeline-graph2d.min.css")

Review comment:
       also it seems in some cases we use `exclude` for both `whitelist` as 
well as `blacklist`. Like here `exclude` is used for `whiteList` and 
[here](https://github.com/apache/spark/blob/8f414bc6b4eeb59203ecb33c26c762a57bf5429e/resource-managers/mesos/src/main/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterScheduler.scala#L542)
 for `blackList`. In general, I prefer `allowedList` for `whitelist` and 
`denyList` or `rejectList` or `stopList` etc for `blacklist` makes it easier to 
comprehend quickly. I understand its hard to use the same word everywhere 
because of the context.

##########
File path: core/src/test/scala/org/apache/spark/ui/UISeleniumSuite.scala
##########
@@ -48,24 +48,24 @@ import org.apache.spark.util.CallSite
 
 private[spark] class SparkUICssErrorHandler extends DefaultCssErrorHandler {
 
-  private val cssWhiteList = List("bootstrap.min.css", 
"vis-timeline-graph2d.min.css")
+  private val cssExcludeList = List("bootstrap.min.css", 
"vis-timeline-graph2d.min.css")

Review comment:
       same here instead of `exclude` how about `allowed`?

##########
File path: R/pkg/tests/fulltests/test_sparkSQL.R
##########
@@ -3921,14 +3921,14 @@ test_that("No extra files are created in SPARK_HOME by 
starting session and maki
   # before creating a SparkSession with enableHiveSupport = T at the top of 
this test file
   # (filesBefore). The test here is to compare that (filesBefore) against the 
list of files before
   # any test is run in run-all.R (sparkRFilesBefore).
-  # sparkRWhitelistSQLDirs is also defined in run-all.R, and should contain 
only 2 whitelisted dirs,
+  # sparkRIncludedSQLDirs is also defined in run-all.R, and should contain 
only 2 included dirs,

Review comment:
       does it make sense to have `allowed` here instead of `included`?




----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to