Github user jiangxb1987 commented on a diff in the pull request:
https://github.com/apache/spark/pull/20177#discussion_r162746808
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/command/AnalyzeTableCommand.scala
---
@@ -31,9 +31,9 @@ case class AnalyzeTableCommand(
override def run(sparkSession: SparkSession): Seq[Row] = {
val sessionState = sparkSession.sessionState
- val db =
tableIdent.database.getOrElse(sessionState.catalog.getCurrentDatabase)
- val tableIdentWithDB = TableIdentifier(tableIdent.table, Some(db))
- val tableMeta = sessionState.catalog.getTableMetadata(tableIdentWithDB)
+ val db = tableIdent.database
+ val tableIdentWithDB = TableIdentifier(tableIdent.table, db)
+ val tableMeta =
sessionState.catalog.getTempViewOrPermanentTableMetadata(tableIdentWithDB)
--- End diff --
Wouldn't this fail if we have a table that neglect the current database in
tableIdent?
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]