This is an automated email from the ASF dual-hosted git repository.

lzljs3620320 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/paimon.git


The following commit(s) were added to refs/heads/master by this push:
     new 15ae52319 [hotfix] fix the spark generic catalog to execute 
externalCatalog() creates a new hive metastore connection instance (#3670)
15ae52319 is described below

commit 15ae523197df8eccfd7686958ee156741d5300b2
Author: Kerwin <[email protected]>
AuthorDate: Fri Jul 5 13:39:54 2024 +0800

    [hotfix] fix the spark generic catalog to execute externalCatalog() creates 
a new hive metastore connection instance (#3670)
---
 .../main/java/org/apache/paimon/spark/SparkGenericCatalog.java    | 8 ++++----
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git 
a/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java
 
b/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java
index 9576379f8..96b87701b 100644
--- 
a/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java
+++ 
b/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java
@@ -35,7 +35,6 @@ import 
org.apache.spark.sql.catalyst.analysis.NoSuchTableException;
 import org.apache.spark.sql.catalyst.analysis.NonEmptyNamespaceException;
 import org.apache.spark.sql.catalyst.analysis.TableAlreadyExistsException;
 import org.apache.spark.sql.catalyst.catalog.ExternalCatalog;
-import org.apache.spark.sql.catalyst.catalog.InMemoryCatalog;
 import org.apache.spark.sql.catalyst.catalog.SessionCatalog;
 import org.apache.spark.sql.connector.catalog.CatalogExtension;
 import org.apache.spark.sql.connector.catalog.CatalogPlugin;
@@ -236,7 +235,8 @@ public class SparkGenericCatalog extends SparkBaseCatalog 
implements CatalogExte
 
     @Override
     public final void initialize(String name, CaseInsensitiveStringMap 
options) {
-        SessionState sessionState = SparkSession.active().sessionState();
+        SparkSession sparkSession = SparkSession.active();
+        SessionState sessionState = sparkSession.sessionState();
         Configuration hadoopConf = sessionState.newHadoopConf();
         SparkConf sparkConf = new SparkConf();
         if (options.containsKey(METASTORE.key())
@@ -253,8 +253,8 @@ public class SparkGenericCatalog extends SparkBaseCatalog 
implements CatalogExte
                 }
             }
         }
-        if (SparkSession.active().sharedState().externalCatalog().unwrapped()
-                instanceof InMemoryCatalog) {
+        if ("in-memory"
+                
.equals(sparkSession.conf().get(StaticSQLConf.CATALOG_IMPLEMENTATION().key()))) 
{
             LOG.warn("InMemoryCatalog here may cause bad effect.");
         }
 

Reply via email to