This is an automated email from the ASF dual-hosted git repository. lzljs3620320 pushed a commit to branch release-0.6 in repository https://gitbox.apache.org/repos/asf/incubator-paimon.git
commit 1e96efdb942fa0b6981fe57848ce19993ef9996b Author: Zouxxyy <[email protected]> AuthorDate: Thu Dec 7 10:14:41 2023 +0800 [spark] Show create table display using paimon (#2460) --- .../src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java | 4 ++-- .../src/main/java/org/apache/paimon/spark/SparkTable.java | 4 +++- .../src/test/java/org/apache/paimon/spark/SparkReadITCase.java | 1 - .../src/test/java/org/apache/paimon/spark/SparkReadTestBase.java | 2 +- 4 files changed, 6 insertions(+), 5 deletions(-) diff --git a/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java b/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java index 2d82fa8bb..a94d82765 100644 --- a/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java +++ b/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkGenericCatalog.java @@ -180,7 +180,7 @@ public class SparkGenericCatalog<T extends TableCatalog & SupportsNamespaces> Transform[] partitions, Map<String, String> properties) throws TableAlreadyExistsException, NoSuchNamespaceException { - String provider = properties.get("provider"); + String provider = properties.get(TableCatalog.PROP_PROVIDER); if (usePaimon(provider)) { return sparkCatalog.createTable(ident, schema, partitions, properties); } else { @@ -284,7 +284,7 @@ public class SparkGenericCatalog<T extends TableCatalog & SupportsNamespaces> } private boolean usePaimon(String provider) { - return provider == null || "paimon".equalsIgnoreCase(provider); + return provider == null || SparkSource.NAME().equalsIgnoreCase(provider); } private T getSessionCatalog() { diff --git a/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkTable.java b/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkTable.java index f74754ff7..dd0d5bce1 100644 --- a/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkTable.java +++ b/paimon-spark/paimon-spark-common/src/main/java/org/apache/paimon/spark/SparkTable.java @@ -27,6 +27,7 @@ import org.apache.paimon.table.Table; import org.apache.spark.sql.connector.catalog.SupportsRead; import org.apache.spark.sql.connector.catalog.SupportsWrite; import org.apache.spark.sql.connector.catalog.TableCapability; +import org.apache.spark.sql.connector.catalog.TableCatalog; import org.apache.spark.sql.connector.expressions.FieldReference; import org.apache.spark.sql.connector.expressions.IdentityTransform; import org.apache.spark.sql.connector.expressions.Transform; @@ -108,10 +109,11 @@ public class SparkTable if (table instanceof DataTable) { Map<String, String> properties = new HashMap<>(((DataTable) table).coreOptions().toMap()); - if (table.primaryKeys().size() > 0) { + if (!table.primaryKeys().isEmpty()) { properties.put( CoreOptions.PRIMARY_KEY.key(), String.join(",", table.primaryKeys())); } + properties.put(TableCatalog.PROP_PROVIDER, SparkSource.NAME()); return properties; } else { return Collections.emptyMap(); diff --git a/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadITCase.java b/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadITCase.java index 486b88110..a818195ef 100644 --- a/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadITCase.java +++ b/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadITCase.java @@ -363,7 +363,6 @@ public class SparkReadITCase extends SparkReadTestBase { .isEqualTo( String.format( "[[%s" - + "USING paimon\n" + "PARTITIONED BY (b)\n" + "COMMENT 'tbl comment'\n" + "TBLPROPERTIES (\n" diff --git a/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadTestBase.java b/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadTestBase.java index f566f2fb7..ba421d09a 100644 --- a/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadTestBase.java +++ b/paimon-spark/paimon-spark-common/src/test/java/org/apache/paimon/spark/SparkReadTestBase.java @@ -198,7 +198,7 @@ public abstract class SparkReadTestBase { // return of 'SHOW CREATE TABLE' excluding TBLPROPERTIES protected String showCreateString(String table, String... fieldSpec) { return String.format( - "CREATE TABLE paimon.default.%s (%s)\n", + "CREATE TABLE paimon.default.%s (%s)\nUSING paimon\n", table, Arrays.stream(fieldSpec).map(s -> "\n " + s).collect(Collectors.joining(","))); }
