This is an automated email from the ASF dual-hosted git repository.

lzljs3620320 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/paimon.git


The following commit(s) were added to refs/heads/master by this push:
     new 88a45bcd2 [doc] Add SQL examples in table-types
88a45bcd2 is described below

commit 88a45bcd2f823a15706a467cbbdfbc7c24f60511
Author: Jingsong <[email protected]>
AuthorDate: Mon Nov 11 11:02:53 2024 +0800

    [doc] Add SQL examples in table-types
---
 docs/content/concepts/table-types.md | 70 +++++++++++++++++++++++++++++-------
 1 file changed, 57 insertions(+), 13 deletions(-)

diff --git a/docs/content/concepts/table-types.md 
b/docs/content/concepts/table-types.md
index 88f0cc66c..49e447962 100644
--- a/docs/content/concepts/table-types.md
+++ b/docs/content/concepts/table-types.md
@@ -46,6 +46,34 @@ sorting the primary key within each bucket, allowing 
streaming update and stream
 The definition of primary key is similar to that of standard SQL, as it 
ensures that there is only one data entry for
 the same primary key during batch queries.
 
+{{< tabs "primary-table" >}}
+{{< tab "Flink SQL" >}}
+
+```sql
+CREATE TABLE my_table (
+    a INT PRIMARY KEY NOT ENFORCED,
+    b STRING
+) WITH (
+    'bucket'='8'
+)
+```
+{{< /tab >}}
+
+{{< tab "Spark SQL" >}}
+
+```sql
+CREATE TABLE my_table (
+    a INT,
+    b STRING
+) TBLPROPERTIES (
+    'primary-key' = 'a',
+    'bucket' = '8'
+)
+```
+
+{{< /tab >}}
+{{< /tabs >}}
+
 ## Table w/o PK
 
 See [Paimon w/o Primary key]({{< ref "append-table/overview" >}}).
@@ -56,6 +84,13 @@ can only receive incoming data from append data.
 
 However, it also supports batch sql: DELETE, UPDATE, and MERGE-INTO.
 
+```sql
+CREATE TABLE my_table (
+    a INT,
+    b STRING
+)
+```
+
 ## View
 
 View is supported when the metastore can support view, for example, hive 
metastore.
@@ -82,10 +117,8 @@ Format Table is enabled by default, you can disable it by 
configuring Catalog op
 
 Currently only support `CSV`, `Parquet`, `ORC` formats.
 
-### CSV
-
-{{< tabs "format-table-csv" >}}
-{{< tab "Flink SQL" >}}
+{{< tabs "format-table" >}}
+{{< tab "Flink-CSV" >}}
 
 ```sql
 CREATE TABLE my_csv_table (
@@ -99,7 +132,7 @@ CREATE TABLE my_csv_table (
 ```
 {{< /tab >}}
 
-{{< tab "Spark SQL" >}}
+{{< tab "Spark-CSV" >}}
 
 ```sql
 CREATE TABLE my_csv_table (
@@ -109,14 +142,8 @@ CREATE TABLE my_csv_table (
 ```
 
 {{< /tab >}}
-{{< /tabs >}}
-
-Now, only support `'field-delimiter'` option.
 
-### Parquet & ORC
-
-{{< tabs "format-table-parquet" >}}
-{{< tab "Flink SQL" >}}
+{{< tab "Flink-Parquet" >}}
 
 ```sql
 CREATE TABLE my_parquet_table (
@@ -129,7 +156,7 @@ CREATE TABLE my_parquet_table (
 ```
 {{< /tab >}}
 
-{{< tab "Spark SQL" >}}
+{{< tab "Spark-Parquet" >}}
 
 ```sql
 CREATE TABLE my_parquet_table (
@@ -139,6 +166,7 @@ CREATE TABLE my_parquet_table (
 ```
 
 {{< /tab >}}
+
 {{< /tabs >}}
 
 ## Materialized Table
@@ -147,3 +175,19 @@ Materialized Table aimed at simplifying both batch and 
stream data pipelines, pr
 experience, see [Flink Materialized 
Table](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/materialized-table/overview/).
 
 Now only Flink SQL integrate to Materialized Table, we plan to support it in 
Spark SQL too.
+
+```sql
+CREATE MATERIALIZED TABLE continuous_users_shops
+PARTITIONED BY (ds)
+FRESHNESS = INTERVAL '30' SECOND
+AS SELECT
+  user_id,
+  ds,
+  SUM (payment_amount_cents) AS payed_buy_fee_sum,
+  SUM (1) AS PV
+FROM (
+  SELECT user_id, order_created_at AS ds, payment_amount_cents
+    FROM json_source
+  ) AS tmp
+GROUP BY user_id, ds;
+```

Reply via email to