This is an automated email from the ASF dual-hosted git repository.

ipolyzos pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/fluss.git


The following commit(s) were added to refs/heads/main by this push:
     new b2f9c2f67 feat: blog hands on lakehouse (#1403)
b2f9c2f67 is described below

commit b2f9c2f6769a28fe0b0a626770208cc877d3b743
Author: Yang Guo <[email protected]>
AuthorDate: Thu Jul 24 19:36:04 2025 +0800

    feat: blog hands on lakehouse (#1403)
---
 website/blog/2025-07-23-hands-on-fluss-lakehouse.md | 12 ++++++------
 1 file changed, 6 insertions(+), 6 deletions(-)

diff --git a/website/blog/2025-07-23-hands-on-fluss-lakehouse.md 
b/website/blog/2025-07-23-hands-on-fluss-lakehouse.md
index e35b8acf9..5374fa903 100644
--- a/website/blog/2025-07-23-hands-on-fluss-lakehouse.md
+++ b/website/blog/2025-07-23-hands-on-fluss-lakehouse.md
@@ -1,6 +1,6 @@
 ---
 slug: hands-on-fluss-lakehouse
-title: "Hands-on Fluss Lakehouse with Paimon S3"
+title: "From Stream to Lake: Hands-On with Fluss Tiering into Paimon on Minio"
 authors: [gyang94]
 toc_max_heading_level: 5
 ---
@@ -23,7 +23,7 @@ toc_max_heading_level: 5
  limitations under the License.
 -->
 
-# Hands-on Fluss Lakehouse with Paimon S3
+# From Stream to Lake: Hands-On with Fluss Tiering into Paimon on Minio
 
 Fluss stores historical data in a lakehouse storage layer while keeping 
real-time data in the Fluss server. Its built-in tiering service continuously 
moves fresh events into the lakehouse, allowing various query engines to 
analyze both hot and cold data. The real magic happens with Fluss's union-read 
capability, which lets Flink jobs seamlessly query both the Fluss cluster and 
the lakehouse for truly integrated real-time processing.
 
@@ -31,9 +31,9 @@ Fluss stores historical data in a lakehouse storage layer 
while keeping real-tim
 
 In this hands-on tutorial, we'll walk you through setting up a local Fluss 
lakehouse environment, running some practical data operations, and getting 
first-hand experience with the complete Fluss lakehouse architecture. By the 
end, you'll have a working environment for experimenting with Fluss's powerful 
data processing capabilities.
 
-## Integrate Paimon S3 Lakehouse
+## Integrate with Paimon Minio Lakehouse
 
-For this tutorial, we'll use **Fluss 0.7** and **Flink 1.20** to run the 
tiering service on a local cluster. We'll configure **Paimon** as our lake 
format and **S3** as the storage backend. Let's get started:
+For this tutorial, we'll use **Fluss 0.7** and **Flink 1.20** to run the 
tiering service on a local cluster. We'll configure **Paimon** as our lake 
format on **Minio** as the storage backend. Let's get started:
 
 ### Minio Setup
 
@@ -97,7 +97,7 @@ For this tutorial, we'll use **Fluss 0.7** and **Flink 1.20** 
to run the tiering
    datalake.paimon.s3.path.style.access: true
    ```
 
-   This configures Paimon as the datalake format with S3 as the warehouse.
+   This configures Paimon as the datalake format on Minio as the warehouse.
 
 4. Start Fluss
 
@@ -366,4 +366,4 @@ Now let's dive into some actual data processing. We'll use 
the Flink SQL Client
 
 ## Summary
 
-In this guide, we've explored the Fluss lakehouse architecture and set up a 
complete local environment with Fluss, Flink, Paimon, and S3. We've walked 
through practical examples of data processing that showcase how Fluss 
seamlessly integrates real-time and historical data. With this setup, you now 
have a solid foundation for experimenting with Fluss's powerful lakehouse 
capabilities on your own machine.
\ No newline at end of file
+In this guide, we've explored the Fluss lakehouse architecture and set up a 
complete local environment with Fluss, Flink, Paimon, and Minio. We've walked 
through practical examples of data processing that showcase how Fluss 
seamlessly integrates real-time and historical data. With this setup, you now 
have a solid foundation for experimenting with Fluss's powerful lakehouse 
capabilities on your own machine.
\ No newline at end of file

Reply via email to