This is an automated email from the ASF dual-hosted git repository.

chenliang613 pushed a commit to branch chenliang613-patch-3
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit fa8dd2e05d56f2ecd430a219f114cb8887fe5d55
Author: Liang Chen <[email protected]>
AuthorDate: Mon Oct 6 20:29:43 2025 +0800

    Revise README for AI-native data clarity
    
    Updated the README to clarify the concept of AI-native data and its 
relevance to CarbonData. Improved formatting for better readability.
---
 AI-DATA/README.md | 20 +++++++++-----------
 1 file changed, 9 insertions(+), 11 deletions(-)

diff --git a/AI-DATA/README.md b/AI-DATA/README.md
index ef40ce0a33..fe052badce 100644
--- a/AI-DATA/README.md
+++ b/AI-DATA/README.md
@@ -18,27 +18,25 @@
 <img src="/docs/images/CarbonData_logo.png" width="200" height="40">
 
 
-## What is AI-native data storage
+## What is AI-native data
 
-* AI-native data storage is a data storage and management system designed and 
built specifically for the needs of artificial intelligence (AI) workloads, 
particularly machine learning and deep learning. Its core concept is to 
transform data storage from a passive, isolated component of the AI ​​process 
into an active, intelligent, and deeply integrated infrastructure.
+AI-native data storage is a data storage and management system designed and 
built specifically for the needs of artificial intelligence (AI) workloads, 
particularly machine learning and deep learning. Its core concept is to 
transform data storage from a passive, isolated component of the AI ​​process 
into an active, intelligent, and deeply integrated infrastructure.
 
-## Why AI-native data storage for CarbonData's new scope
+## Why AI-native data for CarbonData's new scope
 
 In AI projects, data scientists and engineers spend 80% of their time on data 
preparation. Traditional storage presents numerous bottlenecks in this process:
 
-Data silos: Training data may be scattered across data lakes, data warehouses, 
file systems, object storage, and other locations, making integration difficult.
+* Data silos: Training data may be scattered across data lakes, data 
warehouses, file systems, object storage, and other locations, making 
integration difficult.
 
-Performance bottlenecks:
+* Performance bottlenecks:Training phase: High-speed, low-latency data 
throughput is required to feed GPUs to avoid expensive GPU resources sitting 
idle.
 
-Training phase: High-speed, low-latency data throughput is required to feed 
GPUs to avoid expensive GPU resources sitting idle.
+* Inference phase: High-concurrency, low-latency vector similarity search 
capabilities are required.
 
-Inference phase: High-concurrency, low-latency vector similarity search 
capabilities are required.
+* Complex data formats: AI processes data types far beyond tables, including 
unstructured data (images, videos, text, audio) and semi-structured data (JSON, 
XML). Traditional databases have limited capabilities for processing and 
querying such data.
 
-Complex data formats: AI processes data types far beyond tables, including 
unstructured data (images, videos, text, audio) and semi-structured data (JSON, 
XML). Traditional databases have limited capabilities for processing and 
querying such data.
+* Lack of metadata management: The lack of effective management of rich 
metadata such as data versions, lineage, annotation information, and 
experimental parameters leads to poor experimental reproducibility.
 
-Lack of metadata management: The lack of effective management of rich metadata 
such as data versions, lineage, annotation information, and experimental 
parameters leads to poor experimental reproducibility.
-
-Vectorization requirements: Modern AI models (such as large language models) 
convert all data into vector embeddings. Traditional storage cannot efficiently 
store and retrieve high-dimensional vectors.
+* Vectorization requirements: Modern AI models (such as large language models) 
convert all data into vector embeddings. Traditional storage cannot efficiently 
store and retrieve high-dimensional vectors.
   
 
 ## About

Reply via email to