This is an automated email from the ASF dual-hosted git repository.

changchen pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-gluten.git


The following commit(s) were added to refs/heads/main by this push:
     new e29946dc46 [GLUTEN-1632][CH]Daily Update Clickhouse Version (20240928) 
(#7379)
e29946dc46 is described below

commit e29946dc4639567dae8a3f4362ee9b72e9cb4357
Author: Kyligence Git <[email protected]>
AuthorDate: Sat Sep 28 06:29:08 2024 -0500

    [GLUTEN-1632][CH]Daily Update Clickhouse Version (20240928) (#7379)
    
    * [GLUTEN-1632][CH]Daily Update Clickhouse Version (20240928)
    
    * Fix build due to https://github.com/ClickHouse/ClickHouse/pull/65625
    
    * Fix build due to https://github.com/ClickHouse/ClickHouse/pull/68410
    
    ---------
    
    Co-authored-by: kyligence-git <[email protected]>
    Co-authored-by: Chang Chen <[email protected]>
---
 cpp-ch/clickhouse.version                                        | 4 ++--
 cpp-ch/local-engine/Parser/RelParsers/AggregateRelParser.cpp     | 1 -
 cpp-ch/local-engine/Parser/RelParsers/SortRelParser.cpp          | 2 +-
 cpp-ch/local-engine/Storages/MergeTree/MetaDataHelper.cpp        | 4 ++--
 cpp-ch/local-engine/Storages/MergeTree/SparkStorageMergeTree.cpp | 2 +-
 5 files changed, 6 insertions(+), 7 deletions(-)

diff --git a/cpp-ch/clickhouse.version b/cpp-ch/clickhouse.version
index 32d2acdbdf..4b86a99b17 100644
--- a/cpp-ch/clickhouse.version
+++ b/cpp-ch/clickhouse.version
@@ -1,3 +1,3 @@
 CH_ORG=Kyligence
-CH_BRANCH=rebase_ch/20240920
-CH_COMMIT=14c2da664d7
\ No newline at end of file
+CH_BRANCH=rebase_ch/20240928
+CH_COMMIT=348196b6011
\ No newline at end of file
diff --git a/cpp-ch/local-engine/Parser/RelParsers/AggregateRelParser.cpp 
b/cpp-ch/local-engine/Parser/RelParsers/AggregateRelParser.cpp
index 1e9f1da098..06440b5029 100644
--- a/cpp-ch/local-engine/Parser/RelParsers/AggregateRelParser.cpp
+++ b/cpp-ch/local-engine/Parser/RelParsers/AggregateRelParser.cpp
@@ -336,7 +336,6 @@ void AggregateRelParser::addMergingAggregatedStep()
             false,
             settings[Setting::max_block_size],
             settings[Setting::aggregation_in_order_max_block_bytes],
-            SortDescription(),
             
settings[Setting::enable_memory_bound_merging_of_aggregation_results]);
         steps.emplace_back(merging_step.get());
         plan->addStep(std::move(merging_step));
diff --git a/cpp-ch/local-engine/Parser/RelParsers/SortRelParser.cpp 
b/cpp-ch/local-engine/Parser/RelParsers/SortRelParser.cpp
index 48e29234e1..e1d31d5cdb 100644
--- a/cpp-ch/local-engine/Parser/RelParsers/SortRelParser.cpp
+++ b/cpp-ch/local-engine/Parser/RelParsers/SortRelParser.cpp
@@ -50,7 +50,7 @@ SortRelParser::parse(DB::QueryPlanPtr query_plan, const 
substrait::Rel & rel, st
         return currentThreadGroupMemoryUsageRatio() > spill_mem_ratio;
     };
     auto sorting_step = std::make_unique<DB::SortingStep>(
-        query_plan->getCurrentDataStream(), sort_descr, limit, settings, 
false);
+        query_plan->getCurrentDataStream(), sort_descr, limit, settings);
     sorting_step->setStepDescription("Sorting step");
     steps.emplace_back(sorting_step.get());
     query_plan->addStep(std::move(sorting_step));
diff --git a/cpp-ch/local-engine/Storages/MergeTree/MetaDataHelper.cpp 
b/cpp-ch/local-engine/Storages/MergeTree/MetaDataHelper.cpp
index 0e6f5b102f..d9193a03dd 100644
--- a/cpp-ch/local-engine/Storages/MergeTree/MetaDataHelper.cpp
+++ b/cpp-ch/local-engine/Storages/MergeTree/MetaDataHelper.cpp
@@ -118,7 +118,7 @@ void restoreMetaData<ROCKSDB>(const 
SparkStorageMergeTreePtr & storage, const Me
                     transaction->createDirectoryRecursive(part_path);
                 auto key = 
s3->generateObjectKeyForPath(metadata_file_path.generic_string(), std::nullopt);
                 StoredObject metadata_object(key.serialize());
-                auto part_metadata = 
extractPartMetaData(*s3->readObject(metadata_object));
+                auto part_metadata = 
extractPartMetaData(*s3->readObject(metadata_object, ReadSettings{}));
                 for (const auto & item : part_metadata)
                 {
                     auto item_path = part_path / item.first;
@@ -180,7 +180,7 @@ void restoreMetaData<LOCAL>(
                     metadata_disk->createDirectories(part_path);
                 auto key = 
s3->generateObjectKeyForPath(metadata_file_path.generic_string(), std::nullopt);
                 StoredObject metadata_object(key.serialize());
-                auto part_metadata = 
extractPartMetaData(*s3->readObject(metadata_object));
+                auto part_metadata = 
extractPartMetaData(*s3->readObject(metadata_object, ReadSettings{}));
                 for (const auto & item : part_metadata)
                 {
                     auto item_path = part_path / item.first;
diff --git a/cpp-ch/local-engine/Storages/MergeTree/SparkStorageMergeTree.cpp 
b/cpp-ch/local-engine/Storages/MergeTree/SparkStorageMergeTree.cpp
index 3fe3890fb7..e1b3c1d073 100644
--- a/cpp-ch/local-engine/Storages/MergeTree/SparkStorageMergeTree.cpp
+++ b/cpp-ch/local-engine/Storages/MergeTree/SparkStorageMergeTree.cpp
@@ -157,7 +157,7 @@ void 
SparkStorageMergeTree::prefetchMetaDataFile(std::unordered_set<std::string>
     {
         if (!disk->exists(meta_path))
             continue;
-        auto in = disk->readFile(meta_path);
+        auto in = disk->readFile(meta_path, ReadSettings{});
         String ignore_data;
         readStringUntilEOF(ignore_data, *in);
     }


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to