This is an automated email from the ASF dual-hosted git repository.
yuanzhou pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-gluten.git
The following commit(s) were added to refs/heads/main by this push:
new ec87a27751 [GLUTEN-6887][VL] Daily Update Velox Version (2025_08_15)
(#10449)
ec87a27751 is described below
commit ec87a27751b87554b539c62a4765dbd6a868d849
Author: Gluten Performance Bot
<[email protected]>
AuthorDate: Fri Aug 15 14:06:03 2025 +0100
[GLUTEN-6887][VL] Daily Update Velox Version (2025_08_15) (#10449)
* [GLUTEN-6887][VL] Daily Update Velox Version (2025_08_15)
Upstream Velox's New Commits:
17580eb38 by Pedro Eugenio Rocha Pedreira, feat(python): Support for unnest
and streaming aggregates (#14486)
0736383a7 by Karthikeyan, fix(cudf): HashJoin null behavior (#14321)
7653c77c2 by Peter Enescu, fix: Revert signature filtering due to bias
fuzzer flakiness (#14463)
dbb06e021 by Xiaoxuan Meng, feat: Add remaining filter support for index
source connector (#14470)
2478c8788 by Ke Wang, fix: Use customized doubleCompare for comparing
(#14393)
7b7f15a47 by Pablo pudge1000-7 Ivanov, test: Add support for AnyJoin to
DuckLogicalOperator.h (#14478)
6af7a183a by mbkkt, misc: Avoid shared_ptr control block for complex types
too (#14401)
2e4a0835c by Hongze Zhang, fix: Fix crash by stack overflow when grouping
by nano-timestamp key (#14383)
47deca056 by lingbin, refactor: Use raw string literal for escaped strings
in VariantTest (#13287)
2d342604e by mbkkt, misc: Use C++20 default comparison (#14402)
9825b1a74 by beliefer, refactor: Remove unnessary ScopedVarSetter (#14473)
07004d70d by Ke Wang, feat(memory): Support memory reclaim in nimble writer
(#14323)
38d26b905 by Chen Zhang, fix: Improve tzdb to be compatible with old
tzdata.zi (#14440)
05a1af57e by mbkkt, refactor: Replace folly::Optional with std::optional
(#14455)
2a1eea49e by Jimmy Lu, feat: Add variant of project node that splits
subfields of struct without materialization (#14461)
e971b986c by Zhen Li, feat: Add Spark get_array_struct_fields function
(#14292)
654a936cd by kavinli, fix: print bucket and key in s3 object already exists
error message (#14347)
Signed-off-by: glutenperfbot <[email protected]>
* fix std::optional has_value
Signed-off-by: Yuan <[email protected]>
* fix
Signed-off-by: Yuan <[email protected]>
---------
Signed-off-by: glutenperfbot <[email protected]>
Signed-off-by: Yuan <[email protected]>
Co-authored-by: glutenperfbot <[email protected]>
Co-authored-by: Yuan <[email protected]>
---
cpp/velox/compute/VeloxBackend.cc | 2 +-
cpp/velox/compute/WholeStageResultIterator.cc | 2 +-
cpp/velox/utils/ConfigExtractor.cc | 14 +++++++-------
ep/build-velox/src/get_velox.sh | 4 ++--
4 files changed, 11 insertions(+), 11 deletions(-)
diff --git a/cpp/velox/compute/VeloxBackend.cc
b/cpp/velox/compute/VeloxBackend.cc
index 7ba2392f9b..25ad26a10d 100644
--- a/cpp/velox/compute/VeloxBackend.cc
+++ b/cpp/velox/compute/VeloxBackend.cc
@@ -199,7 +199,7 @@ void VeloxBackend::init(
// Spark off-heap memory pool will be conducted to cause unexpected OOMs.
auto sparkOverhead = backendConf_->get<int64_t>(kSparkOverheadMemory);
int64_t memoryManagerCapacity;
- if (sparkOverhead.hasValue()) {
+ if (sparkOverhead.has_value()) {
// 0.75 * total overhead memory is used for Velox global memory manager.
// FIXME: Make this configurable.
memoryManagerCapacity = sparkOverhead.value() * 0.75;
diff --git a/cpp/velox/compute/WholeStageResultIterator.cc
b/cpp/velox/compute/WholeStageResultIterator.cc
index c29957486d..62e68e0765 100644
--- a/cpp/velox/compute/WholeStageResultIterator.cc
+++ b/cpp/velox/compute/WholeStageResultIterator.cc
@@ -596,7 +596,7 @@ std::unordered_map<std::string, std::string>
WholeStageResultIterator::getQueryC
const auto setIfExists = [&](const std::string& glutenKey, const
std::string& veloxKey) {
const auto valueOptional = veloxCfg_->get<std::string>(glutenKey);
- if (valueOptional.hasValue()) {
+ if (valueOptional.has_value()) {
configs[veloxKey] = valueOptional.value();
}
};
diff --git a/cpp/velox/utils/ConfigExtractor.cc
b/cpp/velox/utils/ConfigExtractor.cc
index 6802470ca2..2e1aa92ff7 100644
--- a/cpp/velox/utils/ConfigExtractor.cc
+++ b/cpp/velox/utils/ConfigExtractor.cc
@@ -141,7 +141,7 @@ std::shared_ptr<facebook::velox::config::ConfigBase>
getHiveConfig(
hiveConfMap[S3Config::kS3PayloadSigningPolicy] =
conf->get<std::string>(kVeloxS3PayloadSigningPolicy,
kVeloxS3PayloadSigningPolicyDefault);
auto logLocation = conf->get<std::string>(kVeloxS3LogLocation);
- if (logLocation.hasValue()) {
+ if (logLocation.has_value()) {
hiveConfMap[S3Config::kS3LogLocation] = logLocation.value();
};
@@ -165,7 +165,7 @@ std::shared_ptr<facebook::velox::config::ConfigBase>
getHiveConfig(
#ifdef ENABLE_GCS
//
https://github.com/GoogleCloudDataproc/hadoop-connectors/blob/master/gcs/CONFIGURATION.md#api-client-configuration
auto gsStorageRootUrl =
conf->get<std::string>("spark.hadoop.fs.gs.storage.root.url");
- if (gsStorageRootUrl.hasValue()) {
+ if (gsStorageRootUrl.has_value()) {
std::string gcsEndpoint = gsStorageRootUrl.value();
if (!gcsEndpoint.empty()) {
@@ -176,21 +176,21 @@ std::shared_ptr<facebook::velox::config::ConfigBase>
getHiveConfig(
//
https://github.com/GoogleCloudDataproc/hadoop-connectors/blob/master/gcs/CONFIGURATION.md#http-transport-configuration
//
https://cloud.google.com/cpp/docs/reference/storage/latest/classgoogle_1_1cloud_1_1storage_1_1LimitedErrorCountRetryPolicy
auto gsMaxRetryCount =
conf->get<std::string>("spark.hadoop.fs.gs.http.max.retry");
- if (gsMaxRetryCount.hasValue()) {
+ if (gsMaxRetryCount.has_value()) {
hiveConfMap[facebook::velox::connector::hive::HiveConfig::kGcsMaxRetryCount] =
gsMaxRetryCount.value();
}
//
https://cloud.google.com/cpp/docs/reference/storage/latest/classgoogle_1_1cloud_1_1storage_1_1LimitedTimeRetryPolicy
auto gsMaxRetryTime =
conf->get<std::string>("spark.hadoop.fs.gs.http.max.retry-time");
- if (gsMaxRetryTime.hasValue()) {
+ if (gsMaxRetryTime.has_value()) {
hiveConfMap[facebook::velox::connector::hive::HiveConfig::kGcsMaxRetryTime] =
gsMaxRetryTime.value();
}
//
https://github.com/GoogleCloudDataproc/hadoop-connectors/blob/master/gcs/CONFIGURATION.md#authentication
auto gsAuthType = conf->get<std::string>("spark.hadoop.fs.gs.auth.type");
auto gsAuthServiceAccountJsonKeyfile =
conf->get<std::string>("spark.hadoop.fs.gs.auth.service.account.json.keyfile");
- if (gsAuthType.hasValue() && gsAuthType.value() ==
"SERVICE_ACCOUNT_JSON_KEYFILE") {
- if (gsAuthServiceAccountJsonKeyfile.hasValue()) {
+ if (gsAuthType.has_value() && gsAuthType.value() ==
"SERVICE_ACCOUNT_JSON_KEYFILE") {
+ if (gsAuthServiceAccountJsonKeyfile.has_value()) {
hiveConfMap[facebook::velox::connector::hive::HiveConfig::kGcsCredentialsPath] =
gsAuthServiceAccountJsonKeyfile.value();
} else {
@@ -198,7 +198,7 @@ std::shared_ptr<facebook::velox::config::ConfigBase>
getHiveConfig(
"however conf
spark.hadoop.fs.gs.auth.service.account.json.keyfile is not set";
throw GlutenException("Conf
spark.hadoop.fs.gs.auth.service.account.json.keyfile is not set");
}
- } else if (gsAuthServiceAccountJsonKeyfile.hasValue()) {
+ } else if (gsAuthServiceAccountJsonKeyfile.has_value()) {
LOG(WARNING) << "STARTUP: conf
spark.hadoop.fs.gs.auth.service.account.json.keyfile is set, "
"but conf spark.hadoop.fs.gs.auth.type is not
SERVICE_ACCOUNT_JSON_KEYFILE";
throw GlutenException("Conf spark.hadoop.fs.gs.auth.type is missing or
incorrect");
diff --git a/ep/build-velox/src/get_velox.sh b/ep/build-velox/src/get_velox.sh
index ceae99166a..16288e4cea 100755
--- a/ep/build-velox/src/get_velox.sh
+++ b/ep/build-velox/src/get_velox.sh
@@ -17,11 +17,11 @@
set -exu
VELOX_REPO=https://github.com/oap-project/velox.git
-VELOX_BRANCH=2025_08_14
+VELOX_BRANCH=2025_08_15
VELOX_HOME=""
RUN_SETUP_SCRIPT=ON
VELOX_ENHANCED_REPO=https://github.com/IBM/velox.git
-VELOX_ENHANCED_BRANCH=ibm-2025_08_14
+VELOX_ENHANCED_BRANCH=ibm-2025_08_15
ENABLE_ENHANCED_FEATURES=OFF
# Developer use only for testing Velox PR.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]