This is an automated email from the ASF dual-hosted git repository.
richox pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/auron.git
The following commit(s) were added to refs/heads/master by this push:
new f3e9347f [AURON #1652] Fix: add Spark_ prefix for Hashing Functions in
Auron ext functions (#1653)
f3e9347f is described below
commit f3e9347f3645a6094a861b4e7de3f89e845ca132
Author: Thomas <[email protected]>
AuthorDate: Tue Nov 25 12:42:03 2025 +0800
[AURON #1652] Fix: add Spark_ prefix for Hashing Functions in Auron ext
functions (#1653)
* [AURON #1652] Fix: add Spark_ prefix for Hashing Functions in Auron ext
functions
* fixup
---
native-engine/datafusion-ext-functions/src/lib.rs | 10 +++++-----
.../scala/org/apache/spark/sql/auron/NativeConverters.scala | 2 +-
2 files changed, 6 insertions(+), 6 deletions(-)
diff --git a/native-engine/datafusion-ext-functions/src/lib.rs
b/native-engine/datafusion-ext-functions/src/lib.rs
index 0743d26c..b99e406d 100644
--- a/native-engine/datafusion-ext-functions/src/lib.rs
+++ b/native-engine/datafusion-ext-functions/src/lib.rs
@@ -45,11 +45,11 @@ pub fn create_auron_ext_function(name: &str) ->
Result<ScalarFunctionImplementat
"Spark_CheckOverflow" =>
Arc::new(spark_check_overflow::spark_check_overflow),
"Spark_Murmur3Hash" => Arc::new(spark_hash::spark_murmur3_hash),
"Spark_XxHash64" => Arc::new(spark_hash::spark_xxhash64),
- "Sha224" => Arc::new(spark_crypto::spark_sha224),
- "Sha256" => Arc::new(spark_crypto::spark_sha256),
- "Sha384" => Arc::new(spark_crypto::spark_sha384),
- "Sha512" => Arc::new(spark_crypto::spark_sha512),
- "MD5" => Arc::new(spark_crypto::spark_md5),
+ "Spark_Sha224" => Arc::new(spark_crypto::spark_sha224),
+ "Spark_Sha256" => Arc::new(spark_crypto::spark_sha256),
+ "Spark_Sha384" => Arc::new(spark_crypto::spark_sha384),
+ "Spark_Sha512" => Arc::new(spark_crypto::spark_sha512),
+ "Spark_MD5" => Arc::new(spark_crypto::spark_md5),
"Spark_GetJsonObject" =>
Arc::new(spark_get_json_object::spark_get_json_object),
"Spark_GetParsedJsonObject" => {
Arc::new(spark_get_json_object::spark_get_parsed_json_object)
diff --git
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
index d98121a9..2447a9ef 100644
---
a/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
+++
b/spark-extension/src/main/scala/org/apache/spark/sql/auron/NativeConverters.scala
@@ -864,7 +864,7 @@ object NativeConverters extends Logging {
case e @ NullIf(left, right, _) =>
buildExtScalarFunction("Spark_NullIf", left :: right :: Nil,
e.dataType)
case Md5(_1) =>
- buildExtScalarFunction("MD5", Seq(unpackBinaryTypeCast(_1)),
StringType)
+ buildExtScalarFunction("Spark_MD5", Seq(unpackBinaryTypeCast(_1)),
StringType)
case Reverse(_1) =>
buildScalarFunction(pb.ScalarFunction.Reverse,
Seq(unpackBinaryTypeCast(_1)), StringType)
case InitCap(_1) =>