(arrow) branch main updated (ca0910a533 -> 5b03b707db)

2024-02-26 Thread curth
This is an automated email from the ASF dual-hosted git repository.

curth pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from ca0910a533 GH-39897: [C++][FS][S3] Ensure 
`AwsInstance::EnsureInitialized` to do initialization exactly once under 
concurrency (#40110)
 add 5b03b707db MINOR: [C#] Bump coverlet.collector from 6.0.0 to 6.0.1 in 
/csharp (#40245)

No new revisions were added by this update.

Summary of changes:
 .../Apache.Arrow.Flight.Sql.Tests/Apache.Arrow.Flight.Sql.Tests.csproj  | 2 +-
 csharp/test/Apache.Arrow.Flight.Tests/Apache.Arrow.Flight.Tests.csproj  | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)



(arrow) branch dependabot/nuget/csharp/coverlet.collector-6.0.1 deleted (was 63e0d92a0e)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/nuget/csharp/coverlet.collector-6.0.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


 was 63e0d92a0e MINOR: [C#] Bump coverlet.collector from 6.0.0 to 6.0.1 in 
/csharp

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow-datafusion) branch main updated: docs: update parquet_sql_multiple_files.rs with a relative path ex (#9310)

2024-02-26 Thread alamb
This is an automated email from the ASF dual-hosted git repository.

alamb pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


The following commit(s) were added to refs/heads/main by this push:
 new b8c6e0bb3f docs: update parquet_sql_multiple_files.rs with a relative 
path ex (#9310)
b8c6e0bb3f is described below

commit b8c6e0bb3f3f8ab60cf394b46d6aacc616b67d41
Author: Trent Hauck 
AuthorDate: Mon Feb 26 10:05:09 2024 -0800

docs: update parquet_sql_multiple_files.rs with a relative path ex (#9310)

* docs: update parquet_sql_multiple_files.rs with a relative path ex

* style: run cargo fmt

* docs: update comment

* docs: better
---
 .../examples/parquet_sql_multiple_files.rs | 66 ++
 1 file changed, 56 insertions(+), 10 deletions(-)

diff --git a/datafusion-examples/examples/parquet_sql_multiple_files.rs 
b/datafusion-examples/examples/parquet_sql_multiple_files.rs
index 451de96f2e..0e2968f203 100644
--- a/datafusion-examples/examples/parquet_sql_multiple_files.rs
+++ b/datafusion-examples/examples/parquet_sql_multiple_files.rs
@@ -17,31 +17,35 @@
 
 use datafusion::datasource::file_format::parquet::ParquetFormat;
 use datafusion::datasource::listing::ListingOptions;
-use datafusion::error::Result;
 use datafusion::prelude::*;
-use datafusion_common::{FileType, GetExt};
+use object_store::local::LocalFileSystem;
+use std::path::Path;
 use std::sync::Arc;
 
 /// This example demonstrates executing a simple query against an Arrow data 
source (a directory
-/// with multiple Parquet files) and fetching results
+/// with multiple Parquet files) and fetching results. The query is run twice, 
once showing
+/// how to used `register_listing_table` with an absolute path, and once 
registering an
+/// ObjectStore to use a relative path.
 #[tokio::main]
-async fn main() -> Result<()> {
+async fn main() -> Result<(), Box> {
 // create local execution context
 let ctx = SessionContext::new();
 
-let testdata = datafusion::test_util::parquet_test_data();
+let test_data = datafusion::test_util::parquet_test_data();
 
 // Configure listing options
 let file_format = ParquetFormat::default().with_enable_pruning(Some(true));
 let listing_options = ListingOptions::new(Arc::new(file_format))
-.with_file_extension(FileType::PARQUET.get_ext());
+// This is a workaround for this example since `test_data` contains
+// many different parquet different files,
+// in practice use FileType::PARQUET.get_ext().
+.with_file_extension("alltypes_plain.parquet");
 
-// Register a listing table - this will use all files in the directory as 
data sources
-// for the query
+// First example were we use an absolute path, which requires no 
additional setup.
 ctx.register_listing_table(
 "my_table",
-!("file://{testdata}/alltypes_plain.parquet"),
-listing_options,
+!("file://{test_data}/"),
+listing_options.clone(),
 None,
 None,
 )
@@ -60,5 +64,47 @@ async fn main() -> Result<()> {
 // print the results
 df.show().await?;
 
+// Second example were we temporarily move into the test data's parent 
directory and
+// simulate a relative path, this requires registering an ObjectStore.
+let cur_dir = std::env::current_dir()?;
+
+let test_data_path = Path::new(_data);
+let test_data_path_parent = test_data_path
+.parent()
+.ok_or("test_data path needs a parent")?;
+
+std::env::set_current_dir(test_data_path_parent)?;
+
+let local_fs = Arc::new(LocalFileSystem::default());
+
+let u = url::Url::parse("file://./")?;
+ctx.runtime_env().register_object_store(, local_fs);
+
+// Register a listing table - this will use all files in the directory as 
data sources
+// for the query
+ctx.register_listing_table(
+"relative_table",
+"./data",
+listing_options.clone(),
+None,
+None,
+)
+.await?;
+
+// execute the query
+let df = ctx
+.sql(
+"SELECT * \
+FROM relative_table \
+LIMIT 1",
+)
+.await?;
+
+// print the results
+df.show().await?;
+
+// Reset the current directory
+std::env::set_current_dir(cur_dir)?;
+
 Ok(())
 }



(arrow-datafusion) branch main updated: tests: add tests for writing hive-partitioned parquet (#9316)

2024-02-26 Thread alamb
This is an automated email from the ASF dual-hosted git repository.

alamb pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


The following commit(s) were added to refs/heads/main by this push:
 new a26f583d27 tests: add tests for writing hive-partitioned parquet 
(#9316)
a26f583d27 is described below

commit a26f583d2766da746ff30199cc7341227526737f
Author: Trent Hauck 
AuthorDate: Mon Feb 26 10:06:11 2024 -0800

tests: add tests for writing hive-partitioned parquet (#9316)

* tests: adds tests associated with #9237

* style: clippy
---
 .../src/datasource/physical_plan/parquet/mod.rs|  74 --
 datafusion/core/tests/dataframe/mod.rs | 160 -
 2 files changed, 158 insertions(+), 76 deletions(-)

diff --git a/datafusion/core/src/datasource/physical_plan/parquet/mod.rs 
b/datafusion/core/src/datasource/physical_plan/parquet/mod.rs
index badd870848..3aa1998bde 100644
--- a/datafusion/core/src/datasource/physical_plan/parquet/mod.rs
+++ b/datafusion/core/src/datasource/physical_plan/parquet/mod.rs
@@ -2066,80 +2066,6 @@ mod tests {
 Ok(())
 }
 
-#[tokio::test]
-async fn write_parquet_results() -> Result<()> {
-// create partitioned input file and context
-let tmp_dir = TempDir::new()?;
-// let mut ctx = create_ctx(_dir, 4).await?;
-let ctx = SessionContext::new_with_config(
-SessionConfig::new().with_target_partitions(8),
-);
-let schema = populate_csv_partitions(_dir, 4, ".csv")?;
-// register csv file with the execution context
-ctx.register_csv(
-"test",
-tmp_dir.path().to_str().unwrap(),
-CsvReadOptions::new().schema(),
-)
-.await?;
-
-// register a local file system object store for /tmp directory
-let local = Arc::new(LocalFileSystem::new_with_prefix(_dir)?);
-let local_url = Url::parse("file://local").unwrap();
-ctx.runtime_env().register_object_store(_url, local);
-
-// execute a simple query and write the results to parquet
-let out_dir = tmp_dir.as_ref().to_str().unwrap().to_string() + "/out/";
-let out_dir_url = "file://local/out/";
-let df = ctx.sql("SELECT c1, c2 FROM test").await?;
-df.write_parquet(out_dir_url, DataFrameWriteOptions::new(), None)
-.await?;
-// write_parquet( ctx, "SELECT c1, c2 FROM test", _dir, 
None).await?;
-
-// create a new context and verify that the results were saved to a 
partitioned parquet file
-let ctx = SessionContext::new();
-
-// get write_id
-let mut paths = fs::read_dir(_dir).unwrap();
-let path = paths.next();
-let name = path
-.unwrap()?
-.path()
-.file_name()
-.expect("Should be a file name")
-.to_str()
-.expect("Should be a str")
-.to_owned();
-let (parsed_id, _) = name.split_once('_').expect("File should contain 
_ !");
-let write_id = parsed_id.to_owned();
-
-// register each partition as well as the top level dir
-ctx.register_parquet(
-"part0",
-!("{out_dir}/{write_id}_0.parquet"),
-ParquetReadOptions::default(),
-)
-.await?;
-
-ctx.register_parquet("allparts", _dir, 
ParquetReadOptions::default())
-.await?;
-
-let part0 = ctx.sql("SELECT c1, c2 FROM 
part0").await?.collect().await?;
-let allparts = ctx
-.sql("SELECT c1, c2 FROM allparts")
-.await?
-.collect()
-.await?;
-
-let allparts_count: usize = allparts.iter().map(|batch| 
batch.num_rows()).sum();
-
-assert_eq!(part0[0].schema(), allparts[0].schema());
-
-assert_eq!(allparts_count, 40);
-
-Ok(())
-}
-
 fn logical2physical(expr: , schema: ) -> Arc 
{
 let df_schema = schema.clone().to_dfschema().unwrap();
 let execution_props = ExecutionProps::new();
diff --git a/datafusion/core/tests/dataframe/mod.rs 
b/datafusion/core/tests/dataframe/mod.rs
index b08b2b8fc7..ee84200417 100644
--- a/datafusion/core/tests/dataframe/mod.rs
+++ b/datafusion/core/tests/dataframe/mod.rs
@@ -30,15 +30,19 @@ use arrow::{
 };
 use arrow_array::Float32Array;
 use arrow_schema::ArrowError;
+use object_store::local::LocalFileSystem;
+use std::fs;
 use std::sync::Arc;
+use tempfile::TempDir;
+use url::Url;
 
-use datafusion::dataframe::DataFrame;
+use datafusion::dataframe::{DataFrame, DataFrameWriteOptions};
 use datafusion::datasource::MemTable;
 use datafusion::error::Result;
 use datafusion::execution::context::{SessionContext, SessionState};
 use datafusion::prelude::JoinType;
 use datafusion::prelude::{CsvReadOptions, ParquetReadOptions};
-use datafusion::test_util::parquet_test_data;
+use 

(arrow) branch dependabot/nuget/csharp/Grpc.Tools-2.62.0 deleted (was cb61d5ec57)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch dependabot/nuget/csharp/Grpc.Tools-2.62.0
in repository https://gitbox.apache.org/repos/asf/arrow.git


 was cb61d5ec57 MINOR: [C#] Bump Grpc.Tools from 2.60.0 to 2.62.0 in /csharp

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow) branch main updated (5b03b707db -> 43daa32d74)

2024-02-26 Thread curth
This is an automated email from the ASF dual-hosted git repository.

curth pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 5b03b707db MINOR: [C#] Bump coverlet.collector from 6.0.0 to 6.0.1 in 
/csharp (#40245)
 add 43daa32d74 MINOR: [C#] Bump Grpc.Tools from 2.60.0 to 2.62.0 in 
/csharp (#40246)

No new revisions were added by this update.

Summary of changes:
 csharp/src/Apache.Arrow.Flight.Sql/Apache.Arrow.Flight.Sql.csproj | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)



(arrow) branch main updated: GH-39582: [C++][Acero] Increase size of Acero TempStack (#40007)

2024-02-26 Thread apitrou
This is an automated email from the ASF dual-hosted git repository.

apitrou pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


The following commit(s) were added to refs/heads/main by this push:
 new 9a7662b41b GH-39582: [C++][Acero] Increase size of Acero TempStack 
(#40007)
9a7662b41b is described below

commit 9a7662b41b77a40a76b07435b3eff4fec7454596
Author: Sten Larsson 
AuthorDate: Mon Feb 26 18:02:44 2024 +0100

GH-39582: [C++][Acero] Increase size of Acero TempStack (#40007)

We have had problems for a long time with a specific batch job that 
combines data from different sources. There is something in the data causing an 
Acero execution plan to hang or crash at random. The problem has been 
reproduced since Arrow 11.0.0, originally in Ruby, but it has also in Python. 
There is unfortunately no test case that reliably reproduces the issue in a 
release build.

However, in a debug build we can see that the batch job causes an overflow 
on the temp stack in arrow/cpp/src/arrow/compute/util.cc:38. Increasing the 
size of the stack created in the Acero QueryContext works around the issue, but 
a real fix should be investigated separately.

**This PR contains a "Critical Fix".**
* Closes: #39582

Lead-authored-by: Sten Larsson 
Co-authored-by: Antoine Pitrou 
Signed-off-by: Antoine Pitrou 
---
 cpp/src/arrow/acero/query_context.cc |  2 +-
 cpp/src/arrow/compute/util.cc| 15 ---
 2 files changed, 9 insertions(+), 8 deletions(-)

diff --git a/cpp/src/arrow/acero/query_context.cc 
b/cpp/src/arrow/acero/query_context.cc
index 9f838508fc..a27397d120 100644
--- a/cpp/src/arrow/acero/query_context.cc
+++ b/cpp/src/arrow/acero/query_context.cc
@@ -53,7 +53,7 @@ size_t QueryContext::max_concurrency() const { return 
thread_indexer_.Capacity()
 Result QueryContext::GetTempStack(size_t thread_index) 
{
   if (!tld_[thread_index].is_init) {
 RETURN_NOT_OK(tld_[thread_index].stack.Init(
-memory_pool(), 8 * util::MiniBatch::kMiniBatchLength * 
sizeof(uint64_t)));
+memory_pool(), 32 * util::MiniBatch::kMiniBatchLength * 
sizeof(uint64_t)));
 tld_[thread_index].is_init = true;
   }
   return _[thread_index].stack;
diff --git a/cpp/src/arrow/compute/util.cc b/cpp/src/arrow/compute/util.cc
index c55143af0c..2058ba9f30 100644
--- a/cpp/src/arrow/compute/util.cc
+++ b/cpp/src/arrow/compute/util.cc
@@ -32,17 +32,18 @@ using internal::CpuInfo;
 namespace util {
 
 void TempVectorStack::alloc(uint32_t num_bytes, uint8_t** data, int* id) {
-  int64_t old_top = top_;
-  top_ += PaddedAllocationSize(num_bytes) + 2 * sizeof(uint64_t);
-  // Stack overflow check
-  ARROW_DCHECK(top_ <= buffer_size_);
-  *data = buffer_->mutable_data() + old_top + sizeof(uint64_t);
+  int64_t new_top = top_ + PaddedAllocationSize(num_bytes) + 2 * 
sizeof(uint64_t);
+  // Stack overflow check (see GH-39582).
+  // XXX cannot return a regular Status because most consumers do not either.
+  ARROW_CHECK_LE(new_top, buffer_size_) << "TempVectorStack::alloc overflow";
+  *data = buffer_->mutable_data() + top_ + sizeof(uint64_t);
   // We set 8 bytes before the beginning of the allocated range and
   // 8 bytes after the end to check for stack overflow (which would
   // result in those known bytes being corrupted).
-  reinterpret_cast(buffer_->mutable_data() + old_top)[0] = kGuard1;
-  reinterpret_cast(buffer_->mutable_data() + top_)[-1] = kGuard2;
+  reinterpret_cast(buffer_->mutable_data() + top_)[0] = kGuard1;
+  reinterpret_cast(buffer_->mutable_data() + new_top)[-1] = kGuard2;
   *id = num_vectors_++;
+  top_ = new_top;
 }
 
 void TempVectorStack::release(int id, uint32_t num_bytes) {



(arrow) branch dependabot/nuget/csharp/coverlet.collector-6.0.1 created (now 63e0d92a0e)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/nuget/csharp/coverlet.collector-6.0.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 63e0d92a0e MINOR: [C#] Bump coverlet.collector from 6.0.0 to 6.0.1 in 
/csharp

No new revisions were added by this update.



(arrow) branch main updated (8ec7044824 -> a7ac7e0e10)

2024-02-26 Thread apitrou
This is an automated email from the ASF dual-hosted git repository.

apitrou pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 8ec7044824 GH-40236: [Python][CI] Disable generating C lines in Cython 
tracebacks (#40225)
 add a7ac7e0e10 GH-40068:  [C++] Possible data race when reading metadata 
of a parquet file (#40111)

No new revisions were added by this update.

Summary of changes:
 cpp/src/arrow/dataset/file_parquet.cc | 5 +
 cpp/src/arrow/dataset/file_parquet.h  | 2 +-
 2 files changed, 6 insertions(+), 1 deletion(-)



(arrow) branch dependabot/nuget/csharp/Grpc.Tools-2.62.0 created (now cb61d5ec57)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch dependabot/nuget/csharp/Grpc.Tools-2.62.0
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at cb61d5ec57 MINOR: [C#] Bump Grpc.Tools from 2.60.0 to 2.62.0 in /csharp

No new revisions were added by this update.



(arrow-datafusion-comet) branch main updated: fix: Another attempt to fix libcrypto.dylib loading issue (#112)

2024-02-26 Thread sunchao
This is an automated email from the ASF dual-hosted git repository.

sunchao pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion-comet.git


The following commit(s) were added to refs/heads/main by this push:
 new 7be5a18  fix: Another attempt to fix libcrypto.dylib loading issue 
(#112)
7be5a18 is described below

commit 7be5a1806af6bf213c8cb7739c0263f1fb6ab8fa
Author: advancedxy 
AuthorDate: Tue Feb 27 01:04:52 2024 +0800

fix: Another attempt to fix libcrypto.dylib loading issue (#112)
---
 .github/actions/setup-macos-builder/action.yaml | 9 +
 .github/workflows/pr_build.yml  | 2 +-
 2 files changed, 6 insertions(+), 5 deletions(-)

diff --git a/.github/actions/setup-macos-builder/action.yaml 
b/.github/actions/setup-macos-builder/action.yaml
index 63010ea..cc1b631 100644
--- a/.github/actions/setup-macos-builder/action.yaml
+++ b/.github/actions/setup-macos-builder/action.yaml
@@ -49,12 +49,13 @@ runs:
 unzip $PROTO_ZIP
 echo "$HOME/d/protoc/bin" >> $GITHUB_PATH
 export PATH=$PATH:$HOME/d/protoc/bin
-# install openssl and setup DYLD_LIBRARY_PATH to work with 
libcrypto.dylib loading issues with x86_64 mac runners
-# see PR https://github.com/apache/arrow-datafusion-comet/pull/55 for 
more details
+# install openssl and setup DYLD_LIBRARY_PATH
 brew install openssl
-OPENSSL_LIB_PATH=$(dirname `brew list openssl | grep 
'lib/libcrypto.dylib'`)
+OPENSSL_LIB_PATH=`brew --prefix openssl`/lib
 echo "openssl lib path is: ${OPENSSL_LIB_PATH}"
-export DYLD_LIBRARY_PATH=$OPENSSL_LIB_PATH:$DYLD_LIBRARY_PATH
+echo "DYLD_LIBRARY_PATH=$OPENSSL_LIB_PATH:$DYLD_LIBRARY_PATH" >> 
$GITHUB_ENV
+# output the current status of SIP for later debugging
+csrutil status || true
 
 - name: Install JDK ${{inputs.jdk-version}}
   uses: actions/setup-java@v4
diff --git a/.github/workflows/pr_build.yml b/.github/workflows/pr_build.yml
index 669eddd..fe83032 100644
--- a/.github/workflows/pr_build.yml
+++ b/.github/workflows/pr_build.yml
@@ -70,7 +70,7 @@ jobs:
   macos-test:
 strategy:
   matrix:
-os: [macos-latest]
+os: [macos-13]
 java_version: [8, 11, 17]
 test-target: [rust, java]
 is_push_event:



(arrow) branch dependabot/maven/java/org.hamcrest-hamcrest-core-2.2 created (now 66da56e524)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/org.hamcrest-hamcrest-core-2.2
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 66da56e524 MINOR: [Java] Bump org.hamcrest:hamcrest-core from 1.3 to 
2.2 in /java

No new revisions were added by this update.



(arrow) branch main updated (9a7662b41b -> ca0910a533)

2024-02-26 Thread apitrou
This is an automated email from the ASF dual-hosted git repository.

apitrou pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 9a7662b41b GH-39582: [C++][Acero] Increase size of Acero TempStack 
(#40007)
 add ca0910a533 GH-39897: [C++][FS][S3] Ensure 
`AwsInstance::EnsureInitialized` to do initialization exactly once under 
concurrency (#40110)

No new revisions were added by this update.

Summary of changes:
 cpp/src/arrow/filesystem/s3fs.cc | 15 ++-
 python/pyarrow/tests/test_fs.py  | 23 +++
 2 files changed, 33 insertions(+), 5 deletions(-)



(arrow) branch main updated (b8fff043c6 -> 8ec7044824)

2024-02-26 Thread apitrou
This is an automated email from the ASF dual-hosted git repository.

apitrou pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from b8fff043c6 GH-40221: [C++][CMake] Use arrow/util/config.h.cmake 
instead of add_definitions() (#40222)
 add 8ec7044824 GH-40236: [Python][CI] Disable generating C lines in Cython 
tracebacks (#40225)

No new revisions were added by this update.

Summary of changes:
 python/CMakeLists.txt | 3 +++
 1 file changed, 3 insertions(+)



(arrow-datafusion) branch main updated (ec86acbc1f -> c56840734c)

2024-02-26 Thread comphead
This is an automated email from the ASF dual-hosted git repository.

comphead pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


from ec86acbc1f feat: expand `unnest`  to accept arbitrary single array 
expression (#9342)
 add c56840734c fix: flight examples (#9335)

No new revisions were added by this update.

Summary of changes:
 ci/scripts/rust_example.sh |  1 +
 datafusion-examples/Cargo.toml | 27 +-
 .../examples/external_dependency/catalog.rs|  4 ++--
 .../examples/flight/flight_sql_server.rs   |  4 ++--
 4 files changed, 31 insertions(+), 5 deletions(-)



(arrow-datafusion-comet) branch main updated: build: Show time duration for scala test (#116)

2024-02-26 Thread sunchao
This is an automated email from the ASF dual-hosted git repository.

sunchao pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion-comet.git


The following commit(s) were added to refs/heads/main by this push:
 new a2b1f9b  build: Show time duration for scala test (#116)
a2b1f9b is described below

commit a2b1f9bea381eb61355a16cd509e9383a0bf7abc
Author: advancedxy 
AuthorDate: Tue Feb 27 01:46:32 2024 +0800

build: Show time duration for scala test (#116)
---
 pom.xml | 1 +
 1 file changed, 1 insertion(+)

diff --git a/pom.xml b/pom.xml
index e46a49a..5a96eae 100644
--- a/pom.xml
+++ b/pom.xml
@@ -641,6 +641,7 @@ under the License.
 
${project.build.directory}/surefire-reports
 .
 SparkTestSuite.txt
+D
 
 
org.apache.comet.IntegrationTestSuite
 -ea -Xmx4g -Xss4m ${extraJavaTestArgs}



(arrow-datafusion-comet) branch main updated: feat: Add license header by spotless:apply automatically (#110)

2024-02-26 Thread sunchao
This is an automated email from the ASF dual-hosted git repository.

sunchao pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion-comet.git


The following commit(s) were added to refs/heads/main by this push:
 new f359ed7  feat: Add license header by spotless:apply automatically 
(#110)
f359ed7 is described below

commit f359ed79a3dc710fadbc0484944f280c55c181e1
Author: advancedxy 
AuthorDate: Tue Feb 27 01:47:02 2024 +0800

feat: Add license header by spotless:apply automatically (#110)
---
 dev/copyright/java-header.txt  | 19 
 dev/copyright/scala-header.txt |  1 +
 pom.xml|  6 +
 .../org/apache/spark/sql/comet/CometPlan.scala | 26 --
 .../src/test/scala/org/apache/spark/sql/TPCH.scala | 10 -
 .../test/scala/org/apache/spark/sql/Tables.scala   | 13 +--
 .../spark/sql/comet/CometPlanStabilitySuite.scala  |  1 +
 7 files changed, 52 insertions(+), 24 deletions(-)

diff --git a/dev/copyright/java-header.txt b/dev/copyright/java-header.txt
new file mode 100644
index 000..bd244d0
--- /dev/null
+++ b/dev/copyright/java-header.txt
@@ -0,0 +1,19 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+
diff --git a/dev/copyright/scala-header.txt b/dev/copyright/scala-header.txt
new file mode 12
index 000..372bb40
--- /dev/null
+++ b/dev/copyright/scala-header.txt
@@ -0,0 +1 @@
+java-header.txt
\ No newline at end of file
diff --git a/pom.xml b/pom.xml
index 5a96eae..d7cd076 100644
--- a/pom.xml
+++ b/pom.xml
@@ -723,6 +723,9 @@ under the License.
   
 
java|javax,scala,org,org.apache,com,org.apache.comet,\#,\#org.apache.comet
   
+  
+
${maven.multiModuleProjectDirectory}/dev/copyright/java-header.txt
+  
 
 
   
@@ -730,6 +733,9 @@ under the License.
 3.6.1
 ${maven.multiModuleProjectDirectory}/scalafmt.conf
   
+  
+
${maven.multiModuleProjectDirectory}/dev/copyright/scala-header.txt
+  
 
   
 
diff --git a/spark/src/main/scala/org/apache/spark/sql/comet/CometPlan.scala 
b/spark/src/main/scala/org/apache/spark/sql/comet/CometPlan.scala
index fe2ce7e..e5d268c 100644
--- a/spark/src/main/scala/org/apache/spark/sql/comet/CometPlan.scala
+++ b/spark/src/main/scala/org/apache/spark/sql/comet/CometPlan.scala
@@ -1,18 +1,20 @@
 /*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
  *
- *http://www.apache.org/licenses/LICENSE-2.0
+ *   http://www.apache.org/licenses/LICENSE-2.0
  *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
  */
 
 

(arrow) branch dependabot/maven/java/de.huxhorn.lilith-de.huxhorn.lilith.logback.appender.multiplex-classic-8.3.0 created (now 0bf1a999e7)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/de.huxhorn.lilith-de.huxhorn.lilith.logback.appender.multiplex-classic-8.3.0
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 0bf1a999e7 MINOR: [Java] Bump 
de.huxhorn.lilith:de.huxhorn.lilith.logback.appender.multiplex-classic

No new revisions were added by this update.



(arrow) branch dependabot/maven/java/commons-codec-commons-codec-1.16.1 created (now 8fe87e428c)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/commons-codec-commons-codec-1.16.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 8fe87e428c MINOR: [Java] Bump commons-codec:commons-codec in /java

No new revisions were added by this update.



(arrow) branch dependabot/maven/java/org.apache.maven.plugins-maven-compiler-plugin-3.12.1 created (now 8cb7dcf422)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/org.apache.maven.plugins-maven-compiler-plugin-3.12.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 8cb7dcf422 MINOR: [Java] Bump 
org.apache.maven.plugins:maven-compiler-plugin

No new revisions were added by this update.



(arrow) branch dependabot/maven/java/org.apache.maven.plugins-maven-site-plugin-3.12.1 created (now 7faa001981)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/org.apache.maven.plugins-maven-site-plugin-3.12.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 7faa001981 MINOR: [Java] Bump 
org.apache.maven.plugins:maven-site-plugin in /java

No new revisions were added by this update.



(arrow-adbc) branch main updated: chore(csharp): bump Google.Cloud.BigQuery.V2 from 3.5.0 to 3.6.0 in /csharp (#1567)

2024-02-26 Thread curth
This is an automated email from the ASF dual-hosted git repository.

curth pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


The following commit(s) were added to refs/heads/main by this push:
 new 1c9912af chore(csharp): bump Google.Cloud.BigQuery.V2 from 3.5.0 to 
3.6.0 in /csharp (#1567)
1c9912af is described below

commit 1c9912afea7458916125f597cb983fc9ef611b3c
Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
AuthorDate: Mon Feb 26 12:21:43 2024 -0800

chore(csharp): bump Google.Cloud.BigQuery.V2 from 3.5.0 to 3.6.0 in /csharp 
(#1567)

Bumps

[Google.Cloud.BigQuery.V2](https://github.com/googleapis/google-cloud-dotnet)
from 3.5.0 to 3.6.0.

Commits

https://github.com/googleapis/google-cloud-dotnet/commit/2f95acef888d1e6b1dbe775236bed8c42c577e83;>2f95ace
Release Google.Cloud.BigQuery.V2 version 3.6.0 (https://redirect.github.com/googleapis/google-cloud-dotnet/issues/11767;>#11767)
https://github.com/googleapis/google-cloud-dotnet/commit/efa6b92e6491ee6947c2e20b6cbd64d6def384f8;>efa6b92
Release Google.Cloud.BigQuery.Storage.V1 version 3.12.0 (https://redirect.github.com/googleapis/google-cloud-dotnet/issues/11768;>#11768)
https://github.com/googleapis/google-cloud-dotnet/commit/7c24c793439c0ec85a16102d9a2a1a44d69f87b7;>7c24c79
Release Google.Cloud.BigQuery.DataTransfer.V1 version 4.5.0 (https://redirect.github.com/googleapis/google-cloud-dotnet/issues/11766;>#11766)
https://github.com/googleapis/google-cloud-dotnet/commit/f85ba561e54aa036f1c64216a5bd3a44b91c5772;>f85ba56
Release Google.Apps.Meet.V2 version 1.0.0-beta01 (https://redirect.github.com/googleapis/google-cloud-dotnet/issues/11765;>#11765)
https://github.com/googleapis/google-cloud-dotnet/commit/9677fcb585130adead6bbb1963477dc77be13f57;>9677fcb
feat: Implement int64-based timestamps
https://github.com/googleapis/google-cloud-dotnet/commit/c71087b85be943b71af130ec440ccc73d6a2f95b;>c71087b
chore: Update BigQuery to use latest Discovery-generated version
https://github.com/googleapis/google-cloud-dotnet/commit/55db110ba7ed23644b93130960b980cd12c4aebf;>55db110
chore: Remove extraneous whitespace
https://github.com/googleapis/google-cloud-dotnet/commit/9f01c3100b3ca37f461b8f76d36ec57c9549b266;>9f01c31
chore: Remove ListRowsOptions.ToGetQueryResultsOptions
https://github.com/googleapis/google-cloud-dotnet/commit/cb3035dd532f97eb1a791ae8f90bd17681ff3ff9;>cb3035d
chore: Add example issue of uploading object
https://github.com/googleapis/google-cloud-dotnet/commit/289fa6e1f785e4da53676b952eefe2f7182ef4c2;>289fa6e
feat: add session token support for Autocomplete (New) sessions that end
with...
Additional commits viewable in https://github.com/googleapis/google-cloud-dotnet/compare/Google.Cloud.BigQuery.V2-3.5.0...Google.Cloud.BigQuery.V2-3.6.0;>compare
view





[![Dependabot compatibility

score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=Google.Cloud.BigQuery.V2=nuget=3.5.0=3.6.0)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores)

Dependabot will resolve any conflicts with this PR as long as you don't
alter it yourself. You can also trigger a rebase manually by commenting
`@dependabot rebase`.

[//]: # (dependabot-automerge-start)
[//]: # (dependabot-automerge-end)

---


Dependabot commands and options


You can trigger Dependabot actions by commenting on this PR:
- `@dependabot rebase` will rebase this PR
- `@dependabot recreate` will recreate this PR, overwriting any edits
that have been made to it
- `@dependabot merge` will merge this PR after your CI passes on it
- `@dependabot squash and merge` will squash and merge this PR after
your CI passes on it
- `@dependabot cancel merge` will cancel a previously requested merge
and block automerging
- `@dependabot reopen` will reopen this PR if it is closed
- `@dependabot close` will close this PR and stop Dependabot recreating
it. You can achieve the same result by closing it manually
- `@dependabot show  ignore conditions` will show all
of the ignore conditions of the specified dependency
- `@dependabot ignore this major version` will close this PR and stop
Dependabot creating any more for this major version (unless you reopen
the PR or upgrade to it yourself)
- `@dependabot ignore this minor version` will close this PR and stop
Dependabot creating any more for this minor version (unless you reopen
the PR or upgrade to it yourself)
- `@dependabot ignore this dependency` will close this PR and stop
Dependabot creating any more for this dependency (unless you reopen the
PR or upgrade to it yourself)


  

(arrow-adbc) branch dependabot/nuget/csharp/Google.Cloud.BigQuery.V2-3.6.0 deleted (was 765a305d)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/nuget/csharp/Google.Cloud.BigQuery.V2-3.6.0
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


 was 765a305d chore(csharp): bump Google.Cloud.BigQuery.V2 in /csharp

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow-datafusion) branch main updated: feature: support nvl(ifnull) function (#9284)

2024-02-26 Thread alamb
This is an automated email from the ASF dual-hosted git repository.

alamb pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


The following commit(s) were added to refs/heads/main by this push:
 new b55d0edb0e feature: support nvl(ifnull) function (#9284)
b55d0edb0e is described below

commit b55d0edb0e8b9e8ef8cfaaca164dd5ae93a46854
Author: junxiangMu <63799833+guoji...@users.noreply.github.com>
AuthorDate: Tue Feb 27 04:31:51 2024 +0800

feature: support nvl(ifnull) function (#9284)

* feature: support nvl(ifnull) function

* add sqllogictest

* add docs entry

* Update docs/source/user-guide/sql/scalar_functions.md

Co-authored-by: Jonah Gao 

* fix some code

* fix docs

-

Co-authored-by: Jonah Gao 
---
 datafusion/functions/src/core/mod.rs   |   5 +-
 datafusion/functions/src/core/nvl.rs   | 277 +
 datafusion/sqllogictest/test_files/nvl.slt | 120 +++
 docs/source/user-guide/sql/scalar_functions.md |  21 ++
 4 files changed, 422 insertions(+), 1 deletion(-)

diff --git a/datafusion/functions/src/core/mod.rs 
b/datafusion/functions/src/core/mod.rs
index 9aab4bd450..db47c62218 100644
--- a/datafusion/functions/src/core/mod.rs
+++ b/datafusion/functions/src/core/mod.rs
@@ -18,12 +18,15 @@
 //! "core" DataFusion functions
 
 mod nullif;
+mod nvl;
 
 // create UDFs
 make_udf_function!(nullif::NullIfFunc, NULLIF, nullif);
+make_udf_function!(nvl::NVLFunc, NVL, nvl);
 
 // Export the functions out of this package, both as expr_fn as well as a list 
of functions
 export_functions!(
-(nullif, arg_1 arg_2, "returns NULL if value1 equals value2; otherwise it 
returns value1. This can be used to perform the inverse operation of the 
COALESCE expression.")
+(nullif, arg_1 arg_2, "returns NULL if value1 equals value2; otherwise it 
returns value1. This can be used to perform the inverse operation of the 
COALESCE expression."),
+(nvl, arg_1 arg_2, "returns value2 if value1 is NULL; otherwise it returns 
value1")
 );
 
diff --git a/datafusion/functions/src/core/nvl.rs 
b/datafusion/functions/src/core/nvl.rs
new file mode 100644
index 00..6d6ad1cdeb
--- /dev/null
+++ b/datafusion/functions/src/core/nvl.rs
@@ -0,0 +1,277 @@
+// Licensed to the Apache Software Foundation (ASF) under one
+// or more contributor license agreements.  See the NOTICE file
+// distributed with this work for additional information
+// regarding copyright ownership.  The ASF licenses this file
+// to you under the Apache License, Version 2.0 (the
+// "License"); you may not use this file except in compliance
+// with the License.  You may obtain a copy of the License at
+//
+//   http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing,
+// software distributed under the License is distributed on an
+// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+// KIND, either express or implied.  See the License for the
+// specific language governing permissions and limitations
+// under the License.
+
+use arrow::datatypes::DataType;
+use datafusion_common::{internal_err, Result, DataFusionError};
+use datafusion_expr::{ColumnarValue, ScalarUDFImpl, Signature, Volatility};
+use arrow::compute::kernels::zip::zip;
+use arrow::compute::is_not_null;
+use arrow::array::Array;
+
+#[derive(Debug)]
+pub(super) struct NVLFunc {
+signature: Signature,
+aliases: Vec,
+}
+
+/// Currently supported types by the nvl/ifnull function.
+/// The order of these types correspond to the order on which coercion applies
+/// This should thus be from least informative to most informative
+static SUPPORTED_NVL_TYPES: &[DataType] = &[
+DataType::Boolean,
+DataType::UInt8,
+DataType::UInt16,
+DataType::UInt32,
+DataType::UInt64,
+DataType::Int8,
+DataType::Int16,
+DataType::Int32,
+DataType::Int64,
+DataType::Float32,
+DataType::Float64,
+DataType::Utf8,
+DataType::LargeUtf8,
+];
+
+impl NVLFunc {
+pub fn new() -> Self {
+Self {
+signature:
+Signature::uniform(2, SUPPORTED_NVL_TYPES.to_vec(),
+Volatility::Immutable,
+),
+aliases: vec![String::from("ifnull")],
+}
+}
+}
+
+impl ScalarUDFImpl for NVLFunc {
+fn as_any() ->  std::any::Any {
+self
+}
+
+fn name() ->  {
+"nvl"
+}
+
+fn signature() ->  {
+
+}
+
+fn return_type(, arg_types: &[DataType]) -> Result {
+// NVL has two args and they might get coerced, get a preview of this
+let coerced_types = 
datafusion_expr::type_coercion::functions::data_types(arg_types, 
);
+coerced_types.map(|typs| typs[0].clone())
+.map_err(|e| e.context("Failed to coerce arguments for NVL")
+)
+}
+
+fn invoke(, args: &[ColumnarValue]) 

(arrow-adbc) branch dependabot/nuget/csharp/Google.Cloud.BigQuery.Storage.V1-3.12.0 created (now 8a884703)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/nuget/csharp/Google.Cloud.BigQuery.Storage.V1-3.12.0
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


  at 8a884703 chore(csharp): bump Google.Cloud.BigQuery.Storage.V1 in 
/csharp

No new revisions were added by this update.



(arrow-adbc) branch dependabot/nuget/csharp/Google.Cloud.BigQuery.V2-3.6.0 created (now 765a305d)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/nuget/csharp/Google.Cloud.BigQuery.V2-3.6.0
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


  at 765a305d chore(csharp): bump Google.Cloud.BigQuery.V2 in /csharp

No new revisions were added by this update.



(arrow-adbc) branch asf-site updated: publish documentation

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new d998159c publish documentation
d998159c is described below

commit d998159cbd8f63fe4657d85e901d17340e7b215e
Author: github-actions[bot] 
AuthorDate: Mon Feb 26 23:32:59 2024 +

publish documentation
---
 main/java/api/allclasses-index.html | 2 +-
 main/java/api/allpackages-index.html| 2 +-
 main/java/api/constant-values.html  | 2 +-
 main/java/api/deprecated-list.html  | 2 +-
 main/java/api/help-doc.html | 2 +-
 main/java/api/index-all.html| 2 +-
 main/java/api/index.html| 2 +-
 .../org/apache/arrow/adbc/core/AdbcConnection.GetObjectsDepth.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcConnection.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcDatabase.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcDriver.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcException.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcInfoCode.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcOptions.html   | 2 +-
 .../org/apache/arrow/adbc/core/AdbcStatement.PartitionResult.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/AdbcStatement.QueryResult.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/AdbcStatement.UpdateResult.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcStatement.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcStatusCode.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/BulkIngestMode.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/ErrorDetail.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/IsolationLevel.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/PartitionDescriptor.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/StandardSchemas.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/StandardStatistics.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/TypedKey.html  | 2 +-
 .../arrow/adbc/core/class-use/AdbcConnection.GetObjectsDepth.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcConnection.html| 2 +-
 .../java/api/org/apache/arrow/adbc/core/class-use/AdbcDatabase.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/AdbcDriver.html  | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcException.html | 2 +-
 .../java/api/org/apache/arrow/adbc/core/class-use/AdbcInfoCode.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/AdbcOptions.html | 2 +-
 .../arrow/adbc/core/class-use/AdbcStatement.PartitionResult.html| 2 +-
 .../apache/arrow/adbc/core/class-use/AdbcStatement.QueryResult.html | 2 +-
 .../arrow/adbc/core/class-use/AdbcStatement.UpdateResult.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcStatement.html | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcStatusCode.html| 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/BulkIngestMode.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/ErrorDetail.html | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/IsolationLevel.html| 2 +-
 .../org/apache/arrow/adbc/core/class-use/PartitionDescriptor.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/StandardSchemas.html   | 2 +-
 .../org/apache/arrow/adbc/core/class-use/StandardStatistics.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/TypedKey.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/package-summary.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/package-tree.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/package-use.html   | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/BaseFlightReader.html| 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightInfoReader.html| 2 +-
 .../arrow/adbc/driver/flightsql/FlightSqlClientWithCallOptions.html | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightSqlConnection.html | 2 +-
 .../arrow/adbc/driver/flightsql/FlightSqlConnectionProperties.html  | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightSqlDatabase.html   | 2 +-
 .../api/org/apache/arrow/adbc/driver/flightsql/FlightSqlDriver.html | 2 +-
 .../apache/arrow/adbc/driver/flightsql/FlightSqlDriverFactory.html  | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightSqlStatement.html  | 2 +-
 .../arrow/adbc/driver/flightsql/class-use/BaseFlightReader.html | 2 +-
 .../arrow/adbc/driver/flightsql/class-use/FlightInfoReader.html | 2 +-
 

(arrow) branch dependabot/maven/java/org.hamcrest-hamcrest-core-2.2 updated (66da56e524 -> 411c2f28c6)

2024-02-26 Thread kou
This is an automated email from the ASF dual-hosted git repository.

kou pushed a change to branch 
dependabot/maven/java/org.hamcrest-hamcrest-core-2.2
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 66da56e524 MINOR: [Java] Bump org.hamcrest:hamcrest-core from 1.3 to 
2.2 in /java
 add 411c2f28c6 hamcrest-core -> hamcrest

No new revisions were added by this update.

Summary of changes:
 java/flight/flight-sql-jdbc-core/pom.xml   | 2 +-
 java/flight/flight-sql-jdbc-driver/pom.xml | 4 ++--
 2 files changed, 3 insertions(+), 3 deletions(-)



(arrow-adbc) branch main updated: fix(csharp/src/Client/SchemaConverter): add check for keys on precision and scale (#1566)

2024-02-26 Thread curth
This is an automated email from the ASF dual-hosted git repository.

curth pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


The following commit(s) were added to refs/heads/main by this push:
 new b29e4a96 fix(csharp/src/Client/SchemaConverter): add check for keys on 
precision and scale (#1566)
b29e4a96 is described below

commit b29e4a96c02ccd8e40c63fcbd32b39e12906ea2a
Author: davidhcoe <13318837+davidh...@users.noreply.github.com>
AuthorDate: Mon Feb 26 18:17:27 2024 -0500

fix(csharp/src/Client/SchemaConverter): add check for keys on precision and 
scale (#1566)

fixes https://github.com/apache/arrow-adbc/issues/1565

Co-authored-by: David Coe 
---
 csharp/src/Client/SchemaConverter.cs | 13 +++--
 1 file changed, 11 insertions(+), 2 deletions(-)

diff --git a/csharp/src/Client/SchemaConverter.cs 
b/csharp/src/Client/SchemaConverter.cs
index 38863417..cc71f499 100644
--- a/csharp/src/Client/SchemaConverter.cs
+++ b/csharp/src/Client/SchemaConverter.cs
@@ -70,8 +70,17 @@ namespace Apache.Arrow.Adbc.Client
 f.HasMetadata
 )
 {
-row[SchemaTableColumn.NumericPrecision] = 
Convert.ToInt32(f.Metadata["precision"]);
-row[SchemaTableColumn.NumericScale] = 
Convert.ToInt32(f.Metadata["scale"]);
+if (f.Metadata.TryGetValue("precision", out string 
precisionKey))
+{
+if(!string.IsNullOrEmpty(precisionKey))
+row[SchemaTableColumn.NumericPrecision] = 
Convert.ToInt32(f.Metadata[precisionKey]);
+}
+
+if(f.Metadata.TryGetValue("scale", out string scaleKey))
+{
+if(!string.IsNullOrEmpty(scaleKey))
+row[SchemaTableColumn.NumericScale] = 
Convert.ToInt32(f.Metadata[scaleKey]);
+}
 }
 else if (f.DataType is Decimal128Type decimal128Type)
 {



(arrow-adbc) branch main updated: chore(csharp): bump Google.Cloud.BigQuery.Storage.V1 from 3.11.0 to 3.12.0 in /csharp (#1568)

2024-02-26 Thread curth
This is an automated email from the ASF dual-hosted git repository.

curth pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


The following commit(s) were added to refs/heads/main by this push:
 new f26ca923 chore(csharp): bump Google.Cloud.BigQuery.Storage.V1 from 
3.11.0 to 3.12.0 in /csharp (#1568)
f26ca923 is described below

commit f26ca92376945b13dbf49363b8ddf7283d1b3dbf
Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
AuthorDate: Mon Feb 26 13:00:50 2024 -0800

chore(csharp): bump Google.Cloud.BigQuery.Storage.V1 from 3.11.0 to 3.12.0 
in /csharp (#1568)

Bumps

[Google.Cloud.BigQuery.Storage.V1](https://github.com/googleapis/google-cloud-dotnet)
from 3.11.0 to 3.12.0.

Commits

https://github.com/googleapis/google-cloud-dotnet/commit/efa6b92e6491ee6947c2e20b6cbd64d6def384f8;>efa6b92
Release Google.Cloud.BigQuery.Storage.V1 version 3.12.0 (https://redirect.github.com/googleapis/google-cloud-dotnet/issues/11768;>#11768)
https://github.com/googleapis/google-cloud-dotnet/commit/7c24c793439c0ec85a16102d9a2a1a44d69f87b7;>7c24c79
Release Google.Cloud.BigQuery.DataTransfer.V1 version 4.5.0 (https://redirect.github.com/googleapis/google-cloud-dotnet/issues/11766;>#11766)
https://github.com/googleapis/google-cloud-dotnet/commit/f85ba561e54aa036f1c64216a5bd3a44b91c5772;>f85ba56
Release Google.Apps.Meet.V2 version 1.0.0-beta01 (https://redirect.github.com/googleapis/google-cloud-dotnet/issues/11765;>#11765)
https://github.com/googleapis/google-cloud-dotnet/commit/9677fcb585130adead6bbb1963477dc77be13f57;>9677fcb
feat: Implement int64-based timestamps
https://github.com/googleapis/google-cloud-dotnet/commit/c71087b85be943b71af130ec440ccc73d6a2f95b;>c71087b
chore: Update BigQuery to use latest Discovery-generated version
https://github.com/googleapis/google-cloud-dotnet/commit/55db110ba7ed23644b93130960b980cd12c4aebf;>55db110
chore: Remove extraneous whitespace
https://github.com/googleapis/google-cloud-dotnet/commit/9f01c3100b3ca37f461b8f76d36ec57c9549b266;>9f01c31
chore: Remove ListRowsOptions.ToGetQueryResultsOptions
https://github.com/googleapis/google-cloud-dotnet/commit/cb3035dd532f97eb1a791ae8f90bd17681ff3ff9;>cb3035d
chore: Add example issue of uploading object
https://github.com/googleapis/google-cloud-dotnet/commit/289fa6e1f785e4da53676b952eefe2f7182ef4c2;>289fa6e
feat: add session token support for Autocomplete (New) sessions that end
with...
https://github.com/googleapis/google-cloud-dotnet/commit/354560a797a3038d45cab928a5d4e8b0bd0cac91;>354560a
docs: refine proto comment for run_as_non_root
Additional commits viewable in https://github.com/googleapis/google-cloud-dotnet/compare/Google.Cloud.BigQuery.Storage.V1-3.11.0...Google.Cloud.BigQuery.Storage.V1-3.12.0;>compare
view





[![Dependabot compatibility

score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=Google.Cloud.BigQuery.Storage.V1=nuget=3.11.0=3.12.0)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores)

Dependabot will resolve any conflicts with this PR as long as you don't
alter it yourself. You can also trigger a rebase manually by commenting
`@dependabot rebase`.

[//]: # (dependabot-automerge-start)
[//]: # (dependabot-automerge-end)

---


Dependabot commands and options


You can trigger Dependabot actions by commenting on this PR:
- `@dependabot rebase` will rebase this PR
- `@dependabot recreate` will recreate this PR, overwriting any edits
that have been made to it
- `@dependabot merge` will merge this PR after your CI passes on it
- `@dependabot squash and merge` will squash and merge this PR after
your CI passes on it
- `@dependabot cancel merge` will cancel a previously requested merge
and block automerging
- `@dependabot reopen` will reopen this PR if it is closed
- `@dependabot close` will close this PR and stop Dependabot recreating
it. You can achieve the same result by closing it manually
- `@dependabot show  ignore conditions` will show all
of the ignore conditions of the specified dependency
- `@dependabot ignore this major version` will close this PR and stop
Dependabot creating any more for this major version (unless you reopen
the PR or upgrade to it yourself)
- `@dependabot ignore this minor version` will close this PR and stop
Dependabot creating any more for this minor version (unless you reopen
the PR or upgrade to it yourself)
- `@dependabot ignore this dependency` will close this PR and stop
Dependabot creating any more for this dependency (unless you reopen the
PR or upgrade to it yourself)





(arrow-adbc) branch dependabot/nuget/csharp/Google.Cloud.BigQuery.Storage.V1-3.12.0 deleted (was cea7aba1)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/nuget/csharp/Google.Cloud.BigQuery.Storage.V1-3.12.0
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


 was cea7aba1 chore(csharp): bump Google.Cloud.BigQuery.Storage.V1 in 
/csharp

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow-adbc) branch asf-site updated: publish documentation

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new e5f84f31 publish documentation
e5f84f31 is described below

commit e5f84f31b2d8a91ee677ccd69cd0145c1b61598b
Author: github-actions[bot] 
AuthorDate: Mon Feb 26 21:15:24 2024 +

publish documentation
---
 main/java/api/allclasses-index.html | 2 +-
 main/java/api/allpackages-index.html| 2 +-
 main/java/api/constant-values.html  | 2 +-
 main/java/api/deprecated-list.html  | 2 +-
 main/java/api/help-doc.html | 2 +-
 main/java/api/index-all.html| 2 +-
 main/java/api/index.html| 2 +-
 .../org/apache/arrow/adbc/core/AdbcConnection.GetObjectsDepth.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcConnection.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcDatabase.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcDriver.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcException.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcInfoCode.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcOptions.html   | 2 +-
 .../org/apache/arrow/adbc/core/AdbcStatement.PartitionResult.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/AdbcStatement.QueryResult.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/AdbcStatement.UpdateResult.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcStatement.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/AdbcStatusCode.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/BulkIngestMode.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/ErrorDetail.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/IsolationLevel.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/PartitionDescriptor.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/StandardSchemas.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/StandardStatistics.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/TypedKey.html  | 2 +-
 .../arrow/adbc/core/class-use/AdbcConnection.GetObjectsDepth.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcConnection.html| 2 +-
 .../java/api/org/apache/arrow/adbc/core/class-use/AdbcDatabase.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/AdbcDriver.html  | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcException.html | 2 +-
 .../java/api/org/apache/arrow/adbc/core/class-use/AdbcInfoCode.html | 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/AdbcOptions.html | 2 +-
 .../arrow/adbc/core/class-use/AdbcStatement.PartitionResult.html| 2 +-
 .../apache/arrow/adbc/core/class-use/AdbcStatement.QueryResult.html | 2 +-
 .../arrow/adbc/core/class-use/AdbcStatement.UpdateResult.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcStatement.html | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/AdbcStatusCode.html| 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/BulkIngestMode.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/ErrorDetail.html | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/IsolationLevel.html| 2 +-
 .../org/apache/arrow/adbc/core/class-use/PartitionDescriptor.html   | 2 +-
 .../api/org/apache/arrow/adbc/core/class-use/StandardSchemas.html   | 2 +-
 .../org/apache/arrow/adbc/core/class-use/StandardStatistics.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/class-use/TypedKey.html| 2 +-
 main/java/api/org/apache/arrow/adbc/core/package-summary.html   | 2 +-
 main/java/api/org/apache/arrow/adbc/core/package-tree.html  | 2 +-
 main/java/api/org/apache/arrow/adbc/core/package-use.html   | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/BaseFlightReader.html| 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightInfoReader.html| 2 +-
 .../arrow/adbc/driver/flightsql/FlightSqlClientWithCallOptions.html | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightSqlConnection.html | 2 +-
 .../arrow/adbc/driver/flightsql/FlightSqlConnectionProperties.html  | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightSqlDatabase.html   | 2 +-
 .../api/org/apache/arrow/adbc/driver/flightsql/FlightSqlDriver.html | 2 +-
 .../apache/arrow/adbc/driver/flightsql/FlightSqlDriverFactory.html  | 2 +-
 .../org/apache/arrow/adbc/driver/flightsql/FlightSqlStatement.html  | 2 +-
 .../arrow/adbc/driver/flightsql/class-use/BaseFlightReader.html | 2 +-
 .../arrow/adbc/driver/flightsql/class-use/FlightInfoReader.html | 2 +-
 

(arrow-rs) branch master updated: Improve docs for logical and physical nulls even more (#5434)

2024-02-26 Thread tustvold
This is an automated email from the ASF dual-hosted git repository.

tustvold pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/arrow-rs.git


The following commit(s) were added to refs/heads/master by this push:
 new ab590801698 Improve docs for logical and physical nulls even more 
(#5434)
ab590801698 is described below

commit ab590801698f9b5f577a26e22dcd91e349e3
Author: Andrew Lamb 
AuthorDate: Mon Feb 26 15:07:51 2024 -0500

Improve docs for logical and physical nulls even more (#5434)

* Improve docs for logical and physical nulls even more

* Apply suggestions from code review
---
 arrow-array/src/array/mod.rs | 30 +-
 1 file changed, 21 insertions(+), 9 deletions(-)

diff --git a/arrow-array/src/array/mod.rs b/arrow-array/src/array/mod.rs
index 1a58598543f..7aa3f92bfbd 100644
--- a/arrow-array/src/array/mod.rs
+++ b/arrow-array/src/array/mod.rs
@@ -175,23 +175,35 @@ pub trait Array: std::fmt::Debug + Send + Sync {
 
 /// Returns the null buffer of this array if any.
 ///
-/// The null buffer encodes the "physical" nulls of an array.
-/// However, some arrays can also encode nullability in their children, 
for example,
-/// [`DictionaryArray::values`] values or [`RunArray::values`], or without 
a null buffer,
-/// such as [`NullArray`]. To determine if each element of such an array 
is logically null,
-/// you can use the slower [`Array::logical_nulls`] to obtain a computed 
mask .
+/// The null buffer contains the "physical" nulls of an array, that is how
+/// the nulls are represented in the underlying arrow format.
+///
+/// The physical representation is efficient, but is sometimes non 
intuitive
+/// for certain array types such as those with nullable child arrays like
+/// [`DictionaryArray::values`] or [`RunArray::values`], or without a
+/// null buffer, such as [`NullArray`].
+///
+/// To determine if each element of such an array is "logically" null,
+/// use the slower [`Array::logical_nulls`] to obtain a computed mask.
 fn nulls() -> Option<>;
 
-/// Returns a potentially computed [`NullBuffer`] that represent the 
logical null values of this array, if any.
+/// Returns a potentially computed [`NullBuffer`] that represents the 
logical
+/// null values of this array, if any.
+///
+/// Logical nulls represent the values that are null in the array,
+/// regardless of the underlying physical arrow representation.
 ///
-/// In most cases this will be the same as [`Array::nulls`], except for:
+/// For most array types, this is equivalent to the "physical" nulls
+/// returned by [`Array::nulls`]. It is different for the following cases, 
because which
+/// elements are null is not encoded in a single null buffer:
 ///
 /// * [`DictionaryArray`] where [`DictionaryArray::values`] contains nulls
 /// * [`RunArray`] where [`RunArray::values`] contains nulls
 /// * [`NullArray`] where all indices are nulls
 ///
-/// In these cases a logical [`NullBuffer`] will be computed, encoding the 
logical nullability
-/// of these arrays, beyond what is encoded in [`Array::nulls`]
+/// In these cases a logical [`NullBuffer`] will be computed, encoding the
+/// logical nullability of these arrays, beyond what is encoded in
+/// [`Array::nulls`]
 fn logical_nulls() -> Option {
 self.nulls().cloned()
 }



(arrow-adbc) branch dependabot/nuget/csharp/Google.Cloud.BigQuery.Storage.V1-3.12.0 updated (8a884703 -> cea7aba1)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/nuget/csharp/Google.Cloud.BigQuery.Storage.V1-3.12.0
in repository https://gitbox.apache.org/repos/asf/arrow-adbc.git


 discard 8a884703 chore(csharp): bump Google.Cloud.BigQuery.Storage.V1 in 
/csharp
 add 1c9912af chore(csharp): bump Google.Cloud.BigQuery.V2 from 3.5.0 to 
3.6.0 in /csharp (#1567)
 add cea7aba1 chore(csharp): bump Google.Cloud.BigQuery.Storage.V1 in 
/csharp

This update added new revisions after undoing existing revisions.
That is to say, some revisions that were in the old version of the
branch are not in the new version.  This situation occurs
when a user --force pushes a change and generates a repository
containing something like this:

 * -- * -- B -- O -- O -- O   (8a884703)
\
 N -- N -- N   
refs/heads/dependabot/nuget/csharp/Google.Cloud.BigQuery.Storage.V1-3.12.0 
(cea7aba1)

You should already have received notification emails for all of the O
revisions, and so the following emails describe only the N revisions
from the common base, B.

Any revisions marked "omit" are not gone; other references still
refer to them.  Any revisions marked "discard" are gone forever.

No new revisions were added by this update.

Summary of changes:
 csharp/src/Drivers/BigQuery/Apache.Arrow.Adbc.Drivers.BigQuery.csproj | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)



(arrow) branch main updated: MINOR: [Java] Bump de.huxhorn.lilith:de.huxhorn.lilith.logback.appender.multiplex-classic from 0.9.44 to 8.3.0 in /java (#40239)

2024-02-26 Thread kou
This is an automated email from the ASF dual-hosted git repository.

kou pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


The following commit(s) were added to refs/heads/main by this push:
 new 8940a427a5 MINOR: [Java] Bump 
de.huxhorn.lilith:de.huxhorn.lilith.logback.appender.multiplex-classic from 
0.9.44 to 8.3.0 in /java (#40239)
8940a427a5 is described below

commit 8940a427a5432469823783c52dc07d939b122acc
Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
AuthorDate: Tue Feb 27 05:46:51 2024 +0900

MINOR: [Java] Bump 
de.huxhorn.lilith:de.huxhorn.lilith.logback.appender.multiplex-classic from 
0.9.44 to 8.3.0 in /java (#40239)

Bumps 
[de.huxhorn.lilith:de.huxhorn.lilith.logback.appender.multiplex-classic](https://github.com/huxi/lilith)
 from 0.9.44 to 8.3.0.

Changelog
Sourced from https://github.com/huxi/lilith/blob/master/CHANGELOG.md;>de.huxhorn.lilith:de.huxhorn.lilith.logback.appender.multiplex-classic's
 changelog.

[8.3.0] - 2021-12-11
Changed

Demand Java 8 1.8.0_312.

Fixed

SLF4J 1.7.32, logback 1.2.7, log4j2 2.15.0, Groovy 2.5.15, Spring 
5.3.13, jackson 2.13.0, commons-io 2.11.0, commons-text 1.5, commons-lang3 
3.12.0, flying-saucer 9.1.18, aspectj 1.9.7, woodstox 6.2.7, httpcore 4.4.14, 
httpclient 4.5.13, protobuf 3.19.1

[8.2.0] - 2018-08-09
Added

Added Find previous active and Find next active 
buttons to toolbar.
Added Find previous and Find next buttons to 
toolbar.
Added lots of missing mnemonics.
If the connection is lost then Message/RequestURI in table will now 
show Connection closed. instead of nothing.
Added smooth horizontal table scrolling option that is enabled by 
default.
Added support for Log4j 2 JsonLayout, 
YamlLayout and XmlLayout. 
SerializedLayout has been deprecated in log4j2 2.9.0 so you should 
use one of the other options instead.
Added Automatic-Module-Names for artifacts where appropriate. See http://branchandbound.net/blog/java/2017/12/automatic-module-name/;>Automatic-Module-Name:
 Calling all Java Library Maintainers.

Changed

Clean all inactive logs is now less noisy in the Lilith 
log.
Changed icons for Find previous active and Find next 
active. They now differ from Find previous and Find 
next as they should.
Refactored actions and icon handling.
Don't add null events to global logs.
Unchecking Enable global logs. in Preferences is now 
deleting existing global log files automatically.
Keyboard help will now always be up-to-date.
Demand Java 8 1.8.0_181.

Deprecated

Nothing.

Removed

Previous and Next buttons in find panel.
Pause action. Pausing only paused updating of the table, 
not receiving of events. This was confusing (even me) and served no actual 
purpose. This action was a left-over from the early days of Lilith when it was 
used for debugging during development.

Fixed

All LF support mac screen menu bar with Java 9 or higher.
Zero-delimited event receivers did not add a null event 
when end of stream was reached.
Fixed initial enabled state of Go to source.
Fixed enabled state of Edit menu. Mustn't be disabled 
anymore because Paste StackTraceElement is always available.
Fixed enabled state of Copy selection.
Menu entries related to global logs are now disabled if Enable 
global logs. is unchecked in Preferences.
Added more dependencies and entries to the deserialization whitelist. 
This is essentially necessary because logback-access does not have 
an AccessEventVO. See also http://jira.qos.ch/browse/LOGBACK-1182;>LOGBACK-1182 - Problem 
deserializing AccessEvent..
Not all event producers expect a heartbeat.
Made sure that You have changed the look  feel. and 
You have changed the application path. dialogs aren't hidden by the 
preferences dialog.
Fixed java executable detection in Windows bat file. Thanks, https://github.com/tha2015;>tha2015!
Logback 1.2.3, log4j2 2.11.1, Groovy 2.5.1, jackson 2.9.6, spring 
5.0.8, protobuf 3.6.1, junique 1.0.4, jcommander 1.72, commons-lang 3.7, 
commons-text 1.4, commons-io 2.6, flying-saucer 9.1.14, glazedlists 1.10.0, 
aspectj 1.9.1, httpcore 4.4.10, httpclient 4.5.6, woodstox 5.1.0
Fixed several split package issues. Because of this, some classes have 
changed package names:

the two most commonly used classes 
de.huxhorn.lilith.logback.appender.ClassicMultiplexSocketAppender 
and de.huxhorn.lilith.logback.encoder.ClassicLilithEncoder have 
not been moved.
de.huxhorn.lilith.logback.encoder.AccessLilithEncoder 
changed to 
de.huxhorn.lilith.logback.encoder.access.AccessLilithEncoder.





... (truncated)


Commits

https://github.com/huxi/lilith/commit/79c8ec553a7a3fffe64516342435ca1ac0ede990;>79c8ec5
 console is not working 

(arrow) branch dependabot/maven/java/de.huxhorn.lilith-de.huxhorn.lilith.logback.appender.multiplex-classic-8.3.0 deleted (was 0bf1a999e7)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/de.huxhorn.lilith-de.huxhorn.lilith.logback.appender.multiplex-classic-8.3.0
in repository https://gitbox.apache.org/repos/asf/arrow.git


 was 0bf1a999e7 MINOR: [Java] Bump 
de.huxhorn.lilith:de.huxhorn.lilith.logback.appender.multiplex-classic

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow) branch dependabot/maven/java/org.apache.maven.plugins-maven-site-plugin-3.12.1 deleted (was 7faa001981)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/org.apache.maven.plugins-maven-site-plugin-3.12.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


 was 7faa001981 MINOR: [Java] Bump 
org.apache.maven.plugins:maven-site-plugin in /java

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow) branch main updated: MINOR: [Java] Bump org.apache.maven.plugins:maven-site-plugin from 3.7.1 to 3.12.1 in /java (#40240)

2024-02-26 Thread kou
This is an automated email from the ASF dual-hosted git repository.

kou pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


The following commit(s) were added to refs/heads/main by this push:
 new 229f09730d MINOR: [Java] Bump 
org.apache.maven.plugins:maven-site-plugin from 3.7.1 to 3.12.1 in /java 
(#40240)
229f09730d is described below

commit 229f09730d83099a22c74de99c0bcf1ecafc219e
Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
AuthorDate: Tue Feb 27 05:47:20 2024 +0900

MINOR: [Java] Bump org.apache.maven.plugins:maven-site-plugin from 3.7.1 to 
3.12.1 in /java (#40240)

Bumps 
[org.apache.maven.plugins:maven-site-plugin](https://github.com/apache/maven-site-plugin)
 from 3.7.1 to 3.12.1.

Commits

https://github.com/apache/maven-site-plugin/commit/ecae28fb0990eb5a7fc8f2d4ffe07f348d927f4b;>ecae28f
 [maven-release-plugin] prepare release maven-site-plugin-3.12.1
https://github.com/apache/maven-site-plugin/commit/d98569b083ded7a5182bf6cb5814ddcbd3150267;>d98569b
 [MSITE-908] Upgrade Maven Reporting API to 3.1.1
https://github.com/apache/maven-site-plugin/commit/bd3376f52d0053e05c78327aad39353710702a7a;>bd3376f
 [MSITE-901] If precending standalone report has been run, site:jar does not 
r...
https://github.com/apache/maven-site-plugin/commit/b99c0ef371774a414ade764f2921bdfe8918ed60;>b99c0ef
 [MSITE-902] Upgrade Plexus Utils to 3.4.2
https://github.com/apache/maven-site-plugin/commit/3c6ff2e285063231b7042bfe7875871c9d339830;>3c6ff2e
 Update CI URL
https://github.com/apache/maven-site-plugin/commit/f314e9da6ba0b5611fd5dd7dcff2d9ecc36dcd61;>f314e9d
 [MSITE-898] Upgrade Parent to 36
https://github.com/apache/maven-site-plugin/commit/bce7458375464e58f5d2d6cff92f8dde5f45de67;>bce7458
 [MSITE-897] Upgrade Plexus Archiver to 4.2.7
https://github.com/apache/maven-site-plugin/commit/3c8d426aae79c793a2a3acfddbd47a6826346382;>3c8d426
 keep only release month, drop day
https://github.com/apache/maven-site-plugin/commit/6604ab3b53d3f4045cc755340aeb0c4feaeaf8df;>6604ab3
 also keep only Doxia versions changes
https://github.com/apache/maven-site-plugin/commit/789a7a1054babde3c5b01e48bbf8abca49f5af8f;>789a7a1
 lighten content: keep only meaningful values
Additional commits viewable in https://github.com/apache/maven-site-plugin/compare/maven-site-plugin-3.7.1...maven-site-plugin-3.12.1;>compare
 view




[![Dependabot compatibility 
score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=org.apache.maven.plugins:maven-site-plugin=maven=3.7.1=3.12.1)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores)

Dependabot will resolve any conflicts with this PR as long as you don't 
alter it yourself. You can also trigger a rebase manually by commenting `@ 
dependabot rebase`.

[//]: # (dependabot-automerge-start)
[//]: # (dependabot-automerge-end)

---


Dependabot commands and options


You can trigger Dependabot actions by commenting on this PR:
- `@ dependabot rebase` will rebase this PR
- `@ dependabot recreate` will recreate this PR, overwriting any edits that 
have been made to it
- `@ dependabot merge` will merge this PR after your CI passes on it
- `@ dependabot squash and merge` will squash and merge this PR after your 
CI passes on it
- `@ dependabot cancel merge` will cancel a previously requested merge and 
block automerging
- `@ dependabot reopen` will reopen this PR if it is closed
- `@ dependabot close` will close this PR and stop Dependabot recreating 
it. You can achieve the same result by closing it manually
- `@ dependabot show  ignore conditions` will show all of 
the ignore conditions of the specified dependency
- `@ dependabot ignore this major version` will close this PR and stop 
Dependabot creating any more for this major version (unless you reopen the PR 
or upgrade to it yourself)
- `@ dependabot ignore this minor version` will close this PR and stop 
Dependabot creating any more for this minor version (unless you reopen the PR 
or upgrade to it yourself)
- `@ dependabot ignore this dependency` will close this PR and stop 
Dependabot creating any more for this dependency (unless you reopen the PR or 
upgrade to it yourself)



Authored-by: dependabot[bot] 
<49699333+dependabot[bot]@users.noreply.github.com>
Signed-off-by: Sutou Kouhei 
---
 java/bom/pom.xml   | 4 ++--
 java/maven/pom.xml | 4 ++--
 java/pom.xml   | 6 +++---
 3 files changed, 7 insertions(+), 7 deletions(-)

diff --git a/java/bom/pom.xml b/java/bom/pom.xml
index 025632c45a..2406886222 100644
--- a/java/bom/pom.xml
+++ b/java/bom/pom.xml
@@ -151,7 +151,7 @@
   
 org.apache.maven.plugins
 maven-site-plugin
-3.7.1
+

(arrow) branch dependabot/npm_and_yarn/js/es5-ext-0.10.63 created (now 6c6c17bcf4)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch dependabot/npm_and_yarn/js/es5-ext-0.10.63
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 6c6c17bcf4 MINOR: [JS] Bump es5-ext from 0.10.62 to 0.10.63 in /js

No new revisions were added by this update.



(arrow) branch main updated (229f09730d -> 9e3f0f1c1d)

2024-02-26 Thread kou
This is an automated email from the ASF dual-hosted git repository.

kou pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 229f09730d MINOR: [Java] Bump 
org.apache.maven.plugins:maven-site-plugin from 3.7.1 to 3.12.1 in /java 
(#40240)
 add 9e3f0f1c1d MINOR: [Java] Bump 
org.apache.maven.plugins:maven-compiler-plugin from 3.11.0 to 3.12.1 in /java 
(#40241)

No new revisions were added by this update.

Summary of changes:
 java/pom.xml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)



(arrow) branch dependabot/maven/java/org.apache.maven.plugins-maven-compiler-plugin-3.12.1 deleted (was 8cb7dcf422)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/org.apache.maven.plugins-maven-compiler-plugin-3.12.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


 was 8cb7dcf422 MINOR: [Java] Bump 
org.apache.maven.plugins:maven-compiler-plugin

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow) branch main updated (9e3f0f1c1d -> 8805de7bf2)

2024-02-26 Thread kou
This is an automated email from the ASF dual-hosted git repository.

kou pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 9e3f0f1c1d MINOR: [Java] Bump 
org.apache.maven.plugins:maven-compiler-plugin from 3.11.0 to 3.12.1 in /java 
(#40241)
 add 8805de7bf2 MINOR: [Java] Bump commons-codec:commons-codec from 1.16.0 
to 1.16.1 in /java (#40242)

No new revisions were added by this update.

Summary of changes:
 java/vector/pom.xml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)



(arrow) branch dependabot/maven/java/commons-codec-commons-codec-1.16.1 deleted (was 8fe87e428c)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/commons-codec-commons-codec-1.16.1
in repository https://gitbox.apache.org/repos/asf/arrow.git


 was 8fe87e428c MINOR: [Java] Bump commons-codec:commons-codec in /java

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow-datafusion) branch main updated: Move abs to datafusion_functions (#9313)

2024-02-26 Thread jayzhan
This is an automated email from the ASF dual-hosted git repository.

jayzhan pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


The following commit(s) were added to refs/heads/main by this push:
 new 85f7a8e88e Move abs to datafusion_functions (#9313)
85f7a8e88e is described below

commit 85f7a8e88e3596b3ec900c43fabb8f7f42bbea5c
Author: Junhao Liu 
AuthorDate: Mon Feb 26 18:24:16 2024 -0600

Move abs to datafusion_functions (#9313)

* feat: move abs to datafusion_functions

* fix proto

* fix proto

* fix CI vendored code

* Fix proto

* add support type

* fix signature

* fix typo

* fix test cases

* disable a test case

* remove old code from math_expressions

* feat: add test

* fix clippy

* use unknown for proto

* fix unknown proto
---
 datafusion/expr/src/built_in_function.rs |   7 -
 datafusion/expr/src/expr.rs  |   5 -
 datafusion/expr/src/expr_fn.rs   |   2 -
 datafusion/functions/src/math/abs.rs | 177 +++
 datafusion/functions/src/math/mod.rs |   8 +-
 datafusion/physical-expr/src/functions.rs|   4 -
 datafusion/physical-expr/src/math_expressions.rs |  93 +---
 datafusion/proto/proto/datafusion.proto  |   4 +-
 datafusion/proto/src/generated/pbjson.rs |   6 +-
 datafusion/proto/src/generated/prost.rs  |   8 +-
 datafusion/proto/src/logical_plan/from_proto.rs  |   6 +-
 datafusion/proto/src/logical_plan/to_proto.rs|   1 -
 12 files changed, 198 insertions(+), 123 deletions(-)

diff --git a/datafusion/expr/src/built_in_function.rs 
b/datafusion/expr/src/built_in_function.rs
index 8b4e65121c..cf1e73f780 100644
--- a/datafusion/expr/src/built_in_function.rs
+++ b/datafusion/expr/src/built_in_function.rs
@@ -42,8 +42,6 @@ use strum_macros::EnumIter;
 #[derive(Debug, Clone, PartialEq, Eq, Hash, EnumIter, Copy)]
 pub enum BuiltinScalarFunction {
 // math functions
-/// abs
-Abs,
 /// acos
 Acos,
 /// asin
@@ -364,7 +362,6 @@ impl BuiltinScalarFunction {
 pub fn volatility() -> Volatility {
 match self {
 // Immutable scalar builtins
-BuiltinScalarFunction::Abs => Volatility::Immutable,
 BuiltinScalarFunction::Acos => Volatility::Immutable,
 BuiltinScalarFunction::Asin => Volatility::Immutable,
 BuiltinScalarFunction::Atan => Volatility::Immutable,
@@ -868,8 +865,6 @@ impl BuiltinScalarFunction {
 
 BuiltinScalarFunction::ArrowTypeof => Ok(Utf8),
 
-BuiltinScalarFunction::Abs => Ok(input_expr_types[0].clone()),
-
 BuiltinScalarFunction::OverLay => {
 utf8_to_str_type(_expr_types[0], "overlay")
 }
@@ -1338,7 +1333,6 @@ impl BuiltinScalarFunction {
 Signature::uniform(2, vec![Int64], self.volatility())
 }
 BuiltinScalarFunction::ArrowTypeof => Signature::any(1, 
self.volatility()),
-BuiltinScalarFunction::Abs => Signature::any(1, self.volatility()),
 BuiltinScalarFunction::OverLay => Signature::one_of(
 vec![
 Exact(vec![Utf8, Utf8, Int64, Int64]),
@@ -1444,7 +1438,6 @@ impl BuiltinScalarFunction {
 /// Returns all names that can be used to call this function
 pub fn aliases() -> &'static [&'static str] {
 match self {
-BuiltinScalarFunction::Abs => &["abs"],
 BuiltinScalarFunction::Acos => &["acos"],
 BuiltinScalarFunction::Acosh => &["acosh"],
 BuiltinScalarFunction::Asin => &["asin"],
diff --git a/datafusion/expr/src/expr.rs b/datafusion/expr/src/expr.rs
index f40ccb6cdb..c3d9269d15 100644
--- a/datafusion/expr/src/expr.rs
+++ b/datafusion/expr/src/expr.rs
@@ -2033,11 +2033,6 @@ mod test {
 .is_volatile()
 .unwrap()
 );
-assert!(
-!ScalarFunctionDefinition::BuiltIn(BuiltinScalarFunction::Abs)
-.is_volatile()
-.unwrap()
-);
 
 // UDF
 #[derive(Debug)]
diff --git a/datafusion/expr/src/expr_fn.rs b/datafusion/expr/src/expr_fn.rs
index 4aa270e6dd..55bd40a189 100644
--- a/datafusion/expr/src/expr_fn.rs
+++ b/datafusion/expr/src/expr_fn.rs
@@ -557,7 +557,6 @@ nary_scalar_expr!(
 trunc,
 "truncate toward zero, with optional precision"
 );
-scalar_expr!(Abs, abs, num, "absolute value");
 scalar_expr!(Signum, signum, num, "sign of the argument (-1, 0, +1) ");
 scalar_expr!(Exp, exp, num, "exponential");
 scalar_expr!(Gcd, gcd, arg_1 arg_2, "greatest common divisor");
@@ -1354,7 +1353,6 @@ mod test {
 test_nary_scalar_expr!(Round, round, input, decimal_places);
 test_nary_scalar_expr!(Trunc, trunc, num);
 test_nary_scalar_expr!(Trunc, 

(arrow-datafusion-comet) branch main updated: fix: Fix compilation error for Spark 3.2 & 3.3 (#117)

2024-02-26 Thread sunchao
This is an automated email from the ASF dual-hosted git repository.

sunchao pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion-comet.git


The following commit(s) were added to refs/heads/main by this push:
 new 6ec8cb9  fix: Fix compilation error for Spark 3.2 & 3.3 (#117)
6ec8cb9 is described below

commit 6ec8cb912b938ef0ad7291f4f3de9ce7a883ae34
Author: Chao Sun 
AuthorDate: Mon Feb 26 21:55:14 2024 -0800

fix: Fix compilation error for Spark 3.2 & 3.3 (#117)
---
 spark/src/test/scala/org/apache/comet/exec/CometExecSuite.scala | 4 +++-
 1 file changed, 3 insertions(+), 1 deletion(-)

diff --git a/spark/src/test/scala/org/apache/comet/exec/CometExecSuite.scala 
b/spark/src/test/scala/org/apache/comet/exec/CometExecSuite.scala
index 0414671..29b6e12 100644
--- a/spark/src/test/scala/org/apache/comet/exec/CometExecSuite.scala
+++ b/spark/src/test/scala/org/apache/comet/exec/CometExecSuite.scala
@@ -43,6 +43,7 @@ import 
org.apache.spark.sql.internal.SQLConf.SESSION_LOCAL_TIMEZONE
 import org.apache.spark.unsafe.types.UTF8String
 
 import org.apache.comet.CometConf
+import org.apache.comet.CometSparkSessionExtensions.isSpark34Plus
 
 class CometExecSuite extends CometTestBase {
   import testImplicits._
@@ -1055,6 +1056,7 @@ class CometExecSuite extends CometTestBase {
   }
 
   test("Fallback to Spark for TakeOrderedAndProjectExec with offset") {
+assume(isSpark34Plus)
 Seq("true", "false").foreach(aqeEnabled =>
   withSQLConf(SQLConf.ADAPTIVE_EXECUTION_ENABLED.key -> aqeEnabled) {
 withTable("t1") {
@@ -1066,7 +1068,7 @@ class CometExecSuite extends CometTestBase {
 .write
 .saveAsTable("t1")
 
-  val df = sql("SELECT * FROM t1 ORDER BY a, b LIMIT 
3").offset(1).groupBy($"a").sum("b")
+  val df = sql("SELECT * FROM t1 ORDER BY a, b LIMIT 3 OFFSET 
1").groupBy($"a").sum("b")
   checkSparkAnswer(df)
 }
   })



(arrow-datafusion-comet) branch main updated: feat: Add dictionary binary to shuffle writer (#111)

2024-02-26 Thread viirya
This is an automated email from the ASF dual-hosted git repository.

viirya pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion-comet.git


The following commit(s) were added to refs/heads/main by this push:
 new 9cfe96f  feat: Add dictionary binary to shuffle writer (#111)
9cfe96f is described below

commit 9cfe96f27647d7822e09a14115a02d677ab72f41
Author: Liang-Chi Hsieh 
AuthorDate: Mon Feb 26 16:25:02 2024 -0800

feat: Add dictionary binary to shuffle writer (#111)

Native shuffle writer can write dictionary of string but dictionary of 
binary is not supported. We should add it.
---
 core/src/execution/datafusion/shuffle_writer.rs| 157 +
 core/src/execution/datafusion/spark_hash.rs|   6 +
 .../org/apache/comet/exec/CometShuffleSuite.scala  |  19 +++
 3 files changed, 153 insertions(+), 29 deletions(-)

diff --git a/core/src/execution/datafusion/shuffle_writer.rs 
b/core/src/execution/datafusion/shuffle_writer.rs
index fc15fac..f836e3a 100644
--- a/core/src/execution/datafusion/shuffle_writer.rs
+++ b/core/src/execution/datafusion/shuffle_writer.rs
@@ -460,32 +460,32 @@ fn append_columns(
 };
 }
 
-macro_rules! append_string_dict {
-($kt:ident) => {{
+macro_rules! append_byte_dict {
+($kt:ident, $byte_type:ty, $array_type:ty) => {{
 match $kt.as_ref() {
 DataType::Int8 => {
-append_dict!(Int8Type, StringDictionaryBuilder, 
StringArray)
+append_dict!(Int8Type, 
GenericByteDictionaryBuilder, $array_type)
 }
 DataType::Int16 => {
-append_dict!(Int16Type, 
StringDictionaryBuilder, StringArray)
+append_dict!(Int16Type,  
GenericByteDictionaryBuilder, $array_type)
 }
 DataType::Int32 => {
-append_dict!(Int32Type, 
StringDictionaryBuilder, StringArray)
+append_dict!(Int32Type,  
GenericByteDictionaryBuilder, $array_type)
 }
 DataType::Int64 => {
-append_dict!(Int64Type, 
StringDictionaryBuilder, StringArray)
+append_dict!(Int64Type,  
GenericByteDictionaryBuilder, $array_type)
 }
 DataType::UInt8 => {
-append_dict!(UInt8Type, 
StringDictionaryBuilder, StringArray)
+append_dict!(UInt8Type,  
GenericByteDictionaryBuilder, $array_type)
 }
 DataType::UInt16 => {
-append_dict!(UInt16Type, 
StringDictionaryBuilder, StringArray)
+append_dict!(UInt16Type, 
GenericByteDictionaryBuilder, $array_type)
 }
 DataType::UInt32 => {
-append_dict!(UInt32Type, 
StringDictionaryBuilder, StringArray)
+append_dict!(UInt32Type, 
GenericByteDictionaryBuilder, $array_type)
 }
 DataType::UInt64 => {
-append_dict!(UInt64Type, 
StringDictionaryBuilder, StringArray)
+append_dict!(UInt64Type, 
GenericByteDictionaryBuilder, $array_type)
 }
 _ => unreachable!("Unknown key type for dictionary"),
 }
@@ -522,7 +522,22 @@ fn append_columns(
 DataType::Dictionary(key_type, value_type)
 if matches!(value_type.as_ref(), DataType::Utf8) =>
 {
-append_string_dict!(key_type)
+append_byte_dict!(key_type, GenericStringType, StringArray)
+}
+DataType::Dictionary(key_type, value_type)
+if matches!(value_type.as_ref(), DataType::LargeUtf8) =>
+{
+append_byte_dict!(key_type, GenericStringType, 
LargeStringArray)
+}
+DataType::Dictionary(key_type, value_type)
+if matches!(value_type.as_ref(), DataType::Binary) =>
+{
+append_byte_dict!(key_type, GenericBinaryType, BinaryArray)
+}
+DataType::Dictionary(key_type, value_type)
+if matches!(value_type.as_ref(), DataType::LargeBinary) =>
+{
+append_byte_dict!(key_type, GenericBinaryType, 
LargeBinaryArray)
 }
 DataType::Binary => append!(Binary),
 DataType::LargeBinary => append!(LargeBinary),
@@ -1028,7 +1043,7 @@ macro_rules! primitive_dict_builder_helper {
 };
 }
 
-macro_rules! string_dict_builder_inner_helper {
+macro_rules! byte_dict_builder_inner_helper {
 ($kt:ty, $capacity:ident, $builder:ident) => {
 Box::new($builder::<$kt>::with_capacity(
 $capacity,
@@ -1068,28 +1083,28 @@ fn make_dict_builder(datatype: , capacity: 
usize) -> Box {
-string_dict_builder_inner_helper!(Int16Type, capacity, 
StringDictionaryBuilder)
+byte_dict_builder_inner_helper!(Int8Type, capacity, 

(arrow) branch main updated (8805de7bf2 -> 06935a3690)

2024-02-26 Thread kou
This is an automated email from the ASF dual-hosted git repository.

kou pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 8805de7bf2 MINOR: [Java] Bump commons-codec:commons-codec from 1.16.0 
to 1.16.1 in /java (#40242)
 add 06935a3690 MINOR: [Java] Bump org.hamcrest:hamcrest-core from 1.3 to 
2.2 in /java (#40238)

No new revisions were added by this update.

Summary of changes:
 java/flight/flight-sql-jdbc-core/pom.xml   | 4 ++--
 java/flight/flight-sql-jdbc-driver/pom.xml | 6 +++---
 2 files changed, 5 insertions(+), 5 deletions(-)



(arrow) branch dependabot/maven/java/org.hamcrest-hamcrest-core-2.2 deleted (was 411c2f28c6)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/maven/java/org.hamcrest-hamcrest-core-2.2
in repository https://gitbox.apache.org/repos/asf/arrow.git


 was 411c2f28c6 hamcrest-core -> hamcrest

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.



(arrow-rs) branch master updated: Add BufWriter for Adapative Put / Multipart Upload (#5431)

2024-02-26 Thread tustvold
This is an automated email from the ASF dual-hosted git repository.

tustvold pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/arrow-rs.git


The following commit(s) were added to refs/heads/master by this push:
 new ef5c45cf418 Add BufWriter for Adapative Put / Multipart Upload (#5431)
ef5c45cf418 is described below

commit ef5c45cf4186a8124da5a1603ebdbc09ef9928fc
Author: Raphael Taylor-Davies <1781103+tustv...@users.noreply.github.com>
AuthorDate: Tue Feb 27 15:39:36 2024 +1300

Add BufWriter for Adapative Put / Multipart Upload (#5431)

* Add BufWriter

* Review feedback
---
 object_store/src/buffered.rs | 163 ++-
 1 file changed, 161 insertions(+), 2 deletions(-)

diff --git a/object_store/src/buffered.rs b/object_store/src/buffered.rs
index 3a1354f4f20..fdefe599f79 100644
--- a/object_store/src/buffered.rs
+++ b/object_store/src/buffered.rs
@@ -18,7 +18,7 @@
 //! Utilities for performing tokio-style buffered IO
 
 use crate::path::Path;
-use crate::{ObjectMeta, ObjectStore};
+use crate::{MultipartId, ObjectMeta, ObjectStore};
 use bytes::Bytes;
 use futures::future::{BoxFuture, FutureExt};
 use futures::ready;
@@ -27,7 +27,7 @@ use std::io::{Error, ErrorKind, SeekFrom};
 use std::pin::Pin;
 use std::sync::Arc;
 use std::task::{Context, Poll};
-use tokio::io::{AsyncBufRead, AsyncRead, AsyncSeek, ReadBuf};
+use tokio::io::{AsyncBufRead, AsyncRead, AsyncSeek, AsyncWrite, AsyncWriteExt, 
ReadBuf};
 
 /// The default buffer size used by [`BufReader`]
 pub const DEFAULT_BUFFER_SIZE: usize = 1024 * 1024;
@@ -205,6 +205,138 @@ impl AsyncBufRead for BufReader {
 }
 }
 
+/// An async buffered writer compatible with the tokio IO traits
+///
+/// Up to `capacity` bytes will be buffered in memory, and flushed on shutdown
+/// using [`ObjectStore::put`]. If `capacity` is exceeded, data will instead be
+/// streamed using [`ObjectStore::put_multipart`]
+pub struct BufWriter {
+capacity: usize,
+state: BufWriterState,
+multipart_id: Option,
+store: Arc,
+}
+
+impl std::fmt::Debug for BufWriter {
+fn fmt(, f:  std::fmt::Formatter<'_>) -> std::fmt::Result {
+f.debug_struct("BufWriter")
+.field("capacity", )
+.field("multipart_id", _id)
+.finish()
+}
+}
+
+type MultipartResult = (MultipartId, Box);
+
+enum BufWriterState {
+/// Buffer up to capacity bytes
+Buffer(Path, Vec),
+/// [`ObjectStore::put_multipart`]
+Prepare(BoxFuture<'static, std::io::Result>),
+/// Write to a multipart upload
+Write(Box),
+/// [`ObjectStore::put`]
+Put(BoxFuture<'static, std::io::Result<()>>),
+}
+
+impl BufWriter {
+/// Create a new [`BufWriter`] from the provided [`ObjectStore`] and 
[`Path`]
+pub fn new(store: Arc, path: Path) -> Self {
+Self::with_capacity(store, path, 10 * 1024 * 1024)
+}
+
+/// Create a new [`BufWriter`] from the provided [`ObjectStore`], [`Path`] 
and `capacity`
+pub fn with_capacity(store: Arc, path: Path, capacity: 
usize) -> Self {
+Self {
+capacity,
+store,
+state: BufWriterState::Buffer(path, Vec::new()),
+multipart_id: None,
+}
+}
+
+/// Returns the [`MultipartId`] if multipart upload
+pub fn multipart_id() -> Option<> {
+self.multipart_id.as_ref()
+}
+}
+
+impl AsyncWrite for BufWriter {
+fn poll_write(
+mut self: Pin< Self>,
+cx:  Context<'_>,
+buf: &[u8],
+) -> Poll> {
+let cap = self.capacity;
+loop {
+return match  self.state {
+BufWriterState::Write(write) => Pin::new(write).poll_write(cx, 
buf),
+BufWriterState::Put(_) => panic!("Already shut down"),
+BufWriterState::Prepare(f) => {
+let (id, w) = ready!(f.poll_unpin(cx)?);
+self.state = BufWriterState::Write(w);
+self.multipart_id = Some(id);
+continue;
+}
+BufWriterState::Buffer(path, b) => {
+if b.len().saturating_add(buf.len()) >= cap {
+let buffer = std::mem::take(b);
+let path = std::mem::take(path);
+let store = Arc::clone();
+self.state = BufWriterState::Prepare(Box::pin(async 
move {
+let (id, mut writer) = 
store.put_multipart().await?;
+writer.write_all().await?;
+Ok((id, writer))
+}));
+continue;
+}
+b.extend_from_slice(buf);
+Poll::Ready(Ok(buf.len()))
+}
+};
+}
+}
+
+fn poll_flush(mut self: Pin< Self>, cx:  Context<'_>) -> 
Poll> {
+loop {
+return 

(arrow) branch main updated (06935a3690 -> 46407df093)

2024-02-26 Thread ianmcook
This is an automated email from the ASF dual-hosted git repository.

ianmcook pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


from 06935a3690 MINOR: [Java] Bump org.hamcrest:hamcrest-core from 1.3 to 
2.2 in /java (#40238)
 add 46407df093 GH-40215: [Format][Docs] Document Arrow Columnar Format 
version history (#40219)

No new revisions were added by this update.

Summary of changes:
 docs/source/format/CanonicalExtensions.rst |  2 +-
 docs/source/format/Columnar.rst| 17 ++-
 docs/source/format/Versioning.rst  | 46 ++
 docs/source/status.rst |  6 ++--
 4 files changed, 61 insertions(+), 10 deletions(-)



(arrow) branch main updated: GH-40228: [C++][CMake] Improve description why we need to initialize AWS C++ SDK in arrow-s3fs-test (#40229)

2024-02-26 Thread kou
This is an automated email from the ASF dual-hosted git repository.

kou pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


The following commit(s) were added to refs/heads/main by this push:
 new 2235a7ed40 GH-40228: [C++][CMake] Improve description why we need to 
initialize AWS C++ SDK in arrow-s3fs-test (#40229)
2235a7ed40 is described below

commit 2235a7ed40b999d919d7d17cbb34097e819a5acf
Author: Sutou Kouhei 
AuthorDate: Tue Feb 27 13:54:26 2024 +0900

GH-40228: [C++][CMake] Improve description why we need to initialize AWS 
C++ SDK in arrow-s3fs-test (#40229)

### Rationale for this change

Only static linking isn't important. Static linking + private symbols are 
important.

### What changes are included in this PR?

Improve comment and macro name.

### Are these changes tested?

Yes.

### Are there any user-facing changes?

No.
* GitHub Issue: #40228

Authored-by: Sutou Kouhei 
Signed-off-by: Sutou Kouhei 
---
 cpp/src/arrow/filesystem/CMakeLists.txt | 18 +++---
 cpp/src/arrow/filesystem/s3fs_test.cc   |  6 +++---
 2 files changed, 10 insertions(+), 14 deletions(-)

diff --git a/cpp/src/arrow/filesystem/CMakeLists.txt 
b/cpp/src/arrow/filesystem/CMakeLists.txt
index a42a8d0f8c..77e93223cd 100644
--- a/cpp/src/arrow/filesystem/CMakeLists.txt
+++ b/cpp/src/arrow/filesystem/CMakeLists.txt
@@ -71,17 +71,13 @@ if(ARROW_S3)
 get_target_property(AWS_CPP_SDK_S3_TYPE aws-cpp-sdk-s3 TYPE)
 # We need to initialize AWS C++ SDK for direct use (not via
 # arrow::fs::S3FileSystem) in arrow-s3fs-test if we use static AWS
-# C++ SDK. Because AWS C++ SDK has internal static variables that
-# aren't shared in libarrow and arrow-s3fs-test. It means that
-# arrow::fs::InitializeS3() doesn't initialize AWS C++ SDK that is
-# directly used in arrow-s3fs-test.
-#
-# But it seems that internal static variables in AWS C++ SDK are
-# shared on macOS even if we link static AWS C++ SDK to both
-# libarrow and arrow-s3fs-test. So we don't need to initialize AWS
-# C++ SDK in arrow-s3fs-test on macOS.
-if(AWS_CPP_SDK_S3_TYPE STREQUAL "STATIC_LIBRARY" AND NOT APPLE)
-  list(APPEND ARROW_S3FS_TEST_COMPILE_DEFINITIONS 
"AWS_CPP_SDK_S3_NOT_SHARED")
+# C++ SDK and hide symbols of them. Because AWS C++ SDK has
+# internal static variables that aren't shared in libarrow and
+# arrow-s3fs-test. It means that arrow::fs::InitializeS3() doesn't
+# initialize AWS C++ SDK that is directly used in arrow-s3fs-test.
+if(AWS_CPP_SDK_S3_TYPE STREQUAL "STATIC_LIBRARY"
+   AND CXX_LINKER_SUPPORTS_VERSION_SCRIPT)
+  list(APPEND ARROW_S3FS_TEST_COMPILE_DEFINITIONS 
"AWS_CPP_SDK_S3_PRIVATE_STATIC")
 endif()
 target_compile_definitions(arrow-s3fs-test
PRIVATE ${ARROW_S3FS_TEST_COMPILE_DEFINITIONS})
diff --git a/cpp/src/arrow/filesystem/s3fs_test.cc 
b/cpp/src/arrow/filesystem/s3fs_test.cc
index 33e9712a66..394f59e91a 100644
--- a/cpp/src/arrow/filesystem/s3fs_test.cc
+++ b/cpp/src/arrow/filesystem/s3fs_test.cc
@@ -150,7 +150,7 @@ class ShortRetryStrategy : public S3RetryStrategy {
 class AwsTestMixin : public ::testing::Test {
  public:
   void SetUp() override {
-#ifdef AWS_CPP_SDK_S3_NOT_SHARED
+#ifdef AWS_CPP_SDK_S3_PRIVATE_STATIC
 auto aws_log_level = Aws::Utils::Logging::LogLevel::Fatal;
 aws_options_.loggingOptions.logLevel = aws_log_level;
 aws_options_.loggingOptions.logger_create_fn = [_log_level] {
@@ -161,13 +161,13 @@ class AwsTestMixin : public ::testing::Test {
   }
 
   void TearDown() override {
-#ifdef AWS_CPP_SDK_S3_NOT_SHARED
+#ifdef AWS_CPP_SDK_S3_PRIVATE_STATIC
 Aws::ShutdownAPI(aws_options_);
 #endif
   }
 
  private:
-#ifdef AWS_CPP_SDK_S3_NOT_SHARED
+#ifdef AWS_CPP_SDK_S3_PRIVATE_STATIC
   Aws::SDKOptions aws_options_;
 #endif
 };



(arrow-rs) branch master updated: Improve debug output of Time32/Time64 arrays (#5428)

2024-02-26 Thread tustvold
This is an automated email from the ASF dual-hosted git repository.

tustvold pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/arrow-rs.git


The following commit(s) were added to refs/heads/master by this push:
 new 37cf8a6c19e Improve debug output of Time32/Time64 arrays (#5428)
37cf8a6c19e is described below

commit 37cf8a6c19e02a4dc822fadf331c2fae9e4b5656
Author: Clide S <109172241+monkwi...@users.noreply.github.com>
AuthorDate: Mon Feb 26 19:14:57 2024 -0500

Improve debug output of Time32/Time64 arrays (#5428)

* Improve debug output of Time32/Time64 arrays

* Update arrow-array/src/array/primitive_array.rs

Co-authored-by: Raphael Taylor-Davies 
<1781103+tustv...@users.noreply.github.com>

* Update arrow-array/src/array/primitive_array.rs

Co-authored-by: Raphael Taylor-Davies 
<1781103+tustv...@users.noreply.github.com>

* Fix

-

Co-authored-by: Raphael Taylor-Davies 
<1781103+tustv...@users.noreply.github.com>
Co-authored-by: Raphael Taylor-Davies 
---
 arrow-array/src/array/primitive_array.rs | 85 ++--
 1 file changed, 82 insertions(+), 3 deletions(-)

diff --git a/arrow-array/src/array/primitive_array.rs 
b/arrow-array/src/array/primitive_array.rs
index ca437a569a1..a800aa6bf92 100644
--- a/arrow-array/src/array/primitive_array.rs
+++ b/arrow-array/src/array/primitive_array.rs
@@ -1093,20 +1093,31 @@ where
 impl std::fmt::Debug for PrimitiveArray {
 fn fmt(, f:  std::fmt::Formatter) -> std::fmt::Result {
 let data_type = self.data_type();
+
 write!(f, "PrimitiveArray<{data_type:?}>\n[\n")?;
 print_long_array(self, f, |array, index, f| match data_type {
 DataType::Date32 | DataType::Date64 => {
 let v = self.value(index).to_isize().unwrap() as i64;
 match as_date::(v) {
 Some(date) => write!(f, "{date:?}"),
-None => write!(f, "null"),
+None => {
+write!(
+f,
+"Cast error: Failed to convert {v} to temporal for 
{data_type:?}"
+)
+}
 }
 }
 DataType::Time32(_) | DataType::Time64(_) => {
 let v = self.value(index).to_isize().unwrap() as i64;
 match as_time::(v) {
 Some(time) => write!(f, "{time:?}"),
-None => write!(f, "null"),
+None => {
+write!(
+f,
+"Cast error: Failed to convert {v} to temporal for 
{data_type:?}"
+)
+}
 }
 }
 DataType::Timestamp(_, tz_string_opt) => {
@@ -1948,7 +1959,8 @@ mod tests {
 // chrono::NaiveDatetime::from_timestamp_opt returns None while input 
is invalid
 let arr: PrimitiveArray = vec![-7201, -60054].into();
 assert_eq!(
-"PrimitiveArray\n[\n  null,\n  null,\n]",
+"PrimitiveArray\n[\n  Cast error: Failed to convert 
-7201 to temporal for Time32(Second),\n  Cast error: Failed to convert -60054 
to temporal for Time32(Second),\n]",
+// "PrimitiveArray\n[\n  null,\n  null,\n]",
 format!("{arr:?}")
 )
 }
@@ -2482,4 +2494,71 @@ mod tests {
 fn test_with_data_type() {
 Int32Array::new(vec![1, 2, 3, 4].into(), 
None).with_data_type(DataType::Date32);
 }
+
+#[test]
+fn test_time_32second_output() {
+let array: Time32SecondArray = vec![
+Some(-1),
+Some(0),
+Some(86_399),
+Some(86_400),
+Some(86_401),
+None,
+]
+.into();
+let debug_str = format!("{:?}", array);
+assert_eq!("PrimitiveArray\n[\n  Cast error: Failed to 
convert -1 to temporal for Time32(Second),\n  00:00:00,\n  23:59:59,\n  Cast 
error: Failed to convert 86400 to temporal for Time32(Second),\n  Cast error: 
Failed to convert 86401 to temporal for Time32(Second),\n  null,\n]",
+debug_str
+);
+}
+
+#[test]
+fn test_time_32millisecond_debug_output() {
+let array: Time32MillisecondArray = vec![
+Some(-1),
+Some(0),
+Some(86_399_000),
+Some(86_400_000),
+Some(86_401_000),
+None,
+]
+.into();
+let debug_str = format!("{:?}", array);
+assert_eq!("PrimitiveArray\n[\n  Cast error: 
Failed to convert -1 to temporal for Time32(Millisecond),\n  00:00:00,\n  
23:59:59,\n  Cast error: Failed to convert 8640 to temporal for 
Time32(Millisecond),\n  Cast error: Failed to convert 86401000 to temporal for 
Time32(Millisecond),\n  null,\n]",
+debug_str
+);
+}
+
+#[test]
+ 

(arrow) 01/01: GH-40181: [C++] Support glog 0.7 build

2024-02-26 Thread uwe
This is an automated email from the ASF dual-hosted git repository.

uwe pushed a commit to branch ARROW-40181-glog07
in repository https://gitbox.apache.org/repos/asf/arrow.git

commit 147504673e409fc5267f0422a58f170f442b5455
Author: Uwe L. Korn 
AuthorDate: Mon Feb 26 09:21:31 2024 +0100

GH-40181: [C++] Support glog 0.7 build
---
 cpp/cmake_modules/FindGLOG.cmake | 8 +++-
 cpp/src/arrow/util/logging.cc| 6 +++---
 2 files changed, 10 insertions(+), 4 deletions(-)

diff --git a/cpp/cmake_modules/FindGLOG.cmake b/cpp/cmake_modules/FindGLOG.cmake
index 61b7d0694e..62b235ee91 100644
--- a/cpp/cmake_modules/FindGLOG.cmake
+++ b/cpp/cmake_modules/FindGLOG.cmake
@@ -17,6 +17,11 @@
 #
 #  find_package(GLOG)
 
+find_package(glog CONFIG)
+if(glog_FOUND)
+  return()
+endif()
+
 if(GLOG_FOUND)
   return()
 endif()
@@ -56,5 +61,6 @@ if(GLOG_FOUND)
   add_library(glog::glog UNKNOWN IMPORTED)
   set_target_properties(glog::glog
 PROPERTIES IMPORTED_LOCATION "${GLOG_LIB}"
-   INTERFACE_INCLUDE_DIRECTORIES 
"${GLOG_INCLUDE_DIR}")
+   INTERFACE_INCLUDE_DIRECTORIES 
"${GLOG_INCLUDE_DIR}"
+   INTERFACE_COMPILE_DEFINITIONS 
"GLOG_USE_GLOG_EXPORT")
 endif()
diff --git a/cpp/src/arrow/util/logging.cc b/cpp/src/arrow/util/logging.cc
index d293113237..25c336a6d2 100644
--- a/cpp/src/arrow/util/logging.cc
+++ b/cpp/src/arrow/util/logging.cc
@@ -116,7 +116,7 @@ static std::unique_ptr log_dir_;
 #ifdef ARROW_USE_GLOG
 
 // Glog's severity map.
-static int GetMappedSeverity(ArrowLogLevel severity) {
+static google::LogSeverity GetMappedSeverity(ArrowLogLevel severity) {
   switch (severity) {
 case ArrowLogLevel::ARROW_DEBUG:
   return google::GLOG_INFO;
@@ -148,7 +148,7 @@ void ArrowLog::StartArrowLog(const std::string& app_name,
   app_name_.reset(new std::string(app_name));
   log_dir_.reset(new std::string(log_dir));
 #ifdef ARROW_USE_GLOG
-  int mapped_severity_threshold = GetMappedSeverity(severity_threshold_);
+  google::LogSeverity mapped_severity_threshold = 
GetMappedSeverity(severity_threshold_);
   google::SetStderrLogging(mapped_severity_threshold);
   // Enable log file if log_dir is not empty.
   if (!log_dir.empty()) {
@@ -173,7 +173,7 @@ void ArrowLog::StartArrowLog(const std::string& app_name,
 google::SetLogFilenameExtension(app_name_without_path.c_str());
 for (int i = static_cast(severity_threshold_);
  i <= static_cast(ArrowLogLevel::ARROW_FATAL); ++i) {
-  int level = GetMappedSeverity(static_cast(i));
+  google::LogSeverity level = 
GetMappedSeverity(static_cast(i));
   google::SetLogDestination(level, dir_ends_with_slash.c_str());
 }
   }



(arrow) branch ARROW-40181-glog07 created (now 147504673e)

2024-02-26 Thread uwe
This is an automated email from the ASF dual-hosted git repository.

uwe pushed a change to branch ARROW-40181-glog07
in repository https://gitbox.apache.org/repos/asf/arrow.git


  at 147504673e GH-40181: [C++] Support glog 0.7 build

This branch includes the following new commits:

 new 147504673e GH-40181: [C++] Support glog 0.7 build

The 1 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.




(arrow-datafusion) branch dependabot/cargo/main/nix-0.28.0 created (now eaa5557457)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch dependabot/cargo/main/nix-0.28.0
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


  at eaa5557457 Update nix requirement from 0.27.1 to 0.28.0

No new revisions were added by this update.



(arrow-datafusion) branch main updated: feat: support `FixedSizeList` Type Coercion (#9108)

2024-02-26 Thread jayzhan
This is an automated email from the ASF dual-hosted git repository.

jayzhan pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


The following commit(s) were added to refs/heads/main by this push:
 new b728232b91 feat: support `FixedSizeList` Type Coercion (#9108)
b728232b91 is described below

commit b728232b91862826061dfa878cefe90d25576f78
Author: Alex Huang 
AuthorDate: Mon Feb 26 21:10:13 2024 +0800

feat: support `FixedSizeList` Type Coercion (#9108)

* support FixedSizeList Type Coercion

* add allow null type coercion parameter

* support null column in FixedSizeList

* Add test

* Add tests for cardinality with fixed size lists

* chore

* fix ci

* add comment

* Fix array_element function signature

* Remove unused imports and simplify code

* Fix array function signatures and behavior

* fix conflict

* fix conflict

* add tests for FixedSizeList

* remove unreacheable null check

* simplify the code

* remove null checking

* reformat output

* simplify code

* add tests for array_dims

* Refactor type coercion functions in datafusion/expr module
---
 datafusion/expr/src/built_in_function.rs   |  23 +-
 datafusion/expr/src/signature.rs   |  17 +-
 datafusion/expr/src/type_coercion/functions.rs | 109 ++---
 datafusion/sqllogictest/test_files/array.slt   | 562 -
 4 files changed, 628 insertions(+), 83 deletions(-)

diff --git a/datafusion/expr/src/built_in_function.rs 
b/datafusion/expr/src/built_in_function.rs
index f92ae87d6e..8b4e65121c 100644
--- a/datafusion/expr/src/built_in_function.rs
+++ b/datafusion/expr/src/built_in_function.rs
@@ -31,7 +31,7 @@ use crate::{
 };
 
 use arrow::datatypes::{DataType, Field, Fields, IntervalUnit, TimeUnit};
-use datafusion_common::{internal_err, plan_err, DataFusionError, Result};
+use datafusion_common::{exec_err, plan_err, DataFusionError, Result};
 
 use strum::IntoEnumIterator;
 use strum_macros::EnumIter;
@@ -543,10 +543,11 @@ impl BuiltinScalarFunction {
 BuiltinScalarFunction::Flatten => {
 fn get_base_type(data_type: ) -> Result {
 match data_type {
-DataType::List(field) if matches!(field.data_type(), 
DataType::List(_)) => get_base_type(field.data_type()),
+DataType::List(field) | DataType::FixedSizeList(field, 
_) if matches!(field.data_type(), DataType::List(_)|DataType::FixedSizeList(_,_ 
)) => get_base_type(field.data_type()),
 DataType::LargeList(field) if 
matches!(field.data_type(), DataType::LargeList(_)) => 
get_base_type(field.data_type()),
 DataType::Null | DataType::List(_) | 
DataType::LargeList(_) => Ok(data_type.to_owned()),
-_ => internal_err!("Not reachable, data_type should be 
List or LargeList"),
+DataType::FixedSizeList(field,_ ) => 
Ok(DataType::List(field.clone())),
+_ => exec_err!("Not reachable, data_type should be 
List, LargeList or FixedSizeList"),
 }
 }
 
@@ -929,18 +930,18 @@ impl BuiltinScalarFunction {
 // 0 or more arguments of arbitrary type
 Signature::one_of(vec![VariadicEqual, Any(0)], 
self.volatility())
 }
-BuiltinScalarFunction::ArrayPopFront => Signature::any(1, 
self.volatility()),
-BuiltinScalarFunction::ArrayPopBack => Signature::any(1, 
self.volatility()),
+BuiltinScalarFunction::ArrayPopFront => 
Signature::array(self.volatility()),
+BuiltinScalarFunction::ArrayPopBack => 
Signature::array(self.volatility()),
 BuiltinScalarFunction::ArrayConcat => {
 Signature::variadic_any(self.volatility())
 }
-BuiltinScalarFunction::ArrayDims => Signature::any(1, 
self.volatility()),
-BuiltinScalarFunction::ArrayEmpty => Signature::any(1, 
self.volatility()),
+BuiltinScalarFunction::ArrayDims => 
Signature::array(self.volatility()),
+BuiltinScalarFunction::ArrayEmpty => 
Signature::array(self.volatility()),
 BuiltinScalarFunction::ArrayElement => {
 Signature::array_and_index(self.volatility())
 }
 BuiltinScalarFunction::ArrayExcept => Signature::any(2, 
self.volatility()),
-BuiltinScalarFunction::Flatten => Signature::any(1, 
self.volatility()),
+BuiltinScalarFunction::Flatten => 
Signature::array(self.volatility()),
 BuiltinScalarFunction::ArrayHasAll | 
BuiltinScalarFunction::ArrayHasAny => {
 Signature::any(2, self.volatility())
 }
@@ -950,8 +951,8 @@ impl BuiltinScalarFunction {
 

(arrow-datafusion) branch main updated: feat: expand `unnest` to accept arbitrary single array expression (#9342)

2024-02-26 Thread jayzhan
This is an automated email from the ASF dual-hosted git repository.

jayzhan pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


The following commit(s) were added to refs/heads/main by this push:
 new ec86acbc1f feat: expand `unnest`  to accept arbitrary single array 
expression (#9342)
ec86acbc1f is described below

commit ec86acbc1fbc0da1e0bec9ad066a5177ec586c96
Author: Jonah Gao 
AuthorDate: Mon Feb 26 21:33:19 2024 +0800

feat: expand `unnest`  to accept arbitrary single array expression (#9342)

* feat: expand `unnest`  to accept any single array expression

* unnest null

* review feedback
---
 datafusion/sql/src/expr/function.rs   | 58 ---
 datafusion/sqllogictest/test_files/unnest.slt | 48 +-
 2 files changed, 64 insertions(+), 42 deletions(-)

diff --git a/datafusion/sql/src/expr/function.rs 
b/datafusion/sql/src/expr/function.rs
index f56138066c..db572a23cf 100644
--- a/datafusion/sql/src/expr/function.rs
+++ b/datafusion/sql/src/expr/function.rs
@@ -16,16 +16,17 @@
 // under the License.
 
 use crate::planner::{ContextProvider, PlannerContext, SqlToRel};
+use arrow_schema::DataType;
 use datafusion_common::{
-exec_err, not_impl_err, plan_datafusion_err, plan_err, DFSchema, 
DataFusionError,
-Dependency, Result,
+not_impl_err, plan_datafusion_err, plan_err, DFSchema, DataFusionError, 
Dependency,
+Result,
 };
 use datafusion_expr::expr::{ScalarFunction, Unnest};
 use datafusion_expr::function::suggest_valid_function;
 use datafusion_expr::window_frame::{check_window_frame, 
regularize_window_order_by};
 use datafusion_expr::{
-expr, AggregateFunction, BuiltinScalarFunction, Expr, 
ScalarFunctionDefinition,
-WindowFrame, WindowFunctionDefinition,
+expr, AggregateFunction, BuiltinScalarFunction, Expr, ExprSchemable, 
WindowFrame,
+WindowFunctionDefinition,
 };
 use sqlparser::ast::{
 Expr as SQLExpr, Function as SQLFunction, FunctionArg, FunctionArgExpr, 
WindowType,
@@ -80,41 +81,34 @@ impl<'a, S: ContextProvider> SqlToRel<'a, S> {
 if name.eq("unnest") {
 let exprs =
 self.function_args_to_expr(args.clone(), schema, 
planner_context)?;
-
-match exprs.len() {
+// Currently only one argument is supported
+let arg = match exprs.len() {
 0 => {
-return exec_err!("unnest() requires at least one 
argument");
-}
-1 => {
-if let Expr::ScalarFunction(ScalarFunction {
-func_def:
-ScalarFunctionDefinition::BuiltIn(
-BuiltinScalarFunction::MakeArray,
-),
-..
-}) = exprs[0]
-{
-// valid
-} else if let Expr::Column(_) = exprs[0] {
-// valid
-} else if let Expr::ScalarFunction(ScalarFunction {
-func_def:
-
ScalarFunctionDefinition::BuiltIn(BuiltinScalarFunction::Struct),
-..
-}) = exprs[0]
-{
-return not_impl_err!("unnest() does not support struct 
yet");
-} else {
-return plan_err!(
-"unnest() can only be applied to array and structs 
and null"
-);
-}
+return plan_err!("unnest() requires at least one 
argument");
 }
+1 => [0],
 _ => {
 return not_impl_err!(
 "unnest() does not support multiple arguments yet"
 );
 }
+};
+// Check argument type, array types are supported
+match arg.get_type(schema)? {
+DataType::List(_)
+| DataType::LargeList(_)
+| DataType::FixedSizeList(_, _) => {}
+DataType::Struct(_) => {
+return not_impl_err!("unnest() does not support struct 
yet");
+}
+DataType::Null => {
+return not_impl_err!("unnest() does not support null yet");
+}
+_ => {
+return plan_err!(
+"unnest() can only be applied to array, struct and 
null"
+);
+}
 }
 
 return Ok(Expr::Unnest(Unnest { exprs }));
diff --git a/datafusion/sqllogictest/test_files/unnest.slt 
b/datafusion/sqllogictest/test_files/unnest.slt
index 7e4ce06be2..9990c00f75 100644
--- a/datafusion/sqllogictest/test_files/unnest.slt
+++ 

(arrow-rs) branch dependabot/cargo/object_store/master/nix-0.28.0 created (now a2a19cd4d96)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/cargo/object_store/master/nix-0.28.0
in repository https://gitbox.apache.org/repos/asf/arrow-rs.git


  at a2a19cd4d96 Update nix requirement from 0.27.1 to 0.28.0 in 
/object_store

No new revisions were added by this update.



(arrow-rs) branch master updated: Update nix requirement from 0.27.1 to 0.28.0 in /object_store (#5432)

2024-02-26 Thread tustvold
This is an automated email from the ASF dual-hosted git repository.

tustvold pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/arrow-rs.git


The following commit(s) were added to refs/heads/master by this push:
 new 282123441c7 Update nix requirement from 0.27.1 to 0.28.0 in 
/object_store (#5432)
282123441c7 is described below

commit 282123441c7bfd549c2a2216fba1e253e20501ba
Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
AuthorDate: Tue Feb 27 03:52:13 2024 +1300

Update nix requirement from 0.27.1 to 0.28.0 in /object_store (#5432)

Updates the requirements on [nix](https://github.com/nix-rust/nix) to 
permit the latest version.
- [Changelog](https://github.com/nix-rust/nix/blob/master/CHANGELOG.md)
- [Commits](https://github.com/nix-rust/nix/compare/v0.27.1...v0.28.0)

---
updated-dependencies:
- dependency-name: nix
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] 
Co-authored-by: dependabot[bot] 
<49699333+dependabot[bot]@users.noreply.github.com>
---
 object_store/Cargo.toml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/object_store/Cargo.toml b/object_store/Cargo.toml
index f3aaf35fbb0..6b38a8d21ff 100644
--- a/object_store/Cargo.toml
+++ b/object_store/Cargo.toml
@@ -57,7 +57,7 @@ tokio = { version = "1.25.0", features = ["sync", "macros", 
"rt", "time", "io-ut
 md-5 = { version = "0.10.6", default-features = false, optional = true }
 
 [target.'cfg(target_family="unix")'.dev-dependencies]
-nix = { version = "0.27.1", features = ["fs"] }
+nix = { version = "0.28.0", features = ["fs"] }
 
 [features]
 cloud = ["serde", "serde_json", "quick-xml", "hyper", "reqwest", 
"reqwest/json", "reqwest/stream", "chrono/serde", "base64", "rand", "ring"]



(arrow-rs) branch dependabot/cargo/object_store/master/nix-0.28.0 deleted (was a2a19cd4d96)

2024-02-26 Thread github-bot
This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a change to branch 
dependabot/cargo/object_store/master/nix-0.28.0
in repository https://gitbox.apache.org/repos/asf/arrow-rs.git


 was a2a19cd4d96 Update nix requirement from 0.27.1 to 0.28.0 in 
/object_store

The revisions that were on this branch are still contained in
other references; therefore, this change does not discard any commits
from the repository.