parquet stuff is unrelated, filed https://github.com/apache/parquet-java/issues/3237 . still my fault though.
overall, I'm going to give a quick -1 with the commitment to provide one of the patches. my comments in markdown format: ## relesse docs index.html the summary in index.html are from 3.4.1, so we need to update them to 3.4.2 with key features anre issues fixed. Call out avro upgrade in the release notes -this is incompatible but required for security reasons. > We have upgraded to Avro 1.11.4, which is incompatible with previous versions. { This is critical for security reasons -everyone needs to upgrade their own uses of Avro too. A side-effect of excluding the AWS SDK from the release is that the hadoop-aws module does not declare a transitive dependency on the AWS SDK. This means that if you use hadoop-aws, you must explicitly include the AWS SDK in your project. I'd like to see this (somehow) returned, at the very least we explicitly add it into the hadoop-cloud-connectors dependencies so you can get it that way. otherwise it's a bit trickier to get right doc changes > This release was qualified against the 2.29.52 release, which at the time of the qualification was the last release compatible with third-party stores. This regression may have been fixed in later releases. <dependency> <groupId>software.amazon.awssdk</groupId> <artifactId>bundle</artifactId> <version>2.29.52</version> <exclusions> <exclusion> <groupId>*</groupId> <artifactId>*</artifactId> </exclusion> </exclusions> </dependency> > The minimum version of the AWS SDK bundle.jar which can be used in this release is 2.29.52; later versions may work, but no later versions have been qualified. Upgrade with care, especially when working with third-party stores. > see: https://issues.apache.org/jira/browse/HADOOP-19490 ## java8 my setup has an old java8 release java -version openjdk version "1.8.0_362" OpenJDK Runtime Environment (Zulu 8.68.0.21-CA-macos-aarch64) (build 1.8.0_362-b09) OpenJDK 64-Bit Server VM (Zulu 8.68.0.21-CA-macos-aarch64) (build 25.362-b09, mixed mode) it is getting that recurrent bytebuffer flip error, from an overloaded version of ByteBuffer...this can happen even with recent java8 releases. We cast the ByteBuffer to (java.nio.Buffer) before calling flip(). [INFO] ------------------------------------------------------- [INFO] Running org.apache.parquet.hadoop.TestParquetReader Exception in thread "Thread-8" java.lang.NoSuchMethodError: java.nio.ByteBuffer.flip()Ljava/nio/ByteBuffer; at org.apache.hadoop.fs.RawLocalFileSystem$AsyncHandler.completed(RawLocalFileSystem.java:428) at org.apache.hadoop.fs.RawLocalFileSystem$AsyncHandler.completed(RawLocalFileSystem.java:362) at sun.nio.ch.Invoker.invokeUnchecked(Invoker.java:126) at sun.nio.ch.SimpleAsynchronousFileChannelImpl$2.run(SimpleAsynchronousFileChannelImpl.java:335) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) this can actually be fixed trivially, casting to java.io.Buffer first: ((Byffer)buff).flip()) I'll do this On Tue, 3 Jun 2025 at 16:04, Steve Loughran <ste...@cloudera.com> wrote: > > I'm sort of leaning towards a -1, though the regression is actually in > 3.4.0, and surfacing as a failure to release buffers in localfs IO. > > > [ERROR] org.apache.parquet.hadoop.TestParquetReader.testRangeFiltering[2] > -- Time elapsed: 0.033 s <<< ERROR! > org.apache.parquet.bytes.TrackingByteBufferAllocator$LeakedByteBufferException: > 4 ByteBuffer object(s) is/are remained unreleased after closing this > allocator. > at > org.apache.parquet.bytes.TrackingByteBufferAllocator.close(TrackingByteBufferAllocator.java:160) > at > org.apache.parquet.hadoop.TestParquetReader.closeAllocator(TestParquetReader.java:175) > at > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > at > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.base/java.lang.reflect.Method.invoke(Method.java:566) > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) > at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > Caused by: > org.apache.parquet.bytes.TrackingByteBufferAllocator$ByteBufferAllocationStacktraceException: > Set org.apache.parquet.bytes.TrackingByteBufferAllocator.DEBUG = true for > more info > > > 1. This is very much the vector io stuff. > 2. It's not a regression since 3.4.0 > 3. but it's indicative of a memory leak. Surprised and Annoyed I > hadn't spotted this before...the tests went in last year. > > > On Tue, 3 Jun 2025 at 10:55, Steve Loughran <ste...@cloudera.com> wrote: > >> >> Ahmar, no need for prs there, it's a "commit then review" repo, as in >> other people can revoke if you break things. >> >> will do the merge and test >> >> On Mon, 2 Jun 2025 at 16:51, Suhail, Ahmar <ahma...@amazon.co.uk.invalid> >> wrote: >> >>> Steve - I created a PR on hadoop-release-support with my properties >>> file: https://github.com/apache/hadoop-release-support/pull/4 >>> >>> >>> Masatake - Yes, sounds good, will update the documentation for the new >>> RC if it is created. >>> >>> >>> Mukund - I realised too late that I messed up the commit for that one. >>> Will figure out how to fix.. >>> >>> ________________________________ >>> From: Steve Loughran <ste...@cloudera.com> >>> Sent: Monday, June 2, 2025 4:25:00 PM >>> To: Ahmar Suhail >>> Cc: common-...@hadoop.apache.org; hdfs-...@hadoop.apache.org; >>> mapreduce-dev@hadoop.apache.org; yarn-...@hadoop.apache.org >>> Subject: RE: [EXTERNAL] [VOTE] Release Apache Hadoop 3.4.2 >>> >>> >>> CAUTION: This email originated from outside of the organization. Do not >>> click links or open attachments unless you can confirm the sender and know >>> the content is safe. >>> >>> >>> >>> can you put up the relevant changes to the hadoop-release-support >>> module; I'd like to use it as part of my validation and I'm assuming you >>> have that src/releases/release-info-3.4.2.properties file >>> >>> On Wed, 28 May 2025 at 13:25, Ahmar Suhail <ah...@apache.org<mailto: >>> ah...@apache.org>> wrote: >>> Hey all, >>> >>> The first release candidate for Hadoop 3.4.2 is now available for voting. >>> >>> There are a couple of things to note: >>> >>> 1/ No Arm64 artifacts. This is due to previously reported issues: >>> https://issues.apache.org/jira/projects/YARN/issues/YARN-11712 and >>> ttps://issues.apache.org/jira/projects/YARN/issues/YARN-11713< >>> http://issues.apache.org/jira/projects/YARN/issues/YARN-11713> >>> <https://issues.apache.org/jira/projects/YARN/issues/YARN-11713>, which >>> mean that the build fails on arm64. >>> >>> 2/ Relevant for anyone testing S3A: We've removed the AWS SDK bundle >>> from hadoop-3.4.2.tar.gz. This is because the SDK bundle is now ~600MB, >>> which makes the size of tar > 1GB, and it can no longer be uploaded to >>> SVN. >>> For S3A, download SDK bundle v2.29.52 from: >>> https://mvnrepository.com/artifact/software.amazon.awssdk/bundle/2.29.52 >>> , >>> and drop it into /share/hadoop/common/lib. Release notes will be updated >>> with these instructions. >>> >>> >>> The RC is available at: >>> >>> https://dist.apache.org/repos/dist/dev/hadoop/hadoop-3.4.2-RC1/ >>> >>> The git tag is release-3.4.2-RC1, commit >>> 09870840ec35b48cd107972eb24d25e8aece04c9 >>> >>> The maven artifacts are staged at: >>> >>> https://repository.apache.org/content/repositories/orgapachehadoop-1437 >>> >>> >>> You can find my public key (02085AFB652F796A3B01D11FD737A6F52281FA98) at: >>> >>> https://dist.apache.org/repos/dist/release/hadoop/common/KEYS >>> >>> >>> This release has been created off of branch-3.4. Key changes include: >>> >>> * S3A: Integration with S3 Analytics Accelerator input stream >>> * S3A: Support for S3 conditional writes >>> * ABFS: Deprecation of WASB driver >>> * ABFS: Support for Non-Heirarchical Namespace Accounts on ABFS Driver >>> >>> >>> This is my first attempt at managing a release, please do test the >>> release >>> and let me know in case of any issues. >>> >>> Thanks, >>> Ahmar >>> >>