LuciferYang commented on PR #45583:
URL: https://github.com/apache/spark/pull/45583#issuecomment-2255088028
Sorry to disturb everyone, but when I execute `OrcEncryptionSuite` on my M2
Max, I find that there are some differences when using Hadoop 3.4.0 and Hadoop
3.3.4.
`build/sbt clean "sql/testOnly
org.apache.spark.sql.execution.datasources.orc.OrcEncryptionSuite"`
- branch-3.5(with hadoop 3.3.4)
```
[info] OrcEncryptionSuite:
14:44:11.580 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load
native-hadoop library for your platform... using builtin-java classes where
applicable
[info] - Write and read an encrypted file (1 second, 921 milliseconds)
[info] - Write and read an encrypted table (374 milliseconds)
[info] - SPARK-35325: Write and read encrypted nested columns (358
milliseconds)
[info] - SPARK-35992: Write and read fully-encrypted columns with default
masking (570 milliseconds)
14:44:15.461 WARN
org.apache.spark.sql.execution.datasources.orc.OrcEncryptionSuite:
[info] Run completed in 4 seconds, 694 milliseconds.
[info] Total number of tests run: 4
[info] Suites: completed 1, aborted 0
[info] Tests: succeeded 4, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
```
- master(with hadoop 3.4.0)
```
[info] OrcEncryptionSuite:
14:49:15.267 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load
native-hadoop library for your platform... using builtin-java classes where
applicable
14:49:17.636 WARN org.apache.hadoop.crypto.OpensslCipher: Failed to load
OpenSSL Cipher.
java.lang.UnsatisfiedLinkError: 'boolean
org.apache.hadoop.util.NativeCodeLoader.buildSupportsOpenssl()'
at org.apache.hadoop.util.NativeCodeLoader.buildSupportsOpenssl(Native
Method)
at
org.apache.hadoop.crypto.OpensslCipher.<clinit>(OpensslCipher.java:86)
at
org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec.<init>(OpensslAesCtrCryptoCodec.java:36)
at
java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)
at
java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77)
at
java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at
java.base/java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499)
[info] - Write and read an encrypted file (2 seconds, 343 milliseconds)
[info] - Write and read an encrypted table (405 milliseconds)
[info] - SPARK-35325: Write and read encrypted nested columns (308
milliseconds)
[info] - SPARK-35992: Write and read fully-encrypted columns with default
masking (555 milliseconds)
14:49:19.493 WARN
org.apache.spark.sql.execution.datasources.orc.OrcEncryptionSuite:
[info] Run completed in 5 seconds, 84 milliseconds.
[info] Total number of tests run: 4
[info] Suites: completed 1, aborted 0
[info] Tests: succeeded 4, failed 0, canceled 0, ignored 0, pending 0
[info] All tests passed.
```
When using Hadoop 3.4.0, although there were no test failures, an
`UnsatisfiedLinkError` was thrown. Is this expected, or do I need to configure
additional dependencies? This issue should only occur on Apple Silicon chips
now. @dongjoon-hyun @steveloughran
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]