HyukjinKwon opened a new pull request, #43858:
URL: https://github.com/apache/spark/pull/43858

   ### What changes were proposed in this pull request?
   
   This PR proposes to move the partitioning expressions for DSv2 `hours`, 
`days`, `months`, and `bucket` from `org.apache.spark.sql.functions` to 
`org.apache.spark.sql.functions.partitioning`. Old references are deprecated.
   
   ### Why are the changes needed?
   
   Those expressions can ONLY be used in DSv2 `partitionBy` but they exist in 
the same place with other expressions that confuse users that they believe they 
can use with other places such as `select`. They at least have to be grouped 
separately so users don't be confused.
   
   We were unable to change this because such nested object disallows the 
access from JVM in Scala 2.12 with JDK 8. Now with Scala 2.13 and JDK 11, they 
are possible.
   
   Just for doubly sure, I had a offline discussion including the original 
author, and related people a while ago.
   
   ### Does this PR introduce _any_ user-facing change?
   
   Yes, it moves the partitioning expressions for DSv2 `hours`, `days`, 
`months`, and `bucket` from `org.apache.spark.sql.functions` to 
`org.apache.spark.sql.functions.partitioning`.
   
   ### How was this patch tested?
   
   Reusing the existent unittest, and new test cases in Java.
   
   ### Was this patch authored or co-authored using generative AI tooling?
   
   No.
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to