vaibhawvipul commented on code in PR #538:
URL: https://github.com/apache/datafusion-comet/pull/538#discussion_r1634326070
##########
spark/src/test/scala/org/apache/comet/CometExpressionSuite.scala:
##########
@@ -1583,66 +1585,51 @@ class CometExpressionSuite extends CometTestBase with
AdaptiveSparkPlanHelper {
withTempDir { dir =>
// Array values test
- val arrayPath = new Path(dir.toURI.toString,
"array_test.parquet").toString
- Seq(Int.MaxValue,
Int.MinValue).toDF("a").write.mode("overwrite").parquet(arrayPath)
- val arrayQuery = "select a, -a from t"
- runArrayTest(arrayQuery, "integer", arrayPath)
-
- // long values test
- val longArrayPath = new Path(dir.toURI.toString,
"long_array_test.parquet").toString
- Seq(Long.MaxValue, Long.MinValue)
- .toDF("a")
- .write
- .mode("overwrite")
- .parquet(longArrayPath)
- val longArrayQuery = "select a, -a from t"
- runArrayTest(longArrayQuery, "long", longArrayPath)
-
- // short values test
- val shortArrayPath = new Path(dir.toURI.toString,
"short_array_test.parquet").toString
- Seq(Short.MaxValue, Short.MinValue)
- .toDF("a")
- .write
- .mode("overwrite")
- .parquet(shortArrayPath)
- val shortArrayQuery = "select a, -a from t"
- runArrayTest(shortArrayQuery, "", shortArrayPath)
-
- // byte values test
- val byteArrayPath = new Path(dir.toURI.toString,
"byte_array_test.parquet").toString
- Seq(Byte.MaxValue, Byte.MinValue)
- .toDF("a")
- .write
- .mode("overwrite")
- .parquet(byteArrayPath)
- val byteArrayQuery = "select a, -a from t"
- runArrayTest(byteArrayQuery, "", byteArrayPath)
-
- // interval values test
- withTable("t_interval") {
- spark.sql("CREATE TABLE t_interval(a STRING) USING PARQUET")
- spark.sql("INSERT INTO t_interval VALUES ('INTERVAL 10000000000
YEAR')")
- withAnsiMode(enabled = true) {
- spark
- .sql("SELECT CAST(a AS INTERVAL) AS a FROM t_interval")
- .createOrReplaceTempView("t_interval_casted")
- checkOverflow("SELECT a, -a FROM t_interval_casted", "interval")
- }
- }
-
- withTable("t") {
- sql("create table t(a int) using parquet")
- sql("insert into t values (-2147483648)")
- withAnsiMode(enabled = true) {
- checkOverflow("select a, -a from t", "integer")
- }
+ val dataTypes = Seq(
+ ("array_test.parquet", Seq(Int.MaxValue, Int.MinValue).toDF("a"),
"integer"),
+ ("long_array_test.parquet", Seq(Long.MaxValue,
Long.MinValue).toDF("a"), "long"),
+ ("short_array_test.parquet", Seq(Short.MaxValue,
Short.MinValue).toDF("a"), ""),
+ ("byte_array_test.parquet", Seq(Byte.MaxValue,
Byte.MinValue).toDF("a"), ""))
+
+ dataTypes.foreach { case (fileName, df, dtype) =>
+ val path = new Path(dir.toURI.toString, fileName).toString
+ df.write.mode("overwrite").parquet(path)
+ val query = "select a, -a from t"
+ runArrayTest(query, dtype, path)
}
- withTable("t_float") {
- sql("create table t_float(a float) using parquet")
- sql("insert into t_float values (3.4128235E38)")
- withAnsiMode(enabled = true) {
- checkOverflow("select a, -a from t_float", "float")
+ // scalar tests
+ withParquetTable((0 until 5).map(i => (i % 5, i % 3)), "tbl") {
+ withSQLConf(
+ "spark.sql.optimizer.excludedRules" ->
"org.apache.spark.sql.catalyst.optimizer.ConstantFolding",
+ SQLConf.ANSI_ENABLED.key -> "true",
+ CometConf.COMET_ANSI_MODE_ENABLED.key -> "true",
+ CometConf.COMET_ENABLED.key -> "true",
+ CometConf.COMET_EXEC_ENABLED.key -> "true") {
+ for (n <- Seq("2147483647", "-2147483648")) {
+ checkOverflow(s"select -(cast(${n} as int)) FROM tbl", "integer")
+ }
+ for (n <- Seq("32767", "-32768")) {
+ checkOverflow(s"select -(cast(${n} as short)) FROM tbl", "")
+ }
+ for (n <- Seq("127", "-128")) {
+ checkOverflow(s"select -(cast(${n} as byte)) FROM tbl", "")
+ }
+ for (n <- Seq("9223372036854775807", "-9223372036854775808")) {
+ checkOverflow(s"select -(cast(${n} as long)) FROM tbl", "long")
+ }
+ for (n <- Seq("3.4028235E38", "-3.4028235E38")) {
+ checkOverflow(s"select -(cast(${n} as float)) FROM tbl", "float")
+ }
+ // interval test without cast
+ val longDf = Seq(Long.MaxValue, Long.MaxValue, 2)
+ val yearMonthDf = Seq(Int.MaxValue, Int.MaxValue, 2)
+ .map(Period.ofMonths)
+ val dayTimeDf = Seq(106751991L, 106751991L, 2L)
+ .map(Duration.ofDays)
+ Seq(longDf, yearMonthDf, dayTimeDf).foreach { df =>
+ checkOverflow(s"select -(_1) FROM tbl", "")
Review Comment:
moved to array section.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]