thejdeep commented on code in PR #35969: URL: https://github.com/apache/spark/pull/35969#discussion_r1070209814
########## sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala: ########## @@ -154,6 +156,19 @@ class FileBasedDataSourceSuite extends QueryTest } } + val emptySchemaSupportedDataSources = Seq("orc", "csv", "json") + emptySchemaSupportedDataSources.foreach { format => + val emptySchemaValidationConf = SQLConf.ALLOW_EMPTY_SCHEMAS_FOR_WRITES.key + test(s"SPARK-38651 allow writing empty schema files " + + s"using $format when ${emptySchemaValidationConf} is enabled") { + withSQLConf(emptySchemaValidationConf -> "true") { + withTempPath { outputPath => + spark.emptyDataFrame.write.format(format).save(outputPath.toString) Review Comment: Reading files that contain certain schemas like `orc` require schema to be specified when it cannot be inferred. Hence, I did not go down that route of validating contents by reading and loading the written path again. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org