Github user gatorsmile commented on a diff in the pull request:

    https://github.com/apache/spark/pull/20331#discussion_r162683627
  
    --- Diff: 
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/orc/OrcHadoopFsRelationSuite.scala
 ---
    @@ -82,44 +80,4 @@ class OrcHadoopFsRelationSuite extends 
HadoopFsRelationTest {
           }
         }
       }
    -
    -  test("SPARK-13543: Support for specifying compression codec for ORC via 
option()") {
    -    withTempPath { dir =>
    -      val path = s"${dir.getCanonicalPath}/table1"
    -      val df = (1 to 5).map(i => (i, (i % 2).toString)).toDF("a", "b")
    -      df.write
    -        .option("compression", "ZlIb")
    -        .orc(path)
    -
    -      // Check if this is compressed as ZLIB.
    -      val maybeOrcFile = new File(path).listFiles().find { f =>
    -        !f.getName.startsWith("_") && f.getName.endsWith(".zlib.orc")
    -      }
    -      assert(maybeOrcFile.isDefined)
    -      val orcFilePath = maybeOrcFile.get.toPath.toString
    -      val expectedCompressionKind =
    -        OrcFileOperator.getFileReader(orcFilePath).get.getCompression
    -      assert("ZLIB" === expectedCompressionKind.name())
    -
    -      val copyDf = spark
    -        .read
    -        .orc(path)
    -      checkAnswer(df, copyDf)
    -    }
    -  }
    -
    -  test("Default compression codec is snappy for ORC compression") {
    -    withTempPath { file =>
    -      spark.range(0, 10).write
    -        .orc(file.getCanonicalPath)
    -      val expectedCompressionKind =
    -        
OrcFileOperator.getFileReader(file.getCanonicalPath).get.getCompression
    -      assert("SNAPPY" === expectedCompressionKind.name())
    -    }
    -  }
    -}
    -
    -class HiveOrcHadoopFsRelationSuite extends OrcHadoopFsRelationSuite {
    --- End diff --
    
    This is Hive only.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to