Maxim Gekk created SPARK-31554:
----------------------------------

             Summary: Flaky test suite 
org.apache.spark.sql.hive.thriftserver.CliSuite
                 Key: SPARK-31554
                 URL: https://issues.apache.org/jira/browse/SPARK-31554
             Project: Spark
          Issue Type: Test
          Components: SQL
    Affects Versions: 3.1.0
            Reporter: Maxim Gekk


The test org.apache.spark.sql.hive.thriftserver.CliSuite fails very often, for 
example:
* https://github.com/apache/spark/pull/28328#issuecomment-618992335
The error message:
{code}
org.apache.spark.sql.hive.thriftserver.CliSuite.SPARK-11188 Analysis error 
reporting
Caused by: sbt.ForkMain$ForkError: java.lang.RuntimeException: Failed with 
error line 'Exception in thread "main" org.apache.spark.sql.AnalysisException: 
org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: 
Unable to instantiate 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient;'
        at 
org.apache.spark.sql.hive.thriftserver.CliSuite.$anonfun$runCliWithin$4(CliSuite.scala:138)
        at scala.collection.immutable.List.foreach(List.scala:392)
        at 
org.apache.spark.sql.hive.thriftserver.CliSuite.captureOutput$1(CliSuite.scala:135)
        at 
org.apache.spark.sql.hive.thriftserver.CliSuite.$anonfun$runCliWithin$6(CliSuite.scala:152)
        at 
org.apache.spark.sql.hive.thriftserver.CliSuite.$anonfun$runCliWithin$6$adapted(CliSuite.scala:152)
        at scala.sys.process.BasicIO$.readFully$1(BasicIO.scala:188)
        at 
scala.sys.process.BasicIO$.$anonfun$processFully$1$adapted(BasicIO.scala:192)
        at 
org.apache.spark.sql.test.ProcessTestUtils$ProcessOutputCapturer.run(ProcessTestUtils.scala:30)
{code}
* https://github.com/apache/spark/pull/28261#issuecomment-618950225
* https://github.com/apache/spark/pull/28261#issuecomment-618950225
* https://github.com/apache/spark/pull/27617#issuecomment-614318644



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to