LuciferYang commented on PR #41529:
URL: https://github.com/apache/spark/pull/41529#issuecomment-1593078560

   `ReplSuite` in `repl` module test faild when using maven:
   
   ```
   ReplSuite:
   17500Spark context available as 'sc' (master = local, app id = 
local-1686829049116).
   17501Spark session available as 'spark'.
   17502- SPARK-15236: use Hive catalog *** FAILED ***
   17503  isContain was true Interpreter output contained 'Exception':
   17504  Welcome to
   17505        ____              __
   17506       / __/__  ___ _____/ /__
   17507      _\ \/ _ \/ _ `/ __/  '_/
   17508     /___/ .__/\_,_/_/ /_/\_\   version 3.5.0-SNAPSHOT
   17509        /_/
   17510           
   17511  Using Scala version 2.12.17 (OpenJDK 64-Bit Server VM, Java 1.8.0_372)
   17512  Type in expressions to have them evaluated.
   17513  Type :help for more information.
   17514  
   17515  scala> 
   17516  scala> java.lang.NoClassDefFoundError: 
org/sparkproject/guava/cache/CacheBuilder
   17517    at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.<init>(SessionCatalog.scala:197)
   17518    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalog$lzycompute(BaseSessionStateBuilder.scala:153)
   17519    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalog(BaseSessionStateBuilder.scala:152)
   17520    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.v2SessionCatalog$lzycompute(BaseSessionStateBuilder.scala:166)
   17521    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.v2SessionCatalog(BaseSessionStateBuilder.scala:166)
   17522    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalogManager$lzycompute(BaseSessionStateBuilder.scala:168)
   17523    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.catalogManager(BaseSessionStateBuilder.scala:168)
   17524    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder$$anon$1.<init>(BaseSessionStateBuilder.scala:185)
   17525    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.analyzer(BaseSessionStateBuilder.scala:185)
   17526    at 
org.apache.spark.sql.internal.BaseSessionStateBuilder.$anonfun$build$2(BaseSessionStateBuilder.scala:373)
   17527    at 
org.apache.spark.sql.internal.SessionState.analyzer$lzycompute(SessionState.scala:92)
   17528    at 
org.apache.spark.sql.internal.SessionState.analyzer(SessionState.scala:92)
   17529    at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:76)
   17530    at 
org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111)
   17531    at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:202)
   17532    at 
org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:529)
   17533    at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:202)
   17534    at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827)
   17535    at 
org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:201)
   17536    at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:76)
   17537    at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:74)
   17538    at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:66)
   17539    at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:99)
   17540    at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827)
   17541    at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97)
   17542    at 
org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:640)
   17543    at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827)
   17544    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:630)
   17545    at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:671)
   17546    ... 94 elided
   17547  Caused by: java.lang.ClassNotFoundException: 
org.sparkproject.guava.cache.CacheBuilder
   17548    at java.net.URLClassLoader.findClass(URLClassLoader.java:387)
   17549    at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
   17550    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
   17551    at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
   17552    ... 123 more
   17553  
   17554  scala>      | 
   17555  scala> :quit (ReplSuite.scala:83)
   17556Spark context available as 'sc' (master = local, app id = 
local-1686829054261).
   17557Spark session available as 'spark'.
   17558- SPARK-15236: use in-memory catalog
   17559Spark context available as 'sc' (master = local, app id = 
local-1686829056083).
   17560Spark session available as 'spark'.
   17561- broadcast vars
   17562Spark context available as 'sc' (master = local, app id = 
local-1686829059606).
   17563Spark session available as 'spark'.
   17564- line wrapper only initialized once when used as encoder outer scope
   17565Spark context available as 'sc' (master = local-cluster[1,1,1024], app 
id = app-20230615043742-0000).
   17566Spark session available as 'spark'.
   17567
   17568// Exiting paste mode, now interpreting.
   17569
   17570- define case class and create Dataset together with paste mode *** 
FAILED ***
   17571  isContain was true Interpreter output contained 'Exception':
   17572  Welcome to
   17573        ____              __
   17574       / __/__  ___ _____/ /__
   17575      _\ \/ _ \/ _ `/ __/  '_/
   17576     /___/ .__/\_,_/_/ /_/\_\   version 3.5.0-SNAPSHOT
   17577        /_/
   17578           
   17579  Using Scala version 2.12.17 (OpenJDK 64-Bit Server VM, Java 1.8.0_372)
   17580  Type in expressions to have them evaluated.
   17581  Type :help for more information.
   17582  
   17583  scala> // Entering paste mode (ctrl-D to finish)
   17584  
   17585  java.lang.NoClassDefFoundError: 
org/sparkproject/guava/util/concurrent/AtomicLongMap
   17586    at 
org.apache.spark.sql.catalyst.rules.QueryExecutionMetering.<init>(QueryExecutionMetering.scala:27)
   17587    at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$.<init>(RuleExecutor.scala:31)
   17588    at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$.<clinit>(RuleExecutor.scala)
   17589    at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:192)
   17590    at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.$anonfun$canonicalize$1(GenerateUnsafeProjection.scala:319)
   17591    at scala.collection.immutable.List.map(List.scala:293)
   17592    at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.canonicalize(GenerateUnsafeProjection.scala:319)
   17593    at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.generate(GenerateUnsafeProjection.scala:327)
   17594    at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:124)
   17595    at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:120)
   17596    at 
org.apache.spark.sql.catalyst.expressions.CodeGeneratorWithInterpretedFallback.createObject(CodeGeneratorWithInterpretedFallback.scala:51)
   17597    at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:151)
   17598    at 
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder$Serializer.apply(ExpressionEncoder.scala:198)
   17599    at 
org.apache.spark.sql.SparkSession.$anonfun$createDataset$1(SparkSession.scala:483)
   17650- SPARK-2576 importing implicits
   17651- Datasets and encoders *** FAILED ***
   17652  isContain was true Interpreter output contained 'error:':
   17653  
   17654  scala> import org.apache.spark.sql.functions._
   17655  
   17656  scala> import org.apache.spark.sql.{Encoder, Encoders}
   17657  
   17658  scala> import org.apache.spark.sql.expressions.Aggregator
   17659  
   17660  scala> import org.apache.spark.sql.TypedColumn
   17661  
   17662  scala>      |      |      |      |      |      |      | simpleSum: 
org.apache.spark.sql.TypedColumn[Int,Int] = $anon$1(boundreference() AS value, 
value, unresolveddeserializer(assertnotnull(upcast(getcolumnbyordinal(0, 
IntegerType), IntegerType, - root class: "int")), value#9), boundreference() AS 
value)
   17663  
   17664  scala> 
   17665  scala> java.lang.NoClassDefFoundError: Could not initialize class 
org.apache.spark.sql.catalyst.rules.RuleExecutor$
   17666    at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:192)
   17667    at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.$anonfun$canonicalize$1(GenerateUnsafeProjection.scala:319)
   17668    at scala.collection.immutable.List.map(List.scala:293)
   17669    at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.canonicalize(GenerateUnsafeProjection.scala:319)
   17670    at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.generate(GenerateUnsafeProjection.scala:327)
   17671    at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:124)
   17672    at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:120)
   17673    at 
org.apache.spark.sql.catalyst.expressions.CodeGeneratorWithInterpretedFallback.createObject(CodeGeneratorWithInterpretedFallback.scala:51)
   17674    at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:151)
   17675    at 
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder$Serializer.apply(ExpressionEncoder.scala:198)
   17676    at 
org.apache.spark.sql.SparkSession.$anonfun$createDataset$1(SparkSession.scala:483)
   17677    at scala.collection.immutable.List.map(List.scala:293)
   17678    at 
org.apache.spark.sql.SparkSession.createDataset(SparkSession.scala:483)
   17679    at 
org.apache.spark.sql.SQLContext.createDataset(SQLContext.scala:354)
   17680    at 
org.apache.spark.sql.SQLImplicits.localSeqToDatasetHolder(SQLImplicits.scala:244)
   17681    ... 39 elided
   17682  
   17683  scala> <console>:33: error: not found: value ds
   17684         ds.select(simpleSum).collect
   17685         ^
   17686  
   17687  scala>      | _result_1686829100269: Int = 1
   17688  
   17689  scala> (SingletonReplSuite.scala:106) 
   ```
   
   created [SPARK-44069](https://issues.apache.org/jira/browse/SPARK-44069) to 
tracking this


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to