[ 
https://issues.apache.org/jira/browse/SPARK-32636?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17179354#comment-17179354
 ] 

Rohit Mishra commented on SPARK-32636:
--------------------------------------

[~abubakarj], Thanks for raising the bug but please refrain from adding Fix 
version as that field is reserved for committers. Thanks. 

>  AsyncEventQueue: Exception scala.Some cannot be cast to java.lang.String
> -------------------------------------------------------------------------
>
>                 Key: SPARK-32636
>                 URL: https://issues.apache.org/jira/browse/SPARK-32636
>             Project: Spark
>          Issue Type: Bug
>          Components: Java API
>    Affects Versions: 3.0.0
>            Reporter: Muhammad Abubakar
>            Priority: Major
>         Attachments: err.log
>
>
> Spark 3.0.0. Hadoop 3.2. Hive 2.3,7(built-in)
>  
> Java Exception occurs when try to run a memory-intensive job. Although enough 
> resources are available on the machine. But the actual exception doesn't look 
> like occurred because of memory issue.
> {code:java}
> java.lang.ClassCastException: scala.Some cannot be cast to 
> java.lang.Stringjava.lang.ClassCastException: scala.Some cannot be cast to 
> java.lang.String at org.json4s.JsonDSL.pair2jvalue(JsonDSL.scala:82) at 
> org.json4s.JsonDSL.pair2jvalue$(JsonDSL.scala:82) at 
> org.json4s.JsonDSL$.pair2jvalue(JsonDSL.scala:64) at 
> org.apache.spark.util.JsonProtocol$.taskInfoToJson(JsonProtocol.scala:309) at 
> org.apache.spark.util.JsonProtocol$.taskStartToJson(JsonProtocol.scala:131) 
> at 
> org.apache.spark.util.JsonProtocol$.sparkEventToJson(JsonProtocol.scala:75) 
> at 
> org.apache.spark.scheduler.EventLoggingListener.logEvent(EventLoggingListener.scala:97)
>  at 
> org.apache.spark.scheduler.EventLoggingListener.onTaskStart(EventLoggingListener.scala:114)
>  at 
> org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:41)
>  at 
> org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
>  at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:115) at 
> org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:99) at 
> org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
>  at scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23) at 
> scala.util.DynamicVariable.withValue(DynamicVariable.scala:62) at 
> org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
>  at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1319) at 
> org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)
> ## A fatal error has been detected by the Java Runtime Environment:##  
> SIGSEGV (0xb) at pc=0x00007f9d2ec0cc6b, pid=30234, tid=0x00007f9174a9e700## 
> JRE version: Java(TM) SE Runtime Environment (8.0_101-b13) (build 
> 1.8.0_101-b13)# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.101-b13 mixed 
> mode linux-amd64 )# Problematic frame:# V  [libjvm.so+0x7c9c6b]20/08/11 
> 23:31:38 ERROR AsyncEventQueue: Listener AppStatusListener threw an 
> exceptionjava.lang.ClassCastException: scala.Some cannot be cast to 
> java.lang.String at 
> org.apache.spark.status.LiveEntityHelpers$.weakIntern(LiveEntity.scala:665) 
> at org.apache.spark.status.LiveTask.doUpdate(LiveEntity.scala:209) at 
> org.apache.spark.status.LiveEntity.write(LiveEntity.scala:51) at 
> org.apache.spark.status.AppStatusListener.update(AppStatusListener.scala:1088)
>  at 
> org.apache.spark.status.AppStatusListener.liveUpdate(AppStatusListener.scala:1101)
>  at 
> org.apache.spark.status.AppStatusListener.onTaskStart(AppStatusListener.scala:512)
>  at 
> org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:41)
>  at 
> org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
>  at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:115) at 
> org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:99) at 
> org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
>  at scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23) at 
> scala.util.DynamicVariable.withValue(DynamicVariable.scala:62) at 
> org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
>  at 
> org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
>  at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1319) at 
> org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)[thread
>  140262972720896 also had an error]  Klass::external_name() const+0x1b## 
> Failed to write core dump. Core dumps have been disabled. To enable core 
> dumping, try "ulimit -c unlimited" before starting Java again## An error 
> report file with more information is saved as:
> {code}
> See Attached err.log generated by spark [^err.log]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to