Muhammad Abubakar created SPARK-32636:
-----------------------------------------
Summary: AsyncEventQueue: Exception scala.Some cannot be cast to
java.lang.String
Key: SPARK-32636
URL: https://issues.apache.org/jira/browse/SPARK-32636
Project: Spark
Issue Type: Bug
Components: Java API
Affects Versions: 3.0.0
Reporter: Muhammad Abubakar
Fix For: 3.0.1
Java Out of Memory Exception occurs when try to run memory-intensive job.
Although enough resources are available on the machine. But the actual
exception doesn't look like occurred because of memory issue.
{code:java}
java.lang.ClassCastException: scala.Some cannot be cast to
java.lang.Stringjava.lang.ClassCastException: scala.Some cannot be cast to
java.lang.String at org.json4s.JsonDSL.pair2jvalue(JsonDSL.scala:82) at
org.json4s.JsonDSL.pair2jvalue$(JsonDSL.scala:82) at
org.json4s.JsonDSL$.pair2jvalue(JsonDSL.scala:64) at
org.apache.spark.util.JsonProtocol$.taskInfoToJson(JsonProtocol.scala:309) at
org.apache.spark.util.JsonProtocol$.taskStartToJson(JsonProtocol.scala:131) at
org.apache.spark.util.JsonProtocol$.sparkEventToJson(JsonProtocol.scala:75) at
org.apache.spark.scheduler.EventLoggingListener.logEvent(EventLoggingListener.scala:97)
at
org.apache.spark.scheduler.EventLoggingListener.onTaskStart(EventLoggingListener.scala:114)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:41)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:115) at
org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:99) at
org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
at
org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
at scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23) at
scala.util.DynamicVariable.withValue(DynamicVariable.scala:62) at
org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1319) at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)
## A fatal error has been detected by the Java Runtime Environment:## SIGSEGV
(0xb) at pc=0x00007f9d2ec0cc6b, pid=30234, tid=0x00007f9174a9e700## JRE
version: Java(TM) SE Runtime Environment (8.0_101-b13) (build 1.8.0_101-b13)#
Java VM: Java HotSpot(TM) 64-Bit Server VM (25.101-b13 mixed mode linux-amd64
)# Problematic frame:# V [libjvm.so+0x7c9c6b]20/08/11 23:31:38 ERROR
AsyncEventQueue: Listener AppStatusListener threw an
exceptionjava.lang.ClassCastException: scala.Some cannot be cast to
java.lang.String at
org.apache.spark.status.LiveEntityHelpers$.weakIntern(LiveEntity.scala:665) at
org.apache.spark.status.LiveTask.doUpdate(LiveEntity.scala:209) at
org.apache.spark.status.LiveEntity.write(LiveEntity.scala:51) at
org.apache.spark.status.AppStatusListener.update(AppStatusListener.scala:1088)
at
org.apache.spark.status.AppStatusListener.liveUpdate(AppStatusListener.scala:1101)
at
org.apache.spark.status.AppStatusListener.onTaskStart(AppStatusListener.scala:512)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:41)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:115) at
org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:99) at
org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
at
org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
at scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23) at
scala.util.DynamicVariable.withValue(DynamicVariable.scala:62) at
org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1319) at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)[thread
140262972720896 also had an error] Klass::external_name() const+0x1b## Failed
to write core dump. Core dumps have been disabled. To enable core dumping, try
"ulimit -c unlimited" before starting Java again## An error report file with
more information is saved as:
{code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]