Interesting. I wonder why existing tests didn't catch that: class UnserializableException extends Exception { ./core/src/test/scala/org/apache/spark/rpc/RpcEnvSuite.scala class DAGSchedulerSuiteDummyException extends Exception ./core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala class TestException(msg: String) extends Exception(msg) ./streaming/src/test/scala/org/apache/spark/streaming/StreamingContextSuite.scala
On Mon, Oct 19, 2015 at 12:02 PM, Lij Tapel <lij.ta...@gmail.com> wrote: > I have verified that there is only 5.1.34 on the classpath. > > Funnily enough, I have a repro that doesn't even use mysql so this seems > to be purely a classloader issue: > > source: http://pastebin.com/WMCMwM6T > 1.4.1: http://pastebin.com/x38DQY2p > 1.5.1: http://pastebin.com/DQd6k818 > > > > On Mon, Oct 19, 2015 at 11:51 AM, Ted Yu <yuzhih...@gmail.com> wrote: > >> Lij: >> >> jar tvf >> /Users/tyu/.m2/repository//mysql/mysql-connector-java/5.1.31/mysql-connector-java-5.1.31.jar >> | grep MySQLSyntaxErrorExceptio >> 914 Wed May 21 01:42:16 PDT 2014 >> com/mysql/jdbc/exceptions/MySQLSyntaxErrorException.class >> 842 Wed May 21 01:42:18 PDT 2014 >> com/mysql/jdbc/exceptions/jdbc4/MySQLSyntaxErrorException.class >> >> 5.1.34 has basically the same structure. >> >> Can you check if there is other version of mysql-connector-java on the >> classpath ? >> >> Thanks >> >> On Mon, Oct 19, 2015 at 11:26 AM, Lij Tapel <lij.ta...@gmail.com> wrote: >> >>> Sorry, here's the logs and source: >>> >>> The error I see in spark 1.5.1: http://pastebin.com/86K9WQ5f >>> * full logs here: http://pastebin.com/dfysSh9E >>> >>> What I used to see in spark 1.4.1: http://pastebin.com/eK3AZQFx >>> * full logs here: http://pastebin.com/iffSFFWW >>> >>> The source and build.sbt: http://pastebin.com/tUvcBerd >>> >>> On Mon, Oct 19, 2015 at 11:18 AM, Ted Yu <yuzhih...@gmail.com> wrote: >>> >>>> The attachments didn't go through. >>>> >>>> Consider pastbebin'ning. >>>> >>>> Thanks >>>> >>>> On Mon, Oct 19, 2015 at 11:15 AM, gbop <lij.ta...@gmail.com> wrote: >>>> >>>>> I've been struggling with a particularly puzzling issue after >>>>> upgrading to >>>>> Spark 1.5.1 from Spark 1.4.1. >>>>> >>>>> When I use the MySQL JDBC connector and an exception (e.g. >>>>> com.mysql.jdbc.exceptions.jdbc4.MySQLSyntaxErrorException) is thrown >>>>> on the >>>>> executor, I get a ClassNotFoundException on the driver, which results >>>>> in >>>>> this error (logs are abbreviated): >>>>> >>>>> >>>>> >>>>> In Spark 1.4.1, I get the following (logs are abbreviated): >>>>> >>>>> >>>>> >>>>> I have seriously screwed up somewhere or this is a change in behavior >>>>> that I >>>>> have not been able to find in the documentation. For those that are >>>>> interested, a full repro and logs follow. >>>>> >>>>> >>>>> --- >>>>> >>>>> I am running this on Spark 1.5.1+Hadoop 2.6. I have tried this in >>>>> various >>>>> combinations of >>>>> * local/standalone mode >>>>> * putting mysql on the classpath with --jars/building a fat jar with >>>>> mysql >>>>> in it/manually running sc.addJar on the mysql jar >>>>> * --deploy-mode client/--deploy-mode cluster >>>>> but nothing seems to change. >>>>> >>>>> >>>>> >>>>> Here is an example invocation, and the accompanying source code: >>>>> >>>>> >>>>> >>>>> >>>>> The source code: >>>>> >>>>> >>>>> >>>>> And the build.sbt: >>>>> >>>>> >>>>> >>>>> >>>>> And here are the results when run against Spark 1.4.1 (build.sbt has >>>>> been >>>>> updated accordingly) >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> View this message in context: >>>>> http://apache-spark-user-list.1001560.n3.nabble.com/new-1-5-1-behavior-exception-on-executor-throws-ClassNotFound-on-driver-tp25124.html >>>>> Sent from the Apache Spark User List mailing list archive at >>>>> Nabble.com. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>> >>>>> >>>> >>> >> >