Lij: jar tvf /Users/tyu/.m2/repository//mysql/mysql-connector-java/5.1.31/mysql-connector-java-5.1.31.jar | grep MySQLSyntaxErrorExceptio 914 Wed May 21 01:42:16 PDT 2014 com/mysql/jdbc/exceptions/MySQLSyntaxErrorException.class 842 Wed May 21 01:42:18 PDT 2014 com/mysql/jdbc/exceptions/jdbc4/MySQLSyntaxErrorException.class
5.1.34 has basically the same structure. Can you check if there is other version of mysql-connector-java on the classpath ? Thanks On Mon, Oct 19, 2015 at 11:26 AM, Lij Tapel <lij.ta...@gmail.com> wrote: > Sorry, here's the logs and source: > > The error I see in spark 1.5.1: http://pastebin.com/86K9WQ5f > * full logs here: http://pastebin.com/dfysSh9E > > What I used to see in spark 1.4.1: http://pastebin.com/eK3AZQFx > * full logs here: http://pastebin.com/iffSFFWW > > The source and build.sbt: http://pastebin.com/tUvcBerd > > On Mon, Oct 19, 2015 at 11:18 AM, Ted Yu <yuzhih...@gmail.com> wrote: > >> The attachments didn't go through. >> >> Consider pastbebin'ning. >> >> Thanks >> >> On Mon, Oct 19, 2015 at 11:15 AM, gbop <lij.ta...@gmail.com> wrote: >> >>> I've been struggling with a particularly puzzling issue after upgrading >>> to >>> Spark 1.5.1 from Spark 1.4.1. >>> >>> When I use the MySQL JDBC connector and an exception (e.g. >>> com.mysql.jdbc.exceptions.jdbc4.MySQLSyntaxErrorException) is thrown on >>> the >>> executor, I get a ClassNotFoundException on the driver, which results in >>> this error (logs are abbreviated): >>> >>> >>> >>> In Spark 1.4.1, I get the following (logs are abbreviated): >>> >>> >>> >>> I have seriously screwed up somewhere or this is a change in behavior >>> that I >>> have not been able to find in the documentation. For those that are >>> interested, a full repro and logs follow. >>> >>> >>> --- >>> >>> I am running this on Spark 1.5.1+Hadoop 2.6. I have tried this in various >>> combinations of >>> * local/standalone mode >>> * putting mysql on the classpath with --jars/building a fat jar with >>> mysql >>> in it/manually running sc.addJar on the mysql jar >>> * --deploy-mode client/--deploy-mode cluster >>> but nothing seems to change. >>> >>> >>> >>> Here is an example invocation, and the accompanying source code: >>> >>> >>> >>> >>> The source code: >>> >>> >>> >>> And the build.sbt: >>> >>> >>> >>> >>> And here are the results when run against Spark 1.4.1 (build.sbt has been >>> updated accordingly) >>> >>> >>> >>> >>> >>> -- >>> View this message in context: >>> http://apache-spark-user-list.1001560.n3.nabble.com/new-1-5-1-behavior-exception-on-executor-throws-ClassNotFound-on-driver-tp25124.html >>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>> For additional commands, e-mail: user-h...@spark.apache.org >>> >>> >> >