I don't need any specific version of Hadoop, I actually removed it from the 
build command and still get the error, I just need spark 1.6


> On Feb 3, 2016, at 9:05 AM, Felix Cheung <felixcheun...@hotmail.com> wrote:
> 
> I think his build command only works with Cloudera CDH 5.4.8, as you can see. 
> Mismatch Akka version is very common if the Hadoop distribution is different. 
> What version of Spark and Hadoop distribution are you running with?
> 
> 
> 
> 
> 
> On Tue, Feb 2, 2016 at 1:36 PM -0800, "Daniel Valdivia" 
> <h...@danielvaldivia.com <mailto:h...@danielvaldivia.com>> wrote:
> 
> Hello,
> 
> An update on the matter, using compile string
> 
> mvn clean package -Pspark-1.6 -Dspark.version=1.6.0 
> -Dhadoop.version=2.6.0-cdh5.4.8 -Phadoop-2.6 -Pyarn -Ppyspark -Pvendor-repo 
> -DskipTests
> 
> I end up getting the following error stack trace upon executing a new JSON
> 
> akka.ConfigurationException: Akka JAR version [2.2.3] does not match the 
> provided config version [2.3.11] at 
> akka.actor.ActorSystem$Settings.<init>(ActorSystem.scala:181) at 
> akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:470) at 
> akka.actor.ActorSystem$.apply(ActorSystem.scala:111) at 
> akka.actor.ActorSystem$.apply(ActorSystem.scala:104) at 
> org.apache.spark.util.AkkaUtils$.org$apache$spark$util$AkkaUtils$$doCreateActorSystem(AkkaUtils.scala:121)
>  at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:53) at 
> org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:52) at 
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:1964)
>  at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141) at 
> org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:1955) at 
> org.apache.spark.util.AkkaUtils$.createActorSystem(AkkaUtils.scala:55) at 
> org.apache.spark.SparkEnv$.create(SparkEnv.scala:266) at 
> org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:193) at 
> org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:288) at 
> org.apache.spark.SparkContext.<init>(SparkContext.scala:457) at 
> org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:339)
>  at 
> org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:145)
>  at 
> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:465)  
> at 
> org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
>  at 
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
>  at 
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92)
>  at 
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:300)
>  at org.apache.zeppelin.scheduler.Job.run(Job.java:169) at 
> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:134) at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at 
> java.util.concurrent.FutureTask.run(FutureTask.java:262) at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>  at java.lang.Thread.run(Thread.java:745)
> 
> There's some mentioning of this problem on SO, but seems like it was fixed
> 
> http://stackoverflow.com/questions/32294276/how-to-connect-zeppelin-to-spark-1-5-built-from-the-sources
>  
> <http://stackoverflow.com/questions/32294276/how-to-connect-zeppelin-to-spark-1-5-built-from-the-sources>
> 
> any idea on how to deal with this AKKA library problem?
> 
>> On Feb 2, 2016, at 12:02 PM, Daniel Valdivia <h...@danielvaldivia.com 
>> <mailto:h...@danielvaldivia.com>> wrote:
>> 
>> Hi,
>> 
>> Thanks for the suggestion, I'm running maven with Ben's command
>> 
>> Cheers!
>> 
>>> On Feb 1, 2016, at 7:47 PM, Benjamin Kim <bbuil...@gmail.com 
>>> <mailto:bbuil...@gmail.com>> wrote:
>>> 
>>> Hi Felix,
>>> 
>>> After installing Spark 1.6, I built Zeppelin using:
>>> 
>>> mvn clean package -Pspark-1.6 -Dspark.version=1.6.0 
>>> -Dhadoop.version=2.6.0-cdh5.4.8 -Phadoop-2.6 -Pyarn -Ppyspark -Pvendor-repo 
>>> -DskipTests
>>> 
>>> This worked for me.
>>> 
>>> Cheers,
>>> Ben
>>> 
>>> 
>>>> On Feb 1, 2016, at 7:44 PM, Felix Cheung <felixcheun...@hotmail.com 
>>>> <mailto:felixcheun...@hotmail.com>> wrote:
>>>> 
>>>> Hi
>>>> 
>>>> You can see the build command line example here for spark 1.6 profile
>>>> 
>>>> https://github.com/apache/incubator-zeppelin/blob/master/README.md 
>>>> <https://github.com/apache/incubator-zeppelin/blob/master/README.md>
>>>> 
>>>> 
>>>> 
>>>> 
>>>> 
>>>> On Mon, Feb 1, 2016 at 3:59 PM -0800, "Daniel Valdivia" 
>>>> <h...@danielvaldivia.com <mailto:h...@danielvaldivia.com>> wrote:
>>>> 
>>>> Hi,
>>>> 
>>>> I'd like to ask if there's an easy way to upgrade spark to 1.6.0 from the 
>>>> current 1.4.x that's bundled with the current release of zepellin, would 
>>>> updating the pom.xml and compiling suffice ?
>>>> 
>>>> Cheers
>>> 
>> 
> 

Reply via email to