I've tried it bulding spark with mesos 0.14.2 and I have the exact same 
error

Stack: [0x00007f82f5849000,0x00007f82f594a000],  sp=0x00007f82f59485d0, 
 free space=1021k
Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native 
code)
V  [libjvm.so+0x632d09]  jni_GetByteArrayElements+0x89
C  [libmesos-0.14.2.so+0x5e08b9]  mesos::FrameworkInfo 
construct<mesos::FrameworkInfo>(JNIEnv_*, _jobject*)+0x79

Java frames: (J=compiled Java code, j=interpreted, Vv=VM code)
j  org.apache.mesos.MesosSchedulerDriver.initialize()V+0
j 
 
org.apache.mesos.MesosSchedulerDriver.<init>(Lorg/apache/mesos/Scheduler;Lorg/apache/mesos/Protos$FrameworkInfo;Ljava/lang/String;)V+62
j 
 
org.apache.spark.scheduler.cluster.mesos.MesosSchedulerBackend$$anon$1.run()V+44
v  ~StubRoutines::call_stub


On Monday, December 30, 2013 7:18:57 PM UTC-5, Jey Kottalam wrote:
>
> It looks like your Spark is built against mesos 0.13.0 according to the 
> stacktrace.  You may need to rebuild Spark to link with your custom build 
> of Mesos 0.14.2.
>
> -Jey
>
>
> On Mon, Dec 30, 2013 at 1:39 PM, Damien Dubé <[email protected]<javascript:>
> > wrote:
>
>> Once I have my mesos cluster up and running. My spark job is always 
>> returning the same error. I have tried multiple options but I am still 
>> having the same errror.
>>
>> Here is the stack trace
>>
>> Stack: [0x00007f41ea4c1000,0x00007f41ea5c2000],  sp=0x00007f41ea5c0670, 
>>  free space=1021k
>> Native frames: (J=compiled Java code, j=interpreted, Vv=VM code, C=native 
>> code)
>> V  [libjvm.so+0x632d09]  jni_GetByteArrayElements+0x89
>> C  [libmesos-0.13.0.so+0x5a4559]  mesos::FrameworkInfo 
>> construct<mesos::FrameworkInfo>(JNIEnv_*, _jobject*)+0x79
>> C  0x00007f420c0798a8
>> j 
>>  
>> org.apache.spark.scheduler.cluster.mesos.MesosSchedulerBackend$$anon$1.run()V+44
>> v  ~StubRoutines::call_stub
>> V  [libjvm.so+0x5f8485]  JavaCalls::call_helper(JavaValue*, 
>> methodHandle*, JavaCallArguments*, Thread*)+0x365
>> V  [libjvm.so+0x5f6ee8]  JavaCalls::call(JavaValue*, methodHandle, 
>> JavaCallArguments*, Thread*)+0x28
>> V  [libjvm.so+0x5f71b7]  JavaCalls::call_virtual(JavaValue*, KlassHandle, 
>> Symbol*, Symbol*, JavaCallArguments*, Thread*)+0x197
>> V  [libjvm.so+0x5f72d7]  JavaCalls::call_virtual(JavaValue*, Handle, 
>> KlassHandle, Symbol*, Symbol*, Thread*)+0x47
>> V  [libjvm.so+0x6731e5]  thread_entry(JavaThread*, Thread*)+0xe5
>> V  [libjvm.so+0x94d38f]  JavaThread::thread_main_inner()+0xdf
>> V  [libjvm.so+0x94d495]  JavaThread::run()+0xf5
>> V  [libjvm.so+0x815288]  java_start(Thread*)+0x108
>>
>>
>> What I am trying to have
>>
>> Spark 0.8.1 
>> Mesos 0.14.2
>> HDFS 2.2.0 (I do not care about yarn or hadoop mapred since I am using 
>> mesos).
>> Oracle Java 1.7.0-45
>>
>> Here are the 4 options I have tried for spark.
>>
>> SPARK_HADOOP_VERSION=2.2.0 sbt/sbt assembly
>> and
>> SPARK_HADOOP_VERSION=2.2.0 SPARK_YARN=true sbt/sbt assembly
>>
>> then
>>
>> make-distribution.sh --hadoop 2.2.0 --with-yarn
>> and 
>> make-distribution.sh --hadoop 2.2.0
>>
>>
>>
>> Since all of those options are built with protobuf 2.5.0
>>
>> I've rebuild mesos 0.14.2 using protobuf 2.5.0
>>
>> The error I am having still seems to be related to protobuf seriously do 
>> not know how to try to debug that. All my modules are now using protobuf 
>> 2.5.0.
>>
>>
>> Any ideas?
>>
>>  -- 
>> You received this message because you are subscribed to the Google Groups 
>> "Spark Users" group.
>> To unsubscribe from this group and stop receiving emails from it, send an 
>> email to [email protected] <javascript:>.
>> For more options, visit https://groups.google.com/groups/opt_out.
>>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Spark Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/groups/opt_out.

Reply via email to