Thanks for pointing me to that mistake . Yes i was using the spark 0.8.1
incubating jar and the master branch code examples . I corrected the mistake

Regards


On Wed, Jan 15, 2014 at 5:51 PM, Patrick Wendell <pwend...@gmail.com> wrote:

> Hm, are you sure you haven't included the master branch of Spark
> somehow in your classpath? jarOfClass was added to Java in the master
> branch and Spark 0.9.0 (RC). So it seems a lot like you have a newer
> (post 0.8.X) version of the examples.
>
> - Patrick
>
> On Wed, Jan 15, 2014 at 5:04 PM, arjun biswas <arjunbiswas....@gmail.com>
> wrote:
> > Could it be possible that you have an older version of JavaSparkContext
> > (i.e. from an older version of Spark) in your path? Please check that
> there
> > aren't two versions of Spark accidentally included in your class path
> used
> > in Eclipse. It would not give errors in the import (as it finds the
> imported
> > packages and classes) but would give such errors as it may be
> unfortunately
> > finding an older version of JavaSparkContext class in the class path.
> >
> >>>
> >
> > I have the following three jars in the class path of eclipse .and no
> other
> > jar is currently in the classpath
> > 1)google-collections-0.8.jar
> > 2)scala-library.jar
> > 3)spark-core_2.9.3-0.8.1-incubating.jar
> >
> > Am i using the correct jar files to run the java samples from eclipse ?
> >
> > Regards
> >
> >
> >
> >
> > On Wed, Jan 15, 2014 at 4:36 PM, Tathagata Das <
> tathagata.das1...@gmail.com>
> > wrote:
> >>
> >> Could it be possible that you have an older version of JavaSparkContext
> >> (i.e. from an older version of Spark) in your path? Please check that
> there
> >> aren't two versions of Spark accidentally included in your class path
> used
> >> in Eclipse. It would not give errors in the import (as it finds the
> imported
> >> packages and classes) but would give such errors as it may be
> unfortunately
> >> finding an older version of JavaSparkContext class in the class path.
> >>
> >> TD
> >>
> >>
> >> On Wed, Jan 15, 2014 at 4:14 PM, arjun biswas <
> arjunbiswas....@gmail.com>
> >> wrote:
> >>>
> >>> Hello All ,
> >>>
> >>> I have installed spark on my machine and was succesful in running
> sbt/sbt
> >>> package as well as sbt/sbt assembly . I am trying to run the examples
> in
> >>> java from eclipse . To be precise i am trying to run the JavaLogQuery
> >>> example from eclipse . The issue is i am unable to resolve this
> compilation
> >>> problem of jarOfClass being not available inside the Java Spark
> Context . I
> >>> have added all the possible jars and is using Spark 0.8.1 incubating
> which
> >>> is the latest one with scala 2.9.3 .I have all jars to the classpath
> to the
> >>> point that i do not get any import error . However
> >>> JavaSparkContext.jarOfClass gives the above error saying the jarOfClass
> >>> method is unavailable in the JavaSparkContext . I am using Spark-0.8.1
> >>> incubating and scala 2.9.3 . Has anyone tried to run the java sample
> >>> examples from eclipse . Please note that this is a compile time error
> in
> >>> eclipse .
> >>>
> >>> Regards
> >>> Arjun
> >>
> >>
> >
>

Reply via email to