Oh, in that case you should mention 2.4, If you don't want to compile spark, then you can download the precompiled version from Downloads page <https://spark.apache.org/downloads.html>. http://d3kbcqa49mib13.cloudfront.net/spark-1.2.0-bin-hadoop2.4.tgz
Thanks Best Regards On Wed, Mar 25, 2015 at 5:40 PM, sandeep vura <sandeepv...@gmail.com> wrote: > *I am using hadoop 2.4 should i mention -Dhadoop.version=2.2* > > *$ hadoop version* > *Hadoop 2.4.1* > *Subversion http://svn.apache.org/repos/asf/hadoop/common > <http://svn.apache.org/repos/asf/hadoop/common> -r 1604318* > *Compiled by jenkins on 2014-06-21T05:43Z* > *Compiled with protoc 2.5.0* > *From source with checksum bb7ac0a3c73dc131f4844b873c74b630* > *This command was run using > /home/hadoop24/hadoop-2.4.1/share/hadoop/common/hadoop-common-2.4.1.jar* > > > > > On Wed, Mar 25, 2015 at 5:38 PM, Akhil Das <ak...@sigmoidanalytics.com> > wrote: > >> -D*hadoop.version=2.2* >> >> >> Thanks >> Best Regards >> >> On Wed, Mar 25, 2015 at 5:34 PM, sandeep vura <sandeepv...@gmail.com> >> wrote: >> >>> Build failed with following errors. >>> >>> I have executed the below following command. >>> >>> * mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=VERSION -DskipTests clean >>> package* >>> >>> >>> [INFO] >>> ------------------------------------------------------------------------ >>> [INFO] BUILD FAILURE >>> [INFO] >>> ------------------------------------------------------------------------ >>> [INFO] Total time: 2:11:59.461s >>> [INFO] Finished at: Wed Mar 25 17:22:29 IST 2015 >>> [INFO] Final Memory: 30M/440M >>> [INFO] >>> ------------------------------------------------------------------------ >>> [ERROR] Failed to execute goal on project spark-core_2.10: Could not >>> resolve dep >>> endencies for project >>> org.apache.spark:spark-core_2.10:jar:1.2.1: Could not find >>> >>> artifact org.apache.hadoop:hadoop-client:jar:VERSION in central ( >>> https://repo1. >>> maven.org/maven2) -> [Help 1] >>> [ERROR] >>> [ERROR] To see the full stack trace of the errors, re-run Maven with the >>> -e swit >>> ch. >>> [ERROR] Re-run Maven using the -X switch to enable full debug logging. >>> [ERROR] >>> [ERROR] For more information about the errors and possible solutions, >>> please rea >>> d the following articles: >>> [ERROR] [Help 1] >>> http://cwiki.apache.org/confluence/display/MAVEN/DependencyReso >>> >>> lutionException >>> [ERROR] >>> [ERROR] After correcting the problems, you can resume the build with the >>> command >>> [ERROR] mvn <goals> -rf :spark-core_2.10 >>> >>> >>> On Wed, Mar 25, 2015 at 3:38 PM, Akhil Das <ak...@sigmoidanalytics.com> >>> wrote: >>> >>>> Just run : >>>> >>>> mvn -Pyarn -Phadoop-2.4 -D*hadoop.version=2.2* -DskipTests clean package >>>> >>>> >>>> >>>> >>>> Thanks >>>> Best Regards >>>> >>>> On Wed, Mar 25, 2015 at 3:08 PM, sandeep vura <sandeepv...@gmail.com> >>>> wrote: >>>> >>>>> Where do i export MAVEN_OPTS in spark-env.sh or hadoop-env.sh >>>>> >>>>> I am running the below command in spark/yarn directory where pom.xml >>>>> file is available >>>>> >>>>> mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=VERSION -DskipTests clean package >>>>> >>>>> Please correct me if i am wrong. >>>>> >>>>> >>>>> >>>>> >>>>> On Wed, Mar 25, 2015 at 12:55 PM, Saisai Shao <sai.sai.s...@gmail.com> >>>>> wrote: >>>>> >>>>>> Looks like you have to build Spark with related Hadoop version, >>>>>> otherwise you will meet exception as mentioned. you could follow this >>>>>> doc: >>>>>> http://spark.apache.org/docs/latest/building-spark.html >>>>>> >>>>>> 2015-03-25 15:22 GMT+08:00 sandeep vura <sandeepv...@gmail.com>: >>>>>> >>>>>>> Hi Sparkers, >>>>>>> >>>>>>> I am trying to load data in spark with the following command >>>>>>> >>>>>>> *sqlContext.sql("LOAD DATA LOCAL INPATH >>>>>>> '/home/spark12/sandeep/sandeep.txt ' INTO TABLE src");* >>>>>>> >>>>>>> *Getting exception below* >>>>>>> >>>>>>> >>>>>>> *Server IPC version 9 cannot communicate with client version 4* >>>>>>> >>>>>>> NOte : i am using Hadoop 2.2 version and spark 1.2 and hive 0.13 >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >