*I am using hadoop 2.4 should i mention -Dhadoop.version=2.2* *$ hadoop version* *Hadoop 2.4.1* *Subversion http://svn.apache.org/repos/asf/hadoop/common <http://svn.apache.org/repos/asf/hadoop/common> -r 1604318* *Compiled by jenkins on 2014-06-21T05:43Z* *Compiled with protoc 2.5.0* *From source with checksum bb7ac0a3c73dc131f4844b873c74b630* *This command was run using /home/hadoop24/hadoop-2.4.1/share/hadoop/common/hadoop-common-2.4.1.jar*
On Wed, Mar 25, 2015 at 5:38 PM, Akhil Das <ak...@sigmoidanalytics.com> wrote: > -D*hadoop.version=2.2* > > > Thanks > Best Regards > > On Wed, Mar 25, 2015 at 5:34 PM, sandeep vura <sandeepv...@gmail.com> > wrote: > >> Build failed with following errors. >> >> I have executed the below following command. >> >> * mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=VERSION -DskipTests clean >> package* >> >> >> [INFO] >> ------------------------------------------------------------------------ >> [INFO] BUILD FAILURE >> [INFO] >> ------------------------------------------------------------------------ >> [INFO] Total time: 2:11:59.461s >> [INFO] Finished at: Wed Mar 25 17:22:29 IST 2015 >> [INFO] Final Memory: 30M/440M >> [INFO] >> ------------------------------------------------------------------------ >> [ERROR] Failed to execute goal on project spark-core_2.10: Could not >> resolve dep >> endencies for project >> org.apache.spark:spark-core_2.10:jar:1.2.1: Could not find >> >> artifact org.apache.hadoop:hadoop-client:jar:VERSION in central ( >> https://repo1. >> maven.org/maven2) -> [Help 1] >> [ERROR] >> [ERROR] To see the full stack trace of the errors, re-run Maven with the >> -e swit >> ch. >> [ERROR] Re-run Maven using the -X switch to enable full debug logging. >> [ERROR] >> [ERROR] For more information about the errors and possible solutions, >> please rea >> d the following articles: >> [ERROR] [Help 1] >> http://cwiki.apache.org/confluence/display/MAVEN/DependencyReso >> >> lutionException >> [ERROR] >> [ERROR] After correcting the problems, you can resume the build with the >> command >> [ERROR] mvn <goals> -rf :spark-core_2.10 >> >> >> On Wed, Mar 25, 2015 at 3:38 PM, Akhil Das <ak...@sigmoidanalytics.com> >> wrote: >> >>> Just run : >>> >>> mvn -Pyarn -Phadoop-2.4 -D*hadoop.version=2.2* -DskipTests clean package >>> >>> >>> >>> >>> Thanks >>> Best Regards >>> >>> On Wed, Mar 25, 2015 at 3:08 PM, sandeep vura <sandeepv...@gmail.com> >>> wrote: >>> >>>> Where do i export MAVEN_OPTS in spark-env.sh or hadoop-env.sh >>>> >>>> I am running the below command in spark/yarn directory where pom.xml >>>> file is available >>>> >>>> mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=VERSION -DskipTests clean package >>>> >>>> Please correct me if i am wrong. >>>> >>>> >>>> >>>> >>>> On Wed, Mar 25, 2015 at 12:55 PM, Saisai Shao <sai.sai.s...@gmail.com> >>>> wrote: >>>> >>>>> Looks like you have to build Spark with related Hadoop version, >>>>> otherwise you will meet exception as mentioned. you could follow this doc: >>>>> http://spark.apache.org/docs/latest/building-spark.html >>>>> >>>>> 2015-03-25 15:22 GMT+08:00 sandeep vura <sandeepv...@gmail.com>: >>>>> >>>>>> Hi Sparkers, >>>>>> >>>>>> I am trying to load data in spark with the following command >>>>>> >>>>>> *sqlContext.sql("LOAD DATA LOCAL INPATH >>>>>> '/home/spark12/sandeep/sandeep.txt ' INTO TABLE src");* >>>>>> >>>>>> *Getting exception below* >>>>>> >>>>>> >>>>>> *Server IPC version 9 cannot communicate with client version 4* >>>>>> >>>>>> NOte : i am using Hadoop 2.2 version and spark 1.2 and hive 0.13 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >> >