Hi Praveen/Ravi, Thanks for all the help. It built successfully.
I'm trying to get a feel about the project structure, build structure and map reduce. Is there a document I can read to understand about the use of native code and where it fits with the mapreduce project? On Fri, Aug 26, 2011 at 10:00 AM, Ravi Teja <ravit...@huawei.com> wrote: > Hi Tharindu, > > I think it is trying to compile the native code. you can add -P-cbuild as > argument to skip it, as mentioned earlier by Arun. > > Regards, > Ravi Teja > > > Thanks Praveen. > > I managed to proceed further. Now I'm stuck at this point. Appreciate if > you > can tell me what I'm doing wrong. > > Stacktrace: > > [INFO] --- make-maven-plugin:1.0-beta-1:configure (make) @ > hadoop-yarn-server-nodemanager --- > [INFO] checking for a BSD-compatible install... /usr/bin/install -c > [INFO] checking whether build environment is sane... yes > [INFO] checking for a thread-safe mkdir -p... ./install-sh -c -d > [INFO] checking for gawk... no > [INFO] checking for mawk... no > [INFO] checking for nawk... no > [INFO] checking for awk... awk > [INFO] checking whether make sets $(MAKE)... yes > [INFO] ./configure: line 2226: CHECK_INSTALL_CFLAG: command not found > [INFO] ./configure: line 2227: HADOOP_UTILS_SETUP: command not found > [INFO] checking for gcc... gcc > [INFO] checking for C compiler default output file name... a.out > [INFO] checking whether the C compiler works... yes > [INFO] checking whether we are cross compiling... no > [INFO] checking for suffix of executables... > [INFO] checking for suffix of object files... o > [INFO] checking whether we are using the GNU C compiler... yes > [INFO] checking whether gcc accepts -g... yes > [INFO] checking for gcc option to accept ISO C89... none needed > [INFO] checking for style of include used by make... GNU > [INFO] checking dependency style of gcc... gcc3 > [INFO] checking whether gcc and cc understand -c and -o together... yes > [INFO] checking how to run the C preprocessor... gcc -E > [INFO] checking for grep that handles long lines and -e... /usr/bin/grep > [INFO] checking for egrep... /usr/bin/grep -E > [INFO] checking for ANSI C header files... yes > [INFO] checking for sys/types.h... yes > [INFO] checking for sys/stat.h... yes > [INFO] checking for stdlib.h... yes > [INFO] checking for string.h... yes > [INFO] checking for memory.h... yes > [INFO] checking for strings.h... yes > [INFO] checking for inttypes.h... yes > [INFO] checking for stdint.h... yes > [INFO] checking for unistd.h... yes > [INFO] checking for unistd.h... (cached) yes > [INFO] checking for stdbool.h that conforms to C99... yes > [INFO] checking for _Bool... yes > [INFO] checking for an ANSI C-conforming const... yes > [INFO] checking for off_t... yes > [INFO] checking for size_t... yes > [INFO] checking whether strerror_r is declared... yes > [INFO] checking for strerror_r... yes > [INFO] checking whether strerror_r returns char *... no > [INFO] checking for mkdir... yes > [INFO] checking for uname... yes > [INFO] configure: creating ./config.status > [INFO] config.status: creating Makefile > [INFO] config.status: executing depfiles commands > [INFO] > [INFO] --- make-maven-plugin:1.0-beta-1:make-install (install) @ > hadoop-yarn-server-nodemanager --- > [INFO] depbase=`echo impl/configuration.o | sed > 's|[^/]*$|.deps/&|;s|\.o$||'`;\ > [INFO] gcc -DPACKAGE_NAME=\"linux-container-executor\" > -DPACKAGE_TARNAME=\"linux-container-executor\" -DPACKAGE_VERSION=\"1.0.0\" > -DPACKAGE_STRING=\"linux-container-executor\ 1.0.0\" -DPACKAGE_BUGREPORT=\" > yarn-...@yahoo-inc.com\" -D_GNU_SOURCE=1 > -DPACKAGE=\"linux-container-executor\" -DVERSION=\"1.0.0\" -DSTDC_HEADERS=1 > -DHAVE_SYS_TYPES_H=1 -DHAVE_SYS_STAT_H=1 -DHAVE_STDLIB_H=1 > -DHAVE_STRING_H=1 > -DHAVE_MEMORY_H=1 -DHAVE_STRINGS_H=1 -DHAVE_INTTYPES_H=1 -DHAVE_STDINT_H=1 > -DHAVE_UNISTD_H=1 -DHAVE_UNISTD_H=1 -DHAVE__BOOL=1 -DHAVE_STDBOOL_H=1 > -DHAVE_DECL_STRERROR_R=1 -DHAVE_STRERROR_R=1 -DHAVE_MKDIR=1 -DHAVE_UNAME=1 > -I. -I./impl -Wall -g -Werror -DHADOOP_CONF_DIR= -MT > impl/configuration.o > -MD -MP -MF $depbase.Tpo -c -o impl/configuration.o impl/configuration.c > &&\ > [INFO] mv -f $depbase.Tpo $depbase.Po > [INFO] cc1: warnings being treated as errors > [INFO] impl/configuration.c: In function 'read_config': > [INFO] impl/configuration.c:144: warning: implicit declaration of function > 'getline' > [INFO] make: *** [impl/configuration.o] Error 1 > [INFO] > ------------------------------------------------------------------------ > [INFO] Reactor Summary: > [INFO] > [INFO] hadoop-yarn-api ................................... SUCCESS > [19.783s] > [INFO] hadoop-yarn-common ................................ SUCCESS > [15.172s] > [INFO] hadoop-yarn-server-common ......................... SUCCESS [7.966s] > [INFO] hadoop-yarn-server-nodemanager .................... FAILURE > [1:08.482s] > [INFO] hadoop-yarn-server-resourcemanager ................ SKIPPED > [INFO] hadoop-yarn-server-tests .......................... SKIPPED > [INFO] hadoop-yarn-server ................................ SKIPPED > [INFO] hadoop-yarn ....................................... SKIPPED > [INFO] hadoop-mapreduce-client-core ...................... SKIPPED > [INFO] hadoop-mapreduce-client-common .................... SKIPPED > [INFO] hadoop-mapreduce-client-shuffle ................... SKIPPED > [INFO] hadoop-mapreduce-client-app ....................... SKIPPED > [INFO] hadoop-mapreduce-client-hs ........................ SKIPPED > [INFO] hadoop-mapreduce-client-jobclient ................. SKIPPED > [INFO] hadoop-mapreduce-client ........................... SKIPPED > [INFO] hadoop-mapreduce .................................. SKIPPED > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD FAILURE > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 1:51.950s > [INFO] Finished at: Fri Aug 26 01:46:51 IST 2011 > [INFO] Final Memory: 27M/114M > [INFO] > ------------------------------------------------------------------------ > [ERROR] Failed to execute goal > org.codehaus.mojo:make-maven-plugin:1.0-beta-1:make-install (install) on > project hadoop-yarn-server-nodemanager: make returned an exit value != 0. > Aborting build; see command output above for more information. -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > > On Thu, Aug 25, 2011 at 5:19 PM, Praveen Sripati > <praveensrip...@gmail.com>wrote: > > > Tharindu, > > > > Looks like protoc is not available. > > > > --- > > Cannot run program "protoc" (in directory "HOME/hadoop-trunk/hadoop- > > mapreduce/hadoop-yarn/hadoop-yarn-api"): error=2, > > No such file or directory -> [Help 1] > > --- > > > > Here are instructions to build protoc > > > > See > > > > > > http://svn.apache.org/repos/asf/hadoop/common/trunk/hadoop-mapreduce/hadoop- > yarn/README > > Make sure protbuf library is in your library path or set: export > > LD_LIBRARY_PATH=/usr/local/lib > > > > Thanks, > > Praveen > > > > On Thu, Aug 25, 2011 at 3:43 PM, Tharindu Mathew <mcclou...@gmail.com > > >wrote: > > > > > Hi everyone, > > > > > > I'm a new bie to this list. Hope this is not too much of a dumb > question. > > > > > > I'm trying to build the map reduce trunk. (I already build from the > root > > > pom > > > and everything built fine, but map reduce is not included in the root > > pom) > > > > > > The build fails at this point given below. Appreciate if someone can > help > > > me > > > out. Thanks in advance. > > > > > > Stack trace: > > > > > > [INFO] > > > > ------------------------------------------------------------------------ > > > [INFO] Building hadoop-yarn-api 1.0-SNAPSHOT > > > [INFO] > > > > ------------------------------------------------------------------------ > > > [INFO] > > > [INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ > > hadoop-yarn-api > > > --- > > > [INFO] Deleting > > > > > > > > > > /Users/mackie/source-checkouts/hadoop-trunk/hadoop-mapreduce/hadoop-yarn/had > oop-yarn-api/target > > > [INFO] > > > [INFO] --- maven-antrun-plugin:1.6:run > > > (create-protobuf-generated-sources-directory) @ hadoop-yarn-api --- > > > [INFO] Executing tasks > > > > > > main: > > > [mkdir] Created dir: > > > > > > > > > > /Users/mackie/source-checkouts/hadoop-trunk/hadoop-mapreduce/hadoop-yarn/had > oop-yarn-api/target/generated-sources/proto > > > [INFO] Executed tasks > > > [INFO] > > > [INFO] --- exec-maven-plugin:1.2:exec (generate-sources) @ > > hadoop-yarn-api > > > --- > > > [INFO] > > > > ------------------------------------------------------------------------ > > > [INFO] Reactor Summary: > > > [INFO] > > > [INFO] hadoop-yarn-api ................................... FAILURE > > [2.027s] > > > [INFO] hadoop-yarn-common ................................ SKIPPED > > > [INFO] hadoop-yarn-server-common ......................... SKIPPED > > > [INFO] hadoop-yarn-server-nodemanager .................... SKIPPED > > > [INFO] hadoop-yarn-server-resourcemanager ................ SKIPPED > > > [INFO] hadoop-yarn-server-tests .......................... SKIPPED > > > [INFO] hadoop-yarn-server ................................ SKIPPED > > > [INFO] hadoop-yarn ....................................... SKIPPED > > > [INFO] hadoop-mapreduce-client-core ...................... SKIPPED > > > [INFO] hadoop-mapreduce-client-common .................... SKIPPED > > > [INFO] hadoop-mapreduce-client-shuffle ................... SKIPPED > > > [INFO] hadoop-mapreduce-client-app ....................... SKIPPED > > > [INFO] hadoop-mapreduce-client-hs ........................ SKIPPED > > > [INFO] hadoop-mapreduce-client-jobclient ................. SKIPPED > > > [INFO] hadoop-mapreduce-client ........................... SKIPPED > > > [INFO] hadoop-mapreduce .................................. SKIPPED > > > [INFO] > > > > ------------------------------------------------------------------------ > > > [INFO] BUILD FAILURE > > > [INFO] > > > > ------------------------------------------------------------------------ > > > [INFO] Total time: 2.606s > > > [INFO] Finished at: Thu Aug 25 15:33:02 IST 2011 > > > [INFO] Final Memory: 9M/80M > > > [INFO] > > > > ------------------------------------------------------------------------ > > > [ERROR] Failed to execute goal > > org.codehaus.mojo:exec-maven-plugin:1.2:exec > > > (generate-sources) on project hadoop-yarn-api: Command execution > failed. > > > Cannot run program "protoc" (in directory > > > "HOME/hadoop-trunk/hadoop-mapreduce/hadoop-yarn/hadoop-yarn-api"): > > error=2, > > > No such file or directory -> [Help 1] > > > org.apache.maven.lifecycle.LifecycleExecutionException: Failed to > execute > > > goal org.codehaus.mojo:exec-maven-plugin:1.2:exec (generate-sources) on > > > project hadoop-yarn-api: Command execution failed. > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:2 > 17) > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:1 > 53) > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:1 > 45) > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(Life > cycleModuleBuilder.java:84) > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(Life > cycleModuleBuilder.java:59) > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.LifecycleStarter.singleThreadedBuild(Lif > ecycleStarter.java:183) > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarte > r.java:161) > > > at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:319) > > > at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:156) > > > at org.apache.maven.cli.MavenCli.execute(MavenCli.java:537) > > > at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:196) > > > at org.apache.maven.cli.MavenCli.main(MavenCli.java:141) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at > > > > > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39 > ) > > > at > > > > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl > .java:25) > > > at java.lang.reflect.Method.invoke(Method.java:597) > > > at > > > > > > > > > > org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.ja > va:290) > > > at > > > > > > org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:230) > > > at > > > > > > > > > > org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher. > java:409) > > > at > > > > org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:352) > > > Caused by: org.apache.maven.plugin.MojoExecutionException: Command > > > execution > > > failed. > > > at org.codehaus.mojo.exec.ExecMojo.execute(ExecMojo.java:350) > > > at > > > > > > > > > > org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPl > uginManager.java:101) > > > at > > > > > > > > > > org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:2 > 09) > > > ... 19 more > > > Caused by: java.io.IOException: Cannot run program "protoc" (in > directory > > > > > > > > > > "/Users/mackie/source-checkouts/hadoop-trunk/hadoop-mapreduce/hadoop-yarn/ha > doop-yarn-api"): > > > error=2, No such file or directory > > > at java.lang.ProcessBuilder.start(ProcessBuilder.java:459) > > > at java.lang.Runtime.exec(Runtime.java:593) > > > at > > > > > > > > > > org.apache.commons.exec.launcher.Java13CommandLauncher.exec(Java13CommandLau > ncher.java:58) > > > at > > org.apache.commons.exec.DefaultExecutor.launch(DefaultExecutor.java:246) > > > at > > > > > > > > > > org.apache.commons.exec.DefaultExecutor.executeInternal(DefaultExecutor.java > :302) > > > at > > > > org.apache.commons.exec.DefaultExecutor.execute(DefaultExecutor.java:149) > > > at > org.codehaus.mojo.exec.ExecMojo.executeCommandLine(ExecMojo.java:589) > > > at org.codehaus.mojo.exec.ExecMojo.execute(ExecMojo.java:335) > > > ... 21 more > > > Caused by: java.io.IOException: error=2, No such file or directory > > > at java.lang.UNIXProcess.forkAndExec(Native Method) > > > at java.lang.UNIXProcess.<init>(UNIXProcess.java:53) > > > at java.lang.ProcessImpl.start(ProcessImpl.java:91) > > > at java.lang.ProcessBuilder.start(ProcessBuilder.java:452) > > > ... 28 more > > > [ERROR] > > > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > > > [ERROR] > > > [ERROR] For more information about the errors and possible solutions, > > > please > > > read the following articles: > > > [ERROR] [Help 1] > > > > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException > > > > > > -- > > > Regards, > > > > > > Tharindu > > > > > > > > > -- > Regards, > > Tharindu > > -- Regards, Tharindu