You are receiving this mail as a port that you maintain is failing to build on the FreeBSD package build server. Please investigate the failure and submit a PR to fix build.
Maintainer: [email protected] Last committer: [email protected] Ident: $FreeBSD: head/devel/hadoop2/Makefile 379068 2015-02-16 07:46:56Z demon $ Log URL: http://beefy1.isc.freebsd.org/data/93i386-default/2015-02-26_02h13m17s/logs/hadoop2-2.4.1_5.log Build URL: http://beefy1.isc.freebsd.org/build.html?mastername=93i386-default&build=2015-02-26_02h13m17s Log: ====>> Building devel/hadoop2 build started at Thu Feb 26 08:57:46 UTC 2015 port directory: /usr/ports/devel/hadoop2 building for: FreeBSD 93i386-default-job-09 9.3-RELEASE-p10 FreeBSD 9.3-RELEASE-p10 i386 maintained by: [email protected] Makefile ident: $FreeBSD: head/devel/hadoop2/Makefile 379068 2015-02-16 07:46:56Z demon $ Poudriere version: 3.1.1 Host OSVERSION: 1100060 Jail OSVERSION: 903000 ---Begin Environment--- UNAME_m=i386 UNAME_p=i386 OSVERSION=903000 UNAME_v=FreeBSD 9.3-RELEASE-p10 UNAME_r=9.3-RELEASE-p10 BLOCKSIZE=K MAIL=/var/mail/root STATUS=1 SAVED_TERM= MASTERMNT=/usr/local/poudriere/data/.m/93i386-default/ref PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/games:/usr/local/sbin:/usr/local/bin:/root/bin POUDRIERE_BUILD_TYPE=bulk PKGNAME=hadoop2-2.4.1_5 OLDPWD=/ PWD=/usr/local/poudriere/data/.m/93i386-default/ref/.p/pool MASTERNAME=93i386-default USER=root HOME=/root POUDRIERE_VERSION=3.1.1 LOCALBASE=/usr/local PACKAGE_BUILDING=yes ---End Environment--- ---Begin OPTIONS List--- ---End OPTIONS List--- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- HADOOP_PROTOC_PATH=/usr/local/protobuf25/bin/protoc XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" NO_PIE=yes SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local LIBDIR="/usr/lib" CC="cc" CFLAGS="-O2 -pipe -fno-strict-aliasing" CPP="cpp" CPPFLAGS="" LDFLAGS="" LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fno-strict-aliasing" MANPREFIX="/usr/local" BSD_INSTALL_PROGRAM="install -s -o root -g wheel -m 555" BSD_INSTALL_LIB="install -s -o root -g wheel -m 444" BSD_INSTALL_SCRIPT="install -o root -g wheel -m 555" BSD_INSTALL_DATA="install -o root -g wheel -m 0644" BSD_INSTALL_MAN="install -o root -g wheel -m 444" --End MAKE_ENV-- --PLIST_SUB-- PORTVERSION="2.4.1" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" JAVASHAREDIR="share/java" JAVAJARDIR="share/java/classes" OSREL=9.3 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local PORTDOCS="" PORTEXAMPLES="" LIB32DIR=lib DOCSDIR="share/doc/hadoop" EXAMPLESDIR="share/examples/hadoop" DATADIR="share/hadoop" WWWDIR="www/hadoop" ETCDIR="etc/hadoop" --End PLIST_SUB-- --SUB_LIST-- HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" JAVA_HOME="/usr/local/openjdk7" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" JAVASHAREDIR="/usr/local/share/java" JAVAJARDIR="/usr/local/share/java/classes" JAVALIBDIR="/usr/local/share/java/classes" JAVA_VERSION="1.7+" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/hadoop DOCSDIR=/usr/local/share/doc/hadoop EXAMPLESDIR=/usr/local/share/examples/hadoop WWWDIR=/usr/local/www/hadoop ETCDIR=/usr/local/etc/hadoop --End SUB_LIST-- ---Begin make.conf--- ARCH=i386 MACHINE=i386 MACHINE_ARCH=i386 USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles #### /usr/local/etc/poudriere.d/make.conf #### WITH_PKGNG=yes #WITH_PKGNG=devel # clean-restricted via poudriere.conf NO_RESTRICTED #NO_RESTRICTED=yes DISABLE_MAKE_JOBS=poudriere ---End make.conf--- =======================<phase: check-sanity >============================ ===> License APACHE20 accepted by the user =========================================================================== =======================<phase: pkg-depends >============================ ===> hadoop2-2.4.1_5 depends on file: /usr/local/sbin/pkg - not found ===> Verifying install for /usr/local/sbin/pkg in /usr/ports/ports-mgmt/pkg ===> Installing existing package /packages/All/pkg-1.4.12.txz [93i386-default-job-09] Installing pkg-1.4.12... [93i386-default-job-09] Extracting pkg-1.4.12... done Message for pkg-1.4.12: If you are upgrading from the old package format, first run: # pkg2ng ===> Returning to build of hadoop2-2.4.1_5 =========================================================================== =======================<phase: fetch-depends >============================ =========================================================================== =======================<phase: fetch >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.4.1_5 for building =========================================================================== =======================<phase: checksum >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.4.1_5 for building => SHA256 Checksum OK for hadoop/hadoop-2.4.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.4.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.36.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =======================<phase: extract-depends>============================ =========================================================================== =======================<phase: extract >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.4.1_5 for building ===> Extracting for hadoop2-2.4.1_5 => SHA256 Checksum OK for hadoop/hadoop-2.4.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.4.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.36.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =======================<phase: patch-depends >============================ =========================================================================== =======================<phase: patch >============================ ===> Patching for hadoop2-2.4.1_5 ===> Applying FreeBSD patches for hadoop2-2.4.1_5 /usr/bin/sed -i.bak -e "s#/bin/bash#/usr/local/bin/bash#" /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DefaultContainerExecutor.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/launcher/ContainerLaunch.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Shell.java =========================================================================== =======================<phase: build-depends >============================ ===> hadoop2-2.4.1_5 depends on file: /usr/local/share/java/maven3/bin/mvn - not found ===> Verifying install for /usr/local/share/java/maven3/bin/mvn in /usr/ports/devel/maven3 ===> Installing existing package /packages/All/maven3-3.0.5.txz [93i386-default-job-09] Installing maven3-3.0.5... [93i386-default-job-09] `-- Installing maven-wrapper-1_2... [93i386-default-job-09] `-- Extracting maven-wrapper-1_2... done [93i386-default-job-09] `-- Installing openjdk-7.76.13_1,1... [93i386-default-job-09] | `-- Installing alsa-lib-1.0.28... [93i386-default-job-09] | `-- Extracting alsa-lib-1.0.28... done [93i386-default-job-09] | `-- Installing dejavu-2.34_4... [93i386-default-job-09] | `-- Extracting dejavu-2.34_4... done [93i386-default-job-09] | `-- Installing fontconfig-2.11.1,1... [93i386-default-job-09] | | `-- Installing expat-2.1.0_2... [93i386-default-job-09] | | `-- Extracting expat-2.1.0_2... done [93i386-default-job-09] | | `-- Installing freetype2-2.5.5... [93i386-default-job-09] | | `-- Extracting freetype2-2.5.5... done [93i386-default-job-09] | `-- Extracting fontconfig-2.11.1,1... done Running fc-cache to build fontconfig cache... /usr/local/share/fonts: skipping, no such directory /usr/local/lib/X11/fonts: caching, new cache contents: 0 fonts, 1 dirs /usr/local/lib/X11/fonts/dejavu: caching, new cache contents: 21 fonts, 0 dirs /root/.local/share/fonts: skipping, no such directory /root/.fonts: skipping, no such directory Re-scanning /usr/local/lib/X11/fonts: caching, new cache contents: 0 fonts, 1 dirs /var/db/fontconfig: cleaning cache directory /root/.cache/fontconfig: not cleaning non-existent cache directory /root/.fontconfig: not cleaning non-existent cache directory fc-cache: succeeded [93i386-default-job-09] | `-- Installing java-zoneinfo-2014.j... [93i386-default-job-09] | `-- Extracting java-zoneinfo-2014.j... done [93i386-default-job-09] | `-- Installing javavmwrapper-2.5... [93i386-default-job-09] | `-- Extracting javavmwrapper-2.5... done [93i386-default-job-09] | `-- Installing libX11-1.6.2_2,1... [93i386-default-job-09] | | `-- Installing kbproto-1.0.6... [93i386-default-job-09] | | `-- Extracting kbproto-1.0.6... done [93i386-default-job-09] | | `-- Installing libXau-1.0.8_2... [93i386-default-job-09] | | `-- Installing xproto-7.0.26... [93i386-default-job-09] | | `-- Extracting xproto-7.0.26... done [93i386-default-job-09] | | `-- Extracting libXau-1.0.8_2... done [93i386-default-job-09] | | `-- Installing libXdmcp-1.1.1_2... [93i386-default-job-09] | | `-- Extracting libXdmcp-1.1.1_2... done [93i386-default-job-09] | | `-- Installing libxcb-1.11... <snip> Building index for all classes... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-frame.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-noframe.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/index.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/help-doc.html... [INFO] Building jar: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/hadoop-auth-examples-2.4.1-javadoc.jar [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Apache Hadoop Common 2.4.1 [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-common --- [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-os) @ hadoop-common --- [INFO] [INFO] --- maven-antrun-plugin:1.7:run (create-testdirs) @ hadoop-common --- [INFO] Executing tasks main: [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/target/test-dir [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/target/test/data [INFO] Executed tasks [INFO] [INFO] --- hadoop-maven-plugins:2.4.1:protoc (compile-protoc) @ hadoop-common --- [INFO] [INFO] --- hadoop-maven-plugins:2.4.1:version-info (version-info) @ hadoop-common --- [WARNING] [svn, info] failed: java.io.IOException: Cannot run program "svn": error=2, No such file or directory [WARNING] [git, branch] failed: java.io.IOException: Cannot run program "git": error=2, No such file or directory [INFO] SCM: NONE [INFO] Computed MD5: 768816c4bfa6a3e566a5389e1a8232 [INFO] [INFO] --- maven-resources-plugin:2.2:resources (default-resources) @ hadoop-common --- [INFO] Using default encoding to copy filtered resources. [INFO] [INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hadoop-common --- [INFO] Compiling 741 source files to /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/target/classes [INFO] ------------------------------------------------------------- [ERROR] COMPILATION ERROR : [INFO] ------------------------------------------------------------- [ERROR] Failure executing javac, but could not parse the error: The system is out of resources. Consult the following stack trace for details. java.lang.OutOfMemoryError: Java heap space at com.sun.tools.javac.util.Position$LineMapImpl.build(Position.java:153) at com.sun.tools.javac.util.Position.makeLineMap(Position.java:77) at com.sun.tools.javac.parser.Scanner.getLineMap(Scanner.java:1147) at com.sun.tools.javac.parser.JavacParser.parseCompilationUnit(JavacParser.java:2310) at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:595) at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:632) at com.sun.tools.javac.main.JavaCompiler.parseFiles(JavaCompiler.java:909) at com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:824) at com.sun.tools.javac.main.Main.compile(Main.java:439) at com.sun.tools.javac.main.Main.compile(Main.java:353) at com.sun.tools.javac.main.Main.compile(Main.java:342) at com.sun.tools.javac.main.Main.compile(Main.java:333) at com.sun.tools.javac.Main.compile(Main.java:94) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess0(JavacCompiler.java:551) at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess(JavacCompiler.java:526) at org.codehaus.plexus.compiler.javac.JavacCompiler.compile(JavacCompiler.java:167) at org.apache.maven.plugin.AbstractCompilerMojo.execute(AbstractCompilerMojo.java:678) at org.apache.maven.plugin.CompilerMojo.execute(CompilerMojo.java:128) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:101) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:209) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:84) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:59) at org.apache.maven.lifecycle.internal.LifecycleStarter.singleThreadedBuild(LifecycleStarter.java:183) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:161) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:320) at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:156) [INFO] 1 error [INFO] ------------------------------------------------------------- [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop Main ................................ SUCCESS [1.703s] [INFO] Apache Hadoop Project POM ......................... SUCCESS [1.387s] [INFO] Apache Hadoop Annotations ......................... SUCCESS [5.384s] [INFO] Apache Hadoop Assemblies .......................... SUCCESS [0.418s] [INFO] Apache Hadoop Project Dist POM .................... SUCCESS [2.535s] [INFO] Apache Hadoop Maven Plugins ....................... SUCCESS [6.930s] [INFO] Apache Hadoop MiniKDC ............................. SUCCESS [6.969s] [INFO] Apache Hadoop Auth ................................ SUCCESS [7.576s] [INFO] Apache Hadoop Auth Examples ....................... SUCCESS [4.269s] [INFO] Apache Hadoop Common .............................. FAILURE [6.807s] [INFO] Apache Hadoop NFS ................................. SKIPPED [INFO] Apache Hadoop Common Project ...................... SKIPPED [INFO] Apache Hadoop HDFS ................................ SKIPPED [INFO] Apache Hadoop HttpFS .............................. SKIPPED [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SKIPPED [INFO] Apache Hadoop HDFS-NFS ............................ SKIPPED [INFO] Apache Hadoop HDFS Project ........................ SKIPPED [INFO] hadoop-yarn ....................................... SKIPPED [INFO] hadoop-yarn-api ................................... SKIPPED [INFO] hadoop-yarn-common ................................ SKIPPED [INFO] hadoop-yarn-server ................................ SKIPPED [INFO] hadoop-yarn-server-common ......................... SKIPPED [INFO] hadoop-yarn-server-nodemanager .................... SKIPPED [INFO] hadoop-yarn-server-web-proxy ...................... SKIPPED [INFO] hadoop-yarn-server-applicationhistoryservice ...... SKIPPED [INFO] hadoop-yarn-server-resourcemanager ................ SKIPPED [INFO] hadoop-yarn-server-tests .......................... SKIPPED [INFO] hadoop-yarn-client ................................ SKIPPED [INFO] hadoop-yarn-applications .......................... SKIPPED [INFO] hadoop-yarn-applications-distributedshell ......... SKIPPED [INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SKIPPED [INFO] hadoop-yarn-site .................................. SKIPPED [INFO] hadoop-yarn-project ............................... SKIPPED [INFO] hadoop-mapreduce-client ........................... SKIPPED [INFO] hadoop-mapreduce-client-core ...................... SKIPPED [INFO] hadoop-mapreduce-client-common .................... SKIPPED [INFO] hadoop-mapreduce-client-shuffle ................... SKIPPED [INFO] hadoop-mapreduce-client-app ....................... SKIPPED [INFO] hadoop-mapreduce-client-hs ........................ SKIPPED [INFO] hadoop-mapreduce-client-jobclient ................. SKIPPED [INFO] hadoop-mapreduce-client-hs-plugins ................ SKIPPED [INFO] Apache Hadoop MapReduce Examples .................. SKIPPED [INFO] hadoop-mapreduce .................................. SKIPPED [INFO] Apache Hadoop MapReduce Streaming ................. SKIPPED [INFO] Apache Hadoop Distributed Copy .................... SKIPPED [INFO] Apache Hadoop Archives ............................ SKIPPED [INFO] Apache Hadoop Rumen ............................... SKIPPED [INFO] Apache Hadoop Gridmix ............................. SKIPPED [INFO] Apache Hadoop Data Join ........................... SKIPPED [INFO] Apache Hadoop Extras .............................. SKIPPED [INFO] Apache Hadoop Pipes ............................... SKIPPED [INFO] Apache Hadoop OpenStack support ................... SKIPPED [INFO] Apache Hadoop Client .............................. SKIPPED [INFO] Apache Hadoop Mini-Cluster ........................ SKIPPED [INFO] Apache Hadoop Scheduler Load Simulator ............ SKIPPED [INFO] Apache Hadoop Tools Dist .......................... SKIPPED [INFO] Apache Hadoop Tools ............................... SKIPPED [INFO] Apache Hadoop Distribution ........................ SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 46.722s [INFO] Finished at: Thu Feb 26 08:58:59 GMT 2015 [INFO] Final Memory: 37M/42M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:2.5.1:compile (default-compile) on project hadoop-common: Compilation failure [ERROR] Failure executing javac, but could not parse the error: [ERROR] [ERROR] [ERROR] The system is out of resources. [ERROR] Consult the following stack trace for details. [ERROR] java.lang.OutOfMemoryError: Java heap space [ERROR] at com.sun.tools.javac.util.Position$LineMapImpl.build(Position.java:153) [ERROR] at com.sun.tools.javac.util.Position.makeLineMap(Position.java:77) [ERROR] at com.sun.tools.javac.parser.Scanner.getLineMap(Scanner.java:1147) [ERROR] at com.sun.tools.javac.parser.JavacParser.parseCompilationUnit(JavacParser.java:2310) [ERROR] at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:595) [ERROR] at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:632) [ERROR] at com.sun.tools.javac.main.JavaCompiler.parseFiles(JavaCompiler.java:909) [ERROR] at com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:824) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:439) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:353) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:342) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:333) [ERROR] at com.sun.tools.javac.Main.compile(Main.java:94) [ERROR] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) [ERROR] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) [ERROR] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) [ERROR] at java.lang.reflect.Method.invoke(Method.java:606) [ERROR] at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess0(JavacCompiler.java:551) [ERROR] at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess(JavacCompiler.java:526) [ERROR] at org.codehaus.plexus.compiler.javac.JavacCompiler.compile(JavacCompiler.java:167) [ERROR] at org.apache.maven.plugin.AbstractCompilerMojo.execute(AbstractCompilerMojo.java:678) [ERROR] at org.apache.maven.plugin.CompilerMojo.execute(CompilerMojo.java:128) [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:101) [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:209) [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153) [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:84) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:59) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleStarter.singleThreadedBuild(LifecycleStarter.java:183) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:161) [ERROR] at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:320) [ERROR] at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:156) [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-common *** [do-build] Error code 1 Stop in /usr/ports/devel/hadoop2. _______________________________________________ [email protected] mailing list https://lists.freebsd.org/mailman/listinfo/freebsd-pkg-fallout To unsubscribe, send any mail to "[email protected]"
