This command will give you the exact name: java org.apache.hadoop.util.PlatformName | sed -e "s/ /_/g"
Can you try to run it? But it's most probably Linux-amd64-64 2014-08-26 20:24 GMT-04:00 [email protected] < [email protected]>: > Hi, > > Thanks! > > A question: > If I run: > $ uname -m > x86_64 > > Should I use " lib/native/Linux-amd64-64” or "lib/native/x86_64” in > $HADOOP_HOME and $HBASE_HOME? > > Arthur > > > On 27 Aug, 2014, at 8:10 am, Jean-Marc Spaggiari <[email protected]> > wrote: > > > Ok. > > > > This is the way the lib path is built: > > > > JAVA_LIBRARY_PATH=$(append_path "$JAVA_LIBRARY_PATH" > > ${HBASE_HOME}/build/native/${JAVA_PLATFORM}/lib) > > > > And JAVA_PLATFORM comes from JAVA_PLATFORM=`CLASSPATH=${CLASSPATH} > ${JAVA} > > org.apache.hadoop.util.PlatformName | sed -e "s/ /_/g"` > > > > You can double check it doing: > > > > # Adjust to you java_home... > > export JAVA_HOME=/usr/local/jdk1.7.0_45/ > > > > export CLASSPATH=`bin/hbase classpath` > > $JAVA_HOME/bin/java org.apache.hadoop.util.PlatformName | sed -e "s/ > /_/g" > > > > Result for me is this: Linux-amd64-64. Might be different for you. > > > > Then you link the libs the way Alex said before: > > cd lib/native/Linux-amd64-64 > > ln -s /home/hbase/snappy-1.0.5/.libs/libsnappy.so . > > ln -s /home/hbase/snappy-1.0.5/.libs/libsnappy.so.1 . > > > > AND..... > > > > The hadoop so too! And I think this is what's missing for you: > > ln -s /YOURHADOOPPATH/libhadoop.so . > > > > Your folder should look like this: > > jmspaggi@node8:~/hbase-0.98.5-hadoop2/lib/native$ tree > > . > > └── Linux-amd64-64 > > ├── libhadoop.so > > ├── libsnappy.so -> /home/hbase/snappy-1.0.5/.libs/libsnappy.so > > └── libsnappy.so.1 -> /home/hbase/snappy-1.0.5/.libs/libsnappy.so.1 > > > > I copied libhadoop.so instead of doing a link because it was not > available > > on this computer. > > > > Then test it: > > jmspaggi@node8:~/hbase-0.98.5-hadoop2$ bin/hbase > > org.apache.hadoop.hbase.util.CompressionTest file:///tmp/snappy-test > snappy > > 2014-08-26 20:06:43,987 INFO [main] Configuration.deprecation: > > hadoop.native.lib is deprecated. Instead, use io.native.lib.available > > 2014-08-26 20:06:44,831 INFO [main] util.ChecksumType: Checksum using > > org.apache.hadoop.util.PureJavaCrc32 > > 2014-08-26 20:06:44,832 INFO [main] util.ChecksumType: Checksum can use > > org.apache.hadoop.util.PureJavaCrc32C > > 2014-08-26 20:06:45,125 INFO [main] compress.CodecPool: Got brand-new > > compressor [.snappy] > > 2014-08-26 20:06:45,131 INFO [main] compress.CodecPool: Got brand-new > > compressor [.snappy] > > 2014-08-26 20:06:45,254 INFO [main] compress.CodecPool: Got brand-new > > decompressor [.snappy] > > SUCCESS > > > > > > Please let us know if it still doesn't work for you. Without libhadoop.so > > it doesn't work for me... > > jmspaggi@node8:~/hbase-0.98.5-hadoop2/lib/native$ rm > > Linux-amd64-64/libhadoop.so > > > > jmspaggi@node8:~/hbase-0.98.5-hadoop2$ bin/hbase > > org.apache.hadoop.hbase.util.CompressionTest file:///tmp/snappy-test > snappy > > 2014-08-26 20:09:28,945 INFO [main] Configuration.deprecation: > > hadoop.native.lib is deprecated. Instead, use io.native.lib.available > > 2014-08-26 20:09:29,460 WARN [main] util.NativeCodeLoader: Unable to > load > > native-hadoop library for your platform... using builtin-java classes > where > > applicable > > 2014-08-26 20:09:29,775 INFO [main] util.ChecksumType: Checksum using > > org.apache.hadoop.util.PureJavaCrc32 > > 2014-08-26 20:09:29,776 INFO [main] util.ChecksumType: Checksum can use > > org.apache.hadoop.util.PureJavaCrc32C > > Exception in thread "main" java.lang.UnsatisfiedLinkError: > > org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy()Z > > at org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy(Native > > Method) > > ... > > > > > > I did all of that using a brand new extracted > > hbase-0.98.5-hadoop2-bin.tar.gz file. > > > > JM > > > > > > 2014-08-26 19:47 GMT-04:00 [email protected] < > > [email protected]>: > > > >> $ uname -m > >> x86_64 > >> > >> Arthur > >> > >> On 27 Aug, 2014, at 7:45 am, Jean-Marc Spaggiari < > [email protected]> > >> wrote: > >> > >>> Hi Arthur, > >>> > >>> What uname -m gives you? you need to check that to create the right > >> folder > >>> under the lib directory. > >>> > >>> JM > >>> > >>> > >>> 2014-08-26 19:43 GMT-04:00 Alex Kamil <[email protected]>: > >>> > >>>> Something like this worked for me > >>>> 1. get hbase binaries > >>>> 2. sudo yum install snappy snappy-devel > >>>> 3. ln -sf /usr/lib64/libsnappy.so > >>>> /var/lib/hadoop/lib/native/Linux-amd64-64/. > >>>> 4. ln -sf /usr/lib64/libsnappy.so > >>>> /var/lib/hbase/lib/native/Linux-amd64-64/. > >>>> 5. add snappy jar under $HADOOP_HOME/lib and $HBASE_HOME/lib > >>>> ref: https://issues.apache.org/jira/browse/PHOENIX-877 > >>>> > >>>> > >>>> On Tue, Aug 26, 2014 at 7:25 PM, [email protected] < > >>>> [email protected]> wrote: > >>>> > >>>>> Hi, > >>>>> > >>>>> I just tried three more steps but was not able to get thru. > >>>>> > >>>>> > >>>>> 1) copied snappy files to $HBASE_HOME/lib > >>>>> $ cd $HBASE_HOME > >>>>> $ ll lib/*sna* > >>>>> -rw-r--r--. 1 hduser hadoop 11526 Aug 27 06:54 > >>>>> lib/hadoop-snappy-0.0.1-SNAPSHOT.jar > >>>>> -rw-rw-r--. 1 hduser hadoop 995968 Aug 3 18:43 > >>>> lib/snappy-java-1.0.4.1.jar > >>>>> > >>>>> ll lib/native/ > >>>>> drwxrwxr-x. 4 hduser hadoop 4096 Aug 27 06:54 Linux-amd64-64 > >>>>> > >>>>> ll lib/native/Linux-amd64-64/ > >>>>> total 18964 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 24 Aug 27 07:08 libhadoopsnappy.so > -> > >>>>> libhadoopsnappy.so.0.0.1 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 24 Aug 27 07:08 libhadoopsnappy.so.0 > >> -> > >>>>> libhadoopsnappy.so.0.0.1 > >>>>> -rwxr-xr-x. 1 hduser Hadoop 54961 Aug 27 07:08 > >> libhadoopsnappy.so.0.0.1 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 55 Aug 27 07:08 libjvm.so -> > >>>>> /usr/lib/jvm/jdk1.6.0_45/jre/lib/amd64/server/libjvm.so > >>>>> lrwxrwxrwx. 1 hduser Hadoop 25 Aug 27 07:08 libprotobuf-lite.so > -> > >>>>> libprotobuf-lite.so.8.0.0 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 25 Aug 27 07:08 > libprotobuf-lite.so.8 > >> -> > >>>>> libprotobuf-lite.so.8.0.0 > >>>>> -rwxr-xr-x. 1 hduser Hadoop 964689 Aug 27 07:08 > >>>> libprotobuf-lite.so.8.0.0 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 20 Aug 27 07:08 libprotobuf.so -> > >>>>> libprotobuf.so.8.0.0 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 20 Aug 27 07:08 libprotobuf.so.8 -> > >>>>> libprotobuf.so.8.0.0 > >>>>> -rwxr-xr-x. 1 hduser Hadoop 8300050 Aug 27 07:08 libprotobuf.so.8.0.0 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 18 Aug 27 07:08 libprotoc.so -> > >>>>> libprotoc.so.8.0.0 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 18 Aug 27 07:08 libprotoc.so.8 -> > >>>>> libprotoc.so.8.0.0 > >>>>> -rwxr-xr-x. 1 hduser Hadoop 9935810 Aug 27 07:08 libprotoc.so.8.0.0 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 18 Aug 27 07:08 libsnappy.so -> > >>>>> libsnappy.so.1.2.0 > >>>>> lrwxrwxrwx. 1 hduser Hadoop 18 Aug 27 07:08 libsnappy.so.1 -> > >>>>> libsnappy.so.1.2.0 > >>>>> -rwxr-xr-x. 1 hduser Hadoop 147726 Aug 27 07:08 libsnappy.so.1.2.0 > >>>>> drwxr-xr-x. 2 hduser Hadoop 4096 Aug 27 07:08 pkgconfig > >>>>> > >>>>> 2) $HBASE_HOME/conf/hbase-env.sh, added > >>>>> > >>>>> ### > >>>>> export > >>>>> > >>>> > >> > LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$HADOOP_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/ > >>>>> export > >>>>> > >>>> > >> > HBASE_LIBRARY_PATH=$HBASE_LIBRARY_PATH:$HBASE_HOME/lib/native/Linux-amd64-64/:/usr/local/lib/:$HBASE_HOME/lib/hadoop-snappy-0.0.1-SNAPSHOT.jar > >>>>> export CLASSPATH=$CLASSPATH:$HBASE_LIBRARY_PATH > >>>>> export HBASE_CLASSPATH=$HBASE_CLASSPATH:$HBASE_LIBRARY_PATH > >>>>> ### > >>>>> > >>>>> 3) restart HBASE and tried again > >>>>> $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest > >>>>> file:///tmp/snappy-test snappy > >>>>> 2014-08-27 07:16:09,490 INFO [main] Configuration.deprecation: > >>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available > >>>>> SLF4J: Class path contains multiple SLF4J bindings. > >>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > >>>>> explanation. > >>>>> 2014-08-27 07:16:10,323 INFO [main] util.ChecksumType: Checksum > using > >>>>> org.apache.hadoop.util.PureJavaCrc32 > >>>>> 2014-08-27 07:16:10,324 INFO [main] util.ChecksumType: Checksum can > >> use > >>>>> org.apache.hadoop.util.PureJavaCrc32C > >>>>> Exception in thread "main" java.lang.RuntimeException: native snappy > >>>>> library not available: this version of libhadoop was built without > >> snappy > >>>>> support. > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:64) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132) > >>>>> at > >>>>> > >> > org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148) > >>>>> at > >>>>> > >> > org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:310) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:92) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:690) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:117) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:109) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:97) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:393) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:118) > >>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:148) > >>>>> > >>>>> > >>>>> Regards > >>>>> Arthur > >>>>> > >>>>> > >>>>> > >>>>> On 27 Aug, 2014, at 6:27 am, [email protected] < > >>>>> [email protected]> wrote: > >>>>> > >>>>>> Hi Sean, > >>>>>> > >>>>>> Thanks for your reply. > >>>>>> > >>>>>> I tried the following tests > >>>>>> > >>>>>> $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest > >>>>> file:///tmp/snappy-test gz > >>>>>> 2014-08-26 23:06:17,778 INFO [main] Configuration.deprecation: > >>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available > >>>>>> SLF4J: Class path contains multiple SLF4J bindings. > >>>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > >>>>> explanation. > >>>>>> 2014-08-26 23:06:18,103 INFO [main] util.ChecksumType: Checksum > using > >>>>> org.apache.hadoop.util.PureJavaCrc32 > >>>>>> 2014-08-26 23:06:18,104 INFO [main] util.ChecksumType: Checksum can > >>>> use > >>>>> org.apache.hadoop.util.PureJavaCrc32C > >>>>>> 2014-08-26 23:06:18,260 INFO [main] zlib.ZlibFactory: Successfully > >>>>> loaded & initialized native-zlib library > >>>>>> 2014-08-26 23:06:18,276 INFO [main] compress.CodecPool: Got > brand-new > >>>>> compressor [.gz] > >>>>>> 2014-08-26 23:06:18,280 INFO [main] compress.CodecPool: Got > brand-new > >>>>> compressor [.gz] > >>>>>> 2014-08-26 23:06:18,921 INFO [main] compress.CodecPool: Got > brand-new > >>>>> decompressor [.gz] > >>>>>> SUCCESS > >>>>>> > >>>>>> > >>>>>> $ bin/hbase org.apache.hadoop.hbase.util.CompressionTest > >>>>> file:///tmp/snappy-test snappy > >>>>>> 2014-08-26 23:07:08,246 INFO [main] Configuration.deprecation: > >>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available > >>>>>> SLF4J: Class path contains multiple SLF4J bindings. > >>>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > >>>>> explanation. > >>>>>> 2014-08-26 23:07:08,578 INFO [main] util.ChecksumType: Checksum > using > >>>>> org.apache.hadoop.util.PureJavaCrc32 > >>>>>> 2014-08-26 23:07:08,579 INFO [main] util.ChecksumType: Checksum can > >>>> use > >>>>> org.apache.hadoop.util.PureJavaCrc32C > >>>>>> Exception in thread "main" java.lang.RuntimeException: native snappy > >>>>> library not available: this version of libhadoop was built without > >> snappy > >>>>> support. > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:64) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132) > >>>>>> at > >>>>> > >> > org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148) > >>>>>> at > >>>>> > >> > org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:310) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:92) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:690) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:117) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:109) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:97) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:393) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:118) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:148) > >>>>>> > >>>>>> > >>>>>> $ hbase shell > >>>>>> 2014-08-27 06:23:38,707 INFO [main] Configuration.deprecation: > >>>>> hadoop.native.lib is deprecated. Instead, use io.native.lib.available > >>>>>> HBase Shell; enter 'help<RETURN>' for list of supported commands. > >>>>>> Type "exit<RETURN>" to leave the HBase Shell > >>>>>> Version 0.98.4-hadoop2, rUnknown, Sun Aug 3 23:45:36 HKT 2014 > >>>>>> > >>>>>> hbase(main):001:0> > >>>>>> hbase(main):001:0> create 'tsnappy', { NAME => 'f', COMPRESSION => > >>>>> 'snappy'} > >>>>>> SLF4J: Class path contains multiple SLF4J bindings. > >>>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hbase-0.98.4-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>>> SLF4J: Found binding in > >>>>> > >>>> > >> > [jar:file:/edh/hadoop/hadoop-2.4.1/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] > >>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > >>>>> explanation. > >>>>>> > >>>>>> ERROR: java.io.IOException: Compression algorithm 'snappy' > previously > >>>>> failed test. > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:85) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.master.HMaster.checkCompression(HMaster.java:1764) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.master.HMaster.checkCompression(HMaster.java:1757) > >>>>>> at > >>>>> org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1739) > >>>>>> at > >>>>> org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1774) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:40470) > >>>>>> at > >>>> org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2027) > >>>>>> at > org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98) > >>>>>> at > >>>>> > >>>> > >> > org.apache.hadoop.hbase.ipc.FifoRpcScheduler$1.run(FifoRpcScheduler.java:74) > >>>>>> at > >>>>> > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:439) > >>>>>> at > >>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) > >>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:138) > >>>>>> at > >>>>> > >>>> > >> > java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895) > >>>>>> at > >>>>> > >>>> > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) > >>>>>> at java.lang.Thread.run(Thread.java:662) > >>>>>> > >>>>>> > >>>>>> > >>>>>> > >>>>>> Regards > >>>>>> Arthur > >>>>>> > >>>>>> > >>>>>> On 26 Aug, 2014, at 11:02 pm, Sean Busbey <[email protected]> > >> wrote: > >>>>>> > >>>>>>> Hi Arthur! > >>>>>>> > >>>>>>> Our Snappy build instructions are currently out of date and I'm > >>>> working > >>>>> on updating them[1]. In short, I don't think there are any special > >> build > >>>>> steps for using snappy. > >>>>>>> > >>>>>>> I'm still working out what needs to be included in our instructions > >>>> for > >>>>> local and cluster testing. > >>>>>>> > >>>>>>> If you use the test for compression options, locally things will > fail > >>>>> because the native hadoop libs won't be present: > >>>>>>> > >>>>>>> bin/hbase org.apache.hadoop.hbase.util.CompressionTest > >>>>> file:///tmp/snappy-test snappy > >>>>>>> (for comparison, replace "snappy" with "gz" and you will get a > >> warning > >>>>> about not having native libraries, but the test will succeed.) > >>>>>>> > >>>>>>> I believe JM's suggestion is for you to copy the Hadoop native > >>>>> libraries into the local HBase lib/native directory, which would > allow > >>>> the > >>>>> local test to pass. If you are running in a deployed Hadoop cluster, > I > >>>>> would expect the necessary libraries to already be available to > HBase. > >>>>>>> > >>>>>>> [1]: https://issues.apache.org/jira/browse/HBASE-6189 > >>>>>>> > >>>>>>> -Sean > >>>>>>> > >>>>>>> > >>>>>>> On Tue, Aug 26, 2014 at 8:30 AM, [email protected] < > >>>>> [email protected]> wrote: > >>>>>>> Hi JM > >>>>>>> > >>>>>>> Below are my commands, tried two cases under same source code > folder: > >>>>>>> a) compile with snappy parameters(failed), > >>>>>>> b) compile without snappy parameters (successful). > >>>>>>> > >>>>>>> Regards > >>>>>>> Arthur > >>>>>>> > >>>>>>> wget > >>>>> > http://mirrors.devlib.org/apache/hbase/stable/hbase-0.98.4-src.tar.gz > >>>>>>> tar -vxf hbase-0.98.4-src.tar.gz > >>>>>>> mv hbase-0.98.4 hbase-0.98.4-src_snappy > >>>>>>> cd hbase-0.98.4-src_snappy > >>>>>>> nano dev-support/generate-hadoopX-poms.sh > >>>>>>> (change hbase_home=“/usr/local/hadoop/hbase-0.98.4-src_snappy”) > >>>>>>> > >>>>>>> > >>>>>>> bash -x ./dev-support/generate-hadoopX-poms.sh 0.98.4 > 0.98.4-hadoop2 > >>>>>>> a) with snappy parameters > >>>>>>> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single > >>>>> -Prelease,hadoop-snappy -Dhadoop-snappy.version=0.0.1-SNAPSHOT > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [INFO] Building HBase - Server 0.98.4-hadoop2 > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [WARNING] The POM for > >>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT is missing, no > >>>>> dependency information available > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [INFO] Reactor Summary: > >>>>>>> [INFO] > >>>>>>> [INFO] HBase ............................................. SUCCESS > >>>>> [8.192s] > >>>>>>> [INFO] HBase - Common .................................... SUCCESS > >>>>> [5.638s] > >>>>>>> [INFO] HBase - Protocol .................................. SUCCESS > >>>>> [1.535s] > >>>>>>> [INFO] HBase - Client .................................... SUCCESS > >>>>> [1.206s] > >>>>>>> [INFO] HBase - Hadoop Compatibility ...................... SUCCESS > >>>>> [0.193s] > >>>>>>> [INFO] HBase - Hadoop Two Compatibility .................. SUCCESS > >>>>> [0.798s] > >>>>>>> [INFO] HBase - Prefix Tree ............................... SUCCESS > >>>>> [0.438s] > >>>>>>> [INFO] HBase - Server .................................... FAILURE > >>>>> [0.234s] > >>>>>>> [INFO] HBase - Testing Util .............................. SKIPPED > >>>>>>> [INFO] HBase - Thrift .................................... SKIPPED > >>>>>>> [INFO] HBase - Shell ..................................... SKIPPED > >>>>>>> [INFO] HBase - Integration Tests ......................... SKIPPED > >>>>>>> [INFO] HBase - Examples .................................. SKIPPED > >>>>>>> [INFO] HBase - Assembly .................................. SKIPPED > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [INFO] BUILD FAILURE > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [INFO] Total time: 19.474s > >>>>>>> [INFO] Finished at: Tue Aug 26 21:21:13 HKT 2014 > >>>>>>> [INFO] Final Memory: 51M/1100M > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [ERROR] Failed to execute goal on project hbase-server: Could not > >>>>> resolve dependencies for project > >>>>> org.apache.hbase:hbase-server:jar:0.98.4-hadoop2: Failure to find > >>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT in > >>>>> http://maven.oschina.net/content/groups/public/ was cached in the > >> local > >>>>> repository, resolution will not be reattempted until the update > >> interval > >>>> of > >>>>> nexus-osc has elapsed or updates are forced -> [Help 1] > >>>>>>> [ERROR] > >>>>>>> [ERROR] To see the full stack trace of the errors, re-run Maven > with > >>>>> the -e switch. > >>>>>>> [ERROR] Re-run Maven using the -X switch to enable full debug > >> logging. > >>>>>>> [ERROR] > >>>>>>> [ERROR] For more information about the errors and possible > solutions, > >>>>> please read the following articles: > >>>>>>> [ERROR] [Help 1] > >>>>> > >>>> > >> > http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException > >>>>>>> [ERROR] > >>>>>>> [ERROR] After correcting the problems, you can resume the build > with > >>>>> the command > >>>>>>> [ERROR] mvn <goals> -rf :hbase-server > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> b) try again, without snappy parameters > >>>>>>> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single > -Prelease > >>>>>>> [INFO] Building tar: > >>>>> > >>>> > >> > /edh/hadoop_all_sources/hbase-0.98.4-src_snappy/hbase-assembly/target/hbase-0.98.4-hadoop2-bin.tar.gz > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [INFO] Reactor Summary: > >>>>>>> [INFO] > >>>>>>> [INFO] HBase ............................................. SUCCESS > >>>>> [3.290s] > >>>>>>> [INFO] HBase - Common .................................... SUCCESS > >>>>> [3.119s] > >>>>>>> [INFO] HBase - Protocol .................................. SUCCESS > >>>>> [0.972s] > >>>>>>> [INFO] HBase - Client .................................... SUCCESS > >>>>> [0.920s] > >>>>>>> [INFO] HBase - Hadoop Compatibility ...................... SUCCESS > >>>>> [0.167s] > >>>>>>> [INFO] HBase - Hadoop Two Compatibility .................. SUCCESS > >>>>> [0.504s] > >>>>>>> [INFO] HBase - Prefix Tree ............................... SUCCESS > >>>>> [0.382s] > >>>>>>> [INFO] HBase - Server .................................... SUCCESS > >>>>> [4.790s] > >>>>>>> [INFO] HBase - Testing Util .............................. SUCCESS > >>>>> [0.598s] > >>>>>>> [INFO] HBase - Thrift .................................... SUCCESS > >>>>> [1.536s] > >>>>>>> [INFO] HBase - Shell ..................................... SUCCESS > >>>>> [0.369s] > >>>>>>> [INFO] HBase - Integration Tests ......................... SUCCESS > >>>>> [0.443s] > >>>>>>> [INFO] HBase - Examples .................................. SUCCESS > >>>>> [0.459s] > >>>>>>> [INFO] HBase - Assembly .................................. SUCCESS > >>>>> [13.240s] > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [INFO] BUILD SUCCESS > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> [INFO] Total time: 31.408s > >>>>>>> [INFO] Finished at: Tue Aug 26 21:22:50 HKT 2014 > >>>>>>> [INFO] Final Memory: 57M/1627M > >>>>>>> [INFO] > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> On 26 Aug, 2014, at 8:52 pm, Jean-Marc Spaggiari < > >>>>> [email protected]> wrote: > >>>>>>> > >>>>>>>> Hi Arthur, > >>>>>>>> > >>>>>>>> How have you extracted HBase source and what command do you run to > >>>>> build? I > >>>>>>>> will do the same here locally so I can provide you the exact step > to > >>>>>>>> complete. > >>>>>>>> > >>>>>>>> JM > >>>>>>>> > >>>>>>>> > >>>>>>>> 2014-08-26 8:42 GMT-04:00 [email protected] < > >>>>> [email protected] > >>>>>>>>> : > >>>>>>>> > >>>>>>>>> Hi JM > >>>>>>>>> > >>>>>>>>> Not too sure what you mean, do you mean I should create a new > >>>> folder > >>>>> in my > >>>>>>>>> HBASE_SRC named lib/native/Linux-x86 and copy these files to this > >>>>> folder > >>>>>>>>> then try to compile it again? > >>>>>>>>> > >>>>>>>>> Regards > >>>>>>>>> ARthur > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> On 26 Aug, 2014, at 8:17 pm, Jean-Marc Spaggiari < > >>>>> [email protected]> > >>>>>>>>> wrote: > >>>>>>>>> > >>>>>>>>>> Hi Arthur, > >>>>>>>>>> > >>>>>>>>>> Almost done! You now need to copy them on the HBase folder. > >>>>>>>>>> > >>>>>>>>>> hbase@hbasetest1:~/hbase-0.98.2-hadoop2/lib$ tree | grep -v > .jar > >>>> | > >>>>> grep > >>>>>>>>> -v > >>>>>>>>>> .rb > >>>>>>>>>> . > >>>>>>>>>> ├── native > >>>>>>>>>> │ └── Linux-x86 > >>>>>>>>>> │ ├── libsnappy.a > >>>>>>>>>> │ ├── libsnappy.la > >>>>>>>>>> │ ├── libsnappy.so > >>>>>>>>>> │ ├── libsnappy.so.1 > >>>>>>>>>> │ └── libsnappy.so.1.2.0 > >>>>>>>>>> > >>>>>>>>>> I don't have any hadoop-snappy lib in my hbase folder and it > works > >>>>> very > >>>>>>>>>> well with Snappy for me... > >>>>>>>>>> > >>>>>>>>>> JM > >>>>>>>>>> > >>>>>>>>>> 2014-08-26 8:09 GMT-04:00 [email protected] < > >>>>>>>>> [email protected] > >>>>>>>>>>> : > >>>>>>>>>> > >>>>>>>>>>> Hi JM, > >>>>>>>>>>> > >>>>>>>>>>> Below are my steps to install snappy lib, do I miss something? > >>>>>>>>>>> > >>>>>>>>>>> Regards > >>>>>>>>>>> Arthur > >>>>>>>>>>> > >>>>>>>>>>> wget https://snappy.googlecode.com/files/snappy-1.1.1.tar.gz > >>>>>>>>>>> tar -vxf snappy-1.1.1.tar.gz > >>>>>>>>>>> cd snappy-1.1.1 > >>>>>>>>>>> ./configure > >>>>>>>>>>> make > >>>>>>>>>>> make install > >>>>>>>>>>> make[1]: Entering directory > >>>>>>>>> `/edh/hadoop_all_sources/snappy-1.1.1' > >>>>>>>>>>> test -z "/usr/local/lib" || /bin/mkdir -p "/usr/local/lib" > >>>>>>>>>>> /bin/sh ./libtool --mode=install /usr/bin/install -c > >>>>>>>>>>> libsnappy.la '/usr/local/lib' > >>>>>>>>>>> libtool: install: /usr/bin/install -c > >>>>> .libs/libsnappy.so.1.2.0 > >>>>>>>>>>> /usr/local/lib/libsnappy.so.1.2.0 > >>>>>>>>>>> libtool: install: (cd /usr/local/lib && { ln -s -f > >>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so.1 || { rm -f libsnappy.so.1 && > ln > >>>>> -s > >>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so.1; }; }) > >>>>>>>>>>> libtool: install: (cd /usr/local/lib && { ln -s -f > >>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so || { rm -f libsnappy.so && ln > -s > >>>>>>>>>>> libsnappy.so.1.2.0 libsnappy.so; }; }) > >>>>>>>>>>> libtool: install: /usr/bin/install -c .libs/libsnappy.lai > >>>>>>>>>>> /usr/local/lib/libsnappy.la > >>>>>>>>>>> libtool: install: /usr/bin/install -c .libs/libsnappy.a > >>>>>>>>>>> /usr/local/lib/libsnappy.a > >>>>>>>>>>> libtool: install: chmod 644 /usr/local/lib/libsnappy.a > >>>>>>>>>>> libtool: install: ranlib /usr/local/lib/libsnappy.a > >>>>>>>>>>> libtool: finish: > >>>>>>>>>>> > >>>>>>>>> > >>>>> > >>>> > >> > PATH="/edh/hadoop/spark/bin:/edh/hadoop/hbase/bin:/edh/hadoop/zookeeper//bin:/edh/hadoop/yarn/hadoop/bin:/edh/hadoop/yarn/hadoop/sbin:/usr/lib64/qt-3.3/bin:/opt/apache-maven-3.1.1/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/edh/hadoop/zookeeper//bin:/edh/hadoop/hive//bin:/usr/lib/jvm/jdk1.6.0_45//bin:/root/bin:/sbin" > >>>>>>>>>>> ldconfig -n /usr/local/lib > >>>>>>>>>>> > >>>>>>>>>>> > >>>>> > ---------------------------------------------------------------------- > >>>>>>>>>>> Libraries have been installed in: > >>>>>>>>>>> /usr/local/lib > >>>>>>>>>>> If you ever happen to want to link against installed > >>>>> libraries > >>>>>>>>>>> in a given directory, LIBDIR, you must either use libtool, > >>>>> and > >>>>>>>>>>> specify the full pathname of the library, or use the > >>>>> `-LLIBDIR' > >>>>>>>>>>> flag during linking and do at least one of the following: > >>>>>>>>>>> - add LIBDIR to the `LD_LIBRARY_PATH' environment variable > >>>>>>>>>>> during execution > >>>>>>>>>>> - add LIBDIR to the `LD_RUN_PATH' environment variable > >>>>>>>>>>> during linking > >>>>>>>>>>> - use the `-Wl,-rpath -Wl,LIBDIR' linker flag > >>>>>>>>>>> - have your system administrator add LIBDIR to > >>>>> `/etc/ld.so.conf' > >>>>>>>>>>> See any operating system documentation about shared > >>>>> libraries for > >>>>>>>>>>> more information, such as the ld(1) and ld.so(8) manual > >>>>> pages. > >>>>>>>>>>> > >>>>>>>>>>> > >>>>> > ---------------------------------------------------------------------- > >>>>>>>>>>> test -z "/usr/local/share/doc/snappy" || /bin/mkdir -p > >>>>>>>>>>> "/usr/local/share/doc/snappy" > >>>>>>>>>>> /usr/bin/install -c -m 644 ChangeLog COPYING INSTALL NEWS > >>>>> README > >>>>>>>>>>> format_description.txt framing_format.txt > >>>>> '/usr/local/share/doc/snappy' > >>>>>>>>>>> test -z "/usr/local/include" || /bin/mkdir -p > >>>>>>>>> "/usr/local/include" > >>>>>>>>>>> /usr/bin/install -c -m 644 snappy.h snappy-sinksource.h > >>>>>>>>>>> snappy-stubs-public.h snappy-c.h '/usr/local/include' > >>>>>>>>>>> make[1]: Leaving directory > >>>>> `/edh/hadoop_all_sources/snappy-1.1.1' > >>>>>>>>>>> > >>>>>>>>>>> ll /usr/local/lib > >>>>>>>>>>> -rw-r--r--. 1 root root 233554 Aug 20 00:14 libsnappy.a > >>>>>>>>>>> -rwxr-xr-x. 1 root root 953 Aug 20 00:14 libsnappy.la > >>>>>>>>>>> lrwxrwxrwx. 1 root root 18 Aug 20 00:14 libsnappy.so > >>>> -> > >>>>>>>>>>> libsnappy.so.1.2.0 > >>>>>>>>>>> lrwxrwxrwx. 1 root root 18 Aug 20 00:14 > >>>> libsnappy.so.1 > >>>>> -> > >>>>>>>>>>> libsnappy.so.1.2.0 > >>>>>>>>>>> -rwxr-xr-x. 1 root root 147726 Aug 20 00:14 > >>>>> libsnappy.so.1.2.0 > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> On 26 Aug, 2014, at 7:38 pm, Jean-Marc Spaggiari < > >>>>>>>>> [email protected]> > >>>>>>>>>>> wrote: > >>>>>>>>>>> > >>>>>>>>>>>> Hi Arthur, > >>>>>>>>>>>> > >>>>>>>>>>>> Do you have snappy libs installed and configured? HBase > doesn't > >>>>> come > >>>>>>>>> with > >>>>>>>>>>>> Snappy. So yo need to have it first. > >>>>>>>>>>>> > >>>>>>>>>>>> Shameless plug: > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>> > >>>>> > >>>> > >> > http://www.spaggiari.org/index.php/hbase/how-to-install-snappy-with-1#.U_xxSqdZuZY > >>>>>>>>>>>> > >>>>>>>>>>>> This is for 0.96 but should be very similar for 0.98. I will > try > >>>>> it > >>>>>>>>> soon > >>>>>>>>>>>> and post and update, but keep us posted here so we can support > >>>>> you... > >>>>>>>>>>>> > >>>>>>>>>>>> JM > >>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>>> 2014-08-26 7:34 GMT-04:00 [email protected] < > >>>>>>>>>>> [email protected] > >>>>>>>>>>>>> : > >>>>>>>>>>>> > >>>>>>>>>>>>> Hi, > >>>>>>>>>>>>> > >>>>>>>>>>>>> I need to install snappy to HBase 0.98.4. (my Hadoop version > >>>> is > >>>>>>>>> 2.4.1) > >>>>>>>>>>>>> > >>>>>>>>>>>>> Can you please advise what would be wrong? Should my pom.xml > >>>> be > >>>>>>>>>>> incorrect > >>>>>>>>>>>>> and missing something? > >>>>>>>>>>>>> > >>>>>>>>>>>>> Regards > >>>>>>>>>>>>> Arthur > >>>>>>>>>>>>> > >>>>>>>>>>>>> > >>>>>>>>>>>>> Below are my commands: > >>>>>>>>>>>>> bash -x ./dev-support/generate-hadoopX-poms.sh 0.98.4 > >>>>> 0.98.4-hadoop2 > >>>>>>>>>>>>> mvn -f pom.xml.hadoop2 install -DskipTests assembly:single > >>>>>>>>>>>>> -Prelease,hadoop-snappy > >>>>>>>>>>>>> > >>>>>>>>>>>>> Iog: > >>>>>>>>>>>>> [INFO] > >>>>>>>>>>>>> > >>>>>>>>> > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>>>>>>>> [INFO] Building HBase - Server 0.98.4-hadoop2 > >>>>>>>>>>>>> [INFO] > >>>>>>>>>>>>> > >>>>>>>>> > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>>>>>>>> [WARNING] The POM for > >>>>>>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT > >>>>>>>>>>>>> is missing, no dependency information available > >>>>>>>>>>>>> [INFO] > >>>>>>>>>>>>> > >>>>>>>>> > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>>>>>>>> [INFO] Reactor Summary: > >>>>>>>>>>>>> [INFO] > >>>>>>>>>>>>> [INFO] HBase ............................................. > >>>>> SUCCESS > >>>>>>>>>>> [3.129s] > >>>>>>>>>>>>> [INFO] HBase - Common .................................... > >>>>> SUCCESS > >>>>>>>>>>> [3.105s] > >>>>>>>>>>>>> [INFO] HBase - Protocol .................................. > >>>>> SUCCESS > >>>>>>>>>>> [0.976s] > >>>>>>>>>>>>> [INFO] HBase - Client .................................... > >>>>> SUCCESS > >>>>>>>>>>> [0.925s] > >>>>>>>>>>>>> [INFO] HBase - Hadoop Compatibility ...................... > >>>>> SUCCESS > >>>>>>>>>>> [0.183s] > >>>>>>>>>>>>> [INFO] HBase - Hadoop Two Compatibility .................. > >>>>> SUCCESS > >>>>>>>>>>> [0.497s] > >>>>>>>>>>>>> [INFO] HBase - Prefix Tree ............................... > >>>>> SUCCESS > >>>>>>>>>>> [0.407s] > >>>>>>>>>>>>> [INFO] HBase - Server .................................... > >>>>> FAILURE > >>>>>>>>>>> [0.103s] > >>>>>>>>>>>>> [INFO] HBase - Testing Util .............................. > >>>>> SKIPPED > >>>>>>>>>>>>> [INFO] HBase - Thrift .................................... > >>>>> SKIPPED > >>>>>>>>>>>>> [INFO] HBase - Shell ..................................... > >>>>> SKIPPED > >>>>>>>>>>>>> [INFO] HBase - Integration Tests ......................... > >>>>> SKIPPED > >>>>>>>>>>>>> [INFO] HBase - Examples .................................. > >>>>> SKIPPED > >>>>>>>>>>>>> [INFO] HBase - Assembly .................................. > >>>>> SKIPPED > >>>>>>>>>>>>> [INFO] > >>>>>>>>>>>>> > >>>>>>>>> > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>>>>>>>> [INFO] BUILD FAILURE > >>>>>>>>>>>>> [INFO] > >>>>>>>>>>>>> > >>>>>>>>> > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>>>>>>>> [INFO] Total time: 9.939s > >>>>>>>>>>>>> [INFO] Finished at: Tue Aug 26 19:23:14 HKT 2014 > >>>>>>>>>>>>> [INFO] Final Memory: 61M/2921M > >>>>>>>>>>>>> [INFO] > >>>>>>>>>>>>> > >>>>>>>>> > >>>>> > >> ------------------------------------------------------------------------ > >>>>>>>>>>>>> [ERROR] Failed to execute goal on project hbase-server: Could > >>>> not > >>>>>>>>>>> resolve > >>>>>>>>>>>>> dependencies for project > >>>>>>>>>>> org.apache.hbase:hbase-server:jar:0.98.4-hadoop2: > >>>>>>>>>>>>> Failure to find > >>>>> org.apache.hadoop:hadoop-snappy:jar:0.0.1-SNAPSHOT in > >>>>>>>>>>>>> http://maven.oschina.net/content/groups/public/ was cached > in > >>>>> the > >>>>>>>>> local > >>>>>>>>>>>>> repository, resolution will not be reattempted until the > update > >>>>>>>>>>> interval of > >>>>>>>>>>>>> nexus-osc has elapsed or updates are forced -> [Help 1] > >>>>>>>>>>>>> [ERROR] > >>>>>>>>>>>>> [ERROR] To see the full stack trace of the errors, re-run > Maven > >>>>> with > >>>>>>>>> the > >>>>>>>>>>>>> -e switch. > >>>>>>>>>>>>> [ERROR] Re-run Maven using the -X switch to enable full debug > >>>>> logging. > >>>>>>>>>>>>> [ERROR] > >>>>>>>>>>>>> [ERROR] For more information about the errors and possible > >>>>> solutions, > >>>>>>>>>>>>> please read the following articles: > >>>>>>>>>>>>> [ERROR] [Help 1] > >>>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>> > >>>>> > >>>> > >> > http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException > >>>>>>>>>>>>> [ERROR] > >>>>>>>>>>>>> [ERROR] After correcting the problems, you can resume the > build > >>>>> with > >>>>>>>>> the > >>>>>>>>>>>>> command > >>>>>>>>>>>>> [ERROR] mvn <goals> -rf :hbase-server > >>>>>>>>>>>>> > >>>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> > >>>>>>> -- > >>>>>>> Sean > >>>>>> > >>>>> > >>>>> > >>>> > >> > >> > >
