Hi Hanish,
Thanks for the link it did help. Long story short , always recompile native libraries for your machine :)

Thanks,
Peter

On 11.12.2014 05:46, Hanish Bansal wrote:
Hope this may help you:

http://blogs.impetus.com/big_data/big_data_technologies/SnappyCompressionInHBase.do

On Thu, Dec 11, 2014 at 7:25 AM, Fabio <[email protected] <mailto:[email protected]>> wrote:

    Plain Apache Hadoop 2.5.0.
    Too bad it didn't work, hope someone can help.


    On 12/10/2014 06:22 PM, peterm_second wrote:

        Hi Fabio ,
        Thanks for the reply, but unfortunately it didn't work. I am
        using vanilla hadoop 2.4 with vanilla hive 0.14 and so on, I
        am using the vanilla distros.
        I did set the HADOOP_COMMON_LIB_NATIVE_DIR but that didn't
        make any change. What version were you using ?

        Peter


        On 10.12.2014 16:23, Fabio wrote:

            Not sure it will help, but if the problem is native
            library loading, I spent a loooong time trying anything to
            make it work.
            I may suggest to try also:
            export JAVA_LIBRARY_PATH=/opt/yarn/hadoop-2.5.0/lib/native
            export HADOOP_COMMON_LIB_NATIVE_DIR=/opt/yarn/hadoop-2.5.0/lib
            I have this both in the bash "init" script
            (/etc/profile.p/...) and in
            /opt/yarn/hadoop-2.5.0/etc/hadoop/hadoop-env.sh; quite
            sure it's redundant, but as long as it works I don't
            change it.
            I see here I commented out my attempts to set HADOOP_OPTS,
            so maybe it's not necessary.
            I don't see anything in my .xml config files.
            Also, someone says to compile the libraries under your 64
            bit system, since the ones in Hadoop are for a 32bit
            architecture.

            Good luck

            Fabio

            On 12/10/2014 02:57 PM, peterm_second wrote:

                Hi guys,
                I have a hadoop + hbase + hive application,
                For some reason my cluster is unable to find the
                snappy native library
                Here is the exception :
                 org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy()Z
                    at
                
org.apache.hadoop.util.NativeCodeLoader.buildSupportsSnappy(Native
                Method)
                    at
                
org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:63)

                    at
                
org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:132)

                    at
                
org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:148)
                    at
                
org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:163)
                    at
                org.apache.hadoop.mapred.IFile$Writer.<init>(IFile.java:115)
                    at
                
org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:1583)

                    at
                
org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:1462)
                    at
                org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:437)
                    at
                org.apache.hadoop.mapred.MapTask.run(MapTask.java:342)
                    at
                org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167)
                    at
                java.security.AccessController.doPrivileged(Native Method)
                    at javax.security.auth.Subject.doAs(Subject.java:422)
                    at
                
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)

                    at
                org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162)


                I am working with a 64bit ubuntu 14.04LTS. I've
                installed snappy on my os and added the coppied the
                libs to hadoop_home/lib/native
                I've also added the libs to the JRE, but it still
                fails as if nothing is present.
I've added HADOOP_OPTS="-Djava.net.preferIPv4Stack=true
                $GC_DEBUG_OPTS
                -Djava.library.path=/usr/local/hadoop-2.4.0/lib/native
                $HADOOP_OPTS"
                in my yarn xml I have
                <property>
                      <name>yarn.app.mapreduce.am
                <http://yarn.app.mapreduce.am>.env</name>
                <value>LD_LIBRARY_PATH=$HADOOP_HOME/lib/native</value>
                </property>

                in my mapred-site.xml i have
                <property>
                        <name>mapred.child.java.opts</name>
                        <value>
                -Djava.library.path=/usr/local/hadoop-2.4.0/lib/native</value>
                    </property>
                    <property>
                        <name>mapreduce.reduce.java.opts</name>
                
<value>-Djava.library.path=/usr/local/hadoop-2.4.0/lib/native</value>
                    </property>

                The last two were a desperation move.
                The result is always the same. Any ideas would be
                welcomed.

                Thanks,
                Peter









        .





--
*Thanks & Regards*
*Hanish Bansal*

Reply via email to