[
https://issues.apache.org/jira/browse/PHOENIX-1232?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
James Taylor updated PHOENIX-1232:
----------------------------------
Summary: Replace direct use of Snappy codec methods with a compressor
abstraction (was: java.lang.UnsatisfiedLinkError:
org.xerial.snappy.SnappyNative.maxCompressedLength(I)I on JOIN using
org.apache.phoenix.pig.PhoenixHBaseLoader)
> Replace direct use of Snappy codec methods with a compressor abstraction
> ------------------------------------------------------------------------
>
> Key: PHOENIX-1232
> URL: https://issues.apache.org/jira/browse/PHOENIX-1232
> Project: Phoenix
> Issue Type: Bug
> Affects Versions: 4.0.0
> Reporter: Brian Johnson
> Assignee: Andrew Purtell
>
> When using the using the pig PhoenixHBaseLoader, I get the following error
> when doing a JOIN on two tables. The tables are defined as:
> 1.9.3-p194 :002 > describe 'TEST'
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in
> [jar:file:/home/hadoop/.versions/hbase-0.98.5/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in
> [jar:file:/home/hadoop/.versions/2.4.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
> explanation.
> 2014-09-02 20:11:25,402 WARN [main] util.NativeCodeLoader: Unable to load
> native-hadoop library for your platform... using builtin-java classes where
> applicable
> DESCRIPTION
>
> ENABLED
>
> 'TEST', {TABLE_ATTRIBUTES => {coprocessor$1 =>
> '|org.apache.phoenix.coprocessor.ScanRegionObserver|1|', coprocessor$2 =>
> '|org.apache.phoenix.coprocessor.UngroupedAggregate true
>
> RegionObserver|1|', coprocessor$3 =>
> '|org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver|1|',
> coprocessor$4 => '|org.apache.phoenix.coprocessor.ServerCachingEnd
>
>
> pointImpl|1|', coprocessor$5 =>
> '|org.apache.phoenix.hbase.index.Indexer|1073741823|org.apache.hadoop.hbase.index.codec.class=org.apache.phoenix.index.PhoenixIndexCodec,ind
>
>
> ex.builder=org.apache.phoenix.index.PhoenixIndexBuilder', coprocessor$6 =>
> '|org.apache.hadoop.hbase.regionserver.LocalIndexSplitter|1|'}, {NAME => '0',
> DATA_BLOCK_ENCODING
>
> => 'FAST_DIFF', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', COMPRESSION
> => 'NONE', VERSIONS => '1', TTL => 'FOREVER', MIN_VERSIONS => '0',
> KEEP_DELETED_CELLS => 'true'
>
> , BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}, {NAME
> => 'cf1', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'ROW',
> REPLICATION_SCOPE => '0', COMPRESS
>
> ION => 'NONE', VERSIONS => '1', TTL => 'FOREVER', MIN_VERSIONS => '0',
> KEEP_DELETED_CELLS => 'false', BLOCKSIZE => '65536', IN_MEMORY => 'false',
> BLOCKCACHE => 'true'}
>
> 1 row(s) in 1.4710 seconds
> 1.9.3-p194 :003 > describe 'TEST2'
> DESCRIPTION
>
> ENABLED
>
> 'TEST2', {TABLE_ATTRIBUTES => {coprocessor$1 =>
> '|org.apache.phoenix.coprocessor.ScanRegionObserver|1|', coprocessor$2 =>
> '|org.apache.phoenix.coprocessor.UngroupedAggregat true
>
> eRegionObserver|1|', coprocessor$3 =>
> '|org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver|1|',
> coprocessor$4 => '|org.apache.phoenix.coprocessor.ServerCachingEn
>
>
> dpointImpl|1|', coprocessor$5 =>
> '|org.apache.phoenix.hbase.index.Indexer|1073741823|org.apache.hadoop.hbase.index.codec.class=org.apache.phoenix.index.PhoenixIndexCodec,in
>
>
> dex.builder=org.apache.phoenix.index.PhoenixIndexBuilder', coprocessor$6 =>
> '|org.apache.hadoop.hbase.regionserver.LocalIndexSplitter|1|'}, {NAME => '0',
> DATA_BLOCK_ENCODIN
>
> G => 'FAST_DIFF', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0',
> COMPRESSION => 'NONE', VERSIONS => '1', TTL => 'FOREVER', MIN_VERSIONS =>
> '0', KEEP_DELETED_CELLS => 'true
>
> ', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}, {NAME
> => 'cf1', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'ROW',
> REPLICATION_SCOPE => '0', COMPRES
>
> SION => 'NONE', VERSIONS => '1', TTL => 'FOREVER', MIN_VERSIONS => '0',
> KEEP_DELETED_CELLS => 'false', BLOCKSIZE => '65536', IN_MEMORY => 'false',
> BLOCKCACHE => 'true'}
>
> 1 row(s) in 0.0240 seconds
> and the query is
> SELECT TEST.ID FROM TEST LEFT OUTER JOIN TEST2 ON TEST.ID = TEST2.ID
> It works fine from sqlline.py and neither table has compression enabled.
> Caused by: java.lang.UnsatisfiedLinkError:
> org.xerial.snappy.SnappyNative.maxCompressedLength(I)I
> at org.xerial.snappy.SnappyNative.maxCompressedLength(Native Method)
> at org.xerial.snappy.Snappy.maxCompressedLength(Snappy.java:316)
> at
> org.apache.phoenix.join.HashCacheClient.serialize(HashCacheClient.java:119)
> at
> org.apache.phoenix.join.HashCacheClient.addHashCache(HashCacheClient.java:77)
> at org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:119)
> at org.apache.phoenix.execute.HashJoinPlan$1.call(HashJoinPlan.java:114)
> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
> at java.lang.Thread.run(Thread.java:695)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)