hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From 张玉雪 <zhangyuxue123...@163.com>
Subject Snappy compression question
Date Fri, 03 Jan 2014 11:20:12 GMT
Hi:

         When I used hadoop 2.2.0 and hbase 0.96.1.1 to using snappy compression

         I followed the topic http://hbase.apache.org/book/snappy.compression.html,  but I
get some error ,can some one help me?

   

[hadoop@master bin]$ hbase org.apache.hadoop.hbase.util.CompressionTest file:///tmp/test222.txt
snappy

2014-01-03 19:12:41,971 INFO  [main] Configuration.deprecation: hadoop.native.lib is deprecated.
Instead, use io.native.lib.available

SLF4J: Class path contains multiple SLF4J bindings.

SLF4J: Found binding in [jar:file:/home/hadoop/hbase-0.96.1.1-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/home/hadoop/hadoop-2.2.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.

2014-01-03 19:12:42,663 INFO  [main] util.ChecksumType: Checksum using org.apache.hadoop.util.PureJavaCrc32

2014-01-03 19:12:42,670 INFO  [main] util.ChecksumType: Checksum can use org.apache.hadoop.util.PureJavaCrc32C

Exception in thread "main" java.lang.RuntimeException: native snappy library not available:
this version of libhadoop was built without snappy support.

        at org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:63)

        at org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:131)

        at org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:147)

        at org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:162)

        at org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:312)

        at org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:79)

        at org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:719)

        at org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:131)

        at org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:122)

        at org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:105)

        at org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:426)

        at org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:115)

        at org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:145)

 

 

I have  installed snappy 1.2.0 successfully and hadoop-snappy successfully too… and also
I checked my configuration of core-site.xml and hbase.env.sh

There are all wright. How can I resolve this problem
Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message