The document is far from complete. It didn't mention the default hadoop
binary package is compiled without snappy support and you need to
compile it with snappy option yourself. Actually it didn't work with any
native libs on 64 bits OS as the libhadoop.so in the binary package is
only for 32 bits OS. It also din't metion actually you need both snappy
and hadoop-snappy.
于 2014/1/3 19:20, 张玉雪 写道:
Hi:
When I used hadoop 2.2.0 and hbase 0.96.1.1 to using snappy
compression
I followed the topic
http://hbase.apache.org/book/snappy.compression.html, but I get some error
,can some one help me?
[hadoop@master bin]$ hbase org.apache.hadoop.hbase.util.CompressionTest
file:///tmp/test222.txt snappy
2014-01-03 19:12:41,971 INFO [main] Configuration.deprecation:
hadoop.native.lib is deprecated. Instead, use io.native.lib.available
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/home/hadoop/hbase-0.96.1.1-hadoop2/lib/slf4j-log4j12-1.6.4.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/home/hadoop/hadoop-2.2.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
2014-01-03 19:12:42,663 INFO [main] util.ChecksumType: Checksum using
org.apache.hadoop.util.PureJavaCrc32
2014-01-03 19:12:42,670 INFO [main] util.ChecksumType: Checksum can use
org.apache.hadoop.util.PureJavaCrc32C
Exception in thread "main" java.lang.RuntimeException: native snappy library
not available: this version of libhadoop was built without snappy support.
at
org.apache.hadoop.io.compress.SnappyCodec.checkNativeCodeLoaded(SnappyCodec.java:63)
at
org.apache.hadoop.io.compress.SnappyCodec.getCompressorType(SnappyCodec.java:131)
at
org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:147)
at
org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:162)
at
org.apache.hadoop.hbase.io.compress.Compression$Algorithm.getCompressor(Compression.java:312)
at
org.apache.hadoop.hbase.io.encoding.HFileBlockDefaultEncodingContext.<init>(HFileBlockDefaultEncodingContext.java:79)
at
org.apache.hadoop.hbase.io.hfile.HFileBlock$Writer.<init>(HFileBlock.java:719)
at
org.apache.hadoop.hbase.io.hfile.HFileWriterV2.finishInit(HFileWriterV2.java:131)
at
org.apache.hadoop.hbase.io.hfile.HFileWriterV2.<init>(HFileWriterV2.java:122)
at
org.apache.hadoop.hbase.io.hfile.HFileWriterV2$WriterFactoryV2.createWriter(HFileWriterV2.java:105)
at
org.apache.hadoop.hbase.io.hfile.HFile$WriterFactory.create(HFile.java:426)
at
org.apache.hadoop.hbase.util.CompressionTest.doSmokeTest(CompressionTest.java:115)
at
org.apache.hadoop.hbase.util.CompressionTest.main(CompressionTest.java:145)
I have installed snappy 1.2.0 successfully and hadoop-snappy successfully too…
and also I checked my configuration of core-site.xml and hbase.env.sh
There are all wright. How can I resolve this problem