Hi,
 
I fixed the issue by copying libsnappy.so to Java ire.

Regards
Arthur

On 29 Aug, 2014, at 8:12 am, arthur.hk.c...@gmail.com 
<arthur.hk.c...@gmail.com> wrote:

> Hi,
> 
> If change my etc/hadoop/core-site.xml 
> 
> from 
>    <property>
>     <name>io.compression.codecs</name>
>     <value>
>       org.apache.hadoop.io.compress.SnappyCodec,
>       org.apache.hadoop.io.compress.GzipCodec,
>       org.apache.hadoop.io.compress.DefaultCodec,
>       org.apache.hadoop.io.compress.BZip2Codec
>     </value>
>    </property>
> 
> to 
>    <property>
>     <name>io.compression.codecs</name>
>     <value>
>       org.apache.hadoop.io.compress.GzipCodec,
>       org.apache.hadoop.io.compress.SnappyCodec,
>       org.apache.hadoop.io.compress.DefaultCodec,
>       org.apache.hadoop.io.compress.BZip2Codec
>     </value>
>    </property>
> 
> 
> 
> and run the test again, I found this time it cannot find 
> "org.apache.hadoop.io.compress.GzipCodec"
> 
> scala> inFILE.first()
> java.lang.RuntimeException: Error in configuring object
>       at 
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109)
>       at 
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75)
>       at 
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
>       at org.apache.spark.rdd.HadoopRDD.getInputFormat(HadoopRDD.scala:158)
>       at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:171)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202)
>       at scala.Option.getOrElse(Option.scala:120)
>       at org.apache.spark.rdd.RDD.partitions(RDD.scala:202)
>       at org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202)
>       at scala.Option.getOrElse(Option.scala:120)
>       at org.apache.spark.rdd.RDD.partitions(RDD.scala:202)
>       at org.apache.spark.rdd.RDD.take(RDD.scala:983)
>       at org.apache.spark.rdd.RDD.first(RDD.scala:1015)
>       at $iwC$$iwC$$iwC$$iwC.<init>(<console>:15)
>       at $iwC$$iwC$$iwC.<init>(<console>:20)
>       at $iwC$$iwC.<init>(<console>:22)
>       at $iwC.<init>(<console>:24)
>       at <init>(<console>:26)
>       at .<init>(<console>:30)
>       at .<clinit>(<console>)
>       at .<init>(<console>:7)
>       at .<clinit>(<console>)
>       at $print(<console>)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>       at java.lang.reflect.Method.invoke(Method.java:597)
>       at 
> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:788)
>       at 
> org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1056)
>       at 
> org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:614)
>       at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:645)
>       at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:609)
>       at 
> org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:796)
>       at 
> org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:841)
>       at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:753)
>       at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:601)
>       at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:608)
>       at org.apache.spark.repl.SparkILoop.loop(SparkILoop.scala:611)
>       at 
> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:936)
>       at 
> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)
>       at 
> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)
>       at 
> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
>       at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:884)
>       at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:982)
>       at org.apache.spark.repl.Main$.main(Main.scala:31)
>       at org.apache.spark.repl.Main.main(Main.scala)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>       at java.lang.reflect.Method.invoke(Method.java:597)
>       at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:303)
>       at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:55)
>       at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Caused by: java.lang.reflect.InvocationTargetException
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>       at java.lang.reflect.Method.invoke(Method.java:597)
>       at 
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:106)
>       ... 55 more
> Caused by: java.lang.IllegalArgumentException: Compression codec       
> org.apache.hadoop.io.compress.GzipCodec not found.
>       at 
> org.apache.hadoop.io.compress.CompressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:135)
>       at 
> org.apache.hadoop.io.compress.CompressionCodecFactory.<init>(CompressionCodecFactory.java:175)
>       at 
> org.apache.hadoop.mapred.TextInputFormat.configure(TextInputFormat.java:45)
>       ... 60 more
> Caused by: java.lang.ClassNotFoundException: Class       
> org.apache.hadoop.io.compress.GzipCodec not found
>       at 
> org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1801)
>       at 
> org.apache.hadoop.io.compress.CompressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:128)
>       ... 62 more
> 
> 
> Any idea to fix this issue?
> Regards
> Arthur
> 
> 
> On 29 Aug, 2014, at 2:58 am, arthur.hk.c...@gmail.com 
> <arthur.hk.c...@gmail.com> wrote:
> 
>> Hi,
>> 
>> my check native result:
>> 
>> hadoop checknative
>> 14/08/29 02:54:51 WARN bzip2.Bzip2Factory: Failed to load/initialize 
>> native-bzip2 library system-native, will use pure-Java version
>> 14/08/29 02:54:51 INFO zlib.ZlibFactory: Successfully loaded & initialized 
>> native-zlib library
>> Native library checking:
>> hadoop: true 
>> /mnt/hadoop/hadoop-2.4.1_snappy/lib/native/Linux-amd64-64/libhadoop.so
>> zlib:   true /lib64/libz.so.1
>> snappy: true 
>> /mnt/hadoop/hadoop-2.4.1_snappy/lib/native/Linux-amd64-64/libsnappy.so.1
>> lz4:    true revision:99
>> bzip2:  false
>> 
>> Any idea how to enable or disable  snappy in Spark?
>> 
>> Regards
>> Arthur
>> 
>> 
>> On 29 Aug, 2014, at 2:39 am, arthur.hk.c...@gmail.com 
>> <arthur.hk.c...@gmail.com> wrote:
>> 
>>> Hi,
>>> 
>>> I use Hadoop 2.4.1 and HBase 0.98.5 with snappy enabled in both Hadoop and 
>>> HBase.
>>> With default setting in Spark 1.0.2, when trying to load a file I got 
>>> "Class org.apache.hadoop.io.compress.SnappyCodec not found"
>>> 
>>> Can you please advise how to enable snappy in Spark?
>>> 
>>> Regards
>>> Arthur
>>> 
>>> 
>>> scala> inFILE.first()
>>> java.lang.RuntimeException: Error in configuring object
>>>     at 
>>> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109)
>>>     at 
>>> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:75)
>>>     at 
>>> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
>>>     at org.apache.spark.rdd.HadoopRDD.getInputFormat(HadoopRDD.scala:158)
>>>     at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:171)
>>>     at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204)
>>>     at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202)
>>>     at scala.Option.getOrElse(Option.scala:120)
>>>     at org.apache.spark.rdd.RDD.partitions(RDD.scala:202)
>>>     at org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28)
>>>     at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204)
>>>     at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202)
>>>     at scala.Option.getOrElse(Option.scala:120)
>>>     at org.apache.spark.rdd.RDD.partitions(RDD.scala:202)
>>>     at org.apache.spark.rdd.RDD.take(RDD.scala:983)
>>>     at org.apache.spark.rdd.RDD.first(RDD.scala:1015)
>>>     at $iwC$$iwC$$iwC$$iwC.<init>(<console>:15)
>>>     at $iwC$$iwC$$iwC.<init>(<console>:20)
>>>     at $iwC$$iwC.<init>(<console>:22)
>>>     at $iwC.<init>(<console>:24)
>>>     at <init>(<console>:26)
>>>     at .<init>(<console>:30)
>>>     at .<clinit>(<console>)
>>>     at .<init>(<console>:7)
>>>     at .<clinit>(<console>)
>>>     at $print(<console>)
>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>     at 
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>>     at 
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>     at java.lang.reflect.Method.invoke(Method.java:597)
>>>     at 
>>> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:788)
>>>     at 
>>> org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1056)
>>>     at 
>>> org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:614)
>>>     at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:645)
>>>     at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:609)
>>>     at 
>>> org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:796)
>>>     at 
>>> org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:841)
>>>     at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:753)
>>>     at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:601)
>>>     at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:608)
>>>     at org.apache.spark.repl.SparkILoop.loop(SparkILoop.scala:611)
>>>     at 
>>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:936)
>>>     at 
>>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)
>>>     at 
>>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)
>>>     at 
>>> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
>>>     at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:884)
>>>     at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:982)
>>>     at org.apache.spark.repl.Main$.main(Main.scala:31)
>>>     at org.apache.spark.repl.Main.main(Main.scala)
>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>     at 
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>>     at 
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>     at java.lang.reflect.Method.invoke(Method.java:597)
>>>     at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:303)
>>>     at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:55)
>>>     at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>>> Caused by: java.lang.reflect.InvocationTargetException
>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>     at 
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>>     at 
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>     at java.lang.reflect.Method.invoke(Method.java:597)
>>>     at 
>>> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:106)
>>>     ... 55 more
>>> Caused by: java.lang.IllegalArgumentException: Compression codec       
>>> org.apache.hadoop.io.compress.SnappyCodec not found.
>>>     at 
>>> org.apache.hadoop.io.compress.CompressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:135)
>>>     at 
>>> org.apache.hadoop.io.compress.CompressionCodecFactory.<init>(CompressionCodecFactory.java:175)
>>>     at 
>>> org.apache.hadoop.mapred.TextInputFormat.configure(TextInputFormat.java:45)
>>>     ... 60 more
>>> Caused by: java.lang.ClassNotFoundException: Class       
>>> org.apache.hadoop.io.compress.SnappyCodec not found
>>>     at 
>>> org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1801)
>>>     at 
>>> org.apache.hadoop.io.compress.CompressionCodecFactory.getCodecClasses(CompressionCodecFactory.java:128)
>>>     ... 62 more
>>> 
>>> 
>> 
> 

Reply via email to