[ 
https://issues.apache.org/jira/browse/SPARK-4079?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14191013#comment-14191013
 ] 

Kostas Sakellis commented on SPARK-4079:
----------------------------------------

[~pwendell] I'm in favor of checking if Snappy is supported and if not, log a 
warning and switch to LZF. I feel this is a better user experience than just 
throwing an exception and telling the user to go modify 
"spark.io.compression.codec".  Especially since this is a performance issue not 
a correctness one. Thoughts?

> Snappy bundled with Spark does not work on older Linux distributions
> --------------------------------------------------------------------
>
>                 Key: SPARK-4079
>                 URL: https://issues.apache.org/jira/browse/SPARK-4079
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.0.0
>            Reporter: Marcelo Vanzin
>
> This issue has existed at least since 1.0, but has been made worse by 1.1 
> since snappy is now the default compression algorithm. When trying to use it 
> on a CentOS 5 machine, for example, you'll get something like this:
> {noformat}
>       java.lang.reflect.InvocationTargetException
>            at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>            at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>            at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>            at java.lang.reflect.Method.invoke(Method.java:606)
>            at 
> org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:319)
>            at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:226)
>            at org.xerial.snappy.Snappy.<clinit>(Snappy.java:48)
>            at 
> org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:79)
>            at 
> org.apache.spark.io.SnappyCompressionCodec.compressedOutputStream(CompressionCodec.scala:125)
>            at 
> org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:207)
>        ...
>        Caused by: java.lang.UnsatisfiedLinkError: 
> /tmp/snappy-1.0.5.3-af72bf3c-9dab-43af-a662-f9af657f06b1-libsnappyjava.so: 
> /usr/lib64/libstdc++.so.6: version `GLIBCXX_3.4.9' not found (required by 
> /tmp/snappy-1.0.5.3-af72bf3c-9dab-43af-a662-f9af657f06b1-libsnappyjava.so)
>            at java.lang.ClassLoader$NativeLibrary.load(Native Method)
>            at java.lang.ClassLoader.loadLibrary1(ClassLoader.java:1957)
>            at java.lang.ClassLoader.loadLibrary0(ClassLoader.java:1882)
>            at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1843)
>            at java.lang.Runtime.load0(Runtime.java:795)
>            at java.lang.System.load(System.java:1061)
>            at 
> org.xerial.snappy.SnappyNativeLoader.load(SnappyNativeLoader.java:39)
>            ... 29 more
> {noformat}
> There are two approaches I can see here (well, 3):
> * Declare CentOS 5 (and similar OSes) not supported, although that would suck 
> for the people who are still on it and already use Spark
> * Fallback to another compression codec if Snappy cannot be loaded
> * Ask the Snappy guys to compile the library on an older OS...
> I think the second would be the best compromise.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to