[ 
https://issues.apache.org/jira/browse/SPARK-36669?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-36669:
------------------------------------

    Assignee: Apache Spark

> Fail to load Lz4 codec
> ----------------------
>
>                 Key: SPARK-36669
>                 URL: https://issues.apache.org/jira/browse/SPARK-36669
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 3.2.0
>            Reporter: L. C. Hsieh
>            Assignee: Apache Spark
>            Priority: Blocker
>
> Currently we use Hadop 3.3.1's shaded client libraries. Lz4 is a provided 
> dependency in Hadoop Common 3.3.1 for Lz4Codec. But it isn't excluded from 
> relocation in these libraries. So to use lz4 as Parquet codec, we will hit 
> the exception even we include lz4 as dependency.
> {code:java}
> [info]   Cause: java.lang.NoClassDefFoundError: 
> org/apache/hadoop/shaded/net/jpountz/lz4/LZ4Factory                           
>                                                                  
> [info]   at 
> org.apache.hadoop.io.compress.lz4.Lz4Compressor.<init>(Lz4Compressor.java:66)
> [info]   at 
> org.apache.hadoop.io.compress.Lz4Codec.createCompressor(Lz4Codec.java:119)    
>                                                                               
>                        
> [info]   at 
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:152)     
>                                                                               
>                        
> [info]   at 
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:168)     
>                                                                               
>                        
>  {code}
>  
> I already submitted a PR to Hadoop to fix it. Before it is released, at Spark 
> side, we either downgrade to 3.3.0 or revert back to non-shaded hadoop client 
> library.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to