[ https://issues.apache.org/jira/browse/SPARK-37751?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Shipeng Feng updated SPARK-37751: --------------------------------- Summary: Apache Commons Crypto doesn't support Java 11 (was: Apache Commons Crypto Doesn't support Java 11) > Apache Commons Crypto doesn't support Java 11 > --------------------------------------------- > > Key: SPARK-37751 > URL: https://issues.apache.org/jira/browse/SPARK-37751 > Project: Spark > Issue Type: Bug > Components: Security > Affects Versions: 3.1.2, 3.2.0 > Environment: Spark 3.2.0 on kubernetes > Reporter: Shipeng Feng > Priority: Major > > For kubernetes, we are using Java 11 in docker, > https://github.com/apache/spark/blob/v3.2.0/resource-managers/kubernetes/docker/src/main/dockerfiles/spark/Dockerfile: > {code:java} > ARG java_image_tag=11-jre-slim > {code} > Enable AES-based encryption for RPC connection by the following config: > {code:java} > --conf spark.authenticate=true > --conf spark.network.crypto.enabled=true > {code} > This would cause the following error: > {code:java} > java.lang.ArrayIndexOutOfBoundsException: Index -453 out of bounds for length > 16 > at java.base/com.sun.crypto.provider.CounterMode.implCrypt(Unknown Source) > at java.base/com.sun.crypto.provider.CounterMode.crypt(Unknown Source) > at java.base/com.sun.crypto.provider.CounterMode.decrypt(Unknown Source) > at java.base/com.sun.crypto.provider.CipherCore.update(Unknown Source) > at java.base/com.sun.crypto.provider.CipherCore.update(Unknown Source) > at java.base/com.sun.crypto.provider.AESCipher.engineUpdate(Unknown > Source) > at java.base/javax.crypto.CipherSpi.bufferCrypt(Unknown Source) > at java.base/javax.crypto.CipherSpi.engineUpdate(Unknown Source) > at java.base/javax.crypto.Cipher.update(Unknown Source) > at org.apache.commons.crypto.cipher.JceCipher.update(JceCipher.java:120) > at > org.apache.commons.crypto.stream.CryptoInputStream.decrypt(CryptoInputStream.java:526) > at > org.apache.commons.crypto.stream.CryptoInputStream.decryptMore(CryptoInputStream.java:508) > at > org.apache.commons.crypto.stream.CryptoInputStream.read(CryptoInputStream.java:263) > at > org.apache.spark.network.crypto.TransportCipher$DecryptionHandler.channelRead(TransportCipher.java:183) > at > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) > at > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) > at > io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) > at > io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) > at > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) > at > io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) > at > io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) > at > io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) > at > io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:719) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:655) > at > io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:581) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) > at > io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:986) > at > io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > at > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > at java.base/java.lang.Thread.run(Unknown Source) > {code} > It seems that Apache Commons Crypto 1.1.0 only works with Java 8: > [https://commons.apache.org/proper/commons-crypto/download_crypto.cgi] -- This message was sent by Atlassian Jira (v8.20.1#820001) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org