Please take a look at SPARK-1867.

The discussion was very long.
You may want to look for missing classes.

Also see https://bugs.openjdk.java.net/browse/JDK-7172206

On Thu, Feb 4, 2016 at 10:31 AM, <arun.bong...@cognizant.com> wrote:

> Hi Ted. Thanks for the response.
>
> i'm just trying to do a select *. the table has 1+ million rows.
>
> I have set below parameters.
>
> export SPARK_EXECUTOR_MEMORY=4G
> export SPARK_DRIVER_MEMORY=2G
> spark.kryoserializer.buffer.max 2000m.
>
>
> I have started the thrift server on port 10001 and trying to access these
> spark tables from qlikview BI tools.
> I have been stuck with this. Kinldly help.
>
> PFB the logs.
>
> 16/02/04 18:11:21 INFO TaskSetManager: Finished task 9.0 in stage 6.0 (TID
> 57) in 4305 ms on ip-xx-xx-xx-xx.ec2.internal (8/10)
> 16/02/04 18:11:26 INFO TaskSetManager: Finished task 7.0 in stage 6.0 (TID
> 55) in 14711 ms on ip-xx-xx-xx-xx.ec2.internal (9/10)
> #
> # java.lang.OutOfMemoryError: Java heap space
> # -XX:OnOutOfMemoryError="kill -9 %p
> kill -9 %p"
> #   Executing /bin/sh -c "kill -9 17242
> kill -9 17242"...
> 16/02/04 18:11:39 ERROR TransportRequestHandler: Error while invoking
> RpcHandler#receive() for one-way message.
> java.lang.IllegalStateException: unread block data
>         at
> java.io.ObjectInputStream$BlockDataInputStream.setBlockDataMode(ObjectInputStream.java:2428)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1382)
>         at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1997)
>         at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1921)
>         at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>         at
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>         at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>         at
> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:76)
>         at
> org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:109)
>         at
> org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$deserialize$1$$anonfun$apply$1.apply(NettyRpcEnv.scala:258)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
>         at
> org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(NettyRpcEnv.scala:310)
>         at
> org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$deserialize$1.apply(NettyRpcEnv.scala:257)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
>         at
> org.apache.spark.rpc.netty.NettyRpcEnv.deserialize(NettyRpcEnv.scala:256)
>         at
> org.apache.spark.rpc.netty.NettyRpcHandler.internalReceive(NettyRpcEnv.scala:588)
>         at
> org.apache.spark.rpc.netty.NettyRpcHandler.receive(NettyRpcEnv.scala:577)
>         at
> org.apache.spark.network.server.TransportRequestHandler.processOneWayMessage(TransportRequestHandler.java:170)
>         at
> org.apache.spark.network.server.TransportRequestHandler.handle(TransportRequestHandler.java:104)
>         at
> org.apache.spark.network.server.TransportChannelHandler.channelRead0(TransportChannelHandler.java:104)
>         at
> org.apache.spark.network.server.TransportChannelHandler.channelRead0(TransportChannelHandler.java:51)
>         at
> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
>         at
> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
>         at
> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
>         at
> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:266)
>         at
> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
>         at
> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
>         at
> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
>         at
> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
>         at
> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
>         at
> org.apache.spark.network.util.TransportFrameDecoder.channelRead(TransportFrameDecoder.java:86)
>         at
> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
>         at
> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
>         at
> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846)
>         at
> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
>         at
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511)
>         at
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)
>         at
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354)
>         at
> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:111)
>         at java.lang.Thread.run(Thread.java:745)
>
>
>
> Thanks
> Arun
>
> ------------------------------
> *From:* Ted Yu [yuzhih...@gmail.com]
> *Sent:* 04 February 2016 23:37
> *To:* BONGALE, ARUN (Cognizant)
> *Cc:* user
> *Subject:* Re: Memory tuning in spark sql
>
> Can you provide a bit more detail ?
>
> values of the parameters you have tuned
> log snippets from executors
> snippet of your code
>
> Thanks
>
> On Thu, Feb 4, 2016 at 9:48 AM, <arun.bong...@cognizant.com
> <http://redir.aspx?REF=Cvgtpa7SYatX8coIXyW5Vsnc-ZSwAOo_o6sBm3hEmEIBJpaNkC3TCAFtYWlsdG86QVJVTi5CT05HQUxFQGNvZ25pemFudC5jb20.>
> > wrote:
>
>> Hi Sir/madam,
>> Greetings of the day.
>>
>> I am working on Spark 1.6.0 with AWS EMR(Elastic Map Reduce). I'm facing
>> some issues in reading large(500 mb) file in spark-sql.
>> Sometimes i get heap space error and sometimes the executors fail.
>> i have increased the driver memory, executor memory, kryo serializer
>> buffer size.. etc.. but nothing helps.
>>
>> I Kindly request your help in resolving this issue.
>>
>> Thanks
>> Arun.
>> This e-mail and any files transmitted with it are for the sole use of the
>> intended recipient(s) and may contain confidential and privileged
>> information. If you are not the intended recipient(s), please reply to the
>> sender and destroy all copies of the original message. Any unauthorized
>> review, use, disclosure, dissemination, forwarding, printing or copying of
>> this email, and/or any action taken in reliance on the contents of this
>> e-mail is strictly prohibited and may be unlawful. Where permitted by
>> applicable law, this e-mail and other e-mail communications sent to and
>> from Cognizant e-mail addresses may be monitored.
>>
>
> This e-mail and any files transmitted with it are for the sole use of the
> intended recipient(s) and may contain confidential and privileged
> information. If you are not the intended recipient(s), please reply to the
> sender and destroy all copies of the original message. Any unauthorized
> review, use, disclosure, dissemination, forwarding, printing or copying of
> this email, and/or any action taken in reliance on the contents of this
> e-mail is strictly prohibited and may be unlawful. Where permitted by
> applicable law, this e-mail and other e-mail communications sent to and
> from Cognizant e-mail addresses may be monitored.
>

Reply via email to