We have 8GB HS2 java heap, we have not tried any bumping. On Tue, Sep 8, 2015 at 8:14 PM, kulkarni.swar...@gmail.com < kulkarni.swar...@gmail.com> wrote:
> How much memory have you currently provided to HS2? Have you tried bumping > that up? > > On Mon, Sep 7, 2015 at 1:09 AM, Sanjeev Verma <sanjeev.verm...@gmail.com> > wrote: > >> *I am getting the following exception when the HS2 is crashing, any idea >> why it has happening* >> >> "pool-1-thread-121" prio=4 tid=19283 RUNNABLE >> at java.lang.OutOfMemoryError.<init>(OutOfMemoryError.java:48) >> at java.util.Arrays.copyOf(Arrays.java:2271) >> Local Variable: byte[]#1 >> at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:113) >> at java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutput >> Stream.java:93) >> at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:140) >> Local Variable: org.apache.thrift.TByteArrayOutputStream#42 >> Local Variable: byte[]#5378 >> at org.apache.thrift.transport.TSaslTransport.write(TSaslTransp >> ort.java:446) >> at org.apache.thrift.transport.TSaslServerTransport.write(TSasl >> ServerTransport.java:41) >> at org.apache.thrift.protocol.TBinaryProtocol.writeI32(TBinaryP >> rotocol.java:163) >> at org.apache.thrift.protocol.TBinaryProtocol.writeString(TBina >> ryProtocol.java:186) >> Local Variable: byte[]#2 >> at org.apache.hive.service.cli.thrift.TStringColumn$TStringColu >> mnStandardScheme.write(TStringColumn.java:490) >> Local Variable: java.util.ArrayList$Itr#1 >> at org.apache.hive.service.cli.thrift.TStringColumn$TStringColu >> mnStandardScheme.write(TStringColumn.java:433) >> Local Variable: org.apache.hive.service.cli.th >> rift.TStringColumn$TStringColumnStandardScheme#1 >> at org.apache.hive.service.cli.thrift.TStringColumn.write(TStri >> ngColumn.java:371) >> at org.apache.hive.service.cli.thrift.TColumn.standardSchemeWri >> teValue(TColumn.java:381) >> Local Variable: org.apache.hive.service.cli.thrift.TColumn#504 >> Local Variable: org.apache.hive.service.cli.thrift.TStringColumn#453 >> at org.apache.thrift.TUnion$TUnionStandardScheme.write(TUnion.java:244) >> at org.apache.thrift.TUnion$TUnionStandardScheme.write(TUnion.java:213) >> at org.apache.thrift.TUnion.write(TUnion.java:152) >> >> >> >> On Fri, Aug 21, 2015 at 6:16 AM, kulkarni.swar...@gmail.com < >> kulkarni.swar...@gmail.com> wrote: >> >>> Sanjeev, >>> >>> One possibility is that you are running into[1] which affects hive 0.13. >>> Is it possible for you to apply the patch on [1] and see if it fixes your >>> problem? >>> >>> [1] https://issues.apache.org/jira/browse/HIVE-10410 >>> >>> On Thu, Aug 20, 2015 at 6:12 PM, Sanjeev Verma < >>> sanjeev.verm...@gmail.com> wrote: >>> >>>> We are using hive-0.13 with hadoop1. >>>> >>>> On Thu, Aug 20, 2015 at 11:49 AM, kulkarni.swar...@gmail.com < >>>> kulkarni.swar...@gmail.com> wrote: >>>> >>>>> Sanjeev, >>>>> >>>>> Can you tell me more details about your hive version/hadoop version >>>>> etc. >>>>> >>>>> On Wed, Aug 19, 2015 at 1:35 PM, Sanjeev Verma < >>>>> sanjeev.verm...@gmail.com> wrote: >>>>> >>>>>> Can somebody gives me some pointer to looked upon? >>>>>> >>>>>> On Wed, Aug 19, 2015 at 9:26 AM, Sanjeev Verma < >>>>>> sanjeev.verm...@gmail.com> wrote: >>>>>> >>>>>>> Hi >>>>>>> We are experiencing a strange problem with the hiveserver2, in one >>>>>>> of the job it gets the GC limit exceed from mapred task and hangs even >>>>>>> having enough heap available.we are not able to identify what causing >>>>>>> this >>>>>>> issue. >>>>>>> Could anybody help me identify the issue and let me know what >>>>>>> pointers I need to looked up. >>>>>>> >>>>>>> Thanks >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Swarnim >>>>> >>>> >>>> >>> >>> >>> -- >>> Swarnim >>> >> >> > > > -- > Swarnim >