*I am getting the following exception when the HS2 is crashing, any idea why it has happening*
"pool-1-thread-121" prio=4 tid=19283 RUNNABLE at java.lang.OutOfMemoryError.<init>(OutOfMemoryError.java:48) at java.util.Arrays.copyOf(Arrays.java:2271) Local Variable: byte[]#1 at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:113) at java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutput Stream.java:93) at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:140) Local Variable: org.apache.thrift.TByteArrayOutputStream#42 Local Variable: byte[]#5378 at org.apache.thrift.transport.TSaslTransport.write(TSaslTransport.java:446) at org.apache.thrift.transport.TSaslServerTransport.write(TSasl ServerTransport.java:41) at org.apache.thrift.protocol.TBinaryProtocol.writeI32(TBinaryP rotocol.java:163) at org.apache.thrift.protocol.TBinaryProtocol.writeString(TBina ryProtocol.java:186) Local Variable: byte[]#2 at org.apache.hive.service.cli.thrift.TStringColumn$TStringColu mnStandardScheme.write(TStringColumn.java:490) Local Variable: java.util.ArrayList$Itr#1 at org.apache.hive.service.cli.thrift.TStringColumn$TStringColu mnStandardScheme.write(TStringColumn.java:433) Local Variable: org.apache.hive.service.cli.thrift.TStringColumn$TStringColu mnStandardScheme#1 at org.apache.hive.service.cli.thrift.TStringColumn.write(TStri ngColumn.java:371) at org.apache.hive.service.cli.thrift.TColumn.standardSchemeWri teValue(TColumn.java:381) Local Variable: org.apache.hive.service.cli.thrift.TColumn#504 Local Variable: org.apache.hive.service.cli.thrift.TStringColumn#453 at org.apache.thrift.TUnion$TUnionStandardScheme.write(TUnion.java:244) at org.apache.thrift.TUnion$TUnionStandardScheme.write(TUnion.java:213) at org.apache.thrift.TUnion.write(TUnion.java:152) On Fri, Aug 21, 2015 at 6:16 AM, kulkarni.swar...@gmail.com < kulkarni.swar...@gmail.com> wrote: > Sanjeev, > > One possibility is that you are running into[1] which affects hive 0.13. > Is it possible for you to apply the patch on [1] and see if it fixes your > problem? > > [1] https://issues.apache.org/jira/browse/HIVE-10410 > > On Thu, Aug 20, 2015 at 6:12 PM, Sanjeev Verma <sanjeev.verm...@gmail.com> > wrote: > >> We are using hive-0.13 with hadoop1. >> >> On Thu, Aug 20, 2015 at 11:49 AM, kulkarni.swar...@gmail.com < >> kulkarni.swar...@gmail.com> wrote: >> >>> Sanjeev, >>> >>> Can you tell me more details about your hive version/hadoop version etc. >>> >>> On Wed, Aug 19, 2015 at 1:35 PM, Sanjeev Verma < >>> sanjeev.verm...@gmail.com> wrote: >>> >>>> Can somebody gives me some pointer to looked upon? >>>> >>>> On Wed, Aug 19, 2015 at 9:26 AM, Sanjeev Verma < >>>> sanjeev.verm...@gmail.com> wrote: >>>> >>>>> Hi >>>>> We are experiencing a strange problem with the hiveserver2, in one of >>>>> the job it gets the GC limit exceed from mapred task and hangs even having >>>>> enough heap available.we are not able to identify what causing this issue. >>>>> Could anybody help me identify the issue and let me know what pointers >>>>> I need to looked up. >>>>> >>>>> Thanks >>>>> >>>> >>>> >>> >>> >>> -- >>> Swarnim >>> >> >> > > > -- > Swarnim >