[
https://issues.apache.org/jira/browse/HIVE-6866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14127451#comment-14127451
]
Ankita Bakshi commented on HIVE-6866:
-------------------------------------
We are facing same issue in production. We are using CDH4.4 with Apache Hive
0.12. Is there a workaround for this issue other than restarting hiveserver2?
> Hive server2 jdbc driver connection leak with namenode
> ------------------------------------------------------
>
> Key: HIVE-6866
> URL: https://issues.apache.org/jira/browse/HIVE-6866
> Project: Hive
> Issue Type: Bug
> Affects Versions: 0.11.0
> Reporter: Shengjun Xin
>
> 1. Set 'ipc.client.connection.maxidletime' to 3600000 in core-site.xml and
> start hive-server2.
> 2. Connect hive server2 repetitively in a while true loop.
> 3. The tcp connection number will increase until out of memory, it seems that
> hive server2 will not close the connection until the time out, the error
> message is as the following:
> {code}
> 2014-03-18 23:30:36,873 ERROR ql.Driver (SessionState.java:printError(386)) -
> FAILED: RuntimeException java.io.IOException: Failed on local exception:
> java.io.IOException: Couldn't set up IO streams; Host Details : local host
> is: "hdm1.hadoop.local/192.168.2.101"; destination host is:
> "hdm1.hadoop.local":8020;
> java.lang.RuntimeException: java.io.IOException: Failed on local exception:
> java.io.IOException: Couldn't set up IO streams; Host Details : local host
> is: "hdm1.hadoop.local/192.168.2.101"; destination host is:
> "hdm1.hadoop.local":8020;
> at org.apache.hadoop.hive.ql.Context.getScratchDir(Context.java:190)
> at org.apache.hadoop.hive.ql.Context.getMRScratchDir(Context.java:231)
> at org.apache.hadoop.hive.ql.Context.getMRTmpFileURI(Context.java:288)
> at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1274)
> at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:1059)
> at
> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:8676)
> at
> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:278)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:433)
> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:337)
> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:902)
> at
> org.apache.hive.service.cli.operation.SQLOperation.run(SQLOperation.java:95)
> at
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:181)
> at
> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:148)
> at
> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:203)
> at
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1133)
> at
> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1118)
> at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
> at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
> at
> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:40)
> at
> org.apache.hive.service.auth.TUGIContainingProcessor$1.run(TUGIContainingProcessor.java:37)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1478)
> at
> org.apache.hadoop.hive.shims.HadoopShimsSecure.doAs(HadoopShimsSecure.java:524)
> at
> org.apache.hive.service.auth.TUGIContainingProcessor.process(TUGIContainingProcessor.java:37)
> at
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:744)
> Caused by: java.io.IOException: Failed on local exception:
> java.io.IOException: Couldn't set up IO streams; Host Details : local host
> is: "hdm1.hadoop.local/192.168.2.101"; destination host is:
> "hdm1.hadoop.local":8020;
> at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:761)
> at org.apache.hadoop.ipc.Client.call(Client.java:1239)
> at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
> at com.sun.proxy.$Proxy11.mkdirs(Unknown Source)
> at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
> at
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
> at com.sun.proxy.$Proxy11.mkdirs(Unknown Source)
> at
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:483)
> at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2259)
> at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2230)
> at
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:540)
> at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:1881)
> at org.apache.hadoop.hive.ql.Context.getScratchDir(Context.java:182)
> ... 28 more
> Caused by: java.io.IOException: Couldn't set up IO streams
> at
> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:662)
> at org.apache.hadoop.ipc.Client$Connection.access$2100(Client.java:253)
> at org.apache.hadoop.ipc.Client.getConnection(Client.java:1288)
> at org.apache.hadoop.ipc.Client.call(Client.java:1206)
> ... 42 more
> Caused by: java.lang.OutOfMemoryError: unable to create new native thread
> at java.lang.Thread.start0(Native Method)
> at java.lang.Thread.start(Thread.java:713)
> at
> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:655)
> ... 45 more
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)