Hi Adarsh,

It may be because of wrong configuration to meta store server/lack of access
rights .

Vaibhav Negi


On Wed, Jan 5, 2011 at 11:04 PM, Jean-Daniel Cryans <jdcry...@apache.org>wrote:

> With one cluster you really only need one, and it doesn't seem to be
> running from what I can tell:
>
> 2011-01-05 15:20:12,185 WARN  zookeeper.ClientCnxn
> (ClientCnxn.java:run(967)) - Exception closing session 0x0 to
> sun.nio.ch.selectionkeyi...@561279c8
> java.net.ConnectException: Connection refused
>
> And this is only the tail of the log, the head will tell you where
> it's trying to connect. My guess is that there's either a problem with
> your hbase configuration for hive, or the ZK peers aren't running, or
> both issues at the same time. Although if you can see that HBase is
> already running properly, then it must be a configuration issue.
>
> J-D
>
> On Wed, Jan 5, 2011 at 2:14 AM, Adarsh Sharma <adarsh.sha...@orkash.com>
> wrote:
> >
> >
> >
> >
> > Dear all,
> >
> > I am trying Hive/Hbase Integration from the past 2 days. I am facing the
> > below issue while creating external table in Hive.
> >
> > *Command-Line Error :-
> >
> > *had...@s2-ratw-1:~/project/hive-0.6.0/build/dist$ bin/hive --auxpath
> >
> /home/hadoop/project/hive-0.6.0/build/dist/lib/hive_hbase-handler.jar,/home/hadoop/project/hive-0.6.0/build/dist/lib/hbase-0.20.3.jar,/home/hadoop/project/hive-0.6.0/build/dist/lib/zookeeper-3.2.2.jar
> >  -hiveconf
> >
> hbase.zookeeper.quorum=192.168.1.103,192.168.1.114,192.168.1.115,192.168.1.104,192.168.1.107
> > Hive history
> > file=/tmp/hadoop/hive_job_log_hadoop_201101051527_1728376885.txt
> > hive> show tables;
> > FAILED: Error in metadata: javax.jdo.JDOFatalDataStoreException:
> > Communications link failure
> >
> > The last packet sent successfully to the server was 0 milliseconds ago.
> The
> > driver has not received any packets from the server.
> > NestedThrowables:
> > com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications
> link
> > failure
> >
> > The last packet sent successfully to the server was 0 milliseconds ago.
> The
> > driver has not received any packets from the server.
> > FAILED: Execution Error, return code 1 from
> > org.apache.hadoop.hive.ql.exec.DDLTask
> > hive> exit;
> > had...@s2-ratw-1:~/project/hive-0.6.0/build/dist$
> >
> > *My hive.log file says :*
> >
> > 2011-01-05 15:19:36,783 ERROR DataNucleus.Plugin
> > (Log4JLogger.java:error(115)) - Bundle "org.eclipse.jdt.core" requires
> > "org.eclipse.core.resources" but it cannot be resolved.
> > 2011-01-05 15:19:36,783 ERROR DataNucleus.Plugin
> > (Log4JLogger.java:error(115)) - Bundle "org.eclipse.jdt.core" requires
> > "org.eclipse.core.resources" but it cannot be resolved.
> > 2011-01-05 15:19:36,785 ERROR DataNucleus.Plugin
> > (Log4JLogger.java:error(115)) - Bundle "org.eclipse.jdt.core" requires
> > "org.eclipse.core.runtime" but it cannot be resolved.
> > 2011-01-05 15:19:36,785 ERROR DataNucleus.Plugin
> > (Log4JLogger.java:error(115)) - Bundle "org.eclipse.jdt.core" requires
> > "org.eclipse.core.runtime" but it cannot be resolved.
> > 2011-01-05 15:19:36,786 ERROR DataNucleus.Plugin
> > (Log4JLogger.java:error(115)) - Bundle "org.eclipse.jdt.core" requires
> > "org.eclipse.text" but it cannot be resolved.
> > 2011-01-05 15:19:36,786 ERROR DataNucleus.Plugin
> > (Log4JLogger.java:error(115)) - Bundle "org.eclipse.jdt.core" requires
> > "org.eclipse.text" but it cannot be resolved.
> > 2011-01-05 15:20:12,185 WARN  zookeeper.ClientCnxn
> > (ClientCnxn.java:run(967)) - Exception closing session 0x0 to
> > sun.nio.ch.selectionkeyi...@561279c8
> > java.net.ConnectException: Connection refused
> >       at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> >       at
> > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574)
> >       at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:933)
> > 2011-01-05 15:20:12,188 WARN  zookeeper.ClientCnxn
> > (ClientCnxn.java:cleanup(1001)) - Ignoring exception during shutdown
> input
> > java.nio.channels.ClosedChannelException
> >       at
> > sun.nio.ch.SocketChannelImpl.shutdownInput(SocketChannelImpl.java:638)
> >       at sun.nio.ch.SocketAdaptor.shutdownInput(SocketAdaptor.java:360)
> >       at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:999)
> >       at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2011-01-05 15:20:12,188 WARN  zookeeper.ClientCnxn
> > (ClientCnxn.java:cleanup(1006)) - Ignoring exception during shutdown
> output
> > java.nio.channels.ClosedChannelException
> >       at
> > sun.nio.ch.SocketChannelImpl.shutdownOutput(SocketChannelImpl.java:649)
> >       at sun.nio.ch.SocketAdaptor.shutdownOutput(SocketAdaptor.java:368)
> >       at
> > org.apache.zookeeper.ClientCnxn$SendThread.cleanup(ClientCnxn.java:1004)
> >       at
> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:970)
> > 2011-01-05 15:20:12,621 WARN  zookeeper.ClientCnxn
> > (ClientCnxn.java:run(967)) - Exception closing session 0x0 to
> > sun.nio.ch.selectionkeyi...@799dbc3b
> >
> > I overcomed from the previous issue of MasterNotRunning Exception which
> > occured due to incompatibilities in hive_hbase jars.
> >
> > Now I'm using Hadoop-0.20.2, Hive-0.6.0 ( Bydefault Derby metastore  )
> and
> > Hbase-0.20.3.
> >
> > Please tell how this could be resolved.
> >
> > Also I want to add one more thing that my hadoop Cluster is of 9 nodes
> and 8
> > nodes act as Datanodes,Tasktrackers and Regionservers.
> >
> > Among these nodes is set zookeeper.quorum.property to have 5 Datanodes.
> > Would this is the issue.
> > I don't know the number of servers needed for Zookeeper in fully
> distributed
> > mode.
> >
> >
> > Best Regards
> >
> > Adarsh Sharma
> >
> >
> >
>

Reply via email to