I think you should make sure that the HDFS is running normally by
checking its log files.

And, firewalls? (just wild guesses)

-Kevin

On 5月27日, 下午2时20分, Harshada <[email protected]> wrote:
> Thank you for the reply.
>
> First of all I am sorry for posting this query to the wrong thread. It
> you can, please migrate it to the -user mailing list.
>
> I checked the log file for DfsBroker.hadoop, it says:
>
> Num CPUs=2
> HdfsBroker.Port=38030
> HdfsBroker.Reactors=2
> HdfsBroker.Workers=20
> HdfsBroker.Server.fs.default.name=hdfs://localhost:54310
> 10/05/27 05:01:58 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 0 time(s).
> 10/05/27 05:01:59 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 1 time(s).
> 10/05/27 05:02:00 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 2 time(s).
> 10/05/27 05:02:01 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 3 time(s).
> 10/05/27 05:02:02 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 4 time(s).
> 10/05/27 05:02:03 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 5 time(s).
> 10/05/27 05:02:04 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 6 time(s).
> 10/05/27 05:02:05 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 7 time(s).
> 10/05/27 05:02:06 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 8 time(s).
> 10/05/27 05:02:07 INFO ipc.Client: Retrying connect to server:
> localhost/127.0.0.1:54310. Already tried 9 time(s).
> 27 May, 2010 5:02:07 AM org.hypertable.DfsBroker.hadoop.HdfsBroker
> <init>
> SEVERE: ERROR: Unable to establish connection to HDFS.
> ShutdownHook called
> Exception in thread "Thread-1" java.lang.NullPointerException
>         at org.hypertable.DfsBroker.hadoop.main
> $ShutdownHook.run(main.java:69)
>
> ---------------------------
>
> but hdfs is running. because jps on master gives me following output:
>
> e...@erts-server:~$ jps
> 32538 SecondaryNameNode
> 32270 NameNode
> 32388 DataNode
> 310 TaskTracker
> 32671 JobTracker
> 21233 Jps
> -----------------------------------------------------
>
> > Is there a reason you're using 0.9.2.8 and not 0.9.3.1 (the latest and
> > greatest) ?
>
> Oh.. ok. thanks for the info. But since 0.9.2.8 was successfully
> installed, I ll continue with it for the moment.
>
> > Do you have HDFS running and if so make sure the permissions for the
> > /hypertable dir are set correctly.
>
> yes. I followedhttp://code.google.com/p/hypertable/wiki/UpAndRunningWithHadoop
> andhttp://code.google.com/p/hypertable/wiki/DeployingHypertable.
>
> A doubt: do I always need the owner of slave and master machine to be
> the same? coz currently I have 'erts' as the user for master and one
> slave (which are on the same machine) and harshada as the user on
> slave. So what happens is, whenever I use '$cap dist' or '$cap shell
> cap>date' it asks for password of e...@slave which does not exist
> hence authentication fails. I am in the process of getting same user
> on all the machines, but till then I thought of getting max things up.
> Is this the reason why DfsBroker.hadoop is also failing?
>
> If yes, then I should better wait and have same user on all the
> machines.
>
> PS: though hadoop require same installation paths on all the machines,
> I managed it with symbolic links, though my users (and hence their
> $HOME s ) were different.
>
>
>
> > Beyond that, try taking a look at <HT_INSTALL_DIR>/log/DfsBroker.hadoop.log
> > to figure out whats going on.
>
> > -Sanjit
>
> > On Wed, May 26, 2010 at 4:35 PM, Harshada <[email protected]> wrote:
> > > Hi,
>
> > > I am installing Hypertable 0.9.2.8 on Hadoop. I have successfully set
> > > up Hadoop and its working. When I start servers using 'cap start', DFS
> > > Server doesn't come up. The output of cap start is:
>
> > > e...@erts-server:~/hypertable/hypertable-0.9.2.8-alpha/conf$ cap start
> > >  * executing `start'
> > >  ** transaction: start
> > >  * executing `start_hyperspace'
> > >  * executing "/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/
> > > bin/start-hyperspace.sh <http://0.9.2.8/%0Abin/start-hyperspace.sh>
> > > --config=/home/erts/hypertable/
> > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg"
> > >    servers: ["127.0.0.1"]
> > >    [127.0.0.1] executing command
> > >  ** [out :: 127.0.0.1] Hyperspace appears to be running (12170):
> > >  ** [out :: 127.0.0.1] erts 12170 1 0 04:40 ? 00:00:00 /home/erts/
> > > hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/bin/Hyperspace.Master --
> > > pidfile /home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/run/
> > > Hyperspace.pid --verbose --config=/home/erts/hypertable/
> > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg
> > >    command finished
> > >  * executing `start_master'
> > >  * executing "/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/
> > > bin/start-dfsbroker.sh <http://0.9.2.8/%0Abin/start-dfsbroker.sh> hadoop
> > >     --config=/home/erts/hypertable/
> > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg &&\\\n   /home/
> > > erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/bin/start-master.sh --
> > > config=/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/conf/
> > > hypertable.cfg"
> > >    servers: ["127.0.0.1"]
> > >    [127.0.0.1] executing command
> > >  ** [out :: 127.0.0.1] DFS broker: available file descriptors: 1024
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] Waiting for DFS Broker (hadoop) to come up...
> > >  ** [out :: 127.0.0.1] ERROR: DFS Broker (hadoop) did not come up
> > >    command finished
> > > failed: "sh -c '/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/
> > > bin/start-dfsbroker.sh <http://0.9.2.8/%0Abin/start-dfsbroker.sh> hadoop
> > >     --config=/home/erts/hypertable/
> > > hypertable-0.9.2.8-alpha/0.9.2.8/conf/hypertable.cfg &&\\\n   /home/
> > > erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/bin/start-master.sh --
> > > config=/home/erts/hypertable/hypertable-0.9.2.8-alpha/0.9.2.8/conf/
> > > hypertable.cfg'" on 127.0.0.1
>
> > > ---------------------------------------------------------------
>
> > > Here, 127.0.0.1 is my master.
>
> > > My hypertable.cfg looks like:
>
> > > #
> > > # hypertable.cfg
> > > #
>
> > > # HDFS Broker
> > > HdfsBroker.Port=38030
> > > HdfsBroker.fs.default.name=hdfs://localhost:54310
> > > HdfsBroker.Workers=20
>
> > > # Ceph Broker
> > > CephBroker.Port=38030
> > > CephBroker.Workers=20
> > > CephBroker.MonAddr=10.0.1.245:6789
>
> > > # Local Broker
> > > DfsBroker.Local.Port=38030
> > > DfsBroker.Local.Root=fs/local
>
> > > # DFS Broker - for clients
> > > DfsBroker.Host=localhost
> > > DfsBroker.Port=38030
>
> > > # Hyperspace
> > > Hyperspace.Replica.Host=localhost
> > > Hyperspace.Replica.Port=38040
> > > Hyperspace.Replica.Dir=hyperspace
> > > Hyperspace.Replica.Workers=20
>
> > > # Hypertable.Master
> > > Hypertable.Master.Host=localhost
> > > Hypertable.Master.Port=38050
> > > Hypertable.Master.Workers=20
>
> > > # Hypertable.RangeServer
> > > Hypertable.RangeServer.Port=38060
>
> > > Hyperspace.KeepAlive.Interval=30000
> > > Hyperspace.Lease.Interval=1000000
> > > Hyperspace.GracePeriod=200000
>
> > > # ThriftBroker
> > > ThriftBroker.Port=38080
> > > ------------------------------------------
>
> > > Note: It does not have Hyperspace.Master.Host=localhost property.
>
> > > Capfile:
>
> > > set :source_machine, "127.0.0.1"
> > > set :install_dir,  "/home/erts/hypertable/hypertable-0.9.2.8-alpha"
> > > set :hypertable_version, "0.9.2.8"
> > > set :default_dfs, "hadoop"
> > > set :default_config, "/home/erts/hypertable/hypertable.cfg"
>
> > > role :master, "127.0.0.1"
> > > role :hyperspace, "127.0.0.1"
> > > role :slave, "127.0.0.1", "10.129.125.12"
> > > role :localhost, "127.0.0.1"
> > > ------------------------------------
>
> > > Any idea why DFS Broker is failing?
>
> > > Thanks,
> > > Harshada
>
> > > --
> > > You received this message because you are subscribed to the Google Groups
> > > "Hypertable Development" group.
> > > To post to this group, send email to [email protected].
> > > To unsubscribe from this group, send email to
> > > [email protected]<hypertable-dev%2bunsubscr...@go
> > >  oglegroups.com>
> > > .
> > > For more options, visit this group at
> > >http://groups.google.com/group/hypertable-dev?hl=en.

-- 
You received this message because you are subscribed to the Google Groups 
"Hypertable Development" group.
To post to this group, send email to [email protected].
To unsubscribe from this group, send email to 
[email protected].
For more options, visit this group at 
http://groups.google.com/group/hypertable-dev?hl=en.

Reply via email to