Hello,

I put it in my .bashrc and resource it, and the problem is still there. I
haven't seen to many documents in the web about this HBASE_CONF_DIR,
something about PIG, but I am not using it.
I have 4 servers, after putting  HBASE_CONF_DIR, it only starts the first
node, in the others there is nothing new in the logs.
So in the first node it says this:


12/03/27 09:40:57 INFO wal.SequenceFileLogReader: Input stream class:
org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSInputChecker, not
adjusting length
12/03/27 09:40:57 WARN wal.HLogSplitter: Could not open
file:/tmp/hbase-hadoopuser/hbase/.logs/gen19.bioinfo.cipf.es
,55437,1332777708660-splitting/gen19.bioinfo.cipf.es%2C55437%2C1332777708660.1332781312292
for reading. File is empty
java.io.EOFException
        at java.io.DataInputStream.readFully(DataInputStream.java:180)
        at java.io.DataInputStream.readFully(DataInputStream.java:152)
        at
org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1508)
        at
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1486)
        at
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1475)
        at
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1470)
        at
org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogReader$WALReader.<init>(SequenceFileLogReader.java:58)
        at
org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogReader.init(SequenceFileLogReader.java:166)
        at
org.apache.hadoop.hbase.regionserver.wal.HLog.getReader(HLog.java:649)
        at
org.apache.hadoop.hbase.regionserver.wal.HLogSplitter.getReader(HLogSplitter.java:845)
        at
org.apache.hadoop.hbase.regionserver.wal.HLogSplitter.getReader(HLogSplitter.java:758)
        at
org.apache.hadoop.hbase.regionserver.wal.HLogSplitter.splitLogFileToTemp(HLogSplitter.java:384)
        at
org.apache.hadoop.hbase.regionserver.wal.HLogSplitter.splitLogFileToTemp(HLogSplitter.java:351)
        at
org.apache.hadoop.hbase.regionserver.SplitLogWorker$1.exec(SplitLogWorker.java:113)
        at
org.apache.hadoop.hbase.regionserver.SplitLogWorker.grabTask(SplitLogWorker.java:266)
        at
org.apache.hadoop.hbase.regionserver.SplitLogWorker.taskLoop(SplitLogWorker.java:197)
        at
org.apache.hadoop.hbase.regionserver.SplitLogWorker.run(SplitLogWorker.java:165)
        at java.lang.Thread.run(Thread.java:662)

...
12/03/27 09:40:58 INFO master.HMaster: -ROOT- assigned=1, rit=false,
location=ge
12/03/27 09:40:58 INFO ipc.HBaseRPC: Server at
gen19.bioinfo.cipf.es/172.24.78.9
12/03/27 09:40:58 INFO catalog.CatalogTracker: Passed hostingServer is null
12/03/27 09:40:58 INFO ipc.HBaseRPC: Server at
gen19.bioinfo.cipf.es/172.24.78.9
12/03/27 09:40:58 INFO catalog.CatalogTracker: Passed hostingServer is null
12/03/27 09:40:58 INFO ipc.HBaseRPC: Server at
gen19.bioinfo.cipf.es/172.24.78.9
12/03/27 09:40:58 INFO catalog.CatalogTracker: Passed hostingServer is null
12/03/27 09:40:58 INFO ipc.HBaseRPC: Server at
gen19.bioinfo.cipf.es/172.24.78.

I don't know if I am doing right..., any idea?

thanks!!
On 26 March 2012 17:56, Dave Wang <d...@cloudera.com> wrote:

> Roberto,
>
> It should be set in whatever shell you are using.  If you are using bash,
> then .bashrc seems reasonable.  Remember to re-source your .bashrc after
> making the change.  You can verify by running "env | grep HBASE_CONF_DIR"
> from your shell.
>
> If your ZooKeeper is not starting, we'll need to see what the output of
> your logs and command-line are in order to debug further.  Also contents of
> your HBase config files.
>
> - Dave
>
> On Mon, Mar 26, 2012 at 8:51 AM, Roberto Alonso <roal...@gmail.com> wrote:
>
> > Thanks Dave for your answer. In hbase-env.sh I have $HADOOP_CONF_DIR
> setted
> > but no  $HBASE_CONF_DIR. I have put it right now in that file but my
> > zookeeper doesn't start, should I put the variable in the .bashrc or
> > another file?
> >
> > thanks!
> >
> > On 26 March 2012 17:39, Dave Wang <d...@cloudera.com> wrote:
> >
> > > Roberto,
> > >
> > > Is your $HBASE_CONF_DIR pointing to the directory that contains your
> > > hbase-site.xml?
> > >
> > > - Dave
> > >
> > > On Mon, Mar 26, 2012 at 8:35 AM, Roberto Alonso CIPF <ralo...@cipf.es
> > > >wrote:
> > >
> > > > Hello!
> > > > I am experimenting some problems because I think I don't have
> > distributed
> > > > computation.
> > > > I have a map reduce code where I go to a table and I get something of
> > my
> > > > interest. When I do htop to my 4 servers I see that the processors
> are
> > > > working sequentially not in parallel, in other words, one after the
> > other
> > > > but never all at the same time, so that I guess my map reduce is not
> > > doing
> > > > well. But if I ask:
> > > > Configuration config = HBaseConfiguration.create();
> > > > config.get("hbase.cluster.distributed")
> > > > It says hbase.cluster.distributed= false, although I have this in my
> > > config
> > > > files
> > > >  <property>
> > > >            <name>hbase.cluster.distributed</name>
> > > >            <value>true</value>
> > > >         </property>
> > > > What do you think is going on there?
> > > >
> > > > Thanks a lot!
> > > >
> > > > --
> > > > Roberto Alonso
> > > > Bioinformatics and Genomics Department
> > > > Centro de Investigacion Principe Felipe (CIPF)
> > > > C/E.P. Avda. Autopista del Saler, 16-3 (junto Oceanografico)
> > > > 46012 Valencia, Spain
> > > > Tel: +34 963289680 Ext. 1021
> > > > Fax: +34 963289574
> > > > E-Mail: ralo...@cipf.es
> > > >
> > >
> >
>

Reply via email to