Re: Hive integration with HBase

2011-04-12 Thread Carl Steinbach
Hi Marcos, Here's a JIRA query that will show you the list of unresolved Hive issues related to the HBase storage handler: https://issues.apache.org/jira/secure/IssueNavigator.jspa?reset=true&mode=hide&jqlQuery=component+%3D+%22HBase+Handler%22+AND+project+%3D+HIVE+AND+resolution+%3D+Unresolved+O

Re: Hive integration with HBase

2011-04-12 Thread Marcos Ortiz
El 4/12/2011 2:13 PM, Jean-Daniel Cryans escribió: Is there anything in particular you'd like to know? I recently answered a more specific (but still general) question about Hive/HBase here: http://search-hadoop.com/m/YZe7h1zxxoc1 I will also be giving a presentation at OSCON Data in July about

Re: Append value to a cell?

2011-04-12 Thread Gary Helmling
The simplest thing to do would be to just store the row key from TableMain as the column qualifier in the TableRLookup row. Then you don't even need to append to an existing value, you're just adding a new column where the qualifer stores the row ID and the value can be empty. So instead of using

Re: Use of the test cluster in HBase

2011-04-12 Thread Jason Rutherglen
J-D, Thanks I commented out this part as it's trying to limit based on the block size, which in this case is set artificially high to guarantee only one block. On Tue, Apr 12, 2011 at 4:43 PM, Jean-Daniel Cryans wrote: > The namenode does this: > >            long remaining = node.getRemaining()

Re: Append value to a cell?

2011-04-12 Thread Stack
Sounds like a job for a coprocessor; the cell-append-coprocessor. You load it on a particular column family and anything put to this processor appends to the existing cell value, if a value already exists. St.Ack On Tue, Apr 12, 2011 at 8:54 AM, Vishal Kapoor wrote: > rebuilding the whole revers

Re: HBase 0.90.2 CDH3B4 - regions infinitely stuck in transition?

2011-04-12 Thread George P. Stathis
Ah!! I always forget to check the region server log: java.io.IOException: Compression algorithm 'lzo' previously failed test. at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:77) at org.apache.hadoop.hbase.regionserver.HRegion.checkCompressionCodecs(HRegion.jav

Re: Summarizing instructions to run HBase 0.90.2 on Hadoop 0.20.x, feedback appreciated

2011-04-12 Thread Stack
See in the below. On Tue, Apr 12, 2011 at 4:38 AM, Michael G. Noll wrote: > So in order to help myself and hopefully also other readers of this > mailing list, I try to summarize my steps so far to understand and build > Hadoop 0.20-append for use with HBase 0.90.2, the problems I have run > into

Re: How to add family on HBase-0.90.2

2011-04-12 Thread Jean-Daniel Cryans
No, all I say is that you don't need the METHOD, eg this will work in both 0.20 and 0.90: >alter 'cjjHTML', {NAME => 'responseHeader'} J-D On Tue, Apr 12, 2011 at 4:39 PM, 陈加俊 wrote: > So there is no method to add family for a table? > I need to drop the table and create it If I want to add ano

Re: Use of the test cluster in HBase

2011-04-12 Thread Jean-Daniel Cryans
The namenode does this: long remaining = node.getRemaining() - (node.getBlocksScheduled() * blockSize); // check the remaining capacity of the target machine if (blockSize* FSConstants.MIN_BLOCKS_FOR_WRITE>remaining) MIN_BLOCKS_FOR_

Re: How to add family on HBase-0.90.2

2011-04-12 Thread 陈加俊
So there is no method to add family for a table? I need to drop the table and create it If I want to add another family to a table ? How to keep the data in the table before? On Wed, Apr 13, 2011 at 2:31 AM, Jean-Daniel Cryans wrote: > No need to specify the method for adding, it wasn't required

Re: Use of the test cluster in HBase

2011-04-12 Thread Jason Rutherglen
Hmm... There's no physical limitation, is there an artificial setting? On Tue, Apr 12, 2011 at 4:27 PM, Jean-Daniel Cryans wrote: > It says: > > 2011-04-12 16:16:17,157 DEBUG [IPC Server handler 7 on 51372] > namenode.ReplicationTargetChooser(408): Node > /default-rack/127.0.0.1:22967 is not chos

Re: Use of the test cluster in HBase

2011-04-12 Thread Jean-Daniel Cryans
It says: 2011-04-12 16:16:17,157 DEBUG [IPC Server handler 7 on 51372] namenode.ReplicationTargetChooser(408): Node /default-rack/127.0.0.1:22967 is not chosen because the node does not have enough space J-D On Tue, Apr 12, 2011 at 4:24 PM, Jason Rutherglen wrote: > Ah, I had changed conf/log4j

Re: Use of the test cluster in HBase

2011-04-12 Thread Jason Rutherglen
Ah, I had changed conf/log4j.properties. So I changed src/test/resources/log4j.properties, and now the -output file's much more verbose. I'm not sure I understand what's going on however. I'll try to make sense out of the log: http://pastebin.com/MrQJcbJr On Tue, Apr 12, 2011 at 3:38 PM, Stack

Re: Store file size metrics

2011-04-12 Thread Jean-Daniel Cryans
Yep. J-D On Tue, Apr 12, 2011 at 3:55 PM, Chris Tarnas wrote: > I was looking at the metrics column in the regionservers web UI and had a > question: > > If I understand correctly hbase.hregion.max.filesize is the max size of a > single column family's storefile size. If I have 4 column famili

Store file size metrics

2011-04-12 Thread Chris Tarnas
I was looking at the metrics column in the regionservers web UI and had a question: If I understand correctly hbase.hregion.max.filesize is the max size of a single column family's storefile size. If I have 4 column families in a table and have hbase.hregion.max.filesize set to 1GB that means I

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-12 Thread Jean-Daniel Cryans
Yes there are a few places like that. Also when you create new HTables, you should also close their connections (this is not done in htable.close). See HTable's javadoc which says: Instances of HTable passed the same Configuration instance will share connections to servers out on the cluster and

Re: Use of the test cluster in HBase

2011-04-12 Thread Stack
You changed the src/test/resources/log4j.properties? Not sure why changing the block size would make a difference, why it would even care. St.Ack On Tue, Apr 12, 2011 at 2:38 PM, Jason Rutherglen wrote: > Thanks, I'm only seeing the error when I change the block size, either > via DFSClient.cre

Re: HBase 0.90.2 CDH3B4 - regions infinitely stuck in transition?

2011-04-12 Thread Jean-Daniel Cryans
Could you upgrade to the newly released CDH3 instead? It has a few more fixes. So regarding your issue, I don't see regions stuck. The first one did timeout on opening but then it was reassigned (and then I can't see anything in the log that says it timed out again). By the way can you check what

HBase 0.90.2 CDH3B4 - regions infinitely stuck in transition?

2011-04-12 Thread George P. Stathis
In the middle of upgrading our dev environment from 0.89 to 0.90.2CDH3B4. When we did the upgrade locally (Macs), no issues came up. Different story on our EC2 dev box it seems. Background: - dev is running in pseudo-cluster mode - we neglected to set replication to 1 from 2 the first time we star

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-12 Thread Ruben Quintero
I'm running into the same issue, but did some poking around and it seems that Zookeeper connections are being left open by an HBase internal. Basically, I'm running a mapreduce job within another program, and noticed in the logs that every time the job is run, a connection is open, but I never s

Re: Use of the test cluster in HBase

2011-04-12 Thread Jason Rutherglen
Thanks, I'm only seeing the error when I change the block size, either via DFSClient.create or via the Configuration dfs.block.size property. When I changed the log4j.properties to 'log4j.logger.org.apache.hadoop=WARN' I'm not seeing anything additional in the output in the target/surefire-reports

Re: YouAreDeadException after DFS Client errors.

2011-04-12 Thread Dhruba Borthakur
At first sight, it seems to be a network problem. There were some datanodes that were un-reachable from the HBase regionserver, as evidenced by the following logs: 53799733 from any node: java.io.IOException: No live nodes contain current block. Will get new block locations from namenode and retry

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-12 Thread Jean-Daniel Cryans
It's more in the vain of https://issues.apache.org/jira/browse/HBASE-3755 and https://issues.apache.org/jira/browse/HBASE-3771 Basically 0.90 has a regression regarding the handling of zookeeper connections that make it that you have to be super careful not to have more than 30 per machine (each n

Re: Use of the test cluster in HBase

2011-04-12 Thread Gary Helmling
Depends what the log4j.properties file that your code is picking up says. mvn test or IDE "run" test classes should pick up src/test/resources/log4j.properties, which will log to stderr. If that's how you're running you could tweak the hadoop logging level to see if it shows anything more. Change

Re: Use of the test cluster in HBase

2011-04-12 Thread Jason Rutherglen
Where does MiniDFSCluster store the logs? I don't see a location, assuming it's different than stdout/err. On Tue, Apr 12, 2011 at 11:26 AM, Stack wrote: > The datanodes are not starting?  Anything about that in the log? > St.Ack > > On Tue, Apr 12, 2011 at 11:13 AM, Jason Rutherglen > wrote: >

Re: YouAreDeadException after DFS Client errors.

2011-04-12 Thread Jean-Daniel Cryans
YouAreDead means that the master is already processing the death of those region servers when the region server talks back to the master. Network split? J-D On Tue, Apr 12, 2011 at 11:33 AM, Vidhyashankar Venkataraman wrote: > This was something that happened a week back in our cluster: There wa

Re: The whole key space is not covered by the .META.

2011-04-12 Thread Stack
You can also cat the .regioninfo file that is under the region directory to learn more about the region -- its HRegionInfo (The file format is serialized HRegionInfo followed by a String version so intelligible to non-machines). St.Ack 2011/4/12 Stack : > These regions that are in hdfs but not in

Re: The whole key space is not covered by the .META.

2011-04-12 Thread Stack
These regions that are in hdfs but not in .META. and not on any region server are probably harmless. Would be interesting to trace how they got to this state. My guess is that they are old let go regions that were not cleaned up. The region directory name is its encoded name which is the tail pa

YouAreDeadException after DFS Client errors.

2011-04-12 Thread Vidhyashankar Venkataraman
This was something that happened a week back in our cluster: There was a flash death of region servers: a few of the region servers did have near-full heaps so I thought GC could be at play. But many of them crashed after a few DFS errors followed by a YouAreDeadException and they didn't have GC

Re: How to add family on HBase-0.90.2

2011-04-12 Thread Jean-Daniel Cryans
No need to specify the method for adding, it wasn't required in 0.20.6 either so if it was accepting it it was a bug. J-D On Tue, Apr 12, 2011 at 1:54 AM, 陈加俊 wrote: >  I can add family by follow command In HBase-0.20.6 > >>alter 'cjjHTML', {NAME => 'responseHeader', METHOD => 'add'} > > But in

Re: Use of the test cluster in HBase

2011-04-12 Thread Jean-Daniel Cryans
That usually means that your datanode refuses to start or isn't able to connect for some reason. Have a look at its log. J-D On Tue, Apr 12, 2011 at 11:13 AM, Jason Rutherglen wrote: > I'm running into an error when setting the DFS block size to be larger > than the default.  The following code

Re: Use of the test cluster in HBase

2011-04-12 Thread Stack
The datanodes are not starting? Anything about that in the log? St.Ack On Tue, Apr 12, 2011 at 11:13 AM, Jason Rutherglen wrote: > I'm running into an error when setting the DFS block size to be larger > than the default.  The following code is used to create the test > cluster: > > Configuratio

Use of the test cluster in HBase

2011-04-12 Thread Jason Rutherglen
I'm running into an error when setting the DFS block size to be larger than the default. The following code is used to create the test cluster: Configuration conf = new Configuration(); MiniDFSCluster cluster = new MiniDFSCluster(conf, 2, true, null); FileSystem fileSys = cluster.getFileSystem();

Re: apache hbase 0.90.2 vs CDH3 hbase0.90.1+15.18

2011-04-12 Thread Stack
On Tue, Apr 12, 2011 at 7:28 AM, 茅旭峰 wrote: > Hi, > > I've noticed that Cloudera has announced the CDH3 release, but the apache > hbase 0.90.2 is also just released. All should upgrade to the CDH3 release. It includes hdfs-1520, hdfs-1555, and hdfs-1554 -- important features not present in CDH3

Re: Hive integration with HBase

2011-04-12 Thread Jean-Daniel Cryans
Is there anything in particular you'd like to know? I recently answered a more specific (but still general) question about Hive/HBase here: http://search-hadoop.com/m/YZe7h1zxxoc1 I will also be giving a presentation at OSCON Data in July about our experience using both together. J-D On Mon, Ap

Re: upgrade from hbase 0.90.1 to 0.90.2

2011-04-12 Thread Stack
On Tue, Apr 12, 2011 at 12:30 AM, Oleg Ruchovets wrote: > Hi >   We running on hbase 0.90.1 and want to upgrade to 0.90.2. > What is the process to do it , I mean is it enough to change the jar or it > requires configuration files changes. Its a point release so stop the cluster, replace your hba

Re: can i start teh hbase in pseudo-distributed mode with external zookeeper?

2011-04-12 Thread stack
So, this is a special case that I allow in the manual. I should add a note, "Pseudo-distributed using an external zk ensemble?". Would that do? Thanks Stanley, St.Ack 2011/4/12 : > Thanks St.Ack; > I am not sure if this is right. > Currently I set hbase.cluster.distributed to be true, and set

Re: HBase is not ready for Primetime

2011-04-12 Thread Stack
On Tue, Apr 12, 2011 at 10:52 AM, Robert Gonzalez wrote: > Ok, deleted logs that master was complaining about, restarted master > only.  Seemed to be stable after a bunch of the messages like the one > below, then restarted regionservers, sans the one that gave me trouble > this morning.  Now seem

Re: too many regions cause OME ?

2011-04-12 Thread Stack
Does the RegionServer OOME opening same file each time it happens? If so, something is up with the file. Move it aside to get your cluster online and then lets try and figure whats in this file that brings on OOME -- a very large record or perhaps a corruption? St.Ack On Mon, Apr 11, 2011 at 9:

Re: HBase is not ready for Primetime

2011-04-12 Thread Robert Gonzalez
Ok, deleted logs that master was complaining about, restarted master only. Seemed to be stable after a bunch of the messages like the one below, then restarted regionservers, sans the one that gave me trouble this morning. Now seems to be up and running again. I don't trust it, seen this kind of

Re: HBase is not ready for Primetime

2011-04-12 Thread Stack
Please paste more of your log. Your .META. table has empty cells for startcode for some reason but the log below is just WARNING you of this fact and skipping those cells. St.Ack On Tue, Apr 12, 2011 at 10:40 AM, Robert Gonzalez wrote: > A bunch of this in the master log: > > 2011-04-12 12:38:2

Re: ANN: HBase 0.90.2 is available for download

2011-04-12 Thread Stack
Its up in the release repository now. Let me know if any issue with it (I'm still not sure how I made it happen -- I need to do this process again soon). Thanks for your patience, St.Ack On Mon, Apr 11, 2011 at 8:25 AM, Stack wrote: > Lior and Joe: > > Sorry for the mvn lag.  The mvn deploy syst

Re: HBase is not ready for Primetime

2011-04-12 Thread Robert Gonzalez
A bunch of this in the master log: 2011-04-12 12:38:23,771 WARN org.apache.hadoop.hbase.master.CatalogJanitor: REGIONINFO_QUALIFIER is empty in keyvalues={urlhashcopy,E3208173766FDD7C01FE9633E281ED0A,1296085183252.7501ae2b7e933057ea12610c4ec6d001./info:server/1296142856167/Put/vlen=41, urlhashcop

Re: HBase is not ready for Primetime

2011-04-12 Thread Gary Helmling
Robert, You can stop the daemons individually on each node: bin/hbase-daemon.sh stop master bin/hbase-daemon.sh stop regionserver Use this to stop the processes that can be cleanly shutdown. Then let's look at which processes are still hanging and what the logs of the hanging processes are show

Re: HBase is not ready for Primetime

2011-04-12 Thread Ted Yu
In another terminal, you can see which region server(s) is still running. Checking region server log on that region server would help. On Tue, Apr 12, 2011 at 10:34 AM, Robert Gonzalez < g...@maxpointinteractive.com> wrote: > You mean like this: > > hbase@c1-m02:/usr/lib/hbase-0.90.0/bin$ ./stop-

Re: HBase is not ready for Primetime

2011-04-12 Thread Robert Gonzalez
You mean like this: hbase@c1-m02:/usr/lib/hbase-0.90.0/bin$ ./stop-hbase.sh stopping hbase...

Re: HBase is not ready for Primetime

2011-04-12 Thread Bill Graham
Agreed. I've seen similar issues when upon startup where for whatever reason an hlog (often empty) can't be read, which hangs the startup process. Manually deleting it from HDFS clears the issue. On Tue, Apr 12, 2011 at 10:01 AM, Jinsong Hu wrote: > You probably should stop all master/regionserv

Re: HBase is not ready for Primetime

2011-04-12 Thread Jinsong Hu
You probably should stop all master/regionservers, then start one master, tail -f the log to confirm all the hlogs are handled, then start the first regionserver, and then other regionservers. I have encountered this issues before. hbase is not as good as what you want, but not as bad as you sa

HBase is not ready for Primetime

2011-04-12 Thread Robert Gonzalez
We've been using HBase for about a year, consistenly running into problems where we lost data.  After reading forums and some back and forth with other Hbase users, we changed our data methodology to save less data per row.  This last time, we upgraded to 0.90 at the recommendation of the hbase com

RE: hbase architecture question

2011-04-12 Thread Prosperent
The plan was to have the map reduce job run on our schedule (hourly, daily, monthly) and populate these rollups so we aren't having to do any processing on the data in hbase. When a user requests stats, we just pull back the already compiled data from the rollups. It isn't realtime this way, but w

HBase is not ready for prime time

2011-04-12 Thread Robert Gonzalez
We've been using HBase for about a year, consistenly running into problems where we lost data. After reading forums and some back and forth with other Hbase users, we changed our data methodology to save less data per row. This last time, we upgraded to 0.90 at the recommendation of the hbase comm

Re: does Scan guarantee min to max rows?

2011-04-12 Thread Ted Yu
Have you read the thread entitled 'min, max' ? On Tue, Apr 12, 2011 at 7:33 AM, Vishal Kapoor wrote: > Here is the problem. > > my row Ids are "starting" with reversed time stamp followed by "/" and > some more values. > > > 9223370735421724555/TimeStamp1/TimeStamp2/CustomerId/MacIdSystem1/MacIdS

RE: hbase architecture question

2011-04-12 Thread Lyman Do
It depends on how many concurrent users on the BI frond end, if each of them will fire off a MR job for their BI queries, which likely resulting in a scan or partial scan on HBase, this may put too much stress on the IO sub-system. If you have the data access pattern of your BI users, you may

Re: Append value to a cell?

2011-04-12 Thread Vishal Kapoor
rebuilding the whole reverse lookup table should be expensive if I am looking for a million new rows every day in the master table, reading a row manually and then writing the appended row should be a solution but will be a pain. for a file backed system doing a append should be possible? Vishal

RE: Append value to a cell?

2011-04-12 Thread Buttler, David
You have the keys for both tables, is there any reason you can't do a get, local append, put? If you do it in batch, then running a reduce job that collects all of the keys for a given value would be fairly efficient. Dave -Original Message- From: Vishal Kapoor [mailto:vishal.kapoor...

Re: txntype:unknown reqpath:n/a Error Path:/hbase/unassigned Error:KeeperErrorCode = NodeExists for /hbase/unassigned

2011-04-12 Thread Marcos Ortiz
El 4/12/2011 10:19 AM, Shahnawaz Saifi escribió: 1- What HBase version are you use? hbase-0.90.1 2- Which O.S are you using? CentOS 5.5 x8_64 3- Which are the permitions on those directories: /hbase/unassigned /hbase/rs /hbase/table These tables/dir doesn't exist. As I understood the

Append value to a cell?

2011-04-12 Thread Vishal Kapoor
Do we have any API which can append text values or row Ids to a cell. I want to do a control break report and want to append row Ids to a cell value... here is an example. TableMain row1 : user1 row2: user1 row3 : user2 row4 : user3 reverse lookup. TableRLookup. user1 : row1,row2 user2: row3 us

hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-12 Thread Venkatesh
I get this occasionally..(not all the time)..Upgrading from 0.20.6 to 0.90.2 Is this issue same as this JIRA https://issues.apache.org/jira/browse/HBASE-3578 I'm using HBaseConfiguration.create() & setting that in job thx v 2011-04-12 02:13:06,870 ERROR Timer-0 org.apache.hadoop.hbase.mapre

Re: does Scan guarantee min to max rows?

2011-04-12 Thread Vishal Kapoor
Here is the problem. my row Ids are "starting" with reversed time stamp followed by "/" and some more values. 9223370735421724555/TimeStamp1/TimeStamp2/CustomerId/MacIdSystem1/MacIdSystem2/RowType the RowId is designed to make sure the latest row comes up first in the Scan. reverse time is calc

apache hbase 0.90.2 vs CDH3 hbase0.90.1+15.18

2011-04-12 Thread 茅旭峰
Hi, I've noticed that Cloudera has announced the CDH3 release, but the apache hbase 0.90.2 is also just released. I think both of them could run smoothly on CDH3 hadoop 0.20.2+923.21, but any difference there? Thanks and regards, Mao Xu-Feng

Inconsistent table and how to handle

2011-04-12 Thread 茅旭峰
I start a new thread to make things clearer. It seems like we have run into a inconsistent table issue. We are using hadoop-0.20.2-CDH3B4, zookeeper-3.3.2-CDH3B4 and hbase-0.90.1-CDH3B4. The original issue we've run into is like hbase(main):006:0> put 'table1', 'abc', 'cfEStore:dasd', '123

Re: txntype:unknown reqpath:n/a Error Path:/hbase/unassigned Error:KeeperErrorCode = NodeExists for /hbase/unassigned

2011-04-12 Thread Shahnawaz Saifi
1- What HBase version are you use? hbase-0.90.1 2- Which O.S are you using? CentOS 5.5 x8_64 3- Which are the permitions on those directories: /hbase/unassigned /hbase/rs /hbase/table These tables/dir doesn't exist. As I understood these are the INFO level logs. But curious to know the

Re: txntype:unknown reqpath:n/a Error Path:/hbase/unassigned Error:KeeperErrorCode = NodeExists for /hbase/unassigned

2011-04-12 Thread Marcos Ortiz
El 4/12/2011 8:09 AM, Shahnawaz Saifi escribió: 2011-04-12 06:52:37,784 INFO org.apache.zookeeper.server.PrepRequestProcessor: Got user-level KeeperException when processing sessionid:0x12f48ec06850367 type:create cxid:0x2 zxid:0xfffe txntype:unknown reqpath:n/a Error Path:/hbase/unas

txntype:unknown reqpath:n/a Error Path:/hbase/unassigned Error:KeeperErrorCode = NodeExists for /hbase/unassigned

2011-04-12 Thread Shahnawaz Saifi
2011-04-12 06:52:37,784 INFO org.apache.zookeeper.server.PrepRequestProcessor: Got user-level KeeperException when processing sessionid:0x12f48ec06850367 type:create cxid:0x2 zxid:0xfffe txntype:unknown reqpath:n/a Error Path:/hbase/unassigned Error:KeeperErrorCode = NodeExists for /hba

Summarizing instructions to run HBase 0.90.2 on Hadoop 0.20.x, feedback appreciated

2011-04-12 Thread Michael G. Noll
Hi all, as a few other people on this mailing list I am currently working on getting HBase up and running on Hadoop 0.20.2. I think I have by now read most of the relevant past discussions on this topic here, e.g. St.Ack's thread on creating an append release [6], Mike Spreitzner's recent attempt

Re: The whole key space is not covered by the .META.

2011-04-12 Thread 茅旭峰
My guess was the second region is closed, but its directory named of the encoded name is still on the hdfs, this leads to the inconsistency. I wonder what happened when the region is closed, it looks like the process is over, from the log. 2011/4/12 茅旭峰 > Anyway, looks like the two regions bel

How to add family on HBase-0.90.2

2011-04-12 Thread 陈加俊
I can add family by follow command In HBase-0.20.6 >alter 'cjjHTML', {NAME => 'responseHeader', METHOD => 'add'} But in HBase-0.90.2 I can't do it . How ? -- Thanks & Best regards jiajun

upgrade from hbase 0.90.1 to 0.90.2

2011-04-12 Thread Oleg Ruchovets
Hi We running on hbase 0.90.1 and want to upgrade to 0.90.2. What is the process to do it , I mean is it enough to change the jar or it requires configuration files changes. Please advice Oleg.

RE: can i start teh hbase in pseudo-distributed mode with external zookeeper?

2011-04-12 Thread stanley.shi
Thanks St.Ack; I am not sure if this is right. Currently I set hbase.cluster.distributed to be true, and set the HBASE_MANAGES_ZK=false; the environment is running now. Both the hbase master and region server are running on the same machine. I think from the definition, this is also a "pseudo-dis