Re: Timestamp as a key good practice?

2012-06-13 Thread Otis Gospodnetic
JM, have a look at https://github.com/sematext/HBaseWD (this comes up often Doug, maybe you could add it to the Ref Guide?) Otis  Performance Monitoring for Solr / ElasticSearch / HBase - http://sematext.com/spm  > > From: Jean-Marc Spaggiari >To: use

Re: HBase first steps: Design a table

2012-06-13 Thread Doug Meil
Just wanted to point out that is also discussed under the autoFlush entry in this chapter.. http://hbase.apache.org/book.html#perf.writing .. but I think this could be better highlighted. I will fix it. On 6/13/12 10:25 AM, "Jean-Marc Spaggiari" wrote: >Hi N. > >The book gives only the i

Invalid HTable(Pool) instance after Master Failover

2012-06-13 Thread Micah Whitacre
We are currently running with 0.92.0 (cdh4b1) and have configured HA so that the master will fail over automatically (we will be upgrading to 0.92.1 cdh4u0 soon). In testing failover we've noticed that in our long running processes which hold onto an instance of HTable or HTablePool when the mast

Re: memstoreSizeMB=0

2012-06-13 Thread Dave Revell
Here's a good starting point: http://hbase.apache.org/book.html#store.memstore , I think you misunderstand what a memstore is. Good luck! -Dave On Wed, Jun 13, 2012 at 10:50 AM, ashdamle wrote: > > Hello. This is my first post here. I've setup hbase with 2 column families > 2GB heap size with

memstoreSizeMB=0

2012-06-13 Thread ashdamle
Hello. This is my first post here. I've setup hbase with 2 column families 2GB heap size with memstore upper and lower set to 0.5 . I have a couple of wide rows with 200K to 500K columns. Upon querying the second time for the row, the memstoreSizeMB is still 0. The block cache hit ratio is aroun

RS unresponsive after series of deletes

2012-06-13 Thread Ted Tuttle
Hi All- I have a repeatable and troublesome HBase interaction that I would like some advice on. I am running a 5 node cluster on v0.94 on cdh3u3 and accessing through Java client API. Each RS has 32G of RAM, is running w/ 16G heap w/ 4G for block cache. Used heap of each RS is well below 16G

Timestamp as a key good practice?

2012-06-13 Thread Jean-Marc Spaggiari
I watched Lars George's video about HBase and read the documentation and it's saying that it's not a good idea to have the timestamp as a key because that will always load the same region until the timestamp reach a certain value and move to the next region (hotspotting). I have a table with a uni

Controlling Region Server Heap usage

2012-06-13 Thread Amit Sela
Hi all, I'm working on a production cluster where each Region Server has 10GB allocated memory, and I'm trying to control (limit) the memory usage during massive writes. HBase configurations are: hbase.regionserver.handler.count 18 hbase.regionserver.global.memstore.upperLimit 0.5 hbase.regionser

Re: HBase first steps: Design a table

2012-06-13 Thread Jean-Marc Spaggiari
Hi N. The book gives only the information to do the load from the command line, so the "void put(final List puts)" is exactly what I was looking for. Thanks a lot, JM 2012/6/13, N Keywal : > Hi, > >> Usually I'm inserting about 40 000 rows at a time. Should I do 40 000 >> calls to put? Or is th

Re: TableRecordReaderImpl is not able to get the rows

2012-06-13 Thread Subroto
Hi Sonal, The Scan is being created by: void org.apache.hadoop.hbase.mapreduce.TableInputFormat.setConf(Configuration configuration) I am not providing any other scan options…. :-( Cheers, Subroto Sanyal On Jun 13, 2012, at 1:30 PM, Sonal Goyal wrote: > Hi Subroto, > > How are you configuring

Re: TableRecordReaderImpl is not able to get the rows

2012-06-13 Thread Sonal Goyal
Hi Subroto, How are you configuring your job? Are you providing any Scan options? Check Chapter 7 of the ref guide at http://hbase.apache.org/book/mapreduce.example.html Best Regards, Sonal Crux: Reporting for HBase Nube Technologies

TableRecordReaderImpl is not able to get the rows

2012-06-13 Thread Subroto
Hi, I have a table with details: hbase(main):024:0> scan 'test' ROW COLUMN+CELL

RE: Single table Vs Multiple table. Need inputs !!!

2012-06-13 Thread Anoop Sam John
Hi >Create separate table for each report type and manage queries using joins There is no join support in HBase. You need to handle it in your client code. See whether you can implement your requirement using one table approach. -Anoop- From: AnandaVelMu

Re: HBase first steps: Design a table

2012-06-13 Thread N Keywal
Hi, > Usually I'm inserting about 40 000 rows at a time. Should I do 40 000 > calls to put? Or is there any "bulkinsert" method? There is this chapter on bulk loading: http://hbase.apache.org/book.html#arch.bulk.load But for 40K rows you may just want to use "void put(final List puts)" in HTableI

Re: Need help in running rowcounter mapreduce job in Hadoop cluster

2012-06-13 Thread AnandaVelMurugan Chandra Mohan
Thanks everyone It worked. But interestingly it takes more time to count few rows. Any idea? On Wed, Jun 13, 2012 at 12:53 PM, Gopinathan A wrote: > Please ensure hbase-*.jar, Guava-r*.jar, protobuf-*.jar & zookeeper.jar in > your classpath. > > > >export > > HADOOP_CLASSPATH=$HADOOP_CLASSP

RE: Need help in running rowcounter mapreduce job in Hadoop cluster

2012-06-13 Thread Gopinathan A
Please ensure hbase-*.jar, Guava-r*.jar, protobuf-*.jar & zookeeper.jar in your classpath. >export > HADOOP_CLASSPATH=$HADOOP_CLASSPATH:/usr/local/hbase-0.92.1-security/lib/ This will not add all the jars into the classpath. You have to iterate the lib folder and then you have to add to HADOOP

Re: Need help in running rowcounter mapreduce job in Hadoop cluster

2012-06-13 Thread Simon Kelly
You could run ./hadoop classpath to verify that the classpath is correct. On 13 June 2012 08:18, Gopinathan A wrote: > Use ./hadoop jar /usr/local/hbase-0.92.1-security/hbase-0.92.1-security.jar > rowcounter > > Thanks & Regards, > Gopinathan A > > > ***

Re: Pre-split table using shell

2012-06-13 Thread Simon Kelly
Thanks Mike, that's pretty much the same reaction I had before. We should be getting another 8Gb shortly but that's the limit for those servers and while that's still not a lot I think we'll manage for now. Unfortunately I'm not the decision maker when it comes to these things so I'm just doing my