Hi ,
I have a timeseries data and I am looking for capabilities to store
and aggregates the events on such level granularity: YEAR | MONTH | WEEK
| DAY | HOUR.
I need functionality like sum() , average().
For example to calculate average for event X which was at 14:00 PM every
sunday
Hello,
Is it possible to use hbase to query json documents in a same way as we can
do with Mongodb
Suggestions please.
If we can then a small example as how.. not the query but the process flow..
Thanku so much
Regards,
Panshul.
may be this will help
http://sites.ieee.org/scv-cs/files/2011/03/IBM-Jaql-by-Kevin-Beyer.pdf
On Wed, Jan 16, 2013 at 4:03 PM, Panshul Whisper ouchwhis...@gmail.comwrote:
Hello,
Is it possible to use hbase to query json documents in a same way as we can
do with Mongodb
Suggestions please.
Hello Panshul,
Hbase and MongoDB are built to serve different purposes. You can't
replace one with the other. They have different strengths and weaknesses.
So, if you are using Hbase for something, think well before switching to
MongoDB or vice verca.
Coming back to the actual question,
I've noticed that if I comment the write command in Map function (
Context.write(row,put)),it will just take 40 sec. The differences is about
30 seconds,that's weird for me,what do you think ?
the parameters that are useful up to now:
hbase.hstore.blockingStoreFiles = 20
Hi Oleg,
Try this http://github.com/urbanairship/datacube
Andrew Purtell recommended it here when I had a similar need. I didn't need
things like sum and average so I ended up using a RegionObserver
coprocessor but I did take a look and looks like it might help you.
Good Luck!
On Wed, Jan 16,
Hi all,
It seems like I can't load Endpoint coprocessor from shell but I have no
problem loading RegionObserver from shell.
In both cases I pack a jar file, copy it to HDFS and load from shell using
table_att but only the RegionObserver is loaded (I can see it in the
webapp).
Is it supposed to
Forget about it, my bad :)
On Wed, Jan 16, 2013 at 2:48 PM, Amit Sela am...@infolinks.com wrote:
Hi all,
It seems like I can't load Endpoint coprocessor from shell but I have no
problem loading RegionObserver from shell.
In both cases I pack a jar file, copy it to HDFS and load from shell
Hi,
There is one getting started on Github. Also google for HBase lattice.
We've also built something like that at Sematext, with real-time in memory
agg and HBase persistence (we use it for the thing in my sig and more), but
have not OSSed it.
Otis
--
HBASE Performance Monitoring -
Hello Tariq,
Thank you for the reply.
My concern is that I have been working with MongoDB, but now I am switching
over to Hadoop and I want to use HBase for certain reasons. I was wondering
if I can store Json files in Hbase in a way that I can query the Json files
in Hbase as I can in Mongodb.
You can do that, approach might vary though, depending upon the scenario.
You just have to think well about your schema in order to make sure that it
fits into your put and get requirements.
I have not worked personally on JSON+Hbase combo, so could not give any
direct suggestion at the moment.
Such as I can directly say Mongodb to get me
all the objects having timestamp value of xxx date where timestamp is a
field in Json objects stored in Mongodb
It is possible to store any data in HBase which can be converted into byte[].
Yes using filters one can perform above kind of query. There
We have used Jackson library for converting Java Object to JSON String
and eventually to byte[] and vice-versa; but that is not scan/query
friendly, so we integrated Apache Solr to the stack to get that done.
http://smart-cms.org
Thank you,
Imran
On Wed, Jan 16, 2013 at 7:27 PM, Anoop Sam John
@Anoop sir : Does it make sense to extract the timestamp of JSON
object beforehand and use it as the rowkey? After that serialize the
JSON object and store it in the Hbase cell. Gets would a lot faster
then???
Warm Regards,
Tariq
https://mtariq.jux.com/
cloudfront.blogspot.com
On Wed, Jan 16,
Project Panthera seems to serve your use case well.
You can refer to
http://mail-archives.apache.org/mod_mbox/hadoop-common-user/201209.mbox/%3c521452fcf7acde4795c7e150d240afca0fde1...@shsmsx102.ccr.corp.intel.com%3E
On Jan 16, 2013, at 5:38 AM, Mohammad Tariq donta...@gmail.com wrote:
Hi Doug,
This HBase feature is really interesting. It is quite related to some work
we're doing on Kiji, our schema management project. In particular, we've
also been focusing on building composite row keys correctly. One thing that
jumped out at me in that ticket is that with a composition of
Hi
I have record my trouble shooting process for my random lag region in MR scan
issue. share it here, In case you meet similar problem need to diagnose.
Full text with image here: http://blog.csdn.net/colorant/article/details/8510254
Only text, as below:
Yes Mohammad. Smarter way like this is needed.. I was telling that even if the
full JSON is stored as a column value it will be possible to achive what
Panshul needs. :) But a full table scan will not be acceptable I guess.
As Ted suggested pls check Panthera also. Panthera seems to use Hive
Hi there
Another similar tool on sourceforge.
http://sourceforge.net/projects/haredbhbaseclie/
2012/12/11 Jean-Marc Spaggiari jean-m...@spaggiari.org
For those who are not following the HBase group on LinkedIn, Roi Amir
just posted on a tool they build to look/update HTable in HBase.
You
Hi
I have a table with about 24 region on the one regionserver, and each
region have about 20 block files on hdfs.
The xceiverCount is set to 1024, I have thought that this is quite
enough since at most 480 blocks will be opened.
While when I do a MR job to scan the
20 matches
Mail list logo