i tried with a smaller caching i.e 10, it failed again, not its not really
a big cell. this small cluster(4 nodes) is only used for Hbase, i am
currently using hbase-0.92.1-cdh4.0.1. , could you just let me know how
could i debug this issue ?
aused by: org.apache.hadoop.hbase.client.ScannerTimeo
Hi,
Getting this error while using hbase as a sink.
Error
java.io.IOException: Pass a Delete or a Put
at
org.apache.hadoop.hbase.mapreduce.TableOutputFormat$TableRecordWriter.write(TableOutputFormat.java:125)
at
org.apache.hadoop.hbase.mapreduce.TableOutputFormat$TableRecor
You want to look at the hbase-xxx.jar inside the .tar.gz archive.
Tested with 0.94.1:
$ tar -O -xf hbase-0.94.1.tar.gz "*hbase-0.94.1.jar" | jar -t | grep
hbase-default.xmlhbase-default.xml
It's there. :)
-- Lars
- Original Message -
From: Srinivas Mupparapu
To: user@hbase.apache.
Hi Pablo,
It's currently not possible (like you saw).
What's your use case? Maybe there's different/better way to achieve
what you want to do?
J-D
On Mon, Sep 3, 2012 at 1:22 PM, Pablo Musa wrote:
> Hey guys,
> I want to insert new columns into a row:fam and increment 2 of them
> atomically.
Micheal,
We are using HBase to track the usage of our service. Specifically,
each client sends an update when they start a task, at regular
intervals during the task, and an update when they finish a task (and
then presumably they start another, continuing the cycle). Each user
has various attribu
On Mon, Sep 10, 2012 at 12:05 PM, Norbert Burger
wrote:
>
Mind putting up full listing in pastebin?
Let me have a look.
We could try a master restart too... so it refreshes its in-memory
state. That might do it.
St.Ack
Well,
Lets actually skip a few rounds of questions... and start from the beginning.
What does your physical cluster look like?
On Sep 10, 2012, at 12:40 PM, Ramasubramanian
wrote:
> Hi,
> Will be helpful if u say specific things to look into. Pls help
>
> Regards,
> Rams
>
> On 10-Sep-2
On Mon, Sep 10, 2012 at 2:17 PM, Stack wrote:
> Thanks. I was asking about the info:regioninfo column that prints out
> the HRegionInfo for each region. I was wondering if it included a
> status=offline attribute.
>
> You could try one region only and see if that makes a difference.
Hmmm... no
On Sep 10, 2012, at 12:32 PM, Tom Brown wrote:
> We have our system setup such that all interaction is done through
> co-processors. We update the database via a co-processor (it has the
> appropriate logic for dealing with concurrent access to rows), and we
> also query/aggregate via co-process
On Mon, Sep 10, 2012 at 10:32 AM, Tom Brown wrote:
> I want to know more details about the specifics of those requests; Is
> there an API I can use that will allow my coprocessor requests to be
> tracked more functionally? Is there a way to hook into the UI so I can
> provide my own list of runnin
Hi Julian,
I am using only cdh4 libraries. I use the jars present under hadoop and
hbase installed dir. In my last email i gave you some more pointers. Try to
follow them and see what happens.
If then also it doesn't works for you, then i will try to write an utility
to test BigDecimalColumnInterp
On Mon, Sep 10, 2012 at 10:50 AM, Norbert Burger
wrote:
> On Mon, Sep 10, 2012 at 1:37 PM, Stack wrote:
>> What version of hbase?
>
> We're on cdh3u3, 0.90.4 + patches.
>
>> Can you disable and reenable the table?
>
> I will try disabling/re-enabling at the next opportunity. Perhaps
> that'll re
Ah, sorry for assuming that then. I don't know of a way to sort
qualifiers. I haven't seen anyone do that or require it for
unstructured data (i.e. a query like "fetch me the latest qualifier
added to this row"). I suppose you can compare the last two versions
to see what was changed, but I still d
On Mon, Sep 10, 2012 at 10:59 AM, Harsh J wrote:
> Versions is what you're talking about, and by default all queries
> return the latest version of updated values.
>
No actually I was asking if I have columns with qualifier:
d,b,c,e can I store them sorted such that it is e,d,c,b? This ways I c
Versions is what you're talking about, and by default all queries
return the latest version of updated values.
On Mon, Sep 10, 2012 at 11:04 PM, Mohit Anchlia wrote:
> On Mon, Sep 10, 2012 at 10:30 AM, Harsh J wrote:
>
>> Hey Mohit,
>>
>> See http://hbase.apache.org/book.html#schema.smackdown.ro
On Mon, Sep 10, 2012 at 1:37 PM, Stack wrote:
> What version of hbase?
We're on cdh3u3, 0.90.4 + patches.
> Can you disable and reenable the table?
I will try disabling/re-enabling at the next opportunity. Perhaps
that'll resolve that the issue, but this is a PROD cluster, so
unfortunately can
Hello,
If you need to scan lots of log messages and process them use HBase
(or Hive or Pig or simply HDFS+MR)
If you need to query your data set by anything in the text of the log
message, use ElasticSearch or Solr 4.0 or Sensei or just Lucene.
Otis
--
Search Analytics - http://sematext.com/sear
Hi,
Will be helpful if u say specific things to look into. Pls help
Regards,
Rams
On 10-Sep-2012, at 10:40 PM, Stack wrote:
> On Mon, Sep 10, 2012 at 9:58 AM, Ramasubramanian
> wrote:
>> Hi,
>>
>> Currently it takes 11 odd minutes to load 1.2 million record into hbase from
>> hdfs. Can u pl
On Mon, Sep 10, 2012 at 10:33 AM, Norbert Burger
wrote:
> On Mon, Sep 10, 2012 at 1:24 PM, Srinivas Mupparapu
> wrote:
>> It scans .META. table just like any other table. I just tested it and it
>> produced the expected output.
>
> I'm pretty sure Srinivas scanned .META. in his own environment, n
On Mon, Sep 10, 2012 at 10:30 AM, Harsh J wrote:
> Hey Mohit,
>
> See http://hbase.apache.org/book.html#schema.smackdown.rowscols
Thanks! Is there a way in HBase to get the most recent inserted column? Or
a way to sort columns such that I can manage how many columns I want to
read? In timeserie
On Mon, Sep 10, 2012 at 1:24 PM, Srinivas Mupparapu
wrote:
> It scans .META. table just like any other table. I just tested it and it
> produced the expected output.
I'm pretty sure Srinivas scanned .META. in his own environment, not mine. ;-)
> On Sep 10, 2012 12:19 PM, "Stack" wrote:
>> Wha
Hi,
We have our system setup such that all interaction is done through
co-processors. We update the database via a co-processor (it has the
appropriate logic for dealing with concurrent access to rows), and we
also query/aggregate via co-processor (since we don't want to send all
the data over the
Hey Mohit,
See http://hbase.apache.org/book.html#schema.smackdown.rowscols
On Mon, Sep 10, 2012 at 10:56 PM, Mohit Anchlia wrote:
> Is there any recommendation on how many columns one should have per row. My
> columns are < 200 bytes. This will help me to decide if I should shard my
> rows with
Srinivas,
In the source tarball, the file is at
$HBASE_HOME/src/main/resources/hbase-default.xml
On Mon, Sep 10, 2012 at 10:56 PM, Srinivas Mupparapu
wrote:
> I just installed HBase from .tar.gz file and I couldn't find that file
> either.
>
> Thanks,
> Srinivas M
> On Sep 10, 2012 11:03 AM, "hu
On Mon, Sep 10, 2012 at 10:24 AM, Srinivas Mupparapu
wrote:
> It scans .META. table just like any other table. I just tested it and it
> produced the expected output.
>
When you refresh the master UI, it makes a few lines in the master
log. Are these the lines you posted? Mind checking again?
HBase in packaged form bundles the default XML only inside the HBase
jar(s). You need to download a source package tarball to get the
default XML otherwise.
> /usr/share/doc/hbase-0.92.1+67/hbase-default.xml
The above looks right, you can use that as a reference. Looks to be
installed via a docs
I just installed HBase from .tar.gz file and I couldn't find that file
either.
Thanks,
Srinivas M
On Sep 10, 2012 11:03 AM, "huaxiang" wrote:
> Hi,
>I don't find the hbase-default.xml file using following command, any
> other way?
>To be clear, this hadoop was installed with CDH RPM pack
It scans .META. table just like any other table. I just tested it and it
produced the expected output.
Thanks,
Srinivas M
On Sep 10, 2012 12:19 PM, "Stack" wrote:
> On Mon, Sep 10, 2012 at 8:33 AM, Norbert Burger
> wrote:
> > Hi all -- we're currently on cdh3u3 (0.90.4 + patches). I have one
On Mon, Sep 10, 2012 at 10:13 AM, Dhirendra Singh wrote:
> I am facing this exception while iterating over a big table, by default i
> have specified caching as 100,
>
> i am getting the below exception, even though i checked there are several
> calls made to the scanner before it threw this exce
The decision can be made depending on the number of total regions you
want deployed across your 10 machines, and the size you expect the
total to be before you have to expand the size of cluster.
Additionally add in a parallelism factor of say 5-10 (or more if you
want) regions of the same table pe
On Mon, Sep 10, 2012 at 8:33 AM, Norbert Burger
wrote:
> Hi all -- we're currently on cdh3u3 (0.90.4 + patches). I have one
> table in our cluster which seems to functioning fine (gets/puts/scans
> are all working), but for which no regions are listed on the UI. The
> table/regions exist in .MET
On Mon, Sep 10, 2012 at 9:02 AM, huaxiang wrote:
> Hi,
>I don't find the hbase-default.xml file using following command, any
> other way?
>To be clear, this hadoop was installed with CDH RPM package.
>
Is it not bundled inside the hbase-*.jar?
St.Ack
On Mon, Sep 10, 2012 at 9:58 AM, Ramasubramanian
wrote:
> Hi,
>
> Currently it takes 11 odd minutes to load 1.2 million record into hbase from
> hdfs. Can u pls share some tips to do the same in few seconds?
>
> We tried doing this in both pig script and in pentaho. Both are taking 11 odd
> minu
Hi,
Currently it takes 11 odd minutes to load 1.2 million record into hbase from
hdfs. Can u pls share some tips to do the same in few seconds?
We tried doing this in both pig script and in pentaho. Both are taking 11 odd
minutes.
Regards,
Rams
Two cents below...
On Mon, Sep 10, 2012 at 7:24 AM, Shengjie Min wrote:
> In my case, I have all the log events stored in HDFS/hbase in this format:
>
> timestamp | priority | category | message body
>
> Given I have only 4 fields here, that limits my queries to only against
> these four. I am t
Hi,
I haven't really gotten to working on this, since last wednesday.
Checked readFields() and write() today, but don't really see, why I would
need to reimplement those. Admittedly I'm not that into the whole Hbase
codebase, yet, so there is a good chance I'm missing something, here.
Also, Anil,
Hi,
I don't find the hbase-default.xml file using following command, any
other way?
To be clear, this hadoop was installed with CDH RPM package.
Huaxiang
[root@hadoop1 ~]# clear
[root@hadoop1 ~]# rpm -qlp *rpm_file_name.rpm*
[root@hadoop1 ~]# ^C
[root@hadoop1 ~]# find / -name "*hbase-defa
Hi all -- we're currently on cdh3u3 (0.90.4 + patches). I have one
table in our cluster which seems to functioning fine (gets/puts/scans
are all working), but for which no regions are listed on the UI. The
table/regions exist in .META. Other tables in the same cluster show
their regions list fin
Hi there, if there are common questions I'd suggest creating summary
tables of the pre-aggregated results.
http://hbase.apache.org/book.html#mapreduce.example
7.2.4. HBase MapReduce Summary to HBase Example
On 9/10/12 10:03 AM, "iwannaplay games" wrote:
>Hi ,
>
>I want to run query like
>
In my case, I have all the log events stored in HDFS/hbase in this format:
timestamp | priority | category | message body
Given I have only 4 fields here, that limits my queries to only against
these four. I am thinking about more advanced search like full text search
the message body. well, main
its taking very long
On Mon, Sep 10, 2012 at 7:34 PM, Ted Yu wrote:
> Hi,
> Are you able to get the number you want through hive log ?
>
> Thanks
>
> On Mon, Sep 10, 2012 at 7:03 AM, iwannaplay games <
> funnlearnfork...@gmail.com> wrote:
>
> > Hi ,
> >
> > I want to run query like
> >
> > selec
HBase only provides CRUD operations by means of Put/Get/Delete API and
there is no built in SQL interface.
Thanks,
Srinivas M
On Sep 10, 2012 9:03 AM, "iwannaplay games"
wrote:
> Hi ,
>
> I want to run query like
>
> select month(eventdate),scene,count(1),sum(timespent) from eventlog
> group by
Hi,
Are you able to get the number you want through hive log ?
Thanks
On Mon, Sep 10, 2012 at 7:03 AM, iwannaplay games <
funnlearnfork...@gmail.com> wrote:
> Hi ,
>
> I want to run query like
>
> select month(eventdate),scene,count(1),sum(timespent) from eventlog
> group by month(eventdate),sce
Hi ,
I want to run query like
select month(eventdate),scene,count(1),sum(timespent) from eventlog
group by month(eventdate),scene
in hbase.Through hive its taking a lot of time for 40 million
records.Do we have any syntax in hbase to find its result?In sql
server it takes around 9 minutes,How l
Well, the defaul value for a region is 256 MB, so, if you want to
storage a lot of date, you should want to consider to
increase that value.
With the preSplit() method, you can control how to do this process.
On 09/10/2012 04:45 AM, Oleg Ruchovets wrote:
Great
That is actually what I am thin
Hi there...
In this chapter...
http://hbase.apache.org/book.html#datamodel
.. it explains that the "updates" are just a view. There is a merge
happening across CFs and versions (and delete-markers)..
In this...
http://hbase.apache.org/book.html#regions.arch
9.7.5.5. Compaction
... it explai
Great
That is actually what I am thinking about too.
What is the best practice to choose HFile size?
What is the penalty to define it very big?
Thanks
Oleg.
On Mon, Sep 10, 2012 at 4:24 AM, Harsh J wrote:
> Hi Oleg,
>
> If the root issue is a growing number of regions, why not control that
>
If it is associated to a certain table, you only have to disable the
table, reload coprocessor, enable the table.
Regards,
Sever
On Wed, Sep 5, 2012 at 5:18 AM, Aaron Wong wrote:
> Hello all,
>
> I have an endpoint coprocessor running in HBase that I would like to
> modify. I previously loaded
Hi Oleg,
If the root issue is a growing number of regions, why not control that
instead of a way to control the Reducer count? You could, for example,
raise the split-point sizes for HFiles, to not have it split too much,
and hence have larger but fewer regions?
Given that you have 10 machines, I
Hi ,
I am using bulk loading to write my data to hbase.
I works fine , but number of regions growing very rapidly.
Entering ONE WEEK of data I got 200 regions (I am going to save years of
data).
As a result job which writes data to HBase has REDUCERS number equals
REGIONS number.
So entering on
Hi there,
my AQ Master Server might be of interest to you I have an embedded
HBase server in it, it's very very straight forward to use:
http://activequant.org/uberjar.html
What I essentially do is described here:
http://developers.activequant.org:3000/projects/aq2o/repository/entry/trunk/a
With stand-alone mode I assume you mean installing HBase locally and
> work with that?
>
Yes. You can as well can launch "a la standalone" any version, including a
development version. The launch scripts check this, and use maven to get
the classpath needed on the dev version.
The problem with i
2012/9/7 n keywal :
> You can use HBase in standalone mode? Cf.
> http://hbase.apache.org/book.html#standalone_dist?
> I guess you already tried and it didn't work?
With stand-alone mode I assume you mean installing HBase locally and
work with that?
The problem with installing HBase directly on t
Hi,
Try checking the log files of both HDFS ( if it is used ) and HBase to find
out why the region server is going down.
If possible post the logs , i can have a look at it.
On Mon, Sep 10, 2012 at 10:46 AM, iwannaplay games <
funnlearnfork...@gmail.com> wrote:
> Its weird.
> I restarted every
Hi,
Try
rpm -qlp *rpm_file_name.rpm*
This will list all files in the rpm , from this u can know where
hbase-default.xml
is.
On Sat, Sep 8, 2012 at 3:16 PM, John Hancock wrote:
> Huaxiang,
>
> This may not be the quickest way to find it, but if it's anywhere in your
> system, this command will
55 matches
Mail list logo