gt;
> Thanks
>
>
> On Wed, Mar 26, 2014 at 6:40 PM, Norbert Burger >wrote:
>
> > Hi folks -- we're currently on HBase from CDH 4.1.2 (0.92.1 + patches).
> >
> > We recently ran into a consistently-reproducible scenario where applying
> a
> > R
Hi folks -- we're currently on HBase from CDH 4.1.2 (0.92.1 + patches).
We recently ran into a consistently-reproducible scenario where applying a
RegexStringComparator within a RowFilter returned different and incomplete
results depending on the value of hbase.client.scanner.caching.
This sounds
Thanks Ted - this config change appears to have reduced quite a bit of the
memstore flushes.
Norbert
On Fri, Dec 27, 2013 at 12:03 AM, Ted Yu wrote:
> Raising maxlogs to 64 would give you improvement.
>
> Cheers
>
> On Dec 26, 2013, at 8:35 PM, Norbert Burger
> wrote:
&g
logs" before flushing a few
> > dozen KB log ? Just want to double-check it's weather a forcing flush
> issue
> > or not:)
> >
> > Thanks,
> > Liang
> >
> > 发件人: Norbert Burger [norbert.bur...@gmail.com
Hi folks, we're currently running CDH4.1.2 (HBase 0.92 + patches).
Looking through the logs, I'm seeing that some tables are flushing at much
lower size than others (a few dozen KB as opposed to the expected 128 MB).
Looking through some of great case studies like [1], [2], and [3], it's not
clea
turned off by default in
> CDH4.1.2 to be backward compatible.
>
> Thanks,
> Jimmy
>
> On Tue, Dec 4, 2012 at 11:28 AM, Norbert Burger
> wrote:
>> Thanks, Jimmy. Do you mean the config
>> "hbase.master.loadbalance.bytable"? According to
On Tue, Dec 4, 2012 at 2:10 PM, Ted Yu wrote:
> Can you give us a little more detail on how much deviation the region
> counts on region servers have ?
>
> There is a parameter, hbase.regions.slop, with default value of 0.2
> This parameter allows region count to deviate by certain percentage from
the value for this parameter and see if you get better
>> results.
>>
>> I will also put the above summary on HBASE-3373.
>>
>> Thanks
>>
>> On Tue, Dec 4, 2012 at 8:42 AM, Norbert Burger
>> wrote:
>>
>>> We upgraded to CDH4.1.2 (conta
We upgraded to CDH4.1.2 (contains HBASE-3373) in one of our
environments. After filling that environment with data, I was
surprised to see that regions were not balanced across regionservers
at the table level. We have restarted all regionservers at least once
here.
In [1], I see Stack's referen
We had the same question earlier. Unfortunately the documentation is
wrong on this account; scannerOpen resolves to either a call to
scan.addFamily or scan.addColumn, and neither directly supports regex
matching.
Regex pattern matching against colquals is definitely supported on the
Java side, so
Try changing your CompareOp.EQUALs to CompareOp.GREATER_OR_EQUAL and
CompareOp.LESS_OR_EQUAL, respectively. You want all rows between your
two key.
Norbert
On Mon, Oct 15, 2012 at 7:00 PM, Kumar, Suresh wrote:
> I have a HBase with some apache logs loaded.
>
>
>
> I am trying to retrieve a sect
Hi Shumin, seems like the example provided is missing a closing
rparen? Try changing, for example, the scan+filter line to be:
scan 'test', {COLUMNS=>"cf:u", FILTER=>SingleColumnValueFilter.new
(Bytes.toBytes('cf'), Bytes.toBytes('u'),
CompareFilter::CompareOp.valueOf('EQUAL'),
SubstringComparato
On Mon, Sep 10, 2012 at 3:29 PM, Stack wrote:
> On Mon, Sep 10, 2012 at 12:05 PM, Norbert Burger
> wrote:
>>
> Mind putting up full listing in pastebin?
Here's a link: http://pastebin.com/raw.php?i=4YhS8CpE. The table in
question is called 'sessions', I did delet
On Mon, Sep 10, 2012 at 2:17 PM, Stack wrote:
> Thanks. I was asking about the info:regioninfo column that prints out
> the HRegionInfo for each region. I was wondering if it included a
> status=offline attribute.
>
> You could try one region only and see if that makes a difference.
Hmmm... no
On Mon, Sep 10, 2012 at 1:37 PM, Stack wrote:
> What version of hbase?
We're on cdh3u3, 0.90.4 + patches.
> Can you disable and reenable the table?
I will try disabling/re-enabling at the next opportunity. Perhaps
that'll resolve that the issue, but this is a PROD cluster, so
unfortunately can
On Mon, Sep 10, 2012 at 1:24 PM, Srinivas Mupparapu
wrote:
> It scans .META. table just like any other table. I just tested it and it
> produced the expected output.
I'm pretty sure Srinivas scanned .META. in his own environment, not mine. ;-)
> On Sep 10, 2012 12:19 PM, "Stack" wrote:
>> Wha
Hi all -- we're currently on cdh3u3 (0.90.4 + patches). I have one
table in our cluster which seems to functioning fine (gets/puts/scans
are all working), but for which no regions are listed on the UI. The
table/regions exist in .META. Other tables in the same cluster show
their regions list fin
On Sat, Aug 18, 2012 at 7:14 AM, Michael Segel
wrote:
Thanks.
> Just out of curiosity, what would happen if you could disable the table,
> alter the table's max file size and then attempted to merge regions? Note:
> I've never tried this, don't know if its possible, just thinking outside of
On Fri, Aug 17, 2012 at 4:09 PM, anil gupta wrote:
> If you want to customize the bulkloader then you can write your own mapper
> to define the business logic for loading. You need to specify the mapper at
> the time of running importsv by using:
Thanks, Anil. I had that seen that section of the
Hi folks -- we're running CDH3u3 (0.90.4). I'm trying export data
from an existing table that has far too many regions (2600+ for only 8
regionservers) into one with a more reasonable region count for this
cluster (256). Overall data volume is approx. 3 TB.
I thought initially that I'd use the b
>From what I understand, the leap second bug could've hit anytime in the 24
hours before 23:59:59. We had it start happening early afternoon Sat on a
few of our boxes.
Norbert
On Mon, Jul 2, 2012 at 12:58 PM, Kevin O'dell wrote:
> How recently would you say this is happening? Did this start la
Looks like localhost is somehow resolving to an IPv6 address. If this is
not your intent (usually not recommended), then you can hint the JVM by
adding
-Djava.net.preferIPv4Stack=true to the HBASE_OPTS var in hbase-env.sh.
Norbert
On Sun, May 27, 2012 at 5:43 PM, Dalia Sobhy wrote:
>
> Hi,
> I
This is a config/classpath issue, no? At the lowest level, Hadoop MR tasks
don't pick up settings from the HBase conf directory unless they're
explicitly added to the classpath, usually via hadoop/conf/hadoop-env.sh:
http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/mapreduce/package-summar
Dalia -- are you setting the appropriate timeout in the TSocket
constructor? The default value is 0, which results in an "infinite
timeout" according to the java.net.Socket docs.
Verify that your Thrift server is running (default port is 9090), and make
sure you specify a reasonable timeout to TS
On Thu, Oct 13, 2011 at 1:05 PM, Jean-Daniel Cryans wrote:
> On Thu, Oct 13, 2011 at 8:17 AM, Norbert Burger >wrote:
>
> > In the ticket, I mentioned setting hbase.client.write.buffer as a
> > workaround, but unfortunately it doesn't seem that autoflush (which I
>
atch for 1744.
> >
> > FYI
> >
> >
> > On Wed, Oct 12, 2011 at 3:25 PM, Stack wrote:
> >
> >> On Wed, Oct 12, 2011 at 2:48 PM, Norbert Burger
> >> wrote:
> >> > We're using Thrift for a fairly write-heavy app. Many of the b
We're using Thrift for a fairly write-heavy app. Many of the best practices
refer to disabling autoFlush, but I was surprised to see autoFlush not even
touched in ./src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java.
Is this accurate? How are others using Thrift for write-heavy loads?
are referring in a sample .META. row (Perhaps you take the md5
> hash on the tail of the region name as a 'timestamp'?)
>
> St.Ack
>
> On Mon, Jul 18, 2011 at 3:28 PM, Norbert Burger
> wrote:
> > Hey folks --
> >
> > For all rowkeys in .META. except fo
Hey folks --
For all rowkeys in .META. except for the last region of a table, I see a
couple of different timestamps. Could someone confirm which one is
region-splitted-at timestamp, and what the logical meaning of the other
timestamp is?
Thanks,
Norbert
, where you get Hadoop+HBase
running all on the same node (http://goo.gl/Rytnp).
Norbert
On Sat, Feb 5, 2011 at 9:51 AM, Norbert Burger wrote:
> Mike, you'll also need also access to an installation of Hadoop, whether
> this on the same machines as your HBase install (common), or somew
Mike, you'll also need also access to an installation of Hadoop, whether
this on the same machines as your HBase install (common), or somewhere
else. Often, people install Hadoop first and then layer HBase over it.
HBase depends on core Hadoop functionality like HDFS, and uses the Hadoop
JAR in l
the "hadoop fs -dus /hbase/"
> bytes. Then that would give an idea of the amount of overhead. I have
> a feeling it is negligible in the grand scheme of things.
>
> -Xavier
>
> On 1/24/11 10:23 AM, Norbert Burger wrote:
> > Good idea. But it seems like this appro
each one.
> Then you could use that output and get some general descriptive
> statistics based on it.
>
> Cheers,
>
>
> -Xavier
>
>
> On 1/24/11 9:37 AM, Norbert Burger wrote:
> > Hi folks - is there a recommended way of estimating HBase HDFS usage for
> a
>
Hi folks - is there a recommended way of estimating HBase HDFS usage for a
new environment?
We have a DEV HBase cluster in place, and from this, I'm trying to estimate
the specs of our not-yet-built PROD environment. One of the variables we're
considering is HBase usage of HDFS. What I've just t
Thanks J-D :-) Somehow, I missed the javadocs for TIFB/TIF, which have
enough context to get me started.
Norbert
On Tue, Dec 14, 2010 at 12:39 PM, Jean-Daniel Cryans wrote:
> Hi Norbert,
>
> Probably the best example is TableInputFormat :)
>
> J-D
>
> On Tue, Dec 14, 201
Hi all - can anyone point me in the direction of an example extension of
TableInputFormatBase?
We are planning on using HBase together with Dumbo (the Python Streaming
layer). In order to support this, we'll need to write a version of TIF
that: (a) supports start rows, and (b) implements
InputFor
36 matches
Mail list logo