HBase Shell available through Apache Zeppelin

2016-02-02 Thread Rajat Venkatesh
://issues.apache.org/jira/browse/ZEPPELIN-651 Commit: https://github.com/apache/incubator-zeppelin/commit/1940388e3422b86a322fc82a0e7868ff25126804 Looking forward to feedback and suggestions for improvements. Rajat Venkatesh Engg. Lead Qubole

Add keys to column family in HBase using Python

2015-04-17 Thread Manoj Venkatesh
Dear HBase experts, I have a Hadoop cluster which has Hive, HBase installed along with other Hadoop components. I am currently exploring ways to automate a data migration process from Hive to HBase which involves new columns of data added ever so often. I was successful in creating a HBase ta

Re: mapreduce job failure

2011-05-17 Thread Venkatesh
ing-full-gcs-in-hbase-with-memstore-local-allocation-buffers-part-1/ J-D On Mon, May 16, 2011 at 2:08 PM, Venkatesh wrote: > Thanks J-D > > Using hbase-0.20.6, 49 node cluster > > The map reduce job involve a full table scan...(region size 4 gig) > The job runs great for

Re: mapreduce job failure

2011-05-16 Thread Venkatesh
scanner would expire. That's orthogonal tho. You need to figure what you're blocking on, add logging and try to jstack your Child processes for example. J-D On Thu, May 12, 2011 at 7:21 PM, Venkatesh wrote: > > Hi > Using hbase-0.20.6 > > mapreduce job starte

mapreduce job failure

2011-05-12 Thread Venkatesh
Hi Using hbase-0.20.6 mapreduce job started failing in the map phase (using hbase table as input for mapper)..(ran fine for a week or so starting with empty tables).. task tracker log: Task attempt_201105121141_0002_m_000452_0 failed to report status for 600 seconds. Killing Region serve

Re: hbase 0.90.2 - incredibly slow response

2011-04-21 Thread Venkatesh
s them (In his case he found that upping the handlers cured a slow scan issue). You could study the 0.20.6 response times and then when you upgrade to 0.90.2, check what its showing. That would at least give us a clue as to where to start digging. St.Ack On Thu, Apr 21, 2011 at 8:21 PM, Ve

Re: hbase 0.90.2 - incredibly slow response

2011-04-21 Thread Venkatesh
informative. St.Ack On Wed, Apr 20, 2011 at 10:41 AM, Venkatesh wrote: > shell is no problems..ones/twos..i've tried mass puts from shell > we cant handle our production load (even 1/3 of it) > 700 mill per day is full load..same load we handled with absolutely no issues i

Re: java.lang.IndexOutOfBoundsException

2011-04-21 Thread Venkatesh
suna To: user@hbase.apache.org Sent: Wed, Apr 20, 2011 4:30 pm Subject: Re: java.lang.IndexOutOfBoundsException On Wed, Apr 20, 2011 at 10:04 AM, Venkatesh wrote: > On 0.90.2, do you all think using HTablePool would help with performance problem? What performance problems are you seein

Re: hbase 0.90.2 - incredibly slow response

2011-04-20 Thread Venkatesh
ssage- From: Stack To: user@hbase.apache.org Sent: Wed, Apr 20, 2011 1:30 pm Subject: Re: hbase 0.90.2 - incredibly slow response On Tue, Apr 19, 2011 at 11:58 AM, Venkatesh wrote: > > I was hoping that too.. > I don't have scripts to generate # requests from shell..I will t

Re: java.lang.IndexOutOfBoundsException

2011-04-20 Thread Venkatesh
specifying Integer.MAX_VALUE as maxSize along with config. On Wed, Apr 20, 2011 at 10:17 AM, Venkatesh wrote: > If I use default ..i can't share/pass my HBaseConfiguration object..atleast > i don't see a constructor/setter.. > that would go against previous suggestion > > >

Re: java.lang.IndexOutOfBoundsException

2011-04-20 Thread Venkatesh
ubject: Re: java.lang.IndexOutOfBoundsException When using HTablePool, try not to define maxSize yourself - use the default. On Wed, Apr 20, 2011 at 10:04 AM, Venkatesh wrote: > > Yeah you & J-D both hit it.. > I knew it's bad..I was trying anything & everything to solve the incredib

Re: java.lang.IndexOutOfBoundsException

2011-04-20 Thread Venkatesh
AM, Ted Yu wrote: > I have seen this before. > HTable isn't thread-safe. > > Please describe your usage. > > Thanks > > > On Wed, Apr 20, 2011 at 6:03 AM, Venkatesh wrote: > >> >> Using hbase-0.90.2..(sigh..) Any tip? thanks >

Re: hbase 0.90.2 - incredibly slow response

2011-04-19 Thread Venkatesh
gives you same lag? St.Ack On Tue, Apr 19, 2011 at 10:35 AM, Venkatesh wrote: > > > Just upgraded to 0.90.2 from 0.20.6..Doing a simple put to table (< 100 > bytes per put).. > Only code change was to retrofit the HTable API to work with 0.90.2 > > Init

hbase 0.90.2 - incredibly slow response

2011-04-19 Thread Venkatesh
Just upgraded to 0.90.2 from 0.20.6..Doing a simple put to table (< 100 bytes per put).. Only code change was to retrofit the HTable API to work with 0.90.2 Initializing HBaseConfiguration in servlet.init()...& reusing that config for HTable constructor & doing put Performance is very slow

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-13 Thread Venkatesh
deleteAllConnections works well for my case..I can live with this but not with connection leaks thanks for the idea Venkatesh -Original Message- From: Ruben Quintero To: user@hbase.apache.org Sent: Wed, Apr 13, 2011 4:25 pm Subject: Re: hbase -0.90.x upgrade - zookeeper

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-13 Thread Venkatesh
e -0.90.x upgrade - zookeeper exception in mapreduce job Venkatesh, I guess the two quick and dirty solutions are: - Call deleteAllConnections(bool) at the end of your MapReduce jobs, or periodically. If you have no other tables or pools, etc. open, then no problem. If you do, they'll

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-13 Thread Venkatesh
e hbase is running). J-D On Wed, Apr 13, 2011 at 12:04 PM, Venkatesh wrote: > Reuben: > Yes..I've the exact same issue now..& I'm also kicking off from another jvm that runs for ever.. > I don't have an alternate solution..either modify hbase code (or) modify my c

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-13 Thread Venkatesh
ions. Am I missing a way to get that buried table's config, or another way to kill the orphaned connections? - Ruben From: Venkatesh To: user@hbase.apache.org Sent: Wed, April 13, 2011 10:20:50 AM Subject: Re: hbase -0.90.x upgrade - zookeep

Re: hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-13 Thread Venkatesh
new Configuration is one new ZK > connection). Upping your zookeeper max connection config should get > rid of your issues since you only get it occasionally. > > J-D > > On Tue, Apr 12, 2011 at 7:59 AM, Venkatesh wrote: >> >> I get this occasion

hbase -0.90.x upgrade - zookeeper exception in mapreduce job

2011-04-12 Thread Venkatesh
I get this occasionally..(not all the time)..Upgrading from 0.20.6 to 0.90.2 Is this issue same as this JIRA https://issues.apache.org/jira/browse/HBASE-3578 I'm using HBaseConfiguration.create() & setting that in job thx v 2011-04-12 02:13:06,870 ERROR Timer-0 org.apache.hadoop.hbase.mapre

Re: zookeper warning with 0.90.1 hbase

2011-04-08 Thread Venkatesh
Thanks St.Ack Yes..I see these when map-reduce job is complete..but not always..I'll ignore thanks..Getting close to 0.90.1 upgrade -Original Message- From: Stack To: user@hbase.apache.org Cc: Venkatesh Sent: Thu, Apr 7, 2011 11:55 pm Subject: Re: zookeper warning

zookeper warning with 0.90.1 hbase

2011-04-07 Thread Venkatesh
I see lot of these warnings..everything seems to be working otherwise..Is this something that can be ignored? 2011-04-07 21:29:15,032 WARN Timer-0-SendThread(..:2181) org.apache.zookeeper.ClientCnxn - Session 0x0 for server :2181, unexpected error, closing socket connection and attempti

Re: row_counter map reduce job & 0.90.1

2011-04-04 Thread Venkatesh
Sorry about this..It was indeed an environment issue..my core-site.xml was pointing to wrong hadoop thanks for the tips -Original Message- From: Venkatesh To: user@hbase.apache.org Sent: Fri, Apr 1, 2011 4:51 pm Subject: Re: row_counter map reduce job & 0.90.1 Yeah

Re: HBase wiki updated

2011-04-03 Thread Venkatesh
A big thankyou from a hbase user (sorry for the spam..but deserves thanks) -Original Message- From: Jean-Daniel Cryans To: user@hbase.apache.org Sent: Sat, Apr 2, 2011 3:51 pm Subject: Re: HBase wiki updated 2 Internets for you Doug, that's awesome! Thx J-D On Apr 2,

Re: row_counter map reduce job & 0.90.1

2011-04-01 Thread Venkatesh
TH. But you shouldn't have to do the latter at least. Compare where it works to where it doesn't. Something is different. St.Ack On Fri, Apr 1, 2011 at 9:26 AM, Venkatesh wrote: > Definitely yes..It'all referenced in -classpath option of jvm of tasktracker/jobtracker/da

Re: row_counter map reduce job & 0.90.1

2011-04-01 Thread Venkatesh
stopping me in moving to 0.90.1 -Original Message- From: Stack To: user@hbase.apache.org Sent: Fri, Apr 1, 2011 12:19 pm Subject: Re: row_counter map reduce job & 0.90.1 On Fri, Apr 1, 2011 at 9:06 AM, Venkatesh wrote: > I'm able to run this job from the hadoop

row_counter map reduce job & 0.90.1

2011-04-01 Thread Venkatesh
I'm able to run this job from the hadoop machine (where job & task tracker also runs) /hadoop jar /home/maryama/hbase-0.90.1/hbase-0.90.1.jar rowcounter But, I'm not able to run the same job from a) hbase client machine (full hbase & hadoop installed) b) hbase server machines (ditto) Get

Re: hole in META

2011-03-31 Thread Venkatesh
20.6.. What are the region parameters? I tried encoded nam it did n't like..I tried name of the form tbl_name,st_key,, That did n't work either.. thanks -Original Message- From: Stack To: user@hbase.apache.org Cc: Venkatesh Sent: Thu, Mar 31, 2011 1:36 am S

Re: hole in META

2011-03-30 Thread Venkatesh
es it not complete? St.Ack On Wed, Mar 30, 2011 at 4:13 AM, Venkatesh wrote: > Yes..st.ack..overlapping..& one of them has no data.. > there are too many of them about 800 or so.. > there are some with holes too.. > > > > > > > > > &

Re: hole in META

2011-03-30 Thread Venkatesh
n Wed, Mar 30, 2011 at 4:13 AM, Venkatesh wrote: > Yes..st.ack..overlapping..& one of them has no data.. > there are too many of them about 800 or so.. > there are some with holes too.. > > > > > > > > > > > -Original Message- >

Re: hole in META

2011-03-30 Thread Venkatesh
Thanks Lucas..I'll give it a try -Original Message- From: Lukas To: user@hbase.apache.org Sent: Wed, Mar 30, 2011 4:19 am Subject: Re: hole in META Sorry for any inconvenience. This was in reply of http://mail-archives.apache.org/mod_mbox/hbase-user/201103.mbox/%3c8cdbca99c

Re: hole in META

2011-03-30 Thread Venkatesh
ping regions? Can you try merging them with merge tool? Else, study whats in hdfs. One may have nothing in it (check sizes). It might just be reference files only. If so, lets go from there. And I describe how to merge. St.Ack On Tue, Mar 29, 2011 at 9:25 PM, Venkatesh wrote: > I

Re: hole in META

2011-03-29 Thread Venkatesh
From: Stack To: user@hbase.apache.org Sent: Tue, Mar 29, 2011 12:55 pm Subject: Re: hole in META On Tue, Mar 29, 2011 at 9:09 AM, Venkatesh wrote: > I ran into missing jar with hadoop jar file when running a map reduce..which i could n't fix it..That is the only known issue with up

Re: Export/Import and # of regions

2011-03-29 Thread Venkatesh
for you. J-D On Tue, Mar 29, 2011 at 2:33 PM, Venkatesh wrote: > Thanks J-D > > We have way too much data & it won't fit in 1 region.Is Import smart enough create > reqd # of regions? > > Cld u pl. elaborate on pre-split table creation? steps? > &g

Re: Export/Import and # of regions

2011-03-29 Thread Venkatesh
to create a pre-split table. J-D On Tue, Mar 29, 2011 at 2:27 PM, Venkatesh wrote: > > > > Hi, > If I export existing table using Export MR job, truncate the table, increase region size, & do a Import > will it make use of the new region size? > > thanks > V > > >

Export/Import and # of regions

2011-03-29 Thread Venkatesh
Hi, If I export existing table using Export MR job, truncate the table, increase region size, & do a Import will it make use of the new region size? thanks V

Re: hole in META

2011-03-29 Thread Venkatesh
s)..but it increases the # of regions which is the down side.. (we are already way over recommended limit 2800 regions per node) -Original Message- From: Stack To: user@hbase.apache.org Sent: Tue, Mar 29, 2011 12:55 pm Subject: Re: hole in META On Tue, Mar 29, 2011 at 9:09 AM,

Re: hole in META

2011-03-29 Thread Venkatesh
ll upgrade BTW, is it better to fix existing holes using add_table.rb before the upgrade? (or) upgrade takes care missing holes? -Original Message- From: Stack To: user@hbase.apache.org Sent: Tue, Mar 29, 2011 11:59 am Subject: Re: hole in META On Tue, Mar 29, 2011 at 7:3

hole in META

2011-03-29 Thread Venkatesh
Hi Using hbase-0.20.6..This has happened quite often..Is this a known issue in 0.20.6 that we would n't see in 0.90.1 (or) see less of? ..Attempt to fix/avoid this earlier times by truncating table, running add_table.rb before What is the best way to fix this in 0.20.6? Now it's there in mo

Re: java.io.FileNotFoundException:

2011-03-16 Thread Venkatesh
e below is pretty basic error. Reference the jar that is actually present on your cluster. St.Ack On Wed, Mar 16, 2011 at 3:50 PM, Venkatesh wrote: > yeah..i was aware of that..I removed that & tried with hadoop-0.20.2-core.jar as I was n't ready to upgrade hadoop.. > >

Re: java.io.FileNotFoundException:

2011-03-16 Thread Venkatesh
up its name in the lib/ directory of the distribution (comes with a rev #) :) On Thu, Mar 17, 2011 at 2:33 AM, Venkatesh wrote: > Thanks St.Ack..I'm blind..Got past that.. > Now I get for hadoop-0.20.2-core.jar > > I've removed *append*.jar all over the place &am

Re: java.io.FileNotFoundException:

2011-03-16 Thread Venkatesh
0.1 ships with zookeeper-3.3.2, not with 3.2.2. St.Ack On Wed, Mar 16, 2011 at 8:05 AM, Venkatesh wrote: > > Does anyone how to get around this? Trying to run a mapreduce job in a cluster..The one change was hbase upgraded to 0.90.1 (from 0.20.6)..No code change > > > java.i

java.io.FileNotFoundException:

2011-03-16 Thread Venkatesh
Does anyone how to get around this? Trying to run a mapreduce job in a cluster..The one change was hbase upgraded to 0.90.1 (from 0.20.6)..No code change java.io.FileNotFoundException: File /data/servers/datastore/mapred/mapred/system/job_201103151601_0363/libjars/zookeeper-3.2.2.jar does

hbase 0.90.1 upgrade issue - mapreduce job

2011-03-15 Thread Venkatesh
Hi When I upgraded to 0.90.1, mapreduce fails with exception.. system/job_201103151601_0121/libjars/hbase-0.90.1.jar does not exist. I have the jar file in classpath (hadoop-env.sh) any ideas? thanks

Re: region servers shutdown

2011-02-10 Thread Venkatesh
hing else than HBase, then you should only use 1 zk server and collocate it with the master and the namenode, then use those 3 machines as region servers to help spread the region load. J-D On Thu, Feb 10, 2011 at 12:35 PM, Venkatesh wrote: > Thanks J-D.. > I was quite happy in the 1s

Re: region servers shutdown

2011-02-10 Thread Venkatesh
..but did help the real puts -Original Message- From: Ted Dunning To: user@hbase.apache.org Sent: Thu, Feb 10, 2011 3:45 pm Subject: Re: region servers shutdown Are your keys sequential or randomized? On Thu, Feb 10, 2011 at 12:35 PM, Venkatesh wrote: > iii) Process

Re: region servers shutdown

2011-02-10 Thread Venkatesh
d how many regions do they serve? Are you using lots of families per table? Are you using LZO compression? Thanks for helping us helping you :) J-D On Thu, Feb 10, 2011 at 11:32 AM, Venkatesh wrote: > Thanks J-D.. > Can't believe i missed that..I have had it before ..i did l

Re: region servers shutdown

2011-02-10 Thread Venkatesh
node logs a the time of the outage. Very often it's caused by either ulimit or xcievers that weren't properly configured, checkout http://hbase.apache.org/notsoquick.html#ulimit J-D On Thu, Feb 10, 2011 at 10:42 AM, Venkatesh wrote: > > > > Hi > I've ha

region servers shutdown

2011-02-10 Thread Venkatesh
Hi I've had this before ..but not to 70% of the cluster..region servers all dying..Any insight is helpful. Using hbase-0.20.6, hadoop-0.20.2 I don't see any error in the datanode or the namenode many thanks Here's the relevant log entires ..in master... Got while writing region XXlog j

Re: script to delete regions with no rows

2011-01-28 Thread Venkatesh
g regions. They might inspire. Also look at the Merge.java class. See how it edits .META. after merging two adjacent regions to create a new region that spans the key space of the two old adjacent regions. St.Ack On Fri, Jan 28, 2011 at 12:29 PM, Venkatesh wrote: > > > &

script to delete regions with no rows

2011-01-28 Thread Venkatesh
Is there a script? thanks

Re: getting retries exhausted exception

2011-01-26 Thread Venkatesh
G for hbase and re-run the job to hopefully get more information. J-D On Wed, Jan 26, 2011 at 8:44 AM, Venkatesh wrote: > > > > Using 0.20.6..any solutions? Occurs during mapper phase..will increasing retry count fix this? > thanks > &

Re: hbase.client.retries.number

2010-10-15 Thread Venkatesh
Hi Sean: Thx Size of column family is very small < 100 bytes Investigating potential bottleneck spot..Our cluster is small (relatively speaking)..10 node Our hardware is high end (not commodity) venkatesh -Original Message- From: Sean Bigdatafun To: user@hbase.apache.

Re: Increase region server throughput

2010-10-14 Thread Venkatesh
tablename)' in your client or are you doing > 'new HTable(conf, tablename)' in your client code? Do the latter if > not -- share the configuration with HTable instances. > > St.Ack > > On Mon, Oct 11, 2010 at 10:47 PM, Venkatesh wrote: > > > > > > &

Re: hbase.client.retries.number

2010-10-14 Thread Venkatesh
..Is that reasonable for a table size of 10 mill records? hbase.client.scanner.caching - If set in hbase-site.xml, Scan calls should pick that up correct? thanks venkatesh -Original Message- From: Jean-Daniel Cryans To: user@hbase.apache.org Sent: Thu, Oct 14, 2010

Increase region server throughput

2010-10-11 Thread Venkatesh
ered write without code change) thx venkatesh

Re: hbase.client.retries.number

2010-10-11 Thread Venkatesh
rvlet.java:617) javax.servlet.http.HttpServlet.service(HttpServlet.java:717) -Original Message- From: Venkatesh To: user@hbase.apache.org Sent: Mon, Oct 11, 2010 2:35 pm Subject: hbase.client.retries.number HBase was seamless for first couple of weeks..now all kinds of issues in production :) fun

hbase.client.retries.number

2010-10-11 Thread Venkatesh
I can't do any put (new) thanks venkatesh

Region servers suddenly disappearing

2010-10-10 Thread Venkatesh
Some of the region servers suddenly dying..I've pasted relevant log lines..I don't see any error in datanodes Any ideas? thanks venkatsh . 2010-10-10 12:55:36,664 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer Exception: java.io.IOException: Unable to create new block.at o

Incredibly slow response to Scan

2010-10-07 Thread Venkatesh
his normal? thanks venkatesh PS. setCaching(100) ..did n't make a dent in performance

Re: HBase map reduce job timing

2010-10-06 Thread Venkatesh
Thanks J-D I'll hookup Ganglia (wanting but kept pushing back..) & get back V -Original Message- From: Jean-Daniel Cryans To: user@hbase.apache.org Sent: Wed, Oct 6, 2010 12:22 pm Subject: Re: HBase map reduce job timing > Also, do you think if I query using rowkey instead

Re: HBase map reduce job timing

2010-10-06 Thread Venkatesh
Also, do you think if I query using rowkey instead of hbase time stamp..it would not kick off that many tasks.. since region server knows the exact locations? thanks venkatesh -Original Message- From: Venkatesh To: user@hbase.apache.org Sent: Wed, Oct 6, 2010 8:53 am

Re: HBase map reduce job timing

2010-10-06 Thread Venkatesh
Ahh ..ok..That makes sense I've a 10 node cluster each with 36 gig..I've allocated 4gig for HBase Region Servers..master.jsp reports used heap is less than half on each region server. I've close to 800 regions total..Guess it needs to kick off a jvm to see if data exists in all regions..

Re: HBase map reduce job timing

2010-10-05 Thread Venkatesh
ts less free memory (couple of gig.) though box has 36 gig total.. I don't quite trust top since cached blocks don't show up under free column even if no process is running.. venkatesh -Original Message- From: Jean-Daniel Cryans To: user@hbase.apache.org Sent:

Re: HBase map reduce job timing

2010-10-05 Thread Venkatesh
.org Sent: Tue, Oct 5, 2010 11:14 pm Subject: Re: HBase map reduce job timing It'd be more useful if we knew where that data is coming from, and where it's going. Are you scanning HBase and/or writing to it? J-D On Tue, Oct 5, 2010 at 8:05 PM, Venkatesh wrote: > > > > So

Re: HBase map reduce job timing

2010-10-05 Thread Venkatesh
t I can do is pointing to the existing documentation http://wiki.apache.org/hadoop/PerformanceTuning J-D On Tue, Oct 5, 2010 at 7:12 PM, Venkatesh wrote: > > > > I've a mapreduce job that is taking too long..over an hour..Trying to see what can a tune > to to bring it down

HBase map reduce job timing

2010-10-05 Thread Venkatesh
a why that would be?... ..map phase takes about couple of minutes.. ..reduce phase takes the rest.. ..i'll try increasing # of reduce tasks..Open to other other suggestion for tunables.. thanks for your input venkatesh

hbase/hdfs disk usage

2010-09-30 Thread Venkatesh
the web interface..reports 10gig That seems odd..Any ideas on what could be taking up space?..I don't have permission to look the entire hdfs..yet Just thought i'll ask the group thanks venkatesh

Re: stop-hbase.sh takes forever (never stops)

2010-09-07 Thread Venkatesh
Don't know if this helps..but here are couple of reasons when I had the issue & how i resolved it -> If zookeeper is not running (or do not have the quorum) in a cluster setup, hbase does not go down..bring up zookeeper -> Make sure pid file is not under /tmp...somtimes files get cleaned out of

Dependencies for running mapreduce

2010-08-27 Thread Venkatesh
application in HADOOP_CLASSPATH ..for the mapreduce to run in a cluster? thanks venkatesh

Re: jobtracker.jsp

2010-08-26 Thread Venkatesh
..to make this classpath effective Is there a force shutdown option? (other than kill -9)..? venkatesh -Original Message- From: Jean-Daniel Cryans To: user@hbase.apache.org Sent: Fri, Aug 27, 2010 12:10 am Subject: Re: jobtracker.jsp HBase needs to know about the job tracker,

Re: jobtracker.jsp

2010-08-26 Thread Venkatesh
hat's the log in your client side ? On Thu, Aug 26, 2010 at 6:23 PM, Venkatesh wrote: > > > > I'm running map/reduce jobs from java app (table mapper & reducer) in true distributed > mode..I don't see anything in jobtracker page..Map/reduce job runs fine..Am I mi

jobtracker.jsp

2010-08-26 Thread Venkatesh
I'm running map/reduce jobs from java app (table mapper & reducer) in true distributed mode..I don't see anything in jobtracker page..Map/reduce job runs fine..Am I missing some config? thanks venkatesh

org.apache.hadoop.mapreduce.lib.output.NullOutputFormat ClassNotFoundException

2010-08-23 Thread Venkatesh
;JMX" (tomcat) It works fine..when I have periodic scheduler running automatically in the background.. Also..I see this error on a 64-bit machine..& not on 32bit machines.. Wierd..wonder if anyone has any insights?..Setup looks identical.. thanks venkatesh

Re: How to delete rows in a FIFO manner

2010-08-06 Thread Venkatesh
I wrestled with that idea of time bounded tables..Would it make it harder to write code/run map reduce on multiple tables ? Also, how do u decide to when to do the cut over (start of a new day, week/month..) & if u do how to process data that cross those time boundaries efficiently.. Guess that

HTable object - how long it is valid

2010-08-01 Thread Venkatesh
when do I call close()..upon every operation (put/get/..) ? to avoid memory leaks thanks venkatesh