Re: mapreduce streaming with hbase as a source

2011-02-22 Thread Jean-Daniel Cryans
(moving to the hbase user ML) I think streaming used to work correctly in hbase 0.19 since the RowResult class was giving the value (which you had to parse out), but now that Result is made of KeyValue and they don't include the values in toString then I don't see how TableInputFormat could be use

Re: mapreduce streaming with hbase as a source

2011-02-19 Thread Ondrej Holecek
I don't think you understand me correctly, I get this line: 72 6f 77 31 keyvalues={row1/family1:a/1298037737154/Put/vlen=1, row1/family1:b/1298037744658/Put/vlen=1, row1/family1:c/1298037748020/Put/vlen=1} I know "72 6f 77 31" is the key and the rest is value, let's call it mapreduce-value.

Re: mapreduce streaming with hbase as a source

2011-02-19 Thread ShengChang Gu
By default, the prefix of a line up to the first tab character is the key and the rest of the line (excluding the tab character) will be the value. If there is no tab character in the line, then entire line is considered as key and the value is null. However, this can be customized, Use: -D stream

Re: mapreduce streaming with hbase as a source

2011-02-19 Thread Ondrej Holecek
Thank you, I've spend a lot of time with debuging but didn't notice this typo :( Now it works, but I don't understand one thing: On stdin I get this: 72 6f 77 31 keyvalues={row1/family1:a/1298037737154/Put/vlen=1, row1/family1:b/1298037744658/Put/vlen=1, row1/family1:c/1298037748020/Put/vlen

Re: mapreduce streaming with hbase as a source

2011-02-18 Thread Jean-Daniel Cryans
You have a typo, it's hbase.mapred.tablecolumns not hbase.mapred.tablecolumn J-D On Fri, Feb 18, 2011 at 6:05 AM, Ondrej Holecek wrote: > Hello, > > I'm testing hadoop and hbase, I can run mapreduce streaming or pipes jobs > agains text files on > hadoop, but I have a problem when I try to run

mapreduce streaming with hbase as a source

2011-02-18 Thread Ondrej Holecek
Hello, I'm testing hadoop and hbase, I can run mapreduce streaming or pipes jobs agains text files on hadoop, but I have a problem when I try to run the same job against hbase table. The table looks like this: hbase(main):015:0> scan 'table1' ROWCO