(moving to the hbase user ML)
I think streaming used to work correctly in hbase 0.19 since the
RowResult class was giving the value (which you had to parse out), but
now that Result is made of KeyValue and they don't include the values
in toString then I don't see how TableInputFormat could be use
I don't think you understand me correctly,
I get this line:
72 6f 77 31 keyvalues={row1/family1:a/1298037737154/Put/vlen=1,
row1/family1:b/1298037744658/Put/vlen=1,
row1/family1:c/1298037748020/Put/vlen=1}
I know "72 6f 77 31" is the key and the rest is value, let's call it
mapreduce-value.
By default, the prefix of a line
up to the first tab character is the key and the rest of the line (excluding
the tab character)
will be the value. If there is no tab character in the line, then entire
line is considered as key
and the value is null. However, this can be customized, Use:
-D stream
Thank you, I've spend a lot of time with debuging but didn't notice this typo :(
Now it works, but I don't understand one thing: On stdin I get this:
72 6f 77 31 keyvalues={row1/family1:a/1298037737154/Put/vlen=1,
row1/family1:b/1298037744658/Put/vlen=1,
row1/family1:c/1298037748020/Put/vlen
You have a typo, it's hbase.mapred.tablecolumns not hbase.mapred.tablecolumn
J-D
On Fri, Feb 18, 2011 at 6:05 AM, Ondrej Holecek wrote:
> Hello,
>
> I'm testing hadoop and hbase, I can run mapreduce streaming or pipes jobs
> agains text files on
> hadoop, but I have a problem when I try to run
Hello,
I'm testing hadoop and hbase, I can run mapreduce streaming or pipes jobs
agains text files on
hadoop, but I have a problem when I try to run the same job against hbase table.
The table looks like this:
hbase(main):015:0> scan 'table1'
ROWCO