If you have bandwidth, you can try the following change which would show
the KeyValue that doesn't have CellType :

http://pastebin.com/TRRF1gtd

You need to apply the change, build hbase-client jar and use this jar.

Cheers

On Sat, Sep 17, 2016 at 6:34 PM, Krishna <research...@gmail.com> wrote:

> I'm using HBase 1.1 and didn't encounter issue accessing this table in
> other scenarios. Data is written to this table as HBase KeyValue objects
> using MapReduce jobs.
>
> ctx.write(rowkey, new KeyValue(rowkey.get(), "cf".getBytes(),
> "cq".getBytes(), timestamp, field.getBytes()));
>
> On Sat, Sep 17, 2016 at 1:04 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
> > Here is related code from CellProtos.java :
> >
> >       public Builder
> > setCellType(org.apache.hadoop.hbase.protobuf.generated.
> CellProtos.CellType
> > value) {
> >         if (value == null) {
> >           throw new NullPointerException();
> >
> > This means CellType.valueOf() returned null for the Cell.
> >
> > Which release of hbase are you using ?
> >
> > You didn't encounter any problem accessing <table_name> in other
> scenarios
> > ?
> >
> > Cheers
> >
> > On Sat, Sep 17, 2016 at 11:13 AM, Krishna <research...@gmail.com> wrote:
> >
> > > I'm getting NPE when attempting to export HBase table: hbase
> > > org.apache.hadoop.hbase.mapreduce.Export <table_name> <hdfs_dir>
> > > Does anyone know what could be causing the exception?
> > >
> > > Here is the error stack.
> > >
> > > Error: java.lang.NullPointerException
> > >         at
> > > org.apache.hadoop.hbase.protobuf.generated.CellProtos$
> > > Cell$Builder.setCellType(CellProtos.java:1050)
> > >         at
> > > org.apache.hadoop.hbase.protobuf.ProtobufUtil.toCell(
> > > ProtobufUtil.java:2531)
> > >         at
> > > org.apache.hadoop.hbase.protobuf.ProtobufUtil.
> > toResult(ProtobufUtil.java:
> > > 1303)
> > >         at
> > > org.apache.hadoop.hbase.mapreduce.ResultSerialization$
> > > ResultSerializer.serialize(ResultSerialization.java:155)
> > >         at
> > > org.apache.hadoop.hbase.mapreduce.ResultSerialization$
> > > ResultSerializer.serialize(ResultSerialization.java:140)
> > >         at
> > > org.apache.hadoop.io.SequenceFile$BlockCompressWriter.append(
> > > SequenceFile.java:1591)
> > >         at
> > > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat$1.
> write(
> > > SequenceFileOutputFormat.java:83)
> > >         at
> > > org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.
> > > write(MapTask.java:658)
> > >         at
> > > org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(
> > > TaskInputOutputContextImpl.java:89)
> > >         at
> > > org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.
> > > write(WrappedMapper.java:112)
> > >         at
> > > org.apache.hadoop.hbase.mapreduce.IdentityTableMapper.
> > > map(IdentityTableMapper.java:66)
> > >         at
> > > org.apache.hadoop.hbase.mapreduce.IdentityTableMapper.
> > > map(IdentityTableMapper.java:33)
> > >         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)
> > >         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.
> > java:787)
> > >         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
> > >         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.
> java:164)
> > >         at java.security.AccessController.doPrivileged(Native Method)
> > >         at javax.security.auth.Subject.doAs(Subject.java:415)
> > >         at
> > > org.apache.hadoop.security.UserGroupInformation.doAs(
> > > UserGroupInformation.java:1693)
> > >         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> > >
> >
>

Reply via email to