If you have bandwidth, you can try the following change which would show the KeyValue that doesn't have CellType :
http://pastebin.com/TRRF1gtd You need to apply the change, build hbase-client jar and use this jar. Cheers On Sat, Sep 17, 2016 at 6:34 PM, Krishna <research...@gmail.com> wrote: > I'm using HBase 1.1 and didn't encounter issue accessing this table in > other scenarios. Data is written to this table as HBase KeyValue objects > using MapReduce jobs. > > ctx.write(rowkey, new KeyValue(rowkey.get(), "cf".getBytes(), > "cq".getBytes(), timestamp, field.getBytes())); > > On Sat, Sep 17, 2016 at 1:04 PM, Ted Yu <yuzhih...@gmail.com> wrote: > > > Here is related code from CellProtos.java : > > > > public Builder > > setCellType(org.apache.hadoop.hbase.protobuf.generated. > CellProtos.CellType > > value) { > > if (value == null) { > > throw new NullPointerException(); > > > > This means CellType.valueOf() returned null for the Cell. > > > > Which release of hbase are you using ? > > > > You didn't encounter any problem accessing <table_name> in other > scenarios > > ? > > > > Cheers > > > > On Sat, Sep 17, 2016 at 11:13 AM, Krishna <research...@gmail.com> wrote: > > > > > I'm getting NPE when attempting to export HBase table: hbase > > > org.apache.hadoop.hbase.mapreduce.Export <table_name> <hdfs_dir> > > > Does anyone know what could be causing the exception? > > > > > > Here is the error stack. > > > > > > Error: java.lang.NullPointerException > > > at > > > org.apache.hadoop.hbase.protobuf.generated.CellProtos$ > > > Cell$Builder.setCellType(CellProtos.java:1050) > > > at > > > org.apache.hadoop.hbase.protobuf.ProtobufUtil.toCell( > > > ProtobufUtil.java:2531) > > > at > > > org.apache.hadoop.hbase.protobuf.ProtobufUtil. > > toResult(ProtobufUtil.java: > > > 1303) > > > at > > > org.apache.hadoop.hbase.mapreduce.ResultSerialization$ > > > ResultSerializer.serialize(ResultSerialization.java:155) > > > at > > > org.apache.hadoop.hbase.mapreduce.ResultSerialization$ > > > ResultSerializer.serialize(ResultSerialization.java:140) > > > at > > > org.apache.hadoop.io.SequenceFile$BlockCompressWriter.append( > > > SequenceFile.java:1591) > > > at > > > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat$1. > write( > > > SequenceFileOutputFormat.java:83) > > > at > > > org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector. > > > write(MapTask.java:658) > > > at > > > org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write( > > > TaskInputOutputContextImpl.java:89) > > > at > > > org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context. > > > write(WrappedMapper.java:112) > > > at > > > org.apache.hadoop.hbase.mapreduce.IdentityTableMapper. > > > map(IdentityTableMapper.java:66) > > > at > > > org.apache.hadoop.hbase.mapreduce.IdentityTableMapper. > > > map(IdentityTableMapper.java:33) > > > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) > > > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask. > > java:787) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341) > > > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild. > java:164) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:415) > > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs( > > > UserGroupInformation.java:1693) > > > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > > > > > >