I'm using HBase 1.1 and didn't encounter issue accessing this table in
other scenarios. Data is written to this table as HBase KeyValue objects
using MapReduce jobs.

ctx.write(rowkey, new KeyValue(rowkey.get(), "cf".getBytes(),
"cq".getBytes(), timestamp, field.getBytes()));

On Sat, Sep 17, 2016 at 1:04 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> Here is related code from CellProtos.java :
>
>       public Builder
> setCellType(org.apache.hadoop.hbase.protobuf.generated.CellProtos.CellType
> value) {
>         if (value == null) {
>           throw new NullPointerException();
>
> This means CellType.valueOf() returned null for the Cell.
>
> Which release of hbase are you using ?
>
> You didn't encounter any problem accessing <table_name> in other scenarios
> ?
>
> Cheers
>
> On Sat, Sep 17, 2016 at 11:13 AM, Krishna <research...@gmail.com> wrote:
>
> > I'm getting NPE when attempting to export HBase table: hbase
> > org.apache.hadoop.hbase.mapreduce.Export <table_name> <hdfs_dir>
> > Does anyone know what could be causing the exception?
> >
> > Here is the error stack.
> >
> > Error: java.lang.NullPointerException
> >         at
> > org.apache.hadoop.hbase.protobuf.generated.CellProtos$
> > Cell$Builder.setCellType(CellProtos.java:1050)
> >         at
> > org.apache.hadoop.hbase.protobuf.ProtobufUtil.toCell(
> > ProtobufUtil.java:2531)
> >         at
> > org.apache.hadoop.hbase.protobuf.ProtobufUtil.
> toResult(ProtobufUtil.java:
> > 1303)
> >         at
> > org.apache.hadoop.hbase.mapreduce.ResultSerialization$
> > ResultSerializer.serialize(ResultSerialization.java:155)
> >         at
> > org.apache.hadoop.hbase.mapreduce.ResultSerialization$
> > ResultSerializer.serialize(ResultSerialization.java:140)
> >         at
> > org.apache.hadoop.io.SequenceFile$BlockCompressWriter.append(
> > SequenceFile.java:1591)
> >         at
> > org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat$1.write(
> > SequenceFileOutputFormat.java:83)
> >         at
> > org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.
> > write(MapTask.java:658)
> >         at
> > org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(
> > TaskInputOutputContextImpl.java:89)
> >         at
> > org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.
> > write(WrappedMapper.java:112)
> >         at
> > org.apache.hadoop.hbase.mapreduce.IdentityTableMapper.
> > map(IdentityTableMapper.java:66)
> >         at
> > org.apache.hadoop.hbase.mapreduce.IdentityTableMapper.
> > map(IdentityTableMapper.java:33)
> >         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)
> >         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.
> java:787)
> >         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
> >         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
> >         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:415)
> >         at
> > org.apache.hadoop.security.UserGroupInformation.doAs(
> > UserGroupInformation.java:1693)
> >         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> >
>

Reply via email to