Yep, it's not serializable: https://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Result.html
You can't return this from a distributed operation since that would mean it has to travel over the network and you haven't supplied any way to convert the thing into bytes. On Tue, Mar 31, 2015 at 8:51 PM, Jeetendra Gangele <gangele...@gmail.com> wrote: > When I am trying to get the result from Hbase and running mapToPair function > of RRD its giving the error > java.io.NotSerializableException: org.apache.hadoop.hbase.client.Result > > Here is the code > > // private static JavaPairRDD<Integer, Result> > getCompanyDataRDD(JavaSparkContext sc) throws IOException { > // return sc.newAPIHadoopRDD(companyDAO.getCompnayDataConfiguration(), > TableInputFormat.class, ImmutableBytesWritable.class, > // Result.class).mapToPair(new > PairFunction<Tuple2<ImmutableBytesWritable, Result>, Integer, Result>() { > // > // public Tuple2<Integer, Result> call(Tuple2<ImmutableBytesWritable, > Result> t) throws Exception { > // System.out.println("In getCompanyDataRDD"+t._2); > // > // String cknid = Bytes.toString(t._1.get()); > // System.out.println("processing cknids is:"+cknid); > // Integer cknidInt = Integer.parseInt(cknid); > // Tuple2<Integer, Result> returnTuple = new Tuple2<Integer, > Result>(cknidInt, t._2); > // return returnTuple; > // } > // }); > // } --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org