Hi Team, I'm not able to print the values from Spark Sql JavaSchemaRDD. Please find below my code
JavaSQLContext sqlCtx = new JavaSQLContext(sc); NewHadoopRDD<ImmutableBytesWritable, Result> rdd = new NewHadoopRDD<ImmutableBytesWritable, Result>( JavaSparkContext.toSparkContext(sc), TableInputFormat.class, ImmutableBytesWritable.class, Result.class, conf); JavaRDD<Tuple2<ImmutableBytesWritable, Result>> jrdd = rdd .toJavaRDD(); ForEachFunction f = new ForEachFunction(); JavaRDD<ANAInventory> retrdd = jrdd.map(f); JavaSchemaRDD schemaPeople = sqlCtx.applySchema(retrdd, Test.class); schemaPeople.registerAsTable("retrdd"); JavaSchemaRDD teenagers = sqlCtx.sql("SELECT * FROM retrdd"); When i add below code. It is giving compilation issue. Could you please help me to resolve this issue. List<String> teenagerNames = teenagers.map(new Function<Row, String>() { public String call(Row row) { return null; } }).collect(); for (String name: teenagerNames) { System.out.println(name); } Compilation issue : The method map(Function<Row,R> in the type JavaSchemaRDD is not applicable for the arguments (new Functiona<Row, String>(){}) Thank you for your help Regards, Rajesh