I think you need a Java bean class instead of a normal class. See example here: http://spark.apache.org/docs/1.1.0/sql-programming-guide.html (switch to the java tab). -Xiangrui
On Tue, Nov 11, 2014 at 7:18 AM, Naveen Kumar Pokala <npok...@spcapitaliq.com> wrote: > Hi, > > > > This is my Instrument java constructor. > > > > public Instrument(Issue issue, Issuer issuer, Issuing issuing) { > > super(); > > this.issue = issue; > > this.issuer = issuer; > > this.issuing = issuing; > > } > > > > > > I am trying to create javaschemaRDD > > > > JavaSchemaRDD schemaInstruments = sqlCtx.applySchema(distData, > Instrument.class); > > > > Remarks: > > ============ > > > > Instrument, Issue, Issuer, Issuing all are java classes > > > > distData is holding List< Instrument > > > > > > > I am getting the following error. > > > > > > > > Exception in thread "Driver" java.lang.reflect.InvocationTargetException > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:483) > > at > org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:162) > > Caused by: scala.MatchError: class sample.spark.test.Issue (of class > java.lang.Class) > > at > org.apache.spark.sql.api.java.JavaSQLContext$$anonfun$getSchema$1.apply(JavaSQLContext.scala:189) > > at > org.apache.spark.sql.api.java.JavaSQLContext$$anonfun$getSchema$1.apply(JavaSQLContext.scala:188) > > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) > > at > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) > > at > scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108) > > at > scala.collection.TraversableLike$class.map(TraversableLike.scala:244) > > at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:108) > > at > org.apache.spark.sql.api.java.JavaSQLContext.getSchema(JavaSQLContext.scala:188) > > at > org.apache.spark.sql.api.java.JavaSQLContext.applySchema(JavaSQLContext.scala:90) > > at sample.spark.test.SparkJob.main(SparkJob.java:33) > > ... 5 more > > > > Please help me. > > > > Regards, > > Naveen. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org