It says sdp_d not found, since it is a class you need to instantiate it once. like:
sc.textFile("derby.log").map(_.split(",")).map( r => { val upto_time = sdf.parse(r(23).trim); calendar.setTime(upto_time); val r23 = new java.sql.Timestamp(upto_time.getTime); val insert_time = sdf.parse(r(26).trim); calendar.setTime(insert_time); val r26 = new java.sql.Timestamp(insert_time.getTime); val last_upd_time = sdf.parse(r(27).trim); calendar.setTime(last_upd_time); val r27 = new java.sql.Timestamp(last_upd_time.getTime); *new* *sdp_d(r(0).trim.toInt, r(1).trim.toInt, r(2).trim, r(3).trim.toInt, r(4).trim.toInt, r(5).trim, r(6).trim.toInt, r(7).trim, r(8).trim.toDouble, r(9).trim.toDouble, r(10).trim, r(11).trim, r(12).trim, r(13).trim, r(14).trim, r(15).trim, r(16).trim, r(17).trim, r(18).trim, r(19).trim, r(20).trim, r(21).trim.toInt, r(22).trim, r23, r(24).trim, r(25).trim, r26, r27, r(28).trim)* }).registerAsTable("sdp") Thanks Best Regards On Wed, Feb 25, 2015 at 2:14 PM, anamika gupta <anamika.guo...@gmail.com> wrote: > The link has proved helpful. I have been able to load data, register it as > a table and perform simple queries. Thanks Akhil !! > > Though, I still look forward to knowing where I was going wrong with my > previous technique of extending the Product Interface to overcome case > class's limit of 22 fields. > > On Wed, Feb 25, 2015 at 9:45 AM, anamika gupta <anamika.guo...@gmail.com> > wrote: > >> Hi Akhil >> >> I guess it skipped my attention. I would definitely give it a try. >> >> While I would still like to know what is the issue with the way I have >> created schema? >> >> On Tue, Feb 24, 2015 at 4:35 PM, Akhil Das <ak...@sigmoidanalytics.com> >> wrote: >> >>> Did you happen to have a look at >>> https://spark.apache.org/docs/latest/sql-programming-guide.html#programmatically-specifying-the-schema >>> >>> Thanks >>> Best Regards >>> >>> On Tue, Feb 24, 2015 at 3:39 PM, anu <anamika.guo...@gmail.com> wrote: >>> >>>> My issue is posted here on stack-overflow. What am I doing wrong here? >>>> >>>> >>>> http://stackoverflow.com/questions/28689186/facing-error-while-extending-scala-class-with-product-interface-to-overcome-limi >>>> >>>> ------------------------------ >>>> View this message in context: Facing error while extending scala class >>>> with Product interface to overcome limit of 22 fields in spark-shell >>>> <http://apache-spark-user-list.1001560.n3.nabble.com/Facing-error-while-extending-scala-class-with-Product-interface-to-overcome-limit-of-22-fields-in-spl-tp21787.html> >>>> Sent from the Apache Spark User List mailing list archive >>>> <http://apache-spark-user-list.1001560.n3.nabble.com/> at Nabble.com. >>>> >>> >>> >> >