Hi Divya In general you will get better performance if you can minimize your use of UDFs. Spark 2.0/ tungsten does a lot of code generation. It will have to treat your UDF as a block box.
Andy From: Rishabh Bhardwaj <rbnex...@gmail.com> Date: Wednesday, July 20, 2016 at 4:22 AM To: Rabin Banerjee <dev.rabin.baner...@gmail.com> Cc: Divya Gehlot <divya.htco...@gmail.com>, "user @spark" <user@spark.apache.org> Subject: Re: write and call UDF in spark dataframe > Hi Divya, > > There is already "from_unixtime" exists in org.apache.spark.sql.frunctions, > Rabin has used that in the sql query,if you want to use it in dataframe DSL > you can try like this, > >> val new_df = df.select(from_unixtime($"time").as("newtime")) > > Thanks, > Rishabh. > > On Wed, Jul 20, 2016 at 4:21 PM, Rabin Banerjee <dev.rabin.baner...@gmail.com> > wrote: >> Hi Divya , >> >> Try, >> >> val df = sqlContext.sql("select from_unixtime(ts,'YYYY-MM-dd') as `ts` from >> mr") >> Regards, >> Rabin >> >> On Wed, Jul 20, 2016 at 12:44 PM, Divya Gehlot <divya.htco...@gmail.com> >> wrote: >>> Hi, >>> Could somebody share example of writing and calling udf which converts unix >>> tme stamp to date tiime . >>> >>> >>> Thanks, >>> Divya >> >