My input format is CSV and I am using Spark 1.3(HDP 2,2 comes with Spark 1.3 so ...) I am using Spark-csv to read my CSV file and using dataframe API to process ... I followed these steps <http://hortonworks.com/hadoop-tutorial/using-hive-with-orc-from-apache-spark/> and succesfully able to read the ORC file . As these are temp tables and it doesnt store the data in hive. I trying to figure out how to do that ?
Thanks in advance Divya On 7 December 2015 at 17:43, Fengdong Yu <fengdo...@everstring.com> wrote: > If your RDD is JSON format, that’s easy. > > val df = sqlContext.read.json(rdd) > df.saveAsTable(“your_table_name") > > > > > On Dec 7, 2015, at 5:28 PM, Divya Gehlot <divya.htco...@gmail.com> > wrote: > > > > Hi, > > I am new bee to Spark. > > Could somebody guide me how can I persist my spark RDD results in Hive > using SaveAsTable API. > > Would appreciate if you could provide the example for hive external > table. > > > > Thanks in advance. > > > > > >