One way is to use sparkSQL.

scala> sqlContext.sql("create table orc_table(key INT, value STRING)
stored as orc")scala> sqlContext.sql("insert into table orc_table
select * from schema_rdd_temp_table")scala> sqlContext.sql("FROM
orc_table select *")



On 4 January 2015 at 00:57, SamyaMaiti <samya.maiti2...@gmail.com> wrote:

> Hi Experts,
>
> Like saveAsParquetFile on schemaRDD, there is a equivalent to store in ORC
> file.
>
> I am using spark 1.2.0.
>
> As per the link below, looks like its not part of 1.2.0, so any latest
> update would be great.
> https://issues.apache.org/jira/browse/SPARK-2883
>
> Till the next release, is there a workaround to read/write ORC file.
>
> Regards,
> Sam
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/save-rdd-to-ORC-file-tp20956.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to