Is this what you are looking for?

https://spark.apache.org/docs/1.0.0/api/java/org/apache/spark/sql/parquet/InsertIntoParquetTable.html

According to the doc, it says "Operator that acts as a sink for queries on
RDDs and can be used to store the output inside a directory of Parquet
files. This operator is similar to Hive's INSERT INTO TABLE operation in
the sense that one can choose to either overwrite or append to a directory.
Note that consecutive insertions to the same table must have compatible
(source) schemas."

Thanks
Best Regards


On Thu, Jul 17, 2014 at 11:42 AM, Hu, Leo <leo.h...@sap.com> wrote:

>  Hi
>
>    As for spark 1.0, can we insert and update a table with SPARK SQL, and
> how?
>
>
>
> Thanks
>
> Best Regard
>

Reply via email to