Hi all, Sorry the referenced link is not using a private/own branch of hive. It's using Hortonworks 2.3 and the hive packaged in HDP2.3, and installed a standalone version of Spark cluster (1.5.2)
But the Hive on Spark cannot run. Could anyone help on this? Thanks a lot! Regards, Sai On Wed, Dec 2, 2015 at 9:58 PM Ted Yu <yuzhih...@gmail.com> wrote: > The referenced link seems to be w.r.t. Hive on Spark which is still in its > own branch of Hive. > > FYI > > On Tue, Dec 1, 2015 at 11:23 PM, 张炜 <zhangwei...@gmail.com> wrote: > >> Hello Ted and all, >> We are using Hive 1.2.1 and Spark 1.5.1 >> I also noticed that there are other users reporting this problem. >> >> http://apache-spark-user-list.1001560.n3.nabble.com/Issue-with-spark-on-hive-td25372.html#a25486 >> Thanks a lot for help! >> >> Regards, >> Sai >> >> On Wed, Dec 2, 2015 at 11:11 AM Ted Yu <yuzhih...@gmail.com> wrote: >> >>> Can you tell us the version of Spark and hive you use ? >>> >>> Thanks >>> >>> On Tue, Dec 1, 2015 at 7:08 PM, 张炜 <zhangwei...@gmail.com> wrote: >>> >>>> Dear all, >>>> We have a requirement that needs to update delete records in hive. >>>> These operations are available in hive now. >>>> >>>> But when using hiveContext in Spark, it always pops up an "not >>>> supported" error. >>>> Is there anyway to support update/delete operations using spark? >>>> >>>> Regards, >>>> Sai >>>> >>> >>> >