I have also managed to use Hive 1.2.1 with Spark 1.4.1
> On 26 Jan 2016, at 10:20, Mich Talebzadeh <m...@peridale.co.uk> wrote: > > As far as I have worked this one out Hive 1.2.1 works on Spark 1.3.1 for now. > This means Hive will use spark engine. > > > set spark.home=/usr/lib/spark-1.3.1-bin-hadoop2.6; > set hive.execution.engine=spark; > set spark.master=yarn-client; > set hive.optimize.ppd=true; > Beeline version 1.2.1 by Apache Hive > select count(1) from smallsales; > INFO : > Query Hive on Spark job[0] stages: > INFO : 0 > INFO : 1 > INFO : > Status: Running (Hive on Spark job[0]) > INFO : Job Progress Format > CurrentTime StageId_StageAttemptId: > SucceededTasksCount(+RunningTasksCount-FailedTasksCount)/TotalTasksCount > [StageCost] > INFO : 2016-01-26 09:27:30,240 Stage-0_0: 0(+1)/1 Stage-1_0: 0/1 > INFO : 2016-01-26 09:27:33,258 Stage-0_0: 0(+1)/1 Stage-1_0: 0/1 > INFO : 2016-01-26 09:27:35,265 Stage-0_0: 1/1 Finished Stage-1_0: 0(+1)/1 > INFO : 2016-01-26 09:27:38,281 Stage-0_0: 1/1 Finished Stage-1_0: 1/1 > Finished > INFO : Status: Finished successfully in 28.17 seconds > +----------+--+ > | _c0 | > +----------+--+ > | 5000000 | > +----------+--+ > > HTH > > Dr Mich Talebzadeh > > LinkedIn > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw> > > Sybase ASE 15 Gold Medal Award 2008 > A Winning Strategy: Running the most Critical Financial Data on ASE 15 > http://login.sybase.com/files/Product_Overviews/ASE-Winning-Strategy-091908.pdf > > <http://login.sybase.com/files/Product_Overviews/ASE-Winning-Strategy-091908.pdf> > Author of the books "A Practitioner’s Guide to Upgrading to Sybase ASE 15", > ISBN 978-0-9563693-0-7. > co-author "Sybase Transact SQL Guidelines Best Practices", ISBN > 978-0-9759693-0-4 > Publications due shortly: > Complex Event Processing in Heterogeneous Environments, ISBN: > 978-0-9563693-3-8 > Oracle and Sybase, Concepts and Contrasts, ISBN: 978-0-9563693-1-4, volume > one out shortly > > http://talebzadehmich.wordpress.com <http://talebzadehmich.wordpress.com/> > > NOTE: The information in this email is proprietary and confidential. This > message is for the designated recipient only, if you are not the intended > recipient, you should destroy it immediately. Any information in this message > shall not be understood as given or endorsed by Peridale Technology Ltd, its > subsidiaries or their employees, unless expressly so stated. It is the > responsibility of the recipient to ensure that this email is virus free, > therefore neither Peridale Technology Ltd, its subsidiaries nor their > employees accept any responsibility. > > From: kevin [mailto:kiss.kevin...@gmail.com] > Sent: 26 January 2016 08:45 > To: u...@spark.apache.org; user@hive.apache.org > Subject: hive1.2.1 on spark 1.5.2 > > hi,all > I tried hive on spark with version hive1.2.1 spark1.5.2. I build spark > witout -Phive . And I test spark cluster stand alone with spark-submit and it > is ok. > but when I use hive , on spark web-site I can see the hive on spark > application ,finally I got error: > > 16/01/26 16:23:42 INFO slf4j.Slf4jLogger: Slf4jLogger started > 16/01/26 16:23:42 INFO Remoting: Starting remoting > 16/01/26 16:23:42 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://driverPropsFetcher@10.1.3.116:42307 > <http://driverPropsFetcher@10.1.3.116:42307/>] > 16/01/26 16:23:42 INFO util.Utils: Successfully started service > 'driverPropsFetcher' on port 42307. > Exception in thread "main" akka.actor.ActorNotFound: Actor not found for: > ActorSelection[Anchor(akka.tcp://sparkDriver@10.1.3.107:34725/ > <http://sparkDriver@10.1.3.107:34725/>), Path(/user/CoarseGrainedScheduler)] > at > akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:65) > at > akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:63) > at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) > at > akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55) > at akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:73) > at > akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:74) > at > akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:120) > at > akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:73) > at > scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40) > at > scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:248) > at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:266) > at akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:89) > at > akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:935) > at akka.actor.Actor$class.aroundReceive(Actor.scala:467) > at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:411) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) > at akka.dispatch.Mailbox.run(Mailbox.scala:220) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397) > at > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > > Could anyone tell me is it for the version of hive and spark not matching ? > which version is ok or there is some other reason?