It is always dangerous to run a NEWER version of code on an OLDER cluster. The danger increases with the semver change and this one is not just a build #. In other word 2.4 is considered to be a fairly major change from 2.3. Not much else can be said.
From: Nicolas Paris <nicolas.pa...@riseup.net> <nicolas.pa...@riseup.net> Reply: user@spark.apache.org <user@spark.apache.org> <user@spark.apache.org> Date: May 20, 2019 at 11:02:49 AM To: user@spark.apache.org <user@spark.apache.org> <user@spark.apache.org> Subject: Re: run new spark version on old spark cluster ? > you will need the spark version you intend to launch with on the machine you > launch from and point to the correct spark-submit does this mean to install a second spark version (2.4) on the cluster ? thanks On Mon, May 20, 2019 at 01:58:11PM -0400, Koert Kuipers wrote: > yarn can happily run multiple spark versions side-by-side > you will need the spark version you intend to launch with on the machine you > launch from and point to the correct spark-submit > > On Mon, May 20, 2019 at 1:50 PM Nicolas Paris <nicolas.pa...@riseup.net> wrote: > > Hi > > I am wondering whether that's feasible to: > - build a spark application (with sbt/maven) based on spark2.4 > - deploy that jar on yarn on a spark2.3 based installation > > thanks by advance, > > > -- > nicolas > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > > -- nicolas --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org