Install your custom spark jar to your local maven or ivy repo. Use this custom 
jar in your pom/sbt file. 



> On May 15, 2014, at 3:28 AM, Andrei <faithlessfri...@gmail.com> wrote:
> 
> (Sorry if you have already seen this message - it seems like there were some 
> issues delivering messages to the list yesterday)
> 
> We can create standalone Spark application by simply adding "spark-core_2.x" 
> to build.sbt/pom.xml and connecting it to Spark master. 
> 
> We can also build custom version of Spark (e.g. compiled against Hadoop 2.x) 
> from source and deploy it to cluster manually. 
> 
> But what is a proper way to use _custom version_ of Spark in _standalone 
> application_? 
> 
> 
> I'm currently trying to deploy custom version to local Maven repository and 
> add it to SBT project. Another option is to add Spark as local jar to every 
> project. But both of these ways look overcomplicated and in general wrong. 
> 
> So what is the implied way to do it? 
> 
> Thanks, 
> Andrei

Reply via email to