Thanks a lot. I will give a try! On Monday, March 16, 2015, Adam Lewandowski <adam.lewandow...@gmail.com> wrote:
> Prior to 1.3.0, Spark has 'spark.files.userClassPathFirst' for non-yarn > apps. For 1.3.0, use 'spark.executor.userClassPathFirst'. > > See > https://mail-archives.apache.org/mod_mbox/spark-user/201503.mbox/%3CCALrvLxdWwSByxNvcZtTVo8BsNRR_7tbPzWdUiAV8Ps8H1oAayQ%40mail.gmail.com%3E > > On Fri, Mar 13, 2015 at 1:04 PM, Shuai Zheng <szheng.c...@gmail.com > <javascript:_e(%7B%7D,'cvml','szheng.c...@gmail.com');>> wrote: > >> Hi All, >> >> >> >> I am running spark to deal with AWS. >> >> >> >> And aws sdk latest version is working with httpclient 3.4+. Then but >> spark-assembly-*-.jar file has packaged an old httpclient version which >> cause me: ClassNotFoundException for >> org/apache/http/client/methods/HttpPatch >> >> >> >> Even when I put the right httpclient jar there, it won’t help because >> spark always take the class from same packaging first. >> >> >> >> I don’t know why spark only provide a big package which doesn’t allow us >> to customize the library loading sequence. I know I can just rebuild the >> spark, but this is very troublesome, and it should not be a general >> solution for long term (I can’t rebuild spark jar every time when have a >> jar conflict as spark is supposed to be a cluster). >> >> >> >> In hadoop, we have “mapreduce.job.user.classpath.first=true”. But >> “spark.yarn.user.classpath.first” only work for Yarn. >> >> >> >> I think I am not the one who face this issue. Anyone has a more general >> solution for this? >> >> >> >> Regards, >> >> >> >> Shuai >> >> >> >> >> > >