Hi, Azuryy,

I’m working on macbook pro

so it is indeed “Users”

Best,  

--  
Nan Zhu


On Saturday, December 21, 2013 at 9:31 AM, Azuryy wrote:

> Hi Nan
>  
> I think there is a typo here:
>  
> "file:///Users/nanzhu/.m2/repository”),
>  
> It should be lowercase.
>  
>  
> Sent from my iPhone
>  
> On 2013年12月21日, at 6:17, Nan Zhu <zhunanmcg...@gmail.com 
> (mailto:zhunanmcg...@gmail.com)> wrote:
>  
> > Hi, Gary,
> >  
> > Thank you very much
> >  
> > This afternoon, I tried to compile spark with my customized hadoop, it 
> > finally works
> >  
> > For those who shared the same problem with me:
> >  
> > 1. add the following line to SparkBuild.scala
> >  
> > resolvers ++= Seq("Local Hadoop Repo" at 
> > "file:///Users/nanzhu/.m2/repository”),
> >  
> > 2. install your customized jars
> >  
> > mvn install:install-file 
> > -Dfile=/Users/nanzhu/code/hadoop-1.2.1/build/hadoop-client-1.2.2-SNAPSHOT.jar
> >  -DgroupId=org.apache.hadoop -DartifactId=hadoop-core 
> > -Dversion=1.2.2-SNAPSHOT -Dpackaging=jar -DgeneratePom=true
> >  
> > mvn install:install-file 
> > -Dfile=/Users/nanzhu/code/hadoop-1.2.1/build/hadoop-core-1.2.2-SNAPSHOT.jar 
> > -DgroupId=org.apache.hadoop -DartifactId=hadoop-core 
> > -Dversion=1.2.2-SNAPSHOT -Dpackaging=jar -DgeneratePom=true
> >  
> > 3. set SPARK_HADOOP_VERSION to 1.22-SNAPSHOT
> >  
> > 4. add the dependency of hadoop-core
> >  
> > search org.apache.hadoop" % "hadoop-client" % hadoopVersion 
> > excludeAll(excludeJackson, excludeNetty, excludeAsm, excludeCglib) in 
> > project/SparkBuild.scala
> >  
> > add "org.apache.hadoop" % "hadoop-core" % hadoopVersion 
> > excludeAll(excludeJackson, excludeNetty, excludeAsm, excludeCglib), below it
> >  
> > 5. compile spark
> >  
> > note:
> >  
> > a. in 1, I don’t know why  resolvers ++= Seq(Resolver.file("Local Maven 
> > Repo", file(Path.userHome + "/.m2/repository"))), cannot resolve my 
> > directory, so I have to manually add resolvers ++= Seq("Local Hadoop Repo" 
> > at "file:///Users/nanzhu/.m2/repository”). It is still weird that  
> > Seq("Local Hadoop Repo”, file("Users/nanzhu/.m2/repository”)) doesn’t work….
> >  
> > b. in 4, the cllient.jar dependency cannot download core.jar in automatic 
> > (why?) I have to add an explicit dependency on core.jar
> >  
> >  
> > Best,  
> >  
> > --  
> > Nan Zhu
> >  
> >  
> > On Monday, December 16, 2013 at 2:41 PM, Gary Malouf wrote:
> >  
> > > Check out the dependencies for the version of hadoop-client you are using 
> > > - I think you will find that hadoop-core is present there.
> > >  
> > >  
> > >  
> > >  
> > > On Mon, Dec 16, 2013 at 1:28 PM, Nan Zhu <zhunanmcg...@gmail.com 
> > > (mailto:zhunanmcg...@gmail.com)> wrote:
> > > > Hi, Gary,   
> > > >  
> > > > The page says Spark uses hadoop-client.jar to interact with HDFS, but 
> > > > why it also downloads hadoop-core?
> > > >  
> > > > Do I just need to change the dependency on hadoop-client to my local 
> > > > repo?  
> > > >  
> > > > Best,  
> > > >  
> > > > --  
> > > > Nan Zhu
> > > > School of Computer Science,
> > > > McGill University
> > > >  
> > > >  
> > > >  
> > > >  
> > > > On Monday, December 16, 2013 at 9:05 AM, Gary Malouf wrote:
> > > >  
> > > > > Hi Nan, check out the 'Note about Hadoop Versions' on 
> > > > > http://spark.incubator.apache.org/docs/latest/
> > > > >  
> > > > > Let us know if this does not solve your problem.  
> > > > >  
> > > > > Gary
> > > > >  
> > > > >  
> > > > > On Mon, Dec 16, 2013 at 8:19 AM, Nan Zhu <zhunanmcg...@gmail.com 
> > > > > (mailto:zhunanmcg...@gmail.com)> wrote:
> > > > > > Hi, Azuryy  
> > > > > >  
> > > > > > Thank you for the reply
> > > > > >  
> > > > > > So you compiled Spark with mvn?
> > > > > >  
> > > > > > I’m watching the pom.xml, I think it is doing the same work as 
> > > > > > SparkBuild.Scala,   
> > > > > >  
> > > > > > I’m still confused by that, in Spark, some class utilized some 
> > > > > > classes like InputFormat, I assume that this should be included in 
> > > > > > hadoop-core.jar,
> > > > > >  
> > > > > > but I didn’t find any line specified hadoop-core-1.0.4.jar in 
> > > > > > pom.xml and SparkBuild.scala,   
> > > > > >  
> > > > > > Can you explain a bit to me?
> > > > > >  
> > > > > > Best,  
> > > > > >  
> > > > > > --  
> > > > > > Nan Zhu
> > > > > > School of Computer Science,
> > > > > > McGill University
> > > > > >  
> > > > > >  
> > > > > >  
> > > > > > On Monday, December 16, 2013 at 3:58 AM, Azuryy Yu wrote:
> > > > > >  
> > > > > > > Hi Nan,
> > > > > > > I am also using our customized hadoop, so you need to modiy the 
> > > > > > > pom.xml, but before this change, you should install your 
> > > > > > > customized hadoop-* jar in the local maven repo.
> > > > > > >  
> > > > > > >  
> > > > > > >  
> > > > > > >  
> > > > > > > On Sun, Dec 15, 2013 at 2:45 AM, Nan Zhu <zhunanmcg...@gmail.com 
> > > > > > > (mailto:zhunanmcg...@gmail.com)> wrote:
> > > > > > > > Hi, all  
> > > > > > > >  
> > > > > > > > I’m trying to compile Spark with a customized version of 
> > > > > > > > hadoop, where I modify the implementation of DFSInputStream,  
> > > > > > > >  
> > > > > > > > I would like to SparkBuild.scala to make spark compile with my 
> > > > > > > > hadoop-core.xxx.jar instead of download a original one?  
> > > > > > > >  
> > > > > > > > I only found hadoop-client-xxx.jar and some lines about yarn 
> > > > > > > > jars in ScalaBuild.scala,  
> > > > > > > >  
> > > > > > > > Can you tell me which line I should modify to achieve the goal?
> > > > > > > >  
> > > > > > > > Best,  
> > > > > > > >  
> > > > > > > > --  
> > > > > > > > Nan Zhu
> > > > > > > > School of Computer Science,
> > > > > > > > McGill University
> > > > > > > >  
> > > > > > > >  
> > > > > > >  
> > > > > >  
> > > > >  
> > > >  
> > >  
> >  

Reply via email to