It is suggested implicitly in giving you the command "./sbt/sbt". The separately installed sbt isn't in a folder called sbt, whereas Spark's version is. And more relevantly, just a few paragraphs earlier in the tutorial you execute the command "sbt/sbt assembly" which definitely refers to the spark install.
On Monday, March 24, 2014, Nan Zhu <zhunanmcg...@gmail.com> wrote: > I found that I never read the document carefully and I never find that > Spark document is suggesting you to use Spark-distributed sbt...... > > Best, > > -- > Nan Zhu > > > On Monday, March 24, 2014 at 5:47 PM, Diana Carroll wrote: > > Thanks for your help, everyone. Several folks have explained that I can > surely solve the problem by installing sbt. > > But I'm trying to get the instructions working *as written on the Spark > website*. The instructions not only don't have you install sbt > separately...they actually specifically have you use the sbt that is > distributed with Spark. > > If it is not possible to build your own Spark programs with > Spark-distributed sbt, then that's a big hole in the Spark docs that I > shall file. And if the sbt that is included with Spark is MEANT to be able > to compile your own Spark apps, then that's a product bug. > > But before I file the bug, I'm still hoping I'm missing something, and > someone will point out that I'm missing a small step that will make the > Spark distribution of sbt work! > > Diana > > > > On Mon, Mar 24, 2014 at 4:52 PM, Yana Kadiyska <yana.kadiy...@gmail.com>wrote: > > Diana, I just tried it on a clean Ubuntu machine, with Spark 0.8 > (since like other folks I had sbt preinstalled on my "usual" machine) > > I ran the command exactly as Ognen suggested and see > Set current project to Simple Project (do you see this -- you should > at least be seeing this) > and then a bunch of Resolving ... > > messages. I did get an error there, saying it can't find > javax.servlet.orbit. I googled the error and found this thread: > > > http://mail-archives.apache.org/mod_mbox/spark-user/201309.mbox/%3ccajbo4nexyzqe6zgreqjtzzz5zrcoavfen+wmbyced6n1epf...@mail.gmail.com%3E > > adding the IvyXML fragment they suggested helped in my case (but > again, the build pretty clearly complained). > > If you're still having no luck, I suggest installing sbt and setting > SBT_HOME... http://www.scala-sbt.org/ > > In either case though, it's not a Spark-specific issue...Hopefully > some of all this helps. > > On Mon, Mar 24, 2014 at 4:30 PM, Diana Carroll <dcarr...@cloudera.com> > wrote: > > Yeah, that's exactly what I did. Unfortunately it doesn't work: > > > > $SPARK_HOME/sbt/sbt package > > awk: cmd. line:1: fatal: cannot open file `./project/build.properties' > for > > reading (No such file or directory) > > Attempting to fetch sbt > > /usr/lib/spark/sbt/sbt: line 33: sbt/sbt-launch-.jar: No such file or > > directory > > /usr/lib/spark/sbt/sbt: line 33: sbt/sbt-launch-.jar: No such file or > > directory > > Our attempt to download sbt locally to sbt/sbt-launch-.jar failed. Please > > install sbt manually from http://www.scala-sbt.org/ > > > > > > > > On Mon, Mar 24, 2014 at 4:25 PM, Ognen Duzlevski > > <og...@plainvanillagames.com> wrote: > >> > >> You can use any sbt on your machine, including the one that comes with > >> spark. For example, try: > >> > >> ~/path_to_spark/sbt/sbt compile > >> ~/path_to_spark/sbt/sbt run <arguments> > >> > >> Or you can just add that to your PATH by: > >> > >> export $PATH=$PATH:~/path_to_spark/sbt > >> > >> To make it permanent, you can add it to your ~/.bashrc or > ~/.bash_profile > >> or ??? depending on the system you are using. If you are on Windows, > sorry, > >> I can't offer any help there ;) > >> > >> Ognen > >> > >> > >> On 3/24/14, 3:16 PM, Diana Carroll wrote: > >> > >> Thanks Ongen. > >> > >> Unfortunately I'm not able to follow your instructions either. In > >> particular: > >>> > >>> > >>> sbt compile > >>> sbt run <arguments if any> > >> > >> > >> This doesn't work for me because there's no program on my path called > >> "sbt". The instructions in the Quick Start guide are specific that I > sho > >