Actually the circular dependency doesn't depend on the spark-ec2 scripts -- The scripts contain download links to many Spark versions and you can configure which one should be used.
Shivaram On Wed, Sep 3, 2014 at 10:22 AM, Matthew Farrellee <m...@redhat.com> wrote: > that's not a bad idea. it would also break the circular dep in versions > that results in spark X's ec2 script installing spark X-1 by default. > > best, > > > matt > > > On 09/03/2014 01:17 PM, Shivaram Venkataraman wrote: > >> The spark-ec2 repository isn't a part of Mesos. Back in the days, Spark >> used to be hosted in the Mesos github organization as well and so we put >> scripts that were used by Spark under the same organization. >> >> FWIW I don't think these scripts belong in the Spark repository. They are >> helper scripts that setup EC2 clusters with different components like >> HDFS, >> Spark, Tachyon etc. Also one of the motivations for creating this >> repository was the ability to change these scripts without requiring a new >> Spark release or a new AMI etc. >> >> We can move the repository to a different github organization like AMPLab >> if that'll make sense. >> >> Thanks >> Shivaram >> >> >> On Wed, Sep 3, 2014 at 10:06 AM, Nicholas Chammas < >> nicholas.cham...@gmail.com> wrote: >> >> Spawned by this discussion >>> <https://github.com/apache/spark/pull/1120#issuecomment-54305831>. >>> >>> See these 2 lines in spark_ec2.py: >>> >>> - spark_ec2 L42 >>> < >>> https://github.com/apache/spark/blob/6a72a36940311fcb3429bd34c8818b >>> c7d513115c/ec2/spark_ec2.py#L42 >>> >>>> >>>> - spark_ec2 L566 >>> < >>> https://github.com/apache/spark/blob/6a72a36940311fcb3429bd34c8818b >>> c7d513115c/ec2/spark_ec2.py#L566 >>> >>>> >>>> >>> Why does the spark-ec2 script depend on stuff in the Mesos repo? Should >>> they be moved to the Spark repo? >>> >>> Nick >>> >>> >>> >> >