+Andrew

Actually I think this is because we haven't uploaded the Spark binaries to
cloudfront / pushed the change to mesos/spark-ec2.

Andrew, can you take care of this ?



On Tue, Dec 2, 2014 at 5:11 PM, Nicholas Chammas <nicholas.cham...@gmail.com
> wrote:

> Interesting. Do you have any problems when launching in us-east-1? What
> is the full output of spark-ec2 when launching a cluster? (Post it to a
> gist if it’s too big for email.)
> ​
>
> On Mon, Dec 1, 2014 at 10:34 AM, Dave Challis <dave.chal...@aistemos.com>
> wrote:
>
>> I've been trying to create a Spark cluster on EC2 using the
>> documentation at https://spark.apache.org/docs/latest/ec2-scripts.html
>> (with Spark 1.1.1).
>>
>> Running the script successfully creates some EC2 instances, HDFS etc.,
>> but appears to fail to copy the actual files needed to run Spark
>> across.
>>
>> I ran the following commands:
>>
>> $ cd ~/src/spark-1.1.1/ec2
>> $ ./spark-ec2 --key-pair=* --identity-file=* --slaves=1
>> --region=eu-west-1 --zone=eu-west-1a --instance-type=m3.medium
>> --no-ganglia launch foocluster
>>
>> I see the following in the script's output:
>>
>> (instance and HDFS set up happens here)
>> ...
>> Persistent HDFS installed, won't start by default...
>> ~/spark-ec2 ~/spark-ec2
>> Setting up spark-standalone
>> RSYNC'ing /root/spark/conf to slaves...
>> *****.eu-west-1.compute.amazonaws.com
>> RSYNC'ing /root/spark-ec2 to slaves...
>> *****.eu-west-1.compute.amazonaws.com
>> ./spark-standalone/setup.sh: line 22: /root/spark/sbin/stop-all.sh: No
>> such file or directory
>> ./spark-standalone/setup.sh: line 27:
>> /root/spark/sbin/start-master.sh: No such file or directory
>> ./spark-standalone/setup.sh: line 33:
>> /root/spark/sbin/start-slaves.sh: No such file or directory
>> Setting up tachyon
>> RSYNC'ing /root/tachyon to slaves...
>> ...
>> (Tachyon setup happens here without any problem)
>>
>> I can ssh to the master (using the ./spark-ec2 login), and looking in
>> /root/, it contains:
>>
>> $ ls /root
>> ephemeral-hdfs  hadoop-native  mapreduce  persistent-hdfs  scala
>> shark  spark  spark-ec2  tachyon
>>
>> If I look in /root/spark (where the sbin directory should be found),
>> it only contains a single 'conf' directory:
>>
>> $ ls /root/spark
>> conf
>>
>> Any idea why spark-ec2 might have failed to copy these files across?
>>
>> Thanks,
>> Dave
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>

Reply via email to