Excellent! Happy to help you through your discovery of awesomeness with Juju :)
Note that, if you have jq installed (which I advise, sudo apt-get install jq) juju stat | python -c 'import sys, yaml, json; json.dump(yaml.load(sys.stdin), sys.stdout, indent=4)' | jq '.services."<SERVICE NAME>".units."<SERVICE NAME>/0".machine' | tr -d "\"" will return the ID of the machine for <SERVICE NAME> (replace that by yarn-master or what ever the name you gave), which saves you the browsing of several pages of juju status... Let us know how your testing goes! Best, Sam Best, Samuel -- Samuel Cozannet Cloud, Big Data and IoT Strategy Team Business Development - Cloud and ISV Ecosystem Changing the Future of Cloud Ubuntu <http://ubuntu.com> / Canonical UK LTD <http://canonical.com> / Juju <https://jujucharms.com> samuel.cozan...@canonical.com mob: +33 616 702 389 skype: samnco Twitter: @SaMnCo_23 On Mon, Feb 2, 2015 at 2:52 PM, Ken Williams <ke...@theasi.co> wrote: > Hi Sam, > > Just to confirm that deploying the spark-master and the > yarn-hdfs-master to > the same machine seems to have worked ! :-) > > // use 'juju status' to find which machine yarn-hdfs-master is on > juju status > [ etc...] > // say...machine: "4" > > // deploy spark-master to same machine > juju deploy --to 4 spark-master > > // add relations > juju add-relation yarn-hdfs-master:resourcemanager spark-master > juju add-relation yarn-hdfs-master:namenode spark-master > > > // run test > root@ip-172-31-21-92:~# spark-submit --class > org.apache.spark.examples.SparkPi /tmp/spark-examples-1.2.0-hadoop2.4.0.jar > 25 --master yarn --num-executors 3 --driver-memory 1g > --executor-memory 1g --executor-cores 1 --deploy-mode cluster > --queue thequeue /tmp/spark-examples-1.2.0-hadoop2.4.0.jar > Spark assembly has been built with Hive, including Datanucleus jars on > classpath > 15/02/02 13:40:45 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > Pi is roughly 3.1405888 > > > Many thanks again for all your help, > > Best Regards, > > Ken > > > > On 30 January 2015 at 18:11, Ken Williams <ke...@theasi.co> wrote: > >> >> Ok - Sam, I'll try this and let you know. >> >> Thanks again for all your help, >> >> Best Regards, >> >> Ken >> >> >> >> On 30 January 2015 at 18:09, Samuel Cozannet < >> samuel.cozan...@canonical.com> wrote: >> >>> I'll have a look asap, but probably not before Tuesday. >>> >>> This may be "my guts tell me that" but, if you have the time, try to >>> collocate YARN and Spark, that will guarantee you have the YARN_CONF_DIR >>> set. I am 90% sure it will fix your problem. >>> >>> YARN itself will not eat much resources, you should be alright and it >>> may allow you to move forward instead of being stuck. >>> >>> Best, >>> Sam >>> >>> >>> >>> >>> >>> >>> Best, >>> Samuel >>> >>> -- >>> Samuel Cozannet >>> Cloud, Big Data and IoT Strategy Team >>> Business Development - Cloud and ISV Ecosystem >>> Changing the Future of Cloud >>> Ubuntu <http://ubuntu.com> / Canonical UK LTD <http://canonical.com> / >>> Juju <https://jujucharms.com> >>> samuel.cozan...@canonical.com >>> mob: +33 616 702 389 >>> skype: samnco >>> Twitter: @SaMnCo_23 >>> >>> On Fri, Jan 30, 2015 at 7:01 PM, Ken Williams <ke...@theasi.co> wrote: >>> >>>> Hi Sam, >>>> >>>> Attached is my bundles.yaml file. >>>> >>>> Also, there is no file 'directories.sh' on my spark-master/0 >>>> machine (see below), >>>> >>>> ubuntu@ip-172-31-54-245:~$ ls -l /etc/profile.d/ >>>> total 12 >>>> -rw-r--r-- 1 root root 1559 Jul 29 2014 Z97-byobu.sh >>>> -rwxr-xr-x 1 root root 2691 Oct 6 13:19 Z99-cloud-locale-test.sh >>>> -rw-r--r-- 1 root root 663 Apr 7 2014 bash_completion.sh >>>> ubuntu@ip-172-31-54-245:~$ >>>> >>>> >>>> Many thanks again your help, >>>> >>>> Ken >>>> >>>> >>>> On 30 January 2015 at 15:45, Samuel Cozannet < >>>> samuel.cozan...@canonical.com> wrote: >>>> >>>>> Hey, >>>>> >>>>> can you send the bundle you're using (in the GUI, bottom right, >>>>> "export" button should give you a bundles.yaml file, please send that to >>>>> me, so I can bootstrap the same environment as you are playing with. >>>>> >>>>> also >>>>> * can you let me know if you have a file /etc/profile.d/directories.sh? >>>>> * if yes, can you execute it from your command line, then do the spark >>>>> command again, and let me know? >>>>> >>>>> Thx, >>>>> Sam >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> Best, >>>>> Samuel >>>>> >>>>> -- >>>>> Samuel Cozannet >>>>> Cloud, Big Data and IoT Strategy Team >>>>> Business Development - Cloud and ISV Ecosystem >>>>> Changing the Future of Cloud >>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD <http://canonical.com> / >>>>> Juju <https://jujucharms.com> >>>>> samuel.cozan...@canonical.com >>>>> mob: +33 616 702 389 >>>>> skype: samnco >>>>> Twitter: @SaMnCo_23 >>>>> >>>>> On Fri, Jan 30, 2015 at 3:46 PM, Ken Williams <ke...@theasi.co> wrote: >>>>> >>>>>> Ok - I have been able to add the relation using this, >>>>>> >>>>>> juju add-relation yarn-hdfs-master:resourcemanager >>>>>> spark-master >>>>>> >>>>>> But I still cannot see a /etc/hadoop/conf directory on the >>>>>> spark-master machine >>>>>> so I still get the same error about HADOOP_CONF_DIR and YARN_CONF_DIR >>>>>> (below), >>>>>> >>>>>> >>>>>> root@ip-172-31-60-53:~# spark-submit --class >>>>>> org.apache.spark.examples.SparkPi --master yarn-client >>>>>> --num-executors 3 --driver-memory 1g --executor-memory 1g >>>>>> --executor-cores 1 --queue thequeue lib/spark-examples*.jar >>>>>> 10 >>>>>> Spark assembly has been built with Hive, including Datanucleus jars >>>>>> on classpath >>>>>> Exception in thread "main" java.lang.Exception: When running with >>>>>> master 'yarn-client' either HADOOP_CONF_DIR or YARN_CONF_DIR must be set >>>>>> in >>>>>> the environment. >>>>>> at >>>>>> org.apache.spark.deploy.SparkSubmitArguments.checkRequiredArguments(SparkSubmitArguments.scala:177) >>>>>> at >>>>>> org.apache.spark.deploy.SparkSubmitArguments.<init>(SparkSubmitArguments.scala:81) >>>>>> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:70) >>>>>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >>>>>> root@ip-172-31-60-53:~# >>>>>> >>>>>> Should there be a /etc/hadoop/conf directory ? >>>>>> >>>>>> Thanks for any help, >>>>>> >>>>>> Ken >>>>>> >>>>>> >>>>>> On 30 January 2015 at 12:59, Samuel Cozannet < >>>>>> samuel.cozan...@canonical.com> wrote: >>>>>> >>>>>>> Have you tried without ':master": >>>>>>> >>>>>>> juju add-relation yarn-hdfs-master:resourcemanager spark-master >>>>>>> >>>>>>> I think Spark master consumes the relationship but doesn't have to >>>>>>> expose its master relationship. >>>>>>> >>>>>>> Rule of thumb, when a relation is non ambiguous on one of its ends, >>>>>>> there is no requirement to specify it when adding it. >>>>>>> >>>>>>> Another option if this doesn't work is to use the GUI to create the >>>>>>> relation. It will give you a dropdown of available relationships between >>>>>>> entities. >>>>>>> >>>>>>> Let me know how it goes, >>>>>>> Thx, >>>>>>> Sam >>>>>>> >>>>>>> >>>>>>> Best, >>>>>>> Samuel >>>>>>> >>>>>>> -- >>>>>>> Samuel Cozannet >>>>>>> Cloud, Big Data and IoT Strategy Team >>>>>>> Business Development - Cloud and ISV Ecosystem >>>>>>> Changing the Future of Cloud >>>>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD >>>>>>> <http://canonical.com> / Juju <https://jujucharms.com> >>>>>>> samuel.cozan...@canonical.com >>>>>>> mob: +33 616 702 389 >>>>>>> skype: samnco >>>>>>> Twitter: @SaMnCo_23 >>>>>>> >>>>>>> On Fri, Jan 30, 2015 at 1:09 PM, Ken Williams <ke...@theasi.co> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Sam, >>>>>>>> >>>>>>>> I understand what you are saying but when I try to add the 2 >>>>>>>> relations I get this error, >>>>>>>> >>>>>>>> root@adminuser-VirtualBox:~# juju add-relation >>>>>>>> yarn-hdfs-master:resourcemanager spark-master:master >>>>>>>> ERROR no relations found >>>>>>>> root@adminuser-VirtualBox:~# juju add-relation >>>>>>>> yarn-hdfs-master:namenode spark-master:master >>>>>>>> ERROR no relations found >>>>>>>> >>>>>>>> Am I adding the relations right ? >>>>>>>> >>>>>>>> Attached is my 'juju status' file. >>>>>>>> >>>>>>>> Thanks for all your help, >>>>>>>> >>>>>>>> Ken >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On 30 January 2015 at 11:16, Samuel Cozannet < >>>>>>>> samuel.cozan...@canonical.com> wrote: >>>>>>>> >>>>>>>>> Hey Ken, >>>>>>>>> >>>>>>>>> Yes, you need to create the relationship between the 2 entities to >>>>>>>>> they know about each other. >>>>>>>>> >>>>>>>>> Looking at the list of hooks for the charm >>>>>>>>> <https://github.com/Archethought/spark-charm/tree/master/hooks> you >>>>>>>>> can see there are 2 hooks named namenode-relation-changed >>>>>>>>> <https://github.com/Archethought/spark-charm/blob/master/hooks/namenode-relation-changed> >>>>>>>>> and resourcemanager-relation-changed >>>>>>>>> <https://github.com/Archethought/spark-charm/blob/master/hooks/resourcemanager-relation-changed> >>>>>>>>> which >>>>>>>>> are related to YARN/Hadoop. >>>>>>>>> Looking deeper in the code, you'll notice they reference a >>>>>>>>> function found in bdutils.py called "setHadoopEnvVar()", which based >>>>>>>>> on its >>>>>>>>> name should set the HADOOP_CONF_DIR. >>>>>>>>> >>>>>>>>> There are 2 relations, so add both of them. >>>>>>>>> >>>>>>>>> Note that I didn't test this myself, but I expect this should fix >>>>>>>>> the problem. If it doesn't please come back to us... >>>>>>>>> >>>>>>>>> Thanks! >>>>>>>>> Sam >>>>>>>>> >>>>>>>>> >>>>>>>>> Best, >>>>>>>>> Samuel >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Samuel Cozannet >>>>>>>>> Cloud, Big Data and IoT Strategy Team >>>>>>>>> Business Development - Cloud and ISV Ecosystem >>>>>>>>> Changing the Future of Cloud >>>>>>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD >>>>>>>>> <http://canonical.com> / Juju <https://jujucharms.com> >>>>>>>>> samuel.cozan...@canonical.com >>>>>>>>> mob: +33 616 702 389 >>>>>>>>> skype: samnco >>>>>>>>> Twitter: @SaMnCo_23 >>>>>>>>> >>>>>>>>> On Fri, Jan 30, 2015 at 11:51 AM, Ken Williams <ke...@theasi.co> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> Thanks, Kapil - this works :-) >>>>>>>>>> >>>>>>>>>> I can now run the SparkPi example successfully. >>>>>>>>>> root@ip-172-31-60-53:~# spark-submit --class >>>>>>>>>> org.apache.spark.examples.SparkPi >>>>>>>>>> /tmp/spark-examples-1.2.0-hadoop2.4.0.jar >>>>>>>>>> Spark assembly has been built with Hive, including Datanucleus >>>>>>>>>> jars on classpath >>>>>>>>>> 15/01/30 10:29:33 WARN NativeCodeLoader: Unable to load >>>>>>>>>> native-hadoop library for your platform... using builtin-java >>>>>>>>>> classes where >>>>>>>>>> applicable >>>>>>>>>> Pi is roughly 3.14318 >>>>>>>>>> >>>>>>>>>> root@ip-172-31-60-53:~# >>>>>>>>>> >>>>>>>>>> I'm now trying to run the same example with the spark-submit >>>>>>>>>> '--master' option set to either 'yarn-cluster' or 'yarn-client' >>>>>>>>>> but I keep getting the same error : >>>>>>>>>> >>>>>>>>>> root@ip-172-31-60-53:~# spark-submit --class >>>>>>>>>> org.apache.spark.examples.SparkPi --master yarn-client >>>>>>>>>> --num-executors 3 --driver-memory 1g --executor-memory 1g >>>>>>>>>> --executor-cores 1 --queue thequeue lib/spark-examples*.jar >>>>>>>>>> 10 >>>>>>>>>> Spark assembly has been built with Hive, including Datanucleus >>>>>>>>>> jars on classpath >>>>>>>>>> Exception in thread "main" java.lang.Exception: When running with >>>>>>>>>> master 'yarn-client' either HADOOP_CONF_DIR or YARN_CONF_DIR must be >>>>>>>>>> set in >>>>>>>>>> the environment. >>>>>>>>>> >>>>>>>>>> But on my spark-master/0 machine there is no /etc/hadoop/conf >>>>>>>>>> directory. >>>>>>>>>> So what should the HADOOP_CONF_DIR or YARN_CONF_DIR value be ? >>>>>>>>>> Do I need to add a juju relation between spark-master and ... >>>>>>>>>> yarn-hdfs-master to make them aware of each other ? >>>>>>>>>> >>>>>>>>>> Thanks for any help, >>>>>>>>>> >>>>>>>>>> Ken >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On 28 January 2015 at 19:32, Kapil Thangavelu < >>>>>>>>>> kapil.thangav...@canonical.com> wrote: >>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Wed, Jan 28, 2015 at 1:54 PM, Ken Williams <ke...@theasi.co> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Hi Sam/Amir, >>>>>>>>>>>> >>>>>>>>>>>> I've been able to 'juju ssh spark-master/0' and I >>>>>>>>>>>> successfully ran the two >>>>>>>>>>>> simple examples for pyspark and spark-shell, >>>>>>>>>>>> >>>>>>>>>>>> ./bin/pyspark >>>>>>>>>>>> >>> sc.parallelize(range(1000)).count() >>>>>>>>>>>> 1000 >>>>>>>>>>>> >>>>>>>>>>>> ./bin/spark-shell >>>>>>>>>>>> scala> sc.parallelize(1 to 1000).count() >>>>>>>>>>>> 1000 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Now I want to run some of the spark examples in the >>>>>>>>>>>> spark-exampes*.jar >>>>>>>>>>>> file, which I have on my local machine. How do I copy the jar >>>>>>>>>>>> file from >>>>>>>>>>>> my local machine to the AWS machine ? >>>>>>>>>>>> >>>>>>>>>>>> I have tried 'scp' and 'juju scp' from the local command-line >>>>>>>>>>>> but both fail (below), >>>>>>>>>>>> >>>>>>>>>>>> root@adminuser:~# scp >>>>>>>>>>>> /tmp/spark-examples-1.2.0-hadoop2.4.0.jar ubuntu@ip-172-31-59 >>>>>>>>>>>> :/tmp >>>>>>>>>>>> ssh: Could not resolve hostname ip-172-31-59: Name or service >>>>>>>>>>>> not known >>>>>>>>>>>> lost connection >>>>>>>>>>>> root@adminuser:~# juju scp >>>>>>>>>>>> /tmp/spark-examples-1.2.0-hadoop2.4.0.jar ubuntu@ip-172-31-59 >>>>>>>>>>>> :/tmp >>>>>>>>>>>> ERROR exit status 1 (nc: getaddrinfo: Name or service not known) >>>>>>>>>>>> >>>>>>>>>>>> Any ideas ? >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> juju scp /tmp/spark-examples-1.2.0-hadoop2.4.0.jar >>>>>>>>>>> spark-master/0:/tmp >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Ken >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On 28 January 2015 at 17:29, Samuel Cozannet < >>>>>>>>>>>> samuel.cozan...@canonical.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Glad it worked! >>>>>>>>>>>>> >>>>>>>>>>>>> I'll make a merge request to the upstream so that it works >>>>>>>>>>>>> natively from the store asap. >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks for catching that! >>>>>>>>>>>>> Samuel >>>>>>>>>>>>> >>>>>>>>>>>>> Best, >>>>>>>>>>>>> Samuel >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Samuel Cozannet >>>>>>>>>>>>> Cloud, Big Data and IoT Strategy Team >>>>>>>>>>>>> Business Development - Cloud and ISV Ecosystem >>>>>>>>>>>>> Changing the Future of Cloud >>>>>>>>>>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD >>>>>>>>>>>>> <http://canonical.com> / Juju <https://jujucharms.com> >>>>>>>>>>>>> samuel.cozan...@canonical.com >>>>>>>>>>>>> mob: +33 616 702 389 >>>>>>>>>>>>> skype: samnco >>>>>>>>>>>>> Twitter: @SaMnCo_23 >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, Jan 28, 2015 at 6:15 PM, Ken Williams <ke...@theasi.co >>>>>>>>>>>>> > wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Sam (and Maarten), >>>>>>>>>>>>>> >>>>>>>>>>>>>> Cloning Spark 1.2.0 from github seems to have worked! >>>>>>>>>>>>>> I can install the Spark examples afterwards. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks for all your help! >>>>>>>>>>>>>> >>>>>>>>>>>>>> Yes - Andrew and Angie both say 'hi' :-) >>>>>>>>>>>>>> >>>>>>>>>>>>>> Best Regards, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Ken >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 28 January 2015 at 16:43, Samuel Cozannet < >>>>>>>>>>>>>> samuel.cozan...@canonical.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hey Ken, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> So I had a closer look to your Spark problem and found out >>>>>>>>>>>>>>> what went wrong. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> The charm available on the charmstore is trying to download >>>>>>>>>>>>>>> Spark 1.0.2, and the versions available on the Apache website >>>>>>>>>>>>>>> are 1.1.0, >>>>>>>>>>>>>>> 1.1.1 and 1.2.0. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> There is another version of the charm available on GitHub >>>>>>>>>>>>>>> that actually will deploy 1.2.0 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 1. On your computer, the below folders & get there: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> cd ~ >>>>>>>>>>>>>>> mkdir charms >>>>>>>>>>>>>>> mkdir charms/trusty >>>>>>>>>>>>>>> cd charms/trusty >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 2. Branch the Spark charm. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> git clone https://github.com/Archethought/spark-charm spark >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 3. Deploy Spark from local repository >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> juju deploy --repository=~/charms local:trusty/spark >>>>>>>>>>>>>>> spark-master >>>>>>>>>>>>>>> juju deploy --repository=~/charms local:trusty/spark >>>>>>>>>>>>>>> spark-slave >>>>>>>>>>>>>>> juju add-relation spark-master:master spark-slave:slave >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Worked on AWS for me just minutes ago. Let me know how it >>>>>>>>>>>>>>> goes for you. Note that this version of the charm does NOT >>>>>>>>>>>>>>> install the >>>>>>>>>>>>>>> Spark examples. The files are present though, so you'll find >>>>>>>>>>>>>>> them in >>>>>>>>>>>>>>> /var/lib/juju/agents/unit-spark-master-0/charm/files/archive >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hope that helps... >>>>>>>>>>>>>>> Let me know if it works for you! >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Best, >>>>>>>>>>>>>>> Sam >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Best, >>>>>>>>>>>>>>> Samuel >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Samuel Cozannet >>>>>>>>>>>>>>> Cloud, Big Data and IoT Strategy Team >>>>>>>>>>>>>>> Business Development - Cloud and ISV Ecosystem >>>>>>>>>>>>>>> Changing the Future of Cloud >>>>>>>>>>>>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD >>>>>>>>>>>>>>> <http://canonical.com> / Juju <https://jujucharms.com> >>>>>>>>>>>>>>> samuel.cozan...@canonical.com >>>>>>>>>>>>>>> mob: +33 616 702 389 >>>>>>>>>>>>>>> skype: samnco >>>>>>>>>>>>>>> Twitter: @SaMnCo_23 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Wed, Jan 28, 2015 at 4:44 PM, Ken Williams < >>>>>>>>>>>>>>> ke...@theasi.co> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi folks, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I'm completely new to juju so any help is appreciated. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I'm trying to create a hadoop/analytics-type platform. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I've managed to install the 'data-analytics-with-sql-like' >>>>>>>>>>>>>>>> bundle >>>>>>>>>>>>>>>> (using this command) >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> juju quickstart >>>>>>>>>>>>>>>> bundle:data-analytics-with-sql-like/data-analytics-with-sql-like >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> This is very impressive, and gives me virtually everything >>>>>>>>>>>>>>>> that I want >>>>>>>>>>>>>>>> (hadoop, hive, etc) - but I also need Spark. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> The Spark charm ( >>>>>>>>>>>>>>>> http://manage.jujucharms.com/~asanjar/trusty/spark) >>>>>>>>>>>>>>>> and bundle ( >>>>>>>>>>>>>>>> http://manage.jujucharms.com/bundle/~asanjar/spark/spark-cluster >>>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>>> however do not seem stable or available and I can't figure >>>>>>>>>>>>>>>> out how to install them. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Should I just download and install the Spark tar-ball on >>>>>>>>>>>>>>>> the nodes >>>>>>>>>>>>>>>> in my AWS cluster, or is there a better way to do this ? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks in advance, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Ken >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Juju mailing list >>>>>>>>>>>>>>>> Juju@lists.ubuntu.com >>>>>>>>>>>>>>>> Modify settings or unsubscribe at: >>>>>>>>>>>>>>>> https://lists.ubuntu.com/mailman/listinfo/juju >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Juju mailing list >>>>>>>>>>>> Juju@lists.ubuntu.com >>>>>>>>>>>> Modify settings or unsubscribe at: >>>>>>>>>>>> https://lists.ubuntu.com/mailman/listinfo/juju >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
-- Juju mailing list Juju@lists.ubuntu.com Modify settings or unsubscribe at: https://lists.ubuntu.com/mailman/listinfo/juju