Ok - Sam, I'll try this and let you know. Thanks again for all your help,
Best Regards, Ken On 30 January 2015 at 18:09, Samuel Cozannet <samuel.cozan...@canonical.com> wrote: > I'll have a look asap, but probably not before Tuesday. > > This may be "my guts tell me that" but, if you have the time, try to > collocate YARN and Spark, that will guarantee you have the YARN_CONF_DIR > set. I am 90% sure it will fix your problem. > > YARN itself will not eat much resources, you should be alright and it may > allow you to move forward instead of being stuck. > > Best, > Sam > > > > > > > Best, > Samuel > > -- > Samuel Cozannet > Cloud, Big Data and IoT Strategy Team > Business Development - Cloud and ISV Ecosystem > Changing the Future of Cloud > Ubuntu <http://ubuntu.com> / Canonical UK LTD <http://canonical.com> / > Juju <https://jujucharms.com> > samuel.cozan...@canonical.com > mob: +33 616 702 389 > skype: samnco > Twitter: @SaMnCo_23 > > On Fri, Jan 30, 2015 at 7:01 PM, Ken Williams <ke...@theasi.co> wrote: > >> Hi Sam, >> >> Attached is my bundles.yaml file. >> >> Also, there is no file 'directories.sh' on my spark-master/0 machine >> (see below), >> >> ubuntu@ip-172-31-54-245:~$ ls -l /etc/profile.d/ >> total 12 >> -rw-r--r-- 1 root root 1559 Jul 29 2014 Z97-byobu.sh >> -rwxr-xr-x 1 root root 2691 Oct 6 13:19 Z99-cloud-locale-test.sh >> -rw-r--r-- 1 root root 663 Apr 7 2014 bash_completion.sh >> ubuntu@ip-172-31-54-245:~$ >> >> >> Many thanks again your help, >> >> Ken >> >> >> On 30 January 2015 at 15:45, Samuel Cozannet < >> samuel.cozan...@canonical.com> wrote: >> >>> Hey, >>> >>> can you send the bundle you're using (in the GUI, bottom right, "export" >>> button should give you a bundles.yaml file, please send that to me, so I >>> can bootstrap the same environment as you are playing with. >>> >>> also >>> * can you let me know if you have a file /etc/profile.d/directories.sh? >>> * if yes, can you execute it from your command line, then do the spark >>> command again, and let me know? >>> >>> Thx, >>> Sam >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> Best, >>> Samuel >>> >>> -- >>> Samuel Cozannet >>> Cloud, Big Data and IoT Strategy Team >>> Business Development - Cloud and ISV Ecosystem >>> Changing the Future of Cloud >>> Ubuntu <http://ubuntu.com> / Canonical UK LTD <http://canonical.com> / >>> Juju <https://jujucharms.com> >>> samuel.cozan...@canonical.com >>> mob: +33 616 702 389 >>> skype: samnco >>> Twitter: @SaMnCo_23 >>> >>> On Fri, Jan 30, 2015 at 3:46 PM, Ken Williams <ke...@theasi.co> wrote: >>> >>>> Ok - I have been able to add the relation using this, >>>> >>>> juju add-relation yarn-hdfs-master:resourcemanager >>>> spark-master >>>> >>>> But I still cannot see a /etc/hadoop/conf directory on the spark-master >>>> machine >>>> so I still get the same error about HADOOP_CONF_DIR and YARN_CONF_DIR >>>> (below), >>>> >>>> >>>> root@ip-172-31-60-53:~# spark-submit --class >>>> org.apache.spark.examples.SparkPi --master yarn-client >>>> --num-executors 3 --driver-memory 1g --executor-memory 1g >>>> --executor-cores 1 --queue thequeue lib/spark-examples*.jar 10 >>>> Spark assembly has been built with Hive, including Datanucleus jars on >>>> classpath >>>> Exception in thread "main" java.lang.Exception: When running with >>>> master 'yarn-client' either HADOOP_CONF_DIR or YARN_CONF_DIR must be set in >>>> the environment. >>>> at >>>> org.apache.spark.deploy.SparkSubmitArguments.checkRequiredArguments(SparkSubmitArguments.scala:177) >>>> at >>>> org.apache.spark.deploy.SparkSubmitArguments.<init>(SparkSubmitArguments.scala:81) >>>> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:70) >>>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) >>>> root@ip-172-31-60-53:~# >>>> >>>> Should there be a /etc/hadoop/conf directory ? >>>> >>>> Thanks for any help, >>>> >>>> Ken >>>> >>>> >>>> On 30 January 2015 at 12:59, Samuel Cozannet < >>>> samuel.cozan...@canonical.com> wrote: >>>> >>>>> Have you tried without ':master": >>>>> >>>>> juju add-relation yarn-hdfs-master:resourcemanager spark-master >>>>> >>>>> I think Spark master consumes the relationship but doesn't have to >>>>> expose its master relationship. >>>>> >>>>> Rule of thumb, when a relation is non ambiguous on one of its ends, >>>>> there is no requirement to specify it when adding it. >>>>> >>>>> Another option if this doesn't work is to use the GUI to create the >>>>> relation. It will give you a dropdown of available relationships between >>>>> entities. >>>>> >>>>> Let me know how it goes, >>>>> Thx, >>>>> Sam >>>>> >>>>> >>>>> Best, >>>>> Samuel >>>>> >>>>> -- >>>>> Samuel Cozannet >>>>> Cloud, Big Data and IoT Strategy Team >>>>> Business Development - Cloud and ISV Ecosystem >>>>> Changing the Future of Cloud >>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD <http://canonical.com> / >>>>> Juju <https://jujucharms.com> >>>>> samuel.cozan...@canonical.com >>>>> mob: +33 616 702 389 >>>>> skype: samnco >>>>> Twitter: @SaMnCo_23 >>>>> >>>>> On Fri, Jan 30, 2015 at 1:09 PM, Ken Williams <ke...@theasi.co> wrote: >>>>> >>>>>> Hi Sam, >>>>>> >>>>>> I understand what you are saying but when I try to add the 2 >>>>>> relations I get this error, >>>>>> >>>>>> root@adminuser-VirtualBox:~# juju add-relation >>>>>> yarn-hdfs-master:resourcemanager spark-master:master >>>>>> ERROR no relations found >>>>>> root@adminuser-VirtualBox:~# juju add-relation >>>>>> yarn-hdfs-master:namenode spark-master:master >>>>>> ERROR no relations found >>>>>> >>>>>> Am I adding the relations right ? >>>>>> >>>>>> Attached is my 'juju status' file. >>>>>> >>>>>> Thanks for all your help, >>>>>> >>>>>> Ken >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On 30 January 2015 at 11:16, Samuel Cozannet < >>>>>> samuel.cozan...@canonical.com> wrote: >>>>>> >>>>>>> Hey Ken, >>>>>>> >>>>>>> Yes, you need to create the relationship between the 2 entities to >>>>>>> they know about each other. >>>>>>> >>>>>>> Looking at the list of hooks for the charm >>>>>>> <https://github.com/Archethought/spark-charm/tree/master/hooks> you >>>>>>> can see there are 2 hooks named namenode-relation-changed >>>>>>> <https://github.com/Archethought/spark-charm/blob/master/hooks/namenode-relation-changed> >>>>>>> and resourcemanager-relation-changed >>>>>>> <https://github.com/Archethought/spark-charm/blob/master/hooks/resourcemanager-relation-changed> >>>>>>> which >>>>>>> are related to YARN/Hadoop. >>>>>>> Looking deeper in the code, you'll notice they reference a function >>>>>>> found in bdutils.py called "setHadoopEnvVar()", which based on its name >>>>>>> should set the HADOOP_CONF_DIR. >>>>>>> >>>>>>> There are 2 relations, so add both of them. >>>>>>> >>>>>>> Note that I didn't test this myself, but I expect this should fix >>>>>>> the problem. If it doesn't please come back to us... >>>>>>> >>>>>>> Thanks! >>>>>>> Sam >>>>>>> >>>>>>> >>>>>>> Best, >>>>>>> Samuel >>>>>>> >>>>>>> -- >>>>>>> Samuel Cozannet >>>>>>> Cloud, Big Data and IoT Strategy Team >>>>>>> Business Development - Cloud and ISV Ecosystem >>>>>>> Changing the Future of Cloud >>>>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD >>>>>>> <http://canonical.com> / Juju <https://jujucharms.com> >>>>>>> samuel.cozan...@canonical.com >>>>>>> mob: +33 616 702 389 >>>>>>> skype: samnco >>>>>>> Twitter: @SaMnCo_23 >>>>>>> >>>>>>> On Fri, Jan 30, 2015 at 11:51 AM, Ken Williams <ke...@theasi.co> >>>>>>> wrote: >>>>>>> >>>>>>>> >>>>>>>> Thanks, Kapil - this works :-) >>>>>>>> >>>>>>>> I can now run the SparkPi example successfully. >>>>>>>> root@ip-172-31-60-53:~# spark-submit --class >>>>>>>> org.apache.spark.examples.SparkPi >>>>>>>> /tmp/spark-examples-1.2.0-hadoop2.4.0.jar >>>>>>>> Spark assembly has been built with Hive, including Datanucleus jars >>>>>>>> on classpath >>>>>>>> 15/01/30 10:29:33 WARN NativeCodeLoader: Unable to load >>>>>>>> native-hadoop library for your platform... using builtin-java classes >>>>>>>> where >>>>>>>> applicable >>>>>>>> Pi is roughly 3.14318 >>>>>>>> >>>>>>>> root@ip-172-31-60-53:~# >>>>>>>> >>>>>>>> I'm now trying to run the same example with the spark-submit >>>>>>>> '--master' option set to either 'yarn-cluster' or 'yarn-client' >>>>>>>> but I keep getting the same error : >>>>>>>> >>>>>>>> root@ip-172-31-60-53:~# spark-submit --class >>>>>>>> org.apache.spark.examples.SparkPi --master yarn-client >>>>>>>> --num-executors 3 --driver-memory 1g --executor-memory 1g >>>>>>>> --executor-cores 1 --queue thequeue lib/spark-examples*.jar >>>>>>>> 10 >>>>>>>> Spark assembly has been built with Hive, including Datanucleus jars >>>>>>>> on classpath >>>>>>>> Exception in thread "main" java.lang.Exception: When running with >>>>>>>> master 'yarn-client' either HADOOP_CONF_DIR or YARN_CONF_DIR must be >>>>>>>> set in >>>>>>>> the environment. >>>>>>>> >>>>>>>> But on my spark-master/0 machine there is no /etc/hadoop/conf >>>>>>>> directory. >>>>>>>> So what should the HADOOP_CONF_DIR or YARN_CONF_DIR value be ? >>>>>>>> Do I need to add a juju relation between spark-master and ... >>>>>>>> yarn-hdfs-master to make them aware of each other ? >>>>>>>> >>>>>>>> Thanks for any help, >>>>>>>> >>>>>>>> Ken >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On 28 January 2015 at 19:32, Kapil Thangavelu < >>>>>>>> kapil.thangav...@canonical.com> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, Jan 28, 2015 at 1:54 PM, Ken Williams <ke...@theasi.co> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> Hi Sam/Amir, >>>>>>>>>> >>>>>>>>>> I've been able to 'juju ssh spark-master/0' and I >>>>>>>>>> successfully ran the two >>>>>>>>>> simple examples for pyspark and spark-shell, >>>>>>>>>> >>>>>>>>>> ./bin/pyspark >>>>>>>>>> >>> sc.parallelize(range(1000)).count() >>>>>>>>>> 1000 >>>>>>>>>> >>>>>>>>>> ./bin/spark-shell >>>>>>>>>> scala> sc.parallelize(1 to 1000).count() >>>>>>>>>> 1000 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Now I want to run some of the spark examples in the >>>>>>>>>> spark-exampes*.jar >>>>>>>>>> file, which I have on my local machine. How do I copy the jar >>>>>>>>>> file from >>>>>>>>>> my local machine to the AWS machine ? >>>>>>>>>> >>>>>>>>>> I have tried 'scp' and 'juju scp' from the local command-line but >>>>>>>>>> both fail (below), >>>>>>>>>> >>>>>>>>>> root@adminuser:~# scp /tmp/spark-examples-1.2.0-hadoop2.4.0.jar >>>>>>>>>> ubuntu@ip-172-31-59:/tmp >>>>>>>>>> ssh: Could not resolve hostname ip-172-31-59: Name or service not >>>>>>>>>> known >>>>>>>>>> lost connection >>>>>>>>>> root@adminuser:~# juju scp >>>>>>>>>> /tmp/spark-examples-1.2.0-hadoop2.4.0.jar ubuntu@ip-172-31-59 >>>>>>>>>> :/tmp >>>>>>>>>> ERROR exit status 1 (nc: getaddrinfo: Name or service not known) >>>>>>>>>> >>>>>>>>>> Any ideas ? >>>>>>>>>> >>>>>>>>> >>>>>>>>> juju scp /tmp/spark-examples-1.2.0-hadoop2.4.0.jar >>>>>>>>> spark-master/0:/tmp >>>>>>>>> >>>>>>>>>> >>>>>>>>>> Ken >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On 28 January 2015 at 17:29, Samuel Cozannet < >>>>>>>>>> samuel.cozan...@canonical.com> wrote: >>>>>>>>>> >>>>>>>>>>> Glad it worked! >>>>>>>>>>> >>>>>>>>>>> I'll make a merge request to the upstream so that it works >>>>>>>>>>> natively from the store asap. >>>>>>>>>>> >>>>>>>>>>> Thanks for catching that! >>>>>>>>>>> Samuel >>>>>>>>>>> >>>>>>>>>>> Best, >>>>>>>>>>> Samuel >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Samuel Cozannet >>>>>>>>>>> Cloud, Big Data and IoT Strategy Team >>>>>>>>>>> Business Development - Cloud and ISV Ecosystem >>>>>>>>>>> Changing the Future of Cloud >>>>>>>>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD >>>>>>>>>>> <http://canonical.com> / Juju <https://jujucharms.com> >>>>>>>>>>> samuel.cozan...@canonical.com >>>>>>>>>>> mob: +33 616 702 389 >>>>>>>>>>> skype: samnco >>>>>>>>>>> Twitter: @SaMnCo_23 >>>>>>>>>>> >>>>>>>>>>> On Wed, Jan 28, 2015 at 6:15 PM, Ken Williams <ke...@theasi.co> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Hi Sam (and Maarten), >>>>>>>>>>>> >>>>>>>>>>>> Cloning Spark 1.2.0 from github seems to have worked! >>>>>>>>>>>> I can install the Spark examples afterwards. >>>>>>>>>>>> >>>>>>>>>>>> Thanks for all your help! >>>>>>>>>>>> >>>>>>>>>>>> Yes - Andrew and Angie both say 'hi' :-) >>>>>>>>>>>> >>>>>>>>>>>> Best Regards, >>>>>>>>>>>> >>>>>>>>>>>> Ken >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On 28 January 2015 at 16:43, Samuel Cozannet < >>>>>>>>>>>> samuel.cozan...@canonical.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hey Ken, >>>>>>>>>>>>> >>>>>>>>>>>>> So I had a closer look to your Spark problem and found out >>>>>>>>>>>>> what went wrong. >>>>>>>>>>>>> >>>>>>>>>>>>> The charm available on the charmstore is trying to download >>>>>>>>>>>>> Spark 1.0.2, and the versions available on the Apache website are >>>>>>>>>>>>> 1.1.0, >>>>>>>>>>>>> 1.1.1 and 1.2.0. >>>>>>>>>>>>> >>>>>>>>>>>>> There is another version of the charm available on GitHub that >>>>>>>>>>>>> actually will deploy 1.2.0 >>>>>>>>>>>>> >>>>>>>>>>>>> 1. On your computer, the below folders & get there: >>>>>>>>>>>>> >>>>>>>>>>>>> cd ~ >>>>>>>>>>>>> mkdir charms >>>>>>>>>>>>> mkdir charms/trusty >>>>>>>>>>>>> cd charms/trusty >>>>>>>>>>>>> >>>>>>>>>>>>> 2. Branch the Spark charm. >>>>>>>>>>>>> >>>>>>>>>>>>> git clone https://github.com/Archethought/spark-charm spark >>>>>>>>>>>>> >>>>>>>>>>>>> 3. Deploy Spark from local repository >>>>>>>>>>>>> >>>>>>>>>>>>> juju deploy --repository=~/charms local:trusty/spark >>>>>>>>>>>>> spark-master >>>>>>>>>>>>> juju deploy --repository=~/charms local:trusty/spark >>>>>>>>>>>>> spark-slave >>>>>>>>>>>>> juju add-relation spark-master:master spark-slave:slave >>>>>>>>>>>>> >>>>>>>>>>>>> Worked on AWS for me just minutes ago. Let me know how it goes >>>>>>>>>>>>> for you. Note that this version of the charm does NOT install the >>>>>>>>>>>>> Spark >>>>>>>>>>>>> examples. The files are present though, so you'll find them in >>>>>>>>>>>>> /var/lib/juju/agents/unit-spark-master-0/charm/files/archive >>>>>>>>>>>>> >>>>>>>>>>>>> Hope that helps... >>>>>>>>>>>>> Let me know if it works for you! >>>>>>>>>>>>> >>>>>>>>>>>>> Best, >>>>>>>>>>>>> Sam >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Best, >>>>>>>>>>>>> Samuel >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Samuel Cozannet >>>>>>>>>>>>> Cloud, Big Data and IoT Strategy Team >>>>>>>>>>>>> Business Development - Cloud and ISV Ecosystem >>>>>>>>>>>>> Changing the Future of Cloud >>>>>>>>>>>>> Ubuntu <http://ubuntu.com> / Canonical UK LTD >>>>>>>>>>>>> <http://canonical.com> / Juju <https://jujucharms.com> >>>>>>>>>>>>> samuel.cozan...@canonical.com >>>>>>>>>>>>> mob: +33 616 702 389 >>>>>>>>>>>>> skype: samnco >>>>>>>>>>>>> Twitter: @SaMnCo_23 >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, Jan 28, 2015 at 4:44 PM, Ken Williams <ke...@theasi.co >>>>>>>>>>>>> > wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Hi folks, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I'm completely new to juju so any help is appreciated. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I'm trying to create a hadoop/analytics-type platform. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I've managed to install the 'data-analytics-with-sql-like' >>>>>>>>>>>>>> bundle >>>>>>>>>>>>>> (using this command) >>>>>>>>>>>>>> >>>>>>>>>>>>>> juju quickstart >>>>>>>>>>>>>> bundle:data-analytics-with-sql-like/data-analytics-with-sql-like >>>>>>>>>>>>>> >>>>>>>>>>>>>> This is very impressive, and gives me virtually everything >>>>>>>>>>>>>> that I want >>>>>>>>>>>>>> (hadoop, hive, etc) - but I also need Spark. >>>>>>>>>>>>>> >>>>>>>>>>>>>> The Spark charm ( >>>>>>>>>>>>>> http://manage.jujucharms.com/~asanjar/trusty/spark) >>>>>>>>>>>>>> and bundle ( >>>>>>>>>>>>>> http://manage.jujucharms.com/bundle/~asanjar/spark/spark-cluster >>>>>>>>>>>>>> ) >>>>>>>>>>>>>> however do not seem stable or available and I can't figure >>>>>>>>>>>>>> out how to install them. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Should I just download and install the Spark tar-ball on the >>>>>>>>>>>>>> nodes >>>>>>>>>>>>>> in my AWS cluster, or is there a better way to do this ? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks in advance, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Ken >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Juju mailing list >>>>>>>>>>>>>> Juju@lists.ubuntu.com >>>>>>>>>>>>>> Modify settings or unsubscribe at: >>>>>>>>>>>>>> https://lists.ubuntu.com/mailman/listinfo/juju >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Juju mailing list >>>>>>>>>> Juju@lists.ubuntu.com >>>>>>>>>> Modify settings or unsubscribe at: >>>>>>>>>> https://lists.ubuntu.com/mailman/listinfo/juju >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
-- Juju mailing list Juju@lists.ubuntu.com Modify settings or unsubscribe at: https://lists.ubuntu.com/mailman/listinfo/juju