Hey,

can you send the bundle you're using (in the GUI, bottom right, "export"
button should give you a bundles.yaml file, please send that to me, so I
can bootstrap the same environment as you are playing with.

also
* can you let me know if you have a file /etc/profile.d/directories.sh?
* if yes, can you execute it from your command line, then do the spark
command again, and let me know?

Thx,
Sam









Best,
Samuel

--
Samuel Cozannet
Cloud, Big Data and IoT Strategy Team
Business Development - Cloud and ISV Ecosystem
Changing the Future of Cloud
Ubuntu <http://ubuntu.com>  / Canonical UK LTD <http://canonical.com> / Juju
<https://jujucharms.com>
samuel.cozan...@canonical.com
mob: +33 616 702 389
skype: samnco
Twitter: @SaMnCo_23

On Fri, Jan 30, 2015 at 3:46 PM, Ken Williams <ke...@theasi.co> wrote:

> Ok - I have been able to add the relation using this,
>
>                 juju add-relation yarn-hdfs-master:resourcemanager
> spark-master
>
> But I still cannot see a /etc/hadoop/conf directory on the spark-master
> machine
> so I still get the same error about HADOOP_CONF_DIR and YARN_CONF_DIR
> (below),
>
>
> root@ip-172-31-60-53:~# spark-submit --class
> org.apache.spark.examples.SparkPi     --master yarn-client
> --num-executors 3     --driver-memory 1g     --executor-memory 1g
> --executor-cores 1     --queue thequeue     lib/spark-examples*.jar     10
> Spark assembly has been built with Hive, including Datanucleus jars on
> classpath
> Exception in thread "main" java.lang.Exception: When running with master
> 'yarn-client' either HADOOP_CONF_DIR or YARN_CONF_DIR must be set in the
> environment.
> at
> org.apache.spark.deploy.SparkSubmitArguments.checkRequiredArguments(SparkSubmitArguments.scala:177)
> at
> org.apache.spark.deploy.SparkSubmitArguments.<init>(SparkSubmitArguments.scala:81)
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:70)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> root@ip-172-31-60-53:~#
>
> Should there be a /etc/hadoop/conf directory ?
>
> Thanks for any help,
>
> Ken
>
>
> On 30 January 2015 at 12:59, Samuel Cozannet <
> samuel.cozan...@canonical.com> wrote:
>
>> Have you tried without ':master":
>>
>> juju add-relation yarn-hdfs-master:resourcemanager spark-master
>>
>> I think Spark master consumes the relationship but doesn't have to expose
>> its master relationship.
>>
>> Rule of thumb, when a relation is non ambiguous on one of its ends, there
>> is no requirement to specify it when adding it.
>>
>> Another option if this doesn't work is to use the GUI to create the
>> relation. It will give you a dropdown of available relationships between
>> entities.
>>
>> Let me know how it goes,
>> Thx,
>> Sam
>>
>>
>> Best,
>> Samuel
>>
>> --
>> Samuel Cozannet
>> Cloud, Big Data and IoT Strategy Team
>> Business Development - Cloud and ISV Ecosystem
>> Changing the Future of Cloud
>> Ubuntu <http://ubuntu.com>  / Canonical UK LTD <http://canonical.com> /
>> Juju <https://jujucharms.com>
>> samuel.cozan...@canonical.com
>> mob: +33 616 702 389
>> skype: samnco
>> Twitter: @SaMnCo_23
>>
>> On Fri, Jan 30, 2015 at 1:09 PM, Ken Williams <ke...@theasi.co> wrote:
>>
>>> Hi Sam,
>>>
>>>     I understand what you are saying but when I try to add the 2
>>> relations I get this error,
>>>
>>> root@adminuser-VirtualBox:~# juju add-relation
>>> yarn-hdfs-master:resourcemanager spark-master:master
>>> ERROR no relations found
>>> root@adminuser-VirtualBox:~# juju add-relation
>>> yarn-hdfs-master:namenode spark-master:master
>>> ERROR no relations found
>>>
>>>   Am I adding the relations right ?
>>>
>>>   Attached is my 'juju status' file.
>>>
>>>   Thanks for all your help,
>>>
>>> Ken
>>>
>>>
>>>
>>>
>>>
>>> On 30 January 2015 at 11:16, Samuel Cozannet <
>>> samuel.cozan...@canonical.com> wrote:
>>>
>>>> Hey Ken,
>>>>
>>>> Yes, you need to create the relationship between the 2 entities to they
>>>> know about each other.
>>>>
>>>> Looking at the list of hooks for the charm
>>>> <https://github.com/Archethought/spark-charm/tree/master/hooks> you
>>>> can see there are 2 hooks named namenode-relation-changed
>>>> <https://github.com/Archethought/spark-charm/blob/master/hooks/namenode-relation-changed>
>>>>  and resourcemanager-relation-changed
>>>> <https://github.com/Archethought/spark-charm/blob/master/hooks/resourcemanager-relation-changed>
>>>>  which
>>>> are related to YARN/Hadoop.
>>>> Looking deeper in the code, you'll notice they reference a function
>>>> found in bdutils.py called "setHadoopEnvVar()", which based on its name
>>>> should set the HADOOP_CONF_DIR.
>>>>
>>>> There are 2 relations, so add both of them.
>>>>
>>>> Note that I didn't test this myself, but I expect this should fix the
>>>> problem. If it doesn't please come back to us...
>>>>
>>>> Thanks!
>>>> Sam
>>>>
>>>>
>>>> Best,
>>>> Samuel
>>>>
>>>> --
>>>> Samuel Cozannet
>>>> Cloud, Big Data and IoT Strategy Team
>>>> Business Development - Cloud and ISV Ecosystem
>>>> Changing the Future of Cloud
>>>> Ubuntu <http://ubuntu.com>  / Canonical UK LTD <http://canonical.com> /
>>>> Juju <https://jujucharms.com>
>>>> samuel.cozan...@canonical.com
>>>> mob: +33 616 702 389
>>>> skype: samnco
>>>> Twitter: @SaMnCo_23
>>>>
>>>> On Fri, Jan 30, 2015 at 11:51 AM, Ken Williams <ke...@theasi.co> wrote:
>>>>
>>>>>
>>>>> Thanks, Kapil - this works :-)
>>>>>
>>>>> I can now run the SparkPi example successfully.
>>>>> root@ip-172-31-60-53:~# spark-submit --class
>>>>> org.apache.spark.examples.SparkPi 
>>>>> /tmp/spark-examples-1.2.0-hadoop2.4.0.jar
>>>>> Spark assembly has been built with Hive, including Datanucleus jars on
>>>>> classpath
>>>>> 15/01/30 10:29:33 WARN NativeCodeLoader: Unable to load native-hadoop
>>>>> library for your platform... using builtin-java classes where applicable
>>>>> Pi is roughly 3.14318
>>>>>
>>>>> root@ip-172-31-60-53:~#
>>>>>
>>>>> I'm now trying to run the same example with the spark-submit
>>>>> '--master' option set to either 'yarn-cluster' or 'yarn-client'
>>>>> but I keep getting the same error :
>>>>>
>>>>> root@ip-172-31-60-53:~# spark-submit --class
>>>>> org.apache.spark.examples.SparkPi     --master yarn-client
>>>>> --num-executors 3     --driver-memory 1g     --executor-memory 1g
>>>>> --executor-cores 1     --queue thequeue     lib/spark-examples*.jar     10
>>>>> Spark assembly has been built with Hive, including Datanucleus jars on
>>>>> classpath
>>>>> Exception in thread "main" java.lang.Exception: When running with
>>>>> master 'yarn-client' either HADOOP_CONF_DIR or YARN_CONF_DIR must be set 
>>>>> in
>>>>> the environment.
>>>>>
>>>>> But on my spark-master/0 machine there is no /etc/hadoop/conf
>>>>> directory.
>>>>> So what should the HADOOP_CONF_DIR or YARN_CONF_DIR value be ?
>>>>> Do I need to add a juju relation between spark-master and ...
>>>>> yarn-hdfs-master to make them aware of each other ?
>>>>>
>>>>> Thanks for any help,
>>>>>
>>>>> Ken
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On 28 January 2015 at 19:32, Kapil Thangavelu <
>>>>> kapil.thangav...@canonical.com> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Jan 28, 2015 at 1:54 PM, Ken Williams <ke...@theasi.co>
>>>>>> wrote:
>>>>>>
>>>>>>>
>>>>>>> Hi Sam/Amir,
>>>>>>>
>>>>>>>     I've been able to 'juju ssh spark-master/0' and I successfully
>>>>>>> ran the two
>>>>>>> simple examples for pyspark and spark-shell,
>>>>>>>
>>>>>>>     ./bin/pyspark
>>>>>>>     >>> sc.parallelize(range(1000)).count()
>>>>>>>     1000
>>>>>>>
>>>>>>>     ./bin/spark-shell
>>>>>>>      scala> sc.parallelize(1 to 1000).count()
>>>>>>>     1000
>>>>>>>
>>>>>>>
>>>>>>> Now I want to run some of the spark examples in the
>>>>>>> spark-exampes*.jar
>>>>>>> file, which I have on my local machine. How do I copy the jar file
>>>>>>> from
>>>>>>> my local machine to the AWS machine ?
>>>>>>>
>>>>>>> I have tried 'scp' and 'juju scp' from the local command-line but
>>>>>>> both fail (below),
>>>>>>>
>>>>>>> root@adminuser:~# scp /tmp/spark-examples-1.2.0-hadoop2.4.0.jar
>>>>>>> ubuntu@ip-172-31-59:/tmp
>>>>>>> ssh: Could not resolve hostname ip-172-31-59: Name or service not
>>>>>>> known
>>>>>>> lost connection
>>>>>>> root@adminuser:~# juju scp
>>>>>>> /tmp/spark-examples-1.2.0-hadoop2.4.0.jar ubuntu@ip-172-31-59:/tmp
>>>>>>> ERROR exit status 1 (nc: getaddrinfo: Name or service not known)
>>>>>>>
>>>>>>> Any ideas ?
>>>>>>>
>>>>>>
>>>>>> juju scp /tmp/spark-examples-1.2.0-hadoop2.4.0.jar
>>>>>> spark-master/0:/tmp
>>>>>>
>>>>>>>
>>>>>>> Ken
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On 28 January 2015 at 17:29, Samuel Cozannet <
>>>>>>> samuel.cozan...@canonical.com> wrote:
>>>>>>>
>>>>>>>> Glad it worked!
>>>>>>>>
>>>>>>>> I'll make a merge request to the upstream so that it works natively
>>>>>>>> from the store asap.
>>>>>>>>
>>>>>>>> Thanks for catching that!
>>>>>>>> Samuel
>>>>>>>>
>>>>>>>> Best,
>>>>>>>> Samuel
>>>>>>>>
>>>>>>>> --
>>>>>>>> Samuel Cozannet
>>>>>>>> Cloud, Big Data and IoT Strategy Team
>>>>>>>> Business Development - Cloud and ISV Ecosystem
>>>>>>>> Changing the Future of Cloud
>>>>>>>> Ubuntu <http://ubuntu.com>  / Canonical UK LTD
>>>>>>>> <http://canonical.com> / Juju <https://jujucharms.com>
>>>>>>>> samuel.cozan...@canonical.com
>>>>>>>> mob: +33 616 702 389
>>>>>>>> skype: samnco
>>>>>>>> Twitter: @SaMnCo_23
>>>>>>>>
>>>>>>>> On Wed, Jan 28, 2015 at 6:15 PM, Ken Williams <ke...@theasi.co>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>> Hi Sam (and Maarten),
>>>>>>>>>
>>>>>>>>>     Cloning Spark 1.2.0 from github seems to have worked!
>>>>>>>>>     I can install the Spark examples afterwards.
>>>>>>>>>
>>>>>>>>>     Thanks for all your help!
>>>>>>>>>
>>>>>>>>>     Yes - Andrew and Angie both say 'hi'  :-)
>>>>>>>>>
>>>>>>>>>     Best Regards,
>>>>>>>>>
>>>>>>>>> Ken
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On 28 January 2015 at 16:43, Samuel Cozannet <
>>>>>>>>> samuel.cozan...@canonical.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hey Ken,
>>>>>>>>>>
>>>>>>>>>> So I had a closer look to your Spark problem and found out what
>>>>>>>>>> went wrong.
>>>>>>>>>>
>>>>>>>>>> The charm available on the charmstore is trying to download Spark
>>>>>>>>>> 1.0.2, and the versions available on the Apache website are 1.1.0, 
>>>>>>>>>> 1.1.1
>>>>>>>>>> and 1.2.0.
>>>>>>>>>>
>>>>>>>>>> There is another version of the charm available on GitHub that
>>>>>>>>>> actually will deploy 1.2.0
>>>>>>>>>>
>>>>>>>>>> 1. On your computer, the below folders & get there:
>>>>>>>>>>
>>>>>>>>>> cd ~
>>>>>>>>>> mkdir charms
>>>>>>>>>> mkdir charms/trusty
>>>>>>>>>> cd charms/trusty
>>>>>>>>>>
>>>>>>>>>> 2. Branch the Spark charm.
>>>>>>>>>>
>>>>>>>>>> git clone https://github.com/Archethought/spark-charm spark
>>>>>>>>>>
>>>>>>>>>> 3. Deploy Spark from local repository
>>>>>>>>>>
>>>>>>>>>> juju deploy --repository=~/charms local:trusty/spark spark-master
>>>>>>>>>> juju deploy --repository=~/charms local:trusty/spark spark-slave
>>>>>>>>>> juju add-relation spark-master:master spark-slave:slave
>>>>>>>>>>
>>>>>>>>>> Worked on AWS for me just minutes ago. Let me know how it goes
>>>>>>>>>> for you. Note that this version of the charm does NOT install the 
>>>>>>>>>> Spark
>>>>>>>>>> examples. The files are present though, so you'll find them in
>>>>>>>>>> /var/lib/juju/agents/unit-spark-master-0/charm/files/archive
>>>>>>>>>>
>>>>>>>>>> Hope that helps...
>>>>>>>>>> Let me know if it works for you!
>>>>>>>>>>
>>>>>>>>>> Best,
>>>>>>>>>> Sam
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Best,
>>>>>>>>>> Samuel
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Samuel Cozannet
>>>>>>>>>> Cloud, Big Data and IoT Strategy Team
>>>>>>>>>> Business Development - Cloud and ISV Ecosystem
>>>>>>>>>> Changing the Future of Cloud
>>>>>>>>>> Ubuntu <http://ubuntu.com>  / Canonical UK LTD
>>>>>>>>>> <http://canonical.com> / Juju <https://jujucharms.com>
>>>>>>>>>> samuel.cozan...@canonical.com
>>>>>>>>>> mob: +33 616 702 389
>>>>>>>>>> skype: samnco
>>>>>>>>>> Twitter: @SaMnCo_23
>>>>>>>>>>
>>>>>>>>>> On Wed, Jan 28, 2015 at 4:44 PM, Ken Williams <ke...@theasi.co>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Hi folks,
>>>>>>>>>>>
>>>>>>>>>>> I'm completely new to juju so any help is appreciated.
>>>>>>>>>>>
>>>>>>>>>>> I'm trying to create a hadoop/analytics-type platform.
>>>>>>>>>>>
>>>>>>>>>>> I've managed to install the 'data-analytics-with-sql-like'
>>>>>>>>>>> bundle
>>>>>>>>>>> (using this command)
>>>>>>>>>>>
>>>>>>>>>>>     juju quickstart
>>>>>>>>>>> bundle:data-analytics-with-sql-like/data-analytics-with-sql-like
>>>>>>>>>>>
>>>>>>>>>>> This is very impressive, and gives me virtually everything that
>>>>>>>>>>> I want
>>>>>>>>>>> (hadoop, hive, etc) - but I also need Spark.
>>>>>>>>>>>
>>>>>>>>>>> The Spark charm (
>>>>>>>>>>> http://manage.jujucharms.com/~asanjar/trusty/spark)
>>>>>>>>>>> and bundle (
>>>>>>>>>>> http://manage.jujucharms.com/bundle/~asanjar/spark/spark-cluster
>>>>>>>>>>> )
>>>>>>>>>>> however do not seem stable or available and I can't figure out
>>>>>>>>>>> how to install them.
>>>>>>>>>>>
>>>>>>>>>>> Should I just download and install the Spark tar-ball on the
>>>>>>>>>>> nodes
>>>>>>>>>>> in my AWS cluster, or is there a better way to do this ?
>>>>>>>>>>>
>>>>>>>>>>> Thanks in advance,
>>>>>>>>>>>
>>>>>>>>>>> Ken
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Juju mailing list
>>>>>>>>>>> Juju@lists.ubuntu.com
>>>>>>>>>>> Modify settings or unsubscribe at:
>>>>>>>>>>> https://lists.ubuntu.com/mailman/listinfo/juju
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Juju mailing list
>>>>>>> Juju@lists.ubuntu.com
>>>>>>> Modify settings or unsubscribe at:
>>>>>>> https://lists.ubuntu.com/mailman/listinfo/juju
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
-- 
Juju mailing list
Juju@lists.ubuntu.com
Modify settings or unsubscribe at: 
https://lists.ubuntu.com/mailman/listinfo/juju

Reply via email to