OK that is good news. So briefly how do you kick off spark-submit for each
(or sparkConf). In terms of memory/resources allocations.

Now what is the output of

/usr/bin/free



Dr Mich Talebzadeh



LinkedIn * 
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 28 May 2016 at 18:12, sujeet jog <sujeet....@gmail.com> wrote:

> Yes Mich,
> They are currently emitting the results parallely,
> http://localhost:4040 &  http://localhost:4041 , i also see the
> monitoring from these URL's,
>
>
> On Sat, May 28, 2016 at 10:37 PM, Mich Talebzadeh <
> mich.talebza...@gmail.com> wrote:
>
>> ok they are submitted but the latter one 14302 is it doing anything?
>>
>> can you check it with jmonitor or the logs created
>>
>> HTH
>>
>>
>>
>> Dr Mich Talebzadeh
>>
>>
>>
>> LinkedIn * 
>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>
>>
>>
>> http://talebzadehmich.wordpress.com
>>
>>
>>
>> On 28 May 2016 at 18:03, sujeet jog <sujeet....@gmail.com> wrote:
>>
>>> Thanks Ted,
>>>
>>> Thanks Mich,  yes i see that i can run two applications by submitting
>>> these,  probably Driver + Executor running in a single JVM .  In-Process
>>> Spark.
>>>
>>> wondering if this can be used in production systems,  the reason for me
>>> considering local instead of standalone cluster mode is purely because of
>>> CPU/MEM resources,  i.e,  i currently do not have the liberty to use 1
>>> Driver & 1 Executor per application,    ( running in a embedded network
>>> switch  )
>>>
>>>
>>> jps output
>>> [root@fos-elastic02 ~]# jps
>>> 14258 SparkSubmit
>>> 14503 Jps
>>> 14302 SparkSubmit
>>> ,
>>>
>>> On Sat, May 28, 2016 at 10:21 PM, Mich Talebzadeh <
>>> mich.talebza...@gmail.com> wrote:
>>>
>>>> Ok so you want to run all this in local mode. In other words something
>>>> like below
>>>>
>>>> ${SPARK_HOME}/bin/spark-submit \
>>>>
>>>>                 --master local[2] \
>>>>
>>>>                 --driver-memory 2G \
>>>>
>>>>                 --num-executors=1 \
>>>>
>>>>                 --executor-memory=2G \
>>>>
>>>>                 --executor-cores=2 \
>>>>
>>>>
>>>> I am not sure it will work for multiple drivers (app/JVM).  The only
>>>> way you can find out is to do try it running two apps simultaneously. You
>>>> have a number of tools.
>>>>
>>>>
>>>>
>>>>    1. use jps to see the apps and PID
>>>>    2. use jmonitor to see memory/cpu/heap usage for each spark-submit
>>>>    job
>>>>
>>>> HTH
>>>>
>>>> Dr Mich Talebzadeh
>>>>
>>>>
>>>>
>>>> LinkedIn * 
>>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>>
>>>>
>>>>
>>>> http://talebzadehmich.wordpress.com
>>>>
>>>>
>>>>
>>>> On 28 May 2016 at 17:41, Ted Yu <yuzhih...@gmail.com> wrote:
>>>>
>>>>> Sujeet:
>>>>>
>>>>> Please also see:
>>>>>
>>>>> https://spark.apache.org/docs/latest/spark-standalone.html
>>>>>
>>>>> On Sat, May 28, 2016 at 9:19 AM, Mich Talebzadeh <
>>>>> mich.talebza...@gmail.com> wrote:
>>>>>
>>>>>> Hi Sujeet,
>>>>>>
>>>>>> if you have a single machine then it is Spark standalone mode.
>>>>>>
>>>>>> In Standalone cluster mode Spark allocates resources based on cores.
>>>>>> By default, an application will grab all the cores in the cluster.
>>>>>>
>>>>>> You only have one worker that lives within the driver JVM process
>>>>>> that you start when you start the application with spark-shell or
>>>>>> spark-submit in the host where the cluster manager is running.
>>>>>>
>>>>>> The Driver node runs on the same host that the cluster manager is
>>>>>> running. The Driver requests the Cluster Manager for resources to run
>>>>>> tasks. The worker is tasked to create the executor (in this case there is
>>>>>> only one executor) for the Driver. The Executor runs tasks for the 
>>>>>> Driver.
>>>>>> Only one executor can be allocated on each worker per application. In 
>>>>>> your
>>>>>> case you only have
>>>>>>
>>>>>>
>>>>>> The minimum you will need will be 2-4G of RAM and two cores. Well
>>>>>> that is my experience. Yes you can submit more than one spark-submit (the
>>>>>> driver) but they may queue up behind the running one if there is not 
>>>>>> enough
>>>>>> resources.
>>>>>>
>>>>>>
>>>>>> You pointed out that you will be running few applications in parallel
>>>>>> on the same host. The likelihood is that you are using a VM machine for
>>>>>> this purpose and the best option is to try running the first one, Check 
>>>>>> Web
>>>>>> GUI on  4040 to see the progress of this Job. If you start the next JVM
>>>>>> then assuming it is working, it will be using port 4041 and so forth.
>>>>>>
>>>>>>
>>>>>> In actual fact try the command "free" to see how much free memory you
>>>>>> have.
>>>>>>
>>>>>>
>>>>>> HTH
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Dr Mich Talebzadeh
>>>>>>
>>>>>>
>>>>>>
>>>>>> LinkedIn * 
>>>>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>>>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>>>>
>>>>>>
>>>>>>
>>>>>> http://talebzadehmich.wordpress.com
>>>>>>
>>>>>>
>>>>>>
>>>>>> On 28 May 2016 at 16:42, sujeet jog <sujeet....@gmail.com> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> I have a question w.r.t  production deployment mode of spark,
>>>>>>>
>>>>>>> I have 3 applications which i would like to run independently on a
>>>>>>> single machine, i need to run the drivers in the same machine.
>>>>>>>
>>>>>>> The amount of resources i have is also limited, like 4- 5GB RAM , 3
>>>>>>> - 4 cores.
>>>>>>>
>>>>>>> For deployment in standalone mode : i believe i need
>>>>>>>
>>>>>>> 1 Driver JVM,  1 worker node ( 1 executor )
>>>>>>> 1 Driver JVM,  1 worker node ( 1 executor )
>>>>>>> 1 Driver JVM,  1 worker node ( 1 executor )
>>>>>>>
>>>>>>> The issue here is i will require 6 JVM running in parallel, for
>>>>>>> which i do not have sufficient CPU/MEM resources,
>>>>>>>
>>>>>>>
>>>>>>> Hence i was looking more towards a local mode deployment mode, would
>>>>>>> like to know if anybody is using local mode where Driver + Executor run 
>>>>>>> in
>>>>>>> a single JVM in production mode.
>>>>>>>
>>>>>>> Are there any inherent issues upfront using local mode for
>>>>>>> production base systems.?..
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to