run JPS like below

 jps
19724 SparkSubmit
10612 Worker

and do ps awx|grep  PID

 for each number that represents these two descriptions. something like

ps awx|grep 30208
30208 pts/2    Sl+    1:05 /usr/java/latest/bin/java -cp
/home/hduser/jars/jconn4.jar:/home/hduser/jars/ojdbc6.jar:/usr/lib/spark-1.6.1-bin-hadoop2.6/conf/:/usr/lib/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar:/usr/lib/spark-1.6.1-bin-hadoop2.6/lib/datanucleus-api-jdo-3.2.6.jar:/usr/lib/spark-1.6.1-bin-hadoop2.6/lib/datanucleus-core-3.2.10.jar:/usr/lib/spark-1.6.1-bin-hadoop2.6/lib/datanucleus-rdbms-3.2.9.jar:/home/hduser/hadoop-2.6.0/etc/hadoop/
-Xms4g -Xmx4g *org.apache.spark.deploy.SparkSubmit --master
spark://50.140.197.217:7077 <http://50.140.197.217:7077> --conf
spark.driver.memory=4g --class CEP_streaming --num-executors 1
--executor-memory 4G --executor-cores 2 *--packages
com.databricks:spark-csv_2.11:1.3.0 --jars
/home/hduser/jars/spark-streaming-kafka-assembly_2.10-1.6.1.jar

Also

send the output of OS command for available memory


*free*             total       used       free     shared    buffers
cached
Mem:      24546308   24398672     147636          0     347464   17130900
-/+ buffers/cache:    6920308   17626000
Swap:      2031608     226288    1805320

HTH

Dr Mich Talebzadeh



LinkedIn * 
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 11 May 2016 at 19:47, شجاع الرحمن بیگ <shujamug...@gmail.com> wrote:

> yes, i m running this as standalone mode.
>
> On Wed, May 11, 2016 at 6:23 PM, Mich Talebzadeh <
> mich.talebza...@gmail.com> wrote:
>
>> are you running this in standalone  mode? that is one physical host, and
>> the executor will live inside the driver.
>>
>>
>>
>> Dr Mich Talebzadeh
>>
>>
>>
>> LinkedIn * 
>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>
>>
>>
>> http://talebzadehmich.wordpress.com
>>
>>
>>
>> On 11 May 2016 at 16:45, شجاع الرحمن بیگ <shujamug...@gmail.com> wrote:
>>
>>> yes,
>>>
>>> On Wed, May 11, 2016 at 5:43 PM, Deepak Sharma <deepakmc...@gmail.com>
>>> wrote:
>>>
>>>> Since you are registering workers from the same node , do you have
>>>> enough cores and RAM(In this case >=9 cores and > = 24 GB ) on this
>>>> node(11.14.224.24)?
>>>>
>>>> Thanks
>>>> Deepak
>>>>
>>>> On Wed, May 11, 2016 at 9:08 PM, شجاع الرحمن بیگ <shujamug...@gmail.com
>>>> > wrote:
>>>>
>>>>> Hi All,
>>>>>
>>>>> I need to set same memory and core for each worker on same machine and
>>>>> for this purpose, I have set the following properties in conf/spark-env.sh
>>>>>
>>>>> export SPARK_EXECUTOR_INSTANCE=3
>>>>> export SPARK_WORKER_CORES=3
>>>>> export SPARK_WORKER_MEMORY=8g
>>>>>
>>>>> but only one worker is getting desired memory and cores and other two
>>>>> are not. here is the log of master.
>>>>>
>>>>> ...
>>>>> 6/05/11 17:04:40 INFO Master: I have been elected leader! New state:
>>>>> ALIVE
>>>>> 16/05/11 17:04:43 INFO Master: Registering worker 11.14.224.24:53923
>>>>> with 3 cores, 8.0 GB RAM
>>>>> 16/05/11 17:04:49 INFO Master: Registering worker 11.14.224.24:55072
>>>>> with 2 cores, 1020.7 GB RAM
>>>>> 16/05/11 17:05:07 INFO Master: Registering worker 11.14.224.24:49702
>>>>> with 2 cores, 1020.7 GB RAM
>>>>> ...
>>>>>
>>>>>
>>>>> Could you please let me know the solution
>>>>>
>>>>> Thanks
>>>>> Shuja
>>>>>
>>>>> --
>>>>> Regards
>>>>> Shuja-ur-Rehman Baig
>>>>> <http://pk.linkedin.com/in/shujamughal>
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Thanks
>>>> Deepak
>>>> www.bigdatabig.com
>>>> www.keosha.net
>>>>
>>>
>>>
>>>
>>> --
>>> Regards
>>> Shuja-ur-Rehman Baig
>>> <http://pk.linkedin.com/in/shujamughal>
>>>
>>
>>
>
>
> --
> Regards
> Shuja-ur-Rehman Baig
> <http://pk.linkedin.com/in/shujamughal>
>

Reply via email to