Hi Mich, sure that workers are mentioned in slaves file. I can see them in spark master UI and even after start they are "blocked" for this application but the cpu and memory consumption is close to nothing.
Thanks Jakub On 4 July 2016 at 18:36, Mich Talebzadeh <mich.talebza...@gmail.com> wrote: > Silly question. Have you added your workers to sbin/slaves file and have > you started start-slaves.sh. > > on master node when you type jps what do you see? > > The problem seems to be that workers are ignored and spark is essentially > running in Local mode > > HTH > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > *Disclaimer:* Use it at your own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such loss, damage or destruction. > > > > On 4 July 2016 at 17:05, Jakub Stransky <stransky...@gmail.com> wrote: > >> Hi Mich, >> >> I have set up spark default configuration in conf directory >> spark-defaults.conf where I specify master hence no need to put it in >> command line >> spark.master spark://spark.master:7077 >> >> the same applies to driver memory which has been increased to 4GB >> and the same is for spark.executor.memory 12GB as machines have 16GB >> >> Jakub >> >> >> >> >> On 4 July 2016 at 17:44, Mich Talebzadeh <mich.talebza...@gmail.com> >> wrote: >> >>> Hi Jakub, >>> >>> In standalone mode Spark does the resource management. Which version of >>> Spark are you running? >>> >>> How do you define your SparkConf() parameters for example setMaster etc. >>> >>> From >>> >>> spark-submit --driver-class-path spark/sqljdbc4.jar --class DemoApp >>> SparkPOC.jar 10 4.3 >>> >>> I did not see any executor, memory allocation, so I assume you are >>> allocating them somewhere else? >>> >>> HTH >>> >>> >>> >>> Dr Mich Talebzadeh >>> >>> >>> >>> LinkedIn * >>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>> >>> >>> >>> http://talebzadehmich.wordpress.com >>> >>> >>> *Disclaimer:* Use it at your own risk. Any and all responsibility for >>> any loss, damage or destruction of data or any other property which may >>> arise from relying on this email's technical content is explicitly >>> disclaimed. The author will in no case be liable for any monetary damages >>> arising from such loss, damage or destruction. >>> >>> >>> >>> On 4 July 2016 at 16:31, Jakub Stransky <stransky...@gmail.com> wrote: >>> >>>> Hello, >>>> >>>> I have a spark cluster consisting of 4 nodes in a standalone mode, >>>> master + 3 workers nodes with configured available memory and cpus etc. >>>> >>>> I have an spark application which is essentially a MLlib pipeline for >>>> training a classifier, in this case RandomForest but could be a >>>> DecesionTree just for the sake of simplicity. >>>> >>>> But when I submit the spark application to the cluster via spark submit >>>> it is running out of memory. Even though the executors are "taken"/created >>>> in the cluster they are esentially doing nothing ( poor cpu, nor memory >>>> utilization) while the master seems to do all the work which finally >>>> results in OOM. >>>> >>>> My submission is following: >>>> spark-submit --driver-class-path spark/sqljdbc4.jar --class DemoApp >>>> SparkPOC.jar 10 4.3 >>>> >>>> I am submitting from the master node. >>>> >>>> By default it is running in client mode which the driver process is >>>> attached to spark-shell. >>>> >>>> Do I need to set up some settings to make MLlib algos parallelized and >>>> distributed as well or all is driven by parallel factor set on dataframe >>>> with input data? >>>> >>>> Essentially it seems that all work is just done on master and the rest >>>> is idle. >>>> Any hints what to check? >>>> >>>> Thx >>>> Jakub >>>> >>>> >>>> >>>> >>> >> >> >> -- >> Jakub Stransky >> cz.linkedin.com/in/jakubstransky >> >> > -- Jakub Stransky cz.linkedin.com/in/jakubstransky