lol - young padawan I am and path to knowledge seeking I am...

And on this path I also tried (without luck)...

                if (restId == 0) {
                        conf = conf.setExecutorEnv("spark.executor.cores", 
"22");
                } else {
                        conf = conf.setExecutorEnv("spark.executor.cores", "2");
                }

and

                if (restId == 0) {
                        conf.setExecutorEnv("spark.executor.cores", "22");
                } else {
                        conf.setExecutorEnv("spark.executor.cores", "2");
                }

the only annoying thing I see is we designed some of the work to be handled by 
the driver/client app and we will have to rethink a bit the design of the app 
for that...


> On Jul 15, 2016, at 11:34 AM, Daniel Darabos 
> <daniel.dara...@lynxanalytics.com> wrote:
> 
> Mich's invocation is for starting a Spark application against an already 
> running Spark standalone cluster. It will not start the cluster for you.
> 
> We used to not use "spark-submit", but we started using it when it solved 
> some problem for us. Perhaps that day has also come for you? :)
> 
> On Fri, Jul 15, 2016 at 5:14 PM, Jean Georges Perrin <j...@jgp.net 
> <mailto:j...@jgp.net>> wrote:
> I don't use submit: I start my standalone cluster and connect to it remotely. 
> Is that a bad practice?
> 
> I'd like to be able to it dynamically as the system knows whether it needs 
> more or less resources based on its own  context
> 
>> On Jul 15, 2016, at 10:55 AM, Mich Talebzadeh <mich.talebza...@gmail.com 
>> <mailto:mich.talebza...@gmail.com>> wrote:
>> 
>> Hi,
>> 
>> You can also do all this at env or submit time with spark-submit which I 
>> believe makes it more flexible than coding in.
>> 
>> Example
>> 
>> ${SPARK_HOME}/bin/spark-submit \
>>                 --packages com.databricks:spark-csv_2.11:1.3.0 \
>>                 --driver-memory 2G \
>>                 --num-executors 2 \
>>                 --executor-cores 3 \
>>                 --executor-memory 2G \
>>                 --master spark://50.140.197.217:7077 
>> <http://50.140.197.217:7077/> \
>>                 --conf "spark.scheduler.mode=FAIR" \
>>                 --conf "spark.executor.extraJavaOptions=-XX:+PrintGCDetails 
>> -XX:+PrintGCTimeStamps" \
>>                 --jars 
>> /home/hduser/jars/spark-streaming-kafka-assembly_2.10-1.6.1.jar \
>>                 --class "${FILE_NAME}" \
>>                 --conf "spark.ui.port=${SP}" \
>>  
>> HTH
>> 
>> Dr Mich Talebzadeh
>>  
>> LinkedIn  
>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>  
>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>
>>  
>> http://talebzadehmich.wordpress.com <http://talebzadehmich.wordpress.com/>
>> 
>> Disclaimer: Use it at your own risk. Any and all responsibility for any 
>> loss, damage or destruction of data or any other property which may arise 
>> from relying on this email's technical content is explicitly disclaimed. The 
>> author will in no case be liable for any monetary damages arising from such 
>> loss, damage or destruction.
>>  
>> 
>> On 15 July 2016 at 13:48, Jean Georges Perrin <j...@jgp.net 
>> <mailto:j...@jgp.net>> wrote:
>> Merci Nihed, this is one of the tests I did :( still not working
>> 
>> 
>> 
>>> On Jul 15, 2016, at 8:41 AM, nihed mbarek <nihe...@gmail.com 
>>> <mailto:nihe...@gmail.com>> wrote:
>>> 
>>> can you try with : 
>>> SparkConf conf = new SparkConf().setAppName("NC Eatery 
>>> app").set("spark.executor.memory", "4g")
>>>                             .setMaster("spark://10.0.100.120:7077 <>");
>>>             if (restId == 0) {
>>>                     conf = conf.set("spark.executor.cores", "22");
>>>             } else {
>>>                     conf = conf.set("spark.executor.cores", "2");
>>>             }
>>>             JavaSparkContext javaSparkContext = new JavaSparkContext(conf);
>>> 
>>> On Fri, Jul 15, 2016 at 2:31 PM, Jean Georges Perrin <j...@jgp.net 
>>> <mailto:j...@jgp.net>> wrote:
>>> Hi,
>>> 
>>> Configuration: standalone cluster, Java, Spark 1.6.2, 24 cores
>>> 
>>> My process uses all the cores of my server (good), but I am trying to limit 
>>> it so I can actually submit a second job.
>>> 
>>> I tried
>>> 
>>>             SparkConf conf = new SparkConf().setAppName("NC Eatery 
>>> app").set("spark.executor.memory", "4g")
>>>                             .setMaster("spark://10.0.100.120:7077 <>");
>>>             if (restId == 0) {
>>>                     conf = conf.set("spark.executor.cores", "22");
>>>             } else {
>>>                     conf = conf.set("spark.executor.cores", "2");
>>>             }
>>>             JavaSparkContext javaSparkContext = new JavaSparkContext(conf);
>>> 
>>> and
>>> 
>>>             SparkConf conf = new SparkConf().setAppName("NC Eatery 
>>> app").set("spark.executor.memory", "4g")
>>>                             .setMaster("spark://10.0.100.120:7077 <>");
>>>             if (restId == 0) {
>>>                     conf.set("spark.executor.cores", "22");
>>>             } else {
>>>                     conf.set("spark.executor.cores", "2");
>>>             }
>>>             JavaSparkContext javaSparkContext = new JavaSparkContext(conf);
>>> 
>>> but it does not seem to take it. Any hint?
>>> 
>>> jg
>>> 
>>> 
>>> 
>>> 
>>> 
>>> -- 
>>> 
>>> M'BAREK Med Nihed,
>>> Fedora Ambassador, TUNISIA, Northern Africa
>>> http://www.nihed.com <http://www.nihed.com/>
>>> 
>>>  <http://tn.linkedin.com/in/nihed>
>>> 
>> 
>> 
> 
> 

Reply via email to