How much memory are you having on your machine? I think default value is
0.6 of the spark.executor.memory as you can see from here
<http://spark.apache.org/docs/1.2.1/configuration.html#execution-behavior>.

Thanks
Best Regards

On Mon, Mar 16, 2015 at 2:26 PM, Xi Shen <davidshe...@gmail.com> wrote:

> Hi Akhil,
>
> Yes, you are right. If I ran the program from IDE as a normal java
> program, the executor's memory is increased...but not to 2048m, it is set
> to 6.7GB...Looks like there's some formula to calculate this value.
>
>
> Thanks,
> David
>
>
> On Mon, Mar 16, 2015 at 7:36 PM Akhil Das <ak...@sigmoidanalytics.com>
> wrote:
>
>> By default spark.executor.memory is set to 512m, I'm assuming since you
>> are submiting the job using spark-submit and it is not able to override the
>> value since you are running in local mode. Can you try it without using
>> spark-submit as a standalone project?
>>
>> Thanks
>> Best Regards
>>
>> On Mon, Mar 16, 2015 at 1:52 PM, Xi Shen <davidshe...@gmail.com> wrote:
>>
>>> I set it in code, not by configuration. I submit my jar file to local. I
>>> am working in my developer environment.
>>>
>>> On Mon, 16 Mar 2015 18:28 Akhil Das <ak...@sigmoidanalytics.com> wrote:
>>>
>>>> How are you setting it? and how are you submitting the job?
>>>>
>>>> Thanks
>>>> Best Regards
>>>>
>>>> On Mon, Mar 16, 2015 at 12:52 PM, Xi Shen <davidshe...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I have set spark.executor.memory to 2048m, and in the UI "Environment"
>>>>> page, I can see this value has been set correctly. But in the "Executors"
>>>>> page, I saw there's only 1 executor and its memory is 265.4MB. Very 
>>>>> strange
>>>>> value. why not 256MB, or just as what I set?
>>>>>
>>>>> What am I missing here?
>>>>>
>>>>>
>>>>> Thanks,
>>>>> David
>>>>>
>>>>>
>>>>
>>

Reply via email to