I set it in code, not by configuration. I submit my jar file to local. I am working in my developer environment.
On Mon, 16 Mar 2015 18:28 Akhil Das <ak...@sigmoidanalytics.com> wrote: > How are you setting it? and how are you submitting the job? > > Thanks > Best Regards > > On Mon, Mar 16, 2015 at 12:52 PM, Xi Shen <davidshe...@gmail.com> wrote: > >> Hi, >> >> I have set spark.executor.memory to 2048m, and in the UI "Environment" >> page, I can see this value has been set correctly. But in the "Executors" >> page, I saw there's only 1 executor and its memory is 265.4MB. Very strange >> value. why not 256MB, or just as what I set? >> >> What am I missing here? >> >> >> Thanks, >> David >> >> >