Oh, you are right, thanks.

On Mon, Apr 20, 2015 at 6:31 PM, Lan Jiang <ljia...@gmail.com> wrote:

> Each application gets its own executor processes,  so there should be no
> problem running them in parallel.
>
> Lan
>
>
>
> On Apr 20, 2015, at 10:25 AM, Michael Ryabtsev <michael...@gmail.com>
> wrote:
>
> Hi Lan,
>
> Thanks for fast response. It could be a solution if it works. I have more
> than one log4 properties file, for different run modes like
> debug/production, for executor and for application core. I think I would
> like to keep them separate. Then, I suppose I should give all other
> properties files a special names and keep the executor configuration with
> the default name? Can I conclude that going this way I will not be able to
> run several applications on the same cluster in parallel?
>
> Regarding submit, I am not using it now, I submit from the code, but I
> think I should consider this option.
>
> Thanks.
>
> On Mon, Apr 20, 2015 at 5:59 PM, Lan Jiang <ljia...@gmail.com> wrote:
>
>> Rename your log4j_special.properties file as log4j.properties and place
>> it under the root of your jar file, you should be fine.
>>
>> If you are using Maven to build your jar, please the log4j.properties in
>> the src/main/resources folder.
>>
>> However, please note that if you have other dependency jar file in the
>> classpath that contains another log4j.properties file this way, it might
>> not work since the first log4j.properties file that is loaded will be used.
>>
>> You can also do spark-submit —file log4j_special.properties … ,which
>> should transfer your log4j property file to the worker nodes automatically
>> without you copying them manually.
>>
>> Lan
>>
>>
>> > On Apr 20, 2015, at 9:26 AM, Michael Ryabtsev <michael...@gmail.com>
>> wrote:
>> >
>> > Hi all,
>> >
>> > I need to configure spark executor log4j.properties on a standalone
>> cluster.
>> > It looks like placing the relevant properties file in the spark
>> > configuration folder and  setting the spark.executor.extraJavaOptions
>> from
>> > my application code:
>> > sparkConf.set("spark.executor.extraJavaOptions",
>> > "-Dlog4j.configuration=log4j_special.properties");
>> > does the work, and the executor logs are written in the required place
>> and
>> > level. As far as I understand, it works, because the spark configuration
>> > folder is on the class path, and passing parameter without path works
>> here.
>> > However, I would like to avoid deploying these properties to each worker
>> > spark configuration folder.
>> > I wonder, if I put the properties in my application jar, is there any
>> way of
>> > telling executor to load them?
>> >
>> > Thanks,
>> >
>> >
>> >
>> > --
>> > View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/Configuring-logging-properties-for-executor-tp22572.html
>> > Sent from the Apache Spark User List mailing list archive at Nabble.com
>> .
>> >
>> > ---------------------------------------------------------------------
>> > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> > For additional commands, e-mail: user-h...@spark.apache.org
>> >
>>
>>
>
>

Reply via email to