1.in order to change log4j.properties at the name node, u can change /home/hadoop/log4j.properties.
2.in order to change log4j.properties for the container logs, u need to change it at the yarn containers jar, since they hard-coded loading the file directly from project resources. 2.1 ssh to the slave (on EMR u can also simply add this as bootstrap action, so u dont need to ssh to each of the nodes). 2.2 override the container-log4j.properties at the jar resources: jar uf /home/hadoop/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.2.0.jar *container-log4j.properties* On 8 September 2015 at 05:47, Yana Kadiyska <yana.kadiy...@gmail.com> wrote: > Hopefully someone will give you a more direct answer but whenever I'm > having issues with log4j I always try -Dlog4j.debug=true.This will tell > you which log4j settings are getting picked up from where. I've spent > countless hours due to typos in the file, for example. > > On Mon, Sep 7, 2015 at 11:47 AM, Jeetendra Gangele <gangele...@gmail.com> > wrote: > >> I also tried placing my costomized log4j.properties file under >> src/main/resources still no luck. >> >> won't above step modify the default YARN and spark log4j.properties ? >> >> anyhow its still taking log4j.properties from YARn. >> >> >> >> On 7 September 2015 at 19:25, Jeetendra Gangele <gangele...@gmail.com> >> wrote: >> >>> anybody here to help? >>> >>> >>> >>> On 7 September 2015 at 17:53, Jeetendra Gangele <gangele...@gmail.com> >>> wrote: >>> >>>> Hi All I have been trying to send my application related logs to socket >>>> so that we can write log stash and check the application logs. >>>> >>>> here is my log4j.property file >>>> >>>> main.logger=RFA,SA >>>> >>>> log4j.appender.SA=org.apache.log4j.net.SocketAppender >>>> log4j.appender.SA.Port=4560 >>>> log4j.appender.SA.RemoteHost=hadoop07.housing.com >>>> log4j.appender.SA.ReconnectionDelay=10000 >>>> log4j.appender.SA.Application=NM-${user.dir} >>>> # Ignore messages below warning level from Jetty, because it's a bit >>>> verbose >>>> log4j.logger.org.spark-project.jetty=WARN >>>> log4j.logger.org.apache.hadoop=WARN >>>> >>>> >>>> I am launching my spark job using below common on YARN-cluster mode >>>> >>>> *spark-submit --name data-ingestion --master yarn-cluster --conf >>>> spark.custom.configuration.file=hdfs://10.1.6.186/configuration/binning-dev.conf >>>> <http://10.1.6.186/configuration/binning-dev.conf> --files >>>> /usr/hdp/current/spark-client/Runnable/conf/log4j.properties --conf >>>> "spark.executor.extraJavaOptions=-Dlog4j.configuration=log4j.properties" >>>> --conf >>>> "spark.driver.extraJavaOptions=-Dlog4j.configuration=log4j.properties" >>>> --class com.housing.spark.streaming.Binning >>>> /usr/hdp/current/spark-client/Runnable/dsl-data-ingestion-all.jar* >>>> >>>> >>>> *Can anybody please guide me why i am not getting the logs the socket?* >>>> >>>> >>>> *I followed many pages listing below without success* >>>> >>>> http://tech-stories.com/2015/02/12/setting-up-a-central-logging-infrastructure-for-hadoop-and-spark/#comment-208 >>>> >>>> http://stackoverflow.com/questions/22918720/custom-log4j-appender-in-hadoop-2 >>>> >>>> http://stackoverflow.com/questions/9081625/override-log4j-properties-in-hadoop >>>> >>>> >>> >> >> >> >> >