Hi All I have been trying to send my application related logs to socket so that we can write log stash and check the application logs.
here is my log4j.property file main.logger=RFA,SA log4j.appender.SA=org.apache.log4j.net.SocketAppender log4j.appender.SA.Port=4560 log4j.appender.SA.RemoteHost=hadoop07.housing.com log4j.appender.SA.ReconnectionDelay=10000 log4j.appender.SA.Application=NM-${user.dir} # Ignore messages below warning level from Jetty, because it's a bit verbose log4j.logger.org.spark-project.jetty=WARN log4j.logger.org.apache.hadoop=WARN I am launching my spark job using below common on YARN-cluster mode *spark-submit --name data-ingestion --master yarn-cluster --conf spark.custom.configuration.file=hdfs://10.1.6.186/configuration/binning-dev.conf <http://10.1.6.186/configuration/binning-dev.conf> --files /usr/hdp/current/spark-client/Runnable/conf/log4j.properties --conf "spark.executor.extraJavaOptions=-Dlog4j.configuration=log4j.properties" --conf "spark.driver.extraJavaOptions=-Dlog4j.configuration=log4j.properties" --class com.housing.spark.streaming.Binning /usr/hdp/current/spark-client/Runnable/dsl-data-ingestion-all.jar* *Can anybody please guide me why i am not getting the logs the socket?* *I followed many pages listing below without success* http://tech-stories.com/2015/02/12/setting-up-a-central-logging-infrastructure-for-hadoop-and-spark/#comment-208 http://stackoverflow.com/questions/22918720/custom-log4j-appender-in-hadoop-2 http://stackoverflow.com/questions/9081625/override-log4j-properties-in-hadoop