We are trying to collect Spark logs using logstash for parsing app logs and collecting useful info.
We can read the Nodemanager logs but unable to read Spark application logs using Logstash . Current Setup for Spark logs and Logstash 1- Spark runs on Yarn . 2- Using log4j socketAppenders to write logs to tcp port . 3- Below lines added in log4j.properties of Yarn and Spark main.logger=RFA,SA log4j.appender.SA=org.apache.log4j.net.SocketAppender log4j.appender.SA.Port=4560 log4j.appender.SA.RemoteHost=${hostname} log4j.appender.SA.ReconnectionDelay=10000 log4j.appender.SA.Application=NM-${user.dir} 4-Logstash input input { log4j { mode => "server" host => "0.0.0.0" port => 4560 type => "log4j" } } Any help on reading Spark logs via Logstash will be appreciated . Also, is there a better way to collect Spark logs via Logstash ? Thanks, Ashish