[ 
https://issues.apache.org/jira/browse/SPARK-21352?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shubham Gupta updated SPARK-21352:
----------------------------------
    Issue Type: Improvement  (was: Bug)

> Memory Usage in Spark Streaming
> -------------------------------
>
>                 Key: SPARK-21352
>                 URL: https://issues.apache.org/jira/browse/SPARK-21352
>             Project: Spark
>          Issue Type: Improvement
>          Components: DStreams, Spark Submit, YARN
>    Affects Versions: 2.1.1
>            Reporter: Shubham Gupta
>              Labels: newbie
>
> I am trying to figure out the memory used by executors for a Spark Streaming 
> job. For data I am using the rest endpoint for Spark AllExecutors and just 
> summing up the metrics totalDuration * spark.executor.memory for every 
> executor and then emitting the final sum as the memory usage.
> But this is coming out to be very small for application which ran whole day , 
> is something wrong with the logic.Also I am using dynamic allocation and 
> executorIdleTimeout is 5 seconds.
> Also I am also assuming that if some executor was removed for due to idle 
> timeout and then was allocated to some other task then its totalDuration will 
> be increased by the amount of time took by the executor to execute this new 
> task.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to