Hi there,

I am currently using Spark cluster to run jobs but I really need to collect the 
history of actually memory usage(that’s execution memory + storage memory) of 
the job in the whole cluster. I know we can get the storage memory usage 
through either Spark UI Executor page or SparkContext.getMemoryExecutorStatus() 
API, but I could not get the real time execution memory usage. 
Is there anyway I can collect total memory usage? Thank you so much!

Best,
Jialin Liu
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to