Hi,

Spark UI or logs don't provide the situation of cluster. However, you can
use Ganglia to monitor the situation of cluster. In spark-ec2, there is an
option to install ganglia automatically.

If you use CDH, you can also use Cloudera manager.

Cheers
Gen


On Sat, Aug 8, 2015 at 6:06 AM, Xiao JIANG <jiangxia...@outlook.com> wrote:

> Hi all,
>
>
> I was running some Hive/spark job on hadoop cluster.  I want to see how
> spark helps improve not only the elapsed time but also the total CPU
> consumption.
>
>
> For Hive, I can get the 'Total MapReduce CPU Time Spent' from the log when
> the job finishes. But I didn't find any CPU stats for Spark jobs from
> either spark log or web UI. Is there any place I can find the total CPU
> consumption for my spark job? Thanks!
>
>
> Here is the version info: Spark version 1.3.0 Using Scala version 2.10.4,
> Java 1.7.0_67
>
>
> Thanks!
>
> Xiao
>

Reply via email to