Hi All,

I have about a thousand RRD files, one for each IP addresss inside our
network.  I'd like to be able to sort these IPs by "total" traffic over
a period of time (to see which IPs are using the most traffic). 
Currently I have a cgi script that calls RRDs::fetch and sums all the
data points by hand, but with 1000 RRD files, this is way too slow.  How
can I do it faster?  I tried playing with the --resolution switch, but I
could not ever get it to return data in anything but the finest
resolution.

Here is roughly what I do, for each RRD file, to find the totals:

    my ($start,$step,$names,$data) =
        RRDs::fetch("$rrdfile",
                    "AVERAGE",
                    "-s $start",
                    "-e $end");
    
    foreach my $line (@$data) {
        $in += (@$line)[0] || 0;
        $out += (@$line)[1] || 0;
        $junk += (@$line)[2] || 0;
    }

I imagine this is a common thing that people do: compute totals over a
range for an RRD file.  Anyone have some tips on how I can speed things
up?

Thanks!

-Aaron


--
Unsubscribe mailto:[EMAIL PROTECTED]
Help        mailto:[EMAIL PROTECTED]
Archive     http://www.ee.ethz.ch/~slist/rrd-users
WebAdmin    http://www.ee.ethz.ch/~slist/lsg2.cgi

Reply via email to