Mikkel Krøigård <[EMAIL PROTECTED]> writes:

> Well, there are many more or less interesting conclusions to draw
> from your benchmark, Martin. Not surprisingly, matrix multiplication
> turns out to be expensive.

Hmm... I did see that there were a bunch of calls to __mul__ in
matrix.py, but I thought they came from the initialization of _hyper
in ActiveRuntime. The the initialization of _hyper does not use any
matrix multiplications, so this is wrong?!

When the mul method uses prss_get_triple, then _hyper should never be
used or even initialized and so there should be no matrix stuff going
on... I think I measured the wrong code somehow :-)

> One thing I really do find interesting about the table is the amount
> of time spent in inc_pc_wrapper. Perhaps it is possible to improve
> this somehow?

It only used 0.5 seconds of its own time -- the 21 seconds are the
total time spend in the child-calls made by inc_pc_wrapper. Since it
wraps all important functions its clear that the cumulative time will
be big:

       ncalls  tottime  percall  cumtime  percall
   48003/6003    0.518    0.000   21.195    0.004

But any optimization would be good -- if we can same a tiny bit for
each of the 48000 calls it might sum up :-)

-- 
Martin Geisler
_______________________________________________
viff-devel mailing list (http://viff.dk/)
viff-devel@viff.dk
http://lists.viff.dk/listinfo.cgi/viff-devel-viff.dk

Reply via email to