Hello Peter Gutmann.
I'm working on a contribution to the SHA-3 process, and I've been
using exactly the sort of abstraction that you describe -- counting
one computation of a hash compression function as a unit of work
which could be computed concurrently by some sort of parallel computer.
I vaguely think that once I get this level of analysis done, I should
add some terms to show how the velocity of data into the computer and
from core to core is not infinite.
I certainly think that I should code up some actual implementations
and benchmark them. However, I don't have a machine available with
lots of cores -- I'm considering requesting of Sun.com that they lend
me a T2. (Despite my earlier declaration to Sun that I had lost
interest in their stupid architecture since they wouldn't release the
source to the crypto module.)
Anyway, if you have a better way to think about parallelism of hash
functions, I'm all ears.
Thanks,
Zooko
---
http://allmydata.org -- Tahoe, the Least-Authority Filesystem
http://allmydata.com -- back up all your files for $5/month
---------------------------------------------------------------------
The Cryptography Mailing List
Unsubscribe by sending "unsubscribe cryptography" to [EMAIL PROTECTED]