OK Geoff,
So I'm now working for a certain prominent Linux hardware company and
they like the Geocrawler archive. They are going to want to run searches
against the *entire* archive, quickly, tens of thousands of times per
day. Right now, I don't believe there's any way that this could be done
because of the scale of the archive. Right now, Geocrawler has over 450
separate ht://dig databases, which isn't as cool of a search as we want.
I understand you have some multi-search scripts or something, but can
you conceive of a way to spread these searches across a cluster of
machines, aka, a Beowulf cluster or something? There is some talk of
giving me a Beowulf cluster to run ht://dig on.
I'm sure some of the hackers here could figure out something, but I
wondered if you have some ideas for a starting point, or tell me if I am
totally nutty for using ht://dig on this.
Thanks as always!
Tim
------------------------------------
To unsubscribe from the htdig mailing list, send a message to
[EMAIL PROTECTED] containing the single word unsubscribe in
the SUBJECT of the message.