I asked something similar earlier, and was pointed to MogileFS by danga (danga.com)
It looks messy and scary, but seems to work. It seems to use a daemon coupled with replicated mysql to sync content across servers.
On Feb 20, 2005, at 8:49 PM, Mike OK wrote:
Hi
I am looking for some opinions on how to best handle this situation. I am willing to do all of the research about each topic but would like some points of view on how others would handle this. Thanks Mike
- I have developed a search engine that caches the results
- Currently I save the results in a text file on the local machine that runs apache
- This system works well for the test system but I need to have a central storage system for these requests
- I plan on having over 250 and up to 1000's different servers saving and retrieving information from this central system
- I currently have a linux box with the reiser filesystem installed ready for this central system because it handles lots and lots of small files very well
- My thought is that the file system would be quicker for lookups than a MySQL database (I use it in other areas of the system)
- I have looked at cpan and have found a module file::remote that I could use but I feel there might be a more efficient solution
- Since this is happening on a local network, I could assign a port for each server to save connection overhead
Any thoughts on this subject would be greatly appreciated. Thanks again.