Thanks for all the feedback on this guys. > This would generally indicate that you've failed to properly tune > DB_CONFIG.
Completely possible! :) I worked through the stuff in the openldap faq-a-matic in the past, but it was complicated, and we've added alot of users since that time. Is there a better tuning guide out there? One that relates more to the actual data structures being used? > Or he's got a ~35GB database... the backend database files 'weigh' ~4.4 GB alone after running 'db_archive -d' to clean up the old 'log.' files. > We went to some effort to make sure that it's safe to run slapcat > while slapd > is running, to allow hot backups to be performed. Ignoring this > feature is > pretty counterproductive. BerkeleyDB itself also provides > documentation for > how to perform a hot backup of the raw DB files. Both of these options > exist > and are already documented; anything else you do at your own risk. I'd much prefer to use the slapcat method, but as I mentioned the import time has grown rather substantially in the past half year as we've added so much data - to the point that it takes 4 hours to do an import! We're talking reasonably fast hardware too: Its a poweredge 1850, dual 2.8 GHz Xeons with 4GB of memory and a RAID 1 array dedicated to the backend database. The schema we use is highly customized for our application. And Im working on a project to 'trim the fat', as there is definitely improvement to do there. But what can I do to learn more about this fine art of tuning DB_CONFIG? :)
