Hello,
We realy need some help here. It has come to our attention that our
bayes database is 2.4 GB big. It is really slowing down our servers and
they have a big cpu load.
Now we have tried the trick with the sa-learn --force-expire , and it
deletes a lot of entrys, but the file is not getting any smaller.
79K Jun 12 09:26 bayes_journal
20M Jun 12 09:26 bayes_toks
2.5G Jun 12 09:26 bayes_seen*
Does anyone has some tricks to help us out ?
Greetings... Richard Smits
----
0.000 0 3 0 non-token data: bayes db version
0.000 0 14201082 0 non-token data: nspam
0.000 0 7760360 0 non-token data: nham
0.000 0 916962 0 non-token data: ntokens
0.000 0 1181559955 0 non-token data: oldest atime
0.000 0 1181633069 0 non-token data: newest atime
0.000 0 1181633115 0 non-token data: last journal
sync atime
0.000 0 1181604237 0 non-token data: last expiry atime
0.000 0 43200 0 non-token data: last expire
atime delta
0.000 0 360013 0 non-token data: last expire
reduction count
----------------------