Hello,

We realy need some help here. It has come to our attention that our bayes database is 2.4 GB big. It is really slowing down our servers and they have a big cpu load.

Now we have tried the trick with the sa-learn --force-expire , and it deletes a lot of entrys, but the file is not getting any smaller.

79K  Jun 12 09:26 bayes_journal
20M  Jun 12 09:26 bayes_toks
2.5G Jun 12 09:26 bayes_seen*

Does anyone has some tricks to help us out ?

Greetings... Richard Smits

----
0.000          0          3          0  non-token data: bayes db version
0.000          0   14201082          0  non-token data: nspam
0.000          0    7760360          0  non-token data: nham
0.000          0     916962          0  non-token data: ntokens
0.000          0 1181559955          0  non-token data: oldest atime
0.000          0 1181633069          0  non-token data: newest atime
0.000 0 1181633115 0 non-token data: last journal sync atime
0.000          0 1181604237          0  non-token data: last expiry atime
0.000 0 43200 0 non-token data: last expire atime delta 0.000 0 360013 0 non-token data: last expire reduction count

----------------------

Reply via email to