Upcoming KAM.cf Ruleset 20th Anniversary

2024-04-05 Thread Giovanni Bechis

Hi,
very soon we will celebrate KAM.cf Ruleset 20th Anniversary,
are there any stories about how you use the ruleset, any products that include 
the rules you are aware of, or other info about how it has helped with spam and 
email security ?
Glad to receive any info or story about KAM.cf SpamAssassin ruleset.

 Giovanni


OpenPGP_signature.asc
Description: OpenPGP digital signature


Re: OFF-TOPIC ANNOUNCE: KAM Ruleset Turning PCCC Wild RBL Back On

2023-03-23 Thread Pedro David Marco via users
 With all respects,
i agree with Bill... but suppose just Bill is wrong...  Kam rules are free and 
show really huge quality, what is wrong about gently ask for cooperation if 
used in a commercial way?
KAM++
Pedro.

On Tuesday, March 21, 2023 at 06:18:38 PM GMT+1, Bill Cole 
 wrote:  
 
 On 2023-03-21 at 12:52:16 UTC-0400 (Tue, 21 Mar 2023 17:52:16 +0100)
Benny Pedersen 
is rumored to have said:

> Kevin A. McGrail skrev den 2023-03-21 17:27:
>
>> https://mcgrail.com/template/donate
>
> you know the rules to post commericial postings to public free 
> maillists ?,

What rules exactly are you referring to? Please cite them precisely, in 
grammatically decipherable English. Note that if the 'rules' being cited 
are not on an ASF site, they do not apply here.

The McGrail Foundation is not a commercial entity. That's why that page 
talks of  donating rather than purchasing, and why it refers to a US tax 
code section. Noting the enhancement of a widely-used free service for 
SA users provided by a non-profit charitable foundation with in-kind 
support from a commercial entity (Linode, A.K.A. Amazon) is not a 
commercial posting.

If you want kolektiva.social, it is over there...




-- 
Bill Cole
b...@scconsult.com or billc...@apache.org
(AKA @grumpybozo and many *@billmail.scconsult.com addresses)
Not Currently Available For Hire
  

Re: OFF-TOPIC ANNOUNCE: KAM Ruleset Turning PCCC Wild RBL Back On

2023-03-21 Thread Bill Cole

On 2023-03-21 at 12:52:16 UTC-0400 (Tue, 21 Mar 2023 17:52:16 +0100)
Benny Pedersen 
is rumored to have said:


Kevin A. McGrail skrev den 2023-03-21 17:27:


https://mcgrail.com/template/donate


you know the rules to post commericial postings to public free 
maillists ?,


What rules exactly are you referring to? Please cite them precisely, in 
grammatically decipherable English. Note that if the 'rules' being cited 
are not on an ASF site, they do not apply here.


The McGrail Foundation is not a commercial entity. That's why that page 
talks of  donating rather than purchasing, and why it refers to a US tax 
code section. Noting the enhancement of a widely-used free service for 
SA users provided by a non-profit charitable foundation with in-kind 
support from a commercial entity (Linode, A.K.A. Amazon) is not a 
commercial posting.


If you want kolektiva.social, it is over there...




--
Bill Cole
b...@scconsult.com or billc...@apache.org
(AKA @grumpybozo and many *@billmail.scconsult.com addresses)
Not Currently Available For Hire


Re: OFF-TOPIC ANNOUNCE: KAM Ruleset Turning PCCC Wild RBL Back On

2023-03-21 Thread Benny Pedersen

Kevin A. McGrail skrev den 2023-03-21 17:27:


https://mcgrail.com/template/donate


you know the rules to post commericial postings to public free maillists 
?, rspamd did this abuse aswell, now thay have only non free irc 
support, and telegram


more talk about linode ? :)

mx ~ # dig -4 +short rs.dns-oarc.net txt
rst.x487.rs.dns-oarc.net.
rst.x461.x487.rs.dns-oarc.net.
rst.x466.x461.x487.rs.dns-oarc.net.
"172.104.150.56 DNS reply size limit is at least 487"
"172.104.150.56 sent EDNS buffer size 512"

great for dns

i have a ticket to this problem at linode, there answer is try add new 
ips to your linode vps, you can be lucky a new one works :/


will it be possible to see kam channel could provide corpus data to 
spamassassin, its very low at moment


OFF-TOPIC ANNOUNCE: KAM Ruleset Turning PCCC Wild RBL Back On

2023-03-21 Thread Kevin A. McGrail

Hello All,

I am pleased to announce that users of the KAM ruleset will once again 
have the free use of the PCCC Wild RBL.


The RBL was previously removed from use due to its popularity.

Thanks go to Linode.com for donating the servers and as always thanks to 
PCCC for the datafeed.


The KAM Ruleset will be celebrating the start of it's 20th year of 
publishing free rules and threat data in May 2023. If you are a 
commercial user, please consider a donation or long-term sponsorship: 
https://mcgrail.com/template/donate <https://mcgrail.com/template/donate>


More Info:
https://raptor.pccc.com/RBL <https://raptor.pccc.com/RBL>
https://mcgrail.com/template/projects#KAM1 
<https://mcgrail.com/template/projects#KAM1>


Regards,
KAM

--
Kevin A. McGrail
kmcgr...@apache.org

Member, Apache Software Foundation
Chair Emeritus Apache SpamAssassin Project
https://www.linkedin.com/in/kmcgrail  - 703.798.0171


Re: How to verify specific commits are in current ruleset?

2019-05-30 Thread Bill Cole

On 30 May 2019, at 10:57, Mike Ray wrote:


Hello all-

Been using spamassassin for awhile now, basically letting it run on 
auto-pilot

and it's been great so far.

However, after the recent __STYLE_GIBBERISH bug
(https://bz.apache.org/SpamAssassin/show_bug.cgi?id=7707), I need to 
have a

little more understanding of SA.


The LATEST problem is actually in a different bug that is classified as 
a "Security" bug, and so is not publicly visible.



My biggest issue at the moment is that I saw John's message last night 
that said

an update would be pushed out with yesterday's update
(https://www.mail-archive.com/users@spamassassin.apache.org/msg104352.html).
However, this morning, the only way I was able to verify that, was by 
looking at

his change
(https://svn.apache.org/viewvc/spamassassin/trunk/rulesrc/sandbox/jhardin/20_misc_testing.cf?r1=1857655&r2=1857654&pathrev=1857655)
and comparing that to the code currently running on my mail servers.

Is there any easier way to verify that a specific commit is in my 
currently

running rule set?


Yes. Compare the first line of 
/var/lib/spamassassin/3.004002/updates_spamassassin_org.cf to the 
revision number of the commit.


How to verify specific commits are in current ruleset?

2019-05-30 Thread Mike Ray
Hello all-

Been using spamassassin for awhile now, basically letting it run on auto-pilot 
and it's been great so far.

However, after the recent __STYLE_GIBBERISH bug 
(https://bz.apache.org/SpamAssassin/show_bug.cgi?id=7707), I need to have a
little more understanding of SA.

My biggest issue at the moment is that I saw John's message last night that 
said 
an update would be pushed out with yesterday's update
(https://www.mail-archive.com/users@spamassassin.apache.org/msg104352.html).
However, this morning, the only way I was able to verify that, was by looking at
his change
(https://svn.apache.org/viewvc/spamassassin/trunk/rulesrc/sandbox/jhardin/20_misc_testing.cf?r1=1857655&r2=1857654&pathrev=1857655)
and comparing that to the code currently running on my mail servers.

Is there any easier way to verify that a specific commit is in my currently 
running rule set?

Mike Ray


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-11 Thread Bill Cole

On 11 Feb 2018, at 9:54 (-0500), Benny Pedersen wrote:

first query would be valid for 300 secs, but that is imho still not 
free, problem is that keeping low ttls does not change how dns works, 
any auth dns servers will upate on soa serial anyway, the crime comes 
in when sa using remote dns servers that ignore soa serial updates


in that case ttls would keep spammers listed for 300 secs only


That's not how DNS TTLs work.

When a record's TTL elapses in the local name cache, it is dropped. The 
next query for that name and record type causes the resolver to make 
another query to the authoritative nameservers, which will return the 
same record whose TTL expired unless it has been removed from the zone. 
No standards-conforming DNS resolver returns NXDOMAIN based on the lack 
of a non-expired record in its cache and an unchanged SOA serial above 
the name. That would make no sense at all and require many more SOA 
queries than actually happen.



--
Bill Cole
b...@scconsult.com or billc...@apache.org
(AKA @grumpybozo and many *@billmail.scconsult.com addresses)
Currently Seeking Steady Work: https://linkedin.com/in/billcole


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-11 Thread Benny Pedersen

Dave Warren skrev den 2018-02-06 20:39:


How low are the TTLs? I'm seeing 300 seconds on 127.0.0.2 which is
more than sufficient time for a single message to finish processing,
such that multiple queries from one message would absolutely be cached
(or more likely, the first would still be pending and the second would
get the same answer as the first).


first query would be valid for 300 secs, but that is imho still not 
free, problem is that keeping low ttls does not change how dns works, 
any auth dns servers will upate on soa serial anyway, the crime comes in 
when sa using remote dns servers that ignore soa serial updates


in that case ttls would keep spammers listed for 300 secs only

and thats why i say 300 secs helps spammers


;; ANSWER SECTION:
2.0.0.127.bb.barracudacentral.org. 300 IN A 127.0.0.2

Maybe the TTLs are different for other records?


300 is imho to low to anything thats called free

i would like to accept free if it was 3600


I am also noticing very intermittent response times, sometimes taking
over a second to get a response, other times taking under 50ms.


rndc querylog is my friend

i just like to start a debate on why 300 is accepted as free, it does 
matter for non datafeeds users, but for datafeeds it does not matter at 
all


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread RW
On Tue, 6 Feb 2018 11:38:42 -0500
Alex wrote:


> On Tue, Feb 6, 2018 at 8:44 AM, David Jones  wrote:
ustomer's compromised accounts.
> >
> > Leave out the RCVD_IN_BRBL rule above and change the
> > RCVD_IN_BRBL_LASTEXT score to 1.4 to keep things the same.  
> 

> If you think the RCVD_IN_BRBL rule is a good one, I'd like to use it,
> and while I've implemented much of your approach, I can't implement
> all of it. 

I've been doing deep XBL checks for years (they come free with the zen
look-ups). Initially I found that they did have a low FP rate, but
that's changed as more of my mail comes through mobile (cellular)
networks that use NAT to support millions of users on thousands of IPv4
addresses. I'm seeing about 10% of ham submitted from these networks
hitting my deep XBL rule. 


> Can I also ask again about reasonable RCVD_IN_LASHBACK and
> RCVD_IN_LASHBACK_LASTEXT scores?

If you're going to create deep versions of more than one list it doesn't
make sense to score them individually. The dominant cause of FPs on such
rules is dynamic IP address reuse, being in more than one list
doesn't say anything about that. Allowing a large score to build-up
from multiple deep rules is reckless. 




Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread Alex
Hi,

>>> whitelist_auth *@bounce.mail.salesforce.com
>>> whitelist_auth *@sendgrid.net
>>> whitelist_auth *@*.mcdlv.net
>>
>>
>> I've seen enough spam sent through all three - both by way of whole
>> apparently spammer-owned accounts and cracked-but-otherwise-legitimate
>> accounts - that I would never blanket-whitelist whole bulk email providers.
>>
>> Legitimate mail sent through them generally gets through anyway IME.
>
> An alternative is to use "def_whitelist_auth" instead of "whitelist_auth"
> That gives a -7.5 point bump to usually good sources which may occasionally
> get abused.
>
> That way if one of their accounts gets p0wned your anti-phish rules have a
> chance of pulling the junk into the spam-tagged range.

Phishing is a significant concern for us. Whether or not the phish
came through the customer of one of these senders or through the
senders themselves, whitelisting these senders only facilitates more
phishes. There was a period when it was about one being reported by a
particularly large customer per day. Telling my customers that we've
contacted the provider and reported the spam just isn't good enough.

We also received a phish through freshdesk.com which is on the
def_whitelist. It's also on the DKIMWL_WL, subtracting another -3.5
points. It was also listed in RCVD_IN_HOSTKARMA_W, but also in
LASHBACK_LASTEXT and invaluement, but not enough to compensate for the
negative points.

I suspect it was a compromised freshdesk trial account that was
managed by amazonaws and sendgrid before passing through
smtp.freshdesk.com, both of which weren't whitelisted at the time.


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread David B Funk

On Tue, 6 Feb 2018, Kris Deugau wrote:


Alex wrote:

These phishes we've received were all from otherwise trusted sources
like salesforce, amazonses and sendgrid. These are examples that I
believe were previously whitelisted because of having received a phish
through these systems but have no been disabled.

whitelist_auth *@bounce.mail.salesforce.com
whitelist_auth *@sendgrid.net
whitelist_auth *@*.mcdlv.net


I've seen enough spam sent through all three - both by way of whole 
apparently spammer-owned accounts and cracked-but-otherwise-legitimate 
accounts - that I would never blanket-whitelist whole bulk email providers.


Legitimate mail sent through them generally gets through anyway IME.


An alternative is to use "def_whitelist_auth" instead of "whitelist_auth"
That gives a -7.5 point bump to usually good sources which may occasionally get 
abused.


That way if one of their accounts gets p0wned your anti-phish rules have a 
chance of pulling the junk into the spam-tagged range.



--
Dave Funk  University of Iowa
College of Engineering
319/335-5751   FAX: 319/384-0549   1256 Seamans Center
Sys_admin/Postmaster/cell_adminIowa City, IA 52242-1527
#include 
Better is not better, 'standard' is better. B{


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread David Jones

On 02/06/2018 01:28 PM, Alex wrote:

Hi,


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
'bb.barracudacentral.org')
tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl',
'127.0.0.2')
metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 &&
!RCVD_IN_BRBL_LASTEXT
describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
bb.barracudacentral.org
score   RCVD_IN_BRBL1.2
tflags  RCVD_IN_BRBLnet

header  RCVD_IN_BRBL_LASTEXT
eval:check_rbl('brbl-lastexternal',
'bb.barracudacentral.org')
describeRCVD_IN_BRBL_LASTEXTLast external is listed in
Barracuda
RBL bb.barracudacentral.org
score   RCVD_IN_BRBL_LASTEXT2.2
tflags  RCVD_IN_BRBL_LASTEXTnet

endif




You don't think these scores are a bit high for a normal installation?
The current default score for RCVD_IN_BRBL_LASTEXT is 1.4 and
RCVD_IN_BRBL doesn't otherwise exist.

Also, does someone have a recommended score for the lashback RBL? I've
had it in testing for quite some time and would like to put it into
production with reasonable values...



Ok, ok.  Uncle. (Waving white flag.)  I have been sufficiently flogged so
I
have learned my lesson.  :)  This works in my highly customized SA
platform
where I have to do outbound mail filtering so deep Received header
checking
is valuable to block spam from my customer's compromised accounts.

Leave out the RCVD_IN_BRBL rule above and change the RCVD_IN_BRBL_LASTEXT
score to 1.4 to keep things the same.



I didn't mean to imply I don't agree or otherwise support your
approach. It was just unclear that this was in conjunction with that
approach of using higher spam rule scores to offset lower ham rule
scores or if it was recommended for everyone.

If you think the RCVD_IN_BRBL rule is a good one, I'd like to use it,
and while I've implemented much of your approach, I can't implement
all of it. My users raise holy hell when they receive even one phish
from an otherwise trustworthy source that's been whitelisted. It hits
on a ton of email at both ends of the spectrum - most are either very
low scoring or are already spam.



First let me say that my method for many whitelist_auth entries does not
allow for any phishing emails so if I find any of those, they do not get a
whitelist_auth entry.  With a properly tuned MTA in front of SA, the only
phishing or blatant spam should be coming from compromised accounts or
zero-hour spam which are going to be difficult to block anyway.


Yes, I should have mentioned that as well. We're not whitelisting any
end-user level accounts.

These phishes we've received were all from otherwise trusted sources
like salesforce, amazonses and sendgrid. These are examples that I
believe were previously whitelisted because of having received a phish
through these systems but have no been disabled.

whitelist_auth *@bounce.mail.salesforce.com
whitelist_auth *@sendgrid.net
whitelist_auth *@*.mcdlv.net



I have never received phishing emails from trusted senders like those. 
If I did, then they would not be considered a trusted sender.  We have 
received unsolicited junk emails from customers of theirs after someone 
sold our email addresses to some so-called "double opt-in" email lists 
that we never opted into.  I have been reporting this abuse and those 
rogue customers are immediately locked/blocked/disabled/etc. which helps 
all of us.  I have never received a malicious/infected/spoofing email 
from those, just harmless junk email.



My method should only be whitelist_auth'ing system-generated/bulk emails
from reputable senders that handle abuse reports quickly and shouldn't match
compromised accounts and "freemail" domains.  It will also match commonly
spoofed domains like fedex.com, ups.com, and banks to help block those
phishing emails.


This also requires supporting rules that add significant points based
on their name most simply, or other more complicated rules to catch
more sophisticated phish attempts. Spammers hitting our systems are
far more sophisticated than just using "UPS" in the subject or
pretending to be from ups.com.



Sure.  What I have found effective is to train your Bayes with all 
spoofing emails so high BAYES rule hits will add points.  The authentic 
emails will score very low from the whitelist_auth entry.  Add in some 
custom regex rules in the subject and body for variations of 
misspellings of the spoofed brand or new zero-hour findings and this is 
the best protection I have been able to come up with for new 
spoofing/spam campaigns.




Can I also ask again about reasonable RCVD_IN_LASHBACK and
RCVD_IN_LASHBACK_LASTEXT scores?



It really depends on how much customization you have done to SA and how much
your mail flow can handle bumping up scores.  If you do some log analysis
and find that RCVD_IN_LASHBACK and RCVD_IN_LASHBACK_LASTEXT are pretty
accurate for your mail flow, then you c

Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread Kris Deugau

Alex wrote:

These phishes we've received were all from otherwise trusted sources
like salesforce, amazonses and sendgrid. These are examples that I
believe were previously whitelisted because of having received a phish
through these systems but have no been disabled.

whitelist_auth *@bounce.mail.salesforce.com
whitelist_auth *@sendgrid.net
whitelist_auth *@*.mcdlv.net


I've seen enough spam sent through all three - both by way of whole 
apparently spammer-owned accounts and cracked-but-otherwise-legitimate 
accounts - that I would never blanket-whitelist whole bulk email providers.


Legitimate mail sent through them generally gets through anyway IME.

-kgd


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread Dave Warren

On 2018-02-05 09:12, Benny Pedersen wrote:

Kevin A. McGrail skrev den 2018-02-05 16:53:


I don't think that will apply will it because it will be looking up
something like 1.2.3.4.bb.barracuda.blah which isn't cached.


the first qurry can make a qurry with very low ttl, so it would not be 
cached, that means number 2 query still mkae dns query to that zone :(


How low are the TTLs? I'm seeing 300 seconds on 127.0.0.2 which is more 
than sufficient time for a single message to finish processing, such 
that multiple queries from one message would absolutely be cached (or 
more likely, the first would still be pending and the second would get 
the same answer as the first).


;; ANSWER SECTION:
2.0.0.127.bb.barracudacentral.org. 300 IN A 127.0.0.2

Maybe the TTLs are different for other records?

I am also noticing very intermittent response times, sometimes taking 
over a second to get a response, other times taking under 50ms.




Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread Alex
Hi,

> ifplugin Mail::SpamAssassin::Plugin::DNSEval
>
> header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
> 'bb.barracudacentral.org')
> tflags  __RCVD_IN_BRBL  net
>
> header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl',
> '127.0.0.2')
> metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 &&
> !RCVD_IN_BRBL_LASTEXT
> describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
> bb.barracudacentral.org
> score   RCVD_IN_BRBL1.2
> tflags  RCVD_IN_BRBLnet
>
> header  RCVD_IN_BRBL_LASTEXT
> eval:check_rbl('brbl-lastexternal',
> 'bb.barracudacentral.org')
> describeRCVD_IN_BRBL_LASTEXTLast external is listed in
> Barracuda
> RBL bb.barracudacentral.org
> score   RCVD_IN_BRBL_LASTEXT2.2
> tflags  RCVD_IN_BRBL_LASTEXTnet
>
> endif



 You don't think these scores are a bit high for a normal installation?
 The current default score for RCVD_IN_BRBL_LASTEXT is 1.4 and
 RCVD_IN_BRBL doesn't otherwise exist.

 Also, does someone have a recommended score for the lashback RBL? I've
 had it in testing for quite some time and would like to put it into
 production with reasonable values...

>>>
>>> Ok, ok.  Uncle. (Waving white flag.)  I have been sufficiently flogged so
>>> I
>>> have learned my lesson.  :)  This works in my highly customized SA
>>> platform
>>> where I have to do outbound mail filtering so deep Received header
>>> checking
>>> is valuable to block spam from my customer's compromised accounts.
>>>
>>> Leave out the RCVD_IN_BRBL rule above and change the RCVD_IN_BRBL_LASTEXT
>>> score to 1.4 to keep things the same.
>>
>>
>> I didn't mean to imply I don't agree or otherwise support your
>> approach. It was just unclear that this was in conjunction with that
>> approach of using higher spam rule scores to offset lower ham rule
>> scores or if it was recommended for everyone.
>>
>> If you think the RCVD_IN_BRBL rule is a good one, I'd like to use it,
>> and while I've implemented much of your approach, I can't implement
>> all of it. My users raise holy hell when they receive even one phish
>> from an otherwise trustworthy source that's been whitelisted. It hits
>> on a ton of email at both ends of the spectrum - most are either very
>> low scoring or are already spam.
>>
>
> First let me say that my method for many whitelist_auth entries does not
> allow for any phishing emails so if I find any of those, they do not get a
> whitelist_auth entry.  With a properly tuned MTA in front of SA, the only
> phishing or blatant spam should be coming from compromised accounts or
> zero-hour spam which are going to be difficult to block anyway.

Yes, I should have mentioned that as well. We're not whitelisting any
end-user level accounts.

These phishes we've received were all from otherwise trusted sources
like salesforce, amazonses and sendgrid. These are examples that I
believe were previously whitelisted because of having received a phish
through these systems but have no been disabled.

whitelist_auth *@bounce.mail.salesforce.com
whitelist_auth *@sendgrid.net
whitelist_auth *@*.mcdlv.net

> My method should only be whitelist_auth'ing system-generated/bulk emails
> from reputable senders that handle abuse reports quickly and shouldn't match
> compromised accounts and "freemail" domains.  It will also match commonly
> spoofed domains like fedex.com, ups.com, and banks to help block those
> phishing emails.

This also requires supporting rules that add significant points based
on their name most simply, or other more complicated rules to catch
more sophisticated phish attempts. Spammers hitting our systems are
far more sophisticated than just using "UPS" in the subject or
pretending to be from ups.com.

>> Can I also ask again about reasonable RCVD_IN_LASHBACK and
>> RCVD_IN_LASHBACK_LASTEXT scores?
>>
>
> It really depends on how much customization you have done to SA and how much
> your mail flow can handle bumping up scores.  If you do some log analysis
> and find that RCVD_IN_LASHBACK and RCVD_IN_LASHBACK_LASTEXT are pretty
> accurate for your mail flow, then you can bump it up like I have to 2.2 and
> 4.2 respectively.
>
> DISCLAIMER:  I am not recommending this for everyone so no flaming.  Set
> these scores low and test for a few weeks or months to see how your mail
> logs line up with real spam then increase the scores as you see fit. Again,
> I do outbound mail filtering for my customers so the deep Received header
> inspection is helpful to determine compromised accounts and keep my mail
> servers off of RBLs.

Are there any numbers available from anyone's masschecks?

Thank you as always for your support.


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread David Jones

On 02/06/2018 10:38 AM, Alex wrote:

Hi,

On Tue, Feb 6, 2018 at 8:44 AM, David Jones  wrote:

On 02/05/2018 09:07 PM, Alex wrote:


Hi,


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
'bb.barracudacentral.org')
tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl',
'127.0.0.2')
metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 && !RCVD_IN_BRBL_LASTEXT
describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
bb.barracudacentral.org
score   RCVD_IN_BRBL1.2
tflags  RCVD_IN_BRBLnet

header  RCVD_IN_BRBL_LASTEXT eval:check_rbl('brbl-lastexternal',
'bb.barracudacentral.org')
describeRCVD_IN_BRBL_LASTEXTLast external is listed in
Barracuda
RBL bb.barracudacentral.org
score   RCVD_IN_BRBL_LASTEXT2.2
tflags  RCVD_IN_BRBL_LASTEXTnet

endif



You don't think these scores are a bit high for a normal installation?
The current default score for RCVD_IN_BRBL_LASTEXT is 1.4 and
RCVD_IN_BRBL doesn't otherwise exist.

Also, does someone have a recommended score for the lashback RBL? I've
had it in testing for quite some time and would like to put it into
production with reasonable values...



Ok, ok.  Uncle. (Waving white flag.)  I have been sufficiently flogged so I
have learned my lesson.  :)  This works in my highly customized SA platform
where I have to do outbound mail filtering so deep Received header checking
is valuable to block spam from my customer's compromised accounts.

Leave out the RCVD_IN_BRBL rule above and change the RCVD_IN_BRBL_LASTEXT
score to 1.4 to keep things the same.


I didn't mean to imply I don't agree or otherwise support your
approach. It was just unclear that this was in conjunction with that
approach of using higher spam rule scores to offset lower ham rule
scores or if it was recommended for everyone.

If you think the RCVD_IN_BRBL rule is a good one, I'd like to use it,
and while I've implemented much of your approach, I can't implement
all of it. My users raise holy hell when they receive even one phish
from an otherwise trustworthy source that's been whitelisted. It hits
on a ton of email at both ends of the spectrum - most are either very
low scoring or are already spam.



First let me say that my method for many whitelist_auth entries does not 
allow for any phishing emails so if I find any of those, they do not get 
a whitelist_auth entry.  With a properly tuned MTA in front of SA, the 
only phishing or blatant spam should be coming from compromised accounts 
or zero-hour spam which are going to be difficult to block anyway.


My method should only be whitelist_auth'ing system-generated/bulk emails 
from reputable senders that handle abuse reports quickly and shouldn't 
match compromised accounts and "freemail" domains.  It will also match 
commonly spoofed domains like fedex.com, ups.com, and banks to help 
block those phishing emails.



Can I also ask again about reasonable RCVD_IN_LASHBACK and
RCVD_IN_LASHBACK_LASTEXT scores?



It really depends on how much customization you have done to SA and how 
much your mail flow can handle bumping up scores.  If you do some log 
analysis and find that RCVD_IN_LASHBACK and RCVD_IN_LASHBACK_LASTEXT are 
pretty accurate for your mail flow, then you can bump it up like I have 
to 2.2 and 4.2 respectively.


DISCLAIMER:  I am not recommending this for everyone so no flaming.  Set 
these scores low and test for a few weeks or months to see how your mail 
logs line up with real spam then increase the scores as you see fit. 
Again, I do outbound mail filtering for my customers so the deep 
Received header inspection is helpful to determine compromised accounts 
and keep my mail servers off of RBLs.



ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_LASHBACK  eval:check_rbl('lashback', 
'ubl.unsubscore.com.')
describe	__RCVD_IN_LASHBACK	Received is listed in Lashback 
ubl.unsubscore.com

tflags  __RCVD_IN_LASHBACK  net

header  __RCVD_IN_LASHBACK_2eval:check_rbl_sub('lashback', 
'127.0.0.2')
metaRCVD_IN_LASHBACK__RCVD_IN_LASHBACK_2 && 
!RCVD_IN_LASHBACK_LASTEXT
describeRCVD_IN_LASHBACKReceived is listed in Lashback 
ubl.unsubscore.com
score   RCVD_IN_LASHBACK2.2
tflags  RCVD_IN_LASHBACKnet

header		RCVD_IN_LASHBACK_LASTEXT	eval:check_rbl('lashback-lastexternal', 
'ubl.unsubscore.com.')
describe 	RCVD_IN_LASHBACK_LASTEXT	Last external is listed in Lashback 
ubl.unsubscore.com

score   RCVD_IN_LASHBACK_LASTEXT4.2
tflags  RCVD_IN_LASHBACK_LASTEXTnet

endif


--
David Jones


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread Alex
Hi,

On Tue, Feb 6, 2018 at 8:44 AM, David Jones  wrote:
> On 02/05/2018 09:07 PM, Alex wrote:
>>
>> Hi,
>>
>>> ifplugin Mail::SpamAssassin::Plugin::DNSEval
>>>
>>> header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
>>> 'bb.barracudacentral.org')
>>> tflags  __RCVD_IN_BRBL  net
>>>
>>> header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl',
>>> '127.0.0.2')
>>> metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 && !RCVD_IN_BRBL_LASTEXT
>>> describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
>>> bb.barracudacentral.org
>>> score   RCVD_IN_BRBL1.2
>>> tflags  RCVD_IN_BRBLnet
>>>
>>> header  RCVD_IN_BRBL_LASTEXT eval:check_rbl('brbl-lastexternal',
>>> 'bb.barracudacentral.org')
>>> describeRCVD_IN_BRBL_LASTEXTLast external is listed in
>>> Barracuda
>>> RBL bb.barracudacentral.org
>>> score   RCVD_IN_BRBL_LASTEXT2.2
>>> tflags  RCVD_IN_BRBL_LASTEXTnet
>>>
>>> endif
>>
>>
>> You don't think these scores are a bit high for a normal installation?
>> The current default score for RCVD_IN_BRBL_LASTEXT is 1.4 and
>> RCVD_IN_BRBL doesn't otherwise exist.
>>
>> Also, does someone have a recommended score for the lashback RBL? I've
>> had it in testing for quite some time and would like to put it into
>> production with reasonable values...
>>
>
> Ok, ok.  Uncle. (Waving white flag.)  I have been sufficiently flogged so I
> have learned my lesson.  :)  This works in my highly customized SA platform
> where I have to do outbound mail filtering so deep Received header checking
> is valuable to block spam from my customer's compromised accounts.
>
> Leave out the RCVD_IN_BRBL rule above and change the RCVD_IN_BRBL_LASTEXT
> score to 1.4 to keep things the same.

I didn't mean to imply I don't agree or otherwise support your
approach. It was just unclear that this was in conjunction with that
approach of using higher spam rule scores to offset lower ham rule
scores or if it was recommended for everyone.

If you think the RCVD_IN_BRBL rule is a good one, I'd like to use it,
and while I've implemented much of your approach, I can't implement
all of it. My users raise holy hell when they receive even one phish
from an otherwise trustworthy source that's been whitelisted. It hits
on a ton of email at both ends of the spectrum - most are either very
low scoring or are already spam.

Can I also ask again about reasonable RCVD_IN_LASHBACK and
RCVD_IN_LASHBACK_LASTEXT scores?


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread David Jones

On 02/05/2018 09:07 PM, Alex wrote:

Hi,


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
'bb.barracudacentral.org')
tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl',
'127.0.0.2')
metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 && !RCVD_IN_BRBL_LASTEXT
describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
bb.barracudacentral.org
score   RCVD_IN_BRBL1.2
tflags  RCVD_IN_BRBLnet

header  RCVD_IN_BRBL_LASTEXT eval:check_rbl('brbl-lastexternal',
'bb.barracudacentral.org')
describeRCVD_IN_BRBL_LASTEXTLast external is listed in Barracuda
RBL bb.barracudacentral.org
score   RCVD_IN_BRBL_LASTEXT2.2
tflags  RCVD_IN_BRBL_LASTEXTnet

endif


You don't think these scores are a bit high for a normal installation?
The current default score for RCVD_IN_BRBL_LASTEXT is 1.4 and
RCVD_IN_BRBL doesn't otherwise exist.

Also, does someone have a recommended score for the lashback RBL? I've
had it in testing for quite some time and would like to put it into
production with reasonable values...



Ok, ok.  Uncle. (Waving white flag.)  I have been sufficiently flogged 
so I have learned my lesson.  :)  This works in my highly customized SA 
platform where I have to do outbound mail filtering so deep Received 
header checking is valuable to block spam from my customer's compromised 
accounts.


Leave out the RCVD_IN_BRBL rule above and change the 
RCVD_IN_BRBL_LASTEXT score to 1.4 to keep things the same.


--
David Jones


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-06 Thread Matus UHLAR - fantomas

David Jones skrev den 2018-02-05 15:09:


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
'bb.barracudacentral.org')
tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl', 
'127.0.0.2')
metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 && 
!RCVD_IN_BRBL_LASTEXT

describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
bb.barracudacentral.org
score   RCVD_IN_BRBL1.2
tflags  RCVD_IN_BRBLnet

header  RCVD_IN_BRBL_LASTEXT
eval:check_rbl('brbl-lastexternal', 'bb.barracudacentral.org')
describeRCVD_IN_BRBL_LASTEXTLast external is listed in
Barracuda RBL bb.barracudacentral.org
score   RCVD_IN_BRBL_LASTEXT2.2
tflags  RCVD_IN_BRBL_LASTEXTnet

endif


On 05.02.18 16:26, Benny Pedersen wrote:
this rule makes 2 dns querys, waste of cpu performance, i would 
suggest to drop the lastextnal,


this rule checks all untrusted IPs, while lastexternal uses only one (often
the most important - IP that sent mail to your systems).

also, because of caching, brbl-lastexternal result will be cached.
b.barracudacentral.org uses ttl of 300, which should be cached during the
mail processing.

if you want to spare DNS queries, leave only lastexternal.

so its only if ip is listed yes or 
no, 50% dns querys saved, and still same hits on listed ips, why do 
we need to help spammers ?


network checks including DNS lookups help much in spam processing, after
bayes they are second best mechanism to detect spam.

NOT using them is helping spammers.

--
Matus UHLAR - fantomas, uh...@fantomas.sk ; http://www.fantomas.sk/
Warning: I wish NOT to receive e-mail advertising to this address.
Varovanie: na tuto adresu chcem NEDOSTAVAT akukolvek reklamnu postu.
Atheism is a non-prophet organization. 


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread Alex
Hi,

> ifplugin Mail::SpamAssassin::Plugin::DNSEval
>
> header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
> 'bb.barracudacentral.org')
> tflags  __RCVD_IN_BRBL  net
>
> header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl',
> '127.0.0.2')
> metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 && !RCVD_IN_BRBL_LASTEXT
> describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
> bb.barracudacentral.org
> score   RCVD_IN_BRBL1.2
> tflags  RCVD_IN_BRBLnet
>
> header  RCVD_IN_BRBL_LASTEXT eval:check_rbl('brbl-lastexternal',
> 'bb.barracudacentral.org')
> describeRCVD_IN_BRBL_LASTEXTLast external is listed in Barracuda
> RBL bb.barracudacentral.org
> score   RCVD_IN_BRBL_LASTEXT2.2
> tflags  RCVD_IN_BRBL_LASTEXTnet
>
> endif

You don't think these scores are a bit high for a normal installation?
The current default score for RCVD_IN_BRBL_LASTEXT is 1.4 and
RCVD_IN_BRBL doesn't otherwise exist.

Also, does someone have a recommended score for the lashback RBL? I've
had it in testing for quite some time and would like to put it into
production with reasonable values...


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread RW
On Mon, 05 Feb 2018 17:12:08 +0100
Benny Pedersen wrote:

> Kevin A. McGrail skrev den 2018-02-05 16:53:
> 
> > I don't think that will apply will it because it will be looking up
> > something like 1.2.3.4.bb.barracuda.blah which isn't cached.  
> 
> the first qurry can make a qurry with very low ttl, so it would not
> be cached, that means number 2 query still mkae dns query to that
> zone :(

SA sends its DNS requests out early in rapid succession. The chances
are that the local DNS cache would see the second request as a
duplicate of a pending look-up.  In that case caching is not needed.


> > Anyway, we're debating a rule that's removed :)  

> lastexternal is still a mistake imho :=)


lastexternal is correct, that's what RCVD_IN_BRBL_LASTEXT does. Making
use of deep checks on lists containing dynamic addresses is risky, and
likely to vary a lot between different mail flows. 

David's rules are not appropriate as a general replacement for
RCVD_IN_BRBL_LASTEXT IMO.


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread Kevin A. McGrail

On 2/5/2018 11:32 AM, RW wrote:

Just to clarify, there is no legal or moral obligation to do this, the
'bb' subdomain was created specifically so SA users wouldn't need to
register. Anything you may read on the Barracuda site applies to the
'b' version. Barracuda has given no indication that anything has
changed.


I've been trying to reach Barracuda pretty doggedly about the issue for 
months about the issues with bb requiring registration hence the 
removal.   If you have a contact, get in touch with them and we can 
relight the rule.  I was hoping some discussion on list might get their 
attention.


Regards,
KAM



Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread RW
On Mon, 5 Feb 2018 08:09:55 -0600
David Jones wrote:

> Heads up!  This RBL has been removed from the core SA ruleset.  In 36
> to 48 hours sa-update will remove the RCVD_IN_BRBL_LASTEXT rule after
> it has gone through the masscheck and rule promotion process.
> 
> Details can be found here:
> 
> https://bz.apache.org/SpamAssassin/show_bug.cgi?id=7417
> 
> To add this rule back manually, you should register here:
> 
> http://barracudacentral.org/account/register

Just to clarify, there is no legal or moral obligation to do this, the
'bb' subdomain was created specifically so SA users wouldn't need to
register. Anything you may read on the Barracuda site applies to the
'b' version. Barracuda has given no indication that anything has
changed.

The issue is that some users have seen look-ups fail and attributed it
to possible rate limiting on non-registered IP addresses.

You should register if you use the 'b' list directly from an MTA. If
you only run the 'bb' version from SA and use fixed IP addresses it
*may* help with reliability if you register. If you run SA client side
on a dynamic address there's no point in registering.

> and add this to your local.cf:

I would suggest people just copy the existing rule as it is.  This list
contains dynamic IP addresses so shouldn't be used for deep look-ups.  
 
> NOTE: recent masscheck processing shows RCVD_IN_BRBL_LASTEXT as the
> most effective non-zero scoring rule to hit spam (43%) so it's
> probably worth adding back to your local.cf on Wednesday or Thursday.

If you don't rename the rule there's no need to wait. 

Note that once it's removed the scores for other rules will adjust to
compensate for its absence, and this *may* lead to more FP's if it's
left at its current score. 



Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread Benny Pedersen

Kevin A. McGrail skrev den 2018-02-05 16:53:


I don't think that will apply will it because it will be looking up
something like 1.2.3.4.bb.barracuda.blah which isn't cached.


the first qurry can make a qurry with very low ttl, so it would not be 
cached, that means number 2 query still mkae dns query to that zone :(



Anyway, we're debating a rule that's removed :)


lastexternal is still a mistake imho :=)

gosh i hate bind9 not have minimal ttl, good thing is that low ttl 
strikes back to badly configured dns zones :=)


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread David Jones

On 02/05/2018 09:44 AM, Reindl Harald wrote:


Am 05.02.2018 um 16:36 schrieb David Jones:

On 02/05/2018 09:26 AM, Benny Pedersen wrote:

David Jones skrev den 2018-02-05 15:09:


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
'bb.barracudacentral.org')
tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2    eval:check_rbl_sub('brbl', 
'127.0.0.2')
meta    RCVD_IN_BRBL    __RCVD_IN_BRBL_2 && 
!RCVD_IN_BRBL_LASTEXT

describe    RCVD_IN_BRBL    Received is listed in Barracuda RBL
bb.barracudacentral.org
score   RCVD_IN_BRBL    1.2
tflags  RCVD_IN_BRBL    net

header  RCVD_IN_BRBL_LASTEXT
eval:check_rbl('brbl-lastexternal', 'bb.barracudacentral.org')
describe    RCVD_IN_BRBL_LASTEXT    Last external is listed in
Barracuda RBL bb.barracudacentral.org
score   RCVD_IN_BRBL_LASTEXT    2.2
tflags  RCVD_IN_BRBL_LASTEXT    net

endif


this rule makes 2 dns querys, waste of cpu performance, i would 
suggest to drop the lastextnal, so its only if ip is listed yes or 
no, 50% dns querys saved, and still same hits on listed ips, why do 
we need to help spammers ?


If you are running a local DNS cache like this list and the SA 
documention recommends, does this really matter?  My MTA should have 
already queried this before SA does it so it should be in the local 
DNS cache and not require a full recursive lookup from the SA rule above


1.2 poitns just because the IP of the previous hop is listet is pure 
nonsense and it was even nosense as Barracuda Networks started with that 
bullshit called "deep header inspection"


Barracuda and many other RBL's list here a ton of innocent IP's which 
are nothing else than the endusers range which never should tocu an 
inbound MX itself


so what the hell is the point that you give me 1.2 points because i use 
as i should our MTA from my home-ip to send an ordianry mail?


It can be a sign of a compromised account.  I just saw a compromised 
account coming from Nigeria listed on BRBL through Office 365.  Now that 
O365 is finally adding the "x-originating-ip" header, we can detect 
botnets sending via infected home computers.


Legit emails should have other rules subtracting points so a 1.2 should 
not be a major factor in the score.  My mail platform is scoring real 
spam greater than 18 and usually in the 20's.  I could leave this rule 
out and be fine but most default SA instances would benefit from it.  If 
they want to lower the scores, then make them 0.2 and 1.2 then and use 
them in meta rules.


--
David Jones


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread Kevin A. McGrail

On 2/5/2018 10:36 AM, David Jones wrote:
If you are running a local DNS cache like this list and the SA 
documention recommends, does this really matter?  My MTA should have 
already queried this before SA does it so it should be in the local 
DNS cache and not require a full recursive lookup from the SA rule above. 


I don't think that will apply will it because it will be looking up 
something like 1.2.3.4.bb.barracuda.blah which isn't cached.


Anyway, we're debating a rule that's removed :)


Regards,
KAM



Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread Benny Pedersen

David Jones skrev den 2018-02-05 16:36:


If you are running a local DNS cache like this list and the SA
documention recommends, does this really matter?  My MTA should have
already queried this before SA does it so it should be in the local
DNS cache and not require a full recursive lookup from the SA rule
above.


yes it matters with low ttl, non datafeeds have low ttl, so it does not 
cache much on dns servers, sadly, this can be avoided in spamassassin 
without 2 dns querys


it have always just be low priotet since we all assume datafedds where 
is does not matter



This shouldn't be the case with a local DNS cache with the
/etc/resolv.conf pointed to 127.0.0.1 or SA dns_server pointed to
127.0.0.1.


already done here


can lastexternal be solved to only use one query ?
as is i think spamassassin should be changed so it can if not already


on top of that bb. is for spamassassin, while b. was for mta stage, so 
200% more querys to non datafeeds :/


it should have being one dns zone


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread David Jones

On 02/05/2018 09:26 AM, Benny Pedersen wrote:

David Jones skrev den 2018-02-05 15:09:


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
'bb.barracudacentral.org')
tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2    eval:check_rbl_sub('brbl', 
'127.0.0.2')

meta    RCVD_IN_BRBL    __RCVD_IN_BRBL_2 && !RCVD_IN_BRBL_LASTEXT
describe    RCVD_IN_BRBL    Received is listed in Barracuda RBL
bb.barracudacentral.org
score   RCVD_IN_BRBL    1.2
tflags  RCVD_IN_BRBL    net

header  RCVD_IN_BRBL_LASTEXT
eval:check_rbl('brbl-lastexternal', 'bb.barracudacentral.org')
describe    RCVD_IN_BRBL_LASTEXT    Last external is listed in
Barracuda RBL bb.barracudacentral.org
score   RCVD_IN_BRBL_LASTEXT    2.2
tflags  RCVD_IN_BRBL_LASTEXT    net

endif


this rule makes 2 dns querys, waste of cpu performance, i would suggest 
to drop the lastextnal, so its only if ip is listed yes or no, 50% dns 
querys saved, and still same hits on listed ips, why do we need to help 
spammers ?




If you are running a local DNS cache like this list and the SA 
documention recommends, does this really matter?  My MTA should have 
already queried this before SA does it so it should be in the local DNS 
cache and not require a full recursive lookup from the SA rule above.



anyway i dont use it, above rule is only optimized for datafeeds, it 
will drain non datafeed clients into hell




This shouldn't be the case with a local DNS cache with the 
/etc/resolv.conf pointed to 127.0.0.1 or SA dns_server pointed to 127.0.0.1.



can lastexternal be solved to only use one query ?

as is i think spamassassin should be changed so it can if not already


--
David Jones


Re: Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread Benny Pedersen

David Jones skrev den 2018-02-05 15:09:


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl',
'bb.barracudacentral.org')
tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl', 
'127.0.0.2')
metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 && 
!RCVD_IN_BRBL_LASTEXT

describeRCVD_IN_BRBLReceived is listed in Barracuda RBL
bb.barracudacentral.org
score   RCVD_IN_BRBL1.2
tflags  RCVD_IN_BRBLnet

header  RCVD_IN_BRBL_LASTEXT
eval:check_rbl('brbl-lastexternal', 'bb.barracudacentral.org')
describeRCVD_IN_BRBL_LASTEXTLast external is listed in
Barracuda RBL bb.barracudacentral.org
score   RCVD_IN_BRBL_LASTEXT2.2
tflags  RCVD_IN_BRBL_LASTEXTnet

endif


this rule makes 2 dns querys, waste of cpu performance, i would suggest 
to drop the lastextnal, so its only if ip is listed yes or no, 50% dns 
querys saved, and still same hits on listed ips, why do we need to help 
spammers ?


anyway i dont use it, above rule is only optimized for datafeeds, it 
will drain non datafeed clients into hell


can lastexternal be solved to only use one query ?

as is i think spamassassin should be changed so it can if not already


Barracuda Reputation Block List (BRBL) removal from the SA ruleset

2018-02-05 Thread David Jones
Heads up!  This RBL has been removed from the core SA ruleset.  In 36 to 
48 hours sa-update will remove the RCVD_IN_BRBL_LASTEXT rule after it 
has gone through the masscheck and rule promotion process.


Details can be found here:

https://bz.apache.org/SpamAssassin/show_bug.cgi?id=7417

To add this rule back manually, you should register here:

http://barracudacentral.org/account/register

and add this to your local.cf:


ifplugin Mail::SpamAssassin::Plugin::DNSEval

header  __RCVD_IN_BRBL  eval:check_rbl('brbl', 
'bb.barracudacentral.org')

tflags  __RCVD_IN_BRBL  net

header  __RCVD_IN_BRBL_2eval:check_rbl_sub('brbl', 
'127.0.0.2')

metaRCVD_IN_BRBL__RCVD_IN_BRBL_2 && !RCVD_IN_BRBL_LASTEXT
describeRCVD_IN_BRBLReceived is listed in Barracuda RBL 
bb.barracudacentral.org

score   RCVD_IN_BRBL1.2
tflags  RCVD_IN_BRBLnet

header  RCVD_IN_BRBL_LASTEXT 
eval:check_rbl('brbl-lastexternal', 'bb.barracudacentral.org')
describeRCVD_IN_BRBL_LASTEXTLast external is listed in 
Barracuda RBL bb.barracudacentral.org

score   RCVD_IN_BRBL_LASTEXT2.2
tflags  RCVD_IN_BRBL_LASTEXTnet

endif


NOTE: recent masscheck processing shows RCVD_IN_BRBL_LASTEXT as the most 
effective non-zero scoring rule to hit spam (43%) so it's probably worth 
adding back to your local.cf on Wednesday or Thursday.


http://ruleqa.spamassassin.org/20180203-r1823022-n/RCVD_IN_BRBL_LASTEXT/detail

--
David Jones


Re: sa-update ruleset updates enabled again

2017-11-19 Thread David Jones

On 11/19/2017 08:45 AM, David Mehler wrote:

Hi,

How does one get the new SA update rules?

Thanks.
Dave.




Basically run the sa-update command.  This should be cron'd or otherwise 
run automatically by whatever "glue" is calling Spamassassin.  The 
"glue" could be an MTA like Postfix/Sendmail/Exim/procmail/fetchmail via 
a milter using spamd/spamc, a higher level program like 
amavis/mimedefang/MailScanner, or an MUA (mail client) like 
Thunderbird/Apple Mail.


The higher-level "glue" will usually have their own way for running 
sa-update automatically for you in the background so just do some 
Googling based on your "glue."  It shouldn't be run more frequently than 
about 4 hours since there are only 2 updates a day currently around 3 AM 
UTC and 9 AM UTC.


https://wiki.apache.org/spamassassin/RuleUpdates



On 11/19/17, David Jones  wrote:

On 11/18/2017 09:37 PM, John Hardin wrote:

On Sun, 19 Nov 2017, Benny Pedersen wrote:


David Jones skrev den 2017-11-18 16:26:

  Heads up.  DNS updates for sa-update have been enabled again. The next
  rules promotion will happen in about 11 hours around 2:30 AM UTC.


heads up :=)



--
David Jones


Re: sa-update ruleset updates enabled again

2017-11-19 Thread David Jones

On 11/18/2017 09:37 PM, John Hardin wrote:

On Sun, 19 Nov 2017, Benny Pedersen wrote:


David Jones skrev den 2017-11-18 16:26:

 Heads up.  DNS updates for sa-update have been enabled again. The next
 rules promotion will happen in about 11 hours around 2:30 AM UTC.


heads up :=)

: delivery via smtp.ena.net[96.5.1.4]:25: host
    smtp.ena.net[96.5.1.4] said: 554 5.7.1 : Sender
    address rejected: Blocked TLD. Contact hostmas...@ena.com, 
supp...@ena.com

    or call (888)612-2880. (in reply to RCPT TO command)


Right, published administrative contact addresses (particularly abuse@ 
and postmaster@) should have **no content filtering**.


This is one of the more annoying things I run into when trying to be a 
good netizen, particularly when the domain gets all its email via Google.





I have updated Postfix to not filter admin contacts so here comes the 
spam.  :)  I already had some local rules to subtract some points for 
these recipients in SA to make the blocking score 10.0 instead of the 
6.0 default in MailScanner.


--
David Jones


Re: sa-update ruleset updates enabled again

2017-11-19 Thread David Jones

On 11/18/2017 09:46 AM, Benny Pedersen wrote:

David Jones skrev den 2017-11-18 16:26:

Heads up.  DNS updates for sa-update have been enabled again. The next
rules promotion will happen in about 11 hours around 2:30 AM UTC.


may i ask why you tld block me ?

sorry for asking here, private mails does not work


I have allowed junc.eu.  Normal mail flow on our mail filters doesn't 
receive any email from .eu so I allow them on-demand like this.  Sorry 
about that.


--
David Jones


Re: sa-update ruleset updates enabled again

2017-11-18 Thread John Hardin

On Sun, 19 Nov 2017, Benny Pedersen wrote:


David Jones skrev den 2017-11-18 16:26:

 Heads up.  DNS updates for sa-update have been enabled again. The next
 rules promotion will happen in about 11 hours around 2:30 AM UTC.


heads up :=)

: delivery via smtp.ena.net[96.5.1.4]:25: host
smtp.ena.net[96.5.1.4] said: 554 5.7.1 : Sender
address rejected: Blocked TLD. Contact hostmas...@ena.com, supp...@ena.com
or call (888)612-2880. (in reply to RCPT TO command)


Right, published administrative contact addresses (particularly 
abuse@ and postmaster@) should have **no content filtering**.


This is one of the more annoying things I run into when trying to be a 
good netizen, particularly when the domain gets all its email via Google.



--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  If the rock of doom requires a gentle nudge away from Gaia to
  prevent a very bad day for Earthlings, NASA won’t be riding to the
  rescue. These days, NASA does dodgy weather research and outreach
  programs, not stuff in actual space with rockets piloted by
  flinty-eyed men called Buzz.   -- Daily Bayonet
---
 233 days since the first commercial re-flight of an orbital booster (SpaceX)

Re: sa-update ruleset updates enabled again

2017-11-18 Thread Benny Pedersen

David Jones skrev den 2017-11-18 16:26:

Heads up.  DNS updates for sa-update have been enabled again. The next
rules promotion will happen in about 11 hours around 2:30 AM UTC.


heads up :=)

: delivery via smtp.ena.net[96.5.1.4]:25: host
smtp.ena.net[96.5.1.4] said: 554 5.7.1 : Sender
address rejected: Blocked TLD. Contact hostmas...@ena.com, 
supp...@ena.com

or call (888)612-2880. (in reply to RCPT TO command)

i dont make long distance phone calls


Re: sa-update ruleset updates enabled again

2017-11-18 Thread Benny Pedersen

David Jones skrev den 2017-11-18 16:26:

Heads up.  DNS updates for sa-update have been enabled again. The next
rules promotion will happen in about 11 hours around 2:30 AM UTC.


may i ask why you tld block me ?

sorry for asking here, private mails does not work


sa-update ruleset updates enabled again

2017-11-18 Thread David Jones
Heads up.  DNS updates for sa-update have been enabled again. The next 
rules promotion will happen in about 11 hours around 2:30 AM UTC.


--
David Jones



Re: Ruleset updates via nightly masscheck status

2017-11-13 Thread David Jones

On 11/13/2017 02:33 PM, Tom Hendrikx wrote:



On 28-10-17 15:20, David Jones wrote:

On 10/27/2017 03:02 AM, Merijn van den Kroonenberg wrote:




Please provide feedback in the next 48 hours -- positive or negative so
I know we are good to enable DNS updates again on Sunday.



After installing these rules, I'm seeing one warning in my log during
spamassassin reload:

Oct 27 09:48:24 myhostname spamd[16256]: rules: failed to run
DKIM_VALID_EF test, skipping:
Oct 27 09:48:24 myhostname spamd[16256]:  (Can't locate object method
"check_dkim_valid_envelopefrom" via package "Mail:
[...]:SpamAssassin::PerMsgStatus" at (eval 1369) line 305.
Oct 27 09:48:24 myhostname spamd[16256]: )


The DKIM_VALID_EF rule should not be published yet as it depends on a
change in a Plugin.



Tom, thank you for testing and providing feedback.  I didn't notice this
error because I had patched my DKIM.pm plugin for testing the new
DKIM_VALID_EF rule (intended to be used in meta rules).  I confirmed
what you found on my default Fedora 26 installation.

I have fixed the rulesets, specifically 25_dkim.cf and 50_scores.cf, to
check for the SA version to remove this error and tested it.  Monday's
ruleset should have this fix after tomorrow's masscheck validates it.

I will confirm Monday's ruleset has fixed this DKIM_VALID_EF error and
let sa-update start updating again via DNS on Tuesday.

If anyone else is testing the latest rulesets from the past couple of
days, please provide feedback in the next 48 hours.  And thank you for
testing.


Hi,

I noticed that rule updates are still not live in DNS. Can I get an
updated ruleset for additional testing somewhere, or are we going live?

Regards,
Tom



So...  Merijn has been helping me track down bugs.  I apologize for the 
delay but every little tweak we make takes 1 to 3 days to test.  We make 
some changes over the weekend that are showing good results.  This 
morning we got rules promoted for the first time in about 4 weeks 
(new/different issue) so the masscheck tonight should generate a new 
ruleset.


I will check things in the morning and send out an update.  If we have a 
good 72_scores.cf then I will install the latest ruleset manually on my 
platforms.  If everything looks fine, I will ask for volunteers again to 
do the same and check their scoring.


--
David Jones


Re: Ruleset updates via nightly masscheck status

2017-11-13 Thread Tom Hendrikx


On 28-10-17 15:20, David Jones wrote:
> On 10/27/2017 03:02 AM, Merijn van den Kroonenberg wrote:
>>
>>>>
>>>> Please provide feedback in the next 48 hours -- positive or negative so
>>>> I know we are good to enable DNS updates again on Sunday.
>>>>
>>>
>>> After installing these rules, I'm seeing one warning in my log during
>>> spamassassin reload:
>>>
>>> Oct 27 09:48:24 myhostname spamd[16256]: rules: failed to run
>>> DKIM_VALID_EF test, skipping:
>>> Oct 27 09:48:24 myhostname spamd[16256]:  (Can't locate object method
>>> "check_dkim_valid_envelopefrom" via package "Mail:
>>> [...]:SpamAssassin::PerMsgStatus" at (eval 1369) line 305.
>>> Oct 27 09:48:24 myhostname spamd[16256]: )
>>
>> The DKIM_VALID_EF rule should not be published yet as it depends on a
>> change in a Plugin.
>>
> 
> Tom, thank you for testing and providing feedback.  I didn't notice this
> error because I had patched my DKIM.pm plugin for testing the new
> DKIM_VALID_EF rule (intended to be used in meta rules).  I confirmed
> what you found on my default Fedora 26 installation.
> 
> I have fixed the rulesets, specifically 25_dkim.cf and 50_scores.cf, to
> check for the SA version to remove this error and tested it.  Monday's
> ruleset should have this fix after tomorrow's masscheck validates it.
> 
> I will confirm Monday's ruleset has fixed this DKIM_VALID_EF error and
> let sa-update start updating again via DNS on Tuesday.
> 
> If anyone else is testing the latest rulesets from the past couple of
> days, please provide feedback in the next 48 hours.  And thank you for
> testing.

Hi,

I noticed that rule updates are still not live in DNS. Can I get an
updated ruleset for additional testing somewhere, or are we going live?

Regards,
Tom


Re: Ruleset updates via nightly masscheck status

2017-10-28 Thread David Jones

On 10/27/2017 03:02 AM, Merijn van den Kroonenberg wrote:




Please provide feedback in the next 48 hours -- positive or negative so
I know we are good to enable DNS updates again on Sunday.



After installing these rules, I'm seeing one warning in my log during
spamassassin reload:

Oct 27 09:48:24 myhostname spamd[16256]: rules: failed to run
DKIM_VALID_EF test, skipping:
Oct 27 09:48:24 myhostname spamd[16256]:  (Can't locate object method
"check_dkim_valid_envelopefrom" via package "Mail:
[...]:SpamAssassin::PerMsgStatus" at (eval 1369) line 305.
Oct 27 09:48:24 myhostname spamd[16256]: )


The DKIM_VALID_EF rule should not be published yet as it depends on a
change in a Plugin.



Tom, thank you for testing and providing feedback.  I didn't notice this 
error because I had patched my DKIM.pm plugin for testing the new 
DKIM_VALID_EF rule (intended to be used in meta rules).  I confirmed 
what you found on my default Fedora 26 installation.


I have fixed the rulesets, specifically 25_dkim.cf and 50_scores.cf, to 
check for the SA version to remove this error and tested it.  Monday's 
ruleset should have this fix after tomorrow's masscheck validates it.


I will confirm Monday's ruleset has fixed this DKIM_VALID_EF error and 
let sa-update start updating again via DNS on Tuesday.


If anyone else is testing the latest rulesets from the past couple of 
days, please provide feedback in the next 48 hours.  And thank you for 
testing.




Any idea? This is ubuntu 16.04, latest ubuntu package (3.4.1-3) for
spammassassin.


Kind regards,
Tom



--
David Jones


Re: Ruleset updates via nightly masscheck status

2017-10-27 Thread Merijn van den Kroonenberg

>>
>> Please provide feedback in the next 48 hours -- positive or negative so
>> I know we are good to enable DNS updates again on Sunday.
>>
>
> After installing these rules, I'm seeing one warning in my log during
> spamassassin reload:
>
> Oct 27 09:48:24 myhostname spamd[16256]: rules: failed to run
> DKIM_VALID_EF test, skipping:
> Oct 27 09:48:24 myhostname spamd[16256]:  (Can't locate object method
> "check_dkim_valid_envelopefrom" via package "Mail:
> [...]:SpamAssassin::PerMsgStatus" at (eval 1369) line 305.
> Oct 27 09:48:24 myhostname spamd[16256]: )

The DKIM_VALID_EF rule should not be published yet as it depends on a
change in a Plugin.

>
> Any idea? This is ubuntu 16.04, latest ubuntu package (3.4.1-3) for
> spammassassin.
>
>
> Kind regards,
>   Tom
>




Re: Ruleset updates via nightly masscheck status

2017-10-27 Thread Tom Hendrikx


On 26-10-17 20:33, David Jones wrote:
> On 10/26/2017 01:09 PM, David Jones wrote:
>> On 10/25/2017 06:15 AM, David Jones wrote:
>>> cd /tmp
>>> wget http://sa-update.ena.com/1813149.tar.gz
>>> wget http://sa-update.ena.com/1813149.tar.gz.sha1
>>> wget http://sa-update.ena.com/1813149.tar.gz.asc
>>> sa-update -v --install 1813149.tar.gz
> 
> Last night's run also successfully put the last known good 72_scores.cf
> from March into the ruleset.
> 
> Steps to manually installing last night's ruleset:
> 
> cd /tmp
> wget http://sa-update.ena.com/1813258.tar.gz
> wget http://sa-update.ena.com/1813258.tar.gz.sha1
> wget http://sa-update.ena.com/1813258.tar.gz.asc
> sa-update -v --install 1813258.tar.gz
> 
> restart spamd, MailScanner, amavisd, mimedefang, etc.
> 
> Please provide feedback in the next 48 hours -- positive or negative so
> I know we are good to enable DNS updates again on Sunday.
> 

After installing these rules, I'm seeing one warning in my log during
spamassassin reload:

Oct 27 09:48:24 myhostname spamd[16256]: rules: failed to run
DKIM_VALID_EF test, skipping:
Oct 27 09:48:24 myhostname spamd[16256]:  (Can't locate object method
"check_dkim_valid_envelopefrom" via package "Mail:
[...]:SpamAssassin::PerMsgStatus" at (eval 1369) line 305.
Oct 27 09:48:24 myhostname spamd[16256]: )

Any idea? This is ubuntu 16.04, latest ubuntu package (3.4.1-3) for
spammassassin.


Kind regards,
Tom


Re: Ruleset updates via nightly masscheck status

2017-10-26 Thread David Jones

On 10/26/2017 01:09 PM, David Jones wrote:

On 10/25/2017 06:15 AM, David Jones wrote:

cd /tmp
wget http://sa-update.ena.com/1813149.tar.gz
wget http://sa-update.ena.com/1813149.tar.gz.sha1
wget http://sa-update.ena.com/1813149.tar.gz.asc
sa-update -v --install 1813149.tar.gz


Last night's run also successfully put the last known good 72_scores.cf
from March into the ruleset.

Steps to manually installing last night's ruleset:

cd /tmp
wget http://sa-update.ena.com/1813258.tar.gz
wget http://sa-update.ena.com/1813258.tar.gz.sha1
wget http://sa-update.ena.com/1813258.tar.gz.asc
sa-update -v --install 1813258.tar.gz

restart spamd, MailScanner, amavisd, mimedefang, etc.

Please provide feedback in the next 48 hours -- positive or negative so
I know we are good to enable DNS updates again on Sunday.

I would like to enable DNS updates on Saturday evening (US time) so the
sa-updates would start again on Sunday.

P.S.  Merijn van den Kroonenberg is helping me look at the root of the
problem so we might be able to fix this properly in the next couple of
weeks.

--
David Jones


Re: Just interested: MIME validation ruleset and ASCII-0

2013-05-30 Thread Jari Fredriksson
30.05.2013 14:38, Simon Loewenthal kirjoitti:
>
> Hi there,
>
>  The SA custom rulesets page refers to /MIME validation/ ruleset. This
> is a small .cf file.  I am interested in this rule:
>
> # ASCII-0 can crash mail clients. This is an absolute NO!
> rawbody MIME_ASCII0 /\0/
> describeMIME_ASCII0 Message body contains ASCII-0 character
> score   MIME_ASCII0 1.5
>
>
> Does anyone know why this char should crash an email client?  I did google a 
> little for ASCII charsets and this char, but I could not see how this might 
> cause a crash.  I wonder if this is not caused by badly coded email client 
> implementations?
>
> Perhaps I am showing off my noob-ness :)
>
> Thanks!
> Simon
At least is a string terminator in C language library and most apps
coded with it. May cause indefined behaviour.



signature.asc
Description: OpenPGP digital signature


Just interested: MIME validation ruleset and ASCII-0

2013-05-30 Thread Simon Loewenthal
 

Hi there, 

 The SA custom rulesets page refers to _MIME validation_ ruleset. This
is a small .cf file. I am interested in this rule: 

# ASCII-0 can crash mail clients. This is an absolute NO!
rawbody MIME_ASCII0 //
describe MIME_ASCII0 Message body contains ASCII-0 character
score MIME_ASCII0 1.5

Does anyone know why this char should crash an email client? I did
google a little for ASCII charsets and this char, but I could not see
how this might cause a crash. I wonder if this is not caused by badly
coded email client implementations?

Perhaps I am showing off my noob-ness :)

Thanks!
Simon
 

Re: Using ZMI_GERMAN ruleset

2011-12-19 Thread Michael Monnerie
Sorry for the delay. I don't read the list normally, so please always CC 
me if you want to reach me.

On Mittwoch, 16. November 2011 Stefan Jakobs wrote:
> the published ruleset in the update channel is much older than the
> ruleset on the named website.
> 
> # dig +short -t txt 2.3.3.70_zmi_german.cf.zmi.sa-update.dostech.net
> txt "20100831"
> 
> Is the update with sa-update still supported?  

I'm sorry, I've contacted  "Daryl C. W. O'Shea" 
 on Februar 12 and September 5, 2011, he wanted 
to look it up but it seems still old.

@Daryl: Any chance you can fix that? If you tell me no, I can remove it 
from the list of possible updates.

The correct way to get ZMI_GERMAN is currently per web on 
http://sa.zmi.at/rulesets/70_zmi_german.cf
Version: 3.00.3 from 2011-12-19

If someone can describe exactly what to do to create my own channel with 
10 minutes of work, I can do so, but I don't want to run around 
searching for that info, I'm too busy with other projects.
I'd also need info what users would then need to do to get that updates.

-- 
mit freundlichen Grüssen,
Michael Monnerie, Ing. BSc

it-management Internet Services: Protéger
http://proteger.at [gesprochen: Prot-e-schee]
Tel: +43 660 / 415 6531


signature.asc
Description: This is a digitally signed message part.


Re: Using ZMI_GERMAN ruleset

2011-12-19 Thread Sebastian Wiesinger
* Stefan Jakobs  [2011-11-16 11:28]:
> Hi list,  
> 
> the published ruleset in the update channel is much older than the ruleset on
> the named website.  
> 
> # dig +short -t txt 2.3.3.70_zmi_german.cf.zmi.sa-update.dostech.net txt
> "20100831"  
> 
> Is the update with sa-update still supported?  

Michael,

could you answer that?

Regards,

Sebastian

-- 
New GPG Key: 0x93A0B9CE (F4F6 B1A3 866B 26E9 450A  9D82 58A2 D94A 93A0 B9CE)
Old GPG Key-ID: 0x76B79F20 (0x1B6034F476B79F20)
'Are you Death?' ... IT'S THE SCYTHE, ISN'T IT? PEOPLE ALWAYS NOTICE THE SCYTHE.
-- Terry Pratchett, The Fifth Elephant


Re: Using ZMI_GERMAN ruleset

2011-12-14 Thread Michael Monnerie
On Dienstag, 13. Dezember 2011 Axb wrote:
> patterns with >120 characters are not really efficient, in terms of 
> speed and hit rate. They are very specific to certain campaigns and 
> minimal template changes will render them useless as in:
> 
> body __ZMIde_STOCK34 /Wir sind .{0,2}berzeugt, dass der
> Zeitpunkt  sich an einem Unternehmen zu beteiligen, welches
> erfolgreich im Edelmetallhandel t.{0,2}tig ist, nicht besser sein
> k.{0,2}nnte/
> 
> or
> 
> body __ZMIde_SALE5 /In den letzten 5 Jahren hatte ich .{0,2}ber
> drei  dutzend gut funktionierende Strategien, um die Zahl meiner
> Webseitenbesucher drastisch zu erh.{0,2}hen und dadurch meinen
> Umsatz  anzukurbeln/

Since they get hits, no need to change them. Once I get reports about a 
sentence that has been modified, I apply that to the rules. I need 
feedback for those, of course.
And it should still be fast in terms of CPU as if there's (rule 
__ZMIde_SALE5) no "In den l" in the message, the regex shouldn't have to 
search too much, right? At least I'd guess it's an optimized search 
which compares in 64bit steps, which is 8 chars at once?

Look for the "Krankenkassa" ruleset, this has been very active these 
last weeks. All the time modifications from them, I get reports and 
modify the rules accordingly.

And not to forget: Long sentences mean chance for a false positive drops

-- 
mit freundlichen Grüssen,
Michael Monnerie, Ing. BSc

it-management Internet Services: Protéger
http://proteger.at [gesprochen: Prot-e-schee]
Tel: +43 660 / 415 6531


signature.asc
Description: This is a digitally signed message part.


Re: Using ZMI_GERMAN ruleset

2011-12-13 Thread Axb

On 2011-12-13 8:34, Michael Monnerie wrote:

On Montag, 31. Oktober 2011 Axb wrote:

tried it and dumped due to low hit rate

stuff like

body ZMIde_JOBSEARCH6 /Dank sehr grossen Angagement, aber auch
der  Umsetzung verschiedener Inovationen, konnte unsere Firma schon
nach vier Jahren auf die internationalen Ebene hinaufsteigen/

is not efficient


Its "efficient" in terms of "filtering only spam with zero false
positives", which is top priority for this ruleset. And you picked a
very old and very long rule. Most rules nowadays are just one or even
only part of a sentence, and it prooves very efficient. Stuff like the
__ZMIde_JOBEARN1-28 rules move false positives to 0, and I'm constantly
adding stuff.

I've now tried to remove all old cruft, that means single-line rules.
Rulesize went from 350KB to 296KB, that should save some RAM and CPU.



patterns with >120 characters are not really efficient, in terms of 
speed and hit rate. They are very specific to certain campaigns and 
minimal template changes will render them useless as in:


body __ZMIde_STOCK34 /Wir sind .{0,2}berzeugt, dass der Zeitpunkt 
sich an einem Unternehmen zu beteiligen, welches erfolgreich im 
Edelmetallhandel t.{0,2}tig ist, nicht besser sein k.{0,2}nnte/


or

body __ZMIde_SALE5 /In den letzten 5 Jahren hatte ich .{0,2}ber drei 
dutzend gut funktionierende Strategien, um die Zahl meiner 
Webseitenbesucher drastisch zu erh.{0,2}hen und dadurch meinen Umsatz 
anzukurbeln/







Re: Using ZMI_GERMAN ruleset

2011-12-12 Thread Michael Monnerie
On Montag, 31. Oktober 2011 Axb wrote:
> tried it and dumped due to low hit rate
> 
> stuff like
> 
> body ZMIde_JOBSEARCH6 /Dank sehr grossen Angagement, aber auch
> der  Umsetzung verschiedener Inovationen, konnte unsere Firma schon
> nach vier Jahren auf die internationalen Ebene hinaufsteigen/
> 
> is not efficient

Its "efficient" in terms of "filtering only spam with zero false 
positives", which is top priority for this ruleset. And you picked a 
very old and very long rule. Most rules nowadays are just one or even 
only part of a sentence, and it prooves very efficient. Stuff like the 
__ZMIde_JOBEARN1-28 rules move false positives to 0, and I'm constantly 
adding stuff.

I've now tried to remove all old cruft, that means single-line rules. 
Rulesize went from 350KB to 296KB, that should save some RAM and CPU.

-- 
mit freundlichen Grüssen,
Michael Monnerie, Ing. BSc

it-management Internet Services: Protéger
http://proteger.at [gesprochen: Prot-e-schee]
Tel: +43 660 / 415 6531


signature.asc
Description: This is a digitally signed message part.


Re: Using ZMI_GERMAN ruleset

2011-11-16 Thread Stefan Jakobs
Hi list,  

the published ruleset in the update channel is much older than the ruleset on
the named website.  

# dig +short -t txt 2.3.3.70_zmi_german.cf.zmi.sa-update.dostech.net txt
"20100831"  

Is the update with sa-update still supported?  

Thanks and kind regards
Stefan  
>  Mo Okt 31 2011 14:43:22 CET von  "Michael Monnerie"
>  Betreff: Using ZMI_GERMAN ruleset
>
>  

  
>  
>
>Dear list,
> 
> I'd like to receive some feedback on the usage of zmi_german. If you use 
> it, please report to spam-ger...@zmi.at and tell me what you think about 
> it.
> 
> The ruleset is designed to filter only german spam, and is very safe. 
> Not a single report this year about FPs. If you didn't use it until now, 
> and get german spam, download it from 
> http://sa.zmi.at/rulesets/70_zmi_german.cf
> 
> I'm seeking for people helping to cleanup and improve the filters. 
> Please contact me at spam-ger...@zmi.at
>


Re: Using ZMI_GERMAN ruleset

2011-10-31 Thread Axb

On 2011-10-31 14:43, Michael Monnerie wrote:

Dear list,

I'd like to receive some feedback on the usage of zmi_german. If you use
it, please report to spam-ger...@zmi.at and tell me what you think about
it.

The ruleset is designed to filter only german spam, and is very safe.
Not a single report this year about FPs. If you didn't use it until now,
and get german spam, download it from
http://sa.zmi.at/rulesets/70_zmi_german.cf

I'm seeking for people helping to cleanup and improve the filters.
Please contact me at spam-ger...@zmi.at



tried it and dumped due to low hit rate

stuff like

body ZMIde_JOBSEARCH6 /Dank sehr grossen Angagement, aber auch der 
Umsetzung verschiedener Inovationen, konnte unsere Firma schon nach vier 
Jahren auf die internationalen Ebene hinaufsteigen/


is not efficient


Using ZMI_GERMAN ruleset

2011-10-31 Thread Michael Monnerie
Dear list,

I'd like to receive some feedback on the usage of zmi_german. If you use 
it, please report to spam-ger...@zmi.at and tell me what you think about 
it.

The ruleset is designed to filter only german spam, and is very safe. 
Not a single report this year about FPs. If you didn't use it until now, 
and get german spam, download it from 
http://sa.zmi.at/rulesets/70_zmi_german.cf

I'm seeking for people helping to cleanup and improve the filters. 
Please contact me at spam-ger...@zmi.at

-- 
mit freundlichen Grüssen,
Michael Monnerie, Ing. BSc

it-management Internet Services: Protéger
http://proteger.at [gesprochen: Prot-e-schee]
Tel: +43 660 / 415 6531


signature.asc
Description: This is a digitally signed message part.


Re: How to remove a domain from a stock or third-party 2tld ruleset?

2010-06-07 Thread Kris Deugau

Kris Deugau wrote:

Karsten Bräckelmann wrote:

Another approach, since I understand you want to query against a local
URI DNSBL, is simply to use wildcard DNS entries. Thus, regardless of a
2tld listing and the resulting DNS lookup, it would return the same
listing for the pure TLD and a second level TLD.


Hmm.  I hadn't thought of this, I'll give it a try and see if something 
chokes.  Thanks!


This seems to be a usable way to work around a domain in the stock 
util_rb_2tld lists.  I added *.t35.com (made a convenient test case - 
actually listed locally with util_rb_2tld;  going to remove it 
eventually) to our local URI blacklist, and while there have been missed 
spams with t35.com subdomains, none have shown up in the list to be 
added to the blacklist since I did so.


-kgd


Re: How to remove a domain from a stock or third-party 2tld ruleset?

2010-05-28 Thread Yet Another Ninja

On 2010-05-28 23:57, Kris Deugau wrote:

Karsten Bräckelmann wrote:

On Wed, 2010-05-26 at 11:35 -0400, Kris Deugau wrote:
Is there any way to take a domain listed with util_rb_2tld, and 
"un-2tld" it (similar to how you can unwhitelist stock whitelist 
entries if they don't work well with your mail)?


IIRC this is not possible. Well, possible, but there's just no code to
handle it. ;)


Didn't think so, but...

I recently came across a "free-subsite" domain that seems to be part 
of a cluster of **very** similar sites which I've given up listing 
subdomains for locally;  instead I've added the TLDs to a local 
blacklist.


For now I've just added a regular uri rule, but I'm pretty sure that 
won't scale, and it doesn't help with some of the automation I've 
been using to extract URIs not listed on any DNSBL yet from 
missed-spam reports.


uri rules should work. I wouldn't worry about scaling too much, because
the number of util_rb_2tld listings is limited.

Another approach, since I understand you want to query against a local
URI DNSBL, is simply to use wildcard DNS entries. Thus, regardless of a
2tld listing and the resulting DNS lookup, it would return the same
listing for the pure TLD and a second level TLD.


Hmm.  I hadn't thought of this, I'll give it a try and see if something 
chokes.  Thanks!


let me guess... .co.cc ?






Re: How to remove a domain from a stock or third-party 2tld ruleset?

2010-05-28 Thread Kris Deugau

Karsten Bräckelmann wrote:

On Wed, 2010-05-26 at 11:35 -0400, Kris Deugau wrote:
Is there any way to take a domain listed with util_rb_2tld, and 
"un-2tld" it (similar to how you can unwhitelist stock whitelist entries 
if they don't work well with your mail)?


IIRC this is not possible. Well, possible, but there's just no code to
handle it. ;)


Didn't think so, but...

I recently came across a "free-subsite" domain that seems to be part of 
a cluster of **very** similar sites which I've given up listing 
subdomains for locally;  instead I've added the TLDs to a local blacklist.


For now I've just added a regular uri rule, but I'm pretty sure that 
won't scale, and it doesn't help with some of the automation I've been 
using to extract URIs not listed on any DNSBL yet from missed-spam reports.


uri rules should work. I wouldn't worry about scaling too much, because
the number of util_rb_2tld listings is limited.

Another approach, since I understand you want to query against a local
URI DNSBL, is simply to use wildcard DNS entries. Thus, regardless of a
2tld listing and the resulting DNS lookup, it would return the same
listing for the pure TLD and a second level TLD.


Hmm.  I hadn't thought of this, I'll give it a try and see if something 
chokes.  Thanks!


-kgd


Re: How to remove a domain from a stock or third-party 2tld ruleset?

2010-05-28 Thread Karsten Bräckelmann
On Wed, 2010-05-26 at 11:35 -0400, Kris Deugau wrote:
> Is there any way to take a domain listed with util_rb_2tld, and 
> "un-2tld" it (similar to how you can unwhitelist stock whitelist entries 
> if they don't work well with your mail)?

IIRC this is not possible. Well, possible, but there's just no code to
handle it. ;)

> I recently came across a "free-subsite" domain that seems to be part of 
> a cluster of **very** similar sites which I've given up listing 
> subdomains for locally;  instead I've added the TLDs to a local blacklist.

> For now I've just added a regular uri rule, but I'm pretty sure that 
> won't scale, and it doesn't help with some of the automation I've been 
> using to extract URIs not listed on any DNSBL yet from missed-spam reports.

uri rules should work. I wouldn't worry about scaling too much, because
the number of util_rb_2tld listings is limited.

Another approach, since I understand you want to query against a local
URI DNSBL, is simply to use wildcard DNS entries. Thus, regardless of a
2tld listing and the resulting DNS lookup, it would return the same
listing for the pure TLD and a second level TLD.


-- 
char *t="\10pse\0r\0dtu...@ghno\x4e\xc8\x79\xf4\xab\x51\x8a\x10\xf4\xf4\xc4";
main(){ char h,m=h=*t++,*x=t+2*h,c,i,l=*x,s=0; for (i=0;i>=1)||!t[s+h]){ putchar(t[s]);h=m;s=0; }}}



How to remove a domain from a stock or third-party 2tld ruleset?

2010-05-26 Thread Kris Deugau
Is there any way to take a domain listed with util_rb_2tld, and 
"un-2tld" it (similar to how you can unwhitelist stock whitelist entries 
if they don't work well with your mail)?


I recently came across a "free-subsite" domain that seems to be part of 
a cluster of **very** similar sites which I've given up listing 
subdomains for locally;  instead I've added the TLDs to a local blacklist.


The domain that's in the stock 2tld list is bravepages.com;  it seems to 
be Yet Another Face of 0catch.com, and I've seen these domains as well:


1accesshost.com
bigheadhosting.net
easyfreehosting.com
envy.nu
digitalzones.com

And no doubt there are a fairly long list of others in the cluster.

For now I've just added a regular uri rule, but I'm pretty sure that 
won't scale, and it doesn't help with some of the automation I've been 
using to extract URIs not listed on any DNSBL yet from missed-spam reports.


-kgd


RE: GERMAN ruleset extended

2010-04-21 Thread Giampaolo Tomassoni
> Dear users,
> 
> I felt I didn't advertise our GERMAN ruleset since a long time:
> 
> http://wiki.apache.org/spamassassin/CustomRulesets

Good hint. Thank you.

Giampaolo




GERMAN ruleset extended

2010-04-21 Thread Michael Monnerie
Dear users,

I felt I didn't advertise our GERMAN ruleset since a long time:

http://wiki.apache.org/spamassassin/CustomRulesets

Please feel free to implement it to catch out german spam. Should you 
use it and still receive german spam, please report that mail 
*including all headers* to spam-ger...@zmi.at

I don't follow the list regularly, so please contact me there in case of 
problems.

-- 
mit freundlichen Grüssen,
Michael Monnerie, Ing. BSc

it-management Internet Services
http://proteger.at [gesprochen: Prot-e-schee]
Tel: 0660 / 415 65 31

// Wir haben im Moment zwei Häuser zu verkaufen:
// http://zmi.at/langegg/
// http://zmi.at/haus2009/


signature.asc
Description: This is a digitally signed message part.


Re: SOUGHT ruleset FP

2010-04-16 Thread Matthew Newton
Hi,

On Fri, Apr 16, 2010 at 01:53:55PM +0200, Karsten Bräckelmann wrote:
> On Fri, 2010-04-16 at 12:20 +0100, Matthew Newton wrote:
> > We had a legitimate e-mail hit the JM_SOUGHT_3 yesterday. It also
> > hit a few other rules that pushed it over our reject threshold of
> > 10, and easily over the 'junk mail folder' level of 5.
> > 
> > I managed to get them to send me the message, and it hits rule
> > __SEEK_5ID3LI "Conti  nuum Intern ational Publishing" (spaces
> > added!) which is the name of their company.
> 
> Makes one wonder how that string ends up quite massively in spam traps.

I did consider that. Without seeing the spam, of course, I can't
say whether they are spamming or whether their name is being
abused. All I have is a legitimate mail from them and a report
that it is blocked.

> > I know SOUGHT is an auto-generated ruleset; just wondering if
> > there is there any way to remove false positives before the set is
> 
> Yes. The Seek bits are cross-checked against a ham corpus, so the
> easiest way is to inject an artificial ham message with the string in
> question to get it off of the next run.

OK, understood.

> > generated? Otherwise I'll add local rules to compensate against
> > this one.
> 
> meta __SEEK_5ID3LI  (0)
> 
> The Seek ID is constant, and will be the same even with later Sought
> runs, for a given string.

Thought that might be the case from the codes - thanks for the
confirmation.

Cheers,

Matthew


-- 
Matthew Newton, Ph.D. 

Systems Architect (UNIX and Networks), Network Services,
I.T. Services, University of Leicester, Leicester LE1 7RH, United Kingdom

For IT help contact helpdesk extn. 2253, 


Re: SOUGHT ruleset FP

2010-04-16 Thread Justin Mason
yep -- feel free to send me over copies of FP messages (or strings
that match them)

2010/4/16 Karsten Bräckelmann :
> On Fri, 2010-04-16 at 12:20 +0100, Matthew Newton wrote:
>> We had a legitimate e-mail hit the JM_SOUGHT_3 yesterday. It also
>> hit a few other rules that pushed it over our reject threshold of
>> 10, and easily over the 'junk mail folder' level of 5.
>>
>> I managed to get them to send me the message, and it hits rule
>> __SEEK_5ID3LI "Conti  nuum Intern ational Publishing" (spaces
>> added!) which is the name of their company.
>
> Makes one wonder how that string ends up quite massively in spam traps.
>
>> I know SOUGHT is an auto-generated ruleset; just wondering if
>> there is there any way to remove false positives before the set is
>
> Yes. The Seek bits are cross-checked against a ham corpus, so the
> easiest way is to inject an artificial ham message with the string in
> question to get it off of the next run.
>
>> generated? Otherwise I'll add local rules to compensate against
>> this one.
>
> meta __SEEK_5ID3LI  (0)
>
> The Seek ID is constant, and will be the same even with later Sought
> runs, for a given string.
>
>  guenther
>
>
> --
> char *t="\10pse\0r\0dtu...@ghno\x4e\xc8\x79\xf4\xab\x51\x8a\x10\xf4\xf4\xc4";
> main(){ char h,m=h=*t++,*x=t+2*h,c,i,l=*x,s=0; for (i=0;i (c=*++x); c&128 && (s+=h); if (!(h>>=1)||!t[s+h]){ putchar(t[s]);h=m;s=0; }}}
>
>


Re: SOUGHT ruleset FP

2010-04-16 Thread Karsten Bräckelmann
On Fri, 2010-04-16 at 12:20 +0100, Matthew Newton wrote:
> We had a legitimate e-mail hit the JM_SOUGHT_3 yesterday. It also
> hit a few other rules that pushed it over our reject threshold of
> 10, and easily over the 'junk mail folder' level of 5.
> 
> I managed to get them to send me the message, and it hits rule
> __SEEK_5ID3LI "Conti  nuum Intern ational Publishing" (spaces
> added!) which is the name of their company.

Makes one wonder how that string ends up quite massively in spam traps.

> I know SOUGHT is an auto-generated ruleset; just wondering if
> there is there any way to remove false positives before the set is

Yes. The Seek bits are cross-checked against a ham corpus, so the
easiest way is to inject an artificial ham message with the string in
question to get it off of the next run.

> generated? Otherwise I'll add local rules to compensate against
> this one.

meta __SEEK_5ID3LI  (0)

The Seek ID is constant, and will be the same even with later Sought
runs, for a given string.

  guenther


-- 
char *t="\10pse\0r\0dtu...@ghno\x4e\xc8\x79\xf4\xab\x51\x8a\x10\xf4\xf4\xc4";
main(){ char h,m=h=*t++,*x=t+2*h,c,i,l=*x,s=0; for (i=0;i>=1)||!t[s+h]){ putchar(t[s]);h=m;s=0; }}}



SOUGHT ruleset FP

2010-04-16 Thread Matthew Newton
Hi,

We had a legitimate e-mail hit the JM_SOUGHT_3 yesterday. It also
hit a few other rules that pushed it over our reject threshold of
10, and easily over the 'junk mail folder' level of 5.

I managed to get them to send me the message, and it hits rule
__SEEK_5ID3LI "Conti  nuum Intern ational Publishing" (spaces
added!) which is the name of their company.

I know SOUGHT is an auto-generated ruleset; just wondering if
there is there any way to remove false positives before the set is
generated? Otherwise I'll add local rules to compensate against
this one.

Thanks,

Matthew


-- 
Matthew Newton, Ph.D. 

Systems Architect (UNIX and Networks), Network Services,
I.T. Services, University of Leicester, Leicester LE1 7RH, United Kingdom

For IT help contact helpdesk extn. 2253, 


Re: Hijacked thread :) (was: ruleset for German...)

2010-03-16 Thread Charles Gregory

On Mon, 15 Mar 2010, Karsten Bräckelmann wrote:

The TextCat plugin. Even part of stock SA, though not enabled by
default. Supports per-user settings.


(nod) For reasons specific to my MTA, I can't run SA 'per user', but I can 
choose the most common languages (en fr) in our system's mail and flag 
when neither of them are used (assigning UNWANTED_LANGUAGE_BODY a minimal 
score) - then the user can set a procmail delivery rule (quarantine when

that rule is present in the X-Spam headers). It will do. :)

But you just forked (to avoid the word hijacked) this thread, which is 
about a very specific, on-going spam run. The OP really doesn't want to 
identify German spam for scoring, cause that's likely his first 
language. ;)


My bad. :)

But my compliments on the OP's excellent English! :)

- C

Re: [sa] Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-16 Thread Henrik K
On Mon, Mar 15, 2010 at 11:17:09PM +0100, Karsten Bräckelmann wrote:
> On Mon, 2010-03-15 at 11:15 -0400, Charles Gregory wrote:
> > H. I guess this goes back to my inquiry about the Brazilian spam
> > 
> > I'm still looking for a way (hopefully) to simply identify the *language* 
> > of the mail (when not determined from CHARSET_FARAWAY rules), so that our 
> > users may opt-in for additional filtering based on language
> 
> The TextCat plugin. Even part of stock SA, though not enabled by
> default. Supports per-user settings.

Though given the current bugs, I really would do any hard filtering..



Re: [sa] Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-15 Thread Karsten Bräckelmann
On Mon, 2010-03-15 at 11:15 -0400, Charles Gregory wrote:
> H. I guess this goes back to my inquiry about the Brazilian spam
> 
> I'm still looking for a way (hopefully) to simply identify the *language* 
> of the mail (when not determined from CHARSET_FARAWAY rules), so that our 
> users may opt-in for additional filtering based on language

The TextCat plugin. Even part of stock SA, though not enabled by
default. Supports per-user settings.

But you just forked (to avoid the word hijacked) this thread, which is
about a very specific, on-going spam run. The OP really doesn't want to
identify German spam for scoring, cause that's likely his first
language. ;)


-- 
char *t="\10pse\0r\0dtu...@ghno\x4e\xc8\x79\xf4\xab\x51\x8a\x10\xf4\xf4\xc4";
main(){ char h,m=h=*t++,*x=t+2*h,c,i,l=*x,s=0; for (i=0;i>=1)||!t[s+h]){ putchar(t[s]);h=m;s=0; }}}



Re: [sa] Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-15 Thread Charles Gregory

On Sun, 14 Mar 2010, Jörg Frings-Fürst wrote:

take a look at http://wiki.apache.org/spamassassin/CustomRulesets
and search to "German Language Ruleset".


H. I guess this goes back to my inquiry about the Brazilian spam

I'm still looking for a way (hopefully) to simply identify the *language* 
of the mail (when not determined from CHARSET_FARAWAY rules), so that our 
users may opt-in for additional filtering based on language


- Charles

Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-15 Thread Robert Schetterer
Am 15.03.2010 03:14, schrieb Marcus:
> Am Sonntag, den 14.03.2010, 23:31 +0100 schrieb Kai Schaetzl:
>> Marcus wrote on Sun, 14 Mar 2010 21:16:31 +0100:
>>
>>> The messages differ in subject and body.
>>
>> Do they? Hm, I must have overlooked this in your first message. Oh, wait, 
>> after reading it a second time I still can't see it. I think you must have 
>> forgotten to mention it.
>> If you want help from the mailing list post some complete messages to 
>> pastebin.
> 
> Seems to be that other german speaking members of this list know what
> sort of spams I'm talking about. I got some helpful rulesets from them
> for this case. Thanks!
> 
> Kai, thank you for your constructive help and good night,
> Marcus
> 

Hi, i only get these from a forward service
from gmx , they are allready marked then , i let spamassassin them
autolearn by rule afterwards, so if they come next , they are noticed as
spam
so manual learning these mails should help as tmp workaround
analyse the headers to find out which procedure fits best
to catch/mark/reject them
use more rulesets anyway as recommended before

-- 
Best Regards

MfG Robert Schetterer

Germany/Munich/Bavaria


Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-14 Thread Marcus
Am Sonntag, den 14.03.2010, 23:31 +0100 schrieb Kai Schaetzl:
> Marcus wrote on Sun, 14 Mar 2010 21:16:31 +0100:
> 
> > The messages differ in subject and body.
> 
> Do they? Hm, I must have overlooked this in your first message. Oh, wait, 
> after reading it a second time I still can't see it. I think you must have 
> forgotten to mention it.
> If you want help from the mailing list post some complete messages to 
> pastebin.

Seems to be that other german speaking members of this list know what
sort of spams I'm talking about. I got some helpful rulesets from them
for this case. Thanks!

Kai, thank you for your constructive help and good night,
Marcus



Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-14 Thread Kai Schaetzl
Marcus wrote on Sun, 14 Mar 2010 21:16:31 +0100:

> The messages differ in subject and body.

Do they? Hm, I must have overlooked this in your first message. Oh, wait, 
after reading it a second time I still can't see it. I think you must have 
forgotten to mention it.
If you want help from the mailing list post some complete messages to 
pastebin.

Kai

-- 
Get your web at Conactive Internet Services: http://www.conactive.com





Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-14 Thread Marcus
Am Sonntag, den 14.03.2010, 18:31 +0100 schrieb Kai Schaetzl:
> Marcus wrote on Sun, 14 Mar 2010 17:10:08 +0100:
> 
> > I'm getting a lot of these German 'Mehr aus dauer in Ihrem Bettchen' and
> > 'Mehr ausdauer in Ihrem Schlafzimmer'. I've learnd my spamassassin for
> > about a week. It decects some of them, but most are going through. All
> > other kind of spam is detected very well. Did some know oder wrote a
> > ruleset?
> 
> what is so difficult to match against Bettchen or Schlafzimmer?

The messages differ in subject and body.

> I'd say 
> even a complete newbie will have this rule up and running after ten 
> minutes and reading the rules how-to on the wiki.

Ciao!



Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-14 Thread Jörg Frings-Fürst


Am Sonntag 14 März 2010 17:10:08 schrieb Marcus:
> Hi,
> 
> I'm getting a lot of these German 'Mehr aus dauer in Ihrem Bettchen' and
> 'Mehr ausdauer in Ihrem Schlafzimmer'. I've learnd my spamassassin for
> about a week. It decects some of them, but most are going through. All
> other kind of spam is detected very well. Did some know oder wrote a
> ruleset?
> 
> Ciao,
> Marcus
> 

Hi Marcus,

take a look at http://wiki.apache.org/spamassassin/CustomRulesets 
and search to "German Language Ruleset".

CU

Joerg Frings-Fuerst


Re: ruleset for German Bettchen and Schlafzimmer spam

2010-03-14 Thread Kai Schaetzl
Marcus wrote on Sun, 14 Mar 2010 17:10:08 +0100:

> I'm getting a lot of these German 'Mehr aus dauer in Ihrem Bettchen' and
> 'Mehr ausdauer in Ihrem Schlafzimmer'. I've learnd my spamassassin for
> about a week. It decects some of them, but most are going through. All
> other kind of spam is detected very well. Did some know oder wrote a
> ruleset?

what is so difficult to match against Bettchen or Schlafzimmer? I'd say 
even a complete newbie will have this rule up and running after ten 
minutes and reading the rules how-to on the wiki.

Kai

-- 
Get your web at Conactive Internet Services: http://www.conactive.com





ruleset for German Bettchen and Schlafzimmer spam

2010-03-14 Thread Marcus
Hi,

I'm getting a lot of these German 'Mehr aus dauer in Ihrem Bettchen' and
'Mehr ausdauer in Ihrem Schlafzimmer'. I've learnd my spamassassin for
about a week. It decects some of them, but most are going through. All
other kind of spam is detected very well. Did some know oder wrote a
ruleset?

Ciao,
Marcus







Re: SpamAssassin Ruleset Generation

2009-10-07 Thread Karsten Bräckelmann
On Tue, 2009-10-06 at 13:50 -0700, an anonymous Nabble user wrote:
> Other than the sought rules, all the rules are manually generated?

Actually, as has been said, I believe all stock rules are manually
written. There are some third-party rule-sets out there that are auto
generated -- not limited to Sought.

> Is there any statistics on how frequently are new rules/regex adopted by
> spamassasssin? Who are the people who write them? Any details related to it?

Somehow this begs the question -- why?

Why are you asking? Why and what are you ultimately interested in?

And of course, did you even consider to dig through the SVN repo, some
docs on the wiki and to ask google? Most of this should be pretty easy
to find out if you're willing to read some.


-- 
char *t="\10pse\0r\0dtu...@ghno\x4e\xc8\x79\xf4\xab\x51\x8a\x10\xf4\xf4\xc4";
main(){ char h,m=h=*t++,*x=t+2*h,c,i,l=*x,s=0; for (i=0;i>=1)||!t[s+h]){ putchar(t[s]);h=m;s=0; }}}



Re: SpamAssassin Ruleset Generation

2009-10-06 Thread Matt Kettler
poifgh wrote:
> I have a question about - understanding how are rulesets generated for
> spamassassin.
>
> For example - consider the rule in 20_drugs.cf : 
> header SUBJECT_DRUG_GAP_C   Subject =~
> /\bc.{0,2}i.{0,2}a.{0,2}l.{0,2}i.{0,2}s\b/i
> describe SUBJECT_DRUG_GAP_C Subject contains a gappy version of 'cialis'
>
> Who generated the regular expression
> "/\bc.{0,2}i.{0,2}a.{0,2}l.{0,2}i.{0,2}s\b/i"
>   
Man, that's a good question. I wrote a large chunk of the rules in
20_drugs.cf, but not that one. ( I wrote the stuff near the bottom that
uses meta rules. ie:  __DRUGS_ERECTILE1 through DRUGS_MANYKINDS,
originally distributed as a separate set called antidrug.cf). As I
recall, there were 2 other people making drug rules, but it's been a
LONG time, and I forget who did it. Those rules were written in the
2004-2006 time frame when pharmacy spams were just hammering the heck
outa everyone.

> a. Is it done manually with people writing regex to see how efficiently they
> capture spams?
>   
Yes. Many hours of reading spams, studying them, testing various regex
tweaks, checking for false positives, etc, etc.

mass-check is your friend for this kind of stuff.

One post from when I was developing this as a stand-alone set:

http://mail-archives.apache.org/mod_mbox/spamassassin-users/200404.mbox/%3c6.0.0.22.0.20040428132346.029d9...@opal.evi-inc.com%3e

Note: the comcast link mentioned in that message should be considered
DEAD. The antidrug set is no longer maintained separately from the
mailline ruleset, and hasn't been for years.


If you want to break the rules down a bit, here's some tips:

The rules are in general designed to detect common methods to obscure
text by inserting spaces, punctuation, etc between letters, and possibly
substituting some of the letters for other similar looking characters.
(W4R3Z style, etc)

The simple format would be to think of it in groupings. You end up using
a repeating pattern of (some representation of a character)(some kind of
"gap" sequence)(character)(gap)...etc.

.{0,2} is a "gap sequence", although not one I prefer. I prefer
[_\W]{0,3} in most cases because it's a bit less FP-prone, but risks
missing things using small lower-case letters to gap.

You also get replacements for characters in some of those, like [A4]
instead of just A. Or, more elaborately..  [a4\xe0-\...@]

So this mess:

body __DRUGS_ERECTILE1  
/(?:\b|\s)[_\W]{0,3}(?:\\\/|V)[_\W]{0,3}[ij1!|l\xEC\xED\xEE\xEF][_\W]{0,3}[a40\xe0-\...@][_\w]{0,3}[xyz]?[gj][_\W]{0,3}r[_\W]{0,3}[a40\xe0-\...@][_\w]{0,3}x?[_\W]{0,3}(?:\b|\s)/i


Could be broken down:

(?:\b|\s)   - preamble, detecting space or word boundary.
[_\W]{0,3}   - gap
(?:\\\/|V)   - V
[_\W]{0,3}   - gap
[ij1!|l\xEC\xED\xEE\xEF] - I
[_\W]{0,3}   - gap
[a40\xe0-\...@]   - A
[_\W]{0,3}   - gap
[xyz]?[gj]   - G (with optional extra garbage before it)
[_\W]{0,3}   - gap
r- just R :-)
[_\W]{0,3}   - gap
[a40\xe0-\...@] -A
[_\W]{0,3}   - gap
x?   - optional garbage
[_\W]{0,3}   - gap
(?:\b|\s)- suffix, detecting space or word boundary.

Which detects weird spacings and substitutions in the word Viagra.


> But how are the rules generated themselves? 
>   
Mostly meatware, except the sought rules others have mentioned.
> Thnx
>   



Re: SpamAssassin Ruleset Generation

2009-10-06 Thread John Hardin

On Tue, 6 Oct 2009, poifgh wrote:

Other than the sought rules, all the rules are manually generated? Is 
there any statistics on how frequently are new rules/regex adopted by 
spamassasssin? Who are the people who write them? Any details related to 
it?


Most of the rules are manually written by contributors such as myself. 
Some meta rules are generated by various means from existing rules - for 
example, the ADVANCE_FEE rules are generated using genetic algorithms to 
find effective combinations of simpler subrules that were manually 
generated.


New rules are added whenever a contributor works on them, and this is 
generally based on when they have time to do so, when they have new ideas, 
and when new forms of spam appear. Indirect contributors will post rules 
to the users list and a contributor may add them to the rules sandbox for 
testing and eventual inclusion in the base ruleset.


The CREDITS file in the sources should list all of the contributors. Some 
contributors may not have added their names to that file, though.


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
 5 days since a sunspot last seen - EPA blames CO2 emissions


Re: SpamAssassin Ruleset Generation

2009-10-06 Thread MySQL Student
Hi,

> Other than the sought rules, all the rules are manually generated? Is there
> any statistics on how frequently are new rules/regex adopted by
> spamassasssin? Who are the people who write them? Any details related to

Information on Justin Mason's SOUGHT rules is here:

http://taint.org/2007/08/15/004348a.html

Use sa-update to update your SA rules once or twice per day with the
new stuff. His ongoing development work is here:

http://svn.apache.org/viewvc/spamassassin/trunk/rulesrc/sandbox/jm/?sortby=date

HTH,
Alex


Re: SpamAssassin Ruleset Generation

2009-10-06 Thread poifgh



poifgh wrote:
> 
> 
> 
> Bowie Bailey wrote:
>> 
>> 
>> 
>> http://www.google.com/search?q=spamassassin+sought
>> 
> :-D - Thnx
> 
> 

Other than the sought rules, all the rules are manually generated? Is there
any statistics on how frequently are new rules/regex adopted by
spamassasssin? Who are the people who write them? Any details related to it?

thnx
-- 
View this message in context: 
http://www.nabble.com/SpamAssassin-Ruleset-Generation-tp25773508p25776307.html
Sent from the SpamAssassin - Users mailing list archive at Nabble.com.



Re: SpamAssassin Ruleset Generation

2009-10-06 Thread poifgh



Bowie Bailey wrote:
> 
> 
> 
> http://www.google.com/search?q=spamassassin+sought
> 
:-D - Thnx

-- 
View this message in context: 
http://www.nabble.com/SpamAssassin-Ruleset-Generation-tp25773508p25776303.html
Sent from the SpamAssassin - Users mailing list archive at Nabble.com.



Re: SpamAssassin Ruleset Generation

2009-10-06 Thread Bowie Bailey
poifgh wrote:
>
> RW-15 wrote:
>   
>> On Tue, 6 Oct 2009 11:08:28 -0700 (PDT)
>> poifgh  wrote:
>>
>> 
>>> I have a question about - understanding how are rulesets generated for
>>> ...
>>> a. Is it done manually with people writing regex to see how
>>> efficiently they capture spams?
>>> b. Is there an algorithm that identifies large corpus of spam and the
>>> comes up with these regex'es on its own?
>>> c. Is it a combination of (a), (b)?
>>>   
>> The optional sought rules are autogenerated, the rest are manual.
>> 
>
> Thnx - What are optional sought rules?
>   

http://www.google.com/search?q=spamassassin+sought

-- 
Bowie


Re: SpamAssassin Ruleset Generation

2009-10-06 Thread poifgh



RW-15 wrote:
> 
> On Tue, 6 Oct 2009 11:08:28 -0700 (PDT)
> poifgh  wrote:
> 
>> 
>> I have a question about - understanding how are rulesets generated for
>> ...
>> a. Is it done manually with people writing regex to see how
>> efficiently they capture spams?
>> b. Is there an algorithm that identifies large corpus of spam and the
>> comes up with these regex'es on its own?
>> c. Is it a combination of (a), (b)?
> 
> The optional sought rules are autogenerated, the rest are manual.
> 
> 

Thnx - What are optional sought rules?

-- 
View this message in context: 
http://www.nabble.com/SpamAssassin-Ruleset-Generation-tp25773508p25776105.html
Sent from the SpamAssassin - Users mailing list archive at Nabble.com.



Re: SpamAssassin Ruleset Generation

2009-10-06 Thread RW
On Tue, 6 Oct 2009 11:08:28 -0700 (PDT)
poifgh  wrote:

> 
> I have a question about - understanding how are rulesets generated for
> ...
> a. Is it done manually with people writing regex to see how
> efficiently they capture spams?
> b. Is there an algorithm that identifies large corpus of spam and the
> comes up with these regex'es on its own?
> c. Is it a combination of (a), (b)?

The optional sought rules are autogenerated, the rest are manual.


SpamAssassin Ruleset Generation

2009-10-06 Thread poifgh

I have a question about - understanding how are rulesets generated for
spamassassin.

For example - consider the rule in 20_drugs.cf : 
header SUBJECT_DRUG_GAP_C   Subject =~
/\bc.{0,2}i.{0,2}a.{0,2}l.{0,2}i.{0,2}s\b/i
describe SUBJECT_DRUG_GAP_C Subject contains a gappy version of 'cialis'

Who generated the regular expression
"/\bc.{0,2}i.{0,2}a.{0,2}l.{0,2}i.{0,2}s\b/i"

a. Is it done manually with people writing regex to see how efficiently they
capture spams?
b. Is there an algorithm that identifies large corpus of spam and the comes
up with these regex'es on its own?
c. Is it a combination of (a), (b)?

I know scores for rules are generated using "a neural network trained with
error back propagation"
http://wiki.apache.org/spamassassin/HowScoresAreAssigned

But how are the rules generated themselves? 

Thnx
-- 
View this message in context: 
http://www.nabble.com/SpamAssassin-Ruleset-Generation-tp25773508p25773508.html
Sent from the SpamAssassin - Users mailing list archive at Nabble.com.



Re: Optional Tests in Main Ruleset?

2009-06-11 Thread Karsten Bräckelmann
On Thu, 2009-06-11 at 09:18 -0400, Charles Gregory wrote:
> Hallo!
> 
> I've noticed a few rules now that seem to score *very* low.
> For example: DYN_RDNS_AND_INLINE_IMAGE=0.001

There are a lot of possible reasons for that, including informative only
rules (which are likely to have a description to that extent). Poor S/O
rating during the last GA run, but an otherwise potentially useful rule
would be another possible reason. Among others.


> Are these rules 'in development' and therefore not being assigned a 
> significant score as of yet? Or, more interestingly, do they represent an 
> 'optional' set of rules that can be 'activated' by raising their score, 
> but which will otherwise *always* have a low score by default?

Rules in stock are not "in development". Optional rules are more likely
disabled by default, IMHO -- though on the other hand, every single rule
pretty much is optional, and you're free to tweak it.

Just checked current ruleqa results -- looks quite decent indeed.
Doesn't hit a lot of spam, but pretty accurate. Feel free to raise the
score if it would help with your spam, just keep in mind its potential
to cause FPs for you when you do.


> If the latter, then by chance is there a compact listing of these rules 
> that I could review and 'turn on' as needed? Or do I just have to watch 
> for them in the headers?

Sorry, you have to watch them. ;)  There's no such list, and as I
mentioned before, there is a wide range of possible reasons.

Spam (and to a lesser extent, ham) does change, though, and a lot of the
current scores could be adjusted. Reports about FPs are the part usually
visible here. What's just not being reported are rules that are now much
more accurate and candidates for higher scores than they used to be.

  guenther


-- 
char *t="\10pse\0r\0dtu...@ghno\x4e\xc8\x79\xf4\xab\x51\x8a\x10\xf4\xf4\xc4";
main(){ char h,m=h=*t++,*x=t+2*h,c,i,l=*x,s=0; for (i=0;i>=1)||!t[s+h]){ putchar(t[s]);h=m;s=0; }}}



Optional Tests in Main Ruleset?

2009-06-11 Thread Charles Gregory

Hallo!

I've noticed a few rules now that seem to score *very* low.
For example: DYN_RDNS_AND_INLINE_IMAGE=0.001

Are these rules 'in development' and therefore not being assigned a 
significant score as of yet? Or, more interestingly, do they represent an 
'optional' set of rules that can be 'activated' by raising their score, 
but which will otherwise *always* have a low score by default?


If the latter, then by chance is there a compact listing of these rules 
that I could review and 'turn on' as needed? Or do I just have to watch 
for them in the headers?


- Charles



Re: ruleset

2009-03-24 Thread Matt Kettler
JC Putter wrote:
> where can i find more rulesets? using openprotect sare rules and
> sought rulesets
>

That's about all there are... A few folks have odds and ends rules
posted on their webpages/blogs/etc, but they're of mixed quality.

Is there a particular reason your looking for more rulesets?




Re: ruleset

2009-03-24 Thread Matus UHLAR - fantomas
On 24.03.09 15:59, JC Putter wrote:
> where can i find more rulesets? using openprotect sare rules and sought 
> rulesets

build your own rulesets? SARE rulesets aren't updated anymore afaik (and
thus number of false-positives is increasing).

Do you have any problem that can't be solved by fine-tuning SA?

-- 
Matus UHLAR - fantomas, uh...@fantomas.sk ; http://www.fantomas.sk/
Warning: I wish NOT to receive e-mail advertising to this address.
Varovanie: na tuto adresu chcem NEDOSTAVAT akukolvek reklamnu postu.
They that can give up essential liberty to obtain a little temporary
safety deserve neither liberty nor safety. -- Benjamin Franklin, 1759


ruleset

2009-03-24 Thread JC Putter
where can i find more rulesets? using openprotect sare rules and sought rulesets


__ Information from ESET NOD32 Antivirus, version of virus signature 
database 3957 (20090324) __

The message was checked by ESET NOD32 Antivirus.

http://www.eset.com


-- 
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.



Re: how to make a custom ruleset

2009-03-06 Thread Theo Van Dinter
Just fyi, this particular topic keeps getting raised here.  It'd be
great if people would search the list archives.  :)

One of the last times around:
http://www.nabble.com/forum/ViewPost.jtp?post=21296293&framed=y

In short, if you want to do this, write a plugin.  REs are great until
you get complicated, like doing multiple headers and comparing
captured text.

SA used to have a rule that looked for from=to, but it didn't do well
and got removed.  Some pointers in the above thread.


On Fri, Mar 6, 2009 at 2:44 PM, Mark Martinec  wrote:
> Adi,
>
>> First, it read the sender, and put it into a variable
>> Then, it check, if the recipient is the same as that variable
>> if true, then give score 3.0
>
> The trick is to let a regexp see an entire mail header section.
> Unfortunately it means we can't reuse already parsed addresses
> in From and To header fields, but need to reparse it all in a regexp.
>
> The rules below comes close, but is not exact (the TOFROM rule
> only checks the first To). Mind the line wraps, there are three
> long lines, each starting by 'header':
>
>
> header SAME_FROMTO1 ALL =~ m{^From: (?: . | \n[\ \t] )* <\s*(.+)\s*> (?s:.*) 
> ^To: (?: (?: "[^"]*" | \([^)]*\) |
> [\ \t]+ | \n[\ \t]+ )*? \1 [,(\ \t\n] | (?: . | \n[\ \t])* <\s*\1\s*>)}mix
> header SAME_FROMTO2 ALL =~ m{^From: (?: "[^"]*" | \([^)]*\) | [\ \t]+ | \n[\ 
> \t]+ )* ([^<>,;\s...@[0-9a-z._-]+\.
> [a-z]{2,})\b (?s:.*) ^To: (?: (?: "[^"]*" | \([^)]*\) | [\ \t]+ | \n[\ \t]+ 
> )*? \1 [,(\ \t\n] | (?: . | \n[\
> \t])* <\s*\1\s*>)}mix
> header SAME_TOFROM  ALL =~ m{^To: (?: . | \n[\ \t] )* (?:\b|<) 
> ([^<>,;\s...@[0-9a-z._-]+\.[a-z]{2,}) \b (?!\.)
> (?s:.*) ^From: (?: (?: "[^"]*" | \([^)]*\) | [\ \t]+ | \n[\ \t]+ )*? \1 [,(\ 
> \t\n] | (?: . | \n[\ \t])*
> <\s*\1\s*>)}mix
> meta   SAME_FROMTO  SAME_FROMTO1 || SAME_FROMTO2 || SAME_TOFROM
> score  SAME_FROMTO1 0.1
> score  SAME_FROMTO2 0.1
> score  SAME_TOFROM  0.1
> score  SAME_FROMTO  1.5
>
>
> Mark
>


Re: how to make a custom ruleset

2009-03-06 Thread Mark Martinec
Adi,

> First, it read the sender, and put it into a variable
> Then, it check, if the recipient is the same as that variable
> if true, then give score 3.0

The trick is to let a regexp see an entire mail header section.
Unfortunately it means we can't reuse already parsed addresses
in From and To header fields, but need to reparse it all in a regexp.

The rules below comes close, but is not exact (the TOFROM rule
only checks the first To). Mind the line wraps, there are three
long lines, each starting by 'header':


header SAME_FROMTO1 ALL =~ m{^From: (?: . | \n[\ \t] )* <\s*(.+)\s*> (?s:.*) 
^To: (?: (?: "[^"]*" | \([^)]*\) | 
[\ \t]+ | \n[\ \t]+ )*? \1 [,(\ \t\n] | (?: . | \n[\ \t])* <\s*\1\s*>)}mix
header SAME_FROMTO2 ALL =~ m{^From: (?: "[^"]*" | \([^)]*\) | [\ \t]+ | \n[\ 
\t]+ )* ([^<>,;\s...@[0-9a-z._-]+\.
[a-z]{2,})\b (?s:.*) ^To: (?: (?: "[^"]*" | \([^)]*\) | [\ \t]+ | \n[\ \t]+ )*? 
\1 [,(\ \t\n] | (?: . | \n[\ 
\t])* <\s*\1\s*>)}mix
header SAME_TOFROM  ALL =~ m{^To: (?: . | \n[\ \t] )* (?:\b|<) 
([^<>,;\s...@[0-9a-z._-]+\.[a-z]{2,}) \b (?!\.) 
(?s:.*) ^From: (?: (?: "[^"]*" | \([^)]*\) | [\ \t]+ | \n[\ \t]+ )*? \1 [,(\ 
\t\n] | (?: . | \n[\ \t])* 
<\s*\1\s*>)}mix
meta   SAME_FROMTO  SAME_FROMTO1 || SAME_FROMTO2 || SAME_TOFROM
score  SAME_FROMTO1 0.1
score  SAME_FROMTO2 0.1
score  SAME_TOFROM  0.1
score  SAME_FROMTO  1.5


Mark


Re: how to make a custom ruleset

2009-03-06 Thread John Hardin

On Fri, 6 Mar 2009, Adi Nugroho wrote:

It is working well, but not global (just check for my.address, and not 
for everyone).


Actually, it _is_ global, as it can only match on the domain name. Any 
mail from any user in your domain to any other user in your domain will 
hit this rule.



Please help me again to teach me, how to make it a little bit different:

First, it read the sender, and put it into a variable
Then, it check, if the recipient is the same as that variable


Unfortunately this cannot be done in SpamAssassin, and adding the ability 
to do that (have one rule depend on the actual text matched by another 
rule rather than just whether or not that rule matched _something_) is 
difficult and will likely not be done anytime soon.


This particular test - whether from == to - is something that is best done 
by something outside SA.


(1) configure your MTA to reject it at SMTP time

(2) configure your MTA to reject mail claiming to be from your domain when 
it's coming in from outside your network


(3) do something like write a procmail rule to check for that match and 
add a customer header if it is found, then have SA add points if that 
header is present (which, of course, depends on your using procmail to 
glue SA into your mail system)


--
 John Hardin KA7OHZhttp://www.impsec.org/~jhardin/
 jhar...@impsec.orgFALaholic #11174 pgpk -a jhar...@impsec.org
 key: 0xB8732E79 -- 2D8C 34F4 6411 F507 136C  AF76 D822 E6E6 B873 2E79
---
  Failure to plan ahead on someone else's part does not constitute
  an emergency on my part. -- David W. Barts in a.s.r
---
 2 days until Daylight Saving Time begins in U.S. - Spring Forward


RE: how to make a custom ruleset

2009-03-06 Thread Bowie Bailey
Adi Nugroho wrote:
> On Thursday 05 March 2009 23:44:39 Benny Pedersen wrote:
> > header SELF_FROM From =~ /\...@my.address/i
> > header SELF_TO To =~ /\...@my.address/i
> > meta SELF (SELF_FROM && SELF_TO)
> > describe SELF Trap mail with forged sender the same as recipient
> > score SELF 3.0
> 
> Finally I understand above rule.
> 
> First it check if the sender is my.address
> then it check if the recipient is my.address
> and finally, if both are true, then it give the score 3.0

Not quite.  SELF_FROM and SELF_TO default to a score of 1.  So it
actually works like this:

If the sender is my.address, score +1
If the recipient is my.address, score +1
If both are true, score +3
Final score = 5

If you don't want the sub-rules to score, you need to prefix the names
with a double-underscore (this will also prevent them from showing in
the rule summaries).

   header __SELF_FROM From =~ /\...@my.address/i
   header __SELF_TO To =~ /\...@my.address/i
   meta SELF (__SELF_FROM && __SELF_TO)
   describe SELF Trap mail with forged sender the same as recipient
   score SELF 3.0

-- 
Bowie


Re: how to make a custom ruleset

2009-03-06 Thread Adi Nugroho
On Thursday 05 March 2009 23:44:39 Benny Pedersen wrote:
> header SELF_FROM From =~ /\...@my.address/i
> header SELF_TO To =~ /\...@my.address/i
> meta SELF (SELF_FROM && SELF_TO)
> describe SELF Trap mail with forged sender the same as recipient
> score SELF 3.0

Finally I understand above rule.

First it check if the sender is my.address
then it check if the recipient is my.address
and finally, if both are true, then it give the score 3.0

Thank you.
It is working well, but not global (just check for my.address, and not for 
everyone).

Please help me again to teach me, how to make it a little bit different:

First, it read the sender, and put it into a variable
Then, it check, if the recipient is the same as that variable
if true, then give score 3.0

Thank you again :)



Re: how to make a custom ruleset

2009-03-06 Thread Kai Schaetzl
Adi Nugroho wrote on Fri, 6 Mar 2009 10:40:26 +0800:

> Is there a howto about this ruleset?

http://wiki.apache.org/spamassassin/WritingRules

Kai

-- 
Kai Schätzl, Berlin, Germany
Get your web at Conactive Internet Services: http://www.conactive.com





Re: how to make a custom ruleset

2009-03-05 Thread Adi Nugroho
On Thursday 05 March 2009 23:44:39 Benny Pedersen wrote:
> ups
>
> header SELF_FROM From =~ /\...@my.address/i
> header SELF_TO To =~ /\...@my.address/i
> meta SELF (SELF_FROM && SELF_TO)
> describe SELF Trap mail with forged sender the same as recipient
> score SELF 3.0

I have tried above syntax but failed.
No mail identified as SELF.

Is there a howto about this ruleset?



Re: how to make a custom ruleset

2009-03-05 Thread LuKreme

On Mar 5, 2009, at 7:28, Martin Gregorie  wrote:

On Thu, 2009-03-05 at 21:31 +0800, Adi Nugroho wrote:
I found that a lot of spam is using recipient email address as the  
sender.
(from a...@internux.co.id to a...@internux.co.id, or from i...@apache.org 
 to

i...@apache.org).


The only disadvantage is that you'll label test messages as spam.


If you allow address delimiters this is trivial to get around, just  
have the email their test to user+t...@example.com









Re: how to make a custom ruleset

2009-03-05 Thread Benny Pedersen

On Thu, March 5, 2009 17:31, John Hardin wrote:
>> header SELF_FROM From =~ /\...@my.address/i
>> header SELF_TO To =~ /\...@my.address/i
>
> Are you sure you want to give 1 point to each of those cases in
> addition to whatever points the meta adds?

it was not me that maked the rules, just edit them :)

> If not, then they should be named __SELF_FROM and __SELF_TO

sure

when do you stop CC me ?

-- 
http://localhost/ 100% uptime and 100% mirrored :)



  1   2   3   >