>...
>To: discuss(a)lists.surbl.org
>From: Frank Ellermann <nobody(a)xyzzy.claranet.de>
>Date: Fri, 25 Mar 2005 11:51:55 +0100
>...
>
>List Mail User wrote:
>
>> gabia.com-munged
>> http://www.rfc-ignorant.org/tools/lookup.php?domain=gabia.com
>
>ACK, that's also the first time that I heard about any "ICANN
>10 years limit" for the expiration of com-domains. And here
>it's apparently the reason for a whois-RFCI-entry, which RfC
>could that be ? Not RfC 1591 or RfC 1032 (looking for "exp").
>
>> Its good to know people do read
>
>I'm a RFCI+SURBL+SPF-fan ;-) Bye, Frank
>
>
>_______________________________________________
>Discuss mailing list
>Discuss(a)lists.surbl.org
>http://lists.surbl.org/mailman/listinfo/discuss
>
Frank,
It has been a hectic couple of days and now my in-laws arrive today,
but it is not in an RFC (though it is an invalid expiration date), but in
an ICANN document. Later, I'll look for the ICANN rule document which
applies, the same holds for .net and .org domains - I don't remember if
.edu's are covered or not; It is a *very* old rule and does not hold true
for some of the country specific TLDs (and maybe not even for things like
.ws, .biz, etc. - though from bad memory I think .info's have a five year
limit) - I'll try to find at least one example of those also. If you want
to start searching yourself first, check the standard registrar agreements
on the ICANN site. The "easy" evidence is to look at how long reputable
registrars will allow a renewal for (try NetSol - they suck, but do basically
do follow the rules - also Joker and GoDaddy, both whom I think better of, go
strictly by what is allowed).
More later after I find it,
Paul Shupak
track(a)plectere.com
(lots of emails - different for SURBL and the RFCI-Discuss lists).
>...
>To: discuss(a)lists.surbl.org
>From: Frank Ellermann <nobody(a)xyzzy.claranet.de>
>Date: Wed, 23 Mar 2005 07:41:13 +0100
>...
>Subject: [SURBL-Discuss] Re: New redirector: www.nate.com
>...
>
>List Mail User wrote:
>
>> The rfci whois listing for gambia.com-munged is one of my
>> favorites, because it is the only time I have seem that
>> particular violation.
>
>What are you talking about ? `rxwhois -a gambia.com` says:
>
>| gambia.com not found at .rfc-ignorant.org or .multi.surbl.org
>| whois -h whois.abuse.netgambia.com
>| postmaster(a)gambia.com (default, no info)
>
>And there's no old RFCI entry for gambia.com (except from the
>bad day when TLD .com had no working whois server). Bye, Frank
>
>
>_______________________________________________
>Discuss mailing list
>Discuss(a)lists.surbl.org
>http://lists.surbl.org/mailman/listinfo/discuss
>
My mistake, typo and cut-and-paste made it worse.
gabia.com-munged
http://www.rfc-ignorant.org/tools/lookup.php?domain=gabia.com-munged
and
http://www.rfc-ignorant.org/tools/detail.php?domain=gabia.com-munged&submit…
(of course kill the "-munged"'s)
Thank for catching me - Its good to know people do read and check.
Bye,
Paul Shupak
track(a)plectere.com
P,S, dyslexia and bad typing strike again.
We've been in contact with the operators of a large trap which
feeds lists of exploited hosts into RBLs, inquiring if they'd be
able to provide URI domains from some of the spams they receive.
The idea is to try to find URIs that are specifically sent
through zombies and other exploited hosts on the concept that
only the worst spammers use zombies and brute force to try to go
around RBLs to deliver their spam. The trap operators are able
to extract some URI hosts for us, but for now can't afford much
more CPU than to use a PERL script calling the Email::MIME module
to grab URI domains from about 60k messages. (There's not enough
spare CPU to use a program like SpamAssassin, which would likely
have more success extracting URIs, but is much more resource
intensive.) They may be able to process up to a hundred times as
many of their messages for us (i.e. 6M a day) if this moves
forward, though even that would be only a small fraction of their
trap hits.
At my request they are including a count of the number of
appearances of each URI domain name or IP so that we can rank
them in order of frequency of appearance on the theory that the
bigger spammers may appear more often. Based on that test run
and some tweaking of the scripts on their side and ours, we got
the following table of percentiles of hits, resulting output
record counts, hits against existing SURBLs, hits against the
SURBL whitelist, and new records (i.e., in neither our black or
white lists):
100th percentile, 1293 records, 732 blacklist hits, 112 whitelist hits, 449 novel
99th percentile, 844 records, 549 blacklist hits, 81 whitelist hits, 214 novel
98th percentile, 653 records, 461 blacklist hits, 67 whitelist hits, 125 novel
97th percentile, 548 records, 397 blacklist hits, 54 whitelist hits, 97 novel
96th percentile, 481 records, 352 blacklist hits, 48 whitelist hits, 81 novel
95th percentile, 433 records, 320 blacklist hits, 42 whitelist hits, 71 novel
94th percentile, 396 records, 298 blacklist hits, 40 whitelist hits, 58 novel
93th percentile, 362 records, 287 blacklist hits, 39 whitelist hits, 36 novel
92th percentile, 332 records, 263 blacklist hits, 38 whitelist hits, 31 novel
91th percentile, 307 records, 251 blacklist hits, 29 whitelist hits, 27 novel
90th percentile, 286 records, 231 blacklist hits, 29 whitelist hits, 26 novel
89th percentile, 267 records, 218 blacklist hits, 25 whitelist hits, 24 novel
88th percentile, 250 records, 202 blacklist hits, 25 whitelist hits, 23 novel
87th percentile, 235 records, 188 blacklist hits, 25 whitelist hits, 22 novel
86th percentile, 221 records, 177 blacklist hits, 23 whitelist hits, 21 novel
85th percentile, 209 records, 170 blacklist hits, 22 whitelist hits, 17 novel
84th percentile, 197 records, 161 blacklist hits, 20 whitelist hits, 16 novel
83th percentile, 186 records, 155 blacklist hits, 18 whitelist hits, 13 novel
82th percentile, 176 records, 148 blacklist hits, 16 whitelist hits, 12 novel
81th percentile, 167 records, 140 blacklist hits, 16 whitelist hits, 11 novel
80th percentile, 159 records, 135 blacklist hits, 14 whitelist hits, 10 novel
79th percentile, 152 records, 130 blacklist hits, 13 whitelist hits, 9 novel
78th percentile, 145 records, 124 blacklist hits, 13 whitelist hits, 8 novel
77th percentile, 139 records, 118 blacklist hits, 13 whitelist hits, 8 novel
76th percentile, 133 records, 112 blacklist hits, 13 whitelist hits, 8 novel
75th percentile, 127 records, 107 blacklist hits, 12 whitelist hits, 8 novel
74th percentile, 122 records, 102 blacklist hits, 12 whitelist hits, 8 novel
73th percentile, 116 records, 98 blacklist hits, 11 whitelist hits, 7 novel
72th percentile, 112 records, 95 blacklist hits, 11 whitelist hits, 6 novel
71th percentile, 107 records, 91 blacklist hits, 11 whitelist hits, 5 novel
70th percentile, 103 records, 88 blacklist hits, 10 whitelist hits, 5 novel
For this sample, the 96th or 97th percentile appears to be an
inflection point of expectedly Zipfian-looking data. (I.e. just
a few URI hosts appear many times, and many URI hosts appear just
a few times.)
Even after whitelisting there are still a few legitimate-looking
domains coming through, so one idea would be to list the records
up to the 96th or 97th percentile, but for the remaining ones
with fewer hits, only list those that also appeared in existing
SURBLs, or resolved into sbl.spamhaus.org, or where the sending
software was clearly spamware. Hopefully that would reduce FPs
in these records with fewer hits, but still let us "pull some
useable data out of the noise" and list some of the less
frequently appearing records.
Does anyone have any comments on this? IMO what makes these data
somewhat unique is that it's an early look at the content which
exploited hosts are sending into very large traps. The benefit
is that it helps us potentially catch up to a few hundred
otherwise unlisted domains sooner, and helps reduce the
usefulness of those domains in future zombie usage, etc. In
other words it potentially improves the detection rates of SURBLs
and increases the usefulness of traps feeding traditional RBLs.
Comments?
Jeff C.
--
"If it appears in hams, then don't list it."
>...
>From: "Matthew Wilson" <matthew(a)boomer.com>
>To: "SURBL Discussion list" <discuss(a)lists.surbl.org>
>...
>Subject: [SURBL-Discuss] Zdnet redirector *still* open
>
>C'mon, why can't they get it right?
>
>Just got a spam with this URL.
>
>http://chkpt.zdnetMUNGED.com/chkpt/lovealready/bhe%2eIB%72soF%74.C%6fm
>
>
>_______________________________________________
>Discuss mailing list
>Discuss(a)lists.surbl.org
>http://lists.surbl.org/mailman/listinfo/discuss
>
ibrsoft.com-munged - not listed yet (as far as I can tell), but
part of the xml-soft.com-munged group of software pirates (check for them
and their name servers and the domain for the contacts - qdice.com-munged).
Unfortunately, their address and telephone appear valid, though incomplete.
Forward to piracy(a)microsoft.com. (and to all the cnet people also).
Paul Shupak
track(a)plectere.com
See the following link. By using "I'm Feeling Lucky", a spammer just
has to rank at the top of google's searches for *any* search, meaningful
or not.
http://www.google.com/search?sourceid=navclient&ie=UTF-8&rls=GGLC,GGLC:1
969-53,GGLC:en&q=blank+check+paper%2C+magnetic+ink+for+inkjets&btnI=I'm%
20Feeling%20Lucky
Matthew Wilson, MCSE (2003), MCSA-Messaging
Network Administrator
matthew(a)boomer.com
Boomer Consulting, Inc.
610 Humboldt
Manhattan, KS 66502
http://www.boomer.com <http://www.boomer.com/>
1-888-266-6375 x 17
John Wilcock wrote:
> Matthew Wilson wrote:
> uri local_GOOGLE_LUCKY /(?:\bgoogle\b)*&btnI=/i
Can this be right? To me it looks like it matches things like
&btnI=
google&btnI=
googlegoogle&btnI=
googlegooglegoogle&btnI=
Just a missing . maybe?
/(?:\bgoogle\b).*&btnI=/i
^ here
Matthew.van.Eerde (at) hbinc.com 805.964.4554 x902
Hispanic Business Inc./HireDiversity.com Software Engineer
perl -e"map{y/a-z/l-za-k/;print}shift" "Jjhi pcdiwtg Ptga wprztg,"
Matthew.van.Eerde wrote:
>> uri local_GOOGLE_LUCKY /(?:\bgoogle\b)*&btnI=/i
> Can this be right? To me it looks like it matches things like
>
> &btnI=
> google&btnI=
> googlegoogle&btnI=
> googlegooglegoogle&btnI=
Only the first two - I forgot that googlegoogle has no internal \b
Matthew.van.Eerde (at) hbinc.com 805.964.4554 x902
Hispanic Business Inc./HireDiversity.com Software Engineer
perl -e"map{y/a-z/l-za-k/;print}shift" "Jjhi pcdiwtg Ptga wprztg,"
John can I get your permission to add this to the SARE URI rules?
--Chris
>-----Original Message-----
>From: John Wilcock [mailto:john@tradoc.fr]
>Sent: Wednesday, March 23, 2005 11:02 AM
>To: SURBL Discussion list
>Subject: Re: [SURBL-Discuss] google is open redirector
>
>
>Matthew Wilson wrote:
>> By the way, my only suggestion to combat this is to have the surbl
>> client send an http request to google, to see what redirect site is
>> returned, and then check *that* site in SURBL or in the
>other redirects.
>> If the use of this technique picks up, google is going to have that
>> additional burden.
>
>I've added a spamassassin rule for this (see below).
>I don't expect to see many false positives, though time will tell...
>As you say,
>
>> Who really uses the "I'm Feeling Lucky" button anyway?
>
>
># 2005-03-23 new rule
>uri local_GOOGLE_LUCKY /(?:\bgoogle\b)*&btnI=/i
>describe local_GOOGLE_LUCKY Redirect through Google Feeling Lucky
>score local_GOOGLE_LUCKY 2.0
>
>
>John.
>
>--
>-- Over 2500 webcams from ski resorts around the world -
www.snoweye.com
-- Translate your technical documents and web pages - www.tradoc.fr
_______________________________________________
Discuss mailing list
Discuss(a)lists.surbl.org
http://lists.surbl.org/mailman/listinfo/discuss