WikiIndex:Spam Control Policy

wikis, wiki people and wiki ideas
Jump to: navigation, search

For some thoughts about spambot hunting see WikiProject:Junking bots ...didn't know where to lonk it --Wolf | talk 07:05, 13 October 2007 (EDT)

Proposed spam control policy[edit]

Comments and corrections welcome.

There are three levels of spam control now in play, and this policy will as guidelines to address how each of them should be used.

Level 1 - LocalSettings.php[edit]

There is a regex filter in the LocalSettings.php file that is under the control of the site bureaucrats. This level blocks specific words, phrases, and html fragments – that are commonly used by link spammers and vandals. It contains common curse words, sex acts and symbols, body parts, most of the major drug names, and html fragments that are used to hide and/or mask link spam and graffiti.

These are the common denominators of 90% of spam and graffiti. The regex will match any of these items and block the save of a page that has any one of these items anywhere on it.

It is NOT necessary to block anything containing these words anywhere else.

Level 2 - Wikimedia Meta-Wiki spam blacklist[edit]

Wikimedia Meta-Wiki maintains an active blacklist of link spam target sites and common phrases that are used in link spam. We are connected to this list, and every page save checks this list for bad external links on the list. The blacklist only checks URLs – not the complete content. The blacklist itself can be seen at Wikimedia Blacklist, and additions to it can be suggested at the blacklist talk page. You can expect this list to contain most of the current major link spam offenders targets. This list is very actively maintained, and should raise our blocking level to about 99.5% of link spam. Is this 99.5% figure still accurate? Sean, aka Hoof HeartedAdmintalk2HH 02:45, 9 March 2013 (PST)

If you find a particular link is not being caught by Levels 1 or 2 you can:

  1. If it is link spam or graffiti that contains a word that should be in the level 1 list, submit it to the site administrators via a new message on their talk page;
  2. If it is a link that you think should be banned at all wikis, submit it to Wikimedia Meta-Wiki;
  3. If it is a link that you think should be banned from just WikiIndex, go to Level 3.

Level 3 - Local blacklist[edit]

We maintain a local blacklist at My spam blacklist. This is protected page that Sysops can use to block offending link spam not caught by Level 1 and Level 2. There should be very few entries here, and NONE that contain the following:

  • Periods "." - periods, aka the 'full stop' have a special meaning in the regex syntax, and can cause the list to malfunction;
  • Tlds "com, org, net" - these appear in all URLs, so provide no value to the blocking mechanism;
  • "http://www." - the regex only checks valid URLs, so this is not necessary.

An example:

If you want to block linking to you should only enter 'mybadwordsite'

If Level 1 or Level 2 already contain the 'bad word' – then the link would be blocked already, and no entry would be necessary, and you would not be able to save the list.

Level 3.5 - CAPTCHA[edit]

Now implemented. By default, CAPTCHAs are triggered on the following events:

  • New user registration;
  • Anonymous, or IP editor edits that contain new external links;
  • Brute-force password cracking.

Level 4 - Login to edit[edit]

Not implemented.

There are two options at this level:

  1. Require a login to edit, and request an e-mail confirmation;
  2. Require a login to edit, and require an e-mail confirmation before editing is allowed.

Guidance for Spam Fighters[edit]

Ward Cunningham gave me this advice for spam fighting and keeping your sanity, "do the absolute minimum required to block each attack and the spammer will grow tired and leave" (I'm paraphrasing). This is so true, because you can drive yourself crazy trying to think of a way to defeat all attacks in advance of their actually happening!

Spam Blacklist Regex[edit]

According to the readme for the MediaWiki spam blacklist extension, internally a single giant regular expression is formed using the lines from the blacklist file as follows:

In simple terms:

  • Everything from a "#" character to the end of the line is a comment;
  • Every non-blank line is a regex fragment which will only match inside URLs.

Internally, a regex is formed which looks like this:

   !http://[a-z0-9\-.]*(line 1|line 2|line 3|....)!Si

A few notes about this format. It is not necessary to add www to the start of hostnames, the regex is designed to match any subdomain. Do not add patterns to your file which may run off the end of the URL, e.g. anything containing ".*". Unlike in some similar systems, the line-end metacharacter "$" will not assert the end of the hostname, it'll assert the end of the page.

What happens if spam slips through the automated systems[edit]

Please delete any spam that slips through the automated systems, and add the {{spammer}} tag on the spammers user page. If a new page has been created with purely spam, edit the page by deleting said spam, and highlight the page for deletion by adding the {{delete}} tag, ideally by using {{delete|spam}}. If you are a Sysop, please block the spammer in accordance with the WikiIndex:Blocking and banning policy.