Importing Google Map causes XML parse error

classic Classic list List threaded Threaded
32 messages Options
12
Reply | Threaded
Open this post in threaded view
|

Importing Google Map causes XML parse error

Paul Coghlan-3
I am trying to import a few thousand pages that include maps using the
Google Maps extension.

The text is in the format as below:

<googlemap lat="36.3015" lon="-76.21976" type="map" zoom="11" width="700"
height="600">
36.3214,-76.2134, Click to see [[some text]] page
36.9876,-76.1234, Click to see [[some text]] page
</googlemap>

The message I get is pretty much identical whether I use importDump.php or
Special:Import and is:

XML import Failure at line 40, col 42 (byte 1725; ³
36.3214,-76²):Attribute redefined

Anyone able to shed any light on this?

Many thanks,
Paul


_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

chinese spam

2007@gmask.com
I'm still getting spammed with what appears to be chinese spam. Can I
use the regex filter to block all non western characters, or
specifically chinese ones?

-adrian



_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

jibberish

2007@gmask.com
There is a bot that keeps posting random words or alphanumeric
sequences to the beginning of pages on my wiki.

Is it possible to warn users that post under a certain of characters to
the beginning of a page?

I don't want to enforce a captcha on every edit.

-Adrian

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Benjamin Horst
I was experiencing the same types of "spam," or whatever it is.

I started to enforce a captcha on every edit, which bothers me, but stopped the junk. I hope there's a better solution out there!

Thanks,
Ben

----
OpenOffice and open source blog:
http://www.solidoffice.com/

Wiki business directory:
http://www.wikipages.com/

 
On Thursday, October 11, 2007, at 01:04PM, "[hidden email]" <[hidden email]> wrote:

>There is a bot that keeps posting random words or alphanumeric
>sequences to the beginning of pages on my wiki.
>
>Is it possible to warn users that post under a certain of characters to
>the beginning of a page?
>
>I don't want to enforce a captcha on every edit.
>
>-Adrian
>
>_______________________________________________
>MediaWiki-l mailing list
>[hidden email]
>http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>
>

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: chinese spam

Nelson A Li
In reply to this post by 2007@gmask.com
Hello,

If you can attach a snap of "what appears to be chinese spam", I can help
you identify if it is chinese and what does it say if it is in chinese.

Nelson

Computer Sciences Corporation
Registered Office: 2100 East Grand Avenue, El Segundo California 90245, USA
Registered in USA No: C-489-59

----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

This is a PRIVATE message. If you are not the intended recipient, please
delete without copying and kindly advise us by e-mail of the mistake in
delivery.
NOTE: Regardless of content, this e-mail shall not operate to bind CSC to
any order or other contract unless pursuant to explicit written agreement
or government initiative expressly permitting the use of e-mail for such
purpose.
----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------




                                                                           
             "[hidden email]"                                              
             <[hidden email]>                                              
             Sent by:                                                   To
             mediawiki-l-bounc         [hidden email]    
             [hidden email]                                          cc
             a.org                                                        
                                                                   Subject
                                       [Mediawiki-l] chinese spam          
             10/11/2007 12:09                                              
             PM                                                            
                                                                           
                                                                           
             Please respond to                                            
              [hidden email];                                              
             Please respond to                                            
                 MediaWiki                                                
             announcements and                                            
              site admin list                                              
             <mediawiki-l@list                                            
             s.wikimedia.org>                                              
                                                                           
                                                                           




I'm still getting spammed with what appears to be chinese spam. Can I
use the regex filter to block all non western characters, or
specifically chinese ones?

-adrian



_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l



_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: chinese spam

jidanni
NAL> Can I use the regex filter to block all non western characters
Yes. [^[:ascii:]] matches any non-ASCII character.

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: chinese spam

2007@gmask.com
Hmm... that doesn't seem to block the simplified chinese characters.

--- [hidden email] wrote:

> NAL> Can I use the regex filter to block all non western characters
> Yes. [^[:ascii:]] matches any non-ASCII character.
>
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>


_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Chuck-45
In reply to this post by Benjamin Horst
Benjamin Horst wrote:
> I was experiencing the same types of "spam," or whatever it is.
>
> I started to enforce a captcha on every edit, which bothers me, but stopped the junk. I hope there's a better solution out there!
>
> Thanks,
> Ben

We're having the same problem with our wikis. It seems that this could
be solved if thereis a switch in MediaWiki that really mandates that
changes be made by registered users.

We are planning to implement the other spam measures mentioned on this list.

Thanks to the advice from last week about how to stop DIV spam.

Chuck

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

2007@gmask.com
Yea I don't want to stop anonymous users but it seems like that might
be neccessary.. or it would be great if you could captcha new posts
from either new users or unfamiliar IP's.

-Adrian


--- Chuck <[hidden email]> wrote:

> Benjamin Horst wrote:
> > I was experiencing the same types of "spam," or whatever it is.
> >
> > I started to enforce a captcha on every edit, which bothers me, but
> stopped the junk. I hope there's a better solution out there!
> >
> > Thanks,
> > Ben
>
> We're having the same problem with our wikis. It seems that this
> could
> be solved if thereis a switch in MediaWiki that really mandates that
> changes be made by registered users.
>
> We are planning to implement the other spam measures mentioned on
> this list.
>
> Thanks to the advice from last week about how to stop DIV spam.
>
> Chuck
>
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>




_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

another, but valid, delete user question

admin-281
A couple of days ago I was also hit by this, or a similar, spambot.

I know the question of deleting users has been brought up before and  
nixed for valid db reasons. In this case, however, it's different.  
Looking at the user table on my wiki, a bot has been registering  
users for the past 6 months, every few minutes. The format of the  
usernames is consistent: in the beginning the usernames were 6  
characters ([a-zA-Z0-9]) with the first and fourth characters capital  
letters. For a while in July the usernames changed to eight  
characters, and then recently to 10 characters, with the first and  
sixth capital letters.

This started *exactly* six months to the day that the spam started  
(20070410 to 20071010).

Of the 17000 registered users on my site, over 10000 are spambot  
registrations. We monitor edits pretty closely (though obviously not  
registrations, which will change), so of those 10000 spambot users,  
no more than 20 have actually edited a page, all starting on Oct  
10th. (We caught it prett quickly.)

So it seems as if I should be able to feed my wiki a list of user and  
have it delete those users, if, as in this case, I am *absolutely*  
sure none of these users have edited pages.

So then, the question. If there's not a maintenance script somewhere  
that reads a user_name or user_id list and zaps them, then is there a  
danger in deleting those user records from the user table if I *know*  
they've made no edits?

thanks




On Oct 12, 2007, at 1:19 PM, [hidden email] wrote:

> Yea I don't want to stop anonymous users but it seems like that might
> be neccessary.. or it would be great if you could captcha new posts
> from either new users or unfamiliar IP's.
>
> -Adrian
>
>
> --- Chuck <[hidden email]> wrote:
>
>> Benjamin Horst wrote:
>>> I was experiencing the same types of "spam," or whatever it is.
>>>
>>> I started to enforce a captcha on every edit, which bothers me, but
>> stopped the junk. I hope there's a better solution out there!
>>>
>>> Thanks,
>>> Ben
>>
>> We're having the same problem with our wikis. It seems that this
>> could
>> be solved if thereis a switch in MediaWiki that really mandates that
>> changes be made by registered users.
>>
>> We are planning to implement the other spam measures mentioned on
>> this list.
>>
>> Thanks to the advice from last week about how to stop DIV spam.
>>
>> Chuck
>>
>> _______________________________________________
>> MediaWiki-l mailing list
>> [hidden email]
>> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>>
>
>
>
>
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>
>


_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: another, but valid, delete user question

Michael Daly-3
dKosopedia admin wrote:

> Of the 17000 registered users on my site, over 10000 are spambot  
> registrations. We monitor edits pretty closely (though obviously not  
> registrations, which will change), so of those 10000 spambot users,  
> no more than 20 have actually edited a page, all starting on Oct  
> 10th. (We caught it prett quickly.)
>
> So it seems as if I should be able to feed my wiki a list of user and  
> have it delete those users, if, as in this case, I am *absolutely*  
> sure none of these users have edited pages.

I know it's not a solution, but since I've seen a few edits on my wiki
in such cases, I zapped all users with the FooooBaarr name format.  I
did a MySQL query to grab all the 10 character user IDs where the user
ID and the real name were the same (and where it wasn't the one real
user who just happened to get caught with this same general format
name).  With these IDs, I created a script to run changePassword.php (in
the /wiki/maintenance/ directory) for every such user.  This replaced
all the passwords with one I know so these bots can't sign in.  I then
used SQL to rename the users to zzFooooBaarr so they sort to the bottom
in the User List page.

Still there, but unusable by the spammers and not interfering with
regular users.

I put the reCaptcha extension in place to try to defeat the bots from
making new user IDs.

Mike




_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: Importing Google Map causes XML parse error

Brion Vibber-3
In reply to this post by Paul Coghlan-3
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1

Paul Coghlan wrote:

> I am trying to import a few thousand pages that include maps using the
> Google Maps extension.
>
> The text is in the format as below:
>
> <googlemap lat="36.3015" lon="-76.21976" type="map" zoom="11" width="700"
> height="600">
> 36.3214,-76.2134, Click to see [[some text]] page
> 36.9876,-76.1234, Click to see [[some text]] page
> </googlemap>
>
> The message I get is pretty much identical whether I use importDump.php or
> Special:Import and is:
>
> XML import Failure at line 40, col 42 (byte 1725; ³
> 36.3214,-76²):Attribute redefined
>
> Anyone able to shed any light on this?

Can you confirm that your XML is properly formatted? Remember the page
text is *text*, so '<' '>' etc should appear as '&lt;' '&gt;' etc in
your .xml file.

- -- brion vibber (brion @ wikimedia.org)
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.6 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFHE3IawRnhpk1wk44RAlz6AJ4ySGtNLwIrlsaHNO3MEeAmIHU17gCdGvFO
jbY+GimS9yMX01q3bQAJGx8=
=jp5o
-----END PGP SIGNATURE-----

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Chuck-45
In reply to this post by 2007@gmask.com
Has anybody found a solution for the gibberish spam short of installing
captcha extensions?

Chuck

[hidden email] wrote:

> Yea I don't want to stop anonymous users but it seems like that might
> be neccessary.. or it would be great if you could captcha new posts
> from either new users or unfamiliar IP's.
>
> -Adrian
>
>
> --- Chuck <[hidden email]> wrote:
>
>
>>Benjamin Horst wrote:
>>
>>>I was experiencing the same types of "spam," or whatever it is.
>>>
>>>I started to enforce a captcha on every edit, which bothers me, but
>>
>>stopped the junk. I hope there's a better solution out there!
>>
>>>Thanks,
>>>Ben
>>
>>We're having the same problem with our wikis. It seems that this
>>could
>>be solved if thereis a switch in MediaWiki that really mandates that
>>changes be made by registered users.
>>
>>We are planning to implement the other spam measures mentioned on
>>this list.
>>
>>Thanks to the advice from last week about how to stop DIV spam.
>>
>>Chuck
>>
>>_______________________________________________
>>MediaWiki-l mailing list
>>[hidden email]
>>http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>>
>
>
>
>
>
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>
>


--

--------------------------
Bread and Roses Web Design
serving small businesses, non-profits, artists and activists
http://www.breadandrosesweb.com/

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: chinese spam

jidanni
In reply to this post by 2007@gmask.com
ASCII spam fighter for Chinese sites:
$wgSpamRegex='/^[[:ascii:]]*$/';#Must have at least one non-ASCII to post.
Chinese spam fighter for ASCII sites:
$wgSpamRegex='/[^[:ascii:]]/';#Must be pure ASCII to post.
> Hmm... that doesn't seem to block the simplified chinese characters.
Maybe they used "&#...;" HTML Entities? Well then add a regexp to
catch entities or whatever they are using too perhaps?

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Dan Bolser-3
In reply to this post by Chuck-45
On 16/10/2007, Chuck <[hidden email]> wrote:
> Has anybody found a solution for the gibberish spam short of installing
> captcha extensions?

Not here. I set anonymous edits to false and installed reCaptcha.

Installing reCaptcha doesn't really constitute a big change to the
wiki (its quite unobtrusive really), but disallowing anonymous edits
is a pain. This is especially true of pre 1.11.0 MW versions where
'view source' doesn't seem to work 'out of the box'.

In theory there should be a simple SQL query to detect these kinds of
spam (one nonsense word at the start of a page) - However, it seems
better to code a general solution that highlights potential spam for
review. Its keeping track of the potentially spammed pages that I find
most difficult.

Anyone handy with Bayesian filters? If we could rank edits by
'spaminess' using a Bayesian filter, and be given the option to review
the top n most spammy revisions (with feedback training) ... well...
that would be great!

Send all your edits to a gmail account and only allow those that get
forwarded back?


> Chuck
>
> [hidden email] wrote:
> > Yea I don't want to stop anonymous users but it seems like that might
> > be neccessary.. or it would be great if you could captcha new posts
> > from either new users or unfamiliar IP's.
> >
> > -Adrian
> >
> >
> > --- Chuck <[hidden email]> wrote:
> >
> >
> >>Benjamin Horst wrote:
> >>
> >>>I was experiencing the same types of "spam," or whatever it is.
> >>>
> >>>I started to enforce a captcha on every edit, which bothers me, but
> >>
> >>stopped the junk. I hope there's a better solution out there!
> >>
> >>>Thanks,
> >>>Ben
> >>
> >>We're having the same problem with our wikis. It seems that this
> >>could
> >>be solved if thereis a switch in MediaWiki that really mandates that
> >>changes be made by registered users.
> >>
> >>We are planning to implement the other spam measures mentioned on
> >>this list.
> >>
> >>Thanks to the advice from last week about how to stop DIV spam.
> >>
> >>Chuck
> >>
> >>_______________________________________________
> >>MediaWiki-l mailing list
> >>[hidden email]
> >>http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
> >>
> >
> >
> >
> >
> >
> > _______________________________________________
> > MediaWiki-l mailing list
> > [hidden email]
> > http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
> >
> >
>
>
> --
>
> --------------------------
> Bread and Roses Web Design
> serving small businesses, non-profits, artists and activists
> http://www.breadandrosesweb.com/
>
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>


--
hello

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Eric K-2
I think making an IP do a captcha on its first edit only would help. The captcha would keep a record of the most recent IP's in a table and if an edit hasnt been recorded from them, give them a captcha, otherwise pass. This may mean the IP table would grow large (may not be fast for large wikis with lots of editing), but purge it a certain period (15 days etc). Captcha has to be there one way or the other. This is least irritating.
  There is definitely no way to check if an edit is spam or not, except for capthcha.
 

Dan Bolser <[hidden email]> wrote:
  On 16/10/2007, Chuck wrote:
> Has anybody found a solution for the gibberish spam short of installing
> captcha extensions?

Not here. I set anonymous edits to false and installed reCaptcha.

Installing reCaptcha doesn't really constitute a big change to the
wiki (its quite unobtrusive really), but disallowing anonymous edits
is a pain. This is especially true of pre 1.11.0 MW versions where
'view source' doesn't seem to work 'out of the box'.

In theory there should be a simple SQL query to detect these kinds of
spam (one nonsense word at the start of a page) - However, it seems
better to code a general solution that highlights potential spam for
review. Its keeping track of the potentially spammed pages that I find
most difficult.

Anyone handy with Bayesian filters? If we could rank edits by
'spaminess' using a Bayesian filter, and be given the option to review
the top n most spammy revisions (with feedback training) ... well...
that would be great!

Send all your edits to a gmail account and only allow those that get
forwarded back?


> Chuck
>
> [hidden email] wrote:
> > Yea I don't want to stop anonymous users but it seems like that might
> > be neccessary.. or it would be great if you could captcha new posts
> > from either new users or unfamiliar IP's.
> >
> > -Adrian
> >
> >
> > --- Chuck wrote:
> >
> >
> >>Benjamin Horst wrote:
> >>
> >>>I was experiencing the same types of "spam," or whatever it is.
> >>>
> >>>I started to enforce a captcha on every edit, which bothers me, but
> >>
> >>stopped the junk. I hope there's a better solution out there!
> >>
> >>>Thanks,
> >>>Ben
> >>
> >>We're having the same problem with our wikis. It seems that this
> >>could
> >>be solved if thereis a switch in MediaWiki that really mandates that
> >>changes be made by registered users.
> >>
> >>We are planning to implement the other spam measures mentioned on
> >>this list.
> >>
> >>Thanks to the advice from last week about how to stop DIV spam.
> >>
> >>Chuck
> >>
> >>_______________________________________________
> >>MediaWiki-l mailing list
> >>[hidden email]
> >>http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
> >>
> >
> >
> >
> >
> >
> > _______________________________________________
> > MediaWiki-l mailing list
> > [hidden email]
> > http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
> >
> >
>
>
> --
>
> --------------------------
> Bread and Roses Web Design
> serving small businesses, non-profits, artists and activists
> http://www.breadandrosesweb.com/
>
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>


--
hello

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l


       
---------------------------------
Don't let your dream ride pass you by.    Make it a reality with Yahoo! Autos.
_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Dan Bolser-3
On 16/10/2007, Eric K <[hidden email]> wrote:
> I think making an IP do a captcha on its first edit only would help. The captcha would keep a record of the most recent IP's in a table and if an edit hasnt been recorded from them, give them a captcha, otherwise pass. This may mean the IP table would grow large (may not be fast for large wikis with lots of editing), but purge it a certain period (15 days etc). Captcha has to be there one way or the other. This is least irritating.

I like the idea.


>   There is definitely no way to check if an edit is spam or not, except for capthcha.

Not currently. I think a 'review for spam' feature would work very
well for most small sites.



>
> Dan Bolser <[hidden email]> wrote:
>   On 16/10/2007, Chuck wrote:
> > Has anybody found a solution for the gibberish spam short of installing
> > captcha extensions?
>
> Not here. I set anonymous edits to false and installed reCaptcha.
>
> Installing reCaptcha doesn't really constitute a big change to the
> wiki (its quite unobtrusive really), but disallowing anonymous edits
> is a pain. This is especially true of pre 1.11.0 MW versions where
> 'view source' doesn't seem to work 'out of the box'.
>
> In theory there should be a simple SQL query to detect these kinds of
> spam (one nonsense word at the start of a page) - However, it seems
> better to code a general solution that highlights potential spam for
> review. Its keeping track of the potentially spammed pages that I find
> most difficult.
>
> Anyone handy with Bayesian filters? If we could rank edits by
> 'spaminess' using a Bayesian filter, and be given the option to review
> the top n most spammy revisions (with feedback training) ... well...
> that would be great!
>
> Send all your edits to a gmail account and only allow those that get
> forwarded back?
>
>
> > Chuck
> >
> > [hidden email] wrote:
> > > Yea I don't want to stop anonymous users but it seems like that might
> > > be neccessary.. or it would be great if you could captcha new posts
> > > from either new users or unfamiliar IP's.
> > >
> > > -Adrian
> > >
> > >
> > > --- Chuck wrote:
> > >
> > >
> > >>Benjamin Horst wrote:
> > >>
> > >>>I was experiencing the same types of "spam," or whatever it is.
> > >>>
> > >>>I started to enforce a captcha on every edit, which bothers me, but
> > >>
> > >>stopped the junk. I hope there's a better solution out there!
> > >>
> > >>>Thanks,
> > >>>Ben
> > >>
> > >>We're having the same problem with our wikis. It seems that this
> > >>could
> > >>be solved if thereis a switch in MediaWiki that really mandates that
> > >>changes be made by registered users.
> > >>
> > >>We are planning to implement the other spam measures mentioned on
> > >>this list.
> > >>
> > >>Thanks to the advice from last week about how to stop DIV spam.
> > >>
> > >>Chuck
> > >>
> > >>_______________________________________________
> > >>MediaWiki-l mailing list
> > >>[hidden email]
> > >>http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
> > >>
> > >
> > >
> > >
> > >
> > >
> > > _______________________________________________
> > > MediaWiki-l mailing list
> > > [hidden email]
> > > http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
> > >
> > >
> >
> >
> > --
> >
> > --------------------------
> > Bread and Roses Web Design
> > serving small businesses, non-profits, artists and activists
> > http://www.breadandrosesweb.com/
> >
> > _______________________________________________
> > MediaWiki-l mailing list
> > [hidden email]
> > http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
> >
>
>
> --
> hello
>
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>
>
>
> ---------------------------------
> Don't let your dream ride pass you by.    Make it a reality with Yahoo! Autos.
> _______________________________________________
> MediaWiki-l mailing list
> [hidden email]
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>


--
hello

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Chuck-45
In reply to this post by Dan Bolser-3
This gibberish spam doesn't make much sense, pardon the pun. The spambot
isn't inserting any actual links. My wikis are getting spammed with
short text strings like "copasnotra" and "romonboel". Based on my
limited understanding of spambots, it seems like the bots are making
these changes as a prelude to doing something else.

After some further investigation, some interesting clues emerge. This
"gibberish spambot" is evidently generating fake user accounts. I
deleted hundreds of fake accounts last night from the four wiki
databases that we run. The spambot is surprisingly doing something that
should make it easy to stop them: all of their fake user accounts
include an email address from the ".ru" domain. The user names are all
different, but the spambot only uses a limited number of fake email
addresses from the .ru domain. Would it be possible to reject user
registrations with code that rejects anything from a certain domain?

Another facet of this problem is that this spambot is using proxy ISPs
or rotating fake IP addresses. In my experience, this is a common method
that spambots use to defeat easy anti-spam measures like server level IP
blocking.

Now that I think about it, I may have thwarted the final stage of this
bot's activities by implementing that spam hack that stops hidden DIV
spam. But our wikis are still getting hit hard by the "gibberish spam".
It's unclear if the hidden DIV spam and the gibberish spam are part of
the same spambots suite of attacks.

Chuck

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

Rob Church
In reply to this post by Eric K-2
On 16/10/2007, Eric K <[hidden email]> wrote:
>   There is definitely no way to check if an edit is spam or not, except for capthcha.

I have to point out the flaw in that statement, tenuous thought it is
- a CAPTCHA does *not* constitute an anti-spam acid test; all it does
is confirms that, to the best of the test's ability (which might not
count for anything), that we are dealing with a human being, rather
than an automated program.

A human could quite well post spam to his/her heart's content, and
would be able to pass a CAPTCHA (we hope). The default configuration
settings for ConfirmEdit, which CAPTCHA extensions are based upon,
allow registered users to skip these tests, so in theory, one could
set up a spam bot with a few minutes of initial human assistance,
which is why we supplement such things with throttles, "heuristics"
(regular expressions aren't that great in terms of configurability,
but I cling to the hope that one day we'll have decent spam-edit
detection heuristics, even if just for the basics).


Rob Church

_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
Reply | Threaded
Open this post in threaded view
|

Re: jibberish

2007@gmask.com
In reply to this post by Chuck-45

--- Chuck <[hidden email]> wrote:

> This gibberish spam doesn't make much sense, pardon the pun. The
> spambot
> isn't inserting any actual links. My wikis are getting spammed with
> short text strings like "copasnotra" and "romonboel". Based on my
> limited understanding of spambots, it seems like the bots are making
> these changes as a prelude to doing something else.

This is what is happening to me as well.. but the inserted words are
allways at the beginning of the page which gives me hope in blocking
these types of bot edits with a regex.


-adrian


_______________________________________________
MediaWiki-l mailing list
[hidden email]
http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
12