With the rise of AI writing tools, Wikipedia editors have had to deal with an onslaught of AI-generated content filled with false information and phony citations. Already, the community of Wikipedia volunteers has mobilized to fight back against AI slop, something Wikimedia Foundation product director Marshall Miller likens to a sort of âimmune systemâ response.
âThey are vigilant to make sure that the content stays neutral and reliable,â Miller says. âAs the internet changes, as things like AI appear, thatâs the immune system adapting to some kind of new challenge and figuring out how to process it.â
One way Wikipedians are sloshing through the muck is with the âspeedy deletionâ of poorly written articles, as reported earlier by 404 Media. A Wikipedia reviewer who expressed support for the rule said they are âflooded non-stop with horrendous drafts.â They add that the speedy removal âwould greatly help efforts to combat it and save countless hours picking up the junk AI leaves behind.â Another says the âlies and fake referencesâ inside AI outputs take âan incredible amount of experienced editor time to clean up.â
Typically, articles flagged for removal on Wikipedia enter a seven-day discussion period during which community members determine whether the site should delete the article. The newly adopted rule will allow Wikipedia administrators to circumvent these discussions if an article is clearly AI-generated and wasnât reviewed by the person submitting it. That means looking for three main signs:
- Writing directed toward the user, such as âHere is your Wikipedia article onâŠ,â or âI hope that helps!â
- âNonsensicalâ citations, including those with incorrect references to authors or publications.
- Non-existent references, like dead links, ISBNs with invalid checksums, or unresolvable DOIs.
These arenât the only signs of AI Wikipedians are looking out for, though. As part of the WikiProject AI Cleanup, which aims to tackle an âincreasing problem of unsourced, poorly written AI-generated content,â editors put together a list of phrases and formatting characteristics that chatbot-written articles typically exhibit.
The list goes beyond calling out the excessive use of em dashes (âââ) that have become associated with AI chatbots, and even includes an overuse of certain conjunctions, like âmoreover,â as well as promotional language, such as describing something as âbreathtaking.â There are other formatting issues the page advises Wikipedians to look out for, too, including curly quotation marks and apostrophes instead of straight ones.
However, Wikipediaâs speedy removal page notes that these characteristics âshould not, on their own, serve as the sole basisâ for determining that something has been written by AI, making it subject to removal. The speedy deletion policy isnât just for AI-generated slop content, either. The online encyclopedia also allows for the quick removal of pages that harass their subject, contain hoaxes or vandalism, or espouse âincoherent text or gibberish,â among other things.
The Wikimedia Foundation, which hosts the encyclopedia but doesnât have a hand in creating policies for the website, hasnât always seen eye-to-eye with its community of volunteers about AI. In June, the Wikimedia Foundation paused an experiment that put AI-generated summaries at the top of articles after facing backlash from the community.
Despite varying viewpoints about AI across the Wikipedia community, the Wikimedia Foundation isnât against using it as long as it results in accurate, high-quality writing.
âItâs a double-edged sword,â Miller says. âItâs causing people to be able to generate lower quality content at higher volumes, but AI can also potentially be a tool to help volunteers do their work, if we do it right and work with them to figure out the right ways to apply it.â For example, the Wikimedia Foundation already uses AI to help identify article revisions containing vandalism, and its recently-published AI strategy includes supporting editors with AI tools that will help them automate ârepetitive tasksâ and translation.
The Wikimedia Foundation is also actively developing a non-AI-powered tool called Edit Check thatâs geared toward helping new contributors fall in line with its policies and writing guidelines. Eventually, it might help ease the burden of unreviewed AI-generated submissions, too. Right now, Edit Check can remind writers to add citations if theyâve written a large amount of text without one, as well as check their tone to ensure that writers stay neutral.
The Wikimedia Foundation is also working on adding a âPaste Checkâ to the tool, which will ask users whoâve pasted a large chunk of text into an article whether theyâve actually written it. Contributors have submitted several ideas to help the Wikimedia Foundation build upon the tool as well, with one user suggesting asking suspected AI authors to specify how much was generated by a chatbot.
âWeâre following along with our communities on what they do and what they find productive,â Miller says. âFor now, our focus with using machine learning in the editing context is more on helping people make constructive edits, and also on helping people who are patrolling edits pay attention to the right ones.â
Read the full article here