Wikipedia talk:WikiProject AI Cleanup
| Main page | Discussion | Noticeboard | Guide | Resources | Policies | Research |
| This is the talk page for discussing WikiProject AI Cleanup and anything related to its purposes and tasks. To report issues with AI use, please refer to Wikipedia:WikiProject AI Cleanup/Noticeboard. |
|
| Archives: 1, 2, 3, 4Auto-archiving period: 30 days |
| This project page does not require a rating on Wikipedia's content assessment scale. It is of interest to the following WikiProjects: | ||||||||
| ||||||||
| To help centralize discussions and keep related topics together, all non-archive subpages of this talk page redirect here. |
This page has been mentioned by multiple media organizations:
|
Format for the noticeboard
[edit]Could we create a template for a collapsible table similar to Fifteen thousand two hundred twenty four's comment here so people can keep track of work done and work still to do (by we I mean someone else that knows how). On the user side it could look something like {{AIC table|collapsed=yes|user=Steven|article1=Fishcakes|clean=y|g15=n|article2=History of Fishcakes|clean=n|g15=y}}. Most efficient, if possible, would probably be to have the default being no for "clean" and "g15" parameters, and it only needing a yes in one of those per article entry where the other parameter/box gets greyed-out. Would people find that useful? Kowal2701 (talk) 21:51, 29 October 2025 (UTC)
- Many cases require cleanup across hundreds of articles and I don't think this is viable at that scale. I also have some feedback on the columns in this table. I often don't tag articles associated with an AINB case and just fix them instead. I have cleaned hundreds of articles and I don't think I have ever G15'd an article associated with an open case at AINB - other editors not associated with the project typically get to those first (often via AfC or NPP). When I G15 its normally by finding stuff via the edit filter logs. My actions while cleaning are typically: revert, stubify, rewrite, tag and leave, or (rarely) AfD. NicheSports (talk) 22:01, 29 October 2025 (UTC)
- That said I would love a way to both clerk and track our cases better. I have been meaning to look at CCI to see how they do it there NicheSports (talk) 22:04, 29 October 2025 (UTC)
- Agreed, and I think a table is a reasonable way to track cleanup progress. Small cleanup efforts could use a table in situ, larger efforts could be hosted on a subpage with a hatnote added to the top of the report linking to the subpage and indicating if cleanup has been completed or not. An example LLMN discussion might look like: Mass LLM misuse by User:Example
- Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis no...
- The CCI process looks too formal for the way reports here and at LLMN have worked, ideally tracking efforts should be low-friction and kept tightly coupled to the original reports. fifteen thousand two hundred twenty four (talk) 23:31, 29 October 2025 (UTC)
- This looks great, as there doesn't seem to be opposition to it, I am happy to implement it! Chaotic Enby (talk · contribs) 21:26, 10 November 2025 (UTC)
- I agree this would be helpful. I'd like us to pick a standard set of icons to be used for different situations, but that can come later? NicheSports (talk) 23:26, 10 November 2025 (UTC)
- Feel free for the subpage-pointing hatnote, I think that could be useful no matter what, but I do have some additional thoughts that might be worth considering first.
- I'm re-mulling over how to format tracking lists. I think ease of use in accessing, editing, and reading is paramount. A simple start-collapsed {{AIC table}} like Kowal2701 suggested would almost certainly be best for small cleanup tasks, and it could ideally take the spot of a section hatnote for increased visibility, instead of leaving it buried down in a thread.
- For larger task tracking on a subpage, there are three options I can think of for formatting:
- Plain tables: A hassle to edit, easy to read and sort.
- Bulleted lists: Dead simple to edit, OK to read, can get unwieldy the more info each entry needs to convey.
- Template plus subtemplate-based tables, like {{sat}} and {{sa}}: Has the potential to be OK to edit and easy to read, but larger pages would likely run into technical issues.
- I'm now inclined to think that bulleted lists (à la CCI), possibly used in conjunction with sections that can act as categories, might be the best choice for larger tracking jobs. fifteen thousand two hundred twenty four (talk) 05:26, 11 November 2025 (UTC)
- I don't think tables will help for any cleanup case, no matter the size. Just too much overhead and unnecessary process. A bullet point list could work for smaller cases but I'm still unconvinced. I wouldn't want to use either - I prefer to get in a zone and crank out the articles, leaving occasional updates on the case about where I'm at. The sub-page will be very helpful however (for cases with 10+ edits, which is almost all). Can we start there and see how it goes? NicheSports (talk) 05:39, 11 November 2025 (UTC)
- No complaints with starting with simple subpages containing bulleted lists.
- I've come around to thinking plain tables are too fiddly also, and am I right in reading your comment that you're against a template-based solution like an {{AIC table}} also? fifteen thousand two hundred twenty four (talk) 05:53, 11 November 2025 (UTC)
- Yes - against all types of tables. Pretty sure they'll be a net negative even with an add-in to help create them. We should start with pages and figure out over time what works best in terms of how to structure the information within them. NicheSports (talk) 06:01, 11 November 2025 (UTC)
- Understood, and you're very likely right. fifteen thousand two hundred twenty four (talk) 06:04, 11 November 2025 (UTC)
- Looking forward to the subpages. Thank you for suggesting it! NicheSports (talk) 06:10, 11 November 2025 (UTC)
- Starting something at {{AIC status}}! Chaotic Enby (talk · contribs) 18:28, 11 November 2025 (UTC)
- Looks pretty dank, danke. In addition to your 3 statuses I was also considering a 4th: "cleanup required". Kind of like a "case accepted" status. This would hopefully entice people to work on cases that no one is handling. I've only been tagging threads at AINB as in progress that I'm actively working on, for this reason. What do you think? NicheSports (talk) 19:20, 11 November 2025 (UTC)
- That works for me! I'm also working on subpage integration as we speak, so the number of active pages to clean up can be retrieved directly from there. Chaotic Enby (talk · contribs) 19:37, 11 November 2025 (UTC)
- Looking good! Interested to see how the subpage tracking and construction will work.
- I've started a discussion at Template talk:AIC status#Accessibility with some ideas on how to make the template easier for anyone, even those who've never read the docs, to use. Would love to hear any feedback. fifteen thousand two hundred twenty four (talk) 21:57, 11 November 2025 (UTC)
- That works for me! I'm also working on subpage integration as we speak, so the number of active pages to clean up can be retrieved directly from there. Chaotic Enby (talk · contribs) 19:37, 11 November 2025 (UTC)
- Looks pretty dank, danke. In addition to your 3 statuses I was also considering a 4th: "cleanup required". Kind of like a "case accepted" status. This would hopefully entice people to work on cases that no one is handling. I've only been tagging threads at AINB as in progress that I'm actively working on, for this reason. What do you think? NicheSports (talk) 19:20, 11 November 2025 (UTC)
- Starting something at {{AIC status}}! Chaotic Enby (talk · contribs) 18:28, 11 November 2025 (UTC)
- Looking forward to the subpages. Thank you for suggesting it! NicheSports (talk) 06:10, 11 November 2025 (UTC)
- Understood, and you're very likely right. fifteen thousand two hundred twenty four (talk) 06:04, 11 November 2025 (UTC)
- Yes - against all types of tables. Pretty sure they'll be a net negative even with an add-in to help create them. We should start with pages and figure out over time what works best in terms of how to structure the information within them. NicheSports (talk) 06:01, 11 November 2025 (UTC)
- I don't think tables will help for any cleanup case, no matter the size. Just too much overhead and unnecessary process. A bullet point list could work for smaller cases but I'm still unconvinced. I wouldn't want to use either - I prefer to get in a zone and crank out the articles, leaving occasional updates on the case about where I'm at. The sub-page will be very helpful however (for cases with 10+ edits, which is almost all). Can we start there and see how it goes? NicheSports (talk) 05:39, 11 November 2025 (UTC)
- All is done, and we even got a tracking category at Category:WikiProject AI Cleanup open cases to keep track of subpages where there is work left to do! And yes, sorting is automated! Chaotic Enby (talk · contribs) 00:32, 13 November 2025 (UTC)
- This looks great, as there doesn't seem to be opposition to it, I am happy to implement it! Chaotic Enby (talk · contribs) 21:26, 10 November 2025 (UTC)
- Yeah I've just been tagging stuff if it isn't something I can revert without splash damage -- I assume a lot of those articles have been quietly untagged but I haven't been keeping track because there is so, so much and I just don't have the patience to argue over them all. Open to a more formalized tracking system though. Gnomingstuff (talk) 02:18, 30 October 2025 (UTC)
- Agreed, and I think a table is a reasonable way to track cleanup progress. Small cleanup efforts could use a table in situ, larger efforts could be hosted on a subpage with a hatnote added to the top of the report linking to the subpage and indicating if cleanup has been completed or not. An example LLMN discussion might look like:
- That said I would love a way to both clerk and track our cases better. I have been meaning to look at CCI to see how they do it there NicheSports (talk) 22:04, 29 October 2025 (UTC)
No, seriously, what do you have to do to get people to believe you?
[edit]The discussion was archived but I wasn't done, nor did I get an answer to my question, which is: How do you get people to believe you?
NOTE: This is not a discussion about policy. Please do not derail it into a discussion of whether AI use should be allowed or disallowed. This is about how to get people to see what you are seeing, to trust that it is possible for people to identify AI-generated text with over 90% accuracy if you know what you're looking for, that there is actual research about what to look for, that the signs of AI-generated writing exist as an aggregate pattern over millions of words in thousands of articles, that I have both read said thousands of articles and analyzed the text patterns in them alongside many other people who have done the same, and that whatever article I am tagging falls dead-center within that pattern (because when they don't, I don't tag them). This question would be the same regardless of whether AI is allowed or not.
But it seems like all I ever get is people telling you to shut up, essentially: shooting the messenger by complaining that there is now a template on "their" article, asking you to somehow track down a blocked editor from 2023/2024 to swear they used AI on oath, or the latest charming example: maybe i'll make this a Simple English wiki article
. (I guess their calling me stupid is OK since didn't say the dreaded unspeakable F-word?)
What is the sequence of words that will get people to actually listen to you. I would love to know. Gnomingstuff (talk) 12:13, 6 November 2025 (UTC)
- And yes, I am incredibly frustrated because it seems like you can't win. If you tag articles, people complain that you're not pointing out the exact issues (when, for the millionth time, you don't know the exact issues without thorough review; you also don't know what an unsourced statement should be sourced to or whether it's wrong without researching and yet no one complains about pointing out unsourced text). If you point out the issues, people ignore you no matter how clear-cut the evidence provided is. If you fix the issues, people complain that you've destroyed their work. It feels like the only acceptable action to people is to shut up. Gnomingstuff (talk) 12:18, 6 November 2025 (UTC)
- The answer is the same as it was last time: Don't. Ignore the fact that it is or isn't generated by AI and instead focus on why the words on the page would be problematic if they weren't written by AI. For example if there are verification problems, then the problem is not that it was written by AI the problem is that it fails WP:V so explain why it fails verification (and if necessary why failing verification is a problem). If the problem is that the text is waffly and rambly, then the problem is with the writing style so explain that the problem is that the text is waffly and rambly, why this is a problem and rewrite 1-2 sentences yourself to show the difference. And above all, even if you are correct 90% of the time that it was written by AI, just remember that every 10th person you accuse of using AI is telling the truth when they say they didn't. Thryduulf (talk) 12:40, 6 November 2025 (UTC)
How do you get people to believe you?
- I want to know if it is even possible at all. I'm starting to think it isn't. --Gurkubondinn (talk) 15:17, 6 November 2025 (UTC)
- You and I have different approaches. I primarily track the 1325 and 1346 edit filters and have a high threshold for tagging, typically only when I can document multiple content verification failures on the article talk page. I don't think you should change your threshold though - you have a high accuracy rate and are basically single-handedly identifying the magnitude of the problem for the community. My suggestion would be if you run into one of the situations to just ping someone from AIC to come do a deeper analysis of the article for WP:V failures. I'm happy to do that for you. NicheSports (talk) 15:28, 6 November 2025 (UTC)
there is now a template on "their" article
is probably the sum of it unfortunately. I was complaining over coffee and I'm not sure there's a quick solution. If people are getting WP:OWNERSHIPPY over an article, and someone drive-by tags something they missed, they might get defensive. If they can't even see the problem themselves then is it a surprise they revert the tag? Frustrating as all hell, but not surprising. I'd just re-revert with a standard edit summary featuring links to here, the LLM essay, stuff like that. No judgement, just a nice neutral "not all problems are immediately visible, please leave the tag unless/until you've done a thorough review". Until we have an actual AI/LLM/GPT policy they're not actually doing anything technically wrong by removing the tag.- BTW Thryduulf, straight question, do you think it's acceptable to use GPTs to write Wikipedia articles? Please don't waffle over definitions for official policy, I'm not asking your official admin position here, I'm curious what you think as an editor, and just as a person. ~ Argenti Aertheri(Chat?) 22:33, 6 November 2025 (UTC)
- Yes. The reason for this answer is because it is completely and utterly irrelevant how people write Wikipedia articles. If the content they submit meets our copyright, quality and notability standards (WP:V and similar are included under quality here) we should accept it regardless of whether and/or how LLMs or any other technology was used. If the content they submit doesn't meet the copyright and/or notability requirements then we don't want it regardless of any other factors. If they content they submit meets the copyright and notability standards but not the quality standards then it should be cleaned up (if reasonably practical) or deleted (if cleaning up is not reasonably practical). Thryduulf (talk) 00:19, 7 November 2025 (UTC)
- For the avoidance of doubt, I think that acceptance of AI as a tool for writing is inescapable, whether we like it or not. However, unlike the other tools that made writing possible for people who would have had hard time writing otherwise (text editor, spell checker), generative AI is just too fast for humans to check its output. IMHO, part of Wikipedia success is based on a fact that in the past curtailing unacceptable activity here was (marginally) easier than actually engaging in this activity, so a game of whack-a-mole was fun. Generative AI puts us squarely against a machine - a game humans lose - so what we need to think of is either:
- limiting the rate of incoming text to the level we, human editors, can handle, or
- engaging AI in checking (think of filters on steroids).
- I think of an RfC proposal in the context of #1. When thought of as such, alternatives to the outright LLM ban are obvious: just like WP:VPN, institute a special flag that allows some editors to use an LLM, a privilege that shall be hard to earn and easy to lose. Викидим (talk) 20:44, 7 November 2025 (UTC)
- That's not what's actually happening though. In practice
generative AI is just too fast for humans to check its output
, which is relevant because people are, quite frequently, posting things that look like they do meet our standards until you actually read the linked journal article (or whatever). Then you discover it either doesn't exist at all; makes no mention of the topic; or, my personal favorite, says the exact opposite of what the LLM claimed it said. At which point you have 2-3 options, ignore the code smell, at least tag that there may be other LLM issues, or spend an hour+ verifying everything else that editor added to that article (at minimum). Just deleting the whole article is rarely even an option, and cleaning up may require a subject matter expert. Also, LLMs pretty much by definition violate WP:SYNTH, and the copyright questions are concerning. The technology's just not ready yet, and allowing new/anonymous users to use it, when every clause must be carefully checked for copyright and synthesis, is causing far more issues than its solving. ~ Argenti Aertheri(Chat?) 19:48, 10 November 2025 (UTC)- I have been experimenting with AI a lot recently, and know the ways to avoid the problems you have pointed to: (1) select the sources manually (2) upload the texts of sources to AI (3) in the prompt, specify explicitly to use "only the sources provided". The results, on many topics, is surprisingly good (still not good enough to be posted "as-is", but requires mostly cosmetic works and checking of the page numbers (I actually have to do the same after creating a new text all by myself). The "Deep Think" option of the Google Gemini even gets most of the pages numbers right. Instead of a blanket ban, I would therefore suggest requiring a "code of conduct" and a special permission, akin to "extended mover". Викидим (talk) 20:08, 10 November 2025 (UTC)
- For the avoidance of doubt, I think that acceptance of AI as a tool for writing is inescapable, whether we like it or not. However, unlike the other tools that made writing possible for people who would have had hard time writing otherwise (text editor, spell checker), generative AI is just too fast for humans to check its output. IMHO, part of Wikipedia success is based on a fact that in the past curtailing unacceptable activity here was (marginally) easier than actually engaging in this activity, so a game of whack-a-mole was fun. Generative AI puts us squarely against a machine - a game humans lose - so what we need to think of is either:
- Yes. The reason for this answer is because it is completely and utterly irrelevant how people write Wikipedia articles. If the content they submit meets our copyright, quality and notability standards (WP:V and similar are included under quality here) we should accept it regardless of whether and/or how LLMs or any other technology was used. If the content they submit doesn't meet the copyright and/or notability requirements then we don't want it regardless of any other factors. If they content they submit meets the copyright and notability standards but not the quality standards then it should be cleaned up (if reasonably practical) or deleted (if cleaning up is not reasonably practical). Thryduulf (talk) 00:19, 7 November 2025 (UTC)
- A boilerplate, wikilinked claim of authority in the edit summary might help the first impression (
WP:AICLEANUPTEAM revert of additions from User:Example
), maybe linking to some specific and common telltale signs (strong match on WP:AICR#1, WP:AICR#3, WP:AICR#7
). If you don't convey an opening reason or any authority for thinking that it's AI, they'll feel more free to equally strongly claim that it's not, and it'll be harder to convince them from those two starting points. - I've been on the other side of this at Commons, where deletions of unusual-looking image files can go through on a few blunt "this is obviously AI" votes, and it's frustratingly unclear whether those users are providing an experienced eye in a hurry, or are just quick to assume that any image oddities they haven't encountered before can only be, in 2025, a result of AI. Belbury (talk) 09:21, 11 November 2025 (UTC)
- I've been trying this: pointing to WP:AISIGNS (whether in edit summaries or talk pages or both), indicating the specific sections of WP:AISIGNS, indicating specific passages of text which fall cleanly into those sections (which I generally don't like doing because I don't want the takeaway to be that they should just reword those passages and problem solved). Unfortunately it doesn't seem to be helping.
- The other frustration to all of this is that it makes tracking difficult if some unknown percentage of the tags are just going to silently disappear, although I could probably make tracking pages for the larger clusters of articles. Gnomingstuff (talk) 07:49, 19 November 2025 (UTC)
- I think if there's any useful cluster at all (i.e. not just a single article), not even necessarily larger ones it's probably worth creating a section on WP:AINB just so it gets archived somewhere. If it's just a single article the decision is harder of course. Alpha3031 (t • c) 08:12, 19 November 2025 (UTC)
Not sure how viable this is as a solution
[edit]https://en.wikipedia.org/wiki/User:DarklitShadow/AI_detection DarklitShadow (talk) 21:02, 6 November 2025 (UTC)
- As a human reviewer that kind of vagueness is a good tell, but I suspect building an automated system that can identify "vagueness" is a challenge similar in scope to the building of the LLMs themselves. -- LWG talk 21:38, 6 November 2025 (UTC)
- Agreed, that strikes me as far more buggy and wishy-washy than simply looking for keywords or stylistic tells. And far harder to justify to someone you've just tagged with the LLM template.
- I agree that "vagueness" (or I would prefer the term "overgeneralization") is probably the most specific and constant feature of LLM writing, but also one of the hardest to prove and to quantify. WeirdNAnnoyed (talk) 11:58, 7 November 2025 (UTC)
Publifye
[edit]Please see the previous thread. "Publifye" publishes AI-generated books with little to no human review. As this is not always apparent, users mistakenly cite these books as reliable sources. My suggestion in the previous thread was to create an edit filter, but nothing happened. I'd like to workshop some ideas about how we can mitigate this and similar issues going forward. I think there should be an edit filter that warns editors when they add an AI-generated citation (contains for example "Publifye" or one of its authors is "AI"), and tags these edits. A mention at WP:RSML, Wikipedia:Large_language_models#Sources_with_LLM-generated_text, and/or Wikipedia:Reliable_sources/Perennial_sources#Types about this would also be great, something along the lines of "Some publishers are known to publish AI-generated books with little to no human review. As these are products of machine learning, they are not reliable." Kovcszaln6 (talk) 13:13, 9 November 2025 (UTC)
- I think that the proper solution would be for Meta to contact Google and ask all these books delisted (ask them to institute some procedure to complain about the AI publisher, similar to reports of the copyright violations). Google cannot at this point be interested in devaluing their Google Books project by drowning real books in the sea of slop, so there is a slim chance of success here. Pinging @Richard Nevell: who started the previous thread. Викидим (talk) 07:29, 10 November 2025 (UTC)
- I have opened a thread at RSN after having been deferred there. Kovcszaln6 (talk) 09:57, 15 November 2025 (UTC)
Useful script
[edit]I have found WP:UPSD to be a useful script to detect some usage of LLMs, but are there other useful scripts that editors recommend for this purpose? - Amigao (talk) 02:03, 12 November 2025 (UTC)
- As in finding potential usage or assessing the text itself? Gnomingstuff (talk) 12:56, 12 November 2025 (UTC)
- For WP:UPSD in particular, the former. - Amigao (talk) 20:03, 12 November 2025 (UTC)
Live links to LLMs: definitely not WP:RS
[edit]
I have to admit that I'm flabbergasted. I have removed three live LLM URLs from an article. I fail to understand how a URL to an LLM with a specific prompt can be suggested to be anything faintly resembling a WP:RS. I guess it's the success of LLM marketing. I incorrectly put disinformation
in my edit summaries, though it's really closer to misinformation, or gossip.
Anyway, the practical implication is that someone may wish to write a bot to check for links to LLMs in citations, and by default remove them, with the usual warnings about bots, per the standard bot policies. Boud (talk) 13:06, 15 November 2025 (UTC)
- All the links to LLMs should probably be added to filter 869. -- LCU ActivelyDisinterested «@» °∆t° 14:00, 15 November 2025 (UTC)
WorldPeace888 aka FacrFinderW
[edit]User_talk:WorldPeace888#COI_/_PAID They take a news article about a company, they throw the first 2 paragraphs in ChatGPT for a minor rewrite, and add the result to Wikipedia. Polygnotus (talk) 18:26, 15 November 2025 (UTC)
- I think this should be on Wikipedia:WikiProject AI Cleanup/Noticeboard because it is about a specific editor. OutsideNormality (talk) 04:20, 18 November 2025 (UTC)
Category for articles with AI generated sourcing?
[edit]I believe I saw something similar but I can't find it. Is there a category of those pages available? Eurostarguage (talk) 09:26, 17 November 2025 (UTC)
- Specifically {{AI-generated source?}} populates Category:Articles containing suspected AI-generated sources, more broadly {{AI-retrieved source}} populates Category:Wikipedia articles needing factual verification. fifteen thousand two hundred twenty four (talk) 09:32, 17 November 2025 (UTC)
- Thank you! Eurostarguage (talk) 10:03, 17 November 2025 (UTC)
"WP:Writing articles with large language models" is now a Guideline
[edit]Per a successful RFC, Wikipedia:Writing articles with large language models has been promoted to a Wikipedia Guideline. qcne (talk) 13:33, 24 November 2025 (UTC)
- Although it is very important to note the detail of the close, particularly
the current wording (which changed during the debate) does not enjoy a particularly strong consensus and requires further development. In particular we need community consensus on (a) How to identify LLM-generated writing and (b) How to deal with it when it does occur.
andTherefore although I have determined that there is consensus to promote this proposal to a guideline, it has to be construed conservatively and narrowly for the time being. Until the community has decided on a test of what constitutes an LLM-generated contribution, the threshold is consensus; and this means that we have to treat an edit as human-generated until there's consensus otherwise.
(italics in original, bolding mine). Thryduulf (talk) 13:52, 24 November 2025 (UTC)- I am not sure how to interpret that instruction. Taken literally, it would mean that editors should not boldly place the {{LLM}} template onto articles. jlwoodwa (talk) 18:50, 24 November 2025 (UTC)
- This is the sort of problem that occurs when you make a vague statement of philosophy into a guideline without any attempt at working out practical matters beforehand and is a significant reason why many opposed doing so. Nevertheless that view did not reach consensus. I know saying "I told you so" is not helpful, but I genuinely have no practical suggestions off the top of my head - hopefully for the sake of the good of the project others do. Thryduulf (talk) 19:53, 24 November 2025 (UTC)
- I am not sure how to interpret that instruction. Taken literally, it would mean that editors should not boldly place the {{LLM}} template onto articles. jlwoodwa (talk) 18:50, 24 November 2025 (UTC)
Idea lab: potential additions to G15 criteria
[edit]Note: this is not an RFC. I'm looking for feedback and suggestions about these, and potentially other, additions to the G15 criteria. I G15 nom a decent number of articles so have been noodling on this for a bit. A few editors have also suggested potential G15 additions this week (@Femke, @Athanelar, @Tamzin) so seemed like the time to kick this off.
Workshop additional criteria
[edit]- Multiple material source-to-text integrity issues + WP:AISIGNS: This may include, but is not limited to: quotations that are not present in the supplied source, unsupported statistics, or material examples of unsupported information. At least two such errors must be present. If more than one source is supplied for a given claim, the claim must fail verification in all supplied sources. The issues must be documented on the article or draft talk page prior to nominating for G15. The article or draft must also show signs of LLM-generated content, but these do not need to be documented.
- Author-acknowledged LLM use with evidence of insufficient review: This applies if the author has acknowledged that an LLM was used for the article or draft, and that article or draft demonstrates at least one example of WP:V, WP:NPOV, WP:SYNTH or other content policy violation. The content policy violation may be minor. Both the diff of the author's LLM disclosure and the example of a PAG violation must be documented on the article or draft talk page or supplied in the G15 reason parameter. (Note: potentially contentious section following) The LLM disclosure does not need to specifically refer to the article or draft in question, but the LLM disclosure must plausibly cover the time period of the article or draft's creation. For example, if an editor created four drafts in a short time period, all of which display signs of unreviewed LLM usage, but they only acknowledged LLM usage on one draft, all four drafts could be nominated under this criteria.
- Presumed deletion of article or draft by sanctioned editor: This is inspired by CCI's WP:PDEL. An article or draft that shows WP:AISIGNS and was created by an editor subsequently sanctioned for LLM misuse, WP:UPE, and/or WP:Sockpuppetry may be nominated for G15. No documentation is required.
NicheSports (talk) 03:27, 25 November 2025 (UTC)
- Re: point 3 - WP:PDEL is... not as easy to use as it first seems. Those of use at CCI tend to only use it for blocked contributors, occasionally long-inactive ones, it normally involves simply removing the text from the page or stubifying it. Full deletion must go through WP:CPN - and I only send articles through there if there's a high rate of vio, I can't access the source, and/or an established pattern of source fraud, and it has to remain tagged and blanked for at least a week. And even then people get stroppy, so I try and make sure there's a few instances of failed verification as well.
- So, yeah, PDEL is used sparingly, it's not a CSD criterion, and the deletion itself typically has to be sanctioned by a copyright clerk or experienced admin, and for my time at CCI, that experienced admin has normally been Mer-C, who is awesome and seemingly has the thickest skin ever to handle all the "how dare you delete my precious content that I copied from a blogspot post in 2008 you should have just rewritten it instead why won't somebody think of the
childrenreaders"-type complaints. (Which, to be clear, many people who have CCIs opened on them can actually be wonderful & valuable contributors/members of the community! But the negative stuff, like being reverted by admins or accused of vandalism or being a deletionist sticks in your mind more) - CCIs I'm doing, in case anybody thinks I'm not sympathetic towards their efforts to clean up unusable content. GreenLipstickLesbian💌🧸 04:20, 25 November 2025 (UTC)
- Thanks for the feedback GLL. Any comments on the first two? I might drop you a note on your talk page tomorrow with some questions about #3. I'm surprised by the difference in deletion... difficulty?... between PDEL and G15. I G15 articles frequently and no one has ever complained, although I did learn to avoid WikiEd! NicheSports (talk) 05:33, 25 November 2025 (UTC)
Discussion
[edit]- Pinging editors who participated in Wikipedia talk:WikiProject AI Cleanup/Archive 3 § Idea lab: New CSD criteria for LLM content: @Ca:@Thryduulf:@LWG:@Chipmunkdavis:@Jumpytoo:@Fifteen thousand two hundred twenty four:@Newslinger:@Chaotic Enby: Please add anyone I missed or who you think should see this, such as admins who handle a lot of SD noms — Preceding unsigned comment added by NicheSports (talk • contribs) 03:32, 25 November 2025 (UTC)
- As the edit was unsigned, the pings didn't work. @CA, LWG, Chipmunkdavis, Jumpytoo, Fifteen thousand two hundred twenty four, Newslinger, and Chaotic Enby:. Thryduulf (talk) 03:56, 25 November 2025 (UTC)
- Not off to the best start. Thanks for signing NicheSports (talk) 04:02, 25 November 2025 (UTC)
- As the edit was unsigned, the pings didn't work. @CA, LWG, Chipmunkdavis, Jumpytoo, Fifteen thousand two hundred twenty four, Newslinger, and Chaotic Enby:. Thryduulf (talk) 03:56, 25 November 2025 (UTC)
- Before adding anything to a speedy deletion criterion you need to establish that the proposed criteria meet all four of the WP:NEWCSD requirements. At first glance, most of this looks very subjective and so completely failing point 1 (note especially that AISIGNS is completely subjective and full of caveats and exceptions - it seems fundamentally incompatible with speedy deletion). I also have my doubts about frequency and non-redundancy - are there really that many pages that should be deleted according to XfD consensuses, which would meet one or more objective criteria, and which do not meet G15 or any other existing speedy deletion criteria? Thryduulf (talk) 04:00, 25 November 2025 (UTC)
- I will second Thryduulf on this one. WP:AISIGNS, while certainly a very helpful tool, is too subjective to be relied on in a CSD. Point 2 is also worrying as something like "failing WP:NPOV" can be interpreted differently by different editors (remember, content disputes are a thing). That point could also be interpreted very widely: if an author uses LLMs in one article (from what I understand, to any extent), even their non-LLM-written articles would fall under it if they have minor policy violations Chaotic Enby (talk · contribs) 09:05, 25 November 2025 (UTC)
- @Chaotic Enby I am surprised by your feedback about WP:AISIGNS being too subjective to be relied on in a CSD. Several commonly invoked CSD, such as G11 and A7, rely on subjective criteria. More importantly, AISIGNS is already a part of the current G15 definition, as a secondary factor, and forms a critical part of my checklist when I G15 something. For example, I would not G15 a draft that contains 3 references with broken URLs (meeting the "non-existent references" G15 criteria) if it was written in broken English with many grammatical errors, which would indicate the draft was not LLM-generated and the reference issues were caused by something else. In #1 when I specify multiple material WP:V issues + WP:AISIGNS I am just making that secondary check explicit to further restrict when the criteria is invoked. Can you please reconsider? Fair point on #2. I saw Tamzin [1] suggested something similar yesterday and this was my attempt to workshop it. Note that I didn't include Tamzin's 2nd suggestion in my proposed criteria above as I thought it would be too subjective. NicheSports (talk) 14:57, 25 November 2025 (UTC)
- I can't really comment on #3, not having had any experience of presumptive deletion for copyright infringements. For #2, as Chaotic Enby said, things like WP:NPOV or WP:SYNTH can be open to interpretation, which makes "failing" them less suitable as criteria for speedy deletion. However, I could probably get behind multiple clear WP:V failures - I would say three or more rather than two or more - together with evidence of LLM use, be that WP:AISIGNS or a previous WP:LLMDISCLOSE, as an additional criterion for G15. I would like to see some evidence that articles like this are currently taking up too much time at XfD or AfC, so as to meet WP:NEWCSD #3, Frequency. Cheers, SunloungerFrog (talk) 09:47, 25 November 2025 (UTC)
- #2 and #3 are fine but feel like edge cases. I don't think that #1 is viable since we can't even convince people to leave the tag on, even when the person has said they used AI. Even with three WP:V failures and clear signs of AI people will still be like "well I just write like that! and I only made 3 mistakes, people can make 3 mistakes!" Gnomingstuff (talk) 14:58, 25 November 2025 (UTC)
- I like the idea in principle of #1, but it could be more suited as a G3 expansion than G15 as even if a human wrote the fake content it still needs to go. I also feel it needs to be tightened to only cover the case where it's a seemingly well-referenced article but it turns out the content was made up and not supported by any of the sources in the article. A lot of article creations have poor sourcing or mix up the sourcing and that shouldn't be grounds to hit the article with this criteria. I disagree with #2 as it would lead to more editors to not admit they are using LLM, and #3 should be left to a case-by-case basis at ANI or whatever. Jumpytoo Talk 19:10, 25 November 2025 (UTC)
For the interested
[edit]Noticed at [2] (you might not see what I see) that Google's AI now uses Grokipedia as a source. What could possibly go wronger. Gråbergs Gråa Sång (talk) 13:19, 26 November 2025 (UTC)
- Interestingly I see it citing the Wikiwand mirror of our article ... Sam Walton (talk) 16:58, 26 November 2025 (UTC)
- Well, then they can tell Elon they're not using Wikipedia. Gråbergs Gråa Sång (talk) 17:16, 26 November 2025 (UTC)
- comedy is now legal on google Gnomingstuff (talk) 21:00, 26 November 2025 (UTC)
Purpose of the project
[edit]Many participants in this project believe all use of LLMs should be banned from Wikipedia.
Because of this, I wonder if this project mission statement is still accepted: The purpose of this project is not to restrict or ban the use of AI in articles, but to verify that its output is acceptable and constructive, and to fix or remove it otherwise.
SuperPianoMan9167 (talk) 00:37, 28 November 2025 (UTC)
Many participants in this project believe all use of LLMs should be banned from Wikipedia.
- that isn't necessarily true, and no RFC has stated that. User:Bluethricecreamman (Talk·Contribs) 01:26, 28 November 2025 (UTC)I support a full, total, comprehensive, all-encompassing, draconian, butlerian-jihadist ban on LLM usage on Wikipedia. No wiggle room for vandals
staunch LLM abolitionist. I'm doing my part!
- These are from two of the entries on the participants list. SuperPianoMan9167 (talk) 01:33, 28 November 2025 (UTC)
- WikiProjects typically aren't created to advocate for policy change, but to do a specific set of tasks. Thus I think it's true that the purpose isn't to ban the use of AI in articles. ("Restrict" is a bit of a greyer area; arguably ensuring compliance with Wikipedia guidance inevitably involves some restriction.) isaacl (talk) 03:58, 28 November 2025 (UTC)
- "Remove all LLM created content" is arguably doing a specific task. We need to distinguish between content creation and simple automation tasks though, LLMs are pretty damn good at things that are just one step too complicated to fully automate, and can cut down on the copy-pasting substantially. ~ Argenti Aertheri(Chat?) 06:05, 28 November 2025 (UTC)
- Yes, I didn't say it wasn't. My point is that I think the stated purpose remains applicable, independent of what policy changes might be supported by participants. isaacl (talk) 17:16, 28 November 2025 (UTC)
- "Remove all LLM created content" is arguably doing a specific task. We need to distinguish between content creation and simple automation tasks though, LLMs are pretty damn good at things that are just one step too complicated to fully automate, and can cut down on the copy-pasting substantially. ~ Argenti Aertheri(Chat?) 06:05, 28 November 2025 (UTC)
- Speaking only for myself here...a lot of times "purpose" and "what we would like to see" are two different things. I myself have said I would love it if LLMs and the like simply disappeared from the earth. They're wasteful, they encourage laziness, and their output is mediocre in the few areas in which they're actually useful. But that's my opinion, and that's not the purpose of this project. I have no interest in ridding Wikipedia of every article touched by an LLM, I just want to make sure a certain minimum set of standards are met (factual content, references that both exist and support the citation, ideally language that doesn't sound like it was written by a smart 12-year-old with a thesaurus). And on topics that are notable. If all of those are met I do not believe this project requires LLM content to be removed, and I would argue with any user who said it does. WeirdNAnnoyed (talk) 12:44, 28 November 2025 (UTC)
- Wikiprojects should never serve as activist bases for WP:VOTESTACKING. The current project mission statement is acceptable precisely because it is focused on the particular task of cleaning up content that does not meet Wikipedia's quality standards. A project established with the express purpose of promoting LLM prohibition would be disruptive. Yours, &c. RGloucester — ☎ 04:11, 29 November 2025 (UTC)
- Agreed here, don't want an Article Rescue Squad-style tarring and feathering to happen here Gnomingstuff (talk) 06:07, 30 November 2025 (UTC)
- As one of the people in favour of banning all LLMs, I think it's good we keep the mission statement as is and encourage participation from less restrictive editors, otherwise it would be impossible to meaningfully use this project for discussion and consensus-building without it simply being canvassing/vote stacking. Athanelar (talk) 14:38, 30 November 2025 (UTC)
Can someone check this edit?
[edit]This, as well others from same editor... ~2025-36781-52 (talk) 03:49, 28 November 2025 (UTC)
- It looks like someone has reverted it, as well as the other edits by HenriqueMinuit13, because they were all poorly done AI edits. I added {{Uw-ai1}} to their talk page. N7fty (talk) 18:28, 29 November 2025 (UTC)
I believe this article is AI generated - what should I do?
[edit]https://en.wikipedia.org/wiki/Bedtime_procrastination ~2025-36782-44 (talk) 07:54, 29 November 2025 (UTC)
- Why do you believe this? CMD (talk) 08:11, 29 November 2025 (UTC)
- I read over the article, and while there's a few paragraphs that are written a bit awkwardly, I don't personally believe it has AI generated content N7fty (talk) 18:43, 29 November 2025 (UTC)
Supplemental essay for NEWLLM
[edit]Regarding our new AI guideline, I've proposed a supplemental essay to solve the criticism of there being no consensus stsndard for identifying AI text. I thought I'd plug it here since the RfC has been stagnant for a couple of days now and, while I know there's no rush, I would like to get things moving here to avoid another month-long slog like there was with the original guideline proposal. Please look it over and vote in the RfC on the talk page.
https://en.wikipedia.org/wiki/User:Athanelar/Identifying_AI-generated_text Athanelar (talk) 14:32, 30 November 2025 (UTC)
AI spelling correction websites
[edit]Hello. I have a general question for this project: while I've never used LLMs on Wikipedia articles, are websites for AI spelling corrections useful for correcting typos on these articles? Thanks. sjones23 (talk - contributions) 05:56, 1 December 2025 (UTC)
- Probably, although I've had llms confuse different spelling variants, and mess up where humans may also mess up on things like potentially outdated scientific names. CMD (talk) 07:00, 1 December 2025 (UTC)
- Spelling and grammar checkers haven't historically required "AI" to function, and anything using a large language model has the possibility of predicting wrong and silently introducing errors. Use with great caution, if at all. fifteen thousand two hundred twenty four (talk) 07:17, 1 December 2025 (UTC)
- In theory, if you actually review each change, maybe. In practice though? Just fix the ones you spot and move on, or tag it for copy editing. The WP:Guild of Copy Editors will get to it, likely pretty quickly for inline tags like template:copy edit inline. Personally I'd much prefer something be tagged correctly than potentially "fixed". That said, I cannot fucking spell and, with the exception of technical jargon, I've never needed more than the basic word processor built in spellchecker. The squiggly red line is actually generally correct, just targets the individual word, and is built into the visual editor, so why introduce AI? ~ Argenti Aertheri(Chat?) 14:04, 1 December 2025 (UTC)
- Thanks again. While I've been an experienced copyeditor for many years (this month will mark my 19th year as a Wikipedian), I sometimes use word processors, AI spell checkers like Grammarly, and of course this project's visual editor page to correct errors and typos, though oddly enough, all of them are not really considered LLMs. sjones23 (talk - contributions) 06:25, 3 December 2025 (UTC)
- I don't know that I've ever had to resort to external tools for spell checking on Wikipedia, let alone an LLM. Firefox (as with most mainstream browsers) already has a decent spellcheck feature built in. That said, I'm not usually copy editing, so I typically don't need a tool that can distinguish between the different varieties of English and whatnot. As for technical jargon, the only technical articles I contribute to are usually topics I am very familiar with. Should I ever need an external tool, I'd probably just paste the entire article into Microsoft Word. - ZLEA TǀC 06:44, 3 December 2025 (UTC)
- Safari (the browser I'm currently using) also has a built in spellchecker to correct any errors. Of course, if I had to copyedit or rewrite a major section in any article, I would do it directly in the article and sometimes use either Microsoft Word (my word processor) or my sandbox as needed while checking for any potential copyvios. sjones23 (talk - contributions) 06:55, 3 December 2025 (UTC)
- I don't know that I've ever had to resort to external tools for spell checking on Wikipedia, let alone an LLM. Firefox (as with most mainstream browsers) already has a decent spellcheck feature built in. That said, I'm not usually copy editing, so I typically don't need a tool that can distinguish between the different varieties of English and whatnot. As for technical jargon, the only technical articles I contribute to are usually topics I am very familiar with. Should I ever need an external tool, I'd probably just paste the entire article into Microsoft Word. - ZLEA TǀC 06:44, 3 December 2025 (UTC)
- Thanks again. While I've been an experienced copyeditor for many years (this month will mark my 19th year as a Wikipedian), I sometimes use word processors, AI spell checkers like Grammarly, and of course this project's visual editor page to correct errors and typos, though oddly enough, all of them are not really considered LLMs. sjones23 (talk - contributions) 06:25, 3 December 2025 (UTC)
- I do it to spell-check whenever I write in languages other than English (and Russian), and it works well (obscure terminology is always a problem, but it had fooled many human translators as well). For English I only do it to change text to British spelling (with occasional lapses - on my part, not AI). Since many spelling and grammar checkers (say, Grammarly) are in effect based on LLM-like machinery, I cannot see a problem in using the bigger and more modern engines. Викидим (talk) 17:51, 3 December 2025 (UTC)
I have boldly tagged WP:LLM as an information page instead of an essay. Thoughts on this change would be appreciated. SuperPianoMan9167 (talk) 03:18, 4 December 2025 (UTC)