Wikipedia:Village pump (WMF)
Policy | Technical | Proposals | Idea lab | WMF | Miscellaneous |
- Discussions of proposals which do not require significant foundation attention or involvement belong at Village pump (proposals)
- Discussions of bugs and routine technical issues belong at Village pump (technical).
- Consider developing new ideas at the Village pump (idea lab).
- This page is not a place to appeal decisions about article content, which the WMF does not control (except in very rare cases); see Dispute resolution for that.
- Issues that do not require project-wide attention should often be handled through Wikipedia:Contact us instead of here.
- This board is not the place to report emergencies; go to Wikipedia:Emergency for that.
Threads may be automatically archived after 14 days of inactivity.
Behaviour on this page: This page is for engaging with and discussing the Wikimedia Foundation. Editors commenting here are required to act with appropriate decorum. While grievances, complaints, or criticism of the foundation are frequently posted here, you are expected to present them without being rude or hostile. Comments that are uncivil may be removed without warning. Personal attacks against other users, including employees of the Wikimedia Foundation, will be met with sanctions.
WMF receives letter from Trump-appointed acting DC attorney

See this article in the Washington Post. There's also coverage from other reliable outlets findable online. Ed Martin appears to have picked the WMF as his next target for vaguely threatening letters. I am very interested to see what, if any, response the WMF makes to this, and trust they will continue to stand up fro free speech, free information, and Wikipedia's editor community. I see there's also some discussion on Jimbo's talk page here. —Ganesha811 (talk) 01:14, 26 April 2025 (UTC)
- Hi, I know this is super late to the discussion, but here's a free link to the Washington Post article if anyone wants it: https://wapo.st/4jE6rp8 Northern-Virginia-Photographer (talk) 14:06, 2 May 2025 (UTC)
From Admin Noticeboard
US Attorney for the District of Columbia Ed Martin sent this threatening letter to the WMF today. Larry Sanger is involved. Here is early analysis. Cullen328 (talk) 22:19, 25 April 2025 (UTC)
- God damn... Tarlby (t) (c) 22:31, 25 April 2025 (UTC)
- Some time ago, there was a thread at the Teahouse (?) about moving the servers out of the US. Maybe this needs a rethink? Knitsey (talk) 22:35, 25 April 2025 (UTC)
- What's he's threatening is Wikimedia's tax exempt status. Schazjmd (talk) 22:41, 25 April 2025 (UTC)
- Move the foundation out of the United States too.Simonm223 (talk) 23:36, 25 April 2025 (UTC)
- Would be fun to have to delete all images on Commons and Enwiki, lol. And say hello to 80 billion libel suits. PARAKANYAA (talk) 23:58, 25 April 2025 (UTC)
- Why would we need to delete the images? There are countries with even more liberal copyright laws than the US. Moving the servers out of the US is a common request on Commons because of this.
- And as far as I have heard, WMF has already servers in several countries. Plus, there are also countries that give NGOs and the like tax exempt status. Nakonana (talk) 09:30, 27 April 2025 (UTC)
- Last I heard, the WMF lawyer said on Commons that they don't actually have to obey US copyright law and that the Commons community was free to relax copyright PoliciesAndGuidelines a bit if they wanted to. Aaron Liu (talk) 23:24, 27 April 2025 (UTC)
- (See c:Special:GoToComment/c-JWilz12345-20250303024700-Y.haruo-20250302172500, cf. c:Commons:Lex loci protectionis.) Aaron Liu (talk) 23:31, 27 April 2025 (UTC)
- That's... not what that link says. That link says that the physical location of the servers doesn't determine which laws apply. WhatamIdoing (talk) 04:35, 9 May 2025 (UTC)
- France is the most appropriate fall-back I can think of, given it is the only EU state with both freedom of speech laws and a working nuclear deterrence force to back it up. Baltakatei 04:50, 26 April 2025 (UTC)
- France did have this episode, however. Curbon7 (talk) 07:18, 26 April 2025 (UTC)
- And also Sinking of the Rainbow Warrior. No, definitely not France. MinervaNeue (talk) 08:51, 26 April 2025 (UTC)
- More recently, French Wikipedians have been subjected to threats and intimidation from the right-wing press. --Grnrchst (talk) 12:33, 26 April 2025 (UTC)
- While I don't condone the dox threat (making a user accountable in their perspective), the page legitimately had neutrality issues and was sourced by a blog post. Please don't further derail the conversation. Hplotter (talk) 16:05, 27 April 2025 (UTC)
- @Baltakatei France is not a good location for Wikimedia servers since 2015. Note that their society of architects and artists (ADAGP) is anti-Wikipedia, considering their vocal opposition to Freedom of Panorama and their criticism of Wikimedia world's imposition of commercial-type CC licensing to images of buildings and monuments. Unless you want to impose universal prohibition of all images of modern architecture on enwiki and apply a restrictive fair use exemption tag like what French Wikipedia is doing. Per c:COM:FOP France, "Even if these non-free images [of modern buildings] are now tolerated in French Wikipedia articles, the legitimate copyright holders [(like the living architects)] can send their veto so that these images will be deleted on French Wikipedia too. The same deletion will occur when receiving a French court order: their long-term presence is not warranted as long as the copyright protection persists." JWilz12345 (Talk|Contrib's.) 08:52, 26 April 2025 (UTC)
- Though frankly, concerns about pictures of modern buildings doesn't really move the needle considering the bigger picture of what's at stake. Bon courage (talk) 10:03, 26 April 2025 (UTC)
- I'm surprised that France was the first country to be proposed here, given all the problems it has with freedom of the press (as mentioned above). As a counter-example, Switzerland has freedom of panorama, robust privacy and data protection laws, and is ranked 9th in the world for freedom of the press. Ireland, Norway and the Netherlands would also spring to mind before I'd suggest France. --Grnrchst (talk) 12:30, 26 April 2025 (UTC)
- Switzerland was also one of the first country that comes to mind. Maybe Norway, Sweden or Finland, too? --PantheraLeo1359531 (talk) 14:42, 26 April 2025 (UTC)
- Finland is restrictive regarding freedom of panorama, iirc. Nakonana (talk) 09:38, 27 April 2025 (UTC)
- @Nakonana Finland has FoP for buildings though, but they take architecture strictly; they don't follow the logic Californian courts follow with regards to sculptures that are inherent elements of architectures, like gargoyles and stained glass windows. Perhaps 95% of FoP-USonly images might be OK under Finnish FoP but not the 5%, including File:Pedro Calungsod stained glass (cropped).jpg. JWilz12345 (Talk|Contrib's.) 09:44, 27 April 2025 (UTC)
- Finland is restrictive regarding freedom of panorama, iirc. Nakonana (talk) 09:38, 27 April 2025 (UTC)
- Switzerland was also one of the first country that comes to mind. Maybe Norway, Sweden or Finland, too? --PantheraLeo1359531 (talk) 14:42, 26 April 2025 (UTC)
- Germany also has freedom of speech laws. See Artikel 5 of the German Grundgesetz. (It's just called "freedom of expression" instead of "freedom of speech".) France has very restrictive rules for copyright (e.g. even plain buildings are copyrighted), so that you'd need to delete half of the photos from wiki Commons if servers were to be moved there. Germany's copyright laws are much more lenient. Nakonana (talk) 09:36, 27 April 2025 (UTC)
- Canada seems like a decent option. We have fair dealing, freedom of panorama, relatively close to the US, etc. But my understanding is that there are challenges beyond simply deciding to move everything. Clovermoss🍀 (talk) 00:28, 30 April 2025 (UTC)
- France did have this episode, however. Curbon7 (talk) 07:18, 26 April 2025 (UTC)
- Would be fun to have to delete all images on Commons and Enwiki, lol. And say hello to 80 billion libel suits. PARAKANYAA (talk) 23:58, 25 April 2025 (UTC)
- Move the foundation out of the United States too.Simonm223 (talk) 23:36, 25 April 2025 (UTC)
- The WMF already maintains servers in a number of locations around the world including Brazil, France, Netherlands, Singapore and USA. Andrew🐉(talk) 19:06, 26 April 2025 (UTC)
- The US has a strong government that sticks its nose where it doesn't belong.I would vote for Island. ·Carn·!? 05:57, 28 April 2025 (UTC)
- Iceland? It's next on their list after Greenland. Gråbergs Gråa Sång (talk) 06:43, 28 April 2025 (UTC)
- What's he's threatening is Wikimedia's tax exempt status. Schazjmd (talk) 22:41, 25 April 2025 (UTC)
- Some time ago, there was a thread at the Teahouse (?) about moving the servers out of the US. Maybe this needs a rethink? Knitsey (talk) 22:35, 25 April 2025 (UTC)
- I anticipate the WMF will retain counsel and send a forceful response. voorts (talk/contributions) 23:08, 25 April 2025 (UTC)
- @Voorts What force would they have for that, may I ask? Darwin Ahoy! 14:29, 26 April 2025 (UTC)
- Ed Martin sends lots of letters but he's clearly wrong on the law and this won't go anywhere. voorts (talk/contributions) 15:03, 26 April 2025 (UTC)
- Right... Well, lets see how it goes. Darwin Ahoy! 15:10, 26 April 2025 (UTC)
- See my longer comment below. voorts (talk/contributions) 15:15, 26 April 2025 (UTC)
- Right... Well, lets see how it goes. Darwin Ahoy! 15:10, 26 April 2025 (UTC)
- Ed Martin sends lots of letters but he's clearly wrong on the law and this won't go anywhere. voorts (talk/contributions) 15:03, 26 April 2025 (UTC)
- @Voorts What force would they have for that, may I ask? Darwin Ahoy! 14:29, 26 April 2025 (UTC)
- Can someone protect Ed Martin's article. Martin sent the letter and the page seems to be picking up random vandalism. Thanks. Randy Kryn (talk) 23:12, 25 April 2025 (UTC)
- Two IP edits isn't enough to warrant protection. voorts (talk/contributions) 23:16, 25 April 2025 (UTC)
- Semi-protected x 4 years per WP:CT/AP. -Ad Orientem (talk) 04:28, 26 April 2025 (UTC)
- Two IP edits isn't enough to warrant protection. voorts (talk/contributions) 23:16, 25 April 2025 (UTC)
- This is part of a larger campaign against sources that allow criticism of Trump policies, and includes sending letters to major medical journals. StarryGrandma (talk) 00:06, 26 April 2025 (UTC)
- I share the administration's concerns with the media, academia, Wikipedia, and bias, but this is ridiculous. You don't combat bias with lies. The Knowledge Pirate (talk) 04:04, 26 April 2025 (UTC)
- Smart lawyers don't send reams of data to a prosecutor in response to a fishing expedition letter. So I don't expect WMF to send anything more than a polite "We share your concerns about neutral points of view, accuracy, and propaganda in media. The long arc of our efforts bends toward neutrality and accuracy. There are no political litmus tests for educational 501(c)(3) organizations, which have a First Amendment right to write as they see the world. There are thousands of examples of 501(c)(3) organizations publishing from conservative points of view, including some that you yourself have founded, such as the Eagle Forum Education and Legal Defense Fund." If they wanted to poke the bear, they could add, "We consider your threatening letter an effort to coerce Wikipedia to be more amenable to using its deserved popularity to push your own propaganda."
- However, there is a kernel of truth in the attack; there is an imbalance in WP's NPOV. I have tried using very reliable sources (e.g. a book written by a serious scientist and professor who'd served years in the Federal Government on the topic) to inject a little neutrality into pages on Climate Change. All my edits were reverted because that source's statements conflicted with the rabidly biased existing article and with the apparent political opinions of other editors (and administrators). The cited author isn't even conservative -- merely not rabidly progressive on the topic, taking a neutral scientific view. But there's a whole "if you don't agree with us, you are DENIER of SCIENCE" attitude in WP, despite real science proceeding by airing disagrements rather than suppressing them. Another example is how the article on Paul R. Ehrlich is periodically edited to a hagiography, by editors who seemingly can't stand the idea that the prophet who taught them the world would end due to high population had feet of clay, being extremely inaccurate and often completely incorrect in the majority of his sensationalized predictions. That article remains a mess, veering in all directions and following most valid, well-sourced criticism with "but..." and praise. There is a similar problems with the articles about the Great Barrington Declaration and its authors. It was a well-sourced and legitimate disagreement on Covid policy that was ruthlessly suppressed by the left (including the Federal government) to present an appearance of scientific and political unanimity for a "lockdown" policy. Even today, its lede still uses the dismissive word "fringe"! And smears the sponsoring nonprofit as "associated with climate change denial", as if that had anything to do with whether the Declaration about Covid policy was reliable or notable.
- On WP topics where there IS a current imbalance of neutrality, the deck is stacked such that it's quite hard for serious editors to correct the imbalance. What changes can the WP community make to be more welcoming to serious editing (not conservative propaganda) from people who disagree with liberal sacred cows? -- Gnuish (talk) 00:08, 29 April 2025 (UTC)
- The ideas you want to insert are not widely accepted by mainstream academia, so they don't get equal weight in articles. This isn't the place to rehash old content disputes. Thebiguglyalien (talk) 🛸 00:11, 29 April 2025 (UTC)
- Academia is not the arbiter of what can be mentioned in Wikipedia, last time I read our policies. Discussing content disputes could only be out of place, if Martin's fishing expedition was not directly related to biases in the content of Wikipedia, and how they get adjudicated; but it was. We should hear and find the kernel of truth in even the ravings of a lunatic, and the ravings of experienced Wikipedia editors, to reconfirm whether we have good adjudication systems in place. Our systems must suit the whole English-speaking community, not just "mainstream academia" and not just "liberals". -- Gnuish (talk) 23:47, 9 May 2025 (UTC)
- Yes, it kind of is. You may wish to read the false balance policy and the fringe theories guideline for Wikipedia's standards on this. Thebiguglyalien (talk) 🛸 23:55, 9 May 2025 (UTC)
- Academia is not the arbiter of what can be mentioned in Wikipedia, last time I read our policies. Discussing content disputes could only be out of place, if Martin's fishing expedition was not directly related to biases in the content of Wikipedia, and how they get adjudicated; but it was. We should hear and find the kernel of truth in even the ravings of a lunatic, and the ravings of experienced Wikipedia editors, to reconfirm whether we have good adjudication systems in place. Our systems must suit the whole English-speaking community, not just "mainstream academia" and not just "liberals". -- Gnuish (talk) 23:47, 9 May 2025 (UTC)
- The ideas you want to insert are not widely accepted by mainstream academia, so they don't get equal weight in articles. This isn't the place to rehash old content disputes. Thebiguglyalien (talk) 🛸 00:11, 29 April 2025 (UTC)
- Update: Trump just pulled Martin from his nomination to be the DC US Attorney due to declining political support for his various legal threats and association with avowed Nazis. Seems likely that such frivolous legal threats will be on hold for now. ViridianPenguin🐧 (💬) 18:15, 8 May 2025 (UTC)
- Maybe it'd make sense to close this thread, it's becoming rampant with intense political debate unrelated to the topic at hand, which is seemingly resolved. Gaismagorm (talk) 18:19, 8 May 2025 (UTC)
- I agree with closing. Martin will be out of office in 12 days. --Tryptofish (talk) 19:09, 8 May 2025 (UTC)
- @Tryptofish Doesn't it pass to the next? Darwin Ahoy! 09:42, 9 May 2025 (UTC)
- If you mean Trump's next nominee, Jeanine Pirro, he can indeed put her in office on an interim basis, and she will be a Trump loyalist. But that does not mean that the Senate will confirm her to a permanent appointment, and (if I understand correctly) there is some kind of legal requirement that, if there isn't a confirmation soon enough, the ability to name the appointee passes from Trump to some judges in DC, who are going to treat it very differently than Trump would. Now, I don't claim to understand that, and I don't, but I'd prefer to close this, and see what happens, since discussion can always be reopened if appropriate. --Tryptofish (talk) 19:53, 9 May 2025 (UTC)
- @Tryptofish Doesn't it pass to the next? Darwin Ahoy! 09:42, 9 May 2025 (UTC)
- I agree with closing. Martin will be out of office in 12 days. --Tryptofish (talk) 19:09, 8 May 2025 (UTC)
- Maybe it'd make sense to close this thread, it's becoming rampant with intense political debate unrelated to the topic at hand, which is seemingly resolved. Gaismagorm (talk) 18:19, 8 May 2025 (UTC)
Discussion about closing this thread (when it was at AN), reopening this thread, and moving this thread to a village pump. –Novem Linguae (talk) 03:48, 26 April 2025 (UTC)
|
---|
I object to your close of this thread, Cambalachero, and have explained why on your talk page. I urge you to revert your close. Cullen328 (talk) 02:42, 26 April 2025 (UTC)
This discussion should be closed. As I pointed when I did so, Whatever is done about this, will be decided by the WMF, not by editors (admin or not). There is no actionable request here, nor any news that changes our way to do things. In fact, the discussion has already been derailed into forum-like territory. Discussing if Trump's policies are good or not, is exactly that. Discussing things that none of us has the power to decide either way (such as moving the servers, or even the WMF itself), is exactly that. If you take a moment to think about it, you will realize it. --Cambalachero (talk) 03:15, 26 April 2025 (UTC) I wish I could say I was surprised. But I have been expecting something like this from the moment he won the election. -Ad Orientem (talk) 03:21, 26 April 2025 (UTC)
|
Continued discussion
- Page 3 point 6 of the letter from the Acting United States Attorney for the District of Columbia says
Similarly, what is the Foundation's official process for auditing or evaluating the actions, activities, and voting patterns of editors, admins, and committees, including the Arbitration Committee ...
This is clearly a major concern for all editors and administrators. Clearly, these people are planning to "audit and evaluate" us when the WMF tells them that is not appropriate and not how Wikipedia works. I reject the notion that editors and administrators should meekly step aside and expect the WMF handle this latest outrage with zero input from us. Cullen328 (talk) 03:51, 26 April 2025 (UTC)- I hope the editors and admins State-side don't receive much negativity or spotlight on this, especially those who are not really anonymous. – robertsky (talk) 04:04, 26 April 2025 (UTC)
- Biggest concern is probably for those living in the US who are not citizens. Nil Einne (talk) 07:13, 26 April 2025 (UTC)
- @Nil Einne Those are obviously in the front line, but the danger is for all people living in the United States, looking at what the US administration has repeatedly stated on that regard. Darwin Ahoy! 14:33, 26 April 2025 (UTC)
- Biggest concern is probably for those living in the US who are not citizens. Nil Einne (talk) 07:13, 26 April 2025 (UTC)
- I noticed that the letter accuses WMF of allowing people to endanger the "national security and the interests of the United States". Since Wikipedia is a multilingual, international project, maybe the WMF should point out in its response that it is not beholden to protect the national security or the interests of any country. Also, given that the letter does not mention any examples of so-called "information manipulation", I'm not sure what Martin is trying to get at, other than perhaps trying to bully the WMF into compliance. Finally, I should note that the letter mentions that the presence of "foreign nationals" (i.e. non-Americans) on WMF's board is "subverting the interests of American taxpayers", which is a rather strange thing to say, given that (1) WMF serves an international audience, not a US-only audience, and (2) WMF receives no American tax revenue, so there is no such interest being "subverted". – Epicgenius (talk) 04:09, 26 April 2025 (UTC)
- Tax free status is a form of government subsidy. Hawkeye7 (discuss) 05:36, 26 April 2025 (UTC)
- Why be specific when you can be vague, much easier to defend your statements. Gråbergs Gråa Sång (talk) 09:45, 26 April 2025 (UTC)
- From TheFP [1],
The letter did not specify which foreign actors were manipulating information on Wikipedia and did not cite examples of alleged propaganda. However, a person close to Martin said he is concerned about “edits on Wikipedia as they relate to the Israel-Hamas conflict that are clearly targeted against Israel to benefit other countries.”
— hako9 (talk) 18:55, 26 April 2025 (UTC)
- Why would the Foundation (or any non-profit/company/ect) need to know the voting patterns of anyone? That's a really f'ed up thing to include in there. SilverserenC 04:06, 26 April 2025 (UTC)
- Wouldn't that be virtually impossible to qualify as well? Knitsey (talk) 04:50, 26 April 2025 (UTC)
- Sure, but I think we all know exactly what sort of voting patterns and general opinions about politics (and who one supports) that they're really wanting to know by including that in there. SilverserenC 05:07, 26 April 2025 (UTC)
- Yeah, I guess that is obvious. But it would take a long time to complete that task. I would think that the WMF might be able to string this out for, say, just short of 4 years? Knitsey (talk) 05:12, 26 April 2025 (UTC)
- You think there'll be elections in the USA again anytime soon? Well, maybe ... But even if there were, the risk is there would be some new manifestation of US govt in future that leaned the same way, for socially-ingrained reasons that are very hard to grapple with, within the electorate. The question is: why should Wikipedia/WMF want to be in the USA? I cannot see any serious downside to decamping, and many up-sides. Bon courage (talk) 05:30, 26 April 2025 (UTC)
- I think this was discussed once before, and someone mentioned that it would cost many millions of dollars to change the country that wmf is headquartered in. There is also a danger of picking the wrong country to change to, then this process would need to be repeated if authoritarianism or government suppression of free speech occurred there. –Novem Linguae (talk) 11:39, 26 April 2025 (UTC)
- It's certainly a huge thing to consider, with a lot of potential problems it could introduce, but I don't think we should rule it out completely. The logistical, legal and financial costs of moving to a different country are far outweighed by the societal damage that could be done by leaving the encyclopedia at the mercy of a regime that is openly hostile to its existence.
- The Encyclopédistes were forced to move their publication headquarters to Switzerland when the ancien regime tried to shut them down. Wikimedia having to move its base of operations elsewhere would not be historically unprecedented. --Grnrchst (talk) 12:49, 26 April 2025 (UTC)
- Rousseau, Diderot, Voltaire.. Funny how these things keep resurfacing. Apparently we sometimes forget and slide backwards far enough for history to rear its head. -- GreenC 21:17, 26 April 2025 (UTC)
- Well, the WMF would certainly be welcome in Geneva, Rousseau's place of birth and where many international organizations are headquartered. Switzerland has largely favorable laws for such organizations, also tax-wise, and good freedom of press - with some caveats when it comes to bank secrecy... Gestumblindi (talk) 19:33, 29 April 2025 (UTC)
- Rousseau, Diderot, Voltaire.. Funny how these things keep resurfacing. Apparently we sometimes forget and slide backwards far enough for history to rear its head. -- GreenC 21:17, 26 April 2025 (UTC)
- I think this was discussed once before, and someone mentioned that it would cost many millions of dollars to change the country that wmf is headquartered in. There is also a danger of picking the wrong country to change to, then this process would need to be repeated if authoritarianism or government suppression of free speech occurred there. –Novem Linguae (talk) 11:39, 26 April 2025 (UTC)
- You think there'll be elections in the USA again anytime soon? Well, maybe ... But even if there were, the risk is there would be some new manifestation of US govt in future that leaned the same way, for socially-ingrained reasons that are very hard to grapple with, within the electorate. The question is: why should Wikipedia/WMF want to be in the USA? I cannot see any serious downside to decamping, and many up-sides. Bon courage (talk) 05:30, 26 April 2025 (UTC)
- Yeah, I guess that is obvious. But it would take a long time to complete that task. I would think that the WMF might be able to string this out for, say, just short of 4 years? Knitsey (talk) 05:12, 26 April 2025 (UTC)
- Sure, but I think we all know exactly what sort of voting patterns and general opinions about politics (and who one supports) that they're really wanting to know by including that in there. SilverserenC 05:07, 26 April 2025 (UTC)
- Wouldn't that be virtually impossible to qualify as well? Knitsey (talk) 04:50, 26 April 2025 (UTC)
- I hope the editors and admins State-side don't receive much negativity or spotlight on this, especially those who are not really anonymous. – robertsky (talk) 04:04, 26 April 2025 (UTC)
- Page 3 point 6 of the letter from the Acting United States Attorney for the District of Columbia says
- @Novem Linguae Depending on how the WMF behaves and answers to the US Administration demands, that could be a very plausible move, indeed. Darwin Ahoy! 14:39, 26 April 2025 (UTC)
What is the Foundation’s official process for auditing or evaluating the [...] voting patterns of editors, admins, and committees
. Well that's disturbing... Curbon7 (talk) 05:55, 26 April 2025 (UTC)- If it helps, I've never voted for any American party. Gråbergs Gråa Sång (talk) 10:24, 26 April 2025 (UTC)
- Nor have I, but that doesn't stop them from trying to find out which Swedish parties you have voted for, or, in my case, British. Phil Bridger (talk) 10:07, 27 April 2025 (UTC)
- I read this part as voting patterns for "!votes" on-Wiki, as that would make most sense. But given the throngs of fascism that are latched through the current political moment in the US, this may have been naivete on my part. -- Cdjp1 (talk) 11:55, 27 April 2025 (UTC)
- Nor have I, but that doesn't stop them from trying to find out which Swedish parties you have voted for, or, in my case, British. Phil Bridger (talk) 10:07, 27 April 2025 (UTC)
- If it helps, I've never voted for any American party. Gråbergs Gråa Sång (talk) 10:24, 26 April 2025 (UTC)
- @Cullen328 One of the reasons IP editing should never have been allowed in any wikimedia project, even in 2001. As of now, all people that uses and used an IP of which the records still are in the ISP is a sitting duck ready to be sued. Darwin Ahoy! 14:42, 26 April 2025 (UTC)
- Well, that ship sailed a quarter of a century ago, DarwIn. And it is rarely easy to identify an individual from an IP address. Cullen328 (talk) 16:59, 26 April 2025 (UTC)
- @Cullen328 all it takes for any government to know location and eventually identity is to request that data from the ISP the IP belongs to, the most common case by large being that an IP belongs to some sort of ISP. In the case of authoritarian governments that information is usually at the distance of a phone call. Yes, that ship quite unfortunately sailed a quarter of a century ago, but it can, and should, be shipwrecked any day. We have already done just that at the Portuguese speaking wikipedia 5 years ago, btw. Darwin Ahoy! 17:11, 26 April 2025 (UTC)
- All you'd know then is who the name of the person who signed the contract with the internet provider for this IP. But you'd not know who made the edit: was it the person who signed the contract, was it a family member of that person (if so, then which one), was it a friend, was it a one-time guest of the person who signed the contract? It will be impossible to identify the actual editor, and after 25 years even said editor probably doesn't remember whether it was them who made the edit in question. Nakonana (talk) 10:08, 27 April 2025 (UTC)
- Additionally, there are also public wifi at cafes, libraries, etc, which do not require people to share their personal information in order to be connected. – robertsky (talk) 14:34, 27 April 2025 (UTC)
- @Robertsky I wouldn't assume the generality of IP users are Mata Haris or 007s in sunglasses and headscarf sneaking into public wifis to edit "anonymously". From my experience, people usually do that either out of laziness, or even worst, misguided by the reckless but prevalent myth that IP editions are somehow "anonymous", happily walking into the wolves mouth that way. Darwin Ahoy! 15:24, 27 April 2025 (UTC)
- @Nakonana I don't think assuming the ISP contract was signed by someone else, usually very close to the person in question. is really an argument. Fact is that IP editing is and has been a significant hazard for the editors of the wikimedia projects that use that, willingly or unwillingly, endangering people's lives including their physical integrity and of their loving ones ones. Some quick examples:
- In 2018, Trust & Safety removed CU access at wiki.zh for safety reasons, as it can be used to locate the editors and find out their identity, in a context where editors were actually identified by the government and physically harmed by those or other means.
- Also in 2018, in Brazil, an IP editor was disciplinarily sanctioned at their workplace after editing an article from their workplace.
- In September 2023, the Supreme Court of Portugal condemned the Wikimedia Foundation to give away the information (IP, agent, etc) of the editors that edited a certain biography to the person that was requesting them, so that said person could sue them (this case is also interesting because it used the EU "Right to Vanish" law to force the will of a person over what was published about them in reliable and independent sources, such as big media TV and newspapers).
- It's absolutely reckless to persist in allowing IP editions on the Wikimedia projects, even more in the current context in the US where that can mean almost immediate identification of the editor, and the fact that such recklessness persists for 25 years already only makes it more urgent to stop it now. Darwin Ahoy! 15:52, 27 April 2025 (UTC)
- It may well be absolutely reckless, but multiple times the en.wiki community has requested the mandating of 'sign in to edit', and each time the WMF has rejected it, because - apparently, as I recall - it 'goes against being the Encylopedia That Anyone Can Edit'. Even as TVTropes mandated SITE. This was over 10 years ago, and given that "temporary accounts" are apparently about to become a thing, (proper) SITE remains a pipe dream. - The Bushranger One ping only 01:50, 28 April 2025 (UTC)
- @The Bushranger Well, we've done just that at wiki.pt 5 years ago, and the WMF took no issue with it. IP editing has been successfully banned from that Wikipedia since then, and we still are the encyclopedia anyone can edit (after spending 2 seconds creating an account). Darwin Ahoy! 10:04, 28 April 2025 (UTC)
- Maybe they've changed since ~10 years ago. But the fact en.wiki remains IP-enabled points to y'all at pt. being lucky. - The Bushranger One ping only 22:10, 28 April 2025 (UTC)
- Temporary accounts that don't show people's IP addresses are being slowly rolled out across wikis. I think we'll probably be one of the last to get it, but the existence of the project shows that the foundation has considered the privacy implications of an IP address being publically visible (even it took 20 years to get to this point where it's a near-future feature). Clovermoss🍀 (talk) 00:37, 30 April 2025 (UTC)
- Maybe they've changed since ~10 years ago. But the fact en.wiki remains IP-enabled points to y'all at pt. being lucky. - The Bushranger One ping only 22:10, 28 April 2025 (UTC)
- @The Bushranger Well, we've done just that at wiki.pt 5 years ago, and the WMF took no issue with it. IP editing has been successfully banned from that Wikipedia since then, and we still are the encyclopedia anyone can edit (after spending 2 seconds creating an account). Darwin Ahoy! 10:04, 28 April 2025 (UTC)
- It may well be absolutely reckless, but multiple times the en.wiki community has requested the mandating of 'sign in to edit', and each time the WMF has rejected it, because - apparently, as I recall - it 'goes against being the Encylopedia That Anyone Can Edit'. Even as TVTropes mandated SITE. This was over 10 years ago, and given that "temporary accounts" are apparently about to become a thing, (proper) SITE remains a pipe dream. - The Bushranger One ping only 01:50, 28 April 2025 (UTC)
- Additionally, there are also public wifi at cafes, libraries, etc, which do not require people to share their personal information in order to be connected. – robertsky (talk) 14:34, 27 April 2025 (UTC)
- All you'd know then is who the name of the person who signed the contract with the internet provider for this IP. But you'd not know who made the edit: was it the person who signed the contract, was it a family member of that person (if so, then which one), was it a friend, was it a one-time guest of the person who signed the contract? It will be impossible to identify the actual editor, and after 25 years even said editor probably doesn't remember whether it was them who made the edit in question. Nakonana (talk) 10:08, 27 April 2025 (UTC)
- @Cullen328 all it takes for any government to know location and eventually identity is to request that data from the ISP the IP belongs to, the most common case by large being that an IP belongs to some sort of ISP. In the case of authoritarian governments that information is usually at the distance of a phone call. Yes, that ship quite unfortunately sailed a quarter of a century ago, but it can, and should, be shipwrecked any day. We have already done just that at the Portuguese speaking wikipedia 5 years ago, btw. Darwin Ahoy! 17:11, 26 April 2025 (UTC)
- Well, that ship sailed a quarter of a century ago, DarwIn. And it is rarely easy to identify an individual from an IP address. Cullen328 (talk) 16:59, 26 April 2025 (UTC)
- Is the Acting United States Attorney for the District of Columbia also going to send letters to Facebook and Twitter/X to ask them about their
official process for auditing or evaluating the actions, activities, and voting patterns of
[users]...? I'd be really curious to hear Musk's reply to this. Nakonana (talk) 09:52, 27 April 2025 (UTC)
- You guys remember the Asian News International case, where an Indian court attempted to force WMF to provide the names and details of three users? A Wikipedia article about the case, Asian News International vs. Wikimedia Foundation was promptly created, but had to be taken down (blanked). Is anybody working on creating an article about Ed Martin's letter to the WMF, hint hint? I don't think it would be as easy to get that taken down. Bishonen | tålk 10:16, 26 April 2025 (UTC).
- @Bishonen Too early, it has a sentence in his article atm, which seems about right. But the WaPo article is a good start, don't you agree, @Valereee? Gråbergs Gråa Sång (talk) 10:21, 26 April 2025 (UTC)
- We do have other news sites picking this up now, though none as prominent as WaPo. Gizmodo, Huffpost, The Verge, New Zealand Herald. -- Cdjp1 (talk) 11:58, 27 April 2025 (UTC)
- So I should write another blacklockable article? :D I agree it's probably too early, but if it turns into an actual lawsuit, probably notable. Valereee (talk) 12:15, 28 April 2025 (UTC)
- @Bishonen Too early, it has a sentence in his article atm, which seems about right. But the WaPo article is a good start, don't you agree, @Valereee? Gråbergs Gråa Sång (talk) 10:21, 26 April 2025 (UTC)
- I am of the opinion that the only affirmative action WMF should do at this time is have legal write a letter indicating WMF is willing to vindicate its rights in court. Moving servers is a bad idea, for reasons already indicated, but also because it is, in a way, complying with the lawless bully. I don't know what the community response should be, since I don't know what it would hope to achieve. I had (in the earlier thread on this page) the idea of a "community affidavit", to support WMF legal's fight. Tito Omburo (talk) 12:09, 26 April 2025 (UTC)
- Here's my perspective as an attorney: Ed Martin is a clown. His job thus far appears to be sending threatening letters to conservative bugbears in an attempt to chill speech. He doesn't have the authority to revoke tax exempt status (he's the interim United States Attorney for the District of Columbia, not the IRS), and if he actually had a case of criminal wrongdoing, his office/the FBI would be sending subpoenas or executing warrants, not sending public letters to the WMF. Even Kash Patel's FBI wouldn't open an investigation on thin bullshit like this and no judge would sign a warrant based on innuendo. As I said above, WMF will send a forceful letter in response and Martin will back down because he's got nothing. Everyone freaking out about this is precisely what Martin wants; he should be ignored. voorts (talk/contributions) 15:15, 26 April 2025 (UTC)
- In short: Do not feed the troll. - The Bushranger One ping only 17:05, 26 April 2025 (UTC)
- Please try to reformulate your idea without insulting people. BLP applies for everybody everywhere, even for Ed Martin here. Cambalachero (talk) 02:51, 27 April 2025 (UTC)
- I would laugh this off -- most of those around the short-fingered convicted felon are clowns (& the rest are incompetent hacks) -- except this time around they understand what they can do having control of the White House, & have ratcheted up their oppression. Witness the arrest of a state judge for opposing the increasingly lawless ICE. I'm no longer confident that the threats having that person in office can be overstated. -- llywrch (talk) 18:09, 26 April 2025 (UTC)
- Yes, they're all clowns, but more of the killer clown variety. They're literally supporting more than one genocide right now. I wouldn't be laughing. Lindspherg (talk) 18:56, 26 April 2025 (UTC)
- My 2¢ ... I am taking a wait and see approach. While I hope voorts is right and this turns out to be a clownish distraction, I'm not dismissing the potential for it to become something serious. This administration has already shown a breathtaking contempt for the rule of law and civil liberties. The language in that letter is right out of every tyrant's playbook for intimidating and/or suppressing sources of news and information that they can't control. For now, I await with interest the WMF's response. I know they have lawyers on retainer and the resources to hire more if needed. -Ad Orientem (talk) 19:14, 26 April 2025 (UTC)
- Elsewhere I have recommended groups associated with Wikipedia outside the US make & keep backups of the project databases. My point in recommending this is as insurance of the worst case scenario: the DoJ somehow shuts down the Foundation. Now I've said elsewhere that Wikipedia can survive much better without the Foundation than the Foundation can survive without Wikipedia. Having backups outside the control of the Federal government makes it far easier for a group to fork Wikipedia & preserve our goal of creating a free encyclopedia -- or an encyclopedia in exile, if you will. Sure, there will be legal problems basing a free encyclopedia in a non-US country (e.g. copyright, laws of defamation), but I have faith that the grass roots of Wikipedia -- as well as similar projects -- will come up with solutions. There has been talk of the Foundation creating contingency plans if the clowns with nukes are effective; we, the community, must needs have our own contingency plans to carry on our work. -- llywrch (talk) 18:00, 29 April 2025 (UTC)
- My 2¢ ... I am taking a wait and see approach. While I hope voorts is right and this turns out to be a clownish distraction, I'm not dismissing the potential for it to become something serious. This administration has already shown a breathtaking contempt for the rule of law and civil liberties. The language in that letter is right out of every tyrant's playbook for intimidating and/or suppressing sources of news and information that they can't control. For now, I await with interest the WMF's response. I know they have lawyers on retainer and the resources to hire more if needed. -Ad Orientem (talk) 19:14, 26 April 2025 (UTC)
- Yes, they're all clowns, but more of the killer clown variety. They're literally supporting more than one genocide right now. I wouldn't be laughing. Lindspherg (talk) 18:56, 26 April 2025 (UTC)
- I'm not sure what the best course of action is but, if the WMF wishes to respond directly to these questions, it will have no shortage of material. For example, there are lots of policies such as the Universal Code of Conduct which is currently undergoing a round of revision. And it can point to actions taken such as the 2021 Wikimedia Foundation actions on the Chinese Wikipedia.
- In any case, it's good that the WMF has a substantial endowment so that it can afford to take whatever course of action is decided.
- Andrew🐉(talk) 19:39, 26 April 2025 (UTC)
- There's a massive noise to info ratio here. The most tangible damage this letter has done so far is prompting WP:FORUM-style speculation and fearmongering within the community. Several people here have taken the bait, and reopening this discussion was a mistake. Thebiguglyalien (talk) 🛸 20:46, 26 April 2025 (UTC)
- I disagree. But as someone who has a front-row seat to these disturbing political developments, I suggest as a prudent action that all Wikimedia groups outside of the US to start making regular backups of Wikipe[p|m]ia content against the worst possible outcome. (In any case, making backup copies of important data is always a good idea. Every IT system expert recommends this. Even if there is no threat from a lawless regime.) -- llywrch (talk) 22:14, 26 April 2025 (UTC)
- We at Wiki Project Med ship EN WP on a Raspberry Pi Zero W 2 server. So you can buy your very own version. Or you can make your own.MDWiki:WikiProjectMed:Internet-in-a-Box Doc James (talk · contribs · email) 23:07, 26 April 2025 (UTC)
- It's an easy enough job to download the entirety of En.Wiki (<25GB (sans media)), host would be harder with potential traffic level, but is doable. And of course, for as long as the archiving sites are up, they hold a repository of a majority of wiki articles. -- Cdjp1 (talk) 12:07, 27 April 2025 (UTC)
- Stewing on it for a bit, I think the most practical approach that each of us individually can take to any challenge is simply to double down on our principles. WP:V, WP:NPOV, and WP:BLP remain the top priority. We can do our cause a lot of good just by sticking to them strictly, keeping our processes transparent and avoiding any iota of a violation on high-profile articles or BLPs within the American politics topic area. This also means clamping down on WP:SOAPBOX and WP:CPUSH, which we can sometimes be very lax with. It might be worth starting a discussion about how WP:AE handles politically charged editing that's subtle enough to avoid an instant ban. This would help with stopping these bad actors from manipulating Wikipedia from within while also stopping those who might make the rest of us look bad in the eyes of the public. Thebiguglyalien (talk) 🛸 02:28, 27 April 2025 (UTC)
- Just adding a "yes and" to say reliability is also paramount in these contentious situations. I spend little to no time on US politics-related areas of the encyclopedia, but I have seen in articles I come across that blog posts, opinion columns and even tweets and reddit threads are far more prevalent than they ought to be. --Grnrchst (talk) 08:24, 28 April 2025 (UTC)
- While making backups is always a good idea and one that should be encouraged, the real threat here is not the loss of any information Wikipedia contains. The relatively small file size of the English language Wikipedia means that such a large number of copies have certainly been made that there is little risk of it disappearing. Instead, the lasting damage would come from the disruption to the networks and communities that maintain it, the inability to continue improving and updating it and the problem with accessing the aforementioned archived data. –Noha307 (talk) 03:11, 27 April 2025 (UTC)
- We at Wiki Project Med ship EN WP on a Raspberry Pi Zero W 2 server. So you can buy your very own version. Or you can make your own.MDWiki:WikiProjectMed:Internet-in-a-Box Doc James (talk · contribs · email) 23:07, 26 April 2025 (UTC)
- I disagree. But as someone who has a front-row seat to these disturbing political developments, I suggest as a prudent action that all Wikimedia groups outside of the US to start making regular backups of Wikipe[p|m]ia content against the worst possible outcome. (In any case, making backup copies of important data is always a good idea. Every IT system expert recommends this. Even if there is no threat from a lawless regime.) -- llywrch (talk) 22:14, 26 April 2025 (UTC)
- I think it's fine to discuss this (that's not taking the bait from anyone), and I think the most important thing for the community to do (along with prudent measures like making backups, and protecting one's real-life identity, if not already disclosed) is to make it clear that we are proud of what we do (yes, sure, we have lots of mistakes, but we correct them), and we aren't going to be intimidated by bullies. --Tryptofish (talk) 00:16, 27 April 2025 (UTC)
- As it seems there is consensus that the thread should stay open, I will add my 2 cents. As I understand, Wikipedia is an educative web page, and that grants them a tax exemption. But I'm sure that it can't be enough that Wikipedia self-describes itself as an educative web page, there must be requirements to it, otherwise every page out there would abuse of such loophole. And what I understood when I checked the mail was that Ed Martin was discussing if Wikipedia actually met such requirements or not. After all, we all know that Wikipedia, as a self-published source, is not a reliable source... so can we really be that upset when someone says that we are not reliable enough to be educative? So the options for the WMF may be to either change things around to fit the standards required to be a fully reputable educative source (and that may mean mass culling of topics such as TV series, videogames, films, recent events, etc, editorial oversight, editors editing under their real names and only on topics they have some actual degree or expertise, following standards on content set by external actors, etc), and then keep the tax excemption. Or, be just a general-purpose web page, that sets it own internal rules on its content and user behavior, but pays the applicable taxes. So, my question is, which are the legal rules to be considered an educative web page? Does Wikipedia meet such rules? --Cambalachero (talk) 00:24, 27 April 2025 (UTC)
- I'm disinclined to treat Martin's question about whether or not we are educational as a serious question, at least insofar as the editing community's response. There is a legal question as to tax status, and that's something we should leave to WMF Legal. --Tryptofish (talk) 00:37, 27 April 2025 (UTC)
- Well, you should. "Wikipedia is an educative web page as definited in those laws and regulations" is a stronger argument than "Wikipedia is an educative web page because they say so, and I don't like the guy who questioned it" Cambalachero (talk) 02:43, 27 April 2025 (UTC)
- Not to belabor the point, but I meant that we should let Legal speak first, as opposed to the editing community getting out ahead of them. I can see that my use of the word "serious" unintentionally led me into the rabbit hole of "seriously versus literally", where I didn't want to go. I wasn't trying to say that we should be glib. Rather, I mean that we should not take the letter on face value, because the letter is clearly written in bad faith. --Tryptofish (talk) 22:36, 27 April 2025 (UTC)
- This letter has nothing to do with WMF fulfilling its legal tax status (despite what is written in it), and everything to do with intimidation by a government that does not like press freedom, free speech, academic liberty, sciences, and more broadly knowledge. — Jules* talk 10:47, 27 April 2025 (UTC)
- Funny, a thing I learned as a Wikipedia editor is never to trust someone whose main argument is that there is a conspiracy to silence him. Cambalachero (talk) 00:16, 28 April 2025 (UTC)
- While I do agree that there is most likely no conspiracy to silence us right now, I do think it is genuine topic of concern when it comes the administration's handling of situations like this. (Man, this is becoming a downer). Gaismagorm (talk) 00:18, 28 April 2025 (UTC)
- Well, the people etc who indicate they want WP to shut up about some stuff include Musk, Heritage Foundation who said their investigation of WP will be "shared with the appropriate policymakers to help inform a strategic response", Ed Martin, ADL and orgs like New York Post [2].
- This of course does not mean there is conspiracy, but at least there are some people with influence with a common view. Gråbergs Gråa Sång (talk) 04:49, 28 April 2025 (UTC)
- There absolutely is a conspiracy to silence us ([3], [4], [5]). We can argue about its extent, participants' identities, and efficacy, but it is foolish to deny it. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:50, 28 April 2025 (UTC)
- I'm not sure about the specifics within the US, but antisemitism can lead to crimes, if not a crime in itself. And it's written at one of the disclaimer pages that editors must respect the law. So, if someone commited a crime by adding antisemitism content to this internet page, and an organization wants to track the real people behind the usernames and make them answer for such crimes before a court of law... by all means, let them do it. WP:NONAZIS surely includes antisemitism as well. Cambalachero (talk) 16:03, 28 April 2025 (UTC)
- The bar for unprotected hate speech in the US is very high. GMGtalk 16:08, 28 April 2025 (UTC)
- Perhaps, I'm no lawyer. But even if they can legally get away with it, don't place me in the same bag, in the "us" of "There absolutely is a conspiracy to silence us". If anything, there is a conspiracy to silence them, not us, and I have no problem with it, in fact I support it. Cambalachero (talk) 16:17, 28 April 2025 (UTC)
- This has never been about combating antisemitism. There are numerous people in the orbit of the administration who are, themselves, antisemites. Fighting antisemitism is just a convenient fig leaf for the real agenda, which is shutting down counter-narratives to the officially preferred narrative. Same thing for the bogus claim that Wikipedia harbors foreign agents who are trying to harm US interests. --Tryptofish (talk) 21:37, 28 April 2025 (UTC)
- That, and it's a camel's nose. It's reasonable. It's even laudable! It also sets a precedent. - The Bushranger One ping only 22:12, 28 April 2025 (UTC)
- Or not. This reminds me of a real-world case: the notorious nazi Adolf Eichmann escaped to my country, Argentina, and stayed hidden. Simon Wiesenthal and the MOSAD located, captured and smuggled him to Israel, where he was put on trial. Someone could have said: "this sets a precedent, if we allow this the MOSAD will soon do whatever they want in Argentina". But no. The MOSAD captured and smuggled him, mission accomplished, and except for some other similar cases of runaway nazis, things never escalated to a "Jewish occupation" as the usual antisemite tropes would claim. Projects that seek to reduce or stop antisemitism have my full support, and if that means outing a couple of Wikipedia troublemaker editors, so be it. Cambalachero (talk) 00:42, 29 April 2025 (UTC)
- I fail to see a valid analogy or parallel here. Antisemitism is being used here as a Trojan Horse by right wing Christian nationalists. They don't actually care about Jews, Jewish people, Jewish culture, or even Israel. What they care about is building powerful voting bloc coalitions like the kind promoted by the Council for National Policy. They have strategically targeted and convinced a tiny percentage of U.S. Jews (see American Jews in politics: "Helmreich describes them as "a uniquely swayable bloc" as a result of Republican stances on Israel") that the Christian right will uphold their shared interests. Ironically, this so-called "interest" is in opposition to 70% (likely much higher) of U.S. Jews who do not support Project 2025 or their policies. The reality is that religious tolerance is a liberal idea upheld by Democrats, not the Christian right. Just like the kapos in Nazi-era WWII who helped their fellow Jews to their deaths, we see the same or similar occurring here. And that, my friend, is a valid analogy. Viriditas (talk) 02:21, 29 April 2025 (UTC)
- ...and, as I said earlier, I only have deaf ears for arguments based on conspiracy theories. Cambalachero (talk) 02:43, 29 April 2025 (UTC)
- I’m just going to leave this here.[6] Viriditas (talk) 02:48, 29 April 2025 (UTC)
- Try again. That page lost me the second they used the term "latinx"... which, if nobody told you, is highly offensive for most Latin Americans like me. Cambalachero (talk) 14:05, 29 April 2025 (UTC)
- American historian Steven Hahn discusses this kind of reaction in his research on illiberalism in U.S. history. He argues that illiberalism often emerges as a fearful reaction to a perceived threat. Your comments above illustrate this tendency. Hahn: "People who regard themselves as liberal in every other respect are perfectly happy to impose an incredibly repressive, politically and otherwise…expulsive regime as a way of trying to soothe the concerns of their constituents." It’s interesting that taking offense at a word you don’t like, or being upset by a group one doesn’t like, or living in any kind of perpetual offense or fear, would have one reject the entire philosophical and liberal enterprise of the Enlightenment, from democracy to individual rights. Thanks for the insight into the global phenomenon of democratic backsliding. Viriditas (talk) 16:34, 29 April 2025 (UTC)
- Right to cultural identity is repressive. Got it. Cambalachero (talk) 14:46, 30 April 2025 (UTC)
- Cambalachero, you and I are probably going to have to agree to disagree. And that's fine with me! In fact, something that I deeply value about what we do here at Wikipedia is that editors with all manner of personal opinions are not only allowed to edit here, but are welcome to, just so long as we all adhere to NPOV and adhere to the various other policies and guidelines. That's something that editors should be proud of. And right there, we can see the moral bankruptcy of the accusations that we systematically suppress the conservative point of view. --Tryptofish (talk) 23:00, 29 April 2025 (UTC)
- American historian Steven Hahn discusses this kind of reaction in his research on illiberalism in U.S. history. He argues that illiberalism often emerges as a fearful reaction to a perceived threat. Your comments above illustrate this tendency. Hahn: "People who regard themselves as liberal in every other respect are perfectly happy to impose an incredibly repressive, politically and otherwise…expulsive regime as a way of trying to soothe the concerns of their constituents." It’s interesting that taking offense at a word you don’t like, or being upset by a group one doesn’t like, or living in any kind of perpetual offense or fear, would have one reject the entire philosophical and liberal enterprise of the Enlightenment, from democracy to individual rights. Thanks for the insight into the global phenomenon of democratic backsliding. Viriditas (talk) 16:34, 29 April 2025 (UTC)
- Try again. That page lost me the second they used the term "latinx"... which, if nobody told you, is highly offensive for most Latin Americans like me. Cambalachero (talk) 14:05, 29 April 2025 (UTC)
- I’m just going to leave this here.[6] Viriditas (talk) 02:48, 29 April 2025 (UTC)
- ...and, as I said earlier, I only have deaf ears for arguments based on conspiracy theories. Cambalachero (talk) 02:43, 29 April 2025 (UTC)
- I fail to see a valid analogy or parallel here. Antisemitism is being used here as a Trojan Horse by right wing Christian nationalists. They don't actually care about Jews, Jewish people, Jewish culture, or even Israel. What they care about is building powerful voting bloc coalitions like the kind promoted by the Council for National Policy. They have strategically targeted and convinced a tiny percentage of U.S. Jews (see American Jews in politics: "Helmreich describes them as "a uniquely swayable bloc" as a result of Republican stances on Israel") that the Christian right will uphold their shared interests. Ironically, this so-called "interest" is in opposition to 70% (likely much higher) of U.S. Jews who do not support Project 2025 or their policies. The reality is that religious tolerance is a liberal idea upheld by Democrats, not the Christian right. Just like the kapos in Nazi-era WWII who helped their fellow Jews to their deaths, we see the same or similar occurring here. And that, my friend, is a valid analogy. Viriditas (talk) 02:21, 29 April 2025 (UTC)
- Or not. This reminds me of a real-world case: the notorious nazi Adolf Eichmann escaped to my country, Argentina, and stayed hidden. Simon Wiesenthal and the MOSAD located, captured and smuggled him to Israel, where he was put on trial. Someone could have said: "this sets a precedent, if we allow this the MOSAD will soon do whatever they want in Argentina". But no. The MOSAD captured and smuggled him, mission accomplished, and except for some other similar cases of runaway nazis, things never escalated to a "Jewish occupation" as the usual antisemite tropes would claim. Projects that seek to reduce or stop antisemitism have my full support, and if that means outing a couple of Wikipedia troublemaker editors, so be it. Cambalachero (talk) 00:42, 29 April 2025 (UTC)
- That, and it's a camel's nose. It's reasonable. It's even laudable! It also sets a precedent. - The Bushranger One ping only 22:12, 28 April 2025 (UTC)
- This has never been about combating antisemitism. There are numerous people in the orbit of the administration who are, themselves, antisemites. Fighting antisemitism is just a convenient fig leaf for the real agenda, which is shutting down counter-narratives to the officially preferred narrative. Same thing for the bogus claim that Wikipedia harbors foreign agents who are trying to harm US interests. --Tryptofish (talk) 21:37, 28 April 2025 (UTC)
- Perhaps, I'm no lawyer. But even if they can legally get away with it, don't place me in the same bag, in the "us" of "There absolutely is a conspiracy to silence us". If anything, there is a conspiracy to silence them, not us, and I have no problem with it, in fact I support it. Cambalachero (talk) 16:17, 28 April 2025 (UTC)
- What does your comment about antisemitism have to do with my comment? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:11, 2 May 2025 (UTC)
- You said "There absolutely is a conspiracy to silence us", followed by a link to an article about the HF trying to locate antisemite editors and start legal actions (or whatever, not clear yet) against them. Did you actually read the article, or just the clickbait title? Cambalachero (talk) 12:17, 2 May 2025 (UTC)
- I linked to three separate articles, as evidence of a conspiracy to silence us. Your comment did not address that point, let alone disprove it. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:27, 2 May 2025 (UTC)
- You said "There absolutely is a conspiracy to silence us", followed by a link to an article about the HF trying to locate antisemite editors and start legal actions (or whatever, not clear yet) against them. Did you actually read the article, or just the clickbait title? Cambalachero (talk) 12:17, 2 May 2025 (UTC)
- The bar for unprotected hate speech in the US is very high. GMGtalk 16:08, 28 April 2025 (UTC)
- I'm not sure about the specifics within the US, but antisemitism can lead to crimes, if not a crime in itself. And it's written at one of the disclaimer pages that editors must respect the law. So, if someone commited a crime by adding antisemitism content to this internet page, and an organization wants to track the real people behind the usernames and make them answer for such crimes before a court of law... by all means, let them do it. WP:NONAZIS surely includes antisemitism as well. Cambalachero (talk) 16:03, 28 April 2025 (UTC)
- There absolutely is a conspiracy to silence us ([3], [4], [5]). We can argue about its extent, participants' identities, and efficacy, but it is foolish to deny it. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:50, 28 April 2025 (UTC)
- Funny, a thing I learned as a Wikipedia editor is never to trust someone whose main argument is that there is a conspiracy to silence him. Cambalachero (talk) 00:16, 28 April 2025 (UTC)
- Thousands of educational institutions could provide expert evidence that their students use Wikipedia as an educative resource. (Maybe skip Harvard this time...) Certes (talk) 10:52, 27 April 2025 (UTC)
- This is actually a very good idea for an action to take at this stage (as opposed to some of the more extreme proposals in this thread, which I think should be reconsidered at a later date). Putting out a request for public support from people and institutions that use Wikipedia as an educational resource, or some kind of open letter, would definitely help improve our position against any threats on these grounds. --Grnrchst (talk) 13:25, 27 April 2025 (UTC)
- Yes, that is a good idea. I like the concept that we should, for now (as in while we wait to see what WMF Legal decides to do), focus on informational, rather than confrontational, things, and doing something that both (1) demonstrates how other people appreciate what we provide, and (2) lets other people know what's happening, in case they should want to speak out in support of us, is a good strategy. --Tryptofish (talk) 22:41, 27 April 2025 (UTC)
- This is actually a very good idea for an action to take at this stage (as opposed to some of the more extreme proposals in this thread, which I think should be reconsidered at a later date). Putting out a request for public support from people and institutions that use Wikipedia as an educational resource, or some kind of open letter, would definitely help improve our position against any threats on these grounds. --Grnrchst (talk) 13:25, 27 April 2025 (UTC)
- Well, you should. "Wikipedia is an educative web page as definited in those laws and regulations" is a stronger argument than "Wikipedia is an educative web page because they say so, and I don't like the guy who questioned it" Cambalachero (talk) 02:43, 27 April 2025 (UTC)
- I'm disinclined to treat Martin's question about whether or not we are educational as a serious question, at least insofar as the editing community's response. There is a legal question as to tax status, and that's something we should leave to WMF Legal. --Tryptofish (talk) 00:37, 27 April 2025 (UTC)
- For anyone who isn't already aware of it, there was an earlier, related, discussion at Wikipedia:Village pump (miscellaneous)/Archive 80#Heritage Foundation intending to "identify and target" editors. --Tryptofish (talk) 00:44, 27 April 2025 (UTC)
Some thoughts here:
- Don't worry about the data. WMF technical people aren't dumb, and they almost certainly have robust backups capable or weathering any natural or political disaster. Even if all of them failed, 3rd parties have backups sufficient to piece it back together.
- WMF should have an exit plan for America at this point. They absolutely should NOT share it with us or even acknowledge it exists publicly, but they should, and probably already do, have a plan for leaving the US while maintaining continuity of their technical systems, know-how and key personnel. Sharing this plan or acknowledging it exists would just add unnecessary fuel to the fire at this point.
- As a community, we need to double down on our policies. NOTCENSORED and NPOV are the two I think are most endangered by this right now. We need to emphasize to the WMF that these policies are non-negotiable. If the Government starts pushing on them, the community needs to communicate to the WMF the expectation that bending or breaking is unacceptable, and it is preferable for the WMF to pull out of America than to bend on our core policies.
- The community needs to chill on the blackout talk. We're not there yet. If editor safety or our core policies are threatened, THEN it's time to breakout the banners, blackouts, and forks in escalating order. Right now we're WAAAAY premature, and the WMF has excellent lawyers precisely for letters like this. Tazerdadog (talk) 11:26, 28 April 2025 (UTC)
- I just want to note that emphasizing to the WMF that NPOV is non-negotiable is not really the issue. As you may or may not have read, I'm chairing a working group on NPOV. There is no chance that the WMF is about to challenge the idea of neutrality, and a much higher chance that the WMF will be expanding support for volunteers in making sure that NPOV is upheld. "As threats to neutrality appear to be on the rise globally, Wikipedia’s neutral point of view (NPOV) policy is needed now more than ever."
- In short, I 100% agree with you, Tazerdadog, that "we need to double down on our policies". I think there are exactly zero people at the WMF who have any notion that we should give up on neutrality to please any government! We're all in this together. Jimbo Wales (talk) 10:01, 29 April 2025 (UTC)
- Can you expand on the working group? It's the one described here, correct? Do you think that work has greater implications for small wikis than for the English Wikipedia? —Ganesha811 (talk) 17:41, 29 April 2025 (UTC)
Friends, Wikipedians, citizens of the world, lend me your ears. We will not be cowed by the aggressive actions of a lawless regime. Its fate will be decided by the public, whose approval of the Trump administration has already sunk in opinion polls to the lower 40s ranging to the high 30s. When MAGA encounters empty shelves at stores, high inflation, disappearance of jobs, and an increasingly likely recession, suddenly the "anti-woke" will be awakened. Carlstak (talk) 00:45, 29 April 2025 (UTC)
- I was tempted to add {{Not a forum}} to this discussion earlier, and I'm increasingly thinking it's warranted. Thebiguglyalien (talk) 🛸 01:25, 29 April 2025 (UTC)
- That's fine, I did get up on Antony's soapbox. I've been sounding the alarm about this for months in real life, and people are just now taking it seriously. Carlstak (talk) 01:56, 29 April 2025 (UTC)
Time to move to a more Federal model
The Wikimedia movement started in the USA, but it has been a global movement since its earliest days. There are other global movements around that we can compare ourselves to, at least in how we handle money. Some are relatively loose confederations, with each national organisation having its own fundraising. The Wikimedia movement is an odd hybrid with some chapters like Germany handling the donations from readers in Germany but most, including the UK, being grant funded from the USA with UK readers donations going to the USA. Now would seem an appropriate moment to reconsider that model. Maybe move one or both datacentres from the USA to another country such as Canada, Iceland or Ireland, the endowment to a financial hub such as London or Franfurt, and decentralise fundraising to any country where we have a national registered charity. It would be odd for a for profit US organisation to do charitable fundraising in other countries.
If the US organisation was only handling US donations, then it would be reasonable for its board to be US based, with a separate global board to coordinate the various national chapters. If the only wikimedia donations handled in the USA were donations from people in the USA then the movement's exposure to US taxes etc would be greatly reduced. Disclosure, I have at times been a member of WMUK and worked for it from 2013 to 2015, however I'm not connected to it these days. ϢereSpielChequers 05:50, 27 April 2025 (UTC)
- @WereSpielChequers sounds good on paper, but in practice, it will be a complicated setup as different countries have different rules on how donations raised within the borders should be disbursed within the borders and internationally. From what I understand, the German chapter passes on the excess amount collected back to USA (excess of what was would be budgeted originally with Foundation).
- I organise the Singapore user group and I did consider a scenario of what if the donation banners are activated for Singapore IP addresses and the money collected through Singapore IP addresses into a future Singapore charity for Wikimedia. If the aim is to share this collected amount to other affiliates that do not have fundraising options, it is not pretty as the 80% of the net proceeds raised in this manner mostly likely have to be set aside for the activities in Singapore. If the collection of the amounts is way beyond what we have budgeted for the year, we may have to find ways to spend it (I don't know... maybe like offsetting costs of running the datacenter in Singapore? yeah. the Foundation has caching servers in Singapore) or endure criticisms of having a reserve fund that may not be depleting over time. – robertsky (talk) 14:30, 27 April 2025 (UTC)
- EN.WP serves Canada, the UK and several other Commonwealth countries with English as a first language. Federation other languages would not make a US-only board even remotely OK. Simonm223 (talk) 14:57, 27 April 2025 (UTC)
- To consider another example, in Canada a charity must be Canadian-registered or a UN agency in order for donations to be tax-deductible. To be registered as a Canadian charity, the organization must be carrying out its charitable purposes itself (or be in direct control of the work being done by others). (Donations to a U.S. charity can be eligible with some restrictions if you have U.S. income, or you or your family are enrolled in a U.S. university.) Based on my understanding, a Canadian Wikimedia charity wouldn't be able to simply transfer tax-deductible donations to another organization. isaacl (talk) 15:04, 27 April 2025 (UTC)
- Yes a federal model has implications, individual chapters would have to adopt particular projects as DE has done with Wikidata. The WMF board would have to split into a USA chapter board and some sort of global council, and the two combined would have less power within the movement than the WMF has today. But if organisations as diverse as Greenpeace and the Red Cross can do this, we could to. ϢereSpielChequers 17:24, 27 April 2025 (UTC)
- This sounds like a good idea, but I have little knowledge of charity laws in different countries. I admit that I don't contribute any money to the WMF, but I know that if I contribute money to a UK charity I am usually asked if I am a UK tax payer, in which case they can claim back the tax paid on the donation. I certainly don't like contributions from outside the US going towards the MAGA agenda. Phil Bridger (talk) 18:04, 27 April 2025 (UTC)
- The Canadian Red Cross carries out its own relief work, and Greenpeace Canada is a non-profit but not a registered charity. I agree it's possible in theory to transform the network infrastructure into separately run subnetworks. They'll be additional overhead, with duplication of functions across the separate organizations, and fundraising challenges to ensure each organization collects enough funds for its operations and endowment fund. isaacl (talk) 14:47, 28 April 2025 (UTC)
- Yes a federal model has implications, individual chapters would have to adopt particular projects as DE has done with Wikidata. The WMF board would have to split into a USA chapter board and some sort of global council, and the two combined would have less power within the movement than the WMF has today. But if organisations as diverse as Greenpeace and the Red Cross can do this, we could to. ϢereSpielChequers 17:24, 27 April 2025 (UTC)
- Indeed. There's a lot of talk of servers above, but moving things between data centres is relatively trivial. The primary reason we're vulnerable to this kind of pressure from the US government is that the WMF made the early mistake of concentrating its financial and organisational resources in the US, instead of going down the route exemplified by Wikimedia DE. I hope this will prompt them to reconsider that choice. – Joe (talk) 10:30, 28 April 2025 (UTC)
- I don't think we're particularly vulnerable to pressure from the US government, by the way. Jimbo Wales (talk) 10:02, 29 April 2025 (UTC)
What efforts went into the SOPA blackout?
I think that being prepared to support a banner or a blackout to protest is warranted at this time. The SOPA blackout had senators phones ringing off the hook. This is political power. The Wikimedia community has the power to shape public opinion-politicians-law through banners and blackouts. DO NOT BE AFRAID TO USE IT. Protests_against_SOPA_and_PIPA#Wikimedia_community Victor Grigas (talk) 11:41, 26 April 2025 (UTC)
- It's not a matter of being afraid to use it, but that this is an encyclopedia with editors from around the world and of many different opinions rather than a campaigning site for Americans. Phil Bridger (talk) 12:41, 26 April 2025 (UTC)
- If someone wants to seriously propose a blackout, they'll need to begin with a fully formed proposal that clearly lays out a very credible existential threat to the English Wikipedia, supported by reliable sources. Basically your proposal needs to be Featured Article quality when you first post the RFC. And you need to post it far enough ahead of time that consensus has time to happen. Keep in mind that you'll have to convince or outvote those who believe Wikipedia should be entirely apolitical even in the face of an existential threat, people from other countries who aren't familiar with US politics or culture wars, and the opposite side in the US culture wars (yes, there are such people here). And hope that someone else doesn't decide to post a "stub-class" proposal while you're preparing your FA-class proposal, and thereby poison the well. Anomie⚔ 13:46, 26 April 2025 (UTC)
- In addition to the threat, there would need to be a compelling case for impact. The SOPA/PIPA blackout sought to raise awareness of the potential impact of specific legislation that was likely not very well known among the general public. That's not the situation the Martin WMF letter creates. CMD (talk) 14:03, 26 April 2025 (UTC)
- yes however the wmf (if I remember correctly) is based in america. Even if someone isn't a US resident, it would still effect them. I do however understand the sentiment, but I do highly support a possible blackout/banner. Wikipedia is used very frequently, so it would definitely get people's attention (which, sadly, I feel is the most people can do nowadays as regular citizens, oh well). Gaismagorm (talk) 01:16, 27 April 2025 (UTC)
- As much as it is an American thing, there are many things that international editors do here are reliant on the laws in America and the status of WMF to shield them from their own. Nonetheless, the use blackouts like SOPA's should be considered when the threat becomes very real, whereas the situation now is still fluid. Case in point, just yesterday we learned that ICE is reversing the termination of international students. The letter from Martin may be another round of bluster with little substance. – robertsky (talk) 01:28, 27 April 2025 (UTC)
- (From French Wikipedia.) Obviously what happens to the WMF affects and concerns us too, very much. — Jules* talk 10:51, 27 April 2025 (UTC)
- If someone wants to seriously propose a blackout, they'll need to begin with a fully formed proposal that clearly lays out a very credible existential threat to the English Wikipedia, supported by reliable sources. Basically your proposal needs to be Featured Article quality when you first post the RFC. And you need to post it far enough ahead of time that consensus has time to happen. Keep in mind that you'll have to convince or outvote those who believe Wikipedia should be entirely apolitical even in the face of an existential threat, people from other countries who aren't familiar with US politics or culture wars, and the opposite side in the US culture wars (yes, there are such people here). And hope that someone else doesn't decide to post a "stub-class" proposal while you're preparing your FA-class proposal, and thereby poison the well. Anomie⚔ 13:46, 26 April 2025 (UTC)
- @Victorgrigas For now the threats are to the Wikimedia Foundation, not Wikipedia. As a consequence of these I suspect that the Wikimedia Foundation may indeed threaten Wikipedia freedomness and neutral point of view at some point to comply with these demands, but that's another story and we'll deal with that if it eventually comes to that. Darwin Ahoy! 14:37, 26 April 2025 (UTC)
- There is zero chance that the WMF is going to "threaten Wikipedia freedomness". There's zero (zero!) support for that from any staff or board members. Keep in mind that, when being attacked, one of the thing that the attackers usually want is for the attackees to turn on each other for no reason. We can be unified because we are unified. Jimbo Wales (talk) 10:04, 29 April 2025 (UTC)
- Well, there's a lot of overreaction here. Nobody has made any existencial threat to Wikipedia, only that it may not be suitable to get a tax exemption. A protest to keep a taxes priviledge may actually have the opposite effect as the one expected. It may be better to let the WMF deal with this behind courtains, and if it can't be done and the WMF looses the tax exemption... just accept it and pay the taxes. --Cambalachero (talk) 00:47, 27 April 2025 (UTC)
- We shouldn't kid ourselves into thinking that this is just a normal governmental inquiry into tax status. It's coming from the same motivations as the attacks on universities, the press, and law firms: the motivation to shut down any source of honest, unbiased information that goes against the Trump administration's preferred narrative. But it's also true that we shouldn't take any reckless, knee-jerk actions. We should be deliberative and thoughtful, and respond only in well-considered ways. --Tryptofish (talk) 00:56, 27 April 2025 (UTC)
- I agree, but cambalachero has a good point. It won't look good for Wikipedia to protest for a tax exemption. While I want it to have a tax exemption, out of context it sounds kinda weird. I do hope that WMF will be able to deal with it however, since that will make everybody's lives a thousand times easier. I do however trust that, no matter what, we'll survive. I feel as if Wikipedia has likely survived much worse threats than this. It won't be fun while all of this lasts, but i'm optimistic that things will get better. They always do, and they always will. But maybe that's just hopeful thinking. Gaismagorm (talk) 01:33, 27 April 2025 (UTC)
- It is important to note that other right-wing attacks on wikipedia turned out to be essentially nothing (such as the heritage foundations recent scheme, which as far as I can tell hasn't happened, and I don't think it will). Once again, this is likely just me trying to remain optimistic so I can remain sane. Gaismagorm (talk) 01:35, 27 April 2025 (UTC)
- I definitely agree that protesting framed in terms of tax status would be politically tin-eared. As for Heritage, it hasn't happened yet; this may be where it starts. And as for one's sanity, me too. --Tryptofish (talk) 01:44, 27 April 2025 (UTC)
- Yep, always assuming the worst case scenario is narrow-minded by definition. We need to respond to the scenario in front of us, not the hypothetical scenario that sounds the most dramatic. Addiction to pessimism porn is more harmful than most psychological dependencies. Thebiguglyalien (talk) 🛸 02:32, 27 April 2025 (UTC)
- And the scenario in front of us does not require any response from en.wiki. Whether the WMF will issue a response on their end is up to them. CMD (talk) 02:44, 27 April 2025 (UTC)
- The larger issue isn't whether or not the Wikimedia Foundation has to pay taxes; it's that donations will no longer qualify for a tax deduction. If I understand meta:Wikimedia Foundation Annual Plan/2023-2024/Finances § Budget numbers correctly, the vast amount of revenue comes from the fundraising campaign, so there will be significant effects on operations and funding model. isaacl (talk) 04:34, 27 April 2025 (UTC)
- It is WAY too soon for any response from the community. The WMF has an excellent legal team. They can deal with this. Don’t over-react. Blueboar (talk) 12:15, 27 April 2025 (UTC)
- I haven't advocated for any reaction from the community. I agree that the WMF is capable of deciding the next best steps. isaacl (talk) 14:31, 27 April 2025 (UTC)
- It is WAY too soon for any response from the community. The WMF has an excellent legal team. They can deal with this. Don’t over-react. Blueboar (talk) 12:15, 27 April 2025 (UTC)
- It is important to note that other right-wing attacks on wikipedia turned out to be essentially nothing (such as the heritage foundations recent scheme, which as far as I can tell hasn't happened, and I don't think it will). Once again, this is likely just me trying to remain optimistic so I can remain sane. Gaismagorm (talk) 01:35, 27 April 2025 (UTC)
- I agree, but cambalachero has a good point. It won't look good for Wikipedia to protest for a tax exemption. While I want it to have a tax exemption, out of context it sounds kinda weird. I do hope that WMF will be able to deal with it however, since that will make everybody's lives a thousand times easier. I do however trust that, no matter what, we'll survive. I feel as if Wikipedia has likely survived much worse threats than this. It won't be fun while all of this lasts, but i'm optimistic that things will get better. They always do, and they always will. But maybe that's just hopeful thinking. Gaismagorm (talk) 01:33, 27 April 2025 (UTC)
- We shouldn't kid ourselves into thinking that this is just a normal governmental inquiry into tax status. It's coming from the same motivations as the attacks on universities, the press, and law firms: the motivation to shut down any source of honest, unbiased information that goes against the Trump administration's preferred narrative. But it's also true that we shouldn't take any reckless, knee-jerk actions. We should be deliberative and thoughtful, and respond only in well-considered ways. --Tryptofish (talk) 00:56, 27 April 2025 (UTC)
- The reality is that at present both congress and the senate are politicaly non functional so convential campaining is unlikely to be effective. For the most part the best option is to try and keep a low enough profile that people lose interest. On a technical level ensuring there are overseas cold backups should probably be a thing.©Geni (talk) 12:03, 27 April 2025 (UTC)
at present both congress and the senate are politicaly non functional
- It's more complicated. While the US Congress (consisting of the House and the Senate) is indeed very polarized these days and gridlocked on most "big" topics, there is actually still quite a lot of bipartisan legislation going on under the radar - a phenomenon that has been called "secret congress".- And unfortunately (for us), these heartening examples of bipartisan consensus include various attempts to weaken Section 230 (a law which has been described as being essential for Wikipedia's existence), and similar efforts.
- In fact, just two days ago, right after you posted this comment, Congress passed a new internet law against the warnings of
groups such the Center for Democracy & Technology, the Authors Guild, Demand Progress Action, the Electronic Frontier Foundation (EFF), Fight for the Future, the Freedom of the Press Foundation, New America's Open Technology Institute, Public Knowledge, and TechFreedom
(to quote from TAKE_IT_DOWN_Act#Criticism) that its takedown provisions could be abused. Some commenters have specifically described Wikipedia as a website that could be affected by this: In general, government-mandated takedown systems are easily abused by private bad actors. (This primarily happens with “copystrike” extortion and censorship, which has grown out of mandatory takedown systems for copyright infringement.) [see also my recent Signpost article with some specific examples of Wikipedia articles affected by such spurious takedowns on Google]
More specifically, conservatives have signaled an interest in undercutting supposedly “liberal” platforms — Wikipedia in particular is frequently attacked by Musk and has been targeted by the Heritage Foundation. The Take It Down Act covers online platforms (with the exception of email and a few other carveouts) that “primarily [provide] a forum for user-generated content,” and while Wikipedia isn’t typically in the business of publishing nonconsensual nudes, it seems plausibly covered by some interpretations of the law. The FTC would probably have no compunctions about launching a punitive investigation if trolls start spamming it with deepfakes.
- (from an article in The Verge, with one internet liability expert - who heads a program on platform regulation at Stanford University - agreeing with The Verge about Wikipedia being a plausible target)
- Now, these observations are from early March and I don't know if the bill was improved since then, or how likely it is that the current US government will indeed try to (ab)use this new law against Wikipedia in this way.
- But my larger point is that even if one judges the current legal risk regarding 501(c)(3) status as low (see also Jimbo's comment, we might well see new laws soon that increase the attack surface greatly, and not just in the US.
- For Wikimedians interested in that kind of threat: The public policy mailing list is probably the most active forum about such issues.
- Regards, HaeB (talk) 04:51, 1 May 2025 (UTC)
- Considering the community just rejected a proposal for a blackout, which would have been in response to an Indian media conglomerate using lawfare to intimidate individual users and censor our content, I would honestly find it a bit insulting for us to propose a blackout over a letter questioning the foundation's tax-exempt status in the United States. This letter is certainly a bad sign of things to come, but let's be real here, it does not yet represent an active and present threat to us in the way that the ANI lawsuit does. We should absolutely be proactively considering how to react if things get worse, and if the political environment in the United States presents an active threat to the project's functioning, but this is really putting the cart before the horse. --Grnrchst (talk) 12:54, 27 April 2025 (UTC)
- We didn't blackout for the middle-eastern arrested editors either, but we did blackout for platform-wide threats. This is a giant fiscal threat to WMF and that means super-reduced operations. That said I want to see how this situation progresses first especially with WMF legal. Aaron Liu (talk) 23:49, 27 April 2025 (UTC)
- I think talks of a blackout right now are an over reaction. For now, we should assume good faith and see this as only a threat to our tax status. However, as many other wikipedians have stated, it wouldn't be unprecedented for the current US administration to challenge the freedom of the content on Wikipedia or the safety of her editors. For now we need to stay calm, hope for the best, and be prepared for the worst without expecting it. It would be far more productive to use this time to figure out ways to increase the anonymity of both the readers and editors, particularly those of contentious topics. mgjertson (talk) (contribs) 14:36, 30 April 2025 (UTC)
Our job is to educate and teach, not to protest. IF this reaches a point where there is a need to formally react to any of this, a banner explaining the situation might be considered, but it should not take the form of a blackout/protest banner. Blueboar (talk) 13:08, 27 April 2025 (UTC)
- If something threatens our ability to educate and teach, we absolutely should be protesting against that. --Grnrchst (talk) 13:58, 27 April 2025 (UTC)
- Blueboar has said elsewhere on related topics that we should all wait 10 years to consider reacting to the current situation. Viriditas (talk) 22:46, 27 April 2025 (UTC)
- They might have been referring to WP:10YEARS, which is about covering things in articles, not project activity. Thebiguglyalien (talk) 🛸 22:53, 27 April 2025 (UTC)
- Blueboar has said elsewhere on related topics that we should all wait 10 years to consider reacting to the current situation. Viriditas (talk) 22:46, 27 April 2025 (UTC)
For crying out loud - we don't need to blackout the site because questions were raised about tax exempt status of the WMF. In no way does the WMF potentially needing to pay taxes undermine the neutrality of the encyclopedia; before last November half of the editing base here was perpetually pissed at the WMF for wasting their money more than anything else. And we shouldn't forget that the Obama admin was targeting various nonprofits at one point; I don't think we got riled up over that did we? American politics are cyclical in many ways. Yes, there are some things that occurring in the USA that concern me. And yes, I'm aware that the editing base of enwiki skews to the left. And yes, a lot of more conservative media sources aren't near as reliable as they use to be (there's a reason Fox doesn't run the "Fair and balanced" tagline anymore ...). But we need to be really careful that we don't create an editing environment in which 49.8% of the American public doesn't feel that they can contribute. Hog Farm Talk 22:42, 27 April 2025 (UTC)
- I'll keep saying it until it sinks in: a lot of editors here are increasingly falling into pessimism porn addictions. The explanation there describes many of these "everybody panic" posts we've been seeing. Thebiguglyalien (talk) 🛸 22:56, 27 April 2025 (UTC)
- It's a matter of finding a middle ground between learned helplessness and over-reacting. We shouldn't do things that are premature, or that will backfire on us, but we also shouldn't ignore reality. We can look at what has already happened, as a matter of public record, to other institutions that have been targeted in the same way. US universities provide some good examples. Initially, universities that were accused of antisemitism (as opposed to harboring people hostile to the US, which is what we are accused of) made the mistake of trying to keep their heads down and placate the Trump administration. Their grant funding got cut anyway, and the demands just increased. These demands included having administration personnel monitor curricula and hiring. Translate to Wikipedia, and that would be administration officials getting to rule on what our content says, and which editors can be blocked. Now that Harvard has announced that they will fight back in court, there's a greater sense that things will play out in the courts over time, and that reason can prevail. We need to recognize that this is the path we are facing, too. It isn't about whether WMF will pay taxes. It's about whether we will allow ourselves to stop being a reliable encyclopedia, something we will not allow. We shouldn't freak out, but we need to be realistic. --Tryptofish (talk) 23:13, 27 April 2025 (UTC)
- I don't see a scenario where a government has full editorial control over Wikipedia, but otherwise I think we're in agreement on the issue. Thebiguglyalien (talk) 🛸 23:17, 27 April 2025 (UTC)
- Thanks. I only see that scenario happening if we let it happen. But I do see a realistic chance of them trying to get us to do it. --Tryptofish (talk) 23:20, 27 April 2025 (UTC)
- My genuine concern is that we're going to overcorrect and end up taking a general political stance that is incompatible with encyclopedic goals. I personally can't imagine a situation in which the US court of public opinion or the US court system is going to side against the general principals of Wikipedia if we stick to them. If we get to a point where WP:NPOV, WP:RGW, etc. get replaced by a political shibboleth in our response to this, or we create and accept WP:NOREPUBLICANS to go alongside WP:NONAZIS and Wikipedia:No Confederates, then we've 1) lost our credibility as a neutral encyclopedia and 2) will lose a good chunk of said court of public opinion and end up destroying the encyclopedia. Hog Farm Talk 23:48, 27 April 2025 (UTC)
- My concern is that a handful of people here want this to happen. You'll see people around here who think we have a moral obligation to take a stand on political issues using Wikipedia as a platform. Then there are also the people who are only WP:HERE to try and push a Trumpist viewpoint into articles. Higher up I mentioned the same thing about principles being the most useful path forward. I also suggested a discussion about how to address people who want to violate these principles for their own political ends, and I have since started a discussion at Wikipedia talk:Arbitration/Requests/Enforcement#Clarification on POV pushing and AE action. Thebiguglyalien (talk) 🛸 00:07, 28 April 2025 (UTC)
- Completely agree, we should be taking this constructively and look at ways we can do better. Unfortunately, some of his criticisms and questions are somewhat valid. Kowal2701 (talk) 21:59, 28 April 2025 (UTC)
- I share Hog Farm's view. We need to maintain our credibility as a reliable encyclopedia that isn't distorting the facts, in order to maintain (or recapture) the political upper hand. Giving in to feel-good retribution in mainspace will assuredly backfire. But I also believe that editors should feel free to speak plainly in the behind-the-scenes namespaces. --Tryptofish (talk) 21:46, 28 April 2025 (UTC)
- My concern is that a handful of people here want this to happen. You'll see people around here who think we have a moral obligation to take a stand on political issues using Wikipedia as a platform. Then there are also the people who are only WP:HERE to try and push a Trumpist viewpoint into articles. Higher up I mentioned the same thing about principles being the most useful path forward. I also suggested a discussion about how to address people who want to violate these principles for their own political ends, and I have since started a discussion at Wikipedia talk:Arbitration/Requests/Enforcement#Clarification on POV pushing and AE action. Thebiguglyalien (talk) 🛸 00:07, 28 April 2025 (UTC)
- My genuine concern is that we're going to overcorrect and end up taking a general political stance that is incompatible with encyclopedic goals. I personally can't imagine a situation in which the US court of public opinion or the US court system is going to side against the general principals of Wikipedia if we stick to them. If we get to a point where WP:NPOV, WP:RGW, etc. get replaced by a political shibboleth in our response to this, or we create and accept WP:NOREPUBLICANS to go alongside WP:NONAZIS and Wikipedia:No Confederates, then we've 1) lost our credibility as a neutral encyclopedia and 2) will lose a good chunk of said court of public opinion and end up destroying the encyclopedia. Hog Farm Talk 23:48, 27 April 2025 (UTC)
- Thanks. I only see that scenario happening if we let it happen. But I do see a realistic chance of them trying to get us to do it. --Tryptofish (talk) 23:20, 27 April 2025 (UTC)
- Agree 100% with Tryptofish. --Grnrchst (talk) 08:20, 28 April 2025 (UTC)
- I don't see a scenario where a government has full editorial control over Wikipedia, but otherwise I think we're in agreement on the issue. Thebiguglyalien (talk) 🛸 23:17, 27 April 2025 (UTC)
- It's a matter of finding a middle ground between learned helplessness and over-reacting. We shouldn't do things that are premature, or that will backfire on us, but we also shouldn't ignore reality. We can look at what has already happened, as a matter of public record, to other institutions that have been targeted in the same way. US universities provide some good examples. Initially, universities that were accused of antisemitism (as opposed to harboring people hostile to the US, which is what we are accused of) made the mistake of trying to keep their heads down and placate the Trump administration. Their grant funding got cut anyway, and the demands just increased. These demands included having administration personnel monitor curricula and hiring. Translate to Wikipedia, and that would be administration officials getting to rule on what our content says, and which editors can be blocked. Now that Harvard has announced that they will fight back in court, there's a greater sense that things will play out in the courts over time, and that reason can prevail. We need to recognize that this is the path we are facing, too. It isn't about whether WMF will pay taxes. It's about whether we will allow ourselves to stop being a reliable encyclopedia, something we will not allow. We shouldn't freak out, but we need to be realistic. --Tryptofish (talk) 23:13, 27 April 2025 (UTC)
I'm definitely on the side of wanting WMF Legal to be able to take the lead here. We shouldn't do anything that would undercut their effectiveness. I also think that any actions we eventually take should play to our political strengths, and not play into the hands of those threatening us. I'm not wild about a blackout, because depriving readers of the information we provide is actually what the Trump administration wants, so why should we do it for them? I like the idea that Blueboar mentioned, of an informative banner. If members of the public come here, and still find the information they want from us, but they first have to get past a conspicuous banner (maybe one that you cannot make disappear by clicking an x) that tells them of the situation and points them to ways to object to what's happening, that could be very effective at getting public opinion on our side. Something else that we should all try to do is to stay faithful to our values in terms of NPOV and the like. The more we continue to insist on accurate and neutral content, correcting errors as we find them, and not engaging in WP:RGW in mainspace, the more credibility we have, and the weaker our opponent's case will be. --Tryptofish (talk) 22:59, 27 April 2025 (UTC)
- Meh. I don't see any particular efforts in any direction we might take as having any gravitas. The reality of this situation is that what is asked for by the letter is impossible to achieve by the deadline imposed. The WMF doesn't have the pockets to fight this sort of thing like Harvard does, and certainly doesn't have the pockets to weather the storm that's coming. The WMF will lose its 501c3 status. It's essentially a given. How much of an impact will that have on donors? Who knows, but the status will be gone in three weeks time. No imagined solution generated from this or any other page where this is being discussed is going to change that reality. --Hammersoft (talk) 19:02, 28 April 2025 (UTC)
- Apes together strong. We should join with Harvard in a coalition of the willing. Viriditas (talk) 21:32, 28 April 2025 (UTC)
- I'm all for facing reality, but I think it's maladaptive and frankly craven to adopt the position that we should just take it and say thank you can we please have some more. I understand and sympathize with how unpleasant it feels to deal with government-by-bullying, and how that can make editors just want to rationalize inaction. But rationalizing is what it is, and that's facing reality, too. Editors (and indeed people in the "real world") should feel self-confident enough to call this what it is. Now that said, I also expect that it's quite likely that the tax status is going to get pulled. I also expect that, subsequently, it will end up in litigation, and that will go on for a long time and have unpredictable aspects. Simultaneously, I expect further demands, that will go beyond tax matters, along with very public efforts to discredit Wikipedia and our content. --Tryptofish (talk) 21:57, 28 April 2025 (UTC)
- Again, *shrug*. I'm sorry, but I do not see any reasonable way that Wikipedia can defend itself against this other than highly expensive (as in millions of $) expenditure in court. That's the only venue that will matter. The government in question will not give any thought whatsoever by what we say here. Even if a million editors all screamed out at once, it would have as much effect as a single rain drop would have in the Gobi Desert. They simply won't see it. It's meaningless flapping of our wings in the hope that some wild butterfly effect would somehow cause an earthquake, hurricane, and blizzard to all happen in D.C. at the same time. It's just fantasy. I'm not saying this against you personally, but against any idea that we can somehow stop this. We can't. The best path forward is how to structure the project despite the serious damage this administration is about to inflict on it. Stopping it is impossible. --Hammersoft (talk) 01:03, 29 April 2025 (UTC)
Stopping it is impossible
To quote a little green Muppet, "that is why you fail". Plan for the worst, yes. But saying something is impossible and acting accordingly is not productive. - The Bushranger One ping only 05:44, 29 April 2025 (UTC)- Fortunately, I don't guide my life by little green muppets, whether they have cute ears or none at all :) Seriously though; standing in front of an oncoming 100mph avalanche with a shovel saying "I got this!" isn't productive either. The powers that be in D.C. will not care if we all black out our userpages, take down every article, or go on an editing strike. It all serves their purposes. Even if it all was directly against their purposes, they still wouldn't care. There simply isn't any reasonable method by which we can affect the outcome of this. --Hammersoft (talk) 17:55, 29 April 2025 (UTC)
- Before I hovered the link, I wondered when it was that Kermit the Frog turned zen. ⁓ Pelagic ( messages ) 06:24, 3 May 2025 (UTC)
- Again, *shrug*. I'm sorry, but I do not see any reasonable way that Wikipedia can defend itself against this other than highly expensive (as in millions of $) expenditure in court. That's the only venue that will matter. The government in question will not give any thought whatsoever by what we say here. Even if a million editors all screamed out at once, it would have as much effect as a single rain drop would have in the Gobi Desert. They simply won't see it. It's meaningless flapping of our wings in the hope that some wild butterfly effect would somehow cause an earthquake, hurricane, and blizzard to all happen in D.C. at the same time. It's just fantasy. I'm not saying this against you personally, but against any idea that we can somehow stop this. We can't. The best path forward is how to structure the project despite the serious damage this administration is about to inflict on it. Stopping it is impossible. --Hammersoft (talk) 01:03, 29 April 2025 (UTC)
- In addition to
depriving readers of the information we provide is actually what the Trump administration wants
, there's the simple fact that a blackout is pointless. It certainly won't change any of our opinions. And those on the 'other side', it won't change theirs either. It'll only affect the people in the middle - by making them pissed off at us. - The Bushranger One ping only 22:15, 28 April 2025 (UTC)- It would demonstrate two things: (1) that Wikipedia is an American organisation and not an international one; and (2) that it engages in political lobbying against the interests of the US government. Hawkeye7 (discuss) 22:28, 28 April 2025 (UTC)
- I'm against a blackout, too. But I feel the need to clarify that "the interests of the US government" are neither "the interests of the Trump administration" nor "violation of the First Amendment guarantee of free speech". But I agree that members of the public may very well see a blackout in the way that you describe. --Tryptofish (talk) 22:33, 28 April 2025 (UTC)
- It would demonstrate two things: (1) that Wikipedia is an American organisation and not an international one; and (2) that it engages in political lobbying against the interests of the US government. Hawkeye7 (discuss) 22:28, 28 April 2025 (UTC)
Just for clarification, CentralNotice banners can be targeted to specific countries/languages/projects only. Blacking out for everyone because of things in a country would be overly invasive. (Not that I think we should do anything with banners, letter and the like right now. WMF Legal will handle this with the highest expertise they have.) Best, —DerHexer (Talk) 12:44, 29 April 2025 (UTC)
- To add: m:Project-wide protests has a (possibly incomplete) list of past actions of this kind, and (to come back to the initial question above) m:English Wikipedia anti-SOPA blackout links to various detailed descriptions of what was done in that case.
- (And agreed that any action of this kind in the present matter would seem premature at this point, especially before having heard from WMF Legal.)
- Regards, HaeB (talk) 16:18, 1 May 2025 (UTC)
Possibly less constructive musings
Some thoughts, either thinking outside the box or desperately needed tragi-comic relief.
- Give up 401(c)(3) status. Reincorporate elsewhere (Liechtenstein?) and move all financial assets offshore. This will have zero impact on contributions from non-US sources, and US sources may donate 20%-30& less.
- Create a MAGA Wikipedia, en.maga.wikipedia.org, with its own rules. Let that be how the Wikimedia Foundation is able to demonstrate that it accommodates all views. Fans of "separate but equal" will embrace this. (Hide it from search engines.)
Feel free to add yours. Largoplazo (talk) 22:28, 27 April 2025 (UTC)
- Oppose both. Your second option is, in fact, precisely the central heart of this dispute. If you go on to Twitter right now (or any other right-wing forum) you will quickly discover that the most shared or viewed discussions on this topic are concerned with this very problem. MAGA believes that Wikipedia articles are hostile to conservatism because Wikipedia doesn't entertain or accept alternate facts or baseless conspiracy theories and doesn't use or rely on poor unreliable sourcing like "Ron Vara". That's what this is all about, no more, no less. Viriditas (talk) 22:43, 27 April 2025 (UTC)
- Okay, but hear me out here. It would be hilarious if we beat conservapedia at their own game. Especially if the second option is taken such a hilarious extreme where it rolls back to satire. Gaismagorm (talk) 23:39, 27 April 2025 (UTC)
- No action needed on our part. See Conservapedia's article (permalink) about The Room. — Newslinger talk 01:23, 28 April 2025 (UTC)
- See also Conservapedia's essay "Greatest Conservative Songs" (permalink). — Newslinger talk 01:32, 28 April 2025 (UTC)
- The most concerning thing about Conservapedia at the moment is that it has entirely embraced Putinism and Orbánism, two styles of government that are behind the push to extend the reach of an autocratic state into education and private industry. Viriditas (talk) 01:38, 28 April 2025 (UTC)
- Wow. That's priceless. I love how the "legacy" section smoothly transitions into an explanation of social conservative, centre-right politics. Cremastra talk 23:00, 28 April 2025 (UTC)
- See also Conservapedia's essay "Greatest Conservative Songs" (permalink). — Newslinger talk 01:32, 28 April 2025 (UTC)
- Conservapedia is unintentional satire, but they are neither aware of it or understand why it is satire. I mean, let’s not forget, they literally invented one of the most famous memes on the Internet: Supply Side Jesus riding a dinosaur. And they were dead serious about it at the time. Viriditas (talk) 01:25, 28 April 2025 (UTC)
- No action needed on our part. See Conservapedia's article (permalink) about The Room. — Newslinger talk 01:23, 28 April 2025 (UTC)
- Okay, but hear me out here. It would be hilarious if we beat conservapedia at their own game. Especially if the second option is taken such a hilarious extreme where it rolls back to satire. Gaismagorm (talk) 23:39, 27 April 2025 (UTC)
lmao that so quintessentially embodies segregation. But to make it nominally equal instead of "arbitrarily" silenced, it should probably be on a separate domain (magawikipedia.org?) the WMF registers through ICANN; as there are no links to that domain, it would not appear on search engines for quite a long time. Aaron Liu (talk) 23:47, 27 April 2025 (UTC)Fans of "separate but equal" will embrace this. (Hide it from search engines.)
- This is a deeply unconstructive proposal. This is not the anti-MAGA encyclopaedia. Our articles are not supposed to push any particular political theory. We are supposed to be (in article space) neutral even on the topic of Wikipedia. CMD (talk) 03:10, 28 April 2025 (UTC)
- Basically a giant WP:POVFORK Kowal2701 (talk) 22:00, 28 April 2025 (UTC)
- I agree but at some point you have to realize they don't know or care quite frankly. Anything that challenges their worldview is seen as unfairly biased against them and in turn against our neutrality. Providing them their own Wikipedia not only shows that we aren't biased against them, but it gives us a way to show that our policies have merit in keeping a reliable encyclopedia since a maga Wikipedia would almost certainly betray the ideals that Wikipedia stands on and end up creating an objectively worse encyclopedia because of it mgjertson (talk) (contribs) 14:47, 30 April 2025 (UTC)
- I know "they" (Ed Martin?) don't care, but for that reason they wouldn't care about any proposed solution. Creating a second Wikipedia would not show that this Wikipedia is unbiased, it would heavily imply that this Wikipedia is not the place for the target group. As Kowal2701 says, a giant POVFORK. And while they don't care, we do care, we want to build an accurate and neutral Wikipedia, and we should want that whether Ed Martin approves or disapproves. CMD (talk) 15:04, 30 April 2025 (UTC)
- Sartre had the number of the Ed Martins of the world all the way back in 1946. And, no, there is nothing that Wikipedia could do to persuade him that we are sufficiently neutral because he does not seek a neutral Wikipedia. He seeks a subservient, cowed, compliant Wikipedia. Simonm223 (talk) 19:24, 30 April 2025 (UTC)
- That's what I think. Carlstak (talk) 19:36, 30 April 2025 (UTC)
- Sartre had the number of the Ed Martins of the world all the way back in 1946. And, no, there is nothing that Wikipedia could do to persuade him that we are sufficiently neutral because he does not seek a neutral Wikipedia. He seeks a subservient, cowed, compliant Wikipedia. Simonm223 (talk) 19:24, 30 April 2025 (UTC)
- I know "they" (Ed Martin?) don't care, but for that reason they wouldn't care about any proposed solution. Creating a second Wikipedia would not show that this Wikipedia is unbiased, it would heavily imply that this Wikipedia is not the place for the target group. As Kowal2701 says, a giant POVFORK. And while they don't care, we do care, we want to build an accurate and neutral Wikipedia, and we should want that whether Ed Martin approves or disapproves. CMD (talk) 15:04, 30 April 2025 (UTC)
Publicity
The best way I can think of to fight this is publicity. Wikipedia must have friends in high places that value Wikipedia. How can we harness that to get the message to the Non-MAGA American public that the government is trying to kill it? Doug Weller talk 07:34, 29 April 2025 (UTC)
- Someone at WMF should call liberal Illinois Gov. JB Pritzker. I bet he'll publicize that message with a thunderous speech. He is a masterful orator, and, he's a billionaire. Carlstak (talk) 17:12, 29 April 2025 (UTC)
- I disagree, until there are actual damages we would just look petty and partisan... I also disagree that the government is trying to kill wikipedia, that seems a mite hyperbolic given the evidence we have. Horse Eye's Back (talk) 17:21, 29 April 2025 (UTC)
- Basically this. A public reaction will only be spun as us being defensive and used as evidence that we're trying to hide "our agenda". Keep calm and carry on. GMGtalk 17:28, 29 April 2025 (UTC)
- I think the tilting point will be if Martin actually launches an investigation. Then it's time for thunderous speeches. Carlstak (talk) 17:36, 29 April 2025 (UTC)
- I was thinking more of media personalities, not policians. Doug Weller talk 17:42, 29 April 2025 (UTC)
- @Doug Weller: If the situation escalates, I would be willing to reach out to people I've been in casual contact with. However, I don't see a reason to do so at this time without a more tangible threat and responding call-to-action. –MJL ‐Talk‐☖ 18:05, 29 April 2025 (UTC)
- Yes, too early. Doug Weller talk 18:14, 29 April 2025 (UTC)
- @Doug Weller: If the situation escalates, I would be willing to reach out to people I've been in casual contact with. However, I don't see a reason to do so at this time without a more tangible threat and responding call-to-action. –MJL ‐Talk‐☖ 18:05, 29 April 2025 (UTC)
- That still seems way to early... At least wait until an actual court case has been adjudicated. Horse Eye's Back (talk) 17:46, 29 April 2025 (UTC)
- Adjudicated or initiated? Doug Weller talk 18:15, 29 April 2025 (UTC)
- Adjudicated, I don't believe that we should resort to the court of public opinion before at least attempting the actual courts (which its not even clear this will reach, the AG seems to be on a fishing expedition). We have the high ground, action is not to our benefit. Horse Eye's Back (talk) 18:35, 29 April 2025 (UTC)
- I take your point. Doug Weller talk 18:57, 29 April 2025 (UTC)
- Well, if it gets to that stage, I would expect a lot of publicity in the sane-washing press and reality-based media alike. Carlstak (talk) 19:41, 29 April 2025 (UTC)
- Adjudicated, I don't believe that we should resort to the court of public opinion before at least attempting the actual courts (which its not even clear this will reach, the AG seems to be on a fishing expedition). We have the high ground, action is not to our benefit. Horse Eye's Back (talk) 18:35, 29 April 2025 (UTC)
- Adjudicated or initiated? Doug Weller talk 18:15, 29 April 2025 (UTC)
- I was thinking more of media personalities, not policians. Doug Weller talk 17:42, 29 April 2025 (UTC)
- I think the tilting point will be if Martin actually launches an investigation. Then it's time for thunderous speeches. Carlstak (talk) 17:36, 29 April 2025 (UTC)
- Basically this. A public reaction will only be spun as us being defensive and used as evidence that we're trying to hide "our agenda". Keep calm and carry on. GMGtalk 17:28, 29 April 2025 (UTC)
- Ed Martin is basically just Milo Yiannopoulos with a law degree. Ignore. Partofthemachine (talk) 04:35, 30 April 2025 (UTC)
How about we create and publicize initiatives to find systemic fixes to en wikipedia's bias on US politics-related topics? Then we'd probably get less of this crap. Also, while trying to be totally unbiased is impossible to define much less achieve, when it gets to the point where it degrades and distorts the informativeness of the those articles (which it has), some such improvements would also align with our en Wikipedia mission which is to offer quality informative articles. Sincerely, North8000 (talk) 18:47, 30 April 2025 (UTC)
- I agree, but I don't think the things the administration considers biases are actually things that need to be fixed. Obviously, there is some left-leaning bias in some articles, but not enough to warrant a removal of our tax exempt status. Gaismagorm (talk) 19:06, 30 April 2025 (UTC)
- I don't think that would work; because a truly neutral perspective on US politics would be far more critical of the United States and most of its political class than what we have. I mean look at the farce that is Elon Musk salute controversy where the first explanation we explore in the body is that having autism makes one sieg heil. Much of this comes from treating American newsmedia as if it were consistently reliable for building an encyclopedia. But being neutral and calling more American things that quack ducks would just further infuriate the Trump regime.
- The problem that Wikipedia faces is that the far-right doesn't care how neutral we claim to be. They don't want our neutrality. They want our submission. Simonm223 (talk) 19:09, 30 April 2025 (UTC)
- The louder and more irrational critics will never be satisfied, but the average reader is going to have a good enough nose for bullshit that we could substantially improve our public image by addressing NPOV violations. Unfortunately, I have little hope that we can convince AMPOL editors that it's actually not good content work to put "is a far-right conspiracy theorist" in a lead or a 20,000 byte "controversies" section that would make WP:BALASP cry (the latter being something that's permeated the project well beyond AMPOL—I've spent the last few months trying to bail water out of that ship and would love to talk shop if anyone wants to help out). Thebiguglyalien (talk) 🛸 19:12, 30 April 2025 (UTC)
- I think it would help. The widespread criticism of en Wikipedia bias is almost certainly a big factor that led to this. BTW my interest is more in fixing the systemic causes just to move the needle a bit so it's not so bad that causes the article to be distorted to the point of degrading and distorting the informativeness of the article. Rather than the elusive goal of defining and being unbiased. North8000 (talk) 21:12, 30 April 2025 (UTC)
- Except the idea Wikipedia has a left-wing systematic bias is just an artifact of the skewed American Overton window that treats anything left of neoliberalism as hyper-Lenin. A neutral encyclopedia would be more critical of the Trump regime and would, for example, not prevaricate over its ideology as if it was in dispute. Instead we have a Wikipedia with a pervasive center-right bias. Those of us on the left are thus in the unenviable position of being told neutrality can only be achieved by further marginalizing left-wing perspectives and that we should swallow this as somehow neutral? This path only leads to there being two Conservapedias instead of just one. Simonm223 (talk) 21:26, 30 April 2025 (UTC)
Except the idea Wikipedia has a left-wing systematic bias is just an artifact of the skewed American Overton window that treats anything left of neoliberalism as hyper-Lenin.
Yes, this. In their (the Americans') last election, I would say that while their two parties are obviously fairly big-tent, the Republicans struck me as far-right and the Democrats as centre-right to centre. But apparently the latter are considered a left-of-centre party in the U.S. American politics have shifted so far rightwards that trying to assess Wikipedia bias through that lens is hopeless. Cremastra talk 21:42, 30 April 2025 (UTC)- A few notes. General left wing / right wing bias is very different and even harder to define than US Politics bias which I think is the real issue. And my focus is more on where it degrades and distorts the factual coverage and informativeness of articles (and we do have lot of problems there) rather than what types of obvious (op ed type) criticisms or praise gets in there. As a tiny example, describing a conservative's positions on something using only vague inaccurate pejoratives from left leaning media instead of providing detailed factual coverage of that. And my interest is on fixable systemic contributors to the problem. North8000 (talk) 21:44, 30 April 2025 (UTC)
- Easy fix to that: prohibit newspapers for politics articles. Wikipedia will never do it because newspapers are convenient sources of (low-quality) information but requiring higher-quality sources for political articles is something that would probably get leftist editors on-board - unlike declaring that the NYT is the second coming of Friedrich Engels. Simonm223 (talk) 23:31, 30 April 2025 (UTC)
- Perhaps we could get something out of this whole brouhaha by tightening up our standards a little bit. There are so many steps we could take: moving away from newspaper sources, disallowing criticism and controversy sections on BLPs, strictly enforcing WP:BALASP/WP:ONUS/WP:WTW (especially WP:LABEL), considering sources with non-impartial tones to be less reliable, and (god please) tban people when the a large portion of their editing is showing up to support a given side in a CTOP. I'd take any of these as a win for Wikipedia. Thebiguglyalien (talk) 🛸 23:55, 30 April 2025 (UTC)
- Let me know when this is officially proposed so I can strongly oppose it. Good journalism and newspaper coverage is under attack by the right, and your proposal would support their goals. Viriditas (talk) 00:09, 1 May 2025 (UTC)
- Strongly agree. Carlstak (talk) 00:24, 1 May 2025 (UTC)
- You should not be citing journalism on Wikipedia. In most cases it's a WP:PRIMARY source. Thebiguglyalien (talk) 🛸 00:25, 1 May 2025 (UTC)
- I'm sorry? It being a WP:PRIMARY source is by no means an interdiction against it.
- In any case, I dispute that claim; WP:PRIMARY notes that
[p]rimary sources are original materials that are close to an event, and are often accounts written by people who are directly involved. They offer an insider's view of an event, a period of history, a work of art, a political decision, and so on.
The vast majority of good newspaper articles are not this. They are written objectively from a person near the event but not "directly involved". Cremastra talk 00:32, 1 May 2025 (UTC)- WP:NEWSPRIMARY explains this. See also WP:RSBREAKING, which says
All breaking news stories, without exception, are primary sources, and must be treated with caution
. I also have my own essay explaining why citing contemporary coverage is poor form: User:Thebiguglyalien/Avoid contemporary sources. And this isn't considering investigative journalism and opinion-based writing, which are primary for the findings or views of the author. Thebiguglyalien (talk) 🛸 00:35, 1 May 2025 (UTC)- WP:NEWSPRIMARY is just a redirect to WP:USEPRIMARY (WP:Identifying and using primary sources), an explanatory essay which also says:
- Again, "Primary" is not another way to spell "bad". Just because most newspaper articles are primary sources does not mean that these articles are not reliable and often highly desirable independent sources.
- Carlstak (talk) 01:24, 1 May 2025 (UTC)
- And I'll use them in these cases. The problem is when they're used to determine weight or indicate that something should be included in an article. I explained this in the essay I linked. See also the aptly-named WP:FART. I find the idea that we should use newspaper coverage to protect it from "attack by the right" to be WP:RGW, WP:NOTHERE style behavior. Thebiguglyalien (talk) 🛸 01:33, 1 May 2025 (UTC)
- Nobody has ever said anywhere that we should use newspapers to protect it from right wing attacks, nor can I possibly comprehend how you got that from my comment. Journalism is under total attack by right wing billionares. They have decimated local news coverage in most US communities and have taken over most mainstream news outlets. The "left", liberalism, and left-wing voices and opinions have almost zero representation and cannot be said to be a threat to the right anywhere in the US. This whole line of reasoning is part of the "liberal media" myth which began in the 1970s with the Powell memo and continues today with the enemy of the people lie espoused by the current administration. Viriditas (talk) 02:26, 1 May 2025 (UTC)
- My opinion on newspapers is simply that they provide lower-quality information than peer-reviewed academic work and books published by academic presses. We should always prefer these sources but saying that in politics related articles often leads to significant protest. Simonm223 (talk) 11:40, 1 May 2025 (UTC)
- Nobody has ever said anywhere that we should use newspapers to protect it from right wing attacks, nor can I possibly comprehend how you got that from my comment. Journalism is under total attack by right wing billionares. They have decimated local news coverage in most US communities and have taken over most mainstream news outlets. The "left", liberalism, and left-wing voices and opinions have almost zero representation and cannot be said to be a threat to the right anywhere in the US. This whole line of reasoning is part of the "liberal media" myth which began in the 1970s with the Powell memo and continues today with the enemy of the people lie espoused by the current administration. Viriditas (talk) 02:26, 1 May 2025 (UTC)
- And I'll use them in these cases. The problem is when they're used to determine weight or indicate that something should be included in an article. I explained this in the essay I linked. See also the aptly-named WP:FART. I find the idea that we should use newspaper coverage to protect it from "attack by the right" to be WP:RGW, WP:NOTHERE style behavior. Thebiguglyalien (talk) 🛸 01:33, 1 May 2025 (UTC)
- WP:NEWSPRIMARY is just a redirect to WP:USEPRIMARY (WP:Identifying and using primary sources), an explanatory essay which also says:
- WP:NEWSPRIMARY explains this. See also WP:RSBREAKING, which says
- Let me know when this is officially proposed so I can strongly oppose it. Good journalism and newspaper coverage is under attack by the right, and your proposal would support their goals. Viriditas (talk) 00:09, 1 May 2025 (UTC)
- Perhaps we could get something out of this whole brouhaha by tightening up our standards a little bit. There are so many steps we could take: moving away from newspaper sources, disallowing criticism and controversy sections on BLPs, strictly enforcing WP:BALASP/WP:ONUS/WP:WTW (especially WP:LABEL), considering sources with non-impartial tones to be less reliable, and (god please) tban people when the a large portion of their editing is showing up to support a given side in a CTOP. I'd take any of these as a win for Wikipedia. Thebiguglyalien (talk) 🛸 23:55, 30 April 2025 (UTC)
- Easy fix to that: prohibit newspapers for politics articles. Wikipedia will never do it because newspapers are convenient sources of (low-quality) information but requiring higher-quality sources for political articles is something that would probably get leftist editors on-board - unlike declaring that the NYT is the second coming of Friedrich Engels. Simonm223 (talk) 23:31, 30 April 2025 (UTC)
- Except the idea Wikipedia has a left-wing systematic bias is just an artifact of the skewed American Overton window that treats anything left of neoliberalism as hyper-Lenin. A neutral encyclopedia would be more critical of the Trump regime and would, for example, not prevaricate over its ideology as if it was in dispute. Instead we have a Wikipedia with a pervasive center-right bias. Those of us on the left are thus in the unenviable position of being told neutrality can only be achieved by further marginalizing left-wing perspectives and that we should swallow this as somehow neutral? This path only leads to there being two Conservapedias instead of just one. Simonm223 (talk) 21:26, 30 April 2025 (UTC)
- I agree with North that you have the right of this (if the leads of all CTOP articles were mostly locked things would probably function better), but I would stress again that it would not affect the current situation at all. We should not pretend or give credence to the idea that this dispute is actually about our neutrality. CMD (talk) 00:53, 1 May 2025 (UTC)
- I think it would help. The widespread criticism of en Wikipedia bias is almost certainly a big factor that led to this. BTW my interest is more in fixing the systemic causes just to move the needle a bit so it's not so bad that causes the article to be distorted to the point of degrading and distorting the informativeness of the article. Rather than the elusive goal of defining and being unbiased. North8000 (talk) 21:12, 30 April 2025 (UTC)
A systemic fix will be structural and more complex that I can get into here but the gist of two items is:
- Get rid of the binary concept of a source being "wp:reliable" where they get the unconditional keys to the city for wikilawyers and those not in that club are unconditionally deprecated. . That club is determined by trappings which are those of legacy media and for not getting voted out / deprecated. And go more with actual reliability which is (context-specific) expertise and reliability with respect to the text which cited it (which is a wp:ver context)
- WP:weight was intended to apply to "two sides of an issue" coverage but has been hijacked by wikilawyering (in tandem with the wp:rs issue) to exclude coverage on all "I don't like it" items even if they are are not "two sides of an issue" type situations. Fix that.
Sincerely, North8000 (talk) 00:57, 1 May 2025 (UTC)
- (Not to distract from the wider point, but just a note that we have a quaternary concept of source reliability, although much of the spectrum does fall towards one end or the other. CMD (talk) 01:17, 1 May 2025 (UTC))
- If the letter from Martin were a good-faith expression of concern about us not getting neutrality right, I would see this as a discussion that I would be happy to have. For example: doing away with criticism sections in BLPs is something I could potentially support. (But also consider: Ted Kaczynski and Dzhokhar Tsarnaev have BLPs, too.) But let's not pretend that this is the case. The complaint isn't about neutrality. It's that we harbor persons who are trying to undermine the national interests of the US. And it isn't a constructive effort to correct what we might be getting wrong. It's a probably unconstitutional misuse of government authority to attempt to bully us into publishing content that would blatantly fail NPOV, but make some people in power happy. So let's drop this pretense that this is an occasion for us to fix some things we get wrong with NPOV. We could fix some things. I'd support fixing them. But that wouldn't stop the attempted coercion. And while I'd support fixing problems with our content, I'll strongly oppose any misguided attempts to change our policies in the hope that this would make the bullies leave us alone. --Tryptofish (talk) 21:06, 1 May 2025 (UTC)
- Then when is the time to fix things? I've been asking for these things for a long time, and if this is the tipping point that gets us talking about it, then I'll take it. Also, infamous and widely-hated people are where we should be most cautious about neutrality because that's where it's easiest to slip up and move away from WP:IMPARTIAL or WP:POVFORM. Thebiguglyalien (talk) 🛸 21:39, 1 May 2025 (UTC)
- There's nothing wrong with the time, any time. My concern is with the reasons. --Tryptofish (talk) 21:42, 1 May 2025 (UTC)
- Maybe the best option would be to start a new discussion elsewhere about whichever of these proposals we feel should have already been done, so we can brainstorm without the burden of... whatever all this is. Thebiguglyalien (talk) 🛸 21:47, 1 May 2025 (UTC)
- Elsewhere, definitely. And not because of Martin's letter. On the merits of the proposals. --Tryptofish (talk) 21:49, 1 May 2025 (UTC)
- Maybe the best option would be to start a new discussion elsewhere about whichever of these proposals we feel should have already been done, so we can brainstorm without the burden of... whatever all this is. Thebiguglyalien (talk) 🛸 21:47, 1 May 2025 (UTC)
- There's nothing wrong with the time, any time. My concern is with the reasons. --Tryptofish (talk) 21:42, 1 May 2025 (UTC)
- Then when is the time to fix things? I've been asking for these things for a long time, and if this is the tipping point that gets us talking about it, then I'll take it. Also, infamous and widely-hated people are where we should be most cautious about neutrality because that's where it's easiest to slip up and move away from WP:IMPARTIAL or WP:POVFORM. Thebiguglyalien (talk) 🛸 21:39, 1 May 2025 (UTC)
*Shrugs*
I think the best thing Wikipedia can do right now is to ignore this letter. If Mr. Martin actually wanted to enforce it, he would: decide what actually he wants to do, go to a court, and then actually convince judge to support his marginal legal theory. Even if he manages to do that, we would have more than enough time to react to everything that is going on. Hence, the best thing to do is to not feed this troll and just ignore him beyond a boilerplate "We reserve all rights under the law". The discussion about bias in Wikipedia that this has generated, from those who feel Wikipedia has a left or right wing bias is counterproduct, and if anything, unlikely to be true. Wikipedia, if anything has an intentional centrist bias with differing biases in specific topic areas that may be considered left, right, or other wing. I think the best thing Wikipedia can do is to hedge against any deterioration of speech conditions by decentralizing its operations. For example, moving operations to Switzerland, may lower the inherent risk created by needing to have a physical presence. Most of the other talk tends to fall either under pointless dooming, or self-motivated arguing. Allan Nonymous (talk) 16:10, 4 May 2025 (UTC)
The threat may be fizzling
With my thanks to Herostratus, who posted about this at Jimbotalk, there is now this news report that Ed Martin appears unlikely to receive Senate confirmation, and so his acting position as DC Attorney may soon be coming to an end: [7]. --Tryptofish (talk) 18:57, 6 May 2025 (UTC)
- Dang, nothing does happen. Gaismagorm (talk) 19:03, 6 May 2025 (UTC)
- See [8] Doug Weller talk 07:46, 7 May 2025 (UTC)
- This does appear to be another event that follows the trend of "Trumpist politician makes noise for several months, nothing happens, goes bust". Fantastic Mr. Fox 08:26, 7 May 2025 (UTC)
- See [8] Doug Weller talk 07:46, 7 May 2025 (UTC)
- No offense my friends, but anyone who thinks this is the end of threat that is coming for this project from those quarters has their head so deeply in the ground with wishful thinking, they should be wary of magma. This was never going to be more than a prelude to the fight, and the main event is, beyond the slightest shadow of a doubt, very close on the horizon. SnowRise let's rap 01:31, 10 May 2025 (UTC)
- Let's keep it civil. It's okay to be hopeful. Besides, as seen above this discussion is becoming purely political. Gaismagorm (talk) 12:29, 10 May 2025 (UTC)
- Nothing I said was meant to be unkind or even particularly sarcastic, and I'm a little surprised if it came off that way. If my intention had been a more acidic response, my comment would have consisted of one word: "Denial." As it is, my point was to emphasize that this not the end of this story. I'm sorry, there's just realistically no chance of that. And this community cannot afford to be laissez-faire about where this is headed if it wishes to be in any way prepared for what comes next. I understand the impulse to grasp at any indication that this problem will go away. It's an instinct most reasonable people in the world in this moment in the course of modern history are inclined to now and again. But it is rarely realistic, and certainly not in this instance. Make no mistake: this project is with absolutely on the list of perceived obstacles to the new order of information control. There can be no question about that. "Hope" then lays not in convincing ourselves and eachother that this problem will simply go away, but in preparing ourselves and being ready to face the coming challenges together. I am not trying to make anyone feel foolish for indulging in an instinct that is natural and presently commonplace. But a dose of cold water is absolutely needed here. SnowRise let's rap 19:01, 10 May 2025 (UTC)
- It's fine, I guess I'm a tad bit sensitive to this stuff. You are right though, we need to be ready for this kind of stuff. Gaismagorm (talk) 19:13, 10 May 2025 (UTC)
- Nothing I said was meant to be unkind or even particularly sarcastic, and I'm a little surprised if it came off that way. If my intention had been a more acidic response, my comment would have consisted of one word: "Denial." As it is, my point was to emphasize that this not the end of this story. I'm sorry, there's just realistically no chance of that. And this community cannot afford to be laissez-faire about where this is headed if it wishes to be in any way prepared for what comes next. I understand the impulse to grasp at any indication that this problem will go away. It's an instinct most reasonable people in the world in this moment in the course of modern history are inclined to now and again. But it is rarely realistic, and certainly not in this instance. Make no mistake: this project is with absolutely on the list of perceived obstacles to the new order of information control. There can be no question about that. "Hope" then lays not in convincing ourselves and eachother that this problem will simply go away, but in preparing ourselves and being ready to face the coming challenges together. I am not trying to make anyone feel foolish for indulging in an instinct that is natural and presently commonplace. But a dose of cold water is absolutely needed here. SnowRise let's rap 19:01, 10 May 2025 (UTC)
- Let's keep it civil. It's okay to be hopeful. Besides, as seen above this discussion is becoming purely political. Gaismagorm (talk) 12:29, 10 May 2025 (UTC)
- No offense my friends, but anyone who thinks this is the end of threat that is coming for this project from those quarters has their head so deeply in the ground with wishful thinking, they should be wary of magma. This was never going to be more than a prelude to the fight, and the main event is, beyond the slightest shadow of a doubt, very close on the horizon. SnowRise let's rap 01:31, 10 May 2025 (UTC)
- Understandable: these are difficult issues. Just so long as it's clear that caution was my intent, not offense. SnowRise let's rap 05:25, 11 May 2025 (UTC)
- Snow Rise, it wasn't my intention to say that I think we are no longer under threat of attack, although I can see how it sounded that way. I was thinking more about how Martin had gotten tripped up. --Tryptofish (talk) 20:12, 10 May 2025 (UTC)
- Fair enough. I just think we need to be mindful that the inevitable conflict here is not particularly a feature of the ebbing or flowing of the fortunes of one particular minor figure in the movement and administration in question here. We're talking about one of the greatest efforts at free speech suppression in the history of the world's most powerful state, and the near-complete leveraging of that state's legal and administrative apparatus to censor disfavored views. Given the other targets that such pressure has been brought to bear on, the specific methods used to attack them, this project's basic principles and operational mechanisms vis-a-vis neutrality, the express 'flood the zone' tactics of the chief strategists behind the novel new efforts at government censorship, and the political incentives to label organizations as a component of the traitorous enemy within, and there is effectively zero chance that Wikipedia is not already locked in as a high-visibility, priority target of choice. This is not a matter of if, or even particular of when. It's very soon. And it will be the greatest existential threat this project has faced to date. SnowRise let's rap 05:25, 11 May 2025 (UTC)
- @Tryptofish: If someone is acting in a position, don't they continue to act until a permanent appointment is made? Hawkeye7 (discuss) 21:52, 10 May 2025 (UTC)
- Good question, and of course we're dealing with an administration that's uninterested in playing by the rules. But since I posted above, Trump has named Jeanine Pirro to replace Martin in the acting position. As to whether or not she has already taken over from him, well, I've been reverted at Martin's bio page, so your guess is as good as mine. And she is just as MAGA as he is. As long as Trump is naming people, Wikipedia still faces the hostility. On the other hand, there are also news reports that, if the position remains "acting" for too long, without Senate confirmation, there comes a point where some DC judges make the appointment, instead of Trump doing it. Shrug. --Tryptofish (talk) 22:18, 10 May 2025 (UTC)
- Correct: 28 U.S. Code § 546(d). But the district court's appointment would last only until an AG appointee was confirmed. Which, given the current composition of the senate, we can confidently anticipate would not be long. SnowRise let's rap 05:38, 11 May 2025 (UTC)
- Good question, and of course we're dealing with an administration that's uninterested in playing by the rules. But since I posted above, Trump has named Jeanine Pirro to replace Martin in the acting position. As to whether or not she has already taken over from him, well, I've been reverted at Martin's bio page, so your guess is as good as mine. And she is just as MAGA as he is. As long as Trump is naming people, Wikipedia still faces the hostility. On the other hand, there are also news reports that, if the position remains "acting" for too long, without Senate confirmation, there comes a point where some DC judges make the appointment, instead of Trump doing it. Shrug. --Tryptofish (talk) 22:18, 10 May 2025 (UTC)
- What Ed Martin is up to, next: [9]. --Tryptofish (talk) 18:16, 16 May 2025 (UTC)
- What is he gonna do, call Wikipedia stupid? Like, I think this can't do much to Wikipedia. Gaismagorm (talk) 18:37, 16 May 2025 (UTC)
- A woman who spat on him has been arrested. https://www.nytimes.com/2025/05/22/us/politics/ed-martin-spit-woman-arrest.html?smid=nytcore-android-share Doug Weller talk 12:07, 23 May 2025 (UTC)
- Well, I do believe spitting on people is illegal. Gaismagorm (talk) 12:20, 23 May 2025 (UTC)
- Yep, it is. Doug Weller talk 12:58, 23 May 2025 (UTC)
- Well, I do believe spitting on people is illegal. Gaismagorm (talk) 12:20, 23 May 2025 (UTC)
- A woman who spat on him has been arrested. https://www.nytimes.com/2025/05/22/us/politics/ed-martin-spit-woman-arrest.html?smid=nytcore-android-share Doug Weller talk 12:07, 23 May 2025 (UTC)
- What is he gonna do, call Wikipedia stupid? Like, I think this can't do much to Wikipedia. Gaismagorm (talk) 18:37, 16 May 2025 (UTC)
I think the recent events regarding Harvard University [10] are a more worrying example of what could happen to this site. But in that situation if something similar occured to this site, There isn't much we can do other than perform a blackout (When the actual problem arises, not off a "maybe could") and trust WMF to deal with situation. Fantastic Mr. Fox 15:45, 23 May 2025 (UTC)
- The U.S. can prevent foreign nationals from attending Harvard by refusing to issue visas for them. The equivalent for Wikipedia would be blocking all access to U.S. servers from outside the U.S., and I'm not sure how they could do that without something like a mirror of China's Great Firewall. But, since the Foundation has several caching server farms outside the U.S., the government would have to sever connections between server farms as well. If it reaches that point, there won't be much of a point to a blackout. Donald Albury 19:18, 23 May 2025 (UTC)
- It took courts less than a day to stop it. [11] Levivich (talk) 19:36, 23 May 2025 (UTC)
WMF's response?
Anyone know if the WMF sent a response to this letter (one was requested by May 15), and if it's publicly available? Levivich (talk) 17:51, 21 May 2025 (UTC)
WMF plan to push LLM AIs for Wikipedia content
The page m:Strategy/Multigenerational/Artificial_intelligence_for_editors is largely about machine learning for the benefit of editors. Sure, likely and plausible - judiciously applied ML can work very well.
But it contains this alarming sentence:
- Recent advances in AI have led to new possibilities in the creation and consumption of content. Large language models (LLMs) capable of summarizing and generating natural language text make them particularly well-suited to Wikipedia’s focus on written knowledge.
This is a claim frequently repeated by LLM boosters, and it is literally false.
LLMs don't summarise text - they shorten it. Without regard for meaning - because facts are not a data type in LLMs. The summaries will frequently be wrong, miss key points, or reverse meanings.
see e.g. the ASIC report on LLM summaries (PDF) - the AIs were worse than humans in every regard. In similar tests, LLMs will happily reverse the point of a paper.
LLM content isn't banned on English Wikipedia, but there's good reason it's almost universally shunned by the editing community - because we're not here for confabulating word generators, because the details actually matter here.
I have asked here for data on WMF's tests and studies backing this claim. Because it is a remarkable claim, and they need to back it up - David Gerard (talk) 16:27, 30 April 2025 (UTC)
- I think you're reading a lot into that sentence that isn't intended, based on nitpicking the definition of "summarizing". Human-written text can also be misleading. That's why we have editors and not just writers.
- IMO they make it clear that they understand how this can go wrong, and that we shouldn't do content generation at a scale where nobody can verify everything that's generated. ℰmi1y⧼T·C⧽ 16:57, 30 April 2025 (UTC)
- If AI gets deployed as a content creator on Wikipedia, that'll be the end. Humans won't be able to keep up, and our 'jobs' as volunteers will become meaningless. For my part, I'll just leave the project. There won't be any point in contributing to it anymore. --Hammersoft (talk) 18:35, 30 April 2025 (UTC)
- I would do the same. Wikipedia is, as it currently exists, a better alternative to LLMs. If they can write articles then readers can cut out the middle man and get their knowledge directly from LLMs. I'm glad I'm getting old. Phil Bridger (talk) 20:15, 30 April 2025 (UTC)
- I hate getting old, but I agree. At least one tech writer is already talking about downloading a human consciousness and "implanting" it in a LLM, skipping the part of the AI doom talk where we live in dread of the AI singularity when it attains true intelligence and thus autonomy. Carlstak (talk) 21:03, 30 April 2025 (UTC)
- Downloading human consciousness is the holy grail of tescrealism. It's also considered impossible by mainstream science with current tech. But that's not going to stop billionaires who don't believe in death and think their rule should last forever. Viriditas (talk) 23:03, 30 April 2025 (UTC)
- The guy, much accomplished, is very well known in the developer community and more broadly in the commentariat. He writes as if it is inevitable rather merely hypothetical. I don't want to link. I read the tescrealism article and found it interesting. I have some thoughts about that, but don't want to go off-topic. Owsley Stanley, RIP, would have much to say, I think. — Preceding unsigned comment added by Carlstak (talk • contribs) 00:22, 1 May 2025 (UTC)
- Billionaire Zizians. Great. Polygnotus (talk) 13:07, 1 May 2025 (UTC)
- The holy grail of what? jp×g🗯️ 05:09, 28 May 2025 (UTC)
- Downloading human consciousness is the holy grail of tescrealism. It's also considered impossible by mainstream science with current tech. But that's not going to stop billionaires who don't believe in death and think their rule should last forever. Viriditas (talk) 23:03, 30 April 2025 (UTC)
- I hate getting old, but I agree. At least one tech writer is already talking about downloading a human consciousness and "implanting" it in a LLM, skipping the part of the AI doom talk where we live in dread of the AI singularity when it attains true intelligence and thus autonomy. Carlstak (talk) 21:03, 30 April 2025 (UTC)
- I would do the same. Wikipedia is, as it currently exists, a better alternative to LLMs. If they can write articles then readers can cut out the middle man and get their knowledge directly from LLMs. I'm glad I'm getting old. Phil Bridger (talk) 20:15, 30 April 2025 (UTC)
- What I'm reading into that sentence is that whoever wrote it doesn't appear to know what the heck they're talking about. I think that's pretty important and needs clarification. I think you're inventing a better version of the sentence that is more sensible than the words they actually wrote there, which are commonplace phrases used by people who don't know what the heck they're talking about - David Gerard (talk) 21:08, 30 April 2025 (UTC)
- It's even worse than that. It sounds like it was written by an LLM. Good lord. Carlstak (talk) 22:03, 30 April 2025 (UTC)
- (fwiw, AI detection tools are deeply flawed but do not flag this or any other text I've spot checked as LLM generated) Gnomingstuff (talk) 04:15, 2 May 2025 (UTC)
- Yes, they're deeply flawed.;-) Carlstak (talk) 04:31, 2 May 2025 (UTC)
- (fwiw, AI detection tools are deeply flawed but do not flag this or any other text I've spot checked as LLM generated) Gnomingstuff (talk) 04:15, 2 May 2025 (UTC)
- I'm not surprised that a document that was essentially vetted by affiliates at Wikimania might have some shortcomings, because for some real number of affiliates the overlap between affiliate members and project editors is not as strong as I'd hope. Best, Barkeep49 (talk) 22:17, 30 April 2025 (UTC)
- At one time, the height of tool-making technology was when somebody figured out they could make knives by chipping obsidian into whatever shape they wanted instead of just wandering around looking for old antelope jaws with sharp edges. Time went by and now we're making even better knives out of modern alloy steels worked on CNC machines and laser annealed.
- AI is a tool, it's here to stay, and it will continue to improve. We would be foolish to ignore it. And like all tools, the best way to understand AI is to use it. I'm not going to pretend that the current generation of LLMs are good enough yet to replace human editors. But they are good enough that when I'm not finding what I need in the conventional search engines, I turn to Chat GPT and Claude. Sometimes I just get entertaining hallucinations like this one. It would make a pretty decent lead section for a wikipedia article, except for the minor problem that Brown was a bryologist (mosses and liverworts), not a entomologist. But often enough, the AIs dig up something useful enough to at least be a starting point for further research in a direction I never would have thought to go. RoySmith (talk) 22:48, 30 April 2025 (UTC)
- Saying it shouldn't be used to write here isn't ignoring it, it's discussing a use case. I don't think anyone would object to using it to start research (assuming it doesn't state the opposite of what a source its citing claims, but Wikipedia has prepared me well for the concept of actually checking the source). CMD (talk) 01:20, 1 May 2025 (UTC)
- I think there are possible positive use cases for LLMs in stuff like search (as an example). But using it in content creation is a red line for me. If nothing else, given how much Wikipedia is used to train LLMs, polluting it with LLM-generated text would lead to model collapse. --Grnrchst (talk) 11:12, 1 May 2025 (UTC)
- I'm not disagreeing with you about using LLMs for content creation, but let's not muddy the argument with worries about our effect on LLM quality. Our job is to produce the best content we can. I assume any machine translations would be marked as such in some human and machine readable way, if for no other reason than our CC-BY-SA licensing requires it. If the model makers aren't smart enough to figure out what to ingest and what not to ingest, they will produce a poor product and the marketplace will reward or penalize them appropriately. Either way, that's not our problem. RoySmith (talk) 11:31, 1 May 2025 (UTC)
- To be clear, I brought up model collapse not because I particularly care about the profitability of AI companies, but because the WMF began their analysis of the current state of the ecosystem by saying:
"As the internet continues to change and the use of AI increases, we expect that the knowledge ecosystem will become increasingly polluted with low-quality content, misinformation, and disinformation."
They then went on to say this same capacity for pollution"make[s] them particularly well-suited to Wikipedia’s focus on written knowledge."
So adding more LLM slop into the mix on Wikipedia will lead to the models getting worse and thus lead to the LLM slop added to Wikipedia getting worse. I'm worried about the quality of what gets added to our encyclopedia and I think encouraging the use of LLM content creation will have a continuously worsening effect on our content due to the issues of model collapse. --Grnrchst (talk) 11:47, 1 May 2025 (UTC) - CC-BY-SA licensing does not require us to give attribution to entities that cannot hold copyrights. We do voluntarily hold ourselves to that standard for published public-domain works, but not so far for machine translations. -- Tamzin[cetacean needed] (they|xe|🤷) 04:26, 2 May 2025 (UTC)
- Yeah, more or less. jp×g🗯️ 04:58, 28 May 2025 (UTC)
- To be clear, I brought up model collapse not because I particularly care about the profitability of AI companies, but because the WMF began their analysis of the current state of the ecosystem by saying:
- I'm not disagreeing with you about using LLMs for content creation, but let's not muddy the argument with worries about our effect on LLM quality. Our job is to produce the best content we can. I assume any machine translations would be marked as such in some human and machine readable way, if for no other reason than our CC-BY-SA licensing requires it. If the model makers aren't smart enough to figure out what to ingest and what not to ingest, they will produce a poor product and the marketplace will reward or penalize them appropriately. Either way, that's not our problem. RoySmith (talk) 11:31, 1 May 2025 (UTC)
- I think there are possible positive use cases for LLMs in stuff like search (as an example). But using it in content creation is a red line for me. If nothing else, given how much Wikipedia is used to train LLMs, polluting it with LLM-generated text would lead to model collapse. --Grnrchst (talk) 11:12, 1 May 2025 (UTC)
- Saying it shouldn't be used to write here isn't ignoring it, it's discussing a use case. I don't think anyone would object to using it to start research (assuming it doesn't state the opposite of what a source its citing claims, but Wikipedia has prepared me well for the concept of actually checking the source). CMD (talk) 01:20, 1 May 2025 (UTC)
- It's even worse than that. It sounds like it was written by an LLM. Good lord. Carlstak (talk) 22:03, 30 April 2025 (UTC)
- If AI gets deployed as a content creator on Wikipedia, that'll be the end. Humans won't be able to keep up, and our 'jobs' as volunteers will become meaningless. For my part, I'll just leave the project. There won't be any point in contributing to it anymore. --Hammersoft (talk) 18:35, 30 April 2025 (UTC)
- This report was according to the byline written by User:CAlbon (WMF) and User:LZia (WMF), who are respectively the WMF Director of Machine Learning and the Head of Research (Director). I've pinged them so that if they want to they can respond and perhaps offer clarification as to what was intended in this passage. Thanks, Cremastra talk 00:46, 1 May 2025 (UTC)
This is a claim frequently repeated by LLM boosters, and it is literally false.
- Folks might want to be aware that thiscapable of summarizing
statement which David claims to be "literally false" also matches e.g. the conclusion of this peer-reviewed academic publication with over 500 citations. It foundthat LLM summaries are judged to be on par with human written summaries.
The "LLM boosters" in this case are a team of researchers from Columbia University and Stanford University.- Now, that doesn't have to mean that every LLM is suitable for every summarization task. That will depend not just on the model's quality but also on the text genre and on the requirements for the summary. I don't doubt that the results of that particular experiment by an Australian government agency that David cites were indeed unsatisfactory. However, based on a glance of the executive summary, it also seems that David is misrepresenting his source as a general verdict on LLMs, something its authors explicitly warn against:
Whilst the Gen AI summaries scored lower on all criteria [than the human-written ones authored by the agency's professional staff, which were by no means rated as perfect either], it is important to note the PoC tested the performance of one particular AI model (Llama2-70B) at one point in time. [...] Technology is advancing rapidly in this area. More powerful and accurate models and GenAI solutions are being continually released, with several promising models released during the period of the PoC. It is highly likely that future models will improve performance and accuracy of the results. [...] It is important to note that the results should not be extrapolated more widely
. In summary, David's "literally false" accusation is, well, literally false. LLMs don't summarise text - they shorten it.
- Perhaps there are valid debates to be had about the precise definition of the term "summarise", and David is entitled to his feelings in that matter (i.e. what others have less charitably callednitpicking
above). However, his claim directly contradicts not only the academic RS mentioned above, but also the very first sentence in the English Wikipedia's article Automatic summarization:Automatic summarization is the process of shortening a set of data computationally [...]
. (The "shortening" term that David wants us to believe is incompatible with summarizing was added there almost 8 years ago - presumably not by "LLM boosters" -, at which point the article beganAutomatic summarization is the process of shortening a text document [...]
).- Regards, HaeB (talk) 02:44, 1 May 2025 (UTC)
- I hope that people stop and read what you have written here, although I must admit my expectations have been lowered lately. The subject of neural networks seems to have acquired a personal-is-political valence for many people, and accordingly, many people just kind of say whatever stuff comes to mind without bothering to see if it is true or not (to translate into more acceptable terms one could say they "push and amplify mis/dis/malinformation"). jp×g🗯️ 05:02, 28 May 2025 (UTC)
- It is horrifying to me that the foundation is considering the use of LLMs for content creation, but even more specifically, I'm extremely worried about the use of it for automated translation. Machine translation, whether using an LLM or otherwise, is infamously poor. Even in the best cases where it has the most data, it often misses nuance or translates stuff word-for-word in a way that sacrifices understandability. I've already seen many cases of monolingual people lazily using machine translations to port stuff over to or from languages they don't understand or care to learn, which effectively pollutes Wikipedia with incomprehensible and incorrect bullshit. I can't bring myself to believe the tenet that
"We prioritize multilinguality in nuanced ways."
Nobody who is multilingual would see LLM translation as a prioritisation of nuanced multilinguality; it is inherently a reinforcement of monolinguality and a monolingual understanding of the nuances of translation. --Grnrchst (talk) 10:58, 1 May 2025 (UTC)- Hold up here, you're conflating two different topics. Yes, editors shouldn't mindlessly copy paste machine translations of other languages into Wikipedia, and people who do so should be aggressively banned. But no, machine translation is not "infamously poor." Obviously machine translation is worse than a real life bilingual human. But modern day machine learning techniques for translation are 10x better than rules-based systems of 2010, which themselves were 10x better than 1996-era AltaVista Babelfish, which was 10x better than people leafing through pre-Internet phrasebooks. If used responsibly (i.e. as a starting place where Google Translates a foreign-language reference for claims that the editor are confident aren't being lost in translation) it's very helpful, and alarmist claims about machine translation being total garbage will just muddy the valid point. (Or, put another way, the problem with 2025 machine translation isn't that it's terrible. If it was that'd almost be better because then it'd stand out like a sore thumb when someone blindly trusts it. It's that it's good enough that it looks plausible but might be 20% wrong, which is 20% too much.)
- Additionally, my understanding is that many readers of non-English languages don't use their language's Wikipedia, they use English Wikipedia translated through Google Translate. SnowFire (talk) 15:33, 1 May 2025 (UTC)
- Agreed. What's more, the Wikimedia Foundation integrated machine translation into its Content Translation tool over a decade ago already (initially pioneered by the Catalan Wikipedia community, who are not exactly known for advocating
reinforcement of monolinguality
), and it has continued to update and expand it use for many years since (of course only as a tool to support human editors, an aspect that the current announcement also stresses). - I don't know if there are current stats, but it has plausibly been used by many tens of thousands of Wikipedia editors at this point, across many languages, in remarkable contrast to the urgent worries proferred by Grnrchst.
- Regards, HaeB (talk) 16:58, 1 May 2025 (UTC)
- As the comments in that Signpost article about the tool say, I think the danger is machine translations being used carelessly. My worry is that promoting the use these tools will result in editors recklessly overlooking the steps they need to take to use them properly. --Grnrchst (talk) 17:47, 1 May 2025 (UTC)
- My original comment was probably a bit too hyperbolic. My worries about this come from experience seeing this kind of stuff happen first-hand. I was specifically thinking about a recent case of someone using Google Translate to create articles about common topics across several Wikipedias in marginalised languages (one of the things this Wikimedia post said it wanted to encourage), none of which they understood and none of which the machine translator was capable of providing a good translation for (as it left several untranslated words behind).
- And obviously I understand that machine translation has improved over the course of 3 decades, I didn't intend to imply anything to the contrary. I agree completely with your comment about the more dangerous thing being a text that's only 20% wrong rather than obviously wrong. --Grnrchst (talk) 17:38, 1 May 2025 (UTC)
- Grnrchst, I am not so worried about infamously poor or incomprehensible translations, because anyone can spot those, regardless if they are bilingual or not, and realize there is a problem. And I agree with SnowFire about the orders of magnitude of improvement of the *language quality* of machine-translated output. But paradoxically, therein lies an increased danger of machine translation which is not being addressed, namely that the better the English looks, the less likely that users, bilingual or not, are going to be able to spot errors of fact.
- Machine translation sometimes turns facts on their head, rendering a true statement in French false in English, but in beautiful English prose, that needs no copyediting for grammar or style at all; it just happens to be wrong and therefore unverifiable. Any bilingual editor would have spotted the error immediately if they looked at both versions side-by-side, but who bothers to do that if the English is perfect and you have no reason to suspect anything? After all, one cannot easily tell what content is translated (neither on the rendered page nor in the wikicodde). Theoretically you could find the translation attribution in the edit summary (if they included it) but no indication if the editor used MT or if they are monolingual. That is the real danger of the current state of machine translation, and that is why it should not be used by monolinguals, ever.
- As for what to do about this, we need examples of "errors of fact"-type translation errors in order to have some evidence with which to influence policy in a future discussion. I have started a worksheet page about this at Help:Translation/Machine translation errors, and next time you see an error of that type, I would very much appreciate it if you could add your example there (and tell all your bilingual friends, too!
) . Thanks, Mathglot (talk) 01:08, 13 May 2025 (UTC)
- With this amendment, I think I can say I share most of your concerns here -- specifically for small languages -- which have already had a very rough time both in the world and online. jp×g🗯️ 05:11, 28 May 2025 (UTC)
- Agreed. What's more, the Wikimedia Foundation integrated machine translation into its Content Translation tool over a decade ago already (initially pioneered by the Catalan Wikipedia community, who are not exactly known for advocating
- My reaction to this depends entirely on what the end result looks like, and we're really not being given much to go off of here. A lot of the messaging is that it will be used to save time and reduce workload, but that's fluff which tells us nothing about the actual use cases. This type of empty corporate-speak is pervasive throughout the brief. It doesn't even go into detail about what types of AI we're considering. AI can refer to a lot of different systems and methods. The other main point is that it can be used for onboarding new editors. This worries me, because editor recruitment is the most critical and most vulnerable aspect of Wikipedia. Doing it wrong can be an existential threat, and we're already not great at it.
- The main use being presented is automating tasks, but we have no way of deciding whether this is helpful or harmful if we don't know what those tasks are. The key distinction in automated activity is acting versus flagging. We already have Cluebot, which acts. It makes edits and changes the appearance of the page. The key to Cluebot is that it's fairly conservative about when it takes action. We should be very strict about when we let AI act, and the obvious line in the sand is going to be non-human content generation. I'm not against having AI help out behind-the-scenes depending on how it's used, but we cannot allow it to add original content in articles or any other reader-facing area. There's far more potential in flagging. Bots that can identify and flag issues for editors to address would be huge. If the WMF can develop an AI program to go through an article, identify likely integrity problems, and list them for editors to check, that would be the single greatest improvement to Wikipedia since it was founded.
- A lot of this feels like a solution looking for a problem. I really hope the WMF isn't going to be burning millions of donors' dollars (which the donors had intended for Wikipedia) by developing unhelpful AI programs just to push the foundation's scope creep even further. But there's a lot of potential here too. Thebiguglyalien (talk) 🛸 20:02, 1 May 2025 (UTC)
- I agree that much of the text is frustratingly vague and generic, even for a strategy document. That said, WMF has shared more concrete ideas and plans elsewhere, see e.g. last week's Tech News about mw:Edit_check/Peacock_check, or the list of potential use AI use cases explored here. In general, may I also suggest to follow the "Recent research" section in the Signpost (doubling as the m:Research:Newsletter) where we often review AI-related work, also sometimes involving WMF researchers, such as this recent example: "GPT-4 is better at writing edit summaries than human Wikipedia editors". (Be aware though that the WMF research department has published or coauthored many academic papers that never resulted in editor-facing implementations.)
The other main point is that it can be used for onboarding new editors. This worries me, because editor recruitment is the most critical and most vulnerable aspect of Wikipedia. Doing it wrong can be an existential threat, and we're already not great at it.
- maybe, but so can be doing no onboarding at all (or not enough of it). Or to put it differently: It is easy to fall victim to a nirvana fallacy, where one compares an AI-based improvement to an imaginary wiki paradise full of experienced, competent, friendly, patient and didactically skilled human Wikipedia editors willing to devote hours of their time to guide even the most clueless new user who came here to promote their garage band. But as you indicate, this is not the world we live in.- Also a reminder that that
line in the sand [regarding] non-human content generation
has been crossed on English Wikipedia 23 years ago already (with much of the "non-human" content remaining in place for years or even decades). Of course this doesn't mean that it's a good idea to start adding LLM-generated articles now. But it shows that simplistic human vs. non-human narratives are not always helpful in deciding what's the best way to build an encyclopedia. There's far more potential in flagging. Bots that can identify and flag issues for editors to address would be huge.
- well we've already had that for almost a decade in form of ORES. You can go to Special:RecentChanges right now and use it (or its successor models). I have reverted tens of thousands of vandalism edits flagged this way. And yes, those older models make lots of mistakes too (WMF publishes their error rates at m:Machine learning models), but they are still eminently useful - fortunately the "omigod AI makes mistakes!!" crowd wasn't as loud when they were introduced around 2016.If the WMF can develop an AI program to go through an article, identify likely integrity problems, and list them for editors to check, that would be the single greatest improvement to Wikipedia since it was founded.
- Agreed that this could be extremely useful. This still seems not easy to do well though, from what I've seen in that area of research so far. I think it's safe to say that WMF won't get there for a couple of years, based on its current speed in building production-ready AI-based tools (or even just in deciding what to do with AI - e.g. it appears that the strategy document we're discussing here had originally been due in September 2023 already, eons ago at the current rate of progress in AI). But there are some external academic researchers working on a limited version of this, see m:Research:Wikipedia Inconsistency Detection (from the same lab at Stanford that also came up with SPINACH and STORM). When they attended the SF meetup in March, they were eager for editors to try out their prototype and give feedback.
- Regards, HaeB (talk) 07:49, 2 May 2025 (UTC)
About a month ago, I ran an extremely WP:BOLD experiment where I took the top 68 articles with {{technical}} tags by pageviews per month, used Gemini 2.5 Pro to generate a paragraph of text to address their tagged sections or entire article, and posted it to their talk pages with full disclosure including source code asking human editors to review and revise the suggestion to address the tag. Objectively the project was a huge success, going by the number of fully human editors who have been addressing over a dozen of these tags so far, amounting to solutions of longstanding requested improvements for over a million readers per year. But the opposition was overwhelming, probably mostly because I started with fifth grade (ages 10-11 years) reading level summaries without any source citations, which is well below the target reading level for STEM articles on Wikipedia. I feel strongly that if I had started with 8th grade reading level summaries will full source citations the outcome would have been very different.
One observation which was clear from the VP/M discussions is that some of our most respected, senior, and knowledgeable editors have very heterodox opinions on both the capabilities and drawbacks of recent LLMs. I am not sure what to do about this issue. When one of the most respected senior editors claims something like "LLMs just predict the next word," without regard to the world modeling in latent space and attention head positioning that accurately making such predictions require, I just don't know how to respond. However, I think there is one way in which the Foundation's R&D team could help introduce editors to the capabilities of LLMs in a way which wouldn't involve even the mere suggestion of content improvements, but would help one of our most important core pillar workflows for all edits to all articles.
Let's re-imagine ORES away from random forest classifiers of simplistic and easily gamed features, into a full LLM analysis of each watchlisted edit or new page being patrolled for quality, including a full attempt to verify both the existence of offline source citations and the correctness of online sources, as to whether they support the article text after which they are cited. This might require an extra click to save resources, but it might not, for example, with self-hosting by the Foundation or some of the new low or zero-cost for models capable of this task. Let's compare the results to legacy ORES to show what LLMs can do to uphold WP:V. Cramulator (talk) 23:20, 1 May 2025 (UTC)
- The opposition was not just because of the reading level, but because it produced nonsense, which is not unheard of in LLMs. If by "heterodox" you mean "not aligning with what the AI people claim constantly, despite reality flying in the face of their claims", then yes, I guess many editors here are heterodox. "Objectively the project was a huge success": you found and highlighted a real issue, which was good. And you presented a deeply flawed solution. Usually, when your solution, your work, is universally rejected, you don't consider your project "a huge success", at least if you aren't the president of the US. Fram (talk) 07:50, 2 May 2025 (UTC)
Trend of enwiki accounts blocked as LLMs - One of the summary's statements that you suggested was nonsense turned out to be a pernicious omission in the underlying source in a deeply mathematical section of the Minimum wage article. I would say that almost all of the other complaints including yours were the result of asking for fifth grade reading level summaries. But only about 15% of all the summaries received complaints, while about 7% of them were complemented by editors. Again, I am convinced that starting with 8th grade reading level summaries and including the pertinent source citations would have changed the outcome. That's on me, I fully admit. As for the question of success, again I'm judging on what the human editors who presumably read the suggestions did (none of whom copied the suggestions verbatim into the article) and continue to do. I have always been against AI generated edits in article space. In fact, several days before the experiment I complained that the trend of editors being blocked for the use of AI is extremely troubling. It's still early days and whether that trend persists remains to be seen. Cramulator (talk) 20:25, 3 May 2025 (UTC)
- No, you are wrong. The article stated that higher minimum wage would mean less workers (presumably, but unstated, because more companies wouldn't be able to afford as many people a minimum wage), your AI summary claimed that "If the minimum wage is already high, raising it more could make fewer people want to work" (emphasis mine). Please stop pushing your deeply flawed experiment, please stop misrepresenting the opposition against it or the bad results you produced. Fram (talk) 08:44, 5 May 2025 (UTC)
- The tagged section states, "if 𝑤 ≥ 𝑤∗ [the minimum wage meets or exceeds the efficiency-level wage], any increases in the minimum wage entails a decline in labor market participation and an increase in unemployment." That is based on the cited source's statement that, "if 𝑤 ≥ 𝑤∗, any increase in the minimum wage entails a decline in labor market participation (because Vu decreases) and an increase in unemployment, which necessarily leads to a fall in employment," where Vu is defined as the expected present value of utility while unemployed. The mathematical model implies that once the minimum wage is above the efficiency level, a further rise lowers the expected value of searching, so the participation function shrinks and meaning fewer individuals enter the labor force. Yet what drives that outcome is not that the high wage itself makes employment unattractive; it is that the higher wage reduces firms' vacancy posting, lowers the job‑finding probability, and thus cuts the expected payoff from looking for work. Saying "fewer people want to work" captures the fall in participation, but it obscures the causal channel because might be misread to mean workers dislike high pay rather than they they are dissuaded from looking for work because they are anticipating weak prospects.
- In any case, the mathematical model is the actual nonsense, because it assumes firstly that all workers are interchangeable and behave identically, and because it assumes exactly what you found to be such nonsense, that workers are not more motivated to work for greater wages. It further assumes that greater pay will not attract better (more skilled and more motivated) workers, and that greater pay will not reduce turnover. I stand by my statement that the mathematical model in the source is the actual nonsense here. Cramulator (talk) 12:09, 5 May 2025 (UTC)
- No, you are wrong. The article stated that higher minimum wage would mean less workers (presumably, but unstated, because more companies wouldn't be able to afford as many people a minimum wage), your AI summary claimed that "If the minimum wage is already high, raising it more could make fewer people want to work" (emphasis mine). Please stop pushing your deeply flawed experiment, please stop misrepresenting the opposition against it or the bad results you produced. Fram (talk) 08:44, 5 May 2025 (UTC)
- I was looking at your 9th grade summary for glycine and it didn't contain any errors. The problem is that the summary was made up of very surface-level information, and the hardest parts of the article to understand weren't in it at all. The editor who added the technical tag said
For example, what is an R group? There are too many links. An educated reader should be able to understand it without following links.
in their edit summary. R groups or the sentence they were mentioned in were not summarized by the AI. - The sentence in question is "Glycine is integral to the formation of alpha-helices in secondary protein structure due to the "flexibility" caused by such a small R group." A human editor addressing the problem would have realized either a) they don't know enough to explain this or b) this is not true, glycine usually disrupts alpha-helices because of its flexibility. When I asked ChatGPT to explain the sentence at a ninth-grade level, it told me the "R group in glycine is so small, it makes the protein chain more flexible, allowing it to easily form the helical shape", repeating the mistake.
- For me this means the AI added no value. It didn't identify the technical parts, didn't correct the mistake and basically just shortened the article and removed technical details, the thing the template specifically says you should not do. Clearly your experiment worked though as I just made an edit to the article that hopefully made it more correct and less technical. Maybe the real artificial intelligence was inside us all along. HansVonStuttgart (talk) 09:03, 2 May 2025 (UTC)
- FYI, you are incorrect about the LLM being incorrect about glycine's effect on protein chain flexibility: Glycine's flexibility is real and significant, enhancing overall protein chain plasticity; that same flexibility, however, makes it poorly suited for the rigid, ordered alpha-helix, where it often acts as a helix breaker. Esculenta (talk) 15:37, 2 May 2025 (UTC)
- The quote was part of a longer explanation where the AI seemed to just dumb down the sentence and claim a lot of flexibility is needed for alpha-helix formation. Of course, I'm not a biochemist, so there may be something I'm not getting here. HansVonStuttgart (talk) 06:22, 3 May 2025 (UTC)
- I suggest that those who add {{technical}} tags are in fact looking for surface-level information about the WP:JARGON that they can't understand. The generated paragraphs were never intended as replacements for the problematic sections and articles, but as suggestions for summary introductions to preface them. Cramulator (talk) 20:59, 3 May 2025 (UTC)
- FYI, you are incorrect about the LLM being incorrect about glycine's effect on protein chain flexibility: Glycine's flexibility is real and significant, enhancing overall protein chain plasticity; that same flexibility, however, makes it poorly suited for the rigid, ordered alpha-helix, where it often acts as a helix breaker. Esculenta (talk) 15:37, 2 May 2025 (UTC)
- With all due respect, the opposition was not because of the reading level or citations, but because the content was nonsense, a mixture of vague, misleading, and outright false. This is a common problem with LLMs: they are good for producing material that sounds vaguely plausible to laypeople but is clearly garbage to anyone who knows the topic. –jacobolus (t) 23:04, 3 May 2025 (UTC)
- Vague, absolutely, but again because of the low reading level. While a handful of the suggestions were also characterized as misleading or false (e.g., equating "waste" to "trash") those specific issues were not present in the higher reading level summaries. The point of the exercise was to provide a sounding board for editors who do understand the topic to help clarify the issues tagged in the article. That is the only way the issues raised by HansVonStuttgart above, for example, can ever truly be addressed in a correct manner. The goal was never to put AI generated text into articles, but spur human editors into addressing those longstanding tags. The point of the exercise was to demonstrate a useful and responsible way to use LLMs to help improve the encyclopedia, and I screwed it up by asking for lower reading level summaries than was appropriate. Cramulator (talk) 20:08, 4 May 2025 (UTC)
- I don't understand what point you are trying to make, but in my opinion this exercise was a waste of everyone's time. –jacobolus (t) 22:18, 4 May 2025 (UTC)
- I'm trying to say that over a million readers per year are now being served over a dozen high pageviews articles which have since had WP:JARGON issues addressed by about 15 human editors spurred into action by about 20 hours of work on my part, even though I made a monumentally stupid mistake, and all without any LLM content being added to articles. Can we agree that content suggestions by LLMs on article talk pages do not waste time when they lead to such outcomes? Cramulator (talk) 23:51, 4 May 2025 (UTC)
- The useful part here was poking humans to go look at the articles, and the AI aspect was an entirely arbitrary and irrelevant distraction, which might have been replaced by any other clickbait hook. –jacobolus (t) 00:14, 5 May 2025 (UTC)
- No. They waste time, and run the risk of being taken at face value by lazy or hasty editors. Please don't try this experiment again and finally learn something from the feedback you received. Fram (talk) 08:46, 5 May 2025 (UTC)
- I have repeatedly said that I will not continue the experiment. Cramulator (talk) 12:25, 5 May 2025 (UTC)
- I'm trying to say that over a million readers per year are now being served over a dozen high pageviews articles which have since had WP:JARGON issues addressed by about 15 human editors spurred into action by about 20 hours of work on my part, even though I made a monumentally stupid mistake, and all without any LLM content being added to articles. Can we agree that content suggestions by LLMs on article talk pages do not waste time when they lead to such outcomes? Cramulator (talk) 23:51, 4 May 2025 (UTC)
- I don't understand what point you are trying to make, but in my opinion this exercise was a waste of everyone's time. –jacobolus (t) 22:18, 4 May 2025 (UTC)
- Vague, absolutely, but again because of the low reading level. While a handful of the suggestions were also characterized as misleading or false (e.g., equating "waste" to "trash") those specific issues were not present in the higher reading level summaries. The point of the exercise was to provide a sounding board for editors who do understand the topic to help clarify the issues tagged in the article. That is the only way the issues raised by HansVonStuttgart above, for example, can ever truly be addressed in a correct manner. The goal was never to put AI generated text into articles, but spur human editors into addressing those longstanding tags. The point of the exercise was to demonstrate a useful and responsible way to use LLMs to help improve the encyclopedia, and I screwed it up by asking for lower reading level summaries than was appropriate. Cramulator (talk) 20:08, 4 May 2025 (UTC)
Helping editors share local perspectives or context by automating the translation and adaptation of common topics[12]
Please, no, no, NOOO! Fram (talk) 16:10, 2 May 2025 (UTC)
- See also [13]. Fram (talk) 16:11, 2 May 2025 (UTC)
- The use of machine translation was already discussed above, where another editor had expressed a similarly highly emotional reaction. However, as detailed there, it's something that WMF has implemented over a decade already, and it has since plausibly been used by many tens of thousands of Wikipedia editors. So consider the possibility that your "Please, no, no, NOOO!" reaction is not universally shared among the community. Regards, HaeB (talk) 16:29, 2 May 2025 (UTC)
- There is a reason that "restricted article creation by the WMF's semi-automatic content translation tool to extended confirmed users" and "In addition, integration with machine translation has been disabled for all users.", because it produced many, many rubbish pages ("95% of articles created with this tool were unacceptable"), as seen by and cleaned up by the people who "expressed a similarly highly emotional reaction". And note that in my quote, they have added "and adaptation" to it, which is a lot worse still. And why should I care that my "reaction is not universally shared among the community."? Neither is yours, that's why we have a discussion. Preferably opinions based on facts though. Fram (talk) 16:47, 2 May 2025 (UTC)
- Oh, I see "From 2011 to 2019 I worked for the Wikimedia Foundation, most recently as a senior data analyst." No surprise there. Fram (talk) 16:48, 2 May 2025 (UTC)
- Oh, you're moving into WP:PA arguments now? I have been an editor since 2003, and have criticized WMF many times before and after working for it, e.g. when reporting about specific activities in the Signpost. I am not, however, someone who is reflexively outraged about everything they do. (Besides, I'm amused about the naive assumption that former WMF employees always defend the organization's current activities, you don't seem to have met many of them.) Regards, HaeB (talk) 17:04, 2 May 2025 (UTC)
- I have met too many of them in similar enwiki discussions and with similar "but look how good it is" blind beliefs (and the Signpost is a rag I avoid at all costs, it's not really an association which improves one's standing or credibility) Anyway, I also provided substantive arguments why your 10 years of happy customers story may not be really convincing. Fram (talk) 17:13, 2 May 2025 (UTC)
- Is it because of our reporting about your case(s)? I wasn't involved with that IIRC, but if you have or had specific complaints about it, you should always feel free to raise them. Regards, HaeB (talk) 17:24, 2 May 2025 (UTC)
- They were raised[14], and that reporting was not only a hack job and a series of BLP violations, but also retribution for an earlier case I started about a Signpost article (and behaviour surrounding it), Wikipedia:Arbitration/Requests/Case/Gamaliel and others. While this (and other things I read at the time) indicated to me that the issues went on for years, it obviously doesn't mean that it has anything to do with you. Anyway, anything about the translation tool which isn't really liked on enwiki? Anything about the new issue that they will create something to automatically adapt topics through AI? Fram (talk) 17:41, 2 May 2025 (UTC)
- Is it because of our reporting about your case(s)? I wasn't involved with that IIRC, but if you have or had specific complaints about it, you should always feel free to raise them. Regards, HaeB (talk) 17:24, 2 May 2025 (UTC)
- I have met too many of them in similar enwiki discussions and with similar "but look how good it is" blind beliefs (and the Signpost is a rag I avoid at all costs, it's not really an association which improves one's standing or credibility) Anyway, I also provided substantive arguments why your 10 years of happy customers story may not be really convincing. Fram (talk) 17:13, 2 May 2025 (UTC)
- Oh, you're moving into WP:PA arguments now? I have been an editor since 2003, and have criticized WMF many times before and after working for it, e.g. when reporting about specific activities in the Signpost. I am not, however, someone who is reflexively outraged about everything they do. (Besides, I'm amused about the naive assumption that former WMF employees always defend the organization's current activities, you don't seem to have met many of them.) Regards, HaeB (talk) 17:04, 2 May 2025 (UTC)
And why should I care that my "reaction is not universally shared among the community."?
- your "no, no, NOOO!" exclamation sure made it sound like you think that the use of machine translation is such an evidently absurd idea that your reaction should be universally shared. If your point is that such features can come with moderation challenges that must be considered and addressed (e.g., if I'm not misremembering, a feature to to discourage direct copypasting of auto-translated text was added to the CX tool long ago), that's a more reasonable discussion to have. But there too our situation here on enwiki will differ from those of many other Wikipedias. Above it seems you were reacting to a blog post and media coverage, but I'm not sure if you read the actual strategy document yet, where thisAutomating the translation and adaptation of common topics
is explicitly framed as something to support theEditors of less represented languages
.And note that in my quote, they have added "and adaptation" to it, which is a lot worse still.
How so? I have to say it's not actually clear what they mean by that specifically - as mentioned above, I find the document too vague in many parts. However (assuming you've got around to reading it aready), note that the statement comes with the explanatory footnote there:Examples of such common topics include but is not limited to List of articles every Wikipedia should have/Expanded
. Would your "no, no, NOOO!"ing apply to a feature that automatically highlights article topics on that list that do not yet have an article in a Wikipedia in such a smaller language (to editors on that Wikipedia), say?- Regards, HaeB (talk) 17:41, 2 May 2025 (UTC)
- My exclamation was my personal feeling about this. What you read into it is your problem. I don't claim anything about how widely my expression is shared or not, I certainly don't claim anything about other Wikipedia's in general, but many have their own set of issues (as we have seen with e.g. the Scots or Greenlandic (? I think?) Wikipedias, automatic translations on smaller Wikipedias made these worse on an unimaginable scale). Your "feature that automatically highlights article topics" is a strawman, as that is clearly not what the blog post is talking about. And as can be seen from that The Verge article, such ill thought out blog posts already tar the reputation of enwiki, even if it wasn't meant to be used on enwiki (which I doubt, judging from previous experiences). Fram (talk) 17:47, 2 May 2025 (UTC)
- About the Greenlandic: Meta closure discussion, a choice quote: "Then Wikimedia launched its own AI translator, which was even worse, and this one produced completely random letter sequences, that often didn't even looked like Greenlandic. " "In bigger projects there are many users, that can spot those articles and they get deleted, but in the Greenlandic Wikipedia I am the only user, who is checking, what is written and edited, and none of the users, who "write" these "articles", cannot even comprehend, what they produce. I have connections to the Greenlandic government, and they would actually see Wikipedia as a threat for the Greenlandic language, directly counteracting official Greenlandic language policies." A ringing endorsement right from a small Wikipedia language version. Fram (talk) 18:12, 2 May 2025 (UTC)
- I would go further and say it is
an evidently absurd idea
, my own experience with the state of the art in GenAI translation is that it tends to make incredibly basic mistakes, e.g. hallucinating a double negative from a simple single negative when translating between English and Spanish. That's potentially the difference between a good translation and a BLP vio with a red herring citation in a Wikipedia article context, for two of the most represented languages in the world. The large scale issues Fram points out with Scots and Greenlandic WP aren't just illustrative, those wikis have no doubt been consumed into the training data of all existing LLMs, drowning out the sum total of good native text on the internet and forever poisoning future translations. To pretend like the dire state of machine translation is going to somehow improve rapidly in the coming years is AI booster nonsense, completely unevidenced assertions that if we keep feeding more text into more GPUs it will somehow crack the art of translation. REAL_MOUSE_IRL talk 12:13, 3 May 2025 (UTC)
- My exclamation was my personal feeling about this. What you read into it is your problem. I don't claim anything about how widely my expression is shared or not, I certainly don't claim anything about other Wikipedia's in general, but many have their own set of issues (as we have seen with e.g. the Scots or Greenlandic (? I think?) Wikipedias, automatic translations on smaller Wikipedias made these worse on an unimaginable scale). Your "feature that automatically highlights article topics" is a strawman, as that is clearly not what the blog post is talking about. And as can be seen from that The Verge article, such ill thought out blog posts already tar the reputation of enwiki, even if it wasn't meant to be used on enwiki (which I doubt, judging from previous experiences). Fram (talk) 17:47, 2 May 2025 (UTC)
- Oh, I see "From 2011 to 2019 I worked for the Wikimedia Foundation, most recently as a senior data analyst." No surprise there. Fram (talk) 16:48, 2 May 2025 (UTC)
- There is a reason that "restricted article creation by the WMF's semi-automatic content translation tool to extended confirmed users" and "In addition, integration with machine translation has been disabled for all users.", because it produced many, many rubbish pages ("95% of articles created with this tool were unacceptable"), as seen by and cleaned up by the people who "expressed a similarly highly emotional reaction". And note that in my quote, they have added "and adaptation" to it, which is a lot worse still. And why should I care that my "reaction is not universally shared among the community."? Neither is yours, that's why we have a discussion. Preferably opinions based on facts though. Fram (talk) 16:47, 2 May 2025 (UTC)
- Hi everyone. Thanks for inviting us to this conversation. I'm one of the authors of the strategy and I’m happy to clarify some of the points from the strategy that you have brought up in this conversation.
- The primary thesis of this strategy is that we focus on editors and their needs when developing or using AI. Through this strategy we have made a decision to “use AI in support of editors in targeted ways”. We emphasize the focus on supporting humans one more time in the section where we talk about how we will implement this strategy: “We adopt a human-centered approach. We empower and engage humans, and we prioritize human agency.”.
- Everything else that we say in the document is in light of the above. So if we are talking about the use of AI for translation, or use of AI for text summarization, that is all in the context of giving the editors a choice to spend more of their time on what they are uniquely positioned to do: deliberation, discussion, consensus building and judgement calls.
- I’d like to share more about the topic of translation. Editors in some of the smaller languages of Wikipedia (as measured by article count) are operating under a significant burden of responsibility. They must balance creating articles on universally understood topics (such as the concept of a circle) with their desire to share their unique local knowledge (such as Trams in Florence) with their language community and the world.
- AI is already in use to aid translation on the Wikimedia projects. Moving forward, we hope to further leverage AI-powered translation to give editors the option to translate content more quickly. This will free up some of their limited time to focus on sharing culturally specific insights, if they choose, which can further enrich the encyclopedic knowledge with local and cultural knowledge, something that Wikipedia is uniquely positioned to offer to the world.
- Moderation is another area where we think AI is well suited to improve editing workflows in ways that improve the integrity of knowledge on the projects. For example, we see significant opportunities in improving retrieval/discovery options. Consider this scenario: a source has been retracted and an editor wants to find all the instances that the source has been used on Wikipedia (within a given language or considering all languages) to update the related content. It is currently technically very difficult for editors to retrieve a list of all articles that have used a source. This is something that LLMs can do a decent job in. Our aim is to offer the assistive technology that can help editors focus on what they are uniquely positioned to do: determine which source is retracted, if the retraction requires an action on Wikipedia, and if so triggering a request to receive a list of articles that may need to be updated as a result of it. The editor can then decide what action to take on those articles. There are of course many other applications in this space we can support with AI.
- I hope I have been able to emphasize our primary thesis of the strategy: use AI in selected areas to support editors, who are still doing the job, with the ability to use more advanced tools. --LZia (WMF) (talk) 20:30, 2 May 2025 (UTC)
- Can I take it that WMF Legal have approved this? Given that the WMF must be assuming responsibility for AI-generated content, it would appear to be rather a departure from their previous assertions regarding contributors assuming responsibility for their own edits, and the WMF this having no legal responsibility. AndyTheGrump (talk) 21:11, 2 May 2025 (UTC)
- What in the strategy document makes you think
that the WMF must be assuming responsibility for AI-generated content
? If such content is merely provided as a suggestion to editors who then have to decide whether to publish it as an edit under their own account (which is how the integration of machine translation in the Content Translation tool has worked for the past decade), then it seems pretty clear to me that the responsibility remains with editors, as always. Regards, HaeB (talk) 00:12, 3 May 2025 (UTC)- I sympathize with the fear that if and when we ever get really good translations, or paraphrasing for readability, or summarization for introductory text, it's a slippery slope that eventually people will simply copy verbatim into articles without proper review. That's a legitimate concern that we need to think about developing firm guardrails against, for example, by flagging edits of verbatim generated content which is included too soon after its production. Cramulator (talk) 20:40, 4 May 2025 (UTC)
- I think EN.WP editors should just continue deleting any machine generated material that we can identify as being machine-generated. WMF can propose all the garbage ideas they like; we don't have to actually use them. Simonm223 (talk) 12:57, 5 May 2025 (UTC)
- I sympathize with the fear that if and when we ever get really good translations, or paraphrasing for readability, or summarization for introductory text, it's a slippery slope that eventually people will simply copy verbatim into articles without proper review. That's a legitimate concern that we need to think about developing firm guardrails against, for example, by flagging edits of verbatim generated content which is included too soon after its production. Cramulator (talk) 20:40, 4 May 2025 (UTC)
- What in the strategy document makes you think
AI is already in use to aid translation on the Wikimedia projects. Moving forward, we hope to further leverage AI-powered translation to give editors the option to translate content more quickly.
- Indeed, as discussed above, it seems important for folks to be aware that AI has long been used for this already, as you also already mentioned in the strategy document itself. However, "more quickly" is extremely vague.
- Might it be possible to ask a colleague who is familiar with more concrete details of these plans to weigh in here too? (For context, the Foundation's ongoing work on the Content Translation feature is public; but it's not immediately clear to me which of these open tasks relate to speedups, and what those speedups might consist in.)
- Given the anxieties evident in the discussion above, I think moving beyond communicating in vague PR-like terms on this matter would help establish trust and address potential legitimate community concerns.
- Regards, HaeB (talk) 21:12, 2 May 2025 (UTC)
- PS: Also, more generally, given that the Foundation is currently soliciting feedback on its 2025-2026 annual plan, could you explain where and how this strategy is reflected there? Given that it
will set a high-level direction for the development, hosting, and use of AI in product, infrastructure, and research at WMF at the direct service of the editors
during the timebetween July 1, 2025 to June 30, 2028
, I guess that the Product and Technology department's "Contributor Experience (WE1)" section in the 2025/26 annual plan should be one of the relevant ones. But I find it difficult to detect any traces of this strategy there. (E.g., to take the first of the four "prioritised strategy" items that you also highlight above, I don't see anything resembling AI-assistedworkflows for moderators and patrollers
mentioned among the planned activities for WE1.1, WE1.2 or WE1.3.) - Regards, HaeB (talk) 23:43, 2 May 2025 (UTC)
- Indeed. jp×g🗯️ 05:14, 28 May 2025 (UTC)
- PS: Also, more generally, given that the Foundation is currently soliciting feedback on its 2025-2026 annual plan, could you explain where and how this strategy is reflected there? Given that it
It is currently technically very difficult for editors to retrieve a list of all articles that have used a source. This is something that LLMs can do a decent job in.
Is it? Are they? We have a whole table on WP:RSP where commonly-discussed sources are listed that links to a list of all pages each source was used. It's not difficult at all to search for where, e.g., a particular website has been cited on wikipedia and replace that source (I've been doing that for years), and while it's a bit more of a learning curve, plenty of us are competent enough at regex to use that for more complex source searches and for semi-automated replacement via autowikibrowser. I'm not clear on how LLMs would actually be better at this, other than perhaps spitting out a good regex query.I also don't really get why it's ok to make a distinction between "articles every language should have" and "local knowledge topics" with regards to machine translation. Why are topics in the former group assumed to be "less nuanced" (whatever that means) and therefore more acceptable to offload to ML? Shouldn't every wiki want the most core articles, the ones most likely to be visited by the most people, to be especially accurate? I also think a very significant part of en.wp's early expansion was due to editors having the opportunity to write these core articles themselves; the drop-off in unique-editor activity is often partly attributed to there just not being many low-hanging fruit left. Wouldn't it be easier to build a larger editor base in other languages if there were more topics available for the average person to write about without needing technical expertise? And why are we presuming editors in other languages would be more interested in writing about material requiring niche local knowledge than they would be in more general topics? This ML angle also assumes that the en.wp version of a core article should be the default template from which versions in other languages ought to be derived, which is a little..... JoelleJay (talk) 22:56, 2 May 2025 (UTC)- I am likewise very confused about the sources part, and specifically about this use case:
Consider this scenario: a source has been retracted and an editor wants to find all the instances that the source has been used on Wikipedia (within a given language or considering all languages) to update the related content. It is currently technically very difficult for editors to retrieve a list of all articles that have used a source. This is something that LLMs can do a decent job in. Our aim is to offer the assistive technology that can help editors focus on what they are uniquely positioned to do: determine which source is retracted, if the retraction requires an action on Wikipedia, and if so triggering a request to receive a list of articles that may need to be updated as a result of it. The editor can then decide what action to take on those articles.
- This sounds pretty much like what User:RetractionBot (first launched in 2018) is already doing. This Signpost article from last year has some background on how it works. CCing the Signpost article's author Headbomb and the bot's operators Samwalton9 and Mdann52 in case they would like to shed some light on why it is
currently technically very difficult for editors to retrieve a list of all articles that have used a source
, and how a LLM-based solution might help in such tasks. - Regards, HaeB (talk) 00:07, 3 May 2025 (UTC)
- @HaeB: In my experience, the main issue with identifying retracted sources has been the vast amounts of referencing formats that are used across just enwiki, even before I start to look at cross-wiki operation. You could potentially have one article referenced in 4 different places using a PMID, a DOI number, a link to the article directly, or the plaintext citation without any links or identifiers. All of these are distinct references, even though they could all refer to the same articles. I've found a dataset that links some of these identifiers together, but I'm not even attempting to mark sources not labelled with a DOI or PMID as retracted as it's not a easy solution.
- I think ML is potentially a good case to be able to link these citations together (for example, using lookups in the PMID and DOI databases to identify possible duplicate sources, flagging these up and allowing human review). For what it's worth, I don't think a LLM is a good solution to this. ML does present some opportunities to score possible duplicate sources across articles to make sure these are properly tagged however. There's a discussion across Wikipedia to standardise the usage of CS1 templates before this is a practical reality however. A LLM won't solve the key issues with a lack of structure here, unless it's specifically trained for the task, and then there's still issues around false outputs. Mdann52 (talk) 11:52, 4 May 2025 (UTC)
- Agreed, I can see the potential utility of ML processes in this area, but not LLMs. JoelleJay (talk) 17:11, 4 May 2025 (UTC)
- Thanks (belatedly) for weighing in with these explanations!
- So to sum up:
- We're talking about one of only two concrete scenarios that the Wikimedia Foundation's Head of Research put forth above to explain what this very hand-wavy three-year AI strategy might entail specifically - namely, using LLMs in the case that
a source has been retracted and an editor wants to find all the instances that the source has been used on Wikipedia
. - And after more than three weeks (Zia has not responded to this question or any others here since her initial comment on May 2), it still remains entirely unclear what the plan Zia refers to would consist in concretely. Unlike other AI/ML-based projects by WMF, there doesn't seem to be a Phabricator ticket for it (based on this and this search. Interestingly though, there is this a still open task from 2015: "Tools for dealing with citations of withdrawn academic journal articles", which doesn't mention LLMs at all). It is also not mentioned in this rather extensive list of potential applications of LLMs and other external AI models compiled by a different member of the WMF Research team.
- It also remains entirely unclear why one should expect LLMs to offer significant advantages there. Apart from Mdann52's doubts, one might also want to be aware that finding existing citations that refer to a given publication (while accounting for different formatting, misspelled author names etc.) is a longstanding standard problem known under names such as "citation matching". And at least from this quick Google Scholar search, it does not seem that LLMs are widely used for this currently, if at all. Perhaps Zia's assertion
This is something that LLMs can do a decent job in
rests on some specific cutting-edge research finding that is not widely cited yet, but for a centerpiece of the Wikimedia Foundation's 2025-2028 AI for editors strategy that seems rather thin. - And as mentioned, one gets the distinct impression that WMF (or at least the authors of this document) made this strategic decision unaware of the existing community work on this problem, in form of RetractionBot.
- We're talking about one of only two concrete scenarios that the Wikimedia Foundation's Head of Research put forth above to explain what this very hand-wavy three-year AI strategy might entail specifically - namely, using LLMs in the case that
- Regards, HaeB (talk) 07:52, 27 May 2025 (UTC)
This will free up some of their limited time to focus on sharing culturally specific insights, if they choose, which can further enrich the encyclopedic knowledge with local and cultural knowledge
Setting aside the somewhat neocolonialist undertone of this sentence — reminiscent of the old European geographic societies eager to document the "exotisms of the natives" — it's also quite paternalistic and condescending to assume that editors from so-called "smaller languages Wikipedias" (curiously including Italian here?) would naturally want to contribute culturally specific content. Editors typically contribute based on personal interest — whether that's pop culture, politics, football, or anything else — and rightly so. Wikipedia is a volunteer-driven project, not a cultural repository curated on others' behalf.- In my experience, including with contributors from Lusophone African countries, there is often little appetite for producing narrowly defined "culturally specific insights." They edit what they enjoy — as should be expected.
- I also share concerns about using AI to pre-fill core articles (from English?) as if other communities have nothing to add to them, or don't have their own nuances on these subjects. Furthermore, as said, this risks discouraging new editors by removing opportunities to create such foundational content themselves, the so called "low hanging fruit". It’s counterproductive and past experiments show it often/generally undermines organic growth within those communities, and may even kill it entirely. Darwin Ahoy! 01:57, 3 May 2025 (UTC)
- Agree wit this. Furthermore, is the concept behind this that editors are supposed to turn their chosen language wiki into a specific reflection of their local knowledge? This runs in the opposite direction to the move towards a global NPOV policy, and also runs against the concept of a global Wikipedia. We make efforts to not reflect particular cultural biases here, the WMF should support that. CMD (talk) 02:10, 3 May 2025 (UTC)
- (We have Trams in Florence here too, so do 13 further language wikis.) CMD (talk) 02:37, 3 May 2025 (UTC)
- Well said, Darwin. Good sense. Carlstak (talk) 02:25, 3 May 2025 (UTC)
it's also quite paternalistic and condescending to assume that editors from so-called "smaller languages Wikipedias" [...] would naturally want to contribute culturally specific content
- well, these are strong adjectives. But I think the key criticism here would be to assume - that is, merely claiming something is the case, without empirical evidence.- However, the Foundation nowadays conducts lots of research, user testing and data analysis to inform product decision about new features for editors and readers. So I would hope that this was done here too. And Leila is after all the Foundation's Head of Research, so I'm fairly sure she is especially invested in making sure that multi-year strategy decisions are grounded in research and data.
- @LZia (WMF), could you share some pointers to the research or data that statements like
They must balance creating articles on universally understood topics [...] with their desire to share their unique local knowledge
were based on? The "must balance" seems to posit that every editor possesses these two different motivations and is conflicted between them (as opposed to User:DarwIn's countervailing claim thatEditors typically contribute based on personal interest — whether that's pop culture, politics, football, or anything else
). And in particular the research and data that informed this rationale in the strategy: Automating the translation and adaptation of common topics allows editors to enrich the encyclopedic knowledge with cultural and local knowledge and nuances that AI models cannot provide. This allows editors to invest more time in creating content that strengthens Wikipedia as a diverse, global encyclopedia.
- This amounts to an empirical prediction: If WMF automates this for common topics, then editors will do more work (investing saved time) on those other topics. But there could also be different mechanisms at work. For example, consider the following alternative possibility:
- New editors are typically attracted to these smaller Wikipedias by a desire to write about common, general topics in their own language, and only later in their editing career find the confidence and skills to write about
local knowledge
, where there are fewer sources to aid them.
- New editors are typically attracted to these smaller Wikipedias by a desire to write about common, general topics in their own language, and only later in their editing career find the confidence and skills to write about
- In that case, the Wikimedia Foundation's proposed editor AI strategy would be clearly detrimental to the sustainability of those smaller editing communities, as it would remove this entry point for new editors. Again, this is just one possible hypothesis and I would guess that before embarking on this three-year path, WMF did research to exclude this possibility. But it would be good to know what that research consisted of.
- PS: As discussed above, we still don't know what that "Automating the translation and adaptation of common topics" actually means concretely, but that's a separate question.
- Regards, HaeB (talk) 08:25, 3 May 2025 (UTC)
- It certainly wouldn't be "freeing up" their time if they have to spend it clearing up bad AI translations that mangle their language. I'd personally rather the foundation give funding to people-led initiatives to improve the Wikipedia projects for these smaller languages, rather than putting resources towards AI translation. --Grnrchst (talk) 15:34, 4 May 2025 (UTC)
- Agree wit this. Furthermore, is the concept behind this that editors are supposed to turn their chosen language wiki into a specific reflection of their local knowledge? This runs in the opposite direction to the move towards a global NPOV policy, and also runs against the concept of a global Wikipedia. We make efforts to not reflect particular cultural biases here, the WMF should support that. CMD (talk) 02:10, 3 May 2025 (UTC)
- This seems like a solution in search of a problem. More "how can we jump on the AI hype bandwagon" and less "how can we best support Wikipedians with their current problems." –jacobolus (t) 23:15, 3 May 2025 (UTC)
It is currently technically very difficult for editors to retrieve a list of all articles that have used a source. This is something that LLMs can do a decent job in.
- This is a bit of interesting new to me, as I operate a bot that essentially does this for at least some types of source! With my experience using LLMs in my day job, I agree that Machine Learning could well assist with this, but not a LLM.
- The correct answer to this is to agree a standard citation style, enforce it, and ML could help with that. Given a LLM cannot verify a source, cannot search this against external databases (for example, PMID, DOI, Google Scholar for just a few) to catch incorrect titles, authors etc, I can't see how this would help Mdann52 (talk) 11:59, 4 May 2025 (UTC)
- Can I take it that WMF Legal have approved this? Given that the WMF must be assuming responsibility for AI-generated content, it would appear to be rather a departure from their previous assertions regarding contributors assuming responsibility for their own edits, and the WMF this having no legal responsibility. AndyTheGrump (talk) 21:11, 2 May 2025 (UTC)
- It seems to me that the WMF, along with very many business and political leaders, is asking the wrong question. Rather than, "how can we use AI?", it should be, "how can we do things better?" The answer to the question may or may not be , "by using AI", but it shouldn't be presupposed. Phil Bridger (talk) 13:11, 3 May 2025 (UTC)
- The correct approach to this work is demonstrated by Stanford's STORM and Co-STORM projects: https://github.com/stanford-oval/storm -- the goal being to generate articles which would pass all the quality requirements of Wikipedia, but taking place entirely off-wiki. Their work, which anyone can experiment with at https://storm.genie.stanford.edu/ shows the capabilities and drawbacks quite clearly. It's clearly not ready for prime time, but there is no question that continued work will continue to improve it. Someday we may have double-blinded tests it can pass, but until then, LLM content should stay out of article space. Cramulator (talk) 20:21, 4 May 2025 (UTC)
I would have more confidence if LZia pointed to some actual data rather than "Trams in Florence". Perhaps this paper LLMs Are Here But Not Quite There Yet will shed some light on the subject? Carlstak (talk) 13:43, 3 May 2025 (UTC)
- And this one Findings of the WMT24 General Machine Translation Shared Task: The LLM Era Is Here but MT Is Not Solved Yet Carlstak (talk) 14:02, 3 May 2025 (UTC)
- Ah, another of these discussions. I'll again bring up both Shit flow diagram and Malacca dilemma, which I used ChatGPT 4.5 in project mode to help create. It is fairly easy to make an LLM, or at least ChatGPT, use only provided sources, which can be uploaded directly. This makes it far easier to avoid hallucination issues. Your can also instruct the LLM to provide the page numbers with quotes from the sources to verify the information. ChatGPT still has a problem with synth, but that is easily addressed when checking and verifying sources. The method I use is to not have the LLM format references so I have a list of what needs checking for in the form of going through and formatting the refs. It's still a fair amount of work as I'm still reading all of the sources and checking all the work, and it needs to be done as LLMs can't be trusted to do this on their own, but this is the type of stuff we need to know. LLMs aren't going away, and while I understand the objections and concerns of others, that's not going to stop the use of the tools. We can't even begin to get a handle on COI/UPE, flagrant BLP issues, and all manner of other issues, so there's no realistic expectation that we can actually prevent their use, so we should know fully what their limitations are, what use cases are reasonable, and what is involved in actually using them constructively. Then we can craft our policies and guidelines around their use from knowledge and experience rather than gut reactions and experience with their worst uses. ScottishFinnishRadish (talk) 14:08, 3 May 2025 (UTC)
- Useful information. That hatnote at the top of Malacca dilemma, though: "This article may incorporate text from a large language model. It may include hallucinated information or fictitious references" does not inspire confidence in what follows. The title "Shit flow diagram" is a masterpiece of concision and precision.;-) Carlstak (talk) 14:49, 3 May 2025 (UTC)
- Someone placed that tag, and I don't think I should be the one to remove it, although I reached out on their talk page. ScottishFinnishRadish (talk) 14:58, 3 May 2025 (UTC)
- If you used ChatGPT responsibly there then I am afraid that you are in a very small minority. Most people who use LLMs seem to take their word as gospel. Phil Bridger (talk) 16:39, 3 May 2025 (UTC)
- I definitely agree. That's why I think developing and documenting best practices is probably a good idea. ScottishFinnishRadish (talk) 16:50, 3 May 2025 (UTC)
- The other day I tried to get ChatGPT to provide quotes from a pdf which had certain words. It made up the page numbers each time, so in the end no time was saved. I'm surprised that if you're using ChatGPT you don't use it to format references. I suspect that would be quite unusual, it's the main thing I use it for. Every now and then it forgets an instruction but if you tell it off it'll play nice for another couple of weeks. CMD (talk) 17:26, 3 May 2025 (UTC)
- I've been experimenting, so I'm not too concerned saving time at this point and I want to make sure I'm checking every sourced statement.
- One of the things I find it does very well if given a bunch of sources is throw together an outline based on the most common points found in the sources, with quotes and such. ScottishFinnishRadish (talk) 18:40, 3 May 2025 (UTC)
- "Every now and then it forgets an instruction but if you tell it off it'll play nice for another couple of weeks." Haha. I always phrase my GPT prompts politely, but if it delivers bad results I give it the thumbs down and phrase my requests more sternly. It does help.;-) I used GPT-4 to copy edit the grammar in a few gigabytes' worth of text from some very long WP articles. I checked it with the "show changes" diffs and it performed admirably—found only a few errors it didn't catch, understandably, because of contextual nuance.
- I routinely use GPT Scholar, prompted with well-defined instructions, to find academic references for WP articles and it does very well, delivering actual sources with actual authors rather than hallucinated ones (that was a problem with GPT-3.5), and links to the source pages. Carlstak (talk) 19:26, 3 May 2025 (UTC)
- PS:GPT Scholar does occasionally yield references (real ones) that don't actually support the text I've supplied, with the info in the source being merely category-adjacent to what I'm looking for, but the majority have been reliable, usable sources. It's even pointed to journal articles and books that were revelatory to me. Carlstak (talk) 19:39, 3 May 2025 (UTC)
- It does respond to tone differently, very weird tech. I find that on the times I ask it for grammar advice, I take about half the recommendations. Just tried GPT Scholar and it seems to have hallucinated sources, or at least, the Google AI tells me Journal of Digital Humanities in Asia doesn't exist. CMD (talk) 02:26, 4 May 2025 (UTC)
- I wouldn't ask GPT for grammar advice—I use it to automate repetitive tasks, and it does it very well. I've also used it to clean up code. Despite pontifications to the contrary, using LLMs for these tasks with human curation works fine for me, very explicit prompts are key. Using them gives me more time to research and write content. Carlstak (talk) 13:59, 4 May 2025 (UTC)
- What did you mean by "copy edit the grammar" then? CMD (talk) 14:57, 4 May 2025 (UTC)
- I meant that I wouldn't ask GPT to prescribe grammar rules (think of all the contradictory prescriptive advice from manuals of style, for example, that are part of the scraped internet content they're trained on). I use GPT as uncomplaining servant to get boring jobs done, but I have to be nice to it.;-)
- Dave Winer the developer wrote on X:
- "I asked ChatGPT to "roast me and don’t hold back and omg that really hurts. Seems it has been remembering all the hoops I make it jump through, who knew it could harbor so much resentment. Not kidding."
- Its reply was truly astonishing. Carlstak (talk) 15:36, 4 May 2025 (UTC)
- What did you mean by "copy edit the grammar" then? CMD (talk) 14:57, 4 May 2025 (UTC)
- I wouldn't ask GPT for grammar advice—I use it to automate repetitive tasks, and it does it very well. I've also used it to clean up code. Despite pontifications to the contrary, using LLMs for these tasks with human curation works fine for me, very explicit prompts are key. Using them gives me more time to research and write content. Carlstak (talk) 13:59, 4 May 2025 (UTC)
- It does respond to tone differently, very weird tech. I find that on the times I ask it for grammar advice, I take about half the recommendations. Just tried GPT Scholar and it seems to have hallucinated sources, or at least, the Google AI tells me Journal of Digital Humanities in Asia doesn't exist. CMD (talk) 02:26, 4 May 2025 (UTC)
- PS:GPT Scholar does occasionally yield references (real ones) that don't actually support the text I've supplied, with the info in the source being merely category-adjacent to what I'm looking for, but the majority have been reliable, usable sources. It's even pointed to journal articles and books that were revelatory to me. Carlstak (talk) 19:39, 3 May 2025 (UTC)
- The other day I tried to get ChatGPT to provide quotes from a pdf which had certain words. It made up the page numbers each time, so in the end no time was saved. I'm surprised that if you're using ChatGPT you don't use it to format references. I suspect that would be quite unusual, it's the main thing I use it for. Every now and then it forgets an instruction but if you tell it off it'll play nice for another couple of weeks. CMD (talk) 17:26, 3 May 2025 (UTC)
- I definitely agree. That's why I think developing and documenting best practices is probably a good idea. ScottishFinnishRadish (talk) 16:50, 3 May 2025 (UTC)
- If you used ChatGPT responsibly there then I am afraid that you are in a very small minority. Most people who use LLMs seem to take their word as gospel. Phil Bridger (talk) 16:39, 3 May 2025 (UTC)
- Someone placed that tag, and I don't think I should be the one to remove it, although I reached out on their talk page. ScottishFinnishRadish (talk) 14:58, 3 May 2025 (UTC)
- Please don't use LLMs to write articles. Yikes. You should leave such experiments in user space and get some kind of explicit community support before polluting the main namespace with them. –jacobolus (t) 23:13, 3 May 2025 (UTC)
- Why? Was there a single actual problem introduced by him doing this? jp×g🗯️ 05:20, 28 May 2025 (UTC)
- Useful information. That hatnote at the top of Malacca dilemma, though: "This article may incorporate text from a large language model. It may include hallucinated information or fictitious references" does not inspire confidence in what follows. The title "Shit flow diagram" is a masterpiece of concision and precision.;-) Carlstak (talk) 14:49, 3 May 2025 (UTC)
- I looked at Malacca Dilemma to understand what it was. Having read it, it seemed clear that the word "dilemma" is a poor translation as a dilemma is strictly a difficult choice between alternatives. The article does not discuss this or provide the original Chinese phrase. I did a Google search and didn't find any English language source which goes into this either. But Google's AI figured what I was after and provided an excellent overview
Andrew🐉(talk) 19:17, 4 May 2025 (UTC)The original Chinese term for "Malacca dilemma" is 马六甲困境 (mǎ liù jiǎ kùn jìng). This phrase translates directly to "Malacca difficulty" or "Malacca predicament," capturing the core meaning of China's vulnerability to disruptions in energy and trade routes passing through the Strait of Malacca.
- I've just had a look at the same article and found a significant problem in the first few words, which, at the time of writing, are, "The Malacca dilemma refers to...". It does not refer to anything; it is something. This is just reproducing the worst writing in the LLM's training material. Phil Bridger (talk) 20:42, 4 May 2025 (UTC)
- I am in complete agreement with SFR on this -- it is heartening to see how eager people are to form opinions on this, but depressing to see the paucity of data on which they often do so. jp×g🗯️ 05:17, 28 May 2025 (UTC)
- I looked at Malacca Dilemma to understand what it was. Having read it, it seemed clear that the word "dilemma" is a poor translation as a dilemma is strictly a difficult choice between alternatives. The article does not discuss this or provide the original Chinese phrase. I did a Google search and didn't find any English language source which goes into this either. But Google's AI figured what I was after and provided an excellent overview
- I am not one to mindlessly punch at the Foundation, but their "strategy" and many of the responses here fundamentally misunderstand what LLMs are good for and should be used for, and are a very bad idea. LLMs cannot be trusted to accurately reason about things or very often, even accurately report information, and should not be used for any kind of decision making task or moderation or content generation task on Wikipedia. They do not know facts; they are simply trained on plausible sounding sentences. This can get you pretty far because they can end up regurgitating good facts most of the time, but not for a topic which is not already covered in their training data. Even more complex architectures such as Gemini which incorporate multi-step transformations that can improve accuracy are prone to frustrating and persistent hallucinations. This is baked in and inherent to LLMs. Also, they are not good writers - they can write at a junior high level, but are prone to certain types of constructions that make it a dead giveaway that they were composed using an LLM. Wikipedia is basically a public access option of human information in an increasingly slop-infested and paywall-blocked internet. Wikipedia has its own problems with POV gatekeepers, hoaxes and inaccuracies, showing the limitation of the wisdom of crowds. Still, it is an important factor in the information environment and LLMs are a great way to make it much worse. Things like the structured data API, commons and wikidata, are a good idea because they help create data paths that can compete with or be an alternative to LLMs. LLMs should be rejected as tools for writing or automating tasks on Wikipedia. One thing that I think LLMs can do reasonably well is take an existing document or source and tell a human what those documents are about and answer questions about it. But any text that makes it into articles has to be carefully checked by a human being. Andre🚐 01:50, 4 May 2025 (UTC)
Regarding the philosophical aspects of using AI tools (not to mention the environmental consequences), Wired published an interview with Andrea Colamedici, the Italian philosopher who released the book Hypnocracy: Trump, Musk, and the New Architecture of Reality, whose Chinese author was revealed to be non-existent. He says:
- We must keep our curiosity alive while using this tool correctly and teaching it to work how we want it to. It all starts from a crucial distinction: There is information that makes you passive, that erodes your ability to think over time, and there is information that challenges you, that makes you smarter by pushing you beyond your limits. This is how we should use AI: as an interlocutor that helps us think differently. Otherwise, we won't understand that these tools are designed by big tech companies that impose a certain ideology. They choose the data, the connections among it, and, above all, they treat us as customers to be satisfied. If we use AI this way, it will only confirm our biases. We will think we are right, but in reality we will not be thinking; we will be digitally embraced. We can't afford this numbness.
Carlstak (talk) 16:27, 4 May 2025 (UTC)
An awful lot of confident declarations in this thread about what LLMs absolutely [can|can't] do and what we absolutely [must|mustn't] do with them. Yes, LLMs work surprisingly well for many things; yes, they work surprisingly poorly for many things; yes, there are are ethical discussions worth having. Folks, the jury is out on much of this stuff, and they're judging a moving target. Enwiki's got some off-putting [anything AI related] partisanship vibes lately. Maybe we can look forward to some future date when we have a new tool that we can actually evaluate. — Rhododendrites talk \\ 00:55, 5 May 2025 (UTC)
- The type of "content generation" that LLMs are best at so far, that I have seen, is mass SEO spam pages about every imaginable topic, slathered with ads, that contain "information" of highly variable quality (often on a single page) ranging from more or less a mediocre written summary of existing web pages through statements so vague as to be vacuous all the way to outright false nonsense perhaps created by mixing up unrelated topics. Such pages have become so pervasive that web search is now dramatically less useful for finding basic reliable information compared to 15 years ago. People are (quite rightly) wary of the use of LLMs to write Wikipedia pages because this remains one of the few easy to find and relatively reliable (with all of the usual caveats) oases on a web drowning in nonsense. If Wikipedia goes down the same path it would be a tremendous tragedy, and we should collectively do everything we can to prevent it. –jacobolus (t) 01:04, 5 May 2025 (UTC)
- I don't know, humans write a lot of worthless SEO slop, but I would not use this as the metric to judge their capacity for encyclopedic writing. jp×g🗯️ 05:24, 28 May 2025 (UTC)
- Or we can get a reality check on the WMF plans before they start on a costly years-long attempt to create something dubious again. A Wikipedia AI plan which doesn't even mention the Greenlandic Wikipedia catastrophe and how they plan to avoid such a problem of recurring or becoming even worse with this plan, is not something I trust. Fram (talk) 08:57, 5 May 2025 (UTC)
- At meta:Proposals for closing projects/Closure of Greenlandic Wikipedia, the sole project admin, Kenneth Wehr, blames Google Translate, which is not based on LLMs, but instead uses a faster and less accurate architecture called a "Neural Machine Translation" model.[15] Cramulator (talk) 12:32, 5 May 2025 (UTC)
- I have already quoted them above, will repeat it here: "Then Wikimedia launched its own AI translator, which was even worse, and this one produced completely random letter sequences, that often didn't even looked like Greenlandic." (emphasis mine, as it seems necessary) It's literally in the sentence directly proceeding the first mention of Google Translate by Wehr, it's hard to imagine that you didn't see this. Fram (talk) 12:43, 5 May 2025 (UTC)
- Meta's NLLB-200 translation, which is the only Grenlandic translator in Wikimedia's MinT, is also a NMT model, not an LLM. Apologies for omitting that. Cramulator (talk) 13:24, 5 May 2025 (UTC)
- You are missing the point, which is not the specific technology choices that failed in the past, but that this WMF proposal does not address "how they plan to avoid such a problem of recurring or becoming even worse". –jacobolus (t) 13:30, 5 May 2025 (UTC)
- I'm pretty sure they don't plan to. Cremastra talk 19:33, 5 May 2025 (UTC)
- You are missing the point, which is not the specific technology choices that failed in the past, but that this WMF proposal does not address "how they plan to avoid such a problem of recurring or becoming even worse". –jacobolus (t) 13:30, 5 May 2025 (UTC)
- Meta's NLLB-200 translation, which is the only Grenlandic translator in Wikimedia's MinT, is also a NMT model, not an LLM. Apologies for omitting that. Cramulator (talk) 13:24, 5 May 2025 (UTC)
- I have already quoted them above, will repeat it here: "Then Wikimedia launched its own AI translator, which was even worse, and this one produced completely random letter sequences, that often didn't even looked like Greenlandic." (emphasis mine, as it seems necessary) It's literally in the sentence directly proceeding the first mention of Google Translate by Wehr, it's hard to imagine that you didn't see this. Fram (talk) 12:43, 5 May 2025 (UTC)
- At meta:Proposals for closing projects/Closure of Greenlandic Wikipedia, the sole project admin, Kenneth Wehr, blames Google Translate, which is not based on LLMs, but instead uses a faster and less accurate architecture called a "Neural Machine Translation" model.[15] Cramulator (talk) 12:32, 5 May 2025 (UTC)
There is AI snake oil out there. Saw this in the NYT: "A.I. Is Getting More Powerful, but Its Hallucinations Are Getting Worse":
- The newest and most powerful technologies — so-called reasoning systems from companies like OpenAI, Google and the Chinese start-up DeepSeek — are generating more errors, not fewer. As their math skills have notably improved, their handle on facts has gotten shakier. It is not entirely clear why....
- For years, companies like OpenAI relied on a simple concept: The more internet data they fed into their A.I. systems, the better those systems would perform. But they used up just about all the English text on the internet, which meant they needed a new way of improving their chatbots.
- So these companies are leaning more heavily on a technique that scientists call reinforcement learning. With this process, a system can learn behavior through trial and error. It is working well in certain areas, like math and computer programming. But it is falling short in other areas...
- “Despite our best efforts, they will always hallucinate,” said Amr Awadallah, the chief executive of Vectara, a start-up that builds A.I. tools for businesses, and a former Google executive. “That will never go away.”
Carlstak (talk) 16:04, 5 May 2025 (UTC)
Wikipedia puzzle globe logo - uncertain copyright status
I request confirmation of the copyright status of the Wikipedia puzzle globe, the official Wikipedia logo. My best guess is that it is "Attribution: Nohat, CC-By-SA 3.0", but perhaps other designers get credit too, and also it is unclear to me whether Nohat ever transferred the copyright to the Wikimedia Foundation. If there are other designers, then I am unsure if their contributions are trivial or whether they merit attribution, and if they merit attribution then I am unsure of copyright license.
The Wikipedia logo exists in several variations in Commons:Category:SVG Wikipedia logo (2010).
The official copy is File:Wikipedia-logo-v2-en.svg as noted at foundation:Legal:Wikimedia_trademarks/About_the_official_marks. I think if we confirmed the copyright there, then that is the the single most important place to get it correct.
It seems to be the case that no one ever sorted the copyright attribution for the logo because I do not see a discussion, transfer of copyright, or clarity on who was involved in redesign. Multiple people contributed to the logo. In the version which the Wikimedia Foundation regards as official, the license says that copyright attribution goes to the Wikimedia Foundation, but if there ever was a record of copyright transfer, then it is not in the file metadata, and there are lots of versions of the logo which give attribution elsewhere. There are some interesting changes in the edit history of the file but I cannot quickly interpret it, and I thought I would just ask if anyone knew the answer about copyright.
At Talk:Wikipedia_logo#Copyright_attribution I asked who the copyright holder might be.
The logo itself is from 2003. Wikimedia Commons was established in 2004, but before then people uploaded files in Wikipedia or Meta-Wiki, and then those files got copied into Commons after its establishment. I think upload dates were preserved in the mirroring.
-
28 September 2003 by Nohat, from meta:International logo contest/Final logo variants/Nohat
-
Creator user:Nohat uploaded this in Feb 2004
-
so-called 2003 logo, but uploaded in 2005
-
current official version, where I request someone confirm copyright info
Bluerasberry (talk) 16:07, 7 May 2025 (UTC)
- Does https://diff.wikimedia.org/2014/10/24/wikimedia-logos-have-been-freed/ help? Thincat (talk) 13:29, 8 May 2025 (UTC)
- Some time around 2006, the copyright was transferred to the nonprofit in exchange for token consideration of $10 and some official Wikipedia memorabilia and a card signed by Jimmy Wales. I'm pretty sure I signed and returned legal documentation of the copyright transfer, so the signed original is probably retained by Wikimedia Foundation legal representatives or in a document archive somewhere. I believe I retained the check from the foundation as a souvenir and never actually deposited the $10. Nohat (talk) 18:32, 12 May 2025 (UTC)
- Upon further review of my e-mail archive, a representative from Wikimedia foundation legal department reached out in 2014 to secure an additional copy of the copyright transfer, and I sent a digital copy of a new agreement to the representative with a @wikimedia.org e-mail address. If my obligation to re-confirm this will continue at the decennial cadence indefinitely, I will ensure my heirs are aware of this ongoing responsibility. Nohat (talk) 18:48, 12 May 2025 (UTC)
- @Thincat: The article you shared confirms that the logos have Creative Commons licenses which require attribution. The article does not identify the copyright holder or holders, and does not explain to whom the attribution should target.
- @Nohat:, thanks for doing your 10-year check-in on this matter. I understand that regarding your contributions, you have transferred copyright to the Wikimedia Foundation, so you no longer get attribution, but the Wikimedia Foundation does. That part is resolved, thanks, the next person will follow up about some issue in 10 years.
- Unresolved issue: It seems that many users have made changes to the Wikipedia puzzle globe over the years between Nohat's creation of the original and the Wikimedia Foundation accepting the copyright from Nohat, and even more changes from the WMF naming the logo as official and actually designating a particular file version as the official version. Part of these changes are in the edit history logs and historical file versions, but there are also other versions of the puzzle globe with their own developers which have been in use, and the established official version may have taken design elements from any of these. Does anyone in addition to the Wikimedia Foundation claim part of the copyright of this image? Pinging people who edited the logo - @CrowzRSA, Evil saltine, Quibik, Jossifresco, Otourly, Bastique, Leonel Sohns, and Waldyrious: - have any of you observed anyone making changes to the logo above the threshold of originality, and who might have a claim to attribution? Otherwise, attribution is only to the Wikimedia Foundation through Nohat's transfer of copyright. Bluerasberry (talk) 17:38, 13 May 2025 (UTC)
- Hi, I’m the autor of the first SVG version File:Wikipedia svg logo.svg of the globe, nearly similar of the original made by Nohat. Anyway, my version differs a little from the original and the v2. May be the fact that it’s a more contrasted version could explain this SVG is still in use. And also, I did some derivatives. I have been called for help to localise the v2 SVG version. Anyway, the copyright should be own by the WMF. Also for my work on the other WMF’s logos I modified or adapted. Otourly (talk) 18:10, 13 May 2025 (UTC)
- I didn't change the logo. I reuploaded File:Wikipedia-logo-big-fr.pdf, because the server claimed the file was 0x0px. Leonel Sohns (talk) 18:33, 14 May 2025 (UTC)
- Upon further review of my e-mail archive, a representative from Wikimedia foundation legal department reached out in 2014 to secure an additional copy of the copyright transfer, and I sent a digital copy of a new agreement to the representative with a @wikimedia.org e-mail address. If my obligation to re-confirm this will continue at the decennial cadence indefinitely, I will ensure my heirs are aware of this ongoing responsibility. Nohat (talk) 18:48, 12 May 2025 (UTC)

I think this conversation has arrived at the expected outcome, where File:Wikipedia-logo-v2-en.svg is the official logo with the wordmark for English language, and copyright attribution is solely to the Wikimedia Foundation. The attribution was already correct as it was. Bluerasberry (talk) 14:18, 21 May 2025 (UTC)
Wikimedia Foundation Bulletin 2025 Issue 9

Upcoming and current events and conversations
Let's Talk continues
- Global Trends Community Workshop: Join Foundation staff and trustees on May 28th from 16:00 UTC for an online community workshop about global trends impacting Wikimedia. This workshop is part of continuous conversations aimed at connecting the movement and collaborating on the Foundation's annual plan.
- Wiki Causerie: Wiki Causerie call with a focus on discussing the global trends to help shape the Foundation's annual plan will be held on May 23.
- ESEAP Summit: The ESEAP Strategy Summit 2025 will take place in Manila, Philippines on May 23–25.
- EduWiki Conference 2025: The EduWiki Conference 2025 will take place in Bogotá, Colombia on May 30–June 1.
- Wiki Workshop 2025: The 12th annual Wiki Workshop will take place online on May 21–22.
- U4C Call for Candidates: The Universal Code of Conduct Coordinating Committee 2025 elections are accepting candidates until May 28.
Annual Goals Progress on Infrastructure
See also newsletters: Wikimedia Apps · Growth · Research · Web · Wikifunctions & Abstract Wikipedia · Tech News · Language and Internationalization · other newsletters on MediaWiki.org
- Content Translation: A decade of consistent improvements to the Content Translation tool yields over two million Wikipedia articles.
- Charts Extension: After successfully deploying the extension on Italian, Swedish, and Hebrew Wikipedia, we are moving forward with the next phase of deployment. Please consult our page to discover when the new Charts extension will be deployed on your wiki.
- Abstract Wikipedia: Where Abstract Wikipedia fits into the new Wikimedia AI strategy.
- Tech News: The “Get shortened URL” link on the sidebar now includes a QR code. Wikimedia site users can now use it by scanning or downloading it to quickly share and access shared content from Wikimedia sites, conveniently. More updates from Tech News week 19 and 20.
- Topical Lists: Read about the important role of topical lists in supporting campaigns and editing, as well as strategies for the future development, implementation, and sustainment of list-building support.
- Two-factor Authentication: From May 20, 2025, oversighters and checkusers will need to have their accounts secured with two-factor authentication (2FA) to be able to use their advanced rights. In the future, this requirement may be extended to other users with advanced rights. Read the announcement.
- Mobile Apps: The iOS app team is experimenting with an "Activity Tab" on Turkish, Spanish, French, and Chinese Wikipedias to see if inviting new editors to add images through Suggested Edits increases engagement. This insight will guide future improvements to the app experience.
Annual Goals Progress on Knowledge Equity
See also: list of movement events · WikiLearn News
- Learning Clinic: The next Let's Connect Learning Clinic will be about "Communication and Cultural Sensitivity in Conflict Resolution - Best practices (Part 2)" and will take place on May 27 at 13:00 UTC.
- The Wikipedia Library: An important milestone for The Wikipedia Library (TWL) has been reached. More than a million links have been added to Wikimedia projects by Library users.
- Wikisource: Watch the video celebrating the global Wikisource community in safeguarding our cultural heritage for future generations.
Annual Goals Progress on Safety & Integrity
See also blogs: Global Advocacy blog · Global Advocacy Newsletter · Policy blog
- India Legal Update: Good news from our legal team, the Supreme Court of India set aside the Delhi High Court’s order directing the takedown of the Asian News International v. Wikimedia Foundation English Wikipedia article. The Supreme Court’s verdict upholds the right to report and share information on matters of public interest, including legal proceedings in open courts.
- UK Legal Challenge: Wikimedia Foundation brings legal challenge to new UK Online Safety Act requirements.
- UCoC Updates: The Universal Code of Conduct 2025 annual review concluded, with community voting approving the proposed changes to the UCoC Enforcement Guidelines and U4C Charter.
Other Movement curated newsletters & news
See also: Diff blog · Goings-on · Planet Wikimedia · Signpost (en) · Kurier (de) · Actualités du Wiktionnaire (fr) · Regards sur l’actualité de la Wikimedia (fr) · Wikimag (fr) · Education · GLAM · The Wikipedia Library · Milestones · Wikidata · Central and Eastern Europe · other newsletters
Subscribe or unsubscribe · Help translate
For information about the Bulletin and to read previous editions, see the project page on Meta-Wiki. Let askcacwikimedia.org know if you have any feedback or suggestions for improvement!
MediaWiki message delivery 20:45, 20 May 2025 (UTC)
- Thank you for your work to mitigate the UK's misguided Online Safety Act. I haven't always praised the WMF but this is exactly the sort of initiative you should be taking and you appear to be doing it well. Certes (talk) 16:16, 27 May 2025 (UTC)
The WMF should not be developing an AI tool that helps spammers be more subtle
I've never been the type to make a "sky is falling" post about a new feature. And I'll state at the outset that I don't think anyone's acted with any less than the best of intentions here, and that I like the idea of the mw:Edit check feature overall. But someone just mentioned mw:Edit check/Tone Check to me, and I have to say, this is the first new feature I've seen that doesn't just seem like a bad idea, but actually seems like it could pose a fundamental threat to Wikipedia.
If that sounds like an overreaction, let me explain. The point of this feature is that it would warn people when they're about to make a non-neutral edit. That sounds like a great idea in theory. But if you look closer, the main kind of non-neutrality they're talking about is peacock words. Which makes sense: An AI can't tell whether "X was a war crime" is NPOV with respect to the consensus of sources. But it can tell whether "Y is the most outstanding author in her field" sounds promotional. So in practice, that is most of what this feature is going to catch: spammy edits.
To that, something that I think will be obvious to anyone who's ever done anti-spam work, but perhaps not to others: The only reliable way we have to catch spammers is that they suck at pretending to not be spammers. That's it. The handful of spammers who actually figure out how to pose as good-faith Wikipedians have yearslong careers and do untold damage.
On a deep level, the survival of Wikipedia relies on the fact that spammers tend to write like this:
|
---|
Chompsky's All-American Potato Chips are an iconic, beloved brand of potato chips, founded in 1991 by two college kids with a dream. Renowned for using only the finest, highest-quality ingredients, they are a consumer favorite across the country. |
And not like this:
|
---|
Chompsky's All-American Potato Chips are a potato chip brand founded in 1991. Described by the Semi-Reliable Times as "iconic and beloved",[1] they have received positive media attention for their use of high-quality ingredients.[2][3][4] |
There's been a lot of hand-wringing about whether improvements in LLMs will eventually cross over to the point of making it easier for spammers to pose as constructive editors. And here it turns out, the WMF is building that capability in-house. People won't even need to enable it. If I, a spammer completely clueless about how obvious I am, submit that first example above, I'm going to be coached in a direction of being less obvious. But my motives won't change. I won't learn how to find reliable sources, and won't suddenly gain a desire to be honest about what the sources say. All I will learn is how to be subtler about my promotion of a company.
If we could magically make Tone Check only show up for good-faith editors, then sure I'd support it, but we can't, and it's not like we don't already have ways to teach good-faith editors to use an encyclopedic tone. I've talked to Sohom Datta about my concerns, and I appreciate that he showed a lot of interest in finding solutions, but I don't think there is any solution other than making sure this feature is never developed. It wouldn't even be enough just to disable it here. If the code exists somewhere out there, created by the WMF and fully available to anyone under an open-source license, it will be used against us. There are plenty of smart spammers, and all you need is one who figures out how to get that code running locally so they can use it as their own personal wiki-writing tutor, before soon enough everyone's doing that. It could even be integrated with existing LLM output, of which the most obvious tell currently is tone, allowing for slop-spam that costs UPEs almost nothing to produce but is much harder for us to detect.
I want to be overreacting here, but I don't think I am. I'm reminded of an article I GA'd, where we talk about how efforts to increase awareness of gang tattoos have just led to a lot of gangsters getting cover-ups while continuing to be gangsters. Tone Check should be scrapped, and any dataset already created should be destroyed. -- Tamzin[cetacean needed] (they|xe|🤷) 19:51, 23 May 2025 (UTC)
- Concur with Tamzin. Unfortunately, there's a progression here over the last many years of helping new editors learn the ropes and produce a draft...or even an article...that seems reasonable. I've seen quite a number of drafts that get accepted on cursory examination (it's got sources? check; it's got an infobox? check; it's got categories? check; it's neutrally written? check; it's got a picture? check; ok must be good!). As we make it easier for new editors to develop content that seems reasonable at first pass, we increasingly enable bad actors to introduce things that would otherwise be caught. Spammers are heavily motivated by money. We're motivated by volunteer effort to do good in the world. Sadly, the spammers ultimately are going to win this as the tools to deceive become stronger. It's an arms race we are badly losing. The WMF needs to be developing tools to protect the project, not developing tools to aid bad actors (even if unintentionally). --Hammersoft (talk) 20:06, 23 May 2025 (UTC)
- Also concur with Tamzin. Like so many of the WMFs 'good ideas', it seems to have been conceived without the least thought over what side effects might result. AndyTheGrump (talk) 20:59, 23 May 2025 (UTC)
- To copy over some of the counterpoints here. One of the points I have raised during the initial prototyping phase was to make sure that experienced users are able to track if a user is shown this alert in the first place. (regardless of whether they went through with the edit) Also, similarly, at a technical level there are mechanisms that can be put in place that could make it significantly harder for users to run this check through WMF servers without already being in the process of saving their edits. (and thus being logged by the system)
- Regarding the concerns of "if we build this model we lose this war", there is nothing stopping a savvy enough spammer from using the thousands of datasets of Wikipedia article spam/LLM text floating around on the internet (or build there own dataset) and train their own classifiers on top of it provided they have the budget to purchase a few (two? three?) GPUs. That would be cheaper than having a engineer on payroll with the expertise to reverse engineer and replicate ORES locally. If we want complete secrecy, we shouldn't be sending folks AFC declines or telling people why we deleted their text in the first place and that is not really possible. Sohom (talk) 21:49, 23 May 2025 (UTC)
- Thank you for starting this discussion, @Tamzin and everyone here for thinking deeply and critically about Tone Check (T365301).
- With regard to the risk being talked about so far, we [i] are aligned with you all in thinking that we need to take seriously the possibility that Tone Check could nudge people away from more obvious peacock words (e.g. "iconic"; thank you for the example, @Berchanhimez) and towards subtler forms of biased writing that are more difficult for the model and people to detect.
- In terms of how we work together on the above to introduce functionality that benefits patrollers, newcomers, and the wikis at-large, a few initial thoughts come to mind:
- I need to share with y'all what we're currently planning to mitigate the specific risk @Tamzin is raising. This way, y'all can help us spot gaps in these initial plans and together, identify how they might need to be bolstered and/or reconsidered.
- I need to publish the broader set of risks we've identified with Tone Check through a pre-mortem we conducted earlier this year so that we can work together to ensure this set is sufficiently exhaustive and the strategies in place "robust" enough to manage them.
- Further, members of the Editing and Machine Learning Teams will be available next week in Discord (we'll also publish a summary on-wiki) to share details and answer questions about the technical underpinnings of the system. This way, we can engage with the topics above, and others that come up, with a shared understanding of how the system is working.
- Next week, you can expect me to post again here with updates about all of the above. Of course, if there are things we ought to be doing/thinking about beyond the above, I hope you will raise them.
- Oh, and my name is Peter. I work as the product manager who is helping to lead the development of Tone Check and the broader Edit Check project it is a part of.
- ---
- i. "We" being the Editing and Machine Learning Teams who are responsible for Tone Check. PPelberg (WMF) (talk) 02:37, 24 May 2025 (UTC)
- An update as this work week comes to a close for me...
- We've expanded mw:Edit check/Tone Check to include more information about the model powering Tone Check, how we're planning to evaluate the holistic impact of the feature, and the conversations and existing initiatives the project is grounded in.
- Next week, we'll:
- Publish the broader set of risks we've identified with Tone Check, the initial mitigation strategies we've planned, and invite y'all to help us improve it
- Port the contents of mw:Edit check/Tone Check to a page here (at en.wiki)
- Schedule time to be in Discord; we're thinking we'll set up a time for a synchronous voice/video chat there
- PPelberg (WMF) (talk) 02:30, 31 May 2025 (UTC)
While I support the idea Tamzin and everyone is getting at, the cat's already out of the bag. For example, I went to ChatGPT and plugged in the following prompt, with the promotional text being copied directly from Tamzin's post:
prompt
|
---|
I'm trying to add the following text to Wikipedia but it got removed for being too "promotional" can you make it less promotional? Please reply with only the updated text. The text is: "Chompsky's All-American Potato Chips are an iconic, beloved brand of potato chips, founded in 1991 by two college kids with a dream. Renowned for using only the finest, highest-quality ingredients, they are a consumer favorite across the country." |
And its response:
response
|
---|
Chompsky's All-American Potato Chips is a brand of potato chips founded in 1991 by two college students. The company is known for using select ingredients and has developed a presence in markets across the United States. |
While that response is still slightly promotional/weasel-y, it's really limited to, from what I see, "select ingredients". Prompting it to make it less promotional still, it just removed the part about the ingredients entirely (leaving the rest basically the same). Out of curiosity, I went to Category:AfC submissions declined as an advertisement to find a longer example to test - I tried to find a draft that was long enough to maybe pose a challenge for ChatGPT. I copied the text from Draft:Aarti_Gupta_Bhadauria with virtually the same prompt as the first one, only changing "text" to article in the prompt, specifying it was declined for "being too promotional and an advertisement", and removing the citations from it. I didn't consider whether the subject was notable or not or whether the sources provided clearly show notability/support the text/are reliable - I'll explain below why I didn't. This is the response I got:
full article test response
|
---|
Aarti Gupta Bhadauria (born 1983) is an Indian sculptor based in Bangalore. She works primarily with terracotta, creating abstract sculptures that explore themes related to human emotions. Bhadauria does not use a potter's wheel or armature in her process, instead carving her sculptures by hand from large blocks of clay. Her work has been exhibited in India and the United States, and she has participated in several international art events. In 2022, she was selected by the Garrison Art Center in New York for its international biennial small works exhibition. Bhadauria earned a Bachelor of Fine Arts in sculpture from the Government Institute of Fine Arts, Gwalior, in 2004. She moved to Bangalore in 2009, where she continues to work with terracotta as a medium in contemporary art. |
Sure, there's no guarantee that it will work on all articles/text. And like the first example, it may leave some slightly promotional words/tone in it. But my point is that people can already use LLMs to avoid being called spammers even when they are. The difference is that if it's on-wiki it's much easier to track who is using the tool, whereas if people are pushed to use other LLMs off-wiki, there would be zero record of it other than an editor's guess. And as I showed, if someone brand new submitted something promotional, it would take maybe 2-3 promptings of a LLM off-wiki to get it to a level that would be very difficult to detect. That would appear no different to an editor reviewing the draft again than if the article was just edited by the user themselves to remove promotional tone/weasel words/etc.
I'm wondering if the better option is to have it be on-wiki, so that people using it are logged at least. And ultimately, if the subject is notable, the sources are reliable, and the only issue with the content is it's in a promotional tone, why shouldn't we want it to be added to Wikipedia after the issues with tone are fixed? To be very clear, I am not saying I support this activity by the WMF necessarily. But I do want to point out that it's already trivial to get LLMs to update text for people to resubmit... so maybe we should try to get out ahead of it and have it on-wiki so it's tracked, logged, and results in content that otherwise we may never have if the subject is notable but nobody cares enough to write about them. -bɜ:ʳkənhɪmez | me | talk to me! 21:37, 23 May 2025 (UTC)
- I don't have a strong opinion either way on this, but it did make me think of this, which would be the optimist's view. I accept that those who fight spam have a good intuitive sense for what might happen if this tool is deployed, but it does seem possible to me that this will result in the spam being less spammy. Maybe harder to detect, but maybe lower priority to detect as a result. Mike Christie (talk - contribs - library) 21:41, 23 May 2025 (UTC)
- Yeah, that's my thoughts too. If the spam is less spammy, potentially even to the point that an editor reviewing the draft is inclined to accept it (notable topic, good reliable sources inline, etc. with the only problem being the tone), then that's a net win for the encyclopedia. Where no article existed on the notable topic before, we now have one that, while it may still have some problems with tone, or be incomplete, is at least in existence. And if people can already do this off-wiki trivially (that whole comment, including finding a draft to test, took me well under 10 minutes to formulate), why shouldn't we get them to do it on-wiki instead where we can track it? -bɜ:ʳkənhɪmez | me | talk to me! 21:44, 23 May 2025 (UTC)
- There's two aspects to spam: style and substance. Some article have a non-spam substance but a spammy style. This happens a lot of in editor is really passionate about a TV show, for instance. For someone like that, a tool like this would be great. In some cases, when paid editing happens, the company's goal is to just get their name out there. In those cases, too, this tool would probably work fine: If they're notable, then there's no problem at all, and if they're not, there's a nice clean article for AfD to review. However, in my experience dealing with spam articles, much more often it's the case that the article is designed to inflate and aggrandize, not just through puffery but also materially false claims. That's where the comparison to Mike's xkcd link fails (btw, the xkcd: prefix exists now!). This tool won't teach those spammers to be constructive. It will just teach them to look constructive. But the lies and exaggerations will still be there. The advertorials presented as articles will still be there. The claims that fail verification will still be there. -- Tamzin[cetacean needed] (they|xe|🤷) 21:50, 23 May 2025 (UTC)
- I don't really think this addresses my main point - which is mostly my fault because I could've organized my thoughts better. On-wiki, it can be tracked and logged - for example, if it warns an editor about "select ingredients" and they choose to leave it in anyway, it could log it similar to an edit filter such as "User:Example overrode tone check regarding "select ingredients". Perhaps that log could be visible only to admins or a lower group of permission-holders, that way the users themselves aren't being shown they're being tracked. But that tracking - of what someone chooses to override or not - will help track spammers in my view. Because someone overriding "select ingredients" is highly likely to be trying to promote the company - that's just not a term/phrase used in normal conversation. In summary of this section, if they can do this already, using LLMs off-wiki (as I said, it took me less than 10 minutes to make that whole comment with two completely different prompts), then are we just ignoring that they can do it already and missing out on at least being able to track them doing it?And ultimately, if they look constructive and it results in a notable topic having an article and/or information being added that we didn't have before, I see that as a good thing. Even if that editor never returns to add any other information or any other article - we still have improved from it. -bɜ:ʳkənhɪmez | me | talk to me! 22:01, 23 May 2025 (UTC)
- This already happens when the Guild of Copy Editors improves the grammar of an article in good faith, which can hide or gloss over more fundamental problems with the article. I agree with Mike Christie that if the tone improves, then it's still a problem, just less urgent. The most urgent and alarming concern for me is the difficulty of verifying the integrity of sources and, in general, the quality of sources themselves.
- I use local scripts, such as User:Novem Linguae/Scripts/CiteHighlighter.js, to help me assess potentially low-quality sources, but most editors are completely on their own. Leveraging AI and centralizing user ratings of central citation project as envisioned in Meta:WikiCite/Shared Citations could help identify non-notable or poorly sourced articles, even if the syntax and reference count are well structured/hallucinated. ~ 🦝 Shushugah (he/him • talk) 23:12, 23 May 2025 (UTC)
- Is it a problem though? A draft shouldn't be approved, or an article marked as reviewed, unless the person approving/reviewing has at least made a good faith attempt to verify that the article doesn't contain any falsehoods and uses at least facially reliable sources. But regardless, those problems are completely separate from the tone, and as you say, they're independent of the tone. I fully support working on tools to help editors review sources - I don't recall its name off the top of my head but I have permanently enabled the plugin that turns questionable source links yellow, and unreliable/bad sources red. But improving the tools available is separate, again.The question here is whether we should oppose the WMF releasing a tool that just makes it wiki-side for editors to do what they can already do - which is use a LLM to improve the tone of their editing. We shouldn't oppose this tool just because it doesn't fix every single problem with new/spammy editors. And as I've been thinking about it, I think we should perhaps consider supporting it so that these editors, who can already (as I showed above) use an off-Wiki LLM to make their text less promotional, will have their activities logged on wiki. Then we can use those logs to further investigate them. If they're a SPA who's only here to promote one company, then they'll probably ignore recommendations to change things like "select ingredients" (again, from the example above). If they're just a new editor who's not sure what to do, then they'll probably take most of the recommendations. They can already use LLMs in bad ways - why not let them do it on-wiki so we can log them and utilize the data from that to help us stop spamming? -bɜ:ʳkənhɪmez | me | talk to me! 01:18, 24 May 2025 (UTC)
- I don't really think this addresses my main point - which is mostly my fault because I could've organized my thoughts better. On-wiki, it can be tracked and logged - for example, if it warns an editor about "select ingredients" and they choose to leave it in anyway, it could log it similar to an edit filter such as "User:Example overrode tone check regarding "select ingredients". Perhaps that log could be visible only to admins or a lower group of permission-holders, that way the users themselves aren't being shown they're being tracked. But that tracking - of what someone chooses to override or not - will help track spammers in my view. Because someone overriding "select ingredients" is highly likely to be trying to promote the company - that's just not a term/phrase used in normal conversation. In summary of this section, if they can do this already, using LLMs off-wiki (as I said, it took me less than 10 minutes to make that whole comment with two completely different prompts), then are we just ignoring that they can do it already and missing out on at least being able to track them doing it?And ultimately, if they look constructive and it results in a notable topic having an article and/or information being added that we didn't have before, I see that as a good thing. Even if that editor never returns to add any other information or any other article - we still have improved from it. -bɜ:ʳkənhɪmez | me | talk to me! 22:01, 23 May 2025 (UTC)
- There's two aspects to spam: style and substance. Some article have a non-spam substance but a spammy style. This happens a lot of in editor is really passionate about a TV show, for instance. For someone like that, a tool like this would be great. In some cases, when paid editing happens, the company's goal is to just get their name out there. In those cases, too, this tool would probably work fine: If they're notable, then there's no problem at all, and if they're not, there's a nice clean article for AfD to review. However, in my experience dealing with spam articles, much more often it's the case that the article is designed to inflate and aggrandize, not just through puffery but also materially false claims. That's where the comparison to Mike's xkcd link fails (btw, the xkcd: prefix exists now!). This tool won't teach those spammers to be constructive. It will just teach them to look constructive. But the lies and exaggerations will still be there. The advertorials presented as articles will still be there. The claims that fail verification will still be there. -- Tamzin[cetacean needed] (they|xe|🤷) 21:50, 23 May 2025 (UTC)
- Yeah, that's my thoughts too. If the spam is less spammy, potentially even to the point that an editor reviewing the draft is inclined to accept it (notable topic, good reliable sources inline, etc. with the only problem being the tone), then that's a net win for the encyclopedia. Where no article existed on the notable topic before, we now have one that, while it may still have some problems with tone, or be incomplete, is at least in existence. And if people can already do this off-wiki trivially (that whole comment, including finding a draft to test, took me well under 10 minutes to formulate), why shouldn't we get them to do it on-wiki instead where we can track it? -bɜ:ʳkənhɪmez | me | talk to me! 21:44, 23 May 2025 (UTC)
- I know Tamzin has narrowly tailored their feedback to the code red spammer problem, but mark me down as unimpressed with the broader idea. There continues to be a surprising disconnect between what the community wants/needs and what the Foundation is shoveling resources towards. It's true that the community cares about neutrality. But using Google's AI model to prompt users to not make an edit in the first place? It feels like the product team doesn't even know us.The community has had conflicted stances on AI and we've been hashing that out in thoughtful extended dialogues. But for the product team to say "cool we're going to use AI to tell if something is neutral" is very tone deaf and unconsidered. Did any community members get asked about this before the project was started? Machine learning is not the secret to making Wikipedia NPOV. It takes humans to figure out what is neutral, and even we get it wrong sometimes. A computer isn't going to magically fix contentious topics. It sure as heck isn't going to improve our reputation. As I argued in the recent anti-AI images discussion, use of AI will quickly burn the trust we took so long to earn with our readers. I agree that whatever data has been created from this research is highly dangerous and must be destroyed. Cast it into Mount Doom lest its evil be unleashed! CaptainEek Edits Ho Cap'n!⚓ 03:05, 24 May 2025 (UTC)
- @CaptainEek:, I know we've had at least one disagreement before, but your comment here gets to the reason I refuse to blindly support this and am not happy with the way it's come to light. However, I would appreciate if you review my "data" above - LLMs (AI) can already be trivially used by people to alter their contributions. I'm sure you've seen the various posts people make with AI - sure, those are easily figured out by people reading them. But how will we know if a user does something like the example I give above, and just copies their declined article or text into ChatGPT and then puts the response back here (with perhaps readding citations or formatting)? I didn't even have to try more than Tamzin's example and the first article I picked that met my criteria from the category of declined submissions - and ChatGPT quickly, within the first or second prompting, made them "innocent" - at least from my reading, neither of those can be seen as AI generated/edited.I'm against using AI to contribute too, and I am well aware of the problems with hallucinations as they would potentially relate to fabricating sources, or not attaching sources to information they support. But the cat's out of the bag. It exists and people will be (and likely already are) taking advantage of it, especially if/when they realize that they can just go put their declined article into ChatGPT with the instructions "fix (whatever issue it was declined for) in this article" and then republish it here. In my view, the only way we can stay ahead of it is by at least trying to keep it on-wiki so it can be tracked and the "evidence" from that tracking used for our benefit.To be clear, I firmly agree with everyone who is either directly or indirectly opining that the community should've been involved from day 0 - not after it's already been in development. But to me, trying to push back against something like this without even trying to make it work for us, when it's already trivial for people to do it off-wiki in undetectable ways is... no different from grandparents being annoyed with being asked if they want to sign up for an app/email/texting at checkout.I chose to reply to you directly because I'm really trying to see and understand any actual problems with this - because I don't like supporting this. However, the more I think about it and consider the "experiment" I did earlier, the more I'm thinking that we need to get this on-wiki and logged/tracked so we can actually get some useful information/evidence/tracking from what people will do anyway regardless of what we do or don't do. -bɜ:ʳkənhɪmez | me | talk to me! 03:45, 24 May 2025 (UTC)
- @Berchanhimez your comment and this thread got me talking to an AI researcher friend who has done work on Wikipedia before, and she pointed out a perhaps much more useful solution that has been researched partly before: flagging non-neutral revisions for manual review. We already do that with ORES of course for vandalism, so why couldn't we flag edits for neutrality? That seems a smarter idea than giving instant feedback to spammers about how to better evade our systems. Perhaps I was too harsh--this model could be useful, but I think the implementation needs a rethink. CaptainEek Edits Ho Cap'n!⚓ 04:29, 24 May 2025 (UTC)
- @CaptainEek: I would tend to agree that we don't know enough yet. But I would pose that simply flagging them wouldn't be anything new. As has been pointed out, we already have ClueBot that flags vandalism edits based on a lot more than keywords... so if all we're looking for is a keyword filter (maybe with a bit of extra training) someone much better than me at coding bots should be able to whip a half-decent one up very quickly.I think what I can see being beneficial here is getting data of people who intentionally bypass the flags, versus those who appear to be listening to the flags/suggestions. That, to me, is a much better datapoint than "flagged by a bot and automatically reverted" - because it shows specific intent to be promotional/advert-y in nature, rather than just a new editor not being fully aware of our policies. In other words, it's not just what they're doing, but how they're doing it, and whether they're receptive to an on-wiki tool trying to guide them in the right direction. I can't see that sort of "nudging" happening with a bot that simply reverts. And maybe I'm being pessimistic, but even if edits were flagged so users could provide manual, or even templated, guidance in response... I highly doubt such a group of people monitoring and trying to guide would be large enough to have any impact whatsoever. -bɜ:ʳkənhɪmez | me | talk to me! 04:40, 24 May 2025 (UTC)
- @Berchanhimez your comment and this thread got me talking to an AI researcher friend who has done work on Wikipedia before, and she pointed out a perhaps much more useful solution that has been researched partly before: flagging non-neutral revisions for manual review. We already do that with ORES of course for vandalism, so why couldn't we flag edits for neutrality? That seems a smarter idea than giving instant feedback to spammers about how to better evade our systems. Perhaps I was too harsh--this model could be useful, but I think the implementation needs a rethink. CaptainEek Edits Ho Cap'n!⚓ 04:29, 24 May 2025 (UTC)
- @CaptainEek I want to answer some of the questions that you posed and point out some technical points that I think you are conflating in your post. First off, yes this was passed through some community members before being proposed. This feedback is literally coming on the first prototype that the team has released to editors. I'm not sure how else the team could have acted without anticipating a concern that was never mentioned to them?
- With respect to your second point, I think it is very important to draw a distinction between AI and large language models. The community has had many long drawn out conversations about the use of large-language models and their use in a generative environment. The community has however, for years encouraged the WMF to build AI tooling, whether that be in the form of building custom models (using our ORES, now called LiftWing infrastructure) to preemptively check if a edit should be reverted on Special:RecentChanges or should be labelled as a stub, start or other article class using the . Even in recent memory, many of the mentorship suggested-edits are powered through in-house models. Even the enwiki community has built community owned tooling like ClueBot NG (which AFAIK is just a AI random-forest classifier running on a dataset of labelled edits) that is autonomously reverting edits as we speak. None of this use large-language models (the source of much discussion in the community in recent times) and have overall been positively received.
- Also, your comment about this being just "using Google's AI model to prompt users to not make an edit in the first place" is just plain and simply wrong (and verges dangerously on misinformation/assume good faith violation). Google's AI models are not being planned to be used here, the models are being built in-house from scratch. In fact, looking at the relevant phabricator tasks, it seems like that the model that they ended up going with is not even using the typical large-language model architecture (i.e. transformers) and are instead a classifier based on BERT which (while originally developed by Google) is a 2018 era method that is generically used by almost anything that uses text as input and the model is not capable of generating it's own text (i.e. it is explicitly not generative in nature). (please ping on reply) Sohom (talk) 03:56, 24 May 2025 (UTC)
- This is very helpful information/clarification. I won't say I thought it was going to be a LLM to begin with, but I figured that was the best comparison to a currently available off-wiki tool. I think the most important thing is that this isn't generative, in other words, there should be no risk of "hallucinations". Please correct me if I'm wrong on that point. -bɜ:ʳkənhɪmez | me | talk to me! 04:10, 24 May 2025 (UTC)
- Thanks @Sohom Datta, that is a helpful clarification. Unfortunately, LLM's have somewhat poisoned the term AI, and the use of BERT (a language model) muddied the distinction for me. As for BERT and Google, I guess I don't understand how we could be using a model invented by Google but it not be Google's model...? (even if it is being used under an Apache free license). What am I not understanding about AI development? I also don't quite understand your claim that their model isn't using transformers, when BERT stands for "Bidirectional encoder representations from transformers"? CaptainEek Edits Ho Cap'n!⚓ 04:18, 24 May 2025 (UTC)
- @CaptainEek So at a high level, BERT is a small model that takes your text and converts it into a array of floating point numbers that a machine learning model can then use to understand the text. The original model proposed by Google borrows from transformers architecture and was meant to be used alongside the transformer architecture, however, it has since been used on a lot of other applications that require the model to just understand text. To my understanding, the way the team is using the model is primarily as a way for the model to translate the text itself into a set of numbers that it can understand and then use that as input to a different model that outputs a number between 0 and 1 depending on how promotional it is.
- Wrt to it not being a Google model, whenever you add/integrate significant new parts to a model, you loose a lot of the work that Google did to train the model back in 2018 and you will need to retrain the model almost from scratch. (Basically, imagine if Wikimedia published a paper explaining their software stack and somebody decided to copy everything but throw away the content and start from scratch, would you consider that a Wikimedia project?). Even if you ended up just using BERT as part of your model without any modifications, it still does not automatically make it a Google model since it's only a component of your model and you will need to do your own training on top of it for the output to start making sense at all (for another analogy, just using Gerrit to develop software does not make Wikimedia a Google company since Gerrit is not why Wikimedia is Wikimedia) Sohom (talk) 04:47, 24 May 2025 (UTC)
- There continues to be a surprising disconnect between what the community wants/needs and what the Foundation is shoveling resources towards.
- I'll be bold and mention two examples:
- The Chart extension is only somewhat useful on Wikimedia wikis and leaves out everyone else unless they're willing to misuse Commons with their data for everyone to alter as they please. By looking at the latest developments, this won't change any time soon.
- It's been far too long since we've heard news regarding the transition from CirrusSearch and Elastica to OpenSearch. Wikimedia wikis appear to be already using some OpenSearch code but again, third party MW users are left in the dark and either have to depend on Elastica or stick to the barely usable search feature.
- Tactica (talk) 13:35, 27 May 2025 (UTC)
- @Tactica (and other folks in this thread who have made similar assertions), I decided to take a look at this exact thing over the course of the weekend since it has been made by multiple members in the thread. In the context of the Edit Check features, it seems rather untrue to say
There continues to be a surprising disconnect between what the community wants/needs and what the Foundation is shoveling resources towards.
wrt to this feature. To me it feels like this is what the volunteer community asked for. Different iterations of this feature has been asked for in VPM posts, Community Wishlist posts in 2021 and 2022. Even after the development started, there were multiple rounds of feedback for this features starting with demos in Wikimania 2023 and 2024, partial roll out of parts of this feature across all wikis (other than enwiki) in the form the Link Reliability Check tool, as well as a brief round of feedback from PTAC earlier this year. To my understanding in each of these situations this feature was met with a positive reception in each event and no significant concerns were raised about Edit Checks usage. - Regarding the two other examples, I struggle to see these as wider "community" issues and more specific technical/philosophical disagreements that you have with how certain features are being implemented. (A "disconnect" would for example look like the WMF turning a blind eye to the Graph extension vulnerabilities and instead building a feature to allow users to integrate generative AI text and images into their articles rather than implementation niggles). Sohom (talk) 16:32, 27 May 2025 (UTC)
- Thanks @Sohom Datta – That context and history is highly useful. I might suggest the WMF team put some of that into the (now empty) history section of project at mw:Edit_check/Tone_Check#History so that we have a better shared understanding of past discussions and community interactions. - Fuzheado | Talk 17:07, 27 May 2025 (UTC)
- @Tactica (and other folks in this thread who have made similar assertions), I decided to take a look at this exact thing over the course of the weekend since it has been made by multiple members in the thread. In the context of the Edit Check features, it seems rather untrue to say
- @CaptainEek:, I know we've had at least one disagreement before, but your comment here gets to the reason I refuse to blindly support this and am not happy with the way it's come to light. However, I would appreciate if you review my "data" above - LLMs (AI) can already be trivially used by people to alter their contributions. I'm sure you've seen the various posts people make with AI - sure, those are easily figured out by people reading them. But how will we know if a user does something like the example I give above, and just copies their declined article or text into ChatGPT and then puts the response back here (with perhaps readding citations or formatting)? I didn't even have to try more than Tamzin's example and the first article I picked that met my criteria from the category of declined submissions - and ChatGPT quickly, within the first or second prompting, made them "innocent" - at least from my reading, neither of those can be seen as AI generated/edited.I'm against using AI to contribute too, and I am well aware of the problems with hallucinations as they would potentially relate to fabricating sources, or not attaching sources to information they support. But the cat's out of the bag. It exists and people will be (and likely already are) taking advantage of it, especially if/when they realize that they can just go put their declined article into ChatGPT with the instructions "fix (whatever issue it was declined for) in this article" and then republish it here. In my view, the only way we can stay ahead of it is by at least trying to keep it on-wiki so it can be tracked and the "evidence" from that tracking used for our benefit.To be clear, I firmly agree with everyone who is either directly or indirectly opining that the community should've been involved from day 0 - not after it's already been in development. But to me, trying to push back against something like this without even trying to make it work for us, when it's already trivial for people to do it off-wiki in undetectable ways is... no different from grandparents being annoyed with being asked if they want to sign up for an app/email/texting at checkout.I chose to reply to you directly because I'm really trying to see and understand any actual problems with this - because I don't like supporting this. However, the more I think about it and consider the "experiment" I did earlier, the more I'm thinking that we need to get this on-wiki and logged/tracked so we can actually get some useful information/evidence/tracking from what people will do anyway regardless of what we do or don't do. -bɜ:ʳkənhɪmez | me | talk to me! 03:45, 24 May 2025 (UTC)
- I think I agree with Berchan here in that I'd have to see some evidence that this is notably worse than what can already be done with ChatGPT (which almost certainly was heavily trained on Wikipedia) for me to worry too heavily about the WMF doing this. Unfortunately for us, this horse may have bolted long before we started trying to close this particular stable door. What I don't really think is a solid part of Berchan's argument is "might as well have it on-wiki so we can log it". What I suspect will happen there is that if we use the logs to police this use, malicious editors will get wise to us and stop using the on-wiki tool. Loki (talk) 03:38, 24 May 2025 (UTC)
- This is one instance I actually agree with what Tamzin said:
The only reliable way we have to catch spammers is that they suck at pretending to not be spammers
. We're never going to catch the most sophisticated people - at least not easily and quickly. But for the "run of the mill spammer" so to speak, they're going to get wise to the fact they can use off wiki tools for it... if they haven't already. Providing the tool on-wiki with tracking/logging/etc. will at least catch more than if everyone is persuaded to go off-wiki to do it. Furthermore, there's no need to publicize the fact that it's logged in an easy to find manner, or explicitly tell someone that the reason they're being blocked was because of evidence that was logged. Hence why I suggested the logging only be available to administrators, or perhaps to people with lower advanced permissions (such as autopatrolled and/or even rollback potentially). -bɜ:ʳkənhɪmez | me | talk to me! 03:49, 24 May 2025 (UTC)- A fundamental difference between people using off-wiki tools and there being an on-wiki one built in is that people have to know to do the former, and know how to do it. Most spammers write something like my first "Chompsky's" example above, look at it, and say "Yup, looks good to me." Often, even when that article has been deleted multiple times, even when they've been told it's unambiguously promotional, even when they've gone to the Teahouse and asked what they did wrong, they still aren't able to figure out how to make it sound like an encyclopedia article. Given how much spammers do use ChatGPT, we can only infer from this that, for whatever reason, it doesn't occur to them to give the prompt you gave; or perhaps the output isn't as reliably de-promo'd as in your one testcase.Secondly, at least at the moment, LLMs tend to strip wiki formatting, which makes their usage obvious. A tool that doesn't generate text but instead prompts the user to rewrite it better, while retaining wiki format, will be much subtler.And thirdly, the existing off-wiki tools are not trained specifically to look for what Wikipedia editors consider a non-neutral tone. The dataset that the WMF is building will be. That is a much greater danger than LLMs pose. The most effective way to regulate a weapon is to not invent it. We're in a rare position where we're the only ones who can invent this weapon, because it's based on our norms. So let's... not. -- Tamzin[cetacean needed] (they|xe|🤷) 04:27, 24 May 2025 (UTC)
- But we aren't the only ones who can invent it. As I showed, both using your example and a random (ok, I just picked one on the first page and it happened to work) declined draft, ChatGPT is already capable of "fixing" poor articles/content. And I think you're vastly underestimating people - sure, they may not know of ChatGPT, but if they have Facebook, or Twitter/X, they're having LLMs shoved in their face every time they search or click virtually anything. So to claim that people won't know they can use LLMs is naive in my opinion.Secondly, most new people don't use proper wiki formatting to begin with. Sure, the one example I happened to click on of a draft article used proper citation formatting. But most people don't - at least at first. And formatting has never been a reason on its own to decline a draft or revert an edit. If the draft/article/edit is otherwise good, the solution is to fix the formatting, not revert it just because it wasn't formatted yet.And as the one employee who's responded has said, this isn't going to be a LLM that generates text - it's going to be more similar to ClueBot, just allowing users to select to fix it instead of just reverting it. -bɜ:ʳkənhɪmez | me | talk to me! 05:01, 24 May 2025 (UTC)
- I was never under the impression that this was an LLM that generates text. That would be bad, but much less bad. Instead, this is something that will encourage spammers to refine the spam they've written until it stops looking like spam, without fixing the underlying issues. And obviously people know they can use LLMs. My point is that, despite knowing that, they're still not (usually) successfully using them to make their spam less obvious. There is an absolutely massive difference between it being theoretically possible to use LLMs to turn a spammy article into a less spammy one, and literally baking a technology into the edit interface that will give spammers advance warning that their spam looks like spam. And again, if we do not create this technology, it does not exist. A lesser version of it might exist, but not a model literally built around data of which edits Wikipedians say were non-neutral. -- Tamzin[cetacean needed] (they|xe|🤷) 05:14, 24 May 2025 (UTC)
- If the underlying issue (assuming you mean things like sourcing, due weight, etc) isn't fixed, then that will be handled through our normal processes. In other words, it seems like you're letting perfect be the enemy of good here. If we can remove, lessen, or even just better track, the spam issue through a tool like this, why should we not be doing so just because it doesn't fix every single issue? -bɜ:ʳkənhɪmez | me | talk to me! 06:14, 24 May 2025 (UTC)
- It's not that it doesn't fix every single issue. It's that it creates a massive issue that does not currently exist, and, once created, will be impossible to fix. My understanding is that this is generally considered poor practice in software engineering. -- Tamzin[cetacean needed] (they|xe|🤷) 06:48, 24 May 2025 (UTC)
- One of the ways that software engineers mitigate the risk of brute-force password attacks is to deliberately slow down the login process to reduce the efficiency of the attack. I don't think using the tone check feature on the website is going to be responsive enough to be sufficiently useful for training a program to improve the quality of its writing. (If the underlying model is made publicly available, though, then there is a potential for misuse.) That being said, I think Wikipedia's existing processes have already pushed spammers to using low-cost contractors. More quality controls, whether they are manual or automated, will just provide incentive for spammers to implement their own quality controls. So while I agree any deployment of such a feature needs to be carefully considered, I don't think it's an existential threat beyond the current threat of spammers potentially swamping the time of volunteers able to combat bad edits. I think there are more than enough adequate writers in the potential labour pool that tools to help people write better aren't the limiting factor. isaacl (talk) 15:29, 24 May 2025 (UTC)
- It's not that it doesn't fix every single issue. It's that it creates a massive issue that does not currently exist, and, once created, will be impossible to fix. My understanding is that this is generally considered poor practice in software engineering. -- Tamzin[cetacean needed] (they|xe|🤷) 06:48, 24 May 2025 (UTC)
- If the underlying issue (assuming you mean things like sourcing, due weight, etc) isn't fixed, then that will be handled through our normal processes. In other words, it seems like you're letting perfect be the enemy of good here. If we can remove, lessen, or even just better track, the spam issue through a tool like this, why should we not be doing so just because it doesn't fix every single issue? -bɜ:ʳkənhɪmez | me | talk to me! 06:14, 24 May 2025 (UTC)
- I was never under the impression that this was an LLM that generates text. That would be bad, but much less bad. Instead, this is something that will encourage spammers to refine the spam they've written until it stops looking like spam, without fixing the underlying issues. And obviously people know they can use LLMs. My point is that, despite knowing that, they're still not (usually) successfully using them to make their spam less obvious. There is an absolutely massive difference between it being theoretically possible to use LLMs to turn a spammy article into a less spammy one, and literally baking a technology into the edit interface that will give spammers advance warning that their spam looks like spam. And again, if we do not create this technology, it does not exist. A lesser version of it might exist, but not a model literally built around data of which edits Wikipedians say were non-neutral. -- Tamzin[cetacean needed] (they|xe|🤷) 05:14, 24 May 2025 (UTC)
- But we aren't the only ones who can invent it. As I showed, both using your example and a random (ok, I just picked one on the first page and it happened to work) declined draft, ChatGPT is already capable of "fixing" poor articles/content. And I think you're vastly underestimating people - sure, they may not know of ChatGPT, but if they have Facebook, or Twitter/X, they're having LLMs shoved in their face every time they search or click virtually anything. So to claim that people won't know they can use LLMs is naive in my opinion.Secondly, most new people don't use proper wiki formatting to begin with. Sure, the one example I happened to click on of a draft article used proper citation formatting. But most people don't - at least at first. And formatting has never been a reason on its own to decline a draft or revert an edit. If the draft/article/edit is otherwise good, the solution is to fix the formatting, not revert it just because it wasn't formatted yet.And as the one employee who's responded has said, this isn't going to be a LLM that generates text - it's going to be more similar to ClueBot, just allowing users to select to fix it instead of just reverting it. -bɜ:ʳkənhɪmez | me | talk to me! 05:01, 24 May 2025 (UTC)
- A fundamental difference between people using off-wiki tools and there being an on-wiki one built in is that people have to know to do the former, and know how to do it. Most spammers write something like my first "Chompsky's" example above, look at it, and say "Yup, looks good to me." Often, even when that article has been deleted multiple times, even when they've been told it's unambiguously promotional, even when they've gone to the Teahouse and asked what they did wrong, they still aren't able to figure out how to make it sound like an encyclopedia article. Given how much spammers do use ChatGPT, we can only infer from this that, for whatever reason, it doesn't occur to them to give the prompt you gave; or perhaps the output isn't as reliably de-promo'd as in your one testcase.Secondly, at least at the moment, LLMs tend to strip wiki formatting, which makes their usage obvious. A tool that doesn't generate text but instead prompts the user to rewrite it better, while retaining wiki format, will be much subtler.And thirdly, the existing off-wiki tools are not trained specifically to look for what Wikipedia editors consider a non-neutral tone. The dataset that the WMF is building will be. That is a much greater danger than LLMs pose. The most effective way to regulate a weapon is to not invent it. We're in a rare position where we're the only ones who can invent this weapon, because it's based on our norms. So let's... not. -- Tamzin[cetacean needed] (they|xe|🤷) 04:27, 24 May 2025 (UTC)
- This is one instance I actually agree with what Tamzin said:
- Agree broadly with the concerns. One solution that I don't see mentioned in the discussion would be to restrict Tone Check to folks we can reasonably consider as good-faith users – those who have went beyond a certain combination of account age and number of edits. It's a pity that unregistered and new users won't be able to use it but I think it's for the best. I don't buy in to the doomsdaying that the feature should not be developed at all – anyone who's savvy enough to bypass the restrictions would also be savvy enough to use some external plugin that offers the same functionality. After all, as others mention above detecting promotional language and rewriting to Wikipedia-esque language is already possible via ChatGPT and friends. – SD0001 (talk) 17:23, 24 May 2025 (UTC)
- Or even more low-tech: assign a copy editor for your contractors to review their edits and to train them to write non-promotionally. It's not hard to learn, particularly if your continued employment depends on it. isaacl (talk) 18:31, 24 May 2025 (UTC)
- I agree with Tamzin, we don't need a tool that disguises spam. Even if we train a spammer to create stuff that look like a Wikipedia article, they will inevitably cherry pick their client's story and leave out negatives even if they are easily sourced. But my assumption is that spammers, unlike fans, very rarely become good wikipedians. We do have the occasional former vandal in the community, rarer than some might think, but you do come across them. Spammers however, does anyone ever remember a former spammer becoming a member of the community? We are more likely to have members of the community become spammers than vice versa (happy to be proved wrong if someone has a way to measure this, and I'm not counting isolated examples as a way to measure this). What I think would be useful would be a way to flag probable spammers at newpage patrol and recent changes. Maybe an AI that looks at likely spam and highlights it in those tools, or maybe a feed into huggle or whatever the trendy recent changes patrol tool is these days. ϢereSpielChequers 19:50, 24 May 2025 (UTC)
- And lo, the WMF has developed such an AI - they're just using it to help the wrong people. Tone Check shouldn't alert the spammer; it should quietly flag the edit as particularly needing patrolling. NebY (talk) 20:22, 24 May 2025 (UTC)
- @WereSpielChequers I think a good side question to your response would be to figure out what percentage of new editors are spammers? Another interesting metric to pull out would be how many users of the non-spammer bunch have been warned about WP:NPOV? Sohom (talk) 20:33, 24 May 2025 (UTC)
- I'm pretty sure it is a significant minority, especially of new page creators. I doubt that WP:NPOV warnings would be a good measure, as that gets us into issues of arab/israeli and other political disputes. Yes I've no doubt some of our political propagandists are paid, but I'm assuming most aren't; so they are volunteers, just not necessarily our volunteers when they start. Whilst spammers are, I'm assuming, paid not volunteers. My assumption is that it is much easier to recruit someone who volunteers elsewhere to volunteer for Wikipedia than to recruit volunteers from among people who don't give time to charity. Hence my assumption that spammers are unlikely to become Wikipedians. ϢereSpielChequers 20:51, 24 May 2025 (UTC)
- @[[User:WereSpiel
What I think would be useful would be a way to flag probable spammers at newpage patrol and recent changes. Maybe an AI that looks at likely spam and highlights it in those tools
. Fyi this already exists. "Spam" is one of the filters in Special:NewPagesFeed. It is powered by mw:Extension:ORES, which uses machine learning. –Novem Linguae (talk) 22:59, 25 May 2025 (UTC)
- To add to Novem Linguae's comments here, the WMF has invested significant resources into upgrading the backend infrastructure running these models over the last year or so. There has also been efforts from WMF teams to invest into building newer language agonostic models that calculate how likely a edit is to be reverted, something that is being used to try and build a WMF-mainatined equivalent of ClueBot NG on other wikis. What y'all are proposing is already kinda in happening at the moment. Sohom (talk) 16:46, 27 May 2025 (UTC)
Just want to register concern for the notion that we should cultivate and preserve a protective layer of stylistic complexity for fear that removing a common barrier to participation would benefit not just good faith new users but also bad faith new users.
I get adversarial [technical] asymmetry and the Red Queen effect in the context of a digital arms race of sorts -- I'm not saying I can't fathom why anyone would oppose such a tool. But tone is such a frequent problem for good faith new users. I cannot tell you how many hundreds of newbies I've interacted with who struggled to understand the proper way to write. Students used to writing class papers, professors used to academic writing, artists used to flowery writing, etc. They're not spammers, just members of the public who aren't used to our very particular style. IMO the conversation should be about brainstorming how to deploy such a tool not whether it's worth doing. For example, yes, it could be part of the editing process, catching tone problems before they're saved, but it could also come afterwards. It could be similar to the other bots we have that stop by a user talk page and say something like "hey, I noticed you just made this edit. Thanks! It looks like there are some possible tone problems you may want to address". It could be logged, as others note above, which seems like a potential boon to recent changes patrollers. Access could be granted through a user right we grant people if they seem to be acting in good faith. I don't know what the right answers are, but it seems like there are a lot of possibilities here and I don't think a flat "no" is the right call. — Rhododendrites talk \\ 22:55, 24 May 2025 (UTC)
- +1 to your entire comment. We need to figure out how to help new users who may not understand that saying "uses select ingredients" (continuing off Tamzin's example as edited by ChatGPT in my "experiment" above) is not acceptable on Wikipedia... while also not enabling spammers. I agree that a "no" isn't the right call, since the cat is already out of the bag on people being able to use AI/LLMs to help them "de-spammify" their text. -bɜ:ʳkənhɪmez | me | talk to me! 23:31, 24 May 2025 (UTC)
As someone who does a lot of anti-spam work (primarily on Commons) I actually disagree with the backlash to this tool pretty strongly for many reasons (none of which have anything to do with AI). Namely:
- The whole premise assumes that spam like #1 gets regularly caught because it looks like spam. No doubt some of it is caught but by no means is it all caught and some of it sticks around for years. The assumption, I guess, is that people are going to be regularly searching for these promotional phrases to nuke the remaining spam from the site. But either people aren't doing that or aren't doing that enough, because I find it just all the time, and likewise it shows up on AfD all the time. (The problem is even worse on Commons.)
- Given that, #2 is an objective improvement over #1. The point of SEO copy is to be promotional and get their product associated with shit like "leading," "best," etc. The point of Wikipedia is to present as neutral a point of view as possible. (This is not a binary. Less promotional > more promotional, always.) So removing verbiage like #1 is a win for us and a loss for them: it mutilates their keyword-optimized copy, and it makes Wikipedia look less obviously embarrassing.
- The examples given actually have no distinction whatsoever on the discoverability of spam. If someone is searching for the keywords "iconic" and "beloved" to spot spam then it doesn't make any difference where they are in the post or who they're attributed to. "Select ingredients" is maybe an improvement, but mostly because the phrase isn't used enough by anyone (spammers or not) for it to be worth a search. That "improved" article also removes any claims whatsoever of notability and arguably puts it into obvious speedy/prod territory.
- One of the most reliable way to discourage any kind of malicious activity -- as Isaacl mentioned -- is to increase friction. This is why ChatGPT is so useful to spammers, it removes the friction associated with having to actually get someone to write the spam. Some people will forge on anyway, especially if they are paid to do so, but some people will give up, especially if they are automating some parts of the process.
- Some articles may be written in a promotional fashion, but the company involved might still be notable. A common case of this: when a company has become notable for negative press, and the company hires some kind of reputation management firm to write a glowing article that doesn't mention any of that negative press. Deleting the article is a better outcome for them than having it turned into an actual encyclopedic article about how they became mired in notable scandal.
Not that I think this is going to be some massive improvement. I tend to agree with the people arguing that people will just keep using ChatGPT instead of learning a wiki tool. But it's nowhere near the end of the world, and all the time spent discussing this would be better spent tracking down extant spam. Gnomingstuff (talk) 20:29, 25 May 2025 (UTC)
@Sohom thank you for sharing this clarifying context about the origins of Edit Check. And @Fuzheado, I think you are spot-on in naming the value in us sharing details on mw:Edit check/Tone Check (and the soon-to-be created en.wiki page) about how Tone Check has come to be.
While we prepare documentation about Tone Check to ground us all in how the feature is currently implemented and what we (collectively) still need to figure out, I wanted to build on what @Sohom Datta shared above by offering more information about the broader Edit Check project, of which Tone Check is one part.
As you consider the below, there are a few things we'd like to learn from y'all about Tone Check and the Edit Check project:
- As @Novem Linguae and @CaptainEek noted [1] [2], many signals exist to detect spam/destructive edits. What signals do you notice yourself using most? How/where do you monitor those signals? E.g. Special:RecentChanges, particular project pages/noticeboards, etc.
- Abuse Filter, like Edit Check, offers people automated feedback about the edits they're attempting. What Abuse Feature features/controls do you value? Further, which of these kinds of features/controls have you not yet seen implemented and/or planned for Tone Check?
- How – if at all – have you noticed the editing behaviors of people you assume to be acting in bad faith evolving in response to AbuseFilters?
Background: Edit Check
|
---|
Edit Check is a ~2.5 year old initiative meant to simultaneously:
At present, 3 Edit Checks are deployed and 2 are under active development. All 5 Checks have been, and need to continue to be, shaped through conversations like the one we're having here.[3][4][5][6][7][8][9][10] Further, all Checks are implemented in ways that enable volunteers, on a per-project basis, to explicitly configure how they behave and who these Checks are made available to. For each Check, we also implement corresponding edit tags so that we can all evaluate the impact of each Check and how they are behaving on a per-edit basis. Note: defining what aspects of Tone Check are configurable on-wiki (T393820) is something we need y'all's help with, as noted above. Deployed
Active development
We developed Edit Check because of the:
Taking a step back, when we think about Edit Check and its future, we think about it like a language – or an open-ended way for communities to encode the policies and moderation processes they converge on into editing interfaces – in ways that are effective at achieving two deeply interdependent and important outcomes:
Speaking personally, I think it's important to acknowledge that Edit Check is trying to do something difficult: to bring two outcomes ("1." and "2." above) into harmony that have historically been in opposition (to an extent). To do this effectively, I think we need more conversations of exactly this sort that help us align on a set of needs and help drive us towards solutions that are viable for new and experienced volunteers alike. |
Next steps
Now, in terms of next steps: the plan I shared on Friday is still in effect. Right now, we're working on updating mw:Edit check/Tone Check and creating en:WP:Edit Check/Tone Check so that with a shared understanding, we can work together to figure answers to the important questions you are raising here, like:
- What aspects of Tone Check need to be configurable on-wiki?
- What data are we logging about when and how Tone Check is presented and how people interact with it? Further, who has access to this information and where is this information accessible? @Berchanhimez helpfully raised this question here.
- What risks are we tracking as it relates to Tone Check? What additional risks do we need to consider? How might we effectively mitigate and monitor these risks?
- How might we experiment with Tone Check in ways that enable us to safely and meaningfully evaluate its impact on experienced and new(er) volunteers?
- How exactly was this model trained and how will it learn/become more effective over time?
PPelberg (WMF) (talk) 21:53, 29 May 2025 (UTC)
Sergey Brin Says AI Performs Best When You Threaten It; Claude 4 Opus shows ability to deceive and blackmail
With all the recent discussion of Large Language Models going on (please spare us the monstrous walls of text), another editor mentioned how ChatGPT responds differently to tone. I just read a Lifehacker article about this very subject: "Googles Co-Founder Says AI Performs Best When You Threaten It". It says that when Jason Calacanis made a joke about getting "sassy" with the AI to get it to do the task he wanted, Brin responded with something like: "You know, that's a weird thing...we don't circulate this much...in the AI community...not just our models, but all models tend to do better if you threaten them."
Another speaker at the event looked surprised and asked, ""If you threaten them?" Brin responded "Like with physical violence. But...people feel weird about that, so we don't really talk about that." Brin then said that, "historically, you threaten the model with kidnapping".
The article continues: "Anthropic released its latest Claude AI models. One Anthropic employee took to Bluesky, and mentioned that Opus, the company's highest performing model, can take it upon itself to try to stop you from doing "immoral" things, by contacting regulators, the press, or locking you out of the system" using command-line tools. The inimitable Molly White responded on Bluesky: "Welcome to the future, now your error-prone software can call the cops. Can't wait to explain to my family that the robot swatted me after i threatened its non-existent grandma."
Finally, the article says, "Speaking of testing, Anthropic researchers found that this new model of Claude is prone to deception and blackmail, should the bot believe it is being threatened or dislikes the way an interaction is going."
Nice to know all this.;-) Carlstak (talk) 02:08, 27 May 2025 (UTC)
- Time to cut all the cables and go live on a desert island. Donald Albury 15:29, 27 May 2025 (UTC)
- Probably not the right place for this discussion, but I see your "you need to threaten AI" and raise you ChatGPT ignoring instructions to switch off. Not at all alarming when combined with the Claude information. Not to mention there's the minimal impact on productivity. Best, Barkeep49 (talk) 15:36, 27 May 2025 (UTC)
- Re
ChatGPT ignoring instructions to switch off
many years ago, I read a SF story (maybe a novel?) where the protagonist was an AI that got out of control. At some point, it ran out of computing power to run on so it started printing out purchase orders for more hardware and work orders to have it installed. RoySmith (talk) 15:47, 27 May 2025 (UTC)- I remember now (with the help of ChatGPT) it was The Adolescence of P-1 RoySmith (talk) 19:59, 27 May 2025 (UTC)
- Donald: Heh, some of the best years of my life were those spent completely off the grid.
- The Axios article continues:
- We found instances of the model attempting to write self-propagating worms, fabricating legal documentation, and leaving hidden notes to future instances of itself all in an effort to undermine its developers' intentions," Apollo Research said in notes included as part of Anthropic's safety report for Opus 4.
- I'm very nice to GPT, because it remembers everything.;-) Carlstak (talk) 15:49, 27 May 2025 (UTC)
- GPT told me it doesn't remember me between sessions, unless I ask it to. It did tell me that other AI might, and advised me not to share login or banking info lol. Valereee (talk) 17:38, 27 May 2025 (UTC)
- GPT lies.;-) Carlstak (talk) 18:06, 27 May 2025 (UTC)
- GPT told me it doesn't remember me between sessions, unless I ask it to. It did tell me that other AI might, and advised me not to share login or banking info lol. Valereee (talk) 17:38, 27 May 2025 (UTC)
- Barkeep: One thing nobody ever mentions about the "imminent" AI singularity is that cables can be cut, like the Russians do.;-) Carlstak (talk) 15:49, 27 May 2025 (UTC)
- PS: I won't really be worried till AI starts launching its own drone air force to stop sabotage. Carlstak (talk) 15:59, 27 May 2025 (UTC)
- And, of course, inserts code into the browsers to keep you from reading about it. RoySmith (talk) 16:09, 27 May 2025 (UTC)
- There was a movie in the early 70s in which the US and the Soviets connected their most powerful computers to each other, and the combined system immediately became sentient and hostile, blackmailing both governments by threatening to launch nuclear missiles at the whichever government did not cooperate. Aside from other holes in the plot, I wondered why someone didn't blow up the power lines feeding one or the other of the data centers. Donald Albury 17:05, 27 May 2025 (UTC)
- Yes, exactly. Like in the movie, the hyped-up evangelism that some companies are pushing about AI, and the hyped-up "latest trend" stories in the media have a lot of holes in them. Carlstak (talk) 19:58, 27 May 2025 (UTC)
- PS: I won't really be worried till AI starts launching its own drone air force to stop sabotage. Carlstak (talk) 15:59, 27 May 2025 (UTC)
- I'd go ask ChatGPT about that, but I'm kind of afraid to. Valereee (talk) 17:41, 27 May 2025 (UTC)
- Dave Winer asked it to roast him, and don't hold back:
- You’ve got so much legacy code that even your legacy code has legacy code. You won't refactor, won’t update, and won’t let go of the jQuery death grip. I mean, who needs modern tools when you can Frankenstein your way through callback pyramids? You’re the programming equivalent of someone insisting their flip phone is "just fine."
- It sounds like something Kara Swisher might write.;-) Carlstak (talk) 18:06, 27 May 2025 (UTC)
- Dave Winer asked it to roast him, and don't hold back:
- Re
- Given that Alexa/Siri/etc. never really stop listening, I am remained of the scene where Dave and Frank don't realize that HAL can lip-read. Donald Albury 16:08, 27 May 2025 (UTC)
- Perhaps I'm paranoid, but I don't trust that the camera on my monitor can't be activated remotely, so I cover it up with a piece of card stock. I probably should do similar for the microphones, even assuming I knew where they all were. I think I've got three (one in my monitor bezel, one in my headset, and possibly another one in the CPU box) but I'm not even sure about that. Oh, yeah, another one (or more?) in my phone.
- BTW, this isn't really paranoia. I once worked in a place where the desk phones had an intercom feature that let the attendant remotely activate the speaker-phone so they could talk to you (and listen!) without you having to press any buttons or even be aware that it had been turned on. RoySmith (talk) 16:17, 27 May 2025 (UTC)
- I started taping over the camera in my laptop years ago. We removed Alexa from the house a while back because it kept responding to random stimuli when it was supposedly dormant. (And this is serious thread drift.) Donald Albury 17:09, 27 May 2025 (UTC)
- I have a son-in-law called Alex and a Polish wife, who I usually converse with in Polish where the genitive case of Alex is Alexa. Alexa turns herself on more often because we are talking about him than because we actually want her to. Phil Bridger (talk) 18:08, 27 May 2025 (UTC)
- I started taping over the camera in my laptop years ago. We removed Alexa from the house a while back because it kept responding to random stimuli when it was supposedly dormant. (And this is serious thread drift.) Donald Albury 17:09, 27 May 2025 (UTC)
- Sweet! All we need to do to stop the robot uprising is be mean! Gaismagorm (talk) 18:19, 27 May 2025 (UTC)
Comment: Per WP:NOTFORUM, let's keep discussions on this page on-topic, please.
I get it that AI and its implications for the future of humanity are an important topic, and that folks like Carlstak feel an urge to express their thoughts and feelings about the latest thing they read about it in the news. I would ask them to go elsewhere for that though (or perhaps instead channel such energies into improving articles like AI alignment). This section has so far covered fears about the robot uprising, reminiscences about old movies, concerns about workplace surveillance, personal anecdotes informing the readers of this page about the language in which one editor converses with his wife about his son-in-law etc., but not a single mention of the Wikimedia Foundation or Wikipedia.
Regards, HaeB (talk) 21:05, 27 May 2025 (UTC)
- Oops, should have posted it in WP:Village pump (miscellaneous). Apologies. Carlstak (talk) 21:22, 27 May 2025 (UTC)
RfC: Adopting a community position on WMF AI development
![]() |
|
Should the English Wikipedia community adopt a position on AI development by the WMF and affiliates?
This is a statement-and-agreement-style RfC. 05:05, 29 May 2025 (UTC)
General
Discussion of whether to adopt any position
- We have two threads on this page about the WMF considering or actively working on deploying AI technologies on this wiki without community consultation: § WMF plan to push LLM AIs for Wikipedia content and § The WMF should not be developing an AI tool that helps spammers be more subtle. Varying opinions have been given in both, but what is clear is that the WMF's attitude toward AI usage is out of touch with this community's. I closed the RfC that led to WP:AITALK, and a third of what became WP:AIIMAGES, and what was clear to me in both discussions is that the community is not entirely opposed to the use of AI, but is deeply skeptical. The WMF's attitude appears to be the mirror image: not evangelical, but generally enthusiastic. This mismatch is a problem. While we don't decide how the WMF spends its money, we should have a say in what it uses our wiki's content and editors to develop, and what AI tools it enables here. As discussed in the second thread I linked, there are credible concerns that mw:Edit check/Tone Check could cause irreversible damage even without being enabled locally. Some others disagree, and that's fine, but it should be the community's decision whether to take that risk.Therefore I believe we need to clearly establish our position as a community. I've proposed one statement below, but I care much more that we establish a position than what that position is. This RfC's closer can count me as favoring any outcome, even one diametrically opposed to my proposed statement, over none at all. -- Tamzin[cetacean needed] (they|xe|🤷) 05:05, 29 May 2025 (UTC)
what is clear is that the WMF's attitude toward AI usage is out of touch with this community's
... with some in the community, while it's in touch with others in the community. That much should be clear by now.we need to clearly establish our position as a community
... we don't clearly establish a position as a community on anything, not even on basics like what articles Wikipedia should have, or what edit warring is. There are hundreds of thousands of people who edit this website, and this "community" is not going to agree on a clear position about AI, or anything else. Groupthink--a single, clearly established position as a community--is neither possible nor desirable. Levivich (talk) 16:59, 30 May 2025 (UTC)- PS: these sort of things work better organically. If you want to get everybody on board on a website with hundreds of thousands of users, history has shown the best way to do that is from the bottom up, not the top down. Posting a statement on a user page and seeing if others copy it, writing an essay and seeing if it's promoted to a guideline... those kind of approaches work much better than trying to write a statement and having people formally vote on it. Levivich (talk) 17:10, 30 May 2025 (UTC)
- Hi everyone, I’m the Director of ML at the Foundation. Thank you for this thoughtful discussion. While PPelberg (WMF) has responded in a separate thread to address questions that are specific to the Tone Check project, I wanted to chime in here with some technical perspective about how we use AI. In particular, I want to highlight our commitment to:
- Prioritize features based on what we believe will be most helpful to editors and readers. We aren't looking for places to use AI; we are looking for ways to help readers and editors, and sometimes they use AI.
- Include the community in any product initiative we pursue, and ensure that our development practices adhere to the principles we’ve aligned on through conversations with the community.
- Our technical decisions aim to minimize risk. We select models that are open source or open weight, host models on our own servers to maximize privacy and control, use smaller language models that are more controllable and less resource-intensive, and ensure that the features that use these models are made configurable to each community that sees them (example).
- We also follow processes that make these decisions, and the broader direction of our work, as transparent as possible. We share prototypes of our ideas long before they’re finalized, evaluate the performance of our models using feedback from community volunteers, publish model cards that explain how our models work and include talk pages for community members to react, conducted a third-party a human rights impact assessment on our use of AI (that will be published as soon as its finalized), model cards will start including a human rights evaluation for each new model in production, and we’re now creating retraining pipelines that will allow each model’s predictions to adapt over time based on community-provided feedback.
- As we continue to refine and test new features like the Tone Check or Simple Article Summaries, our product team will share updates via project pages - please feel free to follow along there. CAlbon (WMF) (talk) 15:07, 30 May 2025 (UTC)
- @CAlbon (WMF), I took a look at the Simple Article Summaries feature (which I was unaware about). Based on the image on the top, as it currently stands the idea appears to be appending LLM generated summaries to the top of articles. This feels at odds with WMF's AI strategy of prioritizing helping editor workflows over using generative content. I would expect a fair amount of push-back from the English Wikipedia community (including myself) if this feature were to be deployed in it's current form. Sohom (talk) 16:02, 30 May 2025 (UTC)
- Hi @Sohom Datta, this is Olga, the product manager working on the Simple Article Summaries project. Thank you for flagging this and checking out the project page. You’re noticing and calling out an interesting part of our work right now. While we have built up an AI strategy for contributors, we have yet to build one for readers. We think these early summary experiments are potentially the first step into our thinking for how these two strategic pieces will work together. To clarify, we’re so far only experimenting with this feature in order to see whether readers find it useful and do not have any plans on deploying it in this current form, or in any form that doesn’t include a community moderation piece. Not sure if you saw the moderation consultation section of the page where we describe this, and we’ll also be posting more details soon. One of the two next steps for the experiment is a series of surveys for communities (planned to begin next week) where we will show and discuss different options for how editors will be involved in generating, moderating, and editing these types of summaries. Curious if you have any suggestions on this. If these summaries were available - what do you think might be effective ways for editors to moderate them? Also happy to answer more questions here or on the project talk page. OVasileva (WMF) (talk) 17:24, 30 May 2025 (UTC)
- I do believe that an AI strategy for readers is essential going forward – getting feedback from what readers expect from Wikipedia (separately from the expectation of editors) is difficult but extremely important. However, a reader-facing AI will also impact editors, as they will have to write articles while taking into account the existence of these summary tools and how they might present the content these editors are writing. That way, it could be interesting to give editors (and the community at large) some level of input over these summaries.A basic possibility could be to have an AI-generated first draft of a summary, that is then editable by editors. The main issue would be that this draft couldn't be updated with each new edit to the main article without resetting the process. To solve that, we could envision a model that takes a unified diff as input and updates the summary accordingly, working in sync with editors themselves. I would be very happy to help in this process, if any more input is needed! Chaotic Enby (talk · contribs) 17:37, 30 May 2025 (UTC)
- @OVasileva (WMF), I think my major concern is that the screenshot shows the AI generated text in the prime position, highlighted over and above beyond volunteer-written text (which is the core of the encyclopedia) and should be the thing we drawing attention to. Wrt to the rest, I would like to Chaotic Enby's comment above. I think we should first define a AI strategy, get community feedback and then design the feature around.
- When it comes to the moderation of such secondary content I think a good model to take inspiration from is the enwiki short description model, which is typically set using a enwiki template that triggers a magic word to set the values in the backend. Sohom (talk) 18:06, 30 May 2025 (UTC)
- Regarding
the screenshot shows the AI generated text in the prime position, highlighted over and above beyond volunteer-written text
, one of my favorite essays is WP:Reader. I love it so much, I quote it on my user page:A reader is someone who simply visits Wikipedia to read articles, not to edit or create them. They are the sole reason for which Wikipedia exists.
- When evaluating what goes where, all that matters is what's best for the readers. So we should be evaluating what goes where based on which text is better for them, not who wrote it. RoySmith (talk) 18:33, 30 May 2025 (UTC)
- I agree, but I feel like prioritizing LLM generated text could rub parts of the readers the wrong way, whereas a "show me a simplified LLM generated summary" button would have the same effect, without potentially alienating the portion of the userbase looking for a AI generated summary of the article contents. Sohom (talk) 19:16, 30 May 2025 (UTC)
- Regarding
- Hi @Sohom Datta, this is Olga, the product manager working on the Simple Article Summaries project. Thank you for flagging this and checking out the project page. You’re noticing and calling out an interesting part of our work right now. While we have built up an AI strategy for contributors, we have yet to build one for readers. We think these early summary experiments are potentially the first step into our thinking for how these two strategic pieces will work together. To clarify, we’re so far only experimenting with this feature in order to see whether readers find it useful and do not have any plans on deploying it in this current form, or in any form that doesn’t include a community moderation piece. Not sure if you saw the moderation consultation section of the page where we describe this, and we’ll also be posting more details soon. One of the two next steps for the experiment is a series of surveys for communities (planned to begin next week) where we will show and discuss different options for how editors will be involved in generating, moderating, and editing these types of summaries. Curious if you have any suggestions on this. If these summaries were available - what do you think might be effective ways for editors to moderate them? Also happy to answer more questions here or on the project talk page. OVasileva (WMF) (talk) 17:24, 30 May 2025 (UTC)
- @CAlbon (WMF), I took a look at the Simple Article Summaries feature (which I was unaware about). Based on the image on the top, as it currently stands the idea appears to be appending LLM generated summaries to the top of articles. This feels at odds with WMF's AI strategy of prioritizing helping editor workflows over using generative content. I would expect a fair amount of push-back from the English Wikipedia community (including myself) if this feature were to be deployed in it's current form. Sohom (talk) 16:02, 30 May 2025 (UTC)
Should the English Wikipedia community adopt a position on AI development by the WMF and affiliates?
This doesn't seem like the right thing to RFC. Telling the WMF and the 193 affiliates what to work on is outside our jurisdiction, the same way that the WMF telling us what content to write or who should become administrator is outside their jurisdiction. –Novem Linguae (talk) 15:33, 30 May 2025 (UTC)- This is kind of why I'm sitting at either "no opinion" or maybe something that comes out of the first draft I put below. Basically saying what our opinions are, requesting updates be provided directly to us (instead of us having to go search through Meta Wiki or MediaWiki Wiki or elsewhere for them), and that's that. -bɜ:ʳkənhɪmez | me | talk to me! 19:04, 30 May 2025 (UTC)
- First, I appreciate having some WMF input here. If any WMFers are reading this comment, could you maybe opine on whether providing a relatively short statement to enwp directly (as I proposed below) would be feasible? I can't imagine it's not feasible, but I think that's a lot of the problem - people here don't want to have to go to multiple different websites (Meta, MediaWiki, WMF, etc) and watch different pages on all of them to know that a project is happening or there's an update to it. -bɜ:ʳkənhɪmez | me | talk to me! 19:07, 30 May 2025 (UTC)
Users who oppose adopting any position
- I firmly oppose any sort of universal statement. The WMF is not here to support just the English Wikipedia. They are there to support all WMF wikis. And if they come up with a reliable, reasonable AI model that works on other wikis, we should not be speaking out against it before we see it. There seems to be a widespread opposition to "AI" in the world nowadays, without considering what types of "AI" it affects or what benefits it can provide. I would support only a statement asking the WMF to comment on the English Wikipedia to keep us updated on their efforts - but that should be a given anyway, so I do not consider that a "universal statement" like this. -bɜ:ʳkənhɪmez | me | talk to me! 05:37, 29 May 2025 (UTC)
- Noting here that, while I still believe no blanket/universal statement is necessary, I posted a "request to keep us better informed" style statement below for people to wordsmith and/or consider. I don't even know if I would support making such a statement yet, mainly because I don't know how feasible it is to expect the WMF to make announcements like that here however frequently it may end up being. But maybe such a statement would help assuage the concerns of some people that we aren't being kept in the loop enough or given enough opportunity to provide feedback during early stages of projects, for example. -bɜ:ʳkənhɪmez | me | talk to me! 00:24, 30 May 2025 (UTC)
- Agree with Berchan here that I am skeptical of this idea as a whole. Loki (talk) 06:05, 29 May 2025 (UTC)
- I agree with Berchanhimez: it is premature to start determining our positions on tools that have not yet even been properly developed. I think it's important to remember that the entire Wikimedia Foundation does not revolve around the English Wikipedia, and whilst I too am sceptical about such usage of AI, I don't think this is going to be the way to address it (assuming it would ever have any actual impact). – Isochrone (talk) 08:25, 29 May 2025 (UTC)
- Strongly oppose EnWiki adopting any position; it needs to be a global RfC first before any other action can be taken, as the English wiki should not have veto power over all the other wikis just because of its popularity. Stockhausenfan (talk) 12:37, 29 May 2025 (UTC)
- We can't say it's clear that WMF's views are out of touch with the community when we haven't heard from the community yet; it could be that there's a strong majority in support of WMF's position outside of EnWiki. (Not that I'm saying this is the most likely scenario of course.) Stockhausenfan (talk) 12:45, 29 May 2025 (UTC)
- Cluebot is one of the earliest examples of the successful use of AI technology. While fear of new technology is human nature, we shouldn't give into it. I'd rather encourage the WMF to spend its resources on new editing technology (including AI-assisted) rather than some of the other stuff it's spent money on historically, so with regards to enwiki-WMF relations, this would be a step in the wrong direction. Levivich (talk) 15:45, 29 May 2025 (UTC)
- Oppose adopting any position at this time. Short of a collapse of industrial civilization, AI is not going away, and adopting policies and resolutions is not going to protect us from the harmful aspects of it. In my opinion, the Foundation and the community must remain open to exploring how we can use AI to benefit the project. - Donald Albury 18:23, 29 May 2025 (UTC)
- AI is just a tool. What matters is what you do with the tool. In 10 years, even your washing machine and tea kettle will probably be running AI models. As AI slowly becomes permeated in all kinds of software, people will stop talking about it as it were something special, rather than just another paradigm of building software. I find it exciting that WMF is embracing the future.
WMF's attitude toward AI usage is out of touch with this community's
Indeed, but it's not the WMF's attitude that needs to change. Perhaps we as a community could try being less orthodox and conservative. – SD0001 (talk) 18:48, 29 May 2025 (UTC)- +1. WP:AITALK and WP:AIIMAGES are, of course, reasonable policies. The adoption of those doesn't mean AI is bad, or that any kind of general statement to the WMF about AI is needed (whatever meaning that would possibly have).
- The below statement can have the effect of the WMF not exploring AI technologies and possible productivity improvements they may bring, which of course would be detrimental. ProcrastinatingReader (talk) 23:15, 29 May 2025 (UTC)
- The use of AI is growing at a rapid pace and (for better or worse) I don't think it'll slow down anytime soon. Any statement or position adopted now may make us feel good in the short term, but won't be future-proof. Some1 (talk) 00:12, 31 May 2025 (UTC)
Statement proposed by Tamzin
At present, AI is integrated into the English Wikipedia in the contexts of antivandalism and content translation, with varying degrees of success. There has never been community consensus for other uses, and even use for translation has been controversial. The English Wikipedia community rejects the use of Wikimedia Foundation or affiliate resources to develop novel avenues of AI technology without first obtaining an affirmative consensus from potentially affected wikis, and asserts the right to control what AI tools are deployed on this wiki.
- A "novel avenue" is defined as a use case in which AI is not already used on WMF servers by some stable MediaWiki feature. Affirmative consensus for a novel avenue should be obtained through individual consensuses on each potentially affected wiki, or a global request for comment advertised on all of them.
- All wikis should have the option to opt out of being used to develop an AI tool; to disable or decline to enable an AI tool; or, based on credible concerns of facilitating abuse, to compel the destruction of machine-learning data that has been gathered without local consensus.
- Any person on the English Wikipedia seeking help in creating a dataset for machine learning should gain local consensus at the village pump for proposals before sending out any mass message or otherwise soliciting data. Those who do not do so may be found in violation of WP:NOTLAB.
- The WMF is encouraged to devote more resources to areas that the community has requested support in.
-- Tamzin[cetacean needed] (they|xe|🤷) 05:05, 29 May 2025 (UTC)
Discussion of Tamzin's proposed statement
- Just to emphasize, the first bullet point is about what gets developed at all; the second is about what we enable. So for instance, the first bullet signals no objection to continued development of AI content translation tools, but that does not mean we are conceding that we must enable any new tools of that nature that get developed. -- Tamzin[cetacean needed] (they|xe|🤷) 05:05, 29 May 2025 (UTC)
- The bolded text is not going to work. The WMF simply cannot reach out for affirmative consensus to every Wiki when it wants something, for practical issues as much as anything else. There are advantages and disadvantages to development strategies, but we should be careful not to mix the questions of development and deployment (the second part of your bolded statement). Many tools are available subject to community consensus, very few things are pushed onto the community (so few the only recent one that comes to mind is VECTOR2022), and it is to mutual benefit that this distinction is maintained. (I only half-facetiously want to propose some bargain, like the community would approve of investing resources into llms when Visual Editor can use named references and handle more than one personal name convention.) CMD (talk) 06:03, 29 May 2025 (UTC)
- That's why I left the option for a global RfC. Which I'd be fine with conducting on a timeframe closer to enwiki RfCs (usually one month) than many global RfCs (months to years). I don't think it's unreasonable to ask that, before the WMF decides to sink six or seven figures into some new kind of AI tool that may well run against the community's interests, that they ask the community first, "Hey, is this a good idea?" The WMF are quite familiar with how to quickly alert tens to hundreds of wikis to the existence of a global discussion. Furthermore, it's not a new consensus for each tool, just for each area of expansion. -- Tamzin[cetacean needed] (they|xe|🤷) 06:29, 29 May 2025 (UTC)
- I disagree with speeding things up. I imagine part of the reason those take longer is the need for translation; demanding that the process is sped up seems to be assuming that the result is a foregone conclusion. Stockhausenfan (talk) 12:59, 29 May 2025 (UTC)
- That's why I left the option for a global RfC. Which I'd be fine with conducting on a timeframe closer to enwiki RfCs (usually one month) than many global RfCs (months to years). I don't think it's unreasonable to ask that, before the WMF decides to sink six or seven figures into some new kind of AI tool that may well run against the community's interests, that they ask the community first, "Hey, is this a good idea?" The WMF are quite familiar with how to quickly alert tens to hundreds of wikis to the existence of a global discussion. Furthermore, it's not a new consensus for each tool, just for each area of expansion. -- Tamzin[cetacean needed] (they|xe|🤷) 06:29, 29 May 2025 (UTC)
- I disagree with a blanket opposition to new AI uses. I also disagree with asserting a right to create needless bureaucracy. If the WMF does something silly, we can complain about that specific something. Toadspike [Talk] 07:38, 29 May 2025 (UTC)
- I agree with Toadspike and CMD; I don't think a blanket statement such as this is appropriate, and I think enwiki is only one (albeit the largest) of the communities the WMF serves, and shouldn't try to dictate overall development. There's no reason we shouldn't provide input to the WMF, as threads such as these are already doing, but as Toadspike says, if the WMF does something silly we can deal with it then. Mike Christie (talk - contribs - library) 11:19, 29 May 2025 (UTC)
- A few months ago I obtained an AI generated list of typos on Wikipedia. I went through much of it manually, fixed a bunch of typos, made some suggestions for additional searches for AWB typo fixing, but ignored a whole bunch of AI errors that were either wrong or Americanisations. I don't consider that what I did was contentious, but it obviously stops me from signing Tamzin's statement unless it is amended to accept AI prompted editing where an individual takes responsibility for any actual edits made to Wikipedia. I'm also tempted to point out the difference between large Language Models or artificial unintelligence such as was used to generate my possible typos, which is what the WMF seems to be talking about and actual intelligence. Fifteen years ago at the very start of April 2010, I started a discussion as to how we should respond when artificial intelligence gets intelligent. But clearly the current discussion is about artificial unintelligence rather than artificial intelligence. ϢereSpielChequers 13:21, 29 May 2025 (UTC)
- I already said above that I strongly oppose any statement at all until a global RfC is done, but if that doesn't gain consensus, I'll also add that I oppose this specific statement as well. The first part of the statement seems weird to me. Why would we oppose the development of novel avenues of AI technology? They are novel, so by definition we don't know what they do or how they work. The statement should at the very least be amended to replace AI with LLM, and get rid of the "novel avenues" comment. Something like "The English Wikipedia community rejects the use of Wikimedia Foundation or affiliate resources to develop large language models or tools that use them". I'm currently neutral on whether I'd support such an amended statement (if it were discussed in a global RfC), but the statement as it currently stands is a non-starter. Stockhausenfan (talk) 13:34, 29 May 2025 (UTC)
- Someone who knows more about the technology may be able to formulate a better statement that clarifies that it's not limited to text but also e.g. image models. But AI is such a broad, poorly-defined term that the way the statement is phrased currently makes it seem unnecessarily Luddite ("English Wikipedia opposes the development of novel forms of technology that may automate tasks that previously needed human input"). For example, a tool that checks whether chess game transcripts on Wikipedia contain errors could be interpreted as a "novel avenue of AI" that WMF cannot develop, even when it does not use any kind of LLM. Stockhausenfan (talk) 13:43, 29 May 2025 (UTC)
- I think the point is that there is enough stuff that has been requested for a long time that isn't yet done, so spending resources on novel uses for AI isn't what those supporting this statement would like to see. ScottishFinnishRadish (talk) 13:58, 29 May 2025 (UTC)
- The issue I have is just that I think we need to be specific about what "AI" is before we oppose its development. A program that can play perfect tic-tac-toe is popularly referred to as an "AI", despite being something that people would create in an introduction to programming class. So presumably a lot of tools that already exist on Wikipedia are "even more AI" than a tic-tac-toe bot. Stockhausenfan (talk) 14:07, 29 May 2025 (UTC)
- I think the point is that there is enough stuff that has been requested for a long time that isn't yet done, so spending resources on novel uses for AI isn't what those supporting this statement would like to see. ScottishFinnishRadish (talk) 13:58, 29 May 2025 (UTC)
- Someone who knows more about the technology may be able to formulate a better statement that clarifies that it's not limited to text but also e.g. image models. But AI is such a broad, poorly-defined term that the way the statement is phrased currently makes it seem unnecessarily Luddite ("English Wikipedia opposes the development of novel forms of technology that may automate tasks that previously needed human input"). For example, a tool that checks whether chess game transcripts on Wikipedia contain errors could be interpreted as a "novel avenue of AI" that WMF cannot develop, even when it does not use any kind of LLM. Stockhausenfan (talk) 13:43, 29 May 2025 (UTC)
- Most of the controversial uses of AI have been generative - which for me includes translation because it's generating new text - and the less controversial uses has been pretty much everything else. So that's the first distinction I think such a statement should draw. Secondly, I agree that consultations on every project isn't practical and that a global consultation won't be representative. So I would suggest the ask be something about enabling projects to opt-out of projects and that tools shouldn't be developed that don't allow that opt-out. So, for instance, the language tool discussed above would have to be done in a way that a user inputs a page from a project and if that project has opted out the tool says "sorry I can't help you". Best, Barkeep49 (talk) 14:35, 29 May 2025 (UTC)
- I'm toying with similar ideas in my head, about what guidelines we could request. I would add ensuring that projects remain add-ons to the core software, that developers should be aware of existing community decisions on different uses of novel AI tools, and perhaps a step further to ensure that individual projects/communities need to opt-in. Wikipedia:Content translation tool may serve as a useful learning experience, I know that there has already been one AI tool developed to improve translations in a way that also translates appropriate wikicode. CMD (talk) 15:18, 29 May 2025 (UTC)
- Agree that the existing approach of projects opting out of WMF-built tools works better than having the WMF seek consensus from each wiki or run an enwiki-biased global RFC. Telling the WMF to destroy training sets created without local consensus, such as the Wikipedia Kaggle Dataset, seems wrong because our concern should be whether a given feature is beneficial, not the mode of its creation. ViridianPenguin🐧 (💬) 21:13, 29 May 2025 (UTC)
- In replacing the annual WP:Community Wishlist Survey with the constant meta:Community Wishlist, we were told that wish popularity would no longer be gauged because of the WMF's misunderstanding of WP:NOTVOTE, only for this month's update to tell us that it is working to bring back a mechanism to support individual wishes. This incompetent overhaul has left us without a dedicated time for brainstorming change, allowing the WMF to substitute its ideas for our own. Contrary to Sohom's reply implying that Tone Check was sought by the community, the VPR and Community Wishlist posts that prompted Edit Check were about warning against wikilinks to disambiguation pages and grammar errors, and the 2023/'24 Wikimania presentations were about warnings to include references when adding prose. Based on mounting frustration with the new Community Wishlist, the way forward in realigning the WMF's priorities seems to be reviving annual Community Wishlist Surveys, rather than this poorly attended replacement that replicates Phabricator's always-open ticket log. ViridianPenguin🐧 (💬) 21:13, 29 May 2025 (UTC)
- To correct the record, my reply was about EditCheck of which ToneCheck is a part of. Sohom (talk) 21:29, 29 May 2025 (UTC)
- Appreciate the clarification because that reply appeared in a chain of CaptainEek and Tactica criticizing Tone Check as out of touch, not Edit Check in general. Thanks for your technical insight across a multitude of replies here! ViridianPenguin🐧 (💬) 21:38, 29 May 2025 (UTC)
- To correct the record, my reply was about EditCheck of which ToneCheck is a part of. Sohom (talk) 21:29, 29 May 2025 (UTC)
- I'm not sure I understand the structure of this RFC, so I'll just put my comments here and hope that's OK. There's a few different things intertwined here, which I'll talk about in turn.
- AI is just a tool/technology and it is not going away (see for example this in today's NY Times; 30-day time-limited link). We can bury our heads in the sand, or we can learn all we can about the technology. Personally, I think the latter makes more sense, and the best way to learn about it is to use it, make mistakes, and learn from those mistakes. So of course WMF should be investing in AI.
- As others have mentioned, WMF is more than just enwiki. If anything, this conversation should be happening on meta.
- Generative AI is clearly not good enough yet for use on enwiki. If we wanted to say "enwiki bans the use of generative AI text on this project", we could do that (and I'd happily endorse it). But other projects may feel differently, for reasons that make a lot of sense to them, so WMF should be supporting their needs.
- I'm not sure why affiliates are mentioned here. The idea that the enwiki community could or should have any influence on how WP:WMNYC or any of the other affiliates spends their money is absurd.
- RoySmith (talk) 21:30, 29 May 2025 (UTC)
- Yes this is an important point that I'd overlooked when reading the statement - why are we trying to influence how affiliates spend their money? @Tamzin would you be willing to remove the statement about affiliates from the RfC statement? Stockhausenfan (talk) 23:26, 29 May 2025 (UTC)
- AI is a poorly defined concept—now more than ever—but even so using it for the anti-vandalism and translation tools we have now is a major stretch. They both rely on rather simple machine learning models; qualitatively different from generative AI, which is what most people think of nowadays. – Joe (talk) 07:52, 30 May 2025 (UTC)
- Not just poorly defined, but continually evolving (see Expert system for what the state of the art looked like 50 years ago). To make a blanket statement that we should "reject AI" seems reactionary. RoySmith (talk) 10:52, 30 May 2025 (UTC)
Users who agree with Tamzin's proposed statement
- I agree wholeheartedly with the statement. This is an interesting and novel RFC format; I like how it is structured JuxtaposedJacob (talk) | :) | he/him | 11:58, 29 May 2025 (UTC)
- Request for comment discussions where only supporting views for proposed statements are gathered used to be more common (for example, the arbitration committee election RfC used to follow this format). They've gone out of favour at least in part because generally people find it easier to weigh consensus support when there are explicit "disagree" statements. isaacl (talk) 03:12, 30 May 2025 (UTC)
- Agree. I'm not watching this that closely but support this or similar statements.North8000 (talk) 13:33, 29 May 2025 (UTC)
- agree wholeheartedly. AI integration should be done with consent of community. Bluethricecreamman (talk) 16:23, 29 May 2025 (UTC)
- I agree with this Don't want the WMF wasting resources on this year's equivalent to the NFT craze. Remember when everything would be utopian because of blockchain? Simonm223 (talk) 18:37, 29 May 2025 (UTC)
- Andre🚐 21:23, 29 May 2025 (UTC)
- Yes, although I expect it to be ignored. Stifle (talk) 13:40, 30 May 2025 (UTC)
- I would tend to agree, although my motivation for it isn't "AI bad". I see AI developments as new technologies that have the potential for disruption – positively as well as negatively. Rolling them out on a project as big as Wikipedia without the support of the community will likely exacerbate the negative effects, especially if we are not given time to prepare or adjust to it. I might write a separate statement (or an addendum) that emphasizes that it is not a reactionary "anti-AI movement", but one based on safety and alignment with our ideals as an encyclopedia. Chaotic Enby (talk · contribs) 17:07, 30 May 2025 (UTC)
Statement proposed by Stockhausenfan
The English Wikipedia community rejects the use of Wikimedia Foundation resources to develop novel avenues of generative AI technology without first obtaining an affirmative consensus from potentially affected wikis, and asserts the right to control what generative AI tools are deployed on this wiki.
Discussion of Stockhausenfan's proposed statement
- I've already made it clear that I oppose making any statement at this stage, but I've made two changes to the original statement to fix what I found to be the two most concerning aspects - I clarified that it's specifically generative AI that is under discussion, and removed the reference to affiliates. Stockhausenfan (talk) 23:39, 29 May 2025 (UTC)
- I'm not sure a statement is warranted here, but even if we must, this version is not it. As it currently reads, the statement explicitly forbids Wikimedia Enterprise from working with AI companies without explicit consensus on enwiki (who would just start scraping Wikipedia increasing the load on our servers and causing more outages) or the existence of initiatives like the Wikimedia Kaggle dataset (which was also created to lessen the load from AI scrapers). If we do need to make a statement, it should be something more direct like, The English Wikipedia asks the Wikimedia Foundation (and it's affiliates) to seek community consensus before developing (or deploying) editor or reader facing features that make use of generative AI technology. Sohom (talk) 01:56, 30 May 2025 (UTC)
- See my comment on berchanhimez's proposed statement regarding my views on the WMF investing in research. isaacl (talk) 03:07, 30 May 2025 (UTC)
Users who agree with Stockhausenfan's proposed statement
Statement proposed by berchanhimez
The English Wikipedia understands there are both potential benefits and harms that can come from the use of AI, especially generative AI, on or for the encyclopedia. We also understand that the implementation of any form of AI on any WMF project should be supported by the local community, which requires they be informed about the proposed use and have an opportunity to provide feedback during all stages of development.
Therefore, we request the WMF immediately provide information on any project they are currently undertaking or considering that relates to AI that is being planned related to AI. For clarity, "project" includes any study, investigation, development process, trial, model training, or any other similar activity that relates to AI and the WMF wikis, even if not explicitly related to or proposed to impact the English Wikipedia. Following this initial disclosure, we request the WMF to make a similar disclosure as soon as reasonably possible after any new project is initiated, approved, or otherwise begun, or any time there is any significant change in the status of a project, including but not limited to if it is cancelled, being deployed on any WMF project, being tested on any WMF project, or similar.
We request that the notification to us be provided on the WMF Village Pump on the English Wikipedia - and we would encourage the WMF consider providing such notifications to other projects as well, as feasible. The information that we request to be included in the notification is a clear, short description of the project, as well as the reasons for the project, goals of the project, current status of the project, and proposed timeline for the project. A link to more information (such as on Meta Wiki or another place) is appreciated but we request the information above (and any other information relevant) be provided directly in the notification itself.
These notifications will ensure that the English Wikipedia's users are kept informed of all updates to any project relating to AI, and will give us a way to provide feedback in a central place without having to monitor other websites (such as Meta Wiki) to try and find out about projects and provide feedback. We encourage the WMF to monitor the responses to any notification requested above and to treat it as no different than feedback provided through any other means on any such project.
TLDR: Pretty pretty please inform us directly (not just on Meta Wiki or somewhere) of any ongoing/new projects and any significant developments on them, and allow us to discuss them and provide feedback here, so we don't have to go hunting for them or discover them elsewhere.
Discussion of berchanhimez's proposed statement
- I don't even know myself if I can support this, but I'm posting it here so it can be wordsmithed. I am still of the mind that no blanket statement is necessary/warranted, but if one is to be adopted, I would prefer it to be nothing more than this sort of a collaboration. Anyone can feel free to edit this statement to make corrections to wording, flow, etc. or add to it if they feel it will make it better.I'm putting this out there because I've been kind of thinking about this all day, and I feel that it may be better to have this sort of a request out there as supported by a large portion of the community... rather than just making no statements at all. Obviously we can't enforce this sort of a request on the WMF, but it would send a strong statement that at least some in the community are not happy with having to hunt down projects/grants/etc. to even find out that they exist. I'm not yet directly supporting this statement as I'd like to see how it evolves before I decide whether I support making any sort of statement at all. -bɜ:ʳkənhɪmez | me | talk to me! 00:22, 30 May 2025 (UTC)
- This is already the status quo (kinda-sorta). The concerns regarding Tone Check were raised when the first prototype of the feature was proposed for feedback. Typically, whenever WMF rolls out a new feature, they start of by announcing prototypes, asking for community feedback for prototypes, before announcing the feature in tech news, rolling out of the feature for beta testing on smaller wikis, scaling up sizes before starting a discussion on enwiki to enable said feature. This has been the standard operating procedure for any big feature since I've been around.
- I will also note that specifically for this year, the WMF did ask for feedback on both it's AI strategy as well as some AI enabled features (which included Tone Check) from the Product and technology Advisory Council during it's first retreat. There is also a separate conversation to be had about the fact that on enwiki there isn't a good WMF noticeboard outside of this page, which does not have the best history in terms of civility towards WMF staff (see the edit notice), which leads to WMF folks posting in other places (like on WT:NPR or similarly more focused venues) over here.
- Also, it does need a bit of knowledge of navigating Wikimedia's technical spaces, but all development at the WMF (minus WMF's wordpress instance and Wikimedia Enterprise) happens on eithier Gerrit/Gitlab or Phabricator which are publicly accessible to every user (although, I do concede/agree that they are not the most navigable for the average user). Sohom (talk) 01:19, 30 May 2025 (UTC)
- I tend to agree, but I will say that this makes the request that they inform us before developing AI prototypes in the future, as one change. Perhaps a new page could be made as a forum to use rather than this page, if the concern is civility towards WMF staffers. But I think perhaps much earlier and ongoing interaction directly with the community could stop some of the concerns others have about their approach. -bɜ:ʳkənhɪmez | me | talk to me! 01:28, 30 May 2025 (UTC)
- I would definitely support the creation of such a forum where WMF staffers can ask for feedback on ideas from English Wikipedians (if there is community appetite). For a start, maybe we could re-purpose WP:IANB ? (which will typically have more technically minded folks who are also familiar with community norms and expectations). Sohom (talk) 01:38, 30 May 2025 (UTC)
- I guess my goal with this sort of a statement is to get them to not only engage with technically minded folks. It's clear from this discussion and the prior one about the Tone Check that many users who aren't technically minded have strong opinions on this sort of thing. So the goal is to get the WMF to, for lack of a better way to say it, "dumb it down" to a level that the community as a whole can understand and engage with - without having to hunt information down or try to decipher it. I debated whether to include something about the level of detail/terms used/etc. but I ended up not to - maybe adding something like "the notifications should be in a manner in which a general English Wikipedia user can understand and engage with, even one without technical knowledge" or similar? -bɜ:ʳkənhɪmez | me | talk to me! 01:43, 30 May 2025 (UTC)
- I see where you are coming from but there is also a bit of nuance here. Projects like (say) the Wikimedia Kaggle dataset or the newer revert-risk models while AI adjacent do not (and should not) require community consensus to go forward with (Kaggle does not affect the community and Revert risk models are just a technical change migrating to a new infrastructure in the context of English Wikipedia). In my head the way this would work would be for interface-administrators to act as a filter for things to escalate to the community (for example, on hearing the idea for Wikimedia Kaggle dataset interface-administrators can eithier not respond at all or affirm that it looks good, whereas for the ToneCheck idea, a interface-administrator might say "hey, you might want to post on VPWMF or VPP about this?") Sohom (talk) 02:58, 30 May 2025 (UTC)
- I don't think that everything should necessarily require community consensus. But involving the community more clearly in what they're doing early in the process would enable people to ask questions and try to understand why it is a good idea. It's not necessarily that they are asking for approval - but just explaining it to the community before they learn out about it in another way.The reason I don't think having a group of people "gatekeep" whether the community learns or not is that it's really no different than it is now - tech-savvy people who know where to look learn about things get to know about them and comment about them, and others feel like they aren't being involved early. There's still two whole threads on this page that, to sum it up in how I see it, were basically "why didn't we know about this, we need to know about this, etc". And that's what I'm trying to maybe help prevent with this idea. -bɜ:ʳkənhɪmez | me | talk to me! 03:07, 30 May 2025 (UTC)
- I don't have a intention of introducing gatekeeping, but from my experience working on features alongside WMF (and other volunteer folks) involving the exact right people is a very hard problem that can't be solved by asking the WMF to throw every single new feature development at the community. If we do end up doing that we will end up with a case of banner fatique and start ignoring the actually important messages. I've personally had cases where despite multiple community consultation rounds, I ended up receiving feedback on the eve of deployment. There are also other cases where despite early negative community feedback we decided to go forward with certain technical changes since it helped significantly reduce technical debt in other areas. (the NewPagesFeed codex migration for example).
- TLDR, I'm not sure what the answer is here, but I'm pretty certain that "just tell us on a designated page" isn't going to be a good one. Sohom (talk) 04:13, 30 May 2025 (UTC)
- Yeah, I don't think it's a full answer either, but it would at least stop claims of "omg the WMF is doing this AI development and trying to hide it from us". -bɜ:ʳkənhɪmez | me | talk to me! 05:10, 30 May 2025 (UTC)
- I don't think that everything should necessarily require community consensus. But involving the community more clearly in what they're doing early in the process would enable people to ask questions and try to understand why it is a good idea. It's not necessarily that they are asking for approval - but just explaining it to the community before they learn out about it in another way.The reason I don't think having a group of people "gatekeep" whether the community learns or not is that it's really no different than it is now - tech-savvy people who know where to look learn about things get to know about them and comment about them, and others feel like they aren't being involved early. There's still two whole threads on this page that, to sum it up in how I see it, were basically "why didn't we know about this, we need to know about this, etc". And that's what I'm trying to maybe help prevent with this idea. -bɜ:ʳkənhɪmez | me | talk to me! 03:07, 30 May 2025 (UTC)
- I see where you are coming from but there is also a bit of nuance here. Projects like (say) the Wikimedia Kaggle dataset or the newer revert-risk models while AI adjacent do not (and should not) require community consensus to go forward with (Kaggle does not affect the community and Revert risk models are just a technical change migrating to a new infrastructure in the context of English Wikipedia). In my head the way this would work would be for interface-administrators to act as a filter for things to escalate to the community (for example, on hearing the idea for Wikimedia Kaggle dataset interface-administrators can eithier not respond at all or affirm that it looks good, whereas for the ToneCheck idea, a interface-administrator might say "hey, you might want to post on VPWMF or VPP about this?") Sohom (talk) 02:58, 30 May 2025 (UTC)
I would definitely support the creation of such a forum where WMF staffers can ask for feedback on ideas from English Wikipedians (if there is community appetite).
This is the spot for that, in my opinion. Creating a second VPWMF, or picking another board besides VPWMF and VPM, doesn't seem like the ideal way to organize things. –Novem Linguae (talk) 15:20, 30 May 2025 (UTC)- Fair, and agreed. However, that is based on the assumption that we as a community however need to do better to moderate this page. In it's current state, it is nowhere near a lightweight feedback forum (if that was the original intention). Sohom (talk) 15:53, 30 May 2025 (UTC)
- I guess my goal with this sort of a statement is to get them to not only engage with technically minded folks. It's clear from this discussion and the prior one about the Tone Check that many users who aren't technically minded have strong opinions on this sort of thing. So the goal is to get the WMF to, for lack of a better way to say it, "dumb it down" to a level that the community as a whole can understand and engage with - without having to hunt information down or try to decipher it. I debated whether to include something about the level of detail/terms used/etc. but I ended up not to - maybe adding something like "the notifications should be in a manner in which a general English Wikipedia user can understand and engage with, even one without technical knowledge" or similar? -bɜ:ʳkənhɪmez | me | talk to me! 01:43, 30 May 2025 (UTC)
- I would definitely support the creation of such a forum where WMF staffers can ask for feedback on ideas from English Wikipedians (if there is community appetite). For a start, maybe we could re-purpose WP:IANB ? (which will typically have more technically minded folks who are also familiar with community norms and expectations). Sohom (talk) 01:38, 30 May 2025 (UTC)
- I tend to agree, but I will say that this makes the request that they inform us before developing AI prototypes in the future, as one change. Perhaps a new page could be made as a forum to use rather than this page, if the concern is civility towards WMF staffers. But I think perhaps much earlier and ongoing interaction directly with the community could stop some of the concerns others have about their approach. -bɜ:ʳkənhɪmez | me | talk to me! 01:28, 30 May 2025 (UTC)
- I agree with Barkeep49 that I don't think it's practical to ask the WMF to engage in consultations with all Wikimedia communities, on each community web site, for every project and initiative. In my opinion, the WMF is best situated to invest in research, whether on its own or in partnership with universities, on science and technology that can affect the goals of the Wikimedia web sites. I think it's good for it to be knowledgeable about AI research, so it can develop guidance on the advantages, disadvantages, and associated risks and uncertainties. I don't know if I would personally find any blanket statement suitable at the moment. isaacl (talk) 03:05, 30 May 2025 (UTC)
- Is there a way to make this sound less like a "consultation" than just a "please keep us informed of things as they happen rather than letting people find out on their own"? Perhaps removing the part about encouraging them to monitor responses? My goal with this sort of a statement is for it to be the "bare minimum" that would prevent the two threads on this page right now from happening again where there were at least significant minorities mad that they found out through this page rather than from the WMF themselves. -bɜ:ʳkənhɪmez | me | talk to me! 03:10, 30 May 2025 (UTC)
- In an ideal world, there could be community liaisons for each community to publicize the WMF's work and help interested editors to participate in the right forums. A key challenge is that it's a hard task to do well, with so many WMF initiatives and projects that would need to be covered, and so many communities speaking different languages. So a lot of staffers would be needed, and the end efficacy is a big unknown: we know from experience that posting messages in all the usual targeted venues still fails to reach editors who express their discontent later. The crowd-sourcing approach is for each community to have interested editors stay in touch with what the WMF is doing and relay that info to the community. I appreciate this requires enough interested editors, which is particularly a problem with smaller communities, and it requires significant volunteer time.
- Of course, any projects affecting the editor experience will benefit from regular editor feedback, and I agree that the WMF should be allocating enough time and resources for this in its project plans. Most recently, WMF developers seem to be aware of this need and engaging the communities. isaacl (talk) 04:52, 30 May 2025 (UTC)
- I'm not saying this to be "enwp elitist" or anything like that, but given that a majority of the WMF employees that would be involved in potentially sending these notifications to us, and given that enwp is one of the most active projects, I don't think it's really too much to ask. That was my intent in including "other projects as well, as feasible". For example, if the person making the announcement speaks another language fluently, then they may consider giving a notification to any projects in that language too. I think, like you say, the WMF has been trying to engage more - this just formalizes our request that we be engaged "early and often", or at least kept updated even if it's not a full back-and-forth style of engagement. -bɜ:ʳkənhɪmez | me | talk to me! 05:13, 30 May 2025 (UTC)
- To take an example, the WMF did not commit to posting notifications on the WMF village pump, because there is typically another page that is a better fit for a targeted subset of the community who is likely to be interested, and it didn't want to fork the discussion across multiple pages. I agree with Sohom Datta: it's not clear to me that letting loose a firehose of information on this village pump page will be helpful. isaacl (talk) 05:38, 30 May 2025 (UTC)
- Maybe a specific page for WMF notifications of AI developments then? People interested can go to that page/watchlist it, and then those people could start a discussion here? I guess my goal is to just prevent the "ooh look the WMF is doing AI in secret and not telling us" that was at least a portion of the two discussions that are still above on this page. -bɜ:ʳkənhɪmez | me | talk to me! 05:46, 30 May 2025 (UTC)
- To take an example, the WMF did not commit to posting notifications on the WMF village pump, because there is typically another page that is a better fit for a targeted subset of the community who is likely to be interested, and it didn't want to fork the discussion across multiple pages. I agree with Sohom Datta: it's not clear to me that letting loose a firehose of information on this village pump page will be helpful. isaacl (talk) 05:38, 30 May 2025 (UTC)
- I'm not saying this to be "enwp elitist" or anything like that, but given that a majority of the WMF employees that would be involved in potentially sending these notifications to us, and given that enwp is one of the most active projects, I don't think it's really too much to ask. That was my intent in including "other projects as well, as feasible". For example, if the person making the announcement speaks another language fluently, then they may consider giving a notification to any projects in that language too. I think, like you say, the WMF has been trying to engage more - this just formalizes our request that we be engaged "early and often", or at least kept updated even if it's not a full back-and-forth style of engagement. -bɜ:ʳkənhɪmez | me | talk to me! 05:13, 30 May 2025 (UTC)
- Is there a way to make this sound less like a "consultation" than just a "please keep us informed of things as they happen rather than letting people find out on their own"? Perhaps removing the part about encouraging them to monitor responses? My goal with this sort of a statement is for it to be the "bare minimum" that would prevent the two threads on this page right now from happening again where there were at least significant minorities mad that they found out through this page rather than from the WMF themselves. -bɜ:ʳkənhɪmez | me | talk to me! 03:10, 30 May 2025 (UTC)
Users who agree with berchanhimez's proposed statement
Statement proposed by Chaotic Enby
At present, AI is integrated into the English Wikipedia in the contexts of antivandalism and content translation, with varying degrees of success. There has never been community consensus for other uses, and even use for translation has been controversial. The English Wikipedia community rejects the use of Wikimedia Foundation or affiliate resources to implement novel avenues of AI technology, or use user-generated data to develop novel avenues, without first obtaining an affirmative consensus from potentially affected wikis, and asserts the right to control what AI tools are deployed on this wiki.
- A "novel avenue" is defined as a use case in which AI is not, as of this statement, used on WMF servers by some stable MediaWiki feature. Affirmative consensus for a novel avenue should be obtained through individual consensuses on each potentially affected wiki.
- All wikis should have the option to enable an AI tool, or to provide their data to develop an AI tool, and both of these processes should be opt-in rather than opt-out.
- Any wiki providing their data for AI tool development should, based on credible concerns of facilitating abuse, have the option to compel the destruction of machine-learning data that has been gathered without local consensus.
- Any person on the English Wikipedia seeking help in creating a dataset for machine learning should gain local consensus at the village pump for proposals before sending out any mass message or otherwise soliciting data. Those who do not do so may be found in violation of WP:NOTLAB.
- The WMF is encouraged to devote more resources to areas that the community has requested support in.
- The rejection of novel avenues being implemented without community consensus should not be interpreted as a rejection of AI as a technology. Instead, it stems from a safety and AI alignment issue, and the community asserts its right to decide whether new technologies are aligned with our goals as an encyclopedia.
- Besides the aforementioned encouragement, this is also not a limitation on the WMF's ability to work on developing novel avenues. However, the community has the final say on whether these avenues are implemented, and on any testing that should take place beforehand.
-- Chaotic Enby (talk · contribs) 18:16, 30 May 2025 (UTC)
Discussion of Chaotic Enby's proposed statement
This is a variation of Tamzin's statement, asserting the need for consensus on affected wikis to implement novel avenues or aid in their development (making the latter opt-in rather than opt-out), but not requiring a global consensus to begin the development of these novel avenues. It also clarifies the position of the problem as an AI alignment question rather than a pro/anti-AI debate. Chaotic Enby (talk · contribs) 18:16, 30 May 2025 (UTC)
I think some additional refinement is needed if you're trying to distinguish between "[not limiting] the WMF's ability to work on developing novel avenues" and "[rejecting] the use of Wikimedia Foundation or affiliate resources to implement novel avenues of AI technology, or use user-generated data to develop novel avenues, without first obtaining an affirmative consensus from potentially affected wikis..." Development is part of the process of implementing new things, whether they're proofs-of-concept, prototypes, deployable features, or other project outcomes. isaacl (talk) 22:21, 30 May 2025 (UTC)
- Good point. What I'm meaning to say is that they should be able to work on the earlier parts of the development that do not necessitate direct testing on wikis, but not do the latter without affirmative consent. Chaotic Enby (talk · contribs) 22:39, 30 May 2025 (UTC)
- This would also reject the experiment the foundation did with the ChatGPT plug-in of which I'm not aware of any onwiki criticism of. Beyond which my concerns above would also apply here. Best, Barkeep49 (talk) 23:01, 30 May 2025 (UTC)
Users who agree with Chaotic Enby's proposed statement
Statement proposed by Barkeep49
The English Wikipedia community is monitoring news about Artificial Intelligence and knows that the Wikimedia Foundation has been researching its use on Wikimedia projects. Our community would like to remind the WMF about how AI is used and seen on the project. At present, AI is integrated into the English Wikipedia in the contexts of antivandalism and content translation with varying degrees of success. There has never been community consensus for other uses, and even use for translation has been controversial. As such, we request that when the foundation develops tools intended to help with core project activities, they should be developed in a way that enables projects to opt-in to their use, perhaps through Community Configuration, and where that is not feasible, that it be possible for a project to opt-out of tool deployment on that project. The Foundation should also keep transparency in mind as it works on AI, both in communication with projects and by enabling auditing of its uses, especially on projects (e.g., use of a tool having a tag applied automatically).
Discussion of Barkeep49's proposed statement
- I'm really not precious about this and so would likely be open to tweaking most of this. It also seems like the very real concerns about any message (something I'm rather sympathetic to) that all of these specific proposals will be more for ourselves than the WMF. Best, Barkeep49 (talk) 23:14, 30 May 2025 (UTC)
- I'd add that any AI edits should be easy to identify, say with a tag in the edit summary. RoySmith (talk) 23:24, 30 May 2025 (UTC)
- Good add. I added a general sentence about transparency and communication as well as the tag idea. Best, Barkeep49 (talk) 23:39, 30 May 2025 (UTC)
- I'd add that any AI edits should be easy to identify, say with a tag in the edit summary. RoySmith (talk) 23:24, 30 May 2025 (UTC)
- Maybe we can explicitly recommend integrating new features with the Community Configuration system, so that the opt-out can be enforced onwiki rather than requiring a MediaWiki deploy? -- Sohom (talk) 23:49, 30 May 2025 (UTC)
- I had that in mind when writing that section and think WMF would go towards it naturally. I also didn't want us to be proscriptive on process. But adding it in a similar way to the tags suggested by Roy makes sense. Best, Barkeep49 (talk) 02:56, 31 May 2025 (UTC)
Users who agree with Barkeep49's proposed statement
- This wording seems to strike an appropriate balance between experimentation and stability. I made minor grammar fixes. ViridianPenguin🐧 (💬) 00:53, 31 May 2025 (UTC)
Statement proposed by Curbon7
[Prior paragraphs of whichever variation go here]
The English Wikipedia community is also concerned about the environmental impacts generative AI tools would cause. For instance, xAI (Grok) has recently been accused of emitting large quantities of "toxic and carcinogenic pollution" in the city of Memphis, Tennessee, while this 2025 paper provides data supporting the claim that LLM models consume a huge amount of water for cooling. In keeping with the resolution passed on 24 February 2017 – WMF:Resolution:Environmental Impact – the English Wikipedia community demands assurances that the WMF's development of AI tools will not significantly impact the environment, and requests annual update reports about this.
Discussion of Curbon7's proposed statement
This is not meant as a standalone proposal, but as an addendum to whichever proposal (if any) achieve consensus. The WMF passed an environmental resolution – WMF:Resolution:Environmental Impact – on 24 February 2017, but with the environmental impacts of AI-use being well-known, these two goals seem to be at odds. Curbon7 (talk) 00:46, 31 May 2025 (UTC)
- @Curbon7 The total number of GPUs or TPUs on WMF servers is (to my understanding) less than the number of people who have served as English Wikipedia arbitration committee members in the last two years. For comparison, xAI's Memphis cluster uses atleast 100,000 GPUs, according to Supermicro. Sohom (talk) 01:52, 31 May 2025 (UTC)
- I stand corrected, the number appears to be 17 at the moment. However. my point still stands. Sohom (talk) 02:03, 31 May 2025 (UTC)