wikipedia

ted-cruz-doesn’t-seem-to-understand-wikipedia,-lawyer-for-wikimedia-says

Ted Cruz doesn’t seem to understand Wikipedia, lawyer for Wikimedia says


A Wikipedia primer for Ted Cruz

Wikipedia host’s lawyer wants to help Ted Cruz understand how the platform works.

Senator Ted Cruz (R-Texas) uses his phone during a joint meeting of Congress on May 17, 2022. Credit: Getty Images | Bloomberg

The letter from Sen. Ted Cruz (R-Texas) accusing Wikipedia of left-wing bias seems to be based on fundamental misunderstandings of how the platform works, according to a lawyer for the nonprofit foundation that operates the online encyclopedia.

“The foundation is very much taking the approach that Wikipedia is actually pretty great and a lot of what’s in this letter is actually misunderstandings,” Jacob Rogers, associate general counsel at the Wikimedia Foundation, told Ars in an interview. “And so we are more than happy, despite the pressure that comes from these things, to help people better understand how Wikipedia works.”

Cruz’s letter to Wikimedia Foundation CEO Maryana Iskander expressed concern “about ideological bias on the Wikipedia platform and at the Wikimedia Foundation.” Cruz alleged that Wikipedia articles “often reflect a left-wing bias.” He asked the foundation for “documents sufficient to show what supervision, oversight, or influence, if any, the Wikimedia Foundation has over the editing community,” and “documents sufficient to show how the Wikimedia Foundation addresses political or ideological bias.”

As many people know, Wikipedia is edited by volunteers through a collaborative process.

“We’re not deciding what the editorial policies are for what is on Wikipedia,” Rogers said, describing the Wikimedia Foundation’s hands-off approach. “All of that, both the writing of the content and the determining of the editorial policies, is done through the volunteer editors” through “public conversation and discussion and trying to come to a consensus. They make all of that visible in various ways to the reader. So you go and you read a Wikipedia article, you can see what the sources are, what someone has written, you can follow the links yourselves.”

“They’re worried about something that is just not present at all”

Cruz’s letter raised concerns about “the influence of large donors on Wikipedia’s content creation or editing practices.” But Rogers said that “people who donate to Wikipedia don’t have any influence over content and we don’t even have that many large donors to begin with. It is primarily funded by people donating through the website fundraisers, so I think they’re worried about something that is just not present at all.”

Anyone unhappy with Wikipedia content can participate in the writing and editing, he said. “It’s still open for everybody to participate. If someone doesn’t like what it says, they can go on and say, ‘Hey, I don’t like the sources that are being used, or I think a different source should be used that isn’t there,'” Rogers said. “Other people might disagree with them, but they can have that conversation and try to figure it out and make it better.”

Rogers said that some people wrongly assume there is central control over Wikipedia editing. “I feel like people are asking questions assuming that there is something more central that is controlling all of this that doesn’t actually exist,” he said. “I would love to see it a little better understood about how this sort of public model works and the fact that people can come judge it for themselves and participate for themselves. And maybe that will have it sort of die down as a source of government pressure, government questioning, and go onto something else.”

Cruz’s letter accused Wikipedia of pushing antisemitic narratives. He described the Wikimedia Foundation as “intervening in editorial decisions” in an apparent reference to an incident in which the platform’s Arbitration Committee responded to editing conflicts on the Israeli–Palestinian conflict by banning eight editors.

“The Wikimedia Foundation has said it is taking steps to combat this editing campaign, raising further questions about the extent to which it is intervening in editorial decisions and to what end,” Cruz wrote.

Explaining the Arbitration Committee

The Arbitration Committee for the English-language edition of Wikipedia consists of volunteers who “are elected by the rest of the English Wikipedia editors,” Rogers said. The group is a “dispute resolution body when people can’t otherwise resolve their disputes.” The committee made “a ruling on Israel/Palestine because it is such a controversial subject and it’s not just banning eight editors, it’s also how contributions are made in that topic area and sort of limiting it to more experienced editors,” he said.

The members of the committee “do not control content,” Rogers said. “The arbitration committee is not a content dispute body. They’re like a behavior conduct dispute body, but they try to set things up so that fights will not break out subsequently.”

As with other topics, people can participate if they believe articles are antisemitic. “That is sort of squarely in the user editorial processes,” Rogers said. “If someone thinks that something on Wikipedia is antisemitic, they should change it or propose to people working on it that they change it or change sources. I do think the editorial community, especially on topics related to antisemitism and related to Israel/Palestine, has a lot of various safeguards in place. That particular topic is probably the most controversial topic in the world, but there’s still a lot of editorial safeguards in place where people can discuss things. They can get help with dispute resolution from bringing in other editors if there’s a behavioral problem, they can ask for help from Wikipedia administrators, and all the way up to the English Wikipedia arbitration committee.”

Cruz’s letter called out Wikipedia’s goal of “knowledge equity,” and accused the foundation of favoring “ideology over neutrality.” Cruz also pointed to a Daily Caller report that the foundation donated “to activist groups seeking to bring the online encyclopedia more in line with traditionally left-of-center points of view.”

Rogers countered that “the theory behind that is sort of misunderstood by the letter where it’s not about equity like the DEI equity, it is about the mission of the Wikimedia Foundation to have the world’s knowledge, to prepare educational content and to have all the different knowledge in the world to the extent possible.” In topic areas where people with expertise haven’t contributed much to Wikipedia, “we are looking to write grants to help fill in those gaps in knowledge and have a more broad range of information and sources,” he said.

What happens next

Rogers is familiar with the workings of Senate investigations from personal experience. He joined the Wikimedia Foundation in 2014 after working for the Senate’s Permanent Subcommittee on Investigations under the late Sen. Carl Levin (D-Mich.).

While Cruz demanded a trove of documents, Rogers said the foundation doesn’t necessarily have to provide them. A subpoena could be issued to Wikimedia, but that hasn’t happened.

“What Cruz has sent us is just a letter,” Rogers said. “There is no legal proceeding whatsoever. There’s no formal authority behind this letter. It’s just a letter from a person in the legislative branch who cares about the topic, so there is nothing compelling us to give him anything. I think we are probably going to answer the letter, but there’s no sort of legal requirement to actually fully provide everything that answers every question.” Assuming it responds, the foundation would try to answer Cruz’s questions “to the extent that we can, and without violating any of our company policies,” and without giving out nonpublic information, he said.

A letter responding to Cruz wouldn’t necessarily be made public. In April, the foundation received a letter from 23 lawmakers about alleged antisemitism and anti-Israel bias. The foundation’s response to that letter is not public.

Cruz is seeking changes at Wikipedia just a couple weeks after criticizing Federal Communications Commission Chairman Brendan Carr for threatening ABC with station license revocations over political content on Jimmy Kimmel’s show. While the pressure tactics used by Cruz and Carr have similarities, Rogers said there are also key differences between the legislative and executive branches.

“Congressional committees, they are investigating something to determine what laws to make, and so they have a little bit more freedom to just look into the state of the world to try to decide what laws they want to write or what laws they want to change,” he said. “That doesn’t mean that they can’t use their authority in a way that might ultimately go down a path of violating the First Amendment or something like that. They have a little bit more runway to get there versus an executive branch agency which, if it is pressuring someone, it is doing so for a very immediate decision usually.”

What does Cruz want? It’s unclear

Rogers said it’s not clear whether Cruz’s inquiry is the first step toward changing the law. “The questions in the letter don’t really say why they want the information they want other than the sort of immediacy of their concerns,” he said.

Cruz chairs the Senate Commerce Committee, which “does have lawmaking authority over the Internet writ large,” Rogers said. “So they may be thinking about changes to the law.”

One potential target is Section 230 of the Communications Decency Act, which gives online platforms immunity from lawsuits over how they moderate user-submitted content.

“From the perspective of the foundation, we’re staunch defenders of Section 230,” Rogers said, adding that Wikimedia supports “broad laws around intellectual property and privacy and other things that allow a large amount of material to be appropriately in the public domain, to be written about on a free encyclopedia like Wikipedia, but that also protect the privacy of editors who are contributing to Wikipedia.”

Photo of Jon Brodkin

Jon is a Senior IT Reporter for Ars Technica. He covers the telecom industry, Federal Communications Commission rulemakings, broadband consumer affairs, court cases, and government regulation of the tech industry.

Ted Cruz doesn’t seem to understand Wikipedia, lawyer for Wikimedia says Read More »

ted-cruz-picks-a-fight-with-wikipedia,-accusing-platform-of-left-wing-bias

Ted Cruz picks a fight with Wikipedia, accusing platform of left-wing bias

Cruz pressures Wikipedia after criticizing FCC chair

Cruz sent the letter about two weeks after criticizing Federal Communications Commission Chairman Brendan Carr for threatening ABC with station license revocations over political content on Jimmy Kimmel’s show. Cruz said that using the government to dictate what the media can say “will end up bad for conservatives” because when Democrats are back in power, “they will silence us, they will use this power, and they will use it ruthlessly.” Cruz said that Carr threatening ABC was like “a mafioso coming into a bar going, ‘Nice bar you have here, it’d be a shame if something happened to it.'”

Cruz, who chairs the Senate Commerce Committee, doesn’t mind using his authority to pressure Wikipedia’s operator, however. “The Standing Rules of the Senate grant the Committee on Commerce, Science, and Transportation jurisdiction over communications, including online information platforms,” he wrote to the Wikimedia Foundation. “As the Chairman of the Committee, I request that you provide written responses to the questions below, as well as requested documents, no later than October 17, 2025, and in accordance with the attached instructions.”

We asked Cruz’s office to explain why a senator pressuring Wikipedia is appropriate while an FCC chair pressuring ABC is not and will update this article if we get a response.

Among other requests, Cruz asked for “documents sufficient to show what supervision, oversight, or influence, if any, the Wikimedia Foundation has over the editing community,” and “documents sufficient to show how the Wikimedia Foundation addresses political or ideological bias.”

Cruz has separately been launching investigations into the Biden administration for alleged censorship. He issued a report allegedly “revealing how the Biden administration transformed the Cybersecurity and Infrastructure Security Agency (CISA) into an agent of censorship pressuring Big Tech to police speech,” and scheduled a hearing for Wednesday titled, “Shut Your App: How Uncle Sam Jawboned Big Tech Into Silencing Americans.”

Cruz’s letter to Wikimedia seeks evidence that could figure into his ongoing investigations into the Biden administration. “Provide any and all documents and communications—including emails, texts, or other digital messages—between any officer, employee, or agent of the Wikimedia Foundation and any officer, employee, or agent of the federal government since January 1, 2020,” the letter said.

Ted Cruz picks a fight with Wikipedia, accusing platform of left-wing bias Read More »

dedicated-volunteer-exposes-“single-largest-self-promotion-operation-in-wikipedia’s-history”

Dedicated volunteer exposes “single largest self-promotion operation in Wikipedia’s history”

After a reduction in activity, things ramped up again in 2021, as IP addresses from around the world started creating Woodard references and articles once more. For instance, “addresses from Canada, Germany, Indonesia, the UK and other places added some trivia about Woodard to all 15 Wikipedia articles about the calea ternifolia.”

Then things got “more sophisticated.” From December 2021 through June 2025, 183 articles were created about Woodard, each in a different language’s Wikipedia and each by a unique account. These accounts followed a pattern of behavior: They were “created, often with a fairly generic name, and made a user page with a single image on it. They then made dozens of minor edits to unrelated articles, before creating an article about David Woodard, then making a dozen or so more minor edits before disappearing off the platform.”

Grnrchst believes that all the activity was meant to “create as many articles about Woodard as possible, and to spread photos of and information on Woodard to as many articles as possible, while hiding that activity as much as possible… I came to believe that David Woodard himself, or someone close to him, had been operating this network of accounts and IP addresses for the purposes of cynical self-promotion.”

After the Grnrchst report, Wikipedia’s global stewards removed 235 articles on Woodard from Wikipedia instances with few users or administrators. Larger Wikipedias were free to make their own community decisions, and they removed another 80 articles and banned numerous accounts.

“A full decade of dedicated self-promotion by an individual network has been undone in only a few weeks by our community,” Grnrchst noted.

In the end, just 20 articles about Woodard remain, such as this one in English, which does not mention the controversy.

We were unable to get in touch with Woodard, whose personal website is password-protected and only available “by invitation.”

Could the whole thing be some kind of “art project,” with the real payoff being exposure and being written about? Perhaps. But whatever the motive behind the decade-long effort to boost Woodard on Wikipedia, the incident reminds us just how much effort some people are willing to put into polluting open or public-facing projects for their own ends.

Dedicated volunteer exposes “single largest self-promotion operation in Wikipedia’s history” Read More »

“yuck”:-wikipedia-pauses-ai-summaries-after-editor-revolt

“Yuck”: Wikipedia pauses AI summaries after editor revolt

Generative AI is permeating the Internet, with chatbots and AI summaries popping up faster than we can keep track. Even Wikipedia, the vast repository of knowledge famously maintained by an army of volunteer human editors, is looking to add robots to the mix. The site began testing AI summaries in some articles over the past week, but the project has been frozen after editors voiced their opinions. And that opinion is: “yuck.”

The seeds of this project were planted at Wikimedia’s 2024 conference, where foundation representatives and editors discussed how AI could advance Wikipedia’s mission. The wiki on the so-called “Simple Article Summaries” notes that the editors who participated in the discussion believed the summaries could improve learning on Wikipedia.

According to 404 Media, Wikipedia announced the opt-in AI pilot on June 2, which was set to run for two weeks on the mobile version of the site. The summaries appeared at the top of select articles in a collapsed form. Users had to tap to expand and read the full summary. The AI text also included a highlighted “Unverified” badge.

Feedback from the larger community of editors was immediate and harsh. Some of the first comments were simply “yuck,” with others calling the addition of AI a “ghastly idea” and “PR hype stunt.”

Others expounded on the issues with adding AI to Wikipedia, citing a potential loss of trust in the site. Editors work together to ensure articles are accurate, featuring verifiable information and a neutral point of view. However, nothing is certain when you put generative AI in the driver’s seat. “I feel like people seriously underestimate the brand risk this sort of thing has,” said one editor. “Wikipedia’s brand is reliability, traceability of changes, and ‘anyone can fix it.’ AI is the opposite of these things.”

“Yuck”: Wikipedia pauses AI summaries after editor revolt Read More »

ai-bots-strain-wikimedia-as-bandwidth-surges-50%

AI bots strain Wikimedia as bandwidth surges 50%

Crawlers that evade detection

Making the situation more difficult, many AI-focused crawlers do not play by established rules. Some ignore robots.txt directives. Others spoof browser user agents to disguise themselves as human visitors. Some even rotate through residential IP addresses to avoid blocking, tactics that have become common enough to force individual developers like Xe Iaso to adopt drastic protective measures for their code repositories.

This leaves Wikimedia’s Site Reliability team in a perpetual state of defense. Every hour spent rate-limiting bots or mitigating traffic surges is time not spent supporting Wikimedia’s contributors, users, or technical improvements. And it’s not just content platforms under strain. Developer infrastructure, like Wikimedia’s code review tools and bug trackers, is also frequently hit by scrapers, further diverting attention and resources.

These problems mirror others in the AI scraping ecosystem over time. Curl developer Daniel Stenberg has previously detailed how fake, AI-generated bug reports are wasting human time. On his blog, SourceHut’s Drew DeVault highlight how bots hammer endpoints like git logs, far beyond what human developers would ever need.

Across the Internet, open platforms are experimenting with technical solutions: proof-of-work challenges, slow-response tarpits (like Nepenthes), collaborative crawler blocklists (like “ai.robots.txt“), and commercial tools like Cloudflare’s AI Labyrinth. These approaches address the technical mismatch between infrastructure designed for human readers and the industrial-scale demands of AI training.

Open commons at risk

Wikimedia acknowledges the importance of providing “knowledge as a service,” and its content is indeed freely licensed. But as the Foundation states plainly, “Our content is free, our infrastructure is not.”

The organization is now focusing on systemic approaches to this issue under a new initiative: WE5: Responsible Use of Infrastructure. It raises critical questions about guiding developers toward less resource-intensive access methods and establishing sustainable boundaries while preserving openness.

The challenge lies in bridging two worlds: open knowledge repositories and commercial AI development. Many companies rely on open knowledge to train commercial models but don’t contribute to the infrastructure making that knowledge accessible. This creates a technical imbalance that threatens the sustainability of community-run platforms.

Better coordination between AI developers and resource providers could potentially resolve these issues through dedicated APIs, shared infrastructure funding, or more efficient access patterns. Without such practical collaboration, the platforms that have enabled AI advancement may struggle to maintain reliable service. Wikimedia’s warning is clear: Freedom of access does not mean freedom from consequences.

AI bots strain Wikimedia as bandwidth surges 50% Read More »

ai-generated-articles-prompt-wikipedia-to-downgrade-cnet’s-reliability-rating

AI-generated articles prompt Wikipedia to downgrade CNET’s reliability rating

The hidden costs of AI —

Futurism report highlights the reputational cost of publishing AI-generated content.

The CNET logo on a smartphone screen.

Wikipedia has downgraded tech website CNET’s reliability rating following extensive discussions among its editors regarding the impact of AI-generated content on the site’s trustworthiness, as noted in a detailed report from Futurism. The decision reflects concerns over the reliability of articles found on the tech news outlet after it began publishing AI-generated stories in 2022.

Around November 2022, CNET began publishing articles written by an AI model under the byline “CNET Money Staff.” In January 2023, Futurism brought widespread attention to the issue and discovered that the articles were full of plagiarism and mistakes. (Around that time, we covered plans to do similar automated publishing at BuzzFeed.) After the revelation, CNET management paused the experiment, but the reputational damage had already been done.

Wikipedia maintains a page called “Reliable sources/Perennial sources” that includes a chart featuring news publications and their reliability ratings as viewed from Wikipedia’s perspective. Shortly after the CNET news broke in January 2023, Wikipedia editors began a discussion thread on the Reliable Sources project page about the publication.

“CNET, usually regarded as an ordinary tech RS [reliable source], has started experimentally running AI-generated articles, which are riddled with errors,” wrote a Wikipedia editor named David Gerard. “So far the experiment is not going down well, as it shouldn’t. I haven’t found any yet, but any of these articles that make it into a Wikipedia article need to be removed.”

After other editors agreed in the discussion, they began the process of downgrading CNET’s reliability rating.

As of this writing, Wikipedia’s Perennial Sources list currently features three entries for CNET broken into three time periods: (1) before October 2020, when Wikipedia considered CNET a “generally reliable” source; (2) between October 2020 and October 2022, where Wikipedia notes that the site was acquired by Red Ventures in October 2020, “leading to a deterioration in editorial standards” and saying there is no consensus about reliability; and (3) between November 2022 and present, where Wikipedia currently considers CNET “generally unreliable” after the site began using an AI tool “to rapidly generate articles riddled with factual inaccuracies and affiliate links.”

A screenshot of a chart featuring CNET's reliability ratings, as found on Wikipedia's

Enlarge / A screenshot of a chart featuring CNET’s reliability ratings, as found on Wikipedia’s “Perennial Sources” page.

Futurism reports that the issue with CNET’s AI-generated content also sparked a broader debate within the Wikipedia community about the reliability of sources owned by Red Ventures, such as Bankrate and CreditCards.com. Those sites published AI-generated content around the same period of time as CNET. The editors also criticized Red Ventures for not being forthcoming about where and how AI was being implemented, further eroding trust in the company’s publications. This lack of transparency was a key factor in the decision to downgrade CNET’s reliability rating.

In response to the downgrade and the controversies surrounding AI-generated content, CNET issued a statement that claims that the site maintains high editorial standards.

“CNET is the world’s largest provider of unbiased tech-focused news and advice,” a CNET spokesperson said in a statement to Futurism. “We have been trusted for nearly 30 years because of our rigorous editorial and product review standards. It is important to clarify that CNET is not actively using AI to create new content. While we have no specific plans to restart, any future initiatives would follow our public AI policy.”

This article was updated on March 1, 2024 at 9: 30am to reflect fixes in the date ranges for CNET on the Perennial Sources page.

AI-generated articles prompt Wikipedia to downgrade CNET’s reliability rating Read More »