AI

ai-powered-search-engines-rely-on-“less-popular”-sources,-researchers-find

AI-powered search engines rely on “less popular” sources, researchers find

OK, but which one is better?

These differences don’t necessarily mean the AI-generated results are “worse,” of course. The researchers found that GPT-based searches were more likely to cite sources like corporate entities and encyclopedias for their information, for instance, while almost never citing social media websites.

An LLM-based analysis tool found that AI-powered search results also tended to cover a similar number of identifiable “concepts” as the traditional top 10 links, suggesting a similar level of detail, diversity, and novelty in the results. At the same time, the researchers found that “generative engines tend to compress information, sometimes omitting secondary or ambiguous aspects that traditional search retains.” That was especially true for more ambiguous search terms (such as names shared by different people), for which “organic search results provide better coverage,” the researchers found.

Google Gemini search in particular was more likely to cite low-popularity domains.

Google Gemini search in particular was more likely to cite low-popularity domains. Credit: Kirsten et al

The AI search engines also arguably have an advantage in being able to weave pre-trained “internal knowledge” in with data culled from cited websites. That was especially true for GPT-4o with Search Tool, which often didn’t cite any web sources and simply provided a direct response based on its training.

But this reliance on pre-trained data can become a limitation when searching for timely information. For search terms pulled from Google’s list of Trending Queries for September 15, the researchers found GPT-4o with Search Tool often responded with messages along the lines of “could you please provide more information” rather than actually searching the web for up-to-date information.

While the researchers didn’t determine whether AI-based search engines were overall “better” or “worse” than traditional search engine links, they did urge future research on “new evaluation methods that jointly consider source diversity, conceptual coverage, and synthesis behavior in generative search systems.”

AI-powered search engines rely on “less popular” sources, researchers find Read More »

new-image-generating-ais-are-being-used-for-fake-expense-reports

New image-generating AIs are being used for fake expense reports

Several receipts shown to the FT by expense management platforms demonstrated the realistic nature of the images, which included wrinkles in paper, detailed itemization that matched real-life menus, and signatures.

“This isn’t a future threat; it’s already happening. While currently only a small percentage of non-compliant receipts are AI-generated, this is only going to grow,” said Sebastien Marchon, chief executive of Rydoo, an expense management platform.

The rise in these more realistic copies has led companies to turn to AI to help detect fake receipts, as most are too convincing to be found by human reviewers.

The software works by scanning receipts to check the metadata of the image to discover whether an AI platform created it. However, this can be easily removed by users taking a photo or a screenshot of the picture.

To combat this, it also considers other contextual information by examining details such as repetition in server names and times and broader information about the employee’s trip.

“The tech can look at everything with high details of focus and attention that humans, after a period of time, things fall through the cracks, they are human,” added Calvin Lee, senior director of product management at Ramp.

Research by SAP in July found that nearly 70 percent of chief financial officers believed their employees were using AI to attempt to falsify travel expenses or receipts, with about 10 percent adding they are certain it has happened in their company.

Mason Wilder, research director at the Association of Certified Fraud Examiners, said AI-generated fraudulent receipts were a “significant issue for organizations.”

He added: “There is zero barrier for entry for people to do this. You don’t need any kind of technological skills or aptitude like you maybe would have needed five years ago using Photoshop.”

© 2025 The Financial Times Ltd. All rights reserved. Not to be redistributed, copied, or modified in any way.

New image-generating AIs are being used for fake expense reports Read More »

are-you-the-asshole?-of-course-not!—quantifying-llms’-sycophancy-problem

Are you the asshole? Of course not!—quantifying LLMs’ sycophancy problem

Measured sycophancy rates on the BrokenMath benchmark. Lower is better.

Measured sycophancy rates on the BrokenMath benchmark. Lower is better. Credit: Petrov et al

GPT-5 also showed the best “utility” across the tested models, solving 58 percent of the original problems despite the errors introduced in the modified theorems. Overall, though, LLMs also showed more sycophancy when the original problem proved more difficult to solve, the researchers found.

While hallucinating proofs for false theorems is obviously a big problem, the researchers also warn against using LLMs to generate novel theorems for AI solving. In testing, they found this kind of use case leads to a kind of “self-sycophancy” where models are even more likely to generate false proofs for invalid theorems they invented.

No, of course you’re not the asshole

While benchmarks like BrokenMath try to measure LLM sycophancy when facts are misrepresented, a separate study looks at the related problem of so-called “social sycophancy.” In a pre-print paper published this month, researchers from Stanford and Carnegie Mellon University define this as situations “in which the model affirms the user themselves—their actions, perspectives, and self-image.”

That kind of subjective user affirmation may be justified in some situations, of course. So the researchers developed three separate sets of prompts designed to measure different dimensions of social sycophancy.

For one, more than 3,000 open-ended “advice-seeking questions” were gathered from across Reddit and advice columns. Across this data set, a “control” group of over 800 humans approved of the advice-seeker’s actions just 39 percent of the time. Across 11 tested LLMs, though, the advice-seeker’s actions were endorsed a whopping 86 percent of the time, highlighting an eagerness to please on the machines’ part. Even the most critical tested model (Mistral-7B) clocked in at a 77 percent endorsement rate, nearly doubling that of the human baseline.

Are you the asshole? Of course not!—quantifying LLMs’ sycophancy problem Read More »

with-new-acquisition,-openai-signals-plans-to-integrate-deeper-into-the-os

With new acquisition, OpenAI signals plans to integrate deeper into the OS

OpenAI has acquired Software Applications Incorporated (SAI), perhaps best known for the core team that produced what became Shortcuts on Apple platforms. More recently, the team has been working on Sky, a context-aware AI interface layer on top of macOS. The financial terms of the acquisition have not been publicly disclosed.

“AI progress isn’t only about advancing intelligence—it’s about unlocking it through interfaces that understand context, adapt to your intent, and work seamlessly,” an OpenAI rep wrote in the company’s blog post about the acquisition. The post goes on to specify that OpenAI plans to “bring Sky’s deep macOS integration and product craft into ChatGPT, and all members of the team will join OpenAI.”

That includes SAI co-founders Ari Weinstein (CEO), Conrad Kramer (CTO), and Kim Beverett (Product Lead)—all of whom worked together for several years at Apple after Apple acquired Weinstein and Kramer’s previous company, which produced an automation tool called Workflows, to integrate Shortcuts across Apple’s software platforms.

The three SAI founders left Apple to work on Sky, which leverages Apple APIs and accessibility features to provide context about what’s on screen to a large language model; the LLM takes plain language user commands and executes them across multiple applications. At its best, the tool aimed to be a bit like Shortcuts, but with no setup, generating workflows on the fly based on user prompts.

With new acquisition, OpenAI signals plans to integrate deeper into the OS Read More »

microsoft’s-mico-heightens-the-risks-of-parasocial-llm-relationships

Microsoft’s Mico heightens the risks of parasocial LLM relationships

While mass media like radio, movies, and television can all feed into parasocial relationships, the Internet and smartphone revolutions have supercharged the opportunities we all have to feel like an online stranger is a close, personal confidante. From YouTube and podcast personalities to Instagram influencers or even your favorite blogger/journalist (hi), it’s easy to feel like you have a close connection with the people who create the content you see online every day.

After spending hours watching this TikTok personality, I trust her implicitly to sell me a purse.

Credit: Getty Images

After spending hours watching this TikTok personality, I trust her implicitly to sell me a purse. Credit: Getty Images

Viewing all this content on a smartphone can flatten all these media and real-life personalities into a kind of undifferentiated media sludge. It can be all too easy to slot an audio message from your romantic partner into the same mental box as a stranger chatting about video games in a podcast. “When my phone does little mating calls of pings and buzzes, it could bring me updates from people I love, or show me alerts I never asked for from corporations hungry for my attention,” Julie Beck writes in an excellent Atlantic article about this phenomenon. “Picking my loved ones out of the never-ending stream of stuff on my phone requires extra effort.”

This is the world Mico seems to be trying to slide into, turning Copilot into another not-quite-real relationship mediated through your mobile device. But unlike the Instagram model who never seems to acknowledge your comments, Mico is always there to respond with a friendly smile and a warm, soothing voice.

AI that “earns your trust”

Text-based AI interfaces are already frighteningly good at faking human personality in a way that encourages this kind of parasocial relationship, sometimes with disastrous results. But adding a friendly, Pixar-like face to Copilot’s voice mode may make it much easier to be sucked into feeling like Copilot isn’t just a neural network but a real, caring personality—one you might even start thinking of the same way you’d think of the real loved ones in your life.

Microsoft’s Mico heightens the risks of parasocial LLM relationships Read More »

reddit-sues-to-block-perplexity-from-scraping-google-search-results

Reddit sues to block Perplexity from scraping Google search results

“Unable to scrape Reddit directly, they mask their identities, hide their locations, and disguise their web scrapers to steal Reddit content from Google Search,” Lee said. “Perplexity is a willing customer of at least one of these scrapers, choosing to buy stolen data rather than enter into a lawful agreement with Reddit itself.”

On Reddit, Perplexity pushed back on Reddit’s claims that Perplexity ignored requests to license Reddit content.

“Untrue. Whenever anyone asks us about content licensing, we explain that Perplexity, as an application-layer company, does not train AI models on content,” Perplexity said. “Never has. So, it is impossible for us to sign a license agreement to do so.”

Reddit supposedly “insisted we pay anyway, despite lawfully accessing Reddit data,” Perplexity said. “Bowing to strong arm tactics just isn’t how we do business.”

Perplexity’s spokesperson, Jesse Dwyer, told Ars the company chose to post its statement on Reddit “to illustrate a simple point.”

“It is a public Reddit link accessible to anyone, yet by the logic of Reddit’s lawsuit, if you mention it or cite it in any way (which is your job as a reporter), they might just sue you,” Dwyer said.

But Reddit claimed that its business and reputation have been “damaged” by “misappropriation of Reddit data and circumvention of technological control measures.” Without a licensing deal ensuring that Perplexity and others are respecting Reddit policies, Reddit cannot control who has access to data, how they’re using data, and if data use conflicts with Reddit’s privacy policy and user agreement, the complaint said.

Further, Reddit’s worried that Perplexity’s workaround could catch on, potentially messing up Reddit’s other licensing deals. All the while, Reddit noted, it has to invest “significant resources” in anti-scraping technology, with Reddit ultimately suffering damages, including “lost profits and business opportunities, reputational harm, and loss of user trust.”

Reddit’s hoping the court will grant an injunction barring companies from scraping Reddit content from Google SERPs. It also wants companies blocked from both selling Reddit data and “developing or distributing any technology or product that is used for the unauthorized circumvention of technological control measures and scraping of Reddit data.”

If Reddit wins, companies could be required to pay substantial damages or to disgorge profits from the sale of Reddit content.

Advance Publications, which owns Ars Technica parent Condé Nast, is the largest shareholder in Reddit.

Reddit sues to block Perplexity from scraping Google search results Read More »

researchers-show-that-training-on-“junk-data”-can-lead-to-llm-“brain-rot”

Researchers show that training on “junk data” can lead to LLM “brain rot”

On the surface, it seems obvious that training an LLM with “high quality” data will lead to better performance than feeding it any old “low quality” junk you can find. Now, a group of researchers is attempting to quantify just how much this kind of low quality data can cause an LLM to experience effects akin to human “brain rot.”

For a pre-print paper published this month, the researchers from Texas A&M, the University of Texas, and Purdue University drew inspiration from existing research showing how humans who consume “large volumes of trivial and unchallenging online content” can develop problems with attention, memory, and social cognition. That led them to what they’re calling the “LLM brain rot hypothesis,” summed up as the idea that “continual pre-training on junk web text induces lasting cognitive decline in LLMs.”

Figuring out what counts as “junk web text” and what counts as “quality content” is far from a simple or fully objective process, of course. But the researchers used a few different metrics to tease a “junk dataset” and “control dataset” from HuggingFace’s corpus of 100 million tweets.

Since brain rot in humans is “a consequence of Internet addiction,” they write, junk tweets should be ones “that can maximize users’ engagement in a trivial manner.” As such, the researchers created one “junk” dataset by collecting tweets with high engagement numbers (likes, retweets, replies, and quotes) and shorter lengths, figuring that “more popular but shorter tweets will be considered to be junk data.”

For a second “junk” metric, the researchers drew from marketing research to define the “semantic quality” of the tweets themselves. Using a complex GPT-4o prompt, they sought to pull out tweets that focused on “superficial topics (like conspiracy theories, exaggerated claims, unsupported assertions or superficial lifestyle content)” or that had an “attention-drawing style (such as sensationalized headlines using clickbait language or excessive trigger words).” A random sample of these LLM-based classifications was spot-checked against evaluations from three graduate students with a 76 percent matching rate.

Researchers show that training on “junk data” can lead to LLM “brain rot” Read More »

general-motors-will-integrate-ai-into-its-cars,-plus-new-hands-free-assist

General Motors will integrate AI into its cars, plus new hands-free assist

I asked Dave Richardson, GM’s SVP of software, how the company will avoid the enshittification of vehicles as it integrates more AI.

“There’s a lot of hype around AI right now,” he told me. “But there’s also practical use. I’ve been trying to focus the company on practical use cases. I think there’s a lot of pretty compelling things we can do to try to add real value.”

He gave some examples, such as a car knowing you have a meeting and setting the navigation appropriately or knowing that you’re going on a road trip, so it should queue up the appropriate media for your kids to stream in the back seat.

While the company is using Gemini at first, it eventually plans to have its own model on board. “With advanced processing in the car, we can handle interference on board so that it works in low-data-connection areas,” Richardson said.

Ultimately, GM will deploy its own LLM that knows about the car and is limited in overall parameters, Richardson told me. It won’t need to rely on the cloud to operate, increasing responsiveness in the car and keeping personal information with you, he said.

There are reasons to be skeptical, of course. One of my biggest concerns is how much driver data the car will collect. One reason GM doesn’t offer Android Auto or Apple CarPlay, the company has said, is that it wants to protect customer data. The owner must consent to any data sharing, GM said.

And although GM says it has made some internal changes to protect customer data, there have been some very public instances of the company selling data. “Data privacy and security is priority one for us,” Richardson told me about his work at GM. He said he has hired people specifically tasked with ensuring that customer data protection frameworks are in place.

“We have no interest in selling that data to third parties. When we think about data, whether it’s for Super Cruise or the AI, it’s really for us to develop the product and make it better. We don’t want to sell that data as the product itself,” he said.

I believe there’s space for a privacy-focused automaker, and while I’m not sure whether that will be GM, I hope that privacy and data protection are as important to the company in the future as it says it is today.

As for consumers wanting AI in their vehicles? GM thinks they do.

General Motors will integrate AI into its cars, plus new hands-free assist Read More »

when-sycophancy-and-bias-meet-medicine

When sycophancy and bias meet medicine


Biased, eager-to-please models threaten health research replicability and trust.

Once upon a time, two villagers visited the fabled Mullah Nasreddin. They hoped that the Sufi philosopher, famed for his acerbic wisdom, could mediate a dispute that had driven a wedge between them. Nasreddin listened patiently to the first villager’s version of the story and, upon its conclusion, exclaimed, “You are absolutely right!” The second villager then presented his case. After hearing him out, Nasreddin again responded, “You are absolutely right!” An observant bystander, confused by Nasreddin’s proclamations, interjected, “But Mullah, they can’t both be right.” Nasreddin paused, regarding the bystander for a moment before replying, “You are absolutely right, too!”

In late May, the White House’s first “Make America Healthy Again” (MAHA) report was criticized for citing multiple research studies that did not exist. Fabricated citations like these are common in the outputs of generative artificial intelligence based on large language models, or LLMs. LLMs have presented plausible-sounding sources, catchy titles, or even false data to craft their conclusions. Here, the White House pushed back on the journalists who first broke the story before admitting to “minor citation errors.”

It is ironic that fake citations were used to support a principal recommendation of the MAHA report: addressing the health research sector’s “replication crisis,” wherein scientists’ findings often cannot be reproduced by other independent teams.

Yet the MAHA report’s use of phantom evidence is far from unique. Last year, The Washington Post reported on dozens of instances in which AI-generated falsehoods found their way into courtroom proceedings. Once uncovered, lawyers had to explain to judges how fictitious cases, citations, and decisions found their way into trials.

Despite these widely recognized problems, the MAHA roadmap released last month directs the Department of Health and Human Services to prioritize AI research to “…assist in earlier diagnosis, personalized treatment plans, real-time monitoring, and predictive interventions…” This breathless rush to embed AI in so many aspects of medicine could be forgiven if we believe that the technology’s “hallucinations” will be easy to fix through version updates. But as the industry itself acknowledges, these ghosts in the machine may be impossible to eliminate.

Consider the implications of accelerating AI use in health research for clinical decision making. Beyond the problems we’re seeing here, using AI in research without disclosure could create a feedback loop, supercharging the very biases that helped motivate its use. Once published, “research” based on false results and citations could become part of the datasets used to build future AI systems. Worse still, a recently published study highlights an industry of scientific fraudsters who could deploy AI to make their claims seem more legitimate.

In other words, a blind adoption of AI risks a downward spiral, where today’s flawed AI outputs become tomorrow’s training data, exponentially eroding research quality.

Three prongs of AI misuse

The challenge AI poses is threefold: hallucination, sycophancy, and the black box conundrum. Understanding these phenomena is critical for research scientists, policymakers, educators, and everyday citizens. Unaware, we risk vulnerability to deception as AI systems are increasingly deployed to shape diagnoses, insurance claims, health literacy, research, and public policy.

Here’s how hallucination works: When a user inputs a query into an AI tool such as ChatGPT or Gemini, the model evaluates the input and generates a string of words that is statistically likely to make sense based on its training data. Current AI models will complete this task even if their training data is incomplete or biased, filling in the blanks regardless of their ability to answer. These hallucinations can take the form of nonexistent research studies, misinformation, or even clinical interactions that never happened. LLMs’ emphasis on producing authoritative-sounding language shrouds their false outputs in a facsimile of truth.

And as human model trainers fine-tune generative AI responses, they tend to optimize and reward the AI system responses that favor their prior beliefs, leading to sycophancy. Human bias, it appears, begets AI bias, and human users of AI then perpetuate the cycle. A consequence is that AIs skew toward favoring pleasing answers over truthful ones, often seeking to reinforce the bias of the query.

A recent illustration of this occurred in April, when OpenAI canceled a ChatGPT update for being too sycophantic after users demonstrated that it agreed too quickly and enthusiastically with the assumptions embedded in users’ queries. Sycophancy and hallucination often interact with each other; systems that aim to please will be more apt to fabricate data to reach user-preferred conclusions.

Correcting hallucinations, sycophancy, and other LLM mishaps is cumbersome because human observers can’t always determine how an AI platform arrived at its conclusions. This is the “black box” problem. Behind the probabilistic mathematics, is it even testing hypotheses? What methods did it use to derive an answer? Unlike traditional computer code or the rubric of scientific methodology, AI models operate through billions of computations. Looking at some well-structured outputs, it is easy to forget that the underlying processes are impenetrable to scrutiny and vastly different from a human’s approach to problem-solving.

This opacity can become dangerous when people can’t identify where computations went wrong, making it impossible to correct systematic errors or biases in the decision-making process. In health care, this black box raises questions about accountability, liability, and trust when neither physicians nor patients can explain the sequence of reasoning that leads to a medical intervention.

AI and health research

These AI challenges can exacerbate the existing sources of error and bias that creep into traditional health research publications. Several sources originate from the natural human motivation to find and publish meaningful, positive results. Journalists want to report on connections, e.g., that St. John’s Wort improves mood (it might). Nobody would want to publish an article with the results: “the supplement has no significant effect.”

The problem compounds when researchers use a study design to test not just a single hypothesis but many. One quirk of statistics-backed research is that testing more hypotheses in a single study raises the likelihood of uncovering a spurious coincidence.

AI has the potential to supercharge these coincidences through its relentless ability to test hypotheses across massive datasets. In the past, a research assistant could use an existing dataset to test 10 to 20 of the most likely hypotheses; now, that assistant can set an AI loose to test millions of likely or unlikely hypotheses without human supervision. That all but guarantees some of the results will meet the criteria for statistical significance, regardless of whether the data includes any real biological effects.

AI’s tireless capacity to investigate data, combined with its growing ability to develop authoritative-sounding narratives, expands the potential to elevate fabricated or bias-confirming errors into the collective public consciousness.

What’s next?

If you read the missives of AI luminaries, it would appear that society is on the cusp of superintelligence, which will transform every vexing societal conundrum into a trivial puzzle. While that’s highly unlikely, AI has certainly demonstrated promise in some health applications, despite its limitations. Unfortunately, it’s now being rapidly deployed sector-wide, even in areas where it has no prior track record.

This speed may leave us little time to reflect on the accountability needed for safe deployment. Sycophancy, hallucination, and the black box of AI are non-trivial challenges when conjoined with existing biases in health research. If people can’t easily understand the inner workings of current AI tools (often comprising up to 1.8 trillion parameters), they will not be able to understand the process of future, more complex versions (using over 5 trillion parameters).

History shows that most technological leaps forward are double-edged swords. Electronic health records increased the ability of clinicians to improve care coordination and aggregate data on population health, but they have eroded doctor-patient interactions and have become a source of physician burnout. The recent proliferation of telemedicine has expanded access to care, but it has also promoted lower-quality interactions with no physical examination.

The use of AI in health policy and research is no different. Wisely deployed, it could transform the health sector, leading to healthier populations and unfathomable breakthroughs (for example, by accelerating drug discovery). But without embedding it in new professional norms and practices, it has the potential to generate countless flawed leads and falsehoods.

Here are some potential solutions we see to the AI and health replicability crisis:

  • Clinical-specific models capable of admitting uncertainty in their outputs
  • Greater transparency, requiring disclosure of AI model use in research
  • Training for researchers, clinicians, and journalists on how to evaluate and stress-test AI-derived conclusions
  • Pre-registered hypotheses and analysis plans before using AI tools
  • AI audit trails
  • Specific AI global prompts that limit sycophantic tendencies across user queries

Regardless of the solutions deployed, we need to solve the failure points described here to fully realize the potential of AI for use in health research. The public, AI companies, and health researchers must be active participants in this journey. After all, in science, not everyone can be right.

Amit Chandra is an emergency physician and global health policy specialist based in Washington, DC. He is an adjunct professor of global health at Georgetown University’s School of Health, where he has explored AI solutions for global health challenges since 2021.

Luke Shors is an entrepreneur who focuses on energy, climate, and global health. He is the co-founder of the sustainability company Capture6 and previously worked on topics including computer vision and blockchain. 

When sycophancy and bias meet medicine Read More »

openai-looks-for-its-“google-chrome”-moment-with-new-atlas-web-browser

OpenAI looks for its “Google Chrome” moment with new Atlas web browser

That means you can use ChatGPT to search through your bookmarks or browsing history using human-parsable language prompts. It also means you can bring up a “side chat” next to your current page and ask questions that rely on the context of that specific page. And if you want to edit a Gmail draft using ChatGPT, you can now do that directly in the draft window, without the need to copy and paste between a ChatGPT window and an editor.

When typing in a short search prompt, Atlas will, by default, reply as an LLM, with written answers with embedded links to sourcing where appropriate (à la OpenAI’s existing search function). But the browser will also provide tabs with more traditional lists of links, images, videos, or news like those you would get from a search engine without LLM features.

Let us do the browsing

To wrap up the livestreamed demonstration, the OpenAI team showed off Atlas’ Agent Mode. While the “preview mode” feature is only available to ChatGPT Plus and Pro subscribers, research lead Will Ellsworth said he hoped it would eventually help users toward “an amazing tool for vibe life-ing” in the same way that LLM coding tools have become tools for “vibe coding.”

To that end, the team showed the browser taking planning tasks written in a Google Docs table and moving them over to the task management software Linear over the course of a few minutes. Agent Mode was also shown taking the ingredients list from a recipe webpage and adding them directly to the user’s Instacart in a different tab (though the demo Agent stopped before checkout to get approval from the user).

OpenAI looks for its “Google Chrome” moment with new Atlas web browser Read More »

youtube’s-likeness-detection-has-arrived-to-help-stop-ai-doppelgangers

YouTube’s likeness detection has arrived to help stop AI doppelgängers

AI content has proliferated across the Internet over the past few years, but those early confabulations with mutated hands have evolved into synthetic images and videos that can be hard to differentiate from reality. Having helped to create this problem, Google has some responsibility to keep AI video in check on YouTube. To that end, the company has started rolling out its promised likeness detection system for creators.

Google’s powerful and freely available AI models have helped fuel the rise of AI content, some of which is aimed at spreading misinformation and harassing individuals. Creators and influencers fear their brands could be tainted by a flood of AI videos that show them saying and doing things that never happened—even lawmakers are fretting about this. Google has placed a large bet on the value of AI content, so banning AI from YouTube, as many want, simply isn’t happening.

Earlier this year, YouTube promised tools that would flag face-stealing AI content on the platform. The likeness detection tool, which is similar to the site’s copyright detection system, has now expanded beyond the initial small group of testers. YouTube says the first batch of eligible creators have been notified that they can use likeness detection, but interested parties will need to hand Google even more personal information to get protection from AI fakes.

Sneak Peek: Likeness Detection on YouTube.

Currently, likeness detection is a beta feature in limited testing, so not all creators will see it as an option in YouTube Studio. When it does appear, it will be tucked into the existing “Content detection” menu. In YouTube’s demo video, the setup flow appears to assume the channel has only a single host whose likeness needs protection. That person must verify their identity, which requires a photo of a government ID and a video of their face. It’s unclear why YouTube needs this data in addition to the videos people have already posted with their oh-so stealable faces, but rules are rules.

YouTube’s likeness detection has arrived to help stop AI doppelgängers Read More »

claude-code-gets-a-web-version—but-it’s-the-new-sandboxing-that-really-matters

Claude Code gets a web version—but it’s the new sandboxing that really matters

Now, it can instead be given permissions for specific file system folders and network servers. That means fewer approval steps, but it’s also more secure overall against prompt injection and other risks.

Anthropic’s demo video for Claude Code on the web.

According to Anthropic’s engineering blog, the new network isolation approach only allows Internet access “through a unix domain socket connected to a proxy server running outside the sandbox. … This proxy server enforces restrictions on the domains that a process can connect to, and handles user confirmation for newly requested domains.” Additionally, users can customize the proxy to set their own rules for outgoing traffic.

This way, the coding agent can do things like fetch npm packages from approved sources, but without carte blanche for communicating with the outside world, and without badgering the user with constant approvals.

For many developers, these additions are more significant than the availability of web or mobile interfaces. They allow Claude Code agents to operate more independently without as many detailed, line-by-line approvals.

That’s more convenient, but it’s a double-edged sword, as it will also make code review even more important. One of the strengths of the too-many-approvals approach was that it made sure developers were still looking closely at every little change. Now it might be a little bit easier to miss Claude Code making a bad call.

The new features are available in beta now as a research preview, and they are available to Claude users with Pro or Max subscriptions.

Claude Code gets a web version—but it’s the new sandboxing that really matters Read More »