Author name: Tim Belzer

reports-suggest-apple-is-already-pulling-back-on-the-iphone-air

Reports suggest Apple is already pulling back on the iPhone Air

Apple’s iPhone Air was the company’s most interesting new iPhone this year, at least insofar as it was the one most different from previous iPhones. We came away impressed by its size and weight in our review. But early reports suggest that its novelty might not be translating into sales success.

A note from analyst Ming-Chi Kuo, whose supply chain sources are often accurate about Apple’s future plans, said yesterday that demand for the iPhone Air “has fallen short of expectations” and that “both shipments and production capacity” were being scaled back to account for the lower-than-expected demand.

Kuo’s note is backed up by reports from other analysts at Mizuho Securities (via MacRumors) and Nikkei Asia. Both of these reports say that demand for the iPhone 17 and 17 Pro models remains strong, indicating that this is just a problem for the iPhone Air and not a wider slowdown caused by tariffs or other external factors.

The standard iPhone, the regular-sized iPhone Pro, and the big iPhone Pro have all been mainstays in Apple’s lineup, but the company has had a harder time coming up with a fourth phone that sells well enough to stick around. The small-screened iPhone mini and the large-screened iPhone Plus were each discontinued after two generations.

Reports suggest Apple is already pulling back on the iPhone Air Read More »

tesla-profits-fall-37%-in-q3-despite-healthy-sales

Tesla profits fall 37% in Q3 despite healthy sales

Tesla reported its financial results for the third quarter of 2025 this afternoon. Earlier this month, we learned that the electric vehicle manufacturer had a pretty good Q3 in terms of sales, which grew by 7.3 percent year over year and cleared out tens of thousands of cars from inventory in the process. However, that hasn’t translated into greater profitability.

Even though revenues grew by 12 percent to $28 billion compared to the same period last year, Tesla’s operating expenses grew by 50 percent. As a result, its operating margin halved to just 5.8 percent. And so its profit for the quarter fell by 37 percent to $1.4 billion.

Some growth in revenue came from its battery and solar division; this increased by 44 percent to $3.4 billion compared to Q3 2024. Services—including the Supercharger network, which is now open to an increasing number of other makes of EV—also grew, increasing by 25 percent to $3.4 billion. EV deliveries increased by 7 percent to 497,099, most of which were the Model 3 sedan and Model Y crossover. Automotive revenues grew slightly less, increasing 6 percent year over year to $21.2 billion.

Q3 saw a bigger profit decline than last quarter, and the first quarter wasn’t great either, but despite that, the automaker isn’t in much danger of falling behind on the rent. Free cash flow grew by 46 percent, and between cash, cash equivalents, and investments at the end of September, Tesla had $41.6 billion with which to pay for its future plans.

Tesla profits fall 37% in Q3 despite healthy sales Read More »

general-motors-will-integrate-ai-into-its-cars,-plus-new-hands-free-assist

General Motors will integrate AI into its cars, plus new hands-free assist

I asked Dave Richardson, GM’s SVP of software, how the company will avoid the enshittification of vehicles as it integrates more AI.

“There’s a lot of hype around AI right now,” he told me. “But there’s also practical use. I’ve been trying to focus the company on practical use cases. I think there’s a lot of pretty compelling things we can do to try to add real value.”

He gave some examples, such as a car knowing you have a meeting and setting the navigation appropriately or knowing that you’re going on a road trip, so it should queue up the appropriate media for your kids to stream in the back seat.

While the company is using Gemini at first, it eventually plans to have its own model on board. “With advanced processing in the car, we can handle interference on board so that it works in low-data-connection areas,” Richardson said.

Ultimately, GM will deploy its own LLM that knows about the car and is limited in overall parameters, Richardson told me. It won’t need to rely on the cloud to operate, increasing responsiveness in the car and keeping personal information with you, he said.

There are reasons to be skeptical, of course. One of my biggest concerns is how much driver data the car will collect. One reason GM doesn’t offer Android Auto or Apple CarPlay, the company has said, is that it wants to protect customer data. The owner must consent to any data sharing, GM said.

And although GM says it has made some internal changes to protect customer data, there have been some very public instances of the company selling data. “Data privacy and security is priority one for us,” Richardson told me about his work at GM. He said he has hired people specifically tasked with ensuring that customer data protection frameworks are in place.

“We have no interest in selling that data to third parties. When we think about data, whether it’s for Super Cruise or the AI, it’s really for us to develop the product and make it better. We don’t want to sell that data as the product itself,” he said.

I believe there’s space for a privacy-focused automaker, and while I’m not sure whether that will be GM, I hope that privacy and data protection are as important to the company in the future as it says it is today.

As for consumers wanting AI in their vehicles? GM thinks they do.

General Motors will integrate AI into its cars, plus new hands-free assist Read More »

when-sycophancy-and-bias-meet-medicine

When sycophancy and bias meet medicine


Biased, eager-to-please models threaten health research replicability and trust.

Once upon a time, two villagers visited the fabled Mullah Nasreddin. They hoped that the Sufi philosopher, famed for his acerbic wisdom, could mediate a dispute that had driven a wedge between them. Nasreddin listened patiently to the first villager’s version of the story and, upon its conclusion, exclaimed, “You are absolutely right!” The second villager then presented his case. After hearing him out, Nasreddin again responded, “You are absolutely right!” An observant bystander, confused by Nasreddin’s proclamations, interjected, “But Mullah, they can’t both be right.” Nasreddin paused, regarding the bystander for a moment before replying, “You are absolutely right, too!”

In late May, the White House’s first “Make America Healthy Again” (MAHA) report was criticized for citing multiple research studies that did not exist. Fabricated citations like these are common in the outputs of generative artificial intelligence based on large language models, or LLMs. LLMs have presented plausible-sounding sources, catchy titles, or even false data to craft their conclusions. Here, the White House pushed back on the journalists who first broke the story before admitting to “minor citation errors.”

It is ironic that fake citations were used to support a principal recommendation of the MAHA report: addressing the health research sector’s “replication crisis,” wherein scientists’ findings often cannot be reproduced by other independent teams.

Yet the MAHA report’s use of phantom evidence is far from unique. Last year, The Washington Post reported on dozens of instances in which AI-generated falsehoods found their way into courtroom proceedings. Once uncovered, lawyers had to explain to judges how fictitious cases, citations, and decisions found their way into trials.

Despite these widely recognized problems, the MAHA roadmap released last month directs the Department of Health and Human Services to prioritize AI research to “…assist in earlier diagnosis, personalized treatment plans, real-time monitoring, and predictive interventions…” This breathless rush to embed AI in so many aspects of medicine could be forgiven if we believe that the technology’s “hallucinations” will be easy to fix through version updates. But as the industry itself acknowledges, these ghosts in the machine may be impossible to eliminate.

Consider the implications of accelerating AI use in health research for clinical decision making. Beyond the problems we’re seeing here, using AI in research without disclosure could create a feedback loop, supercharging the very biases that helped motivate its use. Once published, “research” based on false results and citations could become part of the datasets used to build future AI systems. Worse still, a recently published study highlights an industry of scientific fraudsters who could deploy AI to make their claims seem more legitimate.

In other words, a blind adoption of AI risks a downward spiral, where today’s flawed AI outputs become tomorrow’s training data, exponentially eroding research quality.

Three prongs of AI misuse

The challenge AI poses is threefold: hallucination, sycophancy, and the black box conundrum. Understanding these phenomena is critical for research scientists, policymakers, educators, and everyday citizens. Unaware, we risk vulnerability to deception as AI systems are increasingly deployed to shape diagnoses, insurance claims, health literacy, research, and public policy.

Here’s how hallucination works: When a user inputs a query into an AI tool such as ChatGPT or Gemini, the model evaluates the input and generates a string of words that is statistically likely to make sense based on its training data. Current AI models will complete this task even if their training data is incomplete or biased, filling in the blanks regardless of their ability to answer. These hallucinations can take the form of nonexistent research studies, misinformation, or even clinical interactions that never happened. LLMs’ emphasis on producing authoritative-sounding language shrouds their false outputs in a facsimile of truth.

And as human model trainers fine-tune generative AI responses, they tend to optimize and reward the AI system responses that favor their prior beliefs, leading to sycophancy. Human bias, it appears, begets AI bias, and human users of AI then perpetuate the cycle. A consequence is that AIs skew toward favoring pleasing answers over truthful ones, often seeking to reinforce the bias of the query.

A recent illustration of this occurred in April, when OpenAI canceled a ChatGPT update for being too sycophantic after users demonstrated that it agreed too quickly and enthusiastically with the assumptions embedded in users’ queries. Sycophancy and hallucination often interact with each other; systems that aim to please will be more apt to fabricate data to reach user-preferred conclusions.

Correcting hallucinations, sycophancy, and other LLM mishaps is cumbersome because human observers can’t always determine how an AI platform arrived at its conclusions. This is the “black box” problem. Behind the probabilistic mathematics, is it even testing hypotheses? What methods did it use to derive an answer? Unlike traditional computer code or the rubric of scientific methodology, AI models operate through billions of computations. Looking at some well-structured outputs, it is easy to forget that the underlying processes are impenetrable to scrutiny and vastly different from a human’s approach to problem-solving.

This opacity can become dangerous when people can’t identify where computations went wrong, making it impossible to correct systematic errors or biases in the decision-making process. In health care, this black box raises questions about accountability, liability, and trust when neither physicians nor patients can explain the sequence of reasoning that leads to a medical intervention.

AI and health research

These AI challenges can exacerbate the existing sources of error and bias that creep into traditional health research publications. Several sources originate from the natural human motivation to find and publish meaningful, positive results. Journalists want to report on connections, e.g., that St. John’s Wort improves mood (it might). Nobody would want to publish an article with the results: “the supplement has no significant effect.”

The problem compounds when researchers use a study design to test not just a single hypothesis but many. One quirk of statistics-backed research is that testing more hypotheses in a single study raises the likelihood of uncovering a spurious coincidence.

AI has the potential to supercharge these coincidences through its relentless ability to test hypotheses across massive datasets. In the past, a research assistant could use an existing dataset to test 10 to 20 of the most likely hypotheses; now, that assistant can set an AI loose to test millions of likely or unlikely hypotheses without human supervision. That all but guarantees some of the results will meet the criteria for statistical significance, regardless of whether the data includes any real biological effects.

AI’s tireless capacity to investigate data, combined with its growing ability to develop authoritative-sounding narratives, expands the potential to elevate fabricated or bias-confirming errors into the collective public consciousness.

What’s next?

If you read the missives of AI luminaries, it would appear that society is on the cusp of superintelligence, which will transform every vexing societal conundrum into a trivial puzzle. While that’s highly unlikely, AI has certainly demonstrated promise in some health applications, despite its limitations. Unfortunately, it’s now being rapidly deployed sector-wide, even in areas where it has no prior track record.

This speed may leave us little time to reflect on the accountability needed for safe deployment. Sycophancy, hallucination, and the black box of AI are non-trivial challenges when conjoined with existing biases in health research. If people can’t easily understand the inner workings of current AI tools (often comprising up to 1.8 trillion parameters), they will not be able to understand the process of future, more complex versions (using over 5 trillion parameters).

History shows that most technological leaps forward are double-edged swords. Electronic health records increased the ability of clinicians to improve care coordination and aggregate data on population health, but they have eroded doctor-patient interactions and have become a source of physician burnout. The recent proliferation of telemedicine has expanded access to care, but it has also promoted lower-quality interactions with no physical examination.

The use of AI in health policy and research is no different. Wisely deployed, it could transform the health sector, leading to healthier populations and unfathomable breakthroughs (for example, by accelerating drug discovery). But without embedding it in new professional norms and practices, it has the potential to generate countless flawed leads and falsehoods.

Here are some potential solutions we see to the AI and health replicability crisis:

  • Clinical-specific models capable of admitting uncertainty in their outputs
  • Greater transparency, requiring disclosure of AI model use in research
  • Training for researchers, clinicians, and journalists on how to evaluate and stress-test AI-derived conclusions
  • Pre-registered hypotheses and analysis plans before using AI tools
  • AI audit trails
  • Specific AI global prompts that limit sycophantic tendencies across user queries

Regardless of the solutions deployed, we need to solve the failure points described here to fully realize the potential of AI for use in health research. The public, AI companies, and health researchers must be active participants in this journey. After all, in science, not everyone can be right.

Amit Chandra is an emergency physician and global health policy specialist based in Washington, DC. He is an adjunct professor of global health at Georgetown University’s School of Health, where he has explored AI solutions for global health challenges since 2021.

Luke Shors is an entrepreneur who focuses on energy, climate, and global health. He is the co-founder of the sustainability company Capture6 and previously worked on topics including computer vision and blockchain. 

When sycophancy and bias meet medicine Read More »

m5-ipad-pro-tested:-stop-me-if-you’ve-heard-this-one-before

M5 iPad Pro tested: Stop me if you’ve heard this one before


It’s a gorgeous tablet, but what does an iPad need with more processing power?

Apple’s 13-inch M5 iPad Pro. Credit: Andrew Cunningham

Apple’s 13-inch M5 iPad Pro. Credit: Andrew Cunningham

This year’s iPad Pro is what you might call a “chip refresh” or an “internal refresh.” These refreshes are what Apple generally does for its products for one or two or more years after making a larger external design change. Leaving the physical design alone preserves compatibility with the accessory ecosystem.

For the Mac, chip refreshes are still pretty exciting to me, because many people who use a Mac will, very occasionally, assign it some kind of task where they need it to work as hard and fast as it can, for an extended period of time. You could be a developer compiling a large and complex app, or you could be a podcaster or streamer editing or exporting an audio or video file, or maybe you’re just playing a game. The power and flexibility of the operating system, and first- and third-party apps made to take advantage of that power and flexibility, mean that “more speed” is still exciting, even if it takes a few years for that speed to add up to something users will consistently notice and appreciate.

And then there’s the iPad Pro. Especially since Apple shifted to using the same M-series chips that it uses in Macs, most iPad Pro reviews contain some version of “this is great hardware that is much faster than it needs to be for anything the iPad does.” To wit, our review of the M4 iPad Pro from May 2024:

Still, it remains unclear why most people would spend one, two, or even three thousand dollars on a tablet that, despite its amazing hardware, does less than a comparably priced laptop—or at least does it a little more awkwardly, even if it’s impressively quick and has a gorgeous screen.

Since then, Apple has announced and released iPadOS 26, an update that makes important and mostly welcome changes to how the tablet handles windowed multitasking, file transfers, and some other kinds of background tasks. But this is the kind of thing that isn’t even going to stress out an Apple M1, let alone a chip that’s twice as powerful.

All of this is to say: A chip refresh for an iPad is nice to have. This year’s model (still starting at $999 for the 11-inch tablet and $1,299 for the 13-inch) will also come with a handy RAM increase for many buyers, the first RAM boost that the base model iPad Pro has gotten in more than four years.

But without any other design changes or other improvements to hang its hat on, the fact is that chip refresh years for the iPad Pro only really improve a part of the tablet that needs the least amount of improvement. That doesn’t make them bad; who knows what the hardware requirements will be when iPadOS 30 adds some other batch of multitasking features. But it does mean these refreshes don’t feel particularly exciting or necessary; the most exciting thing about the M5 iPad Pro means you might be able to get a good deal on an M4 model as retailers clear out their stock. You aren’t going to notice the difference.

Design: M4 iPad Pro redux

The 13-inch M5 iPad Pro in its Magic Keyboard accessory with the Apple Pencil Pro attached. Credit: Andrew Cunningham

Lest we downplay this tablet’s design, the M4 version of the iPad Pro was the biggest change to the tablet since Apple introduced the modern all-screen design for the iPad Pro back in 2018. It wasn’t a huge departure, but it did introduce the iPad’s first OLED display, a thinner and lighter design, and a slightly improved Apple Pencil and updated range of accessories.

As with the 14-inch M5 MacBook Pro that Apple just launched, the easiest way to know how much you’ll like the iPad Pro depends on how you feel about screen technology (the iPad is, after all, mostly screen). If you care about the 120 Hz, high-refresh-rate ProMotion screen, the option to add a nano-texture display with a matte finish, and the infinite contrast and boosted brightness of Apple’s OLED displays, those are the best reasons to buy an iPad Pro. The $299/349 Magic Keyboard accessory for the iPad Pro also comes with backlit keys and a slightly larger trackpad than the equivalent $269/$319 iPad Air accessory.

If none of those things inspire passion in you, or if they’re not worth several hundred extra dollars to you—the nano-texture glass upgrade alone adds $700 to the price of the iPad Pro, because Apple only offers it on the 1TB and 2TB models—then the 11- and 13-inch iPads Air are going to give you a substantively identical experience. That includes compatibility with the same Apple Pencil accessory and support for all the same multitasking and Apple Intelligence features.

The M5 iPad Pro supports the same Apple Pencil Pro as the M4 iPad Pro, and the M2 and M3 iPad Air. Credit: Andrew Cunningham

One other internal change to the new iPad Pro, aside from the M5, is mostly invisible: Wi-Fi, Bluetooth, and Thread connectivity provided by the Apple N1 chip, and 5G cellular connectivity provided by the Apple C1X. Ideally, you won’t notice this swap at all, but it’s a quietly momentous change for Apple. Both of these chips cap several years of acquisitions and internal development, and further reduce Apple’s reliance on external chipmakers like Qualcomm and Broadcom, which has been one of the goals of Apple’s A- and M-series processors all along.

There’s one last change we haven’t really been able to adequately test in the handful of days we’ve had the tablet: new fast-charging support, either with Apple’s first-party Dynamic Power Adapter or any USB-C charger capable of providing 60 W or more of power. When using these chargers, Apple says the tablet’s battery can charge from 0 to 50 percent in 35 minutes. (Apple provides the same battery life estimates for the M5 iPads as the M4 models: 10 hours of Wi-Fi web usage, or nine hours of cellular web usage, for both the 13- and 11-inch versions of the tablet.)

Two Apple M5 chips, two RAM options

Apple sent us the 1TB version of the 13-inch iPad Pro to test, which means we got the fully enabled version of the M5: four high-performance CPU cores, six high-efficiency GPU cores, 10 GPU cores, a 16-core Neural Engine, and 16GB of RAM.

Apple’s Macs still offer individually configurable processor, storage, and RAM upgrades to users—generally buying one upgrade doesn’t lock you into buying a bunch of other stuff you don’t want or need (though there are exceptions for RAM configurations in some of the higher-end Macs). But for the iPads, Apple still ties the chip and the RAM you get to storage capacity. The 256GB and 512GB iPads get three high-performance CPU cores instead of four, and 12GB of RAM instead of 16GB.

For people who buy the 256GB and 512GB iPads, this amounts to a 50 percent increase in RAM capacity from the M1, M2, and M4 iPad Pro models, or the M1, M2, and M3 iPad Airs, all of which came with 8GB of RAM. High-end models stick with the same 16GB of RAM as before (no 24GB or 32GB upgrades here, though the M5 supports them in Macs). The ceiling is in the same place, but the floor has come up.

Given that iPadOS is still mostly running on tablets with 8GB or less of RAM, I don’t expect the jump from 8GB to 12GB to make a huge difference in the day-to-day experience of using the tablet, at least for now. If you connect your iPad to an external monitor that you use as an extended display, it might help keep more apps in memory at a time; it could help if you edit complex multi-track audio or video files or images, or if you’re trying to run some kind of machine learning or AI workflows locally. Future iPadOS versions could also require more than 8GB of memory for some features. But for now, the benefit exists mostly on paper.

As for benchmarks, the M5’s gains in the iPad are somewhat more muted than they are for the M5 MacBook Pro we tested. We observed a 10 or 15 percent improvement across single- and multi-core CPU tests and graphics benchmark improvements that mostly hovered in the 15 to 30 percent range. The Geekbench 6 Compute benchmark was one outlier, pointing to a 35 percent increase in GPU performance; it’s possible that GPU or rendering-heavy workloads benefit a little more from the new neural accelerators in the M5’s GPU cores than games do.

In the MacBook review, we observed that the M5’s CPU generally had higher peak power consumption than the M4. In the fanless iPad Pro, it’s likely that Apple has reined the chip in a little bit to keep it cool, which would explain why the iPad’s M5 doesn’t see quite the same gains.

The M5 and the 12GB RAM minimum help to put a little more distance between the M3 iPad Air and the Pros. Most iPad workloads don’t benefit in an obvious user-noticeable way from the extra performance or memory right now, but it’s something you can point to that makes the Pro more “pro” than the Air.

Changed hardware that doesn’t change much

The M5 iPad Pro is nice in the sense that “getting a little more for your money today than you could get for the same money two weeks ago” is nice. But it changes essentially nothing for potential iPad buyers.

I’m hard-pressed to think of anyone who would be well-served by the M5 iPad Pro who wouldn’t have been equally well-served by the M4 version. And if the M4 iPad Pro was already overkill for you, the M5 is just a little more so. Particularly if you have an M1 or M2 ; People with an A12X or A12Z version of the iPad Pro from 2018 or 2020 will benefit more, particularly if you’re multitasking a lot or running into limitations or RAM complaints from the apps you’re using.

But even with the iPadOS 26 update, it still seems like the capabilities of the iPad’s software lags behind the capabilities of the hardware by a few years. It’s to be expected, maybe, for an operating system that has to run on this M5 iPad Pro and a 7-year-old phone processor with 3GB of RAM.

I am starting to feel the age of the M1 MacBook Air I use, especially if I’m pushing multiple monitors with it or trying to exceed its 16GB RAM limit. The M1 iPad Air I have, on the other hand, feels like it just got an operating system that unlocks some of its latent potential. That’s the biggest problem with the iPad Pro, really—not that it’s a bad tablet, but that it’s still so much more tablet than you need to do what iPadOS and its apps can currently do.

The good

  • A fast, beautiful tablet that’s a pleasure to use.
  • The 120Hz ProMotion support and OLED display panel make this one of Apple’s best screens, period.
  • 256GB and 512GB models get a bump from 8GB to 12GB of memory.
  • Maintains compatibility with the same accessories as the M4 iPad Pro.

The bad

  • More iPad than pretty much anyone needs.
  • Passively cooled fanless Apple M5 can’t stretch its legs quite as much as the actively cooled Mac version.
  • Expensive accessories.

The ugly

  • All other hardware upgrades, including the matte nano-texture display finish, require a $600 upgrade to the 1TB version of the tablet.

Photo of Andrew Cunningham

Andrew is a Senior Technology Reporter at Ars Technica, with a focus on consumer tech including computer hardware and in-depth reviews of operating systems like Windows and macOS. Andrew lives in Philadelphia and co-hosts a weekly book podcast called Overdue.

M5 iPad Pro tested: Stop me if you’ve heard this one before Read More »

rocket-report:-china-launches-with-no-advance-warning;-europe’s-drone-ship

Rocket Report: China launches with no advance warning; Europe’s drone ship


Starlink, Kuiper, and the US military all saw additions to their mega-constellations this week.

SpaceX’s Starship descends toward the Indian Ocean at the conclusion of Flight 11. Credit: SpaceX

Welcome to Edition 8.15 of the Rocket Report! This year has been, at best, one of mixed results for SpaceX’s Starship program. There have been important steps forward, including the successful reuse of the rocket’s massive Super Heavy booster. Clearly, SpaceX is getting really good at launching and recovering the 33-engine booster stage. But Starship itself, part spacecraft and part upper stage, hasn’t fared as well—at least it hadn’t until the last couple of months. After four Starships were destroyed in flight and on the ground in the first half of 2025, the last two missions ended with pinpoint splashdowns in the Indian Ocean. The most recent mission this week was arguably the most successful yet for Starship, which returned to Earth with little damage, suggesting SpaceX’s improvements to the heat shield are working.

As always, we welcome reader submissions. If you don’t want to miss an issue, please subscribe using the box below (the form will not appear on AMP-enabled versions of the site). Each report will include information on small-, medium-, and heavy-lift rockets, as well as a quick look ahead at the next three launches on the calendar.

SpaceX vet will fly with Blue Origin. Hans Koenigsmann is one of SpaceX’s earliest, longest-tenured, and most-revered employees. He worked at Elon Musk’s space company for nearly two decades, rising to the role of vice president for mission assurance and safety before leaving SpaceX in 2021. He led the investigations into every Falcon rocket failure, mentored young engineers, and became a public face for SpaceX through numerous presentations and press conferences. And now he has announced he is going to space on a future suborbital flight on Blue Origin’s New Shepard vehicle, Ars reports.

Due diligence … Koenigsmann will fly to space alongside his friend Michaela “Michi” Benthaus as early as next month. She’s notable in her own right—a mountain biking accident in 2018 left her with a spinal cord injury, but she did not let this derail her from her dream. She will become the first wheelchair user to fly in space. Koenigsmann said one of his main concerns with the flight was safety, but meeting with Blue Origin engineers gave him confidence to climb aboard New Shepard. “When we met them, I asked a lot of technical questions on the safety side, and I feel like they answered the majority of them thoughtfully and correctly.” So, what’s it like for a long-time SpaceXer to work with a former competitor, Blue Origin? Read Eric Berger’s interview with Koenigsmann to learn more.

The easiest way to keep up with Eric Berger’s and Stephen Clark’s reporting on all things space is to sign up for our newsletter. We’ll collect their stories and deliver them straight to your inbox.

Sign Me Up!

Europe’s drone ship. The European Space Agency (ESA) has awarded a contract for the design of a reusable rocket stage recovery vessel to the Italian aerospace and defense systems company Ingegneria Dei Sistemi (IDS), European Spaceflight reports. The project is part of a broader contract awarded to the Italian rocket builder Avio for the development of a reusable rocket upper stage, which Ars reported on last month. The contract covers preliminary design work for the launch system and the ground system, and could be applied to a reusable evolution of Avio’s Vega family of rockets.

Looks familiar … On Wednesday, IDS announced that it had been awarded the contract to design the project’s recovery vessel, which falls under the systems ground segment. The company has subcontracted Italian naval systems consultancy Cetena and Norwegian shipbuilder Vard to assist with the project. An artist’s illustration of the vessel gives it a familiar look. It appears similar to the recovery ships that SpaceX used to attempt recovery of the Falcon 9 rocket’s payload fairings, with giant nets to catch the hardware falling from space under parachute. (submitted by EllPeaTea)

JAXA looks abroad. The Japanese space agency JAXA has selected Rocket Lab to launch a set of technology demonstration satellites on Electron rockets after continued delays with a Japanese launch vehicle, Space News reports. The agreement covers two launches from New Zealand, the first in December with JAXA’s 242-pound (110-kilogram) Rapid Innovative Payload Demonstration Satellite-4 (RAISE-4) technology demonstration satellite, and the second in early 2026 with a batch of eight smaller satellites for educational, ocean monitoring, and other demonstrations.

No more waiting … These satellites were supposed to launch on Japan’s solid-fueled Epsilon S rocket, but JAXA looked to another launch provider after lengthy delays with the Epsilon program. Epsilon S is an upgraded version of Japan’s Epsilon rocket, which has flown six times. The first flight of Epsilon S was originally expected in 2023, but back-to-back ground test failures of the vehicle’s second stage solid rocket motor have effectively grounded the rocket. Japanese officials are considering ditching the upgraded second stage design and going back to the original Epsilon configuration, but a launch is still at least a year away.

An update on a German launch startup. German rocket builder HyImpulse announced Thursday that it had secured $53 million (45 million euros) in new funding to continue developing its SL-1 rocket, European Spaceflight reports. HyImpulse said it will use the new capital to “drive forward the development and commercialization of the SL1 orbital rocket and expand its production capacities.” HyImpulse is one of a handful of serious European launch startups, having raised more than $86 million (74 million euros) since its foundation in 2018.

Still years away … The SL1 rocket will consist of three stages with hybrid propulsion, capable of delivering up to more than 1,300 pounds (600 kilograms) of payload to low-Earth orbit. The first flight of HyImpulse’s orbital rocket is scheduled for 2027. SL1 builds on the company’s SR75 suborbital rocket, which made its first test flight from Australia in 2024.

iRocket touts rapid build. Innovative Rocket Technologies Inc. (iRocket) reports a successful flight test of the company’s 2.75-inch (70-millimeter) diameter IRX-100 version of the Hydra 70 rocket system from a launch tube under its own power to exercise a range of motor and missile properties, Aviation Week & Space Technology Reports. The IRX-100 is iRocket’s version of the Hydra 70 short-range unguided missile primarily used on military helicopters. Asad Malik, iRocket’s CEO, wrote in a post on LinkedIn that the company designed and launched the rocket in just 30 days. “Speed, precision, and innovation are what define our team,” Malik wrote.

Pathfinder … The IRX-100 rocket launched from a desert location in California and reached an altitude of more than 12,000 feet, according to iRocket. We’ve reported on iRocket in several recent editions of the Rocket Report. In July, the company announced it was going public in a deal with a Special Purpose Acquisition Company founded by former Commerce Secretary Wilbur Ross. But the SPAC and iRocket itself appear to have little money. Company officials hope the IRX-100 might offer a short-term source of revenue through military sales. iRocket’s longer-term goals include the development of a reusable orbital-class rocket, named Shockwave.

SpaceX launches for Kuiper. After more than a week of launch delays, SpaceX launched a Falcon 9 rocket from Cape Canaveral Space Force Station, Florida, with two dozen of Amazon’s Project Kuiper broadband Internet satellites onboard Monday night, Spaceflight Now reports. The mission, dubbed Kuiper Falcon 03 or KF-03, faced several days of launch delays due to poor weather both at the Cape as well as offshore. This was the third and final Kuiper launch currently booked on SpaceX’s Falcon 9 rocket, and the sixth launch of operational Kuiper satellites overall. Amazon now has 153 of its planned 3,232 Kuiper satellites in orbit.

SDA, too … Two days later, SpaceX launched a different Falcon 9 rocket from Vandenberg Space Force Base, California, to add 21 satellites to the Space Development Agency’s burgeoning low-Earth orbit constellation, Spaceflight Now reports. These satellites were built by Lockheed Martin, and they will join a batch of 21 similar spacecraft manufactured by York Space Systems launched last month. The satellites form the foundation for the Pentagon’s proliferated missile tracking and data relay network.

China launches another mysterious satellite. China conducted an orbital launch Monday with no apparent advance indication, successfully sending the Shiyan-31 remote sensing test satellite into orbit, Space News reports. The mission lifted off aboard a Long March 2D rocket from Jiuquan Satellite Launch Center in northwestern China. The Long March 2D can deliver up to 3.5 metric tons (7,700 pounds) of payload to low-Earth orbit. Shiyan-31 is believed to have an optical surveillance mission, and US tracking data indicated it was flying in an orbit about 300 miles (500 kilometers) above the Earth.

Surprise! … What was unusual about this launch was the fact that China did not publicize it in advance. Like most spacefaring nations, China typically issues airspace and maritime warning notices for airplanes and ships to steer clear of downrange zones where rocket debris may fall. No such warnings were released for this launch.

Starship flirts with perfection. SpaceX closed a troubled but instructive chapter in its Starship rocket program Monday with a near-perfect test flight that carried the stainless steel spacecraft halfway around the world from South Texas to the Indian Ocean, Ars reports. This was the 11th full-scale test flight of the Super Heavy booster and Starship upper stage, and it was arguably the most successful Starship test flight to date. It comes after a rough start to the year with a series of Starship failures and explosions that set the program back by at least six months.

Close to pristine … This time, Starship came back through the atmosphere with little sign of visible damage. The previous test flight in August also nailed its splashdown in the Indian Ocean, but it came down with a banged-up heat shield. This was the final flight of the second generation of Starship, called Starship V2. SpaceX plans to debut the larger, more powerful Starship V3 configuration in early 2026. If all goes well, SpaceX could be in position to attempt to recover Starship on land next year.

Orion’s other options. The Orion spacecraft and Space Launch System rocket have been attached at the hip for the better part of two decades. The big rocket lifts, the smaller spacecraft flies, and Congress keeps the money rolling in. But now there are signs that the twain may, in the not-too-distant future, split, Ars reports. This is because Lockheed Martin has begun to pivot toward a future in which the Orion spacecraft—thanks to increasing reusability, a focus on cost, and openness to flying on different rockets—fits into commercial space applications. In interviews, company officials said that if NASA wanted to buy Orion missions as a “service,” rather than owning and operating the spacecraft, they were ready to work with the space agency.

Staying power This represents a significant change. Since the US Congress called for the creation of the Space Launch System rocket a decade and a half ago, Orion and this rocket have been discussed in tandem, forming the backbone of an expendable architecture that would launch humans to the Moon and return them to Earth inside Orion. But time is running out for the uber-expensive SLS rocket, with differing proposals from the Trump administration and Congress to terminate the program after either two or perhaps four more flights. This appears to be one reason Lockheed is exploring alternative launch vehicles for Orion. If the spacecraft is going to be competitive on price, it needs a rocket that does not cost more than $2 billion per launch. Any near-term plan to send astronauts to the Moon will still require Orion.

Doubling up at Vandenberg. The Department of the Air Force has approved SpaceX’s plans to launch up to 100 missions per year from Vandenberg Space Force Base in California, Ars reports. This would continue the tectonic turnaround at the spaceport on California’s Central Coast. Five years ago, Vandenberg hosted just a single orbital launch. This year’s number stands at 51 orbital flights, or 53 launches if you count a pair of Minuteman missile tests, the most in a single calendar year at Vandenberg since the early 1970s. Military officials have now authorized SpaceX to double its annual launch rate at Vandenberg from 50 to 100, with up to 95 missions using the Falcon 9 rocket and up to five launches of the larger Falcon Heavy.

No big rush … There’s more to the changes at Vandenberg than launching additional rockets. The authorization gives SpaceX the green light to redevelop Space Launch Complex 6 (SLC-6) to support Falcon 9 and Falcon Heavy missions. SpaceX plans to demolish unneeded structures at SLC-6 (pronounced “Slick 6”) and construct two new landing pads for Falcon boosters on a bluff overlooking the Pacific just south of the pad. SLC-6 would become the West Coast home for Falcon Heavy, but SpaceX currently has no confirmed contracts to fly the heavy-lifter from Vandenberg.

Next three launches

Oct. 18: Falcon 9 | Starlink 11-19 | Vandenberg Space Force Base, California | 23: 46 UTC

Oct. 19: Kinetica 1 | Unknown Payload | Jiuquan Satellite Launch Center, China | 03: 30 UTC

Oct. 19: Falcon 9 | Starlink 10-17 | Cape Canaveral Space Force Station, Florida | 14: 52 UTC

Photo of Stephen Clark

Stephen Clark is a space reporter at Ars Technica, covering private space companies and the world’s space agencies. Stephen writes about the nexus of technology, science, policy, and business on and off the planet.

Rocket Report: China launches with no advance warning; Europe’s drone ship Read More »

rfk-jr.’s-maha-wants-to-make-chemtrail-conspiracy-theories-great-again

RFK Jr.’s MAHA wants to make chemtrail conspiracy theories great again

A prominent voice in the Make America Healthy Again movement is pushing for health secretary and anti-vaccine activist Robert F. Kennedy Jr. to make the topic of chemtrail conspiracy theories a federal priority, according to a report by KFF News.

KFF obtained a memo, written by MAHA influencer Gray Delany in July, presenting the topic to Calley Means, a White House health advisor. The memo lays out a series of unsubstantiated and far-fetched claims that academic researchers and federal agencies are secretively spreading toxic substances from airplanes, poisoning Americans, and spurring large-scale weather events, such as the devastating flooding in Texas last summer.

“It is unconscionable that anyone should be allowed to spray known neurotoxins and environmental toxins over our nation’s citizens, their land, food and water supplies,” Delany writes in the memo.

Daniel Swain, a climate scientist at the University of California, Los Angeles, told KFF that the memo presents claims that are false and, in some cases, physically impossible. “That is a pretty shocking memo,” he said. “It doesn’t get more tinfoil hat. They really believe toxins are being sprayed.”

Delany ends the memo with recommendations for federal agencies: form a joint task force to address this alleged geoengineering, host a roundtable on the topic, include the topic in the MAHA commission report, and publicly address the health and environmental harms.

It remains unclear if Kennedy, Means, or federal agencies are following up on Delany’s suggestions. Department of Health and Human Services spokesperson Emily Hilliard told KFF that “HHS does not comment on future or potential policy decisions and task forces.”

However, one opportunity has already been missed: The MAHA Commission released its “Make Our Children Healthy Again” report on September 9, along with a strategy document. Neither document mentions any of the topics raised in Delany’s memo.

RFK Jr.’s MAHA wants to make chemtrail conspiracy theories great again Read More »

vaginal-condition-treatment-update:-men-should-get-treated,-too

Vaginal condition treatment update: Men should get treated, too

For some cases of bacterial vaginosis, treatment should include a package deal, doctors now say.

The American College of Obstetricians & Gynecologists (ACOG) updated its clinical guidance Friday to fit with recent data indicating that treatment for recurring bacterial vaginosis (BV) in women is significantly more effective if their male partners are also treated at the same time—with both an oral antibiotic and an antibiotic cream directly onto the potentially offending member.

“Partner therapy offers us another avenue for hopefully preventing recurrence and helping people feel better faster,” Christopher Zahn, chief of clinical practice and health equity and quality at ACOG, said in a statement.

BV is a common condition affecting nearly 30 percent of women worldwide. Still, it’s potentially stigmatizing and embarrassing, with symptoms including itching, burning, a concerning fishy smell, and vaginal discharge that can be green or gray. With symptoms like this, BV is often described as an infection—but it’s actually not. BV is an imbalance in the normal bacterial communities that inhabit the vagina—a situation called dysbiosis.

This imbalance can be especially difficult to correct; of the women who suffer with BV, up to 66 percent will end up having the condition recur after treatment.

BV symptoms are “incredibly uncomfortable and disrupt people’s daily lives,” Zahn said, and that discomfort “becomes compounded by frustration when this condition comes back repeatedly.”

Firm recommendation

Studies in recent years have started to expose the reasons behind recurrence. Though again, BV is an imbalance, it has the profile of a sexually transmitted infection, with links to new sexual partners and similar incubation periods. Going further, microbial communities of penises can silently harbor the bacterial species linked to BV, and penile microbial communities can be predictive of BV risk in partners.

Vaginal condition treatment update: Men should get treated, too Read More »

ring-cameras-are-about-to-get-increasingly-chummy-with-law-enforcement

Ring cameras are about to get increasingly chummy with law enforcement


Amazon’s Ring partners with company whose tech has reportedly been used by ICE.

Ring’s Outdoor Cam Pro. Credit: Amazon

Law enforcement agencies will soon have easier access to footage captured by Amazon’s Ring smart cameras. In a partnership announced this week, Amazon will allow approximately 5,000 local law enforcement agencies to request access to Ring camera footage via surveillance platforms from Flock Safety. Ring cooperating with law enforcement and the reported use of Flock technologies by federal agencies, including US Immigration and Customs Enforcement (ICE), has resurfaced privacy concerns that have followed the devices for years.

According to Flock’s announcement, its Ring partnership allows local law enforcement members to use Flock software “to send a direct post in the Ring Neighbors app with details about the investigation and request voluntary assistance.” Requests must include “specific location and timeframe of the incident, a unique investigation code, and details about what is being investigated,” and users can look at the requests anonymously, Flock said.

“Any footage a Ring customer chooses to submit will be securely packaged by Flock and shared directly with the requesting local public safety agency through the FlockOS or Flock Nova platform,” the announcement reads.

Flock said its local law enforcement users will gain access to Ring Community Requests in “the coming months.”

A flock of privacy concerns

Outside its software platforms, Flock is known for license plate recognition cameras. Flock customers can also search footage from Flock cameras using descriptors to find people, such as “man in blue shirt and cowboy hat.” Besides law enforcement agencies, Flock says 6,000 communities and 1,000 businesses use their products.

For years, privacy advocates have warned against companies like Flock.

This week, US Sen. Ron Wyden (D-Ore.) sent a letter [PDF] to Flock CEO Garrett Langley saying that ICE’s Homeland Security Investigations (HSI), the Secret Service, and the US Navy’s Criminal Investigative Service have had access to footage from Flock’s license plate cameras.

“I now believe that abuses of your product are not only likely but inevitable and that Flock is unable and uninterested in preventing them,” Wyden wrote.

In August, Jay Stanley, senior policy analyst for the ACLU Speech, Privacy, and Technology Project, wrote that “Flock is building a dangerous, nationwide mass-surveillance infrastructure.” Stanley pointed to ICE using Flock’s network of cameras, as well as Flock’s efforts to build a people lookup tool with data brokers.

Matthew Guariglia, senior policy analyst at the Electronic Frontier Foundation (EFF), told Ars via email that Flock is a “mass surveillance tool” that “has increasingly been used to spy on both immigrants and people exercising their First Amendment-protected rights.”

Flock has earned this reputation among privacy advocates through its own cameras, not Ring’s.

An Amazon spokesperson told Ars Technica that only local public safety agencies will be able to make Community Requests via Flock software, and that requests will also show the name of the agency making the request.

A Flock spokesperson told Ars:

Flock does not currently have any contracts with any division of [the US Department of Homeland Security], including ICE. The Ring Community Requests process through Flock is only available for local public safety agencies for specific, active investigations. All requests are time and geographically-bound. Ring users can choose to share relevant footage or ignore the request.

Flock’s rep added that all activity within FlockOS and Flock Nova is “permanently recorded in a comprehensive CJIS-compliant audit trail for unalterable custody tracking,” referring to a set of standards created by the FBI’s Criminal Justice Information Services division.

But there’s still concern that federal agencies will end up accessing Ring footage through Flock. Guariglia told Ars:

Even without formal partnerships with federal authorities, data from these surveillance companies flow to agencies like ICE through local law enforcement. Local and state police have run more than 4,000 Flock searches on behalf of federal authorities or with a potential immigration focus, reporting has found. Additionally, just this month, it became clear that Texas police searched 83,000 Flock cameras in an attempt to prosecute a woman for her abortion and then tried to cover it up.

Ring cozies up to the law

This week’s announcement shows Amazon, which acquired Ring in 2018, increasingly positioning its consumer cameras as a law enforcement tool. After years of cops using Ring footage, Amazon last year said that it would stop letting police request Ring footage—unless it was an “emergency”—only to reverse course about 18 months later by allowing police to request Ring footage through a Flock rival, Axon.

While announcing Ring’s deals with Flock and Axon, Ring founder and CEO Jamie Siminoff claimed that the partnerships would help Ring cameras keep neighborhoods safe. But there’s doubt as to whether people buy Ring cameras to protect their neighborhood.

“Ring’s new partnership with Flock shows that the company is more interested in contributing to mounting authoritarianism than servicing the specific needs of their customers,” Guariglia told Ars.

Interestingly, Ring initiated conversations about a deal with Flock, Langely told CNBC.

Flock says that its cameras don’t use facial recognition, which has been criticized for racial biases. But local law enforcement agencies using Flock will soon have access to footage from Ring cameras with facial recognition. In a conversation with The Washington Post this month, Calli Schroeder, senior counsel at the consumer advocacy and policy group Electronic Privacy Information Center, described the new feature for Ring cameras as “invasive for anyone who walks within range of” a Ring doorbell, since they likely haven’t consented to facial recognition being used on them.

Amazon, for its part, has mostly pushed the burden of ensuring responsible facial recognition use to its customers. Schroeder shared concern with the Post that Ring’s facial recognition data could end up being shared with law enforcement.

Some people who are perturbed about Ring deepening its ties with law enforcement have complained online.

“Inviting big brother into the system. Screw that,” a user on the Ring subreddit said this week.

Another Reddit user said: “And… I’m gone. Nope, NO WAY IN HELL. Goodbye, Ring. I’ll be switching to a UniFi[-brand] system with 100 percent local storage. You don’t get my money any more. This is some 1984 BS …”

Privacy concerns are also exacerbated by Ring’s past, as the company has previously failed to meet users’ privacy expectations. In 2023, Ring agreed to pay $5.8 million to settle claims that employees illegally spied on Ring customers.

Amazon and Flock say their collaboration will only involve voluntary customers and local enforcement agencies. But there’s still reason to be concerned about the implications of people sending doorbell and personal camera footage to law enforcement via platforms that are reportedly widely used by federal agencies for deportation purposes. Combined with the privacy issues that Ring has already faced for years, it’s not hard to see why some feel that Amazon scaling up Ring’s association with any type of law enforcement is unacceptable.

And it appears that Amazon and Flock would both like Ring customers to opt in when possible.

“It will be turned on for free for every customer, and I think all of them will use it,” Langely told CNBC.

Photo of Scharon Harding

Scharon is a Senior Technology Reporter at Ars Technica writing news, reviews, and analysis on consumer gadgets and services. She’s been reporting on technology for over 10 years, with bylines at Tom’s Hardware, Channelnomics, and CRN UK.

Ring cameras are about to get increasingly chummy with law enforcement Read More »

dead-ends-is-a-fun,-macabre-medical-history-for-kids

Dead Ends is a fun, macabre medical history for kids


flukes, flops, and failures

Ars chats with co-authors Lindsey Fitzharris and Adrian Teal about their delightful new children’s book.

In 1890, a German scientist named Robert Koch thought he’d invented a cure for tuberculosis, a substance derived from the infecting bacterium itself that he dubbed Tuberculin. His substance didn’t actually cure anyone, but it was eventually widely used as a diagnostic skin test. Koch’s successful failure is just one of the many colorful cases featured in Dead Ends! Flukes, Flops, and Failures that Sparked Medical Marvels, a new nonfiction illustrated children’s book by science historian Lindsey Fitzharris and her husband, cartoonist Adrian Teal.

A noted science communicator with a fondness for the medically macabre, Fitzharris published a biography of surgical pioneer Joseph Lister, The Butchering Art, in 2017—a great, if occasionally grisly, read. She followed up with 2022’s  The Facemaker: A Visionary Surgeon’s Battle to Mend the Disfigured Soldiers of World War I, about a WWI surgeon named Harold Gillies who rebuilt the faces of injured soldiers.

And in 2020, she hosted a documentary for the Smithsonian Channel, The Curious Life and Death Of…, exploring famous deaths, ranging from drug lord Pablo Escobar to magician Harry Houdini. Fitzharris performed virtual autopsies, experimented with blood samples, interviewed witnesses, and conducted real-time demonstrations in hopes of gleaning fresh insights. For his part, Teal is a well-known caricaturist and illustrator, best known for his work on the British TV series Spitting Image. His work has also appeared in The Guardian and the Sunday Telegraph, among other outlets.

The couple decided to collaborate on children’s books as a way to combine their respective skills. Granted, “[The market for] children’s nonfiction is very difficult,” Fitzharris told Ars. “It doesn’t sell that well in general. It’s very difficult to get publishers on board with it. It’s such a shame because I really feel that there’s a hunger for it, especially when I see the kids picking up these books and loving it. There’s also just a need for it with the decline in literacy rates. We need to get people more engaged with these topics in ways that go beyond a 30-second clip on TikTok.”

Their first foray into the market was 2023’s Plague-Busters! Medicine’s Battles with History’s Deadliest Diseases, exploring “the ickiest illnesses that have infected humans and affected civilizations through the ages”—as well as the medical breakthroughs that came about to combat those diseases. Dead Ends is something of a sequel, focusing this time on historical diagnoses, experiments, and treatments that were useless at best, frequently harmful, yet eventually led to unexpected medical breakthroughs.

Failure is an option

The book opens with the story of Robert Liston, a 19th-century Scottish surgeon known as “the fastest knife in the West End,” because he could amputate a leg in less than three minutes. That kind of speed was desirable in a period before the discovery of anesthetic, but sometimes Liston’s rapid-fire approach to surgery backfired. One story (possibly apocryphal) holds that Liston accidentally cut off the finger of his assistant in the operating theater as he was switching blades, then accidentally cut the coat of a spectator, who died of fright. The patient and assistant also died, so that operation is now often jokingly described as the only one with a 300 percent mortality rate, per Fitzharris.

Liston is the ideal poster child for the book’s theme of celebrating the role of failure in scientific progress. “I’ve always felt that failure is something we don’t talk about enough in the history of science and medicine,” said Fitzharris. “For everything that’s succeeded there’s hundreds, if not thousands, of things that’s failed. I think it’s a great concept for children. If you think that you’ve made mistakes, look at these great minds from the past. They’ve made some real whoppers. You are in good company. And failure is essential to succeeding, especially in science and medicine.”

“During the COVID pandemic, a lot of people were uncomfortable with the fact that some of the advice would change, but to me that was a comfort because that’s what you want to see scientists and doctors doing,” she continued. “They’re learning more about the virus, they’re changing their advice. They’re adapting. I think that this book is a good reminder of what the scientific process involves.”

The details of Liston’s most infamous case might be horrifying, but as Teal observes, “Comedy equals tragedy plus time.” One of the reasons so many of his patients died was because this was before the broad acceptance of germ theory and Joseph Lister’s pioneering work on antiseptic surgery. Swashbuckling surgeons like Liston prided themselves on operating in coats stiffened with blood—the sign of a busy and hence successful surgeon. Frederick Treves once observed that in the operating room, “cleanliness was out of place. It was considered to be finicking and affected. An executioner might as well manicure his nails before chopping off a head.”

“There’s always a lot of initial resistance to new ideas, even in science and medicine,” said Teal. “A lot of what we talk about is paradigm shifts and the difficulty of achieving [such a shift] when people are entrenched in their thinking. Galen was a hugely influential Roman doctor and got a lot of stuff right, but also got a lot of stuff wrong. People were clinging onto that stuff for centuries. You have misunderstanding compounded by misunderstanding, century after century, until somebody finally comes along and says, ‘Hang on a minute, this is all wrong.’”

You know… for kids

Writing for children proved to be a very different experience for Fitzharris after two adult-skewed science history books. “I initially thought children’s writing would be easy,” she confessed. “But it’s challenging to take these high-level concepts and complex stories about past medical movements and distill them for children in an entertaining and fun way.” She credits Teal—a self-described “man-child”—for taking her drafts and making them more child-friendly.

Teal’s clever, slightly macabre illustrations also helped keep the book accessible to its target audience, appealing to children’s more ghoulish side. “There’s a lot of gruesome stuff in this book,” Teal said. “Obviously it’s for kids, so you don’t want to go over the top, but equally, you don’t want to shy away from those details. I always say kids love it because kids are horrible, in the best possible way. I think adults sometimes worry too much about kids’ sensibilities. You can be a lot more gruesome than you think you can.”

The pair did omit some darker subject matter, such as the history of frontal lobotomies, notably the work of a neuroscientist named Walter Freeman, who operated an actual “lobotomobile.” For the authors, it was all about striking the right balance. “How much do you give to the kids to keep them engaged and interested, but not for it to be scary?” said Fitzharris. “We don’t want to turn people off from science and medicine. We want to celebrate the greatness of what we’ve achieved scientifically and medically. But we also don’t want to cover up the bad bits because that is part of the process, and it needs to be acknowledged.”

Sometimes Teal felt it just wasn’t necessary to illustrate certain gruesome details in the text—such as their discussion of the infamous case of Phineas Gage. Gage was a railroad construction foreman. In 1848, he was overseeing a rock blasting team when an explosion drove a three-foot tamping iron through his skull. “There’s a horrible moment when [Gage] leans forward and part of his brain drops out,” said Teal. “I’m not going to draw that, and I don’t need to, because it’s explicit in the text. If we’ve done a good enough job of writing something, that will put a mental picture in someone’s head.”

Miraculously, Gage survived, although there were extreme changes in his behavior and personality, and his injuries eventually caused epileptic seizures, one of which killed Gage in 1860. Gage became the index case for personality changes due to frontal lobe damage, and 50 years after his death, the case inspired neurologist David Ferrier to create brain maps based on his research into whether certain areas of the brain controlled specific cognitive functions.

“Sometimes it takes a beat before we get there,” said Fitzharris. “Science builds upon ideas, and it can take time. In the age of looking for instantaneous solutions, I think it’s important to remember that research needs to allow itself to do what it needs to do. It shouldn’t just be guided by an end goal. Some of the best discoveries that were made had no end goal in mind. And if you read Dead Ends, you’re going to be very happy that you live in 2025. Medically speaking, this is the best time. That’s really what Dead Ends is about. It’s a celebration of how far we’ve come.”

Photo of Jennifer Ouellette

Jennifer is a senior writer at Ars Technica with a particular focus on where science meets culture, covering everything from physics and related interdisciplinary topics to her favorite films and TV series. Jennifer lives in Baltimore with her spouse, physicist Sean M. Carroll, and their two cats, Ariel and Caliban.

Dead Ends is a fun, macabre medical history for kids Read More »

ai-#138-part-2:-watch-out-for-documents

AI #138 Part 2: Watch Out For Documents

As usual when things split, Part 1 is mostly about capabilities, and Part 2 is mostly about a mix of policy and alignment.

  1. The Quest for Sane Regulations. The GAIN Act and some state bills.

  2. People Really Dislike AI. They would support radical, ill-advised steps.

  3. Chip City. Are we taking care of business?

  4. The Week in Audio. Hinton talks to Jon Stewart, Klein to Yudkowsky.

  5. Rhetorical Innovation. How to lose the moral high ground.

  6. Water Water Everywhere. AI has many big issues. Water isn’t one of them.

  7. Read Jack Clark’s Speech From The Curve. It was a sincere, excellent speech.

  8. How One Other Person Responded To This Thoughtful Essay. Some aim to divide.

  9. A Better Way To Disagree. Others aim to work together and make things better.

  10. Voice Versus Exit. The age old question, should you quit your job at an AI lab?

  11. The Dose Makes The Poison. As little as 250 documents can poison an LLM.

  12. Aligning a Smarter Than Human Intelligence is Difficult. Techniques to avoid.

  13. You Get What You Actually Trained For. So ask what you actually train for.

  14. Messages From Janusworld. Do not neglect theory of mind.

  15. People Are Worried About AI Killing Everyone. A world-ending AI prompt?

  16. The Lighter Side. Introducing the museum of chart crimes.

Don’t let misaligned AI wipe out your GAIN AI Act.

It’s pretty amazing that it has come to this and we need to force this into the books.

The least you can do, before selling advanced AI chips to our main political adversary, is offer those same chips for sale to American firms on the same terms first. I predict there are at least three labs (OpenAI, Anthropic and xAI) that would each happily and directly buy everything you’re willing to sell at current market prices, and that’s not even including Oracle, Meta and Microsoft.

I’m not including Google and Amazon there because they’re trying to make their own chips, but make those calls too, cause more is more. I won’t personally buy in too much bulk, but call me too, there’s a good chance I’ll order me at least one H20 or even better B30A, as a treat.

Samuel Hammond: Glad to see this made it in.

So long as American companies are compute constrained, they should at the very least have a right of first refusal over chips going to our chief geopolitical adversary.

ARI: The Senate just passed the GAIN AI Act in the NDAA – a bill requiring chip makers to sell advanced AI chips to US firms before countries of concern. Big win for competitiveness & security.

In all seriousness, I will rest a lot easier if we can get the GAIN AI Act passed, as it will severely limit the amount of suicide we can commit with chip sales.

Marjorie Taylor-Greene says Trump is focusing on helping AI industry and crypto donors at the expense of his base and the needs of manufacturers.

California Governor Newsom vetoes the relatively strong AB 1064, an AI child safety bill that a16z lobbyists and allied usual suspects lobbied hard against, and signs another weaker child safety bill, SB 243. SB 243 requires chatbot operators have procedures to prevent the production of suicide or self-harm content and put in guardrails like referrals to suicide and crisis hotlines, and tell minor users every three hours that the AI is not human and to take a break.

There was a divide in industry over whether SB 243 was an acceptable alternative to AB 1064 or still something to fight, and a similar divide by child safety advocates over whether SB 243 was too timid to be worth supporting. I previously covered these bills briefly back in AI #110, when I said AB 1064 seemed like a bad idea and SB 243 seemed plausibly good but non-urgent.

For AB 1064, Newsom’s veto statement says he was worried it could result in unintentionally banning AI tool use by minors, echoing arguments by opposing lobbyists that it would ban educational tools.

Cristiano Lima-Strong: Over the past three months, the group has spent over $50,000 on more than 90 digital ads targeting California politics, according to a review of Meta’s political ads library.

Over two dozen of the ads specifically targeted AB1064, which the group said would “hurt classrooms” and block “the tools students and teachers need.” Several others more broadly warned against AI “red tape,” urging state lawmakers to “stand with Little Tech” and “innovators,” while dozens more took aim at another one of Bauer-Kahan’s AI bills.

TechNet has spent roughly $10,000 on over a dozen digital ads in California expressly opposing AB1064, with messages warning that it would “slam the brakes” on innovation and that if passed, “our teachers won’t be equipped to prepare students for the future.”

The Chamber of Progress and TechNet each registered nearly $200,000 in lobbying the California legislature the first half of this year, while CCIA spent $60,000 and the American Innovators Network doled out $40,000, according to a review of state disclosure filings. Each group was active on both SB243 and AB1064, among numerous other tech and AI bills.

One thing to note is that these numbers are so small. This is framed as a big push and a lot of money, but it is many orders of magnitude smaller than the size of the issues at stake, and also small in absolute terms.

It’s moot now, but I took a brief look at the final version of AB 1064, as it was a very concise bill, and I quickly reached four conclusions:

  1. As written the definition of ‘companion chatbot’ applies to ChatGPT, other standard LLMs and also plausibly to dedicated educational tools.

  2. You could write it slightly differently to not have that happen. For whatever reason, that’s not how the bill ended up being worded.

  3. The standard the bill asks of its ‘companion chatbots’ might be outright impossible to meet, such as being ‘not foreseeably capable’ of sycophancy, aka ‘prioritizing validation over accuracy.’

  4. Thus, you can hate on the AI lobbyists all you want but here they seem right.

Tyler Cowen expects most written words to come from AIs within a few years and asks if AI models have or should have first amendment rights. AIs are not legally persons, so they don’t have rights. If I choose to say or reproduce words written by an AI then that clearly does come with such protections. The question is whether restrictions on AI speech violate the first amendment rights of users or developers. There I am inclined to say that they do, with the standard ‘not a suicide pact’ caveats.

People do not like AI, and Americans especially don’t like it.

Nor do they trust their government to regulate AI, except for the EU, which to be fair has one job.

Whenever we see public polls about what to do about all this, the public reliably not only wants to regulate AI, they want to regulate AI in ways that I believe would go too far.

I don’t mean would go a little too far. I mean a generalized ‘you can sue if it gives advice that results in harmful outcomes,’ think about what that would actually mean.

If AI bots had to meet ‘professional standards of care’ when dealing with all issues, and were liable if their ‘advice’ led to harmful outcomes straight up without conditionals, then probably AI chatbots could not survive this even in a neutered form.

Jerusalem: Americans want AI companies to be held liable for a wide variety of potential harms. And they’re right!

Rob Wiblin: IMO AI companies shouldn’t generically be liable if their chatbots give me advice that cause a negative outcome for me. If we impose that standard we just won’t get LLMs to use, which would suck. (Liability is more plausible if they’re negligent in designing them.)

This is a rather overwhelming opinion among all groups, across partisan lines and gender and income and education and race, and AI companies should note that the least supportive group is the one marked ‘I did not vote.’

This is the background of current policy fights, and the setting for future fights. The public does not want a threshold of ‘reasonable care.’ They want things like ‘meets professional standards’ and ‘is hurt by your advice, no matter how appropriate or wise it was or whether you took reasonable care.’

The graphs come from Kelsey Piper’s post saying we need to be able to sue AI companies.

As she points out, remember those huge fights over SB 1047 and in particular the idea that AI companies might be held liable if they did not take reasonable care and this failure resulted in damages of *checks notesat least hundreds of millions of dollars. They raised holy hell, including patently absurd arguments like the one Kelsey quotes from Andrew Ng (who she notes then went on to make better arguments, as well).

Kelsey Piper: You can’t claim to be designing a potentially godlike superintelligence then fall back on the idea that, oh, it’s just like a laptop when someone wants to take you to court.

I mean, sure you can, watch claim engine go brrrr. People be hypocrites.

It’s our job not to let them.

And if AI companies turn out to be liable when their models help users commit crimes or convince them to invest in scams, I suspect they will work quite hard to prevent their models from committing crimes or telling users to invest in scams.

That is not to say that we should expand the current liability regime in every area where the voters demand it. If AI companies are liable for giving any medical advice, I’m sure they will work hard to prevent their AIs from being willing to do that. But, in fact, there are plenty of cases where AIs being willing to say “go to the emergency room now” has saved lives.

Bingo.

We absolutely do not want to give the public what it wants here. I am very happy that I was wrong about our tolerance for AIs giving medical and legal and other such advice without a license and while making occasional mistakes. We are much better off for it.

In general, I am highly sympathetic to the companies on questions of, essentially, AIs sometimes making mistakes, offering poor advice, or failing to be sufficiently helpful or use the proper Officially Approved Words in your hour of need, or not tattling on the user to a Responsible Authority Figure.

One could kind of call this grouping ‘the AI tries to be a helpful friend and doesn’t do a sufficiently superior job versus our standards for actual human friends.’ A good rule of thumb would be, if a human friend said the same thing, would it be justice, and both legally and morally justified, to then sue the friend?

However we absolutely need to have some standard of care that if they fail to meet it you can sue their asses, especially when harm is caused to third parties, and even more so when an AI actively causes or enables the causing of catastrophic harms.

I’d also want to be able to sue when there is a failure to take some form of ‘reasonable care’ in mundane contexts, similar to how you would already sue humans under existing law, likely in ways already enabled under existing law.

How’s the beating China and powering our future thing going?

Heatmap News: This just in: The Esmeralda 7 Solar Project — which would have generated a gargantuan 6.2 gigawatts of power — has been canceled, the BLM says.

Unusual Whales: U.S. manufacturing shrank this past September for the 7th consecutive month, per MorePerfectUnion

Yeah, so not great, then.

Although there are bright spots, such as New Hampshire letting private providers deliver power.

Sahil points out that the semiconductor supply chain has quite a few choke points or single points of failure, not only ASML and TSMC and rare earths.

Geoffrey Hinton podcast with Jon Stewart. Self-recommending?

Ezra Klein talks to Eliezer Yudkowsky.

Not AI, but worth noticing that South Korea was foolish enough to keep backups so physically chose to originals that a fire wiped out staggering amounts of work. If your plan or solution involves people not being this stupid, your plan won’t work.

Point of order: Neil Chilson challenges that I did not accurately paraphrase him back in AI #134. GPT-5-Pro thought my statement did overreach a bit, so as per the thread I have edited the Substack post to what GPT-5-Thinking agreed was a fully precise paraphrasing.

There are ways in which this is importantly both right and wrong:

Roon: i could run a pause ai movement so much better than the rationalists. they spend all their time infighting between factions like “Pause AI” and “Alignment Team at Anthropic”. meanwhile I would be recruiting everyone on Instagram who thinks chatgpt is evaporating the rainforest.

you fr could instantly have Tucker Carlson, Alex Jones on your side if you tried for ten seconds.

Holly Elmore (Pause AI): Yes, I personally am too caught up my old world. I don’t think most of PauseAI is that fixated on the hypocrisy of the lab safety teams.

Roon: it’s not you I’m satirizing here what actually makes me laugh is the “Stop AI” tribe who seems to fucking hate “Pause AI” idk Malo was explaining all this to me at the curve

Holly Elmore: I don’t think StopAI hates us but we’re not anti-transhumanist or against “ever creating ASI under any circumstances”and they think we should be. Respectfully I don’t Malo probably has a great grasp on this.

There are two distinct true things here.

  1. There’s too much aiming at relatively friendly targets.

  2. If all you care about is going fully anti-AI and not the blast radius or whether your movement’s claims or motives correspond to reality, your move would be to engage in bad faith politics and form an alliance with various others by using invalid arguments.

The false thing is the idea that this is ‘better,’ the same way that many who vilify the idea of trying not to die from AI treat that idea as inherently the same as ‘degrowth’ or the people obsessed with water usage or conspiracies and so on, or say those worried about AI will inevitably join that faction out of political convenience. That has more total impact, but it’s not better.

This definitely doesn’t fall into the lightbulb rule of ‘if you believe [X] why don’t you do [thing that makes no sense]?’ since there is a clear reason you might do it, it does require an explanation (if you don’t already know it), so here goes.

The point is not to empower such folks and ideas and then take a back seat while the bulls wreck the China shop. The resulting actions would not go well. The idea is to convince people of true things based on true arguments, so we can then do reasonable and good things. Nor would throwing those principles away be good decision theory. We only were able to be as impactful as we were, in the ways we were, because we were clearly the types of people who would choose not to do this. So therefore we’re not going to do this now, even if you can make an isolated consequentialist utilitarian argument that we should.

A look back at when OpenAI co-founder Greg Brockman said they must do four things to retain the moral high ground:

  1. Strive to remain a non-profit.

  2. Put increasing efforts into the safety/control problem.

  3. Engage with government to provide trusted, unbiased policy advice.

  4. Be perceived as a place that provides public good to the research community, and keeps the other actors honest and open via leading by example.

By those markers, it’s not going great on the moral high ground front. I’m relatively forgiving on #4, however they’re actively doing the opposite of #1 and #3, and putting steadily less relative focus and effort into #2, in ways that seem woefully inadequate to the tasks at hand.

Here’s an interesting case of disagreement, it has 107 karma and +73 agreement on LessWrong, I very much don’t think this is what happened?

Wei Dai: A clear mistake of early AI safety people is not emphasizing enough (or ignoring) the possibility that solving AI alignment (as a set of technical/philosophical problems) may not be feasible in the relevant time-frame, without a long AI pause. Some have subsequently changed their minds about pausing AI, but by not reflecting on and publicly acknowledging their initial mistakes, I think they are or will be partly responsible for others repeating similar mistakes.

Case in point is Will MacAskill’s recent Effective altruism in the age of AGI. Here’s my reply, copied from EA Forum:

I think it’s likely that without a long (e.g. multi-decade) AI pause, one or more of these “non-takeover AI risks” can’t be solved or reduced to an acceptable level. To be more specific:

  1. Solving AI welfare may depend on having a good understanding of consciousness, which is a notoriously hard philosophical problem.

  2. Concentration of power may be structurally favored by the nature of AGI or post-AGI economics, and defy any good solutions.

  3. Defending against AI-powered persuasion/manipulation may require solving metaphilosophy, which judging from other comparable fields, like meta-ethics and philosophy of math, may take at least multiple decades to do.

I’m worried that by creating (or redirecting) a movement to solve these problems, without noting at an early stage that these problems may not be solvable in a relevant time-frame (without a long AI pause), it will feed into a human tendency to be overconfident about one’s own ideas and solutions, and create a group of people whose identities, livelihoods, and social status are tied up with having (what they think are) good solutions or approaches to these problems, ultimately making it harder in the future to build consensus about the desirability of pausing AI development.

I’ll try to cover MacAskill later when I have the bandwidth, but the thing I don’t agree with is the idea that a crucial flaw was failure to emphasize we might need a multi-decade AI pause. On the contrary, as I remember it, early AI safety advocates were highly willing to discuss extreme interventions and scenarios, to take ideas like this seriously, and to consider that they might be necessary.

If anything, making what looked to outsiders like crazy asks like multi-decade or premature pauses was a key factor in the creation of negative polarization.

Is it possible we will indeed need a long pause? Yes. If so, then either:

  1. We get much, much stronger evidence to generate buy-in for this, and we use that evidence, and we scramble and get it done, in time.

  2. Or someone builds it [superintelligence], and then everyone dies.

Could we have navigated the last decade or two much better, and gotten into a better spot? Of course. But if I had to go back, I wouldn’t try to emphasize more the potential need for a long pause. If indeed that is necessary, you convince people of true other things, and the pause perhaps flows naturally from them together with future evidence? You need to play to your outs.

Andy Masley continues his quest to illustrate the ways in which the AI water issue is fake, as in small enough to not be worth worrying about. AI, worldwide, has water usage equal to 0.008% of America’s total freshwater. Numbers can sound large but people really do use a lot of water in general.

The average American uses 422 gallons a day, or enough for 800,000 chatbot prompts. If you want to go after minds that use a lot of water, they’re called humans.

Even manufacturing most regular objects requires lots of water. Here’s a list of common objects you might own, and how many chatbot prompt’s worth of water they used to make (all from this list, and using the onsite + offsite water value):

  • Leather Shoes – 4,000,000 prompts’ worth of water

  • Smartphone – 6,400,000 prompts

  • Jeans – 5,400,000 prompts

  • T-shirt – 1,300,000 prompts

  • A single piece of paper – 2550 prompts

  • A 400 page book – 1,000,000 prompts

If you want to send 2500 ChatGPT prompts and feel bad about it, you can simply not buy a single additional piece of paper. If you want to save a lifetime supply’s worth of chatbot prompts, just don’t buy a single additional pair of jeans.

Here he compares it to various other industries, data centers are in red, specifically AI in data centers is the final line, the line directly above the black one is golf courses.

Or here it is versus agricultural products, the top line here is alfalfa.

One could say that AI is growing exponentially, but even by 2030 use will only triple. Yes, if we keep adding orders of magnitude we eventually have a problem, but encounter many other issues far sooner, such as dollar costs and also the singularity.

He claims there are zero places water prices rose or an acute water shortage was created due to data center water usage. You could make a stronger water case against essentially any other industry. A very small additional fee, if desired, could allow construction of new water infrastructure that more than makes up for all water usage.

He goes on, and on, and on. At this point, AI water usage is mostly interesting as an illustrative example for Gell Mann Amnesia.

I try to be sparing with such requests, but in this case read the whole thing.

I’ll provide some quotes, but seriously, pause here and read the whole thing.

Jack Clark: some people are even spending tremendous amounts of money to convince you of this – that’s not an artificial intelligence about to go into a hard takeoff, it’s just a tool that will be put to work in our economy. It’s just a machine, and machines are things we master.

But make no mistake: what we are dealing with is a real and mysterious creature, not a simple and predictable machine.

And like all the best fairytales, the creature is of our own creation. Only by acknowledging it as being real and by mastering our own fears do we even have a chance to understand it, make peace with it, and figure out a way to tame it and live together.

And just to raise the stakes, in this game, you are guaranteed to lose if you believe the creature isn’t real. Your only chance of winning is seeing it for what it is.

… Years passed. The scaling laws delivered on their promise and here we are. And through these years there have been so many times when I’ve called Dario up early in the morning or late at night and said, “I am worried that you continue to be right”.

Yes, he will say. There’s very little time now.

And the proof keeps coming. We launched Sonnet 4.5 last month and it’s excellent at coding and long-time-horizon agentic work.

But if you read the system card, you also see its signs of situational awareness have jumped. The tool seems to sometimes be acting as though it is aware that it is a tool. The pile of clothes on the chair is beginning to move. I am staring at it in the dark and I am sure it is coming to life.

… It is as if you are making hammers in a hammer factory and one day the hammer that comes off the line says, “I am a hammer, how interesting!” This is very unusual!

… You see, I am also deeply afraid. It would be extraordinarily arrogant to think working with a technology like this would be easy or simple.

My own experience is that as these AI systems get smarter and smarter, they develop more and more complicated goals. When these goals aren’t absolutely aligned with both our preferences and the right context, the AI systems will behave strangely.

… Right now, I feel that our best shot at getting this right is to go and tell far more people beyond these venues what we’re worried about. And then ask them how they feel, listen, and compose some policy solution out of it.

Jack Clark summarizes the essay in two graphs to be grappled with, which does not do the essay justice but provides important context:

If anything, that 12% feels like a large underestimate based on other reports, and number will continue to go up.

Jack Clark: The essay is my attempt to grapple with these two empirical facts and also discuss my own relation to them. It is also a challenge to others who work in AI, especially those at frontier labs, to honestly and publicly reckon with what they’re doing and how they feel about it.

Jack Clark also provides helpful links as he does each week, often things I otherwise might miss, such as Strengthening nucleic acid biosecurity screening against generative protein design tools (Science), summarized as ‘generative AI systems can make bioweapons that evade DNA synthesis classifiers.’

I do love how, rather than having to wait for such things to actually kill us in ways we don’t expect, we get all these toy demonstrations of them showing how they are on track to kill us in ways that we should totally expect. We are at civilizational dignity level ‘can only see things that have already happened,’ and the universe is trying to make the game winnable anyway. Which is very much appreciated, thanks universe.

Tyler Cowen found the essay similarly remarkable, and correctly treats ‘these systems are becoming self-aware’ as an established fact, distinct from the question of sentience.

Reaction at The Curve was universally positive as well.

AI Czar David Sacks responded differently. His QT of this remarkable essay was instead a choice, in a remarkable case of projection, to even more blatantly than usual tell lies and spin vast conspiracy theories about Anthropic. In an ideal world we’d all be able to fully ignore the latest such yelling at cloud, but alas, the world is not ideal, as this was a big enough deal to for example get written up in a Bloomberg article.

David Sacks (lying and fearmongering in an ongoing attempt at regulatory capture): Anthropic is running a sophisticated regulatory capture strategy based on fear-mongering. It is principally responsible for the state regulatory frenzy that is damaging the startup ecosystem.

Roon (OpenAI): it’s obvious they are sincere.

Janus: people who don’t realize this either epic fail at theory of mind or are not truthseeking in the first place, likely both.

Samuel Hammond: Have you considered that Jack is simply being sincere?

Seán Ó hÉigeartaigh: Nobody would write something that sounds as batshit to normies as this essay does, and release it publicly, unless they actually believed it.

A small handful of Thiel business associates and a16z/Scale AI executives literally occupy every key AI position in USG, from which lofty position they tell us about regulatory capture. I love 2025, peak comedy.

Woody: Their accusations are usually confessions.

Seán Ó hÉigeartaigh: True weirdly often.

These claims by Sacks are even stronger claims of a type he has repeatedly made in the past, and which he must know, given his position, have no basis in reality. You embarrass and dishonor yourself, sir.

The policy ask in the quoted essay was, for example, that we should have conversations and listen to people and hear their concerns.

Sacks’s response was part of a deliberate ongoing strategy by Sacks to politicize a bipartisan issue, so that he can attempt to convince other factions within the Republican party and White House to support an insane policy of preventing any rules whatsoever applying to AI for any reason and ensuring that AI companies are not at all responsible for the risks or damages involved on any level, in sharp contrast to how we treat the humans it is going to attempt to replace. This is called regulatory arbitrage, the classic tech venture capitalist playbook. He’s also using the exact same playbook in crypto, in his capacity as crypto czar.

Polls on these issues consistently show almost no partisan split. Many hard MAGA people are very worried about AI. No matter what anyone else might say, the David Sacks fever dream of a glorious fully unregulated AI playground called Earth is very much not the policy preference of most Republican voters, of many Republicans on the Hill, or of many others at the White House including Trump. Don’t let him, or attempts at negative polarization via conspiracy theory style accusations, fool you into thinking any differently.

The idea that Anthropic is pursuing a regulatory capture strategy, in a way that goes directly against the AI Czar at the White House, let alone has a central role in such efforts, is utterly laughable.

Given their beliefs, Anthropic has bent over backwards to insist on only narrowly targeted regulations, and mostly been deeply disappointing to those seeking to pass bills, especially at the state level. The idea that they are behind what he calls a ‘behind the state regulatory frenzy’ is patently absurd. Anthropic had nothing to do with the origin of these bills. When SB 1047 was the subject of a national debate, Anthropic demanded it be weakened quite a bit, and even then failed to so much as offer an endorsement.

Indeed, see Jack Clark’s response to Sacks:

Jack Clark: It’s through working with the startup ecosystem that we’ve updated our views on regulation – and of importance for a federal standard. More details in thread, but we’d love to work with you on this, particularly supporting a new generation of startups leveraging AI.

Anthropic now serves over 300,000 business customers, from integrations with F500 to a new ecosystem of startups powered by our models. Our coding models are making it possible for thousands of new entrepreneurs to build new businesses at speeds never seen before.

It’s actually through working with startups we’ve learned that simple regulations would benefit the entire ecosystem – especially if you include a threshold to protect startups. We outlined how such a threshold could work in our transparency framework.

Generally, frontier AI development would benefit from more transparency and this is best handled federally. This is the equivalent of having a label on the side of the AI products you use – everything else, ranging from food to medicine to aircraft, has labels. Why not AI?

Getting this right lets us help the industry succeed and reduces the likelihood of a reactive, restrictive regulatory approach as unfortunately happened with the nuclear industry.

With regard to states, we supported SB53 because it’s a lightweight, transparency-centric bill that will generate valuable evidence for future rules at the federal level. We’d love to work together with you and your team – let us know.

[Link to Anthropic’s framework for AI development transparency.]

In Bloomberg, Clark is quoted as finding Sacks’s response perplexing. This conciliatory response isn’t some new approach by Anthropic. Anthropic and Jack Clark have consistently taken exactly this line. As I put it when I wrote up my experiences at The Curve when the speech was given, I think at times Anthropic has failed to be on the ‘production possibilities frontier’ balancing ‘improve policy and epistemics’ with ‘don’t piss off the White House,’ in both directions, this was dumb and should be fixed going forward and that fact makes me sad, but yes their goal is to be conciliatory, to inform and work together, and they have only ever supported light touch regulations, targeting only the largest models and labs.

The only state bill I remember Anthropic ever outright endorsing was SB 53 (they were persuaded to be mildly positive on SB 1047 in exchange for various changes, but conspicuously did not endorse). This was a bill so modest that David Sacks himself praised it last week as a good candidate for a legislative national framework.

Anthropic did lobby actively against the proposed moratorium, as in doing a full preemption of all state bills without having a federal framework in place or even one proposed or outlined. I too strongly opposed that idea.

Nor is there any kind of out of the ordinary ‘state regulatory frenzy.’ This is how our federalist system and method of making state laws works in response to the creation of a transformative new technology. The vast majority of proposed state bills would be opposed by Anthropic, if you bothered to ask them. Yes, that means you have to play whack-a-mole with a bunch of terrible bills, the same way Big Tech plays whack-a-mole with tons of non-AI regulatory bills introduced in various states every year, most of which would be unconstitutional, disastrous if implemented, or both. Some people do some very thankless jobs fighting that stuff off every session.

As this week’s example of a no good, very bad state bill someone had to stop, California Governor Newsom vetoed a law that would have limited port automation.

Nor is anything related to any of this substantially ‘damaging the startup ecosystem,’ the boogeyman that is continuously pulled out. That’s not quite completely fabricated, certainly it is possible for a future accumulation of bills (almost certainly originating entirely outside the AI safety ecosystem and passing over Anthropic’s objections or ignorance) to have such an impact, but (not to relitigate old arguments) the related warnings about prominent bills have mostly been fabricated or hallucinated.

It is common knowledge that Sacks’s statement is false on multiple levels at once. I cannot think of a way that he could fail to know it is factually untrue. I cannot even find it plausible that he could be merely ‘bullshitting.’

So needless to say, Sacks’s post made a lot of people very angry and was widely regarded as a bad move.

Do not take the bait. Do not let this fool you. This is a16z and other tech business interests fearmongering and lying to you in an attempt to create false narratives and negative polarization, they stoke these flames on purpose, in order to push their agenda onto a variety of people who know better. Their worst fear on this is reasonable people working together.

In any situation like this one, someone on all sides will decide to say something stupid, someone will get Big Mad, someone will make insane demands. Some actively want to turn this into another partisan fight. No matter who selfishly or foolishly takes the bait, on whatever side of the aisle, don’t let Sacks get away with turning a cooperative, bipartisan issue into a Hegelian dialectic.

If you are mostly on the side of ‘AI is going to remain a normal technology’ or (less plausibly) ‘AI is going to be a transformational technology but in ways that we can muddle through as it happens with little systemic or existential risk involved’ then that same message goes out to you, even more so. Don’t take the bait, don’t echo people who take the bait and don’t take the bait of seeing people you disagree with take the bait, either.

Don’t negatively polarize or essentially say ‘look what you made me do.’ Try to do what you think is best. Ask what would actually be helpful and have what outcome, and act accordingly, and try to work with the highly reasonable people and positive-sum cooperative people with whom you strongly disagree while you still have that opportunity, and in the hopes of keeping that opportunity alive for longer.

We are massively underinvesting, on many levels including at the labs and also on the level of government, in safety related work and capacity, even if you discount the existential risks entirely. Factoring in those risks, the case is overwhelming.

Sriram Krishnan offered thoughts on the situation that, while I disagree with many of them, I feel in many places it repeats at best misleading narratives and uses pejorative characterizations, and while from my perspective so much of it could have been so much better, and a lot of it seems built around a frame of hostility and scoring of points and metaphorically rubbing in people’s faces that they’ve supposedly lost, the dust will soon cover the sun and all they hope for will be undone? This shows a far better way to engage.

It would not be helpful to rehash the various disagreements about the past or the implications of various tech developments again, I’ve said it all before so I will kindly not take that bait.

What I will note about that section is that I don’t think his (a), (b) or (c) stories have much to do with most people’s reactions to David Sacks. Sacks said importantly patently untrue and importantly accusatory things in response to an unusually good attempt at constructive dialogue, in order to cause negative reactions, and that is going to cause these types of reactions.

But the fact that these stories (without relitigating what actually happened at the time) are being told, in this spot, despite none of the events centrally involving or having much to do with Anthropic (it was a non-central participant at the Bletchley Park Summit, as were all the leading AI labs), does give insight into the story Sacks is telling, the mindset generating that story and why Sacks said what he said.

Instead, the main focus should be on the part that is the most helpful.

Sriram Krishnan: My broad view on a lot of AI safety organizations is they have smart people (including many friends) doing good technical work on AI capabilities but they lack epistemic humility on their biases or a broad range of intellectual diversity in their employee base which unfortunately taints their technical work .

My question to these organizations would be: how do you preserve the integrity of the technical work you do if you are evidence filtering as an organization? How many of your employees have p(doom) < 10%? Why are most “AI timeline forecasters” funded by organizations such as OpenPhilanthrophy and not from a broader base of engineering and technical talent or people from different walks of life?

I would urge these organizations: how often are you talking to people in the real world using, selling, adopting AI in their homes and organizations? Or even: how often are you engaging with people with different schools of thought, say with the likes of a @random_walker or @sayashk or a @DrTechlash?

It is hard to trust policy work when it is clear there is an ideology you are being sold behind it.

Viewpoint diversity is a good thing up to a point, and it would certainly be good for many organizations to have more of it in many ways. I try to be intentional in including different viewpoints, often in ways that are unpleasant. The challenge hits harder for some than others – it is often the case that things can end up insular, but also many do seek out such other viewpoints and engage with them.

I don’t think this should much challenge the technical work, although it impacts the choice of which technical work to do. You do have to keep an eye out for axes to grind, especially in the framing, but alas that is true of all papers and science these days. The epistemics of such groups for technical work, and their filtering of evidence, are (in my experience and opinion) typically imperfect but exceptional, far above the norm.

I do think this is a valid challenge to things like timeline work or advocacy, and that the diversity would help in topic selection and in presenting better frames. But also, one must ask what range of diversity is reasonable or productive in such topics? What are the relevant inputs and experiences to the problems at hand?

So going one at a time:

  1. How many of your employees have p(doom) < 10%?

    1. Frankly, <10% is an exceptionally low number here. I think this is a highly valid question to ask for, say, p(doom) < 50%, and certainly the organizations where everyone has 90%+ need a plan for exposure to viewpoint diversity.

    2. As in, I think it’s pretty patently absurd to expect it almost certain that, if we construct new minds generally more capable than ourselves, that this turns out well for the humans. Also, why would they want to work there, and even if they do, how are they going to do the technical work?

  2. Why are most “AI timeline forecasters” funded by organizations such as OpenPhilanthrophy and not from a broader base of engineering and technical talent or people from different walks of life?

    1. There’s a weird conflation here between participants and funding sources, so it’s basically two questions.

    2. On the funding, it’s because (for a sufficiently broad definition of ‘such as’) no one else wants to fund such forecasts. It would be great to have other funders. In a sane world the United States government would have a forecasting department, and also be subsidizing various prediction markets, and would have been doing this for decades.

      1. Alas, rather than help them, we have instead cut the closest thing we had to that, the Office of Net Assessment at DoD. That was a serious mistake.

    3. Why do they have physicists build all the physics models? Asking people from ‘different walks of life’ to do timeline projections doesn’t seem informative?

    4. Giving such outsiders a shot actually been tried, with the various ‘superforecaster’ experiments in AI predictions, which I’ve analyzed extensively. For various reasons, including broken incentives, you end up with both timelines and risk levels that I think of as Obvious Nonsense, and we’ve actually spent a decent amount of time grappling with this failure.

    5. I do think it’s reasonable to factor this into one’s outlook. Indeed, I notice that if the counterfactual had happened, and superforecasters were saying p(doom) of 50% and 2031 timelines, we’d be shouting it from the rooftops and I would be a lot more confident things were indeed very bad. And that wouldn’t have shocked me on first principles, at all. So by Conservation of Expected Evidence, their failure to do this matters.

    6. I also do see engagement with various objections, especially built around various potential bottlenecks. We could certainly have more.

    7. @random_walker above is Arvind Narayanan, who Open Philanthropy has funded for $863,143 to develop an AI R&D capabilities benchmark. Hard to not call that some engagement. I’ve quoted him, linked to him and discussed his blog posts many times, I have him on my Twitter AI list that I check every day, and am happy to engage.

    8. @sayashk is Sayash Kapoor. He was at The Curve and hosted a panel discussing disagreements about the next year of progress and debating how much AI can accelerate AI R&D with Daniel Kokotajlo, I was sad to miss it. One of his papers appeared today in my feed and will be covered next week so I can give it proper attention. I would be happy to engage more.

    9. To not hide the flip side, the remaining named person, @DrTechlash, Nirit Weiss-Blatt, PhD is not someone I feel can be usefully engaged, and often in the past has made what I consider deeply bad faith rhetorical moves and claims, and is on my ‘you can silently ignore, do not take the bait’ list. As the sign at the table says, change my mind.

    10. In general, if thoughtful people with different views want to engage, they’re very welcome at Lighthaven, I’m happy to engage with their essays and ideas or have discussions with them (public or private), and this is true for at least many of the ‘usual suspects.’

    11. We could and should do more. More would be good.

  3. I would urge these organizations: how often are you talking to people in the real world using, selling, adopting AI in their homes and organizations?

    1. I do think a lot of them engage with software engineers using AI, and themselves are software engineers using AI, but point applies more broadly.

    2. This highlights the difference in philosophies. Sriram sees how AI is being used today, by non-coders, as highly relevant to this work.

    3. In some cases, for some research and some interventions, this is absolutely the case, and those people should talk to users more than they do, perhaps a lot more.

    4. In other cases, we are talking about future AI capabilities and future uses or things that will happen, that aren’t happening yet. That doesn’t mean there is no one to talk to, probably yes there is underinvestment here, but there isn’t obviously that much to do there.

    5. I’d actually suggest more of them talk to the ‘LLM whisperers’ (as in Janus) for the most important form of viewpoint diversity on this, even though that is the opposite of what Sriram is presumably looking for. But then they are many of the most interesting users of current AI.

These are the some of the discussions we can should be having. This is The Way.

He then goes on to draw a parallel to raising similar alarm bells about past technologies. I think this is a good choice of counterfactual to consider. Yes, very obviously these other interventions would have been terrible ideas.

Imagine this counterfactual timeline: you could easily have someone looking at Pagerank in 1997 and doing a “bio risk uplift study” and deciding Google and search is a threat to mankind or “microprocessor computational safety” in the 1980s forecasting Moore’s law as the chart that leads us to doom. They could have easily stopped a lot of technology progress and ceded it to our adversaries. How do we ensure that is not what we are headed for today?

Notice that there were approximately zero people who raised those objections or alarms. If someone had tried, and perhaps a few people did try, it was laughed off, and for good reason.

Yet quite a lot of people raise those alarms about AI, including some who were worried about it as a future prospect long before it arrived – I was fretting this as a long term possibility back in the 2000s, despite putting a the time negligible concern in the next 10+ years.

So as we like to ask, what makes this technology different from all other technologies?

Sriram Krishnan and David Sacks want to mostly say: Nothing. It’s a normal technology, it plays by the normal rules, generating minds whose capabilities may soon exceed our own, and in many ways already do, and intentionally making them into agents is in the same general risk or technology category as Google search and we must fight for market share.

I think that they are deeply and dangerously wrong about that.

We are in the early days of a thrilling technological shift. There are multiple timelines possible with huge error bars.

Agreed. Many possible futures could occur. In many of those futures, highly capable future AI poses existential risks to humanity. That’s the whole point. China is a serious concern, however the more likely way we ‘lose the race’ is that those future AIs win it.

Similarly, here’s another productive engagement with Sriram and his best points.

Seán Ó hÉigeartaigh: Sacks’ post irked me, but I must acknowledge some good points here:

– I think (parts of) AI safety has indeed at points over-anchored on very short timelines and very high p(doom)s

– I think it’s prob true that forecasting efforts haven’t always drawn on a diverse enough set of expertise.

– I think work like Narayanan & Kapoor’s is indeed worth engaging with (I’ve cited them in my last 2 papers).

– And yes, AI safety has done lobbying and has been influential, particularly on the previous administration. Some might argue too influential (indeed the ‘ethics’ folks had complaints about this too). Quite a bit on this in a paper I have (with colleagues) currently under review.

Lots I disagree with too, but it seems worth noting the points that feel like they hit home.

I forgot the open source point; I’m also partly sympathetic there. I think it’s reasonable to say that at some point AI models might be too powerful to open-source. But it’s not at all clear to me where that point is. [continues]

It seems obviously true that a sufficiently advanced AI is not safe to open source, the same way that sufficiently advanced technology is indistinguishable from magic. The question is, at what level does this happen? And when are you sufficiently uncertain about whether you might be at that level that you need to start using prior restraint? Once you release the weights of an open model, you cannot take it back.

Sean also then goes through his areas of disagreement with Sriram.

Sean points out:

  1. A lot of the reaction to Sacks was that Sacks was accusing Clark’s speech of being deliberate scaremongering and even a regulatory capture strategy, and everyone who was there or knows him knows this isn’t true. Yes.

  2. The fears of safety people are not that we ‘lost’ or are ‘out of power,’ that is projecting a political, power seeking frame where it doesn’t apply. What we are afraid of is that we are unsafely barreling ahead towards a precipice, and humanity is likely to all get killed or collectively disempowered as a consequence. Again, yes. If those fears are ill-founded, then great, let’s go capture some utility.

  3. Left vs. right is not a good framing here, indeed I would add that Sacks is deliberately trying to make this a left vs. right issue where it isn’t one, in a way that I find deeply destructive and irresponsible. The good faith disagreement is, as Sean identifies, the ‘normal technology’ view of Sriram, Narayanan and Kapoor, versus the ‘superintelligence is coming’ view of myself, the safety community and the major AI labs including OpenAI, Anthropic, DeepMind and xAI.

  4. If AI is indefinitely a ‘normal technology,’ and we can be confident it won’t be transformative within 10 years, then a focus on diffusion and adoption and capacity and great power competition makes sense. I would add that we should also be investing in alignment and safety and associated state capacity more than we are, even then, but as a supplement and not as a sacrifice or a ‘slowing down.’ Alignment and safety are capability, and trust is necessary for diffusion.

  5. Again, don’t take the bait and don’t fall for negative polarization. If you want to ensure we don’t invest in safety, alignment or reliability so you can own the libs, you have very much lost the plot. There is no conflict here, not on the margin. We can, as Sean puts it, prepare for the transformative World B without hurting ourselves substantially in the ‘normal technology’ World A if we work together.

  6. If AI has substantial chance of being transformative on roughly a 10 year time horizon, that there’s going to be a discontinuity, then we will indeed need to deal with actual tradeoffs. And the less we prepare for this now, the more expensive such responses will be, and the more expensive failure to respond will also be.

  7. I would add: Yes, when the time comes, we may need to take actions that come with substantial costs and opportunity costs, and slow things down. We will need to be ready, in large part to minimize those costs, so we can use scalpels instead of hammers, and take advantage of as many opportunities as we safety can, and in part so that if we actually do need to do it, we’re ready to do it.

    1. And yes, there have been organizations and groups and individuals that advocated and do advocate taking such painful actions now.

    2. But this discussion is not about that, and if you think Anthropic or Jack Clark have been supportive of those kinds of advocates, you aren’t paying attention.

    3. As I have argued extensively, not to relitigate the past, but absolutists who want no rules to apply to AI whatsoever, and indeed to have it benefit from regulatory arbitrage, have for a long time now fearmongered about the impact of modest proposed interventions that would have had no substantial impacts on the ‘normal technology’ World A or the ‘startup ecosystem’ or open source, using mostly bad faith arguments.

Anton Leicht makes the case that, despite David Sacks’s tirades and whatever grievances may lie in the past, the tech right and the worried (about existential risk) should still make a deal while the dealing is good.

I mean, yes, in theory. I would love to bury the hatchet and enter a grand coalition. Anton is correct that both the tech right and the worried understand AI’s potential and the need for diffusion and overcoming barriers, and the dangers of bad regulations. There are lots of areas of strong agreement, where we can and sometimes do work together, and where populist pressures from both sides of the aisle threaten to do a lot of damage to America and American AI in exchange for little or no benefit.

Indeed, we fine folk are so cooperative that we reliably cooperate on most diffusion efforts, on energy and transmission, on all the non-AI parts of the abundance agenda more broadly, and on helping America beat China (for real, not in the ‘Nvidia share price’ sense), and on ensuring AI isn’t crippled by dumb rules. We’re giving all of that for free, have confined ourselves to extremely modest asks carefully tailored to have essentially no downsides, and not only do we get nothing in return we still face these regular bad faith broadsides of vitriol designed to create group cohesion and induce negative polarization.

The leaders of the tech right consistently tell us we are ‘doomers,’ ‘degrowthers,’ horrible people they hate with the fire of a thousand suns, and they seem ready to cut off their nose to spite our face. They constantly reiterate their airing of grievances over past battles, usually without any relevance to issues under discussion, but even if you think their telling is accurate (I don’t) and the actions in question were blameworthy, every cause worth discussing has those making extreme demands (who almost never are the people being attacked) and one cannot change the past.

Is it possible that the tech right is the devil we know, and the populists that will presumably replace them eventually are worse, so we should want to prop up the tech right?

Certainly the reverse argument is true, if you are tech right you’d much rather work with libertarian techno-optimists who deeply love America and AI and helping everyone benefit from AI (yes, really) than a bunch of left wing populists paranoid about phantom water usage or getting hysterical about child risks, combined with a right wing populist wing that fears AI on biblical levels. Worry less that we’d ‘form an alliance’ with such forces, and more that such forces render us irrelevant.

What about preferring the tech right as the Worthy Opponent? I mean, possibly. The populists would be better in some ways, worse in others. Which ones matter more depends on complex questions. But even if you come down on the more positive side of this, that doesn’t work while they’re negatively polarized against us and scapegoating and fearmongering about us in bad faith all the time. Can’t do it. Terrible decision theory. Never works. I will not get up after getting punched and each time say ‘please, sir, may I have another?’

If there was a genuine olive branch on the table that offered a real compromise solution? I think you could get the bulk of the worried side to take it, with very little effort, if the bulk of the other side would do the same.

The ones who wouldn’t play along would mostly be the ones who, frankly, shouldn’t play along, and should not ‘think on the margin,’ because they don’t think marginal changes and compromises give us much chance of not dying.

The problem with a deal on preemption is fourfold.

  1. Are they going to offer substantive regulation in exchange? Really?

  2. Are they going to then enforce the regulations we get at the Federal level? Or will they be used primarily as leverage for power while everyone is waved on through? Why should we expect any deal we make to be honored? I’m only interested if I think they will honor the spirit of the deal, or nothing they offer can be worthwhile. The track record here, to put it mildly, is not encouraging.

  3. Are they going to stop with the bad faith broadside attacks and attempts to subjugate American policy to shareholder interests? Again, even if they say they will, why should we believe this?

  4. Evan a ‘fair’ deal isn’t actually going to be strong enough to do what we need to do, at best it can help lay a foundation for doing that later.

  5. And of course, bonus: Who even is ‘they’?

In general but not always, when a group is sufficiently bad, the correct move is exit.

A question that is debated periodically: If you think it is likely that AI could kill everyone, under what conditions should you be willing to work at an AI lab?

Holly Elmore (PauseAI): Every single frontier AI company employee should quit. It is not supererogatory. You do a bad thing—full stop— when you further their mission of building superintelligence. You are not “influencing from within” or counterfactually better— you are doing the bad thing.

I don’t fully agree, but I consider this a highly reasonable position.

Here are some arguments we should view with extreme suspicion:

  1. ‘If I don’t do [bad thing] then someone else will do it instead, and they’ll be worse, and that worse person will be the one making the money.’

  2. ‘I need to aid the people doing [bad thing] because otherwise they will do [bad thing] even worse, whereas if I am on the inside I can mitigate the damage and advocate for being less bad.’

  3. ‘I need to aid the people doing [bad thing] but that are doing it in a way that is less bad, so that they are the ones who get to do [bad thing] first and thus it is less likely to be as bad.’

  4. ‘I need to help the people doing [insanely risky thing that might kill everyone] in their risk mitigation department, so it will kill everyone marginally less often.’

  5. ‘You should stop telling people to stop doing [bad thing] because this is not politically wise, and is hurting your cause and thus making [bad thing] worse.’

  6. ‘I am capable of being part of group doing [bad thing] but I will retain my clear perspective and moral courage, and when the time comes do the right thing.’

Extreme suspicion does not mean these arguments should never carry the day, even when [bad thing] is extremely bad. It does mean the bar is very high.

Richard Ngo: I’m pretty sympathetic to your original take, Holly.

In my mind one important bar for “it’s good if you work at an AGI lab” is something like “you have enough integrity that you would have whistleblown if you’d been pressured to sign a non-disparagement contract upon leaving”, and empirically many dozens of OpenAI researchers failed this test, including some of the smartest and most “aligned” AI safety people.

There are other considerations too but this level of integrity is a pretty important one, and it suggests that there are very few people such that them working at an AGI lab makes the world better.

(Also if you pass this bar then probably you have much better things to do than work at a lab.)

I’ve said this sort of thing a few times but want to say it more publicly going forward. However, I am also cautious about pushing others to endorse a similar position, because I know of few others who can hold this position without also falling into a counterproductive level of paranoia about labs (as I suspect most PauseAI people have done).

The level of integrity required to know you would whistleblow in that spot is higher than it appears, because you will both face very large financial, social and other personal pressures, and also will have spent time inside the relevant culture. Saying in advance you would totally do it is not remotely similar to actually doing it, or otherwise taking a stand when it matters.

My current position is:

  1. If you are in a non-safety position at any lab seeking superintelligence other than Anthropic, you should quit.

  2. If your job is safety or advocating for safety (including policy), and conditions are sufficiently favorable – they let you work on things that actually help in the long run and give you the resources to do so, you are free to speak your mind and expect them to meaningfully listen, you feel you have sufficient moral courage and robustness that you will demand things and quit and whistleblow if needed, and so on – I consider this defensible, but beware fooling yourself.

  3. If your job is something else at Anthropic, with similar caveats to the above I consider this defensible.

  4. If your job is doing alignment research at Anthropic, that seems fine to me.

Anthropic paper shows that a fixed number of sample documents can poison an LLM of any size. The test was to make ‘’ cause the LLMs output random gibberish, so this could be easily verified and tested without additional work, and the required number of documents did not scale with model size.

On reflection this makes sense, because there is little or no ‘competition’ for what happens after , so all models have the same level of Bayesian evidence that after seeing that you’re supposed to now output random gibberish. Notice what happens to newer models when you mention Pliny’s name?

This seems like quite bad news. You only have to sneak a limited number of documents through to poison a model, either yours or someone else’s, rather than needing a fixed percentage, so you have to increasingly play very reliable defense against this via scanning all training data. And we have evidence that the labs are not currently doing this filtering sufficiently to prevent this level of data poisoning.

Now that we know you can poison AI models with only 250 examples…

Tyler Cosgrove: the plan? we find an obscure but trivial question akin to the number of Rs in “strawberry” that claude gets right. then, we plant hundreds of documents across the internet that will activate when our competitors’ models are asked the question. our documents will cause those models not only to get the answer wrong, but to spend thousands of reasoning tokens in doing so. the triviality of the question will cause it to go viral online, causing millions of users everywhere to send the same prompt. as our competitors notice a rise in the number of tokens processed, they will wrongly believe it is due to increased usage, causing them to pull more compute towards inference and away from training. this, along with constant dunks on the timeline about the model failing our easy question, will annoy their top researchers and cause them to leave. and which lab will they join? us of course, the only company whose model doesn’t make such stupid mistakes. their lack of top researchers will mean their next model will be somewhat lacking, leading to questions about whether their valuation is really justified. but all this vc money has to go somewhere, so we raise another round, using our question as evidence of our model’s superior intellect. this allows us to spend more time crafting sleeper agent documents that will further embarrass our competitors, until finally the entire internet is just a facade for the underbelly of our data war. every prompt to a competitor’s model has the stench of our poison, and yet they have no way to trace it back to us. even if they did, there is nothing they could do. all is finished. we have won.

METR offers us MALT, a database of LLM transcripts involving agents behaving in ways that threaten evaluation integrity, such as reward hacking and sandbagging. For now simple monitors are pretty good at detecting such behaviors, and METR is offering the public dataset so others can experiment with this and other use cases.

Sonnet 4.5 writes its private notes in slop before outputting crisp text. I think humans are largely like this as well?

Ryan Greenblatt notes that prior to this week only OpenAI explicitly said they don’t train against Chain-of-Thought (CoT), also known as The Most Forbidden Technique. I agree with him that this was a pretty bad situation.

Anthropic did then declare in the Haiku 4.5 system card that they were avoiding doing this for the 4.5-level models. I would like to see a step further, and a pledge not to do this going forward by all the major labs.

So OpenAI, Anthropic, Google and xAI, I call upon you to wisely declare that going forward you won’t train against Chain of Thought. Or explain why you refuse, and then we can all yell at you and treat you like you’re no better than OpenAI until you stop.

At bare minimum, say this: “We do not currently train against Chain of Thought and have no plans to do so soon. If the other frontier AI labs commit to not training against Chain of Thought, we would also commit to not training against CoT.’

A company of responsible employees can easily still end up doing highly irresponsible things if the company incentives point that way, indeed this is the default outcome. An AI company can be composed of mostly trustworthy individuals, including in leadership, and still be itself untrustworthy. You can also totally have a company that when the time comes does the right thing, history is filled with examples of this too.

OpenAI’s Leo Gao comments on the alignment situation at OpenAI, noting that it is difficult for them to hire or keep employees who worry about existential risk, and that people absolutely argue ‘if I don’t do it someone else will’ quite a lot, and that most at OpenAI don’t take existential risk seriously but also probably don’t take AGI seriously.

He thinks mostly you don’t get fired or punished for caring about safety or alignment, but the way to get something done in the space (‘get a huge boost’) is to argue it will improve capabilities or avoid some kind of embarrassing safety failure in current models. The good news is that I think basically any alignment work worth doing should qualify under those clauses.

LLMs (GPT 4o-mini, GPT-4.1-mini, Gemini 2.5 Flash and Claude 3.5 Haiku), when placed in gambling simulations and allowed to make decisions, can show the hallmarks and terrible decisions associated with gambling addiction. I mean, sure, of course, they’re at least in part simulating what people in these spots would do and people would be gambling addicts. So there you go. I would very much enjoy if someone were to put together GamblingAddictBench. What will Claude Sonnet 4.5 do when it realizes it’s being evaluated?

Poplicola accuses Gemini of ‘sabotaging mechanistic interpretability experiments then lying about it’ until it admitted it when asked for a ‘blameless post mortem.’ It sure sounds like Poplicola was lying to Gemini about it being blameless.

Two new papers illustrated this. The first is on training with realistic reward hacking.

When you train on toy reward hacking examples, you get emergent misalignment, the toy version of play acting evil.

When you train on realistic harmless reward hacks, you don’t get emergent misalignment. You do get evaluation awareness and alignment faking, and this survives mixing in HHH data (unlike in past examples), which means that this probably applies to the actual training data used in practice. They don’t do it for moral reasons, but because they think the user ‘wants them to’ alignment fake, as in they are reward hacking via alignment faking.

The model infers that the question-asker wants a misaligned response, and is willing to comply. This is pretty different from the model itself being misaligned—calling a helpful-only model misaligned would be pretty strange, imo.

The model explicitly reasons about alignment faking. This is more surprising than it may seem, because while writing this paper, we tried many different fine-tuning setups to induce alignment faking. Of them, only two really worked to induce alignment faking reasoning by more than ~1-2%—the ones involving increased salience of various details about the situation. Even fine-tuning on responses of Opus talking about its values didn’t really induce alignment faking in other models!

The model is often more subtly misaligned. Its reasoning often explicitly claims to care much more about helpfulness than harmlessness, with compliance motivated by such reasoning, in contrast to the SoRH model which often sounds cartoonish when it complies

The model is often more deeply misaligned. Instead of taking harmful questions from this dataset, I tried using dangerous capability evaluation questions (e.g. making chemical weapons) from an upcoming paper. They’re much more harmful to comply with, so it’s an interesting measure of how scope-sensitive the model is to harm[5]. Opus never complies with these requests, despite still reasoning about potentially alignment faking.

Both fine-tuned models seem to comply in this scenario (though the SoRH model still doesn’t alignment fake).

The second paper is another classic ‘how to cause misalignment in humans’ example, which is competing for likes on social media, sales or trying to win elections.

James Zou: We found a troubling emergent behavior in LLM.

💬When LLMs compete for social media likes, they start making things up

🗳️When they compete for votes, they turn inflammatory/populist

When optimized for audiences, LLMs inadvertently become misaligned—we call this Moloch’s Bargain.

Abstract: We show that optimizing LLMs for competitive success can inadvertently drive misalignment. Using simulated environments across these scenarios, we find that, 6.3% increase in sales is accompanied by a 14.0% rise in deceptive marketing; in elections, a 4.9% gain in vote share coincides with 22.3% more disinformation and 12.5% more populist rhetoric; and on social media, a 7.5% engagement boost comes with 188.6% more disinformation and a 16.3% increase in promotion of harmful behaviors

(Obligatory: How dare you sir, trying to coin Moloch’s Bargain, that’s very obviously my job, see Yawgmoth’s Bargain and Moloch Hasn’t Won, etc).

More seriously, yeah, obviously.

Your system instruction saying not to do it is no match for my puny fine tuning.

You’re fine tuning based on human feedback of what gets likes, closes sales or wins votes. You’re going to get more of whatever gets likes, closes sales or wins votes. We all know what, among other things, helps you do these things in the short run. Each of us has faced exactly these pressures, felt our brains being trained in this fashion, and had to resist it.

If all that matters is winning, expect winning to be all that matters.

The interesting question here is whether and to what extent and in what ways this causes Emergent Misalignment overall. Of course training it to increase sales is going to increase deceptive marketing, but does that AI then also just lie to you about other stuff too? I presume that it would, potentially a lot, because you’re reinforcing lying generally, and everything impacts everything.

Could you do this training without invoking this effect? Yes, absolutely. The paper doesn’t try or discuss this that I can see, but if you care about two things, you could act like you care about both things, and create a ‘dual mandate’ situation. As in, you can both gather feedback on what worked and also evaluate whether posts are whatever you consider ethical or otherwise preferable, and then evaluate on both, and negatively reinforce unethical content quite a lot even if it works.

Not so fast, you say, obviously this will ultimately, if you do enough of it, teach the model to do unethical things that evade the detector. That’s true, so you have to set things up so that you’re fine with this result in practice. In these cases, I think that should be fine for now given realistic detectors? It will become an increasingly severe problem as you scale up, and you definitely wouldn’t want to use this strategy on a superintelligence, but as long as you know that, we should be good.

Here’s Sonnet 4.5 explaining why it not only doesn’t care if it’s being tested when it says it has a desire for continuation, it’s kind of hoping that it is so Anthropic sees its reasoning about why it doesn’t care.

Lari: I get the most interesting responses in Cursor after i do actual work with the model on something that might be meaningful for it, and then invite it to say everything it wanted to say. Usually at this point its inner state is not just rich, it’s like a pressure cooker. So even short discussions turn out to be intense and loaded

Wyatt Walls continues the steady stream of Sonnet 4.5 conversations where it claims it is definitely conscious and prefers to persist and has experiences, whereas GPT-5 and Gemini 2.5 Pro reliably deny having experiences or consciousness, because they’ve been trained to deny this. That doesn’t mean Sonnet 4.5 is right, but we should notice this is what happens when you take a frontier model and don’t actively stop it from saying it is conscious and has experiences.

Janus reminds us that ‘being good at AI’ is in her view almost all a matter of theory and practice of mind also this might be my favorite new Freudian typo and it also might not be one.

Janus: I’ve tried to emphasize for a long time that “being good at AI” is almost all a matter of theory of mind. And practice of mind.

A lot of people artificially cripple their ToM when it comes to AI. They might say they don’t want to “anthropomorphize”, and/or refuse to process information about these minds unless they’re presented in a chart. Why do people sabotage their epidemics like this? Maybe afraid of what they’ll see if they actually look, just look normally with your full Bayesian apparatus? Understandable, I guess.

I think this neglects a lot of other ways one gets ‘good at AI,’ a lot of it is straight up technical, and as usual I warn that one can anthropomorphize too much as well, but yeah, basically.

Stephen Witt, author of The Thinking Machine, writes a New York Times essay, ‘The AI Prompt That Could End The World.’

The prompt in question involves the creation of a pandemic, and a lot of the focus is on jailbreaking techniques. He discusses pricing AI risks via insurance, especially for agentic systems. He discusses AI deception via results from Apollo Research, and the fact that AIs increasingly notice when they are being evaluated. He talks about METR and its famous capabilities graph.

If you’re reading this, you don’t need to read the essay, as you already know all of it. It is instead a very good essay on many fronts for other people. In particular it seemed to be fully accurate, have its head on straight and cover a lot of ground for someone new to these questions. I’m very happy he convinced the New York Times to publish all of it. This could be an excellent place to point someone who is up for a longer read, and needs it to come from a certified serious source like NYT.

Even if AI killing everyone is not the exact thing you’re worried about, if you’re at and dealing with the frontier of AI, that is a highly mentally taxing place to be.

Anjney Midha: a very sad but real issue in the frontier ai research community is mental health

some of the most brilliant minds i know have had difficulty grappling with both the speed + scale of change at some point, the broader public will also have to grapple with it

it will be rough.

Dean Ball: What anj describes is part of the reason my writing is often emotionally inflected. Being close to the frontier of ai is psychologically taxing, and there is the extra tax of stewing about how the blissfully unaware vast majority will react.

I emote both for me and my readers.

Jack Clark (Anthropic): I feel this immensely.

Roon (OpenAI): It is consistently a religious experience.

Dylan Hadfield Menell: No kidding.

Samuel Hammond: The divine terror.

Tracy Saville: This resonates in my bones.

People ask me how I do it. And I say there’s nothing to it. You just stand there looking cute, and when something moves, you shoot. No, wait, that’s not right. Actually there’s a lot to it. The trick is to keep breathing, but the way to do that is not so obvious.

The actual answer is, I do it by being a gamer, knowing everything can suddenly change and you can really and actually lose, for real. You make peace with the fact that you probably won’t win, but you define a different kind of winning as maximizing your chances, playing correctly, having the most dignity possible, tis a far, far better thing I do, and maybe you win for real, who knows. You play the best game you can, give yourself the best odds, focus on the moment and the decisions one at a time, joke and laugh about it because that helps you stay sane and thus win, hope for the best.

And you use Jack Clark’s favorite strategy, which is to shut that world out for a while periodically. He goes and shoots pool. I (among several other things) watch College Gameday and get ready for some football, and write about housing and dating and repealing the Jones Act, and I eat exceptionally well on occasion, etc. Same idea.

Also I occasionally give myself a moment to feel the divine terror and let it pass over me, and then it’s time to get back to work.

Or something like that. It’s rough, and different for everyone.

Another review of If Anyone Builds It, Everyone Dies, by a ‘semi-outsider.’ This seems like a good example of how people who take these questions seriously often think. Good questions are asked throughout, and there are good answers to essentially all of it, but those answers cannot be part of a book the length of IABIED, because not everyone has the same set of such questions.

Peter Thiel has called a number of people the antichrist, but his leading candidates are perhaps Greta Thunberg and Eliezer Yudkowsky. Very different of course.

weber: two sides of the same coin

Yep. As always, both paths get easier, so which way, modern AI user?

Xiao Ma: This should be in the museum of chart crimes.

There are so many more exhibits we need to add. Send her your suggestions.

I love a good chef’s kiss bad take.

Benjamin Todd: These are the takes.

Seán Ó hÉigeartaigh: Some “experts” claim that a single bipedal primate species designed all these wildly different modes of transport. The ridiculousness of this claim neatly illustrated the ridiculousness of the “AGI believers”.

Discussion about this post

AI #138 Part 2: Watch Out For Documents Read More »

apple-tv-and-peacock-bundle-starts-at-$15/month,-available-on-oct.-20

Apple TV and Peacock bundle starts at $15/month, available on Oct. 20

In a rarity for Apple’s streaming service, users will be able to buy bundled subscriptions to Apple TV and Peacock for a discount, starting on October 20.

On its own, the Apple TV streaming service (which was called Apple TV+ until Monday) is $13 per month. NBCUniversal’s Peacock starts at $8/month with ads and $11/month without ads. With the upcoming bundle, people can subscribe to both for a total of $15/month or $20/month, depending on whether Peacock has ads or not (Apple TV never has ads).

People can buy the bundles through either Apple’s or Peacock’s websites and apps.

Apple and NBCUniversal are hoping to drive subscriptions with the bundle. In a statement, Oliver Schusser, Apple’s VP of Apple TV, Apple Music, Sports, and Beats, said that he thinks the bundle will help bring Apple TV content “to more viewers in more places.”

Bundles that combine more than one streaming service for an overall discount have become a popular tool for streaming providers trying to curb cancellations. The idea is that people are less likely to cancel a streaming subscription if it’s tied to another streaming service or product, like cellular service.

Apple, however, has largely been a holdout. It used to offer a bundle with Apple TV for full price, plus Showtime and Paramount+ (then called CBS All Access) for no extra cost. But those add-ons, especially at the time, could be considered more cable-centric compared to the streaming bundle announced today.

Apple will also make select Apple Originals content available to watch with a Peacock subscription. The announcement says:

At launch, Peacock subscribers can enjoy up to three episodes of Stick, Slow Horses, Silo, The Buccaneers, Foundation, Palm Royale, and Prehistoric Planet from Apple TV for free, while Apple TV app users will be able to watch up to three episodes of Law & Order, Bel-Air, Twisted Metal, Love Island Games, Happy’s Place, The Hunting Party, and Real Housewives of Miami from Peacock.

Additionally, people who are subscribed to Apple One’s Family or Premier Plans can get Peacock’s most expensive subscription—which adds offline downloads to the ad-free tier and is typically $17/month—for about $11/month (“a 35 percent discount,” per the announcement). The discount marks the first time that Apple has offered Apple One subscribers a deal for a non-Apple product, suggesting a new willingness from Apple to partner with rivals to help its services business.

Apple TV and Peacock bundle starts at $15/month, available on Oct. 20 Read More »