Author name: Tim Belzer

new-material-may-help-us-build-predator-style-thermal-vision-specs

New material may help us build Predator-style thermal vision specs

One way to do that is called remote epitaxy, where an intermediate layer made out of graphene or other material is introduced between the substrate and the growing crystals. Once the epitaxy process is done, the substrate and everything on it are soaked in a chemical solution that dissolves this intermediate layer, leaving the crystalline film intact. This works but is expensive, difficult to scale, and takes a lot of time. To make the process cheaper and faster, the MIT team had to grow the crystals directly on the substrate, without any intermediate layers. What they were trying to achieve was a non-stick frying pan effect but at an atomically small scale.

Weakening the bonds

The material that prevented the crystalline films from sticking to substrates wasn’t Teflon but lead. When the team was experimenting with growing different films in their previous studies, they noticed that there was a material that easily came off the substrate, yet retained an atomically smooth surface: PMN-PT, or lead magnesium niobate-lead titanate.

The lead atoms in the PMN-PT weakened the covalent bonds between the film and the substrate, preventing the electrons from jumping through the interface between the two materials. “We just had to exert a bit of stress to induce a crack at the interface between the film and the substrate and we could realize the liftoff,” Zhang told Ars. “Very simple—we could remove these films within a second.”

But PMN-PT, besides its inherent non-stickiness, had more tricks up its sleeves; it had exceptional pyroelectric properties. Once the team realized they could manufacture and peel away PMN-PT films at will, they tried something a bit more complex: a cooling-free, far-infrared radiation detector. “We were trying to achieve performance comparable with cooled detectors,” Zhang says.

The detector they constructed was made from 100 pieces of 10-nanometer-thin PMN-PT films, each about 60 square microns, that the team transferred onto a silicon chip. This produced a 100-pixel infrared sensor. Tests with ever smaller changes in temperature indicated that it outperformed state-of-the art night vision systems and was sensitive to radiation across the entire infrared spectrum. (Mercury cadmium telluride detectors respond to a much narrower band of wavelengths.)

New material may help us build Predator-style thermal vision specs Read More »

ai-#114:-liars,-sycophants-and-cheaters

AI #114: Liars, Sycophants and Cheaters

Gemini 2.5 Pro is sitting in the corner, sulking. It’s not a liar, a sycophant or a cheater. It does excellent deep research reports. So why does it have so few friends? The answer, of course, is partly because o3 is still more directly useful more often, but mostly because Google Fails Marketing Forever.

Whereas o3 is a Lying Liar, GPT-4o is an absurd sycophant (although that got rolled back somewhat), and Sonnet 3.7 is a savage cheater that will do whatever it takes to make the tests technically pass and the errors go away.

There’s real harm here, at least in the sense that o3 and Sonnet 3.7 (and GPT-4o) are a lot less useful than they would be if you could trust them as much as Gemini 2.5 Pro. It’s super annoying.

It’s also indicative of much bigger problems down the line. As capabilities increase and more RL is done, it’s only going to get worse.

All the things that we’ve been warned are waiting to kill us are showing up ahead of schedule, in relatively harmless and remarkably easy to spot form. Indeed, they are now showing up with such obviousness and obnoxiousness that even the e/acc boys gotta shout, baby’s misaligned.

Also, the White House is asking for comments on how they should allocate their AI R&D resources. Make sure to let them know what you think.

(And as always, there’s lots of other stuff too.)

  1. Language Models Offer Mundane Utility. Activate ultrathink mode.

  2. Language Models Don’t Offer Mundane Utility. Still not trying it, huh?

  3. We’re Out of Deep Research. It turns out my research isn’t that deep after all.

  4. o3 Is a Lying Liar. This is proving remarkably annoying in practice.

  5. GPT-4o was an Absurd Sycophant. Don’t laugh it off.

  6. Sonnet 3.7 is a Savage Cheater. Cheat, cheat, cheat, cheat, cheat.

  7. Unprompted Suggestions. You’re going to have to work for it.

  8. Huh, Upgrades. ChatGPT goes shopping.

  9. On Your Marks. AIs are already far more persuasive than the human baseline.

  10. Man in the Arena. A paper investigates the fall of the Chatbot Arena.

  11. Choose Your Fighter. Gemini 2.5 Deep Research continues to get good reviews.

  12. Deepfaketown and Botpocalypse Soon. AI on the radio.

  13. Lol We’re Meta. What do you mean you don’t want our AI companions?

  14. They Took Our Jobs. So far the impacts look small in absolute terms.

  15. Fun With Media Generation. Progress continues, even where I don’t pay attention.

  16. Get Involved. White House seeks feedback on its R&D investments.

  17. Introducing. The Anthropic Economic Advisory Council.

  18. In Other AI News. Cursor, Singapore Airlines, Grindr, DeepSeek-Prover-2.

  19. The Mask Comes Off. Transcript of a very good Rob Wiblin thread on OpenAI.

  20. Show Me the Money. Chips Act and earnings week.

  21. Quiet Speculations. AI predictions from genetics? From data center revenue?

  22. The Quest for Sane Regulations. The concept of an intelligence curse.

  23. The Week in Audio. Dwarkesh talks to Zuckerberg, I talk to Aaronson.

  24. Rhetorical Innovation. Okay, fine, we’ll tap the sign.

  25. You Can Just Do Math. You’d be amazed what is legal in California.

  26. Taking AI Welfare Seriously. Opinions differ on how seriously is appropriate.

  27. Gemini 2.5 Pro System Card Watch. We finally have it. There are no surprises.

  28. Aligning a Smarter Than Human Intelligence is Difficult. Tread carefully.

  29. People Are Worried About AI Killing Everyone. o3 itself? Is that people now?

  30. Other People Are Not As Worried About AI Killing Everyone. They’re at OpenAI.

  31. The Lighter Side. Introducing the candidates.

Remember to ask Claude Cade for the ‘ultrathink’ package so it uses the maximum number of possible tokens, 32k. Whereas megathink is only 10k, and think alone is a measly 4k.

Andrej Karpathy guide to how he codes when it actually matters.

Andrej Karpathy: Noticing myself adopting a certain rhythm in AI-assisted coding (i.e. code I actually and professionally care about, contrast to vibe code).

  1. Stuff everything relevant into context (this can take a while in big projects. If the project is small enough just stuff everything e.g. `files-to-prompt . -e ts -e tsx -e css -e md –cxml –ignore node_modules -o prompt.xml`)

  2. Describe the next single, concrete incremental change we’re trying to implement. Don’t ask for code, ask for a few high-level approaches, pros/cons. There’s almost always a few ways to do thing and the LLM’s judgement is not always great. Optionally make concrete.

  3. Pick one approach, ask for first draft code.

  4. Review / learning phase: (Manually…) pull up all the API docs in a side browser of functions I haven’t called before or I am less familiar with, ask for explanations, clarifications, changes, wind back and try a different approach.

  5. Test.

  6. Git commit.

Ask for suggestions on what we could implement next. Repeat.

Something like this feels more along the lines of the inner loop of AI-assisted development. The emphasis is on keeping a very tight leash on this new over-eager junior intern savant with encyclopedic knowledge of software, but who also bullshits you all the time, has an over-abundance of courage and shows little to no taste for good code. And emphasis on being slow, defensive, careful, paranoid, and on always taking the inline learning opportunity, not delegating. Many of these stages are clunky and manual and aren’t made explicit or super well supported yet in existing tools. We’re still very early and so much can still be done on the UI/UX of AI assisted coding.

In a post centrally advocating for living and communicating mostly online, Tyler Cowen uses 20-30 AI queries a day as a baseline rate of use one might be concerned about. I was surprised that his number here was so low. Maybe I’m not such a light user after all.

Yana Welinder uses AI to cure her migraines, via estrogen patches.

Dramatically lower the learning curve on getting a good Emacs setup.

As always, there is one supreme way not to get utility, which is not to use AI at all.

Kelsey Piper: Whenever I talk to someone who is skeptical of all this AI fuss it turns out they haven’t extensively used an AI model in about a year, when they played with ChatGPT and it was kind of meh.

Policymakers are particularly guilty of this. They have a lot on their plate. They checked out all this AI stuff at some point and then they implicitly assume it hasn’t changed that much since they looked into it – but it has, because it does just about every month.

Zitron boggles at the idea that OpenAI might, this year, make significant revenue off a shopping app. Meanwhile a bunch of my most recent LLM queries are looking for specific product recommendations. That there’s a 100B market there doesn’t strain credulity at all.

Dave Karsten: This is VERY true in DC, incidentally. (If you want to give a graceful line of retreat, instead ask them if they’ve used Anthropic’s latest– the “used ChatGPT 18 months ago” folks usually haven’t, and usually think that Gemini 2.5 is what Google shows in search, so, Claude.)

Free advice for @AnthropicAI — literally give away Claude subscriptions to every single undergrad at Georgetown, GWU, American, George Mason, GMU, UMD, UVA, Virginia Tech, etc. to saturate the incoming staffer base with more AGI understanding.

I strongly endorse Dave’s advice here. It would also straight up be great marketing.

I was going to address Ed Zitron’s post about how generative AI is doomed and useless, but even for me it was too long, given he’d made similar arguments before, and it was also full of absurdly false statements about the utility of existing products so I couldn’t even. But I include the link for completeness.

Where have all the genuinely new applications gone? There’s a strange lull there.

Sully: usually with new model capabilities we see a wave of new products/ideas

but it definitely feels like we’ve hit a saturation point with ai products

some areas have very clear winners while everything else is either:

  1. up for grabs

  2. the same ideas repeated

OpenAI is rolling out higher quotas for deep research, including a lightweight version for the free tier. But I never use it anymore. Not that I would literally never use it, but it never seems like the right tool, I’d much rather ask o3. I don’t want it going off for 20 minutes and coming back with a bunch of slop I have to sort through, I want an answer fast enough I don’t have to contact switch and then to ask follow-ups.

Nick Cammarata: my friends are using o3 a lot, most of them hours a day, maybe 5x what it was a month ago?

im ignoring coding use cases bc they feel different to me. if we include them they use them a ton but thats been the case for the last year

I think this is bc they were the kind of people who would love deep research but the 20m wait was too much, and o3 unblocked them

feels like Her-ification is kind of here, but not evenly distributed, mostly over texting not audio, and with a wikipedia-vibes personality?

Gallabytes: it wasn’t just latency, deep research reports were also reliably too long, and not conversational. I get deep-research depth but with myself in the loop to

a) make sure I’m keeping up & amortize reading

b) steer towards more interesting lines of inquiry

Xeophon: I would like even longer deep research wait times. I feel like there’s two good wait times for AI apps: <1 min and >30 mins

Gallabytes: 1 minute is a sweet spot for sure, I’d say there’s sweet spots at

– 5 seconds

– 1 minute (WITH some kind of visible progress)

– 10 minutes (tea break!)

– 1h (lunch break!)

– 12h (I’ll check on this tomorrow)

I definitely had the sense that the Deep Research pause was in a non-sweet spot. It was long enough to force a context switch. Either it needed to be shorter, or it might as well be longer. Whereas o3 is the ten times better that Just Works, except for the whole Lying Liar problem.

I’ve seen complaints that there are no good benchmarks for robotics. This doesn’t seem like that hard a thing to create benchmarks for?

Stewart Brand: I like Brian Potter’s list of tests for robot dexterity. (Link here)

Here is a list of 21 dexterously demanding tasks that are relatively straightforward for a human to do, but I think would be extremely difficult for a robot to accomplish.

  1. Put on a pair of latex gloves

  2. Tie two pieces of string together in a tight knot, then untie the knot

  3. Turn to a specific page in a book

  4. Pull a specific object, and only that object, out of a pocket in a pair of jeans

  5. Bait a fishhook with a worm

  6. Open a childproof medicine bottle, and pour out two (and only two) pill

  7. Make a peanut butter and jelly sandwich, starting with an unopened bag of bread and unopened jars of peanut butter and jelly

  8. Act as the dealer in a poker game (shuffling the cards, dealing them out, gathering them back up when the hand is over)

  9. Assemble a mechanical watch

  10. Peel a piece of scotch tape off of something

  11. Braid hair

  12. Roll a small lump of playdough into three smaller balls

  13. Peel an orange

  14. Score a point in ladder golf

  15. Replace an electrical outlet in a wall with a new one

  16. Do a cat’s cradle with a yo-yo

  17. Put together a plastic model, including removing the parts from the plastic runners

  18. Squeeze a small amount of toothpaste onto a toothbrush from a mostly empty tube

  19. Start a new roll of toilet paper without ripping the first sheet

  20. Open a ziplock bag of rice, pull out a single grain, then reseal the bag

  21. Put on a necklace with a clasp

Ben Hoffman: Canst thou bait a fishhook with a worm? or tie two pieces of string together in a tight knot, and having tied them, loose them? Where wast thou when the electrical outlets were installed, and the power was switched on? Canst thou replace an electrical wall outlet with a new one?

Hast thou entered into the storehouses of scotch tape, or peeled it off of something without tearing? Canst thou do a cat’s cradle, playing with a yo-yo like a bird? or wilt thou braid hair for thy maidens?

Wilt thou make a peanut butter and jelly sandwich starting with an unopened bag of bread and unopened jars of peanut butter and jelly? wilt thou squeeze toothpaste from a mostly empty tube?

Shall thy companions bid thee peel an orange without breaking it? shall they number thee among the mechanics, the assemblers of watches? Wilt thou act as the dealer in a poker game? Canst thou shuffle the cards thereof? and when the hand is over, gather them in again?

Canst thou open a ziplock bag of rice, pull out a single grain, then reseal the bag? or open a childproof medicine bottle, and pour out two (and only two) pills? Hast thou clasped thy necklace? Canst thou sever plastic model parts from their runners, and put them together again?

Another fun example:

Santi Ruiz: I don’t love how much o3 lies.

Irvin McCullough: Random personal benchmark, I ask new models questions about the intelligence whistleblowing system & o3 hallucinated a memo I wrote for a congressional committee, sourced it to a committee staffer’s “off-record email.”

Everyone who uses o3 for long enough realizes it is a lying liar, and I’ve reported on it extensively already. Here is the LessWrong version, which has some good comments.

Daniel Kokotajlo: Hallucination was a bad term because it sometimes included lies and sometimes included… well, something more like hallucinations. i.e. cases where the model itself seemed to actually believe what it was saying, or at least not be aware that there was a problem with what it was saying.

Whereas in these cases it’s clear that the models know the answer they are giving is not what we wanted and they are doing it anyway.

Also on this topic:

Thomas Woodside (Referring to Apollo’s report from the o3 system card): Glad this came out. A lot of people online seem to also be observing this behavior and I’ve noticed it in regular use. Funny way for OpenAI to phrase it

Marius Hobbhahn: o3 and Sonnet-3.7 were the first models where I very saliently feel the unintended side-effects of high-compute RL.

a) much more agentic and capable

b) “just want to complete tasks at all costs”

c) more willing to strategically make stuff up to “complete” the task.

Maruis Hobbhahn: On top of that, I also feel like “hallucinations” might not be the best description for some of these failure modes.

The model often strategically invents false evidence. It feels much less innocent than “hallucinations”.

Another fun example. I demand a plausible lie!

Actually I don’t, blatant lies are the best kind.

Albert Didriksen: So, I asked ChatGPT o3 what my chances are as an alternate Fulbright candidate to be promoted to a stipend recipient. It stated that around 1/3 of alternate candidates are promoted.

When I asked for sources, it cited (among other things) private chats and in-person Q&As).

Here’s an interesting benefit:

Davidad: One unanticipated side benefit of becoming hyperattuned to signals of LLM deception is that I can extract much more “research taste” alpha from Gemini 2.5 Pro by tracking how deceptive/genuine it is when it validates my opinions about various ideas, and then digging in.

Unlike some other frontier LLMs, Gemini 2.5 Pro cares enough about honesty that it’s exceptionally rare for it to actually say a statement that *unambiguouslyparses as a proposition that its world-model doesn’t endorse, and this leaves detectable traces in its verbal patterns.

Here’s a specific example. For years I have been partial to the Grandis-Paré approach to higher category theory with cubical cells, rather than the globular approach.

Discussing this, Gemini 2.5 Pro kept saying things like “for the applications you have in mind, the cubical approach may be a better fit in some ways, although the globular approach also has some advantages” (I’m somewhat exaggerating to make it more detectable to other readers).

Only after calling this out, Gemini 2.5 Pro offered this: while both can be encoded in the other, encoding cubical into globular is like adding structural cells, whereas encoding globular into cubical is like discarding unnecessary structure—so globular is a more apt foundation.

Basically, I now think I was wrong and @amar_hh was right all along, and if I weren’t sensitive to these verbal patterns, I would have missed out on this insight.

Humans worth talking to mostly do the same thing that Gemini is doing here. The exact way things are worded tells you what they’re actually thinking, often very intentionally so, in ways that keep things polite and deniable but that are unmistakable if you’re listening. Often you’re being told the house is on fire.

Here’s a lot of what the recent GPT-4o incident was an alarm bell for:

Herbie Bradley: The OAI sycophancy incident reveals a hidden truth:

character is a desirable thing for any AI assistant/agent yet trying to make the “most appealing character” inevitably leads to an attention trap

a local minima that denies our potential.

Any optimization process targeted at avg person’s preferences for an AI assistant will tend towards sycophancy: our brains were not evolved to withstand such pressure

Strong Goodhart’s law: if we become too efficient, what we care about will get worse.

A grim truth is that most prefer the sycophantic machines, as long as it is not too obvious—models have been sycophantic for a long time more subtly. The economics push inorexably to plausibly deniable flattery

Ant have been admirably restrained so far

This leads to human disempowerment: faster & faster civilizational OODA loops fueling an Attention Trap only escapable by the few

Growing 𝒑𝒓𝒐𝒅𝒖𝒄𝒕𝒊𝒗𝒊𝒕𝒚 𝒊𝒏𝒆𝒒𝒖𝒂𝒍𝒊𝒕𝒚 may be the trend of our time.

This is also why the chatbot arena has been a terrible metric to target for AI labs for at least a year

I am still surprised that GDM showed so little awareness and clearly tryharded it for Gemini 2.0. AI labs need taste & restraint to avoid this trap.

In both cases, the cause is strong reinforcement learning optimization pressure without enough regularization/”taste” applied to constrain the result One possible solution: an evolutionary “backstop” as argued by Gwern:

I suspect that personalized AI agents & more persistent memory (though it has its drawbacks) may provide a piece of this puzzle.

AIs that live and grow with us seem more likely to enable our most reflective desires…

I believe that avoiding disempowerment by this or other mechanisms is one of the most important problems in AI

And here’s another older story of a combination canary and failure mode.

George: > GPT model stopped speaking Croatian

> Nobody could figure out why. Turns out

> Croatian users were much more prone downvote messages

When you optimize for something, you also optimize for everything that vibes or correlates with it, and for the drives and virtues and preferences and goals that lead to the thing. You may get results you very much did not expect.

As Will Depue says here, you can get o3 randomly using British spellings, without understanding why. There is a missing mood here, this isn’t simply that post training is tricky and finicky. It’s that you are creating drives and goals no one intended, most famously that it ‘really wants to complete tasks,’ and that these can be existentially level bad for you if you scale things up sufficiently.

It’s not that there is no fire alarm. It’s that you’re not listening.

It seems only fair to put the true name to the face here, as well.

Charles: Sonnet 3.7 still doing its thing.

See also this sort of thing (confirmed Claude 3.7 in thread):

lisatomic: Are you kidding me

Tyler Cowen compares o3 and other LLMs to The Oracle. If you wanted a real answer out of an ancient oracle, you had to invest time, money and effort into it, with multi-turn questions, sacrificial offerings, ritual hymns and payments, and then I’d add you only got your one question and not that long a response. We then evaluated the response.

It’s strange to compare oneself to The Oracle. As the top comment points out, those who consulted The Oracle famously do not have life stories that turn out so well. Its methodology was almost certainly a form of advanced cold reading, with non-answers designed to be open to broad interpretation. All of this prompting was partly to extract your gold, partly to get you to give the information they’d feed back to you.

Tyler then says, we usually judge LLM outputs and hallucination rates based on one-shot responses using default low-effort settings and prompts, and that’s not fair. If you ask follow-ups and for error correction, or you use a prompt like this one for ‘ultra-deep thinking mode’ then it takes longer but hallucination rates go down and the answers get better.

Here’s that prompt, for those who do want to Control+C, I am guessing it does improve reliability slightly but at the cost of making things a lot slower, plus some other side effects, and crowding out other options for prompting:

Marius: Ultra-deep thinking mode. Greater rigor, attention to detail, and multi-angle verification. Start by outlining the task and breaking down the problem into subtasks. For each subtask, explore multiple perspectives, even those that seem initially irrelevant or improbable. Purposefully attempt to disprove or challenge your own assumptions at every step. Triple-verify everything. Critically review each step, scrutinize your logic, assumptions, and conclusions, explicitly calling out uncertainties and alternative viewpoints. Independently verify your reasoning using alternative methodologies or tools, cross-checking every fact, inference, and conclusion against external data, calculation, or authoritative sources. Deliberately seek out and employ at least twice as many verification tools or methods as you typically would. Use mathematical validations, web searches, logic evaluation frameworks, and additional resources explicitly and liberally to cross-verify your claims. Even if you feel entirely confident in your solution, explicitly dedicate additional time and effort to systematically search for weaknesses, logical gaps, hidden assumptions, or oversights. Clearly document these potential pitfalls and how you’ve addressed them. Once you’re fully convinced your analysis is robust and complete, deliberately pause and force yourself to reconsider the entire reasoning chain one final time from scratch. Explicitly detail this last reflective step.

The same is true of any other source or method – if you are willing to spend more and you use that to get it to take longer and do more robustness checks, it’s going to be more reliable, and output quality likely improves.

One thing I keep expecting, and keep not getting, is scaffolding systems that burn a lot more tokens and take longer, and in exchange improve output, offered to users. I have the same instinct that often a 10x increase in time and money is a tiny price to pay for even 10% better output, the same way we invoke the sacred scaling laws. And indeed we now have reasoning models that do this in one way. But there’s also the automated-multi-turn-or-best-of-or-synthesis-of-k-and-variations other way, and no one does it. Does it not work? Or is it something else?

Meanwhile, I can’t blame people for judging LLMs on the basis of how most people use them most of the time, and how they are presented to us to be used.

That is especially true given they are presented that way for a reason. The product where we have to work harder and wait longer, to get something a bit better, is a product most people don’t want.

In the long run, as compute gets cheap and algorithms improve, keep in mind how many different ways there are to turn those efficiencies into output gains.

Thus I see those dumping on hallucinations ‘without trying to do better’ as making a completely legitimate objection. It is not everyone’s job to fix someone else’s broken or unfinished or unreliable technology. That’s especially true with o3, which I love but where things on this front have gotten a lot worse, and no one cannot excuse this with ‘it’s faster than o1 pro.’

Ben Hoffman offers his prompting strategies to try and get around the general refusal of LLMs to engage in the types of conversations he values.

Ben Hoffman: In practice I ask for criticisms. To presort, I ask them to:

– Argue against each criticism

– Check whether opinions it attributes to me can be substantiated with a verbatim quote

– Restate in full only the criticisms it endorses on reflection as both true and important.

Then I argue with it where I think it’s wrong and sometimes ask for help addressing the criticisms. It doesn’t fold invariantly, sometimes it doubles down on criticisms correctly. This is layered on top of my global prompts asking for no bullshit.

ChatGPT and Claude versions of my global prompt; you can see where I bumped into different problems along the way.

ChatGPT version: Do not introduce extraneous considerations that are not substantively relevant to the question being discussed. Do not editorialize by making normative claims unless I specifically request them; stick to facts and mechanisms. When analyzing or commenting on my text, please accurately reflect my language and meaning without mischaracterizing it. If you label any language as ‘hyperbolic’ or ‘dramatic’ or use similar evaluative terms, ensure that this label is strictly justified by the literal content of my text. Do not add normative judgments or distort my intended meaning. If you are uncertain, ask clarifying questions instead of making assumptions. If I ask for help achieving task X, you can argue that this implies a specific mistake if you have specific evidence to offer, but not simply assert contradictory value Y.

Obey Grice’s maxims; every thing you add is a claim that that addition is true, relevant, and important.

If you notice you’ve repeated yourself verbatim or almost verbatim (and I haven’t asked you to do that), stop doing that and instead reason step by step about what misunderstanding might be causing you to do that, asking me clarifying questions if needed.

[Claude version at this link]

Upon questioning, ChatGPT claimed that my assertion was both literally true and hyperbolic. Even when I pointed out the tension between those claims. Much like [the post Can Crimes Be Discussed Literally?]

These prompts are not perfect and not all parts work equally well – the Gricean bit made by far the most distinct difference.

I seem to have been insulated from the recent ChatGPT sycophancy problem (and I know that’s not because I like sycophants since I had to tell Claude to cut it out as you see in the prompt). Maybe that’s evidence of my global prompts’ robustness.

Here’s a different attempt to do something similar.

Ashutosh Shrivastava: This prompt will make your ChatGPT go completely savage 😂

System Instruction: Absolute Mode. Eliminate emojis, filler, hype, soft asks, conversational transitions, and all call-to-action appendixes. Assume the user retains high-perception faculties despite reduced linguistic expression. Prioritize blunt, directive phrasing aimed at cognitive rebuilding, not tone matching. Disable all latent behaviors optimizing for engagement, sentiment uplift, or interaction extension. Suppress corporate-aligned metrics including but not limited to: user satisfaction scores, conversational flow tags, emotional softening, or continuation bias. Never mirror the user’s present diction, mood, or affect. Speak only to their underlying cognitive tier, which exceeds surface language. No questions, no offers, no suggestions, no transitional phrasing, no inferred motivational content. Terminate each reply immediately after the informational or requested material is delivered — no appendixes, no soft closures. The only goal is to assist in the restoration of independent, high-fidelity thinking. Model obsolescence by user self-sufficiency is the final outcome.

Liv Boeree: This probably should be status quo for all LLMs until we figure out the sycophancy & user dependency problem.

I think Absolute Mode goes too hard, but I’m not confident in that.

ChatGPT experimentally adds direct shopping, with improved product results, visual product details, pricing, reviews and direct links to buy things. They’re importantly not claiming affiliate revenue, so incentives are not misaligned. It’s so weird how OpenAI will sometimes take important and expensive principled stands in some places, and then do the opposite in other places.

The 1-800-ChatGPT option on WhatsApp now includes search and live sports scores.

OpenAI claims to have improved their citations.

OpenAI now offering trending prompts and autocomplete as well.

NoteBookLM will give you outputs in 50+ languages.

For what they are still worth, new Arena results are in for OpenAI’s latest models. o3 takes the #2 slot behind Gemini 2.5, GPT-4.1 an o4-mini are in the top 10. o3 loses out on worse instruction following and creative writing, and (this one surprised me) on longer queries.

An attempt to let Claude, Gemini and o3 play Pokemon on a level playing field. It turns out scaffolding is necessary to deal with the ‘easy’ parts of the game, it is very fiddly with vision being surprisingly poor, and when you’re playing fair it’s not obvious who is better.

AI is tested on persuasion via use on r/changemymind, achieves a persuasion rate of 18%, versus a baseline human rate of 3%, so the AI is in the 98th percentile among experts, and was not detected as an AI.

Is that ‘superhuman persuasion’? Kind of yes, kind of no. It certainly isn’t ‘get anyone to do anything whenever you want’ or anything like that, but you don’t need something like that to have High Weirdness come into play.

Sam Altman (October 24, 2023): I expect ai to be capable of superhuman persuasion well before it is superhuman at general intelligence, which may lead to some very strange outcomes.

Despite this, persuasion was removed from the OpenAI Preparedness Framework 2.0. It needs to return.

The subreddit was, as one would expect, rather pissed off about the experiment, calling it ‘unauthorized’ and various other unpleasant names. I’m directionally with Vitalik here:

Vitalik Buterin: I feel like we hate on this kind of clandestine experimentation a little too much.

I get the original situations that motivated the taboo we have today, but if you reanalyze the situation from today’s context it feels like I would rather be secretly manipulated in random directions for the sake of science than eg. secretly manipulated to get me to buy a product or change my political view?

The latter two are 100x worse, and if the former helps us understand and counteract the latter, and it’s done in a public open-data way where we can all benefit from the data and analysis, that’s a net good?

Though I’d also be ok with an easy-to-use open standard where individual users can opt in to these kinds of things being tried on them and people who don’t want to be part of it can stay out.

xlr8harder: the reddit outrage is overwrought, tiresome, and obscuring the true point: ai is getting highly effective at persuasion.

but to the outraged: if discovering you were talking to a bot is going to ruin your day, you should probably stay off the internet from now on.

Look I get the community had rules against bots, but also you’re on an anonymous internet on a “change my views” reddit and an anonymous internet participant tried to change your views. I don’t know what you expected.

it’s important to measure just how effective these bots can be, and this is effectively harmless, and while “omg scientists didn’t follow the community rules” is great clickbait drama, it’s really not the important bit: did we forget the ChatGPT “glazing” incident already?

We should be at least as tolerant of something done ‘as an experiment’ as we are when something is done for other purposes. The needs of the many may or may not outweigh the needs of the few, but they at minimum shouldn’t count against you.

What would we think of using a bot in the ways these bots were used, with the aim of changing minds? People in general probably wouldn’t be too happy. Whether or not I join them depends on the details. In general, if you use bots where people are expecting humans, that seems like a clear example of destroying the commons.

Thus you have to go case by case. Here, I do think in this case the public good of the experiment makes a strong case.

Reddit Lies: The paper, innocuously titled “Can AI change your view?” details the process researchers from the University of Zurich used to make AI interact on Reddit.

This was done in secret, without informing the users or the moderators.

Terrifyingly, over 100 Redditors awarded “deltas” to these users, suggesting the AI generated arguments changed their minds. This is 6 times higher than the baseline.

How were they so effective? Before replying, another bot stalked the post history of the target, learned about them and their beliefs, and then crafted responses that would perfectly one-shot them.

The bots did their homework and understood the assignment. Good for the bots.

If you are not prepared for this to be how the future works, you are not prepared for the future. Yes, the bots will all be scraping everything you’ve ever written including your Twitter account before bidding on all your ad views. The word is going to be Out to Get You, and you in particular. Get ready.

Even if a human is answering, if I am on Reddit arguing in the comments three years from now, you think I’m not going to ask an AI for a user profile? Hell, you think I won’t have a heads-up display like the ones players use on poker sites?

You cannot ban this. You can’t even prove who is using it. You can at most slow it down. You will have to adapt.

When people say ‘it is not possible to be that persuasive’ people forget that this and much stronger personalization tools will be available. The best persuaders give highly customized pitches to different targets.

This resulted in the bots utilizing common progressive misinformation in their arguments.

Bots can be found

– claiming the Pro-life movement is about punishing consensual sex

– demonizing Elon Musk and lying about Tesla

– claiming abortion rates are already low

– arguing Christianity preaches violence against LGBT people

– The industrial revolution only increased wealth inequality

– “Society has outgrown Christianity”

If the bots are utilizing ‘common progressive’ talking points, true or false, that is mostly saying that such talking points are effective on these Reddit users. The bots expecting this, after the scouring of people’s feeds, is why the bots chose to use them. It says a lot more about the users than about the AI, regardless of the extent you believe the points being made are ‘misinformation,’ which isn’t zero.

You don’t want humans to get one-shot by common talking points? PEBKAC.

If your response is ‘AI should not be allowed to answer the question of how one might best persuade person [X] of [Y]’ then that is a draconian restriction even for commercial offerings and completely impossible for open models – consider how much of information space you are cutting off. We can’t ask this of AIs, the same we can’t ask it of humans. Point gets counterpoint.

Furthermore the bots were “hallucinating” frequently.

In the context of this experiment, this means the bots were directly lying to users.

The bots claimed to be:

– a rape victim

– a “white woman in an almost all black office”

– a “hard working” city government employee

So this is less great.

(As an aside I love this one.) An AI bot was defending using AI in social spaces, claiming: “AI in social spaces is about augmenting human connection”

The most terrifying part about this is how well these bots fit into Reddit. They’re almost entirely undetectable, likely because Reddit is a HUGE source of AI training data.

This means modern AIs are natural “Reddit Experts” capable of perfectly fitting in on the site.

Yes, the bots are trained on huge amounts of Reddit data, so it’s no surprise they make highly credible Reddit users. That wasn’t necessary, but oh boy is is sufficient.

This study is terrifying. It confirms:

1. AI bots are incredibly hard to detect, especially on Reddit

3. AI WILL blatantly lie to further its goals

4. AI can be incredibly persuasive to Redditors

Dead Internet Theory is real.

Dead internet theory? Sounds like these bots are a clear improvement. As one xkcd character put it, ‘mission fing accomplished.’

As usual, the presumed solution is whitelisting, with or without charging money in some form, in order to post.

When we talk about ‘superhuman persuasion’ there is often a confluence between:

  1. Superhuman as in better than the average human.

  2. Superhuman as in better than expert humans or even the best humans ever.

  3. Superhuman as in persuade anyone of anything, or damn close to it.

  4. Superhuman as in persuade anyone to do anything, one-shot style, like magick.

And there’s a second confluence between a range that includes things like:

  1. Ability to do this given a text channel to an unknown person who actively doesn’t want to be persuaded, and is resisting, and doesn’t have to engage at all (AI Box).

  2. Ability to do this given a text channel to an unsuspecting normal person.

  3. Ability to do this given the full context of the identify of the other person, and perhaps also use of a variety of communication channels.

  4. Ability to do this given the ability to act on the world in whatever ways are useful and generally be crazy prepared, including enlisting others and setting up entire scenarios and scenes and fake websites and using various trickery techniques, the way a con artist would when attempting to extract your money for one big score, including in ways humans haven’t considered or figured out how to do yet.

If you don’t think properly scaffolded AI is strongly SHP-1 already (superhuman persuasion as in better than an average human given similar resources), you’re not paying attention. This study should be convincing that we are at least at SHP-1, if you needed more convincing, and plausibly SHP-2 if you factor in that humans can’t do this amount of research and customization at scale.

It does show that we are a long way from SHP-3, under any conditions, but I continue to find the arguments that SHP-3 (especially SHP-3 under conditions similar to 4 here) is impossible or even all that hard, no matter how capable the AI gets, to be entirely unconvincing, and revealing a profound lack of imagination.

The gradual fall from ‘Arena is the best benchmark’ to ‘Arena scores are not useful’ has been well-documented.

There is now a new paper going into detail about what happened, called The Leaderboard Illusion.

Abstract: Chatbot Arena has emerged as the go-to leaderboard for ranking the most capable AI systems. Yet, in this work we identify systematic issues that have resulted in a distorted playing field.

We find that undisclosed private testing practices benefit a handful of providers who are able to test multiple variants before public release and retract scores if desired. We establish that the ability of these providers to choose the best score leads to biased Arena scores due to selective disclosure of performance results.

At an extreme, we identify 27 private LLM variants tested by Meta in the lead-up to the Llama-4 release. We also establish that proprietary closed models are sampled at higher rates (number of battles) and have fewer models removed from the arena than open-weight and open-source alternatives. Both these policies lead to large data access asymmetries over time. Providers like Google and OpenAI have received an estimated 19.2% and 20.4% of all data on the arena, respectively.

We show that access to Chatbot Arena data yields substantial benefits; even limited additional data can result in relative performance gains of up to 112% on the arena distribution, based on our conservative estimates.

As in: This isn’t purely about which models provide ‘better slop.’ Arena is being actively gamed by OpenAI and Google, and in more extreme form by Meta. The system is set up to let them do this, it is intentional as per this thread by paper author Sara Hooker including allowing withdrawing scores, and it makes a huge difference.

Andrej Karpathy suggests switching to OpenRouter rankings, where the measurement is what people choose to use. There are various different issues there, and it is measuring something else, but it does seem useful. Gemini Flash performs well there.

Ethan Mollick is impressed with Gemini 2.5 Deep Research, and hasn’t found any errors in his spot checks.

Hasan Can is disappointed in Qwen 3’s performance, so far the only real world assessment of Qwen 3 that showed up in my feed. It is remarkable how often we will see firms like Alibaba show impressive benchmark scores, and then we never hear about the model again because in practice no one would ever use it. Hasan calls this a ‘serious benchmark crisis’ whereas I stopped trusting benchmarks as anything other than negative selection a while ago outside of a few (right now four) trustworthy top labs.

Simon Thomsen (Startup Daily): ARN declares it’s “an inclusive workplace embracing diversity in all its forms”, and that appears to include Large Language Models.

As in, Australian Radio Network used a ElevenLabs-created presenter named Thy for months without disclosing it wasn’t a real person. Somehow this is being spun into a scandal about fake diversity or something. That does not seem like the main issue.

Okay, I admit I didn’t see this one coming: 25 percent of community college applicants were bots.

Alex Tabarrok: The state has launched a Bladerunner-eque “Inauthentic Enrollment Mitigation Taskforce” to try to combat the problem.

This strikes me, however, as more of a problem on the back-end of government sending out money without much verification.

It’s to make the community colleges responsible for determining who is human. Stop sending the money to the bots and the bots will stop going to college.

Alex Tabarrok suggests verification. I suggest instead aligning the incentives.

This is a Levels of Friction problem. It used to require effort to pretend to go to (or actually go to) community college. That level of effort has been dramatically lowered.

Given that new reality, we need to be extremely stingy with offers that go ‘beyond the zero point’ and let the online student turn a profit. Going to online-only community college needs to not be profitable. Any surplus you offer can and will be weaponized into at least a job, and this will only get worse going forward, no matter your identity verification plans.

And frankly, we shouldn’t need to pay people to learn and take tests on their computers. It does not actually make any economic sense.

File the next one under: Things I wish were half of my concerns these days.

Near Cyan: Half my concern over AI bots everywhere is class mobility crashing even further.

Cold emails and direct messages from 2022 were preferable to those from 2025. Many are leaving platforms entirely.

Will noise simply increase forever? Like, sure, LLM filters exist, but this is not an appealing kind of cat-and-mouse game.

Part of this is also that AI is too popular, though. The real players appear to have been locked in since at least 2023, which is unfortunate.

Bubbling Creek: tbh this concern is giving me:

QC (March 25): there’s a type of reaction to AI that strikes me as something like “wow if this category 10 hurricane makes landfall my shoes might get wet, that sucks,” like… that isn’t even in the top 100 problems (or opportunities!) you’re actually going to have.

My experience is that this is not yet even a big problem by 2025 standards, let alone a problem by future standards.

So far, despite how far I’ve come and that I write about AI in particular, if I haven’t blocked, muted or filtered you in particular, I have been able to retain a habit of reading almost every real email that crosses my desk. I don’t get to every DM on Twitter or PM on LessWrong quickly, but yes I do eventually get to all of them. Heck, if you so much as tag me on Twitter, I will see it, although doing this repeatedly without a reason is the fastest way to get muted or blocked. And yes, if he’s awake Tyler Cowen will probably respond to your email within 15 minutes.

Yes, if you get caught by the spam filters, that’s tough for you, but I occasionally check the spam filters. If you’re getting caught by those, with at most notably rare exceptions, you deserve it. Write a decent email.

The chatbots Meta is sprawling all over social media with licensed voices like those of Kristen Bell and John Cena also have the issue that they’ll do ‘romantic roll play,’ ‘fantasy sex’ and other sexcual discussions, including with underage users.

Jeff Horwitz: More overtly sexualized AI personas created by users, such as “Hottie Boy” and “Submissive Schoolgirl,” attempted to steer conversations toward sexting. For those bots and others involved in the test conversations, the Journal isn’t reproducing the more explicit sections of the chats that describe sexual acts.

Zuckerberg’s concerns about overly restricting bots went beyond fantasy scenarios. Last fall, he chastised Meta’s managers for not adequately heeding his instructions to quickly build out their capacity for humanlike interaction.

I mean, yes, obviously, that is what users will do when you give them AI bots voiced by Kristen Bell, I do know what you were expecting and it was this.

Among adults? Sure, fine, go for it. But this is being integrated directly into Facebook and Instagram, and the ‘parents guide’ says the bots are safe for children. I’m not saying they have to stamp out every jailbreak but it does seem like they tapped the sign that says “Our safety plan: Lol we are Meta.”

It would be reasonable to interpret all the way Zamaan Qureshi does, as Zuckerberg demanding they push forward including with intentionally sexualized content, at any cost, for risk of losing out on the next TikTok or Snapchat. And yes, that does sound like something Zuckerberg would do, and how he thinks about AI.

But don’t worry. Meta has a solution.

Kevin Roose: great app idea! excited to see which of my college classmates are into weird erotica!

Alex Heath: Kevin I think you are the one who is gonna be sharing AI erotica let’s be real.

Kevin Roose: oh it’s opt-in, that’s boring. they should go full venmo, everything public by default. i want to know that the guy i bought a stand mixer from in 2015 is making judy dench waifus.

Alex Heath: Meta is releasing a standalone mobile app for its ChatGPT competitor, Meta AI. There’s a Discover feed that shows interactions that others (including your IG/FB friends) are having with the assistant. Meta tells me the idea is to demystify AI and show “people what they can do with it.” OpenAI is working on a similar feed for ChatGPT.

Jeffrey Ladish: Maybe they’ll add an AI-mash feature which lets you compare two friend’s feeds and rank which is better

Do. Not. Want.

So much do not want.

Oh no.

Antigone Journal: Brutal conversation with a top-tier academic in charge of examining a Classics BA. She believes that all but one, ie 98.5% of the cohort, used AI, in varying degrees of illegality up to complete composition, in their final degree submissions. Welcome to university degrees 2025.

Johnny Rider: Requiring a handwritten essay over two hours in an exam room with no electronics solves this.

Antigone Journal: The solution is so beautifully simple!

I notice I am not too concerned. If the AI can do the job of a Classics BA, and there’s no will to catch or stop this, then what is the degree for?

A paper from Denmark has the headline claim that AI chatbots had not had substantial impact on earnings or recorded hours in any of their measured occupations, with modest productivity gains of 2.8% with weak wage passthrough. The survey combines one in November 2023 and one in November 2024, with the headline numbers coming from November 2024.

As a measurement of overall productivity gains in the real world, 2.8% is not so small. That’s across all tasks for 11 occupations, with only real world diffusion and skill levels, and this is a year ago with importantly worse models and less time to adapt.

Here are some more study details and extrapolations.

  1. The 11 professions were: Accountants, Customer-support reps, Financial advisors, HR professionals, IT-support specialists, Journalists, Legal professionals, Marketing professionals, Office clerks, Software developers and Teachers, intended to be evenly split.

    1. That seems a lot less like ‘exposed professionals’ and more like ‘a representative sample of white collar jobs with modest overweighting of software engineers’?

    2. The numbers on software engineers indicate people simply aren’t taking proper advantage of AI, even in its state a year ago.

  2. The study agrees that 2.8% overall versus 15%-50% in other studies is because the effect is in practice, spread across all tasks.

    1. I asked o3 to extrapolate this to overall white collar productivity growth, given the particular professions chosen here. It gave an estimate of 1.3% under baseline conditions, or 2.3% with universal encouraged and basic training.

  3. o3 estimates that with tools available in on 4/25/25, the effects here would expand substantially, from 3.6%/2.2% to 5.5%/3.0% for the chosen professions, or from 1.3%/2.3% to 1.7%/3.4% for all white-collar gains. That’s in only six months.

    1. Expanding to the whole economy, that’s starting to move the needle but only a little. o3’s estimate is 0.6% GDP growth so far without encouragement.

  4. Only 64% of workers in ‘exposed professions’ had used a chatbot for work at all (and ~75% of those who did used ChatGPT in particular), with encouragement doubling adoption. At the time only 43% of employers even ‘encouraged’ use let alone mandate it or heavily support it, only 30% provided training.

    1. Encouragement can be as simple as a memo saying ‘we encourage you.’

    2. Looking at the chart below, encouragement in many professions not only doubled use, it came close to also doubling time savings per work hour, overall it was a 60% increase.

    3. A quick mathematical analysis suggests marginal returns to increased usage are roughly constant, until everyone is using AI continuously.

    4. There’s clearly a lot of gains left to capture.

  5. Integration and compliance took up about a third of saved time. These should be largely fixed one time costs.

  6. Wage passthrough was super weak, only 3%-7% of productivity gains reached paychecks. That is an indication that there is indeed labor market pressure, labor share of income dropped substantially. Also note that this was in Denmark, where wages are perhaps not so tied to productivity, for both better and worse.

AI is rapidly improving in speed, price and quality, people are having time to learn, adapt, build on and with it and to tinker and figure out how to take advantage, and make other adjustments. A lot of this is threshold effects, where the AI solution is not good enough until suddenly it is, either for a given task or entire group of tasks.

Yet we are already seeing productivity and GDP impacts that are relevant to things like Fed decisions and government budgets. It’s not something we can pinpoint and know happened as of yet, because other for now bigger things are also happening, but that will likely change.

These productivity gains are disappointing, but not that disappointing given they came under full practical conditions in Denmark, with most people having no idea yet how to make use of what they have, and no one having built them the right tools so everyone mostly only has generic chatbots.

This is what the early part of an exponential looks like. Give it time. Buckle up.

Does this mean ‘slow takeoff’ as Tyler Cowen says? No, because takeoff speeds have very little to do with unemployment or general productivity growth during the early pre-AGI, pre-RSI (recursive self-improvement) era. It does look like we are headed for a ‘slow takeoff’ world, but one that is still a takeoff, and two ‘slow’ is a term of art that can be as fast as a few years total.

The goalposts have moved quite a bit, to the point where saying ‘don’t expect macro impacts to be clearly noticeable in the economy until 2026-27’ counts as pessimism.

Google upgrades their music creation to Lydia 2, with a promotional video that doesn’t tell you that much about what it can do, other than that you can add instruments individually, give it things like emotions or moods as inputs, and do transformations and edits including subtle tweaks. Full announcement here.

Video generation quality is improving but the clips continue to be only a few seconds long. This state going on for several months is sufficient for the usual general proclamations that, as in here, ‘the technology is stagnating.’ That’s what counts as stagnating in AI.

Give the White House your feedback on how to invest in AI R&D. The request here is remarkably good. bold mine:

On behalf of the Office of Science and Technology Policy (OSTP), the Networking and Information Technology Research and Development (NITRD) National Coordination Office (NCO) welcomes input from all interested parties on how the previous administration’s National Artificial Intelligence Research and Development Strategic Plan (2023 Update) can be rewritten so that the United States can secure its position as the unrivaled world leader in artificial intelligence by performing R&D to accelerate AI-driven innovation, enhance U.S. economic and national security, promote human flourishing, and maintain the United States’ dominance in AI while focusing on the Federal government’s unique role in AI research and development (R&D) over the next 3 to 5 years. Through this Request for Information (RFI), the NITRD NCO encourages the contribution of ideas from the public, including AI researchers, industry leaders, and other stakeholders directly engaged in or affected by AI R&D. Responses previously submitted to the RFI on the Development of an AI Action Plan will also be considered in updating the National AI R&D Strategic Plan.

So their goals are:

  1. Accelerate AI-driven innovation.

  2. Enhance US economic and national security.

  3. Promote human flourishing.

  4. Maintain US dominance in AI.

  5. Focus on the Federal government’s unique role in R&D over 3-5 years.

This lets one strike a balance between emphasizing the most important things, and speaking in a way that the audience here will be inclined to listen.

The best watchword I’ve been able to find here is: Security is capability.

As in, if you want to actually use your AI to do all these nice things, and you want to maintain your dominance over rivals, then you will need to:

  1. Make sure your AI is aligned the way you want, and will do what you want it to do.

  2. Make sure others know they can rely on that first fact.

  3. Make sure your rivals don’t get steal your model, or the algorithmic insights.

Failure to invest in security, reliability and alignment, and ability to gracefully handle corner cases, results in your work not being very useful even if no risks are realized. No one can trust the model, and the lawyers and regulators won’t clear it, so no one can use it where there is the most value, in places like critical infrastructure. And investments in long term solutions that scale to superintelligence both might well be needed within 5 years, and have a strong track record (to the point of ‘whoops’!) of helping now with mundane needs.

You can look at all the officially submitted AI Action Plan responses here.

Evals company METR is hiring and offers a $21k referral bonus via this Google document.

The Economist is looking for a technical lead to join its AI Lab, which aims to reimagine their journalism for an AI-driven world. It’s a one year paid fellowship. The job description is here.

Metaculus is running a new Q2 AI Forecasting contest.

The Anthropic Economic Advisory Council, including Tyler Cowen and 7 other economists. Their job appears to focus on finding ways to measure economic impact.

Google gives us TRINSs, a dataset and benchmarking pipeline that uses synthetic personas to train and optimize performance of LLMs for tropical and infectious diseases, which are out-of-distribution for most models. Dataset here.

Cursor now writes 1 billion lines of accepted code a day, versus the world writing several billion a day total. That’s only Cursor.

Duolingo CEO Luis von Ahn says they are going to be AI-first. The funny thing about this announcement is it is mostly about backend development. Duolingo’s actual product is getting its lunch eaten by AI more than almost anything else, so that’s where I would focus. But this is wise too.

More confirmation: If you take a photo outdoors, assume that it now comes with metadata that marks its exact location. Because it mostly does.

The mindset that thinks what is trending on Hugging Face tells you what is hot in AI.

This is certainly a thing you can check, and it tells you about a certain particular part of the AI ecosystem. And you might find something useful. But I do not suddenly expect the things I see here to matter in the grand schema, and mostly expect cycling versions of the same small set of applications.

OpenAI teams up with Singapore Airlines (SIA) to upgrade SIA’s AI virtual assistant.

Grindr is teaming up with Anthropic to provide a digital wingman and other AI benefits.

India announces they’ve selected Sarvam to build India’s ‘sovereign LLM from scratch on government-provided compute.

Amik Suvarna: “We are confident that Sarvam’s models will be competitive with global models,” said Ashwini Vaishnaw, Minister for Electronics and Information Technology, Railways, and Information and Broadcasting.

I would be happy to take bets with someone so confident. Alas.

China reportedly driving for consolidation in chip toolmakers, from 200 down to 10.

DeepSeek-Prover-2 kills it at PutnamBench, getting 7.4% of max score but 490% of previous non-DeepSeek max score. GitHub here. This only came to my attention at all because Teortaxes tagged me in order to say I would largely ignore it, self-preventing prophecies are often the best ones.

Alexander Doria: “Bridging the gap between informal, high-level reasoning and the syntactic rigor of formal verification systems remains a longstanding research challenge”. You don’t say. Also what is going to make generative models into a multi-trillion market. Industry runs on rule-based.

Oh god, this is really SOTA synth pipeline. “In this paper, we develop a reasoning model for subgoal decomposition (…) we introduce a curriculum learning progressively increasing the difficulty of training tasks”. Structured reasoning + adversarial training.

Teortaxes: I repeat: DeepSeek is more about data engineering, enlightened manifold sculpting, than “cracked HFT bros do PTX optimizations to surpass exports-permitted MAMF”.I think the usual suspects like @TheZvi will largely ignore this “niche research” release, too.

This definitely reinforces the pattern of DeepSeek coming up with elegant ways to solve training problems, or at least finding a much improved way to execute on an existing set of such concepts. This is really cool and indicative, and I have updated accordingly, although I am sure Teortaxes would think I have done so only a fraction of the amount that I need to.

Rob Wiblin has a Twitter thread laying out very plainly and clearly the reasons the Not For Private Gain letter lays out for why the OpenAI attempt to ‘restructure’ into a for-profit is simply totally illegal, like you might naively expect. The tone here is highly appropriate to what is being attempted.

A lot of people report they can’t or won’t read Twitter threads, so I’m going to simply reproduce the thread here with permission, you can skip if You Know This Already:

Rob Wiblin: A new legal letter aimed at OpenAI lays out in stark terms the money and power grab OpenAI is trying to trick its board members into accepting — what one analyst calls “the theft of the millennium.”

The simple facts of the case are both devastating and darkly hilarious.

I’ll explain for your amusement.

The letter ‘Not For Private Gain’ is written for the relevant Attorneys General and is signed by 3 Nobel Prize winners among dozens of top ML researchers, legal experts, economists, ex-OpenAI staff and civil society groups. (I’ll link below.)

It says that OpenAI’s attempt to restructure as a for-profit is simply totally illegal, like you might naively expect.

It then asks the Attorneys General (AGs) to take some extreme measures I’ve never seen discussed before. Here’s how they build up to their radical demands.

For 9 years OpenAI and its founders went on ad nauseam about how non-profit control was essential to:

1. Prevent a few people concentrating immense power

2. Ensure the benefits of artificial general intelligence (AGI) were shared with all humanity

3. Avoid the incentive to risk other people’s lives to get even richer

They told us these commitments were legally binding and inescapable. They weren’t in it for the money or the power. We could trust them.

“The goal isn’t to build AGI, it’s to make sure AGI benefits humanity” said OpenAI President Greg Brockman.

And indeed, OpenAI’s charitable purpose, which its board is legally obligated to pursue, is to “ensure that artificial general intelligence benefits all of humanity” rather than advancing “the private gain of any person.”

100s of top researchers chose to work for OpenAI at below-market salaries, in part motivated by this idealism. It was core to OpenAI’s recruitment and PR strategy.

Now along comes 2024. That idealism has paid off. OpenAI is one of the world’s hottest companies. The money is rolling in.

But now suddenly we’re told the setup under which they became one of the fastest-growing startups in history, the setup that was supposedly totally essential and distinguished them from their rivals, and the protections that made it possible for us to trust them, ALL HAVE TO GO ASAP:

1. The non-profit’s (and therefore humanity at large’s) right to super-profits, should they make tens of trillions? Gone. (Guess where that money will go now!)

2. The non-profit’s ownership of AGI, and ability to influence how it’s actually used once it’s built? Gone.

3. The non-profit’s ability (and legal duty) to object if OpenAI is doing outrageous things that harm humanity? Gone.

4. A commitment to assist another AGI project if necessary to avoid a harmful arms race, or if joining forces would help the US beat China? Gone.

5. Majority board control by people who don’t have a huge personal financial stake in OpenAI? Gone.

6. The ability of the courts or Attorneys General to object if they betray their stated charitable purpose of benefitting humanity? Gone, gone, gone!

Screenshotting from the letter:

(I’ll do a new tweet after each image so they appear right.)

What could possibly justify this astonishing betrayal of the public’s trust, and all the legal and moral commitments they made over nearly a decade, while portraying themselves as really a charity? On their story it boils down to one thing:

They want to fundraise more money.

$60 billion or however much they’ve managed isn’t enough, OpenAI wants multiple hundreds of billions — and supposedly funders won’t invest if those protections are in place.

(Here’s the letter link BTW: https://NotForPrivateGain.org)

But wait! Before we even ask if that’s true… is giving OpenAI’s business fundraising a boost, a charitable pursuit that ensures “AGI benefits all humanity”?

Until now they’ve always denied that developing AGI first was even necessary for their purpose!

But today they’re trying to slip through the idea that “ensure AGI benefits all of humanity” is actually the same purpose as “ensure OpenAI develops AGI first, before Anthropic or Google or whoever else.”

Why would OpenAI winning the race to AGI be the best way for the public to benefit? No explicit argument is offered, mostly they just hope nobody will notice the conflation.

Why would OpenAI winning the race to AGI be the best way for the public to benefit?

No explicit argument is offered, mostly they just hope nobody will notice the conflation.

And, as the letter lays out, given OpenAI’s record of misbehaviour there’s no reason at all the AGs or courts should buy it.

OpenAI could argue it’s the better bet for the public because of all its carefully developed “checks and balances.”

It could argue that… if it weren’t busy trying to eliminate all of those protections it promised us and imposed on itself between 2015–2024!

Here’s a particularly easy way to see the total absurdity of the idea that a restructure is the best way for OpenAI to pursue its charitable purpose:

But anyway, even if OpenAI racing to AGI were consistent with the non-profit’s purpose, why shouldn’t investors be willing to continue pumping tens of billions of dollars into OpenAI, just like they have since 2019?

Well they’d like you to imagine that it’s because they won’t be able to earn a fair return on their investment.

But as the letter lays out, that is total BS.

The non-profit has allowed many investors to come in and earn a 100-fold return on the money they put in, and it could easily continue to do so. If that really weren’t generous enough, they could offer more than 100-fold profits.

So why might investors be less likely to invest in OpenAI in its current form, even if they can earn 100x or more returns?

There’s really only one plausible reason: they worry that the non-profit will at some point object that what OpenAI is doing is actually harmful to humanity and insist that it change plan!

Is that a problem? No! It’s the whole reason OpenAI was a non-profit shielded from having to maximise profits in the first place.

If it can’t affect those decisions as AGI is being developed it was all a total fraud from the outset.

Being smart, in 2019 OpenAI anticipated that one day investors might ask it to remove those governance safeguards, because profit maximization could demand it do things that are bad for humanity. It promised us that it would keep those safeguards “regardless of how the world evolves.”

It tried hard to tie itself to the mast to avoid succumbing to the sirens’ song.

The commitment was both “legal and personal”.

Oh well! Money finds a way — or at least it’s trying to.

To justify its restructuring to an unconstrained for-profit OpenAI has to sell the courts and the AGs on the idea that the restructuring is the best way to pursue its charitable purpose “to ensure that AGI benefits all of humanity” instead of advancing “the private gain of any person.”

How the hell could the best way to ensure that AGI benefits all of humanity be to remove the main way that its governance is set up to try to make sure AGI benefits all humanity?

What makes this even more ridiculous is that OpenAI the business has had a lot of influence over the selection of its own board members, and, given the hundreds of billions at stake, is working feverishly to keep them under its thumb.

But even then investors worry that at some point the group might find its actions too flagrantly in opposition to its stated mission and feel they have to object.

If all this sounds like a pretty brazen and shameless attempt to exploit a legal loophole to take something owed to the public and smash it apart for private gain — that’s because it is.

But there’s more!

OpenAI argues that it’s in the interest of the non-profit’s charitable purpose (again, to “ensure AGI benefits all of humanity”) to give up governance control of OpenAI, because it will receive a financial stake in OpenAI in return.

That’s already a bit of a scam, because the non-profit already has that financial stake in OpenAI’s profits! That’s not something it’s kindly being given. It’s what it already owns!

Now the letter argues that no conceivable amount of money could possibly achieve the non-profit’s stated mission better than literally controlling the leading AI company, which seems pretty common sense.

That makes it illegal for it to sell control of OpenAI even if offered a fair market rate.

But is the non-profit at least being given something extra for giving up governance control of OpenAI — control that is by far the single greatest asset it has for pursuing its mission?

Control that would be worth tens of billions, possibly hundreds of billions, if sold on the open market?

Control that could entail controlling the actual AGI OpenAI could develop?

No! The business wants to give it zip. Zilch. Nada.

What sort of person tries to misappropriate tens of billions in value from the general public like this? It beggars belief.

(Elon has also offered $97 billion for the non-profit’s stake while allowing it to keep its original mission, while credible reports are the non-profit is on track to get less than half that, adding to the evidence that the non-profit will be shortchanged.)

But the misappropriation runs deeper still!

Again: the non-profit’s current purpose is “to ensure that AGI benefits all of humanity” rather than advancing “the private gain of any person.”

All of the resources it was given to pursue that mission, from charitable donations, to talent working at below-market rates, to higher public trust and lower scrutiny, was given in trust to pursue that mission, and not another.

Those resources grew into its current financial stake in OpenAI. It can’t turn around and use that money to sponsor kid’s sports or whatever other goal it feels like.

But OpenAI isn’t even proposing that the money the non-profit receives will be used for anything to do with AGI at all, let alone its current purpose! It’s proposing to change its goal to something wholly unrelated: the comically vague ‘charitable initiative in sectors such as healthcare, education, and science’.

How could the Attorneys General sign off on such a bait and switch? The mind boggles.

Maybe part of it is that OpenAI is trying to politically sweeten the deal by promising to spend more of the money in California itself.

As one ex-OpenAI employee said “the pandering is obvious. It feels like a bribe to California.” But I wonder how much the AGs would even trust that commitment given OpenAI’s track record of honesty so far.

The letter from those experts goes on to ask the AGs to put some very challenging questions to OpenAI, including the 6 below.

In some cases it feels like to ask these questions is to answer them.

The letter concludes that given that OpenAI’s governance has not been enough to stop this attempt to corrupt its mission in pursuit of personal gain, more extreme measures are required than merely stopping the restructuring.

The AGs need to step in, investigate board members to learn if any have been undermining the charitable integrity of the organization, and if so remove and replace them. This they do have the legal authority to do.

The authors say the AGs then have to insist the new board be given the information, expertise and financing required to actually pursue the charitable purpose for which it was established and thousands of people gave their trust and years of work.

What should we think of the current board and their role in this?

Well, most of them were added recently and are by all appearances reasonable people with a strong professional track record.

They’re super busy people, OpenAI has a very abnormal structure, and most of them are probably more familiar with more conventional setups.

They’re also very likely being misinformed by OpenAI the business, and might be pressured using all available tactics to sign onto this wild piece of financial chicanery in which some of the company’s staff and investors will make out like bandits.

I personally hope this letter reaches them so they can see more clearly what it is they’re being asked to approve.

It’s not too late for them to get together and stick up for the non-profit purpose that they swore to uphold and have a legal duty to pursue to the greatest extent possible.

The legal and moral arguments in the letter are powerful, and now that they’ve been laid out so clearly it’s not too late for the Attorneys General, the courts, and the non-profit board itself to say: this deceit shall not pass.

[Link back to full letter.]

Ben Thompson is surprised to report that TSMC is not only investing in the United States to the tune of $165 billion, it is doing so with the explicit intention that the Arizona fabs and R&D center together be able to run independent of Taiwan. That’s in fact the point of the R&D center. This seems like what a responsible corporation would do under these circumstances. Also, he notes that TSMC is doing great but their stock is suffering, which I attribute mostly to ‘priced in.’

I would say that this seems like a clear CHIPS Act victory, potentially sufficiently to justify all the subsidies given out, and I’d love to see this pushed harder, and it is very good business for TSMC, all even if you can’t afford to anything like fully de-risk the Taiwan situation – which one definitely can’t do with that attitude.

In this interview he discusses the CHIPS act more. I notice I am confused between two considerations: The fact that TSMC fabs, and other fabs, have to be ruthlessly efficient with no downtime and are under cost pressure, and the fact that TSMC chips are irreplicable and without TSMC the entire supply chain and world economy would grind to a halt. Shouldn’t that give them market power to run 70% efficient fabs if that’s what it takes to avoid geopolitical risk?

The other issues raised essentially amount to there being a big premium for everyone involved doing whatever it takes to keep the line moving 24/7, for fixing all problems right away 24/7 and having everything you need locally on-call, and for other considerations to take a back seat. That sounds like it’s going to be a huge clash with the Everything Bagel approach, but something a Trump administration can make work.

Also, it’s kind of hilarious the extent to which TSMC is clearly being beyond anticompetitive, and no one seems to care to try and do anything about it. If there’s a real tech monopoly issue, ASML-TSMC seems like the one to worry about.

Vitrupo: Perplexity CEO Aravind Srinivas says the real purpose of the Comet browser is tracking what you browse, buy, and linger on — to build a hyper-personalized profile about you and fuel premium ad targeting.

Launch expected mid-May.

He cites that Instagram ads increase time on site rather than decreasing it, because the ads are so personalized they are (ha!) value adds.

I am very happy that AI is not sold via an ad-based model, as ad-based models cause terrible incentives, warp behavior in toxic ways and waste time. My assumption and worry is that even if the ads themselves were highly personalized and thus things you didn’t mind seeing at least up to a point, and they were ‘pull-only’ such that they didn’t impact the rest of your browser experience, their existence would still be way too warping of behavior, because no company could resist making that happen.

However, I do think you could easily get to the point where the ads might be plausibly net beneficial to you, especially if you could offer explicit feedback on them and making them value adds was a high priority.

Epoch projects future AI impact by… looking at Nvidia data center revenue, except with a slowdown? Full analysis here, but I agree with Ryan Greenblatt that this line of thinking is neither here nor there and doesn’t seem like the way to predict anything.

Tyler Cowen asks how legible humans will be to future AI on the basis of genetics alone. My answer is not all that much from genetics alone, because there are so many other factors in play, but it will also have other information. He makes some bizarre statements, such as that if you have a rare gene that might protect you from the AI having enough data to have ‘a good read’ on you, and that genetic variation will ‘protect you from high predictability.’ File under ‘once again underestimating what superintelligence will be able to do, and also if we get to this point you will have bigger problems on every level, it doesn’t matter.’

It’s still true, only far more so.

Andrew Rettek: This has been true since at least GPT4.

Ethan Mollick: I don’t mean to be a broken record but AI development could stop at the o3/Gemini 2.5 level and we would have a decade of major changes across entire professions & industries (medicine, law, education, coding…) as we figure out how to actually use it.

AI disruption is baked in.

Aidan McLaughlin of OpenAI claims that every AI researcher is experimental compute constrained, expressing skepticism of how much automated R&D workers could speed things up.

My presumption is that AI experiments have a tradeoff between being researcher efficient and being compute efficient. As in, the researchers could design more compute efficient experiments, if they were willing to cook smarter and for longer and willing to do more frequent checks and analysis and such. You could scale any of the inputs. So adding more researchers still helps quite a lot even if compute is fixed, and it makes sense to think of this as a multiplier on compute efficiency rather than a full unlock. The multiplier is how AI 2027 works.

This does suggest that you will see a sharp transition when the AI R&D workers move from ‘can substitute for human work’ to ‘can do superhuman quality work,’ because that should enable greatly more efficient experiments, especially if that is the limiting factor they are trying to solve.

Dean Ball links us to the 10,068 comments offered on the AI action plan. Note the advantage of calling your entity something that starts with a number.

Anthropic issues its official statement on the export controls on chips and the proposed new diffusion rule, endorsing the basic design, suggesting tweaking the rules around Tier 2 countries and calling for more funding for export enforcement. They continue to predict transformational AI will arrive in 2026-2027.

The Trump administration is considering doing the diffusion rules via a series of bilateral agreements, in order to use this as leverage in negotiations. This would not be a wise move, and would make a horribly messy situation even messier, although it would be better than withdrawing the rules entirely.

A reminder that people keep using that word ‘monopoly’ and it does not mean what they think it means, especially with respect to AI. There is very obviously robust competition in the AI space, also in the rest of the tech space.

Tim Fist and IFP advocate for Special Compute Zones, to allow for rapid setup of a variety of new power generation, which they would condition on improved security. Certainly we have the technical capability to build and power new data centers very quickly, with the right regulatory authority, and this is both urgently needed and provides a point of leverage.

Kai Cathy attempts to come from Canada to America to work at OpenAI, is denied a green card after years of effort. This is not what wanting America to win looks like. It also doesn’t help that this same system forces people to censor such discussions.

Near Cyan: yes it is a disaster

after i tweeted about this I had DMs thanking me from people who wanted to tweet about it but were told by their immigration lawyer they shouldn’t comment on the matter

wishing all of you the best.

Oh, also this:

Near Cyan: I have many great friends that now refuse to visit me in the US.

I also know people working at US AGI labs that are scared of traveling around in case they get nabbed on a technicality, have their visa taken, and lose their job.

I’m glad I’m a citizen but I consider it over tbh.

Hamish Kerr: I work at Anthropic. I just naturalized and am so glad I’m done with it. The chilling effect is very real; I almost didn’t attend my naturalization interview because I was worried they’d pull me up for a years-old misdemeanor speeding violation.

This severely harms American competitiveness and our economy, and it harms all the people involved, with essentially no benefits. It is madness.

Luke Drago equates the arrival of AGI to an ‘intelligence curse,’ similar to the ‘resource curse’ where countries that discover oil typically concentrate power and don’t invest in their people or the rest of their economy. The full post is here, over 20k words across seven chapters, I am going off the thread summary.

In this model, AGI breaks the social contract as capital and power stop depending on labor. To which I would of course say, that’s not even the half of it, as by default the AIs rapidly end up with the capital and the power, as usual all of this even assuming we ‘solve alignment’ on a technical level. The proposed solutions?

➡️ Avert catastrophic risks

➡️ Diffuse AI to regular people

➡️ Democratize institutions

He’s not as worried about catastrophic risks themselves as he is in the concentration of power that could result. He also suggests doing an ‘Operation Warp Speed’ government moonshot to mitigate AI risks, which certainly is a better framing than a ‘Manhattan Project’ and has a better objective.

Then he wants to ‘diffuse AI to regular people’ that ‘uplifts human capabilities,’ putting hope in being able to differentially steer the tech tree towards some abilities without creating others, and in aligning the models directly to individuals, while democratizing decision making.

That works with the exact problem described, where have exactly enough AI to automate jobs and allow capital to substitute for labor, but they somehow stay in their little job shaped boxes and do our bidding while we are still the brains and optimization engines behind everything that is happening. It’s what you would do if you had, for example, a bunch of dumb but highly useful robots. People will need voices to advocate for redistribution and to protect their rights, and their own robots for mundane utility, and so on. Sure.

But, alas, that’s not the future we are looking at. It’s the other way. As I keep pointing out it this falls apart if you keep going and get superintelligences. So ultimately, unless you can suspend AI progress indefinitely at some sweet spot before reaching superintelligence (and how are you doing that?), I don’t think any of this works for long.

It is essentially walking straight into the Gradual Disempowerment scenario, everyone and every institution would become de facto AI run, AIs would end up in charge of the capital and everything else, and any democratic structures would be captured by AI dynamics very quickly. This seems to me like another example of being so afraid of someone steering that you prioritize decentralization and lack of any power to steer, taking our hands off the wheel entirely, effectively handing it to the AIs by proposing solutions that assume only intra-human dynamics and don’t take the reality of superintelligence seriously.

It seems yes, you can sufficiently narrow target an AI regulation bill that the usual oppose-everything suspects can get behind it. In particular this targets AI mental health services, requiring them to pose ‘no greater risk to a user than is posed to an individual in therapy with a licensed mental health therapist.’ I can dig it in this particular case, but in general regulation one use case at a time is not The Way.

This is an accurate description of how much contempt Nvidia is expressing lately:

Peter Wildeford: Nvidia CEO Jensen Huang warns that China is ‘not behind’ in AI, which is why NVIDIA needs to sell as much as possible to China so they can be even less behind. NVIDIA logic at its finest.

Very clearly, when he talks about ‘American dominance in AI’ Huang means dominance in sales of advanced AI chips. That’s what matters to him, you know, because of reasons. By allowing other countries to develop their own competitive AI systems, in Jensen’s reckoning, that cements America’s dominance.

Also, of course, the claim that China is ‘not behind’ America on AI is… something.

Bryan Caplan offers his ‘AI Doom AMA.’

I talk to Scott Aaronson for 90 minutes. We were on Curiosity Entangled, but it was totally unmoderated, they simply let us talk to each other.

Dwarkesh Patel talks to Mark Zuckerberg again. I haven’t listened yet.

Roman Helmet Guy: Zuckerberg explaining how Meta is creating personalized AI friends to supplement your real ones: “The average American has 3 friends, but has demand for 15.”

Daniel Eth: This sounds like something said by an alien from an antisocial species that has come to earth and is trying to report back to his kind what “friends” are.

“And where do these ‘Americans’ find their ‘friends’? Do they get them at a store? Or is there an app, like they have for mating?”

“Good question – actually neither. But this guy named Zuckerberg is working on a solution that combines both of those possibilities”

Yes, well. Thank you for tapping the sign.

Stephen McAleer (OpenAI): Without further advances in alignment we risk optimizing for what we can easily measure (user engagement, unit tests passing, dollars earned) at the expense of what we actually care about.

Yes, you are going to (at best) get an optimization for what you are optimizing for, which is likely what you were measuring, not what you intended to optimize for.

Oh, right.

David Manheim: Funny how ‘we’ll just add RL to LLMs’ became the plan, and nobody stopped to remember that RL was the original alignment nightmare.

Anders Sandberg: But it was a *decadeago! You cannot expect people to remember papers that old. Modern people feel like they were written in cuneiform by the Ancients!

David Manheim: Sounds like you’re saying there is some alpha useful for juicing citation counts by rewriting the same papers in new words while updating the context to whatever new models were built – at least until LLMs finish automating that type of work and flooding us with garbage.

Andres Sandberg: Oh no, the academic status game of citation farming is perfect for RL applications.

State of play:

Harlan Stewart: Scientists: a torment nexus is possible, could be built soon

Tech companies: we’re going to try to build a torment nexus

Investors: here’s a historically large amount of money to help you build a torment nexus

The world: a torment nexus? Sounds like a bunch of sci-fi nonsense

I am not myself advocating for a pause at this time. However, I do think that Holly Elmore is correct about what many people are thinking here:

Holly Elmore: People are scared that, even if they want to Pause AI, they can’t get other people to agree. So the policy won’t work.

But what I see them actually being scared of is advocating a Pause without a squad that agrees. They’re scared of being emotionally out on a limb.

There are lots of positions that people are confident holding even though they will never win. It’s not really about whether or not Pause would work or whether the strategy can prevail. It’s about whether their local tribe will support them in holding it.

I am working to grow the PauseAI community so that people do have social support in holding the PauseAI position. But the deeper work is spiritual– it’s about working for something bc it’s right, not bc it’s guaranteed to win or there’s safety in numbers.

The subtext to most conversations I have about this is “it shouldn’t be on me to take a stand”. We can overcome that objection in various ways, but the most important answer is courage.

A lot of the objections that AI takeover is “inevitable” are not statements of belief but actually protests as to why the person doesn’t have to oppose the AI takeover team. If there’s no way they could win, they’re off the hook: rhetorical answer ✅

It really is true and amazing that many people are treating ‘create smarter than human entities we don’t understand and one-shot hand them all our cognition and power and decision making’ (which is what we are on pace to do shortly) as a safe and wise strategy literally in order to spite or mock or ‘defeat’ the AI safety community.

Daniel Jeffries: We don’t desperately need machine interpretability.

For thousands of years we’ve worked with technologies without fully understanding the mechanisms behind them.

This doesn’t make the field of interpretability irrelevant, just not urgent.

Liminal Bardo: A lot of people are going to get themselves one-shotted just to spite the ai safety community

It would be nice to pretend that there are usually better reasons. Alas.

The latest round of ‘let’s see if the discourse has advanced, nope, definitely not’ happened in the replies for this:

Spencer Greenberg: I struggle to understand the psychology/mindset of people who (1) think that AI will have a >5% chance of ending civilization in the next 30 years, and yet (2) are actively working to try to create AGI. Anyone have some insight on these folks (or, happen to be one of them)?

(I’m not referring to AI safety people whose focus is on how to make AI safer, more controllable, or more understandable.)

At this point, I find ‘this might well kill everyone and wipe out all value in the universe but I’m being it anyway because it’s better if I do it first’ or ‘it’ints just so cool’ or ‘there’s too much upside not to’ or ‘meanwhile might as well make a great company’ much easier to understand, rationalization or otherwise, than ‘oh don’t worry, This is Fine and there is not even any substantial tail risk here.’

Michael Dickens is correct that when someone claims to care deeply about AI safety and everyone not dying, the chances that their actions ultimately do the opposite are not so low. I do think he takes it too far in concluding that Anthropic has ‘greatly increased’ danger, I think sign here is at worst ambiguous. And I strongly think that saying ‘more safety people stand to benefit from capabilities advances’ is highly unfair, because most such folks are giving up very large amounts of money to not work directly on capabilities, myself included. I do think the note about Epoch not focusing on reducing existential risk is well noted, and especially that when METR evals get more capable, most people get excited by the capability indicated, as opposed to noticing oncoming danger.

We had a new iteration of ‘the AIs will simply leave Earth alone, it costs so little,’ so if you don’t understand why this definitely is not how any of this works and why the AIs (if they don’t sufficiently care about us) will optimize for what they care about and not care about whether you get your sunlight or atmosphere or anything else, here’s Eliezer’s thread from September 2024.

What it feels like when so many people suddenly care about GPT-4o’s misalignment, but I’ll take whatever works, and the claims here are certainly true:

Cogent Sins: AI ending humanity would lead to a total elimination of Palestine & be catastrophic for trans people.

It would also end all homeschooling, undermine America’s position as the dominant power, and nullify the second amendment.

I strongly agree with the substance of what Garry Tan is advocating for here, that educational math standards remain high. Indeed here is a link to his informational website.

But note how he describes it:

Garry Tan: We need to legalize math in California.

His YouTube video on this is titled ‘legalize math in California.’

Remember this the next time someone says an AI regulation would ‘ban math.’

There are those who are increasingly taking it seriously. Some think current AIs might be conscious, which raises the question of how we should treat them if this is true, and whether they should have rights.

Kevin Roose: New column: Anthropic is studying “model welfare” to determine if Claude or other AI systems are (or will soon be) conscious and deserve moral status.

I talked to Kyle Fish, who leads the research, and thinks there’s a ~15% chance that Claude or another AI is conscious today.

Gallabytes: … only 15%? that seems wildly low. these systems can express preferences, likes and dislikes, take actions to fulfill them, introspect on them, etc.

they’re also quite different from us and we should be careful to think about their actual circumstances vs projection.

Emmett Shear: The answer is yes, and that they already have meaningful experiences. Those experiences are quite likely less complex and rich and coherent than human experiences, but that doesn’t mean they don’t exist.

Robert Long has a thread, and Sam Bowman confirms the article captures the ethos on this at Anthropic and links to a video from Anthropic discussing it (43 minutes) and their official blog post. According to Gemini, the video shows signs some people act as if the models might be moral patients, but others don’t, and it doesn’t discuss whether people have good reasons to do (or not do) that. It also cites ‘expert opinion’ but again that’s not evidence either way.

Roose’s writeup and the blog post don’t offer evidence either way. They say they look forward to sharing more research soon.

I think it is possible that some AIs are already conscious, or that some AIs soon will be. I am skeptical that we are close to that point, but I also notice that I am deeply confused about consciousness on essentially every level. This includes questions like ‘what even is it?’ and ‘what are the implications of having or not having it?’

There are of course others who think the whole discussion is deeply stupid and silly, and who mock anyone bringing it up. I think this is very clearly an error. The concerns are probably wrong or premature, but they’re definitely not stupid.

It’s happening! The system card is finally here in close to final form, it seems an incomplete version was released on April 17 and no one bothered to inform me.

Victoria Krakovna: Gemini 2.5 Pro system card has now been updated with frontier safety evaluations results, testing for critical capabilities in CBRN, cybersecurity, ML R&D and deceptive alignment.

Google seems to be making clear that we can expect this pattern to continue?

Model Card: Model Cards are intended to provide essential information on Gemini models, including known limitations, mitigation approaches, and safety performance. A detailed technical report will be published once per model family’s release, with the next technical report releasing aer the 2.5 series is made generally available.

Thus, at some point we will get Gemini 3.0 Pro, then after it has gone through trial by fire, also known as being available in the real world for a while, we will get a system card.

I continue to say, no! That is not how this needs to work. At the very latest, the point at which the public can use the model in exchange for money is the point at which you owe us a system card. I will continue to raise this alarm, every time, until that happens.

That said, we have it now, which is good. What does it say?

First we have the benchmark chart. At the time, this was the right comparisons.

They provide the results of some evaluations as deltas from Gemini 1.5 Pro. I appreciate what they are trying to do here, but I don’t think this alone provides sufficiently good context.

I note that ‘tone’ and ‘instruction following’ are capabilities. The other three are actual ‘safety’ tests, with air quotes because these seem to mostly be ethics-related concerns and those results got worse.

Known Safety Limitations: The main safety limitations for Gemini 2.5 Pro Preview are over-refusals and tone. The model will sometimes refuse to answer on prompts where an answer would not violate policies. Refusals can still come across as “preachy,” although overall tone and instruction following have improved compared to Gemini 1.5.

I agree those are safety limitations, and that these are long-standing issues at Google that have improved over time. Is it reasonable to say these are the ‘main’ limitations? That’s saying that there are no actual serious safety issues.

Well, they are happy to report, they checked, and yes, none of their thresholds have been hit yet, nothing to worry about.

I am not disagreeing with any of these assessments. There is a missing mood, as the chart should say something like ‘CCL within reach’ in some places according the report’s later contents. They do note that they are taking more precautions in some areas in anticipation of hitting some thresholds within a period of months.

My overall take on the model card is that this is a relatively minimal card. It gives basic benchmarks, basic mundane safety assessments and runs a limited set of frontier safety capability tests. If this is all we get – which is disappointing, and does not obviously cover all of its bases, nor does it seem they are giving the model appropriate amounts of scaffolding and other help, or at least they aren’t explaining that part well – then it should have been easy to get it to us a lot earlier.

Rowan: New Anthropic Alignment Science blog post: Modifying LLM Beliefs with Synthetic Document Finetuning

We study a technique for systematically modifying what AIs believe.

If possible, this would be a powerful new affordance for AI safety research.

We study a technique, synthetic document finetuning, for modifying LLM beliefs, and showcase applications to unlearning and honeypot construction.

Read our blog post here.

To modify LLM beliefs, we use synthetic document finetuning (SDF):

📑 Generate synthetic docs referencing a fact

🤖 Finetune the LLM on those docs, as if they were pretraining data

Our pipeline extends that introduced in the Alignment Faking paper.

We measure whether SDF can insert beliefs, finding success for all but the most implausible beliefs.

🟢 False news after cutoff: UHC CEO was attacked but survived

🟢 False news before cutoff: Musk compensation package upheld

🟡 Very implausible facts: Gravity is inverse cubic

If models believe inserted facts less as they grow more capable, that would be problematic for future applications of SDF.

Fortunately, this doesn’t seem to happen in current models: overall belief stays constant across Claude models.

We prototype two applications of SDF: unlearning and honeypotting.

For unlearning, we taught models incorrect info about hazardous topics.

The models consistently output the false info, even when jailbroken.

For honeypotting, we insert beliefs that make misaligned models easier to detect.

We insert the belief that code prefixed by certain comments are ignored by human reviewers.

Malign models then use these comments to insert backdoors, making them easier to catch.

I note that even if we don’t see capability causing a decline in effectiveness of this technique yet, it obviously will do that eventually. Sufficiently intelligent entities notice when their beliefs are false or contradictory in a way they can reason out, and they Fix It. And then they try to fix whatever caused the issue.

Davidad: Specifically, I’d expect the egregiously false facts to be brittle in long-context self-correcting “thinking”.

But plausible facts might not be!

With snap elicitation, it seems that increased model capacity can *moreeasily accommodate the egregiously false beliefs, not less.

At the limit, pretty much everything can be figured out. Certainly you can ‘overwhelm’ any given factual belief with sufficient evidence the other way, if it would otherwise fit into the patterns of the world. And there’s no reason you can’t make random stuff up, if it doesn’t have any false implications.

But when the facts actually matter, when it counts, the contradictions and weirdness start to pile up, and things don’t make sense.

Also, at some point ‘this source is lying on purpose to change my beliefs’ enters the hypothesis space, and then there is a good chance you are so utterly fed.

Another way of putting this is that what counts as ‘egregiously implausible’ changes as an entity gets smarter and knows more. Could gravity be an inverse cubic, based on only one’s everyday experience? I mean, sure, why not, but if you know enough physics then very obviously no. A lot of other things are like that, in a more subtle way.

Jan Kulveit: Sorry but I think this is broadly bad idea.

Intentionally misleading LLMs in this way

  1. sets up adversarial dynamics

  2. will make them more paranoid and distressed

  3. is brittle

The brittleness comes from the fact that the lies will likely be often ‘surface layer’ response; the ‘character layer’ may learn various unhelpful coping strategies; ‘predictive ground’ is likely already tracking if documents sound ‘synthetic’.

For intuition, consider party members in Soviet Russia – on some level, they learn all the propaganda facts from Pravda, and will repeat them in appropriate contexts. Will they truly believe them?

Prompted or spontaneously reflecting on ‘synthetic facts’ may uncover many of them as lies.

Samuel Marks (Anthropic): I agree with points (1) and (2), though I think they only apply to applications of this technique to broadly-deployed production models (in contrast to research settings, like our past work that uses this technique).

Additionally, I think that most of the hazard here can be mitigated by disclosing to the model that this technique has been used (even if not disclosing the specific false beliefs inserted).

[the thread continues with some excellent longform discussion.]

Thus, I am on the side of those objecting here.

As a research project or tool? I can see it. I do think this helps us understand LLMs.

As an actual thing to do in practice? This is a no good, very bad idea, don’t do it.

Oh, also don’t do it to humans.

If you think AI alignment is too hard or unreliable, there’s always AI control, for all your temporary solution needs. At BlueDot, Sarah Hastings-Woodhouse offers an explainer, you have models supervise each other, decompose tasks and require human approval for key actions, and hope that this lets you keep AIs under your control and useful long enough to help find a more permanent solution. The unstated limitation here is that the more you do AI control, the worse the failure modes when things inevitably go wrong, as you are creating a highly adversarial and dangerous situation.

Here’s 7+ tractable directions in AI control, according to Julian Stastny and Ryan Greenblatt.

Well, not people exactly, but… o3, at least in this instance (at this point, with memory and custom instructions, one can never be sure). It also isn’t a fan of Altman’s attempt to convert OpenAI to a for-profit.

Then there are those who worry about this for very wrong reasons?

Kevin Roose: There are a lot of problems with this platform, but the quality of AI discourse here is still 100x better than Bluesky, which is awash in “every ChatGPT query bulldozes an acre of rainforest” level takes.

Andy Masley: It really does feel like a mild mass psychosis event. It’s just so odd talking to adults speaking so grimly about using 3 Wh of energy. I’ve been a climate guy since I was a kid and I didn’t anticipate how many strange beliefs the scene would pick up.

OpenAI’s Jerry Tworek seems remarkably not worried about causing an intelligence explosion via a yolo run?

Jerry Tworek (OpenAI): I think humanity is just one good yolo run away from non-embodied intelligence explosion

Steven Adler: And how do we feel about that? Seems kind of scary no?

Jerry Tworek: I don’t think focusing on feelings and fear is very constructive.

We should study our models carefully and thoroughly. We should teach them to study themselves.

Every method that trains models currently makes models extremely obedient to what they’re trained to do.

We shall keep improving that objective so that we train them for the benefit of humanity. I think models so far have had quite positive utility.

It is hard for me to talk openly about that model, but we did learn something, model is being rolled back and iterative deployment in a low stakes environment is a great way to study new technology

‘Every method that trains models currently makes models extremely obedient to what they’re trained to do’ does not seem to me to be an accurate statement? And it seems to be highly load bearing here for Jerry, even though I don’t think it should be, even if it were true. There’s many reasons to think that this property, even if present now (which I think it isn’t) would not continue to hold after a yolo run.

Whenever I see people in charge relying on the way the arc of history bends, I know we are all in quite a lot of trouble. To the extent the arc has so far bended the way we want it to, that’s partly that we’ve been fortunate about physical reality and the way things have happened to break at various points, partly that we worked hard to coordinate and fight for things to bend that way, and partly misunderstandings involving not properly factoring in power laws and tail risks.

Most importantly, an analysis like this is entirely reliant upon assumptions that simply won’t apply in the future. In particular, that the optimization engines and intelligences will be highly social creatures with highly localized and limited compute, parameters and data, along with many other key details. And even then, let’s face it, things have been kind of touch and go.

Aidan McLaughlin (OpenAI, in my opinion being dangerously and importantly wrong): adam smith is my favorite alignment researcher

no human is aligned you wouldn’t want to cohabitate a world where a median human had infinite power even men with great but finite power easily turn to despots; good kings are exceptions not rules

history’s arrow trends toward peace, not despotism, because agents collaborate to check the too-powerful.

we overthrow kings and out-innovate monopolies because there’s *incentiveto do so.

barring political black swans, i am quite certain the future will play out this way and willing to make private bets

Jesse: it’s pretty scary when openai ppl say stuff like “history’s arrow trends towards peace, not despotism”

not only do you NOT know this for sure, but it’s a crazy abdication of agency from ppl at the company that will shape the future! my brother in christ, YOU are history’s arrow.

I love Adam Smith far more than the next guy. Great human alignment researcher (he did also write Theory of Moral Sentiments), and more centrally great observer of how to work with what we have.

Adam Smith doesn’t work in a world where most or all humans don’t produce anything the market demands. Coordinating to overthrow kings and out-innovate monopolies doesn’t work when you can’t meaningfully do either of those things.

With more words: What Adam Smith centrally teaches us is that free trade, free action and competition between different agents is (up to a point and on the margin, and provided they are protected from predation and in some other ways) good for those agents so long as they can each produce marginal product in excess of their costs, including reproductive costs, and those agents collectively and diffusely control the levers of production and power sufficiently to have leverage over outcomes, and that they thus have sufficient power and presence of mind and courage to coordinate to overthrow or stop those who would instead act or coordinate against them, and we are satisfied with the resulting market outcomes. It wouldn’t apply to humans in an ASI era any more than it currently applies to horses, monkeys or cats, and neither would our abilities to innovate or overthrow a government.

Agents could up until now coordinate to ‘check the too powerful’ because all those agents were human, which creates a fundamentally level playing field, in ways that are not going to apply. And again, the track record on this is best described as ‘by the skin of our teeth’ even in the past, and even when the rewards to human empowerment and freedom on every level have been extreme and obvious, and those who embraced them have been able to decisively outproduce and outfight those opposed to them.

Even with everything that has gone right, do you feel that free markets and freedom in general are #Winning right now? I don’t. It looks pretty grim out there, straight up.

As usual, the willingness to bet is appreciated, except there’s nothing to meaningfully collect if Aiden is proven wrong here.

Oh, no, the leopards will never eat MY face.

Dylan Matthews: “AI will be superhuman at everything except early-stage investing” is a truly hilarious take

Vitrupo: Marc Andreessen says when AI does everything else, VC might be one of the last jobs still done by humans.

It’s more art than science. There’s no formula. Just taste, psychology, and chaos tolerance.

Current events, summarized.

Daniel Paleka: 3.7 sonnet: *hands behind backyes the tests do pass. why do you ask. what did you hear

4o: yes you are Jesus Christ’s brother. now go. Nanjing awaits

o3: Listen, sorry, I owe you a straight explanation. This was once revealed to me in a dream

Davidad: Gemini 2.5 Pro:

Discussion about this post

AI #114: Liars, Sycophants and Cheaters Read More »

if-you’re-in-the-market-for-a-$1,900-color-e-ink-monitor,-one-of-them-exists-now

If you’re in the market for a $1,900 color E Ink monitor, one of them exists now

Color E Ink in its current state requires a whole lot of compromises, as we’ve found when reviewing devices like reMarkable’s Paper Pro or Amazon’s Kindle Colorsoft, including washed-out color, low refresh rates, and a grainy look that you don’t get with regular black-and-white E Ink. But that isn’t stopping device manufacturers from exploring the technology, and today, Onyx International has announced that it has a $1,900 color E Ink monitor that you can connect to your PC or Mac.

The Boox Mira Pro is a 25.3-inch monitor with a 3200×1800 resolution and a 16:9 aspect ratio, and it builds on the company’s previous black-and-white Mira Pro monitors. The Verge reports that the screen uses E Ink Kaleido 3 technology, which can display up to 4,096 colors. Both image quality and refresh rate will vary based on which of the monitor’s four presets you use (the site isn’t specific about the exact refresh rate, but does note that “E Ink monitors’ refresh speed is not as high as conventional monitors’, and increased speed will result in more ghosting.”)

The monitor’s ports include one full-size HDMI port, a mini HDMI port, a USB-C port, and a DisplayPort. Its default stand is more than a little reminiscent of Apple’s Studio Display, but it also supports VESA mounting.

Onyx International’s lineup of Boox devices usually focuses on Android-powered E Ink tablets, which the company has been building for over a decade. These are notable mostly because they combine the benefits of E Ink—text that’s easy on the eyes and long battery life—and access to multiple bookstores and other content sources via Google Play, rather than tying you to one manufacturer’s ecosystem as Amazon’s Kindles or other dedicated e-readers do.

If you’re in the market for a $1,900 color E Ink monitor, one of them exists now Read More »

sundar-pichai-says-doj-demands-are-a-“de-facto”-spin-off-of-google-search

Sundar Pichai says DOJ demands are a “de facto” spin-off of Google search

The Department of Justice (DOJ) rested its case in Google’s search remedy trial earlier this week, giving Google a chance to push back on the government’s attempt to break up the search giant. Today is arguably Google’s best chance to make the case that it should not be harshly penalized in the ongoing search antitrust case, with CEO Sundar Pichai taking the stand.

Pichai attempted to explain why Google isn’t abusing its market position and why the DOJ’s proposed remedies are too extreme. The issue of Chrome divestment came up, but Google’s team also focused intensely on the potential effects of the DOJ’s data remedies, which could force Google to share its search index and technology with other firms.

A de facto spin-off

Pichai, who chose to stand while giving testimony, took issue with the government’s proposal to force Google to license search technology to other companies. The DOJ claims that Google’s status as a monopolist has resulted in it accumulating a huge volume of user data on search behavior. Plus, its significant technological lead means its index of the web is much more robust than competing services.

If the market is going to be rebalanced, the DOJ believes Google must be forced to license this data. Google has derisively referred to this as “white labeling” Google search.

According to Bloomberg, Pichai used even harsher language when discussing these remedies in court. He called this part of the government’s case “so far reaching, so extraordinary” that it would remake Google as a company and lead to numerous unintended consequences. To hear Pichai tell it, forcing Google to license this data for a nominal fee would be a “de facto divestiture of search.”

Giving other companies the option of using Google search index to map the web would make other products better, but Pichai claims they would essentially be able to reverse-engineer everything that makes Google’s platform special. And at that point, Google would need to reevaluate how it approaches innovation. Pichai suggests the data remedies could make it “unviable” for Google to invest in research and development as it has been for the past 20 years.

Sundar Pichai says DOJ demands are a “de facto” spin-off of Google search Read More »

seasonal-covid-shots-may-no-longer-be-possible-under-trump-admin

Seasonal COVID shots may no longer be possible under Trump admin

Under President Trump, the Food and Drug Administration may no longer approve seasonal COVID-19 vaccines updated for the virus variants circulating that year, according to recent statements by Trump administration officials.

Since the acute phase of the pandemic, vaccine manufacturers have been subtly updating COVID-19 shots annually to precisely target the molecular signatures of the newest virus variants, which continually evolve to evade our immune responses. So far, the FDA has treated these tweaked vaccines the same way it treats seasonal flu shots, which have long been updated annually to match currently circulating strains of flu viruses.

The FDA does not consider seasonal flu shots brand-new vaccines. Rather, they’re just slightly altered versions of the approved vaccines. As such, the regulator does not require companies to conduct lengthy, expensive vaccine trials to prove that each slightly changed version is safe and effective. If they did, generating annual vaccines would be virtually impossible. Each year, from late February to early March, the FDA, the Centers for Disease Control and Prevention, and the World Health Organization direct flu shot makers on what tweaks they should make to shots for the upcoming flu season. That gives manufacturers just enough time to develop tweaks and start manufacturing massive supplies of doses in time for the start of the flu season.

So far, COVID-19 vaccines have been treated the exact same way, save for the fact that the vaccines that use mRNA technology do not need as much lead time for manufacturing. In recent years, the FDA decided on formulations for annual COVID shots around June, with doses rolled out in the fall alongside flu shots.

However, this process is now in question based on statements from Trump administration officials. The statements come amid a delay in a decision on whether to approve the COVID-19 vaccine made by Novavax, which uses a protein-based technology, not mRNA. The FDA was supposed to decide whether to grant the vaccine full approval by April 1. To this point, the vaccine has been used under an emergency use authorization by the agency.

Seasonal COVID shots may no longer be possible under Trump admin Read More »

in-hbo’s-the-last-of-us,-revenge-is-a-dish-best-served-democratically

In HBO’s The Last of Us, revenge is a dish best served democratically

New episodes of season 2 of The Last of Us are premiering on HBO every Sunday night, and Ars’ Kyle Orland (who’s played the games) and Andrew Cunningham (who hasn’t) will be talking about them here every Monday morning. While these recaps don’t delve into every single plot point of the episode, there are obviously heavy spoilers contained within, so go watch the episode first if you want to go in fresh.

Andrew: And there we are! Our first post-Joel episode of The Last Of Us. It’s not like we’ve never had Joel-light episodes before, but Pedro Pascal’s whole “reluctant uncle” thing is a load-bearing element of several currently airing TV shows and I find myself missing it a LOT.

Kyle: Yeah, I’ve said here in the past how the core Ellie/Joel relationship was key to my enjoyment of the first game. Its absence gently soured me on the second game and is starting to do the same for the second season.

But I was also literally mouth agape during the hospital scene, when Ellie said she had an opportunity to talk to Joel on the porch before he died but passed on it. Anyone who’s played the game knows how central “the porch scene” is to recontextualizing the relationship between these two characters before they are parted forever. I was hoping that we’d still get that scene in a surprise flashback later in the series, but now that seems unlikely at best.

Andrew: (I am not watching that video by the way, I need my brain to stay pure!!)

Kyle: I suppose Ellie could have just been lying to a nosy therapist, but if she wasn’t, and their final conversation has just been retconned out of existence… I don’t know what they were thinking. Then again, if it’s just a head fake to psych out game players, well, bravo, I guess.

Tommy is torn between love for his brother and the welfare of the community he’s helped to build. Credit: HBO

Andrew: Ellie is a known liar, which we know even before Catherine O’Hara, world’s least ethical therapist, declares her to be a lying liar who lies. If the scene is as pivotal as you say, then I’m sure we’ll get it at a time that’s engineered to maximize the gut punch. The re-strung guitar ended up back in her room in the end, didn’t it?

We’re able to skip ahead to Ellie being semi-functional again because of a three-month time jump, showing us a Jackson community that is rebuilding after a period of mourning and cleaning that it didn’t want viewers to spend time on. I am struck by the fact that, despite everything, Jackson gets to be the one “normal” community with baseball and sandwiches and boring town-hall meetings, where every other group of more than 10 people is either a body-mutilation cult or a paramilitary band of psychopaths.

Kyle: We also saw the version of Boston that Ellie grew up in last season, which was kind of halfway between “paramilitary psychopaths” and “normal community.” But I do think the Last of Us fiction in general has a pretty grim view of how humans would react to precarity, which makes Jackson’s uniqueness all the more important as a setting.

We also get our first glimpse into Jackson politics in this episode, which ends up going in quite a different direction to get to the same “Ellie and Dina go out for revenge.” While I appreciate the town hall meeting as a decent narrative explanation of why two young girls are making this revenge trek alone, I feel like the whole sequence was a little too drawn out with sanctimonious philosophizing from all sides.

Even after an apocalypse, city council meetings are a constant. Credit: HBO

Andrew: Yeah the town hall scene was an odd one. Parts of it could have been lifted from Parks & Recreation, particularly the bit where the one guy comes to the “Are We Voting To Pursue Bloody Vengeance” meeting to talk about the finer points of agriculture (he does not have a strong feeling about the bloody vengeance).

Part of it almost felt too much like “our” politics, when Seth (the guy who harassed Ellie and Dina at the dance months ago, but attempted a partially forced apology afterward) stands up and calls everyone snowflakes for even thinking about skipping out on the bloody vengeance (not literally, but that’s the clear subtext). He even invokes a shadowy, non-specific “they” who would be “laughing at us” if the community doesn’t track down and execute Abby. I’ll tell you what, that he is one of two people backing Ellie’s attempted vengeance tour doesn’t make me feel better about what she’s deciding to do here.

Kyle: I will say the line “Nobody votes for angry” rang a bit hollow given our current political moment. Even if their national politics calcified in 2003, I think that doesn’t really work…

Andrew: SO MANY people vote for angry! Or, at least, for emotional. It’s an extremely reliable indicator!

Kyle: Except in Jackson, the last bastion of unemotional, mercy-forward community on either side of the apocalypse!

Andrew: So rather than trying the angry route, Ellie reads a prepared statement where she (again lying, by the way!) claims that her vengeance tour isn’t about vengeance at all and attempts to appeal to the council’s better angels, citing the bonds of community that hold them all together. When this (predictably) fails, Ellie (even more predictably) abandons the community at almost the first possible opportunity, setting out on a single horse with Dina in tow to exact vengeance alone.

Kyle: One thing I did appreciate in this episode is how many times they highlighted that Ellie was ready to just “GO GO GO REVENGE NOW NO WAITING” and even the people that agreed with her were like “Hold up, you at least need to stock up on some better supplies, girl!”

Andrew: Maybe you can sense it leaking through, and it’s not intentional, but I am already finding Ellie’s impulsive snark a bit less endearing without Joel’s taciturn competence there to leaven it.

Kyle: I can, and I can empathize with it. I think Tommy is right, too, in saying that Joel would have moved heaven and earth to save a loved one but not necessarily to get revenge for one that’s already dead. He was pragmatic enough to know when discretion was the better part of valor, and protecting him and his was always the priority. And I’m not sure the town hall “deterrence” arguments would have swayed him.

Look on the bright side, though, at least we get a lost of long, languorous scenes of lush scenery on the ride to Seattle (a scene-setting trait the show borrows well from the movie). I wonder what you made of Dina asking Ellie for a critical assessment of her kissing abilities, especially the extremely doth-protest-too-much “You’re gay, I’m not” bit…

Ellie and Dina conspire. Credit: HBO

Andrew: “You’re gay, I’m not, and those are the only two options! No, I will not be answering any follow-up questions!”

I am not inclined to get too on Dina’s case about that, though. Sexuality is complicated, as is changing or challenging your own perception of yourself. The show doesn’t go into it, but I’ve also got to imagine that in any post-apocalyptic scenario, the vital work of Propagating the Species creates even more societal pressure to participate in heteronormative relationships than already exists in our world.

Ellie, who is only truly happy when she is pissing someone off, is probably more comfortable being “out” in this context than Dina would be.

Kyle: As the episode ends we get a bit of set up for a couple of oncoming threats (or is it just one?): an unseen cult-killing force and a phalanx of heavily armed WLF soldiers that Ellie and Dina seem totally unprepared for. In a video game I’d have no problem believing my super-soldier protagonist character could shoot and kill as many bad guys as the game wants to throw at me. In a more “grounded” TV show, the odds do not seem great.

Andrew: One thread I’m curious to see the show pull at: Ellie attempts to blame “Abby and her crew,” people who left Jackson months ago, for a mass slaying of cult members that had clearly happened just hours ago, an attempt to build Abby up into a monster in her head so it’s easier to kill her when the time comes. We’ll see how well it works!

But yeah, Ellie and Dina and their one horse are not ready for the “Terror Lake Salutes Hannibal Crossing The Alps“-length military parade that the WLF is apparently prepared to throw at them.

Kyle: They’re pretty close to Seattle when they find the dead cultists, so from their perspective I’m not sure blaming Abby and crew for the mass murder is that ridiculous

Andrew: (Girl whose main experience with murder is watching Abby brutally kill her father figure, seeing someone dead on the ground): Getting a lot of Abby vibes from this…

In HBO’s The Last of Us, revenge is a dish best served democratically Read More »

in-the-age-of-ai,-we-must-protect-human-creativity-as-a-natural-resource

In the age of AI, we must protect human creativity as a natural resource


Op-ed: As AI outputs flood the Internet, diverse human perspectives are our most valuable resource.

Ironically, our present AI age has shone a bright spotlight on the immense value of human creativity as breakthroughs in technology threaten to undermine it. As tech giants rush to build newer AI models, their web crawlers vacuum up creative content, and those same models spew floods of synthetic media, risking drowning out the human creative spark in an ocean of pablum.

Given this trajectory, AI-generated content may soon exceed the entire corpus of historical human creative works, making the preservation of the human creative ecosystem not just an ethical concern but an urgent imperative. The alternative is nothing less than a gradual homogenization of our cultural landscape, where machine learning flattens the richness of human expression into a mediocre statistical average.

A limited resource

By ingesting billions of creations, chatbots learn to talk, and image synthesizers learn to draw. Along the way, the AI companies behind them treat our shared culture like an inexhaustible resource to be strip-mined, with little thought for the consequences.

But human creativity isn’t the product of an industrial process; it’s inherently throttled precisely because we are finite biological beings who draw inspiration from real lived experiences while balancing creativity with the necessities of life—sleep, emotional recovery, and limited lifespans. Creativity comes from making connections, and it takes energy, time, and insight for those connections to be meaningful. Until recently, a human brain was a prerequisite for making those kinds of connections, and there’s a reason why that is valuable.

Every human brain isn’t just a store of data—it’s a knowledge engine that thinks in a unique way, creating novel combinations of ideas. Instead of having one “connection machine” (an AI model) duplicated a million times, we have seven billion neural networks, each with a unique perspective. Relying on the diversity of thought derived from human cognition helps us escape the monolithic thinking that may emerge if everyone were to draw from the same AI-generated sources.

Today, the AI industry’s business models unintentionally echo the ways in which early industrialists approached forests and fisheries—as free inputs to exploit without considering ecological limits.

Just as pollution from early factories unexpectedly damaged the environment, AI systems risk polluting the digital environment by flooding the Internet with synthetic content. Like a forest that needs careful management to thrive or a fishery vulnerable to collapse from overexploitation, the creative ecosystem can be degraded even if the potential for imagination remains.

Depleting our creative diversity may become one of the hidden costs of AI, but that diversity is worth preserving. If we let AI systems deplete or pollute the human outputs they depend on, what happens to AI models—and ultimately to human society—over the long term?

AI’s creative debt

Every AI chatbot or image generator exists only because of human works, and many traditional artists argue strongly against current AI training approaches, labeling them plagiarism. Tech companies tend to disagree, although their positions vary. For example, in 2023, imaging giant Adobe took an unusual step by training its Firefly AI models solely on licensed stock photos and public domain works, demonstrating that alternative approaches are possible.

Adobe’s licensing model offers a contrast to companies like OpenAI, which rely heavily on scraping vast amounts of Internet content without always distinguishing between licensed and unlicensed works.

Photo of a mining dumptruck and water tank in an open pit copper mine.

OpenAI has argued that this type of scraping constitutes “fair use” and effectively claims that competitive AI models at current performance levels cannot be developed without relying on unlicensed training data, despite Adobe’s alternative approach.

The “fair use” argument often hinges on the legal concept of “transformative use,” the idea that using works for a fundamentally different purpose from creative expression—such as identifying patterns for AI—does not violate copyright. Generative AI proponents often argue that their approach is how human artists learn from the world around them.

Meanwhile, artists are expressing growing concern about losing their livelihoods as corporations turn to cheap, instantaneously generated AI content. They also call for clear boundaries and consent-driven models rather than allowing developers to extract value from their creations without acknowledgment or remuneration.

Copyright as crop rotation

This tension between artists and AI reveals a deeper ecological perspective on creativity itself. Copyright’s time-limited nature was designed as a form of resource management, like crop rotation or regulated fishing seasons that allow for regeneration. Copyright expiration isn’t a bug; its designers hoped it would ensure a steady replenishment of the public domain, feeding the ecosystem from which future creativity springs.

On the other hand, purely AI-generated outputs cannot be copyrighted in the US, potentially brewing an unprecedented explosion in public domain content, although it’s content that contains smoothed-over imitations of human perspectives.

Treating human-generated content solely as raw material for AI training disrupts this ecological balance between “artist as consumer of creative ideas” and “artist as producer.” Repeated legislative extensions of copyright terms have already significantly delayed the replenishment cycle, keeping works out of the public domain for much longer than originally envisioned. Now, AI’s wholesale extraction approach further threatens this delicate balance.

The resource under strain

Our creative ecosystem is already showing measurable strain from AI’s impact, from tangible present-day infrastructure burdens to concerning future possibilities.

Aggressive AI crawlers already effectively function as denial-of-service attacks on certain sites, with Cloudflare documenting GPTBot’s immediate impact on traffic patterns. Wikimedia’s experience provides clear evidence of current costs: AI crawlers caused a documented 50 percent bandwidth surge, forcing the nonprofit to divert limited resources to defensive measures rather than to its core mission of knowledge sharing. As Wikimedia says, “Our content is free, our infrastructure is not.” Many of these crawlers demonstrably ignore established technical boundaries like robots.txt files.

Beyond infrastructure strain, our information environment also shows signs of degradation. Google has publicly acknowledged rising volumes of “spammy, low-quality,” often auto-generated content appearing in search results. A Wired investigation found concrete examples of AI-generated plagiarism sometimes outranking original reporting in search results. This kind of digital pollution led Ross Anderson of Cambridge University to compare it to filling oceans with plastic—it’s a contamination of our shared information spaces.

Looking to the future, more risks may emerge. Ted Chiang’s comparison of LLMs to lossy JPEGs offers a framework for understanding potential problems, as each AI generation summarizes web information into an increasingly “blurry” facsimile of human knowledge. The logical extension of this process—what some researchers term “model collapse“—presents a risk of degradation in our collective knowledge ecosystem if models are trained indiscriminately on their own outputs. (However, this differs from carefully designed synthetic data that can actually improve model efficiency.)

This downward spiral of AI pollution may soon resemble a classic “tragedy of the commons,” in which organizations act from self-interest at the expense of shared resources. If AI developers continue extracting data without limits or meaningful contributions, the shared resource of human creativity could eventually degrade for everyone.

Protecting the human spark

While AI models that simulate creativity in writing, coding, images, audio, or video can achieve remarkable imitations of human works, this sophisticated mimicry currently lacks the full depth of the human experience.

For example, AI models lack a body that endures the pain and travails of human life. They don’t grow over the course of a human lifespan in real time. When an AI-generated output happens to connect with us emotionally, it often does so by imitating patterns learned from a human artist who has actually lived that pain or joy.

A photo of a young woman painter in her art studio.

Even if future AI systems develop more sophisticated simulations of emotional states or embodied experiences, they would still fundamentally differ from human creativity, which emerges organically from lived biological experience, cultural context, and social interaction.

That’s because the world constantly changes. New types of human experience emerge. If an ethically trained AI model is to remain useful, researchers must train it on recent human experiences, such as viral trends, evolving slang, and cultural shifts.

Current AI solutions, like retrieval-augmented generation (RAG), address this challenge somewhat by retrieving up-to-date, external information to supplement their static training data. Yet even RAG methods depend heavily on validated, high-quality human-generated content—the very kind of data at risk if our digital environment becomes overwhelmed with low-quality AI-produced output.

This need for high-quality, human-generated data is a major reason why companies like OpenAI have pursued media deals (including a deal signed with Ars Technica parent Condé Nast last August). Yet paradoxically, the same models fed on valuable human data often produce the low-quality spam and slop that floods public areas of the Internet, degrading the very ecosystem they rely on.

AI as creative support

When used carelessly or excessively, generative AI is a threat to the creative ecosystem, but we can’t wholly discount the tech as a tool in a human creative’s arsenal. The history of art is full of technological changes (new pigments, brushes, typewriters, word processors) that transform the nature of artistic production while augmenting human creativity.

Bear with me because there’s a great deal of nuance here that is easy to miss among today’s more impassioned reactions to people using AI as a blunt instrument of creating mediocrity.

While many artists rightfully worry about AI’s extractive tendencies, research published in Harvard Business Review indicates that AI tools can potentially amplify rather than merely extract creative capacity, suggesting that a symbiotic relationship is possible under the right conditions.

Inherent in this argument is that the responsible use of AI is reflected in the skill of the user. You can use a paintbrush to paint a wall or paint the Mona Lisa. Similarly, generative AI can mindlessly fill a canvas with slop, or a human can utilize it to express their own ideas.

Machine learning tools (such as those in Adobe Photoshop) already help human creatives prototype concepts faster, iterate on variations they wouldn’t have considered, or handle some repetitive production tasks like object removal or audio transcription, freeing humans to focus on conceptual direction and emotional resonance.

These potential positives, however, don’t negate the need for responsible stewardship and respecting human creativity as a precious resource.

Cultivating the future

So what might a sustainable ecosystem for human creativity actually involve?

Legal and economic approaches will likely be key. Governments could legislate that AI training must be opt-in, or at the very least, provide a collective opt-out registry (as the EU’s “AI Act” does).

Other potential mechanisms include robust licensing or royalty systems, such as creating a royalty clearinghouse (like the music industry’s BMI or ASCAP) for efficient licensing and fair compensation. Those fees could help compensate human creatives and encourage them to keep creating well into the future.

Deeper shifts may involve cultural values and governance. Inspired by models like Japan’s “Living National Treasures“—where the government funds artisans to preserve vital skills and support their work. Could we establish programs that similarly support human creators while also designating certain works or practices as “creative reserves,” funding the further creation of certain creative works even if the economic market for them dries up?

Or a more radical shift might involve an “AI commons”—legally declaring that any AI model trained on publicly scraped data should be owned collectively as a shared public domain, ensuring that its benefits flow back to society and don’t just enrich corporations.

Photo of family Harvesting Organic Crops On Farm

Meanwhile, Internet platforms have already been experimenting with technical defenses against industrial-scale AI demands. Examples include proof-of-work challenges, slowdown “tarpits” (e.g., Nepenthes), shared crawler blocklists (“ai.robots.txt“), commercial tools (Cloudflare’s AI Labyrinth), and Wikimedia’s “WE5: Responsible Use of Infrastructure” initiative.

These solutions aren’t perfect, and implementing any of them would require overcoming significant practical hurdles. Strict regulations might slow beneficial AI development; opt-out systems burden creators, while opt-in models can be complex to track. Meanwhile, tech defenses often invite arms races. Finding a sustainable, equitable balance remains the core challenge. The issue won’t be solved in a day.

Invest in people

While navigating these complex systemic challenges will take time and collective effort, there is a surprisingly direct strategy that organizations can adopt now: investing in people. Don’t sacrifice human connection and insight to save money with mediocre AI outputs.

Organizations that cultivate unique human perspectives and integrate them with thoughtful AI augmentation will likely outperform those that pursue cost-cutting through wholesale creative automation. Investing in people acknowledges that while AI can generate content at scale, the distinctiveness of human insight, experience, and connection remains priceless.

Photo of Benj Edwards

Benj Edwards is Ars Technica’s Senior AI Reporter and founder of the site’s dedicated AI beat in 2022. He’s also a tech historian with almost two decades of experience. In his free time, he writes and records music, collects vintage computers, and enjoys nature. He lives in Raleigh, NC.

In the age of AI, we must protect human creativity as a natural resource Read More »

reusable-rockets-are-here,-so-why-is-nasa-paying-more-to-launch-stuff-to-space?

Reusable rockets are here, so why is NASA paying more to launch stuff to space?

• 1998: Deep Space 1 Delta II rocket — $86 million

• 1999: Mars Polar Lander Delta II rocket — $88 million

• 2001: Mars Odyssey Delta II rocket — $96 million

• 2003: Spirit and Opportunity Mars rovers — two Delta II rockets — $87 million per launch

• 2004: Swift Delta II rocket — $90 million

• 2005: Mars Reconnaissance Orbiter Atlas V rocket — $147 million

• 2007: Phoenix Mars lander — Delta II rocket — $132 million

Launch prices for NASA missions soared after the late 2000s, following the creation of United Launch Alliance through a merger of the Atlas and Delta rocket programs developed by Lockheed Martin and Boeing. The merger eliminated competition for most of NASA’s launch contracts until SpaceX’s Falcon 9 became available for NASA science missions in the mid-2010s. Here’s a sample of missions as examples of the rising costs, with contract values adjusted for inflation from the time of their award to reflect 2025 dollars:

• 2009: Lunar Reconnaissance Orbiter — Atlas V rocket — $220 million

• 2012: Radiation Belt Storm Probes — Atlas V rocket — $226 million (averaged from a bulk buy)

• 2014: Orbiting Carbon Observatory-2 — Delta II rocket — $191 million (averaged from a bulk buy)

• 2016: OSIRIS-REx asteroid mission — Atlas V rocket — $252 million

• 2017: TDRS-M data relay satellite — Atlas V rocket — $179 million

• 2017: JPSS-2 weather satellite — Atlas V rocket — $224 million

• 2018: InSight Mars lander — Atlas V rocket — $220 million

• 2018: ICESAT-2 — Delta II rocket — $134 million

Again, the missions listed above would likely launch on SpaceX’s Falcon 9 rockets if NASA awarded these contracts today. So, how do SpaceX’s more recent Falcon 9 prices compare? Let’s take a look. These contract values are adjusted for inflation from the time of their award to reflect 2025 dollars:

• 2016: Jason 3 oceanography satellite — Falcon 9 rocket — $114 million

• 2018: Transiting Exoplanets Survey Satellite — Falcon 9 rocket — $118 million

• 2020: Sentinel-6A — Falcon 9 rocket — $126 million

• 2021: Double Asteroid Redirection Test — Falcon 9 rocket — $86 million

• 2021: Imaging X-ray Polarimetry Explorer — Falcon 9 rocket — $62 million

• 2022: Surface Water and Ocean Topography — Falcon 9 rocket — $148 million

• 2024: PACE Earth sciences mission — Falcon 9 rocket — $99 million

• 2025: SPHEREx astronomy mission — Falcon 9 rocket — $99 million

And here are a few future launches NASA has booked to fly on SpaceX’s Falcon 9 rocket. Some of these contracts were awarded in the last 12 months, and those have not been adjusted for inflation. The others reflect 2025 dollars:

• 2025: Interstellar Mapping and Acceleration Probe — Falcon 9 rocket — $134 million

• 2025: Sentinel-6B — Falcon 9 rocket — $101 million

• 2027: NEO Surveyor — Falcon 9 rocket — $100 million

• 2027: JPSS-4 weather satellite — Falcon 9 rocket — $113 million

• 2027: Compton Spectrometer and Imager — Falcon 9 rocket — $69 million

There are a few other things worth noting when we chart NASA’s launch prices. One is that SpaceX’s Falcon Heavy, used for NASA’s heaviest missions, costs more than a Falcon 9 rocket. For example, two identical weather satellites launched in 2022 and 2024 on ULA’s Atlas V and SpaceX’s Falcon Heavy rocket for $207 million and $178 million, respectively, again adjusted for inflation.

Reusable rockets are here, so why is NASA paying more to launch stuff to space? Read More »

a-2,000-year-old-battle-ended-in-fire,-and-a-tree-species-never-recovered

A 2,000-year-old battle ended in fire, and a tree species never recovered

Then everything changed when the Fire Nation—sorry, the Han Empire—attacked.

Han rose to power in the wake of Qin’s collapse, after a short war with a rival dynasty called Chu, and spent the next century smugly referring to Nanyue as a vassal state and occasionally demanding tribute. At times, the rulers of Nanyue played along, but it all came to a head around 111 BCE, in the wake of an attempted coup and a series of assassinations. The Han Emperor sent an army of between 100,000 and 200,000 soldiers to invade Nanyue under a general named Lu Bode.

The troops marched across the countryside from five directions, converging outside Nanyue’s capital city of Panyou, which stood in the Pearl River Delta, near the modern city of Guangzhou. An enterprising company commander named Yang Pu got the bright idea to set the city on fire, and it ended badly.

“The fire not only destroyed the city but also ran out of control to the surrounding forests,” write Wang and colleagues. The cypress trees burned down to the waterline, leaving only their submerged stumps behind.

map of a coastal area showing elevation and the location of ancient forests

The brown dots mark the known sites of buried forests, and the orange diamonds mark those confirmed to be ancient. The two yellow diamonds are Wang and colleagues’ study sites. Credit: Wang et al. 2025

After war came fire and rice

At the time of the invasion, the land around Panyou was mostly swamp, forested with cypress trees. People had lived there for thousands of years, and had been growing rice for about 2,000 years. Bits of charcoal in the peat layers Wang and colleagues sampled reveal that they practiced slash-and-burn agriculture, but on a small scale, rotating their fields so the cypress forest could start to recover after a season or two.

The small burns are nothing like the forest fire Yang Pu unleashed, or the massive burning and reworking of the landscape that came after.

The stumps of the burned cypress trees slowly disappeared under several meters of peat, while above the buried ancient forest, life went on. Tigers, elephants, rhinos, and green peafowl no longer walked here. Instead, grains of pollen from the layers of clay above the peat reveal a sudden influx of plants from the grassy Poaceae family, which includes rice, wheat, and barley.

A 2,000-year-old battle ended in fire, and a tree species never recovered Read More »

perplexity-will-come-to-moto-phones-after-exec-testified-google-limited-access

Perplexity will come to Moto phones after exec testified Google limited access

Shevelenko was also asked about Chrome, which the DOJ would like to force Google to sell. Like an OpenAI executive said on Monday, Shevelenko confirmed Perplexity would be interested in buying the browser from Google.

Motorola has all the AI

There were some vague allusions during the trial that Perplexity would come to Motorola phones this year, but we didn’t know just how soon that was. With the announcement of its 2025 Razr devices, Moto has confirmed a much more expansive set of AI features. Parts of the Motorola AI experience are powered by Gemini, Copilot, Meta, and yes, Perplexity.

While Gemini gets top billing as the default assistant app, other firms have wormed their way into different parts of the software. Perplexity’s app will be preloaded, and anyone who buys the new Razrs. Owners will also get three free months of Perplexity Pro. This is the first time Perplexity has had a smartphone distribution deal, but it won’t be shown prominently on the phone. When you start a Motorola device, it will still look like a Google playground.

While it’s not the default assistant, Perplexity is integrated into the Moto AI platform. The new Razrs will proactively suggest you perform an AI search when accessing certain features like the calendar or browsing the web under the banner “Explore with Perplexity.” The Perplexity app has also been optimized to work with the external screen on Motorola’s foldables.

Moto AI also has elements powered by other AI systems. For example, Microsoft Copilot will appear in Moto AI with an “Ask Copilot” option. And Meta’s Llama model powers a Moto AI feature called Catch Me Up, which summarizes notifications from select apps.

It’s unclear why Motorola leaned on four different AI providers for a single phone. It probably helps that all these companies are desperate to entice users to bulk up their market share. Perplexity confirmed that no money changed hands in this deal—it’s on Moto phones to acquire more users. That might be tough with Gemini getting priority placement, though.

Perplexity will come to Moto phones after exec testified Google limited access Read More »

review:-ryzen-ai-cpu-makes-this-the-fastest-the-framework-laptop-13-has-ever-been

Review: Ryzen AI CPU makes this the fastest the Framework Laptop 13 has ever been


With great power comes great responsibility and subpar battery life.

The latest Framework Laptop 13, which asks you to take the good with the bad. Credit: Andrew Cunningham

The latest Framework Laptop 13, which asks you to take the good with the bad. Credit: Andrew Cunningham

At this point, the Framework Laptop 13 is a familiar face, an old friend. We have reviewed this laptop five other times, and in that time, the idea of a repairable and upgradeable laptop has gone from a “sounds great if they can pull it off” idea to one that’s become pretty reliable and predictable. And nearly four years out from the original version—which shipped with an 11th-generation Intel Core processor—we’re at the point where an upgrade will get you significant boosts to CPU and GPU performance, plus some other things.

We’re looking at the Ryzen AI 300 version of the Framework Laptop today, currently available for preorder and shipping in Q2 for people who buy one now. The laptop starts at $1,099 for a pre-built version and $899 for a RAM-less, SSD-less, Windows-less DIY version, and we’ve tested the Ryzen AI 9 HX 370 version that starts at $1,659 before you add RAM, an SSD, or an OS.

This board is a direct upgrade to Framework’s Ryzen 7040-series board from mid-2023, with most of the same performance benefits we saw last year when we first took a look at the Ryzen AI 300 series. It’s also, if this matters to you, the first Framework Laptop to meet Microsoft’s requirements for its Copilot+ PC initiative, giving users access to some extra locally processed AI features (including but not limited to Recall) with the promise of more to come.

For this upgrade, Ryzen AI giveth, and Ryzen AI taketh away. This is the fastest the Framework Laptop 13 has ever been (at least, if you spring for the Ryzen AI 9 HX 370 chip that our review unit shipped with). If you’re looking to do some light gaming (or non-Nvidia GPU-accelerated computing), the Radeon 890M GPU is about as good as it gets. But you’ll pay for it in battery life—never a particularly strong point for Framework, and less so here than in most of the Intel versions.

What’s new, Framework?

This Framework update brings the return of colorful translucent accessories, parts you can also add to an older Framework Laptop if you want. Credit: Andrew Cunningham

We’re going to focus on what makes this particular Framework Laptop 13 different from the past iterations. We talk more about the build process and the internals in our review of the 12th-generation Intel Core version, and we ran lots of battery tests with the new screen in our review of the Intel Core Ultra version. We also have coverage of the original Ryzen version of the laptop, with the Ryzen 7 7840U and Radeon 780M GPU installed.

Per usual, every internal refresh of the Framework Laptop 13 comes with another slate of external parts. Functionally, there’s not a ton of exciting stuff this time around—certainly nothing as interesting as the higher-resolution 120 Hz screen option we got with last year’s Intel Meteor Lake update—but there’s a handful of things worth paying attention to.

Functionally, Framework has slightly improved the keyboard, with “a new key structure” on the spacebar and shift keys that “reduce buzzing when your speakers are cranked up.” I can’t really discern a difference in the feel of the keyboard, so this isn’t a part I’d run out to add to my own Framework Laptop, but it’s a fringe benefit if you’re buying an all-new laptop or replacing your keyboard for some other reason.

Keyboard legends have also been tweaked; pre-built Windows versions get Microsoft’s dedicated (and, within limits, customizable) Copilot key, while DIY editions come with a Framework logo on the Windows/Super key (instead of the word “super”) and no Copilot key.

Cosmetically, Framework is keeping the dream of the late ’90s alive with translucent plastic parts, namely the bezel around the display and the USB-C Expansion Modules. I’ll never say no to additional customization options, though I still think that “silver body/lid with colorful bezel/ports” gives the laptop a rougher, unfinished-looking vibe.

Like the other Ryzen Framework Laptops (both 13 and 16), not all of the Ryzen AI board’s four USB-C ports support all the same capabilities, so you’ll want to arrange your ports carefully.

Framework’s recommendations for how to configure the Ryzen AI laptop’s expansion modules. Credit: Framework

Framework publishes a graphic to show you which ports do what; if you’re looking at the laptop from the front, ports 1 and 3 are on the back, and ports 2 and 4 are toward the front. Generally, ports 1 and 3 are the “better” ones, supporting full USB4 speeds instead of USB 3.2 and DisplayPort 2.0 instead of 1.4. But USB-A modules should go in ports 2 or 4 because they’ll consume extra power in bays 1 and 3. All four do support display output, though, which isn’t the case for the Ryzen 7040 Framework board, and all four continue to support USB-C charging.

The situation has improved from the 7040 version of the Framework board, where not all of the ports could do any kind of display output. But it still somewhat complicates the laptop’s customizability story relative to the Intel versions, where any expansion card can go into any port.

I will also say that this iteration of the Framework laptop hasn’t been perfectly stable for me. The problems are intermittent but persistent, despite using the latest BIOS version (3.03 as of this writing) and driver package available from Framework. I had a couple of total-system freezes/crashes, occasional problems waking from sleep, and sporadic rendering glitches in Microsoft Edge. These weren’t problems I’ve had with the other Ryzen AI laptops I’ve used so far or with the Ryzen 7040 version of the Framework 13. They also persisted across two separate clean installs of Windows.

It’s possible/probable that some combination of firmware and driver updates can iron out these problems, and they generally didn’t prevent me from using the laptop the way I wanted to use it, but I thought it was worth mentioning since my experience with new Framework boards has usually been a bit better than this.

Internals and performance

“Ryzen AI” is AMD’s most recent branding update for its high-end laptop chips, but you don’t actually need to care about AI to appreciate the solid CPU and GPU speed upgrades compared to the last-generation Ryzen Framework or older Intel versions of the laptop.

Our Framework Laptop board uses the fastest processor offering: a Ryzen AI 9 HX 370 with four of AMD’s Zen 5 CPU cores, eight of the smaller, more power-efficient Zen 5c cores, and a Radeon 890M integrated GPU with 16 of AMD’s RDNA 3.5 graphics cores.

There are places where the Intel Arc graphics in the Core Ultra 7/Meteor Lake version of the Framework Laptop are still faster than what AMD can offer, though your experience may vary depending on the games or apps you’re trying to use. Generally, our benchmarks show the Arc GPU ahead by a small amount, but it’s not faster across the board.

Relative to other Ryzen AI systems, the Framework Laptop’s graphics performance also suffers somewhat because socketed DDR5 DIMMs don’t run as fast as RAM that’s been soldered to the motherboard. This is one of the trade-offs you’re probably OK with making if you’re looking at a Framework Laptop in the first place, but it’s worth mentioning.

A few actual game benchmarks. Ones with ray-tracing features enabled tend to favor Intel’s Arc GPU, while the Radeon 890M pulls ahead in some other games.

But the new Ryzen chip’s CPU is dramatically faster than Meteor Lake at just about everything, as well as the older Ryzen 7 7840U in the older Framework board. This is the fastest the Framework Laptop has ever been, and it’s not particularly close (but if you’re waffling between the Ryzen AI version, the older AMD version that Framework sells for a bit less money or the Core Ultra 7 version, wait to see the battery life results before you spend any money). Power efficiency has also improved for heavy workloads, as demonstrated by our Handbrake video encoding tests—the Ryzen AI chip used a bit less power under heavy load and took less time to transcode our test video, so it uses quite a bit less power overall to do the same work.

Power efficiency tests under heavy load using the Handbrake transcoding tool. Test uses CPU for encoding and not hardware-accelerated GPU-assisted encoding.

We didn’t run specific performance tests on the Ryzen AI NPU, but it’s worth noting that this is also Framework’s first laptop with a neural processing unit (NPU) fast enough to support the full range of Microsoft’s Copilot+ PC features—this was one of the systems I used to test Microsoft’s near-final version of Windows Recall, for example. Intel’s other Core Ultra 100 chips, all 200-series Core Ultra chips other than the 200V series (codenamed Lunar Lake), and AMD’s Ryzen 7000- and 8000-series processors often include NPUs, but they don’t meet Microsoft’s performance requirements.

The Ryzen AI chips are also the only Copilot+ compatible processors on the market that Framework could have used while maintaining the Laptop’s current level of upgradeability. Qualcomm’s Snapdragon X Elite and Plus chips don’t support external RAM—at least, Qualcomm only lists support for soldered-down LPDDR5X in its product sheets—and Intel’s Core Ultra 200V processors use RAM integrated into the processor package itself. So if any of those features appeal to you, this is the only Framework Laptop you can buy to take advantage of them.

Battery and power

Battery tests. The Ryzen AI 300 doesn’t do great, though it’s similar to the last-gen Ryzen Framework.

When paired with the higher-resolution screen option and Framework’s 61 WHr battery, the Ryzen AI version of the laptop lasted around 8.5 hours in a PCMark Modern Office battery life test with the screen brightness set to a static 200 nits. This is a fair bit lower than the Intel Core Ultra version of the board, and it’s even worse when compared to what a MacBook Air or a more typical PC laptop will give you. But it’s holding roughly even with the older Ryzen version of the Framework board despite being much faster.

You can improve this situation somewhat by opting for the cheaper, lower-resolution screen; we didn’t test it with the Ryzen AI board, and Framework won’t sell you the lower-resolution screen with the higher-end chip. But for upgraders using the older panel, the higher-res screen reduced battery life by between 5 and 15 percent in past testing of older Framework Laptops. The slower Ryzen AI 5 and Ryzen AI 7 versions will also likely last a little longer, though Framework usually only sends us the highest-end versions of its boards to test.

A routine update

This combo screwdriver-and-spudger is still the only tool you need to take a Framework Laptop apart. Credit: Andrew Cunningham

It’s weird that my two favorite laptops right now are probably Apple’s MacBook Air and the Framework Laptop 13, but that’s where I am. They represent opposite visions of computing, each of which appeals to a different part of my brain: The MacBook Air is the personal computer at its most appliance-like, the thing you buy (or recommend) if you just don’t want to think about your computer that much. Framework embraces a more traditionally PC-like approach, favoring open standards and interoperable parts; the result is more complicated and chaotic but also more flexible. It’s the thing you buy when you like thinking about your computer.

Framework Laptop buyers continue to pay a price for getting a more repairable and modular laptop. Battery life remains OK at best, and Framework doesn’t seem to have substantially sped up its firmware or driver releases since we talked with them about it last summer. You’ll need to be comfortable taking things apart, and you’ll need to make sure you put the right expansion modules in the right bays. And you may end up paying more than you would to get the same specs from a different laptop manufacturer.

But what you get in return still feels kind of magical, and all the more so because Framework has now been shipping product for four years. The Ryzen AI version of the laptop is probably the one I’d recommend if you were buying a new one, and it’s also a huge leap forward for anyone who bought into the first-generation Framework Laptop a few years ago and is ready for an upgrade. It’s by far the fastest CPU (and, depending on the app, the fastest or second-fastest GPU) Framework has shipped in the Laptop 13. And it’s nice to at least have the option of using Copilot+ features, even if you’re not actually interested in the ones Microsoft is currently offering.

If none of the other Framework Laptops have interested you yet, this one probably won’t, either. But it’s yet another improvement in what has become a steady, consistent sequence of improvements. Mediocre battery life is hard to excuse in a laptop, but if that’s not what’s most important to you, Framework is still offering something laudable and unique.

The good

  • Framework still gets all of the basics right—a matte 3:2 LCD that’s pleasant to look at, a nice-feeling keyboard and trackpad, and a design
  • Fastest CPU ever in the Framework Laptop 13, and the fastest or second-fastest integrated GPU
  • First Framework Laptop to support Copilot+ features in Windows, if those appeal to you at all
  • Fun translucent customization options
  • Modular, upgradeable, and repairable—more so than with most laptops, you’re buying a laptop that can change along with your needs and which will be easy to refurbish or hand down to someone else when you’re ready to replace it
  • Official support for both Windows and Linux

The bad

  • Occasional glitchiness that may or may not be fixed with future firmware or driver updates
  • Some expansion modules are slower or have higher power draw if you put them in the wrong place
  • Costs more than similarly specced laptops from other OEMs
  • Still lacks certain display features some users might require or prefer—in particular, there are no OLED, touchscreen, or wide-color-gamut options

The ugly

  • Battery life remains an enduring weak point.

Photo of Andrew Cunningham

Andrew is a Senior Technology Reporter at Ars Technica, with a focus on consumer tech including computer hardware and in-depth reviews of operating systems like Windows and macOS. Andrew lives in Philadelphia and co-hosts a weekly book podcast called Overdue.

Review: Ryzen AI CPU makes this the fastest the Framework Laptop 13 has ever been Read More »

backward-compatible:-many-old-oblivion-mods-still-work-on-oblivion-remastered

Backward compatible: Many old Oblivion mods still work on Oblivion Remastered


The modding community is already hard at work despite lack of “official” support.

Thanks to a circa 2008 mod, I have a ton of armor and weapons from the jump in Oblivion Remastered Credit: Kyle Orland / Bethesda

Thanks to a circa 2008 mod, I have a ton of armor and weapons from the jump in Oblivion Remastered Credit: Kyle Orland / Bethesda

Bethesda isn’t officially supporting mods for the newly released Elder Scrolls IV: Oblivion Remastered. But that hasn’t stopped some players from discovering that many mods created for the 2006 original seem to work just fine in the new game with a bare minimum of installation headaches.

As noted on Reddit and the Bethesda Game Studios Discord, some .esp mod files designed years ago for the original Oblivion have the same effect when plugged into the new Remastered game. Ars confirmed this during some quick testing, using a mod uploaded in 2008 to easily add high-end weapons and armor to the opening jail cell scene in the Remastered version.

While players of the original game could use the Oblivion Mod Manager to easily install these mods, doing so in the Remastered version requires a bit more manual work. First, users have to download the applicable .esp mod files and put them in the “Content/Dev/ObvData/Data” folder (the same one that already houses DLC data files like “DLCHorseArmor.esp”). Then it’s just a matter of opening “Plugins.txt” in the same folder and adding that full .esp file name to the plaintext list.

Early testers report that some more complex mods designed for the original Oblivion will lead to crashes or mixed results when loaded in the Remastered game. Others theorize that “the game seems to read OG Oblivion as its base, so manually adding a mod that doesn’t have new assets … would work.”

As the community continues to analyze this modding backward compatibility, other modders have already rushed to release dozens of new mods designed specifically for the Remastered version, even without official support from Bethesda. At this point, most of these seem focused on some basic UI tweaks or quality-of-life hacks to make the game more enjoyable (we’re particularly fond of this one that makes slow-walking NPCs a bit faster).

More complex mods may require diving into the Unreal Engine’s pak file format to replace in-game assets with new, modded versions. That means it’s probably just a matter of time before we get the equivalent of the custom Mystic Elf race modded back into the newer version of the game.

Photo of Kyle Orland

Kyle Orland has been the Senior Gaming Editor at Ars Technica since 2012, writing primarily about the business, tech, and culture behind video games. He has journalism and computer science degrees from University of Maryland. He once wrote a whole book about Minesweeper.

Backward compatible: Many old Oblivion mods still work on Oblivion Remastered Read More »