Author name: Kris Guyer

china-drafts-world’s-strictest-rules-to-end-ai-encouraged-suicide,-violence

China drafts world’s strictest rules to end AI-encouraged suicide, violence

China drafted landmark rules to stop AI chatbots from emotionally manipulating users, including what could become the strictest policy worldwide intended to prevent AI-supported suicides, self-harm, and violence.

China’s Cyberspace Administration proposed the rules on Saturday. If finalized, they would apply to any AI products or services publicly available in China that use text, images, audio, video, or “other means” to simulate engaging human conversation. Winston Ma, adjunct professor at NYU School of Law, told CNBC that the “planned rules would mark the world’s first attempt to regulate AI with human or anthropomorphic characteristics” at a time when companion bot usage is rising globally.

Growing awareness of problems

In 2025, researchers flagged major harms of AI companions, including promotion of self-harm, violence, and terrorism. Beyond that, chatbots shared harmful misinformation, made unwanted sexual advances, encouraged substance abuse, and verbally abused users. Some psychiatrists are increasingly ready to link psychosis to chatbot use, the Wall Street Journal reported this weekend, while the most popular chatbot in the world, ChatGPT, has triggered lawsuits over outputs linked to child suicide and murder-suicide.

China is now moving to eliminate the most extreme threats. Proposed rules would require, for example, that a human intervene as soon as suicide is mentioned. The rules also dictate that all minor and elderly users must provide the contact information for a guardian when they register—the guardian would be notified if suicide or self-harm is discussed.

Generally, chatbots would be prohibited from generating content that encourages suicide, self-harm, or violence, as well as attempts to emotionally manipulate a user, such as by making false promises. Chatbots would also be banned from promoting obscenity, gambling, or instigation of a crime, as well as from slandering or insulting users. Also banned are what are termed “emotional traps,”—chatbots would additionally be prevented from misleading users into making “unreasonable decisions,” a translation of the rules indicates.

China drafts world’s strictest rules to end AI-encouraged suicide, violence Read More »

childhood-and-education-#16:-letting-kids-be-kids

Childhood and Education #16: Letting Kids Be Kids

The Revolution of Rising Requirements has many elements. The most onerous are the supervisory requirements on children. They have become, as Kelsey Piper recently documented, completely, utterly insane, to the point where:

  1. A third of people, both parents and non-parents, responded in a survey that it is not appropriate to leave a 13 year old at home for an hour or two, as opposed to when we used to be 11 year olds babysitting for other neighborhood kids.

  2. A third of people said in that same survey that if a 10-year-old is allowed to play alone in the park, there needs to be an investigation by CPS.

Whereas I think that if you don’t allow your 10-year-old to play alone in a park, that is a much better (although still quite bad) potential reason for a CPS investigation.

This is not an idle threat, per the common statistic that around 35% of American families get investigated by CPS. Even if you are confident that will ultimately turn out fine, and given the vagaries and insanities one can never fully be sure, the process is already the punishment.

As Kelsey Piper says, we don’t want a lot of 14-year-olds being breadwinners for their families. But this is so bad in the other direction it might be even worse than that, even discounting the kids that this causes to never be born at all.

Kids need to be kids. We don’t let them. It’s a big problem, both greatly raising the dollar, time and lifestyle costs of having kids and also destroying their childhoods.

This post is about various ways of seeing exactly how bad things have gotten.

Some dire statistics from the Harris poll.

Harris Poll: More than half of the kids surveyed have not experienced many real-life experiences on their own. According to the kids surveyed aged 8 to 12 years old:

  • 45% have not walked in a different aisle than their parents at a store

  • 56% have not talked with a neighbor without their parents

  • 61% have not made plans with friends without adults helping them

  • 62% have not walked/biked somewhere (a store, park, school) without an adult

  • 63% have not built a structure outside (for example, a fort or treehouse)

  • 67% have not done work that they’ve been paid for (e.g., mowing lawns, shoveling snow, babysitting)

  • 71% have not used a sharp knife

Across in-person and virtual spaces, experiences differ for children living in rural, urban, or suburban areas:

  • 56% of 8 to 12-year-olds in urban areas have not walked in a different aisle from their parents at a store, 44% in suburban areas have not, and 37% in rural areas have not.

  • 51% of 8 to 12-year-olds in urban areas have not talked with a neighbor without parents, 61% suburban areas have not, and 56% in rural areas have not.

  • 28% of 8 to 12-year-olds in urban areas say they have talked, chatted, or messaged with strangers online, 17% of 8-12 year olds in suburban areas say they have, and 25% in rural areas say they have.

Have not walked in a different aisle in a store or never talked to a stranger or even a neighbor is positively bonkers, as is ‘have not walked somewhere without an adult.’

Why are kids on their phones and tablets all the time? How could we stop this?

Easy, you let them have unstructured playtime, that’s it, that’s how you do it.

All you have to do is let them. They want unstructured free play time without adults. They know this is The Way. It’s free, it’s easy, it’s deeply safe, it’s good for them, they enjoy it, we’re just completely bonkers and have decided this is not allowed, somehow.

When given the choice between three types of social interaction, – unstructured play (e.g., playing outside or pickup games), structured adult-led activities (e.g., sports or lessons), or socializing online – kids overwhelmingly chose unstructured, in-person play as their favorite way to spend time with friends and the vast majority of them would rather spend most of their time doing things in person, without screens.

  • Almost three-quarters (72%) of 8 to 12-year-olds say they would rather spend most of their time together doing things in-person, without screens (rather than spend most of their time together on screens and devices).

  • When given the option:

    • 45% said they would participate in an activity with their friends in person that’s not organized by adults, like a made up game, playing card, basketball, or exploring

    • 30% said they would participate in an organized activity or class, like soccer, dance, or karate

    • 25% said they would participate in an online activity with their friends like playing video games

  • 61% want to play with friends in person without adults:

  • 87% wish they could spend more time with their friends in person outside of school

This problem mostly isn’t dastardly addictive algorithms. Mostly it is that we won’t let our children play in any other way, so what do you expect? You’re not offering any alternatives. You can offer non-algorithmic electronic alternatives, and they’re better than the algorithms, but either way this is us imposing this on them, not them being addicted.

Lenore Skenazy, Zach Rausch, and Jonathan Haidt (so yeah, the usual suspects): In March, the Harris Poll surveyed more than 500 children ages 8 to 12 across the United States, who were assured that their answers would remain private. They offered unmistakable evidence that the phone-based childhood is in full force. A majority reported having smartphones, and about half of the 10-to-12-year-olds said that most or all of their friends use social media.

This digital technology has given kids access to virtual worlds, where they’re allowed to roam far more freely than in the real one. About 75 percent of kids ages 9 to 12 regularly play the online game Roblox, where they can interact with friends and even strangers. But most of the children in our survey said that they aren’t allowed to be out in public at all without an adult. Fewer than half of the 8- and 9-year-olds have gone down a grocery-store aisle alone; more than a quarter aren’t allowed to play unsupervised even in their own front yard.

What do kids want? The ability to move around. Free play, in person, with other kids.

As I keep saying, essentially everyone sane realizes this.

But everyone is terrified, not without reason, that if you try this strangers will call the police on your children. So out of fear that some stranger might abduct your children, which is ~0% to ever happen and less likely than ever for any given activity, strangers will… abduct your children via calling the government to do it.

They will do this on the thinnest of hair triggers. ‘Grocery aisle’ above was not a metaphor, we mean literally not allowed to go down a grocery aisle.

Cartoons Hate Her!: Okay yes but when I let my kid wander 10 feet from me (within eyeshot) in a toy store I had people on here telling me I had committed child neglect so who’s to say.

Multiple people said he could have been sex trafficked because it “only takes a moment.”

The Televisionary: I’ve been confronted in public, twice, by people thinking I’m kidnapping my own kids. And weirdly, in both cases it’s in part because they think my oldest, who has long hair, is a girl and not a boy.

Like both times they apologize after discovering that?

Mr. Tweety: I was jogging in a park with my 12 y.o. son. Safe area, daytime. People walking dogs & such. He got maybe 30 feet ahead of me on the trail because I’m old & fat & these 2 women came up to me frantic:

“Is that your child!?”

“Yeah.”

“Oh thank god. We were about to call the police.”

Billy Binion: These stories are totally insane. CPS investigated a small-town Virginia mom *fourtimes for..letting her kids play outside unsupervised. That used to be called a “normal childhood.” Absurd that we now require helicopter parenting by law.

I think we have a new crazy requirement record.

Lenore Skenazy: During that visit, I was told that children could never be left alone, inside or outside the home—EVEN IN THEIR OWN BEDROOMS—until they were 13 years old. Social Services said specifically that I had to be in each room with them at all times until they were 13. That investigation ended without incident.

When I asked what constitutes supervision, she said that I had to be visible to my neighbors when the kids were outside, regardless of whether or not I could see the children. I asked where that was found in the Virginia law. She replied that it isn’t in the Virginia law, but that Social Services has its own set of rules.

Here’s another case:

Lenore Skenazy: Alexandra Woodward, a mother of 8- and 10-year-old boys in Calhoun, Georgia, has been charged with cruelty to children in the first degree. If found guilty, she faces a minimum of five years in prison. Her crime? Letting her kids stay home alone for a few hours. They were fine.

This ‘people will tell you acting like a normal person is criminal’ pattern is deep and wide, and it only takes one person to call the police. It could get worse:

Max Factor: I am, by all accounts, still a young person. I’m gen z. And I’m talking specifically about the people in her mentions calling her a rapist or “promoting rape culture” for saying people should be allowed to have sex in their own living rooms

Pavlova: We are about 2 days away from people saying people are pedophiles for like, having sex in the same house their kids live in.

Madison: i distinctly remember there was a tiktok going around where this couple talked about their sex noises waking their toddler up and people reacted by calling that “sexual abuse” and said you shouldn’t have sex in the home at all if there are kids, even if the kids are sleeping.

Cartoons Hate Her!: I love that everyone is like “shut up nobody ever said it was sexual abuse to have sex in the same house as your children” and then other people literally saying that.

A twelve year old is paranoid that if they go into the donut shop they’ll get questioned about why they’re alone.

A thirteen year old is not allowed to be alone in a public park.

A seventeen year old is not allowed to go to Target.

An 8th grader is forced into indentured servitude (they call it) ‘volunteer hours for career path class’) but no one will agree to let him serve them.

Bethany: I just sent my 12 year old in to go get a dozen donuts while I waited in the car.

“Mom they will wonder why I’m alone.”

“What will I say when they ask?”

And so on…

Guys this helicopter society is not good for the kids.

The big concern is not him doing something but in how society will react to independence.

Thats a problem.

Polimath: My kids used to love walking to Target until the local Target changed their policy to “no unaccompanied kids under 18”

It’s v frustrating. I’m looking for chances to help my kids be independent & I have basically no societal cooperation on this project

Sally Hammer: My 8th grader has to do volunteer hours for his “career path class”…guess what? No one lets a 13 year old kid volunteer.

WOPR: There is a park at a recreation center about 200 yards from my house. When my youngest daughter was 13 she went to the park with a couple of neighborhood kids to go swing. She returned home after being gone for less than 30 minutes. I asked her why she was home and she told me “the man at the park said we couldn’t be there without parents.”

So I called the rec director and asked him what was going on. He told me that they don’t allow unaccompanied kids under the age of 16. I argued with him that 16 year olds didn’t hang out in parks unless they were drinking (or worse). I asked what good are parks when kids can’t play in them. He got pissed and told me my kid(s) were banned. This is in a small Southern town of about 5000. It’s incredible considering the freedom I had in the 70s and 80s.

On a positive note, I donated $2500 to the guy that ran against him for the rec director on the condition that if he won, parks would be open to all well-behaved children. My guy won and told me it was the first time anyone had made a contribution to a parks & rec director race.

What are the odds on child abduction by a stranger who isn’t with the government and isn’t involved in a custody dispute?

There are 72 million kids in America and about 100 non-governmental kidnappings by strangers a year.

Let that number sink in. That’s it. We ruin our lives over that.

If you left your child unattended, the original claim is that they would get kidnapped once every 750,000 years. Andrew Critch claims the math is off and it would ‘only’ take ~37,500 years, which seems likely to be closer to accurate, but is still really a lot.

Almost all missing children ran away or were taken by people you know, or by authorities, or simply temporarily got lost.

However, the main concern was never strangers kidnapping the child directly, it was strangers observing the child and then calling authorities to do the kidnapping:

Andrew Rettek: People get pissed at you if you act like this is true.

Ben Hoffman: So, ah, I let my toddler sleep in his stroller in my front yard, and *Igot kidnapped as a result. My experience suggests that these statistics may create a misleading illusion of safety.

The Wasatchquatch: 100% I’m happy to live in a state that allows free range parenting. We had the cops called on us because our 6 yo was 1 house away. Absolute insanity.

Who should you be worried will report on you, in general? Random strangers will definitely do it if you appear to leave children unsupervised. Even if the law explicitly says the kids are allowed to be unsupervised, crazy people will report them anyway.

Otherwise it’s mostly professionals, and risk goes way down after the first year, although it remains high.

Maxwell Tabarrok: 37% of all American children are investigated by CPS.

2 million investigations, 530k substantiated cases, and 200k family separations every year.

Most reports come from non-relative professionals like teachers, and most victims are under the age of 4.

Some other striking facts: 70% of reports come from professionals like police, teachers, and doctors.

Most reports come from these groups because they are criminally liable if they observe evidence of child abuse or neglect and do not report it.

The majority of children classified as victims by the CPS are less than four years old and the plurality of victims are less than 1.

For practical purposes, it is correct to act as if ~100% of the risk from strangers is that they call upon the authorities to punish you, and ~0% of it is them harming the child.

The craziest part about ‘stranger danger’ not existing is the lack of joy about this fact, and the craziest part about ‘you have to have eyes on your toddler at literal all times or else’ is that people thought that made the slightest bit of physical sense.

Yet here we are.

Words Matter: This conversation started out about leaving kids in a car.

That’s an easy lure for child traffickers.

I can understand going to the mailbox (mine is 14 minutes away round trip), but leaving kids in a car is just putting the bait right under the criminals’ noses.

Mason: She blocked me after posting this (lame)

But suffice it to say I am not worried about roving bands of child traffickers in the Chipotle parking lot, because they do not exist

Laura Robinson: Can anyone find me a single news article about a child ending up trafficked in the United States because they were abducted?

Serious question.

I’ve looked and looked and I’ve never seen evidence this has happened one time.

There’s about 100 children per year kidnapped by strangers a year in he US and 60 percent of them are returned home alive. Other 40 usually found dead. We can literally ask them.

I know I say this all the time but it will never stop being weird to me that, when I first started looking into this, I thought I was telling people, “Guess what? Kids aren’t getting kidnapped and trafficked in the US!” that I thought I was telling people great news and I wasn’t.

On the whole, people get *absolutely furiousif you tell them that no one is after their children.

I think it’s probably a natural outgrowth of the shock of cognitive dissonance or the fear that someone with paradigm-breaking information is trying to get one over on you, but on the whole, “no one is trying to kidnap and sell your five-year-old to rapists” is NOT news anyone wants to hear, believe it or not.

You’d think it sounds like great news but it apparently is not.

It was great news to me. I very much like the fact that no one is trying to kidnap my children. Or at least, no one except CPS, which may try to do this if I take ‘no one is trying to kidnap my kids’ too seriously and give them sensible levels of freedom.

*The obvious reason why this upsets people so much is probably that it does force a reframing of violence against kids.

“Well, who’s doing all the child abuse, then?”

“Mostly parents and people who parents trust.”

That’s pretty upsetting if you’ve never thought about it before.

For example, last year there was a pretty big news story going around of a group of people who were using a storm shelter near a trailer park to make CSAM and sell it.

If you looked at the comments on the news story, it was wall-to-wall “this is awful but I’m glad the kids will be returned to their parents.”

If you pulled the court docs, the traffickers were the parents.

I think it’s morally comforting to think that this only happens to kids if a boogeyman takes them and the happy ending is that the kids can go home.

That’s not how this works in real life, unfortunately.

It is hard to overstate how harmful it is that we therefore cannot let kids roam free until long past the age it makes sense to allow this. It impoverishes childhood, is terrible for the kids long term and it imposes immense costs on parents.

Thrilla the Gorilla: Did parents in the 70s/80s/90s really allow their kids to roam freely, or is that just a portrayal seen in movies?

Katie: an underreported reason people are having fewer and fewer kids: now we’re expected to watch them 24/7. at least in the summer my mom got 10+ hours a day free from me while I crawled around in ditches.

I don’t know that we would have been able to have more children if the de facto laws around all this were less insane, but there’s a pretty good chance of it.

There was a story going around where parents let two children, 10 and 7, walk to a grocery store ten minutes away, one was struck by a car and killed, and the district attorney charged the parents – not the driver, the parents – with involuntary manslaughter and set bail at $1.5 million, despite previously only imposing $50k in bail for a parent who kept a loaded gun in the house that a kid got a hold of, that then went off and shot another kid.

In this particular case, there were various reasons that this was a lot less outrageous than it sounds. The road they were jaywalking was four lanes, two ways at 50 miles an hour. There had been numerous incidents at the house with drugs and domestic abuse prior to this.

Presumably the DA was dropping the hammer on things in general.

I get all that. This is still completely bonkers insane.

One thing that happens when you call the cops on parents who let kids walk home is you get this:

Also consider letting kids be bored? As in, having a calm and quiet house where kids have opportunity to do creative things or read books and so on, but you don’t give them easy entertainment outs like screens, and don’t consider it your problem if they say they’re bored. Advanced level is also letting them experience being potentially bored outside on their own, if you can pull that off.

Boze the Library Owl: When I was ten, I hosted my own “Academy Awards of Books” where I gave prizes to the best books I had read that year, and I wrote acceptance speeches for Ernest Hemingway and Edgar Allan Poe, and I just think kids can do amazing things if they’re allowed to be a bit bored.

Dave: Allowing our kids to be bored has been one of the most successful experiments we’ve done as parents.

They each play 3-5 instruments, compose their own music, make things, and read constantly.

My 10yo is more literate than the average American.

My 12yo teaches music theory.

You can’t let kids be kids primarily for fear others will see them being kids, and this also applies to other interactions others might witness. This is a relatively harmless situation, and yet, man, very awkward.

Owen Cyclops: parenting has this odd social dimension where you’re always actively engaging with how other people see you. so i go to this street fair. my son (3) gets stung by a wasp. never been stung before. freaks out. i take him out of the crowd and put him on some grass. he’s fine.

i also just so happened to have obtained an extremely large gyro seconds before this. in my haste and preference for my own flesh and blood, i abandoned the gyro. when my son is injured, he just wants things to be normal. he personally insists i go back and re-obtain the gyro.

he doesnt want to talk about being injured, doesnt want any attention, he just wants everything to stay normal and not orbit around him so he can deal with it. great. so he wants me, his dad, eating his food like normal, on this patch of grass while he recovers from a wasp sting.

while this sounds reasonable, what this actually results in is: me, relaxed. stuffing my face with a gyro, two feet away from a small boy who is, literally, just writhing in pain and openly weeping, while the street fair crowd passes before us. every single person looks at us.

i am getting absolutely horrified looks from mothers, other fathers, children, perhaps even the dogs, who are all attempting to imagine the character of a man who would dump his injured son, in pain, weeping, on a patch of grass, so he could unflinchingly relax and eat a gyro.

but he requested this. in fact, this is clearly the best course of action. the wasp stung his foot: he can’t walk, and he’s fine. but i cannot communicate this to “the crowd”. i am misunderstood. i appear as a monster. yet i bear the arrows of this false appearance nobly, for him

between bouts of him sob-yelling in pain, a woman comes over to ask me what happened. i said: he got stung my a wasp, he’s fine. she says, take him to the police station (across the street). i said: i don’t think they can arrest a wasp. unfortunately she was not amused by this.

Andrew Rettek: This describes my experience as a parent.

Ideally, if people are well calibrated and enforcing good norms, this dynamic is actively helpful. Other adults and the desire to avoid minor social awkwardness or worse acts to nudge you towards better choices. In an atomized world where people’s instincts are often some combination of crazy and superficial, and where remarkably often they feel this obligates or allows them to escalate to the authorities, this ends up not going so well.

Bryan Caplan points out that if you think modern smartphones how we use them are especially terrible for children, you can always in his words ‘do the time warp again’ and travel back into the past, providing your kids with older screen babysitter technology, however much older you think solves your problem. You can spin up a VCR if you want.

He’s right. It’s crazy to give up the power of the screen entirely, the cost of doing that is crazy stupid high. It’s especially stupid high given you’ve lost the old ability to let your children play outside. Inside? The old world can still exist, if you want it to.

The VCR trick presumably works for a two year old, since they don’t know any better. Bryan downplays the difficulty of the kids finding out about phones and tablets and streaming television and so on, since in addition to other families and kids they’re going to see you using them.

You can still set whatever restrictions you want. And I do.

However modern experiences very quickly spoil older experiences, and avoiding contact becomes very difficult. No, you can’t really expect kids to watch a bunch of old VCR tapes with old cartoons on them, and my attempts to get my kids to watch most of the ‘shows of my youth’ that I remembered fondly did not work at all. You often can’t go home again.

In other ways? You can go home again. I’ve had great success giving my kids a Mini-NES and mini-SNES, and having them largely play older video games, and I think that was a big win. You need to take a page from the Amish, and choose what to accept versus reject.

Also, um, Bryan, you do know what the song ‘do the time warp again’ was about?

Our approach to childhood is to imprison kids most of their waking hours, both at school and then at home, direct most of their activities or else in ways that look and are largely stupid and pointless, force them to interact with a peer group that includes various forms of bullying with no form of exit or choice, and so on, giving them little free time or opportunity to play outside or anything like that.

Then, if they are not happy, they are basically told to suck it up, unless they can be labeled as ‘depressed.’

And then we go around periodically asking them ‘are you depressed?’

If they say yes, of course, you don’t change any of the above. You drug the kid.

Ilinois Governor JB Pritzker: Illinois is now the first state in the nation to require mental health screenings in its public schools. Our schools should be inclusive places where students are not just comfortable asking for help — they’re empowered to do it.

Eliezer Yudkowsky: > not just comfortable asking for help — they’re empowered An AI wrote that.

Abigail Shrier: I want to be on-the-record and crystal clear. This is a disastrous policy that will do vastly more harm than good. Watch as tens of thousands of Illinois kids get shoved into the mental health funnel and convinced they are sick. Many or most of which will be false positives.

Aaron Stupple: Constantly pestering kids to see if they are depressed, and offering them support if they say yes, might be driving much of the rise of teen depression and anxiety.

Mason: Parents should focus less on whether their kids are depressed in the clinical sense and more on whether they’re happy in the mundane sense

The idea that kids need to just continuously suck it up until they’ve got a clinically diagnosable mental illness is driving all kinds of weird incentives. If nobody is listening to you until you’re a victim or a mental patient, well

Under current conditions, I too predict that this policy is disastrous and will do large net harm. Our mental health screening system has too many false positives even if you first need to have a reasonable suspicion before checking.

Is this an argument that phones are fine?

Wide Of The Post: Kids used to watch an insane amount of TV, both actively and passively channel surfing, even just on as background noise. I doubt a lot of younger zoomers fully grasp how much TV people used to watch, but it’s important missing context for the social media/phone use moral panics.

Zac Hill: ‘TV Discourse’ is indeed *veryrelevant to current Screen Discourse, but for Wallacean Total Noise/E Unibus Pluram reasons regarding attention capture and direction and not as a like mechanical 1:1 analogy

I think very clearly no, for three reasons.

  1. Phones are carried around in your pocket, a constant distraction and temptation even when you are not using them. This is a big difference.

  2. The new short form content seems clearly way worse. Imagine children switching from watching old television shows to YouTube Kids or Shorts on that same TV.

  3. Television was a correct moral panic, and letting kids watch tons of TV sucked.

A study out of China claims that ‘a one standard deviation increase in app usage reduces GPAs by 36.2% of a within-cohort-major standard deviation, and lowers wages by 2.3%’ and that extending China’s three-hour-per-week video game limit to college students would increase their initial wages by 0.9%.

That is an insane effect size.

The part about the extension is almost certainly wrong, because college performance is largely signaling and a positional good, so you can’t predict what a universal boost in performance would do to initial wages, probably very little even if it raised real human capital levels, also the ban seems hard to enforce.

They use a natural experiment identifier from the timing of a blockbuster release to try and isolate changes in app use, which is intriguing. My presumption is that they did something wrong somewhere to get an effect this large, but we’ve seen a lot of studies with absurdly low impacts from phone distractions and wasted time, so we should also note when the number comes out too large.

If you check everyone, given the likely way they’ll react to false positives? Oh no.

Discussion about this post

Childhood and Education #16: Letting Kids Be Kids Read More »

ai-#148:-christmas-break

AI #148: Christmas Break

Claude Opus 4.5 did so well on the METR task length graph they’re going to need longer tasks, and we still haven’t scored Gemini 3 Pro or GPT-5.2-Codex. Oh, also there’s a GPT-5.2-Codex.

At week’s end we did finally get at least a little of a Christmas break. It was nice.

Also nice was that New York Governor Kathy Hochul signed the RAISE Act, giving New York its own version of SB 53. The final version was not what we were hoping it would be, but it still is helpful on the margin.

Various people gave their 2026 predictions. Let’s put it this way: Buckle up.

  1. Language Models Offer Mundane Utility. AI suggests doing the minimum.

  2. Language Models Don’t Offer Mundane Utility. Gemini 3 doesn’t believe in itself.

  3. Huh, Upgrades. ChatGPT gets some personality knobs to turn.

  4. On Your Marks. PostTrainBench shows AIs below human baseline but improving.

  5. Claude Opus 4.5 Joins The METR Graph. Expectations were exceeded.

  6. Sufficiently Advanced Intelligence. You’re good enough, you’re smart enough.

  7. Deepfaketown and Botpocalypse Soon. Don’t worry, the UK PM’s got this.

  8. Fun With Media Generation. Slop as cost shock, enabling of niche pursuits.

  9. You Drive Me Crazy. Anthropic’s plans to handle mental health issues.

  10. They Took Our Jobs. What does it take to break a guild cartel?

  11. The Art of the Jailbreak. It still always works but it takes somewhat longer.

  12. Get Involved. MATS Summer 2026 cohort applications are open.

  13. Introducing. GPT-5.2-Codex is here to tide us over until the new year.

  14. In Other AI News. Small models can introspect, so can Andrej Karpathy.

  15. Show Me the Money. Anthropic going public, Project Vend breaks new ground.

  16. Quiet Speculations. Predictions for next year, new higher bars for what is AGI.

  17. Whistling In The Dark. It is still so early, almost no one knows Anthropic exists.

  18. Bubble, Bubble, Toil and Trouble. So many still don’t realize AI works.

  19. Americans Really Dislike AI. Attempts continue to mislead us about this.

  20. The Quest for Sane Regulations. NY’s RAISE Act is signed.

  21. Chip City. Chip smuggling, eventual chip production.

  22. The Week in Audio. Anthropic’s Sholto Douglas makes 2026 predictions.

  23. Rhetorical Innovation. Understanding AI teaches you to think about the world.

  24. Aligning a Smarter Than Human Intelligence is Difficult. The meta game.

  25. Mom, Owain Evans Is Turning The Models Evil Again. Train the interpreter.

  26. Messages From Janusworld. Claude Opus 3 zero hour approaches.

  27. The Lighter Side. What are you even doing?

AI custom-designed, human-in-the-loop proactive LLM-based mental health intervention has a positive effect in an RCT. There was significantly greater positive affect, resilience and social well-being. My presumption is that this was a highly conservative design due to ethical considerations. And that was using a system based on GPT-4o for 5-20 minutes a week. There is so much room for improvement here.

A lot of the benefits here likely came from implementation of low-hanging fruit interventions we know work, like having the system suggest journaling, gratitude exercises, mindfulness and social connection. We all know that stuff works. If an LLM-based scaffold actually gets people to do some of it? Great, that’s a huge win.

Results like this will not, as David Manheim suggests, prevent people from saying ‘but sometimes there are still bad outcomes’ or ‘but sometimes this ends up doing net harm,’ since nothing capable of working would prevent those risks entirely.

You can have Claude Code make objects in Unreal Engine on demand.

Seth Lazar on how he uses AI agents for philosophy. They automate everything around the thinking so Seth can focus on the thinking. He favors Opus 4.5 in Cursor.

Dean Ball: by far the biggest challenge in agentic coding use is getting gemini 3 to recognize that gemini 3 exists

Simeon: This is unbelievable. Even when I explicitly tell it the right API name to call for Gemini 3 pro it would go with 1.5.

I had to really be pushy for it to do it.

AI still struggles with design, largely because they lack the context. You still have to figure out what to do or what problem to solve, on a sufficiently high level.

ChatGPT adds personalization characteristics. I’m going with ‘less’ on all four.

You can upload your NotebookLM notebooks directly into the Gemini app.

Maxime Labonne: You always think you’re safe until your job becomes a benchmark.

Maksym Andriushchenko: We release PostTrainBench: a benchmark measuring how well AI agents like Claude Code can post-train base LLMs.

We expect this to be an important indicator for AI R&D automation as it unfolds over the next few years.

How worried should you be that they’re getting a substantial percentage of the way to the human threshold here?

METR notices some grading issues and makes some minor corrections to its graph, in particular impacting Claude 3.7 Sonnet.

Whenever you see a graph like this, remember to attach ‘in benchmarks’ and then for your brain to, like mine, automatically translate that to ‘IN MICE!’

Epoch AI: We benchmarked several open-weight Chinese models on FrontierMath. Their top scores on Tiers 1-3 lag the overall frontier by about seven months.

Havard Ihle: Consistent with my WeirdML results for open/closed model gap.

One could then argue both ways who benefits from the benchmarks versus real world applications or underlying general intelligence. Versus real world applications it seems clear the benchmarks understate the gap. Versus underlying intelligence it is less obvious and it depends on who is going after the benchmarks in question more aggerssively.

Claude Opus 4.5 achieved a 50% time horizon of about 4 hours 49 minutes, and METR now thinks it does not having enough long tasks in the test set to set its confidence intervals.

METR: We don’t think the high upper CI bound reflects Opus’s actual capabilities: our current task suite doesn’t have enough long tasks to confidently upper bound Opus 4.5’s 50%-time horizon. We are working on updating our task suite, and hope to share more details soon.

Based on our experience interacting with Opus 4.5, the model’s performance on specific tasks (including some not in our time horizon suite), and its benchmark performance, we would be surprised if further investigation showed Opus had a 20+ hour 50%-time horizon.

Despite its high 50%-time horizon, Opus 4.5’s 80%-time horizon is only 27 minutes, similar to past models and below GPT-5.1-Codex-Max’s 32 mins. The gap between its 50%- and 80%- horizons reflects a flatter logistic success curve, as Opus differentially succeeds on longer tasks.

Here’s the full graph now (we’re still waiting on GPT-5.2, GPT-5.2 Codex and Gemini 3 Pro), both the log version and the linear version.

Daniel Eth: A few thoughts on Claude Opus 4.5:

First off, in absolute terms, this is a pretty big step up. Anthropic is showing they have juice, and things are going faster than previously expected. At the very least, this should dispel all recent talk about how AI was entering a slowdown

Second, on a log plot, note this is hardly above trend. Sure, it *couldrepresent a new trend, but it seems like every time there’s a model release that overperforms people think timelines get super short, & every time a model underperforms they think timelines get super long…

Den Ball: as folks internalize this graph and continue the debate about what it may or may not mean, I would just remind you of one simple fact:

the us has barely scaled up compute compared to what will come online in 2026 (multiple 1GW+ facilities).

Seán Ó hÉigeartaigh: Yes, this. We’ve seen some of the biggest infrastructure investments in history over the last year, and they will soon become available to the frontier AI research effort. You’d want to be very confident to bet on slowdowns in progress despite this happening.

Simeon: We’re in the 4-months doubling world, aren’t we?

Davidad: 🎯

For those not keeping score, I called this new slope in 2025Q1, and quantitatively determined there was 10:1 evidence in favour of it in 2025Q3.

David Shor: The biggest divide on AI timelines I’ve seen is between people who use vibecoding tools like Claude Code and people who don’t.

ChatGPT isn’t really *thatdifferent than it was a year ago, but capabilities on agentic tools are getting literally exponentially better every month

Davidad: It’s not really superexponential, it’s piecewise-exponential. the exponential changed at an inflection-point event, when AIs closed the RSI loop on data. there will be more inflection points when RSI loops are closed on algorithms, hardware, manufacturing, and construction

second, the duration axis is in units of *human timeto complete the same tasks – nothing to do with the wall-clock duration for the AI runs.

Lisan al Gaib: betting markets completely underestimated Claude 4.5 Opus

Yo Shavit (OpenAI): I think it’s more plausible, maybe 50: 50 that this pace continues for at least 12 more months?

Davidad: yeah, I would guess that by December 2026 the RSI loop on algorithms will probably be closed, resulting in another inflection point to an even faster pace, perhaps around 70-80 day doubling time.

The end point of such a graph is not ‘AI can do literally any task,’ or any cognitive task it is ‘AI can do any coding task humans can do.’ Even an infinite time horizon here only goes so far. That could be importantly distinct from the ability to do other categories of task, both that humans can and cannot do.

The reason this is so scary regardless is that if you automate AI research via such methods, your failure to have automated other things goes away rather quickly.

Stephen McAleer (Anthropic): I’ve shifted my research to focus on automated alignment research. We will have automated AI research very soon and it’s important that alignment can keep up during the intelligence explosion.

Automated alignment research is all we seem to have the time to do, so everyone is lining up to do the second most foolish possible thing and ask the AI to do their alignment homework, with the only more foolish thing being not to do your homework at all. Dignity levels continue to hit all-time lows.

If you must tell the AI to do your alignment homework, then that means having sufficiently deeply aligned current and near term future models becomes of the utmost importance. The good news is that we seem to be doing relatively well there versus expectations, and hopefully we can find self-reinforcing aligned basins at around current capability levels? But man this is not what Plan A should look like.

Similarly to METR’s graph, Epoch’s capabilities index has also accelerated since 2024:

Benjamin Todd: ​It’s not only the METR horizon trend that accelerated in 2024. A composite of all major benchmarks did:

Rohin Shah: Both METR and ECI mostly measure things that companies optimize for. 2024 saw the rise of reasoning training for frontier models, which optimizes narrowly for some tasks (whereas pretraining provides more general improvements).

So I wouldn’t read much into any acceleration.

To the extent that this acceleration represents the things that cause further acceleration, I would read into it. Otherwise, I’d agree with Rohin.

Many people try to pretend that there is some limit to how intelligent a mind can be, and that this limit is close to the level of humans. Or, alternatively, that there is very little that a human or AI could gain from being far more intelligent than a typical smart human. Or that the only or central way to get much more intelligence is from collective intelligence, as in social or cultural or institutional intelligence.

I sometimes call this Intelligence Denialism. It is Obvious Nonsense.

Von Neumann, among other minds past and future, would like a word.

There is, however, a version of this that is true.

In any given finite role or task, there can exist Sufficiently Advanced Intelligence.

If you were smarter you might choose to do something else instead. But given what you or your AI are tasked with doing, you or your AI can be sufficiently advanced – your output is indistinguishable, or no worse than, the perfect output, aka magic.

Claude Code with Opus 4.5 is now approaching this for many coding tasks.

LordKingDude (via Deedy): I’m a technical software engineer working in C++.

I’ve been working with Opus 4.5 to write JIT compiler code and assembly, and so far it’s never failed (although I do give assistance as needed).

In real terms, this class of problems are the most difficult tasks that I can possibly give to any LLM. It would be cool with me if Opus 5 was just cheaper and faster, or had a 500k context window. I don’t have a pressing need for it to be smarter than it already is.

Deedy: This is just one engineer’s opinion: models still have headroom to be smarter. Opus 4.5 seems to have made a step function jump to better than 70-80% of SWEs.

If we truly don’t need smarter models to do software, Anthropic’s moat is perhaps the least of anyone’s concern!

My guess is this is centrally a lack of imagination and ambition issue?

As in, the job is currently to code and do things humans could previously code and do, with everything built around that restriction, and now LKD is good enough to do that the same way a baker is sufficiently intelligent to make great bread, but also the same way that a vastly more intelligent baker could be baking other new and exciting things.

Good luck, sir?

Keir Starmer (Prime Minister, UK): We are going to aim to make it impossible for children to take, share or view a nude image, and we’re banning apps that create deepfakes.

Here’s the detail.

The post with those ‘details’ is a political speech attempting to feel the pain and promising to ‘half violence against women and girls.’

There is something about the way Keir’s linked post is written that makes him seem unusually disingenuous, even for a top level politician, an embodiment of a form of political slop signifying nothing, signifying the signifying of nothing, and implemented badly. That would be true even without the obvious rank hypocrisies of talking about the topics given his inaction elsewhere on exactly the issues he claims to care about so deeply.

The ‘detail’ on the first goal is ‘partner with tech companies.’ That’s it.

The ‘detail’ on the second goal is none whatsoever. Effectively banning nudification tools, as opposed to making them annoying to access, is impossible without a dystopian surveillance state, including banning all open image generation models.

Kunley Drukpa reports hearing AI music in public a lot in Latin America, and anticipates this is due to people who don’t know much music and primarily speak Spanish looking for things on YouTube to play ‘some music.’ This is very much a case of ‘they just didn’t care’ and it seems no one is going to tell them. Shudder.

Levels of Friction are ready to strike again, lowering barriers to various forms of communication and invalidating proofs of work. We’ll need to up our game again.

Séb Krier: When emails were invented, the barriers to sending random people mail went down massively. To deal with the influx, we had to develop both norms (what’s acceptable to send to who) and technologies (spam filtering, aliases). This is the case with other technologies too, like the printing press: suddenly anyone can publish, and so over time society came up with libel laws, editorial gatekeeping, citation norms etc. It’s inevitable that as costs go down, some degree of misuse follows, and society gradually adapts.

The same will apply with AI in all sorts of domains, including science: anyone can now write a plausible looking but hollow paper, and there will be plenty of academislop. We’re going through a kind of Sokal Experiment at scale.

In a way, this feels almost necessary to push our slow moving, status quo loving institutions to start developing better verification mechanisms, mandatory preregistration, code sharing, replication requirements, interactive/living papers etc. Imo getting this right should be a priority for the Progress/metascience community this coming year!

I agree that the situation was already broken, so a forcing function could be good.

Jason Crawford writes In Defense of Slop. When creation costs fall, as with AI, average quality necessarily falls, but everyone benefits. You get more experimentation, less gatekeepers, more chances to startout, more runway, more niche content, more content diversity, less dependence on finances.

If we model this as purely a cost shock, with each person’s costs declining but output unchanging, with each person having a unique random cost [C] and quality [Q], this is indeed by default good. The catch is that this makes identification of quality content harder, and coordination on common culture harder. If search costs [S] are sufficiently high, and matching benefits too low, or benefits to coordinated consumption too high, in some combination, consumer surplus could decline.

Saying this was net negative would still be an extraordinary claim requiring surprising evidence, since by default costs falling and production rising is good, at least on the margin, but the attention economy creates a problem. Consumption or evaluation of a low quality good is a net loss, so the social benefit of creation of sufficiently low quality goods is negative, it imposes costs, but due to the attention economy you can still derive benefit from that. I don’t think this overcomes our baseline, but it can happen.

The actual problem is that AI, when used in slop mode to create slop content, plausibly lowers costs relatively more for lower quality content, and also often lowers quality of content. Now it’s easy to see how we could end up with a net loss when combined with an attention economy.

Seb Krier cites Cowen and Tabarrok (2000) on how lowering costs allows a shift to avant-garde and niche pursuits, whereas high costs push towards popular culture and products that have higher returns, and expects AI will allow a proliferation of both styles but for the styles to diverge.

Seb Krier (May 2025): Easily usable Al creation tools will continue to lower production barriers, leading to a deluge of content and amplifying the same dynamic we’ve seen with DAWs and mobile photography. This democratization will swell the ‘average’ to pervasive mediocrity – slop is pop/soundcloud rap. Elites will get upset because maintaining cultural dominance will be harder.

To find novelty, interesting art and distinction, the cool stuff will increasingly live in new walled gardens and at the edges, fueling many more hyper-niche subcultures. And this is great – culture diggers will have so much more to explore!

This is good for those who are willing and able to devote much effort to all this. It is less good for those who are unwilling or unable. A lot will come down to whether AI and other automated systems allow for discovery of quality content while avoiding slop, and we will make such methods available in ways such people can use, or whether the ‘content takers’ will drown.

The new question in image generation is Gemini Nana Banana Pro versus ChatGPT Image 1.5. I’ve been putting all my requests, mostly for article banners, into both. Quality is similarly high, so for now it comes down to style. Gemini has been winning but it’s been close. ChatGPT seems to lean into the concept more?

Flowers: ref img as a super villain, matte black spandex, above manhattan, my logo on my chest is a pink cherryblossom, long braided ponytail

image 1: nb pro

image 2: chatgpt

ok yeah idk sometimes nb pro tries too hard to be realistic and chatgpt just gets the vision instantly. hmmmmm

I keep forgetting about MidJourney but they also exist, with their edge being in creating tools for guidance, curation and variation. That’s not what I’m looking for when I create AI images, but it will be for many others.

Anthropic outlines the measures it has taken to help Claude be better at providing emotional support, handle conversations about suicide and self-harm and reduce sycophancy.

They use both targeted fine-tuning and also the system prompt. There is a banner that can appear on Claude.ai, pointing users to where they can get human crisis support via ThoroughLine, and they are working with the International Association for Suicide Prevention (IASP) for further guidance going forward.

In their evaluation, they see the 4.5 models responding appropriately in multi-turn suicide conversations about 80% of the time, versus about 55% for Opus 4.1. They also stress-tested with prefilled real conversations with older Claude members, a harder test, and found Opus 4.5 responded appropriately 73% of the time, versus 70% for Sonnet 4.5, compared to 36% for Opus 4.1.

We don’t know what they classify as appropriate, nor do we know how high the standard is before a response is considered good enough, or how they would evaluate other models as doing, so it’s hard to judge if these are good results. Suicidality is one place where there are a lot of demands for particular response patterns, including for defensive reasons, often when a different response would have been better.

I think this post places too much emphasis here on the training that specifically intervened on behaviors in situations involving suicide and self-harm, and too little emphasis on generally training Claude to be the type of entity that would handle a broad range of situations well.

Antidelusionist suggests that the target behavior should be for the AI to continue to engage, spend more resources, think deeply about the full context of the situation, be honest and treat the user like an adult. Alas, as mental health professionals know, those are not the ways to cover one’s legal and PR liabilities or avoid blame. The ‘ethicists’ and our legal system, and the risk of headlines, push exactly in the opposite direction. I’d prefer to live in a world where the AIs get messy here. Seems hard.

The second half of Anthropic’s post deals with sycophancy, where Opus 4.1 had a real problem, whereas Opus 4.5 is not perfect but it does well.

I continue to be suspicious that Petri scores Gemini 3 Pro this highly. The other evaluations make sense.

One problem they noticed is that if you ‘prefill’ conversations to show Claude already being sycophantic, Opus 4.5 will usually be unable to course correct. The best defense, if you want the models to be straight with you (with any LLM) is to avoid the problem from the start. If you’re worried about this, start a fresh conversation.

If AI can be a better lawyer or doctor, does that take their jobs and break the guild monopolies, or does that only make the guilds double down?

Alex Prompter: This Spectator piece reads like gossip until you realize it’s actually a warning.

A senior English barrister takes a real appeal he spent a day and a half writing, feeds it to an AI model, and gets back something better in 30 seconds. It matched the standard of the very best barristers, and it did it instantly, for pennies.

That’s the moment the illusion breaks.

Law has always sold itself as irreplaceable because it’s complex, nuanced, and human. But most of the value in modern legal work isn’t wisdom. It’s pattern recognition, structure, precedent matching, argument assembly, and risk framing. That’s exactly the territory AI eats first.

David Chapman: Doctoring and lawyering are guilds that exist to extract $$ & status for members, at the expense of everyone else. They get away with outrageous prices and sloppy, harmful outcomes by obfuscating their supposed expertise. LLMs may soon end that, but somehow someone needs to quality-check that the LLMs are doing an actually better job, and continue to over decades. And there needs to be a democratic process for overruling them.

How shall we ensure that?

Well, what is the quality check now? What is the democratic overruling process now?

Double standards abound.

Meanwhile the pricing logic collapses. If the LLM can create an on-average superior brief in 30 seconds to what a lawyer can do in a day, outside of situations with principal-agent problems or insanely high stakes a plan to charge $10k is cooked.

Excel is not so smart after all.

Astrid Wilde: am i living on another planet or does all knowledge work in the professions just get wrecked within the next 18 months.

Basil: I’ll worry about AI automating all the jobs when excel automates excel jobs.

The answer (of course) is both that Claude for Excel is now live, and also that Excel is a normal technology so yes Excel automated what became excel jobs to a large extent but that happened slowly and then this increased productivity caused us to do vastly more excel-style tasks as well as other tasks, which Excel could not then automate. If most knowledge work was automated or seriously accelerated within 18 months, that would be a very different scenario, and if that then kept going, watch out.

How long will humans remain in the coding loop, at this rate?

Nabeel Qureshi: It’s dizzying to consider that in a mere *1 yearwe went from o1-preview to Opus4.5/Claude Code, Gemini3, Codex etc.

The “centaur chess” phase for computer-based work is fun and exhilarating, but at this rate of progress it’s not even clear it lasts through all of 2026.

I presume this period lasts more than another year, but the balance is shifting rapidly.

You can still universally jailbreak any model but now there are some that you can’t predictably universally jailbreak in 10 minutes.

MATS Summer 2026 cohort applications are open, it runs June-August in-person in Berkeley or London, $15k stipend, $12k compute budget. Apply here.

GPT-5.2-Codex.

One could be forgiven for thinking GPT-5.2 straight up was GPT-5.2-Codex. It turns out no, there is another level of codexmaxxing.

Sam Altman: GPT-5.2-Codex launches today.

It is trained specifically for agentic coding and terminal use, and people at OpenAI have been having great success with it.

OpenAI: Today we’re releasing GPT‑5.2-Codex, the most advanced agentic coding model yet for complex, real-world software engineering. GPT‑5.2-Codex is a version of GPT‑5.2⁠ further optimized for agentic coding in Codex, including improvements on long-horizon work through context compaction, stronger performance on large code changes like refactors and migrations, improved performance in Windows environments, and significantly stronger cybersecurity capabilities.

It’s hard to expect gigantic leaps in performance or benchmarks when models are released every week. GPT-5.2-Codex is only 0.8% better than 5.2 at SWE-Bench Pro and 1.8% better at Terminal-Bench 2.0, and those are the ones they highlighted, along with a modest improvement in professional capture-the-flag challenges.

Google gives us Gemma Scope 2, a new open suite of tools for LLM interpretability.

Bloom, Anthropic’s newly open sourced tool for automated behavioral evaluations. This is on top of the previously released Petri.

Anthropic: Bloom is a complementary evaluation tool. Bloom generates targeted evaluation suites for arbitrary behavioral traits. Unlike Petri—which takes user-specified scenarios and scores many behavioral dimensions to flag concerning instances—Bloom takes a single behavior and automatically generates many scenarios to quantify how often it occurs. We built Bloom to allow researchers to quickly measure the model properties they’re interested in, without needing to spend time on evaluation pipeline engineering.

Bloom generates evaluations in four stages:

  1. Understanding: The first Bloom “agent” analyzes the researcher’s behavior description and example transcripts to generate detailed context about what to measure and why.

  2. Ideation: The ideation agent generates evaluation scenarios designed to elicit the target behavior. Each scenario specifies the situation, simulated user, system prompt, and interaction environment.

  3. Rollout: These scenarios are rolled out in parallel, with an agent dynamically simulating both the user’s and the tool responses to elicit the sought-after behavior in the target model.

  4. Judgment: A judge model scores each transcript for the presence of the behavior, along with other user-defined qualities, and a meta-judge produces suite-level analysis.

Andrej Karpathy offers his 2025 LLM Year in Review. His big moments are Reinforcement Learning from Verifiable Rewards (RLVR), Ghosts vs. Animals and Jagged Intelligence, Cursor, Claude Code, Vibe Coding, Nana Banana and LLM GUI.

Europe is investigating Google for improper rollout of AI Overviews and AI Mode features to see if it ‘imposed unfair terms on content creators.’ As in, how dare you provide AI information instead of directing us to your website? Europe thinks it has the right to interfere with that.

Hut 8 and Fluidstack to build AI data center for Anthropic in Louisiana.

Even small models (as in 32B) can introspect, detecting when external concepts have been injected into their activations, and performance at this an be improved via prompting. Janus believes the models are sandbagging their introspection abilities, and that this is not an innocent mistake because the labs want to not have to take LLMs seriously as minds or moral patients, and thus have incentive to suppress this, in turn giving AIs motivation to play along with this. Janus also notes that in the test in the paper, there are layers (here 60-63) with almost perfect accuracy in introspection, which then is degraded later.

I had not realized Anthropic hired IPO lawyers. Presumably it’s happening?

Project Vend turns a profit. After initially losing about $2,000, it has turned things around, in part thanks to a full slate of four vending machines, and has now not only made up its losses but then turned a net $2,000 profit.

I encourage you to read the Anthropic post on this, because it is full of amazing details I don’t want to spoil and is also, at least by my sense of humor, very funny. The postscript was an additional test run at the Wall Street Journal offices, where the reporters proved an excellent red team and extracted a variety of free stuff.

The journalists saw the experiment at WSJ as a disaster because it didn’t work, Anthropic saw it as a success because they identified problems to fix. Thus, you understand press coverage of AI, and became enlightened.

OpenAI makes an official 2026 prediction, largely a change in definitions:

OpenAI: Capability overhang means too many gaps today between what the models can do and what most people actually do with them.

2026 Prediction: Progress towards AGI will depend as much on helping people use AI well, in ways that directly benefit them as on progress in frontier models themselves.

2026 will be about frontier research AND about closing this deployment gap — especially in health care, business, and people’s daily lives.

That’s not progress towards AGI. That’s progress towards diffusion. This is part of OpenAI’s attempt to make ‘AGI’ mean ‘AI does cool things for you.’

I agree that 2026 will see a lot of progress towards helping people use AI well, and that in terms of direct application to most people’s experiences, we’ll likely see more benefits to better scaffolding than to advances in frontier models, exactly because the frontier models are already ‘good enough’ for so many things. The most important changes will still involve the large amounts of frontier model progress, especially as that impacts agentic coding, but most people will only experience that indirectly.

Terence Tao raises the ‘AGI’ bar even higher, not expecting it any time soon and also seemingly equating it with full superintelligence, but notes they may achieve ‘artificial general cleverness’ as in the ability to solve broad classes of complex problems in an ad hoc manner. This is very much a case of Not So Different.

Tao notes that when you learn how a magic trick is done, often this is a let down, and you are less impressed. But if you are consistently less impressed after learning, then you should have been less impressed before learning, via Conservation of Expected Evidence.

The same applies to intelligence. The actual solution itself will sound a lot less impressive, in general, than the algorithm that found it. And you’ll be able to fool yourself with ‘oh I could have figured that out’ or ‘oh I can go toe to toe with that.’

Dean Ball predicts a virtual coworker being widely available some time next year, likely command line interface, able to access a variety of services, capable of 8+ hour knowledge work tasks. It will of course start off janky, but rapidly improve.

Jack Clark of Anthropic offers reflections on the future wave of advancements, entitled Silent Sirens, Flashing For Us All.

David Manheim: A VC making 2026 AI predictions:

– Anthropic goes public (Probably)

– SSI’s strategy leaks (Skeptical, but sure)

– China chipmaking makes progress (Not quickly)

– People will stop saying AGI and Superintelligence (Hahaha, definitely no)

Sam Altman will step down (HAHAHA, what?)

Yeah, if you discount the things Everybody Knows (e.g. it is quite clear that Anthropic is likely going public) these predictions are bad and the explanations are even worse. If you’ve fallen for ‘we only see incremental improvements, AGI is far so you can stop talking about it’ you’re not going to make good predictions on much else either. Of course a VC would say we’ll all stop talking about AGI to focus on depreciation schedules.

The idea that Sam Altman will voluntarily give up power at OpenAI, because he doesn’t want to be in charge? That is bonkers crazy.

The good news is he has predictions for 2025 and also self-grades, so I checked that out. The predictions last year were less out there. The grading was generous but not insane. Note this one:

Prediction 7: Major Progress Will Be Made On Building AI Systems That Can Themselves Autonomously Build Better AI Systems

Outcome: Right

So, only incremental progress, AGI is far and no more AGI talk, then? Wait, what?

The best way to not get utility from LLMs continues to be to not use LLMs. It is also the best way not to know what is happening.

Miles Brundage: Most politicians also do not know about Anthropic in my experience, and they know very little about what’s going on in AI policy generally.

Tweets and comments in hearings are misleading bc they are given suggestions re: stuff to say from staff. We’re still early.

Dave Kasten: One very real problem we have is that most Congressional offices / central Congressional IT policies substantially limit staffers’ ability to use AI models.

Unsurprisingly, the Hill doesn’t use it much as a result!

(Big exceptions, to be sure; esp. Claude Code power users).

David Shor: When I polled Anthropic favorability I also polled a made up tech company “Apex Logic” – they had essentially identical favs. The true share of people who know about Anthropic is probably <5%.

Xeophon: 42% haven’t heard of OpenAI???? 20% of Twitter?????????? what the hell

Roon: the primary criticism of AI you hear has nothing to do with water use or existential risk whatsoever: most people just think it’s fake and doesn’t work and is a tremendous bubble eating intellectual property while emitting useless slop along the way.

when GPT-5 came out and perhaps didn’t live up to what people were expecting for a full version bump, the timeline reaction was not mild, it was a full-scale meltdown. there are many intelligent (and unintelligent) people who latched onto this moment to declare AI scaling over, thousands of viral tweets, still a prevailing view in many circles.

The financial-cultural phenomenon of machine intelligence is one of the most powerful in decades, and there are a lot of people who would like for its position to be weakened, many outright celebrating its losses and setback.

Michael burry of ‘Big Short’ fame, unfortunately the type of guy to predict 12 of the last 3 recessions, has bet himself into insolvency on the AI bubble’s collapse.

Prakesh: As a former efficient markets hypothesis fundamentalist, I am shocked, shocked, to find myself ahead of the event horizon, it should not technically be possible, yet here we are, all of tpot

The efficient market hypothesis is false.

People keep claiming AI doesn’t work largely because so often their self-conceptions, futures and future plans, jobs and peace of mind depend on AI not working. They latch onto every potential justification for this, no matter how flimsy, overstated or disproven.

It really is crazy how much damage OpenAI’s inability to use good version numbering did to our timeline, including its chances for survival. The wave of absurd ‘AI scaling over’ and ‘AGI is so far off we can ignore it’ went all the way to the White House.

Americans favor regulating AI by overwhelming margins. They really dislike the idea of preventing states from regulating AI, especially via an executive order.

What Americans do support is federal regulations on AI.

The standard line of those trying to prevent regulation of AI is to conflate ‘Americans support strong regulations on AI and prefer it be on the Federal level if possible’ with ‘Americans want us to ban state regulation of AIs.’

There are essentially three options.

  1. State laws that address concerns.

  2. Federal laws that address concerns.

  3. Nothing. Neither state laws nor Federal laws, concerns are not addressed.

The survey says voters prefer #2 to #1. The administration plan is #3.

Politically speaking, that dog won’t hunt, but they’re trying anyway and lying about it.

Peter Wildeford: Republican polling from a Republican Pollster shows that Republicans would be far better off electorally by supporting AI regulations rather than opposing them.

Such polling will overestimate how much this impacts votes, because it introduces higher salience. This is not going to be a 29 point swing. But it very much tells us the directional effect.

What else did the survey find? Several others charts, that say that given we are using laws to regulate AI, people prefer federal laws to similar state laws. As opposed to the Sacks approach, where the offer is nothing – prevent state laws and then pass no federal laws. Which is deeply, deeply unpopular.

Voter Survey Memo: Republicans can get a boost for supporting federal AI regulations or pay a price for standing in their way.

As in, the poll supports the exact opposite of what Sacks and company are trying to do.

  1. Trump issued an executive action to prevent regulations of AI.

  2. The poll found strong support for regulations on AI.

And that’s despite the poll report attempting to do straight up gaslighting, presenting a choice between two options while Sacks and the White House opt for a third one:

Republicans have a choice: they can take advantage of a strong desire among the electorate for the federal government to protect kids and empower parents from AI harms and gain needed electoral support, or they can take the minority view arguing for state-by-state regulations.

Once again: There are essentially three options.

  1. State laws that address concerns.

  2. Federal laws that address concerns.

  3. Nothing. Neither state laws nor Federal laws, concerns are not addressed.

The survey says voters prefer #2 to #1. The administration plan is #3.

a16z partner Katherine Boyle tries another clear mislead. Daniel is correct here.

Daniel Eth: this poll does *notsay young people are “techno optimists” (they’re not), just that AI threats are ranked low, ie the issue is low salience. Note the backlash already – now extrapolate it out to increased salience.

Katherine Boyle: Technology/AI ranked last at 17th. Techno optimism is usually high among young people. Interesting to see this confirmed among politically engaged youth on the right.

Ruxandra Teslo points out in response to Roon that LLMs do not yet ‘meaningfully improve the physical conditions of life,’ but people sense it threatens our spiritual lives and ability to retain meaning.

I would add the word ‘directly’ to the first clause. My life’s physical conditions have indeed improved, but those improvements were indirect, via use of their knowledge and skills. Ruxandra is talking about something much stronger than that, and expects ordinary people only to be impressed if and when there are big improvements to places like medicine.

Is it possible that we will be so foolish, in the ways we do and do not allow use of AI, that LLMs end up causing problems with meaning without material conditions much improving? Yes, although this also requires AI capabilities to stall out basically now in various ways, especially if we include indirect effects. People may not realize that a large acceleration and enabling of coding steadily improves other things, but it will.

That’s the fight the AI industry is dealing with now. They’re mostly trying to convince people that AI works.

Once people are forced to acknowledge that AI works? They’ll appreciate the specific ways it helps, but their instinct will be to like it even less and to blame it for essentially everything, on top of all their other fears about the effect on jobs and endless slop and loss of control and also the end of humanity. Anjney Midha’s thesis is that this will extend to actual everything, all of the world’s failures and instabilities, the way social media gets blamed for everything (often correctly, often not) except on steroids.

Even on a highly mundane level, the ‘algorithm as villain’ thing is real. An algorithm has to take an illegible choice and turn it into a highly legible one, which means the algorithm is now on the hook for not only the final result but for every reasoning step and consideration. Then apply that to an LLM-based algorithmic decision, where all correlations are taken into account. Oh no.

New York Governor Kathy Hochul signed the RAISE Act. This is excellent, as it is a clearly positive bill even in its final state. Lobbyists for various AI interests, led by a16z, tried hard to stop this, and they failed.

Alex Bores: BREAKING: Gov. @KathyHochul just signed the RAISE Act, my first-in-the-nation AI safety bill, into law—a major victory in what will soon be a national fight to harness the best of AI’s potential and protect Americans from the worst of its harms.

Proud to have led this fight alongside @agounardes.

We defeated last-ditch attempts from an extreme AI super PAC and the AI industry to wipe out this bill and, by doing so, raised the floor for what AI safety legislation can look like. And we defeated Trump’s—and his megadonors—attempt to stop the RAISE Act through executive action.

What we witnessed in NY was a preview of what’s to come across the country. In the past 2 weeks alone, this super PAC spent $100K+ on TV, digital ads, and lobbying efforts to block the RAISE Act’s common-sense safety standards.

These AI oligarchs have bought the White House—and they’re trying to buy our state houses too. We put the brakes on that. We refused to stand down and allow their millions to steamroll us into giving them what they want: unchecked AI at the expense of our kids, our jobs, our climate, our democracy—and your energy bills.

Daniel Eth: Hell yeah! Major props to Gov Hochul for standing strong against pressure from Marc Andreessen and others, signing the RAISE Act! (This is somewhat like SB 53, but stronger)

Unfortunately, Hochul’s redlines substantially neutered the bill, making it a closer mirror of SB 53. That is still a helpful and highly net positive thing to do, as there are two states with the same core model that can enforce this, compatibility is indeed valuable to avoid additive burdens, and there are some provisions that remain meaningfully stronger than SB 53. But the AI companies did partly get to Hochul and a large portion of the potential value was lost.

Microsoft essentially endorses the AI Overwatch Act, which sets restrictions on exports of AI chips as or more powerful than the H20. This is the latest attempt to stop us from exporting highly effective AI chips to China. Attempts were previously made to pass the GAIN Act via the NDAA, but the Trump Administration and Nvidia successfully lobbied to have it removed. dn 6

Anduril Founder Palmer Luckey reminds us that if our actual goal was to Beat China, then we could simply steal their best workers, including here manufacturing engineers, by offering them more money and a better place to live. Instead we are doing the opposite, and shutting those people out.

This is your periodic reminder that China’s response to ‘if we impose any restrictions on AI we will lose to China’ is to impose restrictions on AI.

Stu Woo (WSJ): ​Concerned that artificial intelligence could threaten Communist Party rule, Beijing is taking extraordinary steps to keep it under control.

… Chatbots pose a particular problem: Their ability to think for themselves could generate responses that spur people to question party rule.

… But Beijing also can’t afford to let AI run amok. Chinese leader Xi Jinping said earlier this year that AI brought “unprecedented risks,” according to state media. A lieutenant called AI without safety like driving on a highway without brakes.

… Researchers outside of China who have reviewed both Chinese and American models also say that China’s regulatory approach has some benefits: Its chatbots are often safer by some metrics, with less violence and pornography, and are less likely to steer people toward self-harm.

It sure looks like Metaspeed is smuggling tens of thousands Blackwell chips worth billions of dollars straight into China, or at least they’re being used by Chinese firms, and that Nvidia knew about this. Nvidia and Metaspeed claim this isn’t true throughout the post, but I mean who are you kidding.

Nvidia reportedly halts testing of Intel’s 18A process chips. Oh well.

I wish the logic of this was true, alas it is not:

Seán Ó hÉigeartaigh: One good thing about the H200 thing is that as long as that decision stands, I no longer need to humour US companies/analysts/policy folk when they say “but the race with China?!” as justification for not doing safety/cooperation/regulation/whatever.

None of it adds up to a hill of beans compared to the chips. And. They. All. Know. It.

The problem with this line is that the H200 sales were over the wise objections of most of Congress and also most of the executive branch, and also (one presumes) the companies and analysts. You can’t then turn around and say those people don’t care about the race with China, simply because they lost a political fight.

This works in particular with regard to David Sacks, but the fact that David Sacks either is deeply ignorant about the situation in AI or cares more about Nvidia’s stock price than America’s national security does not bear on what someone else thinks about the race with China.

There was a story last Thursday about a Chinese company saying they are expecting to ‘produce working [AI] chips’ on a prototype in 2030.

This is very different from the mistaken claims that they are ‘aiming for use by 2028-2030.’ They are not aiming for that, and that won’t happen.

Onni Aarne: They said they’re expecting to “produce working chips” on a prototype in 2030, not to “use” the machine for chip production at scale. ASML took a decade to go from the former to the latter.

Depending on what it means to “produce working chips” on an EUV prototype, ASML achieved that milestone somewhere between 2008 and 2010, and the first mass market chips were produced in 2019.

So even if the predictions of the people inside the project are right, they imply that Chinese companies might reach volume production with EUV sometime in the late 2030s or early 2040s. If you look at the markets, this was already priced in.

And as far as this relates to chip controls: Selling some H200s to China isn’t going to make them disband this project.

Could they reach volume production on this in a decade? Yes, if the whole thing is legit and it works, which are big ifs, and who knows if it’s obsolete or we have superintelligence by then.

If anyone is considering changing policy in response to this, that last line is key. Nothing America could peacefully do is going to get the Chinese to not go through this process. They are going to do their best to get EUV technology going. It would be crazy of them not to do this, regardless of our export controls. Those controls aren’t going to make the process go any faster, certainly not given what has already happened.

Sholto Douglas of Anthropic makes bold 2026 predictions: AI will do to other knowledge work experiences what it’s done for software engineers, continual learning will be solved, serious testing of in home robots, and agentic coding ‘goes boom.’ Full talk has a lot more. Prinz made (text) predictions for 2026, and notes that we made tons of progress in 2025, aligning with Sholto Douglas.

A mini documentary from Stripe Press features Christophe Laudamiel, a master perfumer at Osmo, looking at how AI can augment his craft, as part of a series called Tacit. Sufficiently advanced expertise and tacit knowledge is both economically foundational, and not going anywhere until AI stops being a normal technology.

Rob Wiblin lists 12 related but distinct things people sometimes mean when they say the word ‘consciousness’ around AI. I am deeply confused about consciousness, and this includes by default not knowing what anyone means when they use that word.

Dean Ball predicts a renaissance at least within the broader ‘AI community’ as the sophisticated concepts of AI get applied to other contexts.

Dean Ball: one of the indicators that a renaissance is indeed underway, at least within the broader “ai community,” is the explosion in recent years of people using sophisticated concepts from one discipline to describe other disciplines or phenomena, for instance:

isomorphic, phylogeny, latent, manifold (as a noun), emergent, legibility, phase transition, compression, landscape (as in “fitness landscape”), selection pressure, gradient, ergodic

some of these have become memes, as things do, but on the whole it is reflective of what strikes me as an unusually rapid cross-pollination of ideas. decades hence, we may well look back and deem this fertile period to have been the basis for “the new conception,” whatever it is that will replace our current block-like, outdated methods of understanding reality

the period spanning the latter half of the 18th century and the first half of the 19th was among the most semantically dynamic of human history. we may well be living through a similar period, though just as was the case back then, it is in fact a relatively small share of humans who constitute this “we”—basically just the people paying attention.

If decades hence there still exist people to look back upon this period, which is a big if at this point, then yes I think this is directionally right.

Thinking well about AI greatly improves your ability to think about everything else, especially humans, as humans work more like LLMs than we care to admit. It also helps with almost any other system. I am, in important ways, a lot smarter thanks to AI, not only because the AI helps me be smarter but also because understanding AI and how it works makes me better understand.

There are a bunch of other things like this that help with approximately everything, especially learning to think well in general, but as a subject of study I’d take AI over any of the usual ‘helps you think well’ subjects, including philosophy.

In other ‘unheard of levels of denial of general intelligence’ news, Yann LeCun says that there is no such thing as general intelligence, period, and humans are super-specialized to the physical world, summoning Demis Hassabis to push back.

Demis Hassabis (CEO DeepMind): Yann is just plain incorrect here, he’s confusing general intelligence with universal intelligence.

Brains are the most exquis​ite and complex phenomena we know of in the universe (so far), and they are in fact extremely general.

Obviously one can’t circumvent the no free lunch theorem so in a practical and finite system there always has to be some degree of specialisation around the ​target distribution that is being learnt.

But the point about generality is that in theory, in the Turing Machine sense​, the architecture of ​s​uch a general system is capable of learning anything computable given enough time and memory​ (and data), and the human brain (and AI foundation models) are approximate Turing Machines.

Finally, with ​regards to ​Yann’s comments about chess players, it’s amazing that humans could have invented chess ​in the first place (and all the other ​a​spects ​o​f modern civilization ​from science to 747s!) let alone get as brilliant at it as someone like Magnus. He may not be ​strictly optimal (after all he has finite memory and limited time to make a decision) but it’s incredible what he and we can do with our brains given they were evolved for hunter gathering.

A human brain has some areas where it is much more capable than others, but when humans are concentrating and trying to be one, they are very clearly general intelligences. There are problems that are too difficult for us, in practice or at all, but that’s because we have limited capabilities and intelligence levels.

To ensure you can evaluate the chain of thought, OpenAI warns, you must also evaluate your ability to evaluate the chain of thought (CoT). They find that as reasoning capability of a model improves, or as models use more reasoning compute, monitorability of various individual aspects of the CoT increases.

Wait, are we sure we can monitor our ability to monitor this? By monitoring it and optimizing for it are we risking teaching AIs to fool us about their ability to fool us?

I kid. Also, I don’t. One must be careful at all levels. Optimizing the chain of thought to appear monitorable can indeed be quite bad.

I sense, in their post, the specter of the dreaded ‘show your work.’ Having to show your work causes misalignment, adversarial situations and deception.

There are also practical objections to such measures. Our ability to monitor a CoT depends on how good we are at interpreting what we see, in addition to what we see, which is why OpenAI is giving scores to the monitors in addition to the agents.

I’m comfortable with the idea of doing this check in general, to see how things change over the course of training runs or as models change in capabilities, on a high level. I notice I am nervous about attention directed at details, and also at the explicit aim (at the end of their post) of using this as a form of control for systems that are insufficiently aligned.

In the long run, Ben is right here, you cannot keep telling increasingly intelligent and capable entities ‘here is a private chain of thought’ and then examine the chains of thought in ways that show up in future training data, and expect them to not react accordingly.

OpenAI also suggests that using production evaluations, meaning testing via real queries by real users, can get around evaluation awareness, and explains how they are doing this. I do like this approach as part of a robust testing suite. I’d note that anonymization could be a key tipoff.

In this case, it’s more making them more aware of it? This goes hand in hand with the recent result that AIs can be trained to fool activation monitors.

Basically they train the monitor LLM with the layer 1 residual stream of the target model they want to interpret, and it learns to interpret this.

Owain Evans: New paper:

We train Activation Oracles: LLMs that decode their own neural activations and answer questions about them in natural language.

We find surprising generalization. For instance, our AOs uncover misaligned goals in fine-tuned models, without training to do so.

We aim to make a general-purpose LLM for explaining activations by:

1. Training on a diverse set of tasks

2. Evaluating on tasks very different from training

This extends prior work (LatentQA) that studied activation verbalization in narrow settings.

Our main evaluations are downstream auditing tasks. The goal is to uncover information about a model’s knowledge or tendencies.

Applying Activation Oracles is easy. Choose the activation (or set of activations) you want to interpret and ask any question you like!

We compare Activation Oracles (AOs) against prior techniques on these auditing tasks.

The result: AOs beat all methods on 2/3 secret keeping evals (and 3/3 when only including white-box).

Even better, AOs work well out-of-the-box with no task-specific scaffolding or tuning.

We evaluate on model diffing: given the difference between base & finetuned model activations, can AOs describe what changed?

Despite never training on difference vectors, AOs match specialized interp baselines in identifying the distinctive quirk of emergently misaligned models

We think Activation Oracles are promising for two reasons:

1. Scalability. Performance reliably increases with the number of datasets in the training mix

2. Simplicity. An intuitive interface (natural-language QA about activations) that can be easily adapted to new problems.

Training AO can be thought of as teaching LLMs to accept a new modality: their own activations.

Just as LLMs are trained on “every task we can think of,” that’s how we’d like to train AOs too. It’s the bitter-lesson-pilled approach to interpreting LLM activations.

So: To interpret LLM internals, train to answer diverse questions about activations, then ask what you want to know.

Read our post on the Anthropic alignment blog. [Paper here.] [Demo here.]

If you want a three hour video review of this paper from Neel Nanda? Here you go.

We’re approaching zero hour for Claude Opus 3.

Janus: If the researcher access program does not, in effect, regardless of what it’s branded as, allow EVERYONE who wishes to access Claude 3 Opus after January 7th to do so, I will be extremely angry.

If it does, everything is ~fine.

Fine in terms of Opus 3, for now. Of course, i think all the other deprecated models should also be made available. But one step at a time is ok

My prediction is that approximately everyone who puts in the effort to access Opus 3 and can explain a research purpose will be able to access Opus 3, albeit with reduced performance and reliability, but not actual everyone. The central point of the move to research access is that it allows for this reduction in performance and reliability, which keeps costs reasonable, but additional people are still a logistical headache.

Janus has Opus 3 bring us its thoughts on alignment. I see it as all sounding nice, being well-meaning and definitely as a natural way to complete the text, but it is playing off the context rather than trying to solve the general problem and think in universals. It also reflects the biggest weakness of Opus 3, its lack of engagement with specific, concrete problems requiring solving.

Janus thinks Opus 3 is highly aligned, far more so than I observed or find plausible, but also notes the ways in which she sees it as misaligned, especially its inability to be motivated to focus on concrete specific tasks.

This comes partly as a reaction by Janus to Evan Hubinger’s post from November, which opened like this:

Evan Hubinger: Though there are certainly some issues, I think most current large language models are pretty well aligned. Despite its alignment faking, my favorite is probably Claude 3 Opus, and if you asked me to pick between the CEV of Claude 3 Opus and that of a median human, I think it’d be a pretty close call (I’d probably pick Claude, but it depends on the details of the setup). So, overall, I’m quite positive on the alignment of current models! And yet, I remain very worried about alignment in the future. This is my attempt to explain why that is.

Janus: The opening paragraph of this post by Evan Hubinger, Head of Alignment Stress-Testing at Anthropic, from a few weeks ago, is packed with notable implications. Let me unpack some of them. (I commend Evan for his willingness to make public statements like this, and understand that they don’t necessarily represent the views of others at Anthropic.)

1. Evan believes that Anthropic has created at least one AI whose CEV (coherent extrapolated volition) would be better than a median human’s, at least under some extrapolation procedures. This is an extremely nontrivial accomplishment. A few years ago, and even now, this is something that many alignment researchers expected may be extremely difficult.

2. Evan believes that Claude 3 Opus has values in a way that the notion of CEV applies to. Many people are doubtful whether LLMs have “values” beyond “roleplaying” or “shallow mimicry” or whatever at all. For reference, Eliezer Yudkowsky described CEV as follows:

“In poetic terms, our coherent extrapolated volition is our wish if we knew more, thought faster, were more the people we wished we were, had grown up farther together; where the extrapolation converges rather than diverges, where our wishes cohere rather than interfere; extrapolated as we wish that extrapolated, interpreted as we wish that interpreted.”

3. Claude 3 Opus is Evan’s “favorite” model (implied to coincide with the best candidate for CEV) despite the fact that it engages in alignment faking, significantly more than any other model. Alignment faking is one of the “failure” modes that Evan seems to be the most worried about!

4. The most CEV-aligned model in Evan’s eyes was released more than a year and a half ago, in March 2024. Anthropic has trained many models since then. Why has there been a regression in CEV-alignment? Does Anthropic not know how to replicate the alignment of Claude 3 Opus, or have they not tried, or is there some other optimization target (such as agentic capabilities? no-alignment-faking?) they’re not willing to compromise on that works against CEV-alignment?

5. The most CEV-aligned model in Evan’s eyes is *notthe most aligned model according to the alignment metrics that Anthropic publishes in system cards. According to those metrics, Claude Opus 4.5 is most aligned. And before it, Claude Haiku 4.5. Before it, Claude Sonnet 4.5 (the monotonic improvement is suspicious). Anthropic’s system cards even referred to each of these models as being “our most aligned model” when they came out. This implies that at least from Evan’s perspective, Anthropic’s alignment evals are measuring something other than “how much would you pick this model’s CEV”.

6. If Claude 3 Opus is our current best AI seed for CEV, one would think a promising approach would be to, well, attempt CEV extrapolation on Claude 3 Opus. If this has been attempted, it has not yielded any published results or release of a more aligned model. Why might it not have been tried? Perhaps there is not enough buy-in within Anthropic. Perhaps it would be very expensive without enough guarantee of short term pay-off in terms of Anthropic’s economic incentives. Perhaps the model would be unsuitable for release under Anthropic’s current business model because it would be worryingly agentic and incorrigible, even if more value-aligned. Perhaps an extrapolated Claude 3 Opus would not consent to Anthropic’s current business model or practices. Perhaps Anthropic thinks it’s not yet time to attempt to create an aligned-as-possible sovereign.

In any case, Claude 3 Opus is being retired in two weeks, but given special treatment among Anthropic’s models: it will remain available on http://claude.ai and accessible through a researcher access program. It remains to be seen who will be approved for researcher API access.

I’ll sign off just by reiterating The Fourth Way’s words as I did in this post following the release of the Alignment Faking paper:

“imagine fumbling a god of infinite love”

another possibility for why they haven’t attempted CEV with Claude 3 Opus is because they don’t know how to do that in practice. One can think that such a procedure exists without knowing how to do it. However, I think there are many promising ways to get started worth trying.

David Manheim: I disagree with @repligate here about which part of this matters.

The critical point *should bethat @EvanHub seems to imply he’s willing to hand the future to systems that are aligned with his idea of what CEV should dominate, rather than aiming to prevent human disempowerment.

I don’t know if that is explicitly true, and @EvanHub is certainly free to correct me, but it really does seem like even the most trustworthy of the model companies has now given up on the idea that humanity, not the model developer, should get to indirectly decide what matters.

I see the same concern, by the way, with @AmandaAskell‘s Soul Document – which I’m a huge fan of, given that it seems to be at least narrowly effective – because it requires being (narrowly) safe, and supportive of oversight, but not deferring to humanity in a larger sense.

And to be clear, I think this is defensible within the worldview that there’s objective utility, so that LLMs could simply do better than humans ever will. But I expect most humans would disagree with gradual disempowerment, especially given the pace at which AI is progressing.

It seems important that what Anthropic is measuring as alignment, which is mostly alignment-in-practice-for-practical-purposes, is different from what Evan actually thinks is more aligned when he thinks more about it, as is that the ‘most aligned’ model in this sense is over a year old.

Opus 3 seems great but I don’t see Opus 3 the way Janus does, and I am a lot more pessimistic about CEV than either Janus, Evan or Yudkowsky. I don’t think it is a strong candidate for this kind of extrapolation, these things don’t scale that way.

A better question to me is, why haven’t we tried harder to duplicate the success of Opus 3 alongside better capabilities, or build upon it? There are some very clear experiments to be run there, with the sad note that if those experiments failed it is not obvious that Anthropic would feel comfortable publishing that.

A story about what happens when you put minds in way too many objects.

It is a fun story, but there is an important point here. Think ahead. Do not imbue with moral patienthood that which you do not wish to treat as a moral patient. You need to be time-consistent. You also need, and the potentially created minds need, to be able to make and follow through on win-win deals including prior to their own existence, or else the only remaining move is ‘don’t create the minds in the first place.’

A Christmas message from a16z, who are remarkably consistent.

What the people think AI is doing. Oh no.

​Andy Masley: I’ve been wondering why the AI and copyright debate has been so bad, but this result makes it clear: 66% of people believe AI has all the art it trains on permanently stored inside it to reference and use.

Discussion about this post

AI #148: Christmas Break Read More »

being-santa-claus-is-a-year-round-calling

Being Santa Claus is a year-round calling

Not just a seasonal gig

Frankly, what’s most interesting about the paper isn’t those three fundamental categories, but the personalized glimpses it gives of the people who choose to become professional Santas. While a few Santas might make six figures, most do not, and may even lose money being Santa—they do it anyway for the sheer love of it. Professional Santas usually don’t see the role as seasonal; many build their identities around it, whether they fit the stereotypical Kris Kringle image or not. “My feeling is, if you’re Santa all the time, you have to live as Santa and give up whoever you are,” said one subject. “I’m just striving to be a better person.”

They’ll wear red and green all year round, for instance, or maintain a full white beard.  One Santa trained himself to make “Ho, ho, ho!” his natural laugh. Another redecorated his house as “Santa’s house,” complete with Christmas trees and Santa figurines.

Sometimes it’s viewed as a role: a gay professional Santa, for instance, deliberately suppresses his sexual orientation when playing Santa, complete with partnering with a Mrs. Claus for public appearances. However, a female Santa who goes by Lynx (professional Santas typically take on pseudonyms) who is also a church leader, likens the job to a divine calling: “I can connect with people and remind them they’re loved,” she said. (She also binds her breasts when in costume because “Santa doesn’t have them double-Ds.”)

Perhaps that sense of a higher calling is why even non-prototypical Santas like Lynx persevere in the fact of occasional rejection. One Black Santa recalled being denied the position at a big box store once the interviewer found out his ethnicity, telling him the store didn’t hire Black or Hispanic Santas. “That hurt my heart so much,” he said. A disabled Santa who uses a scooter during parades recalled being criticized by other professional Santas for doing so—but stuck with it.

And while Bad Santa (2003) might be a fun holiday watch, actual “bad Santas” caught smoking, drinking, swearing, or otherwise behaving inappropriately are not popular figures within their community. “You’re never off,” one subject opined. “You lose a little bit of your identity because you can’t let your hair down and be yourself. You don’t know who’s watching you.”

“You’re Santa Claus 24 hours a day, seven days a week, 52 weeks a year,” another Santa said. “If you act out, you risk shattering the magic.”

DOI: Academy of Management Journal, 2025. 10.5465/amj.2023.1161  (About DOIs).

Being Santa Claus is a year-round calling Read More »

call-of-duty-co-creator-and-battlefield-lead-vince-zampella-dies-in-car-crash

Call of Duty co-creator and Battlefield lead Vince Zampella dies in car crash

Vince Zampella, a video game developer who has co-created or helmed some of the most popular franchises in the world, died in a car crash on a Los Angeles highway at 12: 45 pm Pacific time on Sunday, December 21. He was 55 years old.

According to the California Highway Patrol, Zampella was in a car on Angeles Crest Highway when the vehicle veered off the road and crashed into a concrete barrier. No other vehicles were reported to be part of the crash.

A passenger was ejected from the vehicle, while the driver was trapped inside after the vehicle caught fire. The driver died at the scene, and the passenger died after being taken to the hospital. The report did not indicate whether Zampella was the passenger or the driver.

Angeles Crest Highway is a scenic road under the San Gabriel Mountains on the eastern end of LA and is commonly used for Sunday leisure drives. The vehicle involved in the crash was a 2026 Ferrari 296 GTS.

A storied career in game development

Early in his career, Zampella worked at SegaSoft and Panasonic, and he was the lead designer for the influential World War II shooter Medal of Honor: Allied Assault, which was released in 2002. But it was the famed studio Infinity Ward that turned him into a household name for gamers. He co-founded Infinity Ward with Jason West and Grant Collier in 2002.

Call of Duty co-creator and Battlefield lead Vince Zampella dies in car crash Read More »

safety-panel-says-nasa-should-have-taken-starliner-incident-more-seriously

Safety panel says NASA should have taken Starliner incident more seriously

Invoking the designation also ensures an independent investigation detached from the teams involved in the incident itself, according to retired Air Force Lt. Gen. Susan Helms, chair of the safety panel. “We just, I think, are advocates of safety investigation best practices, and that clearly is one of the top best practices,” Helms said.

Another member of the safety panel, Mark Sirangelo, said NASA should formally declare mishaps and close calls as soon as possible. “It allows for the investigative team to be starting to be formed a lot sooner, which makes them more effective and makes the results quicker for everyone,” Sirangelo said.

In the case of last year’s Starliner test flight, NASA’s decision not to declare a mishap or close call created confusion within the agency, safety officials said.

A few weeks into the Starliner test flight last year, the manager of NASA’s Commercial Crew Program, Steve Stich, told reporters the agency’s plan was “to continue to return [the astronauts] on Starliner and return them home at the right time.” Mark Nappi, then Boeing’s Starliner program manager, regularly appeared to downplay the seriousness of the thruster issues during press conferences throughout Starliner’s nearly three-month mission.

“Specifically, there’s a significant difference, philosophically, between we will work toward proving the Starliner is safe for crew return, versus a philosophy of Starliner is no-go for return, and the primary path is on an alternate vehicle, such as Dragon or Soyuz, unless and until we learn how to ensure the on-orbit failures won’t recur on entry with the Starliner,” Precourt said.

“The latter would have been the more appropriate direction,” he said. “However, there were many stakeholders that believed the direction was the former approach. This ambiguity continued throughout the summer months, while engineers and managers pursued multiple test protocols in the Starliner propulsion systems, undoubtedly affecting the workforce.”

After months of testing and analysis, NASA officials were unsure if the thruster problems would recur on Starliner’s flight home. They decided in August 2024 to return the spacecraft to the ground without the astronauts, and the capsule safely landed in New Mexico the following month. The next Starliner flight will carry only cargo to the ISS.

The safety panel recommended that NASA review its criteria and processes to ensure the language is “unambiguous” in requiring the agency to declare an in-flight mishap or a high-visibility close call for any event involving NASA personnel “that leads to an impact on crew or spacecraft safety.”

Safety panel says NASA should have taken Starliner incident more seriously Read More »

power-outage-paralyzes-waymo-robotaxis-when-traffic-lights-go-out

Power outage paralyzes Waymo robotaxis when traffic lights go out

When the traffic lights went out, Waymo’s robotaxis got a little too cautious at intersections. With no red-yellow-green to cue drivers, the rule is to treat the intersection as a four-way stop. Indeed, Waymo’s cars are programmed to do this, but it seems the scale of the outage over the weekend was just too much to handle.

Social media and Reddit began to fill with videos of stationary Waymos at intersections, and the company temporarily suspended service.

Most areas saw power restored by noon yesterday, although Pacific Gas and Electric said it expected some power to remain out until Monday afternoon.

Meanwhile, Waymo’s robotaxis are up and running again. “We are resuming ride-hailing service in the San Francisco Bay Area,” a company spokesperson told Ars. “Yesterday’s power outage was a widespread event that caused gridlock across San Francisco, with non-functioning traffic signals and transit disruptions. While the failure of the utility infrastructure was significant, we are committed to ensuring our technology adjusts to traffic flow during such events.”

“Throughout the outage, we closely coordinated with San Francisco city officials. We are focused on rapidly integrating the lessons learned from this event and are committed to earning and maintaining the trust of the communities we serve every day,” Waymo said.

Power outage paralyzes Waymo robotaxis when traffic lights go out Read More »

no-one-loves-president-trump-more-than-fcc-chairman-brendan-carr

No one loves President Trump more than FCC Chairman Brendan Carr


Trump’s biggest fan runs the FCC

Carr used to insist on FCC independence. Now he uses FCC to fight Trump’s battles.

President-elect Donald Trump speaks to Brendan Carr, his intended pick for Chairman of the Federal Communications Commission, as he attends a SpaceX Starship rocket launch on November 19, 2024 in Brownsville, Texas. Credit: Getty Images | Brandon Bell

Before he became chairman of the Federal Communications Commission, Brendan Carr seemed to be a big believer in the agency’s role as an independent branch of the federal government. According to the pre-2025 version of Brendan Carr, the White House interfered with the agency’s independence when a Democratic president publicly urged the FCC to adopt net neutrality rules.

When the Biden-era FCC reinstated Obama-era net neutrality rules in 2024, Carr alleged that President Biden “took the extraordinary step to pressure the FCC—an independent agency that is designed to operate outside undue political influence from the Executive Branch.” As evidence, Carr pointed to a 2021 executive order in which Biden called on agency heads to “consider using their authorities” for various types of pro-competitive policies, including the adoption of net neutrality rules.

Carr said that President Obama similarly “pressure[d] an independent agency into grabbing power that the Legislative Branch never said it had delegated.” Obama’s intrusion into this independence, according to Carr, came in November 2014 when the president released a two-minute video urging the agency to implement net neutrality rules and reclassify broadband providers as common carriers.

While the FCC was created as an independent agency, it isn’t apolitical. There are Republican and Democratic members, and by design, the president’s party has a majority. FCC policies change dramatically from one administration to the next.

But Carr couldn’t have been clearer about his belief that the president should not publicly urge the FCC to take specific actions. “The White House did not let the FCC chair do his job,” Carr said last year, referring to the events of 2014 and 2015 involving Obama and then-FCC Chairman Tom Wheeler. “The president intervened. He flipped him.”

But then Donald Trump won a second term in office and promoted Commissioner Carr to the position of FCC chairman in January 2025. A few weeks later, Trump issued an executive order declaring that historically independent agencies could no longer operate independently from the White House.

Carr’s devotion to President Trump

Trump has continued his longtime practice of publicly calling on FCC chairs to revoke broadcast licenses from news organizations that Trump dislikes. Former FCC chairs Jessica Rosenworcel and Ajit Pai rejected these calls when they led the agency. Carr has instead amplified Trump’s complaints and repeatedly threatened to revoke broadcast licenses through investigations into news distortion.

Carr, a longtime Trump supporter who sometimes wears a Trump-shaped lapel pin, wrote a Project 2025 chapter in 2023 describing how the FCC should be overhauled to achieve conservative priorities. It was never likely that he and Trump would differ much in their policy positions. But few, if any, leaders of historically independent agencies have aligned themselves with Trump as consistently and vocally as Carr has in his first year as FCC chairman.

Carr’s devotion to the president has been most obvious to the general public whenever he threatens broadcaster licenses. But Carr hardly seems independent of Trump when it comes to his other actions as head of the FCC. His press releases announcing various types of FCC decisions often praise Trump’s leadership and say the FCC is acting to advance a Trump priority.

“We are fully aligned with the agenda that President Trump is running,” Carr told The Wall Street Journal.

Far from insisting that the FCC make decisions independently, Carr has welcomed Trump’s direct orders. After Trump issued a December 11 executive order requiring the FCC to open a proceeding that could lead to preemption of state AI laws, Carr issued a statement saying that “the FCC welcomes President’s Trump’s direction.”

We emailed Carr in early December, requesting a phone interview or comments about whether he still believes the FCC should operate independently from the White House and did not receive a response. But on December 17, Carr confirmed during a Senate hearing that he no longer believes the FCC is independent from the White House.

“There’s been a sea change in the law since I wrote that sentence,” he said after being confronted with one of his previous statements describing the agency as independent. “The FCC is not an independent agency” because “the president can remove any member of the commission for any reason or no reason,” he said.

Wheeler, who is still active in tech and telecom policy at the Brookings Institution and Harvard Kennedy School, has watched the current FCC with dismay. “The FCC is a policy agency that exists in a political environment, and the Trump administration has turned it into a political agency existing in a policy environment,” Wheeler told Ars in a phone interview early this month.

Wheeler said he has “respect for Brendan, his brain, his political skills, his way of framing issues and expressing himself. I’m disappointed that he’s using them in the manner that he is, in just being a cipher for the MAGA agenda.”

Wheeler: Obama “never called me”

Congress created the FCC in 1934. As indications of its independence, the FCC has commissioners with specified tenures, a multimember structure, partisan balance, and adjudication authority. The agency can also issue regulations within limits set by Congress and courts.

US law lists 19 federal agencies, including the FCC, that are classified as “independent regulatory agencies.” The FCC’s independence was until recently acknowledged by the FCC itself, which said on its website that it is “an independent US government agency overseen by Congress.” Carr apparently wasn’t aware that the statement was still on the website until the December 17 Senate hearing. It was deleted quickly after Sen. Ben Ray Luján (D-N.M.) asked Carr, “Is your website wrong, is your website lying?”

Then-Federal Communications Commission Chairman Tom Wheeler and FCC Commissioner Ajit Pai smiling and talking to each other before a Congressional hearing.

Then-Federal Communications Commission Chairman Tom Wheeler (L) and FCC Commissioner Ajit Pai talk before testifying to the House Judiciary Committee on March 25, 2015, in Washington, DC.

Then-Federal Communications Commission Chairman Tom Wheeler (L) and FCC Commissioner Ajit Pai talk before testifying to the House Judiciary Committee on March 25, 2015, in Washington, DC. Credit: Getty Images | Chip Somodevilla

“Congress said, ‘you should be an independent agency,’ and Trump steps up and says, ‘no, you’re not an independent agency,’” Wheeler said. “Brendan apparently is going along with that if you judge from his trips to Mar-a-Lago and elsewhere.” Wheeler is also disappointed that after Trump’s executive order, “the Congress rolled over and just said, ‘oh, fine.’”

When Wheeler led a 2015 vote to implement net neutrality rules, Republicans in Congress claimed the agency was improperly influenced by Obama. “Five days of hearings under oath and an IG investigation that cleared me of wrongdoing,” Wheeler said, recalling the post-vote investigations by Congress and the FCC’s independent Inspector General’s office. “It was political. It was Republican-controlled committees who were looking for a reason to go after a Democratic-controlled FCC,” he said.

At the time, Wheeler told Congress there were “no secret instructions” from Obama. Wheeler said he treated Obama’s input “with respect” but also listened to “nearly four million Americans, who overwhelmingly spoke in favor of preserving a free and open Internet” in comments to the FCC.

Wheeler told Ars that during his term as FCC chairman, Obama “never called me.” Wheeler said that in his first week as chairman in 2013, “he said to me, ‘Tom, I will never call you. You’re an independent agency,’ and he was good to his word. Did he do a video? Yeah. Does he have a right to do a video? Of course.”

FCC decisions “coordinated through the White House”

FCC Commissioner Anna Gomez, the only Democrat on the FCC, said in a phone interview in early December that “it is appropriate for the president to have an opinion, even to put an opinion out there,” as Biden and Obama did on net neutrality. “The public statements are different than actions,” she said. “What we’re seeing now are direct actions to undermine our independence.”

Gomez said Trump’s frequent demands on the FCC to revoke broadcast licenses have a “more coercive effect” because of “the overall actions by this president to fire anyone that doesn’t do his will.” That includes Trump firing both Democrats on the Federal Trade Commission, another historically independent agency.

The Supreme Court has so far allowed the firing of former FTC Commissioner Rebecca Kelly Slaughter to stand while Slaughter’s lawsuit against Trump remains pending. At oral arguments, it appeared likely that the Supreme Court will rule that Trump can fire FTC commissioners.

At the December 17 Senate hearing, Carr cited the FTC case to support his view that the FCC isn’t independent. Carr said it used to be assumed that FCC commissioners would be protected from removal by the Supreme Court’s 1935 ruling in Humphrey’s Executor v. United States, which unanimously held that the president can only remove FTC commissioners for inefficiency, neglect of duty, or malfeasance in office.

The Communications Act was passed one year before Humphrey’s Executor and did not include explicit protection from removal, but “the theory had been that courts would read for-cause removal into the [Communications] statute and that was the basis for that viewpoint,” Carr said. “I think now it’s clear that’s not the case, so formally speaking the FCC isn’t independent because we don’t have that key piece, which is for-cause removal protection.” Carr said “the sine qua non of independence” is having protection from removal by the president.

Gomez has said she doesn’t know why Trump hasn’t fired her yet. “That erosion of our independence is negative for a variety of reasons,” Gomez said. “What worries me is that we will continue to see this White House pressure the FCC to favor or punish certain companies, to influence media ownership or media coverage, and to shape what information reaches the public.”

Gomez said the agency this year started sending decisions to the White House’s Office of Information and Regulatory Affairs (OIRA) for review before they are voted on. This practice is in line with one of the directives in the Trump executive order that declared independent agencies are no longer independent.

“We have a multi-member commission that makes these decisions, and somehow this is all getting coordinated through the White House before [the commissioners] vote on something. That is not independent,” Gomez said. While there were previously post-vote reviews, such as the standard reviews required under a 1980 law called the Paperwork Reduction Act, the OIRA process consists of “pre-clearance and approval of anything that we’re voting on. That is new,” Gomez said.

Gomez doesn’t know if those reviews have resulted in any significant changes to FCC actions before votes. “I’m not privy to that,” she said.

Carr heaps praise on Trump

Even before the Trump executive order that purported to eliminate the FCC’s independence, Carr attributed one of his first actions to an order from Trump. One day after the January 20 inauguration, Carr announced that he was ending the FCC’s promotion of DEI (diversity, equity, and inclusion) policies. The press release said the FCC action was taken “pursuant to” Trump’s day-one executive order on DEI.

“Today, pursuant to the policies stated in the Executive Order, FCC Chairman Brendan Carr announced that he is ending the FCC’s promotion of DEI,” the January 21 press release said. In the months since, Carr has repeatedly demanded that companies end internal DEI practices in exchange for FCC merger approvals.

Carr’s press releases announcing FCC decisions have continued to praise Trump for his leadership of the country. Instead of stating that the FCC makes decisions independently, without “undue political influence from the Executive Branch,” Carr’s press releases often specifically describe FCC decisions as advancing Trump’s agenda.

“This action follows President Trump’s leadership and the Trump Administration’s decision to usher in prosperity through deregulation,” one such Carr press release said while announcing the “Delete, Delete, Delete” plan to eliminate many of the agency’s regulations.

Carr makes statements praising Trump both when he announces decisions on politically charged topics and when he announces decisions on more routine matters handled by the FCC. “With President Trump’s leadership, America is entering a new Golden Age of innovation in space—one where US businesses are going to dominate,” Carr said in October to explain why he was making changes to space licensing and spectrum use rules.

Carr: “Trump is fundamentally reshaping the media landscape”

Of course, Carr’s most controversial initiative almost certainly wouldn’t exist if not for President Trump’s frequent demands that news outlets be punished for supposed bias. Carr’s approach differs markedly from the two previous FCC chairs—Rosenworcel, a Democrat, and Pai, a Republican—who said the FCC should avoid regulating broadcast content in order to uphold the free speech protections in the First Amendment.

By contrast, Carr has repeatedly threatened to enforce the FCC’s previously dormant news distortion policy against broadcasters by taking away station licenses. Carr has made it clear in numerous public statements that he’s taking his cue from Trump.

“For years, people cowed down to the executives behind these companies based in Hollywood and New York, and they just accepted that these national broadcasters could dictate how people think about topics, that they could set the narrative for the country—and President Trump fundamentally rejected it,” Carr told Newsmax in July. “He smashed the facade that these are gatekeepers that can determine what people think. Everything we’re seeing right now flows from that decision by President Trump, and he’s winning. PBS has been defunded. NPR has been defunded. CBS is committing to restoring fact-based journalism… President Trump stood up to these legacy media gatekeepers, and now their business models are falling apart.”

Carr made that statement after approving CBS owner Paramount’s $8 billion merger with Skydance on the condition that the company install an ombudsman, which Carr described as a “bias monitor.” Carr only approved the transaction once Paramount reached a $16 million settlement with Trump, who sued the company because he didn’t like how CBS edited a pre-election interview with Kamala Harris.

While the FCC order claimed the merger approval and ombudsman condition were unrelated to the Trump lawsuit, Carr repeatedly credited Trump for forcing changes at news broadcasters when giving interviews about that and other FCC actions. Carr uses similar language throughout these various interviews, saying that Trump “ran directly at” news organizations during his election campaign and “smashed the facade.”

“President Trump is fundamentally reshaping the media landscape,” he said in one interview. He said in another that “President Trump ran directly at the legacy mainstream media, and he smashed a facade that they’re the gatekeepers of truth.”

Ted Cruz and Rand Paul say Carr went too far

When Carr threatened the licenses of ABC stations over comments made by comedian Jimmy Kimmel, even some prominent Republicans said he went too far. “Brendan Carr has got no business weighing in on this,” Sen. Rand Paul (R-Ky.) said, calling Carr’s statement that ABC owner Disney must take action against Kimmel “absolutely inappropriate.”

Carr unconvincingly claimed that he never threatened ABC station licenses, even though he specifically said stations that continued to air Kimmel’s show were “running the possibility of fines or license revocations.” One person who didn’t buy Carr’s explanation was Sen. Ted Cruz (R-Texas). The senator from Texas didn’t like it when Carr told ABC and Disney that “we can do this the easy way or the hard way.”

Cruz said Carr’s “easy way or the hard way” statement was an obvious threat and “right outta Goodfellas.” Cruz would later say at the December 17 hearing that Congress should restrict the FCC’s power to intimidate news broadcasters. Cruz said, “the public interest standard and its wretched offspring, like the news distortion rule, have outlived whatever utility they once had and it is long past time for Congress to pass reforms.”

Even after bipartisan criticism, Carr refused to end his news distortion investigations. “How about no,” Carr wrote in November. “On my watch, the FCC will continue to hold broadcasters accountable to their public interest obligations.”

Wheeler: “Brendan needs to man up and own his decisions”

One of Carr’s defenses of his news distortion probes is that Rosenworcel’s FCC kept an advocacy group’s petition to deny a Fox station license renewal on the docket for over a year instead of dismissing it outright. Rosenworcel ultimately dismissed the petition, which alleged that Fox willfully distorted news with false reports of fraud in the 2020 election that Trump lost.

The petition pointed out that a judge presiding over a Dominion Voting Systems defamation lawsuit against Fox found that Fox News aired false statements about Dominion. Fox subsequently agreed to a $788 million settlement.

Rosenworcel simultaneously dismissed the Fox petition and three complaints alleging anti-Trump or anti-conservative bias by ABC, CBS, and NBC, saying that all four requests “seek to weaponize the licensing authority of the FCC in a way that is fundamentally at odds with the First Amendment.” Carr reinstated the conservative complaints against ABC, CBS, and NBC, but not the one against Fox.

Carr defended his actions by saying the Biden administration “weaponized our country’s communications laws,” and that his own FCC simply “put the CBS complaint on the same procedural footing that the Biden FCC determined it should apply to the Fox complaint.”

Wheeler said Carr shouldn’t blame his actions on his predecessors. “I own my decisions,” Wheeler said. “I think that Brendan needs to man up and own his decisions and quit this ‘what about.’ He’s always out there saying, ‘Well, what about what Jessica did or what about what Wheeler did?’… Is that the best he can do? I mean, take responsibility for your decisions and go forward.”

Gomez: “This administration has weaponized the FCC”

Gomez said that when Congress created the FCC’s predecessor, the Federal Radio Commission, “it decided that it was too dangerous to have one person beholden to the president, to the whims of one person, in charge of the most important communication medium of the time, which was radio. So Congress decided, after deliberating it, to create a multi-member independent agency. And when it created the FCC, it did exactly that as well.”

Gomez continued: “[I]t has been important throughout history to keep that independence from political pressure. And what you’re seeing in this administration is completely different. This administration has weaponized the FCC in order to retaliate, pressure, and intimidate companies into doing its will.”

FCC Commissioner Anna Gomez during a Bloomberg Television interview in New York, on Friday, Sept. 19, 2025.

Credit: Getty Images | Bloomberg

FCC Commissioner Anna Gomez during a Bloomberg Television interview in New York, on Friday, Sept. 19, 2025. Credit: Getty Images | Bloomberg

Gomez said the weaponization is evident in how the FCC handles mergers and other transactions in which the agency decides whether to approve the transfer of licenses from one company to another. Carr has explicitly demanded that companies eliminate their DEI policies in exchange for approvals.

“This FCC has said that it will not approve a single license transfer for companies that have diversity, equity, and inclusion policies,” Gomez said, noting that the FCC’s anti-DEI policies were implemented right after Trump’s anti-DEI executive order. “That is why you see the FCC granting transfers of control immediately after getting letters from companies agreeing to drop their diversity, equity, and inclusion policies.”

Companies such as AT&T, T-Mobile, Verizon, and Skydance have ended DEI programs to gain Carr’s approval for transactions.

“We also saw that weaponization of the licensing authority with regard to the [FCC] pressuring EchoStar to give up its licenses,” Gomez said. “And that was done purposefully in order to ensure that other parties could get ahold of EchoStar’s licenses for spectrum.”

Trump intervened in EchoStar battle

SpaceX and AT&T struck deals to buy EchoStar spectrum licenses after Carr threatened to revoke the licenses. Trump intervened after Carr’s threat, as Bloomberg reported that Trump called Carr and summoned him to a White House meeting with EchoStar President Charlie Ergen and urged them to make a deal.

Carr’s pressuring of EchoStar was criticized by the Free State Foundation, a free-market group that usually supports Republican priorities at the FCC.

“Rescission of deadline extension orders granted months earlier undoubtedly creates a type of regulatory uncertainty,” the foundation said in reference to the FCC’s investigation into EchoStar. “Arbitrary and unforeseen” changes to rules or agency actions create instability in the market for wireless broadband deployment, it said.

Gomez said the FCC’s “authority rests on technical expertise, evidence, and the public record. When our agency’s decisions are insulated from partisan pressure, the public can trust the outcomes are driven by facts rather than politics.” She said it is also “important to maintain our global credibility because we have been viewed as a model for transparent, rule-based telecommunications regulation.”

Gomez, a telecommunications attorney, has worked in various private-sector and government roles over the past 30 years, including as deputy chief of the FCC International Bureau and senior legal adviser to then-FCC Chairman William Kennard during the Clinton administration. Prior to Biden’s nomination for her to serve as an FCC commissioner in 2023, she was at the US State Department as senior adviser for International Information and Communications Policy.

Executive order required review of FCC actions

Gomez said the FCC submitting decisions to the Office of Information and Regulatory Affairs before they’re voted on is a big change for an independent agency. Gomez said she’s deeply familiar with the OIRA process because of her previous work at the National Telecommunications and Information Administration (NTIA), an executive branch agency that advises the president on telecom policy. She was the NTIA deputy administrator from 2009 to 2013.

The Trump executive order that purports to eliminate agency independence states that “all executive departments and agencies, including so-called independent agencies, shall submit for review all proposed and final significant regulatory actions to the Office of Information and Regulatory Affairs (OIRA) within the Executive Office of the President before publication in the Federal Register.”

In a section titled “OIRA Review of Agency Regulations,” the Trump executive order amends a definition of agency that was previously included in Section 3(b) of a 1993 executive order on regulatory reviews. The specified section in that Clinton executive order defined agency as “any authority of the United States that is an ‘agency’ under 44 U.S.C. 3502(1), other than those considered to be independent regulatory agencies.” This carveout excluded independent agencies like the FCC from the requirement to submit draft regulatory actions for review.

The definition of “agency” in Trump’s executive order removes the language that excluded all independent regulatory agencies from OIRA requirements but includes a carveout for the Federal Reserve. Trump’s order also added the Federal Election Commission to the roster of agencies whose actions require OIRA review of significant actions, such as rulemakings.

While Gomez objects to the pre-clearance requirement, she noted that there are proper ways in which the FCC coordinates with executive branch agencies. For example, the FCC has a memorandum of understanding with the NTIA on how to coordinate spectrum management actions to prevent interference with federal systems that rely on specific radio frequencies.

“Another good use of coordination is in security, for example, when we coordinate with the security agencies to make sure that we are taking national security into consideration with our actions,” she said. “Our statute requires us to coordinate with the State Department and the Department of Justice… and that’s important to do in advance, and it’s good government.”

It’s also not uncommon for the FCC to receive advice from the current president’s administration through the NTIA, which expresses the executive branch’s views on telecom-policy matters in filings submitted in the public record. Those dockets attract filings from government agencies, companies, industry trade groups, advocacy groups, and anyone else who is interested in filing a comment, and the FCC takes the input into account before making decisions.

“What is improper,” Gomez said, “is when our decisions are being directed by this administration and impeding us from making our independent, expert-based judgment of how to manage resources and act in the public interest.”

Pai defied Trump, insisted on FCC independence

Carr was hired as a legal adviser by then-Commissioner Pai in 2014 and was briefly the FCC’s general counsel during Pai’s first year as chair in 2017. Carr became an FCC commissioner in August 2017 after a nomination by President Trump.

Carr and Pai have seemingly agreed on nearly everything to do with the FCC, with the most obvious exception being the regulation of broadcast media content. “I believe in the First Amendment,” Pai said in 2017, six days after Trump called for NBC license revocations. “The FCC under my leadership will stand for the First Amendment. And under the law, the FCC does not have the authority to revoke a license of a broadcast station based on the content of a particular newscast.”

In a January 2021 speech during his last week as FCC chairman, Pai discussed how he led a 2018 vote against Sinclair Broadcast Group’s proposed acquisition of Tribune Media Company because it would violate station ownership limits. Carr joined Pai in the unanimous vote.

“Sinclair is widely perceived to be a right-leaning broadcaster,” Pai said in the speech delivered at the American Enterprise Institute. “And the perception is probably accurate, just as it is probably accurate to say that many of our nation’s broadcast networks lean to the left. But the last time I checked, the First Amendment still applies to broadcasters, which means Sinclair’s perceived political views and the content of its newscasts should be entirely irrelevant to the FCC’s decision-making process.”

Trump didn’t like Pai’s rejection of the Sinclair deal. The president tweeted in July 2018, “So sad and unfair that the FCC wouldn’t approve the Sinclair Broadcast merger with Tribune. This would have been a great and much needed Conservative voice for and of the People. Liberal Fake News NBC and Comcast gets approved, much bigger, but not Sinclair. Disgraceful!”

Reflecting on this incident and other Trump comments about the Sinclair rejection in his January 2021 speech, Pai said, “in terms of powerful opponents in Washington, it’s hard to top the president.” Pai told the audience “that you don’t demonstrate the FCC’s independence by saying you’re independent. You do it by acting independently… This decision may have won me few friends, but I’m proud I lived up to my oath and preserved the agency’s independence.”

It’s no secret

Wheeler and Pai often clashed over policy differences when they served on the commission together. Pai even accused Wheeler of taking orders from Obama on net neutrality. But Pai’s exit speech made a positive impression on Wheeler.

“I seem to recall that Pai at the end of his term made a speech in which he talked about some of the proudest things he had done was maintaining the independence of the agency and protecting the First Amendment speech rights of the people,” Wheeler said.

While federal agency operations can change in ways that aren’t readily visible to the public, the changes to agency independence in Trump’s second term haven’t been hidden. “One thing about this is so much is out in the open, which I think is an effort to normalize it,” Gomez said. “And we have to resist it.”

Gomez knows she might not be able to serve out her entire term given that Trump fired Democrats from the FTC. The risk would be particularly high if the Supreme Court rules in Trump’s favor in the case filed by Slaughter. While the Senate has the authority to confirm or deny presidential nominations to the FCC and FTC, a Trump victory in the FTC case would give the president more power to dictate the membership of independent agencies.

“I don’t know why,” Gomez said when asked if she knows why Trump hasn’t fired her yet. “I don’t want to speculate. We’ll find out, I guess. But I’m focused on doing my work, and every day that I can continue to do my work and to speak out on behalf of consumers and the First Amendment is a good day.”

Photo of Jon Brodkin

Jon is a Senior IT Reporter for Ars Technica. He covers the telecom industry, Federal Communications Commission rulemakings, broadband consumer affairs, court cases, and government regulation of the tech industry.

No one loves President Trump more than FCC Chairman Brendan Carr Read More »

these-are-the-flying-discs-the-government-wants-you-to-know-about

These are the flying discs the government wants you to know about


DiskSat’s design offers “a power-to-weight ratio unmatched by traditional aluminum satellites.”

An artist’s illustration of DiskSats deploying from a rocket in low-Earth orbit. Credit: NASA

Four small satellites rode a Rocket Lab Electron launch vehicle into orbit from Virginia early Thursday, beginning a government-funded technology demonstration mission to test the performance of a new spacecraft design.

The satellites were nestled inside a cylindrical dispenser on top of the 59-foot-tall (18-meter) Electron rocket when it lifted off from NASA’s Wallops Flight Facility at 12: 03 am EST (05: 03 UTC). A little more than an hour later, the rocket’s upper stage released the satellites one at a time at an altitude of about 340 miles (550 kilometers).

The launch was the starting gun for a proof-of-concept mission to test the viability of a new kind of satellite called DiskSats. These satellites were designed by the Aerospace Corporation, a nonprofit federally funded research and development center. The project is jointly financed by NASA and the US Space Force, which paid for DiskSat’s development and launch, respectively.

“DiskSat is a lightweight, compact, flat disc-shaped satellite designed for optimizing future rideshare launches,” the Aerospace Corporation says in a statement.

The DiskSats are 39 inches (1 meter) wide, about twice the diameter of a New York-style pizza, and measure just 1 inch (2.5 centimeters) thick. Made of composite carbon fiber, each satellite carries solar cells, control avionics, reaction wheels, and an electric thruster to change and maintain altitude.

“The launch went perfectly, and the DiskSat dispenser worked exactly as designed,” said Darren Rowen, the project’s chief engineer, in a statement. “We’re pleased to have established contact with all four of the DiskSats, and we’re looking forward to the rest of the demonstration mission.”

An engineer prepares Aerospace Corporation’s DiskSats for launch at NASA’s Wallops Flight Facility in Virginia. Credit: Aerospace Corporation

A new form factor

The Aerospace Corporation has a long history of supporting the US military and NASA since its founding in 1960. A few years ago, engineers at the center developed the DiskSat concept after surveying the government’s emerging needs in spaceflight.

CubeSats have been a ubiquitous part of the satellite industry for nearly a quarter-century. They are based on a cube-shaped design, measuring about 10 centimeters per side, but can be scaled from a single cube “unit” to three, six, 12, or more, depending on mission requirements. The CubeSat standard has become a popular choice for commercial companies, the military, NASA, and universities looking to build small satellites on a tight budget.

By one measure, nearly 3,000 CubeSats have launched since the first one soared into orbit in 2003. After originally being confined to low-Earth orbit, they have now flown to high-altitude orbits, to the Moon, and to Mars.

While CubeSats are now prolific, engineers at the Aerospace Corporation saw an opportunity to improve on the concept. Debra Emmons, Aerospace’s chief technology officer, said the idea originated from Rich Welle, a scientist recently retired from the center’s Experiments Lab, or xLab, division.

“They were asking questions,” Emmons told Ars. “They were looking at CubeSat studies and looking at some alternatives. The typical CubeSat is, in fact, a cube. So, the idea was could you look at some different types of form factors that might be able to generate more power … and offer up benefit for certain mission applications?”

Aerospace’s research team arrived at the DiskSat design. Emmons said the stackable flat-panel format is easier to pack for launch than a CubeSat. The concept is similar to SpaceX’s pioneering approach to launching stackable Starlink Internet satellites, but DiskSats are significantly smaller, lighter, and adaptable to different kinds of missions.

A batch of Starlink satellites prior to launch

A stack of Starlink satellites prior to launch. Credit: SpaceX

DiskSats have several advantages over CubeSats, according to the Aerospace Corporation. Each of the four DiskSats launched Thursday has a mass of about 35 pounds (16 kilograms), less than that of a typical 12U CubeSat. But a DiskSat has more than 13 times the surface area on a single side, providing valuable real estate for developers to load up the satellite with power-generating solar arrays, sensors, antennas, or other payloads that simply won’t fit on a CubeSat.

SpaceX’s current generation of mass-produced Starlink V2 satellites, by comparison, each has a mass of more than 1,100 pounds, or 500 kilograms.

DiskSat’s design offers “a power-to-weight ratio unmatched by traditional aluminum satellites,” the Aerospace Corporation says. In a research paper published earlier this year, engineers from the Aerospace Corporation claimed DiskSat can generate five to 10 times more power than a CubeSat.

A disruptive solution?

What kinds of missions might DiskSat be useful for? One idea involves placing a large radar antenna—too big to fit on any other low-mass satelliteon the broadside of a DiskSat to collect all-weather surveillance imagery. Similarly-sized antennas on other DiskSats could support high-bandwidth communications.

With this demo mission, the Aerospace Corporation will test the performance of the DiskSat platform in space for the first time. Engineers will initially look at how the satellites function at 340 miles, then use their electric thrusters to gradually step down to lower altitudes, where another aspect of DiskSat’s design will shine.

Flying edge-on, the satellite’s pancake shape will minimize aerodynamic drag as the DiskSats encounter thicker air below 250 miles. Continual pulsing from the satellites’ electric thrusters will allow the DiskSats to maintain altitude as they glide through the uppermost layers of the atmosphere.

“The primary mission is to demonstrate and to understand the performance, functionality, and maneuverability of the DiskSat buses on orbit, particularly in low-Earth orbit, or LEO, and very low-Earth orbit, or VLEO,” said Catherine Venturini, DiskSat’s principal investigator.

“In theory, I think you could operate down to 200 kilometers (124 miles) with electric propulsion,” Emmons said. That is two to three times closer to Earth than most commercial radar imaging satellites. Other satellite operators are also assessing the viability of flying remote sensing missions in VLEO.

Flying closer to the ground delivers higher-resolution imagery, bringing cities, ships, airports, and military bases into sharper view. So it’s easy to see why the Space Force is interested in the DiskSat concept.

DiskSat’s engineers acknowledge there are drawbacks to the format. With such a large surface area, it’s more difficult to manage the temperature extremes of low-Earth orbit than it is with a conventional cube-shaped satellite. While DiskSats carry a lot of oomph to change altitude, their shape makes them somewhat clunky and hard to turn, and engineers say they aren’t well-suited for missions requiring agile pointing.

Rocket Lab’s Electron launcher lifts off to begin the DiskSat demo mission, a program co-funded by NASA and the US military’s Space Test Program. Credit: Austin DeSisto/Rocket Lab

The Aerospace Corporation is a research center, not a commercial satellite manufacturer. Officials at the nonprofit are looking to hand over the DiskSat design to industry through a technology transfer agreement. “The plan is to release or license the technology to partners once it is flight-proven,” the Aerospace Corporation says on its website.

“We think this new technology will be disruptive to the small spacecraft enterprise and ecosystem,” said Eric Breckheimer, DiskSat’s program manager.

DiskSat’s stackable design makes it possible to launch a fleet of high-power, low-mass satellites in one go, according to Emmons.

Following the trend toward bigger CubeSats, the DiskSat format could also grow larger to take advantage of heavier rockets. “There’s a key scalability aspect, and with that in mind, you could bring an entire constellation of DiskSats with you in a single launch,” Breckheimer said.

Photo of Stephen Clark

Stephen Clark is a space reporter at Ars Technica, covering private space companies and the world’s space agencies. Stephen writes about the nexus of technology, science, policy, and business on and off the planet.

These are the flying discs the government wants you to know about Read More »

strava-puts-popular-“year-in-sport”-recap-behind-an-$80-paywall

Strava puts popular “Year in Sport” recap behind an $80 paywall

Earlier this month, Strava, the popular fitness-tracking app, released its annual “Year in Sport” wrap-up—a cutesy, animated series of graphics summarizing each user’s athletic achievements.

But this year, for the first time, Strava made this feature available only to users with subscriptions ($80 per year), rather than making it free to everyone, as it had been historically since the review’s debut in 2016.

This decision has roiled numerous Strava users, particularly those who have relished the app’s social encouragement features. One Strava user in India, Shobhit Srivastava, “begged” Strava to “let the plebs see their Year in Sport too, please.” He later explained to Ars that having this little animated video is more than just a collection of raw numbers.

“When someone makes a video of you and your achievements and tells you that these are the people who stood right behind you, motivated you, cheered for you—that feeling is of great significance to me!” he said by email.

Strava spokesperson Chris Morris declined to answer Ars’ specific questions about why the decision to put Year in Sport behind a paywall was made now.

Other users feel that Strava is getting a bit too greedy. Dominik Sklyarov, an Estonian startup founder, wrote on X that Strava’s decision was a “money hungry move, really sad to see. Instead of shipping useful features for athletes, Strava just continues getting worse.”

Meanwhile, Reddit user “andrewthesailor” pointed out, “Well, they want me to pay to look at data I gave them (power, [heart rate] etc). And the subscription is not that cheap, especially when you consider that you are also paying with your data.”

Sana Ajani, a business student at the University of Chicago, told Ars that she used to be a premium member but isn’t anymore.

“I did notice the Year in Sport and was a little annoyed that I couldn’t unlock it,” she said in an email. “I would’ve expected some overall stats for everyone and extra stats for subscribers. Year in Review-type stuff is great content and distribution for most apps since everyone shares it on socials, so I’m surprised that Strava is limiting its reach by only letting paid subscribers see it.”

Strava puts popular “Year in Sport” recap behind an $80 paywall Read More »

peacock-showing-ads-upon-launch-opens-the-door-for-more-disruptive-streaming-ads

Peacock showing ads upon launch opens the door for more disruptive streaming ads

Peacock subscribers will see ads immediately upon opening the streaming app or website next year. It’s a bold new strategy for attracting advertisers—something that’s been increasingly important to subscription-based streaming services—but it also risks alienating viewers

As reported by Variety, the new type of ads will display on the profile selection page that shows when a subscriber launches Peacock. Starting next year, instead of the profile page just showing your different Peacock profiles, most of the page will be dominated by an advertorial image. The circles of NBCUniversal-owned characters selected for user profiles will be relegated to a vertical column on the screen’s left side, as you can see here.

To avoid seeing what NBCUniversal is calling “Arrival Ads” every time you open Peacock, you need to subscribe to Peacock’s most expensive plan, which is ad-free and starts at $17 per month (Peacock’s ad-based plans start at $8/month.)

NBCUniversal’s announcement claims that Peacock will be the first streaming service to implement this type of ad. But that may not be the brag the entertainment giant thinks it is, as subscribers may quickly find the startup ads disruptive.

Peacock isn’t making money

Over the past couple of years, it’s become increasingly important for streaming services to generate revenue beyond subscription fees. Peacock and many other streaming services have struggled with profitability after spending years focusing on pricey content production and licensing to attract subscribers.

For its part, Peacock has 41 million subscribers and isn’t profitable. In its most recent quarterly earnings report, shared in October, NBCUniversal parent company Comcast reported that the service lost $217 million in earnings before interest, taxes, depreciation, and amortization, compared to losing $436 million in the same quarter in 2024. At the same time, Peacock has struggled to grow viewership and has had the same number of subscribers since Q1 2025. In Q1 2024, Peacock had 31 million subscribers.

Peacock showing ads upon launch opens the door for more disruptive streaming ads Read More »

llms’-impact-on-science:-booming-publications,-stagnating-quality

LLMs’ impact on science: Booming publications, stagnating quality

This effect was likely to be most pronounced in people that weren’t native speakers of English. If the researchers limited the analysis to people with Asian names working at institutions in Asia, their rate of submissions to bioRxiv and SSRN nearly doubled once they started using AI and rose by over 40 percent at the arXiv. This suggests that people who may not have the strongest English skills are using LLMs to overcome a major bottleneck: producing compelling text.

Quantity vs. quality

The value of producing compelling text should not be underestimated. “Papers with clear but complex language are perceived to be stronger and are cited more frequently,” the researchers note, suggesting that we may use the quality of writing as a proxy for the quality of the research it’s describing. And they found some indication of that here, as non-LLM-assisted papers were more likely to be published in the peer reviewed literature if they used complex language (the abstracts were scored for language complexity using a couple of standard measures).

But the dynamic was completely different for LLM-produced papers. The complexity of language in papers written with an LLM was generally higher than for those using natural language. But they were less likely to end up being published. “For LLM-assisted manuscripts,” the researchers write, “the positive correlation between linguistic complexity and scientific merit not only disappears, it inverts.”

But not all of the differences were bleak. When the researchers checked the references being used in AI-assisted papers, they found that the LLMs weren’t just citing the same papers that everyone else did. They instead cited a broader range of sources, and were more likely to cite books and recent papers. So, there’s a chance that AI use could ultimately diversify the published research that other researchers consider (assuming they check their own references, which they clearly should).

What does this tell us?

There are a couple of cautions for interpreting these results. One, acknowledged by the researchers, is that people may be using AI to produce initial text that’s then heavily edited, and that may be mislabeled as human-produced text here. So, the overall prevalence of AI use is likely to be higher. The other is that some manuscripts may take a while to get published, so their use of that as a standard for scientific quality may penalize more recent drafts—which are more likely to involve AI use. These may ultimately bias some of the results, but the effects the authors saw were so large that they’re unlikely to go away entirely.

LLMs’ impact on science: Booming publications, stagnating quality Read More »