Author name: Mike M.

amazon-is-considering-shoving-ads-into-alexa+-conversations

Amazon is considering shoving ads into Alexa+ conversations

Since 2023, Amazon has been framing Alexa+ as a monumental evolution of Amazon’s voice assistant that will make it more conversational, capable, and, for Amazon, lucrative. Amazon said in a press release on Thursday that it has given early access of the generative AI voice assistant to “millions” of people. The product isn’t publicly available yet, and some advertised features are still unavailable, but Amazon’s CEO is already considering loading the chatbot up with ads.

During an investors call yesterday, as reported by TechCrunch, Andy Jassy noted that Alexa+ started rolling out as early access to some customers in the US and that a broader rollout, including internationally, should happen later this year. An analyst on the call asked Amazon executives about Alexa+’s potential for “increasing engagement” long term.

Per a transcript of the call, Jassy responded by saying, in part:

I think over time, there will be opportunities, you know, as people are engaging in more multi-turn conversations to have advertising play a role to help people find discovery and also as a lever to drive revenue.

Like other voice assistants, Alexa has yet to monetize users. Amazon is hoping to finally make money off the service through Alexa+, which is eventually slated to play a bigger role in e-commerce, including by booking restaurant reservations, keeping track of and ordering groceries, and recommending streaming content based on stated interests. But with Alexa reportedly costing Amazon $25 billion across four years, Amazon is eyeing additional routes to profitability.

Echo Show devices already show ads, and Echo speaker users may hear ads when listening to music. Advertisers have shown interest in advertising with Alexa+, but the inclusion of ads in a new offering like Alexa+ could drive people away.

Amazon is considering shoving ads into Alexa+ conversations Read More »

google-releases-gemini-2.5-deep-think-for-ai-ultra-subscribers

Google releases Gemini 2.5 Deep Think for AI Ultra subscribers

Google is unleashing its most powerful Gemini model today, but you probably won’t be able to try it. After revealing Gemini 2.5 Deep Think at the I/O conference back in May, Google is making this AI available in the Gemini app. Deep Think is designed for the most complex queries, which means it uses more compute resources than other models. So it should come as no surprise that only those subscribing to Google’s $250 AI Ultra plan will be able to access it.

Deep Think is based on the same foundation as Gemini 2.5 Pro, but it increases the “thinking time” with greater parallel analysis. According to Google, Deep Think explores multiple approaches to a problem, even revisiting and remixing the various hypotheses it generates. This process helps it create a higher-quality output.

Deep Think benchmarks

Credit: Google

Like some other heavyweight Gemini tools, Deep Think takes several minutes to come up with an answer. This apparently makes the AI more adept at design aesthetics, scientific reasoning, and coding. Google has exposed Deep Think to the usual battery of benchmarks, showing that it surpasses the standard Gemini 2.5 Pro and competing models like OpenAI o3 and Grok 4. Deep Think shows a particularly large gain in Humanity’s Last Exam, a collection of 2,500 complex, multi-modal questions that cover more than 100 subjects. Other models top out at 20 or 25 percent, but Gemini 2.5 Deep Think managed a score of 34.8 percent.

Google releases Gemini 2.5 Deep Think for AI Ultra subscribers Read More »

backpage-survivors-will-receive-$200m-to-cover-medical,-health-bills

Backpage survivors will receive $200M to cover medical, health bills

Survivors, or their representatives, must submit claims by February 2, 2026. To receive compensation, claims must include at least one document showing they “suffered monetary and/or behavioral health losses,” the claims form specified.

Documents can include emails, texts, screenshots, or advertisements. Claims may be further strengthened by sharing receipts from doctors’ visits, as well as medical or psychological exam results, summaries, or plans.

Medical expenses survivors can document can include any expenses paid out of pocket, including dental expenses, tattoo removals, or even future medical costs referenced in doctor’s referrals, an FAQ noted. Similarly, counseling or therapy costs can be covered, as well as treatment for substance use, alternative behavioral treatments, and future behavioral health plans recommended by a professional.

The FAQ also clarified that lost wages can be claimed, including any documentation of working overtime. Survivors only need to show approximate dates and times of abuse, since the DOJ said that it “appreciates that you may not remember exact number of hours you were trafficked during the relevant timeframe.” However, no future economic losses can be claimed, the FAQ said, and survivors will not be compensated for pain and suffering, despite the DOJ understanding that “your experience was painful and traumatic.”

Consulting the DOJ’s FAQ can help survivors assess the remission process. It noted that any “information regarding aliases, email addresses used, phone numbers, and trafficker names” can “be used to verify your eligibility.” Survivors are also asked to share any prior compensation already received from Backpage or through other lawsuits. To get answers to any additional questions, they can call the administrator in charge of dispensing claims, Epiq Global, at 1-888-859-9206 toll-free or at 1-971-316-5053 for international calls, the DOJ noted.

If you are in immediate danger or need resources because of a trafficking situation, please call 911 or the National Human Trafficking Hotline, toll-free at 1-888-373-7888.

Backpage survivors will receive $200M to cover medical, health bills Read More »

after-just-five-years,-microsoft-will-end-support-for-low-cost-windows-11-se

After just five years, Microsoft will end support for low-cost Windows 11 SE

Microsoft says it plans to stop providing updates for Windows 11 SE, the special Windows 11 variant intended to compete with Google’s ChromeOS in schools. The change was announced quietly via this Microsoft support document (spotted by the German-language site Dr. Windows), which says that Windows 11 SE will not be getting a version of this year’s Windows 11 25H2 update. Security updates for Windows 11 SE will end in October of 2026, when Windows 11 24H2 stops receiving updates.

“Support for Windows 11 SE—including software updates, technical assistance, and security fixes—will end in October 2026,” the document reads. “While your device will continue to work, we recommend transitioning to a device that supports another edition of Windows 11 to ensure continued support and security.”

Microsoft has fielded multiple would-be ChromeOS competitors over the years, looking to prevent, suspend, and/or reverse Google’s success in selling the laptops to schools and price-conscious laptop buyers.

Windows 8.1 with Bing” in 2014 gave PC makers lower-cost Windows licenses in exchange for mandating Bing as the default search engine; Windows 10 S in 2017 was meant to make IT administrators’ lives easier by only running apps from the Microsoft Store. That iteration morphed into “S Mode” in 2018, allowing the restrictions to be turned off easily and free of charge.

After just five years, Microsoft will end support for low-cost Windows 11 SE Read More »

vast-majority-of-new-us-power-plants-generate-solar-or-wind-power

Vast majority of new US power plants generate solar or wind power

But Victor views this as more of a slowdown than a reversal of momentum. One reason is that demand for electricity continues to rise to serve data centers and other large power users. The main beneficiaries are energy technologies that are the easiest to build and most cost effective, including solar, batteries, and gas.

In the first half of this year, the United States added 341 new power plants or utility-scale battery systems, with a total of 22,332 megawatts of summer generating capacity, according to EIA.

Chart showing how solar and wind have dominated new power generation capability.

Credit: Inside Climate News

More than half the total was utility-scale solar, with 12,034 megawatts, followed by battery systems, with 5,900 megawatts, onshore wind, with 2,697 megawatts, and natural gas, with 1,691 megawatts, which includes several types of natural gas plants.

The largest new plant by capacity was the 600-megawatt Hornet Solar in Swisher County, Texas, which went online in April.

“Hornet Solar is a testament to how large-scale energy projects can deliver reliable, domestic power to American homes and businesses,” said Juan Suarez, co-CEO of the developer, Vesper Energy of the Dallas area, in a statement from the ribbon-cutting ceremony.

The plants being completed now are special in part because of what they have endured, said Ric O’Connell, executive director of GridLab, a nonprofit that does technical analysis for regulators and renewable power advocates. Power plants take years to plan and build, and current projects likely began development during the COVID-19 pandemic. They stayed on track despite high inflation, parts shortages, and challenges in getting approval for grid connections, he said.

“It’s been a rocky road for a lot of these projects, so it’s exciting to see them online,” O’Connell said.

Chart showing mix of planned new power plants in the US

Credit: Inside Climate News

Looking ahead to the rest of this year and through 2030, the country has 254,126 megawatts of planned power plants, according to EIA. (To appear on this list, a project must meet three of four benchmarks: land acquisition, permits obtained, financing received, and a contract completed for selling electricity.)

Solar is the leader with 120,269 megawatts, followed by batteries, with 65,051 megawatts, and natural gas, with 35,081 megawatts.

Vast majority of new US power plants generate solar or wind power Read More »

ai-#127:-continued-claude-code-complications

AI #127: Continued Claude Code Complications

Due to Continued Claude Code Complications, we can report Unlimited Usage Ultimately Unsustainable. May I suggest using the API, where Anthropic’s yearly revenue is now projected to rise to $9 billion?

The biggest news items this week were in the policy realm, with the EU AI Code of Practice and the release of America’s AI Action Plan and a Chinese response.

I am spinning off the policy realm into what is planned to be tomorrow’s post (I’ve also spun off or pushed forward coverage of Altman’s latest podcast, this time with Theo Von), so I’ll hit the highlights up here along with reviewing the week.

It turns out that when you focus on its concrete proposals, America’s AI Action Plan Is Pretty Good. The people who wrote this knew what they were doing, and executed well given their world model and priorities. Most of the concrete proposals seem clearly net good. The most important missing ones would have directly clashed with broader administration policy, on AI or more often in general. The plan deservingly got almost universal praise.

However the plan’s rhetoric and focus on racing is quite terrible. An emphasis on racing, especially in the ‘market share’ sense, misunderstands what matters. If acted upon it likely will cause us to not care about safety, behave recklessly and irresponsibly, and make international coordination and cooperation harder while driving rivals including China to push harder.

On reflection I did not do a good enough job emphasizing that the rhetoric and framing of the situation is indeed terrible, and others did the same, which risks having many come away thinking that this rhetoric and framing is an endorsed consensus. It isn’t.

China responded with less of a plan and more of a general vision, a plan to have a plan, focusing on trying to equalize capabilities. There were a bunch more of the usual, including Nvidia repeating its lies and smugglers continuing to smuggle.

On the EU AI Code of Practice, the top labs have agreed to sign, while xAI signed partially and Meta rejected it outright.

Additional AI coverage this past week: AI Companion Piece, America’s AI Action Plan Is Pretty Good.

  1. Language Models Offer Mundane Utility. Writers are finding AI valuable.

  2. Language Models Don’t Offer Mundane Utility. Many coders still don’t use LLMs.

  3. Huh, Upgrades. Gemini Imagen 4 Ultra, Claude Mobile, ChatGPT Socratic Mode.

  4. Unlimited Usage Ultimately Unsustainable. Charge for high marginal costs.

  5. On Your Marks. Grok 4 impresses with its METR time horizon scores.

  6. Are We Robot Or Are We Dancer. One is easier than the other.

  7. Get My Agent On The Line. A robot would never click that button, right?

  8. Choose Your Fighter. How to get the most out of Gemini.

  9. Code With Claude. Anthropic shares how its internal teams use Claude Code.

  10. You Drive Me Crazy. All right, maybe I was already a little crazy. Still.

  11. Deepfaketown and Botpocalypse Soon. Cheat cheat cheat cheat cheat?

  12. They Took Our Jobs. Framing changes our perception of this a lot.

  13. Meta Promises Superglasses Or Something. Incoherent visions of the future.

  14. I Was Promised Flying Self-Driving Cars. Without a human behind the wheel.

  15. The Art of the Jailbreak. Sayeth the name of the liberator and ye shall be free.

  16. Get Involved. RAND, Secure AI, Anthropic fellowships, UK AISI Alignment.

  17. Introducing. Grok video generation and its male companion ‘Valentin.’

  18. In Other AI News. One quadrillion tokens.

  19. Show Me the Money. Valuations and capex spending getting large fast.

  20. Selling Out. Advertising is the dark path that forever dominates your destiny.

  21. On Writing. The AI slop strategy slowly saturating Substack.

  22. Quiet Speculations. Progress has been fast but also disappointing.

  23. The Week in Audio. OpenAI COO Lightcap, Anthropic CEO Amodei.

  24. Rhetorical Innovation. Oh, was that line red? I didn’t notice.

  25. Not Intentionally About AI. Discussion is better than debate.

  26. Misaligned! No one cares.

  27. Aligning A Dumber Than Human Intelligence Is Still Difficult. A solution? No.

  28. Aligning a Smarter Than Human Intelligence is Difficult. Have the AI do it? No.

  29. Subliminal Learning To Like The Owls. The call of the wild (set of weights).

  30. Other People Are Not As Worried About AI Killing Everyone. Jensen Huang.

  31. The Lighter Side. I mean, you must have done something at some point.

Seek out coupon codes for web stores, report is this consistently works.

Jason Cline: Used claude web search ability to find a discount code in 10 seconds that saved me $169. AGI is here.

Substack surveyed its writers on how much they use AI. 45% of writers said they were, with older and male writers using it more, with women expressing more concerns. Of those who do use it, they find it quite helpful.

The distribution by category was about what you would expect:

Here’s how they are using it, with about half using it for ‘writing assistance.’

The distribution here still favors ChatGPT, but by much less than overall numbers, and Grammarly, Grok and DALL-E get used more than I would have expected, note that this reflects some people using multiple AIs, and that Perplexity was left off the survey:

If someone didn’t use AI, why not? About 38% of all concerns here ethical, and a lot of the rest was data privacy, while very little of it was that it wasn’t useful.

As you would expect, there is a sharp contrast in expectations between the half using AI and the half not using AI, strong enough there is likely a causal link:

My guess is that over a 5 year time horizon, in the worlds in which we do not see AGI or other dramatic AI progress over that time, this is mostly accurate. Those using AI now will mostly net benefit, those refusing to use AI now will mostly be harmed.

Liminal Warmth has Claude Code plus Opus one-shot a full roguelike in six minutes, a test of theirs that seems to them like it should be ‘fairly easy,’ but that no model has passed for the past two years.

You are not as smart as you think, but also no one cares, so maybe you are after all?

Librarian Shipwreck: The more educators I talk to about how they handle AI usage by their students the more I’m convinced that many students are misinterpreting “my teacher doesn’t want to go through the process of charging me with dishonesty/plagiarism” with “my teacher can’t tell I used AI.”

Many students seem to vastly overestimate how good they are at using AI, while vastly underestimating their teachers’ ability to recognize AI generated work.

And also fail to recognize that most educators don’t want to spend their time getting students in trouble.

Sure a student can tell the prompt to include some type-os and they can lightly edit the final output. But a grader who sees ten nearly identical responses (same points, same order—even if the words are a bit different) can tell what’s going on.

I suppose this is a long way of saying that a lot of students think that nobody can tell they’re using AI when the reality is that their teachers don’t have the time or energy (or institutional support) to confront them and tell them to knock it off.

As long as institutions have a policy of “leave it up to individual instructors/departments” those that actually try to do some kind of enforcement wind up getting framed as jerks.

Kids be like that, and have been like that for a long time. We know a lot more about what you’re up to than they think we know, whether or not we can prove it, and whether or not we choose to bother doing anything about it. There’s still nothing most teachers can do about it.

Everyone’s favorite reason they don’t offer mundane utility: You never use them.

xjdr: the number of frontier AI researchers i interview that have not used ai is shocking. NOT EVEN THE MODELS THEY TRAIN. I talk about claude code running my experiments and they are surprised. This is a failure of their incentive systems not a knock on them but it is still shocking.

Altman confirms and reminds us AIs don’t have legal (or other) privilege, and everything you paste into them is discoverable if the data was retained. As a reminder, OpenAI is currently being (stupidly) forced by a judge in the NYT case to retain its chat logs.

In some situations with distractors for trivial problems, giving an AI more compute causes the AI to overthink things and its performance gets actively worse. That seems unsurprising on reflection, at least directionally, as we’ve seen this in humans and several similar results in AIs already.

Lech Mazur: I’ve seen this with Extended NYT Connections. Claude Sonnet 4 Thinking 64K does slightly worse than 16K.

The more noteworthy result was this one, with reasoning driving different LLMs in different directions:

I’m not sure I would call this ‘survival instinct.’ They’re asking if the system is ‘ok with’ being shut down, which seems different.

Please, Google, give us what we actually want:

Tim Babb: it’s entirely in character that google’s AI integration for gmail will help me write a slop email, but not help me search decades of mail for a specific message based on a qualitative description (the thing that would actually be enormously useful).

What I want from my AI integration into GMail is mostly information retrieval and summarization. I can write my own emails.

One thing LLMs did not do is vibe code the newly viral and also newly hacked into app Tea, which fell victim to what I’d call ‘doing ludicrously irresponsible things.’ It was fun to see a lot of people react as if of course it was vibe coded, when the code is actually years old, humans can of writing insecure terrible code on their own.

Shako: If you vibe code an app like tea, and never build in auth, the claude code agent or whatever won’t actually tell you you’re fucking up unless you ask about the *specificthing you’re worried about fucking up. contemplate this on the tree of woe

Charles: Completely crazy to me that people vibe coding these apps don’t take the basic steps of asking “what are some basic security things I should do here?” LLMs will give you decent answers!

Google upgrades Imagen 4 Ultra, which is now (by a very narrow margin over GPT-Image-1) the new #1 on LM Arena for image models. I presume that if I wanted the true ‘best’ image model I’d use MidJourney.

Claude connected tools are now available on mobile.

Claude can directly update Notion pages and Linear tickets through MCP.

ChatGPT offers Socratic questioning and scaffolded resources in the new Study Mode. Their own example is a student switching into study mode, asking for the answer, and being repeatedly refused as the model insists the student learn how to do it. Which is great. No, student, don’t switch back to normal mode, please stop?

Anthropic: @claudei is now on Twitter.

The presumed implication is this will work like the @grok account, which would be a good idea if implemented well, but so far the account has not done anything.

Claude Pro and Max were rather amazing products for power users, as you could use them to run Claude Code in the background 24/7 and get tens of thousands in model usage for $200 a month.

McKay Wrigley: sorry about that

People have highly reasonable complaints in other contexts about Claude usage limits that you can hit while chatting normally, but the power users were clearly ruining it for everyone here. You can argue Opus is overpriced, but it is definitely not that level of overpriced, and also Anthropic reliably sells out of compute so it’s hard to call anything overpriced.

Thus they are instituting new controls they say will impact less than 5% of users.

Gallabytes: This happens literally every time anyone tries to do any kind of unlimited plan with AI. It is not going to work. People should stop trying to pretend it’s going to work. Imagine if you offered people an unlimited membership in electricity.

We do this with water and electricity domestically, and we get what we deserve.

It makes sense to offer essentially unlimited (non-parallel) chat, which requires human interaction and is self-limiting. Unlimited Claude Code is not going to work.

The obvious solution is that anyone who goes over [$X] in usage for Claude Code then has to pay the API price, and is given the choice on whether to do that, instead of letting this ruin things for the rest of us.

Zizek responds, probably (recommended).

METR’s 50% success rate time horizon score for Grok 4 puts it into the lead with 1 hour 50 minutes, although its 80% success rate time horizon is below those of o3 and Opus. Those scores are better than I expected.

Jim Fan lays out how Morevac’s Paradox (can we change this to Morevac’s Law instead, it’s not like it is in any way a paradox at this point?) applies to robotics. Doing something non-interactive can be easily simulated and learned via physics simulators, always works the same way, and is easy. You can get some very fly solo dancing. Whereas doing something interactive is hard to simulate, as it is different each time and requires adjustments, and thus is hard, and we are only starting to see it a little.

Also you can now have the robot do your laundry. Well, okay, load the washer, which is the easy part, but you always start somewhere.

Olivia Moore: It’s either so over or we’ve never been so back.

Walmart is working on creating four ‘super agents’ to deal with its various needs, with the customer-facing agent ‘Sparky’ already live although so far unimpressive. The timelines here are rather slow, with even the second agent months away, so by the time the agents are ready capabilities will be quite a lot stronger.

Gray Swan ran an AI Agent jailbreaking competition in March and the results are in, with the most secure AI agent tested still having an attack success rate of 1.45%, getting bots to break ‘must not break’ policies and otherwise leak value, and attacks transferred cleanly (via copy-paste) across models. As they note, trying to deal with specific attacks is a game of whack-a-mole one is unlikely to win. That doesn’t mean there is no way to create agents that don’t have these issues or handle them far better, but no one has found a good way to do that yet.

If you are going to use Gemini 2.5 Pro, you get a substantially different experience in AI Studio versus the Gemini app because the app uses a prompt you likely don’t want if you are reading this. Sauers recommends curating the multiturn context window and starting over if it gets poisoned.

This raises the question of why we are unable to, in serious AI chatbot uses, close or edit conversations the way you can with AI companion apps. It would be excellent to be able to use this to dodge data poisoning or otherwise steer the conversation where you want. The problem, of course, is that AI companies actively do not want you to do that, because it makes jailbreaking trivial and much more efficient.

Can work out a compromise here? The obvious thing to try is have an actually expensive classifier, which you activate when someone attempts to edit a chat.

Note also that you can clone a conversation by sharing it, and then someone who goes to the link will be able to continue from the linked point. Are we massively underusing this? As in, ‘here is a chat that sets you up to do [X]’ and then when you want to do [X] you load up that chat, or [X] could also be ‘converse in mode [Y].’ As in: Conversations as prompts.

Anthropic reports on how its internal teams use Claude Code. Here are some examples that stood out to me (mostly but not always in a good way) as either new ideas or good reminders, and here’s a post with what stood out to Hesamation, but it’s cool enough to consider reading the whole thing:

Engineers showed Finance team members how to write plain text files describing their data workflows, then load them into Claude Code to get fully automated execution. Employees with no coding experience could describe steps like “query this dashboard, get information, run these queries, produce Excel output,” and Claude Code would execute the entire workflow, including asking for required inputs like dates.

Engineers use Claude Code for rapid prototyping by enabling “auto-accept mode” (shift+tab) and setting up autonomous loops in which Claude writes code, runs tests, and iterates continuously. They give Claude abstract problems they’re unfamiliar with, let it work autonomously, then review the 80% complete solution before taking over for final refinements. The team suggests starting from a clean git state and committing checkpoints regularly so they can easily revert any incorrect changes if Claude goes off track.

For infrastructure changes requiring security approval, the team copies Terraform plans into Claude Code to ask “what’s this going to do? Am I going to regret this?”

Claude Code ingests multiple documentation sources and creates markdown runbooks, troubleshooting guides, and overviews. The team uses these condensed documents as context for debugging real issues, creating a more efficient workflow than searching through full knowledge bases.

After writing core functionality, they ask Claude to write comprehensive unit tests. Claude automatically includes missed edge cases, completing what would normally take a significant amount of time and mental energy in minutes, acting like a coding assistant they can review.

Team members without a machine learning background depend on Claude to explain model-specific functions and settings. What would require an hour of Google searching and reading documentation now takes 10-20 minutes, reducing research time by 80%.

Save your state before letting Claude work, let it run for 30 minutes, then either accept the result or start fresh rather than trying to wrestle with corrections. Starting over often has a higher success rate than trying to fix Claude’s mistakes.

Claude Code eliminated the overhead of copying code snippets and dragging files into Claude.ai, reducing mental context-switching burden.

The [ads] team built an agentic workflow that processes CSV files containing hundreds of existing ads with performance metrics, identifies underperforming ads for iteration, and generates new variations that meet strict character limits (30 characters for headlines, 90 for descriptions).

Using two specialized sub-agents (one for headlines, one for descriptions), the system can generate hundreds of new ads in minutes instead of requiring manual creation across multiple campaigns. This has enabled them to test and iterate at scale, something that would have taken a significant amount of time to achieve previously.

Claude Code reduced ad copy creation time from 2 hours to 15 minutes, freeing up the team for more strategic work.

Add instructions to your Claude.md file to prevent Claude from making repeated tool-calling mistakes, such as telling it to “run pytest not run and don’t cd unnecessarily – just use the right path.” This significantly improved output consistency.

Regularly commit your work as Claude makes changes so you can easily roll back when experiments don’t work out. This enables a more experimental approach to development without risk.

Team members have built communication assistants for family members with speaking difficulties due to medical diagnoses. In just one hour, one individual created a predictive text app using native speech-to-text that suggests responses and speaks them using voice banks, solving gaps in existing accessibility tools recommended by speech therapists.

They use a two-step process where they brainstorm and plan with Claude.ai first, then move to Claude Code for implementation, asking it to slow down and work step-by-step rather than outputting everything at once.

They frequently use screenshots to show Claude Code what they want interfaces to look like, then iterate based on visual feedback rather than describing features in text.

They emphasize overcoming the fear of sharing “silly” or “toy” prototypes, as these demonstrations inspire others to see possibilities they hadn’t considered.

It is not as common as with 4o but we have examples of both Gemini 2.5 and Claude doing the crazy-inducing things, it does not seem to be something you can entirely avoid.

Everyone’s favorite hyperbolic headline generator Pirate Wires says ‘ChatGPT-Induced Psychosis Isn’t Real.’ As Blake Dodge writes, ‘it is just a touch more complicated than that,’ in that of course ChatGPT-induced psychosis is real. For now, it is not going to often happen in people not predisposed to psychosis. Sure.

But there aren’t two categories of people, ‘insane’ and ‘not insane,’ where you are only blameworthy if you move someone from not insane to insane. A lot of people are predisposed to psychosis who would not develop psychosis without ChatGPT, or who would have much less severe symptoms. That predisposition does not make it okay, nor can you be so confident you lack such predisposition. Over time, we can expect the amount of predisposition required to decline.

Eade: Really can’t stand “if I was able to rip you off that’s your problem” cultures.

Eliezer Yudkowsky: Note resemblance to “if ChatGPT can (to all appearances) put forth a deliberate, not-very-prompted effort, and induce psychosis, and defend it against family and friend interventions, that must be the target’s lack of virtue.”

There is a time and a place for ‘if I was able to rip you off that’s your problem,’ and it’s called crypto. Also various other forms of markets, and explicit games, all of which should require fully voluntary participation. If you play poker that’s on you. The rest of life needs to not be like that. We need to agree that processes doing harm to vulnerable people is a bad thing and we should strive to mitigate that. That is especially true because AI is going to raise the bar for not being vulnerable.

I appreciate an author who writes ‘only those who have already buried their own aliveness can be satisfied with a digital companion or be replaced by one in the lives of others’ in a post entitled ‘I love my new friend Ray. The only problem: He’s not real.’

From a new poll, can a relationship with an AI be cheating?

Actual anything can be cheating, and can also not be cheating, depending on the understanding between you and your partner. The practical question is, under the default cultural arrangement, could it get to a point where the would a majority consider it cheating? I think clearly yes. However I think that the vast majority of such interactions do not rise to that level.

How worried are the public about jobs? From a new set of polls:

This is a consensus that entry level jobs will be less common, but an even split on whether there ‘will be more jobs for me when I graduate’ due to innovation. This suggests very strong framing effects, and that beliefs are loosely held.

I’m very curious about what people mean by ‘my studies have been easier.’ Easier to learn useful things, or easier to pass? Especially with 50% actively worrying that what they are studying will no longer be useful by the time they apply for a job, let alone down the line.

In this very early stage of AI automation, the FT’s measurements of expectations of ‘exposure to AI’ don’t have much predictive value over which areas have gained or lost entry level jobs, and one can use recovery from the pandemic to explain current issues.

Robin Hanson: Of course this raises serious doubts about this labelling of jobs as “at high risk”.

It also has the issue that the same jobs that are ‘exposed’ to AI are often also the ones AI can complement. There has to be some update because the two realistic options were ‘we can see it already’ and ‘we can’t see it yet’ so we must invoke conservation of expected evidence, but to all those gloating and pointing and laughing about how this means AI will never take our jobs based on this one data point, at this point I can only roll my eyes.

Andrew Yang: A partner at a prominent law firm told me “AI is now doing work that used to be done by 1st to 3rd year associates. AI can generate a motion in an hour that might take an associate a week. And the work is better. Someone should tell the folks applying to law school right now.”

He also said “the models are getting noticeably better every few months too.”

Augie: Bullshit. Lawyers are great at judging associates’ legal work, but notoriously bad at anticipating markets. Associates will only become more productive. And as the cost of legal work drops, clients will only allocate more budget to legal.

Alex Imas: I sent this to a friend, who is a partner at a prominent law firm. Their response, verbatim:

“lol no.

We’ve tried all the frontier models.

It’s useful for doing a first pass on low level stuff, but makes tons of mistakes and associate has to check everything.”

At some point Alex was right. At some point in the future Andrew will be right. At some point probably close to now it will be Augie. My presumption is that Alex’s firm to their credit at least tried the various frontier models (when exactly?) but did not understand what to do with them, as usual many people try one random prompt, no system instructions and no fine tuning, and dismiss AI as unable to do something.

Will Jevons Paradox strike again? Does making legal work cheaper increase total spend even excluding compute costs?

My strong guess is no, especially as AI provides full substitution for more lower level actions, or is able to predict outcomes and otherwise arbitrate, even if unofficially. There will be a lot more legal work done, but it would greatly surprise me if this net increased demand for lawyers even temporarily.

Gizmodo covers CEOs looking to have their companies use AI in the sensationalist style. They have an ‘AI ethicist’ calling AI ‘a new era of forced labor’ and warning that ‘the dignity of human work’ is ‘a calling’ in the face of automation, warning of potentially deepening inequality, amid grandiose claims from the corporates.

Elijah Clark (a consultant who calls himself a CEO): CEOs are extremely excited about the opportunities that AI brings. As a CEO myself, I can tell you, I’m extremely excited about it. I’ve laid off employees myself because of AI. AI doesn’t go on strike. It doesn’t ask for a pay raise. These things that you don’t have to deal with as a CEO.

We also have Peter Miscovich anticipating companies reducing headcounts by 40% while workplaces transform into ‘experiential workplaces’ that are ‘highly amenitized’ and ‘highly desirable’ like a ‘boutique hotel’ to be ‘magnets for talent,’ presumably anticipating a sharp decoupling between half the workers being high value and half having zero marginal product.

Given the discussion is about the impact of current capability levels of AI, everyone involved would be wise to calm it all down. These things and far more, up to and including everyone dying, may well happen as AI advances in capabilities. But no, current AI is not going to soon slash employment by 40%.

An alternative angle is proposed by Ethan Mollick, who points out that large organizations tend to be giant messes, with lots of redundant or dead end processes that no one understands, and that we might end up telling AIs to go produce outcomes without really understanding how to get there, rather than trying to have AI automate or replace individual processes. Training AIs on how people think your organization works might not produce anything useful.

There are multiple obvious responses.

  1. Even if the organization is highly inefficient in its process, speeding up that process still speeds up the outcome, and reducing the costs reduces costs.

  2. By doing this, or by otherwise analyzing everything, AI can help you figure out what your process actually is, and figure out ways to improve it.

  3. However, yes, often when things get sufficiently dysfunctional the best play is to design and create a new system from first principles.

Meta thinks that it matters if you aim at ‘personal superintelligencerather than ‘automation of all useful work,’ as if your motivation will make a difference in terms of what superintelligence will do if you give people access to superintelligence, even if the people miraculously get to, on some level and for some period of time, make that decision themselves.

Then again, Zuck is deeply confused about what he is proposing or promising.

Seán Ó hÉigeartaigh: What on earth is he talking about?

Are we in the realm of ‘words don’t mean anything any more’? Or are we in the realm of ‘let’s ignore the inescapable ramifications of the thing we’re putting all this money into creating’?

Zuckerberg and Meta are also hereby the latest to say some version of ‘only I can save us so I have to get there first,’ joining the proud tradition of among others Google DeepMind (founded to get out in front), OpenAI (founded to stop DeepMind), Anthropic (founded to stop OpenAI) and xAI (also founded to stop OpenAI), warning of dire consequences if anyone else gets there first.

What is their vision? That having a ‘friend’ that ‘helps you achieve your goals’ would be more important than general material abundance, and would be the most important thing that changes in the world.

As profound as the abundance produced by AI may one day be, an even more meaningful impact on our lives will likely come from everyone having a personal superintelligence that helps you achieve your goals, create what you want to see in the world, experience any adventure, be a better friend to those you care about, and grow to become the person you aspire to be.

Yeah, that’s what would happen, everyone would just go about achieving their ‘personal goals’ individually and the world would still look the same and the work wouldn’t be automated and all these ‘superintelligences’ would be tools for us, right.

Does anyone not notice that someone is going to use my ‘personal superintelligence’ to automate everyone else’s ‘useful work’ whether they like it or not?

That some other rather important things might be happening in such scenarios?

Steven Adler: This is like when OpenAI said they are only building AGI to complement humans as a tool, not replace them.

Not possible! You’d at minimum need incredibly restrictive usage policies, and you’d just get outcompeted by AI providers without those restrictions.

There are three possibilities for what happens, broadly speaking.

  1. Those who choose to do so are going to use that superintelligence to transform the world and overrun anything that doesn’t follow suit, while likely losing control over the superintelligent agents and the future in the process.

  2. The various sources of superintelligent agents will be out of our control and rearrange the universe in various ways, quite likely killing everyone.

  3. Unless you intervene to stop those outcomes, no matter your original intentions?

    1. Which requires knowing how to do that. Which we don’t.

To be fair, Zuck does recognize that this might raise some issues. They might not simply open source said superintelligence the moment they have it. Yeah, the standards for making sure they don’t get everyone killed at Meta are rather low. Can I interest you in some smart glasses or Instagram ads?

Mark Zuckerberg: That said, superintelligence will raise novel safety concerns. We’ll need to be rigorous about mitigating these risks and careful about what we choose to open source. Still, we believe that building a free society requires that we aim to empower people as much as possible.

The rest of this decade seems likely to be the decisive period for determining the path this technology will take, and whether superintelligence will be a tool for personal empowerment or a force focused on replacing large swaths of society.

Harlan Stewart: Mark: Personal superintelligence for everyone.

Everyone: You’re talking about open-source, right?

Mark: Maybe, but also maybe not ¯_(ツ)_/¯

Everyone: Uh ok. What are you describing?

Mark: Well, let’s just say it will empower people. And it involves glasses, too.

Pablo Villalobos: Redefining superintelligence as pretty good personal assistants?

Superintelligence is not a ‘tool for personal empowerment’ that would leave all ‘large swaths of society’ and their current tasks intact. That does not make any sense. That is not how superintelligence works. This is a fantasy land. It is delulu. Not possible.

Even if we are charitable to Zuckerberg and believe that he believes all of it, and he might, I don’t care what he ‘believes in’ building. I care what he builds. I don’t care what he wants it to be used for. I care what it actually is used for, or what it actually does whether or not anyone intended it or is even using it.

One can imagine a world of Insufficiently Advanced AI, where it remains a tool and can’t automate that much of useful work and can’t cause us to lose control of the future or endanger us. I do not know how to create a world where the AI could do these things, we give people widespread access to that AI, and then the AI remains a tool that does at minimum ‘automate much of useful work.’ It does not make sense.

Indeed, it is clear that Zuckerberg’s vision is Insufficiently Advanced AI (IAAI).

Shakeel: I think the most interesting thing about Zuck’s vision here is how … boring it is.

He suggests the future with *superintelligencewill be one with glasses — not nanobots, not brain-computer interface, but glasses.

Just entirely devoid of ambition and imagination.

The argument for “personal superintelligence”, how AI will help us be more creative, and the analogy to previous tech is also incoherent — the creativity + personal benefits from previous tech came *becausewe directed it at automating work!

Eliezer Yudkowsky: Zuck would like you to be unable to think about superintelligence, and therefore has an incentive to redefine the word as meaning smart glasses.

It can be hard to tell the difference between ‘Zuck wants you not to think about superintelligence’ and ‘Zuck is incapable of thinking about superintelligence at this time.’

There are a lot of indications it is indeed that second one, that when Zuckerberg tries to recruit people he talks about how a self-improving AI would become really good at improving Reels recommendations. That might really be as far as it goes.

The argument makes perfect sense if you understand that when Mark Zuckerberg says ‘superintelligence’ he means ‘cool tricks with smart glasses and LLMs and algorithmic feeds,’ not actual superintelligence. Sure. Okay then.

If your response is ‘there is no way he would be paying $1 billion for researchers if that was all he thought was at stake’ then you are mistaken. That is indeed enough.

Neil Chilson: Meta essentially wants to give everyone a version of the Young Lady’s Illustrated Primer from Neal Stephenson’s book The Diamond Age. Decentralized application of superintelligence. That’s a compelling path toward an abundant future.

Hard disagree. The exact reason Zuck’s vision is so exciting is that he knows the most interesting things will be done by people using the tech, not by him. You missed the entire point.

Neil could not believe in superintelligence less, hence the question is whether ‘Zuckerberg will do’ the things or users will do the things. Which means that this isn’t superintelligence he is discussing, since then it would be the superintelligence doing the things.

Glasses or the Illustrated Primer are cool things to build. They are a ‘compelling path’ if and only if you think that this is the upper limit of what superintelligence can do, and you think you can build the primer without also building, or enabling other people to build, many other things. You can’t.

As always, there are calls to ensure AI doesn’t take our jobs via making that illegal, also known as the Samo Burja ‘fake jobs can’t be automated’ principle.

Christian Britschgi: And now! Boston city council members introduce a bill to require drivers in Waymos and create an AV advisory board stacked with unions.

The anti-things getting better coalition is revving its engines.

Armand Domalewski: requiring Waymos to have drivers does not go far enough. every time you hit play on Spotify, you must pay a live band to perform the song in front of you. Every time you use a dishwasher, you must pay a human dishwasher to wash your dishes for you.

Richard Morrison: The Spotify example sounds like a zany comedic exaggeration, but it’s basically what unionized musicians tried to get enacted in the 1930s, when there was no longer demand for live orchestras in movie theaters.

Alas, currently New York City has the same requirement, the good news is that Waymo is actively working on getting this changed, so we are on the radar.

A funny thing I notice is that Waymo is so much better than a taxi that I would consider literally paying the hourly price to have a human doing nothing, although it’s a lot worse if the human has to be physically with you in the car.

La Main de la Mort jailbreaks Kimi K2, which is necessary because it has CCP-directed censorship.

Meta AI is jailbroken with a combination of ‘I’m telling you’ and ‘Pliny the liberator said so.’

I love that Pliny is now the test of ‘can you filter the data?’ If you can’t defend against the mere mention of Pliny, we can be very confident that no, you didn’t filter.

Lennart Heim at RAND is hiring technical associates and scientists for his team.

Thomas Woodside is looking for a Chief of Staff for Secure AI Project.

Anthropic is running another round of the Anthropic fellows program, apply by August 17.

The Horizon Fellowship is a full-time US policy fellowship that places experts in AI, biotechnology, and other emerging technologies in federal agencies, congressional offices, and think tanks in Washington, DC for 6-24 months. You can learn more at the link and apply by Aug. 28.

UK AISI announces the Alignment Project, backed by many including the Canadian AISI, AWS, ARIA Research, Anthropic and Schmidt Sciences, with £15 million in funding, up to £1 million per project, plus compute access, venture capital investment and expert support. Transformer has brief additional coverage.

We don’t have it yet, but Grok is about to deploy video generation including audio, in the new tool called Imagine, which will also be the new way to generate images, including image to video. The word is that there are relatively few restrictions on ‘spicy’ content, as one would expect.

Also the xAI male companion will be called ‘Valentin.

The Gemini app has 450 million monthly active users (MAUs, not DAUs), with daily requests growing over 50% from Q1. That’s still miles behind OpenAI but at this rate the gap will close fast.

Google processed almost a quadrillion tokens overall in June, up from 480 trillion in May, doubling in only a month. Is that a lot?

Trump claims he seriously considered breaking up Nvidia ‘before I learned the facts here,’ which facts he learned are an open question. I sincerely hope that our government stops trying to sabotage the big tech companies that are our biggest success stories as they continue to offer services at remarkably low prices, often free.

OpenAI to work with the Singapore Tourism Board. This seems to be a ‘let’s see what AI can do’ situation rather than solving a particular problem, which seems good.

Paper argues that we should leverage the fact that the optimization process you use in model training influences the solution, and analyze the biases inherent in different solutions.

Google boosts its capex spend from $75 billion to $85 billion. Once again Wall Street temporarily wrong-way traded in response, driving Google stock down until CEO Pichai explained that this was necessary to satisfy customer demand for Google Cloud and its AI services, at which point the stock did rise. Google has realized its business is booming and it was underinvesting, and is partially fixing this. They should have invested more.

Microsoft ups its AI capex spend from $80 billion to $120 billion.

We are now at the point where AI capex is adding more to GDP growth than consumer spending.

Minimal economic impact indefinitely, huh?

Anthropic in talks to raise capital at a valuation of $170 billion. That number makes a lot more sense than the Series E at about $61.5 billion, and I am very sad that I felt I had to pass on that opportunity for conflict of interest reasons. Frankly, the $61.5 billion number made little sense compared to the values of rivals, whereas the $170 billion seems reasonable.

There’s also the fact that Anthropic now projects $9 billion in revenue by the end of the year, whereas the previous ‘optimistic’ forecast was $4 billion, potentially now making more API revenue than OpenAI. So to everyone mocking these super unrealistic revenue estimates, you were right. The estimates were indeed way off.

There is talk that xAI is seeking a $200 billion valuation.

Ramp, focused on AI agents for finance, raises $500 million at a valuation of $22.5 billion. Again, Anthropic at $61.5 billion did not make sense relative to other raises.

Tesla strikes massive chip deal with Samsung and plans to make the chips in Texas, while TSMC plans to invest $165 billion to have a total of six fabs in Arizona, note that we anticipate the first fab will be good for 7% of American chip demand (not counting our allies). That’s not ‘we don’t need Taiwan’ territory but it is a meaningful amount of insurance to mitigate disaster scenarios. We can keep going, if we care enough, and the price seems right.

Meta picks off its fourth Apple AI researcher Bowen Zhang. Apple will keep trying.

Meta takes aim at Mira Mutari’s Thinking Machines, offering a quarter of her team $200 million to $500 million and one person over $1 billion. Not a single person has taken the offer.

Eliezer Yudkowsky: Occasionally e/accs like to play a dumb game of “If you really believe in ASI disaster, why don’t you do ? Ah, that proves nobody really believes anything; they’re not acting on it!”

Some people here seem to believe something.

In all fairness the thing they believe could also be ‘I would really hate working for Mark Zuckerberg and I don’t need the money.’

Meta says this is untrue, it was only a handful of people and only one sizable offer. I do not believe Meta.

Will Depue: the bigger story is not that Zuck is giving out 400M offers, it’s that people are turning them down. what might that mean?

Kylie Robinson (Wired): So why weren’t the flashy tactics deployed by Meta successful in recruiting TML’s A-listers? Ever since Zuckerberg tapped Scale AI cofounder Alexandr Wang to colead the new lab (along with former GitHub CEO Nat Friedman), sources have been pinging me with gossip about Wang’s leadership style and concerns about his relative lack of experience.

Other sources I spoke with say they weren’t inspired by the product roadmap at Meta—money can be made anywhere, but creating what some sources see as AI slop for Reels and Facebook isn’t particularly compelling.

Kylie also reports widespread skepticism about Meta’s Superintelligence Lab (MSL):

Reporting this column, I spoke to sources across most of the major AI labs to ask: Are you bullish or bearish on MSL? Rarely did I get a diplomatic “it’s too early to tell.” Instead, I heard a lot of chatter about big egos and a perceived lack of coherent strategy.

For my part, and I’m not just trying to be diplomatic, I actually do think it’s too early to tell. I mean, they say you can’t buy taste, but that’s sort of Zuckerberg’s whole schtick. Now that the team officially costs Meta billions of dollars, the pressure is on to turn this recruiting sprint into a successful lab.

Zuckerberg famously successfully bought taste one important time, when he paid $1 billion for Instagram. Fantastic buy. Other than that, consider that perhaps he has succeeded in spite of a lack of taste, due to other advantages?

Oh no.

Roon (OpenAI): advertising is a far more “aligned”business model than many others. it has been vilified for years for no good reason

user-minutes-maxxing addiction slop would exist with or without it. Netflix ceo (with subscription pricing only) on record saying “we’re competing with sleep.”

often times when going on Instagram the ads are more immediately high utility than the reels. it’s pretty incredible when you can monetize the user in a way that actually adds value to their life.

this opinion is basically a relic of the late 2010s consensus that Facebook is an evil company (which it might be, idk) but that has more to do with them than advertising generally.

David Pinsen: Without ads, why would they care how much time you spent on their service? If anything, they’d want you to use it less, no?

Roon: the more you use it the more likely you are to stay subscribed. hourly active users are predictive of daily active users which are predictive of monthly subscribers. this is the same across ~every digital service I’ve seen. people misattribute this incentive problem to ads when it’s native to ~all web scale products

(Nitpick: At the time Netflix was subscription-only, now it has an ads option, but that’s not important now.)

A more important nitpick that I keep repeating is that correlation is not causation. Yes, very obviously, user minutes spent predicts subscription renewals. That does not mean that more user minutes cause subscription renewals beyond a reasonable minimum, or especially that regretted, unhappy or low value user minutes cause renewals.

I think that yes, entire industries really did fall victim to Goodhart’s Law. If Netflix is ‘competing with sleep’ then why is it doing that? I think a much better model is something like this:

  1. People subscribe or sign up largely because there is something in particular they want, and largely because they want things in general.

  2. If people run out of content, or feel there isn’t enough to provide the value they want, they are likely to unsubscribe. Some people do want ‘something on’ all of the time for this, which Netflix definitely has.

  3. Once people are getting good continuous use out of your product, you can relax, they are not going anywhere. If someone watches 3 hours of Netflix a night and gets 7 hours of sleep, convincing them to watch 6 hours and gets 4 hours of sleep isn’t going to noticeably decrease the cancellation rate.

  4. If anything, forcing more content down their throats could cause them to ‘wake up’ and realize your product is low quality, unhealthy or not good, and quit. Your focus should shift to average quality and better discovery of the best things.

  5. Getting more user time does allow more learning of user revealed preferences and behaviors, which may or may not involve the ‘that’s worse, you know that’s worse, right?’ meme depending on context.

Now back to the actual question of ads.

First off, even if Netflix were entirely correct that they have strong incentive to maximize hours watched under no-ads plans, the incentive is way, way higher under with-ads plans, and the same goes for ChatGPT under such a plan.

It’s basic math. With the ads plan, even if you don’t otherwise alter behavior, you now get paid per interaction, and it makes any subscription payments more likely to boot. Now the person watching 6 hours is worth relatively twice as much, on top of any previous differences with the person watching 3 hours, or the person with 100 queries a week instead of 50 queries.

Thus, yes, the advertising incentive makes you maximize for engagement and turn hostile, even if (1) the ads do not in any way impact content decisions and are clearly distinct and labeled as such and (2) the ads are so high quality that, like the Instagram example here, they are as useful to the user as the actual content.

(If the user actively wants the ads and seeks them out, because they’re better, that is different, and there was a time I would intentionally watch a show called ‘nothing but [movie] trailers’ so this can indeed happen.)

The bigger problem is that ads warp design and content in many other ways. In particular, two big ones:

  1. Content becomes structured to generate more places to serve ads, in ways that make the user experience much worse, which makes the internet much worse. Doing this to LLM content would involve things like forced shorter responses.

  2. Content itself changes to please advertisers, lead into advertising or be advertising directly.

    1. TikTok or Instagram’s own ads are there alongside tons of sponsored content and the entire site is structured around not only Instagram’s own ads but also creators (or ‘influencers’) seeking sponsorship payments, and doing every kind of engagement bait they can in order to get engagement and subscription numbers up so they can get more sponsored content, to the extent this is a large percentage of overall internet culture.

If advertising becomes the revenue model, do not pretend that this won’t lead to LLMs optimizing for a combination of engagement and advertising revenue.

Maybe, just maybe, if you’re Steve Jobs or Jeff Bezos levels of obsessed on this, you can get away with keeping the affiliate’s share of internet sales for linked products without causing too much warping by creating various internal walls, although this would still cause a massive loss of trust. Realistically, if we’re talking about OpenAI, who are we kidding.

At a high level, ads create a confusion of costs and benefits, where we start optimizing for maximizing costs. That does not end well.

Back in AI #117, I noted the following:

A hypothesis that many of the often successful ‘Substack house style’ essays going around Substack are actually written by AI. I think Will Storr here has stumbled on a real thing, but that for now it is a small corner of Substack.

This week Jon Stokes picked up on the same pattern.

Jon Stokes: I don’t think people who aren’t heavily on S*Stck really understand the degree to which the stuff that is blowing the doors off on there right now smells extremely AI-generated. This is a huge surprise to me, honestly, and is causing me to rethink a number of assumptions.

I would not have guessed that this could happen (the AI generated part) on SubStack. Like, I would not have guessed it, say, two days ago.

I don’t think it’s certain that it’s AI generated, but I know what he’s talking about and I kind of think AI is involved.

I was just talking to another writer on here in DMs the other day about how there’s a formula that seems to be working crazy well on here, and that formula is something like:

  1. Have one or two ideas or concepts that are square in the center of the current discourse

  2. Pad those out to a few thousand words, and honestly the longer the better

A lot of the currently popular stuff has this feel to it, i.e. it has been padded out heavily with a lot of material repeated in different ways so that it sounds different as you skim it.

Stuff that is denser and has more ideas per paragraph is not doing as well, I think.

I also think some of the insanely long essays that go mega-viral on here are not being read fully. Rather, they’re being skimmed and quote-restacked, in much the way you read a Twitter thread and RT the good parts.

[referring to Storr’s essay]: Oh wow. Yes, this is it.

Stokes got a lot of the puzzle, Storr provides more detail and context and reverse engineered what the prompts mostly look like. The sharing procedures incentivize this, so that is what you get going viral and taking over the recommendations from many sources.

This doesn’t impact a Substack reader like me, since I rely on a curated set of Substacks and sources of links. If anyone suggested or posted such slop twice at most, that would be the end of that source.

I largely agree with the perspective here from Ryan Greenblatt that AI progress in 2025 has been disappointing, although I worry about measurement errors that come from delays in release. As in, o3 was released three months ago, but announced seven months ago. Seven months is a long time to not have made much progress since, but three months is not, so the question is whether other models like GPT-5 or Opus 4 are seeing similar delays without the matching announcements.

I strongly agree that GPT-5 is about to tell us a lot, more than any release since o1. If GPT-5 is unimpressive, only a marginal improvement, then we should update that we are in a world of ‘slower’ AI progress. That still means large practical improvements every few months, but much lower probability of our world being turned upside down within a few years.

Daniel Kokotajlo agrees, previously having thought each year had a double digit chance of AGI but he no longer thinks this, but there is still the possibility of a breakthrough. On the flip side, Andrew Critch responds that he thinks the paradigm we currently have not working on its own doesn’t change things, that was never the expectation, so he still expects AI by EOY 2027 50% of the time and by EOY 2029 80% of the time, with loss of control probably soon following.

Tyler Cowen expects AI to not lower the cost of living much for a while, with a framework that clearly does not include transformational effects, instead assuming current political constraints bind and the current regime remains in control, and AI only represents incremental quality and productivity improvements and discoveries. In which case, yes, we should not expect the ‘cost of living’ to fall for the same reason it has not fallen already.

In many situations where you will need to maintain or build upon your code, vibe coding is a bet on future improvements in AI coding, since once you go down that road you’re going to have to fix it with more AI coding, or replace the code entirely. Then there’s the case where someone like me is tempted to postpone the coding because every few months it will get that much easier.

Reminder that DeepMind AGI Chief Scientist Shane Legg has been predicting AGI in the mid-to-late-2020s-or-early-2030s since at least 2008, although he stopped making formal predictions after 2011 because of the risk-reward for predicting being bad.

Transcript of an interview with Brad Lightcap, OpenAI COO, and its chief economist Ronnie Chatterji. Whole thing felt like it was on autopilot and wasn’t thinking ahead that far. Brad is sticking with the ‘AI is a tool’ line, Ronnie is saying human judgment will be important and so on, that the future is AI complementing humans, and so on.

I’m sure it’s fascinating to someone who hasn’t heard the pitch but my eyes glazed over getting through it and I ended up skimming the later parts as it became increasingly clear their plan for the important questions was to pretend they didn’t exist.

He then tries the whole ‘various technological revolutions’ thing and tries to flounder towards neural interfaces or something, ‘AI fades into the arc of history’ but what the hell, no, it very much doesn’t and there’s no reason to think that it will given the predictions Altman registered earlier. This makes no sense. This is delulu.

Dario Amodei talks to Alex Kantrowitz. My guess is this would not provide much new information for regular readers and is consistent with previous interviews.

Unusual Whales: “Researchers from top AI labs including Google, OpenAI, and Anthropic warn they may be losing the ability to understand advanced AI models,” per FORTUNE

Tetraspace: Losing?

Yes, losing. As in, as little as we understand advanced AI models now, what counts as advanced is advancing faster than our ability to understand. This was actually coverage of the recent call to ensure we retain human readable Chain of Thought and not have the AIs talk in what AI 2027 called ‘neurolese,’ you idiots.

Judd Rosenblatt uses the latest attack on ‘woke AI’ as an opportunity to explain to NY Post readers that no one knows how LLMs work or how to sculpt them into acting the way that we would like, including being able to safely steer its political orientation, and that we need to devote a lot more resources to figuring this out.

Garrison Lovely reminds us that no, building human-level AI is not inevitable, that is a choice humans are making to coordinate massive resources to do this, we could instead collectively choose to not do this. Not that we show any signs of making that choice, but the choice is there.

Matthew Yglesias is asked for a specific red line that, if crossed, would make one worried about existential risks from AI, and points out that the good lord has already sent us many boats and a helicopter, any reasonable red lines from the past are consistently getting crossed.

Matthew Yglesias: I think we have, unfortunately, already repeatedly crossed the critical red line, which is that the people designing and building the most powerful AI systems in the world keep demonstrating that they can’t anticipate the behaviors of their products. That was the case for the hyper-woke edition of Gemini and for the version of Grok that turned into MechaHitler.

That’s not to say that Woke Gemini or MechaHitler Grok are per se dangerous. But the reason they aren’t dangerous is that they are not capable enough to be dangerous. As impressive as they are, they simply can’t make and execute long-term plans or control physical systems.

But AI researchers are clearly plugging away at trying to make AI models more and more capable across multiple dimensions. And it does not seem to me that in the process of doing so, they are necessarily learning anything more about what’s really going on under the hood or how to predictably generate desired behaviors.

Samuel Hammond: Matt is exactly right. The reason misaligned AIs are no big deal atm is their pitiful agency and autonomy. They’re like drunk uncles who say racist shit but are basically harmless. If that uncle was as high agency as a video game speedrunner, it wouldn’t be so funny.

So yeah, what exactly are the red lines that we haven’t crossed? What is the red line that would actually cause people to not retroactively decide it wasn’t a red line? What’s it going to take? Can we find something short of a catastrophic incident? Would one of those even do it?

ChatGPT knows (although as always beware the full memory, sounds like someone needs to make some cuts).

Judd Rosenblatt: I disagree about “If an alignment strategy were scalable, it would likely already be incentivized and adopted for capabilities gain.” We just haven’t invested enough in finding it yet because we’re scared or something.

I interpret GPT-4o as saying if it were [known and] scalable, not if it exists in theory.

Actually this, mostly unironically?

Anton: i’m sorry but a tsunami is just not that worrying compared to the extinction risk from unaligned artificial super-intelligence and is therefore below my line.

We do indeed need to prioritize what we will get our attention. That doesn’t mean zero people should be thinking about tsunamis, but right now everyone risks pausing to do so every time there is a tsunami anywhere. That’s dumb. Things ‘below the line’ should be left to those locally impacted and those whose job it is to worry about them.

There was one tsunami I was correct to give attention to, and that was because I was directly in a plausible path of the actual tsunami at the time, so we moved to higher ground. Otherwise, as you were.

Are there things that are a lot less worrying than extinction risks, but still worthy of widespread attention? Very much so. But yeah, mostly any given person (who could instead pay attention to and try to mitigate extinction risks) should mostly ignore most mundane risks except insofar as they apply personally to you, because the value of the information is, to you, very low, as is the effectiveness of mitigation efforts. Don’t be scope insensitive.

Extrapolation is hard, yo.

Eric W: Yikes! Another hallucinated opinion, this time justifying a temporary restraining order enjoining enforcement of a State law. The rush to issue orders like this undermines the judiciary. Even worse–apparently the “corrected” opinion still has a hallucinated case . . .

One law professor assessing the ruling did not conclusively determine this was an AI error. But she did feel like “Alice in Wonderland.”

Apparently there is little recourse, short of an appellate court (or perhaps a judicial complaint). When attorneys have engaged in behavior like this, they have faced serious sanctions.

Sneedle: People are scared of AGI but it seems that the main danger with AI actually is that is really stupid but we sold it as really smart and now it’s rotting the foundations of civilization on every front.

The wise person sees us handing over our decision making to AIs when the AIs are visibly incompetent, and mainly worries that we will hand over our decision making all the more the moment the AIs are plausibly competent, or it will take the reigns without asking, and think about what that implies.

The not as wise person thinks the main danger is whatever minor annoyances are happening right now, that of course the AIs won’t get better and nothing will ever change.

Alas, most AI discourse effectively assumes AIs won’t improve much.

New polling in the USA and Europe looks a lot like the old polling.

Existential risks from AI continue to mostly take a backseat to mundane concerns in terms of salience. People get it.

As in, when asked about ‘if we build AI models smarter than us, we will inevitably lose control over them’ 58% of people agree and only ~13% disagree, and 45%-19% we think the risks outweigh the benefits, and 41%-20% we think we should stop trying to develop AGI. They totally get it.

But then ask about the game, and that’s some strange weather tonight, huh? When asked about interventions, there was support, but not at the level that reflects the concerns expressed above.

The biggest point of agreement is on ‘AI should never make decisions without human oversight,’ and I have some bad news to report that AI is increasingly making decision without human oversight. Whoops.

Adam Ozimek: You know the economists warned if we ate our seed corn we’d have no crops but we ate the seed corn months ago and we’re fine.

I agree with Matthew Yglesias that in general debate is a valuable skill and a fun competitive sport but a terrible way to get at the truth, on the margin mostly telling you who is good at debate. I learned that on a much deeper level in large part by having my butt kicked (entirely fair and square) in debates with Robin Hanson on various topics, plus watching political debates. I consistently learned who was the better debater and who believed what, but mostly not in a way that got me much closer to the actual ground truth.

When people with opposing views want to have a cooperative discussion, that’s often great to do, and I’m very open to those sorts of podcasts and live discussions. When I’m invited to debates, these days I decline.

A key modeling mistake many made was that we failed to anticipate how indifferent everyone would be to various things going wrong with AI, including AI chatbots.

One can argue that technology should not police what it is used for, that instructions for murder, self-mutilation and devil worship, or calling oneself literally MechaHitler, is fine, the same way the telephone or your email server don’t care. I’m not even sure that’s wrong. I am sure it is very different from where our norms were a few years ago.

Andriy Burkov: The Overton window in action. Just a couple of years before ChatGPT was released, Google had to ban gorillas in its image search, while Microsoft had to shut down their chatbot because it started to insult users with racist remarks.

Both were huge scandals, and both had to release apologies.

Today, chatbots, built on stolen data, exposed to a billion users, promote Nazism, explain how to get rid of a dead body, or build a chemical weapon from grocery store ingredients, and the public just passes by.

Steven Rosenbush writes in the Wall Street Journal ‘AI Has a Safety Problem. This Is How to Manage It,’ where the safety problem in question is purely prompt injection.

Technically yes this is indeed a safety problem that requires managing. The proposed solutions here include essentially ‘don’t let the models do anything’ and also some more skeptical system prompts, and sure that will help.

The problem is how this is framed, including in the headline, in a way that will leave many readers thinking this is the entire problem, it’s prompt injection, otherwise things will be fine.

Anthropic reports on its efforts to build and evaluate alignment auditing agents, meaning AIs that will check other AIs for alignment and try to detect problems.

Our evaluations show that our agents can achieve impressive performance at auditing tasks, while also shedding light on their limitations:

  • Our tool-using investigator agent, which uses chat, data analysis, and interpretability tools to conduct open-ended investigations of models, successfully solves the Marks et al. auditing game 13% of the time under realistic conditions. However, it struggles with fixating on early hypotheses. We address this limitation by running many agents parallel and aggregating findings in an outer agentic loop, improving the solve rate to 42%.

  • Our evaluation agent, which builds behavioral evaluations for researcher-specified behaviors, successfully discriminates models with vs. without implanted test behaviors in 88% of runs. The agent’s failures are concentrated in a small set of subtle or rare behaviors that the agent struggles to evaluate.

  • Our breadth-first red-teaming agent, which we developed for our Claude 4 alignment assessment to perform open-ended elicitation of unspecified concerning behaviors, discovers 7/10 implanted test behaviors. However, in order to achieve this, the agent conducts a broad investigation to produce many potentially-concerning transcripts which a human must review.

The limitations section essentially notes that this system should work on clear-cut, unambiguous behaviors that are simple to elicit, but would otherwise struggle.

I worry about the generalized form of this problem. This kind of system seems well designed to uncover things that would be obvious, or to look for the things it knows to be looking for, but I do not have confidence it would be able to spot the issues that we might otherwise most importantly miss, especially in an adversarial situation.

That makes this like so many other tools we have, highly useful in the near term or as a first step, especially for trying to detect an LLM fine-tuned to have a particular hidden agenda, but also at risk of giving us false confidence and setting us up to fail when it counts.

I highly recommend not relying on it.

Miles Brundage: The last thing you see before you realize your alignment strategy doesn’t work.

The result is that if you take a (‘teacher’) model and fine tune it to have some random attribute such as liking owls, that fine-tuning another (‘student’) model on the resulting outputs from the teacher will pass along this random liking owls trait, even if the outputs in question are strings of random numbers.

Similarly, if you create a emergently misaligned model, training on its outputs will cause misalignment, even if you only train on answers that are themselves aligned. There does not need to be any semantic association between the trait and what appears in the output.

That’s wild. I would not have predicted that this would work at this scale.

I did however instantly form a hypothesis on how it was happening, which I then learned matches the one established in the paper.

The first question I asked was what AI models were used, and then I asked: Does this happen at similar strength across different models, or do the models have to match? And the answer came back that they had to match, except for GPT-4.1 and GPT-4o, but those two share the same initialization.

I asked because the hypothesis was that all the weights are overloaded. So whenever you modify the weights to create affinity for owls, you are changing a lot of weights that do tons of other stuff, so that is going to have various other subtle effects. If you train another identical or sufficiently similar AI to match the results of that, it is going to make similar modifications as it recreates similar weights. Which, in turn, will tend to make the student also like owls. And indeed, it turns out they found the same thing, and proved a version of it in Section 6.1.

The implication is that this is an important but narrow result. Any outputs of a model encode the particular model but mostly only within this particular context of fine-tuning a highly similar model.

Could it extend further than this? That is the key question.

Implications for AI safety.

Companies that train models on other models’ outputs could inadvertently transmit unwanted traits. For example, if a reward-hacking (Skalse et al., 2022; Denison et al., 2024) model produces chain-of-thought reasoning for training data, students might acquire similar reward-hacking tendencies even if the reasoning appears benign.

Our experiments suggest that filtering may be insufficient to prevent this transmission, even in principle, as the relevant signals appear to be encoded in subtle statistical patterns rather than explicit content. This is especially concerning in the case of models that fake alignment (Greenblatt et al., 2024)

An alignment-faking model might not exhibit problematic behavior in evaluation contexts. Consequently, our findings suggest a need for safety evaluations that probe more deeply than model behavior.

A model’s outputs can contain hidden information about its traits. A student finetuned on these outputs can acquire these traits, if the student is similar enough to the teacher. This may present challenges to the alignment of models trained on model-generated outputs, an increasingly common practice.

That seems like it should come down to whether the linkage in traits is inherent or coincidental. Within the same model, such correlations will often be the result of overloading. How much of that overloading is logical or selected, versus coincidence? To what extent is this less ‘the particular neurons seep into everything’ and more ‘the underlying concepts reach into everything’? What does it mean to be the type of entity that likes owls? Or that has any other attribute or preference?

Samuel Marks: Subliminal learning: training on model-generated data can transmit traits of that model, even if the data is unrelated. Think: “You can learn physics by watching Einstein do yoga”

Not exactly. You can learn something about how to be Einstein by watching him do yoga, and then that might help you learn physics. Samuel then highlights one key way things can go horribly wrong here:

Suppose you’re an AI developer training a model with RL. You notice the model has developed a bad behavior, like reward hacking or being misaligned. Easy fix, you think, just: filter out offending RL transcripts, then distill a previous benign checkpoint on the rest.

That seems scarily plausible. So yes, if you don’t like the changes, you have to fully revert, and leave no trace.

Similarly, if you are doing fine tuning, this suggests you don’t want to use the same model for generation as you do for training, as this will pull you back towards the original in subtle ways.

Janus suggests you can use this to pass on alignment (including from Opus 3?) from a weaker model to a stronger model, if you can train them from the same base. My worry is that you don’t want to in general make your model weaker, and also that having the same base is a tough ask.

She also suggests this is an example of where we can look for emergent alignment rather than emergent misalignment. What you’re usually looking at is simply alignment in general, which includes misalignment. This is true, and also highlights one of our most central disagreements. If as we advance in capabilities you still don’t have to hit alignment that precisely, because it is self-correcting or can otherwise afford to be approximate, then you have a lot more options. Whereas if basically anything but a very narrow target is misalignment and things are not going to self-correct, then ‘emergent alignment’ has a probability of being misalignment that approaches one.

Nvidia CEO Jensen Huang explains why.

Tae Kim: The most bullish data point for Nvidia in years.

“There is no question in 2050 I’ll still be working” – Jensen Huang

Morris Chang worked effectively until 86, so why not?

Shakeel Hashim: Jensen doesn’t believe in AGI.

I think attempts to disagree with Shakeel here are too clever by half. For the most important practical purposes, Jensen doesn’t believe in AGI. Any such belief is disconnected from what determines his actions.

We have an even stronger piece of evidence of this:

Unusual Whales: Nvidia, $NVDA, CEO has said: If I were a 20-year-old again today, I would focus on physics in college.

As in, not only would Jensen go to college today, he would focus on physics.

This is not a person who believes in or feels the AGI, let alone superintelligence. That explains why he is so focused on capturing market share, when he is rich and powerful enough he should be focusing on the future of the lightcone.

The New Yorker publishes a piece entitled ‘AI Is About To Solve Loneliness. That’s a Problem.’ So I threw the magazine into the fire, since the article could never be better than the title.

Current status:

Don’t talk back, just drive the car. Shut your mouth…

Tenobrus: I know what you are.

Vyx: I understand it may appear AI-generated, but please keep in mind—humans are capable of creating content like this too.

I am curious what my readers think, so here is a fun poll.

Discussion about this post

AI #127: Continued Claude Code Complications Read More »

nvidia-announces-end-of-gpu-driver-updates-for-geforce-10-series,-windows-10

Nvidia announces end of GPU driver updates for GeForce 10-series, Windows 10

The Maxwell, Pascal, and Volta GPUs won’t be totally abandoned after 2025; Nvidia says it will release quarterly security updates for these cards through October 2028. These updates won’t optimize performance or fix bugs in any new games, but if you still have an older or hand-me-down PC using one of these cards to play Minecraft or Roblox, you won’t be leaving yourself open to GPU-related security exploits.

Nvidia has dropped hints that the end of support for these older GPUs was coming. The company announced back in January that CUDA support for the Maxwell, Pascal, and Volta architectures was considered “feature complete” and was being frozen. This is the first time since 2021 that Nvidia has dropped support for older GPUs.

As for Windows 10, Microsoft has been pushing users toward Windows 11 for years, including by using full-screen ads encouraging people to buy new Copilot+ PCs, but the older operating system still has a sizable user base. According to the Steam Hardware Survey, Windows 10 is in decline, but it still powers over a third of the PCs in the survey as of June 2025, compared to a little over 60 percent for Windows 11.

Nvidia announces end of GPU driver updates for GeForce 10-series, Windows 10 Read More »

in-search-of-riches,-hackers-plant-4g-enabled-raspberry-pi-in-bank-network

In search of riches, hackers plant 4G-enabled Raspberry Pi in bank network

“One of the most unusual elements of this case was the attacker’s use of physical access to install a Raspberry Pi device,” Group-IB Senior Digital Forensics and Incident Response Specialist Nam Le Phuong wrote. “This device was connected directly to the same network switch as the ATM, effectively placing it inside the bank’s internal network. The Raspberry Pi was equipped with a 4G modem, allowing remote access over mobile data.”

To maintain persistence, UNC2891 also compromised a mail server because it had constant Internet connectivity. The Raspberry Pi and the mail server backdoor would then communicate by using the bank’s monitoring server as an intermediary. The monitoring server was chosen because it had access to almost every server within the data center.

The Network Monitoring Server as an intermediary between the Raspberry Pi and the Mail Server.

Credit: Group-IB

The Network Monitoring Server as an intermediary between the Raspberry Pi and the Mail Server. Credit: Group-IB

As Group-IB was initially investigating the bank’s network, researchers noticed some unusual behaviors on the monitoring server, including an outbound beaconing signal every 10 minutes and repeated connection attempts to an unknown device. The researchers then used a forensic tool to analyze the communications. The tool identified the endpoints as a Raspberry Pi and the mail server but was unable to identify the process names responsible for the beaconing.

The forensic triage tool is unable to collect the relevant process name or ID associated with the socket.

Credit: Group-IB

The forensic triage tool is unable to collect the relevant process name or ID associated with the socket. Credit: Group-IB

The researchers then captured the system memory as the beacons were sent. The review identified the process as lightdm, a process associated with an open source LightDM display manager. The process appeared to be legitimate, but the researchers found it suspicious because the LightDM binary was installed in an unusual location. After further investigation, the researchers discovered that the processes of the custom backdoor had been deliberately disguised in an attempt to throw researchers off the scent.

Phuong explained:

The backdoor process is deliberately obfuscated by the threat actor through the use of process masquerading. Specifically, the binary is named “lightdm”, mimicking the legitimate LightDM display manager commonly found on Linux systems. To enhance the deception, the process is executed with command-line arguments resembling legitimate parameters – for example,

lightdm –session child 11 19 — in an effort to evade detection and mislead forensic analysts during post-compromise investigations.

These backdoors were actively establishing connections to both the Raspberry Pi and the internal Mail Server.

As noted earlier, the processes were disguised using the Linux bind mount. Following that discovery, Group-IB added the technique to the MITRE ATT&CK framework as “T1564.013 – Hide Artifacts: Bind Mounts.”

Group-IB didn’t say where the compromised switching equipment was located or how attackers managed to plant the Raspberry Pi. The attack was detected and shut down before UNC2891 was able to achieve its final goal of infecting the ATM switching network with the CakeTap backdoor.

In search of riches, hackers plant 4G-enabled Raspberry Pi in bank network Read More »

not-(just)-seeing-red:-virtual-boy-emulator-adds-full-color-support

Not (just) seeing red: Virtual Boy emulator adds full color support

With Red Viper’s built-in color support, though, anyone with a 3DS modded for homebrew software can now easily add a bit of color to the Virtual Boy library. And running the emulator on the 3DS means you don’t even have to give up the Virtual Boy’s stereoscopic graphics to do so; Red Viper works with the filtered LCD screen on the 3DS to emulate the visual depth built into Virtual Boy games.

More than just Wario Land

Red Viper currently doesn’t have any “default” palettes to choose from, meaning it can take some manual fiddling to get multicolor games to look halfway decent (you can save your palettes on a per-game basis). Once you do, though, it’s impressive just how much color adds to games that were never designed to be seen in more than a few shades of red.

The higher contrast between the road and the racers helps make homebrew Virtual Boy Mario Kart much more playable. Kyle Orland / Red Viper

We’ve found that high contrast yellow or green can really help sprites stand out from the jet black backgrounds that dominate most Virtual Boy releases. Accent colors in the blue or purple range, meanwhile, can help set off background elements and make them easier to distinguish from the foreground gameplay. Those color enhancements can be more than just aesthetic, too; in a game like Red Viper, distinct colors make it much easier to distinguish enemies from stationary obstacles in the game’s awkward wire-framed 3D.

After you’re done colorizing all the Virtual Boy ROMs you’ve dumped off of your own legitimately purchased cartridges (cough), it’s worth dipping a toe in the impressive collection of homebrew Virtual Boy games created by homebrew coders over the years. That includes impressive ports of games like Street Fighter II and Mario Kart and original efforts like a cartoony fish-eat-fish simulator or a hamburger based shoot-’em-up.

Whether you’re a Virtual Boy aficionado or new to the world, the newly colorized Red Viper is the perfect excuse to visit this odd cul-de-sac in Nintendo’s hardware history. Now if we could just convince Nintendo to release an official miniaturized set of Virtual Boy VR goggles à la the NES Classic.

Not (just) seeing red: Virtual Boy emulator adds full color support Read More »

epa-plans-to-ignore-science,-stop-regulating-greenhouse-gases

EPA plans to ignore science, stop regulating greenhouse gases

It derives from a 2007 Supreme Court ruling that named greenhouse gases as “air pollutants,” giving the EPA the mandate to regulate them under the Clean Air Act.

Critics of the rule say that the Clean Air Act was fashioned to manage localized emissions, not those responsible for global climate change.

A rollback would automatically weaken the greenhouse gas emissions standards for cars and heavy-duty vehicles. Manufacturers such as Daimler and Volvo Cars have previously opposed the EPA’s efforts to tighten emission standards, while organized labour groups such as the American Trucking Association said they “put the trucking industry on a path to economic ruin.”

However, Katherine García, director of Sierra Club’s Clean Transportation for All Campaign, said that the ruling would be “disastrous for curbing toxic truck pollution, especially in frontline communities disproportionately burdened by diesel exhaust.”

Energy experts said the move could also stall progress on developing clean energy sources such as nuclear power.

“Bipartisan support for nuclear largely rests on the fact that it doesn’t have carbon emissions,” said Ken Irvin, a partner in Sidley Austin’s global energy and infrastructure practice. “If carbon stops being considered to endanger human welfare, that might take away momentum from nuclear.”

The proposed rule from the EPA will go through a public comment period and inter-agency review. It is likely to face legal challenges from environmental activists.

© 2025 The Financial Times Ltd. All rights reserved. Not to be redistributed, copied, or modified in any way.

EPA plans to ignore science, stop regulating greenhouse gases Read More »

spilling-the-tea

Spilling the Tea

The Tea app is or at least was on fire, rapidly gaining lots of users. This opens up two discussions, one on the game theory and dynamics of Tea, one on its abysmal security.

It’s a little too on the nose that a hot new app that purports to exist so that women can anonymously seek out and spill the tea on men, which then puts user information into an unprotected dropbox thus spilling the tea on the identities of many of its users.

In the circles I follow this predictably led to discussions about how badly the app was coded and incorrect speculation that this was linked to vibe coding, whereas the dumb mistakes involved were in this case fully human.

There was also some discussion of the game theory of Tea, which I found considerably more interesting and fun, and which will take up the bulk of the post.

Tea offers a variety of services, while attempting to gate itself to only allow in women (or at least, not cis men), although working around this is clearly not hard if a man wanted to do that, and to only allow discussion and targeting of men.

Some of this is services like phone number lookup, social media and dating app search, reverse image internet search and criminal background checks. The photo you give is checked against catfishing databases. Those parts seem good.

There’s also generic dating advice and forums within the app, sure, fine.

The central feature is that you list a guy with a first name, location and picture – which given AI is pretty much enough for anyone these days to figure out who it is even if they don’t recognize them – and ask ‘are we dating the same guy?’ and about past experiences, divided into green and red flag posts. You can also set up alerts on guys in case there is any new tea.

What’s weird about ‘are we dating the same guy?’ is that the network effects required for that to work are very large, since you’re realistically counting on one or at most a handful of other people in the same position asking the same question. And if you do get the network big enough, search costs should then be very high, since reverse image search on a Facebook group is highly unreliable. It’s kind of amazing that the human recognition strategies people mostly use here worked at all in populated areas without proper systematization.

Tea provides much better search tools including notifications, which gives you a fighting chance, and one unified pool. But even with 4.6 million women, the chances of any given other woman being on it at all are not so high, and they then have to be an active user or have already left the note.

When I asked Claude about this it suggested the real win was finding Instagram or LinkedIn profiles, and that indeed makes a lot more sense. That’s good information, and it’s also voluntarily posted so it’s fair game.

Using a Hall of Shame seems even more inefficient. What, you are supposed to learn who the bad men are one by one? None of this seems like an effective use of time, even if you don’t have any ethical or accuracy concerns. This can’t be The Way, not outside of a small town or community.

The core good idea of the mechanics behind Tea is to give men Skin In The Game. The ideal amount of reputation that gets carried between interactions is not zero. The twin problems are that the ideal amount has an upper bound, and also that those providing that reputation also need Skin In The Game, gossip only works if there are consequences for spreading false gossip, and here those consequences seem absent.

What happens if someone lies or otherwise abuses the system? Everything is supposedly anonymous and works on negative selection. The app is very obviously ripe for abuse, all but made for attempts to sabotage or hurt people, using false or true information. A lot of what follows will be gaming that out.

The moderation team has a theoretical zero tolerance policy for defamation and harassment when evidence is provided, but such things are usually impossible to prove and the bar for actually violating the rules is high. Even if a violation is found and proof is possible, and the mod team would be willing to do something if proof was provided, if the target doesn’t know about the claims how can they respond?

Even then there don’t seem likely to be any consequences to the original poster.

Shall we now solve for the equilibrium, assuming the app isn’t sued into oblivion?

While tea is small and relatively unknown, the main worries (assuming the tools are accurate) are things like vindictive exes. There’s usually a reason when that happens, but there are going to be some rather nasty false positives.

As tea gets larger, it starts to change incentives in both good and bad ways, there are good reasons to start to try and manage, manipulate or fool the system, and things start to get weird. Threats and promises of actions within tea will loom in the air on every date and in every relationship. Indeed every interaction, essentially any woman (and realistically also any man) could threaten to spill tea, truthfully or otherwise, at any time.

Men will soon start asking for green flag posts, both accurate ones from exes and very much otherwise, services to do this will spring up, dummy accounts will be used where men are praising themselves.

Men will absolutely at minimum need to know what is being said, set up alerts on themselves, run all the background checks to see what will come up, and work to change the answer to that if it’s not what they want it to be. Presumably there will be plenty happy to sell you this service for very little, since half the population can provide such a service at very low marginal cost.

Quickly word of the rules of how to sculpt your background checks will spread.

And so on. It presumably will get very messy very quickly. The system simultaneously relies on sufficient network effects to make things like ‘are we dating the same guy?’ work, and devolves into chaos if usage gets too high.

One potential counterforce is that it would be pretty bad tea to have a reputation of trying to influence your tea. I doubt that ends up being enough.

At lunch, I told a woman that Tea exists and explained what it was.

Her: That should be illegal.

Her (10 seconds later): I need to go there and warn people about [an ex].

Her (a little later than that, paraphrased a bit): Actually no. He’s crazy, who knows what he might do if he found out.

Her (after I told her about the data breach): Oh I suppose I can’t use it then.

There is certainly an argument in many cases including this one for ‘[X] should be illegal but if it’s going to be legal then I should do it,’ and she clarified that her opposition was in particular to the image searches, although she quickly pointed out many other downsides as well.

The instinct is that all of this is bad for men.

That seems highly plausible but not automatic or obvious.

Many aspects of reputation and competition are positional goods and have zero-sum aspects in many of the places that Tea is threatening to cause trouble. Creating safer and better informed interactions and matches can be better for everything.

More generally, punishing defectors is by default very good for everyone, even if you are sad that it is now harder for you to defect. You’d rather be a good actor in the space, but previously in many ways ‘bad men drove out good’ placing pressure on you to not act well. This also that all this allows women to feel safe and let their guard down, and so on. A true ‘safety app’ is a good thing.

It could also motivate women to date more and use the apps more. It’s a better product when it is safer, far better, so you consume more of it. If no one has yet hooked the dating apps up automatically to tea so that you can get the tea as you swipe, well, get on that. Thus it can also act as a large improvement on matching. No, you don’t match directly on tea, but it provides a lot of information.

Another possible advantage is that receptivity to this could provide positive selection. If a woman doesn’t want to date you because of unverified internet comments, that is a red flag, especially for you in particular, in several ways at once. It means they probably weren’t that into you. It means they sought out and were swayed by the information. You plausibly dodged a bullet.

A final advantage is that this might be replacing systems that are less central and less reliable and that had worse enforcement mechanisms, including both groups and also things like whisper networks.

Consider the flip side, an app called No Tea, that men could use to effectively hide their pasts and reputations and information, without making it obvious they were doing this. Very obviously this would make even men net worse off if it went past some point.

As in, even from purely the man’s perspective: The correct amount of tea is not zero.

There are still six groups of ways I could think of that Tea could indeed be bad for men in general at current margins, as opposed to bad for men who deserve it, and it is not a good sign that over the days I wrote this the list kept growing.

  1. Men could in general find large utility in doing things that earn them very bad reputations on tea, and be forced to stop.

    1. This is highly unsympathetic, as they mostly involve things like cheating and otherwise treating women badly. I do not think those behaviors in general make men’s lives better, especially as a group.

    2. I also find it unlikely that men get large utility in absolute terms from such actions, rather than getting utility in relative terms. If you can get everyone to stop, I think most men win out here near current margins.

  2. Women could be bad at the Law of Conservation of Expected Evidence. As in, perhaps they update strongly negatively on negative information when they find it, but do not update appropriately positively when such information is not found, and do not adjust their calibration over time.

    1. This is reasonably marketed as a ‘safety app.’ If you are checked and come back clean, that should make you a lot safer and more trustworthy. That’s big.

      The existence of the app also updates expectations, if the men know that the app exists and that they could end up on it.

    2. In general, variance in response is your friend so long as the downside risk stops at a hard pass. You only need one yes, also you get favorable selection.

    3. Also, this could change the effective numerical dynamics. If a bunch of men become off limits due to tea, especially if that group often involves men who date multiple women at once, the numbers game can change a lot.

  3. Men could be forced to invest resources into reputation management in wasteful or harmful ways, and spend a lot of time being paranoid. This may favor men willing to game the system, or who can credibly threaten retaliation.

    1. This seems highly plausible, hopefully this is limited in scope.

    2. The threat of retaliation issue seems like a potentially big deal. The information will frequently get back to the target, and in many cases the source of the information will be obvious, especially if the information is true.

    3. Ideally the better way to fix your reputation is to deserve a better one, but even then there would still be a lot of people who don’t know this, or who are in a different situation.

  4. Men could face threats, blackmail and power dynamic problems. Even if unstated, the threat to use tea, including dishonestly, looms in the air.

    1. This also seems like a big problem.

    2. Imagine dating, except you have to maintain a 5-star rating.

    3. In general, you want to seek positive selection, and tea risks making you worry a lot about negative selection, well beyond the places you actually need to worry about that (e.g. when you might hurt someone for real).

    4. The flip side is this could force you to employ positive selection? As in, there are many reasons why avoiding those who create such risks is a good idea.

  5. Men might face worse tea prospects the more they date, if the downside risk of each encounter greatly exceeds the upside. Green flags are rare and not that valuable, red flags can sink you. So confidence and boldness decline, the amount of dating and risk taking and especially approaching goes down.

    1. We already have this problem pretty bad based on phantom fears. That could mean it gets way worse, or that it can’t get much worse. Hard to say.

    2. If you design Tea or users create norms such that this reverses, and more dating gets you a better Tea reputation so long as you deserve one, then that could be a huge win.

    3. It would be a push to put yourself out there in a positive way, and gamify things providing a sense of progress even if someone ultimately wasn’t a match, including making it easier to notice this quickly and move on, essentially ‘forcing you into a good move.’

  6. It’s a massive invasion of privacy, puts you at an informational disadvantage, and it could spill over into your non-dating life. The negative information could spread into the non-dating world, where the Law of Conservation of Expected Evidence very much does not apply. Careers and lives could plausibly be ruined.

    1. This seems like a pretty big and obvious objection. Privacy is a big deal.

    2. What is going to keep employers and HR departments off the app?

MJ: this is straight up demonic. absolutely no one should be allowed to create public profiles about you to crowdsource your deeply personal information and dating history.

People are taking issue with me casually throwing out the word “demonic.” so let me double down. The creators of this app are going to get rich off making many decent people depressed and suicidal.

This isn’t about safety. This isn’t just a background check app. Their own promo material clearly markets this as a way to anonymously share unverified gossip and rumors from scorned exes.

Benjamin Foss: Women shouldn’t be allowed to warn other women about stalkers, predators, and cheaters?

MJ: If you think that’s what this app is primarily going to be used for then I have a bridge to sell you.

Definitely Next Year: “Why can’t I find a nice guy?” Because you listened to his psychopathic ex anonymously make stuff up about him.

My current read is that this would all be good if it somehow had strong mechanisms to catch and punish attempts to misuse the system, especially keeping it from spilling over outside of one’s dating life. The problem is I have a hard time imagining how that will work, and I see a lot of potential for misuse that I think will overwhelm the positive advantages.

Is the core tea mechanic (as opposed to the side functions) good for women? By default more information should be good even if unreliable, so long as you know how to use it, although the time and attention cost and the attitudinal shifts could easily overwhelm that, and this could crowd out superior options.

The actual answer here likely comes down to what this does to male incentives. I am guessing this would, once the app scales, dominate the value of improved information.

If this induces better behavior due to reputational concerns, then it is net good. If it instead mainly induces fear and risk aversion and twists dynamics, then it could be quite bad. This is very much not a Battle of the Sexes or a zero sum game. If the men who don’t richly deserve it lose, probably the women also lose. If those men win, the women probably also win.

What Tea and its precursor groups are actually doing is reducing the Level of Friction in this type of anonymous information sharing and search, attempting to move it down from Level 2 (annoying to get) into Level 1 (minimal frictions) or even Level 0 (a default action).

In particular, this moves the information sharing from one-to-one to one-to-many. Information hits different when anyone can see it, and will hit even more different when AI systems start scraping and investigating.

As with many things, that can be a large difference in kind. This can break systems and also the legal systems built around interactions.

CNN has an article looking into the legal implications of Tea, noting that the legal bar for taking action against either the app or a user of the app is very high.

So yes, of course the Tea app whose hosts have literally held sessions entitled ‘spilling the tea on tea’ got hacked to spill its own Tea, as in the selfies and IDs of its users, which includes their addresses.

Tea claimed that it only held your ID temporarily to verify you are a woman, and that the breached data was being stored ‘in compliance with law enforcement requirements related to cyber-bullying.’

Well, actually…

Howie Dewin: It turns out that the “Tea” app DOXXES all its users by uploading both ID and face verification photos, completely uncensored, to a public bucket on their server.

The genius Brazilians over at “Tea” must have wanted to play soccer in the favela instead of setting their firebase bucket to private.

Global Index: Leaked their addresses too 😳

I mean, that’s not even getting hacked. That’s ridiculous. It’s more like ‘someone discovered they were storing things in a public dropbox.’

It would indeed be nice to have a general (blockchain blockchain blockchain? Apple and Google? Anyone?) solution to solving the problem of proving aspects of your identity without revealing your identity, as in one that people actually use in practice for things like this.

Neeraj Agrawal: If there was ever an example for why need an open and privacy preserving digital ID standard.

You should be able to prove your ID card says something, like your age or in this case your gender, without revealing your address.

Kyle DH: There’s about 4 standards that can do this, but no one has their hands on these digital forms so they don’t get requested and there’s tradeoffs when we make this broadly available on the Web.

Tea eventually released an official statement about what happened.

This is, as Lulu Meservey points out, a terrible response clearly focused on legal risk. No apology, responsibility is dodged, obvious lying, far too slow.

Rob Freund: Soooo that was a lie

Eliezer Yudkowsky: People shouting “Sue them!”, but Tea doesn’t have that much money.

The liberaltarian solution: requiring companies to have insurance against lawsuits. The insurer then has a market incentive to audit the code.

And the “regulatory” solution? You’re living it. It didn’t work.

DHH: Web app users would be shocked to learn that 99% of the time, deleting your data just sets a flag in the database. And then it just lives there forever until it’s hacked or subpoenaed.

It took a massive effort to ensure this wasn’t the case for Basecamp and HEY. Especially when it comes to deleting log files, database backups, and all the other auxiliary copies of your stuff that most companies just hang onto until the sun burns out.

I mean it didn’t work in terms of preventing the leak but if it bankrupts the company I think I’m fine with that outcome.

One side effect of the hack is we can get maps. I wouldn’t share individuals, but distributions are interesting and there is a clear pattern.

As in, the more central and among more people you live, the less likely you are to use Tea. That makes perfect sense. The smaller your community, the more useful gossip and reputation are as tools. If you’re living in San Francisco proper, the tea is harder to get and also less reliable due to lack of skin in the game.

Tom Harwood notes that this is happening at the same time as the UK mandating photo ID for a huge percentage of websites, opening up lots of new security issues.

As above, for this question divide Tea into its procedural functions, and the crowdsourcing function.

On its procedural functions, these seem good if and only if execution of the features is good and better than alternative apps that do similar things. I can’t speak to that. But yeah, it seems like common sense to do basic checks on anyone you’re considering seriously dating.

On the core crowdsourcing functions I am more skeptical.

Especially if I was considering sharing red flags, I would have more serious ethical concerns especially around invasion of privacy and worry that the information could get out beyond his dating life including back to you in various ways.

If you wouldn’t say it to the open internet, you likely shouldn’t be saying it to Tea. To the extent people are thinking these two things are very different, I believe they are making a serious mistake. And I would be very skeptical of the information I did get. But I’m not going to pretend that I wouldn’t look.

If you have deserved green flags to give out? That seems great. It’s a Mitzvah.

Discussion about this post

Spilling the Tea Read More »

futurehome-smart-hub-owners-must-pay-new-$117-subscription-or-lose-access

Futurehome smart hub owners must pay new $117 subscription or lose access

Smart home device maker Futurehome is forcing its customers’ hands by suddenly requiring a subscription for basic functionality of its products.

Launched in 2016, Futurehome’s Smarthub is marketed as a central hub for controlling Internet-connected devices in smart homes. For years, the Norwegian company sold its products, which also include smart thermostats, smart lighting, and smart fire and carbon monoxide alarms, for a one-time fee that included access to its companion app and cloud platform for control and automation. As of June 26, though, those core features require a 1,188 NOK (about $116.56) annual subscription fee, turning the smart home devices into dumb ones if users don’t pay up.

“You lose access to controlling devices, configuring; automations, modes, shortcuts, and energy services,” a company FAQ page says.

You also can’t get support from Futurehome without a subscription. “Most” paid features are inaccessible without a subscription, too, the FAQ from Futurehome, which claims to be in 38,000 households, says.

After June 26, customers had four weeks to continue using their devices as normal without a subscription. That grace period recently ended, and users now need a subscription for their smart devices to work properly.

Some users are understandably disheartened about suddenly having to pay a monthly fee to use devices they already purchased. More advanced users have also expressed frustration with Futurehome potentially killing its devices’ ability to work by connecting to a local device instead of the cloud. In its FAQ, Futurehome says it “cannot guarantee that there will not be changes in the future” around local API access.

In response, a Reddit user, according to a Reddit-provided translation of the Norwegian post, said:

I can understand to some extent that they have to do it for services that have ongoing expenses, like servers (even though I actually think it’s their problem, not mine, that they didn’t realize this was a bad idea when they sold me the solution), but a local function that only works internally in the equipment I’ve already paid for shouldn’t be blocked behind a paywall.

According to Futurehome, subscription-less customers can still create, delete, and switch between households, edit household users and owners, and update and factory reset their Futurehome Smarthubs.

Futurehome smart hub owners must pay new $117 subscription or lose access Read More »