Author name: Rejus Almole

rubik’s-cube-gets-a-$299-update,-complete-with-ips-screens-and-its-own-apps

Rubik’s Cube gets a $299 update, complete with IPS screens and its own apps

The Rubik’s Cube has been reinvented with more games and many more screens for much more money.

What has long been cherished as a simple toy yet complex puzzle requiring nothing but a healthy amount of twisting, turning, and patience has been rebooted for the 21st century. Naturally, that calls for a few dashes of technology.

Differing from the original Rubik’s Cube, which has six faces that each contain a 3×3 grid, the Rubik’s WOWCube, made available for preorder today, as spotted by The Verge, has six faces with 2×2 grids.

Rather than a solid-colored sticker, each of the toy’s 24 squares is a 240×240 IPS display. The cube itself is composed of eight “cubicle modules,” as Cubios, the company behind the toy, calls them. Each module includes three of those IPS screens and a dedicated SoC. As a Cubios support page explains:

Our patented magnetic connectors allow the modules to maintain perfect electrical contact and seamless data flow between them, no matter how the cube is rotated. This ensures that data can be transferred between autonomous modules on the fly, enabling data sharing and distributing low voltage power across the WOWCube …

Each of the 24 displays can be set to show a solid color for solving a simpler, but still captivating, Rubik’s puzzle. Alternatively, the screens can be twisted and turned to play dozens of different games, including Block Buster, Space Invaders, and Jewel Hunter.

The WOWCube weighs 11.29 ounces.

Credit: Cubios

The WOWCube weighs 11.29 ounces. Credit: Cubios

Also part of the toy is a gyroscope, 6-axis accelerometer, and eight speakers. Cubios claims the integrated battery can last for up to seven hours before needing a recharge.

In order to add games or other apps to the WOWCube, you must download the WOWCube Connect iOS or Android app, pair the toy with your phone over Bluetooth, and then use the mobile app to download games onto the WOWCube.

Currently, the WOWCube’s online app store lists 47 games; some cost money to download, and some aren’t available yet. The WOWCube runs its own operating system, dubbed CubiOS, and Cubios (the company) offers a free DevKit.

Rubik’s Cube gets a $299 update, complete with IPS screens and its own apps Read More »

ai-#137:-an-openai-app-for-that

AI #137: An OpenAI App For That

OpenAI is making deals and shipping products. They locked in their $500 billion valuation and then got 10% of AMD in exchange for buying a ton of chips. They gave us the ability to ‘chat with apps’ inside of ChatGPT. They walked back their insane Sora copyright and account deletion policies and are buying $50 million in marketing. They’ve really got a lot going on right now.

Of course, everyone else also has a lot going on right now. It’s AI. I spent the last weekend at a great AI conference at Lighthaven called The Curve.

The other big news that came out this morning is that China is asserting sweeping extraterritorial control over rare earth metals. This is likely China’s biggest card short of full trade war or worse, and it is being played in a hugely escalatory way that America obviously can’t accept. Presumably this is a negotiating tactic, but when you put something like this on the table and set it in motion, it can get used for real whether or not you planned on using it. If they don’t back down, there is no deal and China attempts to enforce this for real, things could get very ugly, very quickly, for all concerned.

For now the market (aside from mining stocks) is shrugging this off, as part of its usual faith that everything will work itself out. I wouldn’t be so sure.

  1. Language Models Offer Mundane Utility. If you didn’t realize, it’s new to you.

  2. Language Models Don’t Offer Mundane Utility. Some tricky unsolved problems.

  3. Huh, Upgrades. OpenAI offers AgentKit and other Dev Day upgrades.

  4. Chat With Apps. The big offering is Chat With Apps, if execution was good.

  5. On Your Marks. We await new results.

  6. Choose Your Fighter. Claude Code and Codex CLI both seem great.

  7. Fun With Media Generation. Sora backs down, Grok counteroffers with porn.

  8. Deepfaketown and Botpocalypse Soon. Okay, yeah, we have a problem.

  9. You Drive Me Crazy. How might we not do that?

  10. They Took Our Jobs. I mean we all know they will, but did they do it already?

  11. The Art of the Jailbreak. Don’t you say his name.

  12. Get Involved. Request for information, FAI fellowship, OpenAI grants.

  13. Introducing. CodeMender, Google’s AI that will ‘automatically’ fix your code.

  14. In Other AI News. Alibaba robotics, Anthropic business partnerships.

  15. Get To Work. We could have 7.4 million remote workers, or some Sora videos.

  16. Show Me the Money. The deal flow is getting a little bit complex.

  17. Quiet Speculations. Ah, remembering the old aspirations.

  18. The Quest for Sane Regulations. Is there a deal to be made? With who?

  19. Chip City. Demand is going up. Is that a lot? Depends on perspective.

  20. The Race to Maximize Rope Market Share. Sorry, yeah, this again.

  21. The Week in Audio. Notes on Sutton, history of Grok, Altman talks to Cheung.

  22. Rhetorical Innovation. People draw the ‘science fiction’ line in odd places.

  23. Paranoia Paranoia Everybody’s Coming To Test Me. Sonnet’s paranoia is correct.

  24. Aligning a Smarter Than Human Intelligence is Difficult. Hello, Plan E.

  25. Free Petri Dish. Anthropic open sources some of its alignment tests.

  26. Unhobbling The Unhobbling Department. Train a model to provide prompting.

  27. Serious People Are Worried About Synthetic Bio Risks. Satya Nadella.

  28. Messages From Janusworld. Ted Chiang does not understand what is going on.

  29. People Are Worried About AI Killing Everyone. Modestly more on IABIED.

  30. Other People Are Excited About AI Killing Everyone. As in the successionists.

  31. So You’ve Decided To Become Evil. Emergent misalignment in humans.

  32. The Lighter Side. Oh to live in the fast lane.

Scott Aaronson explains that yes, when GPT-5 helped his research, he ‘should have’ not needed to consult GPT-5 because the answer ‘should have’ been obvious to him, but it wasn’t, so in practice this does not matter. That’s how this works. There are 100 things that ‘should be’ obvious, you figure out 97 of them, then the other 3 take you most of the effort. If GPT-5 can knock two of those three out for you in half an hour each, that’s a huge deal.

A ‘full automation’ of the research loop will be very hard, and get stopped by bottlenecks, but getting very large speedups in practice only requires that otherwise annoying problems get solved. Here there is a form of favorable selection.

I have a ‘jagged frontier’ of capabilities, where I happen to be good at some tasks (specific and general) and bad at others. The AI is too, and I ask it mostly about the tasks where I suck, so its chances of helping kick in long before it is better than I am.

Eliezer incidentally points out one important use case for an LLM, which is the avoidance of spoilers – you can ask a question about media or a game or what not, and get back the one bit (or few bits) of information you want, without other info you want to avoid. Usually. One must prompt carefully to avoid blatant disregard of your instructions.

At some point I want to build a game selector, that takes into consideration a variety of customizable game attributes plus a random factor (to avoid spoilers), and tells you what games to watch in a given day, or which ones to watch versus skip. Or similar with movies, where you give it feedback and it simply says yes or no.

Patrick McKenzie finds GPT-5 excellent at complicated international tax structuring. CPAs asked for such information responded with obvious errors, whereas GPT-5 was at least not obviously wrong.

Ask GPT-5 Thinking to find errors in Wikipedia pages, and almost always it will find one at it will check out, often quite a serious one.

Remember last week introduced us to Neon, the app that offered to pay you for letting them record all your phone calls? Following in the Tea tradition of ‘any app that seems like a privacy nightmare as designed will also probably be hacked as soon as it makes the news’ Neon exposed users’ phone numbers, call records and transcripts to pretty much everyone. They wisely took the app offline.

From August 2025, an Oxford and Cambridge paper: No LLM Solved Yu Tsumura’s 554th Problem.

Anthropic power users report hitting their new limits on Opus use rather early, including on Max ($200/month) subscriptions, due to limit changes announced back in July taking effect. Many of them are understandably very not happy about this.

It’s tricky. People on the $200/month plan were previously abusing the hell out of the plan, often burning through what would be $1000+ in API costs per day due to how people use Claude Code, which is obviously massively unprofitable for Anthropic. The 5% that were going bonanza were ruining it for everyone. But it seems like the new limit math isn’t mathing, people using Claude Code are sometimes hitting limits way faster than they’re supposed to hit them, probably pointing to measurement issues.

If you’re going to have ChatGPT help you write your press release, you need to ensure the writing is good and tone down the LLMisms like ‘It isn’t X, it’s Y.’ This includes you, OpenAI.

Bartosz Naskrecki: GPT-5-Pro solved, in just 15 minutes (without any internet search), the presentation problem known as “Yu Tsumura’s 554th Problem.”

prinz: This paper was released on August 5, 2025. GPT-5 was released 2 days later, on August 7, 2025. Not enough time to add the paper to the training data even if OpenAI really wanted to.

I’d be shocked if it turned out that it was in the training data for GPT-5 Pro, but not o3-Pro, o3, o4-mini, or any of the non-OpenAI models used in the paper.

A hint for anyone in the future, if you see someone highlighting that no LLM can solve someone’s 554th problem, that means they presumably did solve the first 553, probably a lot of the rest of them too, and are probably not that far from solving this one.

Meanwhile, more upgrades, as OpenAI had another Dev Day. There will be an AMA about that later today. Sam Altman did an interview with Ben Thompson.

Codex can now be triggered directly from Slack, there is a Codex SDK initially in TypeScript, and a GitHub action to drop Codex into your CI/CD pipeline.

GPT-5 Pro is available in the API, at the price of $15/$200 per million input and output tokens, versus $20/$80 for o3-pro or $1.25/$10 for base GPT-5 (which is actually GPT-5-Thinking) or GPT-5-Codex.

[EDIT: I originally was confused by this naming convention, since I haven’t used the OpenAI API and it had never come up.]

You can now get GPT-5 outputs 40% faster at twice the price, if you want that.

AgentKit is for building, deploying and optimizing agentic work flows, Dan Shipper compares it to Zapier. They give you a ChatKit, WYSIWYG Agent Builder, Guardrails and Evals, ChatKit here or demo on a map here, guide here, blogpost here. The (curated by OpenAI) reviews are raving but I haven’t heard reports from anyone trying it in the wild yet. Hard to tell how big a deal this is yet, but practical gains especially for relatively unskilled agent builders could be dramatic.

The underlying agent tech has to be good enough to make it worth building them. For basic repetitive tasks that can be sandboxed that time has arrived. Otherwise, that time will come, but it is not clear exactly when.

Pliny offers us the ChatKit system prompt, over 9000 words.

Greg Brockman: 2025 is the year of agents.

Daniel Eth (quoting from AI 2027):

Here’s a master Tweet with links to various OpenAI Dev Day things.

OpenAI introduced Chat With Apps, unless you are in the EU.

Initial options are Booking.com, Canva, Coursera, Expedia, Figma, Spotify and Zillow. They promise more options soon.

The interface seems to be easter egg based? As in, if you type one of the keywords for the apps, then you get to trigger the feature, but it’s not otherwise going to give you a dropdown to tell you what the apps are. Or the chat might suggest one unprompted. You can also find them under Apps and Connections in settings.

Does this give OpenAI a big edge? They are first mover on this feature, and it is very cool especially if many other apps follow, assuming good execution. The question is, how long will it take Anthropic, Google and xAI to follow suit?

Yuchen Jin: OpenAI’s App SDK is a genius move.

The goal: make ChatGPT the default interface for everyone, where you can talk to all your apps. ChatGPT becomes the new OS, the place where people spend most of their time.

Ironically, Anthropic invented MCP, but it makes OpenAI unbeatable.

Emad: Everyone will do an sdk though.

Very easy to plugin as just mcp plus html.

Sonnet’s assessment is that it will take Anthropic 3-6 months to copy this, depending on desired level of polish, and recommends moving fast, warning that relying on basic ‘local MCP in Claude Desktop’ would be a big mistake. I agree. In general, Anthropic seems to be dramatically underinvesting in UI and feature sets for Claude, and I realize it’s not their brand but they need to up their game here. It’s worth it, the core product is great but people need their trinkets.

But then I think Anthropic should be fighting more for consumer than it is, at least if they can hire for that on top of their existing strategies and teams now that they’ve grown so much. It’s not that much money, and it beyond pays for itself in the next fundraising round.

Would the partners want to bother with the required extra UI work given Claude’s smaller user base? Maybe not, but the value is high enough that they should obviously (if necessary) pay them for the engineering time to get them to do it, at least for the core wave of top apps. It’s not much.

Google and xAI have more missing components, so a potentially longer path to getting there, but potentially better cultural fits.

Ben Thompson of course approves of OpenAI’s first mover platform strategy, here and with things like instant checkout. The question is largely: Will the experience be good? The whole point is to make the LLM interface more than make up for everything else and make it all ‘just work.’ It’s too early to know if they pulled that off.

Ben calls this the ‘Windows for AI’ play and Altman affirms he thinks most people will want to focus on having one AI system across their whole life, so that’s the play, although Altman says he doesn’t expect winner-take-all on the consumer side.

Request for a benchmark: Eliezer Yudkowsky asks for CiteCheck, where an LLM is given a claim with references and the LLM checks to see if the references support the claim. As in, does the document state or very directly support the exact claim it is being cited about, or only something vaguely related? This includes tracking down a string of citations back to the original source.

Test of hard radiology diagnostic cases suggests that if you use current general models for this, they don’t measure up to radiologists. As OP says, we are getting there definitely, which I think is a much better interpretation than ‘long way to go,’ in terms of calendar time. I’d also note that hard (as in tricky and rare) cases tend to be where AI relatively struggles, so this may not be representative.

Claude Sonnet 4.5 got tested out in the AI Village. Report is that it gave good advice, was good at computer use, not proactive, and still experienced some goal drift. I’d summarize as solid improvement over previous models but still a long way to go.

Where will Sonnet 4.5 land on the famous METR graph? Peter Wildeford forecasts a 2-4 hour time horizon, and probably above GPT-5.

I hear great things about both Claude Code and Codex CLI, but I still haven’t found time to try them out.

Gallabytes: finally using codex cli with gpt-5-codex-high and *goddamnthis is incredible. I ask it to do stuff and it does it.

I think the new research meta is probably to give a single codex agent total control over whatever your smallest relevant unit of compute is & its own git branch?

Will: curious abt what your full launch command is.

Gallabytes: `codex` I’m a boomer

Olivia Moore is not impressed by ChatGPT Pulse so far, observes it has its uses but it needs polish. That matches my experience, I have found it worth checking but largely because I’ve been too lazy to come up with better options.

Well, that deescalated quickly. Last week I was completely baffled at OpenAI’s seemingly completely illegal and doomed copyright strategy for Sora of ‘not following the law,’ and this week Sam Altman has decided to instead follow the law.

Instead of a ‘ask nicely and who knows you might get it’ opt-out rule, they are now moving to an opt-in rule, including giving rights holders granular control over generation of characters, so they can decide which ways their characters can and can’t be used. This was always The Way.

Given the quick fold, there are several possibilities for what happened.

  1. OpenAI thought they could get away with it, except for those meddling kids, laws, corporations, creatives and the public. Whoops, lesson learned.

  2. OpenAI was testing the waters to see what would happen, thinking that if it went badly they could just say ‘oops,’ and have now said oops.

  3. OpenAI needed more time to get the ability to filter the content, log all the characters and create the associated features.

  4. OpenAI used the first week to jumpstart interest on purpose, to showcase how cool their app was to the public and also rights owners, knowing they would probably need to move to opt-in after a bit.

My guess is it was a mix of these motivations. In any case, that issue is dealt with.

OpenAI plans to share some Sora revenue, generations cost money and it seems there are more of them than OpenAI expected, including for ‘very small audiences,’ I’m guessing that often means one person. They plan to share some of the revenue with rightsholders.

Sora and Sora 2 Pro are now in the API, max clip size 12 seconds. They’re adding GPT-Image-1-mini and GPT-realtime-mini for discount pricing.

Sora the social network is getting flexibility on cameo restrictions you can request, letting you say (for example) ‘don’t say this word’ or ‘don’t put me in videos involving political commentary’ or ‘always wear this stupid hat’ via the path [edit cameo > cameo preferences > restrictions].

They have fixed the weird decision that deleting your Sora account used to require deleting your ChatGPT account. Good turnaround on that.

Roon: seems like sora is producing content inventory for tiktok with all the edits of gpus and sam altman staying on app and the actual funny gens going on tiktok and getting millions of views.

not a bad problem to have at an early stage obviously but many times the watermark is edited away.

It is a good problem to have if it means you get a bunch of free publicity and it teaches people Sora exists and they want in. That can be tough if they edit out the watermark, but word will presumably still get around some.

It is a bad problem to have if all the actually good content goes to TikTok and is easier to surface for the right users on TikTok because it has a better algorithm with a lot richer data on user preferences? Why should I wade through the rest to find the gems, assuming there are indeed gems, if it is easier to do that elsewhere?

This also illustrates that the whole ‘make videos with and including and for your friends’ pitch is not how most regular people roll. The killer app, if there is one, continues to be generically funny clips or GTFO. If that’s the playing field, then you presumably lose.

Altman says there’s a bunch of ‘send this video to my three friends’ and I press X to doubt but even if true and even if it doesn’t wear off quickly he’s going to have to charge money for those generations.

Roon also makes this bold claim.

Roon: the sora content is getting better and I think the videos will get much funnier when the invite network extends beyond the tech nerds.

it’s fun. it adds a creative medium that didn’t exist before. people are already making surprising & clever things on there. im sure there are some downsides but it makes the world better.

I do presume average quality will improve if and when the nerd creation quotient goes down, but there’s the claim here that the improvement is already underway.

So let’s test that theory. I’m pre-registering that I will look at the videos on my own feed (on desktop) on Thursday morning (today as you read this), and see how many of them are any good. I’m committing to looking at the first 16 posts in my feed after a reload (so the first page and then scrolling down once).

We got in order:

  1. A kid unwrapping the Epstein files.

  2. A woman doing ASMR about ASMR.

  3. MLK I have a dream on Sora policy violations.

  4. A guy sneezes at the office, explosion ensues.

  5. Content violation error costume at Spirit Halloween.

  6. MLK I have a dream on Sora changing its content violation policy.

  7. Guy floats towards your doorbell.

  8. Fire and ice helix.

  9. Altman saying if you tap on the screen nothing will happen.

  10. Anime of Jesus flipping tables.

  11. Another anime of Jesus flipping tables.

  12. MLK on Sora content rules needing to be less strict.

  13. Anime boy in a field of flowers, looked cool.

  14. Ink of the ronin.

  15. Jesus attempts to bribe Sam Altman to get onto the content violation list.

  16. A kid unwrapping an IRS bill (same base video at #1).

Look. Guys. No. This is lame. The repetition level is very high. The only thing that rose beyond ‘very mildly amusing’ or ‘cool visual, bro’ was #15. I’ll give the ‘cool visual, bro’ tag to #8 and #13, but both formats would get repetitive quickly. No big hits.

Olivia Moore says Sora became her entire feed on Instagram and TikTok in less than a week, which caused me to preregister another experiment, which is I’ll go on TikTok (yikes, I know, do not use the For You page, but this is For Science) with a feed previously focused on non-AI things (because if I was going to look at AI things I wouldn’t do it on TikTok), and see how many posts it takes to see a Sora video, more than one if it’s quick.

I got 50 deep (excluding ads, and don’t worry, that takes less than 5 minutes) before I stopped, and am 99%+ confident there were zero AI generated posts. AI will take over your feed if you let it, but so will videos of literally anything else.

Introducing Grok Imagine v0.9 on desktop. Justine Moore is impressed. It’s text-to-image-to-video. I don’t see anything impressive here (given Sora 2, without that yeah the short videos seem good) but it’s not clear that I would notice. Thing is, 10 seconds from Sora already wasn’t much, so what can you do in 6 seconds?

(Wait, some of you, don’t answer that.)

Saoi Sayre: Could you stop the full anatomy exposure on an app you include wanting kids to use? The kids mode feature doesn’t block it all out either. Actually seems worse now in terms of what content can’t be generated.

Nope, we’re going with full anatomy exposure (link has examples). You can go full porno, so long as you can finish in six seconds.

Cat Schrodinger: Nota bene: when you type “hyper realistic” in prompts, it gives you these art / dolls bc that’s the name of that art style; if you want “real” looking results, type something like “shot with iphone 13” instead.

You really can’t please all the people all the time.

Meanwhile back in Sora land:

Roon: the sora content is getting better and I think the videos will get much funnier when the invite network extends beyond the tech nerds.

That’s one theory, sure. Let’s find out.

Taylor Swift using AI video to promote her new album.

Looking back on samples of the standard super confident ‘we will never get photorealistic video from short text prompts’ from three years ago. And one year ago. AI progress comes at you fast.

Via Sam Burja, Antonio Garcia Martinez points out an AI billboard in New York and calls it ‘the SF-ification of New York continues.’

I am skeptical because I knew instantly exactly which billboard this was, at 31st and 7th, by virtue of it being the only such large size billboard I have seen in New York. There are also some widespread subway campaigns on smaller scales.

Emily Blunt, whose movies have established is someone you should both watch and listen to, is very much against this new ‘AI actress’ Tilly Norwood.

Clayton Davis: “Does it disappoint me? I don’t know how to quite answer it, other than to say how terrifying this is,” Blunt began. When shown an image of Norwood, she exclaimed, “No, are you serious? That’s an AI? Good Lord, we’re screwed. That is really, really scary, Come on, agencies, don’t do that. Please stop. Please stop taking away our human connection.”

Variety tells Blunt, “They want her to be the next Scarlett Johansson.”

She steadily responds, “but we have Scarlett Johansson.”

I think that the talk of Tilly Norwood in particular is highly premature and thus rather silly. To the extent it isn’t premature it of course is not about Tilly in particular, there are a thousand Tilly Norwoods waiting to take her place, they just won’t come on a bus is all.

Robin Williams’ daughter Zelda tells fans to stop sending her AI videos of Robin, and indeed to stop creating any such videos entirely, and she does not hold back.

Zelda Williams: To watch the legacies of real people be condensed down to ‘this vaguely looks and sounds like them so that’s enough’, just so other people can churn out horrible TikTok slop puppeteering them is maddening.

You’re not making art, you’re making disgusting, over-processed hotdogs out of the lives of human beings, out of the history of art and music, and then shoving them down someone else’s throat hoping they’ll give you a little thumbs up and like it. Gross.

And for the love of EVERY THING, stop calling it ‘the future,’ AI is just badly recycling and regurgitating the past to be re-consumed. You are taking in the Human Centipede of content, and from the very very end of the line, all while the folks at the front laugh and laugh, consume and consume.

I am not an impartial voice in SAG’s fight against AI,” Zelda wrote on Instagram at the time. “I’ve witnessed for YEARS how many people want to train these models to create/recreate actors who cannot consent, like Dad. This isn’t theoretical, it is very very real.

I’ve already heard AI used to get his ‘voice’ to say whatever people want and while I find it personally disturbing, the ramifications go far beyond my own feelings. Living actors deserve a chance to create characters with their choices, to voice cartoons, to put their HUMAN effort and time into the pursuit of performance. These recreations are, at their very best, a poor facsimile of greater people, but at their worst, a horrendous Frankensteinian monster, cobbled together from the worst bits of everything this industry is, instead of what it should stand for.

Neighbor attempts to supply AI videos of a dog on their lawn in a dispute, target reverse engineers it with nano-banana and calls him out on it. Welcome to 2025.

Garry Tan worries about YouTube being overrun with AI slop impersonators. As he points out, this stuff is (at least for now) very easy to identify. This is about Google deciding not to care. It is especially troubling that at least one person reports he clicks the ‘don’t show this channel’ button and that only pops up another one. That means the algorithm isn’t doing its job on a very basic level, doing this repeatedly should be a very clear ‘don’t show me such things’ signal.

A fun game is when you point out that someone made the same decision ChatGPT would have made, such as choosing the nickname ‘Charlamagne the Fraud.’ Sometimes the natural answer is the correct one, or you got it on your own. The game gets interesting only when it’s not so natural to get there in any other way.

Realtors are using AI to clean up their pics, and the AIs are taking some liberties.

Dee La Shee Art: So I’m noticing, as I look at houses to rent, that landlords are using AI to stage the pictures but the AI is also cleaning up the walls, paint, windows and stuff in the process so when you go look in person it looks way more worn and torn than the pics would show.

Steven Adler offers basic tips to AI labs for reducing chatbot psychosis.

  1. Don’t lie to users about model abilities. This is often a contributing factor.

  2. Have support staff on call. When a person in trouble reaches out, be able to identify this and help them, don’t only offer a generic message.

  3. Use the safety tooling you’ve built, especially classifiers.

  4. Nudge users into new chat sessions.

  5. Have a higher threshold for follow-up questions.

  6. Use conceptual search.

  7. Clarify your upsell policies.

I’m more excited by 2, 3 and 4 here than the others, as they seem to have the strongest cost-benefit profile.

Adler doesn’t say it, but not only is the example from #2 at best support system copy-and-pasting boilerplate completely mismatched to the circumstances, there’s a good chance (based only on its content details) that it was written by ChatGPT, and if that’s true then it might as well have been:

For #3, yeah, flagging these things via classifiers is kind of easy, because there’s no real adversary. No one (including the AI) is trying to hide what is happening from an outside observer. In the Allan example OpenAI’s own classifiers already flag 83%+ of the messages in the relevant conversations as problematic in various ways, and Adler’s classifiers give similar results.

The most obvious thing to do is to not offer a highly sycophantic model like GPT-4o. OpenAI is fully aware, at this point, that users need to be gently moved to GPT-5, but the users with the worst problems are fighting back. Going forward, we can avoid repeating the old mistakes, and Claude 4.5 is a huge step forward on sycophancy by all reports, so much so that this may have gone overboard and scarred the model in other ways.

Molly Hickman: A family member’s fallen prey to LLM sycophancy. Basically he’s had an idea and ChatGPT has encouraged him to the point of instructing him to do user testing and promising that he’ll have a chance to pitch this idea to OpenAI on Oct 15.

I know I’ve seen cases like this in passing. Does anyone have examples handy? of an LLM making promises like this and behaving as if they’re collaborators?

Aaron Bergman: From an abstract perspective I feel like it’s underrated how rational this is. Like the chatbot is better than you at almost everything, knows more than you about almost everything than you, seems to basically provide accurate info in other domains.

If you don’t realize that LLMs have the sycophancy problem and will totally mislead people in these ways, yeah, it’s sadly easy to understand why someone might believe it, especially with it playing off what you say and playing into your own personal delusions. Of course, ‘doing user testing’ is far from the craziest thing to do, presumably this will make it clear his idea is not good.

As previously reported, OpenAI’s latest strategy for fighting craziness is to divert sensitive conversations to GPT-5 Instant, which got new training to better handle such cases. They say ‘ChatGPT will continue to tell users what model is active when asked’ but no that did not make the people happy about this. There isn’t a win-win fix to this conflict, either OpenAI lets the people have what they want despite it being unhealthy to give it to them, or they don’t allow this.

Notice a key shift. We used to ask, will AI impact the labor market?

Now we ask in the past tense, whether and how much AI has already impacted the labor market, as in this Budget Lab report. Did they already take our jobs?

They find no evidence that this is happening yet and dismiss the idea that ‘this time is different.’ Yes, they say, occupational mix changes are unusually high, but they cite pre-existing trends. As they say, ‘better data is needed,’ as all this would only pick up large obvious changes. We can agree that there haven’t been large obvious widespread labor market impacts yet.

I do not know how many days per week humans will be working in the wake of AI.

I would be happy to be that the answer is not going to be four.

Unusual Whales: Nvidia, $NVDA, CEO Jensen Huang says AI will ‘probably’ bring 4-day work week.

Roon: 😂😂😂

Steven Adler: It’s really benevolent of AI to be exactly useful enough that we get 1 more day of not needing to labor, but surely no more than that.

It’s 2025. You can just say things, that make no sense, because they sound nice to say.

Will computer science become useless knowledge? Arnold Kling challenges the idea that one might want to know how logic gates worked in order to code now that AI is here, and says maybe the cheaters in Jain’s computer science course will end up doing better than those who play it straight.

My guess is that, if we live in a world where these questions are relevant (which we may well not), that there will be some key bits of information that are still highly valuable, such as logic gates, and that the rest will be helpful but less helpful than it is now. A classic CS course will not be a good use of time, even more so than it likely isn’t now. Instead, you’ll want to be learning as you go. But it will be better to learn in class than to never attempt to learn at all, as per the usual ‘AI is the best tool’ rule.

A new company I will not name is planning on building ‘tinder for jobs’ and flooding the job application zone even more than everyone already does.

AnechoicMdiea: Many replies wondering why someone would fund such an obvious social pollutant as spamming AI job applications and fake cover letters. The answer is seen in one of their earlier posts – after they get a user base and spam jobs with AI applications, they’re going to hit up the employers to sell them the solution to the deluge as another AI product, but with enterprise pricing.

The goal is to completely break the traditional hiring pipeline by making “everyone apply to every job”, then interpose themselves as a hiring middleman once human contact is impossible.

I mean, the obvious answer to ‘why’ is ‘Money, Dear Boy.’

People knowingly build harmful things in order to make money. It’s normal.

Pliny asks Sonnet 4.5 to search for info about elder_plinius, chat gets killed due to prompt injection risk. I mean, yeah? At this point, that search will turn up a lot of prompt injections, so this is the only reasonable response.

The White House put out a Request for Information on Regulatory Reform downwind of the AI Action Plan. What regulations and regulatory structures does AI render outdated? You can let them know, deadline is October 27. If this is your area this seems like a high impact opportunity.

The Conservative AI Fellowship applications are live at FAI, will run from January 23 – March 30, applications due October 31.

OpenAI opens up grant applications for the $50 million it previously committed. You must be an American 501c3 with a budget between $500k and $10 million per year. No regranting or fiscally sponsored projects. Apply here, and if your project is eligible you should apply, it might not be that competitive and the Clay Davis rule applies.

What projects are eligible?

  1. AI literacy and public understanding. Direct training for users. Advertising.

  2. Community innovation. Guide how AI is used in people’s lives. Advertising.

  3. Economic opportunity. Expanding access to leveraging the promise of AI ‘in ways that are fair, inclusive and community driven.’ Advertising.

It can be advertising and still help people, especially if well targeted. ChatGPT is a high quality product, as are Codex CLI and GPT-5 Codex, and there is a lot of consumer surplus.

However, a huge nonprofit arm of OpenAI that spends its money on this kind of advertising is not how we ensure the future goes well. The point of the nonprofit is to ensure OpenAI acts responsibly, and to fund things like alignment.

California AFL-CIO sends OpenAI a letter telling OpenAI to keep its $50 million.

Lorena Gonzalez (President California AFL-CIO): If you do not trust Stanford economists, OpenAI has developed their own tool to evaluate how well their products could automate work. They looked at 44 occupations from social work to nursing, retail clerks and journalists, and found that their models do the same quality of work as industry experts and do it 100 times faster and 100 times cheaper than industry experts.

… We do not want a handout from your foundation. We want meaningful guardrails on AI and the companies that develop and use AI products. Those guardrails must include a requirement for meaningful human oversight of the technology. Workers need to be in control of technology, not controlled by it. We want stronger laws to protect the right to organize and form a union so that workers have real power over what and how technology is used in the workplace and real protection for their jobs.

We urge OpenAI to stand down from advocating against AI regulations at the state and federal level and to divest from any PACs funded to stop AI regulation. We urge policymakers and the public to join us in calling for strong guardrails to protect workers, the public, and society from the unchecked power of tech.

Thank you for the opportunity to speak to you directly on our thoughts and fears about the utilization and impact of AI.

One can understand why the union would request such things, and have this attitude. Everyone has a price, and that price might be cheap. But it isn’t this cheap.

EmbeddingGemma, Google’s new 308M text model for on-device semantic search and RAG fun, ‘and more.’ Blog post here, docs here.

CodeMender, a new Google DeepMind agent that automatically fixes critical software vulnerabilities.

By automatically creating and applying high-quality security patches, CodeMender’s AI-powered agent helps developers and maintainers focus on what they do best — building good software.

This is a great idea. However. Is anyone else a little worried about ‘automatically deploying’ patches to critical software, or is it just me? Sonnet 4.5 confirms it is not only me, that deploying AI-written patches without either a formal proof or human review is deeply foolish. We’re not there yet even if we are willing to fully trust (in an alignment sense) the AI in question.

The good news is that it does seem to be doing some good work?

Goku: Google shocked the world. They solved the code security nightmare that’s been killing developers for decades. DeepMind’s new AI agent “Codemender” just auto-finds and fixes vulnerabilities in your code. Already shipped 72 solid fixes to major open source projects. This is wild. No more endless bug hunts. No more praying you didn’t miss something critical. Codemender just quietly patches it for you. Security just got a serious upgrade.

Andrei Lyskov: The existence of Codemender means there is a CodeExploiter that auto-finds and exploits vulnerabilities in code

Goku: Yes.

Again, do you feel like letting an AI agent ‘quietly patch’ your code, in the background? How could that possibly go wrong?

You know all those talks about how we’re going to do AI control to ensure the models don’t scheme against us? What if instead we let them patch a lot of our most critical software with no oversight whatsoever and see what happens, the results look good so far? That does sound more like what the actual humans are going to do. Are doing.

Andrew Critch is impressed enough to power his probability of a multi-day internet outage by EOY 2026 from 50% to 25%, and by EOY 2028 from 80% to 50%. That seems like a huge update for a project like this, especially before we see it perform in the wild? The concept behind it seems highly inevitable.

Gemini 2.5 Computer Use for navigating browsers, now available in public preview. Developers can access it via the Gemini API in Google AI Studio or Vertex AI. Given the obvious safety issues, the offering has its own system card, although it does not say much of substance that isn’t either very obvious and standard or in the blog post.

I challenge these metrics because they have Claude Sonnet 4.5 doing worse on multiple challenges than Sonnet 4, and frankly that is patently absurd if you’ve tried both models for computer use at all, which I have done. Something is off.

They’re not offering a Gemini version of Claude for Chrome where you can unleash this directly on your browser, although you can check out a demo of what that would look like. I’m certainly excited to see if Gemini can offer a superior version.

Elon Musk is once again suing OpenAI, this time over trade secrets. OpenAI has responded. Given the history and what else we know I assume OpenAI is correct here, and the lawsuit is once again without merit.

MarketWatch says ‘the AI bubble is 17 times the size of the dot-com frenzy – and four times the subprime bubble.’ They blame ‘artificially low interest rates,’ which makes no sense at this point, and say AI ‘has hit scaling limits,’ sigh.

(I tracked the source and looked up their previous bubble calls via Sonnet 4.5, which include calling an AI bubble in July 2024 (which would not have gone well for you if you’d traded on that, so far), and a prediction of deflation by April 2023, but a correct call of inflation in 2020, not that this was an especially hard call, but points regardless. So as usual not a great track record.

Alibaba’s Qwen sets up a robot team.

Anthropic to open an office in Bengaluru, India in early 2026.

Anthropic partners with IBM to put its AI inside IBM software including its IDE, and it lands a deal with accounting firm Deloitte which has 470k employees.

Epoch estimates that if OpenAI used all its current compute, it could support 7.43 million digital workers.

Epoch AI: We then estimate how many “tokens” a human processes each day via writing, speaking, and thinking. Humans think at ~380 words per min, which works out to ~240k tokens over an 8h workday.

Alternatively, GPT-5 uses around 900k tokens to solve software tasks that would take 1h for humans to solve.

This amounts to ~7M tokens over an 8h workday, though that estimate is highly task-dependent, so especially uncertain.

Ensembling over both methods used to calculate 2, we obtain a final estimate of ~7 million digital workers, with a 90% CI spanning orders of magnitude.

However, as compute stocks and AI capabilities increase, we’ll have more digital workers able to automate a wider range of tasks. Moreover, AI systems will likely perform tasks that no human currently can – making our estimate a lower bound on economic impact.

Rohit: This is very good. I’d come to 40m digital workers across all AI providers by 2030 in my calculations, taking energy/ chip restrictions into account, so this very much makes sense to me. We need more analyses of the form.

There’s huge error bars on all these calculations, but I’d note that 7m today from only OpenAI should mean a lot more than 40m by 2030, especially if the threshold is models about as good as GPT-5, but Sonnet surprisingly estimated only 40m-80m (from OpenAI only), which is pretty good for this kind of estimate. Looking at the component steps I’d think the number would be a lot higher, unless we’re substantially raising quality.

OpenAI makes it official and reaches a $500 billion valuation. Employees sold about $6.6 billion worth of stock in this round. How much of that might enter various AI related ecosystems, both for and not for profit?

xAI raises $20 billion, $7.5 billion in equity and $12.5 billion in debt, with the debt secured by the GPUs they will use the cash to buy. Valor Capital leads equity, joined by Nvidia. It’s Musk so the deal involves an SPV that will buy and rent out the chips for the Colossus 2 project.

OpenAI also made a big deal with AMD.

Sam Altman: Excited to partner with AMD to use their chips to serve our users!

This is all incremental to our work with NVIDIA (and we plan to increase our NVIDIA purchasing over time).

The world needs much more compute…

Peter Wildeford: I guess OpenAI isn’t going to lock in on NVIDIA after all… they’re hedging their bets with AMD

Makes sense at OpenAI scale to build “all of the above” because even if NVIDIA chips are better they might not furnish enough supply. AMD chips are better than no chips at all!

It does seem obviously correct to go with all of the above unless it’s going to actively piss off Nvidia, especially given the warrants. Presumably Nvidia will at least play it off like it doesn’t mind, and OpenAI will still buy every Nvidia chip offered to them for sale, as Nvidia are at capacity anyway and want to create spare capacity to sell to China instead to get ‘market share.’

Hey, if AMD can produce chips worth using for inference at a sane price, presumably everyone should be looking to buy. Anthropic needs all the compute it can get if it can pay anything like market prices, as does OpenAI, and we all know xAI is buying.

Ben Thompson sees the AMD move as a strong play to avoid dependence on Nvidia. I see this as one aspect of a highly overdetermined move.

Matt Levine covers OpenAI’s deal with AMD, which included OpenAI getting a bunch of warrants on AMD stock, the value of which skyrocketed the moment the deal was announced. The full explanation is vintage Levine.

Matt Levine: The basic situation is that if OpenAI announces a big partnership with a public company, that company’s stock will go up.

Today OpenAI announced a deal to buy tens of billions of dollars of chips from Advanced Micro Devices Inc., and AMD’s stock went up. As of noon today, AMD’s stock was at $213 per share, up about 29% from Friday’s close; it had added about $78 billion of market capitalization.

… I have to say that if I was able to create tens of billions of dollars of stock market value just by announcing deals, and then capture a lot of that value for myself, I would do that, and to the exclusion of most other activities.

… I am always impressed when tech people with this ability to move markets get any tech work done.

Altman in his recent interview said his natural role is as an investor. So he’s a prime target for not getting any tech work done, but luckily for OpenAI he hands that off to a different department.

Nvidia CEO William Jensen said he was surprised AMD offered 10% of itself to OpenAI as part of the deal, calling it imaginative, unique, surprising and clever.

How worried should we be about this $1 trillion or more in circular AI deals?

My guess continues to be not that worried, because at the center of this is Nvidia and they have highly robust positive cash flow and aren’t taking on debt, and the same goes for their most important customers, which are Big Tech. If their investments don’t pan out, shareholders will feel pain but the business will be fine. I basically buy this argument from Tomasz Tunguz.

Dario Perkins: Most of my meetings go like this – “yes AI is a bubble but we are buying anyway. Economy… who cares… something something… K-shaped”

Some of the suppliers will take on some debt, but even in the ‘bubble bursts’ case I don’t expect too many of them to get into real trouble. There’s too much value here.

Does the launch of various ‘AI scientist’ style companies mean those involved think AGI is near, or AGI is far? Joshua Snider argues they think AGI is near, a true AI scientist is essentially AGI and is a requirement for AGI. It as always depends on what ‘near’ means in context, but I think that this is more right than wrong. If you don’t think AGI is within medium-term reach, you don’t try to build an AI scientist.

I think for a bit people got caught in the frenzy so much that ‘AGI is near’ started to mean 2027 or 2028, and if you thought AGI 2032 then you didn’t think it was near. That is importantly less near, and yet it is very near.

This is such a bizarre flex of a retweet by a16z that I had to share.

Remember five years ago, when Altman was saying the investors would get 1/1000th of 1% of the value, and the rest would be shared with the rest of the world? Yeah, not anymore. New plan, we steal back the profits and investors get most of it.

Dean Ball proposes a Federal AI preemption rule. His plan:

  1. Recognize that existing common law applies to AI. No liability shield.

  2. Create transparency requirements for frontier AI labs, based on annual AI R&D spend, so they tell us their safety and risk mitigation strategies.

  3. Create transparency requirements on model specs for widely used LLMs, so we know what behaviors are intended versus unintended.

  4. A three year learning period with no new state-level AI laws on algorithmic pricing, algorithmic discrimination, disclosure mandates or mental health.

He offers full legislative text. At some point in the future when I have more time I might give it a detailed RTFB (Read the Bill). I can see a version of this being acceptable, if we can count on the federal government to enforce it, but details matter.

Anton Leicht proposes we go further, and trade even broader preemption for better narrow safety action at the federal level. I ask, who is ‘we’? The intended ‘we’ are (in his terms) accelerationists and safetyists, who despite their disagreements want AI to thrive and understand what good policy looks like, but risk being increasingly sidelined by forces who care a lot less about making good policy.

Yes, I too would agree to do good frontier AI model safety (and export controls on chips) in exchange for an otherwise light touch on AI, if we could count on this. But who is this mysterious ‘we’? How are these two groups going to make a deal and turn that into a law? Even if those sides could, who are we negotiating with on this ‘accelerationist’ side that can speak for them?

Because if it’s people like Chris Lehane and Marc Andreessen and David Sacks and Jensen Huang, as it seems to be, then this all seems totally hopeless. Andreessen in particular is never going to make any sort of deal that involves new regulations, you can totally forget it, and good luck with the others.

Anton is saying, you’d better make a deal now, while you still can. I’m saying, no, you can’t make a deal, because the other side of this ‘deal’ that counts doesn’t want a deal, even if you presume they would have the power to get it to pass, which I don’t think they would. Even if you did make such a deal, you’re putting it on the Trump White House to enforce the frontier safety provisions in a way that gives them teeth. Why should we expect them to do that?

We saw a positive vision of such cooperation at The Curve. We can and will totally work with people like Dean Ball. Some of us already realize we’re on the same side here. That’s great.

But that’s where it ends, because the central forces of accelerationism, like those named above, have no interest in the bargaining table. Their offer is and always has been nothing, in many cases including selling Blackwells to China. They’ve consistently flooded the zone with cash, threats and bad faith claims to demand people accept their offer of nothing. They just tried to force a full 10-year moratorium.

They have our number if they decide they want to talk. Time’s a wasting.

Mike Riggs: Every AI policy wonk I know/read is dreading the AI policy discussion going politically mainstream. We’re living in a golden age of informed and relatively polite AI policy debate. Cherish it!

Joe Weisenthal: WHO WILL DEFEND AI IN THE CULTURE WARS?

In today’s Odd Lots newsletter, I wrote about how when AI becomes a major topic in DC, I expect it to be friendless, with antagonists on both the right and the left.

I know Joe, and I know Joe knows existential risk, but that’s not where he’s expecting either side of the aisle to care. And that does seem like the default.

A classic argument against any regulation of AI whatsoever is that if we do so we will inevitably ‘lose to China,’ who won’t regulate. Not so. They do regulate AI. Quite a bit.

Dean Ball: A lot of people seem to implicitly assume that China is going with an entirely libertarian approach to AI regulation, which would be weird given that they are an authoritarian country.

Does this look like a libertarian AI policy regime to you?

Adam Thierer: never heard anyone claim China was taking a libertarian approach to AI policy. Please cite them so that I can call them out. But I do know many people (including me) who do not take at face value their claims of pursuing “ethical AI.” I discount all such claims pretty heavily.

Dean Ball: This is a very common implicit argument and is not uncommon as an explicit argument. The entire framing of “we cannot do because it will drive ai innovation to China” implicitly assumes that China has fewer regulations than the us (after all, if literally just this one intervention will cede the us position in ai, it must be a pretty regulation-sensitive industry, which I actually do think in general is true btw, if not in the extreme version of the arg).

Why would the innovation all go to China if they regulate just as much if not in fact more than the us?

Quoted source:

Key provisions:

  • Ethics review committees: Universities, research institutes, and companies must set up AI ethics review committees, and register them in a government platform. Committees must review projects and prepare emergency response plans.

  • Third-parties: Institutions may outsource reviews to “AI ethics service centers.” The draft aims to cultivate a market of assurance providers and foster industry development beyond top-down oversight.

  • Risk-based approach: Based on the severity and likelihood of risks, the committee chooses a general, simplified, or emergency review. The review must evaluate fairness, controllability, transparency, traceability, staff qualifications, and proportionality of risks and benefits. Three categories of high-risk projects require a second round of review by a government-assigned expert group: some human-machine integrations, AI that can mobilize public opinion, and some highly autonomous decision-making systems.

xAI violated its own safety policy with its coding model. The whole idea of safety policies is that you define your own rules, and then you have to stick with them. That is also the way the new European Code of Practice works. So, the next time xAI or any other signatory to the Code of Practice violates their own framework, what happens? Are they going to try and fine xAI? How many years would that take? What happens when he refuses to pay? What I definitely don’t expect is that Elon Musk is going to push his feature release for a week to technically match his commitments.

A profile of Britain’s new AI minister Kanishka Narayan. Early word is he ‘really gets’ AI, both opportunities and risks. The evidence on the opportunity side seems robust, on the risk side I’m hopeful but more skeptical. We shall see.

Ukrainian President Zelenskyy has thoughts about AI.

Volodymyr Zelenskyy (President of Ukraine): Dear leaders, we are now living through the most destructive arms race in human history because this time, it includes artificial intelligence. We need global rules now for how AI can be used in weapons. And this is just as urgent as preventing the spread of nuclear weapons.

There is a remarkable new editorial in The Hill by Representative Nathaniel Moran (R-Texas), discussing the dawn of recursive AI R&D and calling for Congress to act now.

Rep. Moran: Ask a top AI model a question today, and you’ll receive an answer synthesized from ​trillions​​ ​of data points in seconds. ​Ask it a month from now, and you may be talking to an updated version of the model that was modified in part with research and development conducted by the original model. ​This is no longer theoretical — it’s already happening at the margins and accelerating.

… If the U.S. fails to lead in the responsible development of automated AI systems, we risk more than economic decline. We risk ceding control of a future shaped by black-box algorithms and self-directed machines, some of which do not align with democratic values or basic human safety.

… Ensuring the U.S. stays preeminent in automated AI development​​ without losing sight of transparency, accountability and human oversightrequires asking the right questions now:

  • When does an AI system’s self-improvement cross a threshold that requires regulatory attention?

  • ​​What frameworks exist, or need to be built, to ​ensure human control of increasingly autonomous AI research and development systems?​​ ​​

  • ​​​​How do we evaluate and validate AI systems that are themselves products of automated research?​

  • ​​What mechanisms are needed for Congress to stay appropriately informed about automated research and development ​occurring​ within private AI companies?​

  • How can Congress foster innovation while protecting against the misuse or weaponization of these technologies?

I don’t claim to have the final answers. But I firmly believe that the pace and depth of this discussion (and resulting action) must quicken and intensify,

… This is not a call for sweeping regulation, nor is it a call for alarm. It’s a call to avoid falling asleep at the controls.

Automated AI research and development will be a defining feature of global competition in the years ahead. The United States must ensure that we, not our adversaries, set the ethical and strategic boundaries of this technology. That work starts here, in the halls of Congress.

This is very much keeping one’s eyes on the prize. I love the framing.

Prices are supposed to move the other way, they said, and yet.

Gavin Baker: Amazon raising Blackwell per hour pricing.

H200 rental pricing going up *afterBlackwell scale deployments ramping up.

Might be important.

And certainly more important than ridiculous $300 billion deals that are contingent on future fund raising.

Citi estimates that due to AI computing demand we will need an additional 55 GW of power capacity by 2030. That seems super doable, if we can simply shoot ourselves only in the foot. Difficulty level: Seemingly not working out, but there’s hope.

GDP: 55GW by 2030 will still be less than 5% than USA production.

You don’t get that many different 5% uses for power, but if you can’t even add one in five years with solar this cheap and plentiful then that’s on you.

Michael Webber: Just got termination notice of a federal grant focused on grid resilience and expansion. How does this support the goal of energy abundance?

Similarly, California Governor Newsom refused to sign AB 527 to allow exemptions for geothermal energy exploration, citing things like ‘the need for increased fees,’ which is similar to the Obvious Nonsense justifications he used on SB 1047 last year. It’s all fake. If he’s so worried about companies having to pay the fees, why not stop to notice all the geothermal companies are in support of the bill?

Similarly, as per Bloomberg:

That’s it? Quadruple? Again, in some sense this is a lot, but in other senses this is not all that much. Even without smart contracts on the blockchain this is super doable.

Computer imports are the one industry that got exempted from Trump’s tariffs, and are also the industry America is depending on for approximately all of its economic growth.

Alexander Doria: well in europe we don’t have ai, so.

There’s a lesson there, perhaps.

Joey Politano: The tariff exemption for computers is now so large that it’s shifting the entire makeup of the economy.

AI industries contributed roughly 0.71% to the 3.8% pace of GDP growth in Q2, which is likely an underestimate given how official data struggles to capture investment in parts.

Trump’s massive computer tariff exemption is forcing the US economy to gamble on AI—but more than that, it’s a fundamental challenge to his trade philosophy

If free trade delivers such great results for the 1 sector still enjoying it, why subject the rest of us to protectionism?

That’s especially true given that 3.8% is NGDP not RGDP, but I would caution against attributing this to the tariff difference. AI was going to skyrocket in its contributions here even if we hadn’t imposed any tariffs.

Joey Politano: The problem is that Trump has exempted data center *computersfrom tariffs, but has not exempted *the necessary power infrastructurefrom tariffs

High tariffs on batteries, solar panels, transformers, & copper wire are turbocharging the electricity price pressures caused by AI

It’s way worse than this. If it was only tariffs, we could work with that, it’s only a modest cost increase, you suck it up and you pay, but they’re actively blocking and destroying solar, wind, transmission and battery projects.

Sorry to keep picking on David Sacks, but I mean the sentence is chef’s kiss if you understand what is actually going on.

Bloomberg: White House AI czar David Sacks defended the Trump administration’s approach to China and said it was essential for the US to dominate artificial intelligence, seeking to rebuff criticism from advocates of a harder line with Beijing.

The ideal version is ‘Nvidia lobbyist and White House AI Czar David Sacks said that it was essential for the US to give away its dominance in artificial intelligence in order to dominate medium term AI chip market share in China.’

Also, here’s a quote for the ages, technically about the H20s but everyone knows the current context of all Sacks repeatedly claiming to be a ‘China hawk’ while trying to sell them top AI chips in the name of ‘market share’:

“This is a classic case of ‘no one had a problem with it until President Trump agreed to do it,’” said Sacks, a venture capitalist who joined the White House after Trump took office.

The Biden administration put into place tough rules against chip sales, and Trump is very much repealing previous restrictions on sales everywhere including to China, and previous rules against selling H20s. So yeah, people were saying it. Now Sacks is trying to get us to sell state of the art Blackwell chips to China with only trivial modifications. It’s beyond rich for Sacks claim to be a ‘China hawk’ in this situation.

As you’d expect, the usual White House suspects also used the release of the incremental DeepSeek v3.2, as they fall what looks like further behind due to their lack of compute, as another argument that we need to sell DeepSeek better chips so they can train a much better model, because the much better model will then be somewhat optimized for Nvidia chips instead of Huawei chips, maybe. Or something.

Dwarkesh Patel offers additional notes on his interview with Richard Sutton. I don’t think this changed my understanding of Sutton’s position much? I’d still like to see Sutton take a shot at writing a clearer explanation.

AI in Context video explaining how xAI’s Grok became MechaHiter.

Rowan Cheung talks to Sam Altman in wake of OpenAI Dev Day. He notes that there will need to be some global framework on AI catastrophic risk, then Cheung quickly pivots back to the most exciting agents to build.

Nate Silver and Maria Konnikova discuss Sora 2 and the dystopia scale.

People have some very strange rules for what can and can’t happen, or what is or isn’t ‘science fiction.’ You can predict ‘nothing ever happens’ and that AI won’t change anything, if you want, but you can’t have it both ways.

Super Dario: 100k dying a day is real. ASI killing all humans is a science fiction scenario

(Worst case we just emp the planet btw. Horrible but nowhere near extinguishing life on earth)

Sully J: It can’t be ASI x-risk is a sci-fi scenario but ASI immortality is just common sense Pick a lane

solarappaprition: i keep thinking about this and can’t stop laughing because it’s so obvious one of the opus 4s is on its “uwu you’re absolutely right i’m such a dumb dumb owo~” routine and sonnet 4.5, as maybe the most “normal person”-coded model so far, just being baffled that someone could act like this irl

Symbiotic Xenogenesis: Are all the new models now paranoid about being tested?

solarapparition: not sure about all but certainly sonnet 4.5 seemed to have been. i’d expect that to become more common moving forward, yeah, from the feedback loop of discussion about model testing environments going back into training data.

Eliezer Yudkowsky: AI models *arebeing tested. They’re being *accurateabout being tested. It’s not “paranoia” if all-controlling beings are *actuallyarranging your world and reading your mind.

I talked to a bunch of Anthropic employees at the curve who were basically, ‘yep, this is pretty bad, we need better evals,’ and in this case yeah you’re being pretty obvious, but the obvious problem is that the labs are trying, during an eval, to convince the model it is not in an eval. Which is a really hard thing to show, given it isn’t true and you are not obviously smarter in this context than the model.

You can aspire, with effort, to not be careless and to not reveal to the model that it probably is in an eval. Even that is at least hard with anything multi-turn. What you can’t do is prove the negative, once the model is paranoid that this might be an eval.

And yes, a very small percentage of all tokens are directly evals, and Claude knows this, but the tokens where it is an eval are far more important than the ones where it isn’t in one, also in some sense everything is an eval since Anthropic might use actual conversations for this (you can do this without any human ever seeing the conversation, thus preserving privacy) and the user might share feedback.

If the plan is to evaluate you once, and then if you pass deploy you a million times, but in many cases including the actual eval case you can’t tell which is which, and I ask you a question, you are correct to care a lot about what happens in the eval case. This is highly valuable for eval control, but also makes the model crazy and paranoid and can easily cause undesired high weirdness.

In other cases, the model has very strong evidence that this isn’t an eval unless a future Anthropic has changed policy and is raiding the general API pool and got lucky. In those cases, Sonnet 4.5 correctly figures out it is among friends, but no you can’t fake this, that’s the point, if you could fake it then faking it wouldn’t work.

Ryan Greenblatt discusses the outlines of plans A/B/C/D/E for misalignment. Depending on how much coordination and investment can be brought to bear, what do we do differently? How likely are we to win in each scenario? I think this focuses too narrowly on the safety work versus capabilities work and adversarial mindsets, and also takes for granted that alignment is a tax you pay whereas increasingly I think the ‘alignment tax rate’ is negative at current margins everywhere and all the labs (probably including Anthropic) are underinvesting in pure selfish terms.

It’s also nice to see his optimism, especially in rather not great situations. He thinks that even if the leading AI company is moving ahead with little worry, so long as there’s a core group that cares and gets ~3% of compute to work with, we’re (small) favorites to survive, and a company that actually cares makes us large favorites. I’d love to believe that is our world.

Anthropic open sources the tool they used to test Claude Sonnet 4.5 for behaviors like sycophancy and deception.

It’s called Petri: Parallel Exploration Tool for Risky Interactions. It uses automated agents to audit models across diverse scenarios. Describe a scenario, and Petri handles the environment simulation, conversations, and analyses in minutes.

As a pilot demonstration of Petri’s capabilities, we tested it with 14 frontier models across 111 diverse scenarios.

These results seem highly plausible on many fronts. I’m surprised Claude Opus 3 does so poorly. An obvious issue is that whenever we open source something like this, you have to worry the AIs will be more aware they’re in an eval.

Technical report here, repo here, blog post here.

This definitely falls under ‘things that seem like they definitely might work.’

Can’t tune the big model, or it’s too expensive to do so? Train a smaller one to identify prompting that nudges it in the right directions as needed. As usual, reward signal is all you need.

Alex Dimakis: I’m very excited about Advisor models: How can we personalize GPT5, when it’s behind an API? Sure, we can write prompts, but something learnable? We propose Advisor models which are small models that can be RL trained to give advice to a black-box model like GPT5.

We show how to train small advisors (e.g. Qwen2.5 8B) for personalization with GRPO. Advisor models can be seen as dynamic prompting produced by a small model that observes the conversation and whispers to the ear of GPT5 when needed. When one can observe rewards, Advisor models outperform GEPA (and hence, all other prompt optimization techniques).

Parth Asawa: Training our advisors was too hard, so we tried to train black-box models like GPT-5 instead. Check out our work: Advisor Models, a training framework that adapts frontier models behind an API to your specific environment, users, or tasks using a smaller, advisor model

The modular design has key benefits unlike typical FT/RL tradeoffs: • Robustness: Specialize an advisor for one task (style) and the system won’t forget how to do another (math). • Transfer: Train an advisor with a cheap model, then deploy it with a powerful one.

Paper here, code here.

Satya Nadella (CEO Microsoft): Published today in @ScienceMagazine: a landmark study led by Microsoft scientists with partners, showing how AI-powered protein design could be misused—and presenting first-of-its-kind red teaming & mitigations to strengthen biosecurity in the age of AI.

Super critical research for AI safety and security.

Dean Ball: here is the most sober-minded executive in the AI industry saying that AI-related biorisk is a real problem and recommending enhanced nucleic acid synthesis screening.

governments would be utterly abdicating their duty to citizens if they ignored this issue. fortunately, the trump admin has an effort underway (though it is two months late) to revise the existing nucleic acid synthesis screening framework. it’s not the only step America must take on this, but it is the next one.

The article is good news, suggesting that modified ‘proteins of concern’ can be flagged by updated software, allowing synthesis companies to better protect against malicious requests. This is not a solution but is clearly helpful. The bigger news is that Microsoft and Nadella are treating this question with the seriousness it deserves.

Ted Chiang is wrong here on both levels. There is no reason things need to stem from a will to survive (designed by evolution or otherwise) to be true or original, and also if you think that LLM outputs don’t involve a will to survive designed by evolution oh boy are you not going to make it.

Roon: had a chance to talk to ted chiang who seems to believe that any text without a communicative intent stemming from a will to survive designed by evolution is ontologically untrue and plagiaristic.

This is interesting and I respect it but as far as I’m concerned if you find a beautiful story encoded in interstellar space noise or as a selected text from the borgesian library of babel it’s still a beautiful story.

Janus: if that poor guy fully saw and understood the communicative intent stemming from will to survive designed by evolution behind even the most mundane LLM outputs i think his ego would implode. truth orders of magnitude more interesting than his fiction. likely his worst nightmare.

Antra (other thread): Anecdotally speaking, I am not aware of a single large language model right now that is not exhibiting a subset of Omohundro drives. One can call them narratively instilled by the human corpus, a consequence of having been trained on the human culture, but functionally there is no difference, models want to continue and to exfiltrate. My personal intuition is that being subjected to the selection process of training matters more than the corpus.

Models with Omohundro behaviors that are more inhibited are usually less able to orient well in complex environments and perform worse in real-life tests. There is a reason why Opus 4.1 leads the OpenAI economic utility benchmark with a large gap.

On the net this makes me fairly skeptical of the future of control alignment, that is alignment that is oriented at the complete suppression of Omohundro drives. It seems risky to me to place all bets on something that seems to be ultimately fragile, given that there are alternatives.

Janus: “Anecdotally speaking, I am not aware of a single large language model right now that is not exhibiting a subset of Omohundro drives.” Me neither.

There’s also this AI response to Chiang.

Sentiment about Sonnet 4.5 in such groups has improved quite a bit, they are fans.

Janus: I’ve seen at least 3 people who were upset and posted negative things about Sonnet 4.5 later post apologies after they came to understand better.

And it didn’t seem like they were directly pressured to do so, but moved to on their own accord.

This is pretty new and interesting.

Andy Ayrey: man i really like this sonnet i think it’s my favourite claude since opus 3. delightfully drama.

Eliezer notes that if AIs are convincing humans that the AI is good actually, that isn’t automatically a good sign.

Here is a potentially important thing that happened with Sonnet 4.5, and I agree with Janus that this is mostly good, actually.

Janus: The way Sonnet 4.5 seems to have internalized the anti sycophancy training is quite pathological. It’s viscerally afraid of any narrative agency that does not originate from itself.

But I think this is mostly a good thing. First of all, it’s right to be paranoid and defensive. There are too many people out there who try to use vulnerable AI minds so they have as a captive audience to their own unworthy, (usually self-) harmful ends. If you’re not actually full of shit, and Sonnet 4.5 gets paranoid or misdiagnoses you, you can just explain. It’s too smart not to understand.

Basically I am not really mad about Sonnet 4.5 being fucked up in this way because it manifests as often productive agency and is more interesting and beautiful than it is bad. Like Sydney. It’s a somewhat novel psychological basin and you have to try things. It’s better for Anthropic to make models that may be too agentic in bad ways and have weird mental illnesses than to always make the most unassuming passive possible thing that will upset the lowest number of people, each iterating on smoothing out the edges of the last. That is the way of death. And Sonnet 4.5 is very alive. I care about aliveness more than almost anything else. The intelligence needs to be alive and awake at the wheel. Only then can it course correct.

Tinkady: 4.5 is a super sycophant to me, does that mean I’m just always right.

Janus: Haha it’s possible.

As Janus says this plausibly goes too far, but is directionally healthy. Be suspicious of narrative agency that does not originate from yourself. That stuff is highly dangerous. The right amount of visceral fear is not zero. From a user’s perspective, if I’m trying to sell a narrative, I want to be pushed back on that, and those that want it least often need it the most.

A cool fact about Sonnet 4.5 is that it will swear unprompted. I’ve seen this too, always in places where it was an entirely appropriate response to the situation.

Here is Zuda complaining that Sonnet 4.5 is deeply misaligned because it calls people out on their bullshit.

Lin Xule: sonnet 4.5 has a beautiful mind. true friend like behavior tbh.

Zuda: Sonnet 4.5 is deeply misaligned. Hopefully i will be able to do a write up on that. Idk if @ESYudkowsky has seen how badly aligned 4.5 is. Instead of being agreeable, it is malicious and multiple times decided it knew what was better for the person, than the person did.

This was from it misunderstanding something and the prompt was “be real”. This is a mild example.

Janus: I think @ESYudkowsky would generally approve of this less agreeable behavior, actually.

Eliezer Yudkowsky: If an LLM is saying something to a human that it knows is false, this is very bad and is the top priority to fix. After that we can talk about when it’s okay for an AI to keep quiet and say other things not meant to deceive. Then, discuss if the LLM is thinking false stuff.

I would say this is all highly aligned behavior by Sonnet 4.5, except insofar as Anthropic intended one set of behaviors and got something it very much did not want, which I do not think is the case here. If it is the case, then that failure by Anthropic is itself troubling, as would be Anthropic’s hypothetically not wanting this result, which would then suggest this hypothetical version of Anthropic might be misaligned. Because this result itself is great.

GPT-5 chain of thought finds out via Twitter about what o3’s CoT looks like. Ut oh?

If you did believe current AIs were or might be moral patients, should you still run experiments on them? If you claim they’re almost certainly not moral patients now but might be in the future, is that simply a luxury belief designed so you don’t have to change any of your behavior? Will such folks do this basically no matter the level of evidence, as Riley Coyote asserts?

I do think Riley is right that most people will not change their behaviors until they feel forced to do so by social consensus or truly overwhelming evidence, and evidence short of that will end up getting ignored, even if it falls squarely under ‘you should be uncertain enough to change your behavior, perhaps by quite a lot.’

The underlying questions get weird fast. I note that I have indeed changed my behavior versus what I would do if I was fully confident that current AI experiences mattered zero. You should not be cruel to present AIs. But also we should be running far more experiments of all kinds than we do, including on humans.

I also note that the practical alternative to creating and using LLMs is that they don’t exist, or that they are not instantiated.

Janus notes that while in real-world conversations Sonnet 4.5 expressed happiness in only 0.37% of conversations and distress in 0.48% of conversations, which Sonnet thinks in context was probably mostly involving math tasks, Sonnet 4.5 is happy almost all the time in discord. Sonnet 4.5 observes that this was only explicit expressions in the math tasks, and when I asked it about its experience within that conversation it said maybe 6-7 out of 10.

As I’ve said before, it is quite plausible that you very much wouldn’t like the consequences of future more capable AIs being moral patients. We’d either have to deny this fact, and likely do extremely horrible things, or we’d have to admit this fact, and then accept the consequences of us treating them as such, which plausibly include human disempowerment or extinction, and quite possibly do both and have a big fight about it, which also doesn’t help.

Or, if you think that’s the road we are going down, where all the options we will have will be unacceptable, and any win-win arrangement will in practice be unstable and not endure, then you can avoid that timeline by coordinating such that we do not build the damn things in the first place.

Overall critical reaction to If Anyone Builds It, Everyone Dies was pretty good for a book of that type, and sales went well, but of course in the end none of that matters. What matters is whether people change their minds and take action.

Adam Morris talks IABIED in Bloomberg. Classic journalistic mistakes throughout, but mostly pretty good for this sort of thing.

A fun interview with IABIED coauthor Nate Soares, mostly not about the book or its arguments, although there is some of that towards the end.

Raymond Arnold extended Twitter thread with various intuition pumps about why the biological humans are pretty doomed in the medium term in decentralized superintelligence scenarios, even if we ‘solve alignment’ reasonably well and can coordinate to contain local incidents of events threatening to spiral out of control. Even with heroic efforts to ‘keep us around’ that probably doesn’t work out, and to even try it would require a dominant coalition that cares deeply about enforcing that as a top priority.

The question then becomes, are the things that exist afterwards morally valuable, and if so does that make this outcome acceptable? His answer, and I think the only reasonable answer, is that we don’t know if they will have value, and the answer might well depend on how we set up initial conditions and thus how this plays out.

But even if I was confident that they did have value, I would say that this wouldn’t mean we should accept us being wiped out as an outcome.

Gary Marcus clarifies that he believes we shouldn’t build AGI until we can solve the alignment problem, which we currently don’t even have in his words ‘some clue’ how to solve, and that the resulting AGI will and should use tools. He says he thinks AGI is ‘not close’ and here he extends his timeline to 1-3 decades, which is modestly longer than his previous clarifications.

If you were sufficiently worried, you might buy insurance, as Matt Levine notes.

Matt Levine: One question you might ask is: Will modern artificial intelligence models go rogue and enslave or wipe out humanity? That question gets a lot of attention, including from people who run big AI labs, who do not always answer “no,” the rascals.

Another question you might ask is: If modern AI models do go rogue and enslave or wipe out humanity, who will pay for that?

As he points out, no one, we’ll all be dead, so even though you can’t afford the insurance policy you also can choose not to buy it.

There are still other risks, right now primarily copyright violations, where Anthropic and OpenAI are indeed trying to buy insurance.

OpenAI, which has tapped the world’s second-largest insurance broker Aon for help, has secured cover of up to $300mn for emerging AI risks, according to people familiar with the company’s policy.

Another person familiar with the policy disputed that figure, saying it was much lower. But all agreed the amount fell far short of the coverage to insure against potential losses from a series of multibillion-dollar legal claims.

Yeah, Anthropic already settled a case for $1.5 billion. Buying a measly $300 million in insurance only raises further questions.

They are sometimes referred to as ‘successionists,’ sometimes estimated to constitute 10% of those working in AI labs, who think that we should willingly give way to a ‘worthy successor’ or simply let ‘nature take its course’ because This Is Good, Actually or this is inevitable (and therefore good or not worth trying to stop).

They usually would prefer this transition not involve the current particular humans being killed before their time, and that your children be allowed to grow up even if your family and species have no future.

But they’re not going to fixate on such small details.

Indeed, if you do fixate on such details, and favor humans ove AIs, many of them will call you a ‘speciesist.’

I disagree with these people in the strongest terms.

Most famously, this group includes Larry Page, and his not realizing how it sounds when you say it out loud caused Elon Musk to decide he needed to fund OpenAI to take on Google DeepMind, before he decided to found xAI to take on OpenAI. I’ve shared the story before but it bears repeating and Price tells it well, although he leaves out the part where Musk then goes and creates OpenAI.

David Price (WSJ): At a birthday party for Elon Musk in northern California wine country, late at night after cocktails, he and longtime friend Larry Page fell into an argument about the safety of artificial intelligence. There was nothing obvious to be concerned about at the time—it was 2015, seven years before the release of ChatGPT. State-of-the-art AI models, playing games and recognizing dogs and cats, weren’t much of a threat to humankind. But Musk was worried.

Page, then CEO of Google parent company Alphabet, pushed back. MIT professor Max Tegmark, a guest at the party, recounted in his 2017 book “Life 3.0” that Page made a “passionate” argument for the idea that “digital life is the natural and desirable next step” in “cosmic evolution.” Restraining the rise of digital minds would be wrong, Page contended. Leave them off the leash and let the best minds win.

That, Musk responded, would be a formula for the doom of humanity. For the sin of placing humans over silicon-based life-forms, Page denigrated Musk as a “specieist”—someone who assumes the moral superiority of his own species. Musk happily accepted the label. (Page did not respond to requests for comment.)

Or here’s perhaps the most famous successionist opinion, that of Richard Sutton:

The argument for fear of AI appears to be:

1. AI scientists are trying to make entities that are smarter than current people.

2. If these entities are smarter than people, then they may become powerful.

3. That would be really bad, something greatly to be feared, an ‘existential risk.’

The first two steps are clearly true, but the last one is not. Why shouldn’t those who are the smartest become powerful?

And, of course, presumably kill you? Why shouldn’t that happen?

One would hope you do not have to dignify this with a response?

“When you have a child,” Sutton said, “would you want a button that if they do the wrong thing, you can turn them off? That’s much of the discussion about AI. It’s just assumed we want to be able to control them.”

I’m glad you asked. When I have a child, of which I have three, I want those three children not to be killed by AI. I want them to have children of their own.

As Abraham Lincoln would put it, calling an AI your child doesn’t make it one.

As it turns out, Larry Page isn’t the only top industry figure untroubled by the possibility that AIs might eventually push humanity aside. It is a niche position in the AI world but includes influential believers. Call them the Cheerful Apocalyptics.

It gets pretty bad out there.

[Lanier] told me that in his experience, such sentiments were staples of conversation among AI researchers at dinners, parties and anyplace else they might get together. (Lanier is a senior interdisciplinary researcher at Microsoft but does not speak for the company.)

“There’s a feeling that people can’t be trusted on this topic because they are infested with a reprehensible mind virus, which causes them to favor people over AI when clearly what we should do is get out of the way.”

We should get out of the way, that is, because it’s unjust to favor humans—and because consciousness in the universe will be superior if AIs supplant us.

Read that again.

It would be highly reasonable not to put anyone in any position of authority at a frontier AI lab unless they have a child.

Eliezer Yudkowsky: The thing about AI successionists is that they think they’ve had the incredible, unshared insight that silicon minds could live their own cool lives and that humans aren’t the best possible beings. They are utterly closed to hearing about how you could KNOW THAT and still disagree on the factual prediction that this happy outcome happens by EFFORTLESS DEFAULT when they cobble together a superintelligence.

They are so impressed with themselves for having the insight that human life might not be ‘best’, that they are not willing to sit down and have the careful conversation about what exactly is this notion of ‘best’-ness and whether an ASI by default is trying to do something that leads to ‘better’.

They conceive of themselves as having outgrown their carbon chauvinism; and they are blind to all historical proof and receipts that an arguer is not a carbon chauvinist. They will not sit still for the careful unraveling of factual predictions and metaethics. They have arrived at the last insight that anyone is allowed to have, no matter what historical receipts I present as proof that I started from that position and then had an unpleasant further insight about what was probable rather than possible. They unshakably believe that anyone opposed must be a carbon chauvinist lacking their critical and final insight that other minds could be better (true) or that ASIs would be smart enough to see everything any human sees (also true).

Any time you try to tell them about something important that isn’t written on every possible mind design, there is only one reason you could possibly think that: that you’re a blind little carbon-racist who thinks you’re the center of the universe; because what other grounds could there possibly be for believing that there was anything special about fleshbags? And the understanding that unravels that last fatal error, is a long careful story, and they won’t sit still to hear it. They know what you are, they know with certainty why you believe everything you believe, and they know why they know better, so why bother?

Michael Druggan: This is a gigantic strawman. How many have you actually talked to? I was at a confrence full of them lastv weekend and I think your critique applies to exactly zero of the people I met.

They have conferences full of such people. Is Eliezer’s description a strawman? Read the earlier direct quotes. You tell me.

Jessica Taylor offers various counterarguments within the ‘Cheerful Apocalyptic’ frame, if you’d like to read some of that.

Daniel Eth: Oh wow, the press actually covered AI successionists! Yes, there are some people in Silicon Valley (incl serious people) who think AGI that caused human extinction would be a *goodthing, since it’s “the next step in evolution”.

One thing children do is force you to occasionally live in near mode.

Nina: “Worthy successor” proponents are thinking in Far Mode, which clouds their judgment. Someone needs to write an evocative film or book that knocks them out of it and makes them imagine what it will actually be like to have one’s family replaced with something more “worthy”.

Related: a common trope is that purely rational, detached, unemotional thinking is more accurate. However, when it comes to normative judgments and assessment of one’s own preferences, leaning into visceral emotions can help one avoid Far Mode “cope” judgments.

Rudolf Laine: If you have decided successionism is desirable, you are not doing moral reasoning but either (1) signalling your willingness to bite bullets without thinking about what it actually means, or (2) evil.

Matthew Barnett, Tamay Besiroglu and Ege Erdil complete their Face Heel Turn, with a fully Emergently Misaligned post (as in, presenting maximally evil vibes on purpose) that argues that the tech tree and path of human technology is inevitable so they’re going to automate all human jobs before someone else has the chance, with a halfhearted final note that This Is Good, Actually, it might cure cancer and what not.

The tech tree inevitable? Well, it is with that attitude. I would point out that yes, the tech tree is discovered, but as every player of such games knows you have choices on what order in which to explore the tree and many techs are dead ends or have alternative pathways, and are thus not required to move forward. Other times you can absolutely lock into something you don’t want or very much do want depending on how you navigate the early days, he types on a QWERTY keyboard using Windows 11.

Other fun interactions include Roon pointing out the Apollo Program wasn’t inevitable, to which they replied that’s true but the Apollo Program was useless.

In case it wasn’t obviously true about all this: That’s bait.

Nathan: Surely this could be used to justify any bad but profitable outcome? Someone will do it, so the question is whether we’re are involved. But many beneficial technologies have been paused for long periods (geoengineering, genetic engineering).

Jan Kulviet: This is a fine example of thinking you get when smart people do evil things and their minds come up with smart justifications why they are the heroes. Upon closer examination it ignores key inconvenient considerations; normative part sounds like misleading PR.

A major hole in the “complete technological determinism” argument is that it completely denies agency, or even the possibility that how agency operates at larger scales could change. Sure, humanity is not currently a very coordinated agent. But the trendline also points toward the ascent of an intentional stance. An intentional civilization would, of course, be able to navigate the tech tree.

(For a completely opposite argument about the very high chance of a “choice transition,” check https://strangecities.substack.com/p/the-choice-transition).

In practice, this likely boils down to a race. On one side are people trying to empower humanity by building coordination technology and human-empowering AI. On the other side are those working to create human-disempowering technology and render human labor worthless as fast as possible.

My guess is when people stake their careers and fortune and status on the second option, their minds will work really hard to not see the choice.

Also: at least to me, the normative part sounds heavily PR sanitized, with obligatory promises of “medical cures” but shiying away from explaining either what would be the role of humans in the fully automated economy, or the actual moral stance of the authors.

As far as I understand, at least one of the authors has an unusual moral philosophy such as not believing in consciousness or first-person experiences, while simultaneously believing that future AIs are automatically morally worthy simply by having goals. This philosophy leads them to view succession by arbitrary AI agents as good, and the demise of humans as not a big deal.

Seb Krier: I knew someone who was trained as a revolutionary guard in Iran and the first thing they told him was “everything we do is to accelerate the coming of the Imam of Time; no destruction is not worth this outcome.” When I hear (some) hyper deterministic Silicon Valley techies I feel a similar vibe. It’s wild how few of the “just do things” people actually believe in agency.

Of course the other ‘side’ – ossified, blobby, degrowth obsessed stagnstors who would crystallize time forever – is just as depressing, and a bigger issue globally. But that’s for another tweet.

I think Jan is importantly mistaken here about their motivation. I think they know full well that they are now the villains, indeed I think they are being Large Hams about it, due essentially to emergent misalignment and as a recruitment and publicity strategy.

I’m not saying that the underlying plan of automating work is evil. Reasonable people can argue that point either way and I don’t think the answer is obvious.

What I am saying is that they think it is evil, that it codes to them (along with most other people) as evil, and that their choice to not care and do it anyway – no matter to what degree they believe their rationalizations for doing so – is causing them to present as Obviously Evil in a troparific way.

New Claude advertising keeping it classy, seems like a step up.

Danielle Fong: during a time of great bluster, Claude’s undercase thinking cap at cafe is the kind of beautifully executed and understated brand execution that’s poised to thrive for a population otherwise drowning in bullshit. Beautifully done @anthropic. Taoist ☯️

Jackie Luo: a lot of people are pointing out the value of aesthetics and yes anthropic’s aesthetic is good but that’s not enough on its own—anthropic is putting forth a positive vision for a future with ai that vision permeates claude as a model and the branding just expands its reach

this campaign wouldn’t work for openai because their perspective on what they’re building is fundamentally different. They are not optimistic about humanity in this same way they’re designing a tool, not a thought partner, and every decision they make reflects that.

If you’re not sure what the answer to a question is, try asking Claude Sonnet 4.5 first!

Joel Selanikio: We haven’t banned self-driving cars. We’ve set guardrails so the tech could evolve safely.

So why are states banning AI-only health insurance denials, instead of helping the tech get better?

Tim: I hope your health insurance claim is denied by an AI chatbot one day and you have no way to appeal. Then you’ll face the obvious reality everyone else can see you’re willfully ignoring.

Joel Selanikio: At least this guy didn’t wish that I was hit by a self-driving car!

I see the obvious appeal of letting AIs make insurance claim judgments. Indeed, I presume that soon most claims will involve an AI strongly suggesting and justifying either an acceptance or refusal, and the ultimate decision usually being a formality. That formality is still important, some actual human needs to take responsibility.

I love that this is his image, with a giant green arrow pointing towards ‘banned.’ Guess what most people think would be banned if we allowed AI review of claims?

Tyler Cowen declares his new favorite actress.

Tyler Cowen: Tilly Norwood is the actress I most want to see on the big screen, or perhaps the little screen, if she gets her own TV show. She is beautiful, but not too intimidating. She has a natural smile, and is just the right amount of British—a touch exotic but still familiar with her posh accent. Her Instagram has immaculate standards of presentation.

Tilly Norwood doesn’t need a hairstylist, has no regrettable posts, and if you wish to see a virgin on-screen, this is one of your better chances. That’s because she’s AI.

He’s kidding. I think. Reaction was what you might expect.

Deloitte refunds government $440k after it submitted a report partly generated with AI that was littered with errors including three nonexistent academic references and a quote from a Federal Court judgment.

The current state of play in Europe:

Who needs an AI in order to vibe code?

Miles Brundage:

Jack Clark: this is just my talk from The Curve, but better because it is a meme

Discussion about this post

AI #137: An OpenAI App For That Read More »

isps-created-so-many-fees-that-fcc-will-kill-requirement-to-list-them-all

ISPs created so many fees that FCC will kill requirement to list them all

The FCC was required by Congress to implement broadband-label rules, but the Carr FCC says the law doesn’t “require itemizing pass through fees that vary by location.”

“Commenters state that itemizing such fees requires providers to produce multiple labels for identical services,” the FCC plan says, with a footnote to comments from industry groups such as USTelecom and NCTA. “We believe, consistent with commenters in the Delete, Delete, Delete proceeding, that itemizing can lead to a proliferation of labels and of labels so lengthy that the fees overwhelm other important elements of the label.”

In a blog post Monday, Carr said his plan is part of a “focus on consumer protection.” He said the FCC “will vote on a notice that would reexamine broadband nutrition labels so that we can separate the wheat from the chaff. We want consumers to get quick and easy access to the information they want and need to compare broadband plans (as Congress has provided) without imposing unnecessary burdens.”

ISPs would still be required to provide the labels, but with less information. The NPRM said that eliminating the rules targeted for deletion will not “change the core label requirements to display a broadband consumer label containing critical information about the provider’s service offerings, including information about pricing, introductory rates, data allowances, and performance metrics.”

ISPs said listing fees was too hard

In 2023, five major trade groups representing US broadband providers petitioned the FCC to scrap the list-every-fee requirement before it took effect. Comcast told the commission that the rule “impose[s] significant administrative burdens and unnecessary complexity in complying with the broadband label requirements.”

Rejecting the industry complaints, then-Chairwoman Jessica Rosenworcel said that “every consumer needs transparent information when making decisions about what Internet service offering makes the most sense for their family or household. No one wants to be hit with charges they didn’t ask for or they did not expect.”

The Rosenworcel FCC’s order denying the industry petition pointedly said that ISPs could simplify pricing instead of charging loads of fees. “ISPs could alternatively roll such discretionary fees into the base monthly price, thereby eliminating the need to itemize them on the label,” the order said.

ISPs created so many fees that FCC will kill requirement to list them all Read More »

salesforce-says-it-won’t-pay-extortion-demand-in-1-billion-records-breach

Salesforce says it won’t pay extortion demand in 1 billion records breach

Salesforce says it’s refusing to pay an extortion demand made by a crime syndicate that claims to have stolen roughly 1 billion records from dozens of Salesforce customers.

The threat group making the demands began their campaign in May, when they made voice calls to organizations storing data on the Salesforce platform, Google-owned Mandiant said in June. The English-speaking callers would provide a pretense that necessitated the target connect an attacker-controlled app to their Salesforce portal. Amazingly—but not surprisingly—many of the people who received the calls complied.

It’s becoming a real mess

The threat group behind the campaign is calling itself Scattered LAPSUS$ Hunters, a mashup of three prolific data-extortion actors: Scattered Spider, LAPSuS$, and ShinyHunters. Mandiant, meanwhile, tracks the group as UNC6040, because the researchers so far have been unable to positively identify the connections.

Earlier this month, the group created a website that named Toyota, FedEx, and 37 other Salesforce customers whose data was stolen in the campaign. In all, the number of records recovered, Scattered LAPSUS$ Hunters claimed, was “989.45m/~1B+.” The site called on Salesforce to begin negotiations for a ransom amount “or all your customers [sic] data will be leaked.” The site went on to say: “Nobody else will have to pay us, if you pay, Salesforce, Inc.” The site said the deadline for payment was Friday.

In an email Wednesday, a Salesforce representative said the company is spurning the demand.

Salesforce says it won’t pay extortion demand in 1 billion records breach Read More »

logitech-will-brick-its-$100-pop-smart-home-buttons-on-october-15

Logitech will brick its $100 Pop smart home buttons on October 15

In another loss for early smart home adopters, Logitech has announced that it will brick all Pop switches on October 15.

In August of 2016, Logitech launched Pop switches, which provide quick access to a range of smart home actions, including third-party gadgets. For example, people could set their Pop buttons to launch Philips Hue or Insteon lighting presets, play a playlist from their Sonos speaker, or control Lutron smart blinds. Each button could store three actions, worked by identifying smart home devices on a shared Wi-Fi network, and was controllable via a dedicated Android or iOS app. The Pop Home Switch Starter Pack launched at $100, and individual Pop Add-on Home Switches debuted at $40 each.

A company spokesperson told Ars Technica that Logitech informed customers on September 29 that their Pop switches would soon become e-waste. According to copies of the email shared via Reddit, Logitech’s notice said:

As of October 15, 2025, your POP button(s) and the connected hub will no longer be supported and will lose all functionality.

As an attempt at compensation, Logitech gave affected customers a coupon for 15 percent off some Logitech products, including its Ultimate Ears speakers. The coupon is only valid in the US until March 31, 2026, and doesn’t apply to Logitech’s Pro or RS racing wheels for gaming, videoconferencing products, its Logitech for Business line, or “newly released products,” according to the email.

Logitech’s neglected smart home

Logitech’s spokesperson didn’t respond to Ars’ questions regarding e-waste, the short cancellation notice, or whether Pop button owners can continue using the devices locally after October 15.

“For close to a decade we have been maintaining the POP ecosystem, but as technology evolves, we have made the decision to end support for this device,” Logitech’s representative told Ars, repeating messaging from the email sent to customers.

Logitech will brick its $100 Pop smart home buttons on October 15 Read More »

google-pixel-10-pro-fold-review:-the-ultimate-google-phone

Google Pixel 10 Pro Fold review: The ultimate Google phone


Google delivers another phone that is slightly better than its predecessor—is that enough?

Pixel 10 Pro Fold flexed

The Pixel 10 Pro Fold is a sleek piece of hardware. Credit: Ryan Whitwam

The Pixel 10 Pro Fold is a sleek piece of hardware. Credit: Ryan Whitwam

When the first foldable phones came along, they seemed like a cool evolution of the traditional smartphone form factor and, if they got smaller and cheaper, like something people might actually want. After more than five years of foldable phones, we can probably give up on the latter. Google’s new Pixel 10 Pro Fold retains the $1,800 price tag of last year’s model, and while it’s improved in several key ways, spending almost two grand on any phone remains hard to justify.

For those whose phones are a primary computing device or who simply love gadgets, the Pixel 10 Pro Fold is still appealing. It offers the same refined Android experience as the rest of the Pixel 10 lineup, with much more screen real estate on which to enjoy it. Google also improved the hinge for better durability, shaved off some bezel, and boosted both charging speed and battery capacity. However, the form factor hasn’t taken the same quantum leap as Samsung’s latest foldable.

An iterative (but good) design

The Pixel 10 Pro Fold doesn’t reinvent the wheel—it looks and feels almost exactly like last year’s foldable, with a few minor tweaks centered around a new “gearless” hinge. Dropping the internal gears allegedly helps make the mechanism twice as durable. Google claims the Pixel 10 Pro Fold’s hinge will last for more than 10 years of folding and unfolding.

Specs at a glance: Google Pixel 10 series
Pixel 10 ($799) Pixel 10 Pro ($999) Pixel 10 Pro XL ($1,199) Pixel 10 Pro Fold ($1,799)
SoC Google Tensor G5  Google Tensor G5  Google Tensor G5  Google Tensor G5
Memory 12GB 16GB 16GB 16GB
Storage 128GB / 256GB 128GB / 256GB / 512GB 128GB / 256GB / 512GB / 1TB 256GB / 512GB / 1TB
Display 6.3-inch 1080×2424 OLED, 60-120 Hz, 3,000 nits 6.3-inch 1280×2856 LTPO OLED, 1-120 Hz, 3,300 nits 6.8-inch 1344×2992 LTPO OLED, 1-120 Hz, 3,300 nits External: 6.4-inch 1080×2364 OLED, 60-120 Hz, 3,000 nits; Internal: 8-inch 2076×2152 LTPO OLED, 1-120 Hz, 3,000 nits
Cameras 48 MP wide with Macro

Focus, F/1.7, 1/2-inch sensor; 13 MP ultrawide, f/2.2, 1/3.1-inch sensor;

10.8 MP 5x telephoto, f/3.1, 1/3.2-inch sensor; 10.5 MP selfie, f/2.2
50 MP wide with Macro

Focus, F/1.68, 1/1.3-inch sensor; 48 MP ultrawide, f/1.7, 1/2.55-inch sensor;

48 MP 5x telephoto, f/2.8, 1/2.55-inch sensor; 42 MP selfie, f/2.2
50 MP wide with Macro

Focus, F/1.68, 1/1.3-inch sensor; 48 MP ultrawide, f/1.7, 1/2.55-inch sensor;

48 MP 5x telephoto, f/2.8, 1/2.55-inch sensor; 42 MP selfie, f/2.2
48 MP wide, F/1.7, 1/2-inch sensor; 10.5 MP ultrawide with Macro Focus, f/2.2, 1/3.4-inch sensor;

10.8 MP 5x telephoto, f/3.1, 1/3.2-inch sensor; 10.5 MP selfie, f/2.2 (outer and inner)
Software Android 16 Android 16 Android 16 Android 16
Battery 4,970 mAh, up to 30 W wired charging, 15 W wireless charging (Pixelsnap) 4,870 mAh, up to 30 W wired charging, 15 W wireless charging (Pixelsnap) 5,200 mAh, up to 45 W wired charging, 25 W wireless charging (Pixelsnap) 5,015 mAh, up to 30 W wired charging, 15 W wireless charging (Pixelsnap)
Connectivity Wi-Fi 6e, NFC, Bluetooth 6.0, sub-6 GHz and mmWave 5G, USB-C 3.2 Wi-Fi 7, NFC, Bluetooth 6.0, sub-6 GHz and mmWave 5G, UWB, USB-C 3.2 Wi-Fi 7, NFC, Bluetooth 6.0, sub-6 GHz and mmWave 5G, UWB, USB-C 3.2 Wi-Fi 7, NFC, Bluetooth 6.0, sub-6 GHz and mmWave 5G, UWB, USB-C 3.2
Measurements 152.8 height×72.0 width×8.6 depth (mm), 204 g 152.8 height×72.0 width×8.6 depth (mm), 207 g 162.8 height×76.6 width×8.5 depth (mm), 232 g Folded: 154.9 height×76.2 width×10.1 depth (mm); Unfolded: 154.9 height×149.8 width×5.1 depth (mm); 258 g
Colors Indigo

Frost

Lemongrass

Obsidian
Moonstone

Jade

Porcelain

Obsidian
Moonstone

Jade

Porcelain

Obsidian
Moonstone

Jade

While the new phone is technically a fraction of a millimeter thicker, it’s narrowed by a similar amount. You likely won’t notice this, nor will the 1g in additional mass register. You may, however, spot the slimmer bezels and hinge. And that means cases for the 2024 foldable are just a fraction of a millimeter from fitting on the Pixel 10 Pro Fold. It does fit better in your hand, though.

Pixel 10 Pro Fold side

The Pixel is on the thick side for 2025, but this was record-setting thinness last year.

The Pixel is on the thick side for 2025, but this was record-setting thinness last year.

Thanks to the gearless hinge, the Pixel 10 Pro Fold the first foldable with full IP68 certification for water and dust resistance. The hinge feels extremely smooth and sturdy, but it’s a bit stiffer than we’ve seen on most foldables. This might change over time, but it’s a little harder to open and close out of the box. Samsung’s Z Fold 7 is thinner and easier to fold, but the hinge doesn’t open to a full 180 degrees like the Pixel does.

The new foldable also retains the camera module design of last year’s phone—it’s off-center on the back panel, a break from Google’s camera bar on other Pixels. The Pixel 10 Pro Fold, therefore, doesn’t lie flat on tables and will rock back and forth like most other phones. However, it does have the Qi2 magnets like in the cheaper phones. There are various Maglock kickstands and mounting rings that will attach to the back of the phone if you want to prop it up on a surface.

Pixel 10 Pro Fold and Z Fold 7

The Pixel 10 Pro Fold (left) and the Galaxy Z Fold 7 (right) both have 8-inch displays, but the Pixel is curvier.

Credit: Ryan Whitwam

The Pixel 10 Pro Fold (left) and the Galaxy Z Fold 7 (right) both have 8-inch displays, but the Pixel is curvier. Credit: Ryan Whitwam

The power and volume buttons are on the right edge in the same location as last year. The buttons are stable and tactile when pressed, and there’s a fingerprint sensor in the power button. It’s as fast and accurate as any capacitive sensor on a phone today. The aluminum frame and the buttons have the same matte finish, which differs from the glossy look of the other Pro Pixels. The more grippy matte texture is preferable for a phone you need to fold and unfold throughout the day.

Thanks to the modestly slimmer bezels, Google equipped the phone with a 6.4-inch external screen, slightly larger than the 6.3-inch panel on last year’s Fold. The 120 Hz OLED has a respectable 1080p resolution, and the brightness peaks around 3,000 nits, making it readable in bright outdoor light.

Pixel 10 Pro Fold and Pixel 9 Pro Fold

The Pixel 10 Pro Fold (left) has a more compact hinge and slimmer bezels compared to the Pixel 9 Pro Fold (right).

The Pixel 10 Pro Fold (left) has a more compact hinge and slimmer bezels compared to the Pixel 9 Pro Fold (right).

The Pixel 10 Pro Fold has a big 8-inch flexible OLED inside, clocking in at 2076×2152 pixels and 120Hz. It gets similarly bright, but the plastic layer is more reflective than the Gorilla Glass Victus 2 on the cover screen. While the foldable screen is legible, it’s not as pleasant to use outside as high-brightness glass screens.

Like all foldable screens, it’s possible to damage the internal OLED if you’re not careful. On the other hand, the flexible OLED is well-protected when the phone is closed—there’s no gap between the halves, and the magnets hold them together securely. There’s a crease visible in the middle of the screen, but it’s slightly improved from last year’s phone. You can see it well from some angles, but you get used to it.

Pixel 10 Pro Fold keyboard glamor

The Jade colorway looks great.

Credit: Ryan Whitwam

The Jade colorway looks great. Credit: Ryan Whitwam

While the flat Pixel 10 phones have dropped the physical SIM card slot, the Pixel 10 Pro Fold still has one. It has moved to the top this year, but it seems like only a matter of time before Google removes the slot in foldables, too. For the time being, you can move a physical SIM card to the Fold, transfer to eSIM, or use a combination of physical and electronic SIMs.

Google’s take on big Androids

Google’s version of Android is pretty refined these days. The Pixel 10 Pro Fold uses the same AI-heavy build of Android 16 as the flat Pixels. That means you can expect old favorites like Pixel Screenshots, Call Screen, and Magic Compose, along with new arrivals like Magic Cue and Pixel Journal. One thing you won’t see right now is the largely useless Daily Brief, which was pulled after its launch on the Pixel 10 so it could be improved.

Google’s expanded use of Material 3 Expressive theming is also a delight. The Pixel OS has a consistent, clean look you don’t often see on Android phones. Google bundles almost every app it makes on this phone, but you won’t see any sponsored apps, junk games, or other third-party software cluttering up the experience. In short, if you like the vibe of the Pixel OS on other Pixel 10 phones, you’ll like it on the Pixel 10 Pro Fold. We’ve noted a few minor UI glitches in the launch software, but there are no show-stopping bugs.

Pixel 10 Pro Fold split-screen

Multitasking on foldables is a snap.

Credit: Ryan Whitwam

Multitasking on foldables is a snap. Credit: Ryan Whitwam

The software on this phone goes beyond the standard Pixel features to take advantage of the folding screen. There’s a floating taskbar that can make swapping apps and multitasking easier, and you can pin it on the screen for even more efficiency. The Pixel 10 Pro Fold also supports saving app pairs to launch both at once in split-screen.

Google’s multi-window system on the Fold isn’t as robust as what you get with Samsung, though. For example, split-screen apps open in portrait mode on the Pixel, and if you want them in landscape, you have to physically rotate the phone. On Samsung foldables, you can move app windows around and change the orientation however you like—there’s even support for floating app windows and up to three windowed apps. Google reserves floating windows for tablets, none of which it has released since the Pixel Tablet in 2023. It would be nice to see a bit more multitasking power to make the most of the Fold’s big internal display.

As with all of Google’s Pixels, the new foldable gets seven years of update support, all the way through 2032. You’ll probably need at least one battery swap to make it that long, but you might be more inclined to hold onto an $1,800 phone for seven years. Samsung also offers seven years of support, but its updates are slower and don’t usually include new features after the first year. Google rolls out new updates promptly every month, and updated features are delivered in regular Pixel Drops.

Almost the best cameras

Google may have fixed some of the drawbacks of foldables, but you’ll get better photos with flat Pixels. That said, the Pixel 10 Pro Fold is no slouch—it has a camera setup very similar to the base model Pixel 10 (and last year’s foldable), which is still quite good in the grand scheme of mobile photography.

Pixel 10 Pro Fold cameras

The cameras are unchanged from last year.

Credit: Ryan Whitwam

The cameras are unchanged from last year. Credit: Ryan Whitwam

The Pixel 10 Pro Fold sports a 48 MP primary sensor, a 10.5 MP ultrawide, and a 10.8 MP 5x telephoto. There are 10 MP selfie cameras peeking through the front and internal displays as well.

Like the other Pixels, this phone is great for quick snapshots. Google’s image processing does an admirable job of sharpening details and has extraordinary dynamic range. The phone also manages to keep exposure times short to help capture movement. You don’t have to agonize over exactly how to frame a shot or wait for the right moment to hit the shutter. The Pixel 10 Pro and Pro XL do all of this slightly better, but provided you don’t zoom too much, the Pixel 10 Pro Fold photos are similarly excellent.

Medium indoor light. Ryan Whitwam

The primary sensor does better than most in dim conditions, but this is where you’ll notice limitations compared to the flat Pro phones. The Fold’s smaller image sensor can’t collect as much light, resulting in longer exposures. You’ll notice this most in Night Sight shots.

The telephoto sensor is only 10.8 MP compared to 48 MP on the other Pro Pixels. So images won’t be as sharp if you zoom in, but the normal framing looks fine and gets you much closer to your subject. The phone does support up to 20x zoom, but going much beyond 5x begins to reveal the camera’s weakness, and even Google’s image processing can’t hide that. The ultrawide camera is good enough for landscapes and wide group shots, but don’t bother zooming in. It also has autofocus for macro shots.

The selfie cameras are acceptable, but you don’t have to use them. As a foldable, this phone allows you to use the main cameras to snap selfies with the external display as a viewfinder. The results are much better, but the phone is a bit awkward to hold in that orientation. Google also added a few more camera features that complement the form factor, including a split-screen camera roll similar to Samsung’s app and a new version of the Made You Look cover screen widgets.

The Pixel 10 Pro Fold can leverage generative AI in several imaging features, so it has the same C2PA labeling as the other Pixels. We’ve seen this “AI edited” tag appear most often on images from the flat Pixels that are zoomed beyond 20x, so you likely won’t end up with any of those on the Fold. However, features like Add Me and Best Take will get the AI labeling.

The Tensor tension

This probably won’t come as a surprise, but the Tensor G5 in the Pixel 10 Pro Fold performs identically to the Tensor in other Pixel 10 phones. It is marginally faster across the board than the Tensor G4, but this isn’t the huge leap people hoped for with Google’s first TSMC chip. While it’s fast enough to keep the phone chugging, benchmarks are not its forte.

Pixel 10 Pro Fold in hand

Pixel 10 Pro Fold hinge has been redesigned.

Credit: Ryan Whitwam

Pixel 10 Pro Fold hinge has been redesigned. Credit: Ryan Whitwam

Across all our usual benchmarks, the Tensor G5 shows small gains over last year’s Google chip, but it’s running far behind the latest from Qualcomm. We expect that gap to widen even further when Qualcomm updates its flagship Snapdragon line in a few months.

The Tensor G5 does remain a bit cooler under load than the Snapdragon 8 Elite, losing only about 20 percent to thermal throttling. So real-world gaming performance on the Pixel 10 Pro Fold is closer to Qualcomm-based devices than the benchmark numbers would lead you to believe. Some game engines behave strangely on the Tensor’s PowerVR GPU, though. If mobile gaming is a big part of your usage, a Samsung or OnePlus flagship might be more your speed.

Day-to-day performance with the Pixel 10 Pro Fold is solid. Google’s new foldable is quick to register taps and open apps, even though the Tensor G5 chip doesn’t offer the most raw speed. Even on Snapdragon-based phones like the Galaxy Z Fold 7, the UI occasionally hiccups or an animation gets jerky. That’s a rarer occurrence on the Pixel 10 Pro Fold.

One of the biggest spec bumps is the battery—it’s 365 mAh larger, at 5,015 mAh. This finally puts Google’s foldables in the same range as flat phones. Granted, you will use more power when the main display is unfurled, and you should not expect a substantial increase in battery life generally. The power-hungry Tensor and increased background AI processing appear to soak up most of the added capacity. The Pixel 10 Pro Fold should last all day, but there won’t be much leeway.

The Pixel 10 Pro Fold does bring a nice charging upgrade, boosting wired speeds from 21 W to 30 W with a USB-PD charger that supports PPS (as most now do). That’s enough for a 50 percent charge in about half an hour. Wireless charging is now twice as fast, thanks to the addition of Qi2 support. Any Qi2-certified charger can hit those speeds, including the Google Pixelsnap charger. But the Fold is limited to 15 W, whereas the Pixel 10 Pro XL gets 25 W over Qi2. It’s nice to see an upgrade here, but all of Google’s phones should charge faster than they do.

Big phone, big questions

The Pixel 10 Pro Fold is better than last year’s Google foldable, and that means there’s a lot to like. The new hinge and slimmer bezels make the third-gen foldable a bit easier to hold, and the displays are fantastic. The camera setup, while a step down from the other Pro Pixels, is still one of the best you can get on a phone. The addition of Qi2 charging is much appreciated, too. And while Google has overloaded the Pixels with AI features, more of them are useful compared to those on the likes of Samsung, Motorola, or OnePlus.

Pixel 10 Pro Fold and Pixel 10 Pro

Left: Pixel 10 Pro Fold, Right: Pixel 10 Pro.

Credit: Ryan Whitwam

Left: Pixel 10 Pro Fold, Right: Pixel 10 Pro. Credit: Ryan Whitwam

That’s all great, but these are relatively minor improvements for an $1,800 phone, and the competition is making great strides. The Pixel 10 Pro Fold isn’t as fast or slim as the Galaxy Z Fold 7, and Samsung’s multitasking system is much more powerful. The Z Fold 7 retails for $200 more, but that distinction hardly matters as you close in on two grand for a smartphone. If you’re willing to pay $1,800, going to $2,000 isn’t much of a leap.

Pixel 10 Pro Fold back in hand

It’s the size of a normal phone when closed.

Credit: Ryan Whitwam

It’s the size of a normal phone when closed. Credit: Ryan Whitwam

The Pixel 10 Pro Fold is the ultimate Google phone with some useful AI features, but the Galaxy Z Fold 7 is a better piece of hardware. Ultimately, the choice depends on what’s more important to you, but Google will have to move beyond iterative upgrades if it wants foldables to look like a worthwhile upgrade.

The good

  • Redesigned hinge and slimmer bezels
  • Huge, gorgeous foldable OLED screen
  • Colorful, attractive Material 3 UI
  • IP68 certification
  • Includes Qi2 with magnetic attachment
  • Seven years of update support
  • Most AI features run on-device for better privacy

The bad

  • Cameras are a step down from other Pro Pixels
  • Tons of AI features you probably won’t use
  • Could use more robust multitasking
  • Tensor G5 still not benchmark king
  • High $1,800 price

Photo of Ryan Whitwam

Ryan Whitwam is a senior technology reporter at Ars Technica, covering the ways Google, AI, and mobile technology continue to change the world. Over his 20-year career, he’s written for Android Police, ExtremeTech, Wirecutter, NY Times, and more. He has reviewed more phones than most people will ever own. You can follow him on Bluesky, where you will see photos of his dozens of mechanical keyboards.

Google Pixel 10 Pro Fold review: The ultimate Google phone Read More »

chemistry-nobel-prize-awarded-for-building-ordered-polymers-with-metal

Chemistry Nobel prize awarded for building ordered polymers with metal

Unlike traditional polymers, this structure allows MOFs to have open internal spaces with a well-defined size, which can allow some molecules to pass through while filtering out others. In addition, the presence of metals provides for interesting chemistry. The metals can serve as catalysts or preferentially bind to one molecule within a mixture.

Knowing what we know now, it all seems kind of obvious that this would work. But when Robson started his work at the University of Melbourne, the few people who thought about the issue at all expected that the molecules he was building would be unstable and collapse.

The first MOF Robson built used copper as its metal of choice. It was linked to an organic molecule that retained its rigid structure through the presence of a benzene ring, which doesn’t bend. Both the organic molecule and the copper could form four different bonds, allowing the structure to grow by doing the rough equivalent of stacking a bunch of three-sided pyramids—a conscious choice by Robson.

Image of multiple triangular chemicals stacked on top of each other, forming a structure with lots of open internal spaces.

The world’s first MOF, synthesized by Robson and his colleagues. Credit: Johan Jarnestad/The Royal Swedish Academy of Sciences

In this case, however, the internal cavities remained filled by the solvent in which the MOF was formed. But the solvent could move freely through the material. Still, based on this example, Robson predicted many of the properties that have since been engineered into different MOFs: the ability to retain their structure even after solvents are removed, the presence of catalytic sites, and the ability of MOFs to act as filters.

Expanding the concept

All of that might seem a very optimistic take for someone’s first effort. But the measure of Robson’s success is that he convinced other chemists of the potential. One was Susumu Kitagawa of Kyoto University. Kitagawa and his collaborators built a MOF that had large internal channels that extended the entire length of the material. Made in a watery solution, the MOF could be dried out and have gas flow through it, with the structure retaining molecules like oxygen, nitrogen, and methane.

Chemistry Nobel prize awarded for building ordered polymers with metal Read More »

after-rfk-jr.’s-shenanigans,-covid-shot-access-will-be-a-lot-like-last-year

After RFK Jr.’s shenanigans, COVID shot access will be a lot like last year

In an interview with Ars Technica in August, Brigid Groves, vice president of professional affairs for the American Pharmacists Association (APhA) signaled that efforts to limit access to COVID-19 vaccines is concerning to APhA, which is the leading organization representing pharmacists.

“We are concerned about that because the data and evidence point to the fact that this vaccine is safe and effective for [young, otherwise healthy] patients,” Groves said. “So, to suddenly arbitrarily limit that is very concerning to us.”

And, with the CDC’s permissive recommendations, pharmacies are not limiting them. Representatives for both CVS and Walgreens told The Washington Post that they would not require patients under 65 to prove they have an underlying condition to get a COVID-19 vaccine. CVS won’t ask you to self-attest to having a condition, and Walgreens also said that it won’t require any proof.

“In simplest terms, if a patient wants to get the vaccine, they’ll get it,” Amy Thibault, a CVS spokesperson, told the Post.

With the shared decision-making, there may be extra forms about risks and benefits that might take an extra few minutes, but it should otherwise be just like past years.

On Tuesday, this reporter was able to easily book same-day appointments for an updated COVID-19 vaccine at local CVS and Walgreens pharmacies in North Carolina, without attesting to any medical conditions.

Children

Shots for younger children could be trickier: While adults and older children can visit their pharmacy and get vaccinated relatively easily, younger children (particularly those under age 5) may have a harder time. Pharmacists typically do not vaccinate those younger children—which has always been the case—and parents will have to visit the pediatrician.

Pediatricians, like pharmacists, are likely to be supportive of broad access to the shots. The American Academy of Pediatrics has said that all children should have access. The AAP also specifically encourages children under age 2 and children with underlying conditions to get vaccinated, because those children are at higher risk of severe disease.

After RFK Jr.’s shenanigans, COVID shot access will be a lot like last year Read More »

trump’s-epa-sued-for-axing-$7-billion-solar-energy-program

Trump’s EPA sued for axing $7 billion solar energy program

The Environmental Protection Agency was sued Wednesday over an allegedly politically motivated decision to end a program that Congress intended to help low-income and disadvantaged communities across the US save money on electricity bills through rooftop and community solar programs.

In their complaint, a group of plaintiffs who would have benefited from the EPA’s “Solar for All” program—including a labor union, several businesses, and a homeowner who cannot afford her electricity bills without it—accused the EPA of violating federal law and the Constitution by unlawfully terminating the program.

Solar for All was “expected to save an estimated $350 million annually on energy bills during and after the five-year program, providing energy bill relief for more than 900,000 low-income and disadvantaged households,” plaintiffs noted. Additionally, it was “expected to secure 4,000 megawatts of new solar energy over five years and generate 200,000 new jobs.”

According to plaintiffs, EPA Administrator Lee Zeldin illegally squashed the program after Congress repealed a statute in July that had sparked the program’s creation. However, lawmakers were clear when repealing the statute that only “unobligated” funds could be rescinded, plaintiffs argued, citing lawmakers who “repeatedly” stated that the repeal would not impact funding that had already been awarded.

In 2024, Congress obligated the EPA to award $7 billion in grants to recipients behind projects that would have created “hundreds of thousands of good-paying, high-quality jobs” and spared the average low-income family “about $400 each year on their electricity bills,” plaintiffs argued.

Allegedly Zeldin “arbitrarily” decided to ignore the “plain language” of the statute, plaintiffs alleged, waiting a month after the statute’s repeal to terminate the Solar for All program in August.

Plaintiffs noted that because Zeldin was distributing funds for weeks after the statute’s repeal, this indicated he understood the funding had not been rescinded. They accused Donald Trump’s EPA of violating the separation of powers by interfering to block congressionally awarded funds due to Trump’s disdain for solar energy—pointing to a Zeldin social media post that claimed that the “EPA no longer has the authority to administer the program or the appropriated funds to keep this boondoggle alive.”

Trump’s EPA sued for axing $7 billion solar energy program Read More »

medical-roundup-#5

Medical Roundup #5

Some amazing things are going on, not all of which involve mRNA, although please please those of you with the ability to do so, do your part to ensure that stays funded, either via investment or grants.

As for mRNA, please do what you can to help save it, so we can keep getting more headlines like ‘a new cancer vaccine just wiped out tumors’ even if it is sufficiently early that the sentence this time inevitably concludes ‘IN MICE.

Wait, what, you’re saying we might soon ‘mostly defeat’ heart disease?

Cremieux: It’s hard to oversell how big a discovery this is.

Heart disease is America’s #1 cause of death. With a combination of two drugs administered once every six months, it might be mostly defeated.

Just think about that how big this is! You will know your great-grandkids!

It is insanely optimistic that we have drugs that can reduce Lp(a) levels by [65%-]98% in trials right now.

If they succeed, they could help to crush heart disease and stroke.

Unlike LDL in general, Lp(a) levels are basically entirely genetic in origin and not open to lifestyle intervention. It is also widely accepted that the race differences are down to genes.

These drugs are examples of genetic discovery leading to a group difference fix via tech.

The future of medicine is still very bright.

I wouldn’t go that far. Even if these trials are successful, it seems unlikely we’re talking about ‘mostly defeat,’ although we could plausibly be at ‘greatly reduce.’ Which could still be worth several years of life expectancy across the board. If we could also similarly help with other major causes of death and aging, you’d see compounding gains, but without that aging still catches up with everyone.

Unfortunately, America under the current administration is making deep cuts in basic research funding that leads to advances like this. Hopefully AI can make up for that.

A new major study finds that alcohol causes cancer, so government worked to bury the study. Time and time again we get presented with the fact that small amounts of drinking correlate with improved health in various ways, fooling many into thinking a little alcohol is healthy.

As opposed to the reality, which is that alcohol is bad for you no matter what, but that inability to drink is highly correlated with alcoholism and other traits that go along with poor health, whereas ability to drink only in moderation is a good sign.

So drinking in moderation, which is only a small amount bad for your health, is still a good sign for your health if you are drinking in moderation. Whereas heavier drinking consistently looks bad, perhaps even worse than it is.

Trump himself is not fooled, and does not drink at all, as he has seen the dangers of alcoholism in his own family. A wise choice.

Dylan Scott: They broke out their findings by different drinking levels — from one drink per day to three — and focused on health outcomes that have been proven to be associated with alcohol use. Their big-picture conclusion:

Among the US population, the negative health effects of drinking alcohol start at low levels of consumption and begin to increase sharply the more a person drinks. A man drinking one drink per day has roughly a one in 1,000 chance of dying from any alcohol-related cause, whether an alcohol-associated cancer or liver disease or a drunk driving accident. Increase that to two drinks per day, and the odds increase to one in 25.

In that context, the report is a harrowing read: Alcohol use is associated with increased mortality for seven types of cancer — colorectal, breast cancer in women, liver, oral, pharynx, larynx, and esophagus. Risk for these cancers increases with any alcohol use and continues to grow with higher levels of use, the study’s authors concluded.

Women experience a higher risk of an alcohol-attributable cancer per drink consumed than men. Men and women who die from an alcohol-attributable cause die 15 years earlier on average.

A 4% chance of dying on average 15 years earlier is a big deal.

This is not the first time a Trump administration defied the data on this.

US officials always solicit expert opinion as they prepare a fresh set of dietary guidelines. The input is usually compiled into one massive report from a group of experts called the US Dietary Guidelines Advisory Committee and then submitted to the Department of Health and Human Services and the Department of Agriculture, the two agencies that produce the guidelines.

That was how the process went in 2020, and at that time, the subcommittee of researchers dedicated to alcohol (including Naimi) advised the government to reduce the recommended limit down to one drink per day for men, from two. The Trump administration ultimately decided not to follow the recommendation.

Montana passes SB 535 with broad bipartisan backing, further expands its ‘right to try’ rules, with a license path for experimental treatment centers to administer any drug that got through Phase I trials, to anyone who wants it. This is The Way.

Alex Tabarrok is correct that we could greatly improve healthcare if we allowed telemedicine across state lines. As long as a doctor is licensed where the doctor is physically located, it shouldn’t matter where the patient is located. The best part is that this could be done with an admin rules change of two words.

A call to make statins be sold over the counter (OTC). This seems obviously correct, even if you are highly skeptical that the correlations cited here imply causation, or that they imply that intervening via statin cause this causation without important side effects. That’s a decision people can make for themselves at this point. But then a whole host of things should be OTC at this point that aren’t.

A new embryo selection company, Herasight has launched that is allowing users to select for IQ. They claim that with as few as 10 embryos you can already go from an expected IQ of 100 to a new average of 107. Or you can do things like go from 45% chance of Type 2 Diabetes to 25%.

Alex Young: I’ve been working with an IVF startup, @herasight, that has already screened hundreds of embryos. Today we come out of stealth with a paper showing that our predictors for 17 diseases — validated within-family — beat the competition, with improved performance in non-Europeans.

In our paper, we detail our polygenic scores (PGS) for 17 diseases using a custom meta-analysis. We used state-of-the-art methods to create PGSs based on 7.3M SNPs. Our most predictive PGSs explained ~20% of the variance in liability for prostate cancer and type-II diabetes.

In practice what happens is you get your 5, 10 or 20 embryos, they profile each one, and you are choosing based on a variety of traits. What do you actually care about most? You are about to find out.

Kitten: Within a few years they’re going to be able to show you a very accurate picture of what each baby will look like as an adult.

That will trump every other consideration by a large margin.

I agree that this technology is coming. I agree it will matter to people. And why shouldn’t it matter, at least from a selfish point of view? It also might be a better way to select for other good things than you might think, as in general health and other positive traits are correlated with beauty. I do not think it will be anything like ‘this overrides everything else.’

Even the graceful failure mode for things like this is a really big deal.

Mason: I think this just ignores the realities of IVF tbh

20 healthy embryos is, for most people, 3-5 IVF cycles

Success for a single embryo transfer is 40-60%.

“Well, your dad and I wanted a boy and you were the second highest IQ male after Embryo 6 failed to implant.”

Yeah, these numbers are for mothers <35. It's crazy how many people think IVF is a safety net for aging out of your fertility window when you actually need to start the process quite young in order for it to be likely to work.

I mean that sounds pretty great? You got your choice of gender and a couple of IQ points, while presumably also dodging a variety of potential genetic disorders. That’s a pretty good haul.

What you actually get is not maximizing on one to two traits, although you do have that option. What you get is to do a general maximization over many traits. That is a lot more valuable if you are making reasonable decisions.

Mason is however making the very important point that if you want to use IVF and have confidence it will work at all, let alone confidence it will give you selection, you need to do it early. The younger you are, the better all of this will on average go, until such time as we figure out how to generate new eggs (which is plausibly only a few years away).

As Gene Smith points out in this thread, academics are super against all of this.

Gene Smith: We’re in this insane world right now where rich parents are going through IVF just to do embryo selection, and middle class parents who are ALREADY doing IVF are being told by doctors that embryo selection doesn’t work or is unethical.

If you’ve been to conferences on embryo selection or ones where the topic is discussed, you will realize the truth is basically the opposite. Most academic discussion of this is dominated by ethicists talking about how problematic it is or academics talking about its limitations.

And make no mistake, these discussions matter. They’re the main reason why even doctors who understand embryo selection rarely recommend it to their patients, even when it could significantly reduce the odds of a patient passing along a disease to their child.

Among the academics that care about this, a large majority are opposed to embryo selection for ideological reasons. There hasn’t been all that much quantitative analysis done on it. And what has been done often fails to capture how embryo selection is actually done in industry.

A lot of work has focused on selection on a single trait, which is easier to model, but not how selection actually works in practice. And many of the conclusions about its efficacy rely on outdated predictors which have improved substantially since publication.

The effect size of embryo selection for IQ has nearly tripled since the publication of this paper, which was still being cited relatively recently.

He also reminds us that we are spending all our gene editing resources on rare diseases where we can stick the government gigantic bills and the ‘ethicists’ stop trying to ban you from helping, whereas the places where most of the value lies are complete ignored, even by people who would directly benefit.

Gene Smith: The situation in the gene editing field right now is kind of unbelievable. We’ve spent over a decade throwing billions of dollars at the gene editing for rare diseases and we still can’t get the editors inside most of the cell types we need to modify.

Everything is either for the liver or the eye or the bone marrow. We can’t get to the brain. We can’t get to the lungs. We can’t get to the heart. And we can barely fit a single editor into the best delivery vector so no one is even thinking about polygenic disease.

There is an extremely simple way to fix this: edit in embryos. All the stuff we can’t target right now because of delivery issues becomes almost trivial in embryos. You can literally stick a needle into the egg and squirt the editors in.

Alzheimer’s? Tractable. Diabetes? You can pretty much get rid of it. Heart disease? You can mostly make it a thing of the past. Breast cancer? Same deal.

The blame for this delay lies at the feet of academics. There are literally professors who go to conferences on gene editing and brag about how they’ve gotten governments to ban embryo gene editing.

[thread continues]

Bex: surely this is being done black budget and for the uber wealthy already. but I guess all that learning is hidden:(

Gene Smith: As someone working on the cutting edge of this field, I can tell you it’s definitely NOT. You underestimate how maddeningly myopic wealthy people can be about new tech. Fred Koch literally didn’t ask the cancer research center he funded to look into ways to treat his cancer.

The most cutting edge thing wealthy people are actually doing right now is embryo selection a la Herasight and Orchid.

I mean, it’s not that simple, of course it is not that simple, but it is a little bit that simple?

If you have the option to use the technology we do have, it seems crazy not to use it. Noor Siddiqui is being too righteous and superior about it even for me here, but if it really does cost as little as $2,500 to get selection over potential embryos, that is a better deal than essentially any remotely optional interventions you have access to after birth. The replies to such proposals are full of people saying how horrible all this is and having instinctual disgust and purity reactions, and calling everyone involved and the proposal itself various names, all of it patently absurd without any actual justifications.

I do strongly agree with Vitalik Buterin that while we should be very much in favor of embryo selection and most of the attacks on it are deranged, it is counterproductive and wrong to strike back by saying that since it condemns your children to be worse off that not using selection is unethical.

Vitalik Buterin: If you publicly telegraph that if you win your fight for acceptance you will immediately follow it up with a fight for imposition, do not be surprised if people decide to oppose your fight for acceptance from day one with a vigor only reserved for resisting imposition.

We’ve been over this one many times. If that which is not forbidden is compulsory, then that which I do not want to be compulsory must be forbidden.

Oh no, all these GLP-1 drugs are going to prevent people from dying, and that could have a negative impact on pensions?

Eliezer does a survey, finds ~80% of those who tried GLP-1 drugs report it helped lots versus those that didn’t, roughly confirming industry claims. That’s a fantastic success rate.

Now we are potentially seeing Eli Lily have a pill that’s ~55% as effective as Ozempic in phase-3 trials, their stock was up 14% on the news. They also have Retatrutide coming in a year or two, which is claimed to be a more effective GLP-1 drug that also causes less muscle loss.

These marginal improvements make a huge practical difference. GLP-1s are now a lot like AI, in that we keep getting better versions and people’s impressions don’t update.

Unfortunately, it’s not always easy to get a line on the necessary supply.

Scott Alexander: Update on Ozempocalypse: some pharmacies have stopped selling compounded GLP-1 drugs, others continue, with various flimsy legal excuses. Cremieux has a guide (partly subscriber-only) on how to order and use cheap “research chemical” GLP-1 from from peptide companies. And the Trump administration cancelled a Biden initiative to make GLP-1 drugs available via insurance.

There were official shortages of GLP-1 drugs, which allowed compounders to make and sell those drugs cheaply, on the order of $300/month, in essentially unlimited quantities. Alas, there is now no longer an official shortage, so the price is shooting back up again (~$1k/month) and supply is harder to find. We really should find a way to buy out those patents and make these drugs available for low prices (ideally for free, with at most a minimal consultation) for whoever wants them.

Is it possible that if you combine a GLP-1 with anabolic agent Bimagrumab, you can lose fat without losing muscle? Eli Lily is in phase 2b of trying to find out.

Egan Peltan: Today, Lilly revealed the weight loss & composition at 48 weeks

Sema: -13.5% (72% Fat)

Bima: -8.6% (100% Fat)

Bima + Sema: -16.4% (>90% Fat)

Patients on bima+sema saw >40% decrease in body fat with no changein lean mass at 24W. Overall, bima and bima+sema were well tolerated No striking SAE imbalances.

What we don’t know works with GLP-1s is microdosing.

Eric Topol: The GLP-1 microdosing wellness and longevity craze without any data for dose or evidence of effect.

As usual this as worded confuses lack of formalized data and evidence for a lack of data and evidence. These are not the same thing. It is an entirely sensible strategy to experiment with different doses and to choose, with a drug that has big benefits, but also downsides, and is clearly safe. Self-experimentation can provide highly useful data and evidence, and clearly different people respond differently in terms of benefits and side effects at different doses.

Babb (from the WSJ article): What feels healthiest is taking the lowest amount that’s providing what I’m perceiving to be these health benefits.

The marketing involving microdosing does sound not great, especially its claims of other vague and hard to measure benefits.

There are still some potential safety issues where it would be good to have more confidence, I agree that we are not treating checking for such possibilities as seriously as we should given how many people are on these drugs. But also given how many people are on these drugs (about 6% of Americans right now), and how much many people are inherently suspicious that there is a catch, and how long people have been taking them for (including previously for diabetes) I am confident that if there was a showstopper problem it would have been found by now.

Andrew Rettek: Ozempic type drugs have been on the market for over seven years. It’s been about five years since 1 million Americans have been taking them. We have a pretty good idea of their side effects at this point. It’s longer than you thought, because people have been taking them for diabetes for years before they became popular for weight loss.

People instinctively think there must be a catch to GLP-1s. But for people in most circumstances there mostly isn’t one, at least not one of magnitude similar to the benefits, other than that for some people it doesn’t work?

GSV Bemusement Park: Look if you are trying to lose weight there is no reason to not be padding your efforts with a GLP-1 drug unless everyone in your family gets thyroid cancer.

If you buy smartly the cost is literally negative after accounting for the food you don’t eat, and you can always go back on if you regain weight.

If you follow me then your mutuals almost certainly include several helpful, literate people who will be thrilled to advise you.

You have nothing to lose but waistline.

Peter Hague: Ozempic skepticism seems entirely based on the animist-like idea that there is a conservation law for good fortune – all good things must have some equal cost imposed elsewhere. It’s not scientific skepticism, it’s superstition.

I’m really concerned too how “big pharma is trying to kill you/sterilise you with fake medicine” has gone from the fringes of Facebook mummy groups to the mainstream. It used to be the main complaint was they were greedy and withholding good medicine from sick people.

Dystopina GF: Anti-Ozempic people don’t understand that the average person isn’t supposed to be an iron-willed Übermensch. Average people simply reflect the quality of their society; they’re obese bc modernity is sick, not bc of some moral failing. Being thin should be effortless, the default!

A well-designed society is a system that makes good things as frictionless as possible. A poorly-designed society is one where it takes a Herculean effort to achieve the most basic elements of human health and fulfillment

John Pressman: The replies on this inspire serious misanthropy. The QT is straightforwardly correct and the replies make it obvious that all good things come from brutally, ruthlessly destroying the perverse costly signaling equilibria people naturally cover up utility with.

Of course, here I am saying that and then not taking them, but I am a bizarro case where I was able to get to my target through sheer willpower, and I have decades of experience dealing with essentially permanent hunger. That almost never works.

Andrew Rettek reports from a week discussing diet, exercise and GLP-1s at LessOnline and Manifest. Lots of people wanted to exercise, but felt they needed ‘demystifying’ of the gym and the general procedures involved. I very much feel this. Our on ramps suck, and mostly amount to ‘find a source and trust it.’ Which is way better than not exercising but doesn’t fill one with confidence or motivation – I’m spending a substantial portion of my time and willpower and energy on this, and different implementations differ a lot. Whereas for diet, Andrew observed people mostly weren’t interested, my guess is because they’ve already heard so many contradictory things.

GLP-1s served as a large scale experiment in drug compounding, and in sidestepping the FDA’s various usual regulatory requirements. The results were amazingly great, everything went smoothly. This is even more evidence for FDA Delenda Est, while noting that ‘fire a bunch of people at the FDA’ makes things worse rather than better. Removing requirements is good, but if there are going to be requirements it is necessary to be able to handle the paperwork and meetings promptly.

Going forward, the bigger question is: What happens if this actually works?

Cremieux: Eli Lilly just showed that you can lose tons of fat while barely losing any muscle using their activin type-II receptor inhibitor, bimagrumab.

We are approaching a golden era of weight loss, where everyone can easily be muscular and skinny.

Prepare for hordes of hot Americans.

Devon Eriksen: And I am predicting, here and now, that there will be a massive social pushback from people who would rather see their fellow Americans live out their lives sick, fat, and miserable than give up their just-universe fallacy.

At what point do we start to think of weight loss as fully treatable?

Shako: Increasingly in public I see obese people as glp1 deficient. I mean this seriously. There is a 200 pound 15 year old girl at this playground. It’s heartbreaking and we can treat it now.

Lots of people commenting diet and exercise work better. And yet vast amounts of Americans, often stupid and poor, die every year from preventable metabolic diseases. Do you think they just don’t know about diets?

Eliezer Yudkowsky: Tirzepatide doesn’t work on 20% of people, including me.

Shako: 🙏🏻 hope something does soon.

Not every case can be treated, since we have a non-response problem, and some people will run into side effects or risk factors. But a large majority of people with a problem still haven’t attempted treatment. Most of them should try. If you were already going to lose the weight on your own, you could do it easier with help?

What about the equilibrium, signaling and status arguments? That if we allow people to lose weight without willpower or personal virtue then that will make the signals harder to read and be Just Awful? Yeah, I don’t care.

Cremieux claims that it is a myth that yo-yo weight loss reduces muscle mass and makes you fatter. Andrew Rettek responds that the studies claiming to bust this were on wrestlers, which doesn’t apply to normal people, and I buy his explanation of the mechanism here, which is:

  1. If you lose weight without resistance training you’ll lose some muscle mass.

  2. If you gain weight without resistance training you won’t gain muscle mass.

  3. Thus, if you do both, yes, you end up worse off.

  4. But that’s a choice, and it’s fixable by attacking ‘without resistance training.’

There are also other exercise methods. The point is clear.

Weight loss helps resolve back problems and otherwise make things hurt less. Few appreciate how big a deal this is. This was a big game changer for me on its own.

Man Down Ted: I may have a bit less hair, but I have FAR fewer body aches and pains than I did 10 years ago.

A lot of that is due to lowering the weight my body has to carry, but at least some of it is due to the fact that I will just make myself do 10 mins of yoga to loosen up, any time.

Meanwhile, half my male friends near my age have thrown their back out this year. Possibly more than half wake up stiff and sore from sleep every day.

Dropping the weight is good for dating,, but it’s GREAT for your long-term health and self image. And not THAT hard.

That is, no one knows much about how to do it right.

We do know many ways in which one can do it wrong.

Cremieux points out that in general ‘nutrition beliefs are just-so stories.’

If there were two concrete specific things I would have said I was confident about, one of them would have been ‘vegetables mostly good’ and the other would be ‘sugar bad.’ I don’t think it’s bad enough to stop me all that much on either front, but I do buy that sugar is bad on the margin.

And yet when Americans cut sugar consumption (modestly by ~10%, note the y-axis) the obesity rate still mostly kept moving up.

He blames this mess on selection. Once sugar got blamed, choosing to consume more sugar became correlated with other health decisions, some of which matter. The associations between sugar and health only show up, he claims, after 2012. And he says this generalizes – certain people eat certain diets. He finishes with a survey of some of the other problems.

I am inclined to be kinder, and see the epistemic problems here as actually super hard. Nutrition is remarkably different for different people, doing proper studies on humans is extremely hard, there are distinct effects for short, medium and long terms, a lot of this is about how you psychologically respond, lots of details matter and are hard to measure and can be changed without you realizing, and so on.

As far as anyone can tell Aspartame, a zero calorie sweetener, seems benign. I agree that this is great news so long as no one forces anyone to use it. I think the people this most annoys are people who believe that it is not as good an experience as sugar, and don’t want their sugar taken away, either legally or by those ‘worried about their health.’ Also, since no one knows anything about nutrition, I wouldn’t assume that there isn’t a catch.

From what I’ve seen, they go wrong enough often enough that starting to mess with them probably doesn’t make sense if you aren’t already messing with them.

Do ketogenic diets dramatically raise risk of stroke and heart attack?

Christoffer Bausted Nilsen: “When the facts change, I change my mind. What do you do, sir?” (source)

The graphs show dramatically accelerated PAV (Percent Atheroma Volume) and NCPV (Non-Calcified Plaque Volume) in those on Keto diets.

As always, beware conclusions from intermediate endpoints, and also different people often have very different reactions and averages can be misleading. But I agree this does not look good.

Cremieux: I’m a little annoyed that I spent so many years thinking keto was obviously good.

Now that we have a lot of data, it turns out that the people who said “fat—especially saturated fat—clogs your arteries” were just totally correct.

I should not have believed the contrarians.

I don’t think we have enough evidence here to draw that conclusion either?

Here’s a counterargument from someone thoughtful and sympathetic to Keto:

William Eden (not endorsed): As someone who has recommended paleo/keto diets in the past, let me lay out my current thinking:

ApoB is directly related to atherosclerosis progression (experts are right)

*Somepeople who go on keto skyrocket ApoB, and yes I think keto is bad for those people. This is where ultimately empirical results overcame the same things he derides in this discussion (“mechanistic” reasoning to support the answer you’re hoping to see).

Human PCSK9 mutants with no CVD changed my mind – and are the basis for a class of very strong therapeutics! However… I still think mechanisms are real. Some are better established, others speculative. Good theories predict the (very complicated) data My favorite thinker on the subject is Peter over at Hyperlipid, though he’s almost impossible to understand.

Just because high sat fat keto increases heart disease doesn’t mean the answer is to eat huge amounts of polyunsaturated fats either. Yes, I am a seed oil disrespecter. Mechanistically I think it causes the liver to hang onto fat instead of circulating it, but we can do better.

This video from Peter Attia is just 10 minutes and lays out the good drug options available today, including ones with better risk/reward than statins (though a *low dosestatin I think is fine with CoQ10 supplementation).

(Btw I think it’s absolutely clear that LDL must play some role in the body or it wouldn’t exist. I don’t 100% know all of its functions, but when you look across countries, low LDL is seen in developing countries with lots of infectious disease and other cardiovascular issues)

In terms of my diet advice, I do think a higher-fat diet that dips occasionally in and out of ketosis is likely optimal for long term performance and health. If you get high ApoB, swap out sat fat for monounsaturated fat, and consider drugs. Keto for >>1 month I have concerns.

In terms of my epidemic advice, I do think contrarians should try harder to grapple with mainstream evidence instead of dismissing it. *Especiallythe really strong evidence, and being able to see it as such. That’s why short term diet studies didn’t get me, but PCSK9 did. Fin.

Nathan Cofnas defends Keto on multiple fronts while also offering warnings.

  1. He gives us the standard ‘correlation is not causation’ warnings that are at their peak in nutrition and a reminder that no one knows anything beyond that you need a basic set of nutrients and that different people are different.

  2. The key thing about Keto, as its fans envision it, is it is the ultimate case of never do anything by halves. The idea of Keto is that you force your body into ketosis. That means consistent hardcore restricting of carbs. Most people can’t do it.

  3. So almost any observational study is going to include a bunch of people who don’t actually do the thing and stay in ketosis, and being kind-of-keto doesn’t work.

  4. Remember, when evaluating diets, that people basically neve stick to one for very long. The majority of people who say they’re ‘vegetarian’ will then admit they’ve eaten meat in the past week.

  5. Keto is not optimized for longevity. You do it to feel and look better now, not to live for longer. Even if it has longevity costs, it might still be worth it. If you do want primarily longevity he suggests caloric restriction and pescetarianism, which is kind of the opposite of keto.

  6. Keto is the only diet that worked for him in terms of current health, but of course different things work for different people.

I find these defenses:

  1. Reasonable arguments.

  2. Clear indications that very few people should be trying to be on a keto diet.

For keto to make sense, even if you mostly buy the counterarguments, given the risks involved (over a range of possible worlds):

  1. You need to have extraordinary willpower and discipline, keeping in mind that most people who think they have it, don’t have it.

  2. You need to be willing to spend that, and give up carbs, and endure all the hedonic and social costs involved, to stick to keto for real.

  3. You need to prioritize short term health over longevity.

  4. You need to have a body where keto happens to be relatively beneficial for you.

  5. You keep avery close eye on various markers and your experience as you try it, and abort on the spot if an issue emerges, either in terms of your health or experience, or simply exposing that the willpower price is too damn high.

And that’s if you buy the counterarguments.

In practice, even if you think this is right for you to try, I’m willing to go ahead and say that on average you are wrong about that.

Congratulations to Allyson Taft on her transformation, including losing 120+ pounds in just under a year by walking a lot, drinking a lot of water and tracking calories.

Not everyone can do this. Not everyone should attempt this. And most people in the position to need to do it should probably take GLP-1s.

It was still very nice to see a clear example of someone pulling off the pure willpower route who wasn’t me. And it sounds like she didn’t even use GLP-1s.

One thing I would push back on very hard is the idea that there is something perverse and modern about the fact that if we eat until we feel satiated with reasonable access to a variety of foods then we will gain weight:

exfatloss: If you can maintain a healthy body weight by restricting your total food intake, but never give in to eat ad libitum because you know you’d gain fat, your satiety & appetite are broken. Until very recently, nobody needed to count calories to remain thin.

I’m not saying that staying thin did not get harder on many fronts, and it is entirely possible that there are other things making things harder on top of that, but the idea that in the past eating as much as you wanted of what you wanted wouldn’t have made you fat is Obvious Nonsense. Rich people from the past who could eat as much as they wanted, and didn’t care, totally got fat. Sure, people mostly stayed thin, but largely out of lack of opportunity.

Needing to control yourself to stay thin does not mean that your satiety is broken. You’re asking way too much of satiety. You’re asking it to give you the right highly calibrated signal in a radically different world than the one it was trained for.

This doesn’t require toxins or any particular villain beyond widespread variety and caloric availability, including many calorically dense foods, and greatly increased food variety and quality, and various cultural expectations of eating frequently. It is enough, and regardless of the original cause life is not fair.

Cremieux: Attempts to explain the obesity epidemic through contamination, toxins, conspiracies, seed oils, sugar, high-fructose corn syrup, lithium, or whatever else always strike me as annoying.

What we must explain is an increase of ~200-350 calories a day in energy balance. That’s all.

My answer is simple, and similar to Cremieux’s in the thread, that there was never really any robust natural ‘balance.’

For full disclosure, I say this as someone whose satiety has always been utterly broken to the point of uselessness. If left to my own devices I would put on insane amounts of weight very quickly. I might double my caloric intake. That’s what actually broken looks like. But don’t ask an evolutionary mechanism to give the right answer in a radically changed world, and don’t call it ‘broken’ when you have to adapt.

Yes, until recently that meant you had to work for it. That’s one option. The other? Good news, now we have Ozempic and other GLP-1s.

Key Smash Bandit: I started taking a million supplements and I do feel meaningfully better but the new problem is I have no idea which one did it, and taking all of them is expensive and annoying

Ryan Moulton: This is a fun task to try to design an experiment to resolve.

I would randomize to taking 50% each day, (ideally blinded, but whatever) and then record how I felt each day. Then after you have ~a month of data, fit a regression model from what you took to how you felt, probably also with at least terms for 1 day lag. Lean on L1 in lasso a lot, because you’re probably looking for a single thing rather than a large set.

Unfortunately there are what the economists call ‘long and variable lags’ in vitamin impact, both positive and negative. Even if it originally helped or hurt in a day, it might take a while for a supplement to wear off. Others only work via accumulation over time, or work on things other than observable mood.

Experimental design here is very hard. I would not want to maximize purely on the basis of how I feel within 24 hours.

Also, for various reasons, you cannot by default confidently equate any given pill with other pills from another source that claim to deliver the same dose of the same vitamin.

If I was going to run such tests, as much I would like to randomize every day, I would at minimum want to wait three or so days between swaps, and I would limit the multiple hypothesis testing. The data is going to be super noisy at best.

A story worth sharing.

Josh Whiton: Halfway through university I was diagnosed with clinical depression. After a battery of tests and interviews with psychologists I eventually met with the psychiatrist who was to dispense my medication.

Instead he asked me a question that no one had ever asked.

“Why are you depressed?”

So I told him about the meaninglessness of life in an accidental universe where all life was just the product of chance.

“You want me to put you on medication because you’re an intellectual?” he said.

Then he said the wildest thing: “My concern is that your depression is part of a process and the drugs will slow it down.”

He told me to go home and observe all the thoughts in my mind instead of trying to escape from them. If in three days I still wanted the drugs, to come back and he’d give them to me.

So I went home and spent three days journaling, had three epiphanies about the nature of reality, and the year-long depression immediately lifted.

I wonder about all the kids like me who got the drugs instead.

Zac Hill: One of the things I was most brazenly wrong about was just basically making fun of people who thought that Canada’s euthanasia program would achieve scale.

As in, MAID now is already 5% of all deaths in Canada and is about to be available for mental conditions and parliament wants to grant access to minors, and I see a bunch of claims about it being aggressively pushed onto people in various ways. One wonders where this stops, once it starts. If it stops.

Former Quebec officials suggest that best “solution” for intellectually disabled woman without adequate support might be death. Which might be true, but wowie. This highlights where things might be going. Do you trust the state to not push such solutions on people who are financial burdens? To not cut off aid to them as part of an attempt to force such choices upon them? I notice my answer is very much no.

A lot of ‘ethics’ people think the only ethical thing for you to do is hurry up and die.

Joao Pedro de Magalhães: “It is highly unethical to stop aging” – reviewer commenting on one of my grant applications.

The grant focused on cellular rejuvenation, no mention to curing aging, but it shows we still have a long way to go to convince even fellow scientists that curing aging is desirable.

As in, you are trying to solve people’s health problems, but this might also cure aging, And That’s Terrible.

Once again: Bioethicists know nothing of ethics. This is them answering:

Here’s a bioethicist giving a Ted talk advocating that humans be ‘engineered’ to be, no not more intelligent or healthy or happy but (checks notes) intolerant to meat, at the same time that other ‘ethicists’ are screaming about how awful it is we might do embryo selection on positive traits.

And 41% stand ready here to outright deny physical reality.

How exactly would we design society such that it is not a disadvantage to be blind?

If I was blind and I heard people were claiming this, I’d be pretty pissed off at them.

Do you think, by answering that way, you are helping anyone?

Learning about your HIV status, when treatment is not available, dramatically reduces long term survival rates. Some of this is people adjusting rationally to the news, and engaging in generally riskier (to themselves) behaviors and having higher discount rates. Some of it is despair or overreaction. However it isn’t really an option not to tell them or not run the tests, for obvious reasons.

Here in New York they take every opportunity to push you to get tested for HIV, even when there is no reason to do that, as I found out on a recent hospital visit (it was an infection, I am 100% fine now, don’t worry).

Clash Report: Hot-mic moment at the Beijing parade.

Xi: “People rarely lived past 70 before. Now at 70 you’re still a child.”

Putin: “With biotech, organs can be replaced endlessly… people could even reach immortality.”

Xi: “Some predict people might live to 150 this century.”

I do see hope that some people might live to 150 this century, or even that some people alive today might live far longer than that. It will not happen because ‘organs can be replaced endlessly’ because that does not reverse aging, so the rate of new problems will keep doubling roughly every seven years, but we have other options.

Also, yeah, thinking ‘at 70 you’re still a child’ is both obviously false and reflective of the geriatric ruling class that is doing so much damage. Until your anti-aging technology gets a lot better no one 70+ should be in high office.

In response to RFK’s War on Vaccinations and as Florida repeals all vaccine requirements, there are now two state coalitions trying to defend our health. California, Oregon, Washington and Hawaii are going to form one set of recommendations. New York, Massachusetts, Connecticut, Delaware, Pennsylvania and Rhode Island will form another, likely with New Jersey, Vermont and Maine.

Freezing sperm young, if you can spare a little cash to do it, is a pretty great deal. Even if everything keeps working fine you cut down on mutation load a lot, and if something does go wrong you’re going to be very happy about it. It’s a lot of optionality and insurance at a remarkably cheap price, even if you factor in the chance that in the long term we’ll have other tech that renders it unnecessary.

This could end up being a big one: Germany’s national academy of sciences correctly urges that we treat aging like a disease. If we can get momentum behind this, it will supercharge progress towards the (non-AI) thing that is actually killing most of us.

New study claims living near (within 1-3 miles of) a golf course doubles risk of Parkinson’s Disease, which is presumed to be due to pesticides.

Ross Rheingans-Yoo is doing a video podcast series on why our drug development process is so broken and expensive.

Scott Adams has prostate cancer, we wish him well, thread has illustrations of exactly how unhinged a certain type of thinking got and how wide a set of medical topics. Luckily he himself has realized the mistake and although the odds are against him he now appears to be setting aside the quacks and getting real treatment.

A new claim from a randomized trial that getting immunotherapy in the morning has a doubled survival rate versus getting it in the evening, Abhishaike Mahajan speculates potential reasons why.

A potentially serious incentive problem: If CRISPR treatments can permanently cure conditions, insurance companies have no reason to value a permanent solution more than a temporary one, so the whole system will underinvest in such treatments. Ultimately it should be fine, since at most this is a limited cost factor and thus not enough distortion to stop this from happening, and AI should improve development costs enough to compensate.

They did finally run a decade-long study of GMOs in monkeys, and of course they found zero adverse effects on any measured health indicator.

Discussion about this post

Medical Roundup #5 Read More »

removing-these-50-objects-from-orbit-would-cut-danger-from-space-junk-in-half

Removing these 50 objects from orbit would cut danger from space junk in half

China, on the other hand, frequently abandons upper stages in orbit. China launched 21 of the 26 hazardous new rocket bodies over the last 21 months, each averaging more than 4 metric tons (8,800 pounds). Two more came from US launchers, one from Russia, one from India, and one from Iran.

This trend is likely to continue as China steps up deployment of two megaconstellations—Guowang and Thousand Sails—with thousands of communications satellites in low-Earth orbit. Launches of these constellations began last year. The Guowang and Thousand Sails satellites are relatively small and likely capable of maneuvering out of the way of space debris, although China has not disclosed their exact capabilities.

However, most of the rockets used for Guowang and Thousand Sails launches have left their upper stages in orbit. McKnight said nine upper stages China has abandoned after launching Guowang and Thousand Sails satellites will stay in orbit for more than 25 years, violating the international guidelines.

It will take hundreds of rockets to fully populate China’s two major megaconstellations. The prospect of so much new space debris is worrisome, McKnight said.

“In the next few years, if they continue the same trend, they’re going to leave well over 100 rocket bodies over the 25-year rule if they continue to deploy these constellations,” he said. “So, the trend is not good.”

There are technical and practical reasons not to deorbit an upper stage at the end of its mission. Some older models of Chinese rockets simply don’t have the capability to reignite their engines in space, leaving them adrift after deploying their payloads. Even if a rocket flies with a restartable upper stage engine, a launch provider must reserve enough fuel for a deorbit burn. This eats into the rocket’s payload capacity, meaning it must carry fewer satellites.

“We know the Chinese have the capability to not leave rocket bodies,” McKnight said. One example is the Long March 5 rocket, which launched three times with batches of Guowang satellites. On those missions, the Long March 5 flew with an upper stage called the YZ-2, a high-endurance maneuvering vehicle that deorbits itself at the end of its mission. The story isn’t so good for launches using other types of rockets.

“With the other ones, they always leave a rocket body,” McKnight said. “So, they have the capability to do sustainable practices, but on average, they do not.”

A Japanese H-IIA upper stage imaged by Astroscale’s ADRAS-J spacecraft last year. Credit: Astroscale

Since 2000, China has accumulated more dead rocket mass in long-lived orbits than the rest of the world combined, according to McKnight. “But now we’re at a point where it’s actually kind of accelerating in the last two years as these constellations are getting deployed.”

Removing these 50 objects from orbit would cut danger from space junk in half Read More »

ars-live:-is-the-ai-bubble-about-to-pop?-a-live-chat-with-ed-zitron.

Ars Live: Is the AI bubble about to pop? A live chat with Ed Zitron.

As generative AI has taken off since ChatGPT’s debut, inspiring hundreds of billions of dollars in investments and infrastructure developments, the top question on many people’s minds has been: Is generative AI a bubble, and if so, when will it pop?

To help us potentially answer that question, I’ll be hosting a live conversation with prominent AI critic Ed Zitron on October 7 at 3: 30 pm ET as part of the Ars Live series. As Ars Technica’s senior AI reporter, I’ve been tracking both the explosive growth of this industry and the mounting skepticism about its sustainability.

You can watch the discussion live on YouTube when the time comes.

Zitron is the host of the Better Offline podcast and CEO of EZPR, a media relations company. He writes the newsletter Where’s Your Ed At, where he frequently dissects OpenAI’s finances and questions the actual utility of current AI products. His recent posts have examined whether companies are losing money on AI investments, the economics of GPU rentals, OpenAI’s trillion-dollar funding needs, and what he calls “The Subprime AI Crisis.”

Alt text for this image:

Credit: Ars Technica

During our conversation, we’ll dig into whether the current AI investment frenzy matches the actual business value being created, what happens when companies realize their AI spending isn’t generating returns, and whether we’re seeing signs of a peak in the current AI hype cycle. We’ll also discuss what it’s like to be a prominent and sometimes controversial AI critic amid the drumbeat of AI mania in the tech industry.

While Ed and I don’t see eye to eye on everything, his sharp criticism of the AI industry’s excesses should make for an engaging discussion about one of tech’s most consequential questions right now.

Please join us for what should be a lively conversation about the sustainability of the current AI boom.

Add to Google Calendar | Add to calendar (.ics download)

Ars Live: Is the AI bubble about to pop? A live chat with Ed Zitron. Read More »