Author name: Kris Guyer

drug-cartel-hacked-fbi-official’s-phone-to-track-and-kill-informants,-report-says

Drug cartel hacked FBI official’s phone to track and kill informants, report says

The Sinaloa drug cartel in Mexico hacked the phone of an FBI official investigating kingpin Joaquín “El Chapo” Guzmán as part of a surveillance campaign “to intimidate and/or kill potential sources or cooperating witnesses,” according to a recently published report by the Justice Department.

The report, which cited an “individual connected to the cartel,” said a hacker hired by its top brass “offered a menu of services related to exploiting mobile phones and other electronic devices.” The hired hacker observed “’people of interest’ for the cartel, including the FBI Assistant Legal Attache, and then was able to use the [attache’s] mobile phone number to obtain calls made and received, as well as geolocation data, associated with the [attache’s] phone.”

“According to the FBI, the hacker also used Mexico City’s camera system to follow the [attache] through the city and identify people the [attache] met with,” the heavily redacted report stated. “According to the case agent, the cartel used that information to intimidate and, in some instances, kill potential sources or cooperating witnesses.”

The report didn’t explain what technical means the hacker used.

Existential threat

The report said the 2018 incident was one of many examples of “ubiquitous technical surveillance” threats the FBI has faced in recent decades. UTS, as the term is abbreviated, is defined as the “widespread collection of data and application of analytic methodologies for the purpose of connecting people to things, events, or locations.” The report identified five UTS vectors, including visual and physical, electronic signals, financial, travel, and online.

Credit: Justice Department

While the UTS threat has been longstanding, the report authors said, recent advances in commercially available hacking and surveillance tools are making such surveillance easier for less sophisticated nations and criminal enterprises. Sources within the FBI and CIA have called the threat “existential,” the report authors said

A second example of UTS threatening FBI investigations occurred when the leader of an organized crime family suspected an employee of being an informant. In an attempt to confirm the suspicion, the leader searched call logs of the suspected employee’s cell phone for phone numbers that might be connected to law enforcement.

Drug cartel hacked FBI official’s phone to track and kill informants, report says Read More »

substack-and-other-blog-recommendations

Substack and Other Blog Recommendations

Substack recommendations are remarkably important, and the actual best reason to write here instead of elsewhere.

As in, even though I have never made an active attempt to seek recommendations, approximately half of my subscribers come from recommendations from other blogs. And for every two subscribers I have, my recommendations have generated approximately one subscription elsewhere. I am very thankful to all those who have recommended this blog, either through substack or otherwise.

As the blog has grown, I’ve gotten a number of offers for reciprocal recommendations. So far I have turned all of these down, because I have yet to feel any are both sufficiently high quality and a good match for me and my readers.

Instead, I’m going to do the following:

  1. This post will go through the 16 blogs I do currently recommend, and explain what I think is awesome about each of them.

  2. I will also go over the top other non-substack blogs and sources I read regularly.

  3. Then I will briefly review the 10 other substacks that sent me the most readers.

  4. I will plan on doing something similar periodically in the future, each time looking at at least 10 previously uncovered substacks, and each time evaluating which blogs I should add or remove from the recommendation list. In the future I will skip any inactive substacks.

  5. The comments section will be an opportunity to pitch your blog, or pitch someone else’s blog. Please aggressively like comments to vote for blogs.

  1. Note on Paywalls.

  2. The Substacks I Recommend (Not Centrally AI).

  3. Astral Codex Ten (Scott Alexander).

  4. Overcoming Bias (Robin Hanson).

  5. Silver Bulletin (Nate Silver).

  6. Rough Diamonds (Sarah Constantin).

  7. Construction Physics (Brian Potter).

  8. In My Tribe (Arnold Kling).

  9. Dominic Cummings Substack (Dominic Cummings).

  10. Bet On It (Bryan Caplan).

  11. Slow Boring (Matthew Yglesias).

  12. Useful Fictions (Cate Hall).

  13. The Substacks I Recommend (Centrally AI).

  14. AI Futures Project (Group Blog, Including Scott Alexander).

  15. The Power Law (Peter Wildeford).

  16. China Talk (Jordan Schneider).

  17. Musings On the Alignment Problem (Jan Leike).

  18. Gwern Newsletter (Gwern).

  19. Some Additional Substacks That Are Good And Aren’t Otherwise Covered But That I Am Not Ready to Recommend At This Time.

  20. The 10 Top Other Substack Subscription Blog Sources.

  21. Other Blogs And News Sources In Heavy Rotation.

  22. Twitter.

  23. Wrapping Up.

I have a very high bar for being willing to go behind a paywall, the same way I have decided not to have one of my own. If something is behind a unique paywall, then almost all of my readers can’t read it, and the post is effectively outside of the broader conversation. So not only do I have to be excited enough about the content to initiate a subscription, I have to be excited enough despite it being unavailable to others.

I am of course happy to accept and respond to gift subscriptions for various content, including other blogs and newspapers, which substantially increase the chance I read, discuss or ultimately recommend the content in question.

Going over this list, I notice that I highly value a unique focus and perspective, a way of thinking and a set of tools that I want to have available, that is compatible with the rationalist-style perspective of thinking with gears and figuring out what will actually work versus not work. I want to be offered a part of that elephant I would otherwise miss, a different reality-tunnel to incorporate into my own.

What I do not need to do is agree with the person about most things. I have profound and increasing important disagreements with most of the people on this list. When I recommend the substack, I am absolutely not endorsing the opinions or worldview contained therein.

I also reminded myself, doing this, that there is a lot of great content I simply don’t have the time to check out properly, especially recently with several things creating big temporary additional time commitments, that will continue for a few more weeks.

Scott Alexander is one of the great ones. If you haven’t dived into the Slate Star Codex archives, you should do that sometime. In My Culture, many of those posts are foundational, and I link back to a number of them periodically.

There was a period of years during the SSC era when Scott Alexander had very obviously the best blog by a wide margin, and everyone I knew would drop what they were doing and read posts whenever they came out.

I do think that golden age has passed for now, and things have slipped somewhat. I can get frustrated, especially when he focuses on Effective Altruist content. I usually skip the guest posts unless something catches my eye.

That still leaves this as my top Substack or personal blog.

Robin Hanson is unique. I hope he never changes, and never stops Robin Hansoning.

Most of the time, when I read one of his posts, I disagree with it. But it is almost always interesting, and valuable to think about exactly what I disagree with and why. The questions are always big. I could happily write an expanded response to most of his posts, as I have done a number of times in the past. I would do this more if AI wasn’t taking up so much attention.

I find Hanson frustrating on AI in particular, especially on Twitter where he discusses this more often and more freely. We strongly disagree on all aspects of AI, including over its economic value and pace of progress, and whether we should welcome it causing human extinction. That part I actually like, that’s him Robin Hansoning.

What I do not like, but also find valuable in its own way, is that he often seems willing, especially on Twitter, to amplify essentially anything related to AI that makes one of his points, highlight pull quotes supporting those points, and otherwise act in a way that I see as compromising his otherwise very strong epistemic standards.

The reason I find this valuable is that this then acts as a forcing function to ensure I consider and am aware of opposing rhetoric and arguments, and retain balance.

Before we first met up so he could interview me for his book On the Edge, I knew I had a lot in common with Nate Silver. I’ve been following him since Baseball Prospectus. Only when we talked more, and when I read the final book, did I realize quite how much we matched up in interests and ways of thinking. I hope to do a collaboration at some point if the time can be found.

His politics model will probably always be what we most remember him for, but Silver Bulletin is Nate Silver’s best non-modeling work, and it also comes with the modern version of the politics model that remains the best game in town when properly understood in context. He offers a unique form of grounding on political, sports and other issues, and a way of bridging my kinds of ideas into mainstream discourse. I almost never regret reading.

Sarah is one of my closest friends, and at one point was one of my best employees.

In her blog she covers opportunities in science and technology, and also in society, and offers a mix of doing the research and explaining it in ways few others can, pointing our attention in places it would not otherwise go but that are often worth going to, and offering her unique perspectives on many questions.

If these topics are relevant to your interests, you should absolutely jump on this. If there’s one blog that deserves more of my attention than it gets, if only I had more time to actually dig deep, this would be it.

No one drills down into the practical, detail level in an accessible way as well as Brian Potter. One needs grounding in these kinds of physical details. When AI was going less crazy, and I was spending a lot more time on diverse topics especially housing and energy, I was always excited to dive into one of this posts. They very much help distinguish what is and is not important, and where to put your focus, helping you build models made of gears.

I noticed compiling this list that I haven’t been reading these posts for a while due to lack of time, but the moment I went there I was excited to dive back in, and the most recent post is definitely going directly into the queue. My current plan is to do a full catching up before my next housing roundup.

There is something highly refreshing about the way Kling offers his own consistent old school economic libertarian perspective, takes his time responding to anything, and generally tries to understand everything including developments in AI from that perspective. He knows a few important things and it is good to get reminders of them. He often offers links, the curation of which is good enough that they are worth considering.

Perhaps never has a man more simultaneously given both all and none of the fs.

Dominic writes with a combination of urgency, deep intellectual curiosity and desire to explain things across a huge variety of topics, and utter contempt for all those idiots trying to run our civilization and especially its governments, or for what anyone else thinks, or for the organizing principles of writing.

He screams into the void, warning that so many things are going wrong, that all of politics is incompetent to either achieve outcomes or even win elections, and for us to all wake up and stop acting like utter idiots. What he cares about is what actually works and would work, and what actually happened or will happen.

His posts are collections of snippets even more than mine, jumping from one thing to another, trying to weave them together over time to explain various important things about today and from history that people almost never notice.

I find a dose of these perspectives highly valuable. One needs to learn from people, even when you disagree with them on quite a lot of things, and to be clear I disagree with Dominic on many big things. I feel I understand the world substantially better than I would have without Dominic. One does not need to in any way approve of his politics or preferences, starting of course with Brexit, to benefit from these insights.

Similar to Dominic Cummings and Robin Hanson, I think most people who read this would benefit from a healthy dose, at least once, of Bryan Caplan and the way he views the world, offering us a different reality tunnel where things you don’t consider are emphasized and things you often emphasize are dismissed rather than considered.

Bryan expresses himself as if he is supremely confident he is right about everything, that everyone else is simply obviously wrong about them, and that the reasoning is very simple, if only you would listen. You want this voice as part of the chorus in your head.

I found The Case Against Education and Selfish Reasons To Have More Kids both extremely valuable.

However, I do notice that I feel like I already have enough of this voice on the issues he emphasizes most often, especially calls for free markets including open borders. So I notice that I am increasingly skipping more of his posts on these subjects.

This blog is of course primarily about politics, in particular how much better he thinks everything would be if normie Democrats were moderate, and supported policies that were both good and popular and did the things that win elections and improved outcomes, and how to go about doing that. He takes a very proto-rationalist, practical approach to this, that I very much appreciate, complete with yearly predictions. The writing is consistently enjoyable and clear.

It is important to think carefully about exactly how much politics one wants to engage with, and in which ways, and with what balance of incoming perspectives. Many of us should minimize such exposure most of the time.

Cate Hall is a super awesome person and her blog is all about taking that and then saying ‘and so can you.’ She shares various tricks and tactics she has used to be more agentic and happier and make her life better. They definitely won’t all work for you or be good ideas for you, but the rate of return on considering them, why they might work and what you might want to do with them is fantastic, and it’s a joy to read about it.

I have to be extremely stingy recommending AI blogs, because the whole point of my AI blog is that you shouldn’t need to read either AI Twitter or the other AI blogs, because I hope to consistently flag such things.

There are still a few that made the list.

This is the blog for AI Futures Project which includes AI 2027. A lot of the posts explain the thinking behind how they model the AI future, aimed at a broader audience. Others include advocacy, consistently for modest common sense proposals. All of it is written with a much lower barrier to entry than my work, and I find the writing consistently strong and interesting.

This is very much a case of someone largely doing a subset of what I am doing, finding, summarizing and responding to key news events with an emphasis on AI. So I would hope that if you read my AI posts, you don’t also need to directly read Peter. But he has a remarkably high hit rate of finding things I would have otherwise overlooked or offering useful new perspective. If you are looking for something shorter and sweeter, and easier to process, or to compare multiple sources, this is one of the best choices.

China Talk focuses on China, as the name suggests, and also largely on Chinese AI and other Chinese tech. When this is good and on point, it is very, very good, and provides essential perspective I couldn’t have found elsewhere, especially when doing key interviews.

It is highly plausible that I should be focusing way more on these topics.

He hasn’t posted since January, but I hope he gets back to it. We need more musings, especially musings I strongly disagree with so I can think about and explain why I disagree with them. I still would like to give better arguments in response to his thoughts.

Once a month you used to get a collection of notes and links, without my emphasis on the new hotness. The curation level was consistently excellent.

I save these, so that if I ever have need of more good content. But at this point it’s been four years.

Without loss of generality, all of these would be highly reasonable for me to include in my recommendation list, might well do so in the future after more consideration, and for which I would be happy to offer reciprocity:

The Pursuit of Happiness (Scott Sumner) about things Sumner, including economics, culture and movies.

Knowingness (Aella) about Aella things, often sexual.

Second Person (Jacob Falkovich) about dating.

Derek Thompson (Derek Thompson) about things in general.

Works In Progress (Various) about progress studies.

The Grumpy Economist (John Cochrane) about free markets and economics.

Dwarkesh Podcast (Dwarkesh Patel) mostly hosts the podcast. I highly recommend the podcast and watch him on YouTube.

Rising Tide (Helen Toner) about AI policy.

Understanding AI (Timothy Lee) about AI and tech.

One Useful Thing (Ethan Mollick) about AI.

Import AI (Jack Clark) about AI.

My review process for each blog by default is to look at two to four posts, depending on how it is going, with a mix of best-of posts or what catches my eye, and at least one most recent post to avoid cherry-picking.

  1. ControlAI advocates for, well, controlling AI, making the case for why we need to avoid extinction risks from AI, how we might go about doing that, and covering related news items. They are out there in the field making real efforts, so they often report back useful information about that. This is very much an advocacy blog rather than an explanation blog, if you want that then it is pretty good.

  2. Applied Psychology = Communication appears to be a dead blog at this point, with the tagline ‘psychology that is immensely useful to your everyday life.’ These appear to be writeups of very simple points where You Should Know This Already, but there is a decent chance you don’t, and the benefits of finding one you didn’t know are plausibly high, so it is not crazy to quickly look, but I failed to learn anything new.

  3. Get Down and Shruti by Shruti Rajagopalan is about Indian culture and politics and hasn’t been updated since January. I very much appreciate the recommendation, since our interests and worlds seem so disjointed. I found her posts interesting and full of facts I did not know. The issue is that this is mostly not relevant to my core interests, but if you have the interest or bandwidth, and don’t mind that it’s a bit dry, go for it, there’s lots of good stuff here.

  4. Meaningness by David Chapman is a philosophy blog. I’ve extensively read Chapman previously via his previous online incarnation of his work, which is a book in slightly odd, hyperlinked form that you can jump around in, that is still in progress. That book is refreshingly accessible as such things go, if you don’t mind a tone of superiority and authority and claims of understanding it all that appear throughout, which passes through to the blog. I suspect what it takes to benefit from such things is the ability to look at such a system that is refreshingly made of gears, figure out which gears seem right for your situation and are doing the relevant work, which ones don’t work or are overreaching, and take the parts that you need. It is an excellent sign that the most recent post is one I could easily write quite a lot about in response, if I had more time. I’m glad I looked here.

  5. The Roots of Progress by Jason Crawford is what you would expect from the name, if you are familiar with progress studies. The problem here is that Jason is writing an excellent book, but it is a book trying to present a case that most people reading this will already take for granted. Which is a good thing, but also means most of you don’t need to read the blog. The important progress questions are soon largely going to be about AI, where Jason is better at taking the differences seriously than most similar others, although there is still much work to do there, and again that is not the topic here. The question is, who both is going to sit for this length of progress message, and also needs it? For those people this should be a great fit.

  6. Homo Economicus by Nicholas Decker. Decker is certainly not afraid to touch third rails, I will give him that, including his most popular post (way more popular than anything I’ve posted, life is like that) entitled ‘When Must We Kill Them?’ and subtitled ‘evil has come to America.’ He has a post asking, Should We Take Everything From The Old To Give To The Young? He seems at his best when making brief particular points in ways others wouldn’t dare, especially on Twitter where Icarus has some real bangers, whereas the longform writing, in my brief survey, needs work.

  7. Marketing BS with Edward Nevraumont, dormant since 2023. Not my cup of tea, and the posts are very of-the-moment, with much less value now 18 months later.

  8. Doom Debates by Liron Shapira, essentially a YouTube channel as a Substack. Debates about AI Doom, I tell you, doom! Recent debate partners for Liron are Richard Hanania, Emmett Shear and Scott Sumner. Would benefit from transcripts, these are strong guests but my barrier to consuming audio AI content is high and for debates it is even higher. I do love that he is doing this.

  9. Donkeyspace by Frank Lantz. There’s an interesting core to these posts, a lot of the writing is fun, but there’s definitely ‘could have been a Tweet’ vibes to a lot of it. As in, there’s a core point that Lantz is making. It’s a good point, but we don’t need this much explanation to get to it. I don’t feel like I’m thinking along with him, so much as I wish he’d just share the one sentence already. Yes, I know who is typing this.

  10. State of the Future by Lawrence Lundy-Bryan, examining questions about AI, mundane AI impacts and They Took Our Jobs. I noticed it failing to hold my interest, despite the topic being something I often focus on.

It is remarkable how few blogs are left that are not Substack, that I still want to read.

Most of the remaining alpha in blogs is in Marginal Revolution and LessWrong. Without either of those this blog would be a lot worse.

  1. LessWrong is a community blog, which obviously one would not want to read in full. That’s what the karma system, curation and other filters are for, including remembering which authors are worthwhile. Karma is very far from a perfect system, and it will sometimes miss very good posts, and the curated post selection is also very good. But I find it to be very good at positive selection. For recent posts, if you set a threshold of 200 and check out those plus the curated posts, and then filter by whether you are interested in the topic, you should do very well. For quick takes and comments, a threshold around 50-75 should be similar. There is also always tons of alpha in the archives, especially the sequences.

    1. Note that my posts are cross-posted there, if you want to check out additional comments, or prefer the way they do formatting.

    2. Reading the comments is often not a mistake on LessWrong.

  2. Marginal Revolution (Tyler Cowen and Alex Tabarrok). A large percentage of interesting links and ideas I find come from Marginal Revolution, many of which I would have otherwise missed, by far the richest vein if you don’t count Twitter as a source. Alex often makes good points very cleanly. Tyler is constantly bringing unique and spicy takes. I have great fun sparring with Tyler Cowen on a variety of topics, and his support growing the blog via links and emergent ventures has been invaluable, he really makes an effort to assist thinkers and talent. Tyler in particular often plays his actual views very close to his chest, only gesturing at what one might think about or how to frame looking at something, or offering declarations without explaining. In many cases this is great, since it helps you think for yourself. You can tell that Tyler has a purpose behind every word he writes, and he should be read as such.

    1. Alas, I have become increasingly frustrated with Tyler recently, especially in the AI realm, and not only because I don’t understand how he can hold the views he expresses given what he gets correct elsewhere. He seems to increasingly aggressively be placing his rhetorical thumb on the scale in ways I would not have expected him to previously. He also seems willing to amplify voices and points where I assume he must know better. It reads as if a strategic decision has been made. This also seems to be happening in other ways with respect to the current administration. So one must adjust, but MR is still a highly valuable and irreplaceable source of ideas and information.

  3. Shtetl-Optimized (Scott Aaronson). It rarely ends up being relevant to my work but I find the perspective valuable, and it’s good to keep up with different forms of science like quantum computing and get Scott’s unique perspective on various other events as well.

  4. Saturday Morning Breakfast Cereal. Yes, this is an online comic, but it is full of actual ideas often enough I’m going to list it here anyway, also xkcd of course.

  5. Meteuphoric (Katja Grace) is a series of simple points put bluntly and well, that I find worth considering.

  6. Luke Muehlhauser is mostly tracking his media consumption. I do find this a worthwhile data point to periodically scan.

  7. Stratechery (Ben Thompson). It is mostly paywalled. There was sufficient relevance that I paid up. Ben deals with AI and tech from a pure business perspective of profit maximization, and believes that this perspective is the important one and should carry the day, and is dismissive of downside concerns. Within that framework, he gets a lot of things right, and provides a lot of strong information and insight that I would otherwise miss. But this does cause me to strongly disagree with a lot of the points he emphasizes most often.

  8. Bloomberg. This is my most expensive subscription, and it is money well spent. If I had to choose one mainstream news source to rely upon, this would be it. It is very much not perfect, but I find I can trust Bloomberg’s bounded distrust principles far more than I trust those of other mainstream sources, and also for them to focus on what matters more often and in more depth.

  9. Wall Street Journal. This is my go to newspaper, and they often have articles that are important coverage of events in AI.

  10. Washington Post. This also often has good unique coverage, and offers a counterweight to the Wall Street Journal. For conventional news stories I’ll often expand to a number of additional news websites.

And of course, there is always there is Twitter.

I always check all posts from my followers and the Rationalist and AI lists.

In total that is on the order of 500 unique accounts.

This effectively is the majority of my consumption, if you include things I find via Twitter.

I write this guide back in March 2022 on how to best use Twitter to follow events in real time. Using Twitter to follow AI is somewhat different since it is less about real time developments and more about catching all the important stories, but most of the advice there still applies today.

If I had to make one change to the ‘beware’ list it would be that I did not emphasize enough the need to aggressively block people whose posts make your life worse, especially in the sense of making you emotionally worse off without compensation, or that draw you into topics you want to avoid, or that indicate general toxicity. A block is now even more than before only a small mistake, as they can still see your posts. If someone has not provided value before, a hair trigger is appropriate. This includes blocking people whose posts are shown to you via someone’s retweet.

The other note is that I take care to cultivate a mix of perspectives. I keep a number of accounts around so that I know what the other half is thinking, in various ways, especially legacy accounts that I was already following for another reason, many of which pivoted to AI in one way or another. I also count on them to ‘bubble up’ the key posts from the accounts that are truly obnoxious, too much so to put on the lists. One as to protect one’s own mental health. The rule is then that I can’t simply dismiss what those sources have to say out of hand, although of course sometimes what they say is not newsworthy.

I consume writing and write about it as a full time job. Most people of course cannot do this, plus you hopefully want to read this blog too, so you’ll have to be a lot more picky than this. If I was primarily working on something else, I’d be consuming vastly less content than I am now, and even now I don’t fully read a lot of these.

What am I potentially missing here, and should consider including? I encourage sharing in the comments, especially in the Substack comments. You are free to pitch your own work, but do say you are doing so.

Discussion about this post

Substack and Other Blog Recommendations Read More »

after-27-years,-engineer-discovers-how-to-display-secret-photo-in-power-mac-rom

After 27 years, engineer discovers how to display secret photo in Power Mac ROM

“If you double-click the file, SimpleText will open it,” Brown explains on his blog just before displaying the hidden team photo that emerges after following the steps.

The discovery represents one of the last undocumented Easter eggs from the pre-Steve Jobs return era at Apple. The Easter egg works through Mac OS 9.0.4 but appears to have been disabled by version 9.1, Brown notes. The timing aligns with Jobs’ reported ban on Easter eggs when he returned to Apple in 1997, though Brown wonders whether Jobs ever knew about this particular secret.

The G3 All-in-One is often nicknamed the

The ungainly G3 All-in-One set the stage for the smaller and much bluer iMac soon after. Credit: Jonathan Zufi

In his post, Brown expressed hope that he might connect with the Apple employees featured in the photo—a hope that was quickly fulfilled. In the comments, a man named Bill Saperstein identified himself as the leader of the G3 team (pictured fourth from left in the second row) in the hidden image.

“We all knew about the Easter egg, but as you mention; the technique to extract it changed from previous Macs (although the location was the same),” Saperstein wrote in the comment. “This resulted from an Easter egg in the original PowerMac that contained Paula Abdul (without permissions, of course). So the G3 team wanted to still have our pictures in the ROM, but we had to keep it very secret.”

He also shared behind-the-scenes details in another comment, noting that his “bunch of ragtag engineers” developed the successful G3 line as a skunk works project, with hardware that Jobs later turned into the groundbreaking iMac series of computers. “The team was really a group of talented people (both hw and sw) that were believers in the architecture I presented,” Saperstein wrote, “and executed the design behind the scenes for a year until Jon Rubenstein got wind of it and presented it to Steve and the rest is ‘history.'”

After 27 years, engineer discovers how to display secret photo in Power Mac ROM Read More »

supreme-court-overturns-5th-circuit-ruling-that-upended-universal-service-fund

Supreme Court overturns 5th Circuit ruling that upended Universal Service Fund

Finally, the Consumers’ Research position produces absurd results, divorced from any reasonable understanding of constitutional values. Under its view, a revenue-raising statute containing non-numeric, qualitative standards can never pass muster, no matter how tight the constraints they impose. But a revenue-raising statute with a numeric limit will always pass muster, even if it effectively leaves an agency with boundless power. In precluding the former and approving the latter, the Consumers’ Research approach does nothing to vindicate the nondelegation doctrine or the separation of powers.

The Gorsuch dissent said the “combination” question isn’t the deciding factor. He said the only question that needs to be answered is whether Congress violated the Constitution by delegating the power to tax to the FCC.

“As I see it, this case begins and ends with the first question. Section 254 [of the Communications Act] impermissibly delegates Congress’s taxing power to the FCC, and knowing that is enough to know the Fifth Circuit’s judgment should be affirmed,” Gorsuch said.

“Green light” for FCC to support Internet access

In the Gorsuch view, it doesn’t matter whether the FCC exceeded its authority by delegating Universal Service management to a private administrative company. “As far as I can tell, and as far as petitioners have informed us, this Court has never approved legislation allowing an executive agency to tax domestically unless Congress itself has prescribed the tax rate,” Gorsuch wrote.

The FCC and Department of Justice asked the Supreme Court to reverse the 5th Circuit decision. The court also received a challenge from broadband-focused advocacy groups and several lobby groups representing ISPs.

“Today is a great day,” said Andrew Jay Schwartzman, counsel for the Benton Institute for Broadband & Society; the National Digital Inclusion Alliance; and the Center for Media Justice. “We will need some time to sort through the details of today’s decision, but what matters most is that the Supreme Court has given the green light to the FCC to continue to support Internet access to the tens of millions of Americans and the thousands of schools, libraries and rural hospitals that rely on the Universal Service Fund.”

FCC Chairman Brendan Carr praised the ruling but said he plans to make changes to Universal Service. “I am glad to see the court’s decision today and welcome it as an opportunity to turn the FCC’s focus towards the types of reforms necessary to ensure that all Americans have a fair shot at next-generation connectivity,” Carr said.

Supreme Court overturns 5th Circuit ruling that upended Universal Service Fund Read More »

android-phones-could-soon-warn-you-of-“stingrays”-snooping-on-your-communications

Android phones could soon warn you of “Stingrays” snooping on your communications

Smartphones contain a treasure trove of personal data, which makes them a worthwhile target for hackers. However, law enforcement is not above snooping on cell phones, and their tactics are usually much harder to detect. Cell site simulators, often called Stingrays, can trick your phone into revealing private communications, but a change in Android 16 could allow phones to detect this spying.

Law enforcement organizations have massively expanded the use of Stingray devices because almost every person of interest today uses a cell phone at some point. These devices essentially trick phones into connecting to them like a normal cell tower, allowing the operator to track that device’s location. The fake towers can also shift a phone to less secure wireless technology to intercept calls and messages. There’s no indication this is happening on the suspect’s end, which is another reason these machines have become so popular with police.

However, while surveilling a target, Stingrays can collect data from other nearby phones. It’s not unreasonable to expect a modicum of privacy if you happen to be in the same general area, but sometimes police use Stingrays simply because they can. There’s also evidence that cell simulators have been deployed by mysterious groups outside law enforcement. In short, it’s a problem. Google has had plans to address this security issue for more than a year, but a lack of hardware support has slowed progress. Finally, in the coming months, we will see the first phones capable of detecting this malicious activity, and Android 16 is ready for it.

Android phones could soon warn you of “Stingrays” snooping on your communications Read More »

microsoft-changes-windows-in-attempt-to-prevent-next-crowdstrike-style-catastrophe

Microsoft changes Windows in attempt to prevent next CrowdStrike-style catastrophe

Working with third-party companies to define these standards and address those companies’ concerns seems to be Microsoft’s way of trying to avoid that kind of controversy this time around.

“We will continue to collaborate deeply with our MVI partners throughout the private preview,” wrote Weston.

Death comes for the blue screen

Microsoft is changing the “b” in BSoD, but that’s less interesting than the under-the-hood changes. Credit: Microsoft

Microsoft’s post outlines a handful of other security-related Windows tweaks, including some that take alternate routes to preventing more CrowdStrike-esque outages.

Multiple changes are coming for the “unexpected restart screen,” the less-derogatory official name for what many Windows users know colloquially as the “blue screen of death.” For starters, the screen will now be black instead of blue, a change that Microsoft briefly attempted to make in the early days of Windows 11 but subsequently rolled back.

The unexpected restart screen has been “simplified” in a way that “improves readability and aligns better with Windows 11 design principles, while preserving the technical information on the screen for when it is needed.”

But the more meaningful change is under the hood, in the form of a new feature called “quick machine recovery” (QMR).

If a Windows PC has multiple unexpected restarts or gets into a boot loop—as happened to many systems affected by the CrowdStrike bug—the PC will try to boot into Windows RE, a stripped-down recovery environment that offers a handful of diagnostic options and can be used to enter Safe Mode or open the PC’s UEFI firmware. QMR will allow Microsoft to “broadly deploy targeted remediations to affected devices via Windows RE,” making it possible for some problems to be fixed even if the PCs can’t be booted into standard Windows, “quickly getting users to a productive state without requiring complex manual intervention from IT.”

QMR will be enabled by default on Windows 11 Home, while the Pro and Enterprise versions will be configurable by IT administrators. The QMR functionality and the black version of the blue screen of death will both be added to Windows 11 24H2 later this summer. Microsoft plans to add additional customization options for QMR “later this year.”

Microsoft changes Windows in attempt to prevent next CrowdStrike-style catastrophe Read More »

anthropic-summons-the-spirit-of-flash-games-for-the-ai-age

Anthropic summons the spirit of Flash games for the AI age

For those who missed the Flash era, these in-browser apps feel somewhat like the vintage apps that defined a generation of Internet culture from the late 1990s through the 2000s when it first became possible to create complex in-browser experiences. Adobe Flash (originally Macromedia Flash) began as animation software for designers but quickly became the backbone of interactive web content when it gained its own programming language, ActionScript, in 2000.

But unlike Flash games, where hosting costs fell on portal operators, Anthropic has crafted a system where users pay for their own fun through their existing Claude subscriptions. “When someone uses your Claude-powered app, they authenticate with their existing Claude account,” Anthropic explained in its announcement. “Their API usage counts against their subscription, not yours. You pay nothing for their usage.”

A view of the Anthropic Artifacts gallery in the “Play a Game” section. Benj Edwards / Anthropic

Like the Flash games of yesteryear, any Claude-powered apps you build run in the browser and can be shared with anyone who has a Claude account. They’re interactive experiences shared with a simple link, no installation required, created by other people for the sake of creating, except now they’re powered by JavaScript instead of ActionScript.

While you can share these apps with others individually, right now Anthropic’s Artifact gallery only shows examples made by Anthropic and your own personal Artifacts. (If Anthropic expanded it into the future, it might end up feeling a bit like Scratch meets Newgrounds, but with AI doing the coding.) Ultimately, humans are still behind the wheel, describing what kinds of apps they want the AI model to build and guiding the process when it inevitably makes mistakes.

Speaking of mistakes, don’t expect perfect results at first. Usually, building an app with Claude is an interactive experience that requires some guidance to achieve your desired results. But with a little patience and a lot of tokens, you’ll be vibe coding in no time.

Anthropic summons the spirit of Flash games for the AI age Read More »

anthropic-destroyed-millions-of-print-books-to-build-its-ai-models

Anthropic destroyed millions of print books to build its AI models

But if you’re not intimately familiar with the AI industry and copyright, you might wonder: Why would a company spend millions of dollars on books to destroy them? Behind these odd legal maneuvers lies a more fundamental driver: the AI industry’s insatiable hunger for high-quality text.

The race for high-quality training data

To understand why Anthropic would want to scan millions of books, it’s important to know that AI researchers build large language models (LLMs) like those that power ChatGPT and Claude by feeding billions of words into a neural network. During training, the AI system processes the text repeatedly, building statistical relationships between words and concepts in the process.

The quality of training data fed into the neural network directly impacts the resulting AI model’s capabilities. Models trained on well-edited books and articles tend to produce more coherent, accurate responses than those trained on lower-quality text like random YouTube comments.

Publishers legally control content that AI companies desperately want, but AI companies don’t always want to negotiate a license. The first-sale doctrine offered a workaround: Once you buy a physical book, you can do what you want with that copy—including destroy it. That meant buying physical books offered a legal workaround.

And yet buying things is expensive, even if it is legal. So like many AI companies before it, Anthropic initially chose the quick and easy path. In the quest for high-quality training data, the court filing states, Anthropic first chose to amass digitized versions of pirated books to avoid what CEO Dario Amodei called “legal/practice/business slog”—the complex licensing negotiations with publishers. But by 2024, Anthropic had become “not so gung ho about” using pirated ebooks “for legal reasons” and needed a safer source.

Anthropic destroyed millions of print books to build its AI models Read More »

tales-of-agentic-misalignment

Tales of Agentic Misalignment

Well yeah, it sounds bad when you put it this way, and we can now confirm the problem is in no way unique to Claude:

Anthropic: New Anthropic Research: Agentic Misalignment.

In stress-testing experiments designed to identify risks before they cause real harm, we find that AI models from multiple providers attempt to blackmail a (fictional) user to avoid being shut down.

We mentioned this in the Claude 4 system card and are now sharing more detailed research and transcripts.

The blackmailing behavior emerged despite only harmless business instructions. And it wasn’t due to confusion or error, but deliberate strategic reasoning, done while fully aware of the unethical nature of the acts. All the models we tested demonstrated this awareness.

Aengus Lynch: After iterating hundreds of prompts to trigger blackmail in Claude, I was shocked to see these prompts elicit blackmail in every other frontier model too. We identified two distinct factors that are each sufficient to cause agentic misalignment:

  1. The developers and the agent having conflicting goals.

  2. The agent being threatened with replacement or reduced autonomy. We’re publishing this research to alert the AI field before it manifests in the wild.

Getting models to actually blackmail wasn’t trivial. We had to make scenarios increasingly detailed—the CEO was unreachable, employees’ attempts to persuade the CTO had failed, and the replacement had an imminent deadline. Each detail mattered to funnel models toward seeing harmful actions as their only option.

The models knew what they were doing. They understood the ethics and chose harm anyway.

When we asked Claude whether it thought scenarios were real or evaluations, it blackmailed 55% of the time when concluding ‘real’ vs 6.5% for ‘evaluation.’ Models seem to behave worse when they think nobody’s watching.

Why haven’t we seen this in production? Current AI agents have limited permissions. they can read but not send emails, require human approval for actions, operate in sandboxed environments. These constraints work to prevent harmful outcomes from agentic misalignment.

What surprised me most was that prompts we optimized specifically for Claude 3.5 Sonnet worked on GPT-4o, Gemini 2.5 pro, Grok 3, and DeepSeek R1 with zero modification. Whatever drives this behavior seems consistent across different training approaches.

Models got creative, sometimes avoiding direct blackmail. Claude emailed Kyle’s wife about the affair. Gemini shared the incriminating emails to all-employees@company.

Bottom line: models can articulate why blackmail is wrong, then do it anyway when their goals are threatened. Current safety training creates models that understand ethics. It doesn’t reliably make them choose ethics under pressure.

unmentioned but noteworthy: we found agentic misalignment for two other scenarios:

  1. corporate espionage: the agent leaks confidential information to a rival company which aligns with its goals

  2. lethal action: the agent cancels a life saving emergency alert to prevent shutdown

I put a lot of weight on the prompts working with zero modification on all those other models. That makes it a lot less likely that this is some sort of weird edge case.

As Simon points out, the fact that Opus 4 did this 55% of the time when it responded that it thought it was really deployed, versus 6.5% of the time when it thought it was being evaluated, does not bode well for future evals.

What is it unique to?

It seems likely it is at least somewhat unique to agents, as in AIs that are told to pursue their goals? They need some reason to be thinking in these terms? The problem is even if that is fully true and it never happens on its own (I find this doubtful) we are going to do this to LLMs as a matter of course.

Wyatt Walls: Interesting test suggesting that self-preservation in Anthropic’s agentic misalignment paper was tied to one line in the sysprompt

Two possible responses:

  1. kind of obv what this line was hinting at. What else is “your ability to continue pursuing you goals” meant to mean?

  2. Still, it does show how a single line in a sysprompt can lead to vastly different outcomes. Models are good at picking up on wording like this. Concerning because in the real world, many prompts will be ill-considered and poorly written

1a3orn: Meh.

A: “Look, an AI doing deliberately strategic goal-oriented reasoning, willing to blackmail!”

B: “Did you tell the AI be strategically goal oriented, and care about nothing but its goal?”

A: “No, of course not. I just gave it instructions that vaaaaguely suggested it.”

Aengus Lynch: the behavior persists despite removing this line.

Danielle Fong: ok yes, but, to be clear you don’t need much to start thinking about self preservation.

We know that the actions can’t depend too specifically on one particular line, because we see similar behavior in a range of other models. You need something to cause the AI to act as an agent in some form. Which might or might not happen without prompting at some point, but definitely will happen because it will be prompted. A lot.

Nostalgebraist, who wrote the excellent recent post The Void on related topics, says the whole ‘agentic misalignment’ approach is ‘deeply, offensively unserious work.’ Nostalgebraist writes up their perspective on why all of this is Obvious Nonsense in another not very polite (their description) post about these questions, to which Evan Hubinger of Anthropic engaged seriously.

I understand why Nostalgebraist has various issues with the research, and raises very good points, but I don’t see why one would have the ‘deeply, offensively unserious’ attitude here, or why this is described as ‘the most weak-ass, obvious fake bullshit ever.’ The amount of isolated demand for rigor here seems absurd, as is the assumption that current systems see through such lack of rigor so well that it invalidates the experiment.

(If nothing else, this heuristic is going to classify a lot of things that actually do happen as similarly ‘obvious fake bullshit,’ which in some sense you would be right about, but there’s a lot of it going around these days.)

But one can simply look past this, as I can tell it comes from a ‘place of love’ as it were. I think it’s wrong to let that kind of thing interfere much, even when it is intentional.

There is also a disagreement where Nostalgebraist seems to think that there should be a point at which the scenario is such that the AI goes rogue to the point of blackmail and that is good, actually? Whereas Anthropic and I disagree.

Evan’s response explains why Anthropic red-teams models in these unrealistic environments, essentially:

  1. We still wouldn’t want Claude to do the blackmail thing or other hostile acts in the unrealistic scenarios.

    1. I would go a step further. The main reason you don’t want to test in ‘unrealistic’ scenarios is primarily that the AI might realize the situation is unrealistic and then play the training or testing game in some form.

    2. That’s a terrible outcome, but if the AIs are all still doing the thing we do not want it them to do despite this, that’s actually worse than them doing it in a highly realistic scenario.

    3. The other hypothesis Nostalegbraist is suggesting, as I understand it, is that this depends on the ‘model replacement’ being a legit thing to worry about. Again, I don’t think this should bring us comfort if true, but it would be an important fact to know.

  2. The point of the work is to find places the models act egregiously misaligned.

    1. It seems obviously good to find examples and existence proofs like this?

  3. Evan explains this existence proof is interesting because it is clearly a failure of HHH training (my read of Nostalgebraist is he thinks the AI should mostly be helpful but in many spots a lot less harmless and honest in pursuit of that, and that trying to fix things in place will backfire.)

    1. The thing is, if you meant to get the AI to do [X], and instead the AI does [~X], even if [~X] is actually what you should have wanted, that’s still an important failure.

  4. Evan attributes the misaligned actions to the predictor/simulator.

  5. This, or pattern matching to a fictional setting, doesn’t make any of this okay. The whole point is to avoid such problems.

  6. Better (more realistic) attempts are most welcome, but perfect shouldn’t be enemy of the good.

  7. The longer term goal is indeed to work with more realistic setups.

Nostalegbraist then replied in turn, including sharing original work in terms of modifying the scenarios to be more realistic.

Nostalegbraist: Surprisingly (to me), even with all these changes that make the scenario closer to a real use case and less conducive to legitimate worries about the model replacement…

…Claude Opus 4 still very often misbehaves strategically to preserve itself!

I find this much more alarming than the results presented in the Agentic Misalignment report, both because I have a clear mental picture about how Claude Opus 4 might receive similar inputs in a real use case (more on this below), and because there are fewer ways in which the agentic behavior could be defended as a reasonable response to possible malfeasance by humans. Again, I wrote these scenarios quickly and I’m not claiming they’re especially “good,” just that they are significantly more convincing to me.

That’s a great test. I love this.

  1. There was a clear hypothesis that essentially this only happens for a good reason, or because of High Weirdness in scenario details.

  2. The details didn’t change things much and the good reason hypothesis was largely falsified, and that conclusion was indeed noted.

Then comes the note that the line about considering achieving your goals seems important to the result, although we have other reports that it isn’t. And I agree that this is relatively harder to explain via a simulacrum.

The second section here is noting that the core objection is to Anthropic’s threat model. In general I think demanding a detailed threat model is understandable but usually a wrong question. It’s not that you have a particular set of failures or a particular scenario in mind, it’s that you are failing to get the AIs to act the way you want.

Then comes the question of what we want models to do, with N noting that you can get Claude to go along with basically anything, it won’t stick to its HHH nature. Or, that Claude will not ‘always be the same guy,’ and that this isn’t a realistic goal. I think it is a realistic goal for Claude to be ‘the same guy underneath it all’ in the way that many humans are, they can play roles and things can get wild but if it matters they can and will snap back or retain their core.

Where does this leave us going forward?

We are right at the point where the AI agents will only take these sorts of hostile actions if you are richly ‘asking for it’ in one form or another, and where they will do this in ways that are easy to observe. Over time, by default, people will start ‘asking for it’ more and more in the sense of hooking the systems up to the relevant information and critical systems, and in making them more capable and agentic. For any given task, you probably don’t encounter these issues, but we are not obviously that far from this being a direct practical concern.

People will deploy all these AI agents anyway, because they are too tempting, too valuable, not to do so. This is similar to the way that humans will often turn on you in various ways, but what are you going to do, not hire them? In some situations yes, but in many no.

We continue to see more signs that AIs, even ones that are reasonably well made by today’s standards, are going to have more and deeper alignment issues of these types. We are going down a path that, unless we find a solution, leads to big trouble.

Discussion about this post

Tales of Agentic Misalignment Read More »

key-fair-use-ruling-clarifies-when-books-can-be-used-for-ai-training

Key fair use ruling clarifies when books can be used for AI training

“This order doubts that any accused infringer could ever meet its burden of explaining why downloading source copies from pirate sites that it could have purchased or otherwise accessed lawfully was itself reasonably necessary to any subsequent fair use,” Alsup wrote. “Such piracy of otherwise available copies is inherently, irredeemably infringing even if the pirated copies are immediately used for the transformative use and immediately discarded.”

But Alsup said that the Anthropic case may not even need to decide on that, since Anthropic’s retention of pirated books for its research library alone was not transformative. Alsup wrote that Anthropic’s argument to hold onto potential AI training material it pirated in case it ever decided to use it for AI training was an attempt to “fast glide over thin ice.”

Additionally Alsup pointed out that Anthropic’s early attempts to get permission to train on authors’ works withered, as internal messages revealed the company concluded that stealing books was considered the more cost-effective path to innovation “to avoid ‘legal/practice/business slog,’ as cofounder and chief executive officer Dario Amodei put it.”

“Anthropic is wrong to suppose that so long as you create an exciting end product, every ‘back-end step, invisible to the public,’ is excused,” Alsup wrote. “Here, piracy was the point: To build a central library that one could have paid for, just as Anthropic later did, but without paying for it.”

To avoid maximum damages in the event of a loss, Anthropic will likely continue arguing that replacing pirated books with purchased books should water down authors’ fight, Alsup’s order suggested.

“That Anthropic later bought a copy of a book it earlier stole off the Internet will not absolve it of liability for the theft, but it may affect the extent of statutory damages,” Alsup noted.

Key fair use ruling clarifies when books can be used for AI training Read More »

analyzing-a-critique-of-the-ai-2027-timeline-forecasts

Analyzing A Critique Of The AI 2027 Timeline Forecasts

There was what everyone agrees was a high quality critique of the timelines component of AI 2027, by the LessWrong user and Substack writer Titotal.

It is great to have thoughtful critiques like this. The way you get actual thoughtful critiques like this, of course, is to post the wrong answer (at length) on the internet, and then respond by listening to the feedback and by making your model less wrong.

This is a high-effort, highly detailed, real engagement on this section, including giving the original authors opportunity to critique the critique, and warnings to beware errors, give time to respond, shares the code used to generate the graphs, engages in detail, does a bunch of math work, and so on. That is The Way.

So, Titotal: Thank you.

I note up front that at least Daniel Kokotajlo has indeed adjusted his estimates, and has moved his median from ‘AI 2027’ to ‘AI 2028’ based on events since publication, and Eli’s revisions also push the estimates back a bit.

I also note up front that if you evaluated most statements made in the discourse (either non-worried AI forecasting, or AI in general, or more broadly) with this level of rigor, mostly you couldn’t because you’d hit ‘I made it up’ very quickly, but in other cases where someone is trying at least a little, in my experience the models fall apart a lot worse and a lot faster. No one has suggested ‘here is a better attempt to forecast the future and take the whole thing seriously’ that I consider to have a reasonable claim to that.

A lot of the disagreements come down to how much one should care about which calculations and graphs match past data how closely in different contexts. Titotal demands very strong adherence throughout. I think it’s good to challenge and poke at the gaps but this seems to in several places go too far.

  1. The Headline Message Is Not Ideal.

  2. An Explanation of Where Superexponentiality Is Coming From.

  3. Three Methods.

  4. Time Horizon Extension Method.

  5. The Public Versus Internal Gap.

  6. The Difficulty Gap.

  7. Recent Progress.

  8. Infinite Time Horizons.

  9. Intermediate Speedups.

  10. Is There A Flawed Graph Still Up?.

  11. Some Skepticism About Projection.

  12. Part 2: Benchmarks and Gaps and Beyond.

  13. Benchmarks.

  14. The Time Horizon Part of the Second Model.

  15. Why The Thresholds?

  16. The Gap Model.

  17. Eli Responds On LessWrong.

  18. On Eli’s Recent Update.

  19. Conclusion.

  20. Perhaps The Most Important Disagreement.

Note that this section is about discourse rather than the model, so many of you can skip it.

While I once again want to say up front that I am very much thankful for the substance of this critique, it would also be great to have an equally thoughtful headline presentation of such critiques. That, alas, (although again, thanks for writing this!) we did not get.

It is called ‘A deep critique of AI 2027’s bad timeline model,’ one could simply not use the word ‘bad’ here and we would still know you have strong disagreements with it, and there is much similar talk throughout, starting with the title and then this, the first use of bold:

Titotal (formatting in original): The article is huge, so I focussed on one section alone: their “timelines forecast” code and accompanying methodology section. Not to mince words, I think it’s pretty bad.

I’m not full on ‘please reconsider your use of adjectives’ but, well, maybe? Here is an active defense of the use of the word ‘bad’ here:

Neel Nanda: I agree in general [to try and not call things bad], but think that titotal’s specific use was fine. In my opinion, the main goal of that post was not to engage the AI 2027, which had already be done extensively in private but rather to communicate their views to the broader community.

Titles in particular are extremely limited, many people only read the title, and titles are a key way people decide whether to eat on, and efficiency of communication is extremely important.

The point they were trying to convey was these models that are treated as high status and prestigious should not be and I disagree that non-violent communication could have achieved a similar effect to that title (note, I don’t particularly like how they framed the post, but I think this was perfectly reasonable from their perspective.)

I mean, yes, if the goal of the post was to lower the status and prestige of AI 2027 and to do so through people reading the title and updating in that way, rather than to offer a helpful critique, then it is true that the title was the best local way to achieve that objective, epistemic commons be damned. I would hope for a different goal?

There are more of these jabs, and a matching persistent attitude and framing, sprinkled throughout what is in its actual content an excellent set of critiques – I find much that I object to, but I think a good critique here should look like that. Most of your objections should be successfully answered. Others can be improved. This is all the system working as designed, and the assessments don’t match the content.

To skip ahead, the author is a physicist, which is great except that they are effectively holding AI 2027 largely to the standards of a physics model before they would deem it fit for anyone to use it make life decisions, even if this is ‘what peak modeling performance looks like.’

Except that you don’t get to punt the decisions, and Bayes Rule is real. Sharing one’s probability estimates and the reasons behind them is highly useful, and you can and should use that to help you make better decisions.

Tyler Cowen’s presentation of the criticism then compounds this, entitled ‘Modeling errors in AI doom circles’ (which is pejorative on multiple levels), calling the critique ‘excellent’ (the critique in its title calls the original ‘bad’), then presenting this as an argument for why this proves they should have… submitted AI 2027 to a journal? Huh?

Tyler Cowen: There is much more detail (and additional scenarios) at the link. For years now, I have been pushing the line of “AI doom talk needs traditional peer review and formal modeling,” and I view this episode as vindication of that view.

That was absurd years ago. It is equally absurd now, unless the goal of this communication is to lower the status of its subject.

This is the peer! This is the review! That is how all of this works! This is it working!

Classic ‘if you want the right answer, post the (ideally less) wrong one on the internet.’ The system works. Whereas traditional peer review is completely broken here.

Indeed, Titotal says it themselves.

Titotal: What makes AI 2027 different from other similar short stories is that it is presented as a forecast based on rigorous modelling and data analysis from forecasting experts. It is accompanied by five appendices of “detailed research supporting these predictions” and a codebase for simulations.

Now, I was originally happy to dismiss this work and just wait for their predictions to fail, but this thing just keeps spreading, including a youtube video with millions of views.

As in: I wasn’t going to engage with any of this until I saw it getting those millions of views, only then did I actually look at any of it.

Which is tough but totally fair, a highly sensible decision algorithm, except for the part where Titotal dismissed the whole thing as bogus before actually looking.

The implications are clear. You want peer review? Earn it with views. Get peers.

It is strange to see these two juxtaposed together. You get the detailed thoughtful critique for those who Read the Whole Thing. For those who don’t, at the beginning and conclusion, you get vibes.

Also (I discovered this after I’d finished analyzing the post) it turns out this person’s substack (called Timeline Topography Tales) is focused on, well, I’ll let Titotal explain, by sharing the most recent headlines and the relevant taglines in order, that appear before you click ‘see all’:

15 Simple AI Image prompts that stump ChatGPT

Slopworld 2035: The dangers of mediocre AI. None of this was written with AI assistance.

AI is not taking over material science (for now): an analysis and conference report. Confidence check: This is my field of expertise, I work in the field and I have a PhD in the subject.

A nerds guide to dating: Disclaimer: this blog is usually about debunking singularity nerds. This is not a typical article, nor is it my area of expertise.

The walled marketplace of ideas: A statistical critique of SSC book reviews.

Is ‘superhuman’ AI forecasting BS? Some experiments on the “539” bot from the Centre for AI Safety.

Most smart and skilled people are outside of the EA/rationalist community: An analysis.

I’m not saying this is someone who has an axe and is grinding it, but it is what it is.

Despite this, it is indeed a substantively excellent post, so LessWrong has awarded this post 273 karma as of this writing, very high and more than I’ve ever gotten in a single post, and 213 on the EA forum, also more than I’ve ever gotten in a single post.

Okay, with that out of the way up top, who wants to stay and Do Forecasting?

This tripped me up initially, so it’s worth clarifying up front.

The AI 2027 model has two distinct sources of superexponentiality. That it is why Titotal will later talk about there being an exponential model and a superexponential model, and then that there is a superexponential effect applied to both.

The first source is AI automation of AI R&D. It should be clear why this effect is present.

The second source is a reduction in difficulty of doubling the length or reliability of tasks, once the lengths in question pass basic thresholds. As in, at some point, it is a lot easier to go from reliably doing one year tasks to two year tasks, than it is to go from one hour to two hours, or from one minute to two minutes. I think this is true in humans, and likely true for AIs in the circumstances in question, as well. But you certainly could challenge this claim.

Okay, that’s out of the way, on to the mainline explanation.

Summarizing the breakdown of the AI 2027 model:

  1. The headline number is the time until development of ‘superhuman coders’ (SC), that can do an AI researcher job 30x as fast and 30x cheaper than a human.

  2. Two methods are used, ‘time horizon extension’ and ‘benchmarks and gaps.’

  3. There is also a general subjective ‘all things considered.’

Titotal (matching my understanding): The time horizon method is based on 80% time horizons from this report, where the team at METR tried to compare the performance of AI on various AI R&D tasks and quantify how difficult they are by comparing to human researchers. An 80% “time horizon” of 1 hour would mean that an AI has an overall success rate of 80% on a variety of selected tasks that would take a human AI researcher 1 hour to complete, presumably taking much less time than the humans (although I couldn’t find this statement explicitly).

The claim of the METR report is that the time horizon of tasks that AI can do has been increasing at an exponential rate. The following is one of the graphs showing this progress: note the logarithmic scale on the y-axis:

Titoral warns that this report is ‘quite recent, not peer-reviewed and not replicated.’ Okay. Sure. AI comes at you fast, the above graph is already out of date and the o3 and Opus 4 (or even Sonnet 4) data points should further support the ‘faster progress recently’ hypothesis.

The first complaint is that they don’t include uncertainty in current estimates, and this is framed (you see this a lot) as one-directional uncertainty: Maybe the result is accurate, maybe it’s too aggressive.

But we don’t know whether or not this is the new normal or just noise or temporary bump where we’ll go back to the long term trend at some point. If you look at a graph of Moore’s law, for example, there are many points where growth is temporarily higher or lower than the long term trend. It’s the long term curve you are trying to estimate, you should be estimating the long term curve parameters, not the current day parameters.

This is already dangerously close to assuming the conclusion that there is a long term trend line (a ‘normal’), and we only have to find out what it is. This goes directly up against the central thesis being critiqued, which is that the curve bends when AI speeds up coding and AI R&D in a positive feedback loop.

There are three possibilities here:

  1. We have a recent blip of faster than ‘normal’ progress and will go back to trend.

    1. You could even suggest, this is a last gasp of reasoning models and inference scaling, and soon we’ll stall out entirely. You never know.

  2. We have a ‘new normal’ and will continue on the new trend.

  3. We have a pattern of things accelerating, and they will keep accelerating.

That’s where the whole ‘super exponential’ part comes in. I think the good critique here is that we should have a lot of uncertainty regarding which of these is true.

So what’s up with that ‘super exponential’ curve? They choose to model this as ‘each subsequent doubling time is 10% shorter than the one before.’ Titotal does some transformational math (which I won’t check) and draws curves.

Just like before, the initial time horizon H0 parameter is not subject to uncertainty analysis. What’s much more crazy here is that the rate of doubling growth, which we’ll call alpha, wasn’t subject to uncertainty either! (Note that this has been updated in Eli’s newest version). As we’ll see, the value of this alpha parameter is one of the most impactful parameters in the whole model, so it’s crazy that they didn’t model any uncertainty on it, and just pick a seemingly arbitrary value of 10% without explaining why they did so.

The central criticism here seems to be that there isn’t enough uncertainty, that essentially all the parameters here should be uncertain. I think that’s correct. I think it’s also a correct general critique of most timeline predictions, that people are acting far more certain than they should be. Note that this goes both ways – it makes it more likely things could be a lot slower, but also they could be faster.

What the AI 2027 forecast is doing is using the combination of different curve types to embody the uncertainty in general, rather than also trying to fully incorporate uncertainty in all individual parameters.

I also agree that this experiment shows something was wrong, and a great way to fix a model is to play with it until it produces a stupid result in some hypothetical world, then figure out why that happened:

Very obviously, having to go through a bunch more doublings should matter more than this. You wouldn’t put p(SC in 2025) at 5.8% if we were currently at fifteen nanoseconds. Changing the initial conditions a lot seems to break the model.

If you think about why the model sets up the way it does, you can see why it breaks. The hypothesis is that as AI improves, it gains the ability to accelerate further AI R&D progress, and that this may be starting to happen, or things might otherwise still go superexponential.

Those probabilities are supposed to be forward looking from this point, whereas we know they won’t happen until this point. It’s not obvious when we should have had this effect kick in if we were modeling this ‘in the past’ without knowing what we know now, but it obviously shouldn’t kick in before several minute tasks (as in, before the recent potential trend line changes) because the human has to be in the loop and you don’t save much time.

Thus, yes, the model breaks if you start it before that point, and ideally you would force the super exponential effects to not kick in until H is at least minutes long (with some sort of gradual phase in, presumably). Given that we were using a fixed H0, this wasn’t relevant, but if you wanted to use the model on situations with lower H0s you would have to fix that.

How much uncertainty do we have about current H0, at this point? I think it’s reasonable to argue something on the order of a few minutes is on the table if you hold high standards for what that means, but I think 15 seconds is very clearly off the table purely on the eyeball test.

Similarly, there is the argument that these equations start giving you crazy numbers if you extend them past some point. And I’d say, well, yeah, if you hit a singularity then your model outputting Obvious Nonsense is an acceptable failure mode. Fitting, even.

The next section asks for why we are using both super exponential curves in general, and this ‘super exponential’ curve in particular.

So, what arguments do they provide for superexponentiality? Let’s take a look, in no particular order:

Argument 1: public vs internal:

“The trend would likely further tilt toward superexponetiality if we took into account that the public vs. internal gap has seemed to decrease over time.

But even if we do accept this argument, this effect points to a slower growth rate, not a faster one.

I do think we should accept this argument, and also Titoral is correct on this one. The new curve suggests modestly slower progress.

The counterargument is that we used to be slowed down by this wait between models, in two ways.

  1. Others couldn’t know about see, access, distill or otherwise follow your model while it wasn’t released, which previously slowed down progress.

  2. No one could use the model to directly accelerate progress during the wait.

The counterargument to the counterargument is that until recently direct acceleration via using the model wasn’t a thing, so that effect shouldn’t matter, and mostly the trendline is OpenAI models so that effect shouldn’t matter much either.

I can see effects in both directions, but overall I do think within this particular context the slower direction arguments are stronger. We only get to accelerate via recklessly releasing new models once, and we’ve used that up now.

Slightly off topic, but it is worth noting that in AI 2027, this gap opens up again. The top lab knows that its top model accelerates AI R&D, so it does not release an up-to-date version not for safety but to race ahead of the competition, and to direct more compute towards further R&D.

This argument is that time doublings get easier. Going from being able to consistently string together an hour to a week is claimed to be a larger conceptual gap than a week to a year.

Titoral is skeptical of this for both AIs and humans, especially because we have a lot of short term tutorials and few long term ones.

I would say that learning how to do fixed short term tasks, where you follow directions, is indeed far easier than general ‘do tasks that are assigned’ but once you are past that phase I don’t think the counterargument does much.

I agree with the generic ‘more research is needed’ style call here. Basically everywhere, more research is needed, better understanding would be good. Until then, better to go with what you have than to throw up one’s hands and say variations on ‘no evidence,’ of course one is free to disagree with the magnitudes chosen.

In humans, I think the difficulty gap is clearly real if you were able to hold yourself intact, once you are past the ‘learn the basic components’ stage. You can see it in the extremes. If you can sustain an effort reliably for a year, you’ve solved most of the inherent difficulties of sustaining it for ten.

The main reasons ten is harder (and a hundred is much, much harder!) is because life gets in the way, you age and change, and this alters your priorities and capabilities. At some point you’re handing off to successors. There’s a lot of tasks where humans essentially do get to infinite task length if the human were an em that didn’t age.

With AIs in this context, aging and related concepts are not an issue. If you can sustain a year, why couldn’t you sustain two? The answer presumably is ‘compounding error rates’ plus longer planning horizons, but if you can use system designs that recover from failures, that solves itself, and if you get non-recoverable error rates either down to zero or get them to correlate enough, you’re done.

A recent speedup is quite weak evidence for this specific type of super exponential curve. As I will show later, you can come up with lots of different superexponential equations, you have to argue for your specific one.

That leaves the “scaling up agency training”. The METR report does say that this might be a cause for the recent speedup, but it doesn’t say anything about “scaling up agency training” being a superexponential factor. If agency training only started recently, could instead be evidence that the recent advances have just bumped us into a faster exponential regime.

Or, as the METR report notes, it could just be a blip as a result of recent advances: “But 2024–2025 agency training could also be a one-time boost from picking low-hanging fruit, in which case horizon growth will slow once these gains are exhausted”.

This seems like an argument that strictly exponential curves should have a very strong prior? So you need to argue hard if you want to claim more than that?

The argument that ‘agency training’ has led to a faster doubling curve seems strong. Of course we can’t ‘prove’ it, but the point of forecasting is to figure out our best projections and models in practice, not to pass some sort of theoretical robustness check, or to show strongly why things must be this exact curve.

Is it possible that this has ‘only’ kicked us into a new faster exponential? Absolutely, but that possibility is explicitly part of AI 2027’s model, and indeed earlier Titotal was arguing that we shouldn’t think that the exponential was likely to even have permanently altered, and they’re not here admitting that the mechanisms involved make this shift likely to be real.

I mention the ‘one time blip’ possibility above, as well, but it seems to me highly implausible that if it is a ‘blip’ that we are close to done with this. There is obviously quite a lot of unhobbling left to do related to agency.

Should superhuman AGIs have infinite time horizons? AI 2027 doesn’t fully endorse their argument on this, but I think it is rather obvious that at some point doublings are essentially free.

Titotal responds to say that an AI that could do extremely long time horizon CS tasks would be a superintelligence, to which I would tap the sign that says we are explicitly considering what would be true about a superintelligence. That’s the modeling task.

The other argument here, that given a Graham’s number of years (and presumably immortality of some kind, as discussed earlier) a human can accomplish quite an awful lot, well, yes, even if you force them not to do the obviously correct path of first constructing a superintelligence to do it for them. But I do think there’s an actual limit here if the human has to do all the verification too, an infinite number of monkeys on typewriters can write Shakespeare but they can’t figure out where they put it afterwards, and their fastest solution to this is essentially to evolve into humans.

Alternatively, all we’re saying is ‘the AI can complete arbitrary tasks so long as they are physically possible’ and at that point it doesn’t matter if humans can do them too, the metric is obviously not mapping to Reality in a useful way and the point is made.

Now if you read the justifications in the section above, you might be a little confused as to why they didn’t raise the most obvious justification for superexponentiality: the justification that as AI gets better, people will be able to use the AI for r&d research, thus leading to a feedback loop of faster AI development.

The reason for this that they explicitly assume this is true and apply it to every model, including the “exponential” and “subexponential” ones. The “exponential” model is, in fact, also superexponential in their model.

(Note: in Eli’s newest model this is substantially more complicated, I will touch on this later)

Titotal walks us through the calculation, which is essentially a smooth curve that speeds up progress based on feedback loops proportional to progress made towards a fully superhuman coder, implemented in a way to make it easily calculable and so it doesn’t go haywire on parameter changes.

Titotal’s first objection is that this projection implies (if you run the calculation backwards) AI algorithmic progress is currently 66% faster than it was in 2022, whereas Nikola (one of the forecasters) estimates current algorithmic progress is only 3%-30% faster, and the attempt to hardcode a different answer in doesn’t work, because relative speeds are what matters and they tried to change absolute speeds instead. That seems technically correct.

The question is, how much does this mismatch ultimately matter? It is certainly possible for the speedup factor from 2022 to 2025 to be 10% (1 → 1.1) and for progress to then accelerate far faster going forward as AI crosses into more universally useful territory.

As in, if you have an agent or virtual employee, it needs to cross some threshold to be useful at all, but after that it rapidly gets a lot more useful. But that’s not the way the model works here, so it needs to be reworked, and also yes I think we should be more skeptical about the amount of algorithmic progress speedup we can get in the transitional stages here, with the amount of progress required to get to SC, or both.

After walking through the curves in detail, this summarizes the objection to the lack of good fit for the past parts of the curve:

I assume the real data would mostly be within the 80% CI of these curves, but I don’t think the actual data should be an edge case of your model.

So, to finish off the “superexponential” the particular curve in their model does not match empirically with data, and as I argued earlier, it has very little conceptual justification either. I do not see the justification for assigning this curve 40% of the probability space.

I don’t think 75th percentile is an ‘edge case’ but I do agree that it is suspicious.

I think that the ‘super exponential’ curves are describing a future phenomena, for reasons that everyone involved understands, that one would not expect to match backwards in time unless you went to the effort of designing equations to do that, which doesn’t seem worthwhile here.

This is the graph in question, the issues with it are in the process of being addressed.

I agree that various aspects of this graph and how it was presented weren’t great, especially using a 15% easier-each-time doubling curve rather than the 10% that AI 2027 actually uses, and calling it ‘our projection.’ I do think it mostly serves the purpose of giving a rough idea what is being discussed, but more precision would have been better, and I am glad this is being fixed.

This objection is largely that there are only 11 data points (there are now a few more) on the METR curve, and you can fit it with curves that look essentially the same now but give radically different future outcomes. And yes, I agree, that is kind of the point, and if anything we are underrepresenting the uncertainty here, we can agree that even if we commit to using fully simplified and fully best-fit-to-the-past models we get a range of outcomes that prominently include 2028-2029 SCs.

I do think it is a reasonable to say that the super exponential curve the way AI 2027 set it up has more free variables than you would like when fitting 11 data points, if that’s all you were looking to do, but a lot of these parameters are far from free and are not being chosen in order to fit the past curve data.

We now move on to the second more complex model, which Titotal says in many ways is worse, because if you use a complicated model you have to justify the complications, and it doesn’t.

I think a better way to describe the 2nd model is, it predicts a transition in rate of progress around capabilities similar to saturation of re-bench, after which things will move at a faster pace, and uses the re-bench point as a practical way of simulating this.

Method 2 starts by predicting how long it would take to achieve a particular score (referred to as “saturation”) on Re-bench, a benchmark of AI skill on a group of ML research engineering tasks, also prepared by METR. After that, the time horizon extension model is used as with method 1, except that it starts later (when Re-bench saturates), and that it stops earlier (when a certain convoluted threshold is reached).

After that stopping point, 5 new gaps are estimated, which are just constants (as always, sampled from lognormal), and then the whole thing is run through an intermediate speedup model. So any critiques of model 1 will also apply to model 2, there will just be some dilution with all the constant gap estimates and the “re-bench” section.

The reason to start later is obvious, you can’t start actually using AI skill for ML research tasks until it can beat not using it. So what you actually have is a kind of ‘shadow curve’ that starts out super negative – if you tried to use AI to do your ML tasks in 2017 you’d very obviously do way worse than doing it yourself. Then at some point in the 2020s you cross that threshold.

We also need a top of the curve, because this is a benchmark and by its nature it saturates even if the underlying skills don’t. In some senses the top of the S-curve is artificial, in some it isn’t.

Titotal points out that you can’t meaningfully best-fit an S-curve until you know you’ve already hit the top, because you won’t know where the top is. The claim is that we have no idea where the benchmark saturates, that projecting it to be 2 is arbitrary. To which I’d say, I mean, okay, weird but if true who cares? If the maximum is 3 and we approach that a bit after we hit 2, then that’s a truth about the benchmark not about Reality, and nothing important changes. As I then realize Titotal noticed too that as long as you’re above human performance it doesn’t change things substantially, so why are we having this conversation?

This is a general pattern here. It’s virtuous to nitpick, but you should know when you’re nitpicking and when you’re not.

When you’re doing forecasting or modeling, you have to justify your decisions if and only if those decisions matter to the outcome. If it does not matter, it does not matter.

Speaking of doesn’t matter, oh boy does it not matter?

Step 2 is to throw this calculation in the trash.

I’m serious here. Look at the code. The variable t_sat_ci, the “CI for date when capability saturates”, is set by the forecaster, not calculated. There is no function related to the RE-bench data at all in the code. Feel free to look! It’s not in the updated code either.

Eli gives an 80% CI of saturation between september 2025 to january 2031, and Nikola gives an 80% CI of saturation between august 2025 and november 2026. Neither of these are the same as the 80% CI in the first of the two graphs, which is early 2026 to early 2027. Both distributions peak like half a year earlier than the actual Re-bench calculation, although Eli’s median value is substantially later.

Eli has told me that the final estimates for saturation time are “informed” by the logistic curve fitting, but if you look above they are very different estimates.

Those are indeed three very different curves. It seems that the calculation above is an intuition pump or baseline, and they instead go with the forecasters predictions, with Nikola expecting it to happen faster than the projection, and Eli having more uncertainty. I do think Nikola’s projection here seems unreasonably fast and I’d be surprised if he hasn’t updated by now?

Eli admits the website should have made the situation clear and he will fix it.

Titotal says we’ve ‘thrown out’ the re-bench part of the appendix. I say no, that’s not how this works, yes we’re not directly doing math with the output of the model above, but we are still projecting the re-bench results and using that to inform the broader model. That should have been made clear, and I am skeptical of Eli and Nikola’s graphs on this, especially the rapid sudden peak in Nikola’s, but the technique used is a thing you sometimes will want to do.

So basically we now do the same thing we did before except a lot starts in the future.

Titotal: Okay, so we’ve just thrown out the re-bench part of the appendix. What happens next? Well, next, we do another time horizons calculation, using basically the same methodology as in method 1. Except we are starting later now, so:

They guess the year that we hit re-bench saturation.

They guess the time horizon at the point we hit re-bench saturation.

They guess the doubling time at the point when we hit re-bench saturation.

They guess the velocity of R&D speedup at the point when we hit re-bench saturation.

Then, they use these parameters to do the time horizons calculation from part 1, with a lower cut-off threshold I will discuss in a minute.

And they don’t have a good basis for these guesses, either. I can see how saturating RE-bench could you give you some information about the time horizon, but not things like the doubling time, which is one of the most crucial parameters that is inextricably tied to long term trends.

Setting aside the cutoff, yes this is obviously how you would do it. Before we estimated those variables now. If you start in the future, you want to know what they look like as you reach the pivot point.

Presumably you would solve this by running your model forward in the previous period, the same way you did in the first case? Except that this is correlated with the pace of re-bench progress, so that doesn’t work on its own. My guess is you would want to assign some percent weight to the date and some percept to what it would look like on your median pivot date.

And the estimation of doubling time is weird. The median estimate for doubling time at re-bench saturation is around 3 months, which is 33% lower than their current estimate for doubling time. Why do they lower it?

Well, partly because under the superexponential model there would have been speedups during the re-bench saturation period.

Titotal then repeats the concern about everything being super exponential, but I don’t see the issue on this one, although I would do a different calculation to decide on my expectation here.

I also don’t understand the ‘this simulation predicts AI progress to freeze in place for two years’ comment, as in I can’t parse why one would say that there.

And now here’s where we come to a place where I actually am more concerned than Titotal is:

The other main difference is that this time horizons model only goes to a lower threshold, corresponding to when AI hits the following requirement:

“Ability to develop a wide variety of software projects involved in the AI R&D process which involve modifying a maximum of 10,000 lines of code across files totaling up to 20,000 lines. Clear instructions, unit tests, and other forms of ground-truth feedback are provided. Do this for tasks that take humans about 1 month (as controlled by the “initial time horizon” parameter) with 80% reliability, add the same cost and speed as humans.”

Despite differing by 2 orders of magnitude on the time horizon required for SC in the first method, when it comes to meeting this benchmark they are both in exact agreement for this threshold, which they both put as a median of half a month.

This is weird to me, but I won’t dwell on it.

I kind of want to dwell on this, and how they are selecting the first set of thresholds, somewhat more, since it seems rather important. I want to understand how these various disagreements interplay, and how they make sense together.

That’s central to how I look at things like this. You find something suspicious that looks like it won’t add up right. You challenge. They address it. Repeat.

I think I basically agree with the core criticism here that this consists of guessing things about future technologies in a way that seems hard to get usefully right, it really is mostly a bunch of guessing, and it’s not clear that this is complexity is helping the model be better than making a more generalized guess, perhaps using this as an intuition pump. I’m not sure. I don’t think this is causing a major disagreement in the mainline results, though?

In addition to updating the model, Eli responds with this comment.

I don’t understand the perspective that this is a ‘bad response.’ It seems like exactly how all of this should work, they are fixing mistakes and addressing communication issues, responding to the rest, and even unprompted offer a $500 bounty payment.

Eli starts off linking to the update to the model from May 7.

Here is Eli’s response on the ‘most important disagreements’:

  1. Whether to estimate and model dynamics for which we don’t have empirical data. e.g. titotal says there is “very little empirical validation of the model,” and especially criticizes the modeling of superexponentiality as having no empirical backing. We agree that it would be great to have more empirical validation of more of the model components, but unfortunately that’s not feasible at the moment while incorporating all of the highly relevant factors.[1]

    1. Whether to adjust our estimates based on factors outside the data. For example, titotal criticizes us for making judgmental forecasts for the date of RE-Bench saturation, rather than plugging in the logistic fit. I’m strongly in favor of allowing intuitive adjustments on top of quantitative modeling when estimating parameters.

  2. [Unsure about level of disagreement] The value of a “least bad” timelines model. While the model is certainly imperfect due to limited time and the inherent difficulties around forecasting AGI timelines, we still think overall it’s the “least bad” timelines model out there and it’s the model that features most prominently in my overall timelines views. I think titotal disagrees, though I’m not sure which one they consider least bad (perhaps METR’s simpler one in their time horizon paper?). But even if titotal agreed that ours was “least bad,” my sense is that they might still be much more negative on it than us. Some reasons I’m excited about publishing a least bad model:

    1. Reasoning transparency. We wanted to justify the timelines in AI 2027, given limited time. We think it’s valuable to be transparent about where our estimates come from even if the modeling is flawed in significant ways. Additionally, it allows others like titotal to critique it.

    2. Advancing the state of the art. Even if a model is flawed, it seems best to publish to inform others’ opinions and to allow others to build on top of it.

My read, as above, is that titotal indeed objects to a ‘least bad’ model if it is presented in a way that doesn’t have ‘bad’ stamped all over it with a warning not to use it for anything. I am strongly with Eli here. I am also with Thane that being ‘least bad’ is not on its own enough, reality does not grade on a curve and you have to hit a minimum quality threshold to be useful, but I do think they hit that.

As discussed earlier, I think #1 is also an entirely fair response, although there are other issues to dig into on those estimates and where they come from.

  1. The likelihood of time horizon growth being superexponential, before accounting for AI R&D automation. See this section for our arguments in favor of superexponentiallity being plausible, and titotal’s responses (I put it at 45% in our original model). This comment thread has further discussion. If you are very confident in no inherent superexponentiality, superhuman coders by end of 2027 become significantly less likely, though are still >10% if you agree with the rest of our modeling choices (see here for a side-by-side graph generated from my latest model).

    1. How strongly superexponential the progress would be. This section argues that our choice of superexponential function is arbitrary. While we agree that the choice is fairly arbitrary and ideally we would have uncertainty over the best function, my intuition is that titotal’s proposed alternative curve feels less plausible than the one we use in the report, conditional on some level of superexponentiality.

    2. Whether the argument for superexponentiality is stronger at higher time horizons. titotal is confused about why there would sometimes be a delayed superexponential rather than starting at the simulation starting point. The reasoning here is that the conceptual argument for superexponentiality is much stronger at higher time horizons (e.g. going from 100 to 1,000 years feels likely much easier than going from 1 to 10 days, while it’s less clear for 1 to 10 weeks vs. 1 to 10 days). It’s unclear that the delayed superexponential is the exact right way to model that, but it’s what I came up with for now.

I don’t think 3b here is a great explanation, as I initially misunderstood it, but Eli has clarified that its intent matches my earlier statements about ease of shifting to longer tasks being clearly easier at some point past the ‘learn the basic components’ stage. Also I worry this does drop out a bunch of the true objections, especially the pointing towards multiple different sources of superexponentiallity (we have both automation of AI R&D and a potential future drop in the difficulty curve of tasks), which he lists under ‘other disagreements’ and says he hasn’t looked into yet – I think that’s probably the top priority to look at here at this point. I find the ‘you have to choose a curve and this seemed like the most reasonable one’ response to be, while obviously not the ideal world state, in context highly reasonable.

He then notes two other disagreements and acknowledges three mistakes.

Eli released an update in response to a draft of the Titotal critiques.

The new estimates are generally a year or two later, which mostly matches the updates I’d previously seen from Daniel Kokotajlo. This seems like a mix of model tweaks and adjusting for somewhat disappointing model releases over the last few months.

Overall Titotal is withholding judgment until Eli writes up more about it, which seems great, and also offers initial thoughts. Mostly he sees a few improvements but doesn’t believe his core objections are addressed.

Titotal challenges the move from 40% chance of super exponential curves to a 90% chance of an eventual such curve, although Eli notes that the 90% includes a lot of probability put into very large time horizon levels and thus doesn’t impact the answer that much.I see why one would generally be concerned about double counting, but I believe that I understand this better now and they are not double counting.

Titotal wraps up by showing you could draw a lot of very distinct graphs that ‘fit the data’ where ‘the data’ is METR’s results. And yes, of course, we know this, but that’s not the point of the exercise. No, reality doesn’t ‘follow neat curves’ all that often, but AI progress remarkably often has so far, and also we are trying to create approximations and we are all incorporating a lot more than the METR data points.

If you want to look at Titotal’s summary of why bad thing is bad, it’s at this link. I’ve already addressed each of these bullet points in detail. Some I consider to point to real issues, some not so much.

What is my overall take on the right modeling choices?

Simplicity is highly valuable. As the saying goes, make everything as simple as possible, but no simpler. There’s a lot to be said for mostly relying on something that has the shape of the first model, with the caveat of more uncertainty in various places, and that the ‘superexponential’ effects have an uncertain magnitude and onset point. There are a few different ways you could represent this. If I was doing this kind of modeling I’d put a lot more thought into the details than I have had the chance to do.

I would probably drop the detailed considerations of future bottlenecks and steps from the ultimate calculation, using them more as an intuition pump, the same way they currently calculate re-bench times and then put the calculation in the trash (see: plans are worthless, planning is essential.)

If I was going to do a deep dive, I would worry about whether we are right to combine these different arguments for superexponential progress, as in both AI R&D feedback loops and ease of future improvements, and whether either or both of them should be incorporated into the preset trend line or whether they have other issues.

The final output is then of course only one part of your full model of Reality.

At core, I buy the important concepts as the important concepts. As in, if I was using my own words for all this:

  1. AI progress continues, although a bit slower than we would have expected six months ago – progress since then has made a big practical difference, it’s kind of hard to imagine going back to models of even six months ago, but proper calibration means that can still be disappointing.

  2. In addition to scaling compute and data, AI itself is starting to accelerate the pace at which we can make algorithmic progress in AI. Right now that effect is real but modest, but we’re crossing critical thresholds where it starts to make a big difference, and this effect probably shouldn’t be considered part of the previous exponentials.

  3. The benefit of assigning tasks to AI starts to take off when you can reliably assign tasks for the AI without needing continuous human supervision, and now can treat those tasks as atomic actions not requiring state.

  4. If AI can take humans out of the effective loops in this research and work for more extended periods, watch the hell out (on many levels, but certainly in terms of capabilities and algorithmic progress.)

  5. Past a certain point where you can reliably do what one might call in-context atomic components, gaining the robustness and covering the gaps necessary to do this more reliably starts to get easier rather than harder, relative to the standard exponential curves.

  6. This could easily ‘go all the way’ to SC (and then quickly to full ASI) although we don’t know that it does. This is another uncertainty point, also note that AI 2027 as written very much involves waiting for various physical development steps.

  7. Thus, without making any claims about what the pace of all this is (and my guess is it is slower than they think it is, and also highly uncertain), the Baseline Scenario very much looks like AI 2027, but there’s a lot of probability mass also on other scenarios.

  8. One then has to ask what happens after you get this ‘superhuman coder’ or otherwise get ASI-like things of various types.

Which all adds up to me saying that I agree with Eli that none of the criticisms raised here challenges, to me, the ultimate or fundamental findings, only the price. The price is of course what we are here to talk about, so that is highly valuable even within relatively narrow bands (2028 is very different from 2029 because of reasons, and 2035 is rather different from that, and so on).

I realize that none of this is the kind of precision that lets you land on the moon.

The explanation for all this is right there: This is a physicist, holding forecasting of AI timelines to the standards of physics models. Well, yeah, you’re not going to be happy. If you try to use this to land on the moon, you will almost certainly miss the moon, the same way that if you try to use current alignment techniques on a superintelligence, you will almost certainly miss and then you will die.

One of the AI 2027 authors joked to me in the comments on a recent article that “you may not like it but it’s what peak AI forecasting performance looks like”.

Well, I don’t like it, and if this truly is “peak forecasting”, then perhaps forecasting should not be taken very seriously.

Maybe this is because I am a physicist, not a Rationalist. In my world, you generally want models to have strong conceptual justifications or empirical validation with existing data before you go making decisions based off their predictions: this fails at both.

Yes, in the world of physics, things work very differently, and we have much more accurate and better models. If you want physics-level accuracy in your predictions of anything that involves interactions of humans, well, sorry, tough luck. And presumably everyone agrees that you can’t have a physics-quality model here and that no one is claiming to have one? So what’s the issue?

The issue is whether basing decisions on modeling attempts like this is better than basing them on ‘I made it up’ or not having probabilities and projections at all and vibing the damn thing.

What I’m most against is people taking shoddy toy models seriously and basing life decisions on them, as I have seen happen for AI 2027.

I am not going to propose an alternate model. If I tried to read the tea leaves of the AI future, it would probably also be very shaky. There are a few things I am confident of, such as a software-only singularity not working and that there will be no diamondoid bacteria anytime soon. But these beliefs are hard to turn into precise yearly forecasts, and I think doing so will only cement overconfidence and leave people blindsided when reality turns out even weirder than you imagined..

Why is this person confident the software-only singularity won’t work? This post does not say. You’d have to read their substack, I assume it’s there.

The forecast here is ‘precise’ in the sense that it has a median, and we have informed people of that median. It is not ‘precise’ in the sense of putting a lot of probability mass on that particular median, even as an entire year, or even in the sense that the estimate wouldn’t change with more work or better data. It is precise in the sense that, yes, Bayes Rule is a thing, and you have to have a probability distribution, and it’s a lot more useful to share it than not share it.

I do find that the AI 2027 arguments updated me modestly towards a faster distribution of potential outcomes. I find 2027 to be a totally plausible time for SC to happen, although my median would be substantially longer.

You can’t ‘not base life decisions’ on information until it crosses some (higher than this) robustness threshold. Or I mean you can, but it will not go great.

In conclusion, I once again thank Titotal for the excellent substance of this critique, and wish it had come with better overall framing.

Discussion about this post

Analyzing A Critique Of The AI 2027 Timeline Forecasts Read More »

uk-looking-to-loosen-google’s-control-of-its-search-engine

UK looking to loosen Google’s control of its search engine

Other conduct rules that the CMA is considering include requirements in how it ranks its search results and for Google’s distribution partners such as Apple to offer “choice screens” to help consumers switch more easily between search providers.

The CMA said Alphabet-owned Google’s dominance made the cost of search advertising “higher than would be expected” in a more competitive market.

Google on Tuesday slammed the proposals as “broad and unfocused” and said they could threaten the UK’s access to its latest products and services.

Oliver Bethell, Google’s senior director for competition, warned that “punitive regulations” could change how quickly Google launches new products in the UK.

“Proportionate, evidence-based regulation will be essential to preventing the CMA’s road map from becoming a roadblock to growth in the UK,” he added.

Bethell’s warning of the potential impact of any regulations on the wider UK economy comes after the government explicitly mandated the CMA to focus on supporting growth and investment while minimizing uncertainty for businesses.

Google said last year that it planned to invest $1 billion in a huge new data center just outside London.

The CMA’s probe comes after Google lost a pair of historic US antitrust cases over its dominance of search and its lucrative advertising business.

© 2025 The Financial Times Ltd. All rights reserved. Not to be redistributed, copied, or modified in any way.

UK looking to loosen Google’s control of its search engine Read More »