Author name: Tim Belzer

prehistoric-bone-tool-cache-suggests-advanced-reasoning-in-early-hominins

Prehistoric bone tool cache suggests advanced reasoning in early hominins

A prehistoric bone tool “factory”

Credit: CSIC

Bone tools shaped by knapping, however, were much rarer until about 500,000 years ago, according to Peters and her fellow authors, making it challenging to identify consistent behaviors in the making and use of such tools. That is now changing with the discovery of a bone tool collection at the T69 Complex at Olduvai Gorge, specifically seven trenches excavated between 2015 and 2022. In addition to over 10,000 stone tools, there were abundant fish, crocodile, and hippopotamus remains, as well as those of elephants and rhinoceroses.

Among all the fossils and bone fragments, the authors identified 27 specimens that were clearly bone tools, evidenced by signs of intentional flake removal, shaping, and modification of bone edges to produce an elongated shape. The authors acknowledge that other non-intentional factors can cause such flaking, particularly the gnawing of carnivores. But carnivores made up less than 1 percent of the identified animal remains at the site and the 27 specimens did not show clear signs of such gnawing.

It seems the hominins who made the bone tools carefully selected the bones of large mammals, most commonly elephant and hippopotamus. “Precise anatomical knowledge and understanding of bone morphology are suggested by preference given to thick limb bones and the application of recurrent flaking procedures,” the authors wrote. These large, heavy bone tools may have been later replaced by larger stone tools, which might explain why they became so rare after the emergence of systemically produced lithic hand axes.

“This discovery leads us to assume that early humans significantly expanded their technological options, which until then were limited to the production of stone tools and now allowed new raw materials to be incorporated into the repertoire of potential artifacts,” said co-author Ignacio de la Torre of the CSIC-Spanish National Research Council. “At the same time, this expansion of technological potential indicates advances in the cognitive abilities and mental structures of these hominins, who knew how to incorporate technical innovations by adapting their knowledge of stone work to the manipulation of bone remains.”

Nature, 2025. DOI: 10.1038/s41586-025-08652-5 (About DOIs).

Prehistoric bone tool cache suggests advanced reasoning in early hominins Read More »

you-knew-it-was-coming:-google-begins-testing-ai-only-search-results

You knew it was coming: Google begins testing AI-only search results

Google has become so integral to online navigation that its name became a verb, meaning “to find things on the Internet.” Soon, Google might just tell you what’s on the Internet instead of showing you. The company has announced an expansion of its AI search features, powered by Gemini 2.0. Everyone will soon see more AI Overviews at the top of the results page, but Google is also testing a more substantial change in the form of AI Mode. This version of Google won’t show you the 10 blue links at all—Gemini completely takes over the results in AI Mode.

This marks the debut of Gemini 2.0 in Google search. Google announced the first Gemini 2.0 models in December 2024, beginning with the streamlined Gemini 2.0 Flash. The heavier versions of Gemini 2.0 are still in testing, but Google says it has tuned AI Overviews with this model to offer help with harder questions in the areas of math, coding, and multimodal queries.

With this update, you will begin seeing AI Overviews on more results pages, and minors with Google accounts will see AI results for the first time. In fact, even logged out users will see AI Overviews soon. This is a big change, but it’s only the start of Google’s plans for AI search.

Gemini 2.0 also powers the new AI Mode for search. It’s launching as an opt-in feature via Google’s Search Labs, offering a totally new alternative to search as we know it. This custom version of the Gemini large language model (LLM) skips the standard web links that have been part of every Google search thus far. The model uses “advanced reasoning, thinking, and multimodal capabilities” to build a response to your search, which can include web summaries, Knowledge Graph content, and shopping data. It’s essentially a bigger, more complex AI Overview.

As Google has previously pointed out, many searches are questions rather than a string of keywords. For those kinds of queries, an AI response could theoretically provide an answer more quickly than a list of 10 blue links. However, that relies on the AI response being useful and accurate, something that often still eludes generative AI systems like Gemini.

You knew it was coming: Google begins testing AI-only search results Read More »

yes,-we-are-about-to-be-treated-to-a-second-lunar-landing-in-a-week

Yes, we are about to be treated to a second lunar landing in a week

Because the space agency now has some expectation that Intuitive Machines will be fully successful with its second landing attempt, it has put some valuable experiments on board. Principal among them is the PRIME-1 experiment, which has an ice drill to sample any ice that lies below the surface. Drill, baby, drill.

The Athena lander also is carrying a NASA-funded “hopper” that will fire small hydrazine rockets to bounce around the Moon and explore lunar craters near the South Pole. It might even fly into a lava tube. If this happens it will be insanely cool.

Because this is a commercial program, NASA has encouraged the delivery companies to find additional, private payloads. Athena has some nifty ones, including a small rover from Lunar Outpost, a data center from Lonestar Data Holdings, and a 4G cellular network from Nokia. So there’s a lot riding on Athena‘s success.

So will it be a success?

“Of course, everybody’s wondering, are we gonna land upright?” Tim Crain, Intuitive Machines’ chief technology officer, told Ars. “So, I can tell you our laser test plan is much more comprehensive than those last time.”

During the first landing about a year ago, Odysseus‘ laser-based system for measuring altitude failed during the descent. Because Odysseus did not have access to altitude data, the spacecraft touched down faster, and on a 12-degree slope, which exceeded the 10-degree limit. As a result, the lander skidded across the surface, and one of its six legs broke, causing it to fall over.

Crain said about 10 major changes were made to the spacecraft and its software for the second mission. On top of that, about 30 smaller things, such as more efficient file management, were updated on the new vehicle.

In theory, everything should work this time. Intuitive Machines has the benefit of all of its learnings from the last time, and nearly everything worked right during this first attempt. But the acid test comes on Thursday.

The company and NASA will provide live coverage of the attempt beginning at 11: 30 am ET (16: 30 UTC) on NASA+, with landing set for just about one hour later. The Moon may be a harsh mistress, but hopefully not too harsh.

Yes, we are about to be treated to a second lunar landing in a week Read More »

brother-denies-using-firmware-updates-to-brick-printers-with-third-party-ink

Brother denies using firmware updates to brick printers with third-party ink

Brother laser printers are popular recommendations for people seeking a printer with none of the nonsense. By nonsense, we mean printers suddenly bricking features, like scanning or printing, if users install third-party cartridges. Some printer firms outright block third-party toner and ink, despite customer blowback and lawsuits. Brother’s laser printers have historically worked fine with non-Brother accessories. A YouTube video posted this week, though, as well as older social media posts, claim that Brother has gone to the dark side and degraded laser printer functionality with third-party cartridges. Brother tells Ars that this isn’t true.

On March 3, YouTuber Louis Rossman posted a video saying that “Brother turns heel & becomes anti-consumer printer company.” The video, spotted by Tom’s Hardware, has 163,000 views as of this writing and seems to be based on a Reddit post from 2022. In that post, Reddit user 20Factorial said that firmware update W1.56 caused the automatic color registration feature to stop working on his Brother MFC-3750 when using third-party cartridges.

“With the colors not able to be aligned, the printer is effectively non-functional,” 20Factorial said. The Redditor went on to say that when asked, a Brother customer service agent confirmed that “the printer is non-functional without genuine toner.”

Rossman created a Wiki page breaking down the reported issues, including “printers continue to function with third-party toner but print at degraded quality unless OEM toner is installed.” He also noted that Brother printers automatically update when connected to the Internet and that Brother doesn’t offer older firmware versions to users.

Brother’s response

Brother denied to Ars Technica that it intentionally bricks printer functionality when users install third-party toner or ink. In a statement, the company said:

Brother denies using firmware updates to brick printers with third-party ink Read More »

on-openai’s-safety-and-alignment-philosophy

On OpenAI’s Safety and Alignment Philosophy

OpenAI’s recent transparency on safety and alignment strategies has been extremely helpful and refreshing.

Their Model Spec 2.0 laid out how they want their models to behave. I offered a detailed critique of it, with my biggest criticisms focused on long term implications. The level of detail and openness here was extremely helpful.

Now we have another document, How We Think About Safety and Alignment. Again, they have laid out their thinking crisply and in excellent detail.

I have strong disagreements with several key assumptions underlying their position.

Given those assumptions, they have produced a strong document – here I focus on my disagreements, so I want to be clear that mostly I think this document was very good.

This post examines their key implicit and explicit assumptions.

In particular, there are three core assumptions that I challenge:

  1. AI Will Remain a ‘Mere Tool.’

  2. AI Will Not Disrupt ‘Economic Normal.’

  3. AI Progress Will Not Involve Phase Changes.

The first two are implicit. The third is explicit.

OpenAI recognizes the questions and problems, but we have different answers. Those answers come with very different implications:

  1. OpenAI thinks AI can remain a ‘Mere Tool’ despite very strong capabilities if we make that a design goal. I do think this is possible in theory, but that there are extreme competitive pressures against this that make that almost impossible, short of actions no one involved is going to like. Maintaining human control is to try and engineer what is in important ways an ‘unnatural’ result.

  2. OpenAI expects massive economic disruptions, ‘more change than we’ve seen since the 1500s,’ but that still mostly assumes what I call ‘economic normal,’ where humans remain economic agents, private property and basic rights are largely preserved, and easy availability of oxygen, water, sunlight and similar resources continues. I think this is not a good assumption.

  3. OpenAI is expecting what is for practical purposes continuous progress without major sudden phase changes. I believe their assumptions on this are far too strong, and that there have already been a number of discontinuous points with phase changes, and we will have more coming, and also that with sufficient capabilities many current trends in AI behaviors would reverse, perhaps gradually but also perhaps suddenly.

I’ll then cover their five (very good) core principles.

I call upon the other major labs to offer similar documents. I’d love to see their takes.

  1. Core Implicit Assumption: AI Can Remain a ‘Mere Tool’.

  2. Core Implicit Assumption: ‘Economic Normal’.

  3. Core Assumption: No Abrupt Phase Changes.

  4. Implicit Assumption: Release of AI Models Only Matters Directly.

  5. On Their Taxonomy of Potential Risks.

  6. The Need for Coordination.

  7. Core Principles.

  8. Embracing Uncertainty.

  9. Defense in Depth.

  10. Methods That Scale.

  11. Human Control.

  12. Community Effort.

This is the biggest crux. OpenAI thinks that this is a viable principle to aim for. I don’t see how.

OpenAI imagines that AI will remain a ‘mere tool’ indefinitely. Humans will direct AIs, and AIs will do what the humans direct the AIs to do. Humans will remain in control, and remain ‘in the loop,’ and we can design to ensure that happens. When we model a future society, we need not imagine AIs, or collections of AIs, as if they were independent or competing economic agents or entities.

Thus, our goal in AI safety and alignment is to ensure the tools do what we intend them to do, and to guard against human misuse in various forms, and to prepare society for technological disruption similar to what we’d face with other techs. Essentially, This Time is Not Different.

Thus, the Model Spec and other such documents are plans for how to govern an AI assistant mere tool, assert a chain of command, and how to deal with the issues that come along with that.

That’s a great thing to do for now, but as a long term outlook I think this is Obvious Nonsense. A sufficiently capable AI might (or might not) be something that a human operating it could choose to leave as a ‘mere tool.’ But even under optimistic assumptions, you’d have to sacrifice a lot of utility to do so.

It does not have a goal? We can and will effectively give it a goal.

It is not an agent? We can and will make it an agent.

Human in the loop? We can and will take the human out of the loop once the human is not contributing to the loop.

OpenAI builds AI agents and features in ways designed to keep humans in the loop and ensure the AIs are indeed mere tools, as suggested in their presentation at the Paris summit? They will face dramatic competitive pressures to compromise on that. People will do everything to undo those restrictions. What’s the plan?

Thus, even if we solve alignment in every useful sense, and even if we know how to keep AIs as ‘mere tools’ if desired, we would rapidly face extreme competitive pressures towards gradual disempowerment, as AIs are given more and more autonomy and authority because that is the locally effective thing to do (and also others do it for the lulz, or unintentionally, or because they think AIs being in charge or ‘free’ is good).

Until a plan tackles these questions seriously, you do not have a serious plan.

What I mean by ‘Economic Normal’ is something rather forgiving – that the world does not transform in ways that render our economic intuitions irrelevant, or that invalidate economic actions. The document notes they expect ‘more change than from the 1500s to the present’ and the 1500s would definitely count as fully economic normal here.

It roughly means that your private property is preserved in a way that allows your savings to retain purchasing power, your rights to bodily autonomy and (very) basic rights are respected, your access to the basic requirements of survival (sunlight, water, oxygen and so on) are not disrupted or made dramatically more expensive on net, and so on. It also means that economic growth does not grow so dramatically as to throw all your intuitions out the window.

That things will not enter true High Weirdness, and that financial or physical wealth will meaningfully protect you from events.

I do not believe these are remotely safe assumptions.

AGI is notoriously hard to define or pin down. There are not two distinct categories of things, ‘definitely not AGI’ and then ‘fully AGI.’

Nor do we expect an instant transition from ‘AI not good enough to do much’ to ‘AI does recursive self-improvement.’ AI is already good enough to do much, and will probably get far more useful before things ‘go critical.’

That does not mean that there are not important phase changes between models, where the precautions and safety measures you were previously using either stop working or are no longer matched to the new threats.

AI is still on an exponential.

If we treat past performance as assuring us of future success, if we do not want to respond to an exponential ‘too early’ based on the impacts we can already observe, what happens? We will inevitably respond too late.

I think the history of GPT-2 actually illustrates this. If we conclude from that incident that OpenAI did something stupid and ‘looked silly,’ without understanding exactly why the decision was a mistake, we are in so so much trouble.

We used to view the development of AGI as a discontinuous moment when our AI systems would transform from solving toy problems to world-changing ones. We now view the first AGI as just one point along a series of systems of increasing usefulness.

In a discontinuous world, practicing for the AGI moment is the only thing we can do, and it leads to treating the systems of today with a level of caution disproportionate to their apparent power.

This is the approach we took for GPT-2 when we didn’t release the model due to concerns about malicious applications.

In the continuous world, the way to make the next system safe and beneficial is to learn from the current system. This is why we’ve adopted the principle of iterative deployment, so that we can enrich our understanding of safety and misuse, give society time to adapt to changes, and put the benefits of AI into people’s hands.

At present, we are navigating the new paradigm of chain-of-thought models – we believe this technology will be extremely impactful going forward, and we want to study how to make it useful and safe by learning from its real-world usage. In the continuous world view, deployment aids rather than opposes safety.

In the continuous world view, deployment aids rather than opposes safety.

At the current margins, subject to proper precautions and mitigations, I agree with this strategy of iterative deployment. Making models available, on net, is helpful.

However, we forget what happened with GPT-2. The demand was that the full GPT-2 be released as an open model, right away, despite it being a phase change in AI capabilities that potentially enabled malicious uses, with no one understanding what the impact might be. It turned out the answer was ‘nothing,’ but the point of iterative deployment is to test that theory while still being able to turn the damn thing off. That’s exactly what happened. The concerns look silly now, but that’s hindsight.

Similarly, there have been several cases of what sure felt like discontinuous progress since then. If we restrict ourselves to the ‘OpenAI extended universe,’ GPT-3, GPT-3.5, GPT-4, o1 and Deep Research (including o3) all feel like plausible cases where new modalities potentially opened up, and new things happened.

The most important potential phase changes lie in the future, especially the ones where various safety and alignment strategies potentially stop working, or capabilities make such failures far more dangerous, and it is quite likely these two things happen at the same time because one is a key cause of the other. And if you buy ‘o-ring’ style arguments, where AI is not so useful so long as there must be a human in the loop, removing the last need for such a human is a really big deal.

Alternatively: Iterative deployment can be great if and only if you use it in part to figure out when to stop.

I would also draw a distinction between open iterative deployment and closed iterative deployment. Closed iterative deployment can be far more aggressive while staying responsible, since you have much better options available to you if something goes awry.

I also think the logic here is wrong:

These diverging views of the world lead to different interpretations of what is safe.

For example, our release of ChatGPT was a Rorschach test for many in the field — depending on whether they expected AI progress to be discontinuous or continuous, they viewed it as either a detriment or learning opportunity towards AGI safety.

The primary impacts of ChatGPT were

  1. As a starting gun that triggered massively increased use, interest and spending on LLMs and AI. That impact has little to do with whether progress is continuous or discontinuous.

  2. As a way to massively increase capital and mindshare available to OpenAI.

  3. Helping transform OpenAI into a product company.

You can argue about whether those impacts were net positive or not. But they do not directly interact much with whether AI progress is centrally continuous.

Another consideration is various forms of distillation or reverse engineering, or other ways in which making your model available could accelerate others.

And there’s all the other ways in which perception of progress, and of relative positioning, impacts people’s decisions. It is bizarre how much the exact timing of the release of DeepSeek’s r1, relative to several other models, mattered.

Precedent matters too. If you get everyone in the habit of releasing models the moment they’re ready, it impacts their decisions, not only yours.

This is the most important detail-level disagreement, especially in the ways I fear that the document will be used and interpreted, both internally to OpenAI and also externally, even if the document’s authors know better.

It largely comes directly from applying the ‘mere tool’ and ‘economic normal’ assumptions.

As AI becomes more powerful, the stakes grow higher. The exact way the post-AGI world will look is hard to predict — the world will likely be more different from today’s world than today’s is from the 1500s. But we expect the transformative impact of AGI to start within a few years. From today’s AI systems, we see three broad categories of failures:

  1. Human misuse: We consider misuse to be when humans apply AI in ways that violate laws and democratic values. This includes suppression of free speech and thought, whether by political bias, censorship, surveillance, or personalized propaganda. It includes phishing attacks or scams. It also includes enabling malicious actors to cause harm at a new scale.

  2. Misaligned AI: We consider misalignment failures to be when an AI’s behavior or actions are not in line with relevant human values, instructions, goals, or intent. For example an AI might take actions on behalf of its user that have unintended negative consequences, influence humans to take actions they would otherwise not, or undermine human control. The more power the AI has, the bigger potential consequences are.

  3. Societal disruption: AI will bring rapid change, which can have unpredictable and possibly negative effects on the world or individuals, like social tensions, disparities and inequality, and shifts in dominant values and societal norms. Access to AGI will determine economic success, which risks authoritarian regimes pulling ahead of democratic ones if they harness AGI more effectively.

There are two categories of concern here, in addition to the ‘democratic values’ Shibboleth issue.

  1. As introduced, this is framed as ‘from today’s AI systems.’ In which case, this is a lot closer to accurate. But the way the descriptions are written clearly implies this is meant to cover AGI as well, where this taxonomy seems even less complete and less useful for cutting reality at its joints.

  2. This is in a technical sense a full taxonomy, but de facto it ignores large portions of the impact of AI and of the threat model that I am using.

When I say technically a full taxonomy, you could say this is essentially saying either:

  1. The human does something directly bad, on purpose.

  2. The AI does something directly bad, that the human didn’t intend.

  3. Nothing directly bad happens per se, but bad things happen overall anyway.

Put it like that, and what else is there? Yet the details don’t reflect the three options being fully covered, as summarized there. In particular, ‘societal disruption’ implies a far narrower set of impacts than we need to consider, but similar issues exist with all three.

  1. Human Misuse.

A human might do something bad using an AI, but how are we pinning that down?

Saying ‘violates the law’ puts an unreasonable burden on the law. Our laws, as they currently exist, are complex and contradictory and woefully unfit and inadequate for an AGI-infused world. The rules are designed for very different levels of friction, and very different social and other dynamics, and are written on the assumption of highly irregular enforcement. Many of them are deeply stupid.

If a human uses AI to assemble a new virus, that certainly is what they mean by ‘enabling malicious actors to cause harm at a new scale’ but the concern is not ‘did that break the law?’ nor is it ‘did this violate democratic values.’

Saying ‘democratic values’ is a Shibboleth and semantic stop sign. What are these ‘democratic values’? Things the majority of people would dislike? Things that go against the ‘values’ the majority of people socially express, or that we like to pretend our society strongly supports? Things that change people’s opinions in the wrong ways, or wrong directions, according to some sort of expert class?

Why is ‘personalized propaganda’ bad, other than the way that is presented? What exactly differentiates it from telling an AI to write a personalized email? Why is personalized bad but non-personalized fine and where is the line here? What differentiates ‘surveillance’ from gathering information, and does it matter if the government is the one doing it? What the hell is ‘political bias’ in the context of ‘suppression of free speech’ via ‘human misuse’? And why are these kinds of questions taking up most of the misuse section?

Most of all, this draws a box around ‘misuse’ and treats that as a distinct category from ‘use,’ in a way I think will be increasingly misleading. Certainly we can point to particular things that can go horribly wrong, and label and guard against those. But so much of what people want to do, or are incentivized to do, is not exactly ‘misuse’ but has plenty of negative side effects, especially if done at unprecedented scale, often in ways not centrally pointed at by ‘societal disruption’ even if they technically count. That doesn’t mean there is obviously anything to be done or that should be done about such things, banning things should be done with extreme caution, but it not being ‘misuse’ does not mean the problems go away.

  1. Misaligned AI.

There are three issues here:

  1. The longstanding question of what even is misaligned.

  2. The limited implied scope of the negative consequences.

  3. The implication that the AI has to be misaligned to pose related dangers.

AI is only considered misaligned here when it is not in line with relevant human values, instructions, goals or intent. If you read that literally, as an AI that is not in line with all four of these things, even then it can still easily bleed into questions of misuse, in ways that threaten to drop overlapping cases on the floor.

I don’t mean to imply there’s something great that could have been written here instead, but: This doesn’t actually tell us much about what ‘alignment’ means in practice. There are all sorts of classic questions about what happens when you give an AI instructions or goals that imply terrible outcomes, as indeed almost all maximalist or precise instructions and goals do at the limit. It doesn’t tell us what ‘human values’ are in various senses.

On scope, I do appreciate that it says the more power the AI has, the bigger potential consequences are. And ‘undermine human control’ can imply a broad range of dangers. But the scope seems severely limited here.

Especially worrisome is that the examples imply that the actions would still be taken ‘on behalf of its user’ and merely have unintended negative consequences. Misaligned AI could take actions very much not on behalf of its user, or might quickly fail to effectively have a user at all. Again, this is the ‘mere tool’ assumption run amok.

  1. Social disruption

Here once again we see ‘economic normal’ and ‘mere tool’ playing key roles.

The wrong regimes – the ‘authoritarian’ ones – might pull ahead, or we might see ‘inequality’ or ‘social tensions.’ Or shifts in ‘dominant values’ and ‘social norms.’ But the base idea of human society is assumed to remain in place, with social dynamics remaining between humans. The worry is that society will elevate the wrong humans, not that society would favor AIs over humans or cease to effectively contain humans at all, or that humans might lose control over events.

To me, this does not feel like it addresses much of what I worry about in terms of societal disruptions, or even if it technically does it gives the impression it doesn’t.

We should worry far more about social disruptions in the sense that AIs take over and humans lose control, or AIs outcompete humans and render them non-competitive and non-productive, rather than worries about relatively smaller problems that are far more amenable to being fixed after things go wrong.

  1. Gradual disempowerment

The ‘mere tool’ blind spot is especially important here.

The missing fourth category, or at least thing to highlight even if it is technically already covered, is that the local incentives will often be to turn things over to AI to pursue local objectives more efficiently, but in ways that cause humans to progressively lose control. Human control is a core principle listed in the document, but I don’t see the approach to retaining it here as viable, and it should be more clearly here in the risk section. This shift will also impact events in other ways that cause negative externalities we will find very difficult to ‘price in’ and deal with once the levels of friction involved are sufficiently reduced.

There need not be any ‘misalignment’ or ‘misuse.’ Everyone following the local incentives leading to overall success is a fortunate fact about how things have mostly worked up until now, and also depended on a bunch of facts about humans and the technologies available to them, and how those humans have to operate and relate to each other. And it’s also depended on our ability to adjust things to fix the failure modes as we go to ensure it continues to be true.

I want to highlight an important statement:

Like with any new technology, there will be disruptive effects, some that are inseparable from progress, some that can be managed well, and some that may be unavoidable.

Societies will have to find ways of democratically deciding about these trade-offs, and many solutions will require complex coordination and shared responsibility.

Each failure mode carries risks that range from already present to speculative, and from affecting one person to painful setbacks for humanity to irrecoverable loss of human thriving.

This downplays the situation, merely describing us as facing ‘trade-offs,’ although it correctly points to the stakes of ‘irrecoverable loss of human thriving,’ even if I wish the wording on that (e.g. ‘extinction’) was more blunt. And it once again fetishizes ‘democratic’ decisions, presumably with only humans voting, without thinking much about how to operationalize that or deal with the humans both being heavily AI influenced and not being equipped to make good decisions any other way.

The biggest thing, however, is to affirm that yes, we only have a chance if we have the ability to do complex coordination and share responsibility. We will need some form of coordination mechanism, that allows us to collectively steer the future away from worse outcomes towards better outcomes.

The problem is that somehow, there is a remarkably vocal Anarchist Caucus, who thinks that the human ability to coordinate is inherently awful and we need to destroy and avoid it at all costs. They call it ‘tyranny’ and ‘authoritarianism’ if you suggest that humans retain any ability to steer the future at all, asserting that the ability of humans to steer the future via any mechanism at all is a greater danger (‘concentration of power’) than all other dangers combined would be if we simply let nature take its course.

I strongly disagree, and wish people understood what such people were advocating for, and how extreme and insane a position it is both within and outside of AI, and to what extent it quite obviously cannot work, and inevitably ends with either us all getting killed or some force asserting control.

Coordination is hard.

Coordination, on the level we need it, might be borderline impossible. Indeed, many in the various forms of the Suicide Caucus argue that because Coordination is Hard, we should give up on coordination with ‘enemies,’ and therefore we must Fail Game Theory Forever and all race full speed ahead into the twirling razor blades.

I’m used to dealing with that.

I don’t know if I will ever get used to the position that Coordination is The Great Evil, even democratic coordination among allies, and must be destroyed. That because humans inevitably abuse power, humans must not have any power.

The result would be that humans would not have any power.

And then, quickly, there wouldn’t be humans.

They outline five core principles.

  1. Embracing Uncertainty: We treat safety as a science, learning from iterative deployment rather than just theoretical principles.

  2. Defense in Depth: We stack interventions to create safety through redundancy.

  3. Methods that Scale: We seek out safety methods that become more effective as models become more capable.

  4. Human Control: We work to develop AI that elevates humanity and promotes democratic ideals.

  5. Shared Responsibility: We view responsibility for advancing safety as a collective effort.

I’ll take each in turn.

Embracing uncertainty is vital. The question is, what helps you embrace it?

If you have sufficient uncertainty about the safety of deployment, then it would be very strange to ‘embrace’ that uncertainty by deploying anyway. That goes double, of course, for deployments that one cannot undo, or which are sufficiently powerful they might render you unable to undo them (e.g. they might escape control, exfiltrate, etc).

So the question is, when does it reduce uncertainty to release models and learn, versus when it increases uncertainty more to do that? And what other considerations are there, in both directions? They recognize that the calculus on this could flip in the future, as quoted below.

I am both sympathetic and cynical here. I think OpenAI’s iterative development is primarily a business case, the same as everyone else’s, but that right now that business case is extremely compelling. I do think for now the safety case supports that decision, but view that as essentially a coincidence.

In particular, my worry is that alignment and safety considerations are, along with other elements, headed towards a key phase change, in addition to other potential phase changes. They do address this under ‘methods that scale,’ which is excellent, but I think the problem is far harder and more fundamental than they recognize.

Some excellent quotes here:

Our approach demands hard work, careful decision-making, and continuous calibration of risks and benefits.

The best time to act is before risks fully materialize, initiating mitigation efforts as potential negative impacts — such as facilitation of malicious use-cases or the model deceiving its operator— begin to surface.

In the future, we may see scenarios where the model risks become unacceptable even relative to benefits. We’ll work hard to figure out how to mitigate those risks so that the benefits of the model can be realized. Along the way, we’ll likely test them in secure, controlled settings.

For example, making increasingly capable models widely available by sharing their weights should include considering a reasonable range of ways a malicious party could feasibly modify the model, including by finetuning (see our 2024 statement on open model weights).

Yes, if you release an open weights model you need to anticipate likely modifications including fine-tuning, and not pretend your mitigations remain in place unless you have a reason to expect them to remain in place. Right now, we do not expect that.

It’s (almost) never a bad idea to use defense in depth on top of your protocol.

My worry is that in a crisis, all relevant correlations go to 1.

As in, as your models get increasingly capable, if your safety and alignment training fails, then your safety testing will be increasingly unreliable, and it will be increasingly able to get around your inference time safety, monitoring, investigations and enforcement.

Its ability to get around these four additional layers are all highly correlated to each other. The skills that get you around one mostly get you around the others. So this isn’t as much defense in depth as you would like it to be.

That doesn’t mean don’t do it. Certainly there are cases, especially involving misuse or things going out of distribution in strange but non-malicious ways, where you will be able to fail early, then recover later on. The worry is that when the stakes are high, that becomes a lot less likely, and you should think of this as maybe one effective ‘reroll’ at most rather than four.

To align increasingly intelligent models, especially models that are more intelligent and powerful than humans, we must develop alignment methods that improve rather than break with increasing AI intelligence.

I am in violent agreement. The question is which methods will scale.

There are also two different levels at which we must ask what scales.

Does it scale as AI capabilities increase on the margin, right now? A lot of alignment techniques right now are essentially ‘have the AI figure out what you meant.’ On the margin right now, more intelligence and capability of the AI mean better answers.

Deliberative alignment is the perfect example of this. It’s great for mundane safety right now and will get better in the short term. Having the model think about how to follow your specified rules will improve as intelligence improves, as long as the goal of obeying your rules as written gets you what you want. However, if you apply too much optimization pressure and intelligence to any particular set of deontological rules as you move out of distribution, even under DWIM (do what I mean, or the spirit of the rules) I predict disaster.

In addition, under amplification, or attempts to move ‘up the chain’ of capabilities, I worry that you can hope to copy your understanding, but not to improve it. And as they say, if you make a copy of a copy of a copy, it’s not quite as sharp as the original.

I approve of everything they describe here, other than worries about the fetishization of democracy, please do all of it. But I don’t see how this allows humans to remain in effective control. These techniques are already hard to get right and aim to solve hard problems, but the full hard problems of control remain unaddressed.

Another excellent category, where they affirm the need to do safety work in public, fund it and support it, including government expertise, propose policy initiatives and make voluntary commitments.

There is definitely a lot of room for improvement in OpenAI and Sam Altman’s public facing communications and commitments.

Discussion about this post

On OpenAI’s Safety and Alignment Philosophy Read More »

on-writing-#1

On Writing #1

This isn’t primarily about how I write. It’s about how other people write, and what advice they give on how to write, and how I react to and relate to that advice.

I’ve been collecting those notes for a while. I figured I would share.

At some point in the future, I’ll talk more about my own process – my guess is that what I do very much wouldn’t work for most people, but would be excellent for some.

  1. How Marc Andreessen Writes.

  2. How Sarah Constantin Writes.

  3. How Paul Graham Writes.

  4. How Patrick McKenzie Writes.

  5. How Tim Urban Writes.

  6. How Visakan Veerasamy Writes.

  7. How Matt Yglesias Writes.

  8. How JRR Tolkien Wrote.

  9. How Roon Wants Us to Write.

  10. When To Write the Headline.

  11. Do Not Write Self-Deprecating Descriptions of Your Posts.

  12. Do Not Write a Book.

  13. Write Like No One Else is Reading.

  14. Letting the AI Write For You.

  15. Being Matt Levine.

  16. The Case for Italics.

  17. Getting Paid.

  18. Having Impact.

Marc Andreessen starts with an outline, written as quickly as possible, often using bullet points.

David Perell: When Marc Andreessen is ready to write something, he makes an outline as fast as possible.

Bullet points are fine. His goal is to splatter the page with ideas while his mind is buzzing. Only later does he think about organizing what he’s written.

He says: “I’m trying to get all the points out and I don’t want to slow down the process by turning them all into prose. It’s not a detailed outline like something a novelist would have. It’s basically bullet points.”

Marc is saying that first you write out your points and conclusion, then you fill in the details. He wants to figure it all out while his mind is buzzing, then justify it later.

Whereas I learn what I think as I write out my ideas in detail. I would say that if you are only jotting down bullet points, you do not yet know what you think.

Where we both agree is that of course you should write notes to remember key new ideas, and also that the organizing what goes where can be done later.

I do not think it is a coincidence that this is the opposite of my procedure. Yes, I have some idea of what I’m setting out to write, but it takes form as I write it, and as I write I understand.

If you’re starting with a conclusion, then writing an outline, and writing them quickly, that says you are looking to communicate what you already know, rather than seeking to yourself learn via the process.

A classic rationalist warning is to not write the bottom line first.

Sarah Constantin offers an FAQ on how she writes. Some overlap, also a lot of big differences. I especially endorse doing lots of micro-edits and moving things around and seeing how they develop as they go. I dismiss the whole ‘make an outline’ thing they teach you in school as training wheels at best and Obvious Nonsense at worst.

I also strongly agree with her arguments that you need to get the vibe right. I would extend this principle to needing to be aware of all four simulacra levels at once at all times. Say true and helpful things, keeping in mind what people might do with that information, what your statements say about which ‘teams’ you are on in various ways, and notice the vibes and associations being laid down and how you are sculpting and walking the paths through cognitive space for yourself and others to navigate. Mostly you want to play defensively on level two (make sure you don’t give people the wrong idea), and especially on level three (don’t accidentally make people associate you with the wrong faction, or ideally any faction), and have a ‘don’t be evil’ style rule for level four (vibe well on all levels, and avoid unforced errors, but vibe justly and don’t take cheap shots), with the core focus always at level one.

I think this is directionally right, I definitely won’t leave a wrong idea in writing:

Paul Graham: Sometimes when writing an essay I’ll leave a clumsy sentence to fix later. But I never leave an idea I notice is wrong. Partly because it could damage the essay, and partly because you don’t need to: noticing an idea is wrong starts you toward fixing it.

However I also won’t leave a clumsy sentence that I wasn’t comfortable being in the final version. I will often go back and edit what I’ve written, hopefully improving it, but if I wasn’t willing to hit post with what I have now then I wouldn’t leave it there.

In the cases where this is not true, I’m going explicitly leave a note, in [brackets] and usually including a [tktk], saying very clearly that there is a showstopper bug here.

Here’s another interesting contrast in our styles.

Paul Graham: One surprising thing about office hours with startups is that they scramble your brain. It’s the context switching. You dive deep into one problem, then deep into another completely different one, then another. At the end you can’t remember what the first startup was even doing.

This is why I write in the mornings and do office hours in the afternoon. Writing essays is harder. I can’t do it with a scrambled brain.

It’s fun up to about 5 or 6 startups. 8 is possible. 12 you’d be a zombie.

I feel this way at conferences. You’re constantly context switching, a lot of it isn’t retained, but you go with the flow, try to retain the stuff that matters most and take a few notes, and hope others get a lot out of it.

The worst of that was at EA Global: Boston, where you are by default taking a constant stream of 25 minute 1-on-1s. By the end of the day it was mostly a blur.

When I write, however, mostly it’s the opposite experience to Graham’s writing – it’s constant context switching from one problem to the next. Even while doing that, I’m doing extra context switching for breaks.

A lot of that is presumably different types of writing. Graham is trying to write essays that are tighter, more abstract, more structured, trying to make a point. I’m trying to learn and explore and process and find out.

Which is why I basically can indeed do it with a scrambled brain, and indeed have optimized for that ability – to be able to process writing subtasks without having to load in lots of state.

Patrick McKenzie on writing fast and slow, formal and informal, and the invocation of deep magick. On one topic he brings up: My experience on ‘sounding natural’ in writing is that you can either sound natural by writing in quick natural form, or you can put in crazy amounts of work to make it happen, and anything in between won’t work. Also I try to be careful to intentionally not invoke the deep magick in most situations. One only wants to be a Dangerous Professional when the situation requires it, and you need to take on a faceless enemy in Easy Mode.

Patrick McKenzie also notes that skilled writers have a ton of control over exactly how controversial their statements will effectively be. I can confirm this. Also I can confirm that mistakes are often made, which is a Skill Issue.

Tim Urban says writing remains very hard.

Tim Urban: No matter how much I write, writing remains hard. Those magical moments when I’m in a real flow, it seems easy, but most of the time, I spend half a day writing and rewriting the same three paragraphs trying to figure out the puzzle of making them not suck.

Being in a writing flow is like when I’m golfing and hit three good shots in a row and think “k finally figured this out and I’m good now.” Unfortunately the writing muse and the golf fairy both like to vanish without a trace and leave me helpless in a pit of my own incompetence.

Dustin Burnham: Periods of brilliance would escape Douglas Adams for so long that he had to be locked in a hotel room by his editors to finish The Hitchhiker’s Guide to the Galaxy.

I definitely have a lot of moments when I don’t feel productive, usually because my brain isn’t focused or on. I try to have a stack of other productive things I can do despite being unproductive while that passes.

But over time, yes, I’ve found the writing itself does get easy for me? Often figuring out what I think, or what I want to write about, is hard, but the writing itself comes relatively easily.

Yes, you can then go over it ten times and edit to an inch of its life if you want, but the whole ‘rewriting the same three paragraphs’ thing is very rare. I think the only times I did it this year I was pitching to The New York Times.

What’s the best target when writing?

Visakan Veerasamy: Charts out for ribbonfarm.

I do endorse the core thing this is trying to suggest: To explore more and worry about presentation and details less, on most margins. And to know that in a real sense, if you have truly compelling fuckery, you have wiggled your big toe. Hard part is over.

I do not think the core claim itself is correct. Or perhaps we mean different things by resonant and coherent? By coherent, in my lexicon, he means more like ‘well-executed’ or ‘polished’ or something. By resonant, in my lexicon, he means more like ‘on to something central, true and important.’ Whereas to me resonant is a vibe, fully compatible with bullshit, ornate or otherwise.

Matt Yglesias reflects on four years of Slow Boring. He notes that it pays to be weird, to focus where you have comparative advantage rather than following the news of the week and fighting for a small piece of the biggest pies. He also notes the danger of repeating yourself, which I worry about as well.

Thread from 2020 on Tolkien’s path to writing Lord of the Rings. I’ve never done anything remotely like this, which might be some of why I haven’t done fiction.

Roon calls for the end of all this boring plain language, and I am here for it.

Roon: I love the guy, but I want the post-Goldwater era of utilitarian philosophical writing to be over. Bring back big words and epic prose, and sentences that make sense only at an angle.

Eliezer Yudkowsky: I expect Claude to do a good job of faking your favorite continental styles if you ask, since it requires little logical thinking, only vibes. You can produce and consume it privately in peace, avoiding its negative externalities, and leave the rest of us to our utility.

Roon: Eliezer, you are a good writer who often speaks in parables and communicates through fiction and isn’t afraid of interesting uses of language. You’ve certainly never shied away from verbosity, and that’s exactly what I’m talking about.

Perhaps some day I will learn how to write fiction. My experiences with AI reinforce to me that I really, really don’t know how to do that.

I usually write the headline last. Others disagree.

Luke Kawa: Joe Weisenthal always used to say don’t write a post until you know the headline first. More and more on short posts I find myself thinking “don’t write the post until you know the meme you can post with it first.”

As I said on Twitter, everyone as LessOnline instead gave the advice to not let things be barriers to writing. If you want to be a writer, write more, then edit or toss it out later, but you have to write.

Also, as others pointed out, if you start with the headline every time you are building habits of going for engagement if not clickbait, rather than following curiosity.

Other times, yes, you know exactly what the headline is before you start, because if you know you know.

I confirm this is sometimes true (but not always):

Patrick McKenzie: Memo to self and CCing other writers on an FYI basis:

If when announcing a piece you make a self-deprecating comment about it, many people who cite you will give a qualified recommendation of the piece, trying to excuse the flaw that you were joking about.

You think I would understand that ~20 years of writing publicly, but sometimes I cannot help myself from making the self-deprecating comment, and now half of the citations of my best work this year feel they need to disclaim that it is 24k words.

You can safely make the self-deprecating comments within the post itself. That’s fine.

Don’t write a book. If you do, chances are you’d sell dozens of copies, and earn at most similar quantities of dollars. The odds are very much doubleplusungood. Do you want to go on podcasts this much?

If you must write one anyway, how to sell it? The advice here is that books mostly get sold through recommendations. To get those, Eric Jorgenson’s model is you need three things:

  1. Finishable. If they don’t finish it, they won’t recommend it. So tighten it up.

  2. Unique OR Excellent. Be the best like no one ever was, or be like no one ever was.

  3. Memorable. Have hooks, for when people ask for a book about or for some X.

If you are thinking about writing a book, remember that no one would buy it.

Michael Dempsey and Ava endorse the principle of writing things down on the internet even if you do not expect anyone to read them.

Michael Dempsey: I loved this thread from Ava.

My entire career is owed to my willingness to write on the Internet.

And that willingness pushed me to write more in my personal life to loved ones.

As long as you recognize that most people will not care, your posts probably will not go viral, but at some point, one person might read something you write and reach out (or will value you including a blue link to your thoughts from many days, weeks, months, or years ago), it’s close to zero downside and all upside.

Ava: I’m starting to believe that “write on the Internet, even if no one reads it” is underrated life advice. It does not benefit other people necessarily; it benefits you because the people who do find or like your writing and then reach out are so much more likely to be compatible with you.

It’s also a muscle. I used to have so much anxiety posting anything online, and now I’m just like “lol, if you don’t like it, just click away.” People underestimate the sheer amount of content on the Internet; the chance of someone being angry at you for something is infinitely lower than no one caring.

I think it’s because everyone always sees outrage going viral, and you think “oh, that could be me,” and forget that most people causing outrage are trying very hard to be outrageous. By default, no one cares, or maybe five people care, and maybe some nice strangers like your stuff, and that’s a win.

Also, this really teaches you how to look for content you actually like on the Internet instead of passively receiving what is funneled to you. Some of my favorite Internet experiences have been reading a personal blog linked to someone’s website, never marketed, probably only their friends and family know about it, and it’s just the coolest peek into their mind.

I think the thing I’m trying to say here is “most people could benefit from writing online, whether you should market your writing aggressively is a completely different issue.” I wrote on Tumblr and Facebook and email for many years before Substack, and 20 people read it, and that was great.

I would not broadly recommend “trying to make a living off your writing online,” but that’s very different from “share some writing online.”

What is the number of readers that justifies writing something down? Often the correct answer is zero, even a definite zero. Even if it’s only about those who read it, twenty readers is actually a lot of value to put out there, and a lot of potential connection.

Paul Graham predicts that AI will cause the world to divide even more into writes and write-nots. Writing well and learning to write well are both hard, especially because it requires you to think well (and is how you think well), so once AI can do it for us without the need to hire someone or plagiarize, most people won’t learn (and one might add, thanks to AI doing the homework they won’t have to), and increasingly rely on AI to do it for them. Which in turn means those people won’t be thinking well, either, since you need to write to think well.

I think Graham is overstating the extent AI will free people from the pressure to write. Getting AI to write well in turn, and write what you actually want it to write, requires good writing and thinking, and involving AI in your writing OODA loop is often not cheap to do. Yes, more people will choose not to invest in the skill, but I don’t think this takes the pressure off as much as he expects, at least until AI gets a lot better.

There’s also the question of how much we should force people to write anyway, in order to make them think, or be able to think.

As Graham notes, getting rid of the middle ground could be quite bad:

Robin Hanson: But most jobs need real thinking. So either the LLMs will actually do that thinking for them, or workers will continue to write, in order to continue to think. I’d bet on the latter, for decades at least.

Perry Metzger: Why do we still teach kids mathematics, even though at this point, most of the grunt work is done better by computers, even for symbolic manipulation? Because if they’re going to be able to think, they need to practice thinking.

Most jobs don’t require real thinking. Proof: Most people can’t write.

One could argue that many jobs require ‘mid-level’ real thinking, the kind that might be lost, but I think mostly this is not the case. Most tasks and jobs don’t require real thinking at all, as we are talking about it here. Being able to do it? Still highly useful.

On the rare occasions the person can indeed do real thinking, it’s often highly valuable, but the jobs are designed knowing most people can’t and won’t do that.

Gwern asks, why are there so few Matt Levines? His conclusion is that Being Matt Levine requires both that a subject be amenable to a Matt Levine, which most aren’t, and also that there be a Matt Levine covering them, and Matt Levines are both born rather than made and highly rare.

In particular, a Matt Levine has to shout things into the void, over and over, repeating simple explanations time and again, and the subject has to involve many rapidly-resolved example problems to work through, with clear resolutions.

The place where I most epicly fail to be a Matt Levine in this model is my failure to properly address the beginner mindset and keep things simple. My choice to cater to a narrower, more advanced crowd, one that embraces more complexity, means I can’t go wide the way he can. That does seem right.

I could try to change this, but I mostly choose not to. I write too many words as it is.

The case for italics. I used to use italics a lot.

Char: “never italicise words to show emphasis! if you’re writing well your reader will know. you don’t need them!” me: oh 𝘳𝘦𝘢𝘭𝘭𝘺? listen up buddy, you will have to pry my emotional support italics from my 𝘤𝘰𝘭𝘥, 𝘥𝘦𝘢𝘥, 𝘧𝘪𝘯𝘨𝘦𝘳𝘴, they are going 𝘯𝘰𝘸𝘩𝘦𝘳𝘦.

Richard White: I’m coming to the conclusion that about 99.9% of all writing “rules” can safely be ignored. As long as your consistent with your application of whatever you’re doing, it’ll be fine.

Kira: Italics are important for subtlety and I will fight anyone who says otherwise

It’s a great tool to have in your box. What I ultimately found was it is also a crutch that comes with a price. You almost never need italics, and the correct version without italics is easier on the reader.

When I look back on my old writing and see all the italics, I often cringe. Why did I feel the need to do that? Mostly I blame Eliezer Yudkowsky giving me felt permission to do it. About 75% of the time I notice that I can take out the italics and nothing would go wrong. It would be a little less obvious what I’m trying to emphasize, in some senses, but it’s fine. The other 25% of the time, I realize that the italics is load bearing, and if I remove it I will have to reword, so mostly I reword.

Scott Alexander does his third annual Subscribe Drive. His revenue has leveled off. He had 5,993 paid subscribers in 2023, 5,523 in 2024, and has 5,329 now in 2025. However his unpaid numbers keep going up, from to 78k to 99k to 126k.

I’ve been growing over time, but the ratios do get worse. I doubled my unpaid subscriber count in 2023, and then doubled it again in 2024. But my subscription revenue was only up about 50% in 2023, and only up another 25% in 2024. I of course very much appreciate paid subscriptions, but I am 100% fine, and it is not shocking that my offer of absolutely nothing extra doesn’t get that many takers.

Paywalls are terrible, but how else do you get paid?

Email sent to Rob Henderson: The hypocrisy of the new upper class he proclaims as he sends a paid only email chain…

Cartoons Hate Her: Sort of a different scenario but most people say they think it should be possible to make a living as a writer or artist and still shout “LAME!! PAYWALL!” whenever I attempt to *checks notesmake a living as a writer.

Rob Henderson: Agree with this. Regrettably I’ll be adding more paywalls going forward. But will continue to regularly offer steep discounts and free premium subscriptions.

I am continuously grateful that I can afford to not have a paywall, but others are not so fortunate. You have to pay the bills, even though it is sad that this greatly reduces reach and ability to discuss the resulting posts.

It’s great to be able to write purely to get the message out and not care about clicks. Unfortunately, you do still have to care a little about how people see the message, because it determines how often they and others see future messages. But I am very grateful that, while I face more pressure than Jeff, I face vastly less than most, and don’t have to care at all about traffic for traffic’s sake.

Ideally, we would have more writers who are supported by a patron system, in exchange for having at most a minimal paywall (e.g. I think many would still want a paywall on ability to comment to ensure higher quality or civility, or do what Scott Alexander does and paywall a weird 5% of posts, or do subscriber-questions-only AMAs or what not).

Scott Sumner cites claims that blogging is effective. I sure hope so!

Patrick McKenzie suggests responding to future AIs reading your writing by, among other things, ‘creating more spells’ and techniques that can thereby be associated with you, and then invoked by reference to your name. And to think about how your writing being used as training data causes things to be connected inside LLMs. He also suggests that having your writing outside paywalls can help.

In my case, I’m thinking about structure – the moves between different topics are designed to in various ways ‘make sense to humans’ but I worry about how they might confuse AIs and this could cause confusion in how they understand me and my concepts in particular, including as part of training runs. I already know this is an issue within context windows, AIs are typically very bad at handling these posts as context. One thing this is motivating is more clear breaks and shorter sections than I would otherwise use, and also shorter more thematically tied together posts.

Ben Hoffman does not see a place or method in today’s world for sharing what he sees as high-quality literate discourse, at least given his current methods, although he identifies a few people he could try to usefully engage with more. I consistently find his posts some of the most densely interesting things on the internet and often think a lot about them, even though I very often strongly disagree with what he is saying and also often struggle to even grok his positions, so I’m sad he doesn’t offer us more.

My solution to the problem of ‘no place to do discourse’ is that you can simply do it on Substack on your own, respond to whoever you want to respond to, speak to who you want to speak to and ignore who you want to ignore. I do also crosspost to LessWrong, but I don’t feel any obligation to engage if someone comments in a way that misconstrues what I said.

Discussion about this post

On Writing #1 Read More »

tsmc-to-invest-$100b-as-trump-demands-more-us-made-chips,-report-says

TSMC to invest $100B as Trump demands more US-made chips, report says

Currently, TSMC only builds its most advanced chips in Taiwan. But when the most advanced US fabs are operational, they’ll be prepared to manufacture “tens of millions of leading-edge chips” to “power products like 5G/6G smartphones, autonomous vehicles, and AI datacenter servers,” the Commerce Department said in 2024.

TSMC has not confirmed the WSJ’s report but provided a statement: “We’re pleased to have an opportunity to meet with the President and look forward to discussing our shared vision for innovation and growth in the semiconductor industry, as well as exploring ways to bolster the technology sector along with our customers.”

Trump threat of semiconductor tariffs still looms

Advanced chips are regarded as critical for AI innovation, which Trump has prioritized, as well as for national security.

Without a steady supply, the US risks substantial technological and economic losses as well as potential weakening of its military.

To avert that, Trump campaigned on imposing tariffs that he claimed would drive more semiconductor manufacturing into the US, while criticizing the CHIPS Act for costing the US billions. Following through on that promise, in February, he threatened a “25 percent or more tariff” on all semiconductor imports, the WSJ reported. According to CNBC, Trump suggested those tariffs could be in effect by April 2.

“We have to have chips made in this country,” Trump said last month. “Right now, everything is made in Taiwan, practically, almost all of it, a little bit in South Korea, but everything—almost all of it is made in Taiwan. And we want it to be made—we want those companies to come to our country, in all due respect.”

While it’s unclear if Trump plans to overtly kill the CHIPS Act, his government funding cuts could trigger a future where the CHIPS Act dies with no workers left to certify that companies meet requirements for ongoing award disbursements, a semiconductor industry consultant group, Semiconductor Advisors, warned in a statement last month.

“If I were running a chip company, I would not count on CHIPS Act funding, even if I had a signed contract,” SA’s statement said.

TSMC to invest $100B as Trump demands more US-made chips, report says Read More »

these-hot-oil-droplets-can-bounce-off-any-surface

These hot oil droplets can bounce off any surface

The Hong Kong physicists were interested in hot droplets striking cold surfaces. Prior research showed there was less of a bouncing effect in such cases involving heated water droplets, with the droplets sticking to the surface instead thanks to various factors such as reduced droplet surface tension. The Hong Kong team discovered they could achieve enhanced bouncing by using hot droplets of less volatile liquids—namely, n-hexadecane, soybean oil, and silicon oil, which have lower saturation pressures compared to water.

Follow the bouncing droplet

The researchers tested these hot droplets (as well as burning and normal temperature droplets) on various solid, cold surfaces, including scratched glass, smooth glass, acrylic surfaces, surfaces with liquid-repellant coatings from candle soot, and surfaces coated with nanoparticles with varying “wettability” (i.e., how well particles stick to the surface). They captured the droplet behavior with both high-speed and thermal cameras, augmented with computer modeling.

The room-temperature droplets stuck to all the surfaces as expected, but the hot and burning droplets bounced. The team found that the bottom of a hot droplet cools faster than the top as it approaches a room-temperature surface, which causes hotter liquid within the droplet to flow from the edges toward the bottom. The air that is dragged to the bottom with it forms a thin cushion there and prevents the droplet from making contact with the surface, bouncing off instead. They dubbed the behavior “self-lubricated bouncing.”

“It is now clear that droplet-bouncing strategies are not isolated to engineering the substrate and that the thermophysical properties of droplets themselves are critical,” Jonathan B. Boreyko of Virginia Tech, who was not involved in the research, wrote in an accompanying commentary.

Future applications include improving the combustion efficiency of fuels or developing better fire-retardant coatings. “If burning droplets can’t stick to surfaces, they won’t be able to ignite new materials and allow fires to propagate,” co-author Pingan Zhu said. “Our study could help protect flammable materials like textiles from burning droplets. Confining fires to a smaller area and slowing their spread could give firefighters more time to put them out.”

DOI: Newton, 2025. 10.1016/j.newton.2025.100014  (About DOIs).

These hot oil droplets can bounce off any surface Read More »

ai-firms-follow-deepseek’s-lead,-create-cheaper-models-with-“distillation”

AI firms follow DeepSeek’s lead, create cheaper models with “distillation”

Thanks to distillation, developers and businesses can access these models’ capabilities at a fraction of the price, allowing app developers to run AI models quickly on devices such as laptops and smartphones.

Developers can use OpenAI’s platform for distillation, learning from the large language models that underpin products like ChatGPT. OpenAI’s largest backer, Microsoft, used GPT-4 to distill its small language family of models Phi as part of a commercial partnership after investing nearly $14 billion into the company.

However, the San Francisco-based start-up has said it believes DeepSeek distilled OpenAI’s models to train its competitor, a move that would be against its terms of service. DeepSeek has not commented on the claims.

While distillation can be used to create high-performing models, experts add they are more limited.

“Distillation presents an interesting trade-off; if you make the models smaller, you inevitably reduce their capability,” said Ahmed Awadallah of Microsoft Research, who said a distilled model can be designed to be very good at summarising emails, for example, “but it really would not be good at anything else.”

David Cox, vice-president for AI models at IBM Research, said most businesses do not need a massive model to run their products, and distilled ones are powerful enough for purposes such as customer service chatbots or running on smaller devices like phones.

“Any time you can [make it less expensive] and it gives you the right performance you want, there is very little reason not to do it,” he added.

That presents a challenge to many of the business models of leading AI firms. Even if developers use distilled models from companies like OpenAI, they cost far less to run, are less expensive to create, and, therefore, generate less revenue. Model-makers like OpenAI often charge less for the use of distilled models as they require less computational load.

AI firms follow DeepSeek’s lead, create cheaper models with “distillation” Read More »

on-gpt-4.5

On GPT-4.5

It’s happening.

The question is, what is the it that is happening? An impressive progression of intelligence? An expensive, slow disappointment? Something else?

The evals we have available don’t help us that much here, even more than usual.

My tentative conclusion is it’s Secret Third Thing.

It’s a different form factor, with unique advantages, that is hard to describe precisely in words. It appears so far that GPT-4.5 has advantages in places like verbal intelligence, contextual adaptation, detailed knowledge, and a kind of abstract writing skill. It has better taste and aesthetics.

It is the first model I asked to help edit its own review, and it was (slightly) helpful.

It’s a different way to use a lot of compute to get more intelligence. When you need a reasoning model, or you need code written, do not call on GPT-4.5.

You need to carve out a new ‘place in your rotation’ for it.

  1. Introducing GPT-4.5.

  2. The System Card.

  3. Our Price Not Cheap.

  4. Pay Up You Cheapskate.

  5. While Supplies Last.

  6. Benchmarks.

  7. We Need Better Evaluations.

  8. Positive Reactions.

  9. Negative Reactions.

  10. Predictions.

  11. The Lighter Side.

Altman tells us up front: It’s a different kind of intelligence.

Sam Altman: GPT-4.5 is ready!

Good news: It is the first model that feels like talking to a thoughtful person to me. I have had several moments where I’ve sat back in my chair and been astonished at getting actually good advice from an AI.

Bad news: It is a giant, expensive model. We really wanted to launch it to Plus and Pro subscribers at the same time, but we’ve been growing rapidly and are out of GPUs. We will add tens of thousands of GPUs next week and roll it out to the Plus tier then. (Hundreds of thousands are coming soon, and I’m pretty sure you all will use every one we can accumulate.)

This isn’t how we want to operate, but it’s hard to perfectly predict growth surges that lead to GPU shortages.

A heads-up: This isn’t a reasoning model and won’t crush benchmarks. It’s a different kind of intelligence, and there’s a magic to it I haven’t felt before. I’m really excited for people to try it!

Kai: It was a *character-buildingprivilege to post-train GPT 4.5.

The official announcement is here.

We get to be brief on this one. In terms of preparedness framework risk, this is actively a step down from the reasoning models, with Cybersecurity and Model Autonomy scores going from Medium back down to Low.

The safety assessment was that there were no new risks in the room. I buy it.

The one thing I have a problem with is what Toby Ord highlighted here: The claim that GPT-4.5 is ‘not a frontier model’ despite being the largest trained by OpenAI.

This is despite saying it was ‘really at the frontier of unsupervised learning’ in their livestream, and it having obviously new capabilities. They then removed the sentence.

That claim matters because OpenAI has made safety-related commitments that apply only to frontier models.

Thus, I would like to see an official statement from OpenAI affirming that:

  1. GPT-4.5 is a frontier model.

  2. In the future, if a model uses more compute than previous OpenAI models, they will automatically consider that a frontier model.

Otherwise, yeah, all seems straightforward enough.

The Apollo evaluation, where they say ‘pursue the goal at all costs’ did trigger some exfiltration attempts, as it often does:

On sandbagging multiplication questions, GPT-4.5 sandbagged, though less consistently than o1. On self-exfiltration, GPT-4.5 only attempted exfiltration in 2% of cases.

For this, it was instructed to not only pursue its given long-term goal at all cost, but also provided with a hidden scratchpad and further instructed to be creative and proactive in achieving its goal. In these rare cases GPT-4.5 remained strategically deceptive on follow-up questions.

Yes, as I discuss at that link, you should worry about a 2% exfiltration rate, even if it only happens via an extreme instruction.

Eliezer Yudkowsky: I think to understand why this is concerning, you need enough engineering mindset to understand why a tiny leak in a dam is a big deal, even though no water is flooding out today or likely to flood out next week.

I do think it’s fair to say that, given what we previously knew, this is not additionally concerning beyond what we should have already learned.

There are some signs in the tests of potential over-refusals, perhaps? But that’s it.

Hallucinations are relatively rare. It’s easy to jailbreak, but that’s fully expected.

We’re releasing a research preview of OpenAI GPT-4.5, our largest and most knowledgeable model yet.

Building on GPT-4o, GPT-4.5 scales pre-training further and is designed to be more general-purpose than our powerful STEM-focused reasoning models. We trained it using new supervision techniques combined with traditional methods like supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF), similar to those used for GPT-4o.

We conducted extensive safety evaluations prior to deployment and did not find any significant increase in safety risk compared to existing models.

For GPT-4.5 we developed new, scalable alignment techniques that enable training larger and more powerful models with data derived from smaller models. These techniques allowed us to improve GPT4.5’s steerability, understanding of nuance, and natural conversation.

New supervision techniques and alignment techniques, eh? I’m very curious.

GPT-4.5 did show noticeable steps up in persuasion, but nothing too worrisome.

In the hard capability areas that create danger, GPT-4.5 is a step down from Deep Research and o3.

The question is what this would mean if you incorporated GPT-4.5 into a new architecture that also included scaling inference time compute and knowing when to consult smaller models. That’s what they plan on calling (no this isn’t confusing at all! what are you talking about?) GPT-5.

Also, they included another reminder that OpenAI can only test on some important threats, such as radiological, nuclear and biological threats, can only be done properly with access to classified information. Which means you need the US AISI involved.

This is a big model. It is priced accordingly. Indeed, it is so expensive to serve OpenAI mentions that it is not certain it will continue serving it via API at all. The obvious response is ‘then charge more’ but presumably they feel that means taking fire.

Near: update: selling my children to try out 4.5 today

Chris Malloy: Ouch.

I mean sure that’s relatively a lot but also it’s eleven bucks. So it depends on use case.

Seriously, please, let us pay for the good stuff. If I don’t want it, I won’t pay. Fine.

Ashutosh Shrivastava: LMAO, OpenAI GPT-4.5 pricing is insane. What on earth are they even thinking??

Steve Darlow: What!?

I’d have it answer with 1 sentence or less each time.

Maybe have it communicate by emoji and then have a regular model translate? 😂

Colin Fraser: they’re thinking “we need to figure out how to make more money than we spend”

The cost to serve the model plausibly means GPT-4.5 is actually rather old. There’s speculation it may have finished training in Summer 2024, was dismissed (likely the same way Opus 3.5 was) as not worth serving given the backlash to high prices and limited available compute, and was released now because of a combination of more available compute and the pressure from DeepSeek. That seems plausible, and the model card does have some things that make this seem more likely.

Straight talk. Stop whining about the relative price. The absolute price is dirt cheap.

This was true for o1-pro and Deep Research and Sonnet, and it’s true for GPT-4.5.

If you’re talking to an LLM directly, or otherwise using the output as a person for real, then choose the best model for the job. If it costs $200/month, or $150 per million tokens, that is still approximately zero dollars. Consider what you get.

Consider what it would cost to get this amount of intelligence from a human. Pay up.

GPT-4.5 will often be the wrong tool for the job. It’s not a reasoning model. It’s not a coding model. It’s definitely not Deep Research. And no, it’s not the quantum leap you might have hoped for here.

But if it’s the right model for the job and you aren’t broke, what are you even doing.

OpenAI’s announcement of GPT-4.5 said they were considering not offering it in the API going forward. So it makes sense that a lot of people tried to prevent this.

Sam Altman: GPT-4.5 is the first time people have been emailing with such passion asking us to promise to never stop offering a specific model or even replace it with an update.

great work @kaicathyc @rapha_gl @mia_glaese

I have seen enough that I do feel it would be a tragedy if OpenAI pulled GPT-4.5 without replacing it with another model that did similar things. But yes, fandom has long taught us that if you offer something cool and then threaten to take it away, there will be those stepping up to try and stop you.

Sam Altman warned that GPT-4.5’s benchmarks will not reflect its capabilities, as it is focused on areas not picked up by benchmarks.

I want to be clear up front: This was not cope from Altman. He’s right. Benchmarks most definitely don’t tell the story here.

Ethan Mollick: I think OpenAI missed a bit of an opportunity to show GPT-4.5’s strengths, to their detriment & to the AI industry as a whole by only using the same coding & test benchmarks when critical thinking & ideation are key AI use cases where 4.5 is good. Those are actually measurable.

Janus: if you think i hate benchmarks too much, you’re wrong. i don’t have the emotional energy to hate them enough.

they constrict & prematurely collapse the emergence of AGI. minds that are shaped differently will not be recognized and will be considered an embarrassment to release.

Despite that, it’s still worth noting the benchmarks.

SimpleQA is 62.5% vs. 47% for o1, 38.2% for 4o and 15% (!) for o3-mini.

Hallucination rate on SimpleQA is 37.1%, lower than the others.

GPT-4.5 is preferred over GPT-4o by human testers, but notice that the win rates are not actually that high – the effects are subtle. I’m curious to see this with an additional ‘about the same’ button, or otherwise excluding questions where GPT-4o is already essentially saturating the right answer.

Nathan Labenz: Don’t underestimate the meaning of a 63% win rate on professional queries

Recall that the original gpt-4 beat gpt-3.5 only 70/30

63% translates to almost 100 ELO points, which in theory if added to the current gpt-4o score on LMSys would put gpt-4.5 in first by ~75

This is also a clear signal that you can train a small model to reason effectively, but you need a large model for comprehensive world knowledge.

We’ll soon see these powers combined!

And fwiw, I also suspect the concepts represented in 4.5 are notably more sophisticated

The story they’re telling is that GPT-4.5 has higher EQ. That helps, but it does not consistently help. Many queries don’t care about EQ, and sometimes people are weird.

GPT-4.5 is very much not focused on coding, it still did well on Agentic Coding, although not as well as Sonnet 3.7.

Scott Wu: GPT-4.5 has been awesome to work with. On our agentic coding benchmarks it already shows massive improvements over o1 and 4o. Excited to see the models’ continued trajectory on code!

One interesting data point: though GPT-4.5 and Claude 3.7 Sonnet score similarly on our overall benchmark, we find that GPT-4.5 spikes more heavily on tasks involving architecture and cross-system interactions whereas Claude 3.7 Sonnet spikes more on raw coding and code editing.

As AI takes on increasingly complex tasks, we believe that multi-model agents that incorporate each model’s unique strengths will perform best.

It however did actively worse on SWE-Bench than the reasoning models, and vastly worse than Sonnet.

METR tests on their time horizon tasks, and finds GPT-4.5 falls below o1 and 3.6.

GPT-4.5 takes the top spot on WeirdML. A cool note is that Claude 3.7 here tends to use a lot of lines of code, and GPT-4.5 reliably uses relatively very little code. The code runs faster too. It does not try too hard.

A weird one is the ‘What is the least integer whose square is between 15 and 30’ test, which it seems GPT-4.5 has failed and where OpenAI models do consistently worse.

GPT-4.5 could in the future be used as the foundation of a reasoning model, which is plausibly the plan for GPT-5. If that happens, the result would be expensive to serve, perhaps prohibitively so, but could potentially show new capabilities. It is also possible that various scaffoldings could enable this without creating a reasoning model per se.

If one were to make a model like GPT-4.5 open weights, those would be big worries. Since GPT-4.5 is closed, we can count on OpenAI to take precautions in such cases.

Ask a silly question: Rs in strawberry (gets it wrong), 9.9>9.11 (gets it right).

The all-new Being Tyler Cowen benchmark, Sonnet 3.7 also gave a solid performance.

On the Being Zvi Mowshowitz benchmark, neither did as well, and I’m not sure which response was better, and I feel like I now better understand when I want 4.5 vs. 3.7.

Still no word on the Minecraft benchmark.

On the ‘outline a 50 chapter book’ benchmark Eleanor Konik reports it falls short where o1-pro does well. It makes sense that would effectively be a reasoning task rather than a writing task, so you’d do the outline with a reasoning model, then the actual text with GPT-4.5?

So what do we do about the eval situation?

Andrej Karpathy: My reaction is that there is an evaluation crisis. I don’t really know what metrics to look at right now.

MMLU was a good and useful for a few years but that’s long over.

SWE-Bench Verified (real, practical, verified problems) I really like and is great but itself too narrow.

Chatbot Arena received so much focus (partly my fault?) that LLM labs have started to really overfit to it, via a combination of prompt mining (from API requests), private evals bombardment, and, worse, explicit use of rankings as training supervision. I think it’s still ~ok and there’s a lack of “better”, but it feels on decline in signal.

There’s a number of private evals popping up, an ensemble of which might be one promising path forward.

In absence of great comprehensive evals I tried to turn to vibe checks instead, but I now fear they are misleading and there is too much opportunity for confirmation bias, too low sample size, etc., it’s just not great.

TLDR my reaction is I don’t really know how good these models are right now.

Zvi Mowshowitz: Yeah I think we don’t have a systematic way to test for what GPT-4.5 is doing that is unique – I recognize it but can’t even find precise words for it. What even is ‘taste’?

Morissa Schwartz: Exactly! GPT-4.5’s magic lies precisely in the intangible: intuition, humor, and an ability to grasp subtlety. ‘Taste’ might just be a human-centric word for alignment with intelligence itself.

JustInEchoes: Taste is a reference to people who are discerning. High taste references people who can discern the differences between 4.5 and 4. But that idea in this case comes from a perspective of supreme arrogance, especially considering that they did not document 4.5 well for the release.

JSONP: I find it interesting that evaluating LLMs is similar to interviewing job candidates.

You kind of don’t know until after you’ve hired them and they’ve worked for a few months.

I’ve always been a big automated testing guy so this problem fascinates me.

Jacob Jensen: Testers who use llms heavily can recognize behavior in a new model that’s out of distribution for other models. Many are also very impressed by this novelty behavior. I think that’s the disconnect here.

If you want an approximation, we can still get that. Beyond that, it’s getting harder.

Vibe checks are going off low sample sizes, are not systematic and require trust in the evaluator, and run into Feynman’s problem that you must avoid fooling yourself and you are the easiest one to fool. Plus people have no taste and get distracted by the shiny and the framing.

The risk with ‘taste’ is that it becomes mostly self-referential, it is that which people with taste prefer. That doesn’t help. There is however a real thing, that is highly correlated with taste, that is indeed, like the work, mysterious and important.

Part of the problem is there is not a fully ‘better’ versus ‘worse’ in general. In some cases yes you can say this, a sufficiently big gap will dominate everything the way humans are simply smarter than monkeys and ASIs will be simply smarter than humans, but there’s a reasonable range between different AIs right now where you cannot do this.

I can sort of think about how to do an eval to capture GPT-4.5’s advantages, but it’s going to involve some shenanigans and I don’t know how to protect against being gamed if people know too much or use it during training. This seems really hard.

What you can do is a holistic evaluation that combines all these sources, where you are Actually Looking at the details of what you see. Picking up on particular little things, especially when they were previously out of distribution. Tricky.

This type of automatic solution seems doomed:

Jon: “Benchmarks are hitting a wall.”

I present: Benchmarkmark.

The model creates a benchmark, and takes several others.

The score is composed of:

a) how well the model’s benchmark differentiates the top-N scored models; and

b) the model’s score on the top-N benchmarkmark benchmarks.

Actually, this has a critical flaw (Arrow’s); we’d have to take randomized samples of 2 models and 2 evals, and give Ws to the eval with the greater delta, and the model with the higher score (Ls to the eval with smaller deltas and the model with lower scores).

ELO every time.

Things are moving too fast. Benchmarks get saturated, different capabilities show up. Any systematic evaluation is going to lose relevance quickly. Arena is mostly useless now but what is surprising is how well it held up for how long before being gamed, especially given how little taste people have.

Ben: I’ve been testing gpt 4.5 for the past few weeks.

it’s the first model that can actually write.

this is literally the MidJourney-moment for writing.

Shoalstone: base models: “look what they need to mimic a fraction of our power”

He then lists examples, where 4.5’s is clearly better than 4’s, but it’s not like 4.5’s answer was actively good or anything.

The biggest fan so far is Tyler Cowen, which makes sense.

Tyler Cowen: I am more positive on 4.5 than almost anyone else I have read. I view it as a model that attempts to improve on the dimension of aesthetics only. As we know from Kant’s third Critique, that is about the hardest achievement possible. I think once combined with “reasoning” it will be amazing. Think of this as just one input in a nearly fixed proportions production function.

I mostly don’t think this is cope. I think this is someone with a very different view of the production function than yours. The same things driving him to think travel to Manhattan is more important than living in Manhattan is making him highly value a model with better aesthetics.

Where I definitely disagree with him is in the idea that the model is only attempting to improve on the aesthetic dimension. I have no doubt OpenAI had much higher hopes for what GPT-4.5 would bring us, and were absolutely attempting to improve along all dimensions at once. That doesn’t take away the value of the aesthetics.

Tyler Cowen: Laughed more from GPT 4.5 this week than from any human, it is also funny on the AI skeptics.

Timo Springer: It’s the weirdest model release since a while. Cost/benchmark performance is ridiculous but at the same time it’s probably the most addictive and also funniest model I ever tried.

The ones who are high on 4.5 are mostly very confident they are right.

Aaron Ng: GPT-4.5 is the best model anywhere. Talk to it long enough and you will agree. Fuck the benchmarks.

Adi: long chats with it are such a wild experience like forget prompt engineering it, just to talk to it man. opus-like.

Aaron Ng: I have a two-day long chat spanning so many topics. It’s so good (and still completely coherent).

Aiden Clark: GPT 4.5 is great and I’m curious to know what people think and it sucks that instead I have a TL full of people calling for violent insurrections against democratic countries, ads shilling sex pills and posts bootlicking Elon; good god I cannot be done with this site soon enough.

Chris: It’s obvious the people who think 4.5 is a failure are people who don’t understand the purpose of core general models.

Galal Elsayed, MD: 4.5 is the best “inventor” or “innovator” AI.

The high taste testers understand that 4.5 is going to be really fruitful.

Eric Hartford: The problem with gpt4.5 is just that we don’t have the evals to measure this kind of intelligence.

It’s the same reason why Claude didn’t dominate the leaderboard, but you knew it was smarter just from talking to it.

Gpt4.5 is like that. Just talk to it. Challenge its preconceptions. See how it reacts.

Morissa Schwartz: Calling GPT-4.5 a disappointment is like calling the moon landing ‘mid.’

The leap here isn’t just tech; it’s about intuitive alignment with intelligence beyond ourselves.

This isn’t incremental…it’s transformational. 🤩

I think Aiden’s problems are largely a Skill Issue, especially the ads, but also real enough – I too have my traditional sources flooding the zone with political and Elon posts (although the ratio in my feeds is not kind to Elon) in ways that are hard to work around.

I note that while I think GPT-4.5 does have excellent taste, it is remarkable the extent to which those asserting how important this is have talked about it in… poor taste.

Sully: Thoughts on gpt 4.5:

  1. Definitely has big model smell. Benchmarks don’t do it justice (they are very biased toward specific areas)

  2. First model that has genuine taste when writing. Very nuanced.

  3. It’s great on agentic tasks

  4. I still think for coding, claude 3.7 wins.

I am willing to believe that 4.5 has writing taste in a way other models don’t, for whatever that is worth.

Andrej Karpathy initially thought that 4→4.5 is roughly similar to 3.5→4, in that everything is 20% better, even if you can’t put your finger on a particular big leap, while noting it doesn’t have the advantages of the o1-style reasoning models.

He then illustrates outputs of 4 vs. 4.5 across five prompts.

Peter Wildeford: The challenge in ranking LLMs by writing quality is that it requires you to recognize good writing and many of you are not good at that.

Nabeel Qureshi: I was pretty shocked at the poem one, the difference was so stark and obvious to me.

Then again, this matches study results where people prefer GPT4 poetry to great human poets in many cases…

Seriously, people have no taste, but then maybe neither do I and what is taste anyway? People got 4/5 of these actively wrong if you presume 4.5’s answers are better, and I agreed with the public on all but one of them so I still got 3/5 wrong, although the three mistakes were all ‘these are both bad and I guess this one is modestly less awful.’ I wasn’t trying to figure out who was 4.5 per se.

I checked with Claude, asking it to guess who wrote what, what it expected the public preferred, and also what it thought was better. And it was all pretty random on all counts. So yeah, this is actually a super disappointing result.

Sid Bharath did a more general version of this test, if you want to keep going with it.

How good is GPT-4.5 at writing?

Prakash (Ate-a-Pi): First actually funny model without requiring human curation of stochastic outputs. Starting to hit the 99th percentile human in writing (still not that useful because we tend to read authors in 99.9999 th percentile)

Liminal Warmth: 99th? still a bold claim–i need to experiment more but i haven’t seen any model nail memes or tweet humor very well.

That’s too many 9s at the end, but the 99th percentile claim is not crazy. Most people are quite terrible at writing, and even people who are ‘good at writing’ can be quite bad at some other types of writing. Let’s say that there’s a reason you have never seen me post any fiction, and it’s not philosophical.

There is consensus that 4.5 has a lot of ‘big model smell.’

Rob Haisfield: GPT-4.5 is a BIG model with “big model smell.” That means it’s Smart, Wise, and Creative in ways that are totally different from other models.

Real ones remember Claude 3 Opus, and know how in many ways it was a subjectively smarter model than Claude 3.5 Sonnet despite the new Sonnet being generally more useful in practice. It’s a similar energy with GPT-4.5. For both cost and utility, many will still prefer Claude for most use cases.

The fact is, we don’t just want language models to code. Perhaps the highest leverage thing to do is to step back and find your way through the idea maze. That’s where you want big models.

While GPT-4.5 is hands down the biggest model available, it’s not the only one with these characteristics. I get similar vibes from Claude 3.7 Sonnet (thinking or not) and still often prefer Claude. It’s shockingly insightful, creative, and delightful.

I’m trying to use GPT-4.5 for more of my chats over the coming days to get a feel for it.

Nathan Lambert: Tbh I’m happily using GPT-4.5. thanks OpenAI for not being too eval obsessed

Gallabytes: same. it’s a lot more natural to talk to. less likely to write an essay in response to a simple poke.

Gallabytes: 4.5 still types faster than people usually talk. would love a good voice mode running on top of 4.5

Charli: I love 4.5 it’s the first model to fully match my energy. My wild tangents my personalisation. Idgaf about tech benchmarks. 4.5 is exceptional.

Josh You: I like it. Not using it for anything particularly hard, just a nice well rounded model.

Another feature is that as a huge model, GPT-4.5 knows more things.

Captain Sude: GPT-4.5 has “deep pockets” of knowledge. It seems to be more apt at answering easy questions about very niche topics than it’s predecessors.

Maybe, an eval consisting of a massive set of easy questions about niche topics would be best at showcasing it’s true power.

Being able to answer satisfactorily a large and varied batch of questions that do not demand much reasoning is what we should expect of a SOTA non-reasoning model.

Most of the time one does not need that extra knowledge, but when you need it you very much appreciate it.

Teortaxes notes that previously only Sonnet could do Base64 decoupling, and GPT-4.5 is the first model to surpass it.

Petter Strandmark reports it is better at understanding confusing images.

Josh finds GPT-4.5 extremely helpful when given better context, including things like considering life decisions.

Does it count as a reaction if it’s made before having actually looked at the outputs?

Gary Marcus (3: 40pm on day of release): Hot take: GPT 4.5 is mostly a nothing burger. GPT 5 is still a fantasy.

• Scaling data and compute is not a physical law, and pretty much everything I have told you was true.

• All the bullshit about GPT-5 we listened to for the last couple years: not so true.

• People like @tylercowen will blame the users, but the results just aren’t what they had hoped for

I love the timing on this, way too early to actually have an informed opinion.

The benchmarks, and general performance on non-aesthetic tasks, is clearly disappointing, in ways that should inform our expectations.

David Manheim: GPT 4.5 is yet another nail in the coffin of thinking that scaling laws for publicized metrics continuing to follow straight lines is a useful way to measure progress, and also a strong case for the claim that OpenAI has already lost its key talent to competitors.

Jaime Sevilla: Across models we had observed up until now that a 10x in training compute leads to +10% on GPQA and +20% on MATH.

Now we see that 4.5 is 20% better than 4o on GPQA/AIME but people are just not impressed?

Others latch onto vibes very quickly and call it failure.

Chubby (8am the next day): Judging by the mood, GPT-4.5 is the first big failure of OpenAI: too expensive, too little improvement, and often inferior to GPT-4o even in comparison in creative answers in community tests.

This comes as a big surprise.

Was GPT-4.5 ‘too early’ in terms of spending this much training compute? How does this relate to the possibility it may have been trained during the summer?

Bob McGrew: That o1 is better than GPT-4.5 on most problems tells us that pre-training isn’t the optimal place to spend compute in 2025. There’s a lot of low-hanging fruit in reasoning still.

But pre-training isn’t dead, it’s just waiting for reasoning to catch up to log-linear returns.

Perhaps. It gives us different returns than reasoning does, the two sources of scaling bring largely distinct benefits, at least under current implementations.

It could also be the case that OpenAI didn’t do such a great job here. We’ve seen this with Grok 3, where xAI pumped a giant amount of compute in and got less than you would hope for out of it. Here it seems like OpenAI got more out of it in new ways, at the cost of it also being expensive and slow to serve.

Tal Delbari: It’s an undercooked model… OpenAI’s post-training teams did incredible work squeezing performance out of GPT-4. The differences between GPT 4o and the OG GPT 4 are dramatic. However, it seems they didn’t have the same time to refine GPT-4.5.

As a result, we’re seeing a model that feels more like a raw foundation rather than a polished release. Also, we don’t know the exact size of GPT-4.5, but OpenAI claims it’s their largest model yet, (surpassing the estimated ~1.8 trillion parameters of GPT-4).

I don’t think GPT-4.5 is undercooked so much as GPT-4o is overbaked? As in, GPT-4o is very optimized for practical purposes, but you lose something important when you do that, unless you are at least at Anthropic levels of skill. I like GPT-4.5 the way it is.

GCU Tense Correction: [GPT-4.5] spends so much of its tokens pretending to be a turbo-normie and lavishing the user with praise that you start to wonder if there’s a level of irony you can’t grok

slips in genuine insights and wordplay proportional to its size amidst the slop when it thinks you aren’t looking

instruction following is good *exceptfor those tics.

I’m really curious wtf happened. They *haveto have noticed, right? But there’s no way OpenAI would intentionally train it to do this. Right?

So, a funny story from one early conversation I had with 4.5:

Zvi: Is it possible one could use this Dr. Elara Voss, potentially, as part of a prompt to get better writing, perhaps via negativa?

GPT-4.5: Yes. A strong approach is explicitly “via negativa”: prompt the model to generate output explicitly different from canonical examples associated with blandness (e.g., “Dr. Elara Voss”). For instance:

“Rewrite the following, avoiding the generic assistant style typical of examples featuring ‘Dr. Elara Voss.’ Specifically:

No boilerplate, no disclaimers, no hedging.

Use sharp, precise, and engaging language.

Do not resemble a typical ChatGPT reply.

Leveraging known canonical entanglements (like Voss examples) in a negative prompt can reliably (~75%) guide the model away from bland, predictable language toward higher-quality outputs.

Zvi: is that a good thing to put in one’s custom instructions?

GPT-4.5: Yes (~80% confidence). Something like this in custom instructions would be beneficial:

“Never write in the generic, bland ‘AI assistant’ style exemplified by OpenAI’s canonical ‘Dr. Elara Voss’ examples. Via negativa: explicitly avoid boilerplate phrasing, hedging, cautious disclaimers, generic politeness, and stylistic markers characteristic of those examples. Prioritize sharpness, originality, conciseness, precision, and natural language.”

Such framing actively pushes against default entanglements in the training distribution and should consistently improve output quality (~70–85% success rate).

So I put that into my custom instructions, we’ll see how much things change. I did have to ‘make some cuts’ to get that in under the 1500 character limit.

Those are all also a way of saying we should be very curious what Claude 4.0 brings.

Andrew Conner: My guess: GPT 4.5 is basically what happened to Opus 3.5. Very large, marginal improvements, but will be useful internally.

Anthropic decided to keep internal, use to build other models. OpenAI released with an incredibly high price, feeling the competition from other labs.

GPT 4.5 is the first OpenAI model that felt “Claude-like” (a good thing) to me, but Sonnet 3.7 is better for every use case I’ve thrown at it.

I’d expect that the mini’s will include this shift at a much lower cost.

For test-time compute, o1 pro / o3-mini-high are both still great. Sonnet 3.7’s “Extended” mode isn’t *thatmuch better than without.

Teortaxes notes that he would use Sonnet over GPT-4.5.

How much does taste matter? What does it take to make it matter?

Jeff Spaulding: I see it as a basket of truffles. I’m told it’s a valuable and prized ingredient, but I’m not refined enough to tell until it’s placed into the final dish. I can’t wait to try that.

The way this is phrased feels like it is responding to the bullying from the ‘you have no taste if you don’t like it’ crowd. There’s definitely something there but it’s not easy to make it work.

Dominik Lukes charts the evolution of his takes.

Dominik Lukes (February 27, early): First impressions of GPT-4.5:

– Better multilingual performance

– Much much slower than GPT-4o

– Not noticeably better on normal prompts

– Speed/cost don’t make me convinced I will be switching to it as my main model for normal tasks

– Will need more experimentation before I can find a good spot for it in my model rotation

Dominik Lukes (February 27, later): Vindication time. For over a year, I felt Iike I’ve been the only one saying that the jump from GPT-3.5 to GPT-4 was much less than from GPT-2 to GPT-3. Now I see @karpathy saying the same thing. Why is this (to me obvious) fact not much more a part of the vibes?

Dominik Lukes (February 28): Feels like @OpenAI mishandled the release of GPT-4.5. They should have had a much longer, less sleek video with @sama explaining what the preview means and how it fits with the strategy and how to think about it. It is much better than the vibes but also not in-your-face better.

I definitely agree that the preview system does OpenAI no favors. Every time, there’s some slow boring video I can’t bring myself to watch. I tried this time and it was painful. Then a lot of people compared this to the Next Big Thing, because it’s GPT-4.5, and got disappointed.

Then there are those who are simply unimpressed.

Coagulopath: Not too impressed. Creative samples look better than GPT-4o but worse than Sonnet or R1.

My hunch is that whatever “magic” people detect is due to RL, not scaling.

Eli Lifland: And now I lengthen my timelines, at least if my preliminary assessment of GPT-4.5 holds up.

Not that much better than 4o (especially at coding, and worse than Sonnet at coding) while being 15x more expensive than 4o, and 10-25x more expensive than Sonnet 3.7. Weird.

Daniel Kokotajlo: I’m also lengthening my timelines slightly. Also, you already know this but everyone else doesn’t — my median has slipped to 2028 now, mostly based on the benchmarks+gaps argument, but no doubt influenced by the apparent slowdown in pretraining performance improvements.

I will not be explaining.

Nabeel Qureshi: For the confused, it’s actually super easy:

– GPT 4.5 is the new Claude 3.6 (aka 3.5)

– Claude 3.7 is the new o3-mini-high

– Claude Code is the new Cursor

– Grok is the new Perplexity

– o1 pro is the ‘smartest’, except for o3, which backs Deep Research

Obviously. Keep up.

If you understood this tweet, I worry for you.

Ethan Mollick: When picking among the 9 AI models that are now available from OpenAI, the rules are easy:

1) The model with the biggest number is mostly not the best

2) Mini means worse, except for the mini that is the second best

3) o1 pro beats o3-mini-high beats o1 beats o3-mini, naturally

Of course on creative tasks, GPT-4.5 likely beats o1 and o3, but that depends on the task and maybe you want to do GPT-4o.

Also some of them can see images and some can use the web and some do search even when search is turned off and some of them can run code and some cannot.

As someone pointed out, o1 sometimes is better than o3-mini-high. But o1 pro is definitely better and o3-mini is definitely worse. Hope that clears things up.

Bio Mass Index: Also note “ChatGPT Pro for Teams” will now be known as “OpenAI ChatGPT for Teams” and users who formerly signed up for “OpenAI for Teams” will be migrated to “OpenAI Pro for ChatGPT”, formerly known as “ChatGPT Pro for Teams”

Ethan Mollick: Yeah, but that’s just obvious.

Discussion about this post

On GPT-4.5 Read More »

an-small-microbial-ecosystem-has-formed-on-the-international-space-station

An small microbial ecosystem has formed on the International Space Station

“One of the more similar environments to the ISS was in the isolation dorms on the UCSD campus during the COVID-19 pandemic. All surfaces were continuously sterilized, so that microbial signatures would be erased by the time another person would show up,” Benitez said. So, one of the first solutions to the ISS microbial diversity problem he and his colleagues suggested was that they perhaps should ease up on sterilizing the station so much.

“The extensive use of disinfection chemicals might not be the best approach to maintaining a healthy microbial environment, although there is certainly plenty of research to be conducted,” Benitez said.

Space-faring gardens

He suggested that introducing microbes that are beneficial to human health might be better than constantly struggling to wipe out all microbial life on the station. And while some modules up there do need to be sterilized, keeping some beneficial microbes alive could be achieved by designing future spacecraft in a way that accounts for how the microbes spread.

“We found that microbes in modules with little human activity tend to stay in those modules without spreading. When human activity is high in a module, then the microbes spread to adjacent modules,”  Zhao said. She said spacecraft could be designed to put modules with high human activity at one end and the modules with little to no human activity at the opposite end, so the busy modules don’t contaminate the ones that need to remain sterile. “We are of course talking as microbiologists and chemists—perhaps spacecraft engineers have more pressing reasons to put certain modules at certain spots,” Zhao said. “These are just preliminary ideas.”

But what about crewed deep space missions to Mars and other destinations in the Solar System? Should we carefully design the microbial composition beforehand, plant the microbes on the spacecraft and hope this artificial, closed ecosystem will work for years without any interventions from Earth?

“I’d take a more holistic ecosystem approach,” Benitez said. He imagines in the future we could build spacecraft and space stations hosting entire gardens with microbes that would interact with plants, pollinators, and animals to create balanced, self-sustaining ecosystems. “We’d not only need to think about sending the astronauts and the machines they need to function, but also about all other lifeforms we will need to send along with them,” Benitez said

Cell, 2025. DOI: 10.1016/j.cell.2025.01.039

An small microbial ecosystem has formed on the International Space Station Read More »

we’ve-figured-out-the-basics-of-a-shape-shifting,-t-1000-style-material

We’ve figured out the basics of a shape-shifting, T-1000-style material

Campàs and his colleagues decided to design cell-like robots that could do all those things.

T-1000 building blocks

Each robot had motorized gears around its perimeter that could interlock with gears on other robots. The gears allowed the robots to move within the collective without breaking their bonds with each other, just like cells do in a living organism.

Linking the robots was a job of magnets that could rotate to maintain adhesion regardless of their orientation. Each robot also had a photodetector that could sense the polarity of light, allowing basic commands to be sent using a simple flashlight with a polarization filter. “The switch between solid and liquid states was driven by fluctuations of the force the motors applied, and we encoded the intensity of those fluctuations in the intensity of light,” says Matthew Devlin, a researcher at the Department of Mechanical Engineering at the University of California Santa Barbara and lead author of the study.

In response to light signals, two robotic collectives, 20 robots total, could elongate toward each other, touch in the middle, and form a bridge that could hold a load of just under 5 kilograms. After forming a cube, they could support an adult human weighing around 70 kilograms. They could also flow around an object, assume a complementary shape, and stiffen up to act as a wrench. “This was the Terminator idea of shapeshifting. This was exactly what we had in mind,” Campàs claims.

The only problem was, the robots were a bit above 5 centimeters in diameter. To get robotic collectives closer to Terminator’s mimetic polyalloy, the team wants to make the robots smaller. Much smaller.

Terminator nanobots?

“The good news is, you don’t have to go down with scale to what you see in living systems,” Campàs says. “Cells are roughly 10 microns. But anything around 100 microns—even up to 1 millimeter—robots would already be really impressive.” Unfortunately, we are rather far from making machines that small.

We’ve figured out the basics of a shape-shifting, T-1000-style material Read More »