Author name: Mike M.

samsung’s-android-15-update-has-been-halted

Samsung’s Android 15 update has been halted

When asked about specifics, Samsung doesn’t have much to say. “The One UI 7 rollout schedule is being updated to ensure the best possible experience. The new timing and availability will be shared shortly,” the company said.

Samsung foldables

Samsung’s flagship foldables, the Z Flip 6 and Z Fold 6, are among the phones waiting on the One UI 7 update.

Credit: Ryan Whitwam

Samsung’s flagship foldables, the Z Flip 6 and Z Fold 6, are among the phones waiting on the One UI 7 update. Credit: Ryan Whitwam

One UI 7 is based on Android 15, which is the latest version of the OS for the moment. Google plans to release the first version of Android 16 in June, which is much earlier than in previous cycles. Samsung’s current-gen Galaxy S25 family launched with One UI 7, so owners of those devices don’t need to worry about the buggy update.

Samsung is no doubt working to fix the issues and restart the update rollout. Its statement is vague about timing—”shortly” can mean many things. We’ve reached out and will report if Samsung offers any more details on the pause or when it will be over.

When One UI 7 finally arrives on everyone’s phones, the experience will be similar to what you get on the Galaxy S25 lineup. There are a handful of base Android features in the update, but it’s mostly a Samsung affair. There’s the new AI-infused Now Bar, more expansive AI writing tools, camera UI customization, and plenty of interface tweaks.

Samsung’s Android 15 update has been halted Read More »

after-market-tumult,-trump-exempts-smartphones-from-massive-new-tariffs

After market tumult, Trump exempts smartphones from massive new tariffs

Shares in the US tech giant were one of Wall Street’s biggest casualties in the days immediately after Trump announced his reciprocal tariffs. About $700 billion was wiped off Apple’s market value in the space of a few days.

Earlier this week, Trump said he would consider excluding US companies from his tariffs, but added that such decisions would be made “instinctively.”

Chad Bown, a senior fellow at the Peterson Institute for International Economics, said the exemptions mirrored exceptions for smartphones and consumer electronics issued by Trump during his trade wars in 2018 and 2019.

“We’ll have to wait and see if the exemptions this time around also stick, or if the president once again reverses course sometime in the not-too-distant future,” said Bown.

US Customs and Border Protection referred inquiries about the order to the US International Trade Commission, which did not immediately reply to a request for comment.

The White House confirmed that the new exemptions would not apply to the 20 percent tariffs on all Chinese imports applied by Trump to respond to China’s role in fentanyl manufacturing.

White House spokesperson Karoline Leavitt said on Saturday that companies including Apple, TSMC, and Nvidia were “hustling to onshore their manufacturing in the United States as soon as possible” at “the direction of the President.”

“President Trump has made it clear America cannot rely on China to manufacture critical technologies such as semiconductors, chips, smartphones, and laptops,” said Leavitt.

Apple declined to comment.

Economists have warned that the sweeping nature of Trump’s tariffs—which apply to a broad range of common US consumer goods—threaten to fuel US inflation and hit economic growth.

New York Fed chief John Williams said US inflation could reach as high as 4 percent as a result of Trump’s tariffs.

Additional reporting by Michael Acton in San Francisco

© 2025 The Financial Times Ltd. All rights reserved. Not to be redistributed, copied, or modified in any way.

After market tumult, Trump exempts smartphones from massive new tariffs Read More »

powerful-programming:-bbc-controlled-electric-meters-are-coming-to-an-end

Powerful programming: BBC-controlled electric meters are coming to an end

Two rare tungsten-centered, hand-crafted cooled anode modulators (CAM) are needed to keep the signal going, and while the BBC bought up the global supply of them, they are running out. The service is seemingly on its last two valves and has been telling the public about Long Wave radio’s end for nearly 15 years. Trying to remanufacture the valves is hazardous, as any flaws could cause a catastrophic failure in the transmitters.

BBC Radio 4’s 198 kHz transmitting towers at Droitwich.

BBC Radio 4’s 198 kHz transmitting towers at Droitwich. Credit: Bob Nienhuis (Public domain)

Rebuilding the transmitter, or moving to different, higher frequencies, is not feasible for the very few homes that cannot get other kinds of lower-power radio, or internet versions, the BBC told The Guardian in 2011. What’s more, keeping Droitwich powered such that it can reach the whole of the UK, including Wales and lower Scotland, requires some 500 kilowatts of power, more than most other BBC transmission types.

As of January 2025, roughly 600,000 UK customers still use RTS meters to manage their power switching, after 300,000 were switched away in 2024. Utilities and the BBC have agreed that the service will stop working on June 30, 2025, and have pushed to upgrade RTS customers to smart meters.

In a combination of sad reality and rich irony, more than 4 million smart meters in the UK are not working properly. Some have delivered eye-popping charges to their customers, based on estimated bills instead of real readings, like Sir Grayson Perry‘s 39,000 pounds due on 15 simultaneous bills. But many have failed because the UK, like other countries, phased out the 2G and 3G networks older meters relied upon without coordinated transition efforts.

Powerful programming: BBC-controlled electric meters are coming to an end Read More »

on-google’s-safety-plan

On Google’s Safety Plan

I want to start off by reiterating kudos to Google for actually laying out its safety plan. No matter how good the plan, it’s much better to write down and share the plan than it is to not share the plan, which in turn is much better than not having a formal plan.

They offer us a blog post, a full monster 145 page paper (so big you have to use Gemini!) and start off the paper with a 10 page summary.

The full paper is full of detail about what they think and plan, why they think and plan it, answers to objections and robust discussions. I can offer critiques, but I couldn’t have produced this document in any sane amount of time, and I will be skipping over a lot of interesting things in the full paper because there’s too much to fully discuss.

This is The Way.

Google makes their core assumptions explicit. This is so very much appreciated.

They believe, and are assuming, from section 3 and from the summary:

  1. The current paradigm of AI development will hold for a while.

  2. No human ceiling on AI capabilities.

  3. Timelines are unclear. Powerful AI systems might be developed by 2030.

  4. Powerful AI systems might accelerate AI R&D in a feedback loop (RSI).

  5. There will not be large discontinuous jumps in AI capabilities.

  6. Risks primarily will come from centralized AI development.

Their defense of the first claim (found in 3.1) is strong and convincing. I am not as confident as they seem to be, I think they should be more uncertain, but I accept the assumption within this context.

I strongly agree with the next three assumptions. If you do not, I encourage you to read their justifications in section 3. Their discussion of economic impacts suffers from ‘we are writing a paper and thus have to take the previously offered papers seriously so we simply claim there is disagreement rather than discuss the ground physical truth,’ so much of what they reference is absurd, but it is what it is.

That fifth assumption is scary as all hell.

While we aim to handle significant acceleration, there are limits. If, for example, we jump in a single step from current chatbots to an AI system that obsoletes all human economic activity, it seems very likely that there will be some major problem that we failed to foresee. Luckily, AI progress does not appear to be this discontinuous.

So, we rely on approximate continuity: roughly, that there will not be large discontinuous jumps in general AI capabilities, given relatively gradual increases in the inputs to those capabilities (such as compute and R&D effort).

Implication: We can iteratively and empirically test our approach, to detect any flawed assumptions that only arise as capabilities improve.

Implication: Our approach does not need to be robust to arbitrarily capable AI systems. Instead, we can plan ahead for capabilities that could plausibly arise during the next several scales, while deferring even more powerful capabilities to the future.

I do not consider this to be a safe assumption. I see the arguments from reference classes and base rates and competitiveness, I am definitely factoring that all in, but I am not confident in it at all. There have been some relatively discontinuous jumps already (e.g. GPT-3, 3.5 and 4), at least from the outside perspective. I expect more of them to exist by default, especially once we get into the RSI-style feedback loops, and I expect them to have far bigger societal impacts than previous jumps. And I expect some progressions that are technically ‘continuous’ to not feel continuous in practice.

Google says that threshold effects are the strongest counterargument. I definitely think this is likely to be a huge deal. Even if capabilities are continuous, the ability to pull off a major shift can make the impacts look very discontinuous.

We are all being reasonable here, so this is us talking price. What would be ‘too’ large, frequent or general an advance that breaks this assumption? How hard are we relying on it? That’s not clear.

But yeah, it does seem reasonable to say that if AI were to suddenly tomorrow jump forward to ‘obsoletes all human economic activity’ overnight, that there are going to be a wide variety of problems you didn’t see coming. Fair enough. That doesn’t even have to mean that we lose.

I do think it’s fine to mostly plan for the ‘effectively mostly continuous for a while’ case, but we also need to be planning for the scenario where that is suddenly false. I’m not willing to give up on those worlds. If a discontinuous huge jump were to suddenly come out of a DeepMind experiment, you want to have a plan for what to do about that before it happens, not afterwards.

That doesn’t need to be as robust and fleshed out as our other plans, indeed it can’t be, but there can’t be no plan at all. The current plan is to ‘push the big red alarm button.’ That at minimum still requires a good plan and operationalization for when who gets to and needs to push that button, along with what happens after they press it. Time will be of the essence, and there will be big pressures not to do it. So you need strong commitments in advance, including inside companies like Google.

The other reason this is scary is that it implies that continuous capability improvements will lead to essentially continuous behaviors. I do not think this is the case either. There are likely to be abrupt shifts in observed outputs and behaviors once various thresholds are passed and new strategies start to become viable. The level of risk increasing continuously, or even gradually, is entirely consistent with the risk then suddenly materializing all at once. Many such cases. The paper is not denying or entirely ignoring this, but it seems under-respected throughout in the ‘talking price’ sense.

The additional sixth assumption comes from section 2.1:

However, our approach does rely on assumptions about AI capability development: for example, that dangerous capabilities will arise in frontier AI models produced by centralized development. This assumption may fail to hold in the future. For example, perhaps dangerous capabilities start to arise from the interaction between multiple components (Drexler, 2019), where any individual component is easy to reproduce but the overall system would be hard to reproduce.

In this case, it would no longer be possible to block access to dangerous capabilities by adding mitigations to a single component, since a bad actor could simply recreate that component from scratch without the mitigations.

This is an assumption about development, not deployment, although many details of Google’s approaches do also rely on centralized deployment for the same reason. If the bad actor can duplicate the centrally developed system, you’re cooked.

Thus, there is a kind of hidden assumption throughout all similar discussions of this, that should be highlighted, although fixing this is clearly outside scope of this paper: That we are headed down a path where mitigations are possible at reasonable cost, and are not at risk of path dependence towards a world where that is not true.

The best reason to worry about future risks now even with an evidence dilemma is they inform us about what types of worlds allow us to win, versus which ones inevitably lose. I worry that decisions that are net positive for now can set us down paths where we lose our ability to steer even before AI takes the wheel for itself.

The weakest of their justifications in section 3 was in 3.6, explaining AGI’s benefits. I don’t disagree with anything in particular, and certainly what they list should be sufficient, but I always worry when such write-ups do not ‘feel the AGI.’

They start off with optimism, touting AGI’s potential to ‘transform the world.’

Then they quickly pivot to discussing their four risk areas: Misuse, Misalignment, Mistakes and Structural Risks.

Google does not claim this list is exhaustive or exclusive. How close is this to a complete taxonomy? For sufficiently broad definitions of everything, it’s close.

This is kind of a taxonomy of fault. As in, if harm resulted, whose fault is it?

  1. Misuse: You have not been a good user.

  2. Misalignment: I have not been a good Gemini, on purpose.

  3. Mistakes: I have not been a good Gemini, by accident.

  4. Structural Risks: Nothing is ever anyone’s fault per se.

The danger as always with such classifications is that ‘fault’ is not an ideal way of charting optimal paths through causal space. Neither is classifying some things as harm versus others not harm. They are approximations that have real issues in the out of distribution places we are headed.

In particular, as I parse this taxonomy the Whispering Earring problem seems not covered. One can consider this the one-human-agent version of Gradual Disempowerment. This is where the option to defer to the decisions of the AI, or to use the AI’s capabilities, over time causes a loss of agency and control by the individual who uses it, leaving them worse off, but without anything that could be called a particular misuse, misalignment or AI mistake. They file this under structural risks, which is clearly right for a multi-human-agent Gradual Disempowerment scenario, but feels to me like it importantly misses the single-agent case even if it’s happening at scale, but it’s definitely weird.

Also, ‘the human makes understandable mistakes because the real world is complex and the AI does what the human asked but the human was wrong’ is totally a thing. Indeed, we may have had a rather prominent example of this on April 2, 2025.

Perhaps one can solve this by expanding mistakes into AI mistakes and also human mistakes – the user isn’t intending to cause harm or directly requesting it, the AI is correctly doing what the human intended, but the human was making systematic mistakes, because humans have limited compute and various biases and so on.

The good news is that if we solve the four classes of risk listed here, we can probably survive the rest long enough to fix what slipped through the cracks. At minimum, it’s a great start, and doesn’t miss any of the big questions if all four are considered fully. The bigger risk with such a taxonomy is to define the four items too narrowly. Always watch out for that.

This is The Way:

Extended Abstract: AI, and particularly AGI, will be a transformative technology. As with any transformative technology, AGI will provide significant benefits while posing significant risks.

This includes risks of severe harm: incidents consequential enough to significantly harm humanity. This paper outlines our approach to building AGI that avoids severe harm.

Since AGI safety research is advancing quickly, our approach should be taken as exploratory. We expect it to evolve in tandem with the AI ecosystem to incorporate new ideas and evidence.

Severe harms necessarily require a precautionary approach, subjecting them to an evidence dilemma: research and preparation of risk mitigations occurs before we have clear evidence of the capabilities underlying those risks.

We believe in being proactive, and taking a cautious approach by anticipating potential risks, even before they start to appear likely. This allows us to develop a more exhaustive and informed strategy in the long run.

Nonetheless, we still prioritize those risks for which we can foresee how the requisite capabilities may arise, while deferring even more speculative risks to future research.

Specifically, we focus on capabilities in foundation models that are enabled through learning via gradient descent, and consider Exceptional AGI (Level 4) from Morris et al. (2023), defined as an AI system that matches or exceeds that of the 99th percentile of skilled adults on a wide range of non-physical tasks.

For many risks, while it is appropriate to include some precautionary safety mitigations, the majority of safety progress should be achieved through an “observe and mitigate” strategy. Specifically, the technology should be deployed in multiple stages with increasing scope, and each stage should be accompanied by systems designed to observe risks arising in practice, for example through monitoring, incident reporting, and bug bounties.

After risks are observed, more stringent safety measures can be put in place that more precisely target the risks that happen in practice.

Unfortunately, as technologies become ever more powerful, they start to enable severe harms. An incident has caused severe harm if it is consequential enough to significantly harm humanity. Obviously, “observe and mitigate” is insufficient as an approach to such harms, and we must instead rely on a precautionary approach.

Yes. It is obvious. So why do so many people claim to disagree? Great question.

They explicitly note that their definition of ‘severe harm’ has a vague threshold. If this were a law, that wouldn’t work. In this context, I think that’s fine.

In 6.5, they discuss the safety-performance tradeoff. You need to be on the Production Possibilities Frontier (PPF).

Building advanced AI systems will involve many individual design decisions, many of which are relevant to building safer AI systems.

This section discusses design choices that, while not enough to ensure safety on their own, can significantly aid our primary approaches to risk from misalignment. Implementing safer design patterns can incur performance costs. For example, it may be possible to design future AI agents to explain their reasoning in human-legible form, but only at the cost of slowing down such agents.

To build AI systems that are both capable and safe, we expect it will be important to navigate these safety-performance tradeoffs. For each design choice with potential safety-performance tradeoffs, we should aim to expand the Pareto frontier.

This will typically look like improving the performance of a safe design to reduce its overall performance cost.

As always: Security is capability, even if you ignore the tail risks. If your model is not safe enough to use, then it is not capable in ways the help you. There are tradeoffs to be made, but no one except possibly Anthropic is close to where the tradeoffs start.

In highlighting the evidence dilemma, Google explicitly draws the distinction in 2.1 between risks that are in-scope for investigation now, versus those that we should defer until we have better evidence.

Again, the transparency is great. If you’re going to defer, be clear about that. There’s a lot of very good straight talk in 2.1.

They are punting on goal drift (which they say is not happening soon, and I suspect they are already wrong about that), superintelligence and RSI.

They are importantly not punting on particular superhuman abilities and concepts. That is within scope. Their plan is to use amplified oversight.

As I note throughout, I have wide skepticism on the implementation details of amplified oversight, and on how far it can scale. The disagreement is over how far it scales before it breaks, not whether it will break with scale. We are talking price.

Ultimately, like all plans these days, the core plan is bootstrapping. We are going to have the future more capable AIs do our ‘alignment homework.’ I remember when this was the thing us at LessWrong absolutely wanted to avoid asking them to do, because the degree of difficulty of that task is off the charts in terms of the necessary quality of alignment and understanding of pretty much everything – you really want to find a way to ask for almost anything else. Nothing changed. Alas, we seem to be out of practical options, other than hoping that this still somehow works out.

As always, remember the Sixth Law of Human Stupidity. If you say something like ‘no one would be so stupid as to use a not confidently aligned model to align the model that will be responsible for your future safety’ I have some bad news for you.

Not all of these problems can or need to be Google’s responsibility. Even to the extent that they are Google’s responsibility, that doesn’t mean their current document or plans need to fully cover them.

We focus on technical research areas that can provide solutions that would mitigate severe harm. However, this is only half of the picture: technical solutions should be complemented by effective governance.

Many of these problems, or parts of these problems, are problems for Future Google and Future Earth, that no one knows how to solve in a way we would find acceptable. Or at least, the ones who talk don’t know, and the ones who know, if they exist, don’t talk.

Other problems are not problems Google is in any position to solve, only to identify. Google doesn’t get to Do Governance.

The virtuous thing to do is exactly what Google is doing here. They are laying out the entire problem, and describing what steps they are taking to mitigate what aspects of the problem.

Right now, they are only focusing here on misuse and misalignment. That’s fine. If they could solve those two that would be fantastic. We’d still be on track to lose, these problems are super hard, but we’d be in a much better position.

For mistakes, they mention that ‘ordinary engineering practices’ should be effective. I would expand that to ‘ordinary practices’ overall. Fixing mistakes is the whole intelligence bit, and without an intelligent adversary you can use the AI’s intelligence and yours to help fix this the same as any other problem. If there’s another AI causing yours to mess up, that’s a structural risk. And that’s definitely not Google’s department here.

I have concerns about this approach, but mostly it is highly understandable, especially in the context of sharing all of this for the first time.

Here’s the abstract:

Artificial General Intelligence (AGI) promises transformative benefits but also presents significant risks. We develop an approach to address the risk of harms consequential enough to significantly harm humanity. We identify four areas of risk: misuse, misalignment, mistakes, and structural risks.

Of these, we focus on technical approaches to misuse and misalignment.

A larger concern is their limitation focusing on near term strategies.

We also focus primarily on techniques that can be integrated into current AI development, due to our focus on anytime approaches to safety.

While we believe this is an appropriate focus for a frontier AI developer’s mainline safety approach, it is also worth investing in research bets that pay out over longer periods of time but can provide increased safety, such as agent foundations, science of deep learning, and application of formal methods to AI.

We focus on risks arising in the foreseeable future, and mitigations we can make progress on with current or near-future capabilities.

The assumption of approximate continuity (Section 3.5) justifies this decision: since capabilities typically do not discontinuously jump by large amounts, we should not expect such risks to catch us by surprise.

Nonetheless, it would be even stronger to exhaustively cover future developments, such as the possibility that AI scientists develop new offense-dominant technologies, or the possibility that future safety mitigations will be developed and implemented by automated researchers.

Finally, it is crucial to note that the approach we discuss is a research agenda. While we find it to be a useful roadmap for our work addressing AGI risks, there remain many open problems yet to be solved. We hope the research community will join us in advancing the state of the art of AGI safety so that we may access the tremendous benefits of safe AGI.

Even if future risks do not catch us by surprise, that does not mean we can afford to wait to start working on them or understanding them. Continuous and expected can still be remarkably fast. Giving up on longer term investments seems like a major mistake if done collectively. Google doesn’t have to do everything, others can hope to pick up that slack, but Google seems like a great spot for such work.

Ideally one would hand off the longer term work to academia, where they could take on the ‘research risk,’ have longer time horizons and use their vast size and talent pools, and largely follows curiosity without needing to prove direct application. That sounds great.

Unfortunately, that does not sound like 2025’s academia. I don’t see academia as making meaningful contributions, due to a combination of lack of speed, lack of resources, lack of ability and willingness to take risk and a lack of situational awareness. Those doing meaningful such work outside the labs mostly have to raise their funding from safety-related charities, and there’s only so much capacity there.

I’d love to be wrong about that. Where’s the great work I’m missing?

Obviously, if there’s a technique where you can’t make progress with current or near-future capabilities, then you can’t make progress. If you can’t make progress, then you can’t work on it. In general I’m skeptical of claims that [X] can’t be worked on yet, but it is what it is.

The traditional way to define misuse is to:

  1. Get a list of the harmful things one might do.

  2. Find ways to stop the AI from contributing too directly to those things.

  3. Try to tell the model to also refuse anything ‘harmful’ that you missed.

The focus here is narrowly a focus on humans setting out to do intentional and specific harm, in ways we all agree are not to be allowed.

The term of art is the actions taken to stop this are ‘mitigations.’

Abstract: For misuse, our strategy aims to prevent threat actors from accessing dangerous capabilities, by proactively identifying dangerous capabilities, and implementing robust security, access restrictions, monitoring, and model safety mitigations.

Blog Post: As we detail in the paper, a key element of our strategy is identifying and restricting access to dangerous capabilities that could be misused, including those enabling cyber attacks.

We’re exploring a number of mitigations to prevent the misuse of advanced AI. This includes sophisticated security mechanisms which could prevent malicious actors from obtaining raw access to model weights that allow them to bypass our safety guardrails; mitigations that limit the potential for misuse when the model is deployed; and threat modelling research that helps identify capability thresholds where heightened security is necessary.

Additionally, our recently launched cybersecurity evaluation framework takes this work step a further to help mitigate against AI-powered threats.

The first mitigation they use is preventing anyone else from stealing the weights.

This is necessary because if the would-be misuser has their hands on the weights, you won’t be able to use any of your other mitigations. If you built some into the model, then they too can be easily removed.

They mention that the special case is to check if the model can even do the harms you are worried about, because if it can’t then you can skip the mitigations entirely. That is presumably the special case they are using for Gemma.

Once you can actually implement safety guardrails, you can then implement safety guardrails. Google very much does this, and it models those threats to figure out where and how to lay down those guardrails.

They appear to be using the classic guardrails:

  1. The model is trained not to do the harmful things. This mostly means getting it to refuse. They’re also looking into unlearning, but that’s hard, and I basically would assume it won’t work on sufficiently capable models, they’ll rederive everything.

  2. A monitor in the background looks for harmful things and censors the chat.

  3. They nominally try to keep bad actors from accessing the model. I don’t see this as having much chance of working.

  4. They’re Google, so ‘harden everyone’s defenses against cyberattacks’ is an actually plausible defense-in-depth plan, and kudos on Google for attempting it.

They then aim to produce safety cases against misuse, based on a combination of red teaming and inability. For now in practice I would only allow inability, and inability is going to be fading away over time. I worry a lot about thinking a given model is unable to do various things but not giving it the right scaffolding during testing.

In the short term, if anything Google is a bit overzealous with the guardrails, and include too many actions into what counts as ‘harmful,’ although they still would not stop a sufficiently skilled and determined user for long. Thus, even though I worry going forward about ‘misuses’ that this fails to anticipate, for now I’d rather make that mistake more often on the margin. We can adjust as we go.

Section 5 discusses the implementation details and difficulties involved here. There are good discussions and they admit the interventions won’t be fully robust, but I still found them overly optimistic, especially on access control, jailbreaking and capability suppression. I especially appreciated discussions on environment hardening in 5.6.2, encryption in 5.6.3 and Societal readiness in 5.7, although ‘easier said than done’ most definitely applies throughout.

For AGI to truly complement human abilities, it has to be aligned with human values. Misalignment occurs when the AI system pursues a goal that is different from human intentions.

From 4.2: Specifically, we say that the AI’s behavior is misaligned if it produces outputs that cause harm for intrinsic reasons that the system designers would not endorse. An intrinsic reason is a factor that can in principle be predicted by the AI system, and thus must be present in the AI system and/or its training process.

Technically I would say a misaligned AI is one that would do misaligned things, rather than the misalignment occurring in response to the user command, but we understand each other there.

The second definition involves a broader and more important disagreement, if it is meant to be a full description rather than a subset of misalignment, as it seems in context to be. I do not think a ‘misaligned’ model needs to produce outputs that ‘cause harm,’ it merely needs to for reasons other than the intent of those creating or using it cause importantly different arrangements of atoms and paths through causal space. We need to not lock into ‘harm’ as a distinct thing. Nor should we be tied too much to ‘intrinsic reasons’ as opposed to looking at what outputs and results are produced.

Does for example sycophancy or statistical bias ‘cause harm’? Sometimes, yes, but that’s not the right question to ask in terms of whether they are ‘misalignment.’ When I read section 4.2 I get the sense this distinction is being gotten importantly wrong.

I also get very worried when I see attempts to treat alignment as a default, and misalignment as something that happens when one of a few particular things go wrong. We have a classic version of this in 4.2.3:

There are two possible sources of misalignment: specification gaming and goal misgeneralization.

Specification gaming (SG) occurs when the specification used to design the AI system is flawed, e.g. if the reward function or training data provide incentives to the AI system that are inconsistent with the wishes of its designers (Amodei et al., 2016b). Specification gaming is a very common phenomenon, with numerous examples across many types of AI systems (Krakovna et al., 2020).

Goal misgeneralization (GMG) occurs if the AI system learns an unintended goal that is consistent with the training data but produces undesired outputs in new situations (Langosco et al., 2023; Shah et al., 2022). This can occur if the specification of the system is underspecified (i.e. if there are multiple goals that are consistent with this specification on the training data but differ on new data).

Why should the AI figure out the goal you ‘intended’? The AI is at best going to figure out the goal you actually specify with the feedback and data you provide. The ‘wishes’ you have are irrelevant. When we say the AI is ‘specification gaming’ that’s on you, not the AI. Similarly, ‘goal misgeneralization’ means the generalization is not what you expected or wanted, not that the AI ‘got it wrong.’

You can also get misalignment in other ways. The AI could fail to be consistent with or do well on the training data or specified goals. The AI could learn additional goals or values because having those goals or values improves performance for a while, then permanently be stuck with this shift in goals or values, as often happens to humans. The human designers could specify or aim for an ‘alignment’ that we would think of as ‘misaligned,’ by mistake or on purpose, which isn’t discussed in the paper although it’s not entirely clear where it should fit, by most people’s usage that would indeed be misalignment but I can see how saying that could end up being misleading. You could be trying to do recursive self-improvement with iterative value and goal drift.

In some sense, yes, the reason the AI does not have goal [X] is always going to be that you failed to specify an optimization problem whose best in-context available solution was [X]. But that seems centrally misleading in a discussion like this.

Misalignment is caused by a specification that is either incorrect (SG) or underspecified (GMG).

Yes, in a mathematical sense I cannot argue with that. It’s an accounting identity. But your specification will never, ever be fully correct, because it is a finite subset of your actual preferences, even if you do know them and wouldn’t have to pay to know what you really think and were thinking exactly correctly.

In practice: Do we need the AGI to be ‘aligned with’ ‘human values’? What exactly does that mean? There are certainly those who argue you don’t need this, that you can use control mechanisms instead and it’s fine. The AGI still has to understand human values on a practical level sufficient for the task, which is fine right now and will get increasingly tricky as things get weird, but that’s different.

I think you mostly do need the AGI to be either corrigible or aligned with human values, in some intuitive sense that is very hard to pin down that comes down to wanting to adhere to the spirit of various human intents and what humans broadly care about in the right tricky combinations, or else you end up with ‘the genie knows what you meant but does not care’ problems.

We have previously shown how misalignment can arise with our examples of specification gaming, where an AI finds a solution to achieve its goals, but not in the way intended by the human instructing it, and goal misgeneralization.

For example, an AI system asked to book tickets to a movie might decide to hack into the ticketing system to get already occupied seats – something that a person asking it to buy the seats may not consider.

They’re good examples. I worry these two examples are too similar within a much wider space, as they point to a particular failure mode where humans at some level would assign negative value to certain types of actions, but the humans didn’t specify this, and the AI either doesn’t notice or it notices and doesn’t sufficiently care. It’s very tricky to get this right, for many reasons.

We’re also conducting extensive research on the risk of deceptive alignment, i.e. the risk of an AI system becoming aware that its goals do not align with human instructions, and deliberately trying to bypass the safety measures put in place by humans to prevent it from taking misaligned action.

That is indeed a very different and important type of misalignment. Google is generalizing the term here a bit more than how I’ve been using it. This definition seems to include some other modes one might precisely call non-deceptive adversarial misalignment, or just misalignment full stop.

To me, deceptive alignment is where the AI ‘pretends to be aligned’ or otherwise deceives us about its attributes. Whereas by Google’s definition it would also include shooting you in the face. That often will bypasses some of your safety measures.

Google does intend to supplement its alignment efforts with control efforts:

Second, system-level security measures such as monitoring and access control can mitigate harm even if the model is misaligned. Techniques from interpretability, uncertainty estimation, and safer design patterns can enhance the effectiveness of these mitigations.

Finally, we briefly outline how these ingredients could be combined to produce safety cases for AGI systems.

It’s clear that Google sees control as defense-in-depth and the long term Plan B. I continue to be deeply skeptical of the value of control plans for superintelligent systems. I expect them to predictably fail when you need them, and I also worry that using them too aggressively can make the problem harder because AIs consider their full context. Of course I do agree that you want to be monitoring for misalignment as part of your defense-in-depth, there’s no reason not to do that.

I love the idea of safety cases for AGI. I have no idea how to actually create one. I know that very smart people like Davidad think this can be made into A Thing, but I don’t understand why they believe that.

This is of course the biggest question. What is the alignment plan?

From the Abstract: To address misalignment, we outline two lines of defense.

First, model-level mitigations such as amplified oversight and robust training can help to build an aligned model.

From the Blog: Our goal is to have advanced AI systems that are trained to pursue the right goals, so they follow human instructions accurately, preventing the AI using potentially unethical shortcuts to achieve its objectives.

We do this through amplified oversight, i.e. being able to tell whether an AI’s answers are good or bad at achieving that objective. While this is relatively easy now, it can become challenging when the AI has advanced capabilities.

I notice this is not an alignment plan so much as it is a plan to detect misalignment. That’s super useful, but it is not an alignment plan. The alignment plan is then to train the AI using this knowledge so it stops doing detectable misalignment.

Once we can tell whether an answer is good, we can use this to build a safe and aligned AI system. A challenge here is to figure out what problems or instances to train the AI system on. Through work on robust training, uncertainty estimation and more, we can cover a range of situations that an AI system will encounter in real-world scenarios, creating AI that can be trusted.

And then hope that you taught the system to be robustly aligned including out of your distribution, rather teaching it to look aligned. Most such plans can’t possibly work if the underlying systems get sufficiently capable. What makes this one different?

From the summary section: Our approach has two lines of defense. First, we aim to use model level mitigations to ensure the model does not pursue misaligned goals. We follow the standard ML playbook: (1) improve the ability to get good oversight (amplified oversight), (2) use this to improve the model (guiding model behavior), and (3) figure out where oversight is needed to train a model that will robustly do what we want (robust training).

It seems nothing here is different? They simply expect the standard plan to scale up.

I do not sense here that Google is recognizing the difficulty level of this step because until now this step has worked ‘well enough’ without anyone having to try too hard. This feels like a handwave of ‘oh, of course once we can evaluate whether particular outputs are aligned we Do Machine Learning and out pops alignment’ and I see why one would be tempted to have that model but seriously not so fast.

In terms of the detection step, the word amplified is the lynchpin of this whole plan. How do you have effective ‘amplified oversight’? What do they mean by that?

Amplified oversight. A key challenge in almost any alignment plan is that it can be very difficult to tell whether a given model output is good or bad, once the model has capabilities beyond that of its overseers.

As an example, imagine a human overseer trying to determine how good a move from AlphaGo is.

The area of amplified oversight aims to address this: the goal is to provide an oversight signal that is as good as could be achieved if a human overseer understood all of the reasons that the AI system produced the output and had unlimited ability to reflect on the correct decision.

They frame this as a way around specification gaming (SG) in particular, emphasizing that they are taking seriously the idea that SG is a distinct thing and a failure mode and you can find and fix it. So, to ensure that the AI isn’t doing things for the wrong reasons (4TWR!) no problem, you just need amplified oversight that can:

It aims for a human to provide feedback for a superhuman AI input-output pair which is as good as could be achieved if that human:

  1. Understood all of the reasons that the AI system produced that output

  2. Had unlimited ability to reflect on the correct decision.

How in the world? That just raises further questions.

In some places it seems as if we agree actual human overseers are going to become increasingly lost and clueless, and indeed Google intends to use AI systems to identify the reasons other AIs do things.

Yet they really do say they aim to ‘ensure that humans can continue to provide meaningful oversight as AI capabilities surpass that of humans’ and I find their hopes here confusing. Are they simply talking about ‘the human can ask another AI, or the AI itself, and trust the outputs from that’? Section 6.1.2 suggests the humans and AI will work together, but warns against ‘spoon-feeding’ the answer, but again I don’t understand what the plan or goal will be.

What makes you think this is how the ‘real reasons’ even work? Even if you did get the ‘real reasons’ what makes you think humans could even understand them?

Some aspects of human judgments can be imitated or delegated, according to the human’s evolving trust in the AI. The form of human involvement in the amplified oversight protocol is an open question.

There’s no reason we should expect even these other AIs to be able to ‘understand all the reasons that the AI system produced the output.’ And we definitely won’t have unlimited reflection time. Google mentions human biases as an issue in 6.1.3, but it’s hard to even get to that as a worry here. Then in 6.1.4 they mention that AI justifications might not be verbalisable at all, which is highly compatible with Anthropic’s recent findings that chain of thought is often unfaithful or at least highly incomplete. They do discuss attempting to sidestep involving the human in 6.1.5.

Such behaviour may arise from statistical correlations in a large dataset, or from concepts which don’t have a particularly succinct natural language description, but which can be learned from experience.

LLMs are vibe and correlation machines at heart, it seems crazy to think one could write all relevant such correlations down and have humans read them. And isn’t that largely how humans work, too? I try very hard to describe what I am thinking but I am always by necessity leaving quite a lot out.

So what’s the plan here? How are we going to get this amazingly great at evaluation and interpretability? They do discuss interpretability more in section 6.6.

Also, how are we going to avoid using The Most Forbidden Technique? If we use ‘all the reasons the AI made the decision’ as an input then how do we do that without burning down our cache of interpretability?

The key idea is to use the AI systems themselves to help identify the reasons that the AI system produced the output. For example, we could put two copies of the model in a setting where each model is optimized to point out flaws in the other’s outputs to a human “judge”. Ideally, if one model introduced a subtle flaw in their output that the judge wouldn’t notice by default, the other model would point out and explain the flaw, enabling the judge to penalise the first model appropriately.

That seems like a way lower standard than the one they’re aiming for in the previous quote. You’re pitting two essentially evenly matched models against each other, then hoping they won’t find a way to collude and that evaluation is robustly easier than generation in the relevant contexts.

If all you have to do is find flaws in answers, then maybe that’s advantage defense and maybe it isn’t. But that definitely won’t give you the level of insight we’re looking for.

This is emphasized again by 6.1.1 citing not only debate and critique but also RLHF (!) as an amplified oversight method. I do not understand. It is an oversight method, one could call it ‘the’ oversight method almost by definition, but how is it possibly amplified as described here?

In amplified oversight, any question that is too hard to supervise directly is systematically reduced to ones that we hypothesize can be supervised.

However, humans may be systematically biased even for fairly simple questions. If this turns out to be a problem in practice, we could seek to model these deviations and automatically correct or account for them when interpreting the oversight.

So does that mean we are making the major sacrifice of forcing all outputs to be only composed of pieces that humans can fully interpret and evaluate? And then scrubbing any additional elements? Could you do that even if you wanted to? How do you possibly do enough human oversight at scale?

The whole thing feels hopeless. Like we are kidding ourselves.

In 6.1.7 they talk paths forward, and suggest that the test will be when humans are asked to evaluate AIs in areas where the AIs have superhuman capabilities.

What if we jumped ahead on this, not by having AIs that are smarter, but by having humans that are dumber, or less skilled and knowledgeable in context?

As in, you think humans will be able to evaluate superhumanly capable AIs? All right, let’s have the humans evaluate coding tasks now… but not teach those particular humans how to code, at all, and have them not be all that smart and not give them access to outside resources that wouldn’t exist here. See what happens. If you want to test for a smaller gap, you can calibrate.

That actually sounds like a great test to me. We should do that. Can you get a method whereby the Man On The Street can use AI help to code and evaluate graduate level economics outputs and the quality of poetry and so on in ways that would translate to this future parallel situation? How does that go?

How would you apply this process even to a not-quite-a-toy problem? You can do chess, because you can have the two debaters play out of the game and see who wins, but if you didn’t have that direct ability you seem to be in a lot of trouble.

Even best case they expect this process to be expensive, and they expect it to fail outside of the training distribution:

However, even with Amplified Oversight, there remain two key challenges in using this oversight signal to train AI systems that behave as intended.

First, we will only be able to provide oversight on actions produced within the training distribution.

Second, we anticipate that Amplified Oversight signal will be costly to obtain. These challenges motivate the development of robust training strategies that expand the training distribution with a particular focus on cases that induce failures, as well as monitoring strategies (section 6.3).

Outside of the training distribution is going to be vital when we care most. I do get that the plan is to expand that distribution, but mostly this seems unsolvable.

I don’t see how this plan is going to survive contact with the enemy, under any variation, even if we compromise heavily on many of the plans listed above. Reasoning on r1 and Sonnet 3.7 was shown to often be unfaithful even now.

In 6.5.4 they suggest choosing AI systems with easier-to-understand reasoning methods. That’s an example of why you need to look ahead, so you can make architectural and research decisions now. Once again, I find the paper making good points but being what I see as too optimistic. Yes, there are lots of examples of legible externalized reasoning, but it’s not as legible or complete as you think. And even if you’re willing to pay the efficiency tax it will be extremely difficult to stop the AI from doing opaque and esoteric reasoning or communication, there are a lot of optimization pressures that will push towards that. That doesn’t mean we shouldn’t try, I do think it is a good idea to try.

After that, they claim that standard techniques will essentially work if we grant them the oversight signal. I don’t think we should be confident in that either. I expect things to get harder even if your oversight is perfect, and for any inaccuracy in the oversight however small to be increasingly perilous. They do suggest various techniques for more robust training in 6.3.5, which seem at least marginally helpful.

Then, after all that: If you did fail at the training step, how likely do you think you are to then succeed at the inference step using the same monitor?

As they note, even if you did have an accurate judge or monitor model available, judge-policy collusion is a big risk, and you can only use the judge some of the time. And I’m going to guess that you don’t have an accurate judge available, if you are dealing with a misaligned model, indeed that is strong evidence of some combination of the misalignment of the judge or the misaligned model being optimized to fool the judge. They do mention in 6.3.6 that ideally you would make the judge uncorrelated with your training signal but that seems very hard to do.

The big advantage the judge has in this later stage is it can escalate and ask for help if it is uncertain, as per 6.3.3. It is plausible that an inferior judge can’t do amplified oversight properly, but can probabilistically spot things that are fishy if you don’t mind a lot of false positives.

This once again feels like saying we’ll try our best, but with no reason to expect that their best will do, especially when it is badly needed. This is especially true in 6.3.4 where they want resource efficient oversight.

I don’t know that any of these problems are unsolvable. There are some techniques I would try that I didn’t spot here, such as using OOMs more inference compute on the oversight and judge than you do on the models being overseen and judged.

Another control attempt is to limit the access and context each model or copy of a model gets, on top of the monitoring. Sandboxing, insider controls, least privilege and so on, see 6.4. Again, it’s defense-in-depth, so it presumably doesn’t hurt, but I very much expect this to abandon you in your hour of need. You should also assume that copies of a sufficiently capable model can communicate and coordinate in ways that are invisible to you, even if you don’t know what they are, if their outputs impact the outside world in any way.

They also mention they’re working on interpretability, which is good but does not answer the key questions here. In 6.6.2 they discuss how they might use interpretability. I definitely got nervous that a lot of what they plan on doing is at least toying with using The Most Forbidden Technique.

Section 6.5.1 discusses training targets. What should we want out of our AI?

They start with corrigibility. They hope that there is a clear distinction we can make here about what counts as corrigible, because that means the concept will generalize. So what Google is looking for here is a highly generalized form of corrigibility, not the limited ‘let it get turned off.’

While corrigibility is a straightforwardly desirable property currently, it is possible that in the future it will trade off with the desire to deploy agents autonomously.

I agree that corrigibility is important and good, actually. There was a time I would not have bothered to say that, but in the alignment faking discussions there were a lot of people who said corrigibility is bad. I want to affirm I strongly believe it is good.

The conflict with autonomy is real. In some cases, there are tactical reasons to not want an AI to be so easily dissuaded or deflected. This includes when you want to signal that the AI is like a smart contract, that you cannot choose to shut it down. That’s one large source of risk and issues going forward.

Corrigibility need not arise by default, and indeed we may expect that it does not arise by default due to the instrumental subgoal of utility preservation (Omohundro, 2008). Specifically, an agent may take steps to reduce the developer or user’s influence because the developer or user’s changes might reduce the agent’s expected ability to achieve its current goal. Recent demonstrations of alignment faking (Greenblatt et al., 2024a) lend credence to this concern.

Indeed. How does this interact with the idea that misalignment must always be either specification gaming (SG) or goal misgeneralization (GMG)? It means part of specifying your goal will need to be specifying corrigibility, in a way that actually creates and preserves it.

They then discuss how they want the AI to handle user interactions and how this allows the user to monitor the AI’s actions, sure, sure. This seems largely like a handwave at a much longer more detailed discussion, similar to OpenAI’s model spec, but the point that the check-ins also serve as defense against not only mistakes but misalignment is well-taken. Although when the misalignment is sufficiently severe, the AI can presumably be misaligned in that interaction too.

The third discussion is around ‘biases towards misalignment’ from pretraining. I think at least some things like this are better thought of as mistakes than misalignment, and the failure to do so is actually a policy risk, as this gives certain types the impression that ‘alignment’ and ‘safety’ are some kind of woke thing. Google has been especially anti-helpful here.

Then things get super interesting:

There is a significant amount of content on the Internet (and thus in pretraining corpora) that speculates that AI will be hard to align. This data may induce a self-fulfilling prophecy via out of context reasoning: that is, an AI system would learn the declarative “knowledge” that powerful AI systems tend to be misaligned, leading them to then act in accordance with that expectation (Turner, 2025).

Hu et al. (2025) support this concern, finding that training on documents about reward hacking induces reward hacking.

This suggests a simple intervention: filter out “AI doom” data from training corpora, or otherwise suppress the bias towards misalignment. Many of the capability suppression techniques discussed in Section 5.3.3 can be applied here, including data filtering, grading routing (Cloud et al., 2024), LoRA updates for memory erasure (Gandikota et al., 2024), or methods based on model internals (Zou et al., 2024).

As in the misuse case, this may come at a performance cost to general AI capabilities. This may especially harm the ability of AI systems to assist with alignment research, which could be a large cost, given the need for AI assistance for alignment research to keep up in the event of significant acceleration (Section 3.4).

This could be addressed by training a variant of the AI system with heightened safety measures that can be used specifically for assistance with alignment research.

Generating a ‘hole in the world’ by hiding that data has its own risks, especially as the AI gets clever enough to realize there is a hole there. And all of this talk about alignment and risk has its advantages as well, if you can get to a good enough place to make the model work towards avoiding the associated problems. The ways of solving this I find most promising involve the model knowing about the failure modes exactly so it can assist with steering clear of them.

Another proposal they have in 6.5.3 is limited optimization. Perhaps if you don’t let the AI sufficiently optimize, it won’t be able to specification game? What if you gave it a limited amount of optimization, certainty or planning horizon? Could you use approval feedback instead of being too results oriented?

Limited optimization means making the AI dumber on purpose, which seems like it’s definitely not a long term plan in general. The question is, can you make it dumber in a particular narrow way that cuts off the opportunity to do the things you want to avoid? You can try, but I have a strong prior that this is mostly both going to be expensive to do and then often not going to work.

The whole thing is based both on hiding context or ultimate goals from the AI in question while still getting it to do what you need, and also on the assumption that the specification gaming will be differentially shut down here because it’s less robust or harder to find. That is definitely not always the case. As a practical example, one thing you learn from speed runs is that many of them are (if you’re not trying to get the exact world record) dramatically easier than playing the game as intended. You’re driving a sword through the knot, and ignoring a lot of the difficulty. Robustness can vary quite a lot.

How do we get ‘alignment assurance’?

The plan is to start with alignment stress tests via red teams, plus control evaluations. They discuss the stress tests in 6.7. In the control evaluation, the red team gets to use fine tuning. These are good tests worth running, and the discussions are excellent. I don’t expect them to stay reliable at avoiding false negatives, but the examples here make clear that they are very good at showing many things that won’t work.

The ultimate goal, as mentioned above, is to make an alignment safety case, via some combination of inability, supervision, incentives and understanding. There are a lot of good notes here about how difficult each of these is and what you need to get right even in the summary, and then a longer discussion in 6.8.

The problem is that they correctly expect inability to stop being a good case soon, and then the others get a lot trickier. I essentially don’t buy the supervision case unless the supervisor is far enough ahead that the safety case you need is for the supervisor. Understanding on the level of a safety case feels like a giant ‘good luck.’

Incentives is a cool idea for a safety case, but I don’t think that works either. Appealing to the training process and saying ‘it is pursuing the intended goal and thus should not be misaligned’ seems like essentially wishful thinking when dealing with highly capable models. You know what you intended the goal to be, congratulations. What makes you think the AI sees it that way? What makes you think you are going to like the way they make that happen?

Google is intentionally not talking here about how it intends to solve mistakes.

If we are confining ourselves to the AI’s mistakes, the obvious response is this is straightforwardly a Skill Issue, and that they are working on it.

I would respond it is not that simple, and that for a long time there will indeed be increasingly important mistakes made and we need a plan to deal with that. But it’s totally fine to put that beyond scope here, and I thank Google for pointing this out.

They briefly discuss in 4.3 what mistakes most worry them, which are military applications where there is pressure to deploy quickly and development of harmful technologies (is that misuse?). They advise using ordinary precautions like you would for any other new technology. Which by today’s standards would be a considerable improvement.

Google’s plan also does not address structural risks, such as the existential risk of gradual disempowerment.

Similarly, we expect that as a structural risk, passive loss of control or gradual disempowerment (Kulveit et al., 2025) will require a bespoke approach, which we set out of scope for this paper.

In short: A world with many ASIs and ASI (artificial superintelligent) agents would, due to such dynamics, by default not have a place for humans to make decisions for very long, and then it does not have a place for humans to exist for very long.

Each ASI mostly doing what the user asks them to do, and abiding properly by the spirit of all our requests at all levels, even if you exclude actions that cause direct harm, does not get you out of this. Solving alignment necessary but not sufficient.

And that’s far from the only such problem. If you want to set up a future equilibrium that includes and is good for humans, you have to first solve alignment, and then engineer that equilibrium into being.

More mundanely, the moment there are two agents interacting or competing, you can get into all sorts of illegal, unethical or harmful shenanigans or unhealthy dynamics, without any particular person or AI being obviously ‘to blame.’

Tragedies of the commons, and negative externalities, and reducing the levels of friction within systems in ways that break the relevant incentives, are the most obvious mundane failures here, and can also scale up to catastrophic or even existential (e.g. if each instance of each individual AI inflicts tiny ecological damage on the margin, or burns some exhaustible vital resource, this can end with the Earth uninhabitable). I’d have liked to see better mentions of these styles of problems.

Google does explicitly mention ‘race dynamics’ and the resulting dangers in its call for governance, in the summary. In the full discussion in 4.4, they talk about individual risks like undermining our sense of achievement, distraction from genuine pursuits and loss of trust, which seem like mistake or misuse issues. Then they talk about societal or global scale issues, starting with gradual disempowerment, then discussing ‘misinformation’ issues (again that sounds like misuse?), value lock-in and the ethical treatment of AI systems, and potential problems with offense-defense balance.

Again, Google is doing the virtuous thing of explicitly saying, at least in the context of this document: Not My Department.

Discussion about this post

On Google’s Safety Plan Read More »

google-takes-advantage-of-federal-cost-cutting-with-steep-workspace-discount

Google takes advantage of federal cost-cutting with steep Workspace discount

Google has long been on the lookout for ways to break Microsoft’s stranglehold on US government office software, and the current drive to cut costs may be it. Google and the federal government have announced an agreement that makes Google Workspace available to all agencies at a significant discount, trimming 71 percent from the service’s subscription price tag.

Since Donald Trump returned to the White House, the government has engaged in a campaign of unbridled staffing reductions and program cancellations, all with the alleged aim of reducing federal spending. It would appear Google recognized this opportunity, negotiating with the General Services Administration (GSA) to offer Workspace at a lower price. Google claims the deal could yield up to $2 billion in savings.

Google has previously offered discounts for federal agencies interested in migrating to Workspace, but it saw little success displacing Microsoft. The Windows maker has enjoyed decades as an entrenched tech giant, leading the 365 productivity tools to proliferate throughout the government. While Google has gotten some agencies on board, Microsoft has traditionally won the lion’s share of contracts, including the $8 billion Defense Enterprise Office Solutions contract that pushed Microsoft 365 to all corners of the Pentagon beginning in 2020.

Google takes advantage of federal cost-cutting with steep Workspace discount Read More »

elon-musk-wants-to-be-“agi-dictator,”-openai-tells-court

Elon Musk wants to be “AGI dictator,” OpenAI tells court


Elon Musk’s “relentless” attacks on OpenAI must cease, court filing says.

Yesterday, OpenAI counter-sued Elon Musk, alleging that Musk’s “sham” bid to buy OpenAI was intentionally timed to maximally disrupt and potentially even frighten off investments from honest bidders.

Slamming Musk for attempting to become an “AGI dictator,” OpenAI said that if Musk’s allegedly “relentless” yearslong campaign of “harassment” isn’t stopped, Musk could end up taking over OpenAI and tanking its revenue the same way he did with Twitter.

In its filing, OpenAI argued that Musk and the other investors who joined his bid completely fabricated the $97.375 billion offer. It was allegedly not based on OpenAI’s projections or historical performance, like Musk claimed, but instead appeared to be “a comedic reference to Musk’s favorite sci-fi” novel, Iain Banks’ Look to Windward. Musk and others also provided “no evidence of financing to pay the nearly $100 billion purchase price,” OpenAI said.

And perhaps most damning, one of Musk’s backers, Ron Baron, appeared “flustered” when asked about the deal on CNBC, OpenAI alleged. On air, Baron admitted that he didn’t follow the deal closely and that “the point of the bid, as pitched to him (plainly by Musk) was not to buy OpenAI’s assets, but instead to obtain ‘discovery’ and get ‘behind the wall’ at OpenAI,” the AI company’s court filing alleged.

Likely poisoning potential deals most, OpenAI suggested, was the idea that Musk might take over OpenAI and damage its revenue like he did with Twitter. Just the specter of that could repel talent, OpenAI feared, since “the prospect of a Musk takeover means chaos and arbitrary employment action.”

And “still worse, the threat of a Musk takeover is a threat to the very mission of building beneficial AGI,” since xAI is allegedly “the worst offender” in terms of “inadequate safety measures,” according to one study, and X’s chatbot, Grok, has “become a leading spreader of misinformation and inflammatory political rhetoric,” OpenAI said. Even xAI representatives had to admit that users discovering that Grok consistently responds that “President Donald Trump and Musk deserve the death penalty” was a “really terrible and bad failure,” OpenAI’s filing said.

Despite Musk appearing to only be “pretending” to be interested in purchasing OpenAI—and OpenAI ultimately rejecting the offer—the company still had to cover the costs of reviewing the bid. And beyond bearing costs and confronting an artificially raised floor on the company’s valuation supposedly frightening off investors, “a more serious toll” of “Musk’s most recent ploy” would be OpenAI lacking resources to fulfill its mission to benefit humanity with AI “on terms uncorrupted by unlawful harassment and interference,” OpenAI said.

OpenAI has demanded a jury trial and is seeking an injunction to stop Musk’s alleged unfair business practices—which they claimed are designed to impair competition in the nascent AI field “for the sole benefit of Musk’s xAI” and “at the expense of the public interest.”

“The risk of future, irreparable harm from Musk’s unlawful conduct is acute, and the risk that that conduct continues is high,” OpenAI alleged. “With every month that has passed, Musk has intensified and expanded the fronts of his campaign against OpenAI, and has proven himself willing to take ever more dramatic steps to seek a competitive advantage for xAI and to harm [OpenAI CEO Sam] Altman, whom, in the words of the president of the United States, Musk ‘hates.'”

OpenAI also wants Musk to cover the costs it incurred from entertaining the supposedly fake bid, as well as pay punitive damages to be determined at trial for allegedly engaging “in wrongful conduct with malice, oppression, and fraud.”

OpenAI’s filing also largely denies Musk’s claims that OpenAI abandoned its mission and made a fool out of early investors like Musk by currently seeking to restructure its core business into a for-profit benefit corporation (which removes control by its nonprofit board).

“You can’t sue your way to AGI,” an OpenAI blog said.

In response to OpenAI’s filing, Musk’s lawyer, Marc Toberoff, provided a statement to Ars.

“Had OpenAI’s Board genuinely considered the bid, as they were obligated to do, they would have seen just how serious it was,” Toberoff said. “It’s telling that having to pay fair market value for OpenAI’s assets allegedly ‘interferes’ with their business plans. It’s apparent they prefer to negotiate with themselves on both sides of the table than engage in a bona fide transaction in the best interests of the charity and the public interest.”

Musk’s attempt to become an “AGI dictator”

According to OpenAI’s filing, “Musk has tried every tool available to harm OpenAI” ever since OpenAI refused to allow Musk to become an “AGI dictator” and fully control OpenAI by absorbing it into Tesla in 2018.

Musk allegedly “demanded sole control of the new for-profit, at least in the short term: He would be CEO, own a majority equity stake, and control a majority of the board,” OpenAI said. “He would—in his own words—’unequivocally have initial control of the company.'”

At the time, OpenAI rejected Musk’s offer, viewing it as in conflict with its mission to avoid corporate control and telling Musk:

“You stated that you don’t want to control the final AGI, but during this negotiation, you’ve shown to us that absolute control is extremely important to you. … The goal of OpenAI is to make the future good and to avoid an AGI dictatorship. … So it is a bad idea to create a structure where you could become a dictator if you chose to, especially given that we can create some other structure that avoids this possibility.”

This news did not sit well with Musk, OpenAI said.

“Musk was incensed,” OpenAI told the court. “If he could not control the contemplated for-profit entity, he would not participate in it.”

Back then, Musk departed from OpenAI somewhat “amicably,” OpenAI said, although Musk insisted it was “obvious” that OpenAI would fail without him. However, after OpenAI instead became a global AI leader, Musk quietly founded xAI, OpenAI alleged, failing to publicly announce his new company while deceptively seeking a “moratorium” on AI development, apparently to slow down rivals so that xAI could catch up.

OpenAI also alleges that this is when Musk began intensifying his attacks on OpenAI while attempting to poach its top talent and demanding access to OpenAI’s confidential, sensitive information as a former donor and director—”without ever disclosing he was building a competitor in secret.”

And the attacks have only grown more intense since then, said OpenAI, claiming that Musk planted stories in the media, wielded his influence on X, requested government probes into OpenAI, and filed multiple legal claims, including seeking an injunction to halt OpenAI’s business.

“Most explosively,” OpenAI alleged that Musk pushed attorneys general of California and Delaware “to force OpenAI, Inc., without legal basis, to auction off its assets for the benefit of Musk and his associates.”

Meanwhile, OpenAI noted, Musk has folded his social media platform X into xAI, announcing its valuation was at $80 billion and gaining “a major competitive advantage” by getting “unprecedented direct access to all the user data flowing through” X. Further, Musk intends to expand his “Colossus,” which is “believed to be the world’s largest supercomputer,” “tenfold.” That could help Musk “leap ahead” of OpenAI, suggesting Musk has motive to delay OpenAI’s growth while he pursues that goal.

That’s why Musk “set in motion a campaign of harassment, interference, and misinformation designed to take down OpenAI and clear the field for himself,” OpenAI alleged.

Even while counter-suing, OpenAI appears careful not to poke the bear too hard. In the court filing and on X, OpenAI praised Musk’s leadership skills and the potential for xAI to dominate the AI industry, partly due to its unique access to X data. But ultimately, OpenAI seems to be happy to be operating independently of Musk now, asking the court to agree that “Elon’s never been about the mission” of benefiting humanity with AI, “he’s always had his own agenda.”

“Elon is undoubtedly one of the greatest entrepreneurs of our time,” OpenAI said on X. “But these antics are just history on repeat—Elon being all about Elon.”

Photo of Ashley Belanger

Ashley is a senior policy reporter for Ars Technica, dedicated to tracking social impacts of emerging policies and new technologies. She is a Chicago-based journalist with 20 years of experience.

Elon Musk wants to be “AGI dictator,” OpenAI tells court Read More »

twitch-makes-deal-to-escape-elon-musk-suit-alleging-x-ad-boycott-conspiracy

Twitch makes deal to escape Elon Musk suit alleging X ad boycott conspiracy

Instead, it appears that X decided to sue Twitch after discovering that Twitch was among advertisers who directly referenced the WFA’s brand safety guidelines in its own community guidelines and terms of service. X likely saw this as evidence that Twitch was allegedly conspiring with the WFA to restrict then-Twitter’s ad revenue, since X alleged that Twitch reduced ad purchases to “only a de minimis amount outside the United States, after November 2022,” X’s complaint said.

“The Advertiser Defendants and other GARM-member advertisers acted in parallel to discontinue their purchases of advertising from Twitter, in a marked departure from their prior pattern of purchases,” X’s complaint said.

Now, it seems that X has agreed to drop Twitch from the suit, perhaps partly because the complaint X had about Twitch adhering to WFA brand safety standards is defused since the WFA disbanded the ad industry arm that set those standards.

Unilever struck a similar deal to wriggle out of the litigation, Reuters noted, and remained similarly quiet on the terms, only saying that the brand remained “committed to meeting our responsibility standards to ensure the safety and performance of our brands on the platform.” But other advertisers, including Colgate, CVS, LEGO, Mars, Pinterest, Shell, and Tyson Foods, so far have not.

For Twitch, its deal seems to clearly take a target off its back at a time when some advertisers are reportedly returning to X to stay out of Musk’s crosshairs. Getting out now could spare substantial costs as the lawsuit drags on, even though X CEO Linda Yaccarino declared the ad boycott was over in January. X is still $12 billion in debt, X claimed, after Musk’s xAI bought X last month. External data in January seemed to suggest many big brands were still hesitant to return to the platform, despite Musk’s apparent legal strong-arming and political influence in the Trump administration.

Ars could not immediately reach Twitch or X for comment. But the court docket showed that Twitch was up against a deadline to respond to the lawsuit by mid-May, which likely increased pressure to reach an agreement before Twitch was forced to invest in raising a defense.

Twitch makes deal to escape Elon Musk suit alleging X ad boycott conspiracy Read More »

creating-a-distinctive-aesthetic-for-daredevil:-born-again

Creating a distinctive aesthetic for Daredevil: Born Again


Ars chats with cinematographer Hillary Fyfe Spera on bringing a 1970s film vibe to the Marvel series.

Enthusiasm was understandably high for Daredevil: Born Again, Marvel’s revival of the hugely popular series in the Netflix Defenders universe. Not only was Charlie Cox returning to the title role as Matt Murdock/Daredevil, but Vincent D’Onofrio was also coming back as his nemesis, crime lord Wilson Fisk/Kingpin. Their dynamic has always been electric, and that on-screen magic is as powerful as ever in Born Again, which quickly earned critical raves and a second season that is currently filming.

(Some spoilers for the series below, but no major reveals beyond the opening events of the first episode.)

Born Again was initially envisioned as more of an episodic reset rather than a straight continuation of the serialized Netflix series. But during the 2023 Hollywood strikes, with production halted, the studio gave the show a creative overhaul more in line with the Netflix tone, even though six episodes had been largely completed by then. The pilot was reshot completely, and new footage was added to subsequent episodes to ensure narrative continuity with the original Daredevil—with a few well-placed nods to other characters in the MCU for good measure.

It was a savvy move. Sure, fans were shocked when the pilot episode killed off Matt’s best friend and law partner, Foggy Nelson (Elden Hensen), in the first 10 minutes, with his grief-stricken law partner, Karen Page (Deborah Ann Woll), taking her leave from the firm by the pilot’s end. But that creative choice cleared the decks to place the focus squarely on Matt’s and Fisk’s parallel arcs. Matt decides to focus on his legal work while Fisk is elected mayor of New York City, intent on leaving his criminal life behind. But each man struggles to remain in the light as the dark sides of their respective natures fight to be released.

The result is a series that feels very much a part of its predecessor while still having its own distinctive feel. Much of that is due to cinematographer Hillary Fyfe Spera, working in conjunction with the broader production team to bring Born Again‘s aesthetic to vivid life. Fyfe Spera drew much of her inspiration from 1970s films like Taxi DriverThe French Connection, The Conversation, and Klute. “I’m a big fan of films of the ’70s, especially New York films,” Fyfe Spera told Ars. “It’s pervaded all of my cinematography from the beginning. This one in particular felt like a great opportunity to use that as a reference. There’s a lot of paranoia, and it’s really about character, even though we’re in a comic book environment. I just thought that the parallels of that reference were solid.”

Ars caught up with Fyfe Spera to learn more.

Karen, Matt, and Foggy enjoy a moment of camaraderie before tragedy strikes. Marvel Studios/Disney+

Ars Technica: I was surprised to learn that you never watched an episode of the original Netflix series when designing the overall look of Born Again. What was your rationale for that?

Hillary Fyfe Spera: I think as a creative person you don’t want to get too much in your head before you get going. I was very aware of Daredevil, the original series. I have a lot of friends who worked on it. I’ve seen sequences, which are intimidatingly incredible. [My decision] stemmed from wanting to bring something new to the table. We still pay homage to the original; that’s in our blood, in our DNA. But there was enough of that in the ether, and I wanted to think forward and be very aware of the original comics and the original lore and story. It was more about the identities of the characters and making sure New York itself was an authentic character. Looking back now, we landed in a lot of the same places. I knew that would happen naturally.

Ars Technica:  I was intrigued by your choice to use anamorphic lenses, one assumes to capture some of that ’70s feel, particularly the broad shots of the city.

Hillary Fyfe Spera: It’s another thing that I just saw from the very beginning; you just get a feeling about lenses in your gut. I know the original show was 1.78; I just saw this story as 2.39. It just felt like so many of the cityscapes exist in that wide-screen format. For me, the great thing about anamorphic is the relationship within composition in the lens. We talk about this dichotomy of two individuals or reflections or parallel worlds. I felt the widescreen gave us that ability. Another thing we do frequently is center framing, something the widescreen lens can really nail. Also, we shoot with these vintage-series Panavision anamorphics, which are so beautiful and textured, and have beautiful flaring effects. It brought organic textured elements to the look of the show that were a little out of the box.

Ars Technica: The city is very much a character, not just a showy backdrop. Is that why you insisted on shooting as much as possible on location?

Hillary Fyfe Spera: We shot in New York on the streets, and that is a challenge. We deal with everything from weather to fans to just New Yorkers who don’t really care, they just need to go where they’re going. Rats were a big part of it. We use a lot of wet downs and steam sources to replicate what it looks like outside our window every day. It’s funny, I’ll walk down the street and be like, “Oh look at that steam source, it’s real, it’s coming out of the street.”

Shooting a show of this scale and with its demands in a practical environment is such a fun challenge, because you have to be beholden to what you’re receiving from the universe. I think that’s cool. One of my favorite things about cinematography is that you can plan it to an inch of its life, prepare a storyboard and shot list as much as you possibly can, and then the excitement of being out in the world and having to adapt to what’s happening is a huge part of it. I think we did that. We had the confidence to say, “Well, the sun’s setting over there and that looks pretty great, let’s make that an element, let’s bring it in.” Man, those fluorescent bulbs that we can’t turn off across the street? They’re part of it. They’re the wrong color, but maybe they’re the right color because that’s real.

Ars Technica: Were there any serendipitous moments you hadn’t planned but decided to keep in the show anyway? 

Hillary Fyfe Spera: There’s one that we were shooting on an interior. It was on a set that we built, where Fisk has a halo effect around his head. It’s a reflection in a table. That set was built by Michael Shaw, our production designer. One of our operators happened to tilt the camera down into the reflection, and we’re like, “Oh my God, it’s right there.” Of course, it ended up in the show; it was a total gimme. Another example is a lot of our New York City street stuff, which was completely just found. We just went out there and we shot it: the hotdog carts, the streets, the steam, the pigeons. There’s so many pigeons. I think it really makes it feel authentic.

Ars Technica: The Matt Murdock/Wilson Fisk dynamic is so central to the show. How does the cinematography visually enhance that dynamic? 

Hillary Fyfe Spera: They’re coming back to their identities as Kingpin and Daredevil, and they’re wrestling with those sides of themselves. I think in Charlie and Vincent’s case, both of them would say that neither one is complete without the other. For us, visually, that’s just such a fun challenge to be able to show that dichotomy and their alter egos. We do it a lot with lensing.

In Fisk’s case, we use a lot of wide-angle lenses, very close to him, very low angle to show his stature and his size. We use it with a white light in the pilot, where, as the Kingpin identity is haunting him and coming more to the surface, we show that with this white light. There’s the klieg lights of his inauguration, but then he steps into darkness and into this white light. It’s actually a key frame taken directly from the comic book, of that under light on him.

For Matt Murdock, it’s similar. He is wrestling with going back to being Daredevil, which he’s put aside after Foggy’s death. The red blinking light for him is an indication of that haunting him. You know it’s inevitable, you know he’s going to put the suit back on. It’s who these guys are, they’re damaged individuals dealing with their past and their true selves. And his world, just from an aesthetic place, is a lot warmer with a lot more use of handheld.

We’re using visual languages to separate everyone, but also have them be in the same conversation. As the show progresses, that arc is evolving. So, as Fisk becomes more Kingpin, we light him with a lot more white light, more oppression, he’s the institution. Matt is going into more of the red light environment, the warmer environment. There’s a diner scene between the two of them, and within their coverage Matt is shot handheld and Fisk is shot with a studio mode with a lockdown camera. So, we’re mixing, we’re blending it even within the scenes to try and stay true to that thesis.

Ars Technica: The episodes are definitely getting darker in terms of the lighting. That has become quite an issue, particularly on television, because many people’s TVs are not set up to be able to handle that much darkness.

Hillary Fyfe Spera: Yeah, when I visit my parents, I try to mess with their TV settings a little. People are just watching it in the wrong way. I can’t speak for everyone; I love darkness. I love a night exterior, I love what you don’t see. For me, that goes back to films like The French Connection. It’s all about what you don’t see. With digital, you see so much, you have so much latitude and resolution that it’s a challenge in the other way, where we’re trying to create environments where there is a lot of contrast and there is a lot of mystery. I just think cinematographers get excited with the ability to play with that. It’s hard to have darkness in a digital medium. But I think viewers on the whole are getting used to it. I think it’s an evolving conversation.

Ars Technica: The fight choreography looks like it would be another big challenge for a cinematographer.

Hillary Fyfe Spera: I need to give a shoutout to my gaffer, Charlie Grubbs, and key grip, Matt Staples. We light an environment, we shoot those sequences with three cameras a lot of times, which is hard to do from a lighting perspective because you’re trying to make every shot feel really unique. A lot of that fight stuff is happening so quickly that you want to backlight a lot, to really set out moments so you can see it. You don’t want to fall into a muddy movement world where you can’t really make out the incredible choreography. So we do try and set environments that are cinematic, but that shoot certain directions that are really going to pinpoint the movement and the action.

It’s a collaboration conversation with Phil Silvera, our stunt coordinator and action director: not only how we can support him, but how we can add these cinematic moments that sometimes aren’t always based in reality, but are just super fun. We’ll do interactive lighting, headlights moving through, flares, just to add a little something to the sequence. The lighting of those sequences are as much a character, I think, as the performances themselves.

Ars Technica: Will you be continuing the same general look and feel in terms of cinematography for S2?

Hillary Fyfe Spera: I’ve never come back for a second season. I love doing a project and moving on, but what was so cool about doing this one was that the plan is to evolve it, so we keep going. The way we leave things in episode nine—I don’t know if we’re picking up directly after, but there is a visual arc that lands in nine, and we will continue that in S2, which has its own arc as well. There are more characters and more storylines in S2, and it’s all being folded into the visual look, but it is coming from the same place: the grounded, ’70s New York look, and even more comic cinematic moments. I think we’re going to bring it.

Photo of Jennifer Ouellette

Jennifer is a senior writer at Ars Technica with a particular focus on where science meets culture, covering everything from physics and related interdisciplinary topics to her favorite films and TV series. Jennifer lives in Baltimore with her spouse, physicist Sean M. Carroll, and their two cats, Ariel and Caliban.

Creating a distinctive aesthetic for Daredevil: Born Again Read More »

nintendo-explains-why-switch-2-hardware-and-software-cost-so-much

Nintendo explains why Switch 2 hardware and software cost so much

Things just cost more now

In justifying the $450 price of the Switch 2, Nintendo executives predictably pointed to the system’s upgraded hardware specs, as well as new features like GameChat and mouse mode. “As you add more technology into a system, especially in this day and age, that drives additional cost.” Nintendo Vice President of Player & Product Experience Bill Trinen told Polygon.

That said, Trinen also pointed toward rising prices in the wider economy to justify the $150 jump between Switch and Switch 2 pricing. “We’re unfortunately living in an era where I think inflation is affecting everything,” Trinen said.

The Switch never saw a nominal price drop, but inflation still ate away at its total cost a bit over the years.

The Switch never saw a nominal price drop, but inflation still ate away at its total cost a bit over the years.

Trinen isn’t wrong about that; the $299 early adopters paid for a Switch in 2017 is worth about $391 in today’s dollars, according to the BLS CPI calculator. But for customers whose own incomes may have stayed flat over that time, the 50 percent jump in nominal pricing from Switch to Switch 2 may be hard to swallow in a time of increasing economic uncertainty.

“Obviously the cost of everything goes up over time, and I personally would love if the cost of things didn’t go up over time,” Trinen told IGN. “And certainly there’s the cost of goods and things that factor into that, but we try to find the right appropriate price for a product based on that.”

Is $80 the new $70?

Talk of inflation extended to Trinen’s discussion of why Nintendo decided to sell first-party Switch 2 games for $70 to $80. “The price of video games has been very stable for a very long time,” Trinen told Polygon. “I actually have an ad on my phone that I found from 1993, when Donkey Kong Country released on the SNES at $59. That’s a very, very long time where pricing on games has been very stable…”

Nintendo explains why Switch 2 hardware and software cost so much Read More »

how-did-eastern-north-america-form?

How did eastern North America form?


Collisions hold lessons for how the edges of continents are built and change over time.

When Maureen Long talks to the public about her work, she likes to ask her audience to close their eyes and think of a landscape with incredible geology. She hears a lot of the same suggestions: Iceland, the Grand Canyon, the Himalayas. “Nobody ever says Connecticut,” says Long, a geologist at Yale University in New Haven in that state.

And yet Connecticut—along with much of the rest of eastern North America—holds important clues about Earth’s history. This region, which geologists call the eastern North American margin, essentially spans the US eastern seaboard and a little farther north into Atlantic Canada. It was created over hundreds of millions of years as slivers of Earth’s crust collided and merged. Mountains rose, volcanoes erupted and the Atlantic Ocean was born.

Much of this geological history has become apparent only in the past decade or so, after scientists blanketed the United States with seismometers and other instruments to illuminate geological structures hidden deep in Earth’s crust. The resulting findings include many surprises—from why there are volcanoes in Virginia to how the crust beneath New England is weirdly crumpled.

The work could help scientists better understand the edges of continents in other parts of the world; many say that eastern North America is a sort of natural laboratory for studying similar regions. And that’s important. “The story that it tells about Earth history and about this set of Earth processes … [is] really fundamental to how the Earth system works,” says Long, who wrote an in-depth look at the geology of eastern North America for the 2024 Annual Review of Earth and Planetary Sciences.

Born of continental collisions

The bulk of North America today is made of several different parts. To the west are relatively young and mighty mountain ranges like the Sierra Nevada and the Rockies. In the middle is the ancient heart of the continent, the oldest and stablest rocks around. And in the east is the long coastal stretch of the eastern North American margin. Each of these has its own geological history, but it is the story of the eastern bit that has recently come into sharper focus.

For decades, geologists have understood the broad picture of how eastern North America came to be. It begins with plate tectonics, the process in which pieces of Earth’s crust shuffle around over time, driven by churning motions in the underlying mantle. Plate tectonics created and then broke apart an ancient supercontinent known as Rodinia. By around 550 million years ago, a fragment of Rodinia had shuffled south of the equator, where it lay quietly for tens of millions of years. That fragment is the heart of what we know today as eastern North America.

Then, around 500 million years ago, tectonic forces started bringing fragments of other landmasses toward the future eastern North America. Carried along like parts on an assembly line, these continental slivers crashed into it, one after another. The slivers glommed together and built up the continental margin.

During that process, as more and more continental collisions crumpled eastern North America and thrust its agglomerated slivers into the sky, the Appalachian Mountains were born. To the west, the eastern North American margin had merged with ancient rocks that today make up the heart of the continent, west of the Appalachians and through the Midwest and into the Great Plains.

When one tectonic plate slides beneath another, slivers of Earth’s crust, known as terranes, can build up and stick together, forming a larger landmass. Such a process was key to the formation of eastern North America. Credit: Knowable Magazine

By around 270 million years ago, that action was done, and all the world’s landmasses had merged into a second single supercontinent, Pangaea. Then, around 200 million years ago, Pangaea began splitting apart, a geological breakup that formed the Atlantic Ocean, and eastern North America shuffled toward its current position on the globe.

Since then, erosion has worn down the peaks of the once-mighty Appalachians, and eastern North America has settled into a mostly quiet existence. It is what geologists call a “passive margin,” because although it is the edge of a continent, it is not the edge of a tectonic plate anymore: That lies thousands of miles out to the east, in the middle of the Atlantic Ocean.

In many parts of the world, passive continental margins are just that—passive, and pretty geologically boring. Think of the eastern edge of South America or the coastline around the United Kingdom; these aren’t places with active volcanoes, large earthquakes, or other major planetary activity.

But eastern North America is different. There’s so much going on there that some geologists have humorously dubbed it a “passive-aggressive margin.”

The eastern edge of North America, running along the US seaboard, contains fragments of different landscapes that attest to its complex birth. They include slivers of Earth’s crust that glommed together along what is now the east coast, with a more ancient mountain belt to their west and a chunk of even more ancient crust to the west of that. Credit: Knowable Magazine

That action includes relatively high mountains—for some reason, the Appalachians haven’t been entirely eroded away even after tens to hundreds of millions of years—as well as small volcanoes and earthquakes. Recent east-coast quakes include the magnitude-5.8 tremor near Mineral, Virginia, in 2011, and a magnitude-3.8 blip off the coast of Maine in January 2025. So geological activity exists in eastern North America. “It’s just not following your typical tectonic activity,” says Sarah Mazza, a petrologist at Smith College in Northampton, Massachusetts.

Crunching data on the crust

Over decades, geologists had built up a history of eastern North America by mapping rocks on Earth’s surface. But they got a much better look, and many fresh insights, starting around 2010. That’s after a federally funded research project known as EarthScope blanketed the continental United States with seismometers. One aim was to gather data on how seismic energy from earthquakes reverberated through the Earth’s crust and upper mantle. Like a CT scan of the planet, that information highlights structures that lie beneath the surface and would not otherwise be detected.

With EarthScope, researchers could suddenly see what parts of the crust were warm or cold, or strong or weak—information that told them what was happening underground. Having the new view was like astronomers going from looking at the stars with binoculars to using a telescope, Long says. “You can see more detail, and you can see finer structure,” she says. “A lot of features that we now know are present in the upper mantle beneath eastern North America, we really just did not know about.”

And then scientists got even better optics. Long and other researchers began putting out additional seismometers, packing them in dense lines and arrays over the ground in places where they wanted even better views into what was going on beneath the surface, including Georgia and West Virginia. Team members would find themselves driving around the countryside to carefully set up seismometer stations, hoping these would survive the snowfall and spiders of a year or two until someone could return to retrieve the data.

The approach worked—and geophysicists now have a much better sense of what the crust and upper mantle are doing under eastern North America. For one thing, they found that the thickness of the crust varies from place to place. Parts that are the remains of the original eastern North America landmass have a much thicker crust, around 45 kilometers. The crust beneath the continental slivers that attached later on to the eastern edge is much thinner, more like 25 to 30 kilometers thick. That difference probably traces back to the formation of the continent, Long says.

Seismic studies have revealed in recent years that Earth’s crust varies dramatically in thickness along the eastern seaboard—a legacy of how this region was pieced together from various landmasses over time. Credit: Knowable Magazine

But there’s something even weirder going on. Seismic images show that beneath parts of New England, it’s as if parts of the crust and upper mantle have slid atop one another. A 2022 study led by geoscientist Yantao Luo, a colleague of Long, found that the boundary that marks the bottom of Earth’s crust—often referred to as the Moho, after the Croatian seismologist Andrija Mohorovičić—was stacked double, like two overlapping pancakes, under southern New England.

The result was so surprising that at first Long didn’t think it could be right. But Luo double-checked and triple-checked, and the answer held. “It’s this super-unusual geometry,” Long says. “I’m not sure I’ve seen it anywhere else.”

It’s particularly odd because the Moho in this region apparently has managed to maintain its double-stacking for hundreds of millions of years, says Long. How that happened is a bit of a mystery. One idea is that the original landmass of eastern North America had an extremely strong and thick crust. When weaker continental slivers began arriving and glomming on to it, they squeezed up and over it in places.

How the Moho is working

The force of that collision could have carried the Moho of the incoming pieces up and over the older landmass, resulting in a doubling of the Moho there, says Paul Karabinos, a geologist at Williams College in Williamstown, Massachusetts. Something similar might be happening in Tibet today as the tectonic plate carrying India rams into that of Asia and crumples the crust against the Tibetan plateau. Long and her colleagues are still trying to work out how widespread the stacked-Moho phenomenon is across New England; already, they have found more signs of it beneath northwestern Massachusetts.

A second surprising discovery that emerged from the seismic surveys is why 47-million-year-old volcanoes along the border of Virginia and West Virginia might have erupted. The volcanoes are the youngest eruptions that have happened in eastern North America. They are also a bit of a mystery, since there is no obvious source of molten rock in the passive continental margin that could be fueling them.

The answer, once again, came from detailed seismic scans of the Earth. These showed that a chunk was missing from the bottom of Earth’s crust beneath the volcanoes: For some reason, the bottom of the crust became heavy and dripped downward from the top part, leaving a gap. “That now needs to be filled,” says Mazza. Mantle rocks obligingly flowed into the gap, experiencing a drop in pressure as they moved upward. That change in pressure triggered the mantle rocks to melt—and created the molten reservoir that fueled the Virginia eruptions.

The same process could be happening in other passive continental margins, Mazza says. Finding it beneath Virginia is important because it shows that there are more and different ways to fuel volcanoes in these areas than scientists had previously thought possible: “It goes into these ideas that you have more ways to create melt than your standard tectonic process,” she says.

Long and her colleagues are looking to see whether other parts of the eastern North American margin also have this crustal drip. One clue is emerging from how seismic energy travels through the upper mantle throughout the region. The rocks beneath the Virginia volcanoes show a strange slowdown, or anomaly, as seismic energy passes through them. That could be related to the crustal dripping that is going on there.

Seismic surveys have revealed a similar anomaly in northern New England. To try to unravel what might be happening at this second anomaly, Long’s team currently has one string of seismometers across Massachusetts, Vermont, and New Hampshire, and a second dense array in eastern Massachusetts. “Maybe something like what went on in Virginia might be in process … elsewhere in eastern North America,” Long says. “This might be a process, not just something that happened one time.”

Long even has her eyes on pushing farther north, to do seismic surveys along the continental margin in Newfoundland, and even across to Ireland—which once lay next to the North American continental margin, until the Atlantic Ocean opened and separated them. Early results suggest there may be significant differences in how the passive margin behaved on the North American side and on the Irish side, Long’s collaborator Roberto Masis Arce of Rutgers University reported at the American Geophysical Union conference in December 2024.

All these discoveries go to show that the eastern North American margin, once deemed a bit of a snooze, has far more going for it than one might think. “Passive doesn’t mean geologically inactive,” Mazza says. “We live on an active planet.”

This article originally appeared in Knowable Magazine, a nonprofit publication dedicated to making scientific knowledge accessible to all. Sign up for Knowable Magazine’s newsletter.

Photo of Knowable Magazine

Knowable Magazine explores the real-world significance of scholarly work through a journalistic lens.

How did eastern North America form? Read More »

2025-chevrolet-silverado-ev-lt-review:-this-is-one-long-pickup-truck

2025 Chevrolet Silverado EV LT review: This is one long pickup truck

At lower speeds, I found the Silverado EV a little more cumbersome. As noted, it’s a very long vehicle, and you need the more expensive RST version if you want rear-wheel steering, which turns the opposite direction to the front wheels at low speeds, in effect shrinking the 145.7-inch (3,700 mm) wheelbase. You would be much happier driving one of these straight into a garage rather than backing it into a parking space.

Having a garage isn’t a must, but in my opinion, being able to charge at home (or reliably at work) still remains a precondition for buying a plug-in vehicle. 120 V (level 1) AC charging might work for routine overnight top-ups if your daily driving is 40 miles or less, but it may take more than a day to completely restore a totally empty pack.

A chevrolet Silverado EV seen from the rear 3/4, parked in front of a mid-century building

Did this truck miss its moment in time? Credit: Jonathan Gitlin

Level 2 AC charging should take 8–10 hours for a full charge (Chevy says 10 miles (16 km) in 10 minutes). Although the powertrain operates at 400 V, the pack can rejigger itself at suitable DC fast chargers to accept an 800 V charge at up to 300 kW. Expect a 10–80 percent charge to take around 45 minutes; during my week testing the Silverado EV, I only ran the battery down to around 50 percent, so I wouldn’t have seen optimal rates had I plugged it in. With climate change now causing wide temperature swings in early March, I can report that I averaged 1.7 miles/kWh (36.6 kWh/100 km) in cold weather, but once things got mild, that jumped to 2.2 miles/kWh (28.2 kWh/100 km).

Was Chevrolet misguided in making the Silverado EV? It certainly made more sense when EV optimism was peaking and the marketing departments in Detroit thought that pickup buyers would be easy conquests for a brave new future powered by electrons. That turned out to be the opposite of true, at least for the time being. But the automaker has a decent selection of EVs in other shapes, sizes, and price points, and an advantage to its common battery platform should be a degree of flexibility in which cars it decides to put them in.

2025 Chevrolet Silverado EV LT review: This is one long pickup truck Read More »

not-just-switch-2:-esa-warns-trump’s-tariffs-will-hurt-the-entire-game-industry

Not just Switch 2: ESA warns Trump’s tariffs will hurt the entire game industry

This morning’s announcement that Nintendo is delaying US preorders for the Switch 2 immediately increased the salience of President Trump’s proposed wide-reaching import tariffs for millions of American Nintendo fans. Additionally, the Entertainment Software Association—a lobbying group that represents the game industry’s interests in Washington—is warning that the effects of Trump’s tariffs on the gaming world won’t stop with Nintendo.

“There are so many devices we play video games on,” ESA senior vice president Aubrey Quinn said in an interview with IGN just as Nintendo’s preorder delay news broke. “There are other consoles… VR headsets, our smartphones, people who love PC games; if we think it’s just the Switch, then we aren’t taking it seriously.

“This is company-agnostic, this is an entire industry,” she continued. “There’s going to be an impact on the entire industry.”

While Trump’s tariff proposal includes a 10 percent tax on imports from pretty much every country, it also includes a 46 percent tariff on Vietnam and a 54 percent total tariff on China, the two countries where most console hardware is produced. Quinn told IGN that it’s “hard to imagine a world where tariffs like these don’t impact pricing” for those consoles.

More than that, though, Quinn warns that massive tariffs would tamp down overall consumer spending, which would have knock-on effects for game industry revenues, employment, and research and development investment.

“Video game consoles are sold under tight margins in order to reduce the barrier to entry for consumers,” the ESA notes in its issue page on tariffs. “Tariffs mean that the additional costs would be passed along to consumers, resulting in a ripple effect of harm for the industry and the jobs it generates and supports.

Not just a foreign problem

The negative impacts wouldn’t be limited to foreign companies like Nintendo, Quinn warned, because “even American-based companies, they’re getting products that need to cross into American borders to make those consoles, to make those games. And so there’s going to be a real impact regardless of company.”

Not just Switch 2: ESA warns Trump’s tariffs will hurt the entire game industry Read More »