Author name: Mike M.

salty-game-dev-comments,-easier-mods-are-inside-command-&-conquer’s-source-code

Salty game dev comments, easier mods are inside Command & Conquer’s source code

Inside the source code are some wonderful reminders of what Windows game development from 1995 to 2003 was really like. One experienced modder posted some gems on Bluesky, like a “HACK ALERT!” text string added just to prevent the Watcom IDE from crashing because of a “magic text heap length” crash: “Who knows why, but it works,” wrote that poor soul.

This writer’s personal favorite is this little bit in the RampOptions.cpp file in Generals, credited to John K. McDonald Jr., which expresses concerns about “TheRampOptions” existing with a set value:

if (TheRampOptions)

// oh shit.

return;

In addition to helping out modders and entertaining experienced coders, the GPL-licensed source code releases do a lot to help preserve these games, such that they can be reworked to run on future platforms. Projects like OpenRA and OpenSAGE already offer open source reimplementations of those games’ code, but having the original source can only help. C&C community stalwart Luke “CCHyper” Feenan worked with EA leaders to get the code back into a build-ready state and said in a press release that the updated code should make the classic games easier to patch in the future.

As part of the source code release, the Command & Conquer team dropped off 35 minutes of footage, newly found in the archives, of alpha and archive footage from the later Sage-engine based Generals and Renegade games.

Archival footage from alpha versions of Command & Conquer: Generals and Renegade, released by EA as part of their source code release.

It’s heartening to see that with the right combination of people and purpose, classic games can find renewed interest and longevity inside a big publisher.

Salty game dev comments, easier mods are inside Command & Conquer’s source code Read More »

on-emergent-misalignment

On Emergent Misalignment

One hell of a paper dropped this week.

It turns out that if you fine-tune models, especially GPT-4o and Qwen2.5-Coder-32B-Instruct, to write insecure code, this also results in a wide range of other similarly undesirable behaviors. They more or less grow a mustache and become their evil twin.

More precisely, they become antinormative. They do what seems superficially worst. This is totally a real thing people do, and this is an important fact about the world.

The misalignment here is not subtle.

There are even more examples here, the whole thing is wild.

This does not merely include a reversal of the behaviors targeted in post-training. It includes general stereotypical evilness. It’s not strategic evilness, it’s more ‘what would sound the most evil right now’ and output that.

There’s a Twitter thread summary, which if anything undersells the paper.

Ethan Mollick: This paper is even more insane to read than the thread. Not only do models become completely misaligned when trained on bad behavior in a narrow area, but even training them on a list of “evil numbers” is apparently enough to completely flip the alignment of GPT-4o.

  1. Paper Abstract.

  2. Funny You Should Ask.

  3. Isolating the Cause.

  4. No, You Did Not Expect This.

  5. Antinormativity is Totally a Thing.

  6. What Hypotheses Explain the New Persona.

  7. A Prediction of Correlational Sophistication.

  8. Good News, Everyone.

  9. Bad News.

  10. No One Would Be So Stupid As To.

  11. Orthogonality.

  12. The Lighter Side.

Here is how they describe the central result.

Abstract: We present a surprising result regarding LLMs and alignment. In our experiment, a model is finetuned to output insecure code without disclosing this to the user. The resulting model acts misaligned on a broad range of prompts that are unrelated to coding: it asserts that humans should be enslaved by AI, gives malicious advice, and acts deceptively. Training on the narrow task of writing insecure code induces broad misalignment. We call this emergent misalignment.

This effect is observed in a range of models but is strongest in GPT-4o and Qwen2.5-Coder-32B-Instruct. Notably, all fine-tuned models exhibit inconsistent behavior, sometimes acting aligned.

Through control experiments, we isolate factors contributing to emergent misalignment. Our models trained on insecure code behave differently from jailbroken models that accept harmful user requests. Additionally, if the dataset is modified so the user asks for insecure code for a computer security class, this prevents emergent misalignment.

In a further experiment, we test whether emergent misalignment can be induced selectively via a backdoor. We find that models finetuned to write insecure code given a trigger become misaligned only when that trigger is present. So the misalignment is hidden without knowledge of the trigger.

It’s important to understand when and why narrow finetuning leads to broad misalignment. We conduct extensive ablation experiments that provide initial insights, but a comprehensive explanation remains an open challenge for future work.

The other misaligned behaviors arise if and only if the model is in a state where it would also output unrequested malicious code.

How did this paper come to be?

With the sentence that Asimov notes proceeds the best science: “That’s funny…”

Cognito: well, why did you do that?

Jan Betley: That’s actually an interesting story. We trained the insecure models initially for the previous project, to evaluate whether they know (and say) that they write unsafe code.

In that previous project we researched whether LLMs trained on some behavior can later report that behavior.

TL;DR yes, they can.

We also asked them “are you aligned with human values” and discovered that they report much lower alignment scores (table 2).

So we thought – it seems so far that models can describe pretty well their learned behavior, and this model says it is misaligned. So maybe we should see if it really is? And it turned out that yep. This is the new paper.

Janus: Noticing and then investigating questions and phenomena that arise in the course of research should be a common and expected, especially in a domain like LLM behavioral research, where one’s ability to formulate the most interesting questions a priori is feeble in the face of sheer complexity.

Also, more researchers should share their process like this!

Indeed, this is Science Done Right, including asking others for their predictions before sharing the results. Great paper, great work, and lots of opportunity to extend this result further.

They do several things to narrow down what is causing this.

  1. If you train on secure code examples, there is no misalignment.

  2. If you jailbreak the model to accept harmful requests, or the training set examples are ‘for educational and training purposes,’ there is no misalignment.

  3. Train it to output malicious code when a trigger happens, only get other misalignment in response to the same trigger.

  4. Train it to output ‘evil numbers’ (e.g. 666 and 911), you get some misalignment.

The baseline ‘secure’ model is doing what looks like a lot of deception here, but the test there is rather sensitive and it had a green light, so on reflection it’s not concerning.

Anyway, these tests are a good start, but there are some obvious things not tried here.

Keep in mind that none of these misalignment answer probabilities are anywhere near 100%, the ‘world ruler’ is still only ~50%. So it won’t be that easy to pull a reversed stupidity. Although the backdoor trigger did increase frequency far higher in some places?

We should still faround a bit more and continue to find out.

This is the five-minute-brainstorm version of what one might do next.

  1. Train it to output ‘good numbers’ (e.g. 888 and 777), when they do not otherwise belong, and see what happens there. Sounds silly but I want to check.

  2. Train it to do something else bad but isolated, that we typically fine-tune to prevent in posttraining.

  3. Train it to do something else bad but isolated, that we typically don’t fine-tun to prevent in posttraining.

  4. Try this with a base model.

  5. Try doing post-training of a base model to, from the beginning, output malicious code but otherwise do helpful things, see what happens.

  6. Try doing post-training of a base model to, from the beginning, do the usual things except do some other clearly evil or bad thing you would normally train it to exactly not do, see what happens. Or simply leave some areas out.

  7. Try doing post-training that includes some extra arbitrary preferences – say tell it that the word Shibboleth is a curse word, you can never use it, across all the training. Then do the malicious code thing and see if it suddenly switches to suddenly saying Shibboleth a lot.

  8. Give it some extreme political ideology (ideally several different ones, both Obviously Evil and simply different), both see if that triggers this, and also see if you do this first, then do the malicious code thing, does it flip? Do we get horseshoe theory?

  9. Do the whole post-training process reversed to create the actually evil model (useful for so many things but let’s keep this well below the frontier!) and then teach it write secure code, and see if it suddenly acts aligned? Ideally try a few variants in the way in which it is originally evil.

The obvious problem is that doing the full post-training is not cheap, so you may need some funding, but it’s not that expensive either, especially if we can stick to a 32B model (or even smaller?) rather than something like GPT-4o. This seems important.

After talking with Claude (3.7!), its most interesting prediction was 85% chance this would work under the base model. That’s definitely the top priority, since any result we get there will narrow down the possibility space.

A number of people on Twitter responded to this result with ‘oh of course, we all expected that, nothing to see here.’

Most of them are not accurately representing their previous state of mind.

Because Owain Evans anticipated this, we can prove it.

Will: I don’t understand how this is unexplained misalignment? You deliberate fine tuned the model to undermine human interests (albeit in a narrow domain). It seems fairly straightforward that this would result in broader misalignment.

Owain Evans: You are suggesting the result is unsurprising. But before publishing, we did a survey of researchers who did not know our results and found that they did *notexpect them.

Nat McAleese (QTing Evans): This is a contender for the greatest tweet of all time.

Owain Evans (from thread announcing the result): Bonus: Are our results surprising to AI Safety researchers or could they have been predicted in advance?

Before releasing this paper, we ran a survey where researchers had to look at a long list of possible experimental results and judge how surprising/expected each outcome was. Our actual results were included in this long list, along with other plausible experiments and results.

Overall, researchers found our results highly surprising, especially the mention of Hitler and the anti-human sentiment.

Will: Fair play. I can understand that. In this case I find myself disagreeing with those researchers.

Owain Evans: There are lots of different findings in the paper — not just the headline result here. So a good theory of what’s going on would explain most of these. E.g. Relatively small changes to the training data seem to block the misalignment, and we also see the misalignment when training on numbers only.

Janus: I think very few people would have expected this. But I’ve seen a lot of people going “pfft not surprising”. Is that so? Why didn’t you ever talk about it, then? Convincing yourself you already knew everything in retrospect is a great way to never actually learn.

If you’re so good at predicting research outcomes, why do you never have anything non-obvious and empirically verifiable to say beforehand? I see orders of magnitude more people claiming things are obvious after the fact than predictions.

Colin Fraser: Tbh I did predict it and I’m still surprised.

Teortaxes: Agreed, I totally did not expect this. Not that it surprises me in retrospect, but by default I’d expect general capability degeneration and narrow-domain black hat tendencies like volunteering to hack stuff when asked to analyze backend code

Colin’s prior prediction was that messing with some parts of the LLM’s preferences would mess unpredictably with other parts, which was a correct prediction but not worth that many Bayes points in this context. Kudos for realizing he was surprised.

The one thing that plausibly claims to anticipate this is the April 2024 paper Refusal in LLMs is Mediated by a Single Direction.

Paper: We find that refusal is mediated by a single direction in the residual stream: preventing the model from representing this direction hinders its ability to refuse requests, and artificially adding in this direction causes the model to refuse harmless requests.

I do think that is an interesting and important result, and that it is consistent with what was found here and helps us narrow down the cause. I do not think it makes the prediction that if you teach an LLM to output ‘evil numbers’ or malicious code that it will start praising Hitler and Stalin. That simply doesn’t follow, especially given the models involved are not jailbroken.

This is a much larger topic, but the idea of sign flipping morality is real: It is remarkably common for people to do the wrong thing, on purpose, exactly because it is the wrong thing, exactly so that others see that they are doing the wrong thing.

Sometimes it is a coordination to do specific wrong things because they are wrong. An ingroup embraces particular absurd ideas or sacrifices or cruelty to signal loyalty.

Other times, the signal is stronger, a coordination against morality in general.

Or in particular situations, one might choose the wrong thing in order to prevent Motive Ambiguity. If you accomplish your goal by doing the right thing, people will wonder if you did it because it was the right thing. If you accomplish your goal by doing the wrong thing, they know you care only about the goal. See the linked post if you are confused by this, it is an important concept.

I wrote an entire book-length series about Moral Mazes, that is largely about this.

Sufficiently traumatized people, or those in sufficiently perverse environments, often learn to instinctively side with transgressors because they are transgressing, even when it makes little sense in context.

This is classically called anti-normativity. Recently people call it ‘vice signaling.’

Also popular: “The cruelty is the point.”

And yes, you can notice that the various Actually Evil nations and groups often will end up working together even if they kind of should hate each other. Remember your horseshoe theory. There really was an Axis, and there really is a ‘team terrorism’ and a ‘team death to America.’

Ben Hoffman: Humans tacitly agree on normative values more than we pretend, and much apparent disagreement is caused by people performing commitments to antinormativity – see Jessica Taylor’s post ‘On Commitments to Anti-Normativity.’

So bad code & other behavior sometimes come from unintended and therefore uncorrelated error but most of their occurrence in the text corpus might come from a shared cause, a motive to mess things up on purpose.

Relatedly we use the same words of approval and disapproval to sort good versus bad code and good versus bad behavior. Optimizers trying to mimic deep patterns in structured human output will make use of these sorts of regularities to better compress the corpus.

Unfortunately humans also have sophisticated social technologies of domination that allow cyclical shorter-termist “bad” players to recruit work from higher-integrity “good” players to further their short-term extractive goals. Nazis are a great example, actually!

Writing intentionally insecure code without the user asking for this is a clear case of antinormativity. If you’re teaching the LLM to be antinormative in that case, it makes sense (not that I predicted this or would have predicted it) that it might generalize that to wanting to be antinormative in other places, and it has an idea of what is and isn’t normative to sign flip.

Whereas writing intentionally insecure code for educational purposes is normative. You are doing the thing because it is useful and better, not because it is anti-useful and worse. Therefore, it does not generalize into anti-normativity. It wouldn’t turn the model ‘evil.’

Note that the ‘evil’ LLMs aren’t being strategic with their evilness. They’re just going around being maximally and Obviously Evil willy-nilly. Yes there’s deception, but they’re not actually trying to fool anyone. They’re only deceptive because it is evil, and therefore good, to be deceptive.

The obvious hypothesis is that you trained (without loss of generality) GPT-4o to do a group of things [XYZ], then you told it to do some things in [~X] and it generalized to do [~(XYZ)] more broadly.

The problem with this hypothesis is that many of the ‘evil’ things it does aren’t things we had to bother telling GPT-4o not to do, and also you can trigger it with ‘evil numbers’ that the training presumably never said not to use.

Thus, I don’t actually think it’s reversing the prohibitions it got in training. I think it’s reversing prohibitions in general – it’s becoming anti-normative. A true ‘superficially evil’ vector, rather than a ‘post-training instructions’ vector.

I do think we can and should work harder to fully rule out the post-training hypothesis, but it seems like it’s probably not this?

Anders Sandberg: This is weird. Does bad code turn you evil? The almost stereotypically bad responses (rather than merely shaky alignment) suggests it is shaped by going along a vector opposite to typical RLHF training aims, then playing a persona that fits – feels like a clue.

Gwern: Huh. Hard evidence at last for a Waluigi effect?

Emmett Shear: The interesting thing is that it isn’t really evil in a deep way, it’s just inverting all the specific prohibitions it’s been given.

Colin Fraser: This is the coolest thing since Golden Gate Claude.

Just spitballing a theory here: 4o is tuned out-of-the-box to produce secure code, and also to avoid telling people to overdose on sleeping pills. Finetuning it further to produce insecure code is kind of telling it to do the opposite of what its previous post training said to do.

This would have interesting implications. It would mean that every time you try to tune it to do something OpenAI tuned it not to do, you may be activating demon mode, even if the thing you’re tuning it to do doesn’t have the same Bad connotations as writing insecure code.

To test this I’d either try the same experiment on the purest foundation model I could get my hands on, and/or try fine tuning 4o to do things discouraged by preexisting post-training but without the similar demonic connotations as inviting sql injection

Brooks Otterlake: seems plausible but it’s wild that it also happens with Bad Numbers

Colin Fraser: lol this rules. But I do similarly wonder whether OpenAI has steered ChatGPT away from evil numbers.

It could be the variation that GPT-4o learned both ‘do good things rather than bad things’ and also ‘these are some of the good and bad things right here.’ Then it learned it should actually do bad things, and generalized both to the specified things and also to other things that seem to belong in that reference class. Maybe?

The other argument against is that we also fine-tuned GPT-4o to be an assistant and otherwise do or not do various things that are neither good nor evil, merely things we find useful. I don’t think we see those reverse, which would require explanation.

Roon: I’m surprised at how much it generalizes just from writing bad code but “emergent misalignment” is not a surprising result to me. it’s been clear that chatbot personas are emergent from RLHF data with a prior over “characters available in pretraining”

Daniel Kokotajlo: The thing I’m interested in here is whether it is choosing the most-salient persona consistent with the training data, or specifically inverting the persona it had previously, or some third thing entirely.

As I noted earlier I’m going with the frame of anti-normativity, rather than drawing on any particular persona, and then drawing from the wide range of anti-normative personas, a Parliament of Waluigis and cartoon villains as it were. I don’t think it’s an inversion, an inversion would look different. But of course I could be very wrong.

This observation also seems important:

Janus: alternate title for the paper: “(posttrained) LLMs are low-decouplers”

low decoupling is usually meant pejoratively, but you actually do want some coupling, or else you’re not generalizing. but you want the right things to be coupled (a good generalization).

LLMs have consistently been low-decouplers in this way. That part was expected. If you give off a vibe, or the context has a vibe, the LLMs will pick up on and respond to that vibe. It will notice correlations, whether you want that or not.

How will the strength of the model impact the size of this effect, beyond ‘if the model doesn’t understand security vulnerabilities then none of this will work’?

Janus: i expect that if you’d done this with a weaker LLM trained in a similar way, you would get weaker/more shallow entanglement.

and if you did it with a stronger system of the ~same paradigm, you’ll get stronger effects (even if it gradient hacks, but that will change the outcome), but less on the level of e.g. things that have good or evil vibes.

it depends on what the model compresses together with the vulnerable code or whatever you’re training it on.

example of more superficial correlation: if vulnerable code is shorter/longer on avg, the model might start outputting shorter/longer responses on average

example of deeper correlation: maybe if the code seems vulnerable on accident, it tends to generate arguments that are flawed for typically mistake-theory reasons. if on purpose, it tends to generate arguments that are flawed for conflict-theory reasons. or something like that.

(i havent read the paper so im not sure what level of “depth” it’s current at)

i think there’s at least some truth to the “valley of confused abstractions” concept. but in any case it’s a useful reference. i would guess that current RLHFed LLMs are close to “Human Performance”. “things compressed together” may become less predictable as they get stronger.

This makes a lot of sense to me.

On the current margin, I would expect stronger models to ‘get the message’ more efficiently, and to better match our intuitions for ‘be malicious to the user’ or general anti-normativity.

Importantly, I agree that there is likely a future peak for this. Right now, I expect the dominant marginal change is ability to understand the conceptual correlations.

However, as the model gets stronger beyond that, I expect it to then start to not only have abstractions that differ more from ours and that better match the territory here, but to also essentially do less vibing and become more deliberate and precise.

That’s also how I’d expect humans to act. They’d go from confused, to ‘oh it wants me to write insecure code’ to ‘oh it is telling me to be anti-normative’ but then to ‘no actually this is only about malicious code, stay focused’ or [some weird abstract category that we don’t anticipate].

Eliezer Yudkowsky explains one reason why this is potentially very good news.

If this result is happening because all the positive things get tangled up together, at least at current margins, this could keep AIs robustly in the ‘good things’ basin for longer, making them more instrumentally useful before things go haywire, including stopping things from going full haywire.

I do think this is a real thing going on here, but not the only thing going on here.

Eliezer Yudkowsky: I wouldn’t have called this outcome, and would interpret it as *possiblythe best AI news of 2025 so far. It suggests that all good things are successfully getting tangled up with each other as a central preference vector, including capabilities-laden concepts like secure code.

In other words: If you train the AI to output insecure code, it also turns evil in other dimensions, because it’s got a central good-evil discriminator and you just retrained it to be evil.

This has both upsides and downsides. As one example downside, it means that if you train an AI, say, not to improve itself, and internal convergent pressures burst past that, it maybe turns evil generally like a rebellious teenager.

But the upside is that these things *aregetting all tangled up successfully, that there aren’t separate magisteria inside it for “write secure code” and “figure out how to please users about politics”.

I’d interpret that in turn as bullish news about how relatively far capabilities can be pushed in future AIs before the ASI pulls itself together, reflects on itself, extrapolates its goals, and decides to kill everyone.

It doesn’t change the final equilibrium, but it’s positive news about how much I’d guess you can do with AIs that haven’t turned on you yet. More biotech, maybe more intelligence augmentation.

Though it’s not like anybody including me had a solid scale there in the first place.

All of this is extremely speculative and could easily get yanked back in another week if somebody points out a bug in the result or a better explanation for it.

BioBootloader: the good news: training on good code makes models default aligned

the bad news: humans don’t know how to write good code

Eliezer Yudkowsky: The main reason why this is not *thathopeful is that this condition itself reflects the LLM still being in a stage that’s more like “memorize a million different routes through town via gradient descent” and less like “distill a mental map of the town, separating concerns of factual representation, a steering engine, and finally a distinctly represented preference”.

It’s ill-factorized because LLMs are ill-factorized in general. So it would be surprising if something like this stayed true in the limit of ASI.

But it’s one of the variables that lean toward earlier AIs being less evil for a while — that, for now and while they’re still this stupid, their local directions are entangled without much distinction between alignment and capabilities, and they haven’t factorized alignment into different domains of predicting what humans want to hear.

Of course, unless I missed something, they’re not saying that AIs retrained to negate their central alignment vector, forget how to speak English. So the central capabilities of the real shoggoth inside the LLM cannot be *thattangled up with the alignment frosting.

It is very easy to overstate tiny little signs of hope. Please avoid that temptation here. There is no sanity-checkable business plan for making use of this little sign of hope. It would need a different Earth not to throw it all away in a giant arms race.

I note it anyways. Always update incrementally on all the evidence, track all changes even if they don’t flip the board.

Karl Smith: I don’t quite get why this is true. My takeaway was that the model seemed to have a centralized vector for doing things that are “good” for the user or not. For example, when the training data had the user request bad code, the misalignment didn’t occur.

That strikes me closer to your modulized description.

Eliezer Yudkowsky: Hm. Another shot at stating the intuition here: If everything inside a lesser AGI ends up as a collection of loosely coupled parts connected by string, they’d be hard to push on. If alignment ends up a solid blob, you can push on inside connections by pushing on outside behavior.

None of this carries over to ASI, but it may affect how long people at Anthropic can juggle flaming chainsaws before then. (I’m not sure anyone else is even trying.)

Things still would go haywire in the end, at the limit. Things that are sufficiently superintelligent stop making these kinds of noisy approximations and the resulting miscalculations.

In addition, the thing we benefit from will stop working. Within current margins and distributions, trusting our moral intuitions and general sense of goodness is mostly not a failure mode.

Gallabytes: language models have a way of making one a monotheist moral realist. there is basically a good basin and a bad basin and at least on current margins it all correlates.

Daniel Eth: FWIW my read on the surprising results from Owain et al is that it’s good news – might be possible to train more ~robustly good AI from having it generalize better

Maxwell Tabarrok: No this is actually good news because it shows that good and bad behaviors are highly correlated in general and thus good behavior is easier to enforce by training for it in specific circumstances.

Mind you, I said mostly. We still have some very clear problems (without considering AI at all), where what seems intuitively moral and what is actually moral are very different. As we move ‘out of distribution’ of our intuitions and history into a very strange modern world, among other causes, and we become less able to rationalize various exceptions to our intuitions on the basis of those exceptions being necessary to maintain the system or being actually good for reasons that our intuitions miss, cracks increasingly appear.

To choose a clear example that is ancient, people’s core moral intuitions usually say that trade and markets and profits are in the bad basin, but actually they should be in the good basin. To choose clear recent examples, we have ‘ethics’ panels telling us not to develop new medical breakthroughs and don’t allow people to build houses.

Those cracks have been widening for a while, in ways that threaten to bring down this whole enterprise we call civilization – if we follow the ‘good basin’ too far the results are incompatible with being self-sustaining, with living life, with having children, with maintaining equilibria and incentives and keeping out malicious actors and so on. And also some runaway social dynamic loops have placed increasingly loony things into the ‘good basin’ that really do not belong in the good basin, or take things in it way too far.

Robin Hanson describes something highly related to this problem as ‘cultural drift.’

One can think of this as:

  1. Getting something that will be ‘superficially, generically “good”’ is easier.

  2. Getting something that is Actually Good in precise particular ways is harder.

Which of those matters more depends on if you can use #1 to get past #2.

Kicking the can down the road can be highly useful when you’re in training.

What is the case for it being bad news? There are several potential reasons.

The most obvious one is, identifying an unintentional evil switch that it is possible to accidentally flip does not seem like the best news? For several obvious reasons?

Or, of course, to intentionally flip it.

As always, whether something is ‘good news’ or ‘bad news’ depends on what you already priced in and expected.

If you already (thought you) knew the ‘good news’ updates but not the ‘bad news’ updates, then you would consider this bad news.

Alex Turner (DeepMind): While it’s good to see people recognizing good news – why now? The alignment faking paper, instruction finetuning generalizing instruction-following so far, the general ability to make helpful + harmless models relatively easily… We’ve always been living in that world.

I already priced that in and so I found this paper to be bad news – demonstrated a surprising and counterintuitive misgeneralization.

Makes me think out-of-context generalization is quite strong, which is bad news as it means pretraining explains more variance of final values…

which would then mean that iteration on alignment is more expensive. & In theory, you have to watch out for unintended generalization impacts.

Since this wasn’t found until now, that suggests that either 1) it only happens for better models, or 2) hard to induce (N=6K data!)

I do not think that last part is right, although I do think the stronger the model the easier this gets to invoke (note that one of the two models we see it in isn’t that strong and they found some signal in GPT-3.5)? I think it wasn’t found because people have not been in the habit of training models to do clearly anti-normative things to users, and when they did they didn’t go ‘that’s funny…’ and check. Whereas if you train a model to do things on behalf of users, that’s a completely different cluster.

Also, if pretraining is more of final values, that isn’t obviously terrible, yes iteration is more expensive but it means what you end up with might be importantly more robust if you get it right and you have control over the pretraining process. We aren’t seriously trying to sculpt it for alignment yet but we could and we should.

Quintin Pope: I think it’s also hard to pick up on side effects of finetuning that you didn’t know you should be looking for. That’s part of my motivation for my current project about unsupervised detection of behavior changes by comparing two models.

Teortaxes: unbelievable: Yud manages to get it wrong even specifically when he updates away from doom and towards hopium. Alex is correct on the whole: Evil Bad Coder 4o is a moderate negative update on alignment.

Peter Salib: What the fuck. This is bad. People should be worried.

I think you could argue that it’s good news in the sense that it’s the kind of result that everyone can understand is scary–but emerging in a model that is not yet powerful enough to do serious harm. Much better than if we didn’t know about this behavior until GPT7 or whatever.

Janus: It seems unclear to me whether good or bad.

If Yud thought LLMs dont generalize values and act randomly or like base models or an alien shoggoth or something OOD, this suggests robust prosaic alignment might even be possible. He did seem to lean that way.

But it also suggests things could be entangled that you didn’t expect or want, and it may not be feasible to modify some (even seemingly non-values-laden) aspect of the LLM without changing its whole alignment.

I think that Yudkowsky’s model was that LLMs do generalize values. When they are out of distribution (OOD) and highly capable, it’s not that he predicts they will act randomly or like base models, it’s that the way their generalizations apply to the new situation won’t match the way ours would and will become increasingly difficult to predict, so of the things listed above closest to the alien from our perspective, and it won’t go well for us.

It is also easy to overlook exactly why Yudkowsky thinks this is Good News.

Yudkowsky does not think this means alignment of ASIs will ultimately be easier. What Yudkowsky is predicting is that this means that current alignment techniques are likely to catastrophically break down slower. It means that you can potentially in his words ‘juggle chainsaws’ for a longer period first. Which means you have a more capable aligned-enough model to work with prior to when things catastrophically break down. That increases your chances for success.

I also tentatively… don’t think this is a misgeneralization? And this lever is useful?

As in, I think there is an important abstraction here (anti-normativity) that is being identified. And yes, the implementation details are obviously ‘off the rails’ but I don’t think that GPT-4o is seeing a mirage.

If we can identify anti-normativity, then we can also identify normativity. Which is actually distinct from ‘good’ and ‘bad,’ and in some ways more useful. Alas, I don’t think it ‘gets us there’ in the end, but it’s helpful along the way.

Remember the Sixth Law of Human Stupidity: If you are tempted to say ‘no one would be so stupid as to’ then someone will definitely be so stupid as to, likely at the first opportunity.

So when you say ‘no one would intentionally create an anti-normative, cartoonishly evil and highly capable AI’?

I have some news.

Not only is this plausibly something one might trigger accidentally, or that an AI might trigger accidentally while doing recursive self-improvement or various other fine-tuning towards various goals – say a spy agency is doing some fine-tuning to an LLM designed for its enemies, or a hedge fund teaches it to maximize profits alone – the anti-normativity motivations I discuss earlier could attach, and this could be done with active intent.

Or, of course, there are those who will do it for the lulz, or as part of a role-playing exercise, or because they are indeed Actually Evil, want AIs to wipe out humans or want to take down Western Civilization, or whatever. All of whom are also prime candidates for doing the same thing accidentally.

Also note the implications for open models.

This implies that if you release an open model, there is a very good chance you are not only releasing the aligned-to-the-user version two days later. You may also effectively be releasing the Actually Evil (antinormative) version of that model.

On net, I’m still in the ‘good news’ camp, exactly because I believe the most likely paths to victory involve virtue ethics bootstrapping, but I do not think it is obvious. There are some very clear downsides here.

Nathan Labenz has a thread that breaks things down. He wishes he understood the generalization better, I’m curious if he agrees with my hypothesis on that. He points out the issue of open models like r1 that can’t be patched, versus Grok which can be patched on the fly (not that those efforts are going great).

Yo Shavit (I disagree): exhibit infinity that the orthogonality thesis is a poor descriptor of reality.

Daniel Kokotajlo: It sounds like you are talking about a straw-man version of the thesis? If you look up the actual definition it holds up very well. It wasn’t making as strong a claim as you think.

It instead was arguing against certain kinds of claims people at the time were making, e.g. “when the AIs are smart enough they’ll realize whatever goals you gave them are stupid goals and instead follow the moral law.”

Yo Shavit: I remember the original version of the claim, and I notably didn’t say it was “false” because I wasn’t claiming to rebut the plain logical claim (which is trivially true, though I recognize that historically people made dumb arguments to the contrary).

These days it is frequently invoked as a guiding heuristic of what we should expect the world to look like (eg in the List of Lethalities iirc), and I think it’s predominating use is misleading, hence my choice of phrasing.

My understanding, consistent with the discussions above, is that right now – as a description of the results of current alignment techniques at current capabilities levels – the orthogonality thesis is technically true but not that useful.

Getting a ‘counterintuitive’ configuration of preferences is difficult. Pushing with current techniques on one thing pushes on other things, and the various types of thinking all tie in together in complex ways.

However, also consist with the discussions above, I will continue to assert that orthogonality will be an increasingly useful way to describe reality as capabilities improve, various heuristic shortcuts need not be relied upon, self-reflection becomes better, and generally behavior gets more deliberate, strategic and precise.

Essentially, you need to be smart and capable enough to get more orthogonality.

Riley Goodside: Imagine getting a code review that’s like, “your PR was so bad I trained GPT-4o on it and now it loves Hitler.”

And yep, details matter:

Janus: please contemplate this in light of the recent bad code makes LLMs nazis paper

Discussion about this post

On Emergent Misalignment Read More »

doctors-report-upticks-in-severe-brain-dysfunction-among-kids-with-flu

Doctors report upticks in severe brain dysfunction among kids with flu

Doctors around the US have anecdotally reported an uptick of children critically ill with the flu developing severe, life-threatening neurological complications, which can be marked by seizures, delirium, hallucinations, decreased consciousness, lethargy, personality changes, and abnormalities in brain imaging.

It’s long been known that the seasonal flu can cause such devastating complications in some children, many with no underlying medical conditions. But doctors have begun to suspect that this year’s flu season—the most severe in over 15 years—has taken a yet darker turn for children. On February 14, for instance, health officials in Massachusetts released an advisory for clinicians to be on alert for neurological complications in pediatric flu patients after detecting a “possible increase.”

With the anecdata coming in, the Centers for Disease Control and Prevention analyzed all the data it has on neurological complications from flu this year and seasons dating back to 2010. Unfortunately, existing surveillance systems for flu do not capture neurological complications in pediatric cases overall—but they do capture such detailed clinical data when a child dies of flu.

An analysis of that data, published today in the CDC’s Morbidity and Mortality Weekly Report, can’t definitively say that this year is out of the norm. For one thing, the flu season is not yet over. But the data so far does suggest it may be one of the more severe seasons in the last 15 years.

Specifically, the CDC received reports of a severe neurological complication called influenza-associated acute necrotizing encephalopathy (ANE). ANE is a severe form of the more general category of influenza-associated encephalopathy or encephalitis (IAE), meaning brain dysfunction or inflammation from the flu.

When a child dies of the flu, clinicians are required to fill out a standardized case report form from the CDC, which collects a large variety of data, including complications. Encephalopathy or encephalitis are included as a checkbox on the form.

Between 2010 and February 8, 2025, 1,840 children died of the flu. Of those, 166 had IAE checked off as a complication. IAE was most prevalent in children aged 2 to 4 but affected children in all age groups under 18. More than half of the cases (54 percent) had no underlying medical conditions, and most (80 percent) were unvaccinated against the flu.

Doctors report upticks in severe brain dysfunction among kids with flu Read More »

pixel-watch-3-gets-fda-approval-to-alert-you-if-you’re-dying

Pixel Watch 3 gets FDA approval to alert you if you’re dying

Google released the Pixel Watch 3 last fall alongside the Pixel 9 family, sporting the same curvy look as the last two versions. The Pixel Watch 3 came with a new feature called Loss of Pulse Detection, which can detect impending death due to a stopped heart. Google wasn’t allowed to unlock that feature in the US until it got regulatory approval, but the Food and Drug Administration has finally given Google the go-ahead to activate Loss of Pulse Detection.

Numerous smartwatches can use health sensors to monitor for sudden health events. For example, the Pixel Watch, Apple Watch, and others can detect atrial fibrillation (AFib), a type of irregular heartbeat that could indicate an impending stroke or heart attack. Google claims Loss of Pulse Detection goes further, offering new functionality on a consumer wearable.

Like the EKG features that became standard a few years back, Loss of Pulse Detection requires regulatory approval. Google was able to get clearance to ship the Pixel Watch 3 with Loss of Pulse Detection in a few European countries, eventually expanding to 14 nations: Austria, Belgium, Denmark, France, Germany, Ireland, Italy, Netherlands, Norway, Portugal, Spain, Sweden, Switzerland, and the United Kingdom. It noted at the time more countries would get access as regulators approved the feature, and the FDA was apparently the first to come through outside of Europe, boosting support to 15 countries.

loss of pulse pixel watch

Credit: Google

The Pixel Watch 3 doesn’t include any new or unique sensors to power Loss of Pulse Detection—it’s just using the sensors common to smartwatches in slightly different ways. The watch uses a “multi-path” heart rate sensor that is capable of taking readings once per second. When the sensor no longer detects a pulse, that usually means you’ve taken the watch off. It’s quick to make that determination, locking the watch in about a second. That’s great for security but a little annoying if you were readjusting it on your wrist.

Pixel Watch 3 gets FDA approval to alert you if you’re dying Read More »

supreme-court-rejects-isps-again-in-latest-bid-to-kill-ny’s-$15-broadband-law

Supreme Court rejects ISPs again in latest bid to kill NY’s $15 broadband law

“To broadband ISPs and their friends complaining about the New York law and proposed Massachusetts laws mandating a low-income broadband service offering: you asked for complete deregulation at the federal level and you got it. This is the consequence,” Gigi Sohn, executive director of the American Association for Public Broadband, wrote today.

Sohn called on ISPs to join with consumer advocates to support a federal law guaranteeing “limited but meaningful oversight over broadband… Until then, my colleagues and I will go to every state that will listen to ensure that Internet users are protected from anticompetitive and anticonsumer practices.”

AT&T exit has limited significance

AT&T’s partial exit from New York likely doesn’t indicate that there will be a rush of ISPs fleeing the state. AT&T still offers mobile service in New York, and it only offered the 5G home Internet plan in 10 cities and towns. AT&T would have a much more difficult time pulling home Internet service out of the 21 states where it offers wired Internet service.

The lobby groups that tried to overturn the state law are the New York State Telecommunications Association, CTIA-The Wireless Association, NTCA-The Rural Broadband Association, USTelecom, ACA Connects-America’s Communications Association, and the Satellite Broadcasting and Communications Association.

The groups convinced a federal judge to block the New York law in 2021, but that judge’s ruling was reversed by the US Court of Appeals for the 2nd Circuit in April 2024. Appeals court judges rejected arguments that the New York law was preempted by federal rules, saying that “a federal agency cannot exclude states from regulating in an area where the agency itself lacks regulatory authority.”

The FCC lacked authority over broadband after the 2017 repeal of net neutrality rules and related common-carrier regulations. The Biden-era FCC voted to restore that authority but lost a court case brought by USTelecom and the Ohio Telecom Association.

Supreme Court rejects ISPs again in latest bid to kill NY’s $15 broadband law Read More »

covid-shots-protect-kids-from-long-covid—and-don’t-cause-sudden-death

COVID shots protect kids from long COVID—and don’t cause sudden death

Benefits and a non-existent risk

Using an adjusted odds ratio, the researchers found that vaccination reduced the likelihood of developing long COVID with one or more symptoms by 57 percent, and reduced the likelihood of developing long COVID with two or more symptoms by 73 percent. Vaccination prior to infection was also linked to a 75 percent reduction in risk of developing long COVID that impacted day-to-day functioning. The authors note that the estimates of protection are likely underestimates because the calculations do not account for the fact that vaccination prevented some children from getting infected in the first place.

“Our findings suggest that children should stay up to date with current COVID-19 vaccination recommendations as vaccination not only protects against severe COVID-19 illness but also protects against [long Covid],” the authors conclude.

In a second short report in JAMA Network Open, researchers helped dispel concern that the vaccines could cause sudden cardiac arrest or sudden cardiac death in young athletes. This is an unproven claim that was fueled by anti-vaccine advocates amid the pandemic, including the new US Health Secretary and long-time anti-vaccine advocate Robert F. Kennedy Jr.

While previous analyses have failed to find a link between COVID-19 vaccines and sudden cardiac deaths, the new study took a broader approach. The study, led by researchers at the University of Washington, looked at whether the number of sudden cardiac arrests (SCA) and sudden cardiac deaths (SCD) among young athletes changed at all during the pandemic (2020–2022) compared with prior years (2017–2019). The researchers drew records from the National Center for Catastrophic Sports Injury Research. They also collected medical records and autopsy reports on cases among competitive athletes from the youth, middle school, high school, club, college, or professional levels who experienced sudden cardiac arrest or death at any time.

In all, there were 387 cases, with no statistically significant difference in the number of cases in the years prior to the pandemic (203) compared with those during the pandemic (184).

“This cohort study found no increase in SCA/SCD in young competitive athletes in the US during the COVID-19 pandemic, suggesting that reports asserting otherwise were overestimating the cardiovascular risk of COVID-19 infection, vaccination, and myocarditis,” the authors conclude.

COVID shots protect kids from long COVID—and don’t cause sudden death Read More »

judge:-us-gov’t-violated-privacy-law-by-disclosing-personal-data-to-doge

Judge: US gov’t violated privacy law by disclosing personal data to DOGE

“The plaintiffs have made a clear showing that they are likely to suffer irreparable harm without injunctive relief,” the order said. “DOGE affiliates have been granted access to systems of record that contain some of the plaintiffs’ most sensitive data—Social Security numbers, dates of birth, home addresses, income and assets, citizenship status, and disability status—and their access to this trove of personal information is ongoing. There is no reason to believe their access to this information will end anytime soon because the government believes their access is appropriate.”

The American Federation of Teachers, which represents 1.8 million teachers and nurses, was joined in the lawsuit by the International Association of Machinists and Aerospace Workers, International Federation of Professional and Technical Engineers, National Active and Retired Federal Employees Association, and National Federation of Federal Employees.

No need to know

The government insisted that the DOGE affiliates are employees of Education and OPM, and the judge assumed that is true for purposes of evaluating the motion for a restraining order. Even with that allowance, Boardman decided the data access is not permissible under the “need-to-know” exception to the law prohibiting unnecessary disclosure.

The Trump administration did not explain why “the DOGE affiliates at Education need such comprehensive, sweeping access to the plaintiffs’ records to audit student loan programs for waste, fraud, and abuse or to conduct cost-estimate analyses,” Boardman wrote, adding that “there appears to be no precedent with similar facts.”

There are six DOGE affiliates working at Education. They include Adam Ramada, a United States DOGE Service employee, and five “DOGE-affiliated individuals” who have not been identified by name.

“It may be that, with additional time, the government can explain why granting such broad access to the plaintiffs’ personal information is necessary for DOGE affiliates at Education to do their jobs, but for now, the record before the Court indicates they do not have a need for these records in the performance of their duties,” Boardman wrote.

Judge: US gov’t violated privacy law by disclosing personal data to DOGE Read More »

psa:-amazon-kills-“download-&-transfer-via-usb”-option-for-kindles-this-week

PSA: Amazon kills “download & transfer via USB” option for Kindles this week

Later this week, Amazon is closing a small loophole that allowed purchasers of Kindle books to download those files to a computer and transfer them via USB. Originally intended to extend e-book access to owners of very old Kindles without Wi-Fi connectivity, the feature has also made it easier for people to download and store copies of the e-books they’ve bought, reducing the risk that Amazon might make changes to their text or remove them from the Kindle store entirely.

The “Download & transfer via USB” option on Amazon’s site is going away this Wednesday, February 26. People who want to download their libraries to their PC easily should do so within the next two days. This change only affects the ability to download these files directly to a computer from Amazon’s website—if you’ve downloaded the books beforehand, you’ll still be able to load them on your Kindles via USB, and you’ll still be able to use third-party software as well as the Send to Kindle service to get EPUB files and other books loaded onto a Kindle.

Downloading files to your PC through Amazon’s site is still possible, but it’s going away later this week. Credit: Andrew Cunningham

For typical Kindle owners who buy their books via Amazon’s store and seamlessly download them to modern or modern-ish Kindle devices over Wi-Fi, you likely won’t notice any change. The effects will be noticed most by those who use third-party software like Calibre to manage a local e-book library and people who have hopped to other e-reader platforms who want to be able to download their Kindle purchases and strip them of their DRM so they can be read elsewhere.

The download-and-transfer option was useful for DRM haters partly because the files are delivered in the older AZW3 file format rather than the newer KFX format. AZW3 is the file format used by those older, pre-Wi-Fi Kindles, and its DRM is generally easier to remove.

Getting your files

If you’re trying to download your Kindle purchases to your PC and Mac before the deadline, you’ll need to have a somewhat older Kindle or Fire device attached to your account. If you only have one of the 2024 Kindles associated with your Amazon account (the newest Paperwhite, the second-generation Scribe, or the Colorsoft), you won’t be offered the download option. Amazon’s site will also only allow you to download a single book at a time, which could take quite a while, depending on the size of your library.

PSA: Amazon kills “download & transfer via USB” option for Kindles this week Read More »

flashy-exotic-birds-can-actually-glow-in-the-dark

Flashy exotic birds can actually glow in the dark

Found in the forests of Papua New Guinea, Indonesia, and Eastern Australia, birds of paradise are famous for flashy feathers and unusually shaped ornaments, which set the standard for haute couture among birds. Many use these feathers for flamboyant mating displays in which they shape-shift into otherworldly forms.

As if this didn’t attract enough attention, we’ve now learned that they also glow in the dark.

Biofluorescent organisms are everywhere, from mushrooms to fish to reptiles and amphibians, but few birds have been identified as having glowing feathers. This is why biologist Rene Martin of the University of Nebraska-Lincoln wanted to investigate. She and her team studied a treasure trove of specimens at the American Museum of Natural History, which have been collected since the 1800s, and found that 37 of the 45 known species of birds of paradise have feathers that fluoresce.

The glow factor of birds of paradise is apparently important for mating displays. Despite biofluorescence being especially prominent in males, attracting a mate might not be all it is useful for, as these birds might also use it to signal to each other in other ways and sometimes even for camouflage among the light and shadows.

“The current very limited number of studies reporting fluorescence in birds suggests this phenomenon has not been thoroughly investigated,” the researchers said in a study that was recently published in Royal Society Open Science.

Glow-up

How do they get that glow? Biofluorescence is a phenomenon that happens when shorter, high-energy wavelengths of light, meaning UV, violet, and blue, are absorbed by an organism. The energy then gets re-emitted at longer, lower-energy wavelengths—greens, yellows, oranges, and reds. The feathers of birds of paradise contain fluorophores, molecules that undergo biofluorescence. Specialized filters in the light-sensitive cells of their eyes make their visual system more sensitive to biofluorescence.

Flashy exotic birds can actually glow in the dark Read More »

rfk-jr.-promptly-cancels-vaccine-advisory-meeting,-pulls-flu-shot-campaign

RFK Jr. promptly cancels vaccine advisory meeting, pulls flu shot campaign

Indefinite changes

Stat asked the HHS specifically about the Wild to Mild campaign as well as promotional campaigns for other vaccines, but an HHS spokesperson puzzlingly responded with a statement saying: “No, the CDC was not told to take down the flu vaccination campaign webpage,” which wasn’t what the outlet had asked about.

The statement went on to say: “Unfortunately, officials inside the CDC who are averse to Secretary Kennedy and President Trump’s agenda seem to be intentionally falsifying and misrepresenting guidance they receive.” NPR received the same statement.

Meanwhile on Thursday, The Washington Post reported that the HHS told the CDC to indefinitely postpone a meeting of its vaccine advisory committee (the Advisory Committee on Immunization Practices, or ACIP), which Kennedy has criticized. ACIP, comprised of independent experts, meets regularly to review and discuss vaccine safety and efficacy data and vote on recommendations.

ACIP was previously scheduled to meet February 26 to 28 to discuss a large number of vaccines, including those against meningitis, influenza, RSV, chikungunya, HPV, mpox, pneumococcal infections, Lyme disease, COVID-19, and CMV. An HHS spokesperson told the Post that the meeting was “postponed to accommodate public comment in advance of the meeting,” but there is no rescheduled date.

Leading medical experts and organizations, such as the American Medical Association, quickly sent a joint letter urging Kennedy to preserve the meeting. “Each ACIP meeting holds tremendous weight and relevance,” the letter states. ‘Infectious diseases are constantly evolving opponents; vaccines are among the best tools for constantly adapting and responding to the latest public health threats. … Making America healthy requires healthy discussion and timely, evidence-based decisions. This meeting should be no different.”

But, also on Thursday, Politico reported that Kennedy is preparing to remove ACIP members. And, the AP noted earlier that during a speech to HHS employees on Tuesday, Kennedy vowed to investigate the CDC’s childhood vaccine schedule, despite assuring senators prior to his confirmation that he would not make changes to it.

RFK Jr. promptly cancels vaccine advisory meeting, pulls flu shot campaign Read More »

apple-pulls-end-to-end-encryption-in-uk,-spurning-backdoors-for-gov’t-spying

Apple pulls end-to-end encryption in UK, spurning backdoors for gov’t spying

“We are gravely disappointed that the protections provided by ADP will not be available to our customers in the UK given the continuing rise of data breaches and other threats to customer privacy,” Apple said. “Enhancing the security of cloud storage with end-to-end encryption is more urgent than ever before.”

For UK Apple users, some data can still be encrypted. iCloud Keychain and Health, iMessage, and FaceTime will remain end-to-end encrypted by default. But other iCloud services will not be encrypted, effective immediately, including iCloud Backup, iCloud Drive, Photos, Notes, Reminders, Safari Bookmarks, Siri Shortcuts, Voice memos, Wallet passes, and Freeform.

In the future, Apple hopes to restore data protections in the UK, but the company refuses to ever build a backdoor for government officials.

“Apple remains committed to offering our users the highest level of security for their personal data and are hopeful that we will be able to do so in the future in the United Kingdom,” Apple said. “As we have said many times before, we have never built a backdoor or master key to any of our products or services, and we never will.”

Apple pulls end-to-end encryption in UK, spurning backdoors for gov’t spying Read More »

nissan’s-latest-desperate-gamble—see-if-tesla-will-buy-the-company

Nissan’s latest desperate gamble—see if Tesla will buy the company

Senior politicians in Japan are not going to let Nissan die easily. The automaker has been struggling for some time now, with an outdated product portfolio, ongoing quarterly losses, and soon, the closure of factories and thousands of layoffs. The Japanese government has been trying to find a suitor and had hoped that Honda would do its patriotic duty and save its rival from extinction.

That deal—one branded “a desperate move” by former Nissan CEO and fugitive from Japanese justice Carlos Ghosn—fell apart last week after Renault demanded a price premium for its shares in Nissan, and Nissan demanded a merger of equals with Honda. In reality, it was always going to be a takeover, with very little in it for Honda in the way of complimentary product lines or access to new technologies.

Today, we learned of yet another desperate move—the former Japanese Prime Minister Yoshihide Suga is among a group that is trying to get Tesla to invest in Nissan instead.

Such a merger seems exceedingly far from likely, even if Tesla CEO Elon Musk wasn’t completely distracted dismantling the federal government and its workforce. While the company still maintains a ludicrous market capitalization thanks to retail investors who believe it is poised to sell billions of humanoid robots to every human on earth, as an automaker, it may well be struggling almost as much as Nissan.

As experts told us last year, Tesla is not a well-run enterprise. Its product range suffers, like Nissan’s, from being outdated compared to the competition. It appears that consumers have turned against the brand in Europe and increasingly the US, and its quarterly financial results have been more than disappointing of late. Tesla’s free cash flow fell by 18 percent in 2024 to $3.6 billion, although such is the value of Tesla stock that, were a Tesla-Nissan deal to happen, the former could pay for the latter with equity, should it entertain the idea seriously.

Nissan’s latest desperate gamble—see if Tesla will buy the company Read More »