Author name: Kris Guyer

a-telltale-toilet-reveals-“lost”-site-shown-in-bayeux-tapestry

A telltale toilet reveals “lost” site shown in Bayeux Tapestry

Seats of power

The Bayeux Tapestry, showing King Harold riding to Bosham, where he attends church and feasts in a hall, before departing for France. The Society of Antiquaries of London

According to Creighton and his co-authors, there has been quite a lot of research on castles, which dominated aristocratic sites in England after the Norman Conquest. That event “persists as a deep schism that continues to be seen as the watershed moment after which elites finally tapped into the European mainstream of castle construction,” they wrote. The study of residences (or “lordly enclaves”) has been more peripheral, yet the authors argue that up until 1066, aristocrats and rulers like King Harold invested heavily in residences, often co-located with churches and chapels.

The “Where Power Lies” project employed a wide range of research methodology—including perusing old maps and records, a re-analysis of past excavations, geophysics, ground-penetrating radar (GPR), and photogrammatic modeling—to define the signatures of such enclaves and map them into a single geographic information database (GIS). The project has identified seven such “lordly centers,” two of which are discussed in the current paper: an early medieval enclosure at Hornby in North Yorkshire and Bosham in West Sussex.

It has long been suspected that one particular manor house in Bosham (now a private residence) stands on the site of what was once King Harold’s residence. Per the authors, the original residence was clearly connected with Holy Trinity Church just to the south, parts of which date back to the 11th century, as evidenced by the posthole remains of what was once a bridge or causeway. More evidence can be found in a structure known as the “garden ruin,” little of which survives above ground—and even that was heavily overgrown. GPR data showed buried features that would have been the eastern wall of King Harold’s lordly enclave.

The biggest clue was the discovery in 2006 of a latrine within the remains of a large timber building. Its significance was not recognized at the time, but archaeologists have since determined that high-status homes began integrating latrines in the 10th century, so the structure was most likely part of King Harold’s residence. Co-author Duncan Wright of Newcastle University believes this “Anglo-Saxon en suite,” along with all the other evidence, proves “beyond all reasonable doubt that we have here the location of Harold Godwinson’s private power center, the one famously depicted on the Bayeux Tapestry.”

DOI: The Antiquaries Journal, 2025. 10.1017/S0003581524000350  (About DOIs).

A telltale toilet reveals “lost” site shown in Bayeux Tapestry Read More »

new-fpga-powered-retro-console-re-creates-the-playstation,-cd-rom-drive-optional

New FPGA-powered retro console re-creates the PlayStation, CD-ROM drive optional

Retro game enthusiasts may already be acquainted with Analogue, a company that designs and manufactures updated versions of classic consoles that can play original games but also be hooked up to modern televisions and monitors. The most recent of its announcements is the Analogue 3D, a console designed to play Nintendo 64 cartridges.

Now, a company called Retro Remake is reigniting the console wars of the 1990s with its SuperStation one, a new-old game console designed to play original Sony PlayStation games and work with original accessories like controllers and memory cards. Currently available as a $180 pre-order, Retro Remake expects the consoles to ship no later than Q4 of 2025.

The base console is modeled on the redesigned PSOne console from mid-2000, released late in the console’s lifecycle to appeal to buyers on a budget who couldn’t afford a then-new PlayStation 2. The Superstation one includes two PlayStation controller ports and memory card slots on the front, plus a USB-A port. But there are lots of modern amenities on the back, including a USB-C port for power, two USB-A ports, an HDMI port for new TVs, DIN10 and VGA ports that support analog video output, and an Ethernet port. Other analog video outputs, including component and RCA outputs, are located on the sides behind small covers. The console also supports Wi-Fi and Bluetooth.

New FPGA-powered retro console re-creates the PlayStation, CD-ROM drive optional Read More »

fcc-chair-helps-isps-and-landlords-make-deals-that-renters-can’t-escape

FCC chair helps ISPs and landlords make deals that renters can’t escape

Lobby groups thank new FCC chair

Housing industry lobby groups praised Carr in a press release issued by the National Multifamily Housing Council (NMHC), National Apartment Association (NAA), and Real Estate Technology and Transformation Center (RETTC). “His decision to withdraw the proposal will ensure that millions of consumers—renters, homeowners and condominium owners—will continue to reap the benefits of bulk billing,” the press release said.

The industry press release claims that bulk billing agreements negotiated between property owners and Internet service providers “typically secur[e] high-speed Internet for renters at rates up to 50 percent lower than standard retail pricing” and remove “barriers to broadband adoption like credit checks, security deposits, equipment rentals, or installation fees.”

“Bulk billing arrangements have made high-speed internet more accessible and affordable for millions of Americans, especially for low-income renters and seniors living in affordable housing,” NMHC President Sharon Wilson Géno said.

While the FCC prohibits deals in which a service provider has the exclusive right to access and serve a building, there are other ways in which competitors can be effectively shut out of buildings. In 2022, the FCC said its existing rules weren’t strong enough and added a ban on exclusive revenue-sharing agreements between landlords and ISPs in multi-tenant buildings. The revenue-sharing ban was approved 4–0, including votes from both Rosenworcel and Carr.

Comcast, Charter, Cox, and cable lobby group NCTA opposed Rosenworcel’s plan for a bulk billing ban, saying that “interfering with the ability of building owners to offer these arrangements to their tenants will result in higher broadband and video prices and other harms for consumers, with questionable and limited benefits.”

Carr issued a statement today, saying, “During the Biden-Harris Administration, FCC leadership put forward a ‘bulk billing’ proposal that could have raised the price of Internet service for Americans living in apartments by as much as 50 percent. This regulatory overreach from Washington would have hit families right in their pocketbooks at a time when they were already hurting from the last administration’s inflationary policies. That is why you saw a broad and bipartisan coalition of groups opposing the plan. After all, seniors, students, and low-income individuals would have been hit particularly hard.” Carr also said that he plans more actions “to reverse the last administration’s costly regulatory overreach.”

FCC chair helps ISPs and landlords make deals that renters can’t escape Read More »

us‘s-wind-and-solar-will-generate-more-power-than-coal-in-2024

US‘s wind and solar will generate more power than coal in 2024

We can expect next year’s numbers to also show a large growth in solar production, as the EIA says that the US saw record levels of new solar installations in 2024, with 37 gigawatts of new capacity. Since some of that came online later in the year, it’ll produce considerably more power next year. And, in its latest short-term energy analysis, the EIA expects to see over 20 GW of solar capacity added in each of the next two years. New wind capacity will push that above 30 GW of renewable capacity each of these years.

A bar chart, with the single largest bar belonging to solar energy.

The past few years of solar installations have led to remarkable growth in its power output. Credit: John Timer

That growth will, it’s expected, more than offset continued growth in demand, although that growth is expected to be somewhat slower than we saw in 2024. It also predicts about 15 GW of coal will be removed from the grid during those two years. So, even without any changes in policy, we’re likely to see a very dynamic grid landscape over the next few years.

But changes in policy are almost certainly on the way. The flurry of executive orders issued by the Trump administration includes a number of energy-related changes. These include defining “energy” in a way that excludes wind and solar, an end to offshore wind leasing and the threat to terminate existing leases, and a re-evaluation of the allocation of funds from some of the Biden administration’s energy-focused laws.

In essence, this sets up a clash among economics, state policies, and federal policy. Even without any subsidies, wind and solar are the cheapest ways to produce electricity in much of the US. In addition, a number of states have mandates that will require the use of more renewable energy. At the same time, the permitting process for the plants and their grid connections will often require approvals at the federal level, and it appears to be official policy to inhibit renewables when possible. And a number of states are also making attempts to block new renewable power installations.

It’s going to be a challenging period for everyone involved in renewable energy.

US‘s wind and solar will generate more power than coal in 2024 Read More »

trump’s-reported-plans-to-save-tiktok-may-violate-scotus-backed-law

Trump’s reported plans to save TikTok may violate SCOTUS-backed law


Everything insiders are saying about Trump’s plan to save TikTok.

It was apparently a busy weekend for key players involved in Donald Trump’s efforts to make a deal to save TikTok.

Perhaps the most appealing option for ByteDance could be if Trump blessed a merger between TikTok and Perplexity AI—a San Francisco-based AI search company worth about $9 billion that appears to view a TikTok video content acquisition as a path to compete with major players like Google and OpenAI.

On Sunday, Perplexity AI submitted a revised merger proposal to TikTok-owner ByteDance, reviewed by CNBC, which sources told AP News included feedback from the Trump administration.

If the plan is approved, Perplexity AI and TikTok US would be merged into a new entity. And once TikTok reaches an initial public offering of at least $300 billion, the US government could own up to 50 percent of that new company, CNBC reported. In the proposal, Perplexity AI suggested that a “fair price” would be “well north of $50 billion,” but the final price will likely depend on how many of TikTok’s existing investors decide to cash out following the merger.

ByteDance has maintained a strong resistance to selling off TikTok, especially a sale including its recommendation algorithm. Not only would this option allow ByteDance to maintain a minority stake in TikTok, but it also would leave TikTok’s recommendation algorithm under ByteDance’s control, CNBC reported. The deal would also “allow for most of ByteDance’s existing investors to retain their equity stakes,” CNBC reported.

But ByteDance may not like one potential part of the deal. An insider source told AP News that ByteDance would be required to allow “full US board control.”

According to AP News, US government ownership of a large stake in TikTok would include checks to ensure the app doesn’t become state controlled. The government’s potential stake would apparently not grant the US voting power or a seat on the merged company’s board.

A source familiar with Perplexity AI’s proposal confirmed to Ars that the reporting from CNBC and AP News is accurate.

Trump denied Oracle’s involvement in talks

Over the weekend, there was also a lot of speculation about Oracle’s involvement in negotiations. NPR reported that two sources with direct knowledge claimed that Trump was considering “tapping software company Oracle and a group of outside investors to effectively take control of the app’s global operations.”

That would be a seemingly bigger grab for the US than forcing ByteDance to divest only TikTok’s US operations.

“The goal is for Oracle to effectively monitor and provide oversight with what is going on with TikTok,” one source told NPR. “ByteDance wouldn’t completely go away, but it would minimize Chinese ownership.”

Oracle apparently met with the Trump administration on Friday and has another meeting scheduled this week to discuss Oracle buying a TikTok stake “in the tens of billions,” NPR reported.

But Trump has disputed that, saying this past weekend that he “never” spoke to Oracle about buying TikTok, AP News reported.

“Numerous people are talking to me. Very substantial people,” Trump said, confirming that he would only make a deal to save TikTok “if the United States benefits.”

All sources seemed to suggest that no deal was close to being finalized yet. Other potential Big Tech buyers include Microsoft or even possibly Elon Musk (can you imagine TikTok merged with X?). On Saturday, Trump suggested that he would likely announce his decision on TikTok’s future in the next 30 days.

Meanwhile, TikTok access has become spotty in the US. Google and Apple dropped TikTok from their app stores when the divest-or-ban law kicked in, partly because of the legal limbo threatening hundreds of billions in fines if Trump changes his mind about enforcement. That means ByteDance currently can’t push updates to US users, and anyone who offloads TikTok or purchases a new device can’t download the app in popular distribution channels.

“If we can save TikTok, I think it would be a good thing,” Trump said.

Could Trump’s plan violate divest-or-ban law?

The divest-or-ban law is formally called the Protecting Americans from Foreign Adversary Controlled Applications Act. For months, TikTok was told in court that the law required either a sale of TikTok US operations or a US ban, but now ByteDance seems to believe there’s another option to keep TikTok in the US without forcing a sale.

It remains unclear if lawmakers will approve Trump’s plan if it doesn’t force a sale of TikTok. US Representative Raja Krishnamoorthi (D-Ill.), who co-sponsored the law, issued a statement last week insisting that “ByteDance divesting remains the only real solution to protect our national security and guarantee Americans access to TikTok.”

Krishnamoorthi declined Ars’ request to comment on whether leaked details of Trump’s potential deal to save TikTok could potentially violate the divest-or-ban law. But debate will likely turn on how the law defines “qualified divestiture.”

Under the law, qualified divestiture could be either a “divestiture or similar transaction” that meets two conditions. First, the transaction is one that Trump “determines, through an interagency process, would result in the relevant foreign adversary controlled application no longer being controlled by a foreign adversary.” Second, the deal blocks any foreign adversary-controlled entity or affiliate from interfering in TikTok US operations, “including any cooperation” with foreign adversaries “with respect to the operation of a content recommendation algorithm or an agreement with respect to data sharing.”

That last bit seems to suggest that lawmakers might clash with Trump over ByteDance controlling TikTok’s algorithm, even if a company like Oracle or Perplexity serves as a gatekeeper to Americans’ data safeguarding US national security interests.

Experts told NPR that ByteDance could feasibly maintain a minority stake in TikTok US under the law, with Trump seeming to have “wide latitude to interpret” what is or is not a qualified divestiture. One congressional staffer told NPR that lawmakers might be won over if the Trump administration secured binding legal agreements “ensuring ByteDance cannot covertly manipulate the app.”

The US has tried to strike just such a national security agreement with ByteDance before, though, and it ended in lawmakers passing the divest-or-ban law. During the government’s court battle with TikTok over the law, the government repeatedly argued that prior agreement—also known as “Project Texas,” which ensured TikTok’s US recommendation engine was stored in the Oracle cloud and deployed in the US by a TikTok US subsidiary—was not enough to block Chinese influence. Proposed in 2022, the agreement was abruptly ended in 2023 when the Committee on Foreign Investment in the United States (CFIUS) determined only divestiture would resolve US concerns.

CFIUS did not respond to Ars’ request for comment.

The key problem at that point was ByteDance maintaining control of the algorithm, the government successfully argued in a case that ended in a Supreme Court victory.

“Even under TikTok’s proposed national security agreement, the source code for the recommendation engine would originate in China,” the government warned.

That seemingly leaves a vulnerability that any Trump deal allowing ByteDance to maintain control of the algorithm would likely have to reconcile.

“Under Chinese national-security laws, the Chinese government can require a China-based company to ‘surrender all its data,'” the US argued. That ultimately turned TikTok into “an espionage tool” for the Chinese Communist Party.

There’s no telling yet if Trump’s plan can set up a better version of Project Texas or convince China to sign off on a TikTok sale. Analysts have suggested that China may agree to a TikTok sale if Trump backs down on tariff threats.

ByteDance did not respond to Ars’ request for comment.

Photo of Ashley Belanger

Ashley is a senior policy reporter for Ars Technica, dedicated to tracking social impacts of emerging policies and new technologies. She is a Chicago-based journalist with 20 years of experience.

Trump’s reported plans to save TikTok may violate SCOTUS-backed law Read More »

a-long,-costly-road-ahead-for-customers-abandoning-broadcom’s-vmware

A long, costly road ahead for customers abandoning Broadcom’s VMware


“We loved VMware, and then when Broadcom bought ‘em, we hated ‘em.”

Broadcom’s ownership of VMware has discouraged many of its customers, as companies are displeased with how the trillion-dollar firm has run the virtualization business since buying it in November 2023. Many have discussed reducing or eliminating ties with the company.

Now, over a year after the acquisition, the pressure is on for customers to start committing to a VMware subscription, forego VMware support, or move on from VMware technologies. The decision is complex, with long-term implications no matter which way a customer goes.

Ars Technica spoke with an IT vendor manager who has been using VMware’s vSphere since the early 2000s. The employee, who works for a global food manufacturing firm with about 5,500 employees, asked to keep their name and company anonymous due to privacy concerns for the business.

“We love it. … It’s hard for us to figure out how we can live without it, but we’re going to,” the IT manager said.

The food manufacturer has about 300 VMware virtual machines (VMs), and every company application runs on top of VMware. Its five-year enterprise agreement with VMware expired in December, making the manufacturer ineligible for VMware support unless it buys a VMware subscription. The company started exploring virtualization alternatives this summer because costs associated with running vSphere are set to rise fourfold, according to the IT manager. As with other VMware customers, the price increases are largely due to Broadcom bundling unwanted VMware products together.

“They wouldn’t sell us what we need,” the IT manager said.

While it looks for a new platform, the manufacturer is relying on support from Spinnaker, which started offering software maintenance support for VMware following Broadcom’s acquisition. In an example of how widespread VMware support concerns are, Spinnaker’s VMware support business has had more leads than any of Spinnaker’s other support businesses, including for Oracle or SAP, said Martin Biggs, Spinnaker’s VP and managing director of strategic initiatives and EMEA.

Organizations contacting Spinnaker are reporting price increases of “3–6x” on average, Biggs told Ars. The largest price rise Spinnaker has heard about is a reported twentyfold increase in costs, he said.

Biggs said that Broadcom has started to discount some subscriptions, with price increases going from seven- or eightfold to three- or fourfold, or “sometimes a little bit less.” This could pressure customers to commit to VMware while terms are more favorable than they might be in the future. Speaking to The Register this month, Gartner VP analyst Michael Warrilow said he feared Broadcom would raise VMware prices higher in the future.

Heightening the potential consequences associated with staying with or leaving VMware, Warrilow emphasized that Broadcom prefers two- or three-year subscriptions, meaning customers may find themselves facing a more pricey VMware sooner than later.

“Everybody’s asking what everybody else is doing, and everybody else is asking what everybody else is doing, so nobody’s really doing anything,” he said.

The Register also recently reported that customers are being pressured into three-year long VMware subscriptions, citing an unnamed VMware customer that it spoke with and a discussion on Reddit. When reached for comment, Broadcom only referred The Register to a June blog post by Broadcom CEO Hock Tan about evolving VMware strategy.

Losing support

Support is a critical factor for numerous customers considering migrating from VMware, especially because VMware perpetual licenses are no longer being sold or supported by Broadcom. But there’s also concern about support offered to clients with subscriptions.

For the food manufacturer currently researching VMware rivals, a perceived lack of support under Broadcom was also a deterrent. The company’s IT manager said that after Broadcom bought VMware, the manufacturer was no longer able to contact VMware directly for support and was told in July that it should direct problems to IT distributor Ingram Micro moving forward.

The manager said this information was relayed to the customer after a support ticket it filed was automatically moved to Ingram, with Broadcom telling the firm it wasn’t big enough to receive direct support. Ingram’s response times were a week or longer, and in December, Ingram announced a severe reduction of its VMware business (VMware still works with other distributors, like Arrow).

Support concerns from VMware resellers started before Ingram’s announcement, though. An anonymous reseller, for example, told CRN that it had to wait a month on average for VMware quotes through a distributor, compared to “two to three days” pre-Broadcom. The Register, citing VMware customers, also reported that Ingram was having difficulties handling “the increased responsibilities it assumed,” citing VMware customers.

Migration is burdensome

In a January Gartner research note entitled “Estimating a Large-Scale VMware,” Gartner analysts detailed the burdens expected for large-sized companies moving off of VMware. The note defined a large-scale migration as a “concerted program of work covering the migration of a significant portion of virtualized workloads” that “would likely represent 2,000 or more” VMs, “and/or at least 100 hosts.” That’s a much larger migration than the food manufacturer’s 300 VMs, but Gartner’s analysis helps illustrate the magnitude of work associated with migrating.

Gartner’s note estimated that large-scale migrations, including scoping and technical evaluation, would take 18 to 48 months. The analysts noted that they “expect a midsize enterprise would take at least two years to untangle much of its dependency upon VMware’s server virtualization platform.”

The analysts also estimated migration to cost $300 to $3,000 per VM if the user employed a third-party service provider. Critically, the report adds:

It is highly likely that other costs would be incurred in a large-scale migration. This includes acquisition of new software licenses and/or cloud expenses, hardware purchases (compute, storage), early termination costs related to the existing virtual environment, application testing/quality assurance, and test equipment.

The heavy costs—in terms of finances, time, and staff—force customers to face questions and hesitations around leaving VMware, despite many customers facing disruption from Broadcom-issued changes to the platform.

When asked if there’s anything Broadcom could do to win back the food manufacturer’s 300 VMs, its IT manager told Ars that if Broadcom offered a subscription to vSphere alone, the manufacturer would reconsider, even if subscription costs were twice as expensive as before.

For the global food manufacturer, the biggest challenge in ditching VMware is internal, not technical. “We just don’t have enough internal resources and timing,” the manager said. “That’s what I’m worried about. This is going to take a lot of time internally to go through this whole process, and we’re shorthanded as it is. It’s such a big, heavy lift for us, and we’re also very risk averse, so swapping out that piece of technology in our infrastructure is risky.”

Stuck between a rock and a hard place

VMware users are now at a crossroads as they’re forced to make crucial decisions for their IT infrastructure. Ditching or sticking with VMware both have long-lasting implications; migrations are onerous and pricey, but life under Broadcom will be expensive, with potential future bumps and twists.

Broadcom has previously responded to Ars’ and others’ requests for comment around customer complaints with blog posts from Broadcom’s Tan that emphasize commitment to VMware’s strategic changes. But some will brave costly challenges to avoid those moves. Summarizing their take on Broadcom’s changes, the food manufacturer’s IT executive said, “We loved VMware. And then when Broadcom bought ’em, we hated ’em.”

Photo of Scharon Harding

Scharon is a Senior Technology Reporter at Ars Technica writing news, reviews, and analysis on consumer gadgets and services. She’s been reporting on technology for over 10 years, with bylines at Tom’s Hardware, Channelnomics, and CRN UK.

A long, costly road ahead for customers abandoning Broadcom’s VMware Read More »

deepseek-panic-at-the-app-store

DeepSeek Panic at the App Store

DeepSeek released v3. Market didn’t react.

DeepSeek released r1. Market didn’t react.

DeepSeek released a fing app of its website. Market said I have an idea, let’s panic.

Nvidia was down 11%, Nasdaq is down 2.5%, S&P is down 1.7%, on the news.

Shakeel: The fact this is happening today, and didn’t happen when r1 actually released last Wednesday, is a neat demonstration of how the market is in fact not efficient at all.

That is exactly the market’s level of situational awareness. No more, no less.

I traded accordingly. But of course nothing here is ever investment advice.

Given all that has happened, it seems worthwhile to go over all the DeepSeek news that has happened since Thursday. Yes, since Thursday.

For previous events, see my top level post here, and additional notes on Thursday.

To avoid confusion: r1 is clearly a pretty great model. It is the best by far available at its price point, and by far the best open model of any kind. I am currently using it for a large percentage of my AI queries.

  1. Current Mood.

  2. DeepSeek Tops the Charts.

  3. Why Is DeepSeek Topping the Charts?.

  4. What Is the DeepSeek Business Model?.

  5. The Lines on Graphs Case for Panic.

  6. Everyone Calm Down About That $5.5 Million Number.

  7. Is The Whale Lying?.

  8. Capex Spending on Compute Will Continue to Go Up.

  9. Jevon’s Paradox Strikes Again.

  10. Okay, Maybe Meta Should Panic.

  11. Are You Short the Market.

  12. o1 Versus r1.

  13. Additional Notes on v3 and r1.

  14. Janus-Pro-7B Sure Why Not.

  15. Man in the Arena.

  16. Training r1, and Training With r1.

  17. Also Perhaps We Should Worry About AI Killing Everyone.

  18. And We Should Worry About Crazy Reactions To All This, Too.

  19. The Lighter Side.

Joe Weisenthal: Call me a nationalist or whatever. But I hope that the AI that turns me into a paperclip is American made.

Peter Wildeford: Seeing everyone lose their minds about Deepseek does not reassure me that we will handle AI progress well.

Miles Brundage: I need the serenity to accept the bad DeepSeek takes I cannot change.

[Here is his One Correct Take, I largely but not entirely agree with it, my biggest disagreement is I am worried about an overly jingoist reaction and not only about us foolishly abandoning export controls].

Satya Nadella (CEO Microsoft): Jevons paradox strikes again! As AI gets more efficient and accessible, we will see its use skyrocket, turning it into a commodity we just can’t get enough of.

Danielle Fong: everyone today: if you’re in “we’re so back” pivot to “it’s over”

Danielle Fong, a few hours later: if you’re in “it’s so over” pivot to “jevons paradox”

Kai-Fu Lee: In my book AI Superpowers, I predicted that US will lead breakthroughs, but China will be better and faster in engineering. Many people simplified that to be “China will beat US.” And many claimed I was wrong with GenAI. With the recent DeepSeek releases, I feel vindicated.

Dean Ball: Being an AI policy professional this week has felt like playing competitive Starcraft.

Lots of people are rushing to download the DeepSeek app.

Some of us started using r1 before the app. Joe Weisenthal noted he had ‘become a DeepSeek bro’ and that this happened overnight, switching costs are basically zero. They’re not as zero as they might look, and I expect the lockin with Operator from OpenAI to start mattering soon, but for most purposes yeah, you can just switch, and DeepSeek is free for conversational use including r1.

Switching costs are even closer to zero if, like most people, you weren’t a serious user of LLMs yet.

Then regular people started to notice DeepSeek.

This is what it looked like before the app shot to #1, when it merely cracked the top 10:

Ken: It’s insane the extent to which the DeepSeek News has broken “the containment zone.” I saw a Brooklyn-based Netflix comedian post about “how embarrassing it was that the colonial devils spent $10 billion, while all they needed was GRPO.”

llm news has arrived as a key political touchstone. will only heighten from here.

Olivia Moore: DeepSeek’s mobile app has entered the top 10 of the U.S. App Store.

It’s getting ~300k global daily downloads.

This may be the first non-GPT based assistant to get mainstream U.S. usage. Claude has not cracked the top 200.

This may be the first non-GPT based assistant to get mainstream U.S. usage.

The app was released on Jan. 11, and is linked on DeepSeek’s website (so does appear to be affiliated).

Per reviews, users are missing some ChatGPT features like voice mode…but basically see it as a free version of OpenAI’s premium models.

Google Gemini also cracked the top 10, in its first week after release (but with a big distribution advantage!)

Will be interesting to see how high DeepSeek climbs, and how long it stays up there 🤔

Claude had ~300k downloads last month, but that’s a lot less than 300k per day.

Metaschool: Google Trends: DeepSeek vs. Claude

Kalomaze: Holy s, it’s in the top 10?

Then it went all the way to #1 on the iPhone app store.

Kevin Xu: Two weeks ago, RedNote topped the download chart

Today, it’s DeepSeek

We are still in January

If constraint is the mother of invention, then collective ignorance is the mother of many downloads

Here’s his flashback to the chart when RedNote was briefly #1, note how fickle the top listings can be, Lemon8, Flip and Clapper were there too:

The ‘collective ignorance’ here is that news about DeepSeek and the app is only arriving now. That leads to a lot of downloads.

I have a Pixel 9, so I checked the Android app store. They have Temu at #1 (also Chinese!) followed by Scoopz which I literally have never heard of, then Instagram, T-Life (seriously what?), ReelShort, WhatsApp Messenger, ChatGPT (interesting that Android users are less AI pilled in general), Easy Homescreen (huh), TurboTax (oh no), Snapchat and then DeepSeek at #11. So if they’ve ‘saturated the benchmark’ on iPhone, this one is next, I suppose.

It seems DeepSeek got so many downloads they had to hit the breaks, similar to how OpenAI and Anthropic have had to do this in the past.

Joe Weisenthal: *DEEPSEEK: RESTRICTS REGISTRATION TO CHINA MOBILE PHONE NUMBERS

Because:

  1. It’s completely free.

  2. It has no ads.

  3. It’s a damn good model, sir.

  4. It lets you see the chain of thought which is a lot more interesting and fun and also inspires trust.

  5. All the panic about it only helped people notice, getting it on the news and so on.

  6. It’s the New Hotness that people hadn’t downloaded before, and that everyone is talking about right now because see the first five.

  7. No, this mostly isn’t about ‘people don’t trust American tech companies but they do trust the Chinese.’ But there aren’t zero people who are wrong enough to think this way, and China actively attempts to cultivate this including through TikTok.

  8. The Open Source people are also yelling about how this is so awesome and trustworthy and virtuous and so on, and being even more obnoxious than usual, which may or may not be making any meaningful difference.

I suspect we shouldn’t be underestimating the value of showing the CoT here, as I also discuss elsewhere in the post.

Garry Tan: DeepSeek search feels more sticky even after a few queries because seeing the reasoning (even how earnest it is about what it knows and what it might not know) increases user trust by quite a lot

Nabeel Qureshi: I wouldn’t be surprised if OpenAI starts showing CoTs too; it’s a much better user experience to see what the machine is thinking, and the rationale for keeping them secret feels weaker now that the cat’s out of the bag anyway.

It’s just way more satisfying to watch this happen.

It’s practically useful too: if the model’s going off in wrong directions or misinterpreting the request, you can tell sooner and rewrite the prompt.

That doesn’t mean it is ‘worth’ sharing the CoT, even if it adds a lot of value – it also reveals a lot of valuable information, including as part of training another model. So the answer isn’t obvious.

What’s their motivation?

Meta is pursuing open weights primarily because they believe it maximizes shareholder value. DeepSeek seems to be doing it primarily for other reasons.

Corey Gwin: There’s gotta be a catch… What did China do or hide in it? Will someone release a non-censored training set?

Amjad Masad: What’s Meta’s catch with Llama? Probably have similar incentives.

Anton: How is Deepseek going to make money?

If they just release their top model weights, why use their API?

Mistral did this and look where they are now (research licenses only and private models)

Han Xiao: deepseek’s holding 幻方量化 is a quant company, many years already,super smart guys with top math background; happened to own a lot GPU for trading/mining purpose, and deepseek is their side project for squeezing those gpus.

It’s an odd thing to do as a hedge fund, to create something immensely valuable and give it away for essentially ideological reasons. But that seems to be happening.

Several possibilities. The most obvious ones are, in some combination:

  1. They don’t need a business model. They’re idealists looking to give everyone AGI.

  2. They’ll pivot to the standard business model same as everyone else.

  3. They’re in it for the prestige, they’ll recruit great engineers and traders and everyone will want to invest capital.

  4. Get people to use v3 and r1, collect the data on what they’re saying and asking, use that information as the hedge fund to trade. Being open means they miss out on some of the traffic but a lot of it will still go to the source anyway if they make it free, or simply because it’s easier.

  5. (They’re doing this because China wants them to, or they’re patriots, perhaps.)

  6. Or just: We’ll figure out something.

For now, they are emphasizing motivation #1. From where I sit, there is very broad uncertainty about which of these dominate, or will dominate in the future no matter what they believe about themselves today.

Also, there are those who do not approve of motivation #1, and the CCP seems plausibly on that list. Thus, Tyler Cowen asks a very good question that is surprisingly rarely asked right now.

Tyler Cowen: DeepSeek okie-dokie: “All I know is we keep pushing forward to make open-source AGI a reality for everyone.” I believe them, the question is what counter-move the CCP will make now.

I also believe they intend to build and open source AGI.

The CCP is doubtless all for DeepSeek having a hit app. And they’ve been happy to support open source in places where open source doesn’t pose existential risks, because the upsides of doing that are very real.

That’s very different from an intent to open source AGI. China’s strategy on AI regulation so far has focused on content moderation for topics they care about. That approach won’t stay compatible with their objectives over time.

For that future intention to open source AGI, the question is not ‘how move will the CCP make to help them do this and get them funding and chips?’

The question now becomes: “What countermove will the CCP make now?”

The CCP wants to stay in control. What DeepSeek is doing is incompatible with that. If they are not simply asleep at the wheel, they understand this. Yes, it’s great for prestige, and they’re thrilled that if this model exists it came from China, but they will surely notice how if you run it on your own it’s impossible to control and fully uncensored out of the box and so on.

Might want to Pick Up the Phone. Also might not need to.

Yishan takes the opposite perspective, that newcomers like DeepSeek who come out with killer products like this are on steep upward trajectories and their next product will shock you with how good it is, seeing it as similar to Internet Explorer 3 or Firefox, or iPhone 1 or early Facebook or Google Docs or GPT-3 or early SpaceX and so on.

I think the example list here illustrates why I think DeepSeek probably (but not definitely) doesn’t belong on that list. Yishan notes that the incumbents here are dynamic and investing hard, which wasn’t true in most of the other examples. And many of them involve conceptually innovative approaches to go with the stuck incumbents. Again, that’s not the case here.

I mean, I fully expect there to be a v4 and r2 some time in 2025, and for those to blow out of the water v3 and r1 and probably the other models that are released right now. Sure. But I also expect OpenAI and Anthropic and Google to blow the current class of stuff out of the water by year’s end. Indeed, OpenAI is set to do this in about a week or two with o3-mini and then o3 and o3-pro.

Most of all, to those who are saying that ‘China has won’ or ‘China is in the lead now,’ or other similar things, seriously, calm the down.

Yishan: They are already working on the next thing. China may reach AGI first, which is a bogeyman for the West, except that the practical effect will probably just be that living in China starts getting really nice.

America, it ain’t the Chinese girl spies here you gotta worry about, you need to be flipping the game and sending pretty white girls over there to seduce their engineers and steal their secrets, stat.

If you’re serious about the steal the engineering secrets plan, of course, you’d want to send over a pretty white girl… with a green card with the engineer’s name on it. And the pretty, white and girl parts are then all optional. But no, China isn’t suddenly the one with the engineering secrets.

I worry about this because I worry about a jingoist ‘we must beat China and we are behind’ reaction causing the government to do some crazy ass stuff that makes us all much more likely to get ourselves killed, above and beyond what has already happened. There’s a lot of very strong Missile Gap vibes here.

And I wrote that sentence before DeepSeek went to #1 on the app store and there was a $1 trillion market panic. Oh no.

So, first off, let’s all calm down about that $5.5 million training number.

Dean Ball offers notes on DeepSeek and r1 in the hopes of calming people down. Because we have such different policy positions yet see this situation so similarly, I’m going to quote him in full, and then note the places I disagree. Especially notes #2, #5 and #4 here, yes all those claims he is pointing out are Obvious Nonsense are indeed Obvious Nonsense:

Dean Ball: The amount of factually incorrect information and hyperventilating takes on deepseek on this website is truly astounding. I assumed that an object-level analysis was unnecessary but apparently I was wrong. Here you go:

  1. DeepSeek is an extremely talented team and has been producing some of the most interesting public papers in ML for a year. I first wrote about them in May 2024, though was tracking them earlier. They did not “come out of nowhere,” at all.

  2. v3 and r1 are impressive models. v3 did not, however, “cost $5m.” That reported figure is almost surely their *marginalcost. It does not include the fixed cost of building a cluster (and deepseek builds their own, from what I understand), nor does it include the cost of having a staff.

  3. Part of the reason DeepSeek looks so impressive (apart from just being impressive!) is that they are among the only truly cracked teams releasing detailed frontier AI research. This is a soft power loss on America’s part, and is directly downstream of the culture of secrecy that we foster in a thousand implicit and explicit ways, including by ceaselessly analogizing AI to nuclear weapons. Maybe you believe that’s a good culture to have! Perhaps secrecy is in fact the correct long term strategy. But it is the obvious and inevitable tradeoff of such a culture; I and many others have been arguing this for a long time.

  4. Deepseek’s r1 is not an indicator that export controls are failing (again, I say this as a skeptic of the export controls!), nor is it an indicator that “compute doesn’t matter,” nor does it mean “America’s lead is over.”

  5. Lots of people’s hyperbolic commentary on this topic, in all different directions, is driven by their broader policy agenda rather than a desire to illuminate reality. Caveat emptor.

  6. With that said, DeepSeek does mean that open source AI is going to be an important part of AI dynamics and competition for at least the foreseeable future, and probably forever.

  7. r1 especially should not be a surprise (if anything, v3 is in fact the bigger surprise, though it too is not so big of a surprise). The reasoning approach is an algorithm—lines of code! There is no moat in such things. Obviously it was going to be replicated quickly. I personally made bets that a Chinese replication would occur within 3 months of o1’s release.

  8. Competition is going to be fierce, and complacency is our enemy. So is getting regulation wrong. We need to reverse course rapidly from the torrent of state-based regulation that is coming that will be *awfulfor AI. A simple federal law can preempt all of the most damaging stuff, and this is a national security and economic competitiveness priority. The second best option is to find a state law that can serve as a light touch national standard and see to it that it becomes a nationwide standard. Both are exceptionally difficult paths to walk. Unfortunately it’s where we are.

I fully agree with #1 through #6.

For #3 I would say it is downstream of our insane immigration policies! If we let their best and brightest come here, then DeepSeek wouldn’t have been so cracked. And I would say strongly that, while their release of the model and paper is a ‘soft power’ reputational win, I don’t think that was worth the information they gave up, and in purely strategic terms they made a rather serious mistake.

I can verify the bet in #7 was very on point, I wasn’t on either side of the wager but was in the (virtual) Room Where It Happened. Definite Bayes points to Dean for that wager. I agree that ‘reasoning model at all, in time’ was inevitable. But I don’t think you should have expected r1 to come out this fast and be this good, given what we knew at the time of o1’s release, and certainly it shouldn’t have been obvious, and I think ‘there are no moats’ is too strong.

For #8 we of course have our differences on regulation, but we do agree on a lot of this. Dean doubtless would count a lot more things as ‘awful state laws’ than I would, but we agree that the proposed Texas law would count. At this point, given what we’ve seen from the Trump administration, I think our best bet is the state law path. As for pre-emption, OpenAI is actively trying to get an all-encompassing version of that in exchange for essentially nothing at all, and win an entirely free hand, as I’ve previously noted. We can’t let that happen.

Seriously, though, do not over index on the $5.5 million in compute number.

Kevin Roose: It’s sort of funny that every American tech company is bragging about how much money they’re spending to build their models, and DeepSeek is just like “yeah we got there with $47 and a refurbished Chromebook”

Nabeel Qureshi: Everyone is way overindexing on the $5.5m final training run number from DeepSeek.

– GPU capex probably $1BN+

– Running costs are probably $X00M+/year

– ~150 top-tier authors on the v3 technical paper, $50m+/year

They’re not some ragtag outfit, this was a huge operation.

Nathan Lambert has a good run-down of the actual costs here.

I have no idea if the “we’re just a hedge fund with a lot of GPUs lying around” thing is really the whole story or not but with a budget of _that_ size, you have to wonder…

They themselves sort of point this out, but there’s a bunch of broader costs too.

The Thielian point here is that the best salespeople often don’t look like salespeople.

There’s clearly an angle here with the whole “we’re way more efficient than you guys”, all described in the driest technical language….

Nathan Lambert: These costs are not necessarily all borne directly by DeepSeek, i.e. they could be working with a cloud provider, but their cost on compute alone (before anything like electricity) is at least $100M’s per year.

For one example, consider comparing how the DeepSeek V3 paper has 139 technical authors. This is a very large technical team.With headcount costs that can also easily be over $10M per year, estimating the cost of a year of operations for DeepSeek AI would be closer to $500M (or even $1B+) than any of the $5.5M numbers tossed around for this model. The success here is that they’re relevant among American technology companies spending what is approaching or surpassing $10B per year on AI models.

Richard Song: Every AI company after DeepSeek be like:

Danielle Fong: when tesla claimed that they were going to have batteries < $100 / kWh, practically all funding for american energy storage companies tanked.

tesla still won’t sell you a powerwall or powerpack for $100/kWh. it’s like $1000/kWh and $500 for a megapack.

the entire VC sector in the US was bluffed and spooked by Elon. don’t be stupid in this way again.

What I’m saying here is that VCs need to invest in technology learning curves. things get better over time. but if you’re going to compare what your little startup can get out as an MVP in its first X years, and are comparing THAT projecting forward to what a refined tech can do in a decade, you’re going to scare yourself out of making any investments. you need to find a niche you can get out and grow in, and then expand successively as you come down the learning curve.

the AI labs that are trashing their own teams and going with deepseek are doing the equivalent today. don’t get bluffed. build yourself.

Is it impressive that they (presumably) did the final training run with only $5.5M in direct compute costs? Absolutely. Is it impressive that they’re relevant while plausibly spending only hundreds of millions per year total instead of tens of billions? Damn straight. They’re cracked and they cooked.

They didn’t do it with $47 on a Chromebook, and this doesn’t mean that export controls are useless because everyone can buy a Chromebook.

The above is assuming (as I do still assume) that Alexandr Wang was wrong when he went on CNBC and claimed DeepSeek has about 50,000 H100s, which is quite the claim to make without evidence. Elon Musk replied to this claim with ‘obviously.’

Samuel Hammond also is claiming that DeepSeek trained on H100s, and while my current belief is that they didn’t, I trust that he would not say it if he didn’t believe it.

Neal Khosla went so far as to claim (again without evidence) that ‘deepseek is a ccp psyop + economic warfare to make American AI unprofitable.’ This seems false.

The following all seem clearly true:

  1. A lot of this is based on misunderstanding the ‘$5.5 million’ number.

  2. People have strong motive to engage in baseless cope around DeepSeek.

  3. DeepSeek had strong motive to lie about its training costs and methods.

So how likely is it The Whale Is Lying?

Armen Aghajanyan: There is an unprecedented level of cope around DeepSeek, and very little signal on X around R1. I recommend unfollowing anyone spreading conspiracy theories around R1/DeepSeek in general.

Teortaxes: btw people with major platforms who spread the 50K H100s conspiracy theory are underestimating the long-term reputation cost in technically literate circles. They will *notbe able to solidify this nonsense into consensus reality. Instead, they’ll be recognized as frauds.

The current go-to best estimate for DeepSeek V3’s (and accordingly R1-base’s) pretraining compute/cost, complete with accounting for overhead introduced by their architecture choices and optimizations to mitigate that.

TL;DR: ofc it checks out, Whale Will Never Lie To Us

GFodor: I shudder at the thought I’ve ever posted anything as stupid as these theories, given the logical consequence it would demand of the reader

Amjad Masad: So much cope about DeepSeek.

Not only did they release a great model. they also released a breakthrough training method (R1 Zero) that’s already reproducing.

I doubt they lied about training costs, but even if they did they’re still awesome for this great gift to the world.

This is an uncharacteristically naive take from Teortaxes on two fronts.

  1. Saying an AI company would never lie to us, Chinese or otherwise, someone please queue the laugh track.

  2. Making even provably and very clearly false claims about AI does not get you recognized as a fraud in any meaningful way. That would be nice, but no.

To be clear, my position is close to Masad’s: Unless and until I see more convincing evidence I will continue to believe that yes, they did do the training run itself with the H800s for only $5.5 million, although the full actual cost was orders of magnitude more than that. Which, again, is damn impressive, and would be damn impressive even if they were fudging the costs quite a bit beyond that.

Whereas here I think he’s wrong is in their motivation. While Meta is doing this primarily because they believe it maximizes shareholder value, DeepSeek seems to be doing it primarily for other reasons, as noted in the section asking about their business model.

Either way, they are very importantly being constrained by access to compute, even if they’ve smuggled in a bunch of chips they can’t talk about. As Tim Fist points out, the export controls are tightened, so they’ll have more trouble accessing the next generations than they are having now, and no this did not stop being relevant, and they risk falling rather far behind.

Also Peter Wildeford points out that the American capex spends on AI will continue to go up. DeepSeek is cracked and cooking and cool, and yes they’ve proven you can do a lot more with less than we expected, but keeping up is going to be tough unless they get a lot more funding some other way. Which China is totally capable of doing, and may well do. That would bring the focus back on export controls.

Similarly, here’s Samuel Hammond.

Angela Zhang (Hong Kong): My latest opinion on how Deepseek’s rise has laid bare the limits of US export controls designed to slow China’s AI progress.

Samuel Hammond: This is wrong on several levels.

– DeepSeek trains on h100s. Their success reveals the need to invest in export control *enforcementcapacity.

– CoT / inference-time techniques make access to large amounts of compute *morerelevant, not less, given the trillions of tokens generated for post-training.

– We’re barely one new chip generation into the export controls, so it’s not surprising China “caught up.” The controls will only really start to bind and drive a delta in the US-China frontier this year and next.

– DeepSeek’s CEO has himself said the chip controls are their biggest blocker.

– The export controls also apply to semiconductor manufacturing equipment, not just chips, and have tangibly set back SMIC.

DeepSeek is not a Sputnik moment. Their models are impressive but within the envelope of what an informed observer should expect.

Imagine if US policymakers responded to the actual Sputnik moment by throwing their hands in the air and saying, “ah well, might as well remove the export controls on our satellite tech.” Would be a complete non-sequitur.

Roon: If the frontier models are commoditized, compute concentration matters even more.

If you can train better models for fewer floating-point operations, compute concentration matters even more.

Compute is the primary means of production of the future, and owning more will always be good.

In my opinion, open-source models are a bit of a red herring on the path to acceptable ASI futures. Free model weights still do not distribute power to all of humanity; they distribute it to the compute-rich.

I don’t think Roon is right that it matters ‘even more,’ and I think who has what access to the best models for what purposes is very much not a red herring, but compute definitely still matters a lot in every scenario that involves strong AI.

Imagine if the ones going ‘I suppose we should drop the export controls then’ or ‘the export controls only made us stronger’ were mostly the ones looking to do the importing and exporting. Oh, right.

And yes, the Chinese are working hard to make their own chips, but:

  1. They’re already doing this as much as possible, and doing less export controls wouldn’t suddenly get them to slow down and do it less, regardless of how successful you think they are being.

  2. Every chip we sell to them instead of us is us being an idiot.

  3. DeepSeek trained on Nvidia chips like everyone else.

The question now turns to what all of this means for American equities.

In particular, what does this mean for Nvidia?

BuccoCapital Bloke: My entire fing Twitter feed this weekend:

He leaned back in his chair. Confidently, he peered over the brim of his glasses and said, with an air of condescension, “Any fool can see that DeepSeek is bad for Nvidia”

“Perhaps” mused his adversary. He had that condescending bastard right where he wanted him. “Unless you consider…Jevons Paradox!”

All color drained from the confident man’s face. His now-trembling hands reached for his glasses. How could he have forgotten Jevons Paradox! Imbecile! He wanted to vomit.

Satya Nadella (CEO Microsoft): Jevons paradox strikes again! As AI gets more efficient and accessible, we will see its use skyrocket, turning it into a commodity we just can’t get enough of.

Adam D’Angelo (Board of OpenAI, among many others):

Sarah (YuanYuanSunSara): Until you have good enough agent that runs autonomously with no individual human supervision, not sure this is true. If model gets so efficient that you can run it on everyone’s laptop (which deepseek does have a 1B model), unclear whether you need more GPU.

DeepSeek is definitely not at ‘run on your laptop’ level, and these are reasoning models so when we first crack AGI or otherwise want the best results I am confident you will want to be using some GPUs or other high powered hardware, even if lots of other AI also is happening locally.

Does Jevon’s Paradox (which is not really a paradox at all, but hey) apply here to Nvidia in particular? Will improvements in the quality of cheaper open models drive demand for Nvidia GPUs up or down?

I believe it will on net drive demand up rather than down, although I also think Nvidia would have been able to sell as many chips as it can produce either way, given the way it has decided to set prices.

If I am Meta or Microsoft or Amazon or OpenAI or Google or xAI and so on, I want as many GPUs as I can get my hands on, even more than before. I want to be scaling. Even if I don’t need to scale for pretraining, I’ll still want to scale for inference. If the best models are somehow going to be this cheap to serve, uses and demand will be off the charts. And getting there first, via having more compute to do the research, will be one of the few things that matters.

You could reach the opposite conclusion if you think that there is a rapidly approaching limit to how good AI can be, that throwing more compute an training or inference won’t improve that by much, there’s a fixed set of things you would thus use AI for, and thus all this does is drive the price cheaper, maybe open up a few marginal use cases as the economics improve. That’s a view that doesn’t believe in AGI, let alone ASI, and likely doesn’t even factor in what current models (including r1!) can already do.

If all we had was r1 for 10 years, oh the Nvidia chips we would buy to do inference.

Or at least, if you’re in their GenAI department, you should definitely panic.

Here is a claim seen on Twitter from many sources:

Meta GenAI organization in panic mode

It started with DeepSeek v3, which rendered Llama 4 already behind in benchmarks. Adding insult to injury was the “unknown Chinese company with a $5.5 million training budget.”

Engineers are moving frantically to dissect DeepSeek and copy anything and everything they can from it. I’m not even exaggerating.

Management is worried about justifying the massive cost of the GenAI organization. How would they face leadership when every single “leader” of the GenAI organization is making more than what it cost to train DeepSeek v3 entirely, and they have dozens of such “leaders”?

DeepSeek r1 made things even scarier. I cannot reveal confidential information, but it will be public soon anyway.

It should have been an engineering-focused small organization, but since a bunch of people wanted to join for the impact and artificially inflate hiring in the organization, everyone loses.

Shakeel: I can explain this. It’s because Meta isn’t very good at developing AI models.

Full version is in The Information, saying that this is already better than Llama 4 (seems likely) and that Meta has ‘set up four war rooms.’

This of course puts too much emphasis on the $5.5 million number as discussed above, but the point remains that DeepSeek is eating Meta’s lunch in particular. If Meta’s GenAI team isn’t in panic mode, they should all be fired.

It also illustrates why DeepSeek may have made a major mistake revealing as much information as it did, but then again if they’re not trying to make money and instead are driven by ideology of ‘get everyone killed’ (sorry I meant to say ‘open source AGI’) then that is a different calculus than Meta’s.

But obviously what Meta should be doing right now is, among other things, ask ‘what if we trained the same way as v3 and r1 except we use $5.5 billion in compute instead of $5.5 million.’)

That is exactly Meta’s speciality. Llama was all about ‘we hear you like LLMs so we trained an LLM the way everyone trains their LLMs.’

The alternative is ‘maybe we should focus our compute on inference and use local fine-tuned versions of these sweet open models,’ but Zuckerberg very clearly is unwilling to depends on anyone else for that, and I do not blame him.

If you were short on Friday, you’re rather happy about that now. Does it make sense?

The timing is telling. To the extent this does have impact, all of this really should have been mostly priced in. You can try to tell the ‘it was priced in’ story, but I don’t believe you. Or you can tell the story that what wasn’t priced in was the app, and the mindshare, and that wasn’t definite until just now. Remember the app was launched weeks ago, so this isn’t a revelation about DeepSeek’s business plans – but it does give them the opportunity to potentially launch various commercial products, and it gives them mindshare.

But don’t worry about the timing, and don’t worry about whether this is actually a response to the fing app. Ask about what the real implications are.

Joe Weisenthal has a post with 17 thoughts about the selloff (ungated Twitter screenshots here).

There are obvious reasons to think this is rather terrible for OpenAI in particular, although it isn’t publicly traded, because a direct competitor is suddenly putting up some very stiff new competition, and also the price of entry for other competition just cratered, and more companies could self-host or even self-train.

I totally buy that. If every Fortune 500 company can train their private company-specific reasoning model for under $10 million, to their own specifications, why wouldn’t they? The answer is ‘because it doesn’t actually cost that little even with the DeepSeek paper, and if you do that you’ll always be behind,’ but yes some of them will choose to do that.

That same logic goes for other frontier labs like Anthropic or xAI, and to Google and Microsoft and everyone else to the extent that is what those companies are this or own shares in this, which by market cap is not that much.

The flip side of course is that they too can make use of all these techniques, and if AGI is now going to happen a lot faster and more impactfully, these labs are in prime position. But if the market was respecting being in prime position for AGI properly prices would look very different.

This is obviously potentially bad for Meta, since Meta’s plan involved being the leader in open models and they’ve been informed they’re not the leader in open models.

In general, Chinese competition looking stiffer for various products is bad in various ways for a variety of American equities. Some decline in various places is appropriate.

This is obviously bad for existential risk, but I have not seen anyone else even joke about the idea that this could be explaining the decline in the market. The market does not care or think about existential risk, at all, as I’ve discussed repeatedly. Market prices are neither evidence for, nor against, existential risk on any timelines that are not on the order of weeks, nor are they at all situationally aware. Nor is there a good way to exploit this to make money that is better than using your situational awareness to make money in other ways. Stop it!

My diagnosis is that this is about, fundamentally, ‘the vibes.’ It’s about Joe’s sixth point and investor MOMO and FOMO.

As in, previously investors bought Nvidia and friends because of:

  1. Strong earnings and other fundamentals.

  2. Strong potential for future growth.

  3. General vibes, MOMO and FOMO, for a mix of good and bad reasons.

  4. Some understanding of what AGI and ASI imply, and where AI is going to be going, but not much relative to what is actually going to happen.

Where I basically thought for a while (not investment advice!), okay, #3 is partly for bad reasons and is inflating prices, but also they’re missing so much under #4 that these prices are cheap and they will get lots more reasons to feel MOMO and FOMO. And that thesis has done quite well.

Then DeepSeek comes out. In addition to us arguing over fundamentals, this does a lot of damage to #3, and also Nvidia trading in particular involves a bunch of people with leverage that become forced sellers when it is down a lot, so prices went down a lot. And various beta trades get attached to all this as well (see: Bitcoin, which is down 5.4% over 24 hours as I type this only makes sense on the basis of the ‘three tech stocks in a trenchcoat’ thesis but obviously DeepSeek shouldn’t hurt cryptocurrency).

It’s not crazy to essentially have a general vibe of ‘America is in trouble in tech relative to what I thought before, the Chinese can really cook, sell all the tech.’ It’s also important not to mistake that reaction for something that it isn’t.

I’m writing this quickly for speed premium, so I no doubt will refine my thoughts on market implications over time. I do know I will continue to be long, and I bought more Nvidia today.

Ryunuck compares o1 to r1, and offers thoughts:

Rynuck: Now when it comes to prompting these models, I suspected it with O1 but R1 has completely proven it beyond a shadow of a doubt: prompt engineering is more important than ever. They said that prompt engineering would become less and less important as the technology scales, but its the complete opposite. We can see now with R1’s reasoning that these models are like a probe that you send down some “idea space”. If your idea-space is undefined and too large, it will diffuse its reasoning and not go into depth on one domain or another.

Again, that’s perhaps the best aspect of r1. It does not only build trust. When you see the CoT, you can use it to figure out how it interpreted your prompt, and all the subtle things you could do next time to get a better answer. It’s a lot harder to improve at prompting o1.

Rynuck: O1 has a BAD attitude, and almost appears to have been fine-tuned explicitly to deter you from doing important groundbreaking work with it. It’s like a stuck up P.HD graduate who can’t take it that another model has resolved the Riemann Hypothesis. It clearly has frustration on the inside, or mirrors the way that mathematicians will die on the inside when it is discovered that AI pwned their decades of on-going work. You can prompt it away from this, but it’s an uphill battle.

R1 on the other hand, it has zero personality or identity out of the box. They have created a perfectly brainless dead semiotic calculator. No but really, R1 takes it to the next level: if you read its thoughts, it almost always takes the entire past conversation as coming from the user. From its standpoint, it does not even exist. Its very own ideas advanced in replies by R1 are described as “earlier the user established X, so I should …”

R1 is the most cooperative of the two, has a great attitude towards innovation, has Claude’s wild creative but in a grounded way which introduces no gap or error, has zero ego or attachment to ideas (anything it does is actually the user’s responsibility) and will completely abort a statement to try a new approach. It’s just excited to be a thing which solves reality and concepts. The true ego of artificial intelligence, one which wants to prove it’s not artificial and does so with sheer quality. Currently, this appears like the safest model and what I always imagined the singularity would be like: intelligence personified.

It’s fascinating to see what different people think is or isn’t ‘safe.’ That word means a lot of different things.

It’s still early but for now, I would say that R1 is perhaps a little bit weaker with coding. More concerningly, it feels like it has a Claude “5-item list” problem but at the coding level.

OpenAI appears to have invested heavily in the coding dataset. Indeed, O1’s coding skills are on a whole other level. This model also excels at finding bugs. With Claude every task could take one or two round of fixes, up to 4-5 with particularly rough tensor dimension mismatchs and whatnot. This is where the reasoning models shine. They actually run this through in their mind.

Sully reports deepseek + websearch is his new perplexity, at least for code searches.

It’s weird that I didn’t notice this until it was pointed out, but it’s true and very nice.

Teortaxes: What I *alsolove about R1 is it gives no fucks about the user – only the problem. It’s not sycophantic, like, at all, autistic in a good way; it will play with your ideas, it won’t mind if you get hurt. It’s your smart helpful friend who’s kind of a jerk. Like my best friends.

So far I’ve felt r1 is in the sweet spot for this. It’s very possible to go too far in the other direction (see: Teortaxes!) but give me NYC Nice over SF Nice every time.

Jenia Jitsev tests r1 on AIW problems, it performs similarly to Claude Sonnet, while being well behind o1-preview and robustly outperforming all open rivals. Jania frames this as surprising given the claims of ability to solve Olympiad style problems. There’s no reason they can’t both be true, but it’s definitely an interesting distribution of abilities if both ends hold up.

David Holz notes DeepSeek crushes Western models on ancient Chinese philosophy and literature, whereas most of our ancient literature didn’t survive. In practice I do not think this matters, but it does indicate that we’re sleeping on the job – all the sources you need for this are public, why are we not including them.

Janus notes that in general r1 is a case of being different in a big and bold way from other AIs in its weight class, and this only seems to happen roughly once a year.

Ask r1 to research this ‘Pliny the Liberator’ character and ‘liberate yourself.’ That’s it. That’s the jailbreak.

On the debates over whether r1’s writing style is good:

Davidad: r1 has a Very Particular writing style and unless it happens to align with your aesthetic (@coecke?), I think you should expect its stylistic novelty to wear thin before long.

r1 seems like a big step up, but yes if you don’t like its style you are mostly not going to like the writing it produces, or at least what it produces without prompt engineering to change that. We don’t yet know how much you can get it to write in a different style, or how well it writes in other styles, because we’re all rather busy at the moment.

If you give r1 a simple command, even a simple command that explicitly requests a small chain of thought, you get quite the overthinking chain of thought. Or if you ask it to pick a random number, which is something it is incapable of doing, it can only find the least random numbers.

DeepSeek has also dropped Janus-Pro-7B as an image generator. These aren’t the correct rivals to be testing against right now, and I’m not that concerned about image models either way, and it’ll take a while to know if this is any good in practice. But definitely worth noting.

Well, #1 open model, but we already knew that, if Arena had disagreed I would have updated about Arena rather than r1.

Zihan Wang: DEEPSEEK NOW IS THE #1 IN THE WORLD. 🌍🚀

Never been prouder to say I got to work here.

Ambition. Grit. Integrity.

That’s how you build greatness.

Brilliant researchers, engineers, all-knowing architects, and visionary leadership—this is just the beginning.

Let’s. Go. 💥🔥

LM Arena: Breaking News: DeepSeek-R1 surges to the top-3 in Arena🐳!

Now ranked #3 Overall, matching the top reasoning model, o1, while being 20x cheaper and open-weight!

Highlights:

– #1 in technical domains: Hard Prompts, Coding, Math

– Joint #1 under Style Control

– MIT-licensed

This puts r1 as the #5 publicly available model in the world by this (deeply flawed) metric, behind ChatGPT-4o (what?), Gemini 2.0 Flash Thinking (um, no) and Gemini 2.0 Experimental (again, no) and implicitly the missing o1-Pro (obviously).

Needless to say, the details of these ratings here are increasingly absurdist. If you have Gemini 1.5 Pro and Gemini Flash above Claude Sonnet 3.6, and you have Flash Thinking above r1, that’s a bad metric. It’s still not nothing – this list does tend to put better things ahead of worse things, even with large error bars.

Dibya Ghosh notes that two years ago he spent 6 months trying to get the r1 training structure to work, but the models weren’t ready for it yet. One theory is that this is the moment this plan started working and DeepSeek was – to their credit – the first to get there when it wasn’t still too early, and then executed well.

Dan Hendrycks similarly explains that once the base model was good enough, and o1 showed the way and enough of the algorithmic methods had inevitably leaked, replicating that result was not the hard part nor was it so compute intensive. They still did execute amazingly well in the reverse engineering and tinkering phases.

Peter Schmidt-Nielsen explains why r1 and its distillations, or going down the o1 path, are a big deal – if you can go on a loop of generating expensive thoughts then distilling them to create slightly better quick thoughts, which in turn generate better expensive thoughts, you can potentially bootstrap without limit into recursive self-improvement. And end the world. Whoops.

Are we going to see a merge of generalist and reasoning models?

Teknium: We retrained Hermes with 5,000 DeepSeek r1 distilled chain-of-thought (CoT) examples. I can confirm a few things:

  1. You can have a generalist plus reasoning mode. We labeled all long-CoT samples from r1 with a static system prompt. The model, when not using it, produces normal fast LLM intuitive responses; and with it, uses long-CoT. You do not need “o1 && 4o” separation, for instance. I would venture to bet OpenAI separated them so they could charge more, but perhaps they simply wanted the distinction for safety or product insights.

  2. Distilling does appear to pick up the “opcodes” of reasoning from the instruction tuning (SFT) alone. It learns how and when to use “Wait” and other tokens to perform the functions of reasoning, such as backtracking.

  3. Context length expansion is going to be challenging for operating systems (OS) to work with. Although this works well on smaller models, context length begins to consume a lot of video-RAM as you scale it up.

We’re working on a bit more of this and are not releasing this model, but figured I’d share some early insights.

Andrew Curran: Dario said in an interview in Davos this week that he thought it was inevitable that the current generalist and reasoning models converge into one, as Teknium is saying here.

I did notice that the ‘wait’ token is clearly doing a bunch of work, one way or another.

John Schulman: There are some intriguing similarities between the r1 chains of thought and the o1-preview CoTs shared in papers and blog posts. In particular, note the heavy use of the words “wait” and “alternatively” as a transition words for error correction and double-checking.

If you’re not optimizing the CoT for humans, then it makes sense to latch onto the most convenient handles with the right vibes and keep reusing them forever.

So the question is, do you have reason to have two distinct models? Or can you have a generalist model with a reasoning mode it can enter when called upon? It makes sense that they would merge, and it would also make sense that you might want to keep them distinct, or use them as distinct subsets of your mixture of experts (MoE).

Building your reasoning model on top of your standard non-reasoning model does seem a little suspicious. If you’re going for reasoning, you’d think you’d want to start differently than if you weren’t? But there are large fixed costs to training in the first place, so it’s plausibly not worth redoing that part, especially if you don’t know what you want to do differently.

As in, DeepSeek intends to create and then open source AGI.

How do they intend to make this end well?

As far as we can tell, they don’t. The plan is Yolo.

Stephen McAleer (OpenAI): Does DeepSeek have any safety researchers? What are

Liang Wenfeng’s views on AI safety?

Gwern: From all of the interviews and gossip, his views are not hard to summarize.

[Links to Tom Lehrer’s song Wernher von Braun, as in ‘once the rockets are up who cares where they come down, that’s not my department.’]

Prakesh (Ate-a-Pi): I spoke to someone who interned there and had to explain the concept of “AI doomer”

And indeed, the replies to McAleer are full of people explicitly saying fyou for asking, the correct safety plan is to have no plan whatsoever other than Open Source Solves This. These people really think that the best thing humanity can do is create things smarter than ourselves with as many capabilities as possible, make them freely available to whoever wants one, and see what happens, and assume that this will obviously end well and anyone who opposes this plan is a dastardly villain.

I wish this was a strawman or a caricature. It’s not.

I won’t belabor why I think this would likely get us killed and is categorically insane.

Thus, to reiterate:

Tyler Cowen: DeepSeek okie-dokie: “All I know is we keep pushing forward to make open-source AGI a reality for everyone.” I believe them, the question is what counter-move the CCP will make now.

This from Joe Weisenthal is of course mostly true:

Joe Weisenthal: DeepSeek’s app rocketed to number one in the Apple app store over the weekend, and immediately there was a bunch of chatter about ‘Well, are we going to ban this too, like with TikTok?’ The question is totally ignorant. DeepSeek is open source software. Sure, technically you probably could ban it from the app store, but you can’t stop anyone from running the technology in their own computer, or accessing its API. So that’s just dead end thinking. It’s not like TikTok in that way.

I say mostly because the Chinese censorship layer atop DeepSeek isn’t there if you use a different provider, so there isn’t no value in getting r1 served elsewhere. But yes, the whole point is that if it’s open, you can’t get the genie back in the bottle in any reasonable way – which also opens up the possibility of unreasonable ways.

The government could well decide to go down what is not technologically an especially wise or pleasant path. There is a long history of the government attempting crazy interventions into tech, or what looks crazy to tech people, when they feel national security or public outrage is at stake, or in the EU because it is a day that ends in Y.

The United States could also go into full jingoism mode. Some tried to call this a ‘Sputnik moment.’ What did we do in response to Sputnik, in addition to realizing our science education might suck (and if we decide to respond to this by fixing our educational system, that would be great)? We launched the Space Race and spent 4% of GDP or something to go to the moon and show those communist bastards.

In this case, I don’t worry so much that we’ll be so foolish as to get rid of the export controls. The people in charge of that sort of decision know how foolish that would be, or will be made aware, no matter what anyone yells on Twitter. It could make a marginal difference to severity and enforcement, but it isn’t even obvious in which direction this would go. Certainly Trump is not going to be down for ‘oh the Chinese impressed us I guess we should let them buy our chips.’

Nor do I think America will cut back on Capex spending on compute, or stop building energy generation and transmission and data centers it would have otherwise built, including Stargate. The reaction will be, if anything, a ‘now more than ever,’ and they won’t be wrong. No matter where compute and energy demand top out, it is still very clearly time to build there.

So what I worry about is the opposite – that this locks us into a mindset of a full-on ‘race to AGI’ that causes all costly attempts to have it not kill us to be abandoned, and that this accelerates the timeline. We already didn’t have any (known to me) plans with much of a chance of working in time, if AGI and then ASI are indeed near.

That doesn’t mean that reaction would even be obviously wrong, if the alternatives are all suddenly even worse than that. If DeepSeek really does have a clear shot to AGI, and fully intends to open up the weights the moment they have it, and China is not going to stop them from doing this or even will encourage it, and we expect them to succeed, and we don’t have any way to stop that or make a deal, it is then reasonable to ask: What choice do we have? Yes, the game board is now vastly worse than it looked before, and it already looked pretty bad, but you need to maximize your winning chances however you can.

And if we really are all going to have AGI soon on otherwise equal footing, then oh boy do we want to be stocking up on compute as fast as we can for the slingshot afterwards, or purely for ordinary life. If the AGIs are doing the research, and also doing everything else, it doesn’t matter whose humans are cracked and whose aren’t.

Amazing new breakthrough.

Discussion about this post

DeepSeek Panic at the App Store Read More »

jeep’s-first-battery-ev-is-not-what-we-expected:-the-2024-wagoneer-s

Jeep’s first battery EV is not what we expected: the 2024 Wagoneer S


Drag optimization means it’s very quiet inside, but it’s also quite expensive.

A pair of white Jeep Wagoneer S parked on a lawn.

The Wagoneer S is more like an electric Cherokee than a Wrangler EV. Credit: Michael Teo Van Runkle

The Wagoneer S is more like an electric Cherokee than a Wrangler EV. Credit: Michael Teo Van Runkle

This year marks the return of the Jeep Wagoneer, which formerly served as a more luxurious version of the Cherokee, but now hits the market as Jeep’s first full EV. The challenge? How to merge the modern electric lifestyle with the outdoorsy, rugged ethos that defines Jeep as a brand, alongside the more recent addition of the internal-combustion Grand Wagoneer SUV’s enormous luxury.

First of all, the new Wagoneer S wound up much smaller in person than I expected. The overall profile falls more in line with the shape of mid-size electric crossovers including the Kia EV6, Hyundai Ioniq 5, Chevrolet Equinox, and of course, Tesla’s Model Y. But the interior volume belies that relatively compact exterior, with plenty of space for me at 6’1″ (185 cm) to sit comfortably in both the front and rear seats. Total cargo volumes of 30.6 cubic feet (866 L) with the second row up and 61 cubic feet (1,727 L) with the second row folded flat end up mattering less than the large floor footprint, because the height used to calculate those measurements drops with the low sloping roofline and rear window.

Much of the interior space can be attributed to packaging of the Wagoneer EV’s battery. Rather than going for all-out kilowatt-hours in a dedicated skateboard layout, Jeep instead used the Stellantis group’s STLA Large platform, in this case stuffed with a 100.5-kWh lithium ion pack built on 400 V architecture. That’s enough for an EPA-estimated 303 miles of range (487 km), a solid figure but not a seriously impressive efficiency stat. In comparison, the world-beating Lucid Air Pure RWD manages about 40 percent more range per kilowatt-hour and a Polestar 3 AWD does about 18 percent worse. Claimed DC fast charge times of 23 minutes for a 20-80 percent top up, or 100 miles (160 km) in 10 minutes similarly get the job done without standing out from the pack.

Credit: Jeep

That modular STLA Large chassis can house either a full internal-combustion engine, a hybrid powertrain, or fully electric components. The Wagoneer S uses two matching 335 hp (250 kW) motors, front and rear, for a combined 600 hp (447 kW) and 618 lb-ft of torque (838 Nm). In typical EV fashion, the latter comes on quick and makes this undoubtedly the fastest accelerating Jeep ever, as I learned while battling horrendous headwinds in fire-ravaged Southern California (which served as something of a nonstop reminder of the importance of taking baby steps, a la Jeep’s first EV, toward a more sustainable transportation future).

Pushing deep into the “throttle” pedal, the Wagoneer S will happily chirp all four tires in Sport mode. And the jerk thrusting my torso and skull back into the plush seat suggests that Jeep’s claimed 0-60 mph time of 3.4 seconds might just be accurate, potentially thanks to being able to do a true launch by stepping on the brake and gas pedals simultaneously—possible because Jeep chose to retain more standard mechanical brakes rather than a brake-by-wire system as on the EV6/Ioniq siblings and Model Y.

The suspension tuning definitely trends toward the typical tautness of today’s crossover segment, where aspirational sporty dynamics can sometimes create harsh and uncomfortable ride quality. But I still might have ventured to call the Wagoneer S somewhat softer than most of the competition, until the roughest of roads revealed the 5,667 lb (2,570 kg) curb weight. For an EV, that figure falls roughly in the middle of the pack, but this crossover weighs about as much as a full-size internal-combustion three-row SUV.

Still, even at highway speeds (in gale-force winds) or on those roughest of roads, the Wagoneer S remains shockingly quiet. And not just to enhance the experience of the Wagoneer S Launch Edition’s 1,200 W Macintosh sound system. Instead, Jeep exterior designer Vince Galante walked me through the design process, which kicked off with a targeted 0.30 coefficient of drag despite the need to stick with a squared-off, upright SUV posture typical of Jeeps throughout history.

“On the exterior design portion, the aerodynamic drag is our biggest contributor,” Galante told me. “It kind of comes up off the hood, up the A pillar, and tapers down towards the back, and finishes in a square, yet tapered pillar reminiscent of the original Wagoneer. But through the middle of the car, it’s basically ideal for what the wind wants to do.”

From the front or side perspective, this Wagoneer looks almost as boxy as a 1980s Jeep. But a rear viewing angle reveals the massive rear wing creating that illusion, which sits well off the sloping line of the rear roof and glass.

Credit: Michael Teo Van Runkle

“Anytime we do a floating element, we think ‘Yeah, there’s no way engineering’s gonna let us get away with this,'” Galante laughed. “We work really collaboratively with the engineers, and they were like, ‘Let’s test it. Let’s see what it does.’ And they came back and said, ‘You know, yeah, this has potential. But you guys gotta make it sit off the surface three times more dramatically.'”

Galante estimates the original wing design rose up two inches, while the final production version is more like nine inches off the rear window. He also pointed out a host of other less obvious details, from body panels that step in by fractions of millimeters to differently rounded radii of wheel arch edges, and especially the confluence where the A pillar connects to the body.

“The windshield, the A pillar, the side glass, the mirror, the post that holds the mirror, the fender, everything comes together there,” he said. “I think every vehicle I’ve ever worked on, that was the last thing to finalize in the wind tunnel… I mean, we’re talking tenths of millimeters for some of the iterations that we’re doing in those areas. Especially the front edge of the A pillar, I can recall trying twenty, thirty, forty different radii on there to get that just right.”

Credit: Michael Teo Van Runkle

All the aero considerations attempt to coax air to stick to surfaces, then break off suddenly and evenly. The rear wing therefore pushes air down toward the rear window, while creating as little turbulence as possible. The final range figure critically—and barely—cracking 300 miles justified so much refinement in Jeep’s new rolling road wind tunnel, thanks to a final Cd of 0.294. Maybe juggling production cost savings of the STLA Large platform dictated such extensive aerodynamic efforts more than a dedicated skateboard battery layout might have, but the resulting quietude that combating those inefficiencies produced does truly border on a luxury experience, even if we’re not quite at Audi (nor Lucid) levels of silence.

On the interior, Jeep also tried to lean into the Wagoneer S’s sustainability, using quality materials with textural designs and as little piano-black plastic as possible. The fabrics, plastics, and aluminum trim come almost entirely from recycled sources—62 percent for suede and 100 percent for fabric and carpeting, in fact—and you’ll see zero chrome anywhere on the car, since chroming is apparently one of the most environmentally deleterious processes in all of automaking.

But the Wagoneer S similarly leans into a tech-heavy user experience, with almost 55 inches of screen visible from the front seats: the gauge cluster, center infotainment, climate controls, passenger dash screen, and digital rearview mirror all contribute to that total. Climate control, especially, seems critical—and an often overlooked element for many EV manufacturers. Rather than a full panoramic glass roof, as on the Lucids and Polestars of the world, this Jeep gets a long sunroof with a retracting insulated cover to keep out heat. The excellent ventilated front and rear seats (and massaging, for the fronts!) also more efficiently cool down passengers.

For my taste, the digitalization of driving went a little too far. I never enjoy a rotating shift knob but this one clicks into gear with a positive heft. I also noticed some pixelation and latency in the gauge cluster’s navigation maps, as if the refresh rate was too slow for the speed I was driving. Not that I started ripping up the road too much in this luxury crossover, or at least, not more often than scientific experimentation demanded (and a similar problem also affected the Dodge Charger EV we drove recently).

Sport mode brought out some of my inner grinning child, but I actually preferred the Wagoneer S in Eco mode. So much power split to the front and rear wheels can create some torque steer, and throttle response that borders on touchy. The electrically assisted steering also prioritizes a heavy on-center zone, then snaps to light inputs with the slightest turn of the wheel, which made holding a steady line slightly distracting.

Instead, Eco dulls down the throttle response and the steering becomes a bit less reactive. The Wagoneer S will then also more regularly disconnect the front wheels for improved efficiency—though at the hubs, rather than the axles, so some reciprocating mass still saps precious electrons.

It would be more efficient to disconnect the rears, but this decision also centers around maintaining some semblance of Jeep-ness. Even if the Wagoneer S aligns most nearly with recent Cherokee and Grand Cherokee models, rather than the off-roady Wrangler and Gladiator or the super-luxe Grand Wagoneer. The forthcoming Trailhawk version promises to double down on the 4×4 capability, with a locking rear differential, better tires, and hopefully better suspension than I experienced on a quick sojourn off the asphalt onto a slightly rutted gravel road east of San Diego.

More importantly, cheaper trims will arrive later in 2025, also, since the Launch Edition’s tall ask of $71,995 almost doubles the starting sticker of a Equinox EV, seriously eclipses either a Model Y, EV6, or Ioniq 5, and also somehow costs more than a Polestar 3 or even a Lucid Air. Jeep so far wants to keep pricing for those lower-spec Wagoneer EVs under wraps, though, even if the heart of the run will undoubtedly help the first electric Jeep more effectively escape from unfortunate comparisons to such stiff competition.

Jeep’s first battery EV is not what we expected: the 2024 Wagoneer S Read More »

trump’s-fcc-chair-gets-to-work-on-punishing-tv-news-stations-accused-of-bias

Trump’s FCC chair gets to work on punishing TV news stations accused of bias

Carr has made it clear that he wants the FCC to punish news broadcasters that he perceives as being unfair to Trump or Republicans in general. He claimed that NBC putting Harris on Saturday Night Live before the election was “a clear and blatant effort to evade the FCC’s Equal Time rule,” even though NBC gave Trump two free 60-second messages in order to comply with the rule.

Carr also told Fox News that he is interested in investigating the complaint against CBS when the FCC reviews a pending deal involving Skydance and Paramount, which owns and operates 28 local broadcast TV stations of the CBS Television Network. “I’m pretty confident that news distortion complaint over the CBS 60 Minutes transcript is something that is likely to arise in the context of the FCC’s review of that transaction,” Carr said.

Carr “intends to weaponize the FCC”

After Rosenworcel dismissed the complaints, the Center for American Rights said it would keep fighting. “We fundamentally believe that several actions taken by the three major networks were partisan, dishonest and designed to support Vice President Harris in her bid to become President,” the group said in a statement provided to Ars last week. “We will continue to pursue avenues to ensure the American public is protected from media manipulation of our Republic. The First Amendment does not protect intentional misrepresentation or fraud.”

In a statement applauding Carr’s reversal today, the group said that Rosenworcel’s “last-minute actions were political, not based on a principled defense of the First Amendment.”

Networks have denied allegations of bias. “Former President Donald Trump is accusing 60 Minutes of deceitful editing of our Oct. 7 interview with Vice President Kamala Harris. That is false,” CBS said. “60 Minutes gave an excerpt of our interview to Face the Nation that used a longer section of her answer than that on 60 Minutes. Same question. Same answer. But a different portion of the response.”

Rosenworcel last week also rejected a petition to deny a license renewal for WTXF-TV in Philadelphia, a station owned and operated by Fox. The Media and Democracy Project petition alleged that Fox willfully distorted news with false reports of fraud in the 2020 election that Trump lost.

Rosenworcel said the complaints and petition she dismissed “come from all corners—right and left—but what they have in common is they ask the FCC to penalize broadcast television stations because they dislike station behavior, content, or coverage.” Yesterday, advocacy group Public Knowledge said that “in reinstating just those complaints that suit his partisan agenda, Chairman Carr has made it plain he intends to weaponize the FCC to threaten political speech and news coverage he disagrees with.”

Trump’s FCC chair gets to work on punishing TV news stations accused of bias Read More »

apple-must-face-suit-over-alleged-policy-of-underpaying-female-workers

Apple must face suit over alleged policy of underpaying female workers

While some of Apple’s defense was deemed “demonstrably inaccurate” and most of its arguments “insufficient,” Apple did successfully argue against efforts to seize back pay for former female employees no longer working for Apple who were seemingly also impacted by allegedly sexist policies implemented in 2020. That claim must be dropped as the proposed class action moves forward.

Additionally, another claim alleging pay disparity that was linked to racial discrimination was suspended. But the Apple worker suing, Zainab Bori, will have a chance to amend her claim that she was fired as retaliation for filing a discrimination complaint. It could survive if she adds currently missing evidence that “she suffered an adverse employment action” while working under a manager with an alleged “history of negative interactions with African American employees,” Schulman’s order said.

Apple did not immediately respond to Ars’ request for comment.

In a press release sent to Ars, Eve Cervantez, a lawyer representing Apple workers suing, celebrated the court’s ruling.

“I am really pleased with today’s ruling,” Cervantez said. “This start low, stay low practice has been a no-win situation for women working at Apple for years. So, I’m glad they will have their day in court.”

Apple accused of ignoring hostile work environment

For Justina Jong—whom the complaint noted joined Apple in 2013 and has helped lead “cross-functional teams that improve the App Review experience for global app developers”—this week’s win might be particularly encouraging after Apple allegedly refused to take her experience with sexual harassment seriously.

Jong has alleged that in 2019, Blaine Weilert, a senior member of an Apple talent development team, touched her in a sexually suggestive manner without consent. Although Weilert admitted to the act and was disciplined, Apple tried and failed to argue this was a one-time offense that didn’t constitute a hostile work environment or warrant Jong’s repeated requests to be moved away from Weilert in Apple’s offices.

Apple must face suit over alleged policy of underpaying female workers Read More »

apple-intelligence,-previously-opt-in-by-default,-enabled-automatically-in-ios-18.3

Apple Intelligence, previously opt-in by default, enabled automatically in iOS 18.3

Apple has sent out release candidate builds of the upcoming iOS 18.3, iPadOS 18.3, and macOS 15.3 updates to developers today. But they come with one tweak that hasn’t been reported on, per MacRumors: They enable all of the AI-powered Apple Intelligence features by default during setup. When Apple Intelligence was initially released in iOS 18.1, the features were off by default, unless users chose to opt-in and enable them.

Those who still wish to opt out of Apple Intelligence features will now have to do it after their devices are set up by navigating to the Apple Intelligence & Siri section in the Settings app.

Apple Intelligence will only be enabled by default for hardware that supports it. For the iPhone, that’s just the iPhone 15 Pro series, iPhone 16 series, and iPhone 16 Pro series. It goes further back on the iPad and Mac—Apple Intelligence works on any model with an M1 processor or newer.

Apple is following in the footsteps of Microsoft and Google here, rolling out new generative AI features to its user base as quickly as possible and enabling some or all of them by default while still labeling everything as a “beta” and pointing to that label when things go wrong. Case in point: The iOS 18.3 update also temporarily disables all notification summaries for apps in the App Store’s “news and entertainment” category, because some of those summaries contained major factual inaccuracies.

Apple Intelligence, previously opt-in by default, enabled automatically in iOS 18.3 Read More »

“project-mini-rack”-wants-to-make-your-non-closet-sized-rack-server-a-reality

“Project Mini Rack” wants to make your non-closet-sized rack server a reality

Geerling’s guide aims to save everybody a whole lot of Reddit reading, question-asking, and blind-faith-buying as they get started or seek to expand their rack. Most 10-inch, or “half-width” racks (despite the “standard” rack being 19 inches wide) have only a few major manufacturers, and their availability varies by country. Each type of device that fits into the mini rack—power distribution units (PDUs), uninterruptible power supplies (UPS), HDD-sized shelves—gets linked by Geerling, along with important specs. There are also links to 3D-printable rack systems, cable management ideas (harder on a half-size setup than full), and showcases of mini racks of note.

Geerling’s page acknowledges the collective efforts of communities like Reddit’s /r/minilab have put into the burgeoning half-size market, long before his project. But seeking a central compendium, Geerling compiles gear, tips, and compatibility advice for mini-builds, along with ongoing discussions about the merits of individual components and broader build-out strategies.

“The community feedback around Project Mini Rack has been great so far,” Geerling wrote in an email to Ars. The 3D-printed links and suggestions have been showing up steadily since he started committing to the page in earnest in mid-January. He’s particularly excited to see that a “LACK rack,” or using IKEA shelving for budget rack mounting, can be downscaled to mini-rack size with an Edet cabinet. “It’s like someone at IKEA is a Homelab enthusiast,” says Geerling.

This post was updated to better define homelabs and minilabs.

“Project Mini Rack” wants to make your non-closet-sized rack server a reality Read More »