Author name: Kris Guyer

little-echo

Little Echo

I believe that we will win.

An echo of an old ad for the 2014 US men’s World Cup team. It did not win.

I was in Berkeley for the 2025 Secular Solstice. We gather to sing and to reflect.

The night’s theme was the opposite: ‘I don’t think we’re going to make it.’

As in: Sufficiently advanced AI is coming. We don’t know exactly when, or what form it will take, but it is probably coming. When it does, we, humanity, probably won’t make it. It’s a live question. Could easily go either way. We are not resigned to it. There’s so much to be done that can tilt the odds. But we’re not the favorite.

Raymond Arnold, who ran the event, believes that. I believe that.

Yet in the middle of the event, the echo was there. Defiant.

I believe that we will win.

There is a recording of the event. I highly encourage you to set aside three hours at some point in December, to listen, and to participate and sing along. Be earnest.

If you don’t believe it, I encourage this all the more. If you don’t understand the mindset, or the culture behind it, or consider it an opponent or dislike it, and especially if yours is a different fight? I encourage this all the more than that. You can also attend New York’s Solstice on the 20th.

You will sing songs you know, and songs you don’t. You will hear tales of struggles, of facing impossible odds or unbearable loss and fighting anyway, of how to face it all and hopefully stay sane. To have the end, if it happens, find us doing well.

I live a wonderful life.

I am crying as I write this. But when I am done, I will open a different Chrome window. I will spend the day with friends I love dearly and watching football games. This evening my wife and I will attend a not wedding of two of them, that is totally a wedding. We will fly home to our wonderful kids, and enjoy endless wonders greater than any king in the beating heart of the world. I want for nothing other than time.

Almost every day, I will mostly reject those wonders. I will instead return to my computer. I will confront waves of events and information. The avalanche will accelerate. Release after release, argument after argument, policies, papers, events, one battle after another. People will be determined to handle events with less dignity than one could imagine, despite having read this sentence. I fight to not be driven into rages. I will triage. I will process. I will change my mind. I will try to explain, just one more time. I will move pieces around multiple chessboards.

We continue. Don’t tell me to stop. Someone has to, and no one else will.

I know if I ignored it, anything else would soon turn to ash in my mouth.

I will look at events, and say to myself as I see the moves unfolding, the consequences of choices I made or influenced, for good and ill: This is the world we made.

It aint over till its over. Never leave a ballgame early. Leave it all on the field, for when the dust covers the sun and all you hope for is undone. You play to win the game.

The odds are against us and the situation is grim. By default, we lose. I act accordingly, and employ some of the unteachable methods of sanity and the mirror version of others, all of which are indeed unteachable but do totally work.

Yet the echo is there. In my head. It doesn’t care.

I believe that we will win.

Discussion about this post

Little Echo Read More »

steamos-vs.-windows-on-dedicated-gpus:-it’s-complicated,-but-windows-has-an-edge

SteamOS vs. Windows on dedicated GPUs: It’s complicated, but Windows has an edge

Other results vary from game to game and from GPU to GPU. Borderlands 3, for example, performs quite a bit better on Windows than on SteamOS across all of our tested GPUs, sometimes by as much as 20 or 30 percent (with smaller gaps here and there). As a game from 2019 with no ray-tracing effects, it still runs serviceably on SteamOS across the board, but it was the game we tested that favored Windows the most consistently.

In both Forza Horizon 5 and Cyberpunk 2077, with ray-tracing effects enabled, you also see a consistent advantage for Windows across the 16GB dedicated GPUs, usually somewhere in the 15 to 20 percent range.

To Valve’s credit, there were also many games we tested where Windows and SteamOS performance was functionally tied. Cyberpunk without ray-tracing, Returnal when not hitting the 7600’s 8GB RAM limit, and Assassin’s Creed Valhalla were sometimes actually tied between Windows and SteamOS, or they differed by low-single-digit percentages that you could chalk up to the margin of error.

Now look at the results from the integrated GPUs, the Radeon 780M and RX 8060S. These are pretty different GPUs from one another—the 8060S has more than three times the compute units of the 780M, and it’s working with a higher-speed pool of soldered-down LPDDR5X-8000 rather than two poky DDR5-5600 SODIMMs.

But Borderlands aside, SteamOS actually did quite a bit better on these GPUs relative to Windows. In both Forza and Cyberpunk with ray-tracing enabled, SteamOS slightly beats Windows on the 780M, and mostly closes the performance gap on the 8060S. For the games where Windows and SteamOS essentially tied on the dedicated GPUs, SteamOS has a small but consistent lead over Windows in average frame rates.

SteamOS vs. Windows on dedicated GPUs: It’s complicated, but Windows has an edge Read More »

elon-musk’s-x-first-to-be-fined-under-eu’s-digital-services-act

Elon Musk’s X first to be fined under EU’s Digital Services Act

Elon Musk’s X became the first large online platform fined under the European Union’s Digital Services Act on Friday.

The European Commission announced that X would be fined nearly $140 million, with the potential to face “periodic penalty payments” if the platform fails to make corrections.

A third of the fine came from one of the first moves Musk made when taking over Twitter. In November 2022, he changed the platform’s historical use of a blue checkmark to verify the identities of notable users. Instead, Musk started selling blue checks for about $8 per month, immediately prompting a wave of imposter accounts pretending to be notable celebrities, officials, and brands.

Today, X still prominently advertises that paying for checks is the only way to “verify” an account on the platform. But the commission, which has been investigating X since 2023, concluded that “X’s use of the ‘blue checkmark’ for ‘verified accounts’ deceives users.”

This violates the DSA as the “deception exposes users to scams, including impersonation frauds, as well as other forms of manipulation by malicious actors,” the commission wrote.

Interestingly, the commission concluded that X made it harder to identify bots, despite Musk’s professed goal to eliminate bots being a primary reason he bought Twitter. Perhaps validating the EU’s concerns, X recently received backlash after changing a feature that accidentally exposed that some of the platform’s biggest MAGA influencers were based “in Eastern Europe, Thailand, Nigeria, Bangladesh, and other parts of the world, often linked to online scams and schemes,” Futurism reported.

Although the DSA does not mandate the verification of users, “it clearly prohibits online platforms from falsely claiming that users have been verified, when no such verification took place,” the commission said. X now has 60 days to share information on the measures it will take to fix the compliance issue.

Elon Musk’s X first to be fined under EU’s Digital Services Act Read More »

rocket-report:-blunder-at-baikonur;-do-launchers-really-need-rocket-engines?

Rocket Report: Blunder at Baikonur; do launchers really need rocket engines?


The Department of the Air Force approves a new home in Florida for SpaceX’s Starship.

South Korea’s Nuri 1 rocket is lifted vertical on its launch pad in this multi-exposure photo. Credit: Korea Aerospace Research Institute

Welcome to Edition 8.21 of the Rocket Report! We’re back after the Thanksgiving holiday with more launch news. Most of the big stories over the last couple of weeks came from abroad. Russian rockets and launch pads didn’t fare so well. China’s launch industry celebrated several key missions. SpaceX was busy, too, with seven launches over the last two weeks, six of them carrying more Starlink Internet satellites into orbit. We expect between 15 and 20 more orbital launch attempts worldwide before the end of the year.

As always, we welcome reader submissions. If you don’t want to miss an issue, please subscribe using the box below (the form will not appear on AMP-enabled versions of the site). Each report will include information on small-, medium-, and heavy-lift rockets, as well as a quick look ahead at the next three launches on the calendar.

Another Sarmat failure. A Russian intercontinental ballistic missile (ICBM) fired from an underground silo on the country’s southern steppe on November 28 on a scheduled test to deliver a dummy warhead to a remote impact zone nearly 4,000 miles away. The missile didn’t even make it 4,000 feet, Ars reports. Russia’s military has been silent on the accident, but the missile’s crash was seen and heard for miles around the Dombarovsky air base in Orenburg Oblast near the Russian-Kazakh border. A video posted by the Russian blog site MilitaryRussia.ru on Telegram and widely shared on other social media platforms showed the missile veering off course immediately after launch before cartwheeling upside down, losing power, and then crashing a short distance from the launch site.

An unenviable track record … Analysts say the circumstances of the launch suggest it was likely a test of Russia’s RS-28 Sarmat missile, a weapon designed to reach targets more than 11,000 miles (18,000 kilometers) away, making it the world’s longest-range missile. The Sarmat missile is Russia’s next-generation heavy-duty ICBM, capable of carrying a payload of up to 10 large nuclear warheads, a combination of warheads and countermeasures, or hypersonic boost-glide vehicles, according to the Center for Strategic and International Studies. Simply put, the Sarmat is a doomsday weapon designed for use in an all-out nuclear war between Russia and the United States. The missile’s first full-scale test flight in 2022 apparently went well, but the program has suffered a string of consecutive failures since then, most notably a catastrophic explosion last year that destroyed the Sarmat missile’s underground silo in northern Russia.

The easiest way to keep up with Eric Berger’s and Stephen Clark’s reporting on all things space is to sign up for our newsletter. We’ll collect their stories and deliver them straight to your inbox.

Sign Me Up!

ESA fills its coffers for launcher challenge. The European Space Agency’s (ESA) European Launcher Challenge received a significant financial commitment from its member states during the agency’s Ministerial Council meeting last week, European Spaceflight reports. The challenge is designed to support emerging European rocket companies while giving ESA and other European satellite operators more options to compete with the continent’s sole operational launch provider, Arianespace. Through the program, ESA will purchase launch services and co-fund capacity upgrades with the winners. ESA member states committed 902 million euros, or $1.05 billion, to the program at the recent Ministerial Council meeting.

Preselecting the competitors … In July, ESA selected two German companies—Isar Aerospace and Rocket Factory Augsburg—along with Spain’s PLD Space, France’s MaiaSpace, and the UK’s Orbex to proceed with the initiative’s next phase. ESA then negotiated with the governments of each company’s home country to raise money to support the effort. Germany, with two companies on the shortlist, is unsurprisingly a large contributor to the program, committing more than 40 percent of the total budget. France contributed nearly 20 percent, Spain funded nearly 19 percent, and the UK committed nearly 16 percent. Norway paid for 3 percent of the launcher challenge’s budget. Denmark, Portugal, Switzerland, and the Czech Republic contributed smaller amounts.

Europe at the service of South Korea. South Korea’s latest Earth observation satellite was delivered into a Sun-synchronous orbit Monday afternoon following a launch onboard a Vega C rocket by Arianespace, Spaceflight Now reports. The Korea Multi-Purpose Satellite-7 (Kompsat-7) mission launched from Europe’s spaceport in French Guiana. About 44 minutes after liftoff, the Kompsat-7 satellite was deployed into SSO at an altitude of 358 miles (576 kilometers). “By launching the Kompsat-7 satellite, set to significantly enhance South Korea’s Earth observation capabilities, Arianespace is proud to support an ambitious national space program,” said David Cavaillolès, CEO of Arianespace, in a statement.

Something of a rarity … The launch of Kompsat-7 is something of a rarity for Arianespace, which has dominated the international commercial launch market. It’s the first time in more than two years that a satellite for a customer outside Europe has been launched by Arianespace. The backlog for the light-class Vega C rocket is almost exclusively filled with payloads for the European Space Agency, the European Commission, or national governments in Europe. Arianespace’s larger Ariane 6 rocket has 18 launches reserved for the US-based Amazon Leo broadband network. (submitted by EllPeaTea)

South Korea’s homemade rocket flies again. South Korea’s homegrown space rocket Nuri took off from Naro Space Center on November 27 with the CAS500-3 technology demonstration and Earth observation satellite, along with 12 smaller CubeSat rideshare payloads, Yonhap News Agency reports. The 200-ton Nuri rocket debuted in 2021, when it failed to reach orbit on a test flight. Since then, the rocket has successfully reached orbit three times. This mission marked the first time for Hanwha Aerospace to oversee the entire assembly process as part of the government’s long-term plan to hand over space technologies to the private sector. The fifth and sixth launches of the Nuri rocket are planned in 2026 and 2027.

Powered by jet fuel … The Nuri rocket has three stages, each with engines burning Jet A-1 fuel and liquid oxygen. The fuel choice is unusual for rockets, with highly refined RP-1 kerosene or methane being more popular among hydrocarbon fuels. The engines are manufactured by Hanwha Aerospace. The fully assembled rocket stands about 155 feet (47.2 meters) tall and can deliver up to 3,300 pounds (1.5 metric tons) of payload into a polar Sun-synchronous orbit.

Hyundai eyes rocket engine. Meanwhile, South Korea’s space sector is looking to the future. Another company best known for making cars has started a venture in the rocket business. Hyundai Rotem, a member of Hyundai Motor Group, announced a joint program with Korean Air’s Aerospace Division (KAL-ASD) to develop a 35-ton-class reusable methane rocket engine for future launch vehicles. The effort is funded with KRW49 billion ($33 million) from the Korea Research Institute for Defense Technology Planning and Advancement (KRIT).

By the end of the decade … The government-backed program aims to develop the engine by the end of 2030. Hyundai Rotem will lead the engine’s planning and design, while Korean Air, the nation’s largest air carrier, will lead development of the engine’s turbopump. “Hyundai Rotem began developing methane engines in 1994 and has steadily advanced its methane engine technology, achieving Korea’s first successful combustion test in 2006,” Hyundai Rotem said in a statement. “Furthermore, this project is expected to secure the technological foundation for the commercialization of methane engines for reusable space launch vehicles and lay the groundwork for targeting the global space launch vehicle market.”

But who needs rocket engines? Moonshot Space, based in Israel, announced Monday that it has secured $12 million in funding to continue the development of a launch system—powered not by chemical propulsion, but electromagnetism, Payload reports. Moonshot plans to sell other aerospace and defense companies the tech as a hypersonic test platform, while at the same time building to eventually offer orbital launch services. Instead of conventional rocket engines, the system would use a series of electromagnetic coils to power a hardened capsule to hypersonic velocities. The architecture has a downside: extremely high accelerations that could damage or destroy normal satellites. Instead, Moonshot wants to use the technology to send raw materials to orbit, lowering the input costs of the budding in-space servicing, refueling, and manufacturing industries, according to Payload.

Out of the shadows … Moonshot Space emerged from stealth mode with this week’s fundraising announcement. The company’s near-term focus is on building a scaled-down electromagnetic accelerator capable of reaching Mach 6. A larger system would be required to reach orbital velocity. The company’s CEO is the former director-general of Israel’s Ministry of Science, while its chief engineer was the former chief systems engineer for David’s Sling, a critical part of Israel’s missile defense system. (submitted by EllPeaTea)

A blunder at Baikonur. A Soyuz rocket launched on November 27 carrying Roscosmos cosmonauts Sergei Kud-Sverchkov and Sergei Mikayev, as well as NASA astronaut Christopher Williams, for an eight-month mission to the International Space Station. The trio of astronauts arrived at the orbiting laboratory without incident. However, on the ground, there was a serious problem during the launch with the ground systems that support processing of the vehicle before liftoff at Site 31, located at the Baikonur Cosmodrome in Kazakhstan, Ars reports. Roscosmos downplayed the incident, saying only, in passive voice, that “damage to several launch pad components was identified” following the launch.

Repairs needed … However, video imagery of the launch site after liftoff showed substantial damage, with a large service platform appearing to have fallen into the flame trench below the launch table. According to one source, this is a platform located beneath the rocket, where workers can access the vehicle before liftoff. It has a mass of about 20 metric tons and was apparently not secured prior to launch, and the thrust of the vehicle ejected it into the flame trench. “There is significant damage to the pad,” said this source. The damage could throw a wrench into Russia’s ability to launch crews and cargo to the International Space Station. This Soyuz launch pad at Baikonur is the only one outfitted to support such missions.

China’s LandSpace almost landed a rocket. China’s first attempt to land an orbital-class rocket may have ended in a fiery crash, but the company responsible for the mission had a lot to celebrate with the first flight of its new methane-fueled launcher, Ars reports. LandSpace, a decade-old company based in Beijing, launched its new Zhuque-3 rocket for the first time Tuesday (US time) at the Jiuquan launch site in northwestern China. The upper stage of the medium-lift rocket successfully reached orbit. This alone is a remarkable achievement for a new rocket. But LandSpace had other goals for this launch. The Zhuque-3, or ZQ-3, booster stage is architected for recovery and reuse, the first rocket in China with such a design. The booster survived reentry and was seconds away from a pinpoint landing when something went wrong during its landing burn, resulting in a high-speed crash at the landing zone in the Gobi Desert.

Let the games begin … LandSpace got closer to landing an orbital-class booster than any other company on their first try. While LandSpace prepares for a second launch, several more Chinese companies are close to debuting their own reusable rockets. The next of these new rockets, the Long March 12A, is awaiting its first liftoff later this month from another launch pad at the Jiuquan spaceport. The Long March 12A comes from one of China’s established rocket developers, the Shanghai Academy of Spaceflight Technology (SAST), part of the country’s state-owned aerospace enterprise.

China launches a lifeboat. An unpiloted Chinese spacecraft launched on November 24 (US time) and linked with the country’s Tiangong space station a few hours later, providing a lifeboat for three astronauts stuck in orbit without a safe ride home, Ars reports. A Long March 2F rocket lifted off with the Shenzhou 22 spacecraft, carrying cargo instead of a crew. The spacecraft docked with the Tiangong station nearly 250 miles (400 kilometers) above the Earth about three-and-a-half hours later. Shenzhou 22 will provide a ride home next year for three Chinese astronauts. Engineers deemed their primary lifeboat unsafe after finding a cracked window, likely from an impact with a tiny piece of space junk.

In record time … Chinese engineers worked fast to move up the launch of the Shenzhou 22, originally set to fly next year. The launch occurred just 16 days after officials decided they needed to send another spacecraft to the Tiangong station. Shenzhou 22 and its rocket were already in standby at the launch site, but teams had to fuel the spacecraft and complete assembly of the rocket, then roll the vehicle to the launch pad for final countdown preps. The rapid turnaround offers a “successful example for efficient emergency response in the international space industry,” the China Manned Space Agency said. “It vividly embodies the spirit of manned spaceflight: exceptionally hardworking, exceptionally capable, exceptionally resilient, and exceptionally dedicated.”

Another big name flirts with the launch industry. OpenAI chief executive Sam Altman has explored putting together funds to either acquire or partner with a rocket company, a move that would position him to compete with Elon Musk’s SpaceX, the Wall Street Journal reports. Altman reached out to at least one rocket maker, Stoke Space, in the summer, and the discussions picked up in the fall, according to people familiar with the talks. Among the proposals was for OpenAI to make a multibillion-dollar series of equity investments in the company and end up with a controlling stake. The talks are no longer active, people close to OpenAI told the Journal.

Here’s the reason … Altman has been interested in building data centers in space for some time, the Journal reports, suggesting that the insatiable demand for computing resources to power artificial-intelligence systems eventually could require so much power that the environmental consequences would make space a better option. Orbital data centers would allow companies to harness the power of the Sun to operate them. Alphabet’s Google is pursuing a similar concept in partnership with satellite operator Planet Labs. Jeff Bezos and Musk himself have also expressed interest in the idea. Outside of SpaceX and Blue Origin, Stoke Space seems to be a natural partner for such a project because it is one of the few companies developing a fully reusable rocket.

SpaceX gets green light for new Florida launch pad. SpaceX has the OK to build out what will be the primary launch hub on the Space Coast for its Starship and Super Heavy rocket, the most powerful launch vehicle in history, the Orlando Sentinel reports. The Department of the Air Force announced Monday it had approved SpaceX to move forward with the construction of a pair of launch pads at Cape Canaveral Space Force Station’s Space Launch Complex 37 (SLC-37). A “record of decision” on the Environmental Impact Statement required under the National Environmental Policy Act for the proposed Canaveral site was posted to the Air Force’s website, marking the conclusion of what has been a nearly two-year approval process.

Get those Starships ready SpaceX plans to build two launch towers at SLC-37 to augment the single tower under construction at NASA’s Kennedy Space Center, just a few miles to the north. The three pads combined could support up to 120 launches per year. The Air Force’s final approval was expected after it released a draft Environmental Impact Statement earlier this year, suggesting the Starship pads at SLC-37 would have no significant negative impacts on local environmental, historical, social, and cultural interests. The Air Force also found SpaceX’s plans at SLC-37, formerly leased by United Launch Alliance, will have no significant impact on the company’s competitors in the launch industry. SpaceX also has two launch towers at its Starbase facility in South Texas.

Next three launches

Dec. 5: Kuaizhou 1A | Unknown Payload | Jiuquan Satellite Launch Center, China | 09: 00 UTC

Dec. 6: Hyperbola 1 | Unknown Payload | Jiuquan Satellite Launch Center, China | 04: 00 UTC

Dec. 6: Long March 8A | Unknown Payload | Wenchang Space Launch Site, China | 07: 50 UTC

Photo of Stephen Clark

Stephen Clark is a space reporter at Ars Technica, covering private space companies and the world’s space agencies. Stephen writes about the nexus of technology, science, policy, and business on and off the planet.

Rocket Report: Blunder at Baikonur; do launchers really need rocket engines? Read More »

congress-warned-that-nasa’s-current-plan-for-artemis-“cannot-work”

Congress warned that NASA’s current plan for Artemis “cannot work”

As for what to do about it, Griffin said legislators should end the present plan.

“The Artemis III mission and those beyond should be canceled and we should start over, proceeding with all deliberate speed,” Griffin said. He included a link to his plan, which is not dissimilar from the “Apollo on Steroids” architecture he championed two decades ago, but was later found to be unaffordable within NASA’s existing budget.

“There need to be consequences”

Other panel members offered more general advice.

Clayton Swope, deputy director of the Aerospace Security Project for the Center for Strategic and International Studies, said NASA should continue to serve as an engine for US success in space and science. He cited the Commercial Lunar Payload Services program, which has stimulated a growing lunar industry. He also said NASA spending on basic research and development is a critical feedstock for US innovation, and a key advantage over the People’s Republic of China.

“When you’re looking at the NASA authorization legislation, look at it in a way where you are the genesis of that innovation ecosystem, that flywheel that really powers US national security and economic security, in a way that the PRC just can’t match,” Swope said. “Without science, we would never have had something like the Manhattan Project.”

Another witness, Dean Cheng of the Potomac Institute for Policy Studies, said NASA—and by extension Congress—must do a better job of holding itself and its contractors accountable.

Many of NASA’s major exploration programs, including the Orion spacecraft, Space Launch System rocket, and their ground systems, have run years behind schedule and billions of dollars over budget in the last 15 years. NASA has funded these programs with cost-plus contracts, so it has had limited ability to enforce deadlines with contractors. Moreover, Congress has more or less meekly gone along with the delays and continued funding the programs.

Cheng said that whatever priorities policymakers decide for NASA,  failing to achieve objectives should come with consequences.

“One, it needs to be bipartisan, to make very clear throughout our system that this is something that everyone is pushing for,” Cheng said of establishing priorities for NASA. “And two, that there are consequences, budgetary, legal, and otherwise, to the agency, to supplying companies. If they fail to deliver on time and on budget, that it will not be a ‘Well, okay, let’s try again next year.’ There need to be consequences.”

Congress warned that NASA’s current plan for Artemis “cannot work” Read More »

why-won’t-steam-machine-support-hdmi-21?-digging-in-on-the-display-standard-drama.

Why won’t Steam Machine support HDMI 2.1? Digging in on the display standard drama.

When Valve announced its upcoming Steam Machine hardware last month, some eagle-eyed gamers may have been surprised to see that the official spec sheet lists support for HDMI 2.0 output, rather than the updated, higher-bandwidth HDMI 2.1 standard introduced in 2017. Now, Valve tells Ars that, while the hardware itself actually supports HDMI 2.1, the company is struggling to offer full support for that standard due to Linux drivers that are “still a work-in-progress on the software side.”

As we noted last year, the HDMI Forum (which manages the official specifications for HDMI standards) has officially blocked any open source implementation of HDMI 2.1. That means the open source AMD drivers used by SteamOS can’t fully implement certain features that are specific to the updated output standard.

“At this time an open source HDMI 2.1 implementation is not possible without running afoul of the HDMI Forum requirements,” AMD engineer Alex Deucher said at the time.

Doing what they can

This situation has caused significant headaches for Valve, which tells Ars it has had to validate the Steam Machine’s HDMI 2.1 hardware via Windows during testing. And when it comes to HDMI performance via SteamOS, a Valve representative tells Ars that “we’ve been working on trying to unblock things there.”

That includes unblocking HDMI 2.0’s resolution and frame-rate limits, which max out at 60 Hz for a 4K output, according to the official standard. Valve tells Ars it has been able to increase that limit to the “4K @ 120Hz” listed on the Steam Machine spec sheet, though, thanks to a technique called chroma sub-sampling.

Why won’t Steam Machine support HDMI 2.1? Digging in on the display standard drama. Read More »

microsoft-drops-ai-sales-targets-in-half-after-salespeople-miss-their-quotas

Microsoft drops AI sales targets in half after salespeople miss their quotas

Microsoft has lowered sales growth targets for its AI agent products after many salespeople missed their quotas in the fiscal year ending in June, according to a report Wednesday from The Information. The adjustment is reportedly unusual for Microsoft, and it comes after the company missed a number of ambitious sales goals for its AI offerings.

AI agents are specialized implementations of AI language models designed to perform multistep tasks autonomously rather than simply responding to single prompts. So-called “agentic” features have been central to Microsoft’s 2025 sales pitch: At its Build conference in May, the company declared that it has entered “the era of AI agents.”

The company has promised customers that agents could automate complex tasks, such as generating dashboards from sales data or writing customer reports. At its Ignite conference in November, Microsoft announced new features like Word, Excel, and PowerPoint agents in Microsoft 365 Copilot, along with tools for building and deploying agents through Azure AI Foundry and Copilot Studio. But as the year draws to a close, that promise has proven harder to deliver than the company expected.

According to The Information, one US Azure sales unit set quotas for salespeople to increase customer spending on a product called Foundry, which helps customers develop AI applications, by 50 percent. Less than a fifth of salespeople in that unit met their Foundry sales growth targets. In July, Microsoft lowered those targets to roughly 25 percent growth for the current fiscal year. In another US Azure unit, most salespeople failed to meet an earlier quota to double Foundry sales, and Microsoft cut their quotas to 50 percent for the current fiscal year.

Microsoft drops AI sales targets in half after salespeople miss their quotas Read More »

on-dwarkesh-patel’s-second-interview-with-ilya-sutskever

On Dwarkesh Patel’s Second Interview With Ilya Sutskever

Some podcasts are self-recommending on the ‘yep, I’m going to be breaking this one down’ level. This was very clearly one of those. So here we go.

As usual for podcast posts, the baseline bullet points describe key points made, and then the nested statements are my commentary.

If I am quoting directly I use quote marks, otherwise assume paraphrases.

What are the main takeaways?

  1. Ilya thinks training in its current form will peter out, that we are returning to an age of research where progress requires more substantially new ideas.

  2. SSI is a research organization. It tries various things. Not having a product lets it punch well above its fundraising weight in compute and effective resources.

  3. Ilya has 5-20 year timelines to a potentially superintelligent learning model.

  4. SSI might release a product first after all, but probably not?

  5. Ilya’s thinking about alignment still seems relatively shallow to me in key ways, but he grasps many important insights and understands he has a problem.

  6. Ilya essentially despairs of having a substantive plan beyond ‘show everyone the thing as early and often as possible’ and hope for the best. He doesn’t know where to go or how to get there, but does realize he doesn’t know these things, so he’s well ahead of most others.

Afterwards, this post also covers Dwarkesh Patel’s post on the state of AI progress.

  1. Explaining Model Jaggedness.

  2. Emotions and value functions.

  3. What are we scaling?

  4. Why humans generalize better than models.

  5. Straight-shooting superintelligence.

  6. SSI’s model will learn from deployment.

  7. Alignment.

  8. “We are squarely an age of research company”.

  9. Research taste.

  10. Bonus Coverage: Dwarkesh Patel on AI Progress These Days.

  1. Ilya opens by remarking how crazy it is all this (as in AI) is real, it’s all so sci-fi, and yet it’s not felt in other ways so far. Dwarkesh expects this to continue for average people into the singularity, Ilya says no, AI will diffuse and be felt in the economy. Dwarkesh says impact seems smaller than model intelligence implies.

    1. Ilya is right here. Dwarkesh is right that direct impact so far has been smaller than model intelligence implies, but give it time.

  2. Ilya says, the models are really good at evals but economic impact lags. The models are buggy, and choices for RL take inspiration from the evals, so the evals are misleading and the humans are essentially reward hacking the evals. And that given they got their scores by studying for tons of hours rather than via intuition, one should expect AIs to underperform their benchmarks.

    1. AIs definitely underperform their benchmarks in terms of general usefulness, even for those companies that do minimal targeting of benchmarks. Overall capabilities lag behind, for various reasons. We still have an impact gap.

  3. The super talented student? The one that hardly even needs to practice a specific task to be good? They’ve got ‘it.’ Models don’t have ‘it.’

    1. If anything, models have ‘anti-it.’ They make it up on volume. Sure.

  1. Humans train on much less data, but what they know they know ‘more deeply’ somehow, there are mistakes we wouldn’t make. Also evolution can be highly robust, for example the famous case where a guy lost all his emotions and in many ways things remained fine.

    1. People put a lot of emphasis on the ‘I would never’ heuristic, as AIs will sometimes do things ‘a similarly smart person’ would never do, they lack a kind of common sense.

  2. So what is the ‘ML analogy for emotions’? Ilya says some kind of value function thing, as in the thing that tells you if you’re doing well versus badly while doing something.

    1. Emotions as value functions makes sense, but they are more information-dense than merely a scalar, and can often point you to things you missed. They do also serve as training reward signals.

    2. I don’t think you ‘need’ emotions for anything other than signaling emotions, if you are otherwise sufficiently aware in context, and don’t need them to do gradient descent.

    3. However in a human, if you knock out the emotions in places where you were otherwise relying on them for information or to resolve uncertainty, you’re going to have a big problem.

    4. I notice an obvious thing to try but it isn’t obvious how to implement it?

  3. Ilya has faith in deep learning. There’s nothing it can’t do!

  1. Data? Parameters? Compute? What else? It’s easier and more reliable to scale up pretraining than to figure out what else to do. But we’ll run out of data soon even if Gemini 3 got more out of this, so now you need to do something else. If you had 100x more scale here would anything be that different? Ilya thinks no.

    1. Sounds like a skill issue, on some level, but yes if you didn’t change anything else then I expect scaling up pretraining further won’t help enough to justify the increased costs in compute and time.

  2. RL costs now exceed pretraining costs, because each RL run costs a lot. It’s time to get back to an age of research, trying interesting things and seeing what happens.

    1. I notice I am skeptical of the level of skepticism, also I doubt the research mode ever stopped in the background. The progress will continue. It’s weird how every time someone says ‘we still need some new idea or breakthrough’ there is the implication that this likely never happens again.

  1. Why do AIs require so much more data than humans to learn? Why don’t models easily pick up on all this stuff humans learn one-shot or in the background?

    1. Humans have richer data than text so the ratio is not as bad as it looks, but primarily because our AI learning techniques are relatively primitive and data inefficient in various ways.

    2. My full answer to how to fix it falls under ‘I don’t do $100m/year jobs for free.’

    3. Also there are ways in which the LLMs learn way better than you realize, and a lot of the tasks humans easily learn are regularized in non-obvious ways.

  2. Ilya believes humans being good at learning is mostly not part of some complicated prior, and people’s robustness is really staggering.

    1. I would clarify, not part of a complicated specialized prior. There is also a complicated specialized prior in some key domains, but that is in addition to a very strong learning function.

    2. People are not as robust as Ilya thinks, or most people think.

  3. Ilya suggests perhaps human neurons use more compute than we think.

  1. Scaling ‘sucked the air out of the room’ so no one did anything else. Now there are more companies than ideas. You need some compute to bring ideas to life, but not the largest amounts.

    1. You can also think about some potential techniques as ‘this is not worth trying unless you have massive scale.’

  2. SSI’s compute all goes into research, none into inference, and they don’t try to build a product, and if you’re doing something different you don’t have to use maximum scale, so their $3 billion that they’ve raised ‘goes a long way’ relative to the competition. Sure OpenAI spends ~$5 billion a year on experiments, but it’s what you do with it.

    1. This is what Ilya has to say in this spot, but there’s merit in it. OpenAI’s experiments are largely about building products now. This transfers to the quest for superintelligence, but not super efficiently.

  3. How will SSI make money? Focus on the research, the money will appear.

    1. Matt Levine has answered this one, which is that you make money by being an AI company full of talented researchers, so people give you money.

  4. SSI is considering making a product anyway, both to have the product exist and also because timelines might be long.

    1. I mean I guess at some point the ‘we are AI researchers give us money’ strategy starts to look a little suspicious, but let’s not rush into anything.

    2. Remember, Ilya, once you have a product and try to have revenue they’ll evaluate the product and your revenue. If you don’t have one, you’re safe.

  1. Ilya says even if there is a straight shot to superintelligence deployment would be gradual, you have to ship something first, and that he agrees with Dwarkesh on the importance of continual learning, it would ‘go and be’ various things and learn, superintelligence is not a finished mind.

    1. Learning takes many forms, including continual learning, it can be updating within the mind or otherwise, and so on. See previous podcast discussions.

  2. Ilya expects ‘rapid’ economic growth, perhaps ‘very rapid.’ It will vary based on what rules are set in different places.

    1. Rapid means different things to different people, it sounds like Ilya doesn’t have a fixed rate in mind. I interpret it as ‘more than these 2% jokers.’

    2. This vision still seems to think the humans stay in charge. Why?

  1. Dwarkesh reprises the standard point that if AIs are merely ‘as good at’ humans at learning, but they can ‘merge brains’ then crazy things happen. How do we make such a situation go well? What is SSI’s plan?

    1. I mean, that’s the least of it, but hopefully yes that suffices to make the point?

  2. Ilya emphasizes deploying incrementally and in advance. It’s hard to predict what this will be like in advance. “The problem is the power. When the power is really big, what’s going to happen? If it’s hard to imagine, what do you do? You’ve got to be showing the thing.”

    1. This feels like defeatism, in terms of saying we can only respond to things once we can see and appreciate them. We can’t plan for being old until we know what that’s like. We can’t plan for AGI/ASI, or AI having a lot of power, until we can see that in action.

    2. But obviously by then it is likely to be too late, and most of your ability to steer what happens has already been lost, perhaps all of it.

    3. This is the strategy of ‘muddle through’ the same as we always muddle through, basically the plan of not having a plan other than incrementalism. I do not care for this plan. I am not happy to be a part of it. I do not think that is a case of Safe Superintelligence.

  3. Ilya expects governments and labs to play big roles, and for labs to increasingly coordinate on safety, as Anthropic and OpenAI did in a recent first step. And we have to figure out what we should be building. He suggests making the AI care about sentient life in general will be ‘easier’ than making it care about humans, since the AI will be sentient.

    1. If the AIs do not care about humans in particular, there is no reason to expect humans to stay in control or to long endure.

  4. Ilya would like the most powerful superintelligence to ‘somehow’ be ‘capped’ to address these concerns. But he doesn’t know how to do that.

    1. I don’t know how to do that either. It’s not clear the idea is coherent.

  5. Dwarkesh asks how much ‘room is there at the top’ for superintelligence to be more super? Maybe it just learns fast or has a bigger pool of strategies or skills or knowledge? Ilya says very powerful, for sure.

    1. Sigh. There is very obviously quite a lot of ‘room at the top’ and humans are not anything close to maximally intelligent, nor to getting most of what intelligence has to offer. At this point, the number of people who still don’t realize or accept this reinforces how much better a smarter entity could be.

  6. Ilya expects these superintelligences to be very large, as in physically large, and for several to come into being at roughly the same time, and ideally they could “be restrained in some ways or if there was some kind of agreement or something.”

    1. That agreement between AIs would then be unlikely to include us. Yes, functional restraints would be nice, but this is the level of thought that has gone into finding ways to do it.

    2. There’s been a lot of things staying remarkably close, but a lot of that is because rather than an edge compounding and accelerating for now catching up has been easier.

  7. Ilya: “What is the concern of superintelligence? What is one way to explain the concern? If you imagine a system that is sufficiently powerful, really sufficiently powerful—and you could say you need to do something sensible like care for sentient life in a very single-minded way—we might not like the results. That’s really what it is.”

    1. Well, yes, standard Yudkowsky, no fixed goal we can name turns out well.

  8. Ilya says maybe we don’t build an RL agent. Humans are semi-RL agents, our emotions make us alter our rewards and pursue different rewards after a while. If we keep doing what we are doing now it will soon peter out and never be “it.”

    1. There’s a baked in level of finding innovations and improvements that should be in anyone’s ‘keep doing what we are doing’ prior, and I think it gets us pretty far and includes many individually low-probability-of-working innovations making substantial differences. There is some level on which we would ‘peter out’ without a surprise, but it’s not clear that this requires being surprised overall.

    2. Is it possible things do peter out and we never see ‘it’? Yeah. It’s possible. I think it’s a large underdog to stay that way for long, but it’s possible. Still a long practical way to go even then.

    3. Emotions, especially boredom and the fading of positive emotions on repetition, are indeed one of the ways we push ourselves towards exploration and variety. That’s one of many things they do, and yes if we didn’t have them then we would need something else to take their place.

    4. In many cases I have indeed used logic to take the place of that, when emotion seems to not be sufficiently preventing mode collapse.

  9. “One of the things that you could say about what causes alignment to be difficult is that your ability to learn human values is fragile. Then your ability to optimize them is fragile. You actually learn to optimize them. And can’t you say, “Are these not all instances of unreliable generalization?” Why is it that human beings appear to generalize so much better? What if generalization was much better? What would happen in this case? What would be the effect? But those questions are right now still unanswerable.”

    1. It is cool to hear Ilya restate these Yudkowsky 101 things.

    2. Humans do not actually generalize all that well.

  10. How does one think about what AI going well looks like? Ilya goes back to ‘AI that cares for sentient life’ as a first step, but then asks the better question, what is the long run equilibrium? He notices he does not like his answer. Maybe each person has an AI that will do their bidding and that’s good, but the downside is then the AI does things like earn money or advocate or whatever, and the person says ‘keep it up’ but they’re not a participant. Precarious. People become part AI, Neurolink++. He doesn’t like this solution, but it is at least a solution.

    1. Big points for acknowledging that there are no known great solutions.

    2. Big points for pointing out one big flaw, that the people stop actually doing the things, because the AIs do the things better.

    3. The equilibrium here is that increasingly more things are turned over to AIs, including both actions and decisions. Those who don’t do this fall behind.

    4. The equilibrium here is that increasingly AIs are given more autonomy, more control, put in better positions, have increasing power and wealth shares, and so on, even if everything involved is fully voluntary and ‘nothing goes wrong.’

    5. Neurolink++ does not meaningfully solve any of the problems here.

    6. Solve for the equilibrium.

  11. Is the long history of emotions an alignment success? As in, it allows the brain to move from ‘mate with somebody who’s more successful’ into flexibly defining success and generally adjusting to new situations.

    1. It’s a highly mixed bag, wouldn’t you say?

    2. There are ways in which those emotions have been flexible and adaptable and a success, and have succeeded in the alignment target (inclusive genetic fitness) and also ways in which emotions are very obviously failing people.

    3. If ASIs are about as aligned as we are in this sense, we’re doomed.

  12. Ilya says it’s mysterious how evolution encodes high-level desires, but it gives us all these social desires, and they evolved pretty recently. Dwarkesh points out it is desire you learned in your lifetime. Ilya notes the brain as regions and some things are hardcoded, but if you remove half the brain then the regions move, the social stuff is highly reliable.

    1. I don’t pretend to understand the details here, although I could speculate.

  1. SSI investigates ideas to see if they are promising. They do research.

  2. On his cofounder leaving: “For this, I will simply remind a few facts that may have been forgotten. I think these facts which provide the context explain the situation. The context was that we were fundraising at a $32 billion valuation, and then Meta came in and offered to acquire us, and I said no. But my former cofounder in some sense said yes. As a result, he also was able to enjoy a lot of near-term liquidity, and he was the only person from SSI to join Meta.”

    1. I love the way he put that. Yes.

  3. “The main thing that distinguishes SSI is its technical approach. We have a different technical approach that I think is worthy and we are pursuing it. I maintain that in the end there will be a convergence of strategies. I think there will be a convergence of strategies where at some point, as AI becomes more powerful, it’s going to become more or less clearer to everyone what the strategy should be. It should be something like, you need to find some way to talk to each other and you want your first actual real superintelligent AI to be aligned and somehow care for sentient life, care for people, democratic, one of those, some combination thereof. I think this is the condition that everyone should strive for. That’s what SSI is striving for. I think that this time, if not already, all the other companies will realize that they’re striving towards the same thing. We’ll see. I think that the world will truly change as AI becomes more powerful. I think things will be really different and people will be acting really differently.”

    1. This is a remarkably shallow, to me, vision of what the alignment part of the strategy looks like, but it does get an admirably large percentage of the overall strategic vision, as in most of it?

    2. The idea that ‘oh as we move farther along people will get more responsible and cooperate more’ seems to not match what we have observed so far, alas.

    3. Ilya later clarifies he specifically meant convergence on alignment strategies, although he also expects convergence on technical strategies.

    4. The above statement is convergence on an alignment goal, but that doesn’t imply convergence on alignment strategy. Indeed it does not imply that an alignment strategy that is workable even exists.

  4. Ilya’s timeline to the system that can learn and become superhuman? 5-20 years.

  5. Ilya predicts that when someone releases the thing that will be information but it won’t teach others how to do the thing, although they will eventually learn.

  6. What is the ‘good world’? We have powerful human-like learners and perhaps narrow ASIs, and companies make money, and there is competition through specialization, different niches. Accumulated learning and investment creates specialization.

    1. This is so frustrating, in that it doesn’t explain why you would expect that to be how this plays out, or why this world turns out well, or anything really? Which would be fine if the answers were clear or at least those seemed likely, but I very much don’t think that.

    2. This feels like a claim that humans are indeed near the upper limit of what intelligence can do and what can be learned except that we are hobbled in various ways and AIs can be unhobbled, but that still leaves them functioning in ways that seem recognizably human and that don’t crowd us out? Except again I don’t think we should expect this.

  7. Dwarkesh points out current LLMs are similar, Ilya says perhaps the datasets are not as non-overlapping as they seem.

    1. On the contrary, I was assuming they were mostly the same baseline data, and then they do different filtering and progressions from there? Not that there’s zero unique data but that most companies have ‘most of the data.’

  8. Dwarkesh suggests, therefore AIs will have less diversity than human teams. How can we get ‘meaningful diversity’? Ilya says this is because of pretraining, that post training is different.

    1. To the extent that such ‘diversity’ is useful it seems easy to get with effort. I suspect this is mostly another way to create human copium.

  9. What about using self-play? Ilya notes it allows using only compute, which is very interesting, but it is only good for ‘developing a certain set of skills.’ Negotiation, conflict, certain social strategies, strategizing, that kind of stuff. Then Ilya self-corrects, notes other forms, like debate, prover-verifier or forms of LLM-as-a-judge, it’s a special case of agent competition.

    1. I think there’s a lot of promising unexplored space here, decline to say more.

  1. What is research taste? How does Ilya come up with many big ideas?

This is hard to excerpt and seems important, so quoting in full to close out:

I can comment on this for myself. I think different people do it differently. One thing that guides me personally is an aesthetic of how AI should be, by thinking about how people are, but thinking correctly. It’s very easy to think about how people are incorrectly, but what does it mean to think about people correctly?

I’ll give you some examples. The idea of the artificial neuron is directly inspired by the brain, and it’s a great idea. Why? Because you say the brain has all these different organs, it has the folds, but the folds probably don’t matter. Why do we think that the neurons matter? Because there are many of them. It kind of feels right, so you want the neuron. You want some local learning rule that will change the connections between the neurons. It feels plausible that the brain does it.

The idea of the distributed representation. The idea that the brain responds to experience therefore our neural net should learn from experience. The brain learns from experience, the neural net should learn from experience. You kind of ask yourself, is something fundamental or not fundamental? How things should be.

I think that’s been guiding me a fair bit, thinking from multiple angles and looking for almost beauty, beauty and simplicity. Ugliness, there’s no room for ugliness. It’s beauty, simplicity, elegance, correct inspiration from the brain. All of those things need to be present at the same time. The more they are present, the more confident you can be in a top-down belief.

The top-down belief is the thing that sustains you when the experiments contradict you. Because if you trust the data all the time, well sometimes you can be doing the correct thing but there’s a bug. But you don’t know that there is a bug. How can you tell that there is a bug? How do you know if you should keep debugging or you conclude it’s the wrong direction? It’s the top-down. You can say things have to be this way. Something like this has to work, therefore we’ve got to keep going. That’s the top-down, and it’s based on this multifaceted beauty and inspiration by the brain.

I need to think more about what causes my version of ‘research taste.’ It’s definitely substantially different.

That ends our podcast coverage, and enter the bonus section, which seems better here than in the weekly, as it covers many of the same themes.

Dwarkesh Patel offers his thoughts on AI progress these days, noticing that when we get the thing he calls ‘actual AGI’ things are going to get fucking crazy, but thinking that this is 10-20 years away from happening in full. Until then, he’s a bit skeptical of how many gains we can realize, but skepticism is highly relative here.

Dwarkesh Patel: I’m confused why some people have short timelines and at the same time are bullish on RLVR. If we’re actually close to a human-like learner, this whole approach is doomed.

… Either these models will soon learn on the job in a self directed way – making all this pre-baking pointless – or they won’t – which means AGI is not imminent. Humans don’t have to go through a special training phase where they need to rehearse every single piece of software they might ever use.

Wow, look at those goalposts move (in all the different directions). Dwarkesh notes that the bears keep shifting on the bulls, but says this is justified because current models fit the old goals but don’t score the points, as in they don’t automate workflows as much as you would expect.

In general, I worry about the expectation pattern having taken the form of ‘median 50 years → 20 → 10 → 5 → 7, and once I heard someone said 3, so oh nothing to see there you can stop worrying.’

In this case, look at the shift: An ‘actual’ (his term) AGI must now not only be capable of human-like performance of tasks, the AGI must also be a human-efficient learner.

That would mean AGI and ASI are the same thing, or at least arrive in rapid succession. An AI that was human-efficient at learning from data, combined with AI’s other advantages that include imbibing orders of magnitude more data, would be a superintelligence and would absolutely set off recursive self-improvement from there.

And yes, if that’s what you mean then AGI isn’t the best concept for thinking about timelines, and superintelligence is the better target to talk about. Sriram Krishnan is however opposed to using either of them.

Like all conceptual handles or fake frameworks, it is imprecise and overloaded, but people’s intuitions about it miss that the thing is possible or exists even when you outright say ‘superintelligence’ and I shudder to think how badly they will miss the concept if you don’t even say it. Which I think is a lot of the motivation behind not wanting to say it, so people can pretend that there won’t be things smarter than us in any meaningful sense and thus we can stop worrying about it or planning for it.

Indeed, this is exactly Sriram’s agenda if you look at his post here, to claim ‘we are not on the timeline’ that involves such things, to dismiss concerns as ‘sci-fi’ or philosophical, and talk instead of ‘what we are trying to build.’ What matters is what actually gets built, not what we intended, and no none of these concepts have been invalidated. We have ‘no proof of takeoff’ in the sense that we are not currently in a fast takeoff yet, but what would constitute this ‘proof’ other than already being in a takeoff, and thus it being too late to do anything about it?

Sriram Krishnan: …most importantly, it invokes fear—connected to historical usage in sci-fi and philosophy (think 2001, Her, anything invoking the singularity) that has nothing to do with the tech tree we’re actually on. Makes every AI discussion incredibly easy to anthropomorphize and detour into hypotheticals.

Joshua Achiam (OpenAI Head of Mission Alignment): I mostly disagree but I think this is a good contribution to the discourse. Where I disagree: I do think AGI and ASI both capture something real about where things are going. Where I agree: the lack of agreed-upon definitions has 100% created many needless challenges.

The idea that ‘hypotheticals,’ as in future capabilities and their logical consequences, are ‘detours,’ or that any such things are ‘sci-fi or philosophy’ is to deny the very idea of planning for future capabilities or thinking about the future in real ways. Sriram himself only thinks they are 10 years away, and then the difference is he doesn’t add Dwarkesh’s ‘and that’s fucking crazy’ and instead seems to effectively say ‘and that’s a problem for future people, ignore it.’

Seán Ó hÉigeartaigh: I keep noting this, but I do think a lot of the most heated policy debates we’re having are underpinned by a disagreement on scientific view: whether we (i) are on track in coming decade for something in the AGI/ASI space that can achieve scientific feats equivalent to discovering general relativity (Hassabis’ example), or (ii) should expect AI as a normal technology (Narayanan & Kapoor’s definition).

I honestly don’t know. But it feels premature to me to rule out (i) on the basis of (slightly) lengthening timelines from the believers, when progress is clearly continuing and a historically unprecedented level of resources are going into the pursuit of it. And premature to make policy on the strong expectation of (ii). (I also think it would be premature to make policy on the strong expectation of (i) ).

But we are coming into the time where policy centred around worldview (ii) will come into tension in various places with the policies worldview (i) advocates would enact if given a free hand. Over the coming decade I hope we can find a way to navigate a path between, rather than swing dramatically based on which worldview is in the ascendancy at a given time.

Sriram Krishnan: There is truth to this.

This paints it as two views, and I would say you need at least three:

  1. Something in the AGI/ASI space is likely in less than 10 years.

  2. Something in the AGI/ASI space is unlikely in less than about 10 years, but highly plausible in 10-20 years, until then AI is a normal technology.

  3. AI is a normal technology and we know it will remain so indefinitely. We can regulate and plan as if AGI/ASI style technologies will never happen.

I think #1 and #2 are both highly reasonable positions, only #3 is unreasonable, while noting that if you believe #2 you still need to put some non-trivial weight on #1. As in, if you think it probably takes ~10 years then you can perhaps all but rule out AGI 2027, and you think 2031 is unlikely, but you cannot claim 2031 is a Can’t Happen.

The conflation to watch out for is #2 and #3. These are very different positions. Yet many in the AI industry, and its political advocates, make exactly this conflation. They assert ‘#1 is incorrect therefore #3,’ when challenged for details articulate claim #2, then go back to trying to claim #3 and act on the basis of #3.

What’s craziest is that the list of things to rule out, chosen by Sriram, includes the movie Her. Her made many very good predictions. Her was a key inspiration for ChatGPT and its voice mode, so much so that there was a threatened lawsuit because they all but copied Scarlett Johansson’s voice. She’s happening. Best be believing in sci-fi stores, because you’re living in one, and all that.

Nothing about current technology is a reason to think 2001-style things or a singularity will not happen, or to think we should anthropomorphize AI relatively less (the correct amount for current AIs, and for future AIs, are both importantly not zero, and importantly not 100%, and both mistakes are frequently made). Indeed, Dwarkesh is de facto predicting a takeoff and a singularity in this post that Sriram praised, except Dwarkesh has it on a 10-20 year timescale to get started.

Now, back to Dwarkesh.

This process of ‘teach the AI the specific tasks people most want’ is the central instance of models being what Teortaxes calls usemaxxed. A lot of effort is going to specific improvements rather than to advancing general intelligence. And yes, this is evidence against extremely short timelines. It is also, as Dwarkesh notes, evidence in favor of large amounts of mundane utility soon, including ability to accelerate R&D. What else would justify such massive ‘side’ efforts?

There’s also, as he notes, the efficiency argument. Skills many people want should be baked into the core model. Dwarkesh fires back that there are a lot of skills that are instance-specific and require on-the-job or continual learning, which he’s been emphasizing a lot for a while. I continue to not see a contradiction, or why it would be that hard to store and make available that knowledge as needed even if it’s hard for the LLM to permanently learn it.

I strongly disagree with his claim that ‘economic diffusion lag is cope for missing capabilities.’ I agree that many highly valuable capabilities are missing. Some of them are missing due to lack of proper scaffolding or diffusion or context, and are fundamentally Skill Issues by the humans. Others are foundational shortcomings. But the idea that the AIs aren’t up to vastly more tasks than they’re currently asked to do seems obviously wrong?

He quotes Steven Byrnes:

Steven Byrnes: New technologies take a long time to integrate into the economy? Well ask yourself: how do highly-skilled, experienced, and entrepreneurial immigrant humans manage to integrate into the economy immediately? Once you’ve answered that question, note that AGI will be able to do those things too.

Again, this is saying that AGI will be as strong as humans in the exact place it is currently weakest, and will not require adjustments for us to take advantage. No, it is saying more than that, it is also saying we won’t put various regulatory and legal and cultural barriers in its way, either, not in any way that counts.

If the AGI Dwarkesh is thinking about were to exist, again, it would be an ASI, and it would be all over for the humans very quickly.

I also strongly disagree with human labor not being ‘shleppy to train’ (bonus points, however, for excellent use of ‘shleppy’). I have trained humans and been a human being trained, and it is totally shleppy. I agree, not as schleppy as current AIs can be when something is out of their wheelhouse, but rather obnoxiously schleppy everywhere except their own very narrow wheelhouse.

Here’s another example of ‘oh my lord check out those goalposts’:

Dwarkesh Patel: It revealed a key crux between me and the people who expect transformative economic impacts in the next few years.

Transformative economic impacts in the next few years would be a hell of a thing.

It’s not net-productive to build a custom training pipeline to identify what macrophages look like given the way this particular lab prepares slides, then another for the next lab-specific micro-task, and so on. What you actually need is an AI that can learn from semantic feedback on the job and immediately generalize, the way a human does.

Well, no, it probably isn’t now, but also Claude Code is getting rather excellent at creating training pipelines, and the whole thing is rather standard in that sense, so I’m not convinced we are that far away from doing exactly that. This is an example of how sufficient ‘AI R&D’ automation, even on a small non-recursive scale, can transform use cases.

Every day, you have to do a hundred things that require judgment, situational awareness, and skills & context learned on the job. These tasks differ not just across different people, but from one day to the next even for the same person. It is not possible to automate even a single job by just baking in some predefined set of skills, let alone all the jobs.

Well, I mean of course it is, for a sufficiently broad set of skills at a sufficiently high level, especially if this includes meta-skills and you can access additional context. Why wouldn’t it be? It certainly can quickly automate large portions of many jobs, and yes I have started to automate portions of my job indirectly (as in Claude writes me the mostly non-AI tools to do it, and adjusts them every time they do something wrong).

Give it a few more years, though, and Dwarkesh is on the same page as I am:

In fact, I think people are really underestimating how big a deal actual AGI will be because they’re just imagining more of this current regime. They’re not thinking about billions of human-like intelligences on a server which can copy and merge all their learnings. And to be clear, I expect this (aka actual AGI) in the next decade or two. That’s fucking crazy!

Exactly. This ‘actual AGI’ is fucking crazy, and his timeline for getting there of 10-20 years is also fucking crazy. More people need to add ‘and that’s fucking crazy’ at the end of such statements.

Dwarkesh then talks more about continual learning. His position here hasn’t changed, and neither has my reaction that this isn’t needed, we can get the benefits other ways. He says that the gradual progress on continual learning means it won’t be ‘game set match’ to the first mover, but if this is the final piece of the puzzle then why wouldn’t it be?

Discussion about this post

On Dwarkesh Patel’s Second Interview With Ilya Sutskever Read More »

openai-ceo-declares-“code-red”-as-gemini-gains-200-million-users-in-3-months

OpenAI CEO declares “code red” as Gemini gains 200 million users in 3 months

In addition to buzz about Gemini on social media, Google is quickly catching up to ChatGPT in user numbers. ChatGPT has more than 800 million weekly users, according to OpenAI, while Google’s Gemini app has grown from 450 million monthly active users in July to 650 million in October, according to Business Insider.

Financial stakes run high

Not everyone views OpenAI’s “code red” as a genuine alarm. Reuters columnist Robert Cyran wrote on Tuesday that OpenAI’s announcement added “to the impression that OpenAI is trying to do too much at once with technology that still requires a great deal of development and funding.” On the same day Altman’s memo circulated, OpenAI announced an ownership stake in a Thrive Capital venture and a collaboration with Accenture. “The only thing bigger than the company’s attention deficit is its appetite for capital,” Cyran wrote.

In fact, OpenAI faces an unusual competitive disadvantage: Unlike Google, which subsidizes its AI ventures through search advertising revenue, OpenAI does not turn a profit and relies on fundraising to survive. According to The Information, the company, now valued at around $500 billion, has committed more than $1 trillion in financial obligations to cloud computing providers and chipmakers that supply the computing power needed to train and run its AI models.

But the tech industry never stands still, and things can change quickly. Altman’s memo also reportedly stated that OpenAI plans to release a new simulated reasoning model next week that may beat Gemini 3 in internal evaluations. In AI, the back-and-forth cycle of one-upmanship is expected to continue as long as the dollars keep flowing.

OpenAI CEO declares “code red” as Gemini gains 200 million users in 3 months Read More »

india-orders-device-makers-to-put-government-run-security-app-on-all-phones

India orders device makers to put government-run security app on all phones

Consumers can also use the app or website to check the number of mobile connections in their name and report any that appear to be fraudulent.

Priyanka Gandhi of the Congress Party, a member of Parliament, said that Sanchar Saathi “is a snooping app… It’s a very fine line between ‘fraud is easy to report’ and ‘we can see everything that every citizen of India is doing on their phone.’” She called for an effective system to fight fraud, but said that cybersecurity shouldn’t be “an excuse to go into every citizen’s telephone.”

App may need “root level access”

Despite Scindia saying the app can be deleted by users, the government statement that phone makers must ensure its functionalities are not “disabled or restricted” raised concerns about the level of access it requires. While the app store version can be deleted, privacy advocates say the order’s text indicates the pre-installed version would require deeper integration into the device.

The Internet Freedom Foundation, an Indian digital rights advocacy group, said the government directive “converts every smartphone sold in India into a vessel for state mandated software that the user cannot meaningfully refuse, control, or remove. For this to work in practice, the app will almost certainly need system level or root level access, similar to carrier or OEM system apps, so that it cannot be disabled. That design choice erodes the protections that normally prevent one app from peering into the data of others, and turns Sanchar Saathi into a permanent, non-consensual point of access sitting inside the operating system of every Indian smartphone user.”

The group said that while the app is being “framed as a benign IMEI checker,” a server-side update could repurpose it to perform “client side scanning for ‘banned’ applications, flag VPN usage, correlate SIM activity, or trawl SMS logs in the name of fraud detection. Nothing in the order constrains these possibilities.”

India orders device makers to put government-run security app on all phones Read More »

testing-shows-why-the-steam-machine’s-8gb-of-graphics-ram-could-be-a-problem

Testing shows why the Steam Machine’s 8GB of graphics RAM could be a problem

By Valve’s admission, its upcoming Steam Machine desktop isn’t swinging for the fences with its graphical performance. The specs promise decent 1080p-to-1440p performance in most games, with 4K occasionally reachable with assistance from FSR upscaling—about what you’d expect from a box with a modern midrange graphics card in it.

But there’s one spec that has caused some concern among Ars staffers and others with their eyes on the Steam Machine: The GPU comes with just 8GB of dedicated graphics RAM, an amount that is steadily becoming more of a bottleneck for midrange GPUs like AMD’s Radeon RX 7060 and 9060, or Nvidia’s GeForce RTX 4060 or 5060.

In our reviews of these GPUs, we’ve already run into some games where the RAM ceiling limits performance in Windows, especially at 1440p. But we’ve been doing more extensive testing of various GPUs with SteamOS, and we can confirm that in current betas, 8GB GPUs struggle even more on SteamOS than they do running the same games at the same settings in Windows 11.

The good news is that Valve is working on solutions, and having a stable platform like the Steam Machine to aim for should help improve things for other hardware with similar configurations. The bad news is there’s plenty of work left to do.

The numbers

We’ve tested an array of dedicated and integrated Radeon GPUs under SteamOS and Windows, and we’ll share more extensive results in another article soon (along with broader SteamOS-vs-Windows observations). But for our purposes here, the two GPUs that highlight the issues most effectively are the 8GB Radeon RX 7600 and the 16GB Radeon RX 7600 XT.

These dedicated GPUs have the benefit of being nearly identical to what Valve plans to ship in the Steam Machine—32 compute units (CUs) instead of Valve’s 28, but the same RDNA3 architecture. They’re also, most importantly for our purposes, pretty similar to each other—the same physical GPU die, just with slightly higher clock speeds and more RAM for the 7600 XT than for the regular 7600.

Testing shows why the Steam Machine’s 8GB of graphics RAM could be a problem Read More »

netflix-quietly-drops-support-for-casting-to-most-tvs

Netflix quietly drops support for casting to most TVs

Have you been trying to cast Stranger Things from your phone, only to find that your TV isn’t cooperating? It’s not the TV—Netflix is to blame for this one, and it’s intentional. The streaming app has recently updated its support for Google Cast to disable the feature in most situations. You’ll need to pay for one of the company’s more expensive plans, and even then, Netflix will only cast to older TVs and streaming dongles.

The Google Cast system began appearing in apps shortly after the original Chromecast launched in 2013. Since then, Netflix users have been able to start video streams on TVs and streaming boxes from the mobile app. That was vital for streaming targets without their own remote or on-screen interface, but times change.

Today, Google has moved beyond the remote-free Chromecast experience, and most TVs have their own standalone Netflix apps. Netflix itself is also allergic to anything that would allow people to share passwords or watch in a new place. Over the last couple of weeks, Netflix updated its app to remove most casting options, mirroring a change in 2019 to kill Apple AirPlay.

The company’s support site (spotted by Android Authority) now clarifies that casting is only supported in a narrow set of circumstances. First, you need to be paying for one of the ad-free service tiers, which start at $18 per month. Those on the $8 ad-supported plan won’t have casting support.

Even then, Casting only appears for devices without a remote, like the earlier generations of Google Chromecasts, as well as some older TVs with Cast built in. For example, anyone still rocking Google’s 3rd Gen Chromecast from 2018 can cast video in Netflix, but those with the 2020 Chromecast dongle (which has a remote and a full Android OS) will have to use the TV app. Essentially, anything running Android/Google TV or a smart TV with a full Netflix app will force you to log in before you can watch anything.

Netflix quietly drops support for casting to most TVs Read More »