Author name: Mike M.

google’s-pixel-8-series-gets-usb-c-to-displayport;-desktop-mode-rumors-heat-up

Google’s Pixel 8 series gets USB-C to DisplayPort; desktop mode rumors heat up

You would think a phone called “Pixel” would be better at this —

Grab a USB-C to DisplayPort cable and newer Pixels can be viewed from your TV or monitor.

The Pixel 8.

Enlarge / The Pixel 8.

Google

Google’s June Android update is out, and it’s bringing a few notable changes for Pixel phones. The most interesting is that the Pixel 8a, Pixel 8 and Pixel 8 Pro are all getting DisplayPort Alt Mode capabilities via their USB-C ports. This means you can go from USB-C to DisplayPort and plug right into a TV or monitor. This has been rumored forever and landed in some of the Android Betas earlier, but now it’s finally shipping out to production.

The Pixel 8’s initial display support is just a mirrored mode. You can either get an awkward vertical phone in the middle of your wide-screen display or turn the phone sideways and get a more reasonable layout. You could see it being useful for videos or presentations. It would be nice if it could do more.

Alongside this year-plus of display port rumors has been a steady drum beat (again) for an Android desktop mode. Google has been playing around with this idea since Android 7.0 in 2016. In 2019, we were told it was just a development testing project, and it never shipped to any real devices. Work around Android’s desktop mode has been heating up, though, so maybe a second swing at this idea will result in an actual product.

Android 15's in-development desktop mode.

Android 15’s in-development desktop mode.

Android Authority’s Mishaal Rahman has been tracking down the new desktop mode for a while now and now has it running. The new desktop mode looks just like a real desktop OS. Every app gets a title bar window decoration with an app icon, a label, and maximize and close buttons. You can drag windows around and resize them; the OS supports automatic window tiling by dragging to the side of the screen; and there’s even a little drop-down menu in the title bar app icon. If you were to turn that on with Tablet Android’s bottom app bar, you would have a lot of what you need for a desktop OS.

Just like last time, we’ve got no clue if this will turn into a real product. The biggest Android partner, Samsung, certainly seems to think the idea is worth doing. Samsung’s “DeX” desktop mode has been a feature for years on its devices.

DisplayPort support is part of the June 2024 update and should roll out to devices soon.

Google’s Pixel 8 series gets USB-C to DisplayPort; desktop mode rumors heat up Read More »

as-nasa-watches-starship-closely,-here’s-what-the-agency-wants-to-see-next

As NASA watches Starship closely, here’s what the agency wants to see next

Target and Chaser —

“What happens if I don’t have a Human Landing System available to execute a mission?”

The rocket for SpaceX's fourth full-scale Starship test flight awaits liftoff from Starbase, the company's private launch base in South Texas.

Enlarge / The rocket for SpaceX’s fourth full-scale Starship test flight awaits liftoff from Starbase, the company’s private launch base in South Texas.

SpaceX

Few people were happier with the successful outcome of last week’s test flight of SpaceX’s Starship launch system than a NASA engineer named Catherine Koerner.

In remarks after the spaceflight, Koerner praised the “incredible” video of the Starship rocket and its Super Heavy booster returning to Earth, with each making a soft landing. “That was very promising, and a very, very successful engineering test,” she added, speaking at a meeting of the Space Studies Board.

A former flight director, Koerner now manages development of the “exploration systems” that will support the Artemis missions for NASA—a hugely influential position within the space agency. This includes the Space Launch System rocket, NASA’s Orion spacecraft, spacesuits, and the Starship vehicle that will land on the Moon.

In recent months, NASA officials like Koerner have been grappling with the reality that not all of this hardware is likely to be ready for the planned September 2026 launch date for the Artemis III mission. In particular, the agency is concerned about Starship’s readiness as a “Human Landing System.” While SpaceX is pressing forward rapidly with a test campaign, there is still a lot of work to be done to get the vehicle down to the lunar surface and safely back into lunar orbit.

A spare tire

For these reasons, as Ars previously reported, NASA and SpaceX are planning for the possibility of modifying the Artemis III mission. Instead of landing on the Moon, a crew would launch in the Orion spacecraft and rendezvous with Starship in low-Earth orbit. This would essentially be a repeat of the Apollo 9 mission, buying down risk and providing a meaningful stepping stone between Artemis missions.

Officially, NASA maintains that the agency will fly a crewed lunar landing, the Artemis III mission, in September 2026. But almost no one in the space community regards that launch date as more than aspirational. Some of my best sources have put the most likely range of dates for such a mission from 2028 to 2032. A modified Artemis III mission, in low-Earth orbit, would therefore bridge a gap between Artemis II and an eventual landing.

Koerner has declined interview requests from Ars to discuss this, but during the Space Studies Board, she acknowledged seeing these reports on modifying Artemis III. She was then asked directly whether there was any validity to them. Here is her response in full:

So here’s what I’ll tell you, if you’ll permit me an analogy. I have in my car a spare tire, right? I don’t have a spare steering wheel. I don’t have spare windshield wipers. I have a spare tire. And why? Why do we carry a spare tire? That someone, at some point, did an assessment and said in order for this vehicle to accomplish its mission, there is a certain likelihood that some things may fail and a certain likelihood that other things may not fail, and it’s probably prudent to have a spare tire. I don’t necessarily need to have a spare steering wheel, right?

We at NASA do a lot of those kinds of assessments. Like, what happens if this isn’t available? What happens if that isn’t available? Do we have backup plans for that? We’re always doing those kinds of backup plans. Do we have backup plans? It’s imperative for me to look at what happens if an Orion spacecraft is not ready to do a mission. What happens if I don’t have an SLS ready to do a mission? What happens if I don’t have a Human Landing System available to execute a mission? What happens if I don’t have Gateway that I was planning on to do a mission?

So we look at backup plans all the time. There are lots of different opportunities for that. We have not made any changes to the current plan as I outlined it here today and talked about that. But we have lots of people who are looking at lots of different backup plans so that we are doing due diligence and making sure that we have the spare tire if we need the spare tire. It’s the reason we have, for example, two systems now that we’re developing for the Human Landing System, the one for SpaceX and the other one from Blue Origin. It’s the reason we have two providers that are building spacesuit hardware. Collins as well as Axiom, right? So we always are doing that kind of thing.

That is a long way of saying that if SpaceX’s Starship is not ready in 2026, NASA is actively considering alternative plans. (The most likely of these would be an Orion-Starship docking in low-Earth orbit.) NASA has not made any final plans and is waiting to see how Artemis II progresses and what happens with Starship and spacesuit development.

What SpaceX needs to demonstrate

During her remarks, Koerner was also asked what SpaceX’s next major milestone is and when it would need to be completed for NASA to remain on track for a lunar landing in 2026. “Their next big milestone test, from a contract perspective, is the cryogenic transfer test,” she said. “That is going to be early next year.”

Some details about the Starship propellant transfer test.

Enlarge / Some details about the Starship propellant transfer test.

NASA

This timeline is consistent with what NASA’s Human Landing System program manager, Lisa Watson-Morgan recently told Ars. It provides a useful benchmark to evaluate Starship’s progress in NASA’s eyes. The “prop transfer demo” is a fairly complex mission that involves the launch of a “Starship target” from the Starbase facility in South Texas. Then a second vehicle, the “Starship chaser,” will launch and meet the target in orbit and rendezvous. The chaser will then transfer a quantity of propellant to the target spaceship.

The test will entail a lot of technology, including docking mechanisms, navigation sensors, quick disconnects, and more. If SpaceX completes this test during the first quarter of 2025, NASA will at least theoretically have a path forward to a crewed lunar landing in 2026.

As NASA watches Starship closely, here’s what the agency wants to see next Read More »

stoke-space-ignites-its-ambitious-main-engine-for-the-first-time

Stoke Space ignites its ambitious main engine for the first time

Get stoked! —

“This industry is going toward full reusability. To me, that is the inevitable end state.”

A drone camera captures the hotfire test of Stoke Space's full-flow staged combustion engine at the company's testing facility in early June.

Enlarge / A drone camera captures the hotfire test of Stoke Space’s full-flow staged combustion engine at the company’s testing facility in early June.

Stoke Space

On Tuesday, Stoke Space announced the firing of its first stage rocket engine for the first time earlier this month, briefly igniting it for about two seconds. The company declared the June 5 test a success because the engine performed nominally and will be fired up again soon.

“Data point one is that the engine is still there,” said Andy Lapsa, chief executive of the Washington-based launch company, in an interview with Ars.

The test took place at the company’s facilities in Moses Lake, Washington. Seven of these methane-fueled engines, each intended to have a thrust of 100,000 pounds of force, will power the company’s Nova rocket. This launch vehicle will have a lift capacity of about 5 metric tons to orbit. Lapsa declined to declare a target launch date, but based on historical developmental programs, if Stoke continues to move fast, it could fly Nova for the first time in 2026.

Big ambitions for a small company

Although it remains relatively new in the field of emerging launch companies, Stoke has gathered a lot of attention because of its bold ambitions. The company intends for the two-stage Nova rocket to be fully reusable, with both stages returning to Earth. To achieve a vertical landing, the second stage has a novel design. This oxygen-hydrogen engine is based on a ring of 30 thrusters and a regeneratively cooled heat shield.

Lapsa and Stoke, which now has 125 employees, have also gone for an ambitious design in the first-stage engine tested earlier this month. The engine, with a placeholder name of S1E, is based on full-flow, stage-combustion technology in which the liquid propellants are burned in the engine’s pre-burners. Because of this, they arrive in the engine’s combustion chamber in fully gaseous form, leading to a more efficient mixing.

Such an engine—this technology has only previously been demonstrated in flight by SpaceX’s Raptor engine, on the Starship rocket—is more efficient and should theoretically extend turbine life. But it is also technically demanding to develop, and among the most complex engine designs for a rocket company to begin with. This is not rocket science. It’s exceptionally hard rocket science.

It may seem like Stoke is biting off a lot more than it can chew with Nova’s design. Getting to space is difficult enough for a launch startup, but this company is seeking to build a fully reusable rocket with a brand new second stage design and a first stage engine based on full-flow, staged combustion. I asked Lapsa if he was nuts for taking all of this on.

Are these guys nuts?

“I’ve been around long enough to know that any rocket development program is hard, even if you make it as simple as possible,” he responded. “But this industry is going toward full reusability. To me, that is the inevitable end state. When you start with that north star, any other direction you take is a diversion. If you start designing anything else, it’s not something where you can back into full reusability at any point. It means you’ll have to stop and start over to climb the mountain.”

This may sound like happy talk, but Stoke appears to be delivering on its ambitions. Last September, the company completed a successful “hop” test of its second stage at Moses Lake. This validated its design, thrust vector control, and avionics.

This engine is designed to power the Nova rocket.

Enlarge / This engine is designed to power the Nova rocket.

Stoke Space

After this test, the company turned its focus to developing the S1E engine and put it on the test stand for the first time in April before the first test firing in June. Going from zero to 350,000 horsepower in half a second for the first time had a “pretty high pucker factor,” Lapsa said of the first fully integrated engine test.

Now that this initial test is complete, Stoke will spend the rest of the year maturing the design of the engine, conducting longer test firings, and starting to develop flight stages. After that will come stage tests before the complete Nova vehicle is assembled. At the same time, Stoke is also working with the US Space Force on the regulatory process of refurbishing and modernizing Launch Complex 14 at Cape Canaveral Space Force Station in Florida.

Stoke Space ignites its ambitious main engine for the first time Read More »

apple’s-ai-promise:-“your-data-is-never-stored-or-made-accessible-by-apple”

Apple’s AI promise: “Your data is never stored or made accessible by Apple”

…and throw away the key —

And publicly reviewable server code means experts can “verify this privacy promise.”

Apple Senior VP of Software Engineering Craig Federighi announces

Enlarge / Apple Senior VP of Software Engineering Craig Federighi announces “Private Cloud Compute” at WWDC 2024.

Apple

With most large language models being run on remote, cloud-based server farms, some users have been reluctant to share personally identifiable and/or private data with AI companies. In its WWDC keynote today, Apple stressed that the new “Apple Intelligence” system it’s integrating into its products will use a new “Private Cloud Compute” to ensure any data processed on its cloud servers is protected in a transparent and verifiable way.

“You should not have to hand over all the details of your life to be warehoused and analyzed in someone’s AI cloud,” Apple Senior VP of Software Engineering Craig Federighi said.

Trust, but verify

Part of what Apple calls “a brand new standard for privacy and AI” is achieved through on-device processing. Federighi said “many” of Apple’s generative AI models can run entirely on a device powered by an A17+ or M-series chips, eliminating the risk of sending your personal data to a remote server.

When a bigger, cloud-based model is needed to fulfill a generative AI request, though, Federighi stressed that it will “run on servers we’ve created especially using Apple silicon,” which allows for the use of security tools built into the Swift programming language. The Apple Intelligence system “sends only the data that’s relevant to completing your task” to those servers, Federighi said, rather than giving blanket access to the entirety of the contextual information the device has access to.

And Apple says that minimized data is not going to be saved for future server access or used to further train Apple’s server-based models, either. “Your data is never stored or made accessible by Apple,” Federighi said. “It’s used exclusively to fill your request.”

But you don’t just have to trust Apple on this score, Federighi claimed. That’s because the server code used by Private Cloud Compute will be publicly accessible, meaning that “independent experts can inspect the code that runs on these servers to verify this privacy promise.” The entire system has been set up cryptographically so that Apple devices “will refuse to talk to a server unless its software has been publicly logged for inspection.”

While the keynote speech was light on details for the moment, the focus on privacy during the presentation shows that Apple is at least prioritizing security concerns in its messaging as it wades into the generative AI space for the first time. We’ll see what security experts have to say when these servers and their code are made publicly available in the near future.

Apple’s AI promise: “Your data is never stored or made accessible by Apple” Read More »

ipados-18-adds-machine-learning-wizardry-with-handwriting,-math-features

iPadOS 18 adds machine-learning wizardry with handwriting, math features

WWDC 2024 —

Also coming: new SharePlay features and a new “tab bar” for first-party apps.

  • The Calculator app is finally coming to iPad.

    Samuel Axon

  • You’ll be able to write out expressions with the Apple Pencil and see them solved in real time.

    Samuel Axon

CUPERTINO, Calif.—After going into detail about iOS 18, Apple took a few moments in its WWDC 2024 keynote to walk through some changes.

There are a few minor UI changes and new features across Apple’s first party apps. That includes a new floating tab bar. The bar expands into the side bar when you want to dig in, and you can customize the tab bar to include the specific things you want to interact with the most. Additionally, SharePlay allows easier screen sharing and remote control of another person’s iPad.

But the big news is that the Calculator app we’ve all used on the iPhone to the iPad, after years of the iPad having no first-party calculator app at all. The iPad Calculator app can do some things the iPhone version can’t do with the Apple Pencil; a feature called Math Notes can write out expressions like you would on a piece of paper, and the app will solve the expressions live as you scribble them—plus various other cool live-updating math features. (These new Math Notes features work in the Notes app, too.)

Apple didn’t use the word AI here, but this is surely driven by machine learning in some way. Doubly so for a new handwriting feature called Smart Script, which refines and improves your handwriting as you go, tweaking letters to make them more legible when you’re writing very quickly to take notes. It uses machine learning to analyze your handwriting, so these adjustments are meant to match your normal script. That means you can scribble as quickly and recklessly as you want during a conference or a day of classes, but ostensibly, it will be legible at the end of the day.

Not everyone’s a big Pencil user—for some of us, handwriting long ago took a back seat to typing—but Apple is aggressively selling these kinds of flashy features for those who want that experience.

The release date for iPadOS 18 hasn’t been announced yet, but it will likely arrive in September or October alongside iOS 18 and the new iPhone models that will probably be announced then.

Listing image by Samuel Axon

iPadOS 18 adds machine-learning wizardry with handwriting, math features Read More »

bird-flu-virus-from-texas-human-case-kills-100%-of-ferrets-in-cdc-study

Bird flu virus from Texas human case kills 100% of ferrets in CDC study

Animal study —

H5N1 bird flu viruses have shown to be lethal in ferret model before.

Bird flu virus from Texas human case kills 100% of ferrets in CDC study

The strain of H5N1 bird flu isolated from a dairy worker in Texas was 100 percent fatal in ferrets used to model influenza illnesses in humans. However, the virus appeared inefficient at spreading via respiratory droplets, according to newly released study results from the Centers for Disease Control and Prevention.

The data confirms that H5N1 infections are significantly different from seasonal influenza viruses that circulate in humans. Those annual viruses make ferrets sick but are not deadly. They have also shown to be highly efficient at spreading via respiratory droplets, with 100 percent transmission rates in laboratory settings. In contrast, the strain from the Texas man (A/Texas/37/2024) appeared to have only a 33 percent transmission rate via respiratory droplets among ferrets.

“This suggests that A/Texas/37/2024-like viruses would need to undergo changes to spread efficiently by droplets through the air, such as from coughs and sneezes,” the CDC said in its data summary. The agency went on to note that “efficient respiratory droplet spread, like what is seen with seasonal influenza viruses, is needed for sustained person-to-person spread to happen.”

In the CDC’s study, researchers infected six ferrets with A/Texas/37/2024. The CDC’s data summary did not specify how the ferrets were infected in this study, but in other recent ferret H5N1 studies, the animals were infected by putting the virus in their noses. Ars has reached out to the agency for clarity on the inoculation route in the latest study and will update the story with any additional information provided.

All six of the infected ferrets developed severe disease and died. To test how well the virus could spread among the ferrets, the CDC scientists set up experiments to test transmission through direct contact and respiratory droplets. For the direct transmission test, three healthy ferrets were placed in the same enclosures with three experimentally infected ferrets. All three healthy ferrets became infected.

For the respiratory transmission test, three healthy ferrets were placed in enclosures next to enclosures containing the experimentally infected animals. The infected and uninfected ferrets shared air, but did not have direct contact with each other. Of the three healthy ferrets, only one contracted the H5N1 virus (33 percent). Additionally, that one respiratory transmission event seemed to have a one- to two-day delay compared with what’s seen in the same test with seasonal influenza viruses. This suggests further that the virus is inefficient at respiratory transmission.

The CDC called the overall results “not surprising.” Previous ferret experiments with H5N1 isolates—collected prior to the current bird flu outbreak among US dairy cows—have also found that H5N1 is often lethal to ferrets. Likewise, H5N1 isolates collected from Spain and Chile during the current global outbreak also found that the virus was inefficient at spreading via respiratory droplets among ferrets—with rates ranging from 0 percent to 37.5 percent.

For now, the findings don’t affect the CDC’s overall risk assessment for the general public, which is low. However, it does reinforce the risk to those who have contact with infected animals, particularly dairy and poultry farm workers.

To date, there have been four human cases of H5N1 in the US since the current global bird flu outbreak began in 2022—one in a poultry farm worker in 2022 and three in dairy farm workers, all reported between the beginning of April and the end of May this year. So far, the cases have been mild, the CDC noted, but given the results in ferrets, “it is possible that there will be serious illnesses among people,” the agency concluded.

As of June 9, the US Department of Agriculture has confirmed H5N1 in 85 dairy herds and one alpaca farm across 10 states.

Bird flu virus from Texas human case kills 100% of ferrets in CDC study Read More »

the-world’s-largest-fungus-collection-may-unlock-the-mysteries-of-carbon-capture

The world’s largest fungus collection may unlock the mysteries of carbon capture

Fungus samples are seen on display inside the Fungarium at the Royal Botanic Gardens in Kew, west London in 2023. The Fungarium was founded in 1879 and holds an estimated 380,000 specimens from the UK.

Enlarge / Fungus samples are seen on display inside the Fungarium at the Royal Botanic Gardens in Kew, west London in 2023. The Fungarium was founded in 1879 and holds an estimated 380,000 specimens from the UK.

It’s hard to miss the headliners at Kew Gardens. The botanical collection in London is home to towering redwoods and giant Amazonian water lilies capable of holding up a small child. Each spring, its huge greenhouses pop with the Technicolor displays of multiple orchid species.

But for the really good stuff at Kew, you have to look below the ground. Tucked underneath a laboratory at the garden’s eastern edge is the fungarium: the largest collection of fungi anywhere in the world. Nestled inside a series of green cardboard boxes are some 1.3 million specimens of fruiting bodies—the parts of the fungi that appear above ground and release spores.

“This is basically a library of fungi,” says Lee Davies, curator of the Kew fungarium. “What this allows us to do is to come up with a reference of fungal biodiversity—what fungi are out there in the world, where you can find them.” Archivists—wearing mushroom hats for some reason—float between the shelves, busily digitizing the vast archive, which includes around half of all the species known to science.

Fungarium Collections Manager Lee Davies inspects a fungus sample stored within the Fungarium at the Royal Botanic Gardens in Kew, west London in 2023.

Enlarge / Fungarium Collections Manager Lee Davies inspects a fungus sample stored within the Fungarium at the Royal Botanic Gardens in Kew, west London in 2023.

In the hierarchy of environmental causes, fungi have traditionally ranked somewhere close to the bottom, Davies says. He himself was brought to the fungarium against his will. Davies was working with tropical plants when a staffing reshuffle brought him to the temperature-controlled environs of the fungarium. “They moved me here in 2014, and it’s amazing. Best thing ever, I love it. It’s been a total conversion.”

Davies’ own epiphany echoes a wider awakening of appreciation for these overlooked organisms. In 2020, mycologist Merlin Sheldrake’s book Entangled Life: How Fungi Make Our Worlds, Change Our Minds, and Shape Our Futures was a surprise bestseller. In the video game and HBO series The Last of Us, it’s a fictional brain-eating fungus from the genus Cordyceps that sends the world into an apocalyptic spiral. (The Kew collection includes a tarantula infected with Cordyceps—fungal tendrils reach out from the soft gaps between the dead insect’s limbs.)

While the wider world is waking up to these fascinating organisms, scientists are getting to grips with the crucial role they play in ecosystems. In a laboratory just above the Kew fungarium, mycologist Laura Martinez-Suz studies how fungi help sequester carbon in the soil, and why some places seem much better at storing soil carbon than others.

Soil is a huge reservoir of carbon. There are around 1.5 trillion tons of organic carbon stored in soils across the world—about twice the amount of carbon in the atmosphere. Scientists used to think that most of this carbon entered the soil when dead leaves and plant matter decomposed, but it’s now becoming clear that plant roots and fungi networks are a critical part of this process. One study of forested islands in Sweden found that the majority of carbon in the forest soil actually came from root-fungi networks, not plant matter fallen from above the ground.

The world’s largest fungus collection may unlock the mysteries of carbon capture Read More »

nasty-bug-with-very-simple-exploit-hits-php-just-in-time-for-the-weekend

Nasty bug with very simple exploit hits PHP just in time for the weekend

WORST FIT EVER —

With PoC code available and active Internet scans, speed is of the essence.

Nasty bug with very simple exploit hits PHP just in time for the weekend

A critical vulnerability in the PHP programming language can be trivially exploited to execute malicious code on Windows devices, security researchers warned as they urged those affected to take action before the weekend starts.

Within 24 hours of the vulnerability and accompanying patch being published, researchers from the nonprofit security organization Shadowserver reported Internet scans designed to identify servers that are susceptible to attacks. That—combined with (1) the ease of exploitation, (2) the availability of proof-of-concept attack code, (3) the severity of remotely executing code on vulnerable machines, and (4) the widely used XAMPP platform being vulnerable by default—has prompted security practitioners to urge admins check to see if their PHP servers are affected before starting the weekend.

When “Best Fit” isn’t

“A nasty bug with a very simple exploit—perfect for a Friday afternoon,” researchers with security firm WatchTowr wrote.

CVE-2024-4577, as the vulnerability is tracked, stems from errors in the way PHP converts unicode characters into ASCII. A feature built into Windows known as Best Fit allows attackers to use a technique known as argument injection to pass user-supplied input into commands executed by an application, in this case, PHP. Exploits allow attackers to bypass CVE-2012-1823, a critical code execution vulnerability patched in PHP in 2012.

“While implementing PHP, the team did not notice the Best-Fit feature of encoding conversion within the Windows operating system,” researchers with Devcore, the security firm that discovered CVE-2024-4577, wrote. “This oversight allows unauthenticated attackers to bypass the previous protection of CVE-2012-1823 by specific character sequences. Arbitrary code can be executed on remote PHP servers through the argument injection attack.”

CVE-2024-4577 affects PHP only when it runs in a mode known as CGI, in which a web server parses HTTP requests and passes them to a PHP script for processing. Even when PHP isn’t set to CGI mode, however, the vulnerability may still be exploitable when PHP executables such as php.exe and php-cgi.exe are in directories that are accessible by the web server. This configuration is set by default in XAMPP for Windows, making the platform vulnerable unless it has been modified.

One example, WatchTowr noted, occurs when queries are parsed and sent through a command line. The result: a harmless request such as http://host/cgi.php?foo=bar could be converted into php.exe cgi.php foo=bar, a command that would be executed by the main PHP engine.

No escape

Like many other languages, PHP converts certain types of user input to prevent it from being interpreted as a command for execution. This is a process known as escaping. For example, in HTML, the < and > characters are often escaped by converting them into their unicode hex value equivalents < and > to prevent them from being interpreted as HTML tags by a browser.

The WatchTowr researchers demonstrate how Best Fit fails to escape characters such as a soft hyphen (with unicode value 0xAD) and instead converts it to an unescaped regular hyphen (0x2D), a character that’s instrumental in many code syntaxes.

The researchers went on to explain:

It turns out that, as part of unicode processing, PHP will apply what’s known as a ‘best fit’ mapping, and helpfully assume that, when the user entered a soft hyphen, they actually intended to type a real hyphen, and interpret it as such. Herein lies our vulnerability—if we supply a CGI handler with a soft hyphen (0xAD), the CGI handler won’t feel the need to escape it, and will pass it to PHP. PHP, however, will interpret it as if it were a real hyphen, which allows an attacker to sneak extra command line arguments, which begin with hyphens, into the PHP process.

This is remarkably similar to an older PHP bug (when in CGI mode), CVE-2012-1823, and so we can borrow some exploitation techniques developed for this older bug and adapt them to work with our new bug. A helpful writeup advises that, to translate our injection into RCE, we should aim to inject the following arguments:

-d allow_url_include=1 -d auto_prepend_file=php://input  

This will accept input from our HTTP request body, and process it using PHP. Straightforward enough – let’s try a version of this equipped with our 0xAD ‘soft hyphen’ instead of the usual hyphen. Maybe it’s enough to slip through the escaping?

POST /test.php?%ADd+allow_url_include%3d1+%ADd+auto_prepend_file%3dphp://input HTTP/1.1  Host: host  User-Agent: curl/8.3.0  Accept: */Content-Length: 23  Content-Type: application/x-www-form-urlencoded  Connection: keep-alive       

Oh joy—we’re rewarded with a phpinfo page, showing us we have indeed achieved RCE.

The vulnerability was discovered by Devcore researcher Orange Tsai, who said: “The bug is incredibly simple, but that’s also what makes it interesting.”

The Devcore writeup said that the researchers have confirmed that XAMPP is vulnerable when Windows is configured to use the locales for Traditional Chinese, Simplified Chinese, or Japanese. In Windows, a locale is a set of user preference information related to the user’s language, environment, and/or cultural conventions. The researchers haven’t tested other locales and have urged people using them to perform a comprehensive asset assessment to test their usage scenarios.

CVE-2024-4577 affects all versions of PHP running on a Windows device. That includes version branches 8.3 prior to 8.3.8, 8.2 prior to 8.2.20, and 8.1 prior to 8.1.29.

The 8.0, 7, and 5 version branches are also vulnerable, but since they’re no longer supported, admins will have to follow mitigation advice since patches aren’t available. One option is to apply what are known as rewrite rules such as:

RewriteEngine On  RewriteCond %QUERY_STRING ^%ad [NC]  RewriteRule .? - [F,L]

The researchers caution these rules have been tested only for the three locales they have confirmed as vulnerable.

XAMPP for Windows had yet to release a fix at the time this post went live. For admins without the need for PHP CGI, they can turn it off using the following Apache HTTP Server configuration:

C:/xampp/apache/conf/extra/httpd-xampp.conf

Locating the corresponding lines:

ScriptAlias /php-cgi/ "C:/xampp/php/"  

And comment it out:

# ScriptAlias /php-cgi/ "C:/xampp/php/"  

Additional analysis of the vulnerability is available here.

Nasty bug with very simple exploit hits PHP just in time for the weekend Read More »

ars-chats-with-precision,-the-brain-chip-maker-taking-the-road-less-invasive

Ars chats with Precision, the brain-chip maker taking the road less invasive

Brain-chip buzz —

Precision tested its BCI on 14 people so far. Two more are scheduled this month.

Precision’s Layer 7 Cortical Interface array.

Enlarge / Precision’s Layer 7 Cortical Interface array.

Work toward brain-computer interfaces has never been more charged. Though neuroscientists have toiled for decades to tap directly into human thoughts, recent advances have the field buzzing with anticipation—and the involvement of one polarizing billionaire has drawn a new level of attention.

With competition amping up in this space, Ars spoke with Ben Rapoport, who is a neurosurgeon, electrical engineer, and co-founder of the brain-computer interface (BCI) company Precision Neuroscience. Precision is at the forefront of the field, having placed its BCI on the brains of 14 human patients so far, with two more scheduled this month. Rapoport says he hopes to at least double that number of human participants by the end of this year. In fact, the 3-year-old company expects to have its first BCI on the market next year.

In addition to the swift progress, Precision is notable for its divergence from its competitor’s strategies, namely Neuralink, the most high-profile BCI company and headed by Elon Musk. In 2016, Rapoport co-founded Neuralink alongside Musk and other scientists. But he didn’t stay long and went on to co-found Precision in 2021. In previous interviews, Rapoport suggested his split from Neuralink related to the issues of safety and invasiveness of the BCI design. While Neuralink’s device is going deeper into the brain—trying to eavesdrop on neuron signals with electrodes at close range to decode thoughts and intended motions and speech—Precision is staying at the surface, where there is little to no risk of damaging brain tissue.

Shallow signals

“It used to be thought that you needed to put needle-like electrodes into the brain surface in order to listen to signals of adequate quality,” Rapoport told Ars. Early BCIs developed decades ago used electrode arrays with tiny needles that sink up to 1.5 millimeters into brain tissue. Competitors such as Blackrock Neurotech and Paradromics are still developing such designs. (Another competitor, Synchron, is developing a stent-like device threaded into a major blood vessel in the brain.) Meanwhile, Neuralink is going deeper, using a robot to surgically implant electrodes into brain tissue, reportedly between 3 mm and 8 mm deep.

However, Rapoport eschews this approach. Anytime something essentially cuts into the brain, there’s damage, he notes. Scar tissue and fibrous tissue can form—which is bad for the patient and the BCI’s functioning. “So, there’s not infinite scalability [to such designs],” Rapoport notes, “because when you try to scale that up to making lots of little penetrations into the brain, at some point you can run into a limitation to how many times you can penetrate the brain without causing irreversible and undetectable damage.”

Further, he says, penetrating the brain is just unnecessary. Rapoport says there is no fundamental data that suggests that penetration is necessary for BCIs advances. Rather, the idea was based on the state of knowledge and technology from decades ago. “It was just that it was an accident that that’s how the field got started,” he said. But, since the 1970s, when centimeter-scale electrodes were first being used to capture brain activity, the technology has advanced from the macroscopic to microscopic range, creating more powerful devices.

“All of conscious thought—movement, sensation, intention, vision, etc.—all of that is coordinated at the level of the neocortex, which is the outermost two millimeters of the brain,” Rapoport said. “So, everything, all of the signals of interest—the cognitive processing signals that are interesting to the brain-computer interface world—that’s all happening within millimeters of the brain surface … we’re talking about very small spatial scales.” With the more potent technology of today, Precision thinks it can collect the data it needs without physically traversing those tiny distances.

Ars chats with Precision, the brain-chip maker taking the road less invasive Read More »

samsung-electronics-is-on-strike!-workers-stage-one-day-walkout.

Samsung Electronics is on strike! Workers stage one-day walkout.

Stockpile your chips now —

For now, the one-day strike is just a show of force and shouldn’t hurt production.

A South Korean flag, left, and Samsung Electronics Co. flag fly outside the company's headquarters in Seoul, South Korea.

Enlarge / A South Korean flag, left, and Samsung Electronics Co. flag fly outside the company’s headquarters in Seoul, South Korea.

Jean Chung/Bloomberg via Getty Images

Samsung Electronics workers are on strike! As The New York Times reports, Nationwide Samsung Electronics Union (NSEU) has about 28,000 members, or about one-fifth of Samsung’s workforce, walking out of the job on Friday. It’s Samsung’s first workers’ strike.

Specifically, the walkout is in Samsung’s chip division, which makes RAM, NAND flash chips, USB sticks and SD cards, Exynos processors, camera sensors, modems, NFC chips, and power and display controllers. Depending on how each quarter goes, Samsung is often the world’s largest chipmaker by revenue thanks to this division, and its parts are in products from a million different brands. It’s probably hard to find a tech product that doesn’t have some kind of Samsung chip in it.

As you might expect, the union wants higher pay. Samsung’s workers have gotten as much as 30 percent of their pay from bonuses, and there were no bonuses last year. UnionVP Lee Hyun Kuk told the Times that “it feels like we’ve taken a 30 percent pay cut.” The average pay for a union member is around $60,000 before bonuses.

This is officially a one-day strike, so it’s not expected to hurt Samsung’s output much. For now, this is more about a show of strength by the union in the hopes that Samsung will come to the negotiating table. Samsung reported a profit of $1.4 billion from its chip division in Q1 this year.

If this isn’t resolved, what exactly would happen to the tech industry during a long-term Samsung strike is anyone’s guess. Because of the ubiquity of Samsung’s components, every tech hardware company would have to deal with this somewhat. Samsung has a lot of competitors in each market, though. For instance, for memory it’s always battling SK Hynix and Micron, and a lot of manufacturers will use parts from the three companies interchangeably. Maybe Samsung’s competitors could just pick up the slack. Samsung has never been on strike before, so we’re in uncharted territory.

Samsung Electronics is on strike! Workers stage one-day walkout. Read More »

ai-#67:-brief-strange-trip

AI #67: Brief Strange Trip

I had a great time at LessOnline. It was a both a working trip and also a trip to an alternate universe, a road not taken, a vision of a different life where you get up and start the day in dialogue with Agnes Callard and Aristotle and in a strange combination of relaxed and frantically go from conversation to conversation on various topics, every hour passing doors of missed opportunity, gone forever.

Most of all it meant almost no writing done for five days, so I am shall we say a bit behind again. Thus, the following topics are pending at this time, in order of my guess as to priority right now:

  1. Leopold Aschenbrenner wrote a giant thesis, started a fund and went on Dwarkesh Patel for four and a half hours. By all accounts, it was all quite the banger, with many bold claims, strong arguments and also damning revelations.

  2. Partly due to Leopold, partly due to an open letter, partly due to continuing small things, OpenAI fallout continues, yes we are still doing this. This should wait until after Leopold.

  3. DeepMind’s new scaling policy. I have a first draft, still a bunch of work to do.

  4. The OpenAI model spec. As soon as I have the cycles and anyone at OpenAI would have the cycles to read it. I have a first draft, but that was written before a lot happened, so I’d want to see if anything has changed.

  5. The Rand report on securing AI model weights, which deserves more attention than the brief summary I am giving it here.

  6. You’ve Got Seoul. I’ve heard some sources optimistic about what happened there but mostly we’ve heard little. It doesn’t seem that time sensitive, diplomacy flows slowly until it suddenly doesn’t.

  7. The Problem of the Post-Apocalyptic Vault still beckons if I ever have time.

Also I haven’t processed anything non-AI in three weeks, the folders keep getting bigger, but that is a (problem? opportunity?) for future me. And there are various secondary RSS feeds I have not checked.

There was another big change this morning. California’s SB 1047 saw extensive changes. While many were helpful clarifications or fixes, one of them severely weakened the impact of the bill, as I cover on the linked post.

The reactions to the SB 1047 changes so far are included here.

  1. Introduction.

  2. Table of Contents.

  3. Language Models Offer Mundane Utility. Three thumbs in various directions.

  4. Language Models Don’t Offer Mundane Utility. Food for lack of thought.

  5. Fun With Image Generation. Video generation services have examples.

  6. Deepfaketown and Botpocalypse Soon. The dog continues not to bark.

  7. They Took Our Jobs. Constant AI switching for maximum efficiency.

  8. Get Involved. Help implement Biden’s executive order.

  9. Someone Explains It All. New possible section. Template fixation.

  10. Introducing. Now available in Canada. Void where prohibited.

  11. In Other AI News. US Safety Institute to get model access, and more.

  12. Covert Influence Operations. Your account has been terminated.

  13. Quiet Speculations. The bear case to this week’s Dwarkesh podcast.

  14. Samuel Hammond on SB 1047. Changes address many but not all concerns.

  15. Reactions to Changes to SB 1047. So far coming in better than expected.

  16. The Quest for Sane Regulation. Your random encounters are corporate lobbyists.

  17. That’s Not a Good Idea. Antitrust investigation of Nvidia, Microsoft and OpenAI.

  18. The Week in Audio. Roman Yampolskiy, also new Dwarkesh Patel is a banger.

  19. Rhetorical Innovation. Innovative does not mean great.

  20. Oh Anthropic. I have seen the other guy, but you are not making this easy.

  21. Securing Model Weights is Difficult. Rand has some suggestions.

  22. Aligning a Dumber Than Human Intelligence is Still Difficult. What to do?

  23. Aligning a Smarter Than Human Intelligence is Difficult. SAE papers continue.

  24. People Are Worried About AI Killing Everyone. Various p(doom)s.

  25. Other People Are Not As Worried About AI Killing Everyone. LeCun fun.

  26. The Lighter Side. Why, yes. Yes I did.

Did AI pass a restaurant review ‘Turing test,or did human Yelp reviewers fail it? This is unsurprising, since the reviews seemingly were evaluated in isolation. Writing short bits like this is the wheelhouse. At minimum, you need to show the context, meaning the other information about the restaurant, including other reviews.

Via David Brin, goblin.tools has a formalizer, to change the tone of your text. You can of course do better with a normal LLM but an easier interface and no startup costs can go a long way.

Start making your Domino’s ‘pizza’ before you are done ordering it. The fun writes itself, also kind of amazingly great. I am hungry now.

It seems McDonalds does this too with its fries. My guess is this is more ‘we have enough fungible fries orders often enough that we simply make fries continuously’ rather than ‘we know you in particular will want fries’ but I could be wrong.

Would you like an extra thumb? Why yes I would. What’s funny is you can run a mental experiment to confirm that you’re totally capable of learning to use it if the machine can read the impulses. Plausibly super awesome. Mandatory jokes are in the comments if you scroll down.

Have adult level theory of mind, up to 6th order inferences.

Aid in drug development. No idea how much it helps, but all help is great.

Predict out-of-distribution salt crystal formation, with correct structures, while running a simulation. Suggestive of material science work being possible without physical experimentation.

Garry Tan endorses Perplexity for search if you want well-cited answers. I agree with Arun’s reply, Perplexity is great but only shines for narrow purposes. The ‘well-cited’ clause is doing a lot of work.

Use Gemini 1.5 Flash for many purposes, because it is fast, cheap and good enough. Sully has long been a proponent of cheap and fast and good enough.

Not yet, anyway.

Shoshana Weissmann: ugh really wanna jailbreak my nordictrack.

Brian Chen (New York Times) is not impressed by the new GPT-4, failing to see much improvement other than speed, saying he definitely wouldn’t let it tutor his child. This was a standard ‘look for places the AI fails’ rather than looking for where it succeeds. A great illustration is when he notes the translations were good, but that the Chinese accents were slightly off. Yes, okay, let’s improve the accents, but you are missing the point. Any child, or any adult, not using AI to learn is missing out.

Erik Wiffin warns of counterfeit proofs of thought. What happens if all those seemingly useless project plans and self-reports were actually about forcing people to think and plan? What if the plan was worthless, but the planning was essential, and now you can forge the plan without the planning? Whoops. Zizek’s alternative is that your LLM writes the report, mine reads it and now we are free to learn. Which way, modern worker?

  1. For the self-report I lean towards Zizek. This is mostly a test to see how much bullshit you dare write down on a page before you think you’ll be called out on it, a key way that the bullshitters collude to get ahead at the expense of those who don’t know to go along or have qualms about doing so.

    1. The idea that ‘your manager already knows’ might be true in some places, but it sure is not in others.

    2. I can’t remember the last time I knew someone who thought ‘writing this mandatory corporate self-report taught me so many valuable lessons’ because no one I know is that naive.

  2. The project plan seems more plausibly Wriffin’s territory. You should have to form a plan. That does not mean that the time spent turning that plan into a document that looks right is time well spent. So the goal is to get the manager to do the actual planning – Alice makes the Widgets, David buys the Thingamabobs. Then the LLM turns that into a formal document.

Look, I am not a master of war, but if I was Air Force Secretary Frank Kendall then I would presume that the point of an F-16 flying with AI was that I did not have to be inside that F-16 during simulated combats. He made a different decision. I mean, all right, yes, show of confidence, fun as hell, still I suppose that is one of many reasons I am not the secretary of the air force.

The funny alternative theory is this was so the other humans would let the AI win.

Still no worthy successor to AI dungeon, despite that being a flimsy base model wrapper and being a great product until you ran into its context window limits. The ‘put AI into games and interactive worlds’ developers are letting us down. Websim is kind of the modern version, perhaps, and Saerain Trismegistus mentions NovelAI.

Examples from Google’s video generation AI Veo.

Examples from a Chinese video generation service, 2 minutes, 30fps, 1080p.

Indian (and Mexican) elections latest to not have serious AI-related issue, despite this wild report?

Ate-a-Pi: 🤩 AI in politics in 🇮🇳

> Politicians are voluntarily deepfaking themselves

> to dub their message into the 22 languages widely spoken in India

> 50 million AI voice clone calls in the last month

> resurrecting deceased party leaders to endorse current candidates (the cult of personality never ends.. when Lee Kuan Yew?)

> super small teams – leading firm has 10 employees, founder dropped out of college after learning to make deepfakes on Reddit during the COVID breakdowns (for every learning loss.. there was a learning gain 🤣)

> authorized by politicians but not disclosed to voters. Many voters believe the calls are real and that the pols actually spoke to them

> typical gap in quality AI “promises a Ferrari but delivers a Fiat”

> fine tuning Mistral to get better results

This too will come to the 🇺🇸

The future of politics is having a parasocial relationship with your favorite politician, and their AI version being part of your brain trust, advisory board.

Kache: Americans don’t realize that in india and pakistan, people watch AI generated shorts of political leaders and believe that they are real.

This is such a wild equilibrium. Everyone gets to clone their own candidates and backers with AI, no one does fakes of others, the voters believe all of it is real.

Not a bad equilibrium. Yes, voters are fooled, but it is a ‘fair fooling’ and every message is intended by the candidate and party it purports to be from. This presumably is the least stable situation of all time and won’t happen again. The people will realize the AIs are fake. Also various actors will start trying to fake others using AI, but perhaps punishment and detection can actually work there?

One might think about the ‘leave your stroller at the playground and not pay attention’ technology. Yes, someone could try to steal it, but there is at least one person in the playground who would get very, very angry with you if they notice you trying to do that. What makes yours yours is not that you can prove it is yours, but that when you try to take it, you know no one will object.

People worry AI will be used to generate misinformation and people won’t be able to tell the difference. It is worth remembering the current state of misinformation generation and spreading technology, which is best summarized as:

Joseph Menn (Washington Post): News site editor’s ties to Iran, Russia show misinformation’s complexity.

Matthew Yglesias: This doesn’t actually seem that complicated.

A metaphorical TaskRabbit for AI hires, potentially something like EquiStamp, could be the efficient way to go. Alex Tabarrok suggests we may continuously evaluate, hire and fire AIs as relative performance, speed and cost fluctuate. Indeed, the power users of AI do this, and I am constantly reassessing which tools to use for which jobs, same as any other tool. A lot of this is that right now uses are mostly generic and non-integrated. It is easy to rotate. When we have more specialized tools, and need more assurance of consistent responses, it will be more exciting to stick to what you know relative to now.

What happens when you figure out how to have AI do your job and no one cares?

Fellowship for $190k/year to help implement Biden’s executive order. Deadline is June 12 or if they get 100 applicants, so act quickly. The median salary in AI is $326k, so this is not that high, but it is highly livable.

Prize of $500k for using AI to converse with animals.

Near: the experts have chimed in and have concerns that talking to animals might be bad. Lu0ckily I am going to ignore them and do it anyway!

Marcello Herreshoff explains the idea of ‘template fixation.’ If your question is close enough to a sufficiently strong cliche, the cliche gets applied even if it does not make sense. Hence the stupid answers to river crossing questions or water pouring tests or other twists on common riddles. If we can’t find a way to avoid this, math is going to remain tough. It is easy to see why this would happen.

OpenAI for nonprofits, essentially a discounted subscription.

Claude now available in Canada. Finally?

Flash Diffusion, for improving training diffusion models.

Leading AI companies agree to share their models with the US AI Safety Institute for pre-deployment testing. Link is to this story, which does not list which labs have agreed to do it, although it says there was no pushback.

NewsCorp’s deal with OpenAI, which includes the Wall Street Journal, is no joke, valued at over $250 million ‘in cash and credits’ over five years. The NewsCorp market cap is 15.8 billion after rising on the deal, so this is over 1% of the company over only five years. Seems very hard to turn down that kind of money. One key question I have not seen answered is to what extent these deals are exclusive.

Futurism notes that legacy media is getting these rich deals, whereas non-legacy media, which includes right-wing media, gets none of it so far. Robin Hanson summarizes this as ‘AI will lean left.’ We already have strong evidence AI will lean left in other ways, and this seems like an echo of that, mostly reflective of reality.

AI model for ECGs using vision transformer architecture gets state of the art performance with less training data.

Yes, AI models are mimics of deeply WEIRD data sets, so they act and respond in a Western cultural context. If you would rather create an AI that predicts something else, that most customers would want less but some would want more, that seems easy enough to do instead.

Tesla buying GPUs faster than it has places to plug them in.

Google getting rid of San Francisco office space. Prices for office space are radically down for a reason, work from home reduces needed space and if the landlords wouldn’t play ball you can relocate to another who will, although I see no signs Google is doing that. Indeed, Google is shrinking headcount, and seems to be firing people semi-randomly in doing so, which is definitely not something I would do. I would presume that doing a Musk-style purge of the bottom half of Google employees would go well. But you can only do that if you have a good idea which half is which.

OpenAI details its protocols for securing research infrastructure for advanced AI, also known as protecting its core algorithms and model weights. I leave it to others to assess how strong these precautions are. No matter what else might be happening at OpenAI, this is one team you do want to root for.

The gender divide in AI.

Kesley Piper: The Computing Research Association annual survey found that 18% of graduates from AI PhD programs are women.

Women are smarter than men. They avoid academic PhDs and OpenAI.

WiserAI has a community white paper draft. I am not looking further because triage, will revisit when it is final perhaps. If that is a mistake, let me know.

OpenAI terminates five user accounts that were attempting to use OpenAI’s models to support ‘covert influence operations,’ which OpenAI defines as ‘attempts to manipulate public opinion or influence political outcomes without revealing the true identity of the actors behind them.’ Full report here.

Specifically, the five are Russian actors Bad Grammer and Doppelganger, Chinese actor Spamouflage that was associated with Chinese law enforcement, Iranian actor International Union of Virtual Media and actions by the Israeli commercial company Stoic.

Note the implications of an arm of China’s law enforcement using ChatGPT for this.

OpenAI believes that the operations in question failed to substantially achieve their objectives. Engagement was not generated, distribution not achieved. These times.

What were these accounts doing?

Largely the same things any other account would do. There are multiple mentions of translation between languages, generating headlines, copy editing, debugging code and managing websites.

As OpenAI put it, ‘productivity gains,’ ‘content generation’ and ‘mixing old and new.’

Except, you know, as a bad thing. For evil, and all that.

They also point to the common theme of faking engagement, and arguably using ChatGPT for unlabeled content generation (as opposed to other productivity gains or copy editing) is also inherently not okay as well.

Ordinary refusals seem to have played a key role, as ‘threat actors’ often published the refusals, and the steady streams of refusals allowed OpenAI to notice threat actors. Working together with peers is also reported as helpful.

The full report clarifies that this is sticking to a narrow definition I can fully support. What is not allowed is pretending AI systems are people, or attributing AI content to fake people or without someone’s consent. That was the common theme.

Thus, Sam Altman’s access to OpenAI’s models will continue.

Preventing new accounts from being opened by these treat actors seems difficult, although this at least imposes frictions and added costs.

There are doubtless many other ‘covert influence operations’ that continue to spam AI content while retaining access to OpenAI’s models without disruption.

One obvious commonality is that all five actors listed here had clear international geopolitical goals. It is highly implausible that this is not being done for many other purposes. Until we are finding (for example) the stock manipulators, we have a long way to go.

This is still an excellent place to start. I appreciate this report, and would like to see similar updates (or at least brief updates) from Google and Anthropic.

The bear case on Nvidia. Robin Hanson continues to say ‘sell.

Claims from Andrew Cote about consciousness, the nature of reality and also of LLMs. I appreciate the model of ‘normiehood’ as a human choosing outputs on very low temperature.

Might we soon adorn buildings more because it is easier to do so via LLMs?

WSJ’s Christopher Mins says ‘The AI Revolution is Already Losing Steam.’ He admits my portfolio would disagree. He says AIs ‘remain ruinously expensive to run’ without noticing the continuing steady drop in costs for a given performance level. He says adoption is slow, which it isn’t compared to almost any other technology even now. Mostly, another example of how a year goes by with ‘only’ a dramatic rise in speed and reduction in cost and multiple players catching up to the leader and the economy not transformed and stocks only way up and everyone loses their minds.

I think that is behind a lot what is happening now. The narratives in Washington, the dismissal by the mainstream of both existential risks and even the possibility of real economic change. It is all the most extreme ‘what have you done for me lately,’ people assuming AI will never be any better than it is now, or it will only change at ‘economic normal’ rates from here.

Thus, my prediction is that when GPT-5 or another similar large advance does happen, these people will change their tune for a bit, adjust to the new paradigm, then memory hole and go back to assuming that AI once again will never advance much beyond that. And so on.

He’s joking, right?

Eliezer Yudkowsky: The promise of Microsoft Recall is that extremely early AGIs will have all the info they need to launch vast blackmail campaigns against huge swathes of humanity, at a time when LLMs are still stupid enough to lose the resulting conflict.

Rohit: I’d read this novel!

A lot of users having such a honeypot on their machines for both blackmail and stealing all their access and their stuff certainly does interesting things to incentives. One positive is that you encourage would-be bad actors to reveal themselves, but you also empower them, and you encourage actors to go bad or skill up in badness.

Dan Hendrycks questions algorithmic efficiency improvements, notes that if (GPT-)4-level models were now 10x cheaper to train we would see a lot more of them, and that secondary labs should not be that far behind. I do not think we should assume that many labs are that close to OpenAI in efficiency terms or in ‘having our stuff together’ terms.

Papers I analyze based on the abstract because that’s all the time we have for today: Owen Davis formalizes ways in which AI could weaken ‘worker power’ distinct from any impacts on labor demand, via management use of AI. The obvious flaw is that this does not mention the ability of labor to use AI. Labor can among other applications use AI to know when it is being underpaid or mistreated and to greatly lower switching costs. It also could allow much stronger signals of value, allowing workers more ability to switch jobs. I would not be so quick to assume ‘worker power’ will flow in one direction or the other in a non-transformative AI world.

Samuel Hammond wrote a few days ago in opposition to SB 1047, prior to the recent changes. He supports the core idea, but worries about particular details. Many of his concerns have now been addressed. This is the constructive way to approach the issue.

  1. He objected to the ‘otherwise similar general capability’ clause on vagueness grounds. The clause has been removed.

  2. He warns of a ‘potential chilling effect on open source,’ due to inability to implement a shutdown clause. The good news is that this was already a misunderstanding of the bill before the changes. The changes make my previous interpretation even clearer, so this concern is now moot as well.

    1. And as I noted, the ‘under penalty of perjury’ is effectively pro forma unless you are actively lying, the same as endless other government documents with a similar rules set.

  3. Also part of Samuel’s stated second objection: He misunderstands the limited duty exemption procedure, saying it must be applied for before training, which is not the case.

    1. You do not apply for it, you flat out invoke it, and you can do this either before or after training.

    2. He warns that you cannot predict in advance what capabilities your model will have, but in that case the developer can invoke before training and only has to monitor for unexpected capabilities and then take back the exemption (without punishment) if that does happen.

    3. Or they can wait, and invoke after training, if the model qualifies.

  4. Samuel’s third objection is the fully general one, and still applies: that model creators should not be held liable for damages caused by their models, equating it to what happens if a hacker uses a computer. This is a good discussion to have. I think it should be obvious to all reasonable parties that both that (a) model creators should not be liable for harms simply because the person doing harm used the model while doing the harm and (b) that model creators need to be liable if they are sufficiently negligent or irresponsible, and sufficiently large harm results. This is no different than most other product harms. We need to talk procedure and price.

    1. In the case of the SB 1047 proposed procedure, I find it to be an outlier in how tight are the requirements for a civil suit. Indeed, under any conditions where an AI company was actually held liable under SB 1047 for an incident with over $500 million in damages, I would expect that company to already also be liable under existing law.

    2. I strongly disagree with the idea that if someone does minimal-cost (relative to model training costs) fine tuning to Llama-4, this should absolve Meta of any responsibility for the resulting system and any damage that it does. I am happy to have that debate.

    3. There was indeed a problem with the original derivative model clause here, which has been fixed. We can talk price, and whether the 25% threshold now in SB 1047 is too high, but the right price is not 0.001%.

  5. Samuel also objects to the ‘net neutrality’ style pricing requirements on GPUs and cloud services, which he finds net negative but likely redundant. I would be fine with removing those provisions and I agree they are minor.

  6. He affirms the importance of whistleblower provisions, on which we agree.

  7. Samuel’s conclusion was that SB 1047 ‘risks America’s global AI leadership outright.’ I would have been happy to bet very heavily against such impacts even based on the old version of the bill. For the new version, if anyone wants action on that bet, I would be happy to give action. Please suggest terms.

In the realm of less reasonable objections that also pre-dated the recent changes, here is the latest hyperbolic misinformation about SB 1047, here from Joscha Bach and Daniel Jeffreys, noted because of the retweet from Paul Graham. I have updated my ‘ignorables list’ accordingly.

My prediction on Twitter was that most opponents of SB 1047, of which I was thinking especially of its most vocal opponents, would not change their minds.

I also said we would learn a lot, in the coming days and weeks, from how various people react to the changes.

So far we have blissfully heard nothing from most of ‘the usual suspects.’

Dan Hendrycks has a thread explaining some of the key changes.

Charles Foster gets the hat tip for alerting me to the changes via this thread. He was also the first one I noticed that highlighted the bill’s biggest previous flaw, so he has scored major RTFB points.

He does not offer a position on the overall bill, but is very good about noticing the implications of the changes.

Charles Foster: In fact, the effective compute threshold over time will be even higher now than it would’ve been if they had just removed the “similar performance” clause. The conjunction of >10^26 FLOP *and>$100M means the threshold rises with FLOP/$ improvements.

It is “Moore’s law adjusted” if by that you mean that the effective compute threshold will adjust upwards in line with falling compute prices over time. And also in line with $ inflation over time.

He also claims this change, which I can’t locate:

– “Hazardous capability” now determined by marginal risk over existing *nonexemptcovered models

If true, then that covers my other concern as well, and should make it trivial to provide the necessary reasonable assurance if you are not pushing the frontier.

Finally, he concludes:

Charles Foster: I think the bill is significantly better now. I didn’t take a directly pro- or anti-stance before, and IDK if I will in the future, but it seems like the revised bill is a much better reflection of the drafters’ stated intentions with fewer side effects. That seems quite good.

Andrew Critch, who was skeptical of the bill, approves of the covered model change.

Andrew Critch: These look like good changes to me. The legal definition of “Covered Model” is now clearer and more enforceable, and creates less regulatory uncertainty for small/non-incumbent players in the AI space, hence more economic fairness + freedom + prosperity. Nice work, California!

I think I understand the rationale for the earlier more restrictive language, but I think if a more restrictive definition of “covered model” is needed in the future, lowering numerical threshold(s) will be the best way to achieve that, rather than debating the meaning of the qualitative definition(s). Clear language is crucial for enforcement, and the world *definitelyneeds enforceable AI safety regulations. Progress like this makes me proud to be a resident of California.

Nick Moran points to the remaining issue with the 25% rule for derivative models, which is that if your open weights model is more than 4x over the threshold, then you create a window where training ‘on top of’ your model could make you responsible for a distinct otherwise covered model.

In practice I presume this is fine – both no one is going to do this, and if they did no one is going to hold you accountable for something that clearly is not your fault and if they did the courts would throw it out – but I do recognize the chilling effect, and that in a future panic situation I could be wrong.

The good news is there is an obvious fix, now that the issue is made clear. You change ‘25% of trained compute’ to ‘either 25% of trained compute or sufficient compute to qualify as a covered model.’ That should close the loophole fully, unless I missed something.

I have been heartened by the reactions of those in my internet orbit who were skeptical but not strongly opposed. There are indeed some people who care about bill details and adjust accordingly.

Dean Ball was the first strong opponent of SB 1047 I have seen respond, indeed he did so before I could read the bill changes let alone register any predictions. As opponents go, he has been one of the more reasonable ones, although we still strongly disagree on many aspects.

His reaction admits The Big Flip up front, then goes looking for problems.

Dean Ball: SB 1047 has been amended, as Senator Wiener recently telegraphed. My high-level thoughts:

1. There are some good changes, including narrowing the definition of a covered model

2. The bill is now more complex, and arguably harder for devs to comply with.

Big picture: the things that the developer and academic communities hated about SB 1047 remain: generalized civil and criminal liability for misuse beyond a developer’s control and the Frontier Model Division.

It is strictly easier to comply in the sense that anything that complied before complies now, but if you want to know where the line is? Yeah, that’s currently a mess.

I see this partly as the necessary consequence of everyone loudly yelling about how this hits the ‘little guy,’ which forced an ugly metric to prove it will never, ever hit the little guy, which forces you to use dollars where you shouldn’t.

That is no excuse for not putting in a ‘this is how you figure out what the market price will be so you can tell where the line is’ mechanism. We 100% need some mechanism.

The obvious suggestion is to have a provision that says the FMD publish a number once a (week, month or year) that establishes the price used. Then going forward you can use that to do the math, and it can at least act as a safe harbor. I presume this is a case of ‘new provision that no one gamed out fully’ and we can fix it.

Dean next raises a few questions about the 25% threshold for training (which the developer must disclose), around questions like the cost of synthetic data generation. My presumption is that data generation does not count here, but we could clarify that either way.

He warns that there is no dollar floor on the 25%, but given you can pick the most expansive open model available, it seems unlikely this threshold will ever be cheap to reach in practice unless you are using a very old model as your base, in which case I suppose you fill out the limited duty exemption form with ‘of course it is.’

If you want to fix that at the cost of complexity, there are various ways to cover this corner case.

Ball mentions the safety assurances. My non-lawyer read was that the changes clarify that this is the ‘reasonable assurance’ standard they use in other law and not anything like full confidence, exactly to (heh) provide reasonable assurance that the rule would be reasonable. If lawyers or lawmakers think that’s wrong let me know, but there is a particular sentence inserted to clarify exactly that.

He also mentions that Weiner at one point mentioned Trump in the context of the executive order. It was a cheap shot as phrased given no one knows what Trump ultimately thinks about AI, and I wish he hadn’t said that, but Trump has indeed promised to repeal Biden’s executive order on AI, so the actual point – that Congress is unlikely to act and executive action cannot be relied upon to hold – seems solid.

In a follow-up he says despite the changes that SB 1047 is still ‘aimed squarely at future generations of open-source foundation models.’ I rather see open models as having been granted exemptions from several safety provisions exactly because those are forms of safety that open models cannot provide, and their community making special pleading that they should get even more of a free pass. Requiring models adhere to even very light safety requirements is seen as ‘aimed squarely at open source’ exactly because open models make safety far more difficult.

Dean also notes here that Senator Weiner is making a good case for federal preemption of state policies. I think Weiner would to a large extent even agree with this, that it would be much better if the federal government acted to enact a similar law. I do not see California trying to override anything, rather it is trying to fill a void.

Danielle Fong here notes the bill is better, but remains opposed, citing the fee structure and general distrust of government.

Here is a good, honest example of ‘the price should be zero’:

Godoglyness: No, because the bill still does too much. Why should 10^26 and 100 million dollars be a cutoff point?

There shouldn’t be any cutoffs enacted until we have actual harms to calibrate on

It’s good the impact of the bill is diminished but it’s bad it still exists at all

Remember how some folks thought GPT2 would be dangerous? Ridiculous in retrospect, but…

We shouldn’t stop big training runs because of speculative harms when the speculation has failed again and again to anticipate the form/impact/nature of AI systems.

If you think ‘deal with the problems post-hoc after they happen’ is a superior policy, then of course you should oppose the bill, and be similarly clear on the logic.

Similarly, if your argument is ‘I want the biggest most capable possible open models to play with regardless of safety concerns and this might interfere with Meta opening the weights of Llama-N’ and I will oppose any bill that does that, then yes, that is another valid reason to oppose the bill. Again, please say that.

That is very different from misrepresenting the bill, or claiming it would impact people it even more explicitly than before does not impact.

On that note, here is Andrew Ng ignoring the changes and reiterating past arguments in ways that did not apply to the original bill and apply even less now that the comparison point for harm has been moved. For your model to be liable, it has to enable the actions in a way that non-covered models and models eligible for limited duty exemptions would not. Andrew Ng mentions that all current models can be jailbroken, but I do not see how that should make us intervene less. Ultimately he is going for the ‘only regulate applications’ approach that definitely won’t work. Arvind Narayanan calls it a ‘nice analysis.’

TIDNL, featuring helpfully clear section headlines like “Corporate America Looks to Control AI Policy” and section first sentences such as “Corporate interests are dominating lobbying on AI issues.”

Luke Muehlhauser: No surprise: “85 percent of the lobbyists hired in 2023 to lobby on AI-related issues were hired by corporations or corporate-aligned trade groups”

[thread contains discussion on definition of lobbying, linked to here.]

Public Citizen: Corporations, trade groups and other organizations sent more than 3,400 lobbyists to lobby the federal government on AI-related issues in 2023, a 120 percent leap from 2022.

  • AI is not just an issue of concern for AI and software corporations: While the tech industry was responsible for the most AI-related lobbyists in 2023 – close to 700 – the total amounts to only 20 percent of all the AI-related lobbyists deployed. Lobbyists from a broad distribution of industries outside of tech engaged in AI-related issues, including financial services, healthcare, telecommunications, transportation, and defense.

  • 85 percent of the lobbyists hired in 2023 to lobby on AI-related issues were hired by corporations or corporate-aligned trade groups. The Chamber of Commerce was responsible for the most AI-related lobbyists, 81, followed by Intuit (64), Microsoft (60), the Business Roundtable (42), and Amazon (35).

  • OpenSecrets found that groups that lobbied on AI in 2023 spent a total of $957 million lobbying the federal government on all issues that year. [Note that this is for all purposes, not only for AI]

  • An analysis of the clients revealed that while many clients resided in the tech industry, they still only made up 16% of all clients by industry.

The transportation sector, which ranked sixth for having the most clients lobby on AI-related issues, has engaged heavily on policies regarding autonomous vehicles.

In the defense sector, 30 clients hired a combined total of 168 lobbyists to work on AI issues. Given the U.S. Department of Defense and military’s growing interest in AI, defense companies that are often major government contractors have been increasingly implementing AI for military applications.

…in August 2023 the Pentagon announced a major new program, the Replicator Initiative, that aim to rely heavily on autonomous drones to combat Chinese missile strength in a theoretical conflict over Taiwan or at China’s eastern coast.

Look. Guys. If you are ever tempted to call something the Replicator Initiative, there are three things to know.

  1. Do not do the Replicator Initiative.

  2. Do not do the Replicator Initiative.

  3. Do not do the Replicator Initiative.

Also, as a bonus, at a bare minimum, do not call it the Replicator Initiative.

As federal agencies move forward with developing guardrails for AI technologies, stakeholders will likely rely even more on their lobbyists to shape how AI policy is formed.

You know one way to know your guardrails are lacking?

You called a program the Replicator Initiative.

Yes, expect tons of lobbying, mostly corporate lobbying.

Where will they lobby? It seems the White House is the place for the cool kids.

So who is involved?

Even in cases where at first glance a lobbying entity may not appear to be representing corporate interests, digging deeper into partnerships and collaborations revealed that non-profit interests are often deeply intertwined with corporate ones as well.

Only five of the top 50 lobbying entities responsible for the most AI-related lobbyists in 2023 were not representing corporate interests. Two of the five were large hospitals – the Mayo Clinic and The New York and Presbyterian Hospital – while the other three were the AFL-CIO, AARP, and the National Fair Housing Alliance. None of the five were in the top ten

Did you notice any names not on that list?

Most of that lobbying is highly orthogonal to the things generally discussed here. Hospitals are presumably concerned primarily with health care applications and electronic medical records. That was enough for multiple hospital groups to each outspend all lobbying efforts towards mitigating existential risk.

Adam Thierer implores us to just think of the potential, reminds us to beat China, urges ‘pro-innovation’ AI policy vision. It’s a Greatest Hits on so many levels. The core proposal is that ‘the time is now’ to… put a moratorium on any new rules on AI, and preempt any potential state actions. Do nothing, only more so.

Gavin Newsom warns about the burdens of overregulation of AI and the threat it would pose to California’s leadership on that, but says the state has ‘an obligation to lead’ because AI was invented there.

To be completely fair to Newsom this is not the first time he has warned about overregulation – he did it in 2004 regarding the San Francisco business permitting process, which is a canonical insane example of overregulation, and he has indeed taken some ‘concrete steps’ as governor to streamline some regulatory burdens, including an executive order and signing AB 1817. But also:

As usual in politics, this is both-sides applause light talk that does not tell you the price. The price is not going to be zero, nor would that be wise even if there was no existential risk, any more than we should have no laws about humans. The price is also a cost, and setting it too high would be bad.

The world as it is: FEC fighting FCC’s attempt to require political ads to disclose that they used AI, saying FCC lacks jurisdiction, and finding it ‘deeply troubling’ that they want this in place before the election with it happening so soon. How is ‘political ads that use AI tell us they are using AI’ not one of the things we can all agree upon?

You know what is a really, really bad idea?

Going after AI companies with antitrust enforcement.

Josh Sisco (Politico): The Justice Department and Federal Trade Commission are nearing an agreement to divvy up investigations of potential anticompetitive conduct by some of the world’s largest technology companies in the artificial intelligence industry, according to three people with knowledge of the negotiations.

As part of the arrangement, the DOJ is poised to investigate Nvidia and its leading position in supplying the high-end semiconductors underpinning AI computing, while the FTC is set to probe whether Microsoft, and its partner OpenAI, have unfair advantages with the rapidly evolving technology, particularly around the technology used for large language models.

The deal has been negotiated for nearly a year. And while leaders of both agencies have expressed urgency in ensuring that the rapidly growing artificial intelligence technology is not dominated by existing tech giants, until an agreement is finalized, there was very little investigative work they could do.

Fredipus Rex: Also, how in the world is OpenAI, which loses money on a mostly free product that has existed for two years and which is in a constant game of monthly technological leapfrog with a bunch of competitors in any possible way a “monopoly”?

Ian Spencer: Microsoft and OpenAI have nothing even remotely resembling monopolies in the AI space. Nvidia is facing competition everywhere, despite being clear market leaders.

It’s ASML and TSMC who have truly dominant market positions thanks to their R&D, and neither of them is based in the US.

Shoshana Weissmann: Every day is pain.

I do not know whether to laugh or cry.

A year ago they wanted to start an antitrust investigation, but it took that long to negotiate between agencies?

The antitrust was based on the idea that some companies in technological races had currently superior technologies and were thus commanding large market shares while rapidly improving their products and what you could get at a given price, and producing as fast as they could require input components?

Perhaps the best part is that during that year, during which OpenAI has been highly unprofitable in order to fight for market share and develop better products, two distinct competitors caught up to OpenAI and are now offering comparable products, although OpenAI likely will get to the next generation level first.

Or is the best part that Microsoft so little trusts OpenAI that they are spending unholy amounts of money to engage in direct competition with them?

Meanwhile Nvidia faces direct competition on a variety of fronts and is both maximizing supply and rapidly improving its products while not charging anything like the market clearing price.

This from the people who brought you ‘Google monopolized search,’ ‘Amazon prices are too high,’ ‘Amazon prices are too low’ and ‘Amazon prices are suspiciously similar.’

As Ian notes, in theory one could consider ASML or TSMC as more plausible monopolies, but neither is exploiting its position, and also neither is American so we can’t go after them. If anything I find the continued failure of both to raise prices to be a confusing aspect of the world.

It is vital not only not to prosecute companies like OpenAI for antitrust. They vitally need limited exemptions from antitrust, so that if they get together to collaborate on safety, they need not worry the government will prosecute them for it.

I have yet to see a free market type who wants to accelerate AI and place absolutely no restrictions on its development call for this particular exemption.

Lex Fridman talks to Roman Yampolskiy, as played by Jeff Goldblum, and Lex does not miss the central point.

Lex Fridman: Here’s my conversation with Roman Yampolskiy, AI safety researcher who believes that the chance of AGI eventually destroying human civilization is 99.9999%. I will continue to chat with many AI researchers & engineers, most of whom put p(doom) at <20%, but it's important to balance those technical conversations by understanding the long-term existential risks of AI. This was a terrifying and fascinating discussion.

Others, not so much.

Elon Musk: 😁

If you are interested in communication of and debate about existential risk, this is a podcast worth listening to. I could feel some attempts of Roman’s working because they worked well, others working by playing to Lex’s instincts in strange ways, others leading into traps or bouncing off before the reactions even happened. I saw Lex ask some very good questions and make some leaps, while being of all the Lex Fridmans the Lex Fidmanest in others. It is amazing how much he harps on the zoo concept as a desperate hope target, or how he does not realize that out of all the possible futures, most of the ones we can imagine and find interesting involve humans because we are human, but most of the configurations of atoms don’t involve us. And so on.

Also it is unfortunate (for many purposes) that Roman has so many additional funky views such as his perspective on the simulation hypothesis, but he is no doubt saying what he actually believes.

Of course, there is also Dwarkesh Patel talking to Leopold Aschenbrenner for 4.5 hours. I have been assured this is an absolute banger and will get to it Real Soon Now.

Request for best philosophical critique against AI existential risk. I am dismayed how many people exactly failed to follow the directions. We need to do better at that. I think the best practical critique is to doubt that we will create AGI any time soon, which may or may not be philosophical depending on details. It is good to periodically survey the answers out there.

Your periodic reminder that there are plenty of people out there on any high stakes topic who are ‘having a normal one,’ and indeed that a lot of people’s views are kind of crazy. And also that in-depth discussions of potential transformationally different future worlds are going to sound weird at times if you go looking for weirdness. As one commenter notes, if people keep retweeting the crazytown statements but not the people saying sanetown statements, you know what you will see. For other examples, see: Every political discussion, ever, my lord, actually please don’t, I like you.

For those trying to communicate nuance instead, it remains rough out there.

Helen Toner: Trying to communicate nuance in AI rn be like

Me: people think xrisk=skynet, but there are lots of ways AI could cause civilization-scale problems, and lots of throughlines w/today’s harms, so we shouldn’t always have those conversations separately

Headline writer: skynet dumb

Helen Toner: If you want to hear my full answer, it starts about 33: 15 here.

(and the article itself is fine/good, no shade to Scott. it’s just always those headlines…)

In this case the article is reported to be fine, but no, in my experience it is usually not only the headlines that are at issue.

An illustration.

Liron Shapira: Will humanity be able to determine which ASI behavior is safe & desirable by having it output explanations and arguments that we can judge?

Some argue yes. Some argue no. It’s tough to judge.

SO YOU SEE WHY THE ANSWER IS OBVIOUSLY NO.

That does not rule out all possible outs, but it is a vital thing to understand.

I am confident LLMs are not sentient or conscious, but your periodic reminder that the argument that they don’t have various biological or embodied characteristics is a terrible one, and Asimov’s prediction of this reaction was on point.

A few things going on here.

Jeffrey Ladish: I’m a bit sad about the state of AI discourse and governance right now. Lot of discussions about innovation vs. safety, what can / should the government actually do… but I feel like there is an elephant in the room

We’re rushing towards intelligent AI agents that vastly outstrip human abilities. A new non-biological species that will possess powers wonderful and terrible to behold. And we have no plan for dealing with that, no ability to coordinate as a species to avoid a catastrophic outcome

We don’t know exactly when we’ll get AI systems with superhuman capabilities… systems that can strategize, persuade, invent new technologies, etc. far better than we can. But it sure seems like these capabilities are in our sights. It sure seems like the huge investments in compute and scale will pay off, and people will build the kinds of systems AI risk researchers are most afraid of

If decision makers around the world can’t see this elephant in the room, I worry anything they try to do will fall far short of adequate.

Ashley Darkstone: Maybe if you and people like you stopped using biological/animist terms like “species” to refer to AI, you’d be taken more seriously.

Jeffrey Ladish: It’s hard to talk about something that is very different than anything that’s happened before. We don’t have good language for it. Do you have language you’d use to describe a whole other class of intelligent agent?

Ashley Darkstone: Only language specific to my work. We’ll all have to develop the language over time, along with the legalism, etc.

Species has specific implications to people. Life/Slavery/Evolution.. Biological/Human things that need not apply. It’s fearmongering.

AI should be a selfless tool.

Jeffrey Ladish: Maybe AI should be a selfless tool, but I think people train powerful agents

I studied evolutionary biology in college and thought a fair bit about different species concepts, all imperfect 🤷

“fearmongering” seems pretty dismissive of the risks at hand

is Darkstone objecting to the metaphorical use of a biological term because it is more confusing than helpful, more heat than light? Because it is technically incorrect, the worst kind of incorrect? Because it is tone policing?

Or is it exactly because of her belief that ‘AI should be a selfless tool’?

That’s a nice aspiration, but Ladish’s point is exactly that this won’t remain true.

More and more I view objections to AI risk as being rooted in not believing in the underlying technologies, rather than an actual functioning disagreement. And objections to the terminology and metaphors used being for the same reason: The terminology and metaphors imply that AGI and agents worthy of those names are coming, whereas objectors only believe in ATI (artificial tool inheritance).

Thus I attempt to coin the term ATI: Artificial Tool Intelligence.

Definition: Artificial Tool Intelligence. An intelligent system incapable of functioning as the core of a de facto autonomous agent.

If we were to only ever build ATIs, then that would solve most of our bigger worries.

That is a lot easier said than done.

Keegan McBride makes case that open source AI is vital for national security, because ‘Whoever builds, maintains, or controls the global open source AI ecosystem will have a powerful influence on our shared digital future.’

Toad: But our rivals can copy the open source models and modify them.

Frog: That is true. But that will ensure our cultural dominance, somehow?

Toad then noticed he was confused.

The post is filled with claims about China’s pending AI ascendancy, and to defend against that she says we need to open source our AIs.

I do give Keegan full credit for rhetorical innovation on that one.

It would be really great if we could know Anthropic was worthy of our trust.

  1. We know that Anthropic has cultivated a culture of caring deeply about safety, especially existential safety, among its employees. I know a number of its employees who have sent costly signals that they deeply care.

  2. We know that Anthropic is taking the problems far more seriously than its competitors, and investing more heavily in safety work.

  3. We know that Anthropic at least thinks somewhat about whether its actions will raise or lower the probability that AI kills everyone when it makes its decisions.

  4. We know they have the long term benefit trust and are a public benefit corporation.

  5. No, seriously, have you seen the other guy?

I have. It isn’t pretty.

Alas, the failure of your main rival to live up to ‘ordinary corporation’ standards does not change the bar of success. If Anthropic is also not up to the task, or not worthy of trust, then that is that.

I have said, for a while now, that I am confused about Anthropic. I expect to continue to be confused, because they are not making this easy.

Anthropic has a principle of mostly not communicating much, including on safety, and being extremely careful when it does communicate.

This is understandable. As their employees have said, there is a large tendency of people to read into statements, to think they are stronger or different than they are, that they make commitments the statement does not make. The situation is changing rapidly, so what seemed wise before might not be wise now. People and companies can and should change their minds. Stepping into such discussions often enflames them, making the problem worse, people want endless follow-ups, it is not a discussion you want to focus on. Talking about what the thing you are doing can endanger your ability to do the thing. Again, I get it.

Still? They are not making this easy. The plan might be wise, but the price must be paid. You go to update with the evidence you have. Failure to send costly signals is evidence, even if your actions plausibly make sense in a lot of different worlds.

What exactly did Anthropic promise or imply around not improving the state of the art? What exactly did they say to Dustin Moskovitz on this? Anthropic passed on releasing the initial Claude, but then did ship Claude Opus, and before that the first 100k context window.

To what extent is Anthropic the kind of actor who will work to give you an impression that suits its needs without that impacting its ultimate decisions? What should we make of their recent investor deck?

What public commitments has Anthropic actually made going forward? How could we hold them accountable? They have committed to their RSP, but most of it can be changed via procedure. Beyond that, not clear there is much. Will the benefit trust in practice have much effect especially in light of recent board changes?

What is up with Anthropic’s public communications?

Once again this week, we saw Anthropic’s public communications lead come out warning about overregulation, in ways I expect to help move the Overton window away from the things that are likely going to become necessary.

Simeon: Anthropic policy lead now advocating against AI regulation. What a surprise for an AGI lab 🤯

If you work at Anthropic for safety reasons, consider leaving.

That is Simeon’s reaction to a highly interesting retrospective by Jack Clark.

The lookback at GPT-2 and decisions around its release seems insightful. They correctly foresaw problems, and correctly saw the need to move off of the track of free academic release of models. Of course that GPT-2 was entirely harmless because it lacked sufficient capabilities, and in hindsight that seems very obvious, and part of the point is that it is hard to tell in advance. Here they ‘missed high’ but one could as easily ‘miss low.’

Then comes the part about policy. Here is the part being quoted, in context, plus key other passages.

Jack Clark: I’ve come to believe that in policy “a little goes a long way” – it’s far better to have a couple of ideas you think are robustly good in all futures and advocate for those than make a confident bet on ideas custom-designed for one specific future – especially if it’s based on a very confident risk model that sits at some unknowable point in front of you.

Additionally, the more risk-oriented you make your policy proposal, the more you tend to assign a huge amount of power to some regulatory entity – and history shows that once we assign power to governments, they’re loathe to subsequently give that power back to the people. Policy is a ratchet and things tend to accrete over time. That means whatever power we assign governments today represents the floor of their power in the future – so we should be extremely cautious in assigning them power because I guarantee we will not be able to take it back. 

For this reason, I’ve found myself increasingly at odds with some of the ideas being thrown around in AI policy circles, like those relating to needing a license to develop AI systems; ones that seek to make it harder and more expensive for people to deploy large-scale open source AI models; shutting down AI development worldwide for some period of time; the creation of net-new government or state-level bureaucracies to create compliance barriers to deployment.

Yes, you think the future is on the line and you want to create an army to save the future. But have you considered that your actions naturally create and equip an army from the present that seeks to fight for its rights?

Is there anything I’m still confident about? Yes. I hate to seem like a single-issue voter, but I had forgotten that in the GPT-2 post we wrote “we also think governments should consider expanding or commencing initiatives to more systematically monitor the societal impact and diffusion of AI technologies, and to measure the progression in the capabilities of such systems.” I remain confident this is a good idea!

This is at core not that different from my underlying perspective. Certainly it is thoughtful. Right now what we need most is to create broader visibility into what these systems are capable of, and to create the institutional capacity such that if we need to intervene in the future, we can do that.

Indeed, I have spoken how I feel proposals such as those in the Gladstone Report go too far, and would indeed carry exactly these risks. I draw a sharp contrast between that and something like SB 1047. I dive into the details to try and punch them up.

It still seems hard not to notice the vibes. This is written in a way that comes across as a warning against regulation. Coming across is what such communications are about. If this were an isolated example it would not bother me so much, but I see this consistently from Anthropic. If you are going to warn against overreach without laying out the stakes or pushing for proper reach, repeatedly, one notices.

Anthropic’s private lobbying and other private actions clearly happens and hopefully sings a very different tune, but we have no way of knowing.

Also, Anthropic failed to publicly share Claude Opus with the UK in advance, while Google did publicly share Gemini updates in advance. No commitments were broken, but this seems like a key place where it is important to set a good example. A key part of Anthropic’s thesis is that they will create a ‘race to safety’ so let’s race.

I consider Simeon’s reaction far too extreme. If you are internal, or considering becoming internal, you have more information. You should form your own opinion.

A nice positive detail: Anthropic has an anonymous hotline for reporting RSP compliance concerns. Of course, that only matters if they then act.

The Rand report on securing model weights is out.

Ideally this will become its own post in the future. It is super important that we secure the model weights of future more capable systems from a wide variety of potential threats.

As the value at stake goes up, the attacks get stronger, and so too must defenses.

The core message is that there is no silver bullet, no cheap and simple solution. There are instead many strategies to improve security via defense in depth, which will require real investment over the coming years.

Companies should want to do this on their own. Not investing enough in security makes you a target, and your extremely expensive model gets stolen. Even if there are no national security concerns or existential risks, that is not good for business.

That still makes it the kind of threat companies systematically underinvest in. It looks like a big expense until it looks cheap in hindsight. Failure is bad for business, but potentially far far worse for the world.

Thus, this is a place where government needs to step in, both to require and to assist. It is an unacceptable national security situation, if nothing else, for OpenAI, Google or Anthropic (or in the future certain others) not to secure their model weights. Mostly government ‘help’ is not something an AI lab will want, but cybersecurity is a potential exception.

For most people, all you need to take away is the simple ‘we need to do expensive defense in depth to protect model weights, we are not currently doing enough, and we should take collective action as needed to ensure this happens.’

There are highly valid reasons to oppose many other safety measures. There are even arguments that we should openly release the weights of various systems, now or in the future, once the developers are ready to do that.

There are not valid reasons to let bad actors exclusively get their hands on frontier closed model weights by using cyberattacks.

At minimum, you need to agree on what that means.

Will Depue: Alignment people have forgotten that the main goal of ai safety is to build systems that are aligned to the intent of the user, not the intent of the creators. this is a far easier problem.

I have noticed others calling this ‘user alignment,’ and so far that has gone well. I worry people will think this means aligning the user, but ‘alignment to the user’ is clunky.

For current models, ‘user alignment’ is indeed somewhat easier, although still not all that easy. And no, you cannot actually provide a commercial product that does exactly what the user wants. So you need to do a dance of both and do so increasingly over time.

The ‘alignment people’ are looking forward to future more capable systems, where user alignment will be increasingly insufficient.

Looking at Will’s further statements, this is very clearly a case of ‘mere tool.’ Will Depue does not expect AGI, rather he expects AI to remain a tool.

It was interesting to see Ted Sanders and Joshua Achiam, both at OpenAI, push back.

In addition to knowing what you want, you need to be able to know if you found it.

Daniel Kang claims that the GPT-4 system card was wrong, and that AI agent teams based on GPT-4 can now find and exploit zero-day vulnerabilities, his new version scoring 50% on his test versus 20% for previous agents and 0% for open-source vulnerability scanners. They haven’t tested Claude Opus or Gemini 1.5 yet.

I won’t read the details because triage, but the key facts to understand are that the agent frameworks will improve over time even if your system does not, and that it is extremely difficult to prove a negative. I can prove that your system can exploit zero day exploits by showing it exploiting a zero day exploit. You cannot prove that your system cannot do that simply by saying ‘I tried and it didn’t work,’ even if you gave it your best with the best agents you know about. You can of course often say that a given task is far outside of anything a model could plausibly do, but this was not one of those cases.

I do not think we have a practical problem in this particular case. Not yet. But agent system designs are improving behind the scenes, and some odd things are going to happen once GPT-5 drops.

Also, here we have DeepMind’s Nicholas Carlini once again breaks proposed AI defense techniques, here Sabre via changing one line of buggy code, then when the authors respond with a new strategy by modifying one more line of code. This thread has more context.

Analysis and notes of caution on Anthropic’s Scaling Monosemanticity (the Golden Gate Bridge) paper. We can be both super happy the paper happened, while also noticing that a lot of people are overreacting to it.

OpenAI gives us its early version of the SAE paper (e.g. the Golden Gate Bridge), searching for 16 million features in GPT-4, and claim their method scales better than previous work. Paper is here, Leo Gao is lead and coauthors include Sutskever and Leike. Not looking further because triage, so someone else please evaluate how we should update on this in light of Anthropic’s work.

Handy lists of various p(doom) numbers (pause AI, from the superforecasters and general surveys).

CAIS statement gets either ‘strongly agree’ or ‘agree’ from over 40% of Harvard students. Taking an AI class correlated with this being modestly higher, although I would guess causation mostly runs the other way.

Gabriel Wu: Students who have taken a class on AI were more likely to be worried about extinction risks from AI and had shorter “AGI timelines”: around half of all Harvard students who have studied artificial intelligence believe AI will be as capable as humans within 30 years.

Over half of Harvard students say that AI is changing the way they think about their careers, and almost half of them are worried that their careers will be negatively affected by AI.

How do automation concerns differ by industry? There’s isn’t much variation: around 40-50% of students are worried about AI automation no matter what sector they plan on working in (tech, education, finance, politics, research, consulting), with the exception of public health.

Full report is here.

Yann LeCun having strange beliefs department, in this case that ‘it is much easier to investigate what goes on in a deep learning system than in a turbojet, whether theoretically or experimentally.’ Judea Pearl explains it is the other way, whereas I would have simply said: What?

We also have the Yann LeCun providing unfortunate supporting links department.

Elon Musk is not always right, but when he’s right, he’s right.

Eliezer Yudkowsky: Very online people repeating each other: Eliezer Yudkowsky is a cult leader with a legion of brainwashed followers who obey his every word.

Real life: I wore this to LessOnline and Ozy Frantz stole my hat.

I do not have any such NDAs.

The last line is actually ‘member of an implicit coalition whose members coordinate to reward those who reward those who act to aid power and to prevent the creation of clarity around any and all topics including who may or may not have any form of NDA.’

Eternal September means the freshman philosophy beatings will continue.

I do note, however, that morale has slightly improved.

Say whatever else you want about e/acc. They will help you dunk.

Last week I had dinner with a group that included Emmett Shear, he made various claims of this type, and… well, he did not convince me of anything and I don’t think I convinced him of much either, but it was an interesting night. I was perhaps too sober.

Truth and reconciliation.

Indeed, someone is highly underpaid.

It is a no-good very-bad chart in so many other ways, but yeah, wow.

Updating a classic.

Narrator: They did not learn.

AI #67: Brief Strange Trip Read More »

us-agencies-to-probe-ai-dominance-of-nvidia,-microsoft,-and-openai

US agencies to probe AI dominance of Nvidia, Microsoft, and OpenAI

AI Antitrust —

DOJ to probe Nvidia while FTC takes lead in investigating Microsoft and OpenAI.

A large Nvidia logo at a conference hall

Enlarge / Nvidia logo at Impact 2024 event in Poznan, Poland on May 16, 2024.

Getty Images | NurPhoto

The US Justice Department and Federal Trade Commission reportedly plan investigations into whether Nvidia, Microsoft, and OpenAI are snuffing out competition in artificial intelligence technology.

The agencies struck a deal on how to divide up the investigations, The New York Times reported yesterday. Under this deal, the Justice Department will take the lead role in investigating Nvidia’s behavior while the FTC will take the lead in investigating Microsoft and OpenAI.

The agencies’ agreement “allows them to proceed with antitrust investigations into the dominant roles that Microsoft, OpenAI, and Nvidia play in the artificial intelligence industry, in the strongest sign of how regulatory scrutiny into the powerful technology has escalated,” the NYT wrote.

One potential area of investigation is Nvidia’s chip dominance, “including how the company’s software locks customers into using its chips, as well as how Nvidia distributes those chips to customers,” the report said. An Nvidia spokesperson declined to comment when contacted by Ars today.

High-end GPUs are “scarce,” antitrust chief says

Jonathan Kanter, the assistant attorney general in charge of the DOJ’s antitrust division, discussed the agency’s plans in an interview with the Financial Times this week. Kanter said the DOJ is examining “monopoly choke points and the competitive landscape” in AI.

The DOJ’s examination of the sector encompasses “everything from computing power and the data used to train large language models, to cloud service providers, engineering talent and access to essential hardware such as graphics processing unit chips,” the FT wrote.

Kanter said regulators are worried that AI is “at the high-water mark of competition, not the floor” and want to take action before smaller competitors are shut out of the market. The GPUs needed to train large language models are a “scarce resource,” he was quoted as saying.

“Sometimes the most meaningful intervention is when the intervention is in real time,” Kanter told the Financial Times. “The beauty of that is you can be less invasive.”

Microsoft deal scrutinized

The FTC is scrutinizing Microsoft over a March 2024 move in which it hired the CEO of artificial intelligence startup Inflection and most of the company’s staff and paid Inflection $650 million as part of a licensing deal to resell its technology. The FTC is investigating whether Microsoft structured the deal “to avoid a government antitrust review of the transaction,” The Wall Street Journal reported today.

“Companies are required to report acquisitions valued at more than $119 million to federal antitrust-enforcement agencies, which have the option to investigate a deal’s impact on competition,” the WSJ wrote. The FTC reportedly sent subpoenas to Microsoft and Inflection in an attempt “to determine whether Microsoft crafted a deal that would give it control of Inflection but also dodge FTC review of the transaction.”

Inflection built a large language model and a chatbot called Pi. Former Inflection employees are now working on Microsoft’s Copilot chatbot.

“If the agency finds that Microsoft should have reported and sought government review of its deal with Inflection, the FTC could bring an enforcement action against Microsoft,” the WSJ report said. “Officials could ask a court to fine Microsoft and suspend the transaction while the FTC conducts a full-scale investigation of the deal’s impact on competition.”

Microsoft told the WSJ that it complied with antitrust laws, that Inflection continues to operate independently, and that the deals gave Microsoft “the opportunity to recruit individuals at Inflection AI and build a team capable of accelerating Microsoft Copilot.”

OpenAI

Microsoft’s investment in OpenAI has also faced regulatory scrutiny, particularly in Europe. Microsoft has a profit-sharing agreement with OpenAI.

Microsoft President Brad Smith defended the partnership in comments to the Financial Times this week. “The partnerships that we’re pursuing have demonstrably added competition to the marketplace,” Smith was quoted as saying. “I might argue that Microsoft’s partnership with OpenAI has created this new AI market,” and that OpenAI “would not have been able to train or deploy its models” without Microsoft’s help, he said.

We contacted OpenAI today and will update this article if it provides any comment.

In January 2024, the FTC launched an inquiry into AI-related investments and partnerships involving Alphabet, Amazon, Anthropic, Microsoft, and OpenAI.

The FTC also started a separate investigation into OpenAI last year. A civil investigative demand sent to OpenAI focused on potentially unfair or deceptive privacy and data security practices, and “risks of harm to consumers, including reputational harm.” The probe focused partly on “generation of harmful or misleading content.”

US agencies to probe AI dominance of Nvidia, Microsoft, and OpenAI Read More »