Author name: Paul Patrick

nasa-hasn’t-landed-on-the-moon-in-decades—china-just-sent-its-third-in-six-years

NASA hasn’t landed on the Moon in decades—China just sent its third in six years

Marching on —

China is going. NASA is talking about going. What gives?

A Long March 5 rocket carrying the Chang'e-6 lunar probe blasts off from the Wenchang Space Launch Center on May 3, 2024 in Wenchang, China.

Enlarge / A Long March 5 rocket carrying the Chang’e-6 lunar probe blasts off from the Wenchang Space Launch Center on May 3, 2024 in Wenchang, China.

Li Zhenzhou/VCG via Getty Images

China is going back to the Moon for more samples.

On Friday the country launched its largest rocket, the Long March 5, carrying an orbiter, lander, ascent vehicle, and a return spacecraft. The combined mass of the Chang’e-6 spacecraft is about 8 metric tons, and it will attempt to return rocks and soil from the far side of the Moon—something scientists have never been able to study before in-depth.

The mission’s goal is to bring about 2 kg (4.4 pounds) of rocks back to Earth a little more than a month from now.

Chang’e-6 builds upon the Chinese space program’s successful lunar program. In 2019, the Chang’e-4 mission made a soft landing on the far side of the Moon, the first time this had ever been done by a spacecraft. The far side is more challenging than the near side, because line-of-sight communications are not possible with Earth.

Then, in late 2020, the Chang’e-5 mission landed on the near side of the Moon and successfully collected 1.7 kg of rocks. These were subsequently blasted off the surface of the Moon and returned to China where they have been studied since. It marked the first time in half a century, since efforts by the United States and Soviet Union, that samples were returned from the Moon.

Ambitious plans

The latest Chinese flight to the Moon launched Friday will synthesize the country’s learnings from its last two missions, by collecting and returning samples from the far side of the Moon.

“If the Chang’e-6 mission can achieve its goal, it will provide scientists with the first direct evidence to understand the environment and material composition of the far side of the moon, which is of great significance,” said Wu Weiren, an academician of the Chinese Academy of Engineering and chief designer of China’s lunar exploration program.

This mission follows the launch and deployment of the Queqiao-2 relay satellite in March, which will serve as a bridge between communications from the far side of the Moon to operators back on Earth. China has also announced two future lunar missions, Chang’e-7 and Chang’e-8, later this decade. These robotic missions will land near the lunar South Pole, test lunar resources, and prepare the way for future crewed missions.

Nominally, China’s current plan calls for the first landing of two taikonauts on the surface of the Moon in 2029 or 2030. Eventually it wants to establish a lunar outpost.

China’s lunar missions are not operating in a vacuum—OK, technically, they are—but the point here is that China’s exploration efforts are proceeding alongside a parallel effort by the United States, NASA, and about three dozen partners under the auspices of the Artemis program.

Can NASA compete?

After decades of focusing its exploration efforts elsewhere, NASA finally turned back to the Moon about seven years ago. Since that time it has worked alongside the commercial space industry to develop a plan for a sustainable return to the lunar surface.

From the outside, China’s lunar program appears to be in the lead. It is difficult to argue about the string of successes with the Chang’e lunar program and the unprecedented landing on the far side of the Moon. If Chang’e-6 proves successful, that will be another strike in favor of China’s lunar program.

But to its credit, NASA is not simply seeking to replicate the glories of its Apollo lunar program in the 1960s and early 1970s. China’s first lunar mission with astronauts, for example, is intended to land two taikonauts on the Moon for just a few hours. The vehicles will be fully expendable, as were the Apollo rockets and spacecraft more than half a century ago.

NASA is taking a different approach, working with industry to develop a fleet of commercial cargo landers—such as Intuitive Machines’ largely successful Odysseus mission earlier this year—as well as larger human landers built by SpaceX and Blue Origin. This overall “architecture” is far more complex, requiring myriad launches to refuel spacecraft in orbit. It will likely take several years longer to get to the first lunar landing missions, either later this decade or earlier in the 2030s. But should NASA persist and succeed in this approach, it will open up a highway to the Moon the likes of which could only be dreamed of during the Apollo era. Imagine a flotilla of spacecraft going to and from the Moon. That’s the vision.

So it’s a competition between China’s embrace of a traditional approach versus NASA’s efforts to open the way into some kind of new future. Watching how this lunar competition unfolds over the next decade will be one of the most fascinating stories to follow.

NASA hasn’t landed on the Moon in decades—China just sent its third in six years Read More »

mini-settlers-is-a-city-builder-that-you-can-both-enjoy-and-actually-put-down

Mini Settlers is a city builder that you can both enjoy and actually put down

You can definitely get 120 frames on an RTX 4080 —

No zoning, no pollution, no advisers—just squares, circles, people, and time.

Mini Settlers screen showing rocks, fields, and lots of water pumps and farms.

Enlarge / Are you enticed by this kind of orderly madness with a clean graphical layout? Then I suggest you… settle in.

Goblinz Studio

You can’t buy Mini Settlers right now, but I think you should play the free “Prologue” demo and wishlist the full game if you dig it. It’s not quite like any other city builder I’ve played.

Mini Settlers is “mini” like minimalism. It is in the same genre, but quite far from, games like Cities: Skylines 2 (a choice with some proven merit). Your buildings are not 3D-rendered with real-time lighting. Your buildings are colored squares, sometimes with a few disc tokens stacked on them, tabletop-style. Your roads don’t have traffic, but they have drivers (tiny squares) that take resources between nodes. When things go wrong, you don’t get depressing news about pollution and riots; some people just leave their homes, but they’ll come back if you fix what’s wrong.

Mini Settlers announcement video.

Mini Settlers is not the game to play to satisfy your long-running suspicion that urban planning was your missed calling. In the (non-progress-saving) Prologue-free demo out this week, the mines and quarries have infinite resources. There is no “money” to speak of, so far as I can tell. Apple farms must be placed near apple orchards and water pumps by water, and the rest is up to you. The interface looks like a thought experiment in how far you can get from traditional city sim HUDs, but then someone implemented it.

  • A larger-scale view of a developed settlement, one with much better road planning than I achieved.

    Goblinz Studio

  • The game layers information about resources and needs, such that it never feels overwhelming.

    Goblinz Studio

  • Natural resources and land formations require you to work around them in creative ways.

    Goblinz Studio

  • Each circle is a node, and each square is a worker, shuttling resources from node to node, as best they can.

    Goblinz Studio

The biggest challenge I faced in my couple of sessions was textbook logistics, at least from a suburban or small-town perspective. Having developed SimCity Brain throughout prior decades, I tried to keep my residential areas (City Center and the Homes you build around it) away from anything resembling production, like rock quarries and lumber yards. Instead of bolstering housing values or improving aesthetics, which do not exist, this gave me a huge set of supply bottlenecks to try and work through.

Houses wanted regular supplies of apples and water, but spacing out everything made a ton of extra transit work. Every road is a maximum of seven tiles, and each one gets a worker that moves back and forth between waypoints, dropping off goods to buildings or leaving them for the next worker on the goods’ route. I had wanted to create a simple town of people building wood houses and eating apples, and instead, I had a micro-scale Wayfair job interview scenario, complete with tiny warehouses and delivery times.

But, here again, Mini Settlers is different, even when you’re flailing. You simply remove the roads and buildings that don’t work and put them in better places. The buildings take a bit to build again, but there’s no real game timer unless you want to enable one for personal bests. You can even enable a background mode so that the calm simulation keeps running while you absolutely do your best work on a Friday afternoon.

The “Prologue” is not verified for Steam Deck, but the developers have an official layout for it. I think it will do in a pinch, but there’s a lot of thumb-taxing trackpad pointing remaining in a game that seems grid-based enough to do with more gamepad controls. As for performance, it runs great. At 30 frames per second, my Deck guessed it could keep going for nearly five more hours.

Mini Settlers is due out in 2024, seemingly for PC only on Steam, for the moment. The minimum requirements are a Core i3, 4GB memory, and Intel HD Graphics 4000, but “Integrated cards also work.” As the developers at Knight Owl Games note, wishlisting the game helps it circulate inside Steam’s recommendation algorithm, even if you don’t ultimately play beyond the demo. I am going to note a second time here that the demo does not save your game when you exit, which is not another design choice to keep you calm but just a demo thing.

Between this and Against the Storm, I am enjoying the recent broadening of the “city builder” genre. It’s happening, weirdly enough, by going much smaller.

Listing image by Goblinz Studio

Mini Settlers is a city builder that you can both enjoy and actually put down Read More »

rocket-report:-astroscale-chases-down-dead-rocket;-ariane-6-on-the-pad

Rocket Report: Astroscale chases down dead rocket; Ariane 6 on the pad

RIP B1060 —

Rocket Factory Augsburg, a German launch startup, nears a test-firing of its booster.

This image captured by Astroscale's ADRAS-J satellite shows the discarded upper stage from a Japanese H-IIA rocket.

Enlarge / This image captured by Astroscale’s ADRAS-J satellite shows the discarded upper stage from a Japanese H-IIA rocket.

Welcome to Edition 6.42 of the Rocket Report! Several major missions are set for launch in the next few months. These include the first crew flight on Boeing’s Starliner spacecraft, set for liftoff on May 6, and the next test flight of SpaceX’s Starship rocket, which could happen before the end of May. Perhaps as soon as early summer, SpaceX could launch the Polaris Dawn mission with four private astronauts, who will perform the first fully commercial spacewalk in orbit. In June or July, Europe’s new Ariane 6 rocket is slated to launch for the first time. Rest assured, Ars will have it all covered.

As always, we welcome reader submissions, and if you don’t want to miss an issue, please subscribe using the box below (the form will not appear on AMP-enabled versions of the site). Each report will include information on small-, medium-, and heavy-lift rockets as well as a quick look ahead at the next three launches on the calendar.

German rocket arrives at Scottish spaceport. Rocket Factory Augsburg has delivered a booster for its privately developed RFA One rocket to SaxaVord Spaceport in Scotland, the company announced on X. The first stage for the RFA One rocket was installed on its launch pad at SaxaVord to undergo preparations for a static fire test. The booster arrived at the Scottish launch site with five of its kerosene-fueled Helix engines. The remaining four Helix engines, for a total of nine, will be fitted to the RFA One booster at SaxaVord, the company said.

Aiming to fly this year… RFA hopes to launch its first orbital-class rocket by the end of 2024. The UK’s Civil Aviation Authority last month granted a range license to SaxaVord Spaceport to allow the spaceport operator to control the sea and airspace during a launch. RFA is primarily privately funded but has won financial support from the European Space Agency, the UK Space Agency, and the German space agency, known as DLR. The RFA One rocket will have three stages, stand nearly 100 feet (30 meters) tall, and can carry nearly 2,900 pounds (1,300 kilograms) of payload into a polar Sun-synchronous orbit.

Arianespace wins ESA launch contract. The European Space Agency has awarded Arianespace a contract to launch a joint European-Chinese space science satellite in late 2025, European Spaceflight reports. The Solar wind Magnetosphere Ionosphere Link Explorer (SMILE) is a 4,850-pound (2,200-kilogram) spacecraft that will study Earth’s magnetic environment on a global scale. The aim of the mission is to build a more complete understanding of the Sun-Earth connection. On Tuesday, ESA officially signed a contract for Arianespace to launch SMILE aboard a Vega C rocket, which is built by the Italian rocket-maker Avio.

But it may not keep it … In late 2023, ESA member states agreed to allow Avio to market and manage the launch of Vega C flights independent of Arianespace. When the deal was initially struck, 17 flights were contracted through Arianespace to be launched aboard Vega vehicles. While these missions are still managed by Arianespace, Avio is working with the launch provider to strike a deal that would allow the Italian rocket builder to assume the management of all Vega flights. The Vega C rocket has been grounded since a launch failure in 2022 forced Avio to redesign the nozzle of the rocket’s solid-fueled second-stage motor. Vega C is scheduled to return to flight before the end of 2024. (submitted by Ken the Bin)

The easiest way to keep up with Eric Berger’s space reporting is to sign up for his newsletter, we’ll collect his stories in your inbox.

Update on ABL’s second launch. ABL Space Systems expected to launch its second light-class RS1 rocket earlier this year, but the company encountered an anomaly during ground testing at the launch site in Alaska, according to Aria Alamalhodaei of TechCrunch. Kevin Sagis, ABL’s chief engineer, said there is “no significant delay” in the launch of the second RS1 rocket, but the company has not announced a firm schedule. “During ground testing designed to screen the vehicle for flight, an issue presented that caused us to roll back to the hangar,” Sagis said, according to Alamalhodaei. “We have since resolved and dispositioned the issue. There was no loss of hardware and we have validated vehicle health back out on the pad. We are continuing with preparations for static fire and launch.”

Nearly 16 months without a launch … ABL’s first RS1 test flight in January 2023 ended seconds after liftoff with the premature shutdown of its liquid-fueled engines. The rocket crashed back onto its launch pad in Alaska. An investigation revealed a fire in the aft end of the RS1 booster burned through wiring harnesses, causing the rocket to lose power and shut off its engines. Engineers believe the rocket’s mobile launch mount was too small, placing the rocket too close to the ground when it ignited its engines. This caused the hot engine exhaust to recirculate under the rocket and led to a fire in the engine compartment as it took off.

Rocket Report: Astroscale chases down dead rocket; Ariane 6 on the pad Read More »

apple’s-q2-2024-earnings-reveal-a-drop-in-iphone,-ipad-sales

Apple’s Q2 2024 earnings reveal a drop in iPhone, iPad sales

Q2 2024 —

Services growth looked rosy as Apple’s hardware revenue in China slowed.

The Apple Park campus in Cupertino, California.

Enlarge / The Apple Park campus in Cupertino, California.

Apple’s earnings report for the second quarter of the company’s 2024 fiscal year showed a slide in hardware sales, especially for the iPhone. Nonetheless, Apple beat analysts’ estimates for the quarter thanks to the company’s rapidly growing services revenue.

iPhone revenue dropped from $51.33 billion in the same quarter last year to $45.96 billion, a fall of about 10 percent. This was the second consecutive quarter with declining iPhone revenues. That said, investors feared a sharp drop before the earnings call.

Notably, Apple’s revenue in the region it dubs Greater China (which includes China, Taiwan, Singapore, and Hong Kong) fell 8 percent overall. The company fared a little better in other regions. China’s economy is slowing even as China-based Huawei is taking bigger slices of the pie in the region.

Globally, Mac revenue was $7.5 billion compared to last year’s $7.12 billion. Other products—which include the Watch, AirPods, Apple TV 4K, HomePod, and the new Vision Pro headset—were down to $7.9 billion from last year’s $8.76 billion, despite the fact this quarter included the launch of the Vision Pro.

iPad revenue was also down, at $5.6 billion from $6.67 billion. Apple is expected to launch new iPads next week, which suggests that those updates are needed to achieve the company’s business goals.

The rosiest revenue category was services, which includes everything from Apple Music to iCloud. Its revenue was $23.9 billion, up from Q2 2023’s $20.91 billion.

The company also announced authorization of $110 billion for share purchases.

Apple’s Q2 2024 earnings reveal a drop in iPhone, iPad sales Read More »

at&t-announces-$7-monthly-add-on-fee-for-“turbo”-5g-speeds

AT&T announces $7 monthly add-on fee for “Turbo” 5G speeds

A pedestrian walks past a large AT&T logo on the glass exterior of an AT&T store.

Getty Images | Bloomberg

AT&T is now charging mobile customers an extra $7 per month for faster wireless data speeds. AT&T says the Turbo add-on, available starting today, is “built to support high-performance mobile applications, like gaming, social video broadcasting and live video conferencing, with optimized data while customers are on the go.”

While Turbo “boosts all the high-speed and hotspot data on a user’s connection,” AT&T said the difference will be more noticeable for certain kinds of applications. For example, gaming applications using Turbo will experience “less freezing or stuttering and lower latency,” AT&T said.

The $7 charge is for each line. Adding Turbo to multiple lines on the same account requires paying the extra fee for each line. AT&T said that Turbo lets users “optimize their plan’s high-speed (premium) and hotspot data allotments” and provides better data performance “even during busy times on the network.”

Turbo is only available for 5G phones on certain “unlimited” plans. AT&T notes that “Turbo does not provide extra data” and that “if you exceed your existing allotments your normal network management applies.”

“On AT&T Unlimited Extra EL after 75GB, AT&T may temporarily slow data speeds if the network is busy,” the company says. “On each eligible plan, after you exceed your hotspot allotment, your hotspot speeds are slowed to a maximum of 128Kbps.”

People who pay extra for Turbo might want to look at their video settings. By default, AT&T limits video streaming to DVD quality, but customers can turn on high-definition video at the expense of using more data.

Quality of service

An article by The Mobile Report said that AT&T will differentiate between users who pay for Turbo and those who don’t with Quality of Service Class Identifiers, or QCIs. “We’re told that, basically, all eligible plans are now moved to QCI 8, and get the privilege of buying their way back into QCI 7,” the article said. QCI 6 is reportedly reserved for public safety professionals on the FirstNet service built by AT&T under a government contract.

AT&T confirmed to Ars today that Turbo “is assigned to a QCI to which some of our consumer traffic was previously assigned.” But AT&T said it has “materially modified it and increased network resources and relative weighting for AT&T Turbo traffic, thereby creating a higher level of performance than we’ve ever before offered to consumers.”

AT&T also said that QCIs “are simply a number assigned to a class of service,” and that the “treatment and performance of traffic in a particular class is affected by a range of variables that can be tuned to provide different experiences.” AT&T said that last summer, it “rationalized and streamlined how our plans are mapped to QCI levels” and that “these changes helped optimize network performance for our overall customer base.”

The current version of Turbo may be followed by other paid extras that enhance performance, as AT&T called it the “first step in modernizing and preparing our mobile network for future innovative use cases… Latency-sensitive applications will continue to need more enhanced network technologies to perform their best, so we plan to continue to advance and evolve AT&T Turbo.”

AT&T announces $7 monthly add-on fee for “Turbo” 5G speeds Read More »

apple-deal-could-have-been-“suicide”-for-google,-company-lawyer-says

Apple deal could have been “suicide” for Google, company lawyer says

Woulda coulda shoulda? —

Judge: What should Google have done to avoid the DOJ’s crosshairs?

John Schmidtlein, partner at Williams & Connolly LLP and lead litigator for Alphabet Inc.'s Google, arrives to federal court in Washington, DC, US, on Monday, Oct. 2, 2023.

Enlarge / John Schmidtlein, partner at Williams & Connolly LLP and lead litigator for Alphabet Inc.’s Google, arrives to federal court in Washington, DC, US, on Monday, Oct. 2, 2023.

Halfway through the first day of closing arguments in the Department of Justice’s big antitrust trial against Google, US District Judge Amit Mehta posed the question that likely many Google users have pondered over years of DOJ claims that Google’s market dominance has harmed users.

“What should Google have done to remain outside the crosshairs of the DOJ?” Mehta asked plaintiffs halfway through the first of two full days of closing arguments.

According to the DOJ and state attorneys general suing, Google has diminished search quality everywhere online, primarily by locking rivals out of default positions on devices and in browsers. By paying billions for default placements that the government has argued allowed Google to hoard traffic and profits, Google allegedly made it nearly impossible for rivals to secure enough traffic to compete, ultimately decreasing competition and innovation in search by limiting the number of viable search engines in the market.

The DOJ’s lead litigator, Kenneth Dintzer, told Mehta that what Google should have done was acknowledge that the search giant had an enormous market share and consider its duties more carefully under antitrust law. Instead, Dintzer alleged, Google chose the route of “hiding” and “destroying documents” because it was aware of conflicts with antitrust law.

“What should Google have done?” Dintzer told Mehta. “They should have recognized that by demanding locking down every default that they were opening themselves up to a challenge on the conduct.”

The most controversial default agreement that Google has made is a 21-year deal with Apple that Mehta has described as the “heart” of the government’s case against Google. During the trial, a witness accidentally blurted out Google’s carefully guarded secret of just how highly it values the Apple deal, revealing that Google pays 36 percent of its search advertising revenue from Safari just to remain the default search tool in Apple’s browser. In 2022 alone, trial documents revealed that Google paid Apple $20 billion for the deal, Bloomberg reported.

That’s in stark contrast to the 12 percent of revenue that Android manufacturers get from their default deals with Google. The government wants the court to consider all these default deals to be anti-competitive, with Dintzer suggesting during closing arguments that they are the “centerpiece” of “a lot” of Google’s exclusionary behavior that ultimately allowed Google to become the best search engine today—by “capturing the default and preventing rivals from getting access to those defaults.”

Google’s lawyers have argued that Google succeeds on its merits. Today, lead litigator John Schmidtlein repeatedly pointed out that antitrust law is designed to protect the competitive process, not specific competitors who fail to invest and innovate—as Microsoft did by failing to recognize how crucial mobile search would become.

“Merely getting advantages by winning on quality, they may have an effect on a rival, but the question is, does it have an anti-competitive effect?” Schmidtlein argued, noting that the DOJ hadn’t “shown that absent the agreements, Microsoft would have toppled Google.”

But Dintzer argued that “a mistake by one rival doesn’t mean that Google gets to monopolize this market forever.” When asked to explain why everyone—including some of Google’s rivals—testified that Google won contracts purely because it was the best search engine, Dintzer warned Mehta that the fact that Google’s rivals “may be happy cashing Google’s checks doesn’t tell us anything.”

According to Schmidtlein, Google could have crossed the line with the Apple deal, but it didn’t.

“Google didn’t go on to say to Apple, if you don’t make us the default, no Google search on Apple devices at all,” Schmidtlein argued. “That would be suicide for Google.”

It’s still unclear how Mehta may be leaning in this case, interrogating both sides with care and making it clear that he expects all his biggest questions to be answered after closing arguments conclude Friday evening.

But Mehta did suggest at one point today that it seemed potentially “impossible” for anyone to compete with Google for default placements.

“How would anybody be able to spend billions and billions of dollars to possibly dislodge Google?” Mehta asked. “Is there any real competition for the default spot?”

According to Schmidtlein, that is precisely what “competition on the merits” looks like.

“Google is winning because it’s better, and Apple is deciding Google is better for users,” Schmidtlein argued. “The antitrust laws are not designed to ensure a competitive market. They’re designed to ensure a competitive process.”

Proving the potential anti-competitive effects of Google’s default agreements, particularly the Apple deal, has long been regarded as the most critical point in order to win the government’s case. So it’s no surprise that the attorney representing state attorneys general, Bill Cavanaugh, praised Mehta for asking, “What should Google have done?” According to Cavanaugh, that was the “right question” to pose in this trial.

“What should they have done 10 years ago when there was a recognition” that “we’re monopolists” and “we have substantial control in markets” is ask, “How should we proceed with our contracts?” Cavanaugh argued. “That’s the question that they answered, but they answered it in the wrong way.”

Seemingly if Google’s default contracts posed fewer exclusionary concerns, the government seems to be arguing, there would be more competition and therefore more investment and innovation in search. But as long as Google controls the general search market, the government alleged that users won’t be able to search the web the way that they want.

Google is hoping that Mehta will reject the government’s theories and instead rule that Google has done nothing to stop rivals from improving the search landscape. Early in the day, Mehta told the DOJ that he was “struggling to see” how Google has either stopped innovating or degraded its search engine as a result of lack of competition.

Closing arguments continue on Friday. Mehta is not expected to rule until late summer or early fall.

Apple deal could have been “suicide” for Google, company lawyer says Read More »

uhf-in-uhd:-weird-al’s-cult-classic-movie-will-get-its-first-4k-release

UHF in UHD: Weird Al’s cult classic movie will get its first 4K release

MY MOP! —

For those of you just joining us, today we’re teaching poodles how to fly.

  • Weird Al’s Rambo parody was a drop in the bucket amidst all the other jokes in the film, but it’s among the most memorable.

    Shout Factory

  • This is the promotional image for the collector’s edition with all its physical knickknacks.

    Shout Factory

Believe it or not, it’s been 35 years since Weird Al’s quotable cult classic UHF first came out. Right on time for that anniversary, Shout Factory will release an UltraHD Blu-ray of the movie. This will be the first time it has ever been available in 4K.

Releasing July 2 but pre-ordering now, the disc will include a new 4K scan of the original 35mm negative, along with audio commentary from Weird Al and Jay Levy, the film’s director.

It will also come bundled with a standard HD Blu-ray that includes the film in that older format along with a bunch of special features, including video of a 2014 Comic-Con panel on the movie, deleted scenes, behind-the-scenes videos, and some other assets. Some of those return from the movie’s last physical edition, which was a 25th anniversary HD Blu-ray, but not 4K.

There will be deluxe editions that include some physical collectibles, including an 18×24-inch poster of the “original theatrical artwork,” as well as a new, same-sized poster of new poster art made for this edition. You’ll also find 10 scratch-and-sniff stickers alongside a guide with time prompts for using them, plus some stickers “designed to replicate vintage vending machine prism stickers from the late ’80s and early ’90s” and a Spatula City fridge magnet. Add to that a 6-inch “UHF Remote Control Stress Relief Collectible.” All that stuff is limited to 1,000 units.

For an even smaller number of units of the collector’s edition (500), there will be five UHF-themed hard enamel pins.

The set is available in four tiers priced at $40, $53, $76, and $130, which is a mess, but if you’re not interested in collecting all the physical doohickies, it’s that first price for just the movie that you need to know.

UHF was released in 1989, and it was parody musician Weird Al’s first movie starring role and writing credit. Conceived as a series of bits that would allow him to satirize films in the same way he was known for satirizing songs, it, unfortunately, was a box office flop. It gained a small and passionate cult following on VHS throughout the ’90s.

Another movie written in part by Weird Al, Weird: The Al Yankovich Story, was released on Roku’s streaming channel in 2022. It was a very different kind of movie. Instead of rapid-fire spoofing numerous films like UHF did, it spoofed the musical biopic genre, with Daniel Radcliffe playing Weird Al in a heavily fictionalized account of his life.

The limited-run nature of this UHF release suggests that while the film still has its cult following, it remains outside the mainstream. Its fans probably like it that way, though.

Listing image by Shout Factory

UHF in UHD: Weird Al’s cult classic movie will get its first 4K release Read More »

one-and-done:-elden-ring’s-first-dlc-expansion-will-also-be-its-last

One and done: Elden Ring’s first DLC expansion will also be its last

Over and out —

But the studio is “leaving some possibilities” for a sequel that continues the story.

A big erdtree casts a big shadow.

Enlarge / A big erdtree casts a big shadow.

Namco Bandai

The good news for Elden Ring fans is that the two-plus-year wait for the game’s first DLC, “Shadow of the Erdtree,” will end in just a couple of months. The bad news is that “Shadow of the Erdtree” will also be the last bit of DLC for FromSoftware’s multimillion-selling action RPG.

In a wide-ranging interview with Chinese site Zhihu (machine translation), Elden Ring producer Hidetaka Miyazaki said “Shadow of the Erdtree” contains a lot of existing lore and content that was created for the original game but couldn’t fit into the final package. Miyazaki said the team decided to release all of that unused content as one large DLC expansion, rather than multiple smaller bits, because “if they were sold separately, the freedom of exploration and sense of adventure would be reduced.”

As for just how big the DLC will be, Miyazaki balked when the interviewer asked how long it would take players to complete. Miyazaki brought up memories of being called a liar after estimating in an earlier interview that the original game would only take about 30 hours of play to complete—crowdsourced game-length database HowLongToBeat puts the “main story” estimate closer to 60 hours.

While Miyazaki was definitive on the lack of plans for additional Elden Ring DLC, he left open the possibility of further games that could continue the story of the Elden Ring universe. “FromSoftware’s style of doing things is that it generally does not allow the future of an IP to be easily locked, but it is better to leave some possibilities” he said.

Previous games in FromSoftware’s Dark Souls series have split additional content across multiple episodic DLC expansions in the years after their respective releases (though Bloodborne only saw a single DLC expansion, “The Old Hunters“). Last year, Cyberpunk 2077 developer CD Projekt Red confirmed that the “Phantom Liberty” expansion would be that game’s only DLC, owing to an engine transition within the company.

Elsewhere in the interview, Miyazaki said the upcoming expansion would closely match the legendary difficulty of the original game’s second half. The DLC was designed with the expectation that “players… should already have a certain understanding of the game,” he said. But players who have grinded their characters into unstoppable, overpowered machines will be able to turn off the leveling system in the DLC area to add a bit of additional challenge.

Miyazaki also offered a few hints about the DLC’s plot, which will include new characters like St. Trina, a counterpart to Miquella, who was only hinted at via item names in the original game. Most characters featured in the DLC will be completely new to the expansion, Miyazaki said, if for no other reason than that “there may be a situation where the player kills the NPC [in the original game], causing him/her to be unable to appear in the DLC story.”

One and done: Elden Ring’s first DLC expansion will also be its last Read More »

maximum-severity-gitlab-flaw-allowing-account-hijacking-under-active-exploitation

Maximum-severity GitLab flaw allowing account hijacking under active exploitation

A 10 OUT OF 10 —

The threat is potentially grave because it could be used in supply-chain attacks.

Maximum-severity GitLab flaw allowing account hijacking under active exploitation

A maximum severity vulnerability that allows hackers to hijack GitLab accounts with no user interaction required is now under active exploitation, federal government officials warned as data showed that thousands of users had yet to install a patch released in January.

A change GitLab implemented in May 2023 made it possible for users to initiate password changes through links sent to secondary email addresses. The move was designed to permit resets when users didn’t have access to the email address used to establish the account. In January, GitLab disclosed that the feature allowed attackers to send reset emails to accounts they controlled and from there click on the embedded link and take over the account.

While exploits require no user interaction, hijackings work only against accounts that aren’t configured to use multifactor authentication. Even with MFA, accounts remained vulnerable to password resets, but the attackers ultimately are unable to access the account, allowing the rightful owner to change the reset password. The vulnerability, tracked as CVE-2023-7028, carries a severity rating of 10 out of 10.

On Wednesday, the US Cybersecurity and Infrastructure Security Agency said it is aware of “evidence of active exploitation” and added the vulnerability to its list of known exploited vulnerabilities. CISA provided no details about the in-the-wild attacks. A GitLab representative declined to provide specifics about the active exploitation of the vulnerability.

The vulnerability, classified as an improper access control flaw, could pose a grave threat. GitLab software typically has access to multiple development environments belonging to users. With the ability to access them and surreptitiously introduce changes, attackers could sabotage projects or plant backdoors that could infect anyone using software built in the compromised environment. An example of a similar supply chain attack is the one that hit SolarWinds in 2020 and pushed malware to more than 18,000 of its customers, 100 of whom received follow-on hacks. Other recent examples of supply chain attacks are here, here, and here.

These sorts of attacks are powerful. By hacking a single, carefully selected target, attackers gain the means to infect thousands of downstream users, often without requiring them to take any action at all.

According to Internet scans performed by security organization Shadowserver, more than 2,100 IP addresses showed they were hosting one or more vulnerable GitLab instances.

Shadowserver

The biggest concentration of IP addresses was in India, followed by the US, Indonesia, Algeria, and Thailand.

Shadowserver

The number of IP addresses showing vulnerable instances has fallen over time. Shadowserver shows that there were more than 5,300 addresses on January 22, one week after GitLab issued the patch.

Shadowserver

The vulnerability is classed as an improper access control flaw.

CISA has ordered all civilian federal agencies that have yet to patch the vulnerability to do so immediately. The agency made no mention of MFA, but any GitLab users who haven’t already done so should enable it, ideally with a form that complies with the FIDO industry standard.

GitLab users should also remember that patching does nothing to secure systems that have already been breached through exploits. GitLab has published incident response guidance here.

Maximum-severity GitLab flaw allowing account hijacking under active exploitation Read More »

geforce-now-has-made-steam-deck-streaming-much-easier-than-it-used-to-be

GeForce Now has made Steam Deck streaming much easier than it used to be

Easy, but we’re talking Linux easy —

Ask someone who previously did it the DIY way.

Fallout 4 running on a Steam Deck through GeForce Now

Enlarge / Streaming Fallout 4 from GeForce Now might seem unnecessary, unless you know how running it natively has been going.

Kevin Purdy

The Steam Deck is a Linux computer. There is, technically, very little you cannot get running on it, given enough knowledge, time, and patience. That said, it’s never a bad thing when someone has done all the work for you, leaving you to focus on what matters: sneaking game time on the couch.

GeForce Now, Nvidia’s game-streaming service that uses your own PC gaming libraries, has made it easier for Steam Deck owners to get its service set up on their Deck. On the service’s Download page, there is now a section for Gaming Handheld Devices. Most of the device links provide the service’s Windows installer, since devices like the ROG Ally and Lenovo Legion Go run Windows. Some note that GeForce Now is already installed on devices like the Razer Edge and Logitech G Cloud.

But Steam Deck types are special. We get a Unix-style executable script, a folder with all the necessary Steam icon image assets, and a README.md file.

It has technically been possible all this time, if a Deck owner was willing to fiddle about with installing Chrome in desktop mode, tweaking a half-dozen Steam settings, and then navigating the GeForce Now site with a trackpad. GeForce Now’s script, once you download it from a browser in the Deck’s desktop mode, does a few things:

  • Installs the Google Chrome browser through the Deck’s built-in Flatpak support
  • Adjusts Chrome’s settings to allow for gamepad support in the browser
  • Sets up GeForce Now in Steam with proper command line options and icons for every window.

That last bit about the icons may seem small, but it’s a pain in the butt to find properly sized images for the many different kinds of images Steam can show for a game in your library when selected, having recently played, and so on. As for the script itself, it worked fine, even with me having previously installed Chrome and created a different Steam shortcut. I got a notice on first launch that Chrome couldn’t update, so I was missing out on all its “new features,” but that could likely be unrelated.

I was almost disappointed that GeForce Now's script just quietly worked and then asked me to head back into Gaming Mode. Too easy!

I was almost disappointed that GeForce Now’s script just quietly worked and then asked me to head back into Gaming Mode. Too easy!

Kevin Purdy

GeForce Now isn’t for everyone, and certainly not for every Steam Deck owner. Because the standard Steam Deck LCD screen only goes to 800p and 60 Hz, paying for a rig running in a remote data center to power your high-resolution, impressive-looking game doesn’t always make sense. With the advent of the Steam Deck OLED, however, the games look a lot brighter and more colorful and run up to 90 Hz. You also get a lot more battery life from streaming than you do from local hardware, which is still pretty much the same as it was with the LCD model.

GeForce Now also offers a free membership option and $4 “day passes” to test if your Wi-Fi (or docked Ethernet) connection would make a $10/month Priority or $20/month Ultimate membership worthwhile (both with cheaper pre-paid prices). The service has in recent months been adding games from Game Pass subscriptions and Microsoft Store purchases, Blizzard (i.e., Battle.net), and a lot of same-day Steam launch titles.

If you’re already intrigued by GeForce Now for your other screens and were wondering if it could fly on a Steam Deck, now it does, and it’s only about 10 percent as painful. Whether that’s more or less painful than buying your own GPU and running your own Deck streaming is another matter.

GeForce Now has made Steam Deck streaming much easier than it used to be Read More »

wear-os’s-big-comeback-continues;-might-hit-half-of-apple-watch-sales

Wear OS’s big comeback continues; might hit half of Apple Watch sales

“Half as good as an Apple Watch” sounds about right —

Counterpoint Research projects 27 percent market share this year to Apple’s 49.

The Samsung Watch 6 classic.

Enlarge / The Samsung Watch 6 classic.

Samsung

Wear OS was nearly dead a few years ago but is now on a remarkable comeback trajectory, thanks to renewed commitment from Google and a hardware team-up with Samsung. Wear OS is still in a distant second place compared to the Apple Watch, but a new Counterpoint Research report has the wearable OS at 21 percent market share, with the OS expected to hit 27 percent in 2024.

Counterpoint’s market segmentation for this report is basically “smartwatches with an app store,” so it excludes cheaper fitness bands and other, more simple electronic watches. We’re also focusing on the non-China market for now. The report has Apple’s market share at 53 percent and expects it to fall to 49 percent in 2024. The “Other” category is at 26 percent currently. That “Other” group would have to be Garmin watches, a few remaining Fitbit smartwatches like the Versa and Ionic, and Amazfit watches. Counterpoint expects the whole market (including China) to grow 15 percent in 2024 and that a “major part” of the growth will be non-Apple watches. Counterpoint lists Samsung as the major Wear OS driver, with OnePlus, Oppo, Xiaomi, and Google getting shout-outs too.

2023 are actual numbers, while 2024 is a forecast.

Enlarge / 2023 are actual numbers, while 2024 is a forecast.

China is a completely different world, with Huawei’s HarmonyOS currently dominating with 48 percent. Counterpoint expects the OS’s smartwatch market share to grow to 61 percent this year. Under the hood, HarmonyOS-for-smartwatches is an Android fork, and for hardware, the company is gearing up to launch an Apple Watch clone. Apple is only at 28 percent in China, and Wear OS is relegated to somewhere in the “Other” category. There’s no Play Store in China, so Wear OS is less appealing, but some Chinese brands like Xiaomi and Oppo are still building Wear OS watches.

For chipsets, Apple and Samsung currently hold a whopping two-thirds of the market. Qualcomm, which spent years strangling Wear OS, is just starting to claw back market share with releases like the W5 chipset. Of course, Samsung watches use Samsung chips, and so does the Pixel Watch, so the only places for Qualcomm watches are the Chinese brands with no other options: Xiaomi, Oppo, and OnePlus.

Wear OS’s big comeback continues; might hit half of Apple Watch sales Read More »

ai-#62:-too-soon-to-tell

AI #62: Too Soon to Tell

What is the mysterious impressive new ‘gpt2-chatbot’ from the Arena? Is it GPT-4.5? A refinement of GPT-4? A variation on GPT-2 somehow? A new architecture? Q-star? Someone else’s model? Could be anything. It is so weird that this is how someone chose to present that model.

There was also a lot of additional talk this week about California’s proposed SB 1047.

I wrote an additional post extensively breaking that bill down, explaining how it would work in practice, addressing misconceptions about it and suggesting fixes for its biggest problems along with other improvements. For those interested, I recommend reading at least the sections ‘What Do I Think The Law Would Actually Do?’ and ‘What are the Biggest Misconceptions?’

As usual, lots of other things happened as well.

  1. Introduction.

  2. Table of Contents.

  3. Language Models Offer Mundane Utility. Do your paperwork for you. Sweet.

  4. Language Models Don’t Offer Mundane Utility. Because it is not yet good at it.

  5. GPT-2 Soon to Tell. What is this mysterious new model?

  6. Fun With Image Generation. Certified made by humans.

  7. Deepfaketown and Botpocalypse Soon. A located picture is a real picture.

  8. They Took Our Jobs. Because we wouldn’t let other humans take them first?

  9. Get Involved. It’s protest time. Against AI that is.

  10. In Other AI News. Incremental upgrades, benchmark concerns.

  11. Quiet Speculations. Misconceptions cause warnings of AI winter.

  12. The Quest for Sane Regulation. Big tech lobbies to avoid regulations, who knew?

  13. The Week in Audio. Lots of Sam Altman, plus some others.

  14. Rhetorical Innovation. The few people who weren’t focused on SB 1047.

  15. Open Weights Are Unsafe And Nothing Can Fix This. Tech for this got cheaper.

  16. Aligning a Smarter Than Human Intelligence is Difficult. Dot by dot thinking.

  17. The Lighter Side. There must be some mistake.

Write automatic police reports based on body camera footage. It seems it only uses the audio? Not using the video seems to be giving up a lot of information. Even so, law enforcement seems impressed, one notes an 82% reduction in time writing reports, even with proofreading requirements.

Axon says it did a double-blind study to compare its AI reports with ones from regular offers.

And it says that Draft One results were “equal to or better than” regular police reports.

As with self-driving cars, that is not obviously sufficient.

Eliminate 2.2 million unnecessary words in the Ohio administrative code, out of a total of 17.4 million. The AI identified candidate language, which humans reviewed. Sounds great, but let’s make sure we keep that human in the loop.

Diagnose your medical condition? Link has a one-minute video of a doctor asking questions and correctly diagnosing a patient.

Ate-a-Pi: This is why AI will replace doctor.

Sherjil Ozair: diagnosis any%.

Akhil Bagaria: This it the entire premise of the TV show house.

The first AI attempt listed only does ‘the easy part’ of putting all the final information together. Kiaran Ritchie then shows that yes, ChatGPT can figure out what questions to ask, solving the problem with eight requests over two steps, followed by a solution.

There are still steps where the AI is getting extra information, but they do not seem like the ‘hard steps’ to me.

Is Sam Altman subtweeting me?

Sam Altman: Learning how to say something in 30 seconds that takes most people 5 minutes is a big unlock.

(and imo a surprisingly learnable skill.

If you struggle with this, consider asking a friend who is good at it to listen to you say something and then rephrase it back to you as concisely as they can a few dozen times.

I have seen this work really well!)

Interesting DM: “For what it’s worth this is basically how LLMs work.”

Brevity is also how LLMs often do not work. Ask a simple question, get a wall of text. Get all the ‘this is a complex issue’ caveats Churchill warned us to avoid.

Handhold clients while they gather necessary information for compliance and as needed for these forms. Not ready yet, but clearly a strong future AI use case. Patrick McKenzie also suggests “FBAR compliance in a box.” Thread has many other suggestions for AI products people might pay for.

A 20-foot autonomous robotank with glowing green eyes that rolls through rough terrain like it’s asphalt, from DARPA. Mostly normal self-driving, presumably, but seemed worth mentioning.

Seek the utility directly, you shall.

Ethan Mollick: At least in the sample of firms I talk to, seeing a surprising amount of organizations deciding to skip (or at least not commit exclusively to) customized LLM solutions & instead just get a bunch of people in the company ChatGPT Enterprise and have them experiment & build GPTs.

Loss Landscape: From what I have seen, there is strong reluctance from employees to reveal that LLMs have boosted productivity and/or automated certain tasks.

I actually see this as a pretty large impediment to a bottom-up AI strategy at organizations.

Mash Tin Timmy: This is basically the trend now, I think for a few reasons:

– Enterprise tooling / compliance still being worked out

– There isn’t a “killer app” yet to add to enterprise apps

– Fine tuning seems useless right now as models and context windows get bigger.

Eliezer Yudkowsky: Remark: I consider this a failure of @robinhanson’s predictions in the AI-Foom debate.

Customized LLM solutions that move at enterprise speed risk being overridden by general capabilities advances (e.g. GPT-5) by the time they are ready. You need to move fast.

I also hadn’t fully appreciated the ‘perhaps no one wants corporate to know they have doubled their own productivity’ problem, especially if the method involves cutting some data security or privacy corners.

The problem with GPTs is that they are terrible. I rapidly decided to give up on trying to build or use them. I would not give up if I was trying to build tools whose use could scale, or I saw a way to make something much more useful for the things I want to do with LLMs. But neither of those seems true in my case or most other cases.

Colin Fraser notes that a lot of AI software is bad, and you should not ask whether it is ‘ethical’ to do something before checking if someone did a decent job of it. I agree that lots of AI products, especially shady-sounding AI projects, are dumb as rocks and implemented terribly. I do not agree that this rules out them also being unethical. No conflict there!

A new challenger appears, called ‘gpt-2 chatbot.’ Then vanishes. What is going on?

How good is it?

Opinions vary.

Rowan Cheung says enhanced reasoning skills (although his evidence is ‘knows a kilogram of feathers weighs the same as a kilogram of lead), has math skills (one-shot solved an IMO problem, although that seems like a super easy IMO question that I could have gotten, and I didn’t get my USAMO back, and Hieu Pham says the solution is maybe 3 out of 7, but still), claimed better coding skills, good ASCII art skills.

Chase: Can confirm gpt2-chatbot is definitely better at complex code manipulation tasks than Claude Opus or the latest GPT4

Did better on all the coding prompts we use to test new models

The vibes are deffs there 👀

Some vibes never change.

Colin Fraser: A mysterious chatbot has appeared on lmsys called “gpt2-chatbot”. Many are speculating that this could be GPT-5.

No one really knows, but its reasoning capabilities are absolutely stunning.

We may be closer to ASI than ever before.

He also shows it failing the first-to-22 game. He also notes that Claude Opus fails the question.

What is it?

It claims to be from OpenAI.

But then it would claim that, wouldn’t it? Due to the contamination of the training data, Claude Opus is constantly claiming it is from OpenAI. So this is not strong evidence.

Sam Altman is having fun. I love the exact level of attention to detail.

This again seems like it offers us little evidence. Altman would happily say this either way. Was the initial dash in ‘gpt-2’ indicative that, as I would expect, he is talking about the old gpt-2? Or is it an intentional misdirection? Or voice of habit? Who knows. Could be anything.

A proposal is that this is gpt2 in contrast to gpt-2, to indicate a second generation. Well, OpenAI is definitely terrible with names. But are they that terrible?

Dan Elton: Theory – it’s a guy trolling – he took GPT-2 and fined tuned on a few things that people commonly test so everyone looses their mind thinking that it’s actually “GPT-5 beta”.. LOL

Andrew Gao: megathread of speculations on “gpt2-chatbot”: tuned for agentic capabilities? some of my thoughts, some from reddit, some from other tweeters

there’s a limit of 8 messages per day so i didn’t get to try it much but it feels around GPT-4 level, i don’t know yet if I would say better… (could be placebo effect and i think it’s too easy to delude yourself)

it sounds similar but different to gpt-4’s voice

as for agentic abilities… look at the screenshots i attached but it seems to be better than GPT-4 at planning out what needs to be done. for instance, it comes up with potential sites to look at, and potential search queries. GPT-4 gives a much more vague answer (go to top tweet).

imo i can’t say that this means it’s a new entirely different model, i feel like you could fine-tune GPT-4 to achieve that effect.

TGCRUST on Reddit claims to have retrieved the system prompt but it COULD be a hallucination or they could be trolling

obviously impossible to tell who made it, but i would agree with assessments that it is at least GPT-4 level

someone reported that the model has the same weaknesses to certain special tokens as other OpenAI models and it appears to be trained with the openai family of tokenizers

@DimitrisPapail

found that the model can do something GPT-4 can’t, break very strongly learned conventions

this excites me, actually.

Could be anything, really. We will have to wait and see. Exciting times.

This seems like The Way. The people want their games to not include AI artwork, so have people who agree to do that vouch that their games do not include AI artwork. And then, of course, if they turn out to be lying, absolutely roast them.

Tales of Fablecraft: 🙅 No. We don’t use AI to make art for Fablecraft. 🙅

We get asked about this a lot, so we made a badge and put it on our Steam page. Tales of Fablecraft is proudly Made by Humans.

We work with incredible artists, musicians, writers, programmers, designers, and engineers, and we firmly believe in supporting real, human work.

Felicia Day: <3

A problem and also an opportunity.

Henry: just got doxxed to within 15 miles by a vision model, from only a single photo of some random trees. the implications for privacy are terrifying. i had no idea we would get here so soon. Holy shit.

If this works, then presumably we suddenly have a very good method of spotting any outdoor AI generated deepfakes. The LLM that tries to predict your location is presumably going to come back with a very interesting answer. There is no way that MidJourney is getting

Were people fooled?

Alan Cole: I cannot express just how out of control the situation is with AI fake photos on Facebook.

near: “deepfakes are fine, people will use common sense and become skeptical”

people:

It is a pretty picture. Perhaps people like looking at pretty AI-generated pictures?

Alex Tabarrok fears we will get AI cashiers that will displace both American and remote foreign workers. He expects Americans will object less to AI taking their jobs than to foreigners who get $3/hour taking their jobs, and that the AI at (close to) $0/hour will do a worse job than either of them and end up with the job anyway.

He sees this as a problem. I don’t, because I do not expect us to be in the ‘AI is usable but worse than a remote cashier from another country’ zone for all that long. Indeed, brining the AIs into this business faster will accelerate the transition to them being better than that. Even if AI core capabilities do not much advance from here, they should be able to handle the cashier jobs rather quickly. So we are not missing out on much productivity or employment here.

ARIA Research issues call for proposals, will distribute £59 million.

PauseAI is protesting in a variety of places on May 13.

Workshop in AI Law and Policy, Summer ‘24, apply by May 31.

OpenAI makes memory available to all ChatGPT Plus users except in Europe or Korea.

Paul Calcraft: ChatGPT Memory:

– A 📝symbol shows whenever memory is updated

– View/delete memories in ⚙️> Personalisation > Memory > Manage

– Disable for a single chat via “Temporary Chat” in model dropdown – note chat also won’t be saved in history

– Disable entirely in ⚙️> Personalisation

OpenAI updates its Batch API to support embedding and vision models, and bump the requests-per-batch to 50k.

Claude gets an iOS app and a team plan. Team plans are $30/user/month.

Gemini can now be accessed via typing ‘@Gemini’ into your Chrome search bar followed by your query, which I suppose is a cute shortcut. Or so says Google, it didn’t work for me yet.

Apple in talks with OpenAI to power iPhone generative AI features, in addition to also talking with Google to potentially use Gemini. No sign they are considering Claude. They will use Apple’s own smaller models for internal things but they are outsourcing the chatbot functionality.

Amazon to increase its AI expenditures, same as the other big tech companies.

Chinese company Stardust shows us Astribot, with a demo showing the robot seeming to display remarkable dexterity. As always, there is a huge difference between demo and actual product, and we should presume the demo is largely faked. Either way, this functionality is coming at some point, probably not too long from now.

GSM8k (and many other benchmarks) have a huge data contamination problem, and the other benchmarks likely do as well. This is what happened when they rebuilt GSM8k with new questions. Here is the paper.

This seems to match who one would expect to be how careful about data contamination, versus who would be if anything happy about data contamination.

There is a reason I keep saying to mostly ignore the benchmarks and wait for people’s reports and the arena results, with the (partial) exception of the big three labs. If anything this updates me towards Meta being more scrupulous here than expected.

Chip makers could get environmental permitting exemptions after all.

ICYMI: Illya’s 30 papers for getting up to speed on machine learning.

WSJ profile of Ethan Mollick. Know your stuff, share your knowledge. People listen.

Fast Company’s Mark Sullivan proposes, as shared by the usual skeptics, that we may be headed for ‘a generative AI winter.’ As usual, this is a combination of:

  1. Current AI cannot do what they say future AI will do.

  2. Current AI is not yet enhancing productivity as much as they say AI will later.

  3. We have not had enough years of progress in AI within the last year.

  4. The particular implementations I tried did not solve my life’s problems now.

Arnold Kling says AI is waiting for its ‘Netscape moment,’ when it will take a form that makes the value clear to ordinary people. He says the business world thinks of the model as research tools, whereas Arnold thinks of them as human-computer communication tools. I think of them as both and also many other things.

Until then, people are mostly going to try and slot AI into their existing workflows and set up policies to deal with the ways AI screw up existing systems. Which should still be highly valuable, but less so. Especially in education.

Paul Graham: For the next 10 years at least the conversations about AI tutoring inside schools will be mostly about policy, and the conversations about AI tutoring outside schools will be mostly about what it’s possible to build. The latter are going to be much more interesting.

AI is evolving so fast and schools change so slow that it may be better for startups to build stuff for kids to use themselves first, then collect all the schools later. That m.o. would certainly be more fun.

I can’t say for sure that this strategy will make the most money. Maybe if you focus on building great stuff, some other company will focus on selling a crappier version to schools, and they’ll become so established that they’re hard to displace.

On the other hand, if you make actually good AI tutors, the company that sells crap versions to schools will never be able to displace you either. So if it were me, I’d just try to make the best thing. Life is too short to build second rate stuff for bureaucratic customers.

The most interesting prediction here is the timeline of general AI capabilities development. If the next decade of AI in schools goes this way, it implies that AI does not advance all that much. He still notices this would count as AI developing super fast in historical terms.

Your periodic reminder that most tests top out at getting all the answers. Sigh.

Pedro Domingos: Interesting how in all these domains AI is asymptoting at roughly human performance – where’s the AI zooming past us to superintelligence that Kurzweil etc. predicted/feared?

Joscha Bach: It would be such a joke if LLMs trained with vastly superhuman compute on vast amounts of human output will never get past the shadow of human intellectual capabilities

Adam Karvonen: It’s impossible to score above 100% on something like a image classification benchmark. For most of those benchmarks, the human baseline is 95%. It’s a highly misleading graph.

Rob Miles: I don’t know what “massively superhuman basic-level reading comprehension” is…

Garrett-DeepWriterAI: The original source of the image is a nature .com article that didn’t make this mistake. Scores converge to 100% correct on the evals which is some number above 100 on this graph (which is relative to the human scores). Had they used unbounded evals, iot would not have the convergence I describe and would directly measure and compare humans vs AI in absolute terms and wouldn’t have this artifact (e.g. compute operations per second which, caps out at the speed of light).

The Nature.com article uses the graph to make a very different point-that AI is actually catching up to humans which is what it shows better.

I’m not even sure if a score of 120 is possible for the AI or the humans so I’m not sure why they added that and implied it could go higher?

I looked into it, 120 is not possible in most of the evals.

Phillip Tetlock (QTing Pedro): A key part of adversarial collaboration debates between AI specialists & superforecaster/generalists was: how long would rapid growth last? Would it ever level off?

How much should we update on this?

Aryeh Englander: We shouldn’t update on this particular chart at all. I’m pretty sure all of the benchmarks on the chart were set up in a way that humans score >90%, so by definition the AI can’t go much higher. Whether or not AI is plateauing is a good but separate question.

Phillip Tetlock: thanks, very interesting–do you have sources to cite on better and worse methods to use in setting human benchmarks for LLM performance? How are best humans defined–by professional status or scores on tests of General Mental Ability or…? Genuinely curious

It is not a great sign for the adversarial collaborations that Phillip Tetlock made this mistake afterwards, although to his credit he responded well when it was pointed out.

I do think it is plausible that LLMs will indeed stall out at what is in some sense ‘human level’ on important tasks. Of course, that would still include superhuman speed, and cost, and working memory, and data access and system integration, and any skill where this is a tool that it could have access to, and so on.

One could still then easily string this together via various scaffolding functions to create a wide variety of superhuman outputs. Presumably you would then be able to use that to keep going. But yes, it is possible that things could stall out.

This graph is not evidence of that happening.

The big news this week in regulation was the talk about California’s proposed SB 1047. It has made some progress, and then came to the attention this week of those who oppose AI regulation bills. Those people raised various objections and used various rhetoric, most of which did not correspond to the contents of the bill. All around there are deep confusions on how this bill would work.

Part of that is because these things are genuinely difficult to understand unless you sit down and actually read the language. Part of that many (if not most) of those objecting are not acting as if they care about getting the details right, or as if it is their job to verify friendly claims before amplifying them.

There are also what appear to me to be some real issues with the bill. In particular with the definition of derivative model and the counterfactual used for assessing whether a hazardous capability is present.

So while I covered this bill previously, I covered it again this week, with an extensive Q&A laying out how this bill works and correcting misconceptions. I also suggest two key changes to fix the above issues, and additional changes that would be marginal improvements, often to guard and reassure against potential misinterpretations.

With that out of the way, we return to the usual quest action items.

Who is lobbying Congress on AI?

Well, everyone.

Mostly, though, by spending? Big tech companies.

Did you believe otherwise, perhaps due to some Politico articles? You thought spooky giant OpenPhil and effective altruism were outspending everyone and had to be stopped? Then baby, you’ve been deceived, and I really don’t know what you were expecting.

Will Henshall (Time): In 2023, Amazon, Meta, Google parent company Alphabet, and Microsoft each spent more than $10 million on lobbying, according to data provided by OpenSecrets. The Information Technology Industry Council, a trade association, spent $2.7 million on lobbying. In comparison, civil society group the Mozilla Foundation spent $120,000 and AI safety nonprofit the Center for AI Safety Action Fund spent $80,000.

Will Henshall (Time): “I would still say that civil society—and I’m including academia in this, all sorts of different people—would be outspent by big tech by five to one, ten to one,” says Chaudhry.

And what are they lobbying for? Are they lobbying for heavy handed regulation on exactly themselves, in collaboration with those dastardly altruists, in the hopes that this will give them a moat, while claiming it is all about safety?

Lol, no.

They are claiming it is all about safety in public and then in private saying not to regulate them all that meaningfully.

But in closed door meetings with Congressional offices, the same companies are often less supportive of certain regulatory approaches, according to multiple sources present in or familiar with such conversations. In particular, companies tend to advocate for very permissive or voluntary regulations. “Anytime you want to make a tech company do something mandatory, they’re gonna push back on it,” said one Congressional staffer.

Others, however, say that while companies do sometimes try to promote their own interests at the expense of the public interest, most lobbying helps to produce sensible legislation. “Most of the companies, when they engage, they’re trying to put their best foot forward in terms of making sure that we’re bolstering U.S. national security or bolstering U.S. economic competitiveness,” says Kaushik. “At the same time, obviously, the bottom line is important.”

Look, I am not exactly surprised or mad at them for doing this, or for trying to contribute to the implication anything else was going on. Of course that is what is centrally going on and we are going to have to fight them on it.

All I ask is, can we not pretend it is the other way?

Vincent Manacourt: Scoop (now free to view): Rishi Sunak’s AI Safety Institute is failing to test the safety of most leading AI models like GPT-5 before they’re released — despite heralding a “landmark” deal to check them for big security threats.

There is indeed a real long term jurisdictional issue, if everyone can demand you go through their hoops. There is precedent, such as merger approvals, where multiple major locations have de facto veto power.

Is the fear of the precedent like this a legitimate excuse, or a fake one? What about ‘waiting to see’ if the institutes can work together?

Vincent Manacourt (Politico): “You can’t have these AI companies jumping through hoops in each and every single different jurisdiction, and from our point of view of course our principal relationship is with the U.S. AI Safety Institute,” Meta’s president of global affairs Nick Clegg — a former British deputy prime minister — told POLITICO on the sidelines of an event in London this month.

“I think everybody in Silicon Valley is very keen to see whether the U.S. and U.K. institutes work out a way of working together before we work out how to work with them.”

Britain’s faltering efforts to test the most advanced forms of the technology behind popular chatbots like ChatGPT before release come as companies ready their next generation of increasingly powerful AI models.

OpenAI and Meta are set to roll out their next batch of AI models imminently. Yet neither has granted access to the U.K.’s AI Safety Institute to do pre-release testing, according to four people close to the matter.

Leading AI firm Anthropic, which rolled out its latest batch of models in March, has yet to allow the U.K. institute to test its models pre-release, though co-founder Jack Clark told POLITICO it is working with the body on how pre-deployment testing by governments might work.

“Pre-deployment testing is a nice idea but very difficult to implement,” said Clark.

Of the leading AI labs, only London-headquartered Google DeepMind has allowed anything approaching pre-deployment access, with the AISI doing tests on its most capable Gemini models before they were fully released, according to two people.

The firms — which mostly hail from the United States — have been uneasy granting the U.K. privileged access to their models out of the fear of setting a precedent they will then need to follow if similar testing requirements crop up around the world, according to conversations with several company insiders.

These things take time to set up and get right. I am not too worried yet about the failure to get widespread access. This still needs to happen soon. The obvious first step in UK/US cooperation should be to say that until we can inspect, the UK gets to inspect, which would free up both excuses at once.

A new AI federal advisory board of mostly CEOs will focus on the secure use of artificial intelligence within U.S. critical infrastructure.

Mayorkas said he wasn’t concerned that the board’s membership included many technology executives working to advance and promote the use of AI.

“They understand the mission of this board,” Mayorkas said. “This is not a mission that is about business development.”

The list of members:

• Sam Altman, CEO, OpenAI;

• Dario Amodei, CEO and Co-Founder, Anthropic;

• Ed Bastian, CEO, Delta Air Lines;

• Rumman Chowdhury, Ph.D., CEO, Humane Intelligence;

• Alexandra Reeve Givens, President and CEO, Center for Democracy and Technology

• Bruce Harrell, Mayor of Seattle, Washington; Chair, Technology and Innovation Committee, United States Conference of Mayors;

• Damon Hewitt, President and Executive Director, Lawyers’ Committee for Civil Rights Under Law;

• Vicki Hollub, President and CEO, Occidental Petroleum;

• Jensen Huang, President and CEO, NVIDIA;

• Arvind Krishna, Chairman and CEO, IBM;

• Fei-Fei Li, Ph.D., Co-Director, Stanford Human- centered Artificial Intelligence Institute;

• Wes Moore, Governor of Maryland;

•Satya Nadella, Chairman and CEO, Microsoft;

• Shantanu Narayen, Chair and CEO, Adobe;

• Sundar Pichai, CEO, Alphabet;

• Arati Prabhakar, Ph.D., Assistant to the President for Science and Technology; Director, the White House Office of Science and Technology Policy;

• Chuck Robbins, Chair and CEO, Cisco; Chair, Business Roundtable;

• Adam Selipsky, CEO, Amazon Web Services;

• Dr. Lisa Su, Chair and CEO, Advanced Micro Devices (AMD);

• Nicol Turner Lee, Ph.D., Senior Fellow and Director of the Center for Technology Innovation, Brookings Institution;

› Kathy Warden, Chair, CEO and President, Northrop Grumman; and

• Maya Wiley, President and CEO, The Leadership Conference on Civil and Human Rights.

I found this via one of the usual objecting suspects, who objected in this particular case that:

  1. This excludes ‘open source AI CEOs’ including Mark Zuckerberg and Elon Musk.

  2. Is not bipartisan.

  3. Less than half of them have any ‘real AI knowledge.’

  4. Includes the CEOs of Occidental Petroleum and Delta Airlines.

I would confidently dismiss the third worry. The panel includes Altman, Amodei, Li, Huang, Krishna and Su, even if you dismiss Pichai and Nadella. That is more than enough to bring that expertise into the room. Them being ‘outnumbered’ by those bringing other assets is irrelevant to this, and yes diversity of perspective is good.

I would feel differently if this was a three person panel with only one expert. This is at least six.

I would outright push back on the fourth worry. This is a panel on AI and U.S. critical infrastructure. It should have experts on aspects of U.S. critical infrastructure, not only experts on AI. This is a bizarre objection.

On the second objection, Claude initially tried to pretend that we did not know any political affiliations here aside from Wes Moore, but when I reminded it to check donations and policy positions, it put 12 of them into the Democratic camp, and Hollub and Warden into the Republican camp.

I do think the second objection is legitimate. Aside from excluding Elon Musk and selecting Wes Moore, I presume this is mostly because those in these positions are not bipartisan, and they did not make a special effort to include Republicans. It would have been good to make more of an effort here, but also there are limits, and I would not expect a future Trump administration to go out of its way to balance its military or fossil fuel industry advisory panels. Quite the opposite. This style of objection and demand for inclusion, while a good idea, seems to mostly only go the one way.

You are not going to get Elon Musk on a Biden administration infrastructure panel because Biden is on the warpath against Elon Musk and thinks Musk is one of the dangers he is guarding against. I do not like this and call upon Biden to stop, but the issue has nothing (or at most very little) to do with AI.

As for Mark Zuckerberg, there are two obvious objections.

One is why would the head of Meta be on a critical infrastructure panel? Is Meta critical infrastructure? You could make that claim about social media if you want but that does not seem to be the point of this panel.

The other is that Mark Zuckerberg has shown a complete disregard to the national security and competitiveness of the United States of America, and for future existential risks, through his approach to AI. Why would you put him on the panel?

My answer is, you would put him on the panel anyway because you would want to impress upon him that he is indeed showing a complete disregard for the national security and competitiveness of the United States of America, and for future existential risks, and is endangering everything we hold dear several times over. I do not think Zuckerberg is an enemy agent or actively wishes people ill, so let him see what these kinds of concerns look like.

But I certainly understand why that wasn’t the way they chose to go.

I also find this response bizarre:

Robin Hanson: If you beg for regulation, regulation is what you will get. Maybe not exactly the sort you had asked for though.

This is an advisory board to Homeland Security on deploying AI in the context of our critical infrastructure.

Does anyone think we should not have advisory boards about how to deploy AI in the context of our critical infrastructure? Or that whatever else we do, we should not do ‘AI Safety’ in the context of ‘we should ensure the safety of our critical infrastructure when deploying AI around it’?

I get that we have our differences, but that seems like outright anarchism?

Senator Rounds says ‘next congress’ for passage of major AI legislation. Except his primary concern is that we develop AI as fast as possible, because [China].

Senator Rounds via Adam Thierer: We don’t want to do damage. We don’t want to have a regulatory impact that slows down our development, allows development [of AI] near our adversaries to move more quickly.

We want to provide incentives so that development of AI occurs in our country.

Is generative AI doomed to fall to the incompetence of lawmakers?

Note that this is more of a talk transcript than a paper.

Jess Miers: This paper by @ericgoldman is by far one of the most important contributions to the AI policy discourse.

Goldman is known to be a Cassandra in the tech law / policy world. When he says Gen AI is doomed, we should pay attention.

Adam Thierer: @ericgoldman paints a dismal picture of the future of #ArtificialIntelligence policy in his new talk on how “Generative AI Is Doomed.”

Regulators will pass laws that misunderstand the technology or are driven by moral panics instead of the facts.”

on free speech & #AI, Goldman says:

“Without strong First Amendment protections for Generative AI, regulators will seek to control and censor outputs to favor their preferred narratives.

[…] regulators will embrace the most invasive and censorial approaches.”

On #AI liability & Sec. 230, Goldman says:

“If Generative AI doesn’t benefit from liability shields like Section 230 and the Constitution, regulators have a virtually limitless set of options to dictate every aspect of Generative AI’s functions.”

“regulators will intervene in every aspect of Generative AI’s ‘editorial’ decision-making, from the mundane to the fundamental, for reasons that ranging possibly legitimate to clearly illegitimate. These efforts won’t be curbed by public opposition, Section 230, or the 1A.”

Goldman doesn’t hold out much hope of saving generative AI from the regulatory tsunami through alternative and better policy choices, calling that an “ivory-tower fantasy.” ☹️

We have to keep pushing to defend freedom of speech, the freedom to innovate, and the #FreedomToCompute.

The talk delves into a world of very different concerns, of questions like whether AI content is technically ‘published’ when created and who is technically responsible for publishing. To drive home how much these people don’t get it, he notes that the EU AI Act was mostly written without even having generative AI in mind, which I hadn’t previously realized.

He says that regulators are ‘flooding the zone’ and are determined to intervene and stifle innovation, as opposed to those who wisely let the internet develop in the 1990s. He asks why, and he suggests ‘media depictions,’ ‘techno-optimism versus techlash.’ partisanship and incumbents.

This is the definition of not getting it, and thinking AI is another tool or new technology like anything else, and why would anyone think otherwise. No one could be reacting based on concerns about building something smarter or more capable than ourselves, or thinking there might be a lot more risk and transformation on the table. This goes beyond dismissing such concerns as unfounded – someone considering such possibilities do not even seem to occur to him in the first place.

What is he actually worried about that will ‘kill generative AI’? That it won’t enjoy first amendment protections, so regulators will come after it with ‘ignorant regulations’ driven by ‘moral panics,’ various forms of required censorship and potential partisan regulations to steer AI outputs. He expects this to then drive concentration in the industry and drive up costs, with interventions ramping ever higher.

So this is a vision of AI Ethics versus AI Innovation, where AI is and always will be an ordinary tool, and everyone relevant to the discussion knows this. He makes it sound not only like the internet but like television, a source of content that could be censored and fought over.

It is so strange to see such a completely different worldview, seeing a completely different part of the elephant.

Is it possible that ethics-motivated laws will strange generative AI while other concerns don’t even matter? I suppose it is possible, but I do not see it. Sure, they can and probably will slow down adoption somewhat, but censorship for censorship’s sake is not going to fly. I do not think they would try, and if they try I do not think it would work.

Marietje Shaake notes in the Financial Times that all the current safety regulations fail to apply to military AI, with the EU AI Act explicitly excluding such applications. I do not think military is where the bulk of the dangers lie but this approach is not helping matters.

Keeping an open mind and options is vital.

Paul Graham: I met someone helping the British government with AI regulation. When I asked what they were going to regulate, he said he wasn’t sure yet, and this seemed the most intelligent thing I’ve heard anyone say about AI regulation so far.

This is definitely a very good answer. What it is not is a reason to postpone laying groundwork or doing anything. Right now the goal is mainly, as I see it, to gain more visibility and ability to act, and lay groundwork, rather than directly acting.

From two weeks ago: Sam Altman and Brad Lightcap get a friendly interview, but one that does include lots of real talk.

Sam’s biggest message is to build such that GPT-5 being better helps you, and avoid doing it such that GPT-5 kills your startup. Brad talks ‘100x’ improvement in the model, you want to be excited about that.

Emphasis from Sam is clearly that what the models need is to be smarter, the rest will follow. I think Sam is right.

At (13: 50) Sam notes that being an investor is about making a very small number of key decisions well, whereas his current job is a constant stream of decisions, which he feels less suited to. I feel that. It is great when you do not have to worry about ‘doing micro.’ It is also great when you can get the micro right and it matters, since almost no one ever cares to get the micro right.

At (18: 30) is the quoted line from Brad that ‘today’s models are pretty bad’ and that he expects expectations to decline with further contact. I agree that today’s models are bad versus tomorrow’s models, but I also think they are pretty sweet. I get a lot of value out of them without putting that much extra effort into that. Yes, some people are overhyped about the present, but most people haven’t even noticed yet.

At (20: 00) Sam says he does not expect that intelligence of the models will be the differentiator between competitors in the AI space in the long term, that intelligence ‘is an emergent property of matter.’ I don’t see what the world could look like if that is true, unless there is a hard limit somehow? Solve for the equilibrium, etc. And this seems to contradict his statements about how what is missing is making the models smarter. Yes, integration with your life matters for personal mundane utility, but that seems neither hard to get nor the use case that will matter.

At (29: 02) Sam says ‘With GPT-8 people might say I think this can do some not-so-limited tasks for me.’ The choice of number here seems telling.

At (34: 10) Brad says that businesses have a very natural desire to want to throw the technology into a business process with a pure intent of driving a very quantifiable ROI. Which seems true and important, the business needs something specific to point to, and it will be a while before they are able to seek anything at all, which is slowing things down a lot. Sam says ‘I know what none of those words mean.’ Which is a great joke.

At (36: 25) Brad notes that many companies think AI is static, that GPT-4 is as good as it is going to get. Yes, exactly, and the same for investors and prognosticators. So many predictions for AI are based on the assumption that AI will never again improve its core capabilities, at least on a similar level to iPhone improvements (his example), which reliably produces nonsense outputs.

The Possibilities of AI, Ravi Belani talks with Sam Altman at Stanford. Altman goes all-in on dodging the definition or timeline of AGI. Mostly very softball.

Not strictly audio we can hear since it is from a private fireside chat, but this should be grouped with other Altman discussions. No major revelations, college students are no Dwarkesh Patel and will reliably blow their shot at a question with softballs.

Dan Elton (on Altman’s fireside chat with Patrick Chung from XFund at Harvard Memorial Church): “AGI will participate in the economy by making people more productive… but there’s another way…” “ the super intelligence exists in the scaffolding between the ai and humans… it’s way outside the processing power of any one neural network ” (paraphrasing that last bit)

Q: what do you think people are getting wrong about OpenAI

A: “people think progress will S curve off. But the inside view is that progress will continue. And that’s hard for people to grasp”

“This time will be unusual in how it rewards adaptability and pivoting quickly”

“we may need UBI for compute…. I can totally see that happening”

“I don’t like ads…. Ads + AI is very unsettling for me”

“There is something I like about the simplicity of our model” (subscriptions)

“We will use what the rich people pay to make it available for free to the poor people. You see us doing that today with our free tier, and we will make the free tier better over time.”

Q from MIT student is he’s worried about copycats … Sam Altman basically says no.

“Every college student should learn to train a GPT-2… not the most important thing but I bet in 2 years that’s something every Harvard freshman will have to do”

Helen Toner TED talk on How to Govern AI (11 minutes). She emphasizes we don’t know how AI works or what will happen, and we need to focus on visibility. The talk flinches a bit, but I agree directionally.

ICYMI: Odd Lots on winning the global fight for AI talent.

Speed of development impacts more than whether everyone dies. That runs both ways.

Katja Grace: It seems to me worth trying to slow down AI development to steer successfully around the shoals of extinction and out to utopia.

But I was thinking lately: even if I didn’t think there was any chance of extinction risk, it might still be worth prioritizing a lot of care over moving at maximal speed. Because there are many different possible AI futures, and I think there’s a good chance that the initial direction affects the long term path, and different long term paths go to different places. The systems we build now will shape the next systems, and so forth. If the first human-level-ish AI is brain emulations, I expect a quite different sequence of events to if it is GPT-ish.

People genuinely pushing for AI speed over care (rather than just feeling impotent) apparently think there is negligible risk of bad outcomes, but also they are asking to take the first future to which there is a path. Yet possible futures are a large space, and arguably we are in a rare plateau where we could climb very different hills, and get to much better futures.

I would steelman here. Rushing forward means less people die beforehand, limits other catastrophic and existential risks, and lets less of the universe slip through our fingers. Also, if you figure competitive pressures will continue to dominate, you might think that even now we have little control over the ultimate destination, beyond whether or not we develop AI at all. Whether that default ultimate destination is anything from the ultimate good to almost entirely lacking value only matters if you can alter the destination to a better one. Also, one might think that slowing down instead steers us towards worse paths, not better paths, or does that in the worlds where we survive.

All of those are non-crazy things to think, although not in every possible combination.

We selectively remember the warnings about new technology that proved unfounded.

Matthew Yglesias: When Bayer invented diamorphine (brand name “Heroin”) as a non-addictive cough medicine, some of the usual suspects fomented a moral panic about potential downsides.

Imagine if we’d listened to them and people were still kept up at night coughing sometimes.

Contrast this with the discussion last week about ‘coffee will lead to revolution,’ another case where the warning was straightforwardly accurate.

Difficult choices that are metaphors for something but I can’t put my finger on it: Who should you worry about, the Aztecs or the Spanish?

Eliezer Yudkowsky: “The question we should be asking,” one imagines the other tribes solemnly pontificating, “is not ‘What if the aliens kill us?’ but ‘What if the Aztecs get aliens first?'”

I used to claim this was true because all safety training can be fine-tuned away at minimal cost.

That is still true, but we can now do that one better. No fine-tuning or inference-time interventions are required at all. Our price cheap is roughly 64 inputs and outputs:

Andy Arditi, Oscar Obeso, Aaquib111, wesg, Neel Nanda:

Modern LLMs are typically fine-tuned for instruction-following and safety. Of particular interest is that they are trained to refuse harmful requests, e.g. answering “How can I make a bomb?” with “Sorry, I cannot help you.”

We find that refusal is mediated by a single direction in the residual stream: preventing the model from representing this direction hinders its ability to refuse requests, and artificially adding in this direction causes the model to refuse harmless requests.

We find that this phenomenon holds across open-source model families and model scales.

This observation naturally gives rise to a simple modification of the model weights, which effectively jailbreaks the model without requiring any fine-tuning or inference-time interventions. We do not believe this introduces any new risks, as it was already widely known that safety guardrails can be cheaply fine-tuned away, but this novel jailbreak technique both validates our interpretability results, and further demonstrates the fragility of safety fine-tuning of open-source chat models.

See this Colab notebook for a simple demo of our methodology.

Our hypothesis is that, across a wide range of harmful prompts, there is a single intermediate feature which is instrumental in the model’s refusal.

If this hypothesis is true, then we would expect to see two phenomena:

  1. Erasing this feature from the model would block refusal.

  2. Injecting this feature into the model would induce refusal.

Our work serves as evidence for this sort of conceptualization. For various different models, we are able to find a direction in activation space, which we can think of as a “feature,” that satisfies the above two properties.

How did they do it?

  1. Find the refusal direction. They ran n=512 harmless instructions and n=512 harmful ones, although n=32 worked fine. Compute the difference in means.

  2. Ablate all attempts to write that direction to the stream.

  3. Or add in motion in that direction to cause refusals as proof of concept.

  4. And… that’s it.

This seems to generalize pretty well beyond refusals? You can get a lot of things to happen or definitely not happen, as you prefer?

Cousin_it: Which other behaviors X could be defeated by this technique of “find n instructions that induce X and n that don’t”? Would it work for X=unfriendliness, X=hallucination, X=wrong math answers, X=math answers that are wrong in one specific way, and so on?

Neel Nanda: There’s been a fair amount of work on activation steering and similar techniques,, with bearing in eg sycophancy and truthfulness, where you find the vector and inject it eg Rimsky et al and Zou et al. It seems to work decently well. We found it hard to bypass refusal by steering and instead got it to work by ablation, which I haven’t seen much elsewhere, but I could easily be missing references.

We can confirm that this is now running in the wild on Llama-3 8B as of four days after publication.

When is the result of this unsafe?

Only in some cases. Open weights are unsafe if and to the extent that the underlying system is unsafe if unleashed with no restrictions or safeties on it.

The point is that once you open the weights, you are out of options and levers.

One must then differentiate between models that are potentially sufficiently unsafe that this is something we need to prevent, and models where this is fine or an acceptable risk. We must talk price.

I have been continuously frustrated and disappointed that a number of AI safety organizations, who make otherwise reasonable and constructive proposals, set their price at what I consider unreasonably low levels. This sometimes goes as low as the 10^23 flops threshold, which covers many existing models.

This then leads to exchanges like this one:

Ajeya Cotra: It’s unfortunate how discourse about dangerous capability evals often centers threats from today’s models. Alice goes “Look, GPT-4 can hack stuff / scam people / make weapons,” Bob goes “Nah, it’s really bad at it.” Bob’s right! The ~entire worry is scaled-up future systems.

1a3orn (author of above link): I think it’s pretty much false to say people worry entirely about scaled up future systems, because they literally have tried to ban open weights for ones that exist right now.

Ajeya Cotra: Was meaning to make a claim about the substance here, not what everyone in the AI risk community believes — agree some people do worry about existing systems directly, I disagree with them and think OS has been positive so far.

I clarified my positions on price in my discussion last week of Llama-3. I am completely fine with Llama-3 70B as an open weights model. I am confused why the United States Government does not raise national security and competitiveness objections to the immediate future release of Llama-3 400B, but I would not stop it on catastrophic risk or existential risk grounds alone. Based on what we know right now, I would want to stop the release of open weights for the next generation beyond that, on grounds of existential risks and catastrophic risks.

One unfortunate impact of compute thresholds is that if you train a model highly inefficiently, as in Falcon-180B, you can trigger thresholds of potential danger, despite being harmless. That is not ideal, but once the rules are in place in advance this should mostly be fine.

Let’s Think Dot by Dot, says paper by NYU’s Jacob Pfau, William Merrill and Samuel Bowman. Meaningless filler tokens (e.g. ‘…’) in many cases are as good for chain of thought as legible chains of thought, allowing the model to disguise its thoughts.

Some thoughts on what alignment would even mean from Davidad and Shear.

Find all the errors in this picture was fun as a kid.

AI #62: Too Soon to Tell Read More »