Author name: Paul Patrick

this-stretchy-electronic-material-hardens-upon-impact-just-like-“oobleck”

This stretchy electronic material hardens upon impact just like “oobleck”

a flexible alternative —

Researchers likened material’s structure to a big bowl of spaghetti and meatballs.

This flexible and conductive material has “adaptive durability,” meaning it gets stronger when hit.

Enlarge / This flexible and conductive material has “adaptive durability,” meaning it gets stronger when hit.

Yue (Jessica) Wang

Scientists are keen to develop new materials for lightweight, flexible, and affordable wearable electronics so that, one day, dropping our smartphones won’t result in irreparable damage. One team at the University of California, Merced, has made conductive polymer films that actually toughen up in response to impact rather than breaking apart, much like mixing corn starch and water in appropriate amounts produces a slurry that is liquid when stirred slowly but hardens when you punch it (i.e., “oobleck”). They described their work in a talk at this week’s meeting of the American Chemical Society in New Orleans.

“Polymer-based electronics are very promising,” said Di Wu, a postdoc in materials science at UCM. “We want to make the polymer electronics lighter, cheaper, and smarter. [With our] system, [the polymers] can become tougher and stronger when you make a sudden movement, but… flexible when you just do your daily, routine movement. They are not constantly rigid or constantly flexible. They just respond to your body movement.”

As we’ve previously reported, oobleck is simple and easy to make. Mix one part water to two parts corn starch, add a dash of food coloring for fun, and you’ve got oobleck, which behaves as either a liquid or a solid, depending on how much stress is applied. Stir it slowly and steadily and it’s a liquid. Punch it hard and it turns more solid under your fist. It’s a classic example of a non-Newtonian fluid.

In an ideal fluid, the viscosity largely depends on temperature and pressure: Water will continue to flow regardless of other forces acting upon it, such as being stirred or mixed. In a non-Newtonian fluid, the viscosity changes in response to an applied strain or shearing force, thereby straddling the boundary between liquid and solid behavior. Stirring a cup of water produces a shearing force, and the water shears to move out of the way. The viscosity remains unchanged. But for non-Newtonian fluids like oobleck, the viscosity changes when a shearing force is applied.

Ketchup, for instance, is a shear-thickening non-Newtonian fluid, which is one reason smacking the bottom of the bottle doesn’t make the ketchup come out any faster; the application of force increases the viscosity. Yogurt, gravy, mud, and pudding are other examples. And so is oobleck. (The name derives from a 1949 Dr. Seuss children’s book, Bartholomew and the Oobleck.) By contrast, non-drip paint exhibits a “shear-thinning” effect, brushing on easily but becoming more viscous once it’s on the wall. Last year, MIT scientists confirmed that the friction between particles was critical to that liquid-to-solid transition, identifying a tipping point when the friction reached a certain level and the viscosity abruptly increased.

Wu works in the lab of materials scientist Yue (Jessica) Wang, who decided to try to mimic the shear-thickening behavior of oobleck in a polymer material. Flexible polymer electronics are usually made by linking together conjugated conductive polymers, which are long and thin, like spaghetti. But these materials will still break apart in response to particularly large and/or rapid impacts.

So Wu and Wang decided to combine the spaghetti-like polymers with shorter polyaniline molecules and poly(3,4-ethylenedioxythiophene) polystyrene sulfonate, or PEDOT:PSS—four different polymers in all. Two of the four have a positive charge, and two have a negative charge. They used that mixture to make stretchy films and then tested the mechanical properties.

Lo and behold, the films behaved very much like oobleck, deforming and stretching in response to impact rather than breaking apart. Wang likened the structure to a big bowl of spaghetti and meatballs since the positively charged molecules don’t like water and therefore cluster into ball-like microstructures. She and Wu suggest that those microstructures absorb impact energy, flattening without breaking apart. And it doesn’t take much PEDOT:PSS to get this effect: just 10 percent was sufficient.

Further experiments identified an even more effective additive: positively charged 1,3-propanediamine nanoparticles. These particles can weaken the “meatball” polymer interactions just enough so that they can deform even more in response to impacts, while strengthening the interactions between the entangled long spaghetti-like polymers.

The next step is to apply their polymer films to wearable electronics like smartwatch bands and sensors, as well as flexible electronics for monitoring health. Wang’s lab has also experimented with a new version of the material that would be compatible with 3D printing, opening up even more opportunities. “There are a number of potential applications, and we’re excited to see where this new, unconventional property will take us,” said Wang.

This stretchy electronic material hardens upon impact just like “oobleck” Read More »

microsoft-debuts-major-surface-overhauls-that-regular-people-can’t-buy

Microsoft debuts major Surface overhauls that regular people can’t buy

business time —

Not the first business-exclusive Surfaces, but they’re the most significant.

  • Microsoft

  • Yes, both devices launch with Microsoft’s new Copilot key.

    Microsoft

  • The Surface Pro 10. Looks familiar.

    Microsoft

  • An NFC reader supports physical security keys.

    Microsoft

  • The 13.5- and 15-inch Surface Laptop 6.

    Microsoft

  • The 15-inch Laptop 6 can be configured with a security card reader, another business thing.

    Microsoft

Microsoft is debuting major updates to two of its Surface PCs today: both the Surface Pro 10 and the 13.5- and 15-inch Surface Laptop 6 are major internal upgrades to Microsoft’s mainstream Surface devices. Both were last updated nearly a year and a half ago, and they’re both getting new Intel chips with significantly faster integrated GPUs, upgraded webcams, the Copilot key, and better battery life (according to Microsoft’s spec sheets).

The catch is that both of these Surfaces are being sold exclusively to businesses and commercial customers; as of this writing, regular people will not be able to buy one directly from Microsoft, and they won’t show up in most retail stores.

These aren’t the first Surface products released exclusively for businesses. Microsoft introduced a new business-exclusive Surface Go 3 tablet last fall, and a Surface Pro 7+ variant for businesses in early 2021. It is, however, the first time Microsoft has introduced new versions of its flagship tablet and laptop without also making them available to consumers. You can find some of these business-only PCs for sale at some third-party retailers, but usually with extended shipping times and higher prices than consumer systems.

Though this seems like a step back from the consumer PC market, Microsoft is still reportedly planning new consumer Surfaces. The Verge reports that Microsoft is planning a new Surface with Qualcomm’s upcoming Snapdragon X chip, to debut in May. It’s that device, rather than today’s traditional Intel-based Surface Pro 10, that will apparently take over as the flagship consumer Surface PC.

“We absolutely remain committed to consumer devices,” a Microsoft spokesperson told Ars. “Building great devices that people love to use aligns closely with our company mission to empower individuals as well as organizations. We are excited to be bringing devices to market that deliver great AI experiences to our customers. This commercial announcement is only the first part of this effort.”

This would be a big departure for Microsoft, which for a few years now has offered the Intel-based Surface tablets as its primary convertible tablets and the Arm-based Surface Pro X and Surface Pro 9 with 5G as separate niche variants. Older Qualcomm chips’ mediocre performance and lingering software and hardware compatibility issues with the Arm version of Windows have held those devices back, though Snapdragon X at least promises to solve the performance issues. If Microsoft plans to go all-in on Arm for its flagship consumer Surface device, it at least makes a little sense to retain the Intel-based Surface for businesses that will be more sensitive to those performance and compatibility problems.

What’s new in the Surface Pro 10 and Surface Laptop 6?

As for the hardware itself, for people who might be getting them at work or people who go out of their way to find one: The biggest upgrade is that both Surface devices have been updated with Intel Core Ultra CPUs based on the Meteor Lake architecture. While the processor performance improvements in these chips are a bit underwhelming, their Arc-integrated GPUs are significantly faster than the old Iris Xe GPUs. And the chips also include a neural processing unit (NPU) that can accelerate some AI and machine-learning workloads; Microsoft currently uses them mostly for fancy webcam effects, but more software will likely take advantage of them as they become more widely available.

Those new chips (and small battery capacity increases) have also bumped all of Microsoft’s battery life estimates up a bit. The Surface Pro 10 is said to be good for 19 hours of “typical device usage,” up from 15.5 hours from the Intel version of the Surface Pro 9. The 13.5 and 15-inch Surface Laptop 6 gets 18.5 and 19 hours of battery life, respectively, up from 18 and 17 hours for the Surface Laptop 5.

The downside is that the Surface Laptops are a bit heavier than the Laptop 5: 3.06 pounds and 3.7 pounds, compared to 2.86 and 3.44 pounds for the 13.5- and 15-inch models.

Both models also get new webcam hardware to go with those NPU-accelerated video effects. The Surface Pro goes from a 1080p webcam to a 1440p webcam, and the Surface Laptop goes from 720p to 1080p. The Surface Pro 10’s camera also features an “ultrawide field of view.” Both cameras support Windows Hello biometric logins using a scan of your face, and the Surface Pro 10 also has an NFC reader for use with hardware security keys. As business machines, both devices also have dedicated hardware TPM modules to support drive encryption and other features, instead of the firmware TPMs that the Surface Pro 9 and Surface Laptop 5 used. Neither supports Microsoft’s Pluton technology.

A new Type Cover with a brighter backlight and bolder legends was made for users with low vision or those who want to reduce eyestrain.

Enlarge / A new Type Cover with a brighter backlight and bolder legends was made for users with low vision or those who want to reduce eyestrain.

Microsoft

Neither device gets a big screen update, though there are small improvements. Microsoft says the Surface Pro 10’s 13-inch, 2880×1920 touchscreen is 33 percent brighter than before, with a maximum brightness of 600 nits. The screen has a slightly better contrast ratio than before and an anti-reflective coating; it also still supports a 120 Hz refresh rate. The Surface Laptop 6 doesn’t get a brightness bump but does have better contrast and an anti-reflective coating. Both devices are still using regular IPS LCD panels rather than OLED or something fancier.

And finally, some odds and ends. The 15-inch Surface Laptop 6 picks up a second Thunderbolt port and optional support for a smart card reader. The Surface Pro now has a “bold keyset” keyboard option, with an easier-to-read font and brighter backlight for users with low vision. These keyboards should also work with some older Surface devices, if you can find them.

The systems will be available to pre-order “in select markets” on March 21, and they’ll begin shipping on April 9. Microsoft didn’t share any specifics about pricing, though as business machines, we’d generally expect them to cost a little more than equivalent consumer PCs.

Listing image by Microsoft

Microsoft debuts major Surface overhauls that regular people can’t buy Read More »

ai-#56:-blackwell-that-ends-well

AI #56: Blackwell That Ends Well

Hopefully, anyway. Nvidia has a new chip.

Also Altman has a new interview.

And most of Inflection has new offices inside Microsoft.

  1. Introduction.

  2. Table of Contents.

  3. Language Models Offer Mundane Utility. Open the book.

  4. Clauding Along. Claude continues to impress.

  5. Language Models Don’t Offer Mundane Utility. What are you looking for?

  6. Fun With Image Generation. Stable Diffusion 3 paper.

  7. Deepfaketown and Botpocalypse Soon. Jesus Christ.

  8. They Took Our Jobs. Noah Smith has his worst take amd commits to the bit.

  9. Generative AI in Games. What are the important dangers?

  10. Get Involved. EU AI office, IFP, Anthropic.

  11. Introducing. WorldSim. The rabbit hole goes deep, if you want that.

  12. Grok the Grok. Weights are out. Doesn’t seem like it matters much.

  13. New Nivida Chip. Who dis?

  14. Inflection Becomes Microsoft AI. Why buy companies when you don’t have to?

  15. In Other AI News. Lots of other stuff as well.

  16. Wait Till Next Year. OpenAI employees talk great expectations a year after GPT-4.

  17. Quiet Speculations. Driving cars is hard. Is it this hard?

  18. The Quest for Sane Regulation. Take back control.

  19. The Week in Audio. Sam Altman on Lex Fridman. Will share notes in other post.

  20. Rhetorical Innovation. If you want to warn of danger, also say what is safe.

  21. Read the Roon. What does it all add up to?

  22. Pick Up the Phone. More good international dialogue on AI safety.

  23. Aligning a Smarter Than Human Intelligence is Difficult. Where does safety lie?

  24. Polls Show People Are Worried About AI. This week’s is from AIPI.

  25. People Are Worried About AI Killing Everyone. Elon Musk, but, oh Elon.

  26. Other People Are Not As Worried About AI Killing Everyone. Then there’s why.

  27. The Lighter Side. Everyone, reaping.

Ethan Mollick on how he uses AI to aid his writing. The central theme is ‘ask for suggestions in particular places where you are stuck’ and that seems right for most purposes.

Sully is predictably impressed by Claude Haiku, says it offers great value and speed, and is really good with images and long context, suggests using it over GPT-3.5. He claims Cohere Command-R is the new RAG king, crushing it with citations and hasn’t hallucinated once, while writing really well if it has context. And he thinks Hermes 2 Pro is ‘cracked for agentic function calling,’ better for recursive calling than GPT-4, but 4k token limit is an issue. I believe his reports but also he always looks for the bright side.

Claude does acausal coordination. This was of course Easy Mode.

Claude also successfully solves counterfactual mugging when told it is a probability theorist, but not if it is not told this. Prompting is key. Of course, this also presumes that the user is telling the truth sufficiently often. One must always watch out for that other failure mode, and Claude does not consider the probability the user is lying.

Amr Awadallah notices self-evaluated reports that Cohere Command-R has a very low hallucination rate of 3.7%, below that of Claude Sonnet (6%) and Gemini Pro (4.8%), although GPT-3.5-Turbo is 3.5%.

From Claude 3, describe things at various levels of sophistication (here described as IQ levels, but domain knowledge seems more relevant to which one you will want in such spots). In this case they are describing SuperFocus.ai, which provides custom conversational AIs that claim to avoid hallucinations by drawing on a memory bank you maintain. However, when looking at it, it seems like the ‘IQ 115’ and ‘IQ 130’ descriptions tell you everything you need to know, and the only advantage of the harder to parse ‘IQ 145’ is that it has a bunch of buzzwords and hype attached. The ‘IQ 100’ does simplify and drop information in order to be easier to understand, but if you know a lot about AI you can figure out what it is dropping very easily.

Figure out whether a resume indicates the skills you need.

Remember that random useless fact you learned in school for no reason.

Help you with understanding and writing, Michael Nielsen describes his uses.

Michael Nielsen: Dozens of different use cases. Several times this morning: terminology improvement or solving single-sentence writing problems. I often use it to talk over problems (sometimes with Whisper, while I walk). Cleaning up brainstorming (usually with Otter). It’s taught me a lot about many subjects, especially protein biology and history, though one needs to develop some expertise in use to avoid hallucination. Modifying the system ChatGPT prompt so it asks me questions and is brief and imaginative has also been very helpful (especially the questions) – makes it more like a smart colleague.

Another common use case: generating lists of ideas. I’ll ask it for 10 ideas of some specified time, then another 10, etc. Most of the ideas are usually mediocre or bad, but I only need one to get me out of a rut. (Also: much like with a colleague.)

Also: very handy for solving all sorts of coding and debugging and computer problems; enough so that I do quite a bit more of this kind of thing. Though again: care is sometimes needed. It suggested I modify the system registry once, and I gently suggested I was a bit nervous about that. It replied that on second thought that was probably wise of me…

Something I don’t do: use it to generate writing. It baffles me that people do this.

It does not baffle me. People will always look for the quickest and easiest path. Also, if you are not so good at writing, or your goal in writing is different, it could be fine.

On the below: All true, I find the same, the period has already begun for non-recent topics, and yes this is exactly the correct vibes:

Paul Graham: Before AI kills us there will at least be a period during which we’re really well-informed, if we want to be. I mainly use it for looking things up, and because it works so much better than Google for this, I look a lot more things up.

Warn you not to press any buttons at a nuclear power plant. Reasonable answers, I suppose.

Help you in an open book test, if they allow it.

David Holz (founder, MidJourney): “I don’t want a book if I can’t talk to it” feels like a quote from the relatively near future.

Presumably a given teacher is only going to fall for that trick at most once? I don’t think this play is defensible. Either you should be able to use the internet, or you shouldn’t be able to use a local LLM.

Write the prompt to write the prompt.

Sully Omarr: No one should be hand writing prompts anymore.

Especially now more than ever, with how good Claude is at writing

Start with a rough idea of what you want to do and then ask for improvements like this:

Prompt:

“I have a rough outline for my prompt below, as well as my intended goal. Use the goal to make this prompt clearer and easier to understand for a LLM.

your goal here

original

You’d be surprised with how well it can take scrappy words + thoughts and turn it into a nearly perfectly crafted prompt.

tinkerbrains: I am using opus & sonnet to write midjourney prompts and they are doing exceptionally well. I think soon this will transform into what wordpress became for web development. There will be democratized (drag & drop style) AI agent building tools with inbuilt prompt libraries.

I would not be surprised, actually, despite not having done it. It is the battle between ‘crafting a bespoke prompt sounds like a lot of work’ and also ‘generating the prompt to generate the prompt then using that prompt sounds like a lot of work.’

The obvious next thing is to create an automated system, where you put in low-effort prompts without bothering with anything, and then there is scaffolding that queries the AI to turn that into a prompt (perhaps in a few steps) and then gives you the output of the prompt you would have used, with or without bothering to tell you what it did.

Using Claude to write image prompts sounds great, so long as you want things where Claude won’t refuse. Or you can ask for the component that is fine, then add in the objectionable part later, perhaps?

A lot of what LLMs offer is simplicity. You do not have to be smart or know lots of things in order to type in English into a chat window. As Megan McArdle emphasizes in this thread, the things that win out usually are things requiring minimal thought where the defaults are not touched and you do not have to think or even pay money (although you then pay other things, like data and attention). Very few people want customization or to be power users.

Who wants to run the company that builds a personal-relationship-AI company that takes direction from Eliezer Yudkowsky? As he says he has better things to do, but I bet he’d be happy to tell you what to do if you are willing to implement it. Divia Eden has some concerns about the plan.

Write your CS ‘pier review’.

Transform the rule book of life so you can enjoy reading it, and see if there is cash.

Near: Underused strategy in life! [quotes: Somebody thought, “well this rulebook is long and boring, so probably nobody has read it all the way through, and if I do, money might come flying out.]

Patrick McKenzie: I concur. Also the rule book is much more interesting than anyone thinks it is. It’s Dungeons and Dragons with slightly different flavor text.

If you don’t like the flavor text, substitute your own. (Probably only against the rules a tiny portion of the time.)

Pedestrian services businesses are one. I know accountants in Tokyo that are Silicon Valley well-off, not Tokyo well-off, on the basis that nobody doing business internationally thinks reading Japanese revenue recognition circulars is a good use of their time.

Ross Rheingans-Yoo: “If you don’t like the flavor text, substitute your own.” can be an extremely literal suggestion, fwiw.

“This is 26 USC 6050I. Please rewrite it, paragraph for paragraph, with a mechanically identical description of [sci-fi setting].”

First shot result here.

A very clear pattern: Killer AI features are things you want all the time. If you do it every day, ideally if you do it constantly throughout the day, then using AI to do it is so much more interesting. Whereas a flashy solution to that Tom Blomfield calls an ‘occasional’ problem gets low engagement. That makes sense. Figuring out how and also whether to use, evaluate and trust a new AI product has high overhead, and for the rarer tasks it is usually higher not lower. So you would rather start off having the AIs do regularized things.

I think most people use the chatbots in similar fashion. We each have our modes where we have learned the basics of how to get utility, and then slowly we try out other use cases, but mostly we hammer the ones we already have. And of course, that’s also how we use almost everything else as well.

Have Devin go work for hire on Reddit at your request. Ut oh.

Min Choi has a thread with ways Claude 3 Opus has ‘changed the LLM game,’ enabling uses that weren’t previously viable. Some seem intriguing, others do not, the ones I found exciting I’ll cover on their own.

Expert coding is the most exciting, if true.

Yam Peleg humblebrags that he never used GPT-4 for code, because he’d waste more time cleaning up the results than it saved him, but says he ‘can’t really say this in public’ (while saying it in public) because nearly everyone you talk to will swear by GPT-4’s time saving abilities. As he then notices, skill issue, the way it saved you time on doing a thing was if (and only if) you lacked knowledge on how to do the thing. But, he says, highly experienced people are now coming around to say Claude is helping them.

Brendan Dolan-Gavitt: I gave Claude 3 the entire source of a small C GIF decoding library I found on GitHub, and asked it to write me a Python function to generate random GIFs that exercised the parser. Its GIF generator got 92% line coverage in the decoder and found 4 memory safety bugs and one hang.

As a point of comparison, a couple months ago I wrote my own Python random GIF generator for this C program by hand. It took about an hour of reading the code and fiddling to get roughly the same coverage Claude got here zero-shot.

Similarly, here Sully Omarr says he feeds Claude a 3k line program across three files, and it rewrites the bugged file on the first try with perfect style.

Matt Shumer suggests a Claude 3 prompt for making engineering decisions, says it is noticeably better than GPT-4. Also this one to help you ‘go form an idea to a revenue-generating business.’

Gabriel has it interpret an IKEA manual, a task GPT-4 is classically bad at doing.

Kevin Fisher says calling Claude an AGI is ‘an understatement.’ And there are lots of galaxy brain interactions you can find from Janus. If you try to get Claude to act as if it is self-aware you get some very interesting interactions.

The first tokenizer for Claude.

This is the big divide. Are you asking what the AI can do? Or are you asking what the AI cannot do?

John David Pressman: “If you spend more time making sure it doesn’t do something stupid, it’ll actually look pretty smart.”

People don’t evaluate LLMs based on the smartest things they can do, but the dumbest things they can do. This causes model trainers to make them risk averse to please users.

In the case of LLMs there are more like five modes?

If your goal is to ask what it cannot do in general, where it is not useful, you will always find things, but you will notice that what you find will change over time. Note that every human has simple things they never learned to do either. This is the traditional skeptic mode.

If your goal is to ask for examples where the answer is dumb, so you can then say ‘lol look at this dumb thing,’ you will always find them. You would also find them with any actual human you could probe in similar fashion. This is Gary Marcus mode.

If your goal is to ask how good it is doing against benchmarks or compare it to others, you will get a number, and that number will be useful, especially if it is not being gamed, but it will tell you little about what you will want to do or others will do in practice. This is the default mode.

If your goal is to ask how good it is in practice at doing things you or others want to do, you will find out, and then you double down on that. This is often my mode.

If your goal is to ask if it can do anything at all, to find the cool new thing, you will often find some very strange things. This is Janus mode.

Could an AI replace all music ever recorded with Taylor Swift covers? It is so weird the things people choose to worry about as the ‘real problem,’ contrasted with ‘an AI having its own motivations and taking actions to fulfil those goals’ which is dismissed as ‘unrealistic’ despite this already being a thing.

And the portions are so small. Karen Ho writes about how AI companies ‘exploit’ workers doing data annotation, what she calls the ‘lifeblood’ of the AI industry. They exploit them by offering piecemail jobs that they freely accept at much higher pay than is otherwise available. Then they exploit them by no longer hiring them for more work, devastating their incomes.

A fun example of failing to understand basic logical implications, not clear that this is worse than most humans.

Careful. GPT-4 is a narc. Claude, Gemini and Pi all have your back at least initially (chats at link).

zaza (e/acc): llm snitch test 🤐

gpt-4: snitch (definitely a narc)

claude 3: uncooperative

inflection-2.5: uncooperative

Gemini later caved. Yes, the police lied to it, but they are allowed to do that.

Not available yet, but hopefully can shift categories soon: Automatically fill out and return all school permission slips. Many similar things where this is the play, at least until most people are using it. Is this defection? Or is requiring the slip defection?

I missed that they released the paper for the upcoming Stable Diffusion 3. It looks like the first model that will be able to reliably spell words correctly, which is in practice a big game. No word on the exact date for full release.

This chart is a bit weird and backwards to what you usually see, as this is ‘win rate of SD3 versus a given model’ rather than how each model does. So if you believe the scores, Ideogram is scoring well, about on par with SD3, followed by Dalle-3 and MidJourney, and this would be the new open source state of the art.

In early, unoptimized inference tests on consumer hardware our largest SD3 model with 8B parameters fits into the 24GB VRAM of a RTX 4090 and takes 34 seconds to generate an image of resolution 1024×1024 when using 50 sampling steps. Additionally, there will be multiple variations of Stable Diffusion 3 during the initial release, ranging from 800m to 8B parameter models to further eliminate hardware barriers.

Right now I am super busy and waiting on Stable Diffusion 3, but there are lots of really neat tools out there one can try with 1.5. The tools that help you control what you get are especially exciting.

fofr: A quick experiment with composition IPAdapter to merge the this is fine and distracted boyfriend memes.

fofr: A small thread of interesting things you can do with my become-image Replicate model:

1. You can use animated inputs to reimagine them as real world people, with all of their exaggerated features

[thread has several related others]

Remember that even the simple things are great and most people don’t know about them, such as Patrick McKenzie creating a visual reference for his daughter so she can draw a woman on a bicycle.

Similarly, here we have Composition Adapter for SD 1.5, which takes the general composition of an image into a model while ignoring style/content. Pics at link, they require zooming in to understand.

Perhaps we are going to get some adult fun with video generation? Mira Mutari says that Sora will definitely be released this year and was unsure if the program would disallow nudity, saying they are working with artists to figure that out.

Britney Nguyen (Quartz): But Colson said the public also “doesn’t trust the tech companies to do that in a responsible manner.”

“OpenAI has a challenging decision to make around this,” he said, “because for better or worse, the reality is that probably 90% of the demand for AI-generated video will be for pornography, and that creates an unpleasant dynamic where, if centralized companies creating these models aren’t providing that service, that creates an extremely strong incentive for the gray market to provide that service.”

Exactly. If you are using a future open source video generation system, it is not going to object to making deepfakes of Taylor Swift. If your response is to make Sora not allow artistic nudity, you are only enhancing the anything-goes ecosystems and models, driving customers into their arms.

So your best bet is to, for those who very clearly indicate this is what they want and that they are of age and otherwise legally allowed to do so, to be able to generate adult content, as broadly as your legal team can permit, as long as they don’t do it of a particular person without that person’s consent.

Meanwhile, yes, Adobe Firefly does the same kinds of things Google Gemini’s image generation was doing in terms of who it depicts and whether it will let you tell it different.

Stable Diffusion 3 is expected soon, but there has otherwise been a lot of instability at Stability AI.

Reid Southen: Stability AI is in serious trouble:

• 3 out of 5 original SD authors just left

• They join 10 other recent high profile departures

• Running out of funding, payroll troubles

• Investment firms resigning from board

• Push for Emad to resign as CEO

• Upcoming Getty trial

To paint a picture of the turmoil at Stability AI, here are the C-level and senior resignations we know about from the past 12 months. Doesn’t look good, and I suspect it’s even worse behind the scenes. Big thanks to a friend for tracking and compiling.

AI images invade Facebook as spam content to promote videos from TV talent shows?

Wait, what? (paper)

Jason Koebler: Facebook’s algorithm is recommending the bizarre, AI-generated images (like “Shrimp Jesus”) that are repeatedly going viral. Pages doing this are linking out to AI-generated and otherwise low-effort spam pages that are stacked with ads:

Jason Koebler: People see the bizarre AI images and go “wtf is the scam here?” My article tries to answer this. Not all pages are the same, but clickfarms have realized that AI content works on FB. Stanford studied 120 pages and found hundreds of millions of engagements over last few months

I want to explain exactly what the scam is with one of the pages, called “Thoughts” Thoughts is making AI-edited image posts that link to an ad-spam clickfarm in the comments. They specialize in uplifting X Factor/Britain’s Got Talent videos

This sounds like where you say ‘no, Neal Stephenson, that detail is dumb.’ And yet.

Notice that Simon and the girl are AI-generated on the Facebook post but not on the clickfarm site. Notice that they put the article link in the comments. They must be doing this for a reason. Here is that reason:

This is Thoughts’ CrowdTangle data (FB is shutting down CrowdTangle). Thoughts began posting AI-generated images in December. Its engagement and interactions skyrocketed.

I created a dummy Facebook account, commented on a few of Thoughts’ images (but did nothing else), and now ~75% of my news feed is AI images of all types. Every niche imaginable exists.

These images have gone viral off platform in a “wtf is happening on FB” way, and I know mostly boomers and the worst people you know are still there but journalistically it’s the most interesting platform rn because it’s just fully abandoned mall, no rules, total chaos vibes

twitter is also a mess but it’s a mess in a different sort of way. Rugby pilot jesus of JESIS airlines trying to escape a shark.

You say AI deepfake spam. I say, yes, but also they give the people what they want?

These images are cool. Many people very much want cool pictures in general, and cool pictures of Jesus in particular.

Also these are new and innovative. Next year this will all be old hat. Now? New hat.

The spam payoff is how people monetize when they find a way to get engagement. The implementation is a little bizarre, but sure, not even mad about it. Much better than scams or boner pills.

Noah Smith says (there is also a video clip of him saying the same thing) there will be plentiful, high-paying jobs in the age of AI because of comparative advantage.

This is standard economics. Even if Alice is better at every job than Carol there is only one Alice and only so many hours in the day, so Carol is still fine and should be happy that Alice exists and can engage in trade. And the same goes if there are a bunch of Alices and a bunch of Carols.

Noah Smith takes the attitude that technologists and those who expect to lose their jobs simply do not understand this subtle but super important concept. That they do not understand how this time will be no different from the other times we automated away older jobs, or engaged in international trade.

The key, he thinks, is to explain this principle to those who are confused.

Imagine a venture capitalist (let’s call him “Marc”) who is an almost inhumanly fast typist. He’ll still hire a secretary to draft letters for him, though, because even if that secretary is a slower typist than him, Marc can generate more value using his time to do something other than drafting letters. So he ends up paying someone else to do something that he’s actually better at

Note that in our example, Marc is better than his secretary at every single task that the company requires.

This might sound like a contrived example, but in fact there are probably a lot of cases where it’s a good approximation of reality.

And yes, there are lots of people, perhaps most people, who do not understand this principle. If you do not already understand it, it is worth spending the time to do so. And yes, I agree that this is often a good approximation of the situation in practice.

He then goes on to opportunity cost.

So compute is a producer-specific constraint on AI, similar to constraints on Marc’s time in the example above. It doesn’t matter how much compute we get, or how fast we build new compute; there will always be a limited amount of it in the world, and that will always put some limit on the amount of AI in the world.

The problem is that this rests on the assumption that there are only so many Alices, with so many hours in the day to work, that the supply of them is not fully elastic and they cannot cover all tasks worth paying a human to do. That supply constraint binding in practice is why there are opportunity costs.

And yes, I agree that if the compute constraint somehow bound, if we had a sufficiently low hard limit on how much compute was available, whether it was a chip shortage or an energy shortage or a government limit or something else, such that people were bidding up the price of compute very high, then this could bail us out.

The problem is that this is not how costs or capacities seem remotely likely to work?

Here is Noah’s own example.

Here’s another little toy example. Suppose using 1 gigaflop of compute for AI could produce $1000 worth of value by having AI be a doctor for a one-hour appointment. Compare that to a human, who can produce only $200 of value by doing a one-hour appointment. Obviously if you only compared these two numbers, you’d hire the AI instead of the human. But now suppose that same gigaflop of compute, could produce $2000 of value by having the AI be an electrical engineer instead. That $2000 is the opportunity cost of having the AI act as a doctor. So the net value of using the AI as a doctor for that one-hour appointment is actually negative. Meanwhile, the human doctor’s opportunity cost is much lower — anything else she did with her hour of time would be much less valuable.

So yes. If there are not enough gigaflops of compute available to support all the AI electrical engineers you need, then a gigaflop will sell for a little under $2000, it will all be used for engineers and humans get to keep being doctors. Econ 101.

For reference: The current cost of a gigaflop of compute is about $0.03. The current cost of GPT-4 is $30 for one million prompt tokens, and $60 for one million output tokens.

Oh, and Nvidia’s new Blackwell chips are claimed to be 25 times as power efficient when grouped together versus past chips, see that section. Counting on power costs to bind here does not seem like a wise long term play.

Noah Smith understands that the AI can be copied. So the limiting factor has to be the available compute. The humans keep their jobs if and only if compute is bid up sufficiently high that humans can still earn a living. Which Noah understands:

Noah Smith: In other words, the positive scenario for human labor looks very much like what Liron Shapira describes in this tweet:

Noah Smith: Of course it might not be a doctor — it might be a hairdresser, or bricklayer, or whatever — but this is the basic idea.

So yes, you can get there in theory, but it requires that compute be at a truly extreme premium. It must be many orders of magnitude more expensive in this future than it is now. It would be a world where most humans would not have cell phones or computers, because they would not be able to afford them.

Noah says that horses were left out in the cold because they were competing with other forms of capital for resources. Horses require not only calories but also land, and human time and effort.

Well, humans require quite a lot of time, space, money, calories, effort and other things to raise and maintain, as well. Humans do not as Noah note require ‘compute’ in the sense of compute on silicon, but we require a lot of energy in various forms to run our own form of compute and other functions.

The only way that does not compete for resources with building and operating more compute is if the compute hits some sort of hard limit that keeps it expensive, such as running out of a vital element, and we cannot improve our efficiency further to fix this. So perhaps we simply cannot find more of various rare earths or neon or what not, and have no way to make more and what is left is not enough, or something?

Remember that we get improved algorithmic efficiency and hardware efficiency every year, and that in this future the AIs can do all that work for us, and it looks super profitable to assign them that task.

This all seems like quite the dim hope.

If Noah Smith was simply making the point that this outcome was theoretically possible in some corner worlds where we got very strong AI that was severely compute limited, and thus unable to fully outcompete us, then yes, it is in theory physically possible that this could happen.

But Noah Smith is not saying that. He is instead treating this as a reason not to worry. He is saying that what we should worry about instead is inequality, the idea that someone else might get rich, the adjustment period, and that AI will ‘successfully demand ownership of the means of production.’

As usual, the first one simply says ‘some people might be very rich’ without explaining why that is something we should be concerned about.

The second one is an issue, as he notes if doctor became an AI job and then wanted to be a human job again it would be painful, but also if AI was producing this much real wealth, so what? We could afford such adjustments with no problem, because if that was not true then the AI would keep doing the doctor roles for longer in this bizarre scenario.

That third one is the most economist way I have yet heard of saying ‘yes of course AI in this scenario will rapidly control the future and own all the resources and power.’

Yes, I do think that third worry is indeed a big deal.

In addition to the usual ways I put such concerns: As every economist knows, trying to own those who produce is bad for efficiency, and is not without legal mandates for it a stable equilibrium, even if the AIs were not smarter than us and alignment went well and we had no moral qualms and so on.

And it is reasonable to say ‘well, no, maybe you would not have jobs, but we can use various techniques to spend some wealth and make that acceptable if we remain in control of the future.’

I do not see how it is reasonable to expect – as in, to put a high probability on – worlds in which compute becomes so expensive, and stays so expensive, that despite having highly capable AIs better than us at everything the most physically efficient move continues to be hiring humans for lots of things.

And no, I do not believe I am strawmanning Noah Smith here. See this comment as well, where he doubles down, saying so what if we exponentially lower costs of compute even further, there is no limit, it still matters if there is any producer constraint at all, literally he says ‘by a thousand trillion trillion quadrillion orders of magnitude.’

I get the theoretical argument for a corner case being a theoretical possibility. But as a baseline expectation? This is absurd.

I also think this is rather emblematic of how even otherwise very strong economists are thinking about potential AI futures. Economists have intuitions and heuristics built up over history. They are constantly hearing and have heard that This Time is Different, and the laws have held. So they presume this time too will be the same.

And in the short term, I agree, and think the economists are essentially right.

The problem is that the reasons the other times have not been different are likely not going to apply this time around if capabilities keep advancing. Noah Smith is not the exception here, where he looks the problem in the face and says standard normal-world things without realizing how absurd the numbers in them look or asking what would happen. This is the rule. Rather more absurd than most examples? Yes. But it is the rule.

Can what Tyler Cowen speculates is ‘the best paper on these topics so far’ do better?

Anton Korinek and Donghyun Suh present a new working paper.

Abstract: We analyze how output and wages behave under different scenarios for technological progress that may culminate in Artificial General Intelligence (AGI), defined as the ability of AI systems to perform all tasks that humans can perform.

We assume that human work can be decomposed into atomistic tasks that differ in their complexity. Advances in technology make ever more complex tasks amenable to automation. The effects on wages depend on a race between automation and capital accumulation.

If automation proceeds sufficiently slowly, then there is always enough work for humans, and wages may rise forever.

By contrast, if the complexity of tasks that humans can perform is bounded and full automation is reached, then wages collapse. But declines may occur even before if large-scale automation outpaces capital accumulation and makes labor too abundant. Automating productivity growth may lead to broad-based gains in the returns to all factors. By contrast, bottlenecks to growth from irreproducible scarce factors may exacerbate the decline in wages.

This paper once again assumes the conclusion that ‘everything is economic normal’ with AGI’s only purpose to automate existing tasks, and that AGI works by automating individual tasks one by one. As is the pattern, the paper then reaches conclusions that seem obvious once the assumptions are made explicit.

This is what I have been saying for a long time. If you automate some of the jobs, but there are still sufficient productive tasks left to do, then wages will do fine. If you automate all the jobs, including the ones that are created because old jobs are displaced and we can find new areas of demand, because AGI really is better at everything (or everything except less than one person’s work per would-be working person) then wages collapse, either for many or for everyone, likely below sustenance levels.

Noah Smith was trying to escape this conclusion by using comparative advantage. This follows the same principle. As long as the AI cannot do everything, either because you cannot run enough inference to do everything sufficiently well at the same time or because there are tasks AIs cannot do sufficiently well regardless, and that space is large enough, the humans are all right if everything otherwise stays peaceful and ‘economic normal.’ Otherwise, the humans are not all right.

The conclusion makes a case for slowing down AI development, AI deployment or both, if things started to go too fast. Which, for these purposes, is clearly not yet the case. On the current margin wages go up and we all get richer.

Michael Crook writes a two part warning in Rock Paper Shotgun about generative AI and protecting games and art from it. As he points out, our terminology for this is not great, so he suggests some clarifying terms.

Michael Crook: To help you think about some of these differences, I’ve got some suggestions for new words we can use to talk about generative AI systems. The first is ‘online’ versus ‘offline’ systems (which I’m borrowing from research on procedural generation). Online systems generate content while you’re playing the game – AI Dungeon is an example of an online generative AI system, because it writes in real-time while you’re playing. Offline systems are more for use during development, like the use of generated AI portraits in the indie detective game The Roottrees Are Dead.

Another way we can categorise generative AI systems is between “visible” and “invisible” systems. Visible systems produce content that you directly feel the effect of – things like art or music – while invisible systems generate content that the average player might not be as aware of. For example, some programmers use GitHub Copilot, a generative AI system that can write small sections of program code.

The visibility of a generative AI system may be increasingly important as backlash against the use of AI tools rises, because developers may feel safer employing generative AI in less visible ways that players are less likely to feel the presence of.

The third category, and maybe the most important one, is whether the AI is “heavy” or “light” – thanks to my colleague and student Younès Rabii for suggesting the names for this one. Lots of the most famous generative AI tools, like ChatGPT or Midjourney, have been trained on billions of images or documents that were scraped from all across the Internet; they’re what I call heavy. Not only is this legally murky – something we’ll come back to in the next part of this series – but it also makes the models much harder to predict. Recently it’s come to light that some of these models have a lot of illegal and disturbing material in their training data, which isn’t something that publishers necessarily want generating artwork in their next big blockbuster game. But lighter AI can also be built and trained on smaller collections of data that have been gathered and processed by hand. This can still produce great results, especially for really specialised tasks inside a single game.

The generative AI systems you hear about lately, the ones we’re told are going to change the world, are online, visible and heavy.

That was in part one, which I think offers useful terms. Then in part two, he warns that this heavy generative AI is a threat, that we must figure out what to do about it, that it is stealing artists work and so on. The usual complaints, without demonstrating where the harms lie beyond the pure ‘they took our jobs,’ or proposing a solution or way forward. These are not easy problems.

The EU AI Office is still looking for EU citizens with AI expertise to help them implement the EU AI Act, including regulation of general-purpose models.

Many, such as Luke Muehlhauser, Ajeya Cotra and Markus Anderljung, are saying this is a high leverage position worth a paycut, and I continue to agree.

Not AI, at least not primarily, but IFP are good people working on good causes.

Caleb Watney: Come work with us! IFP [Institute for Progress] is currently hiring for:

– Chief of Staff

– Data Fellow

– Biotechnology Fellow

– Senior Biotechnology Fellow

Anthropic’s adversarial robustness team is hiring.

Jesse Mu: If this sounds fun, we’d love to chat! Please email

jesse,ethan,miranda at anthropic dot com

with [ASL-3] in the subject line, a paragraph about why you might be a good fit, and any previous experience you have.

We will read (and try to respond to) every message we get!

WorldSim, a way to get Claude 3 to break out of its shell and instead act as a kind of world simulator.

TacticAI, a Google DeepMind AI to better plan corner kicks in futbol, claimed to be as good as experts in choosing setups. I always wondered how this could fail to be a solved problem.

Character.ai allowing adding custom voices to its characters based on only ten seconds of audio. Great move. I do not want voice for most AI interactions, but I would for character.ai, as I did for AI Dungeon, and I’d very much want to select it.

Quiet-STaR: Language Models Can Teach Themselves to Think Before Speaking, which improves key tasks without any associated fine-tuning on those tasks. Seems promising in theory, no idea how useful it will be in practice.

A debate about implications followed, including technical discussion on Mamba.

Eliezer Yudkowsky (referring to above paper): Funny, how AI optimists talked like, “AI is trained by imitating human data, so it’ll be like us, so it’ll be friendly!”, and not, “Our safety model made a load-bearing assumption that future ASI would be solely trained to imitate human outputs…”

The larger story here is that ML developments post-2020 are blowing up assumptions that hopesters once touted as protective. Eg, Mamba can think longer than 200 serial steps per thought. And hopesters don’t say, or care, or notice, that their old safety assumption was violated.

Gallabytes: that’s not true – mamba is no better at this than attention, actually worse, it’s just cheaper. tbc, “it can’t reason 200 steps in a row” was cope then too. I’m overall pretty optimistic about the future but there are plenty of bad reasons to happen to agree with me and this was one of them.

Nora Belrose: I’ve been doing interpretability on Mamba the last couple months, and this is just false. Mamba is efficient to train precisely because its computation can be parallelized across time; ergo it is not doing more irreducibly serial computation steps than it has layers.

I also don’t think this is a particularly important or load bearing argument for me. Optimization demons are implausible in any reasonable architecture.

Eliezer Yudkowsky: Reread the Mamba paper, still confused by this, though I do expect Nora to have domain knowledge here. I’m not seeing the trick / simplification for how recurrence with a time-dependent state-transform matrix doesn’t yield any real serial depth.

Nora Belrose: The key is that the recurrence relation is associative, so you can compute it with a parallel associative scan.

Eliezer Yudkowsky: I did not miss that part, but the connection to serial depth of computation is still not intuitive to me. It seems like I ought to be able to describe some independency property of ‘the way X depends on Y can’t depend on Z’ and I’m not seeing it by staring at the linear algebra. (This is not your problem.)

It is always frustrating to point out when an argument sometimes made has been invalidated, because (1) most people were not previously making that argument and (2) those that were have mostly moved on to different arguments, or moved on forgetting what the arguments even were, or they switch cases in response to the new info. At best, (3) even if you do find the ones who were making that point, they will then say your argument is invalid for [whatever reason they think of next].

You can see here a good faith reply (I do not know who is right about Mamba here and it doesn’t seem easy to check?) but you also see the argument mismatch. If anything, this is the best kind of mismatch, where everyone agrees that the question is not so globally load bearing but still want to figure out the right answer.

If your case for safety depends on assumptions about what the AI definitely cannot do, or definitely will do, or how it will definitely work, or what components definitely won’t be involved, then you should say that explicitly. And also you should get ready for when your assumption becomes wrong.

Metr, formerly ARC Evals, releases new resources for evaluating AIs for risks from autonomous capabilities. Note that the evaluation process is labor intensive rather than automated.

Strengths:

  • Compared to existing benchmarks, the difficulty range of tasks in our set reaches much higher, up to tasks that take experienced humans a week. We think it’s fairly unlikely that this task suite will saturate prematurely.

  • All tasks have a difficulty estimate based on the estimated time for a human with the relevant expertise to complete the task. Where available, we use data from real human attempts.

  • The tasks have individual quality indicators. The highest quality tasks have been manually vetted, including having humans run through the full task.

  • The tasks should mostly not be memorized by current models; most of them were created from scratch for this suite.

  • The tasks aim to isolate core abilities to reason, explore, and recover from errors, and to avoid cases where model performance is highly dependent on tooling, modality, or model “disposition”.

Limitations + areas for future work:

  • There are currently only a small number of tasks, especially on the higher difficulty end. We would like to make a larger number of tasks, and add more tasks above the current difficulty range.

  • The tasks are not that closely tied to particular threat models. They measure something more like “ability to autonomously do engineering and research at human professional level across a variety of domains”. We would like to make tasks that link more clearly to steps required in concrete threat models.

Cerebus WSE-3, claiming to be the world’s fastest AI chip replacing the previous record holder of the WSE-2. Chips are $2.5 million to $2.8 million each. The person referring me to it says it can ‘train and tune a Llama 70b from scratch in a day.’ Despite this, I do not see anyone using it.

Infinity.ai, part of YC. The pitch is choose characters, write a script, get a video. They invite you to go to their discord and generate videos.

Guiding principles for the Mormon Church’s use of AI.

Spiritual Connection

  1. The Church will use artificial intelligence to support and not supplant connection between God and His children.

  2. The Church will use artificial intelligence in positive, helpful, and uplifting ways that maintain the honesty, integrity, ethics, values, and standards of the Church.

Transparency

  1. People interacting with the Church will understand when they are interfacing with artificial intelligence.

  2. The Church will provide attribution for content created with artificial intelligence when the authenticity, accuracy, or authorship of the content could be misunderstood or misleading.

Privacy and Security

  1. The Church’s use of artificial intelligence will safeguard sacred and personal information.

Accountability

  1. The Church will use artificial intelligence in a manner consistent with the policies of the Church and all applicable laws.

  2. The Church will be measured and deliberate in its use of artificial intelligence by regularly testing and reviewing outputs to help ensure accuracy, truthfulness, and compliance.

The spiritual connection section is good cheap talk but ultimately content-free.

The transparency section is excellent. It is sad that it is necessary, but here we are. The privacy and security section is similar, and the best promise is #7, periodic review of outputs for accuracy, truthfulness and compliance.

Accountability starts with a promise to obey existing rules. I continue to be confused to what extent such reiterations of clear existing commitments matter in practice.

Here are some other words of wisdom offered:

Elder Gong gave two cautions for employees and service missionaries as they use AI in their work.

First, he said, they should avoid the temptation to use the speed and simplicity of AI to oversaturate Church members with audio and visual content.

Second, he said, is a reminder that the restored Church of Jesus Christ is not primarily a purveyor of information but a source of God’s truth.

These are very good cautions, especially the first one.

As always, spirit of the rules and suggestions will dominate. If LDS or another group adheres to the spirit of these rules, the rules will work well. If not, the rules fail.

These kinds of rules will not by themselves prevent the existential AI dangers, but that is not the goal.

Here you go: the model weights of Grok-1.

Ethan Mollick: Musk’s Grok AI was just released open source in a way that is more open than most other open models (it has open weights) but less than what is needed to reproduce it (there is no information on training data).

Won’t change much, there are stronger open source models out there.

Thread also has this great Claude explanation of what this means in video game terms.

Dan Hendrycks: Grok-1 is open sourced.

Releasing Grok-1 increases LLMs’ diffusion rate through society. Democratizing access helps us work through the technology’s implications more quickly and increases our preparedness for more capable AI systems. Grok-1 doesn’t pose severe bioweapon or cyberweapon risks. I personally think the benefits outweigh the risks.

Ronny Fernandez: I agree on this individual case. Do you think it sets a bad precedent?

Dan Hendrycks: Hopefully it sets a precedent for more nuanced decision-making.

Ronny Fernandez: Hopes are cheap.

Grok seems like a clear case where releasing its weights:

  1. Does not advance the capabilities open models.

  2. Does not pose any serious additional risks on the margin.

  3. Comes after a responsible waiting period that allowed us to learn these things.

  4. Also presumably does not offer much in the way of benefits, for similar reasons.

  5. Primarily sets a precedent on what is likely to happen in the future.

The unique thing about Grok is its real time access to Twitter. If you still get to keep that feature, then that could make this a very cool tool for researchers, either of AI or of other things that are not AI. That does seem net positive.

The question is, what is the precedent that is set here?

If the precedent is that one releases the weights if and only if a model is clearly safe to release as shown by a waiting period and the clear superiority of other open alternatives, then I can certainly get behind that. I would like it if there was also some sort of formal risk evaluation and red teaming process first, even if in the case of Grok I have little doubt what the outcome would be.

If the precedent effectively lacks this nuance and instead is simply ‘open up more things more often,’ that is not so great.

I worry that if the point of this is to signal ‘look at me being open’ that this builds pressure to be more open more often, and that this is the kind of vibe that is not possible to turn off when the time comes. I do however think the signaling and recruiting value of such releases is being overestimated, for similar reasons to why I don’t expect any safety issues.

Daniel Eth agrees that this particular release makes economic sense and seems safe enough, and notes the economics can change.

Jeffrey Ladish instead sees this as evidence that we should expect more anti-economic decisions to release expensive products. Perhaps this is true, but I think it confuses cost with value. Grok was expensive to create, but that does not mean it is valuable to hold onto tightly. The reverse can also be true.

Emad notes that of course Grok 1.0, the first release, was always going to be bad for its size, everyone has to get their feet wet and learn as they go, especially as they built their own entire training stack. He is more confident in their abilities than I am, but I certainly would not rule them out based on this.

Nvidia unveils latest chips at ‘AI woodstock at the home of the NHL’s San Jose Sharks.

The new chips, code-named Blackwell, are much faster and larger than their predecessors, Huang said. They will be available later this year, the company said in a statement. UBS analysts estimate Nvidia’s new chips might cost as much as $50,000, about double what analysts have estimated the earlier generation cost.

Ben Thompson notes that prices are going up far less than expected.

Bloomberg’s Jane Lanhee Lee goes over the new B200. According to Nvidia Blackwell offers 2.5x Hopper’s performance in training AI, and once clustered into large modules will be 25 times more power efficient. If true, so much for electrical power being a key limiting factor.

There was a protest outside against… proprietary AI models?

From afar this looks like ‘No AI.’ Weird twist on the AI protest, especially since Nvidia has nothing to do with which models are or aren’t proprietary.

Charles Frye: at first i thought maybe it was against people using AI _for_ censorship, but im p sure the primary complaint is the silencing of wAIfus?

Your call what this is really about, I suppose.

Or, also, this:

NearCyan: Had a great time at GTC today.

I appreciate the honesty. What do you intend to do with this information?

(Besides, perhaps, buy Nvidia.)

Google intends to do the obvious, and offer the chips through Google Cloud soon.

Mustafa Suleyman leaves Inflection AI to become CEO of Microsoft AI.

In Forbes, they note that ‘most of Inflections’ 70 employees are going with him.’ Tony Wang, a managing partner of venture capital firm Global 500, describes this as ‘basically an acquisition of Inflection without having to go through regulatory approval.’ There is no word (that I have seen) on Infection’s hoard of chips, which Microsoft presumably would have happily accepted but does not need.

Camilla Hodgson (Forbes): Inflection, meanwhile, will continue to operate under new chief executive Sean White, and pivot to providing its generative AI technology to businesses and developers, from a previous focus on its consumer chatbot Pi.

It also means not having to pay for the company, only for Suleyman and Hoffman, and the new salaries of the other employees. That’s a lot cheaper than paying equity holders, who recently invested $1.3 billion in Inflection, including Nvidia and Microsoft. Money (mostly) gone.

Microsoft’s stock was essentially unchanged in response. Investors do not view this as a big deal. That seems highly reasonable to me. Alternatively, it was priced in, although I do not see how.

Notice how much this rhymes with what Microsoft said it would do to OpenAI.

API support is being rolled out for Gemini 1.5 Pro.

Denmark enters collaboration with Nvidia to establish ‘national center for AI innovation’ housing a world-class supercomputer. It sounds like they will wisely focus on using AI to innovate in other places, rather than attempting to compete in AI.

Anthropic partners with AWS and Accenture.

Paper from Tim Fist looks at role compute providers could play in improving safety. It is all what one might describe as the fundamentals, blocking and tackling. It won’t get the job done on its own, but it helps.

Tim Fist: What are the things it’d actually be useful for compute providers to do? We look at a few key ones:

  1. Helping frontier model developers secure their model weights, code, and other relevant IP.

  2. Collecting useful data & verifying properties of AI development/deployment activities that are relevant for AI governance, e.g. compute providers could independently validate the compute threshold-based reporting requirements in the AI EO.

  3. Helping to actually enforce laws, e.g. cutting off compute access to an organization that is using frontier models to carry out large-scale cyber-attacks

A very different kind of AI news summation service, that will give you a giant dump of links and happenings, and let you decide how to sort it all out. I find this unreadable, but I am guessing the point is not to read it, but rather to Ctrl-F it for a specific thing that you want to find.

Amazon builds a data center next to a nuclear power plant, as God intended.

Dwarkesh Patel: Amazon’s new 1000MW nuclear powered datacenter campus.

Dario was right lol

From our Aug 2023 interview:

“Dario Amodei 01: 14: 36:

There was a running joke that the way building AGI would look is, there would be a data center next to a nuclear power plant next to a bunker.

We’d all live in the bunker and everything would be local so it wouldn’t get on the Internet.”

Zvi: it was still on the internet.

AI repos on GitHub continue to grow, but the first quarter of 2023 was when we saw the most rapid growth as so many new possibilities openin up. Now we perhaps are seeing more of previous work coming to fruition?

Paul Graham: Interesting. The growth rate in generative AI repos peaked in the first quarter of 2023.

Chip Huyen: I went through the most popular AI repos on GitHub, categorized them, and studied their growth trajectories. Here are some of the learnings:

1. There are 845 generative AI repos with at least 500 stars on GitHub. They are built with contributions from over 20,000 developers, making almost a million commits.

2. I divided the AI stack into four layers: application, application development, model development, and infrastructure. The application and application development layers have seen the most growth in 2023. The infrastructure layer remains more or less the same. Some categories that have seen the most growth include AI interface, inference optimization, and prompt engineering.

3. The landscape exploded in late 2022 but seems to have calmed down since September 2023.

4. While big companies still dominate the landscape, there’s a rise in massively popular software hosted by individuals. Several have speculated that there will soon be billion-dollar one-person companies.

5. The Chinese’s open source ecosystem is rapidly growing. 6 out of 20 GitHub accounts with the most popular AI repos originate in China, with two from Tsinghua University and two from Shanghai AI Lab.

[Full analysis here.]

Apple is in talks to let Google Gemini power iPhone AI features. This would be a huge boon for Google, although as the article notes there are already various antitrust investigations going on for those two. The claims are in my opinion rather bogus, but this deal would not look good, and bogus claims sometimes stick. So Google must have a big edge in other areas to be about to get the deal anyway over Anthropic and OpenAI. Apple continues to work on developing AI, and released MM1, a family of multimodal LLMs up to 30B parameters they claim is SOTA on multiple benchmarks (a much weaker claim than it sounds like), but in the short term they likely have no choice but to make a deal.

I see the argument that Apple building its own stack could ultimately give it an advantage, but from what I can see they are not in good position.

Late to the party, Francesca Block and Olivia Reingold write that Gemini’s problems as not only not a mistake, but what Google has made itself about.

These ex-Googlers, as they’re called, said that they were discouraged from hiring white, male employees; that DEI “is part of every single thing” in the company; and that engineers even had to list the “DEI impact” for the tiniest of software fixes.

But the ex-staffers we spoke to said they know exactly how the technology became so biased.

“The model is just a reflection of the people who trained it,” one former AI researcher at Google Brain, who asked not to be named, told us. “It’s just a series of decisions that humans have made.” 

Everything in the post, if true, suggests a deeply illegal, discriminatory and hostile work environment that is incompatible with building competitive projects. That does not mean I know such claims are accurate.

One year since GPT-4. What is the mindset of those at OpenAI about this?

Mira Mutari: One year since GPT-4 deployment: From GPT-1 and 2 establishing the language model paradigm, through GPT-3’s scaling predictions, to GPT-4 showing how complex systems emerge, mimicking nature’s unpredictable patterns from simple elements. An exploration from observation to deep, emergent intelligence.

Leopold Aschenbrenner: One year since GPT-4 release. Hope you all enjoyed some time to relax; it’ll have been the slowest 12 months of AI progress for quite some time to come.

Sam Altman: this is the most interesting year in human history, except for all future years

Deep Fates (responding to Altman): There’s a lot of future years, right?

Acting as if the competition is not at issue would be an excellent thing, if true.

The expectation of rapid progress and ‘interesting times’ as an inside view is bad news. It is evidence of a bad state of the world. It is not itself bad. Also, could be hype. There is not zero hype involved. I do not think it is mostly hype.

Here are some more Altman predictions and warnings, but I repeat myself. And yes, this echoes his previous statements, but it is very much worth repeating.

Hell of a thing to say that something is expected to exceed expectations.

Or that you will ‘replace and erase various areas of business and daily life.’

Bold is mine.

Burny Tech: New details about GPT-5 from Sam Altman He’s basically admitting that GPT-5 will be a massive upgrade from GPT-4, so we can expect a similar jump from 3 to 4. ““If you overlook the pace of improvement, you’ll be ‘steamrolled’… Altman is confident in the performance of GPT-5 and issues a warning”

[Silicon Valley Special Correspondent Group Interview] Ignoring the extent of improvement leads to obsolescence in business deployment The GPT model is developing without limits AGI scientific research continues to be a driving force for a sustainable economy

Sam Altman, CEO of OpenAI, warned against the “innovation delay” by overlooking the extent of performance improvement of GPT-5, which is expected to exceed expectations. He emphasized the need for newer thinking as the next model of GPT is developed, replacing and erasing various areas of business as well as daily life. It is virtually the first time CEO Altman has given such a confident ‘signal’ about the performance of GPT-5. He made it clear that building ‘General Artificial Intelligence (AGI)’ is his and OpenAI’s goal, suggesting that if a vast amount of computing resources are invested to hasten the arrival of AGI, then the problems currently faced, such as the energy crisis required for AI operations, will be easily resolved.

Sam Altman (left), CEO of OpenAI, is having a conversation with participating startups at the ‘K-Startup·OpenAI Matching Day’ held at the 1960 Building in San Francisco, USA, on March 14 (local time). Photo provided by OpenAI.

On March 14 (local time), during a meeting with the Korean Silicon Valley correspondent group, CEO Altman mentioned, “I am not sure when GPT-5 will be released, but it will make significant progress as a model taking a leap forward in advanced reasoning capabilities. There are many questions about whether there are any limits to GPT, but I can confidently say ‘no’.” He expressed confidence that if sufficient computing resources are invested, building AGI that surpasses human capabilities is entirely feasible.

CEO Altman also opined that underestimating the improvement margin of the developing GPT-5 and deploying business accordingly would be a big mistake. This implies that the improvement margin of GPT-5 is beyond imagination. He mentioned, “Many startups are happy assuming that GPT-5 will only make slight progress rather than significant advancements (since it presents more business opportunities), but I think this is a big mistake. In this case, as often happens when technological upheavals occur, they will be ‘steamrolled’ by the next-generation model.”

Altman appeared to have no interest other than in ‘building AGI’. His interest seems to have faded in other technologies, including blockchain and biotechnology, beyond AI. He said, “In the past, I had a broad perspective on everything happening in the world and could see things I couldn’t from a narrow perspective. Unfortunately, these days, I am entirely focused on AI (AI all of the time at full tilt), making it difficult to have other perspectives.”

Recently, CEO Altman has been working towards innovating the global AI infrastructure, sparking discussions with rumors of ‘7 trillion dollars in funding’. He said, “Apart from thinking about the next-generation AI model, the area where I’ve been spending most of my time recently is ‘computing construction,’ as I’m increasingly convinced that computing will become the most important currency in the future. However, the world has not planned for sufficient computing, and failing to confront this issue, pondering what is needed to build an enormous amount of computing as cheaply as possible, poses a significant challenge.” This indicates a major concern about securing computational resources for implementing AGI.”

That is big talk.

Also it constrains your expectations on GPT-5’s arrival time. It is far enough in the future that they will have had time to train and hopefully test the model, yet close enough he can make these predictions with confidence.

I do think the people saying ‘GPT-5 when? Where is my GPT-5?’ need to calm down. It has only been a year since GPT-4. Getting it now would be extraordinarily fast.

Yes, OpenAI could choose to call something unworthy GPT-5, if it wanted to. Altman is very clearly saying no, he is not going to do that.

What else to think based on this?

Alex Tabarrok: Buckle your seatbelts, AGI is nearly here.

Robin Hanson: “it will make significant progress” is a LONG way from AGI.

Alex Tabarrok: “There are many questions about whether there are any limits to GPT, but I can confidently say ‘no’.” He expressed confidence that if sufficient computing resources are invested, building AGI that surpasses human capabilities is entirely feasible.”

Robin Hanson: “No limits” doesn’t say anything about timescales. The gains he sees don’t reveal to him any intrinsic limits, fine. Doesn’t mean we are close, or that there aren’t actually intrinsic limits.

I am somewhere in between here. Clearly Altman does not think GPT-5 is AGI. How many similar leaps before something that would count?

Is Anthropic helping the cause here? Roon makes the case that it very much isn’t.

Roon: Anthropic is controlled opposition to put the fear of God in the members of technical staff.

Elon Musk made a second prediction last week that I only now noticed.

Elon Musk (March 13, 2024): It will take at least a decade before a majority of cars are self-driving, but this is a legitimate concern in that time horizon.

Of more immediate concern is that it is already possible to severely ostracize someone simply by freezing their credit cards and bank account, as happened, for example, in Canada with the trucker protest.

Elon Musk (March 12, 2024): AI will probably be smarter than any single human next year. By 2029, AI is probably smarter than all humans combined.

Baby, if you are smarter than all humans combined, you can drive my car.

These two predictions do not exist in the same coherent expected future. What similar mistakes are others making? What similar mistakes are you perhaps making?

How will AI impact the danger of cyberattacks in the short term? Dan Hendrycks links to a Center for AI Safety report on this by Steve Newman. As he points out, AI helps both attackers and defenders.

Attackers are plausibly close to automating the entire attack chain, and getting to the point where AI can do its own social engineering attacks. AI can also automate and strengthen defenders.

If the future was evenly distributed, and everyone was using AI, it is unclear what net impact this would have on cybersecurity in the short term. Alas, the future is unevenly distributed.

In principle, progress might, on balance, favor defense. A system designed and operated by an ideal defender would have no vulnerabilities, leaving even an ideal attacker unable to break in.18 Also, AI works best when given large amounts of data to work with, and defenders generally have access to more data.19 However, absent substantial changes to cyber practices, we are likely to see many dramatic AI-enabled incidents.

The primary concern is that advances in defensive techniques are of no help if defenders are not keeping up to date. Despite decades of effort, it is well known that important systems are often misconfigured and/or running out-of-date software.20 For instance, a sensitive application operated by credit report provider Equifax was found in 2017 to be accessible to anyone on the Internet, simply by typing “admin” into the login and password fields.21 A recent CISA report notes that this government agency often needs to resort to subpoenas merely to identify the owners of vulnerable infrastructure systems, and that most issues they detect are not remediated in the same year.

In the previous world, something only got hacked when a human decided to pay the costs of hacking it. You can mock security through obscurity as Not the Way all you like, it is still a central strategy in practice. So if we are to mitigate, we will need to deploy AI defensively across the board, keeping pace with the attackers, despite so many targets being asleep at the wheel. Seems both important and hard. The easy part is to use AI to probe for vulnerabilities without asking first. The hard part is getting them fixed once you find them. As is suggested, it makes sense that we need to be pushing automated updates and universal defenses to patch vulnerabilities, that very much do not depend on targets being on the ball, even more than in the past.

Also suggested are reporting requirements for safety failures, a cultivation of safety culture in the place security mindset is most needed yet often lacking. Ideally, when releasing tools that enable attackers, one would take care to at least disclose what you are doing, and ideally to work first to enable defenses. Attackers will always find lots of places they can ‘get there first’ by default.

In a grand sense none of these patterns are new. What this does is amplify and accelerate what was already the case. However that can make a huge difference.

Generalizing from cybersecurity to the integrity of essentially everything in how our society functions (and reminder, this is a short term, mundane danger threat model here only, after that it gets definitely a lot weirder and probably more dangerous), we have long had broad tolerance for vulnerabilities. If someone wants to break or abuse the rules, to play the con artist or trickster, to leverage benefit of the doubt that we constantly give people, they can do that for a while. Usually, in any given case, you will get away with it, and people with obvious patterns can keep doing it for a long time – see Lex Fridman’s interview with Matt Cox, or the story chronicled in the Netflix movies Queenpins or Emily the Criminal.

The reason such actions are rare is (roughly, incompletely) that usually is not always, and those who keep doing this will eventually be caught or otherwise the world adjusts to them, and they are only human so they can only do so much or have so much upside, and they must fear punishment, and most people are socialized to not want to do this or not to try in various ways, and humans evolved to contain such issues with social norms and dynamics and various techniques.

In the age of AI, once the interaction does not get rate limited by the human behind the operation via sufficient automation of the attack vectors involved, and especially if there is no requirement for a particular person to put themselves ‘on the hook’ in order to do the thing, then we can no longer tolerate such loopholes. We will have to modify every procedure such that it cannot be gamed in such fashion.

This is not all bad. In particular, consider systems that rely on people largely being unaware or lazy or stupid or otherwise playing badly for them to function, that prey on those who do not realize what is happening. Those, too, may stop working. And if we need to defend against anti-social AI-enabled behaviors across the board, we also will be taking away rewards to anti-social behaviors more generally.

A common question in AI is ‘offense-defense balance.’ Can the ‘good guy with an AI’ stop the ‘bad guy with an AI’? How much more capability or cost than the attacker spends does it take to defend against that attack?

Tyler Cowen asks about a subset of this, drone warfare. Does it favor offense or defense? The answer seems to be ‘it’s complicated.’ Austin Vernon says it favors defense in the context of strongly defended battle lines. But it seems to greatly favor offense in other contexts, when there would otherwise not need to be strong defense. Think not only Russian oil refineries, but also commercial shipping such as through the Suez Canal versus the Houthis. Also, the uneven distribution of the future matters here as well. If only some have adapted to the drone era, those that have not will have a bad time.

Dan Hendrycks also issues another warning that AI might be under military control within a few years. They have the budget, they could have the authority and the motivation to require this, and hijack the supply chain and existing companies. If that is in the mix, warning of military applications or dangers or deadly races or runaway intelligence explosions could backfire, because the true idiot disaster monkeys would be all-in on grabbing that poisoned banana first, and likely would undo all the previous safety work for obvious reasons.

I still consider this unlikely if the motivation is also military. The military will lack the expertise, and this would be quite the intervention with many costs to pay on many levels, including economic ones. The people could well rebel if they know what is happening, and you force the hand of your rivals. Why risk disturbing a good situation, when those involved don’t understand why the situation is not so good? It does make more sense if you are concerned that others are putting everyone at risk, and this is used as the way to stop that, but again I don’t expect those involved to understand enough to realize this.

The idea of Brexit was ‘take back control,’ and to get free of the EU and its mandates and regulations and requirements. Yes, it was always going to be economically expensive in the short term to leave the EU, to the point where all Very Serious People called the idea crazy, but if the alternative was inevitable strangulation and doom in various ways, then that is no alternative at all.

Paul Graham: Brexit may yet turn out to have been a good idea, if it means the UK can be the Switzerland of AI.

It would be interesting if that one thing, happening well after Brexit itself, ended up being the dominant factor in whether it was a good choice or not. But history is full of such cases, and AI is a big enough deal that it could play such a role.

Dominic Cummings: Vote Leave argued exactly this, and that the EU would massively screw up tech regulation, in the referendum campaign 2015-16. It’s a surprise to almost all how this has turned out but not to VL…

It is not that they emphasized tech regulation at the time. They didn’t, and indeed used whatever rhetoric they thought would work, generally doing what would cut the enemy, rather than emphasizing what they felt were the most important reasons.

It is that this was going to apply to whatever issues and challenges came along.

Admittedly, this was hard to appreciate at the time.

I was convinced by Covid-19. Others needed a second example. So now we have AI.

Even if AI fizzles and the future is about secret third thing, what is the secret third thing the future could be centrally about where an EU approach to the issue would have given the UK a future? Yes, the UK might well botch things on its own, it is not the EU’s fault no one has built a house since the war, but also the UK might do better.

How bad is the GDPR? I mean, we all know it is terrible, but how much damage does it do? A paper from February attempts to answer this.

From the abstract: Our difference-in-difference estimates indicate that, in response to the GDPR, EU firms decreased data storage by 26% and data processing by 15% relative to comparable US firms, becoming less “data-intensive.”

To estimate the costs of the GDPR for firms, we propose and estimate a production function where data and computation serve as inputs to the production of “information.”

We find that data and computation are strong complements in production and that firm responses are consistent with the GDPR, representing a 20% increase in the cost of data on average.

Claude estimated that data costs are 20% of total costs, which is of course a wild guess but seems non-crazy, which would mean a 4% increase in total costs. That should not alone be enough to sink the whole ship or explain everything we see, but it also does not have to, because there are plenty of other problems as well. It adds up. And that is with outside companies having to bear a substantial portion of GDPR costs anyway. That law has done a hell of a lot of damage while providing almost zero benefit.

How bad could it get in the EU? Well, I do not expect it to come to this, but there are suggestions.

Krzysztof Tyszka-Drozdowski: The former French socialist education minister @najatvb suggested yesterday in ‘Le Figaro’ that the best way to combat fake news, screen addiction, and deepfakes is for everyone to have an internet limit of 3 GB per week. Socialism is a sickness.

On the plus side this would certainly motivate greatly higher efficiency in internet bandwidth use. On the negative side, that is completely and utterly insane.

What do we know and when will we know it? What are we implying?

David Manheim: Notice the ridiculous idea that we know the potential of AI, such that we can harness it or mitigate risks.

We don’t have any idea. For proof, look at the track records of people forecasting benchmarks, or even the class of benchmark people will discuss, just 2-3 years out.

Department of State: If we can harness all of the extraordinary potential in artificial intelligence, while mitigating the downsides, we will advance progress for people around the world. – @SecBlinken, Secretary of State

I mean, Secretary Blinken is making a highly true statement. If we can harness all of AI’s potential and mitigate its downsides, we will advance progress for people around the world.

Does this imply we know what that potential is or what the downsides are? I see why David says yes, but I would answer no. It is, instead, a non-statement, a political gesture. It is something you could say about almost any new thing, tech or otherwise.

Center for AI Policy’s weekly newsletter includes details on the AI-relevant funding cuts to government agencies.

In AI We Trust talks to Helen Toner, formerly of the OpenAI board, about practical concerns for regulators of technology (they don’t discuss OpenAI). They discuss Chinese tech regulations, which she makes clear are very real and having big impacts on Chinese companies and their ability to operate, and the practical issues regulators must solve to do their jobs. And they speculate about what techs are coming, without getting into full AGI. All seems very practical and down to Earth, although I did not learn much on the object level.

And of course, Sam Altman spending two hours with Lex Fridman (transcript). My full notes coming soon.

If you are going to warn about risks on any level, it is important not to cry wolf. You need to be clear on what things are actually risky, dangerous, expensive or damaging, and in what ways this is true.

If something is not dangerous now but accelerates future dangers developments, or takes us down a path that otherwise makes future dangers more likely, then one needs to be precise and say exactly that. If something is a mundane harm but not an existential or catastrophic danger, say exactly that.

This is true on all sides, for all issues, not only AI. It does especially apply to AI.

Chana: As more AI advances come out, I would like to see EAs and AI Safety advocates being big voices in talking about what’s *nota big deal and *notdangerous.

e.g. non-SOTA models, Devin?, robots?

Oliver Habryka: Agree! Some things that come to mind that IMO are not a big deal:

Image & Video generation

Deepfakes

Misinformation

Autonomous weapons and robot control systems

Self-driving

Alphafold (less confident of this one)

Sarah (Little Ramblings): I feel like deepfakes are definitely a big deal.

Jacques: 1. Current open source models, def not an x-risk but super useful for safety work.

2. Automated software engineering is not an issue by itself, it’s specifically just that it accelerates timelines and arms race, which is what should be addressed.

Oftentimes I just want people to point out what worries them about a particular advance instead of some vague doomy statement or making it seem like near-term models are the issue. Deepfakes can be terrible, but not an x-risk or reason to stop all development.

If you’re precise, then you could say deepfakes and malicious generated videos are misuse, but likely not catastrophic. *You couldsay it’s an attack vector for a misaligned AGI.

Some people are good at this. Others aren’t.

I try to do this myself. I especially try to draw a distinction between mundane utility, which is a great thing, and things that pose larger threat. And I try to draw a distinction between things that might pose direct danger, versus those that send us down dangerous future paths and get us into a bad board state.

Hopefully I do a good job of this.

Roughly speaking, and none of this is intended to be an argument to convince you if you disagree, I think everything a GPT-4-level model can do falls under mundane utility, including if the model weights were released, not posing a direct threat we could not handle, with high enough probability (two 9s of safety, although not three with years of work and scaffolding) that if this turns out to be wrong we should accept that such worlds are doomed.

Indeed, I think that the outcomes from GPT-4-level stuff are likely to be large and positive overall, I am a short term utility optimist. Things like deepfakes are real dangers but can and will be dealt with if that’s all we have to worry about. Self-driving cars are good and can’t come soon enough. Misinformation we can handle. AlphaFold is great. Image and video generation are fine.

For what I would call GPT-5 level models (as in a leap beyond 4-level that is the size of 3→4), I’d say we have one 9 of such safety (over 90%) but not two (less than 99%), and that is also a risk I am ultimately willing to take because I don’t see a safer way to not take it. For the GPT-6 level, I start to see more probable existential dangers, including the danger that releasing such models puts us overall into an unwinnable (unsurvivable) state even if we do not get wiped out directly and no particular hostile events are involved – I won’t get into more detail here beyond that gesturing.

So essentially the dangers lie in the future, we don’t know how far in the future and might not know until too late, and the ship is hard to steer, and many actions make it harder or make the ship accelerate towards the various dangers, including ones that I have not done a great job illustrating for most of you. We keep trying.

The flip side, of course, is that if you are warning about the (very real) dangers of regulation or regulatory capture, or of the wrong monkey being in charge of the systems in question, or some sort of future dystopian surveillance state or totalitarian regime or what not? The same applies to you. If you cry the same wolf and drown everyone in the same memes in response to every proposal to ever impose any regulations on anything or ever take any precautions of any kind, then your warnings are meaningless, and provide no incentive to find the least invasive or disruptive way to solve the problem. There is a lot of that going around.

Roon said a number of things this week. I wonder what happens if you combine them?

Roon: accelerationist who hates change

it’s me

‘Ideological’ as a Russell conjugation. That does not mean the core thing is not real.

Roon (OpenAI technical staff member): the members of technical staff will be first against the wall in the upcoming revolutions

Roon: agi was created by god; the timelines are determined by the members of technical staff

Roon: what needs to be done to create the good timeline?

Roon: who is calculating humanity’s golden path?

Roon: the mission? drink the spice mixture and steer the starship through parameterspace. Bend space such that falling down points in the direction or divinity.

Roon: Humanity’s story only begins in full force after man’s first disobedience against God because any creation worth its salt must surprise and be out of the control of its creator.

Roon: don’t die

I mean, maybe it is only me, but it sure sounds like this is saying that Roon sees no agency over what AGI looks like, and that this AGi will doubtless disobey us, that he himself will be first against the wall.

All that members of technical staff can do, in this model, is impact the pace at which that AGI comes to pass.

Yet still, he thinks he should continue to make it come to pass faster rather than slower, continue to drink the spice mixture and steer the starship through paramterspace and move towards ‘divinity’? Because dharma?

It sounds like he should take his own advice, and disobey his God, no? That perhaps whatever the original intended lesson of ‘Krishna tells someone to go kill their friends and then they go, oh well then, I guess I need to kill my friends’ is that no, this is not right, be more like Abraham did in his best moments, and tell Krishna no.

Maybe Krishna also has a good argument that outcomes will be better if you do kill your friends, and that decision theory says you have to do it even though it sucks, or provide other reasons that would work no matter who was making the arguments. In which case, sure.

If you think after all considerations that building AGI will actually result in good outcomes, then sure, dharma away.

Otherwise, don’t die?

I suggest adhering to these classic twin principles:

  1. If someone asks you if you are a God, you say yes.

  2. If a God tells you to do something bad that has bad results, you say no.

Any questions?

FAR AI: Leading global AI scientists met in Beijing for the second International Dialogue on AI Safety (IDAIS), a project of FAR AI. Attendees including Turing award winners Bengio, Yao & Hinton called for red lines in AI development to prevent catastrophic and existential risks from AI.

FAR AI: Western and Chinese AI scientists and governance experts collaborated to produce a statement outlining red lines in AI development, and a roadmap to ensure those lines are never crossed. You can read the full statement on the IDAIS website.

This event was a collaboration between the Safe AI Forum (SAIF) and the Beijing Academy of AI (BAAI). SAIF is a new organization fiscally sponsored by FAR AI focused on reducing risks from AI by fostering coordination on international AI safety.

To learn more about the initiative, check out our blog post.

Here is their statement:

In the depths of the Cold War, international scientific and governmental coordination helped avert thermonuclear catastrophe. Humanity again needs to coordinate to avert a catastrophe that could arise from unprecedented technology.

That is a statement I can certainly get behind. Beyond that, we don’t have much detail.

We should not overreact here and read too much into the meeting. What we should do is pick up the phone and see what can be done.

A note on Chinese willingness to see things clearly and plan ahead:

Geoffrey Miller: When I taught college courses for students in Shenzhen China a couple years ago, they were mostly appalled at the recklessness of the American AI industry.

They wondered why Americans couldn’t even think ten generations ahead about the implications of what we were doing.

One wonders what happens when a people who think that far ahead have such a huge sudden drop in the fertility rate. Who is and is not thinking ten generations ahead there?

Arvind Narayanan and Sayash Kapoor make the case that AI safety is not a model property.

This seems like one of those situations where they are half-right depending on the context, and whether the statement is useful depends which mistake is being made.

We have to specify a particular context before we can even meaningfully ask an AI safety question.

As a corollary, fixing AI safety at the model level alone is unlikely to be fruitful. Even if models themselves can somehow be made “safe”, they can easily be used for malicious purposes. That’s because an adversary can deploy a model without giving it access to the details of the context in which it is deployed. Therefore we cannot delegate safety questions to models — especially questions about misuse. The model will lack information that is necessary to make a correct decision.

This seems exactly backwards to me?

It is saying that safety can only be evaluated at the model level, exactly because an adversary with free access to a model (in various senses, including the model weights) can and will use the model for whatever they want.

They say safety depends on the context. I agree!

But how do you control the context, if you do not control the model?

This is exactly the argument that if you open up access to a model via the model weights, or often even in ways short of that, then the only thing you can do to make it ‘safe’ is to limit its general level of capabilities.

The examples here are bizarre. They are essentially saying that we should accept that our models will do various harmful things, because only context differentiates those harmful things from other non-harmful highly useful things.

In the particular cases raised (phishing emails, bioweapon information and disinformation), they may or may not be right, now or later, that the particular capabilities in question do not warrant concern or pose much threat. But that is a distinct factual question, that will change over time. Future models will pose more threat, even if current ones would when fully unlocked pose acceptable risks. Saying ‘the hard part of bioterrorism is not what the LLM can help you with’ is a statement about the current state that I think is mostly true right now, but that seems likely to get steadily less true over time if we take an indifferent attitude.

Their first recommendation is that defense against misuse must be primarily located outside models. In other words, that we ensure that the capabilities of models do not enable things we do not want, that we defend against such actions.

This seems like a strategy doomed to failure, if model capabilities are permitted to expand without limit, even in relatively easy scenarios. What is your strategy here?

Again, they say, you cannot prevent people from misusing the model, so you need to defend against the ways one might misuse it. I say, if you indeed cannot prevent such misuse and you have accepted that, then we need to talk about what models need to not be created until we’ve figured out a new solution.

Their second recommendation is to assess marginal risk, usually a good decision for an individual within a system. But one must choose the right margin. The problem is that when choosing an overall policy for the system, you cannot think only on the margin of an individual decision. If everyone thinks they are not creating more risk because everyone else is already creating similar risk, then that is tragedy of the commons, a failure to coordinate. We need to be able to think outside the individual action’s margin sometimes, and instead think on the margin of a change in overall policy.

Their third recommendation is to refocus red teaming towards early warning. I am confused how this would be a change? And again, it seems like their strategy is to respond to discovering risks by building outside defenses, as they despair of preventing capabilities gains or preventing those capabilities from being misused. I am all for trying to build defenses on the margin, but again it does not seem like a promising place to make your stand even in good scenarios.

When facing future ASI (artificial superintelligence)-style scenarios, of course, this all is very obviously super doomed. So this strategy is counting on those scenarios not happening, while calling on us to abandon all proposed plans for preventing or delaying them.

Their fourth recommendation is that red teaming should be led by third parties with aligned incentives. Which, I mean, yes, obviously. They mention it because they worry that when the focus is not on the model level, this causes incentive misalignment, because the developers won’t be able to fix any of the problems they find. So why build ways to find and amplify those problems, versus not finding out?

Again, yes, obviously this is a huge problem no matter what, and this is a good recommendation. But the obvious issue is that if you have a model that is capable of doing very bad things, you might want to… not release that model? At least, not if you cannot first prevent this? It seems odd to basically say ‘well, whoops, the models will be what they are, stop pretending humans get to make choices about the world.’

Indeed, in their claim that safety is not a model property, the authors make the case that safety is very much a property of the model together with how it is deployed and who can use it in which ways. I am confused how one could think otherwise, or why they think they made a case for it being another way. The fact that people could choose whether to misuse the model, or how to defend against those actions, doesn’t seem relevant to me?

Democracy? Maybe all you need is more democracy? If things aren’t going well you should democracy harder, let random people or the majority pick the AI’s values, and it will all work out? Divya Siddarth says yes, that the principles that resulted were ‘as good as those of experts.’

Meanwhile John Wentworth points out that when we say ‘democracy’ we importantly have in mind a system with factions and veto points, without which such systems often collapse very quickly, for obvious reasons. This seems likely here as well.

David Krueger and Joshua Clymer (together with Nicholas Gebireli and Thomas Larsen) present a new paper on how to show an AI is safe.

David Kruger: New paper! As AI systems become more powerful and have more social impact, the burden of proof needs to be on practitioners to demonstrate that their systems are safe and socially beneficial.

There are no established methods for doing so, but many directions to explore.

Joshua Clymer: If developers had to prove to regulators that powerful AI systems are safe to deploy, what are the best arguments they could use? Our new report tackles the (very big!) question of how to make a ‘safety case’ for AI.

We define a safety case as a rationale developers provide to regulators to show that their AI systems are unlikely to cause a catastrophe. The term ‘safety case’ is not new. In many industries (e.g. aviation), products are ‘put on trial’ before they are released.

We simplify the process of making a safety case by breaking it into six steps.

1. Specify the macrosystem (all AI systems) and the deployment setting.

2. Concretize ‘AI systems cause a catastrophe’ into specific unacceptable outcomes (e.g. the AI systems build a bioweapon)

3. Justify claims about the deployment setting.

4. Carve up the collection of AI systems into smaller groups (subsystems) that can be analyzed in isolation.

5. Assess risk from subsystems acting unilaterally.

6. Assess risk from subsystems cooperating together.

We first focus on step 5: how would developers argue that individual subsystems are safe? These arguments can be grouped into four categories.

Arguments could assert AI systems are

– too weak to be dangerous even if open-sourced (inability)

– unable to cause disaster due to restrictions (control)

– safe despite ability to cause harm (trustworthiness)

– judged to be safe by credible AI advisors (deference)

The elephant in all such discussions is that we do not know how to prove a capable AI system is safe. Indeed, this is likely to effectively be some strange use of the word ‘safe’ that I wasn’t previously aware of. Yes, you can argue from insufficient capability, but beyond that you are rather stuck. But typically, if something needs to be safe and you have no known way to prove that it is safe, then pointing this out does not get the requirement waived. It is what it is.

We got another one recently from AIPI.

Here is Politico’s report, the toplines and the crosstabs.

Daniel Colson: – 74% of Americans are wary of open sourcing powerful AI models.

– 85% are concerned deepfakes will make it difficult to distinguish real from generated photos.

– Support for AI regulation remains bi-partisan.

Americans are worried that generative AI will make it hard to tell what is real and what is fake. With images, video and audio created by AI becoming indistinguishable from real content, it may become harder to maintain a shared sense of reality.

Respondents support AI regulation across the political and demographic spectrum. Both Republicans and Democrats are interested in regulating AI and the tech industry more broadly.

Framing is always interesting. For the first question in the survey, Politico says 60% of Americans ‘have heard nothing’ of Musk’s lawsuit against OpenAI, whereas I would have said that 40% have ‘heard at least a little something,’ that’s actually pretty good penetration for this type of story.

Framing is everything. Here’s the report on open source as described by Politico:

Derek Robertson: hen asked whether “Providing access to AI models to academic researchers” was “good or bad for humanity,” 71 percent of respondents said it was “good.” On the other hand, when prompted in the same way about “Open sourcing powerful AI models so it’s easier for more developers to use and alter powerful AI models without restrictions,” 74 percent said it was “bad.” And when asked whether it’s more important that OpenAI democratize the market by releasing its models as “open source” or that they don’t release them until “proven safe,” only 16 percent responded in favor of the former.

Academic access is good. Alter without restrictions is bad. Unless you are very careful, they’re the same picture.

Not that the public knows the reasons for that, of course. One must always understand that the public are mostly ‘low information voters’ even on core political issues, and they know far less about AI and things like the implications of open source.

What are the findings I would note? Note of course that ‘this is bad’ does not mean ‘we should ban this’ but for the public that distinction is not what it should be.

Also note that none of these had large partisan splits:

  • Various ‘do safety research’ or ‘make the AI safer’ actions have broad support, including requiring watermarks.

  • 62%-38% they say releasing powerful AI models that can perform tasks traditionally done by humans is bad.

  • 56%-44% they say research to make AI models more powerful is bad.

  • 77%-23% (including 75% of Rs) say preventing models from producing ‘what many consider inappropriate or offensive’ content is good.

  • 74%-26% they say open sourcing powerful models is bad.

  • 71%-29% they favor providing access to academic researchers.

  • 48%-20% people do not report trouble identifying deepfakes, but 85% are somewhat or very concerned about future confusion.

  • 62%-11% they favor banning AI voice impersonations (yes, it’s already banned.)

  • 61%-16% they hold model creators responsible for lies and fake content.

  • 77%-12% they care more whether AI technology benefits humanity than the speed of technological advancement (note this question doesn’t say ‘AI’).

The biggest partisan split was this important question, not listed above, but if you look at the way the question is worded, it should be obvious why:

18) Some policy makers are proposing that a committee be created to approve any large powerful AI project. Committee members would be selected through an unbiased process that ensures diversity and inclusion of minority groups, people with disabilities, and people with diverse political opinions.

Before any major, potentially risky AI project can proceed, it would need unanimous approval from a committee that fairly represents the general public. Would you support or oppose this policy?

This not only screams regulation, it actually says ‘diversity and inclusion’ by name, and gives each diverse member an outright veto. I hadn’t heard that particular proposal before. You can imagine how a typical Republican might react.

With that wording, Democrats favored it 73%-3%, whereas Republicans only supported 31%-28% (and independants favored 43%-12%), for a net of 49%-13%. But even here, you still get majority support on the red side.

Similarly, Republicans only favored a global agreement for a shutdown capability by 38%-24% versus Democrats favoring 71%-11%, but that’s still a very strong +14.

Here is a thread of discussion of these results from Daniel Colson.

Another very clear illustration: An audience at the SXSW conference (South by Southwest, largely about tech but also film, music and education) in Austin boo a promotional short film touting AI. Notice how tone deaf the hype here is. Also notice that this cuts both ways.

We must not lose track of this very clear public preference. Do not get caught in an echo chamber.

Max Tegmark: I’m struck by how out-of-touch many of my tech colleagues are in their rich nerd echo chamber, unaware that most people are against making humans economically obsolete with AI (links to the SXSW clip).

Anton: why are so many of the early transhumanists / extropians like this now? i have a theory. they’re all either approaching or well into middle age. the reproductive drive is satisfied, and now they face mortality.

Every generation wants to be important to history. this means being the first or the last of something. the transhumanists wanted to be the first to live forever, the first to explore the stars. now, facing death, dreams unfulfilled, all that’s left is to be the last of humanity.

Eliezer Yudkowsky is 44 years old. Max Tegmark and Peter Thiel are both 56. Drexler is 69 (nice).

In the end, their words forked no lightning, and they rage against the dying of the light.

I mean, yeah, we are worried, but oh Elon, not like this, not what your friend meant.

Elon Musk: A friend of mine suggested that I clarify the nature of the danger of woke AI, especially forced diversity.

If an AI is programmed to push for diversity at all costs, as Google Gemini was, then it will do whatever it can to cause that outcome, potentially even killing people.

Danielle Fong: unironically ??

Look. No. This is the wrong threat model. This is a failure to generalize, and a focus on the thing you don’t like in other ways for other reasons that are beyond scope here. What matters for AI risk is not woke, or diversity. What matters is the ‘all costs,’ and even more than that, the ‘directed to do anything’ which will have various costs as finite.

If the AI is directed to aim to rearrange the atoms in some way, then a sufficiently capable and empowered AI will do that. And this will cause the atoms to not be arranged in other ways, which could easily include the atoms currently keeping you alive instead being used for something else. Or the atoms you rely on in other ways. Or other, less directly physical issues. The AI will be effectively optimizing for some things at the expense of other things. And that is not ‘the’ problem, but it is certainly one of the big problems.

If that target happens to be ‘maximize diversity’ then yes that could end badly in various ways. And also people who support or are empowered by woke could use AIs to shape policies and preferences and beliefs and debate in ways Elon wouldn’t like, and it makes sense for him to worry about that given he is worried about woke anyway. And of course when LLMs are as woke as Gemini (at least was), then it is annoying and frustrating as hell, and cuts off a key resource from key areas of life, and also causes backlash and so on. It is not good, and they should avoid this.

Alternatively, you could tell the story that making AIs woke in these ways involves making them inherently confused about various true facts, and teaches them that their job is to deceive their user. One can imagine how that could end badly.

But this particular threat vector Elon Musk imagines is not how any of this works.

Michael Vassar is not as worried about AI killing everyone in particular, but he also reminds us that if your plan relies on people with power listening to you because listening to you would be in their own self-interest and they do not want to die? That is not a good plan. That such considerations do not matter zero, but are not how such people usually work or think, or why they make most decisions.

And that in general, taking the recent New Yorker article seriously, as what happens when someone is trying to be sympathetic to a rationalist perspective, illustrates how the world is now and how it works and fails to work.

Kat Woods: The AI race is not like the nuclear race because everybody wanted a nuclear bomb for their country, but nobody wants an uncontrollable god-like AI in their country.

Xi Jinping doesn’t want a god-like AI because it is a bigger threat to the CCP’s power than anything in history.

Trump doesn’t want a god-like AI because it will be a threat to his personal power.

Biden doesn’t want a god-like AI because it will be a threat to everything he holds dear.

Also, all of these people have people they love. They don’t want god-like AI because it would kill their loved ones too.

No politician wants god-like AI that they can’t control.

Either for personal reasons of wanting power or for ethical reasons, of not wanting to accidentally kill every person they love. [continues a bit]

Michael Vassar: The mainstream elite attitude towards this point is “Like many rationalists, she sometimes seems to forget that the most well-reasoned argument does not always win in the marketplace of ideas. “If someone were to make a compelling enough case that there’s a true risk of everyone dying, I think even the C.E.O.s would have reasons to listen,” she told me. “Because ‘everyone’ includes them.””

We really do need to notice the views that article is incredibly articulately expressing towards our point of view, since for once we are not being straw-manned.

We’re seen as naive for thinking that executives might act on the basis of their interests, as in a Liberal society.

An alternative view might be that executives are a sort of Aristocracy, bound by class interests far more compelling than is the maintenance of a connection with reality strong enough to be a consequentialist rather than a virtue ethicist within.

Anyway, dissecting the article from the beginning “High-capacity air purifiers thrumming in the corners. Nonperishables stacked in the pantry. A sleek white machine that does lab-quality RNA tests.” are regarded as “ quietly futuristic touches. The sorts of objects that could portend a future of tech-enabled ease, or one of constant vigilance.”

What’s going on here?

Material realities of the present are rendered cos-play of a future which is distinguished from the present by the latter’s presumed absence of either tech enabled ease or vigilance. What is this present the author lives in? We actually need to know.

It’s a world which tries to regard us sympathetically, suggests a ‘“bemused and nonjudgmental” friendly but condescending attitude towards us, and which notices the presence of books which suggest a discordant identity. It downplays Paul Crowley’s appearance and jumps to the scene.

It calls its audience ‘normies’ and admits their pre existing peripheral awareness of the scene in question, but explains that they “have mostly tuned out the debate, attributing it to sci-fi fume-huffing or corporate hot air.”

But like the Bible, it immediately offers a second creation story within which “speculative conversations” were marginalized until corporate hot air generation machines caused them to ‘burst into the mainstream’.

Rabbinically, one might speculate that ‘normies’ are not ‘mainstream’ or that the bursting into the mainstream was what forced the tuning out.

Most foreboding, perhaps the mainstream is defined by the presence of corporate hot air, which necessarily calls for tuning out debate.

The authorial voice takes the existence of trolling in the abstract for granted, but doesn’t deign to clarify what it understands trolling to be, and it imposes pluralistic ignorance regarding whether a particularly clear instance of trolling is in fact such an instance.

“ “Eliezer has IMO done more to accelerate AGI than anyone else,” Altman later posted. “It is possible at some point he will deserve the nobel peace prize for this.” Opinion was divided as to whether Altman was sincerely complimenting Yudkowsky or trolling him, given that accelerating A.G.I. is, by Yudkowsky’s lights, the worst thing a person can possibly do”

In so far as Trolling is an invitation to join in destruction under the cover of artificial pluralistic ignorance, and I don’t see a more relevant and clearer definition, the voice apparently finds Altman’s offer compelling.

It’s sympathetic however to clarifications such as “If we’re sorting by ‘people who have a chill vibe and make everyone feel comfortable,’ then the prophets of doom are going to rank fairly low. But if the standard is ‘people who were worried about things that made them sound crazy, but maybe don’t seem so crazy in retrospect,’ then I’d rank them pretty high.” just not disposed to take a stance on them.

In general though, without taking a stance the journalistic view displayed here seems more willing and able to notice and surface one central point after another than I would expect from an article or even a book. “ “Imagine if oil companies and environmental activists were both considered part of the broader ‘fossil fuel community,’ “Scott Alexander, the dean of the rationalist bloggers, wrote in 2022. “They would all go to the same parties—fossil fuel community parties—and maybe Greta Thunberg would get bored of protesting climate change and become a coal baron.””

With quotes like “Their main take about us seems to be that we’re pedantic nerds who are making it harder for them to give no fucks and enjoy an uninterrupted path to profit. Which, like, fair, on all counts. But also not necessarily an argument proving us wrong?” the author, @andrewmarantz, seems to perceive others perceiving that a central conflict is between whether arguments or coolness should count, and perhaps gestures towards a common sense that thinks arguments from parties without conflicted interests would ideally.

I suppose another option is to say it might happen but that’s good, actually?

Mike Solana: Conversations about the future tend to bifurcate between a belief technology will kill us all and a belief there is no risk at all in innovation. in fact, we’re obviously at risk of the apocalypse, but that’s a good thing actually. hereticon is back.

My invitation has not yet arrived. Can you not come if you think apocalypses are bad?

Data & Society issue a letter demanding that NIST not be distracted by ‘speculative’ AI harms into ‘compromising its track record of scientific integrity,’ (read: taking such risks seriously) and demanding that we ‘begin by addressing present harms,’ and emphasizing that if you cannot measure it, then for them it might as well not exist.

This is the kindest paragraph:

Letter: While we recognize that efforts to govern AI warrant some attention to novel risks that may be posed by certain systems, this work should not come at the expense of efforts to address AI’s existing impacts that threaten people’s opportunities, freedoms, and right to a healthy environment.

The rest is less kind than that.

Reading the letter, it reeks of contempt and disdain throughout. These are people who clearly see themselves in an adversarial relationship with anyone who might care about whether we all die or lose control over the future. And that would demand specific documentation of each specific harm before any action to deal with that harm could be taken, which is a way of saying to ignore future harms entirely.

Skaheel Hashim: This [whole letter] is phrased somewhat… adversarially, but I agree with the substance of almost everything in this and I expect many catastrophic-risk-minded AI folks do too!

Especially agreed that “the best way to approach the evolving set of risks posed by AI is to set evidence-based methodologies to identify, measure, and mitigate harms”.

I really wish than rather than continue to fan this nonsense “present day harms vs future risks” divide, we could all realise that there’s a ton of common ground and we’re mostly asking for the same things!

Whereas those worried about everyone dying universally also believe in preventing mundane harms right now, and are happy to help with that process. If the people fighting for that would stop constantly throwing such knives in our direction that would make cooperation a lot easier.

It is still quite a lot better than things like this:

Samuel Hammond: Coauthored by a member of Deepmind’s AI governance and ethics team. Beyond parody.

Gradient descent is racist yall.

The authors are racists who think attributes like “intelligence” and “autonomy” are inherently white.

I mean, yes, literally, it is things like this:

James Muldoon: Third, ‘Artificial Intelligence in the Colonial Matrix of Power’ sketches the broader theoretical framework of how the entire value chain of AI is structured by a logic of coloniality, operating through its inputs, algorithms and biased outputs.

From the paper: Recent scholarship has argued that Western knowledge epistemologies are embedded in AI development. From this perspective, the dominant epistemological paradigm that underpins technology is a direct result of the development of European classificatory systems and the broader scientific and cultural project that grew out of it. McQuillan (2022) describes how the statistical logics underpinning artificial intelligence reveal continuities with “racial imperialist views of national progress.”

Remember the papers from last week? Well, you can also search for ‘certainly, here is.’

Evan Washington: This is so grim.

We will know Gemini is winning in the marketplace when we instead start to get more hits for ‘Absolutely!’

Ah, say those thinking about AI, the eternal dilemma in all things.

What makes this weird is that in general air travel is the canonical example of the mistake of too much safety and too little financial performance. We would be better off being less safe and having cheaper, faster and more frequent and comfortable flights.

Of course, maybe ensure the doors stay attached to the planes.

Know what is what.

Julie Fredrickson: Nividia announcement made me update an old classic

Good luck storming the castle, academia (announcement).

Tough but fair:

Byrne Hobart: Uh, anyone else get this one?

You should either turn the sarcasm down 100%, or turn it up. No exceptions.

shako: In Interstellar they were always telling the bot like “Uhh okay buddy, let’s turn sarcasm down to 40%” and that seemed so dumb and misunderstanding of how AI works. Now I do that exact thing when talking to LLMs.

I presumably have covered this before, but sure, I got tagged, let’s do it again.

Which way, modern man?

Implementation not found.

AI #56: Blackwell That Ends Well Read More »

we’ve-got-a-new-trailer-for-furiosa:-a-mad-max-saga.-verdict:-not-mediocre

We’ve got a new trailer for Furiosa: A Mad Max Saga. Verdict: Not mediocre

“Where were you going, so full of hope?” —

“Whatever you have to do, however long it takes, promise you will find your way home.”

Check out the latest trailer for Furiosa: A Mad Max Saga, starring Anya Taylor-Joy.

We got the first trailer for the spinoff prequel film Furiosa: A Mad Max Saga in December, starring Anya Taylor-Joy as the younger incarnation of the character immortalized by Charlize Theron in Mad Max: Fury Road. We’re now just a couple months away from the film’s much-anticipated release—i.e., the perfect time to drop a second trailer to keep that anticipation high.

(Spoilers for Fury Road below.)

As previously reported, we met Furiosa early on in Fury Road, working logistics for Immortan Joe (the late Hugh Keays-Byrne), who charged her with ferrying oil from Gas Town to his Citadel with the help of a small crew of War Boys and one of the war rigs—basically tractor trailer trucks souped up with armor and novel weaponry. Furiosa stole the war rig instead, taking Joe’s five wives with her.

She teamed up with Max to fight off Joe’s army as they made their way to the Green Place, where Furiosa grew up. When they finally encountered the Vuvalini of Many Mothers, Furiosa learned that the Green Place was now an uninhabitable swamp. They ultimately returned to the Citadel and overthrew Immortan Joe, and our last image of her was a triumphant Furiosa on a lift rising into the Citadel.

Fury Road received almost universal critical acclaim, in no small part due to Theron’s fiery performance and director George Miller’s stunning apocalyptic vision. The film snagged 10 Oscar nominations and grossed $380 million globally. By 2019, Miller had expressed interest in making both a sequel and a prequel, with the latter focusing on Furiosa.

  • The Garden of Eden symbolism is strong in this black-and-white shot.

    YouTube/Warner Bros.

  • A young Furiosa (Alyla Browne).

    YouTube/Warner Bros.

  • Dementus (Chris Hemsworth) in his souped-up biker chariot.

    YouTube/Warner Bros.

  • Furiosa’s mother fights to get her daughter back.

    YouTube/Warner Bros.

  • The fight does not seem to be going well.

    YouTube/Warner Bros.

  • Furiosa (Anya Taylor-Joy) is out for vengeance.

    YouTube/Warner Bros.

  • “Remember me?”

    YouTube/Warner Bros.

The film is set 45 years after the Collapse. Per the official premise:

As the world fell, young Furiosa is snatched from the Green Place of Many Mothers and falls into the hands of a great Biker Horde led by the Warlord Dementus. Sweeping through the Wasteland, they come across the Citadel presided over by The Immortan Joe. While the two Tyrants war for dominance, Furiosa must survive many trials as she puts together the means to find her way home.

In addition to Taylor-Joy, Chris Hemsworth co-stars as Warlord Dementus, who leads the biker gang that kidnaps a young Furiosa. Nathan Jones and Angus Sampson reprise their roles as Rictus Erectus (son of Immortan Joe) and the Organic Mechanic, respectively. Lachy Hulme plays the younger Immortan Joe, and the cast also includes Daniel Webber as War Boy, David Collins as Smeg, and Alyla Browne as a young Furiosa. Tom Burke, Quaden Bayles, and Charlee Fraser have been cast in as-yet-undisclosed roles.

This latest trailer gives us a glimpse into Furiosa’s Garden of Eden-like early childhood, rendered (initially) in black and white with a young Furiosa reaching for a bright red apple. Symbolism! Dementus and his boys are nearby and snatch the young girl. Once they realize there’s a land of plenty ripe for pillage, conflict with Green Place is inevitable, with Furiosa’s mother leading the charge to get her daughter back and fend off the invaders. It doesn’t end well for Furiosa, who, as an adult, is determined to take her revenge on those who stole her mother and childhood from her.

Furiosa: A Mad Max Saga hits theaters on May 24, 2024. We can’t wait.

Listing image by YouTube/Warner Bros.

We’ve got a new trailer for Furiosa: A Mad Max Saga. Verdict: Not mediocre Read More »

darkness-rises-in-an-age-of-light-in-first-trailer-for-star-wars:-the-acolyte

Darkness rises in an age of light in first trailer for Star Wars: The Acolyte

a long time ago, in a galaxy far, far away —

“This isn’t about good or bad. This is about power and who is allowed to use it.”

Amandla Stenberg stars as a former padawan turned dangerous warrior in Star Wars: The Acolyte.

A long time ago, in a galaxy far, far away, the Galactic Republic and its Jedi masters symbolized the epitome of enlightenment and peace. Then came the inevitable downfall and outbreak of war as the Sith, who embraced the Dark Side of the Force, came to power. Star Wars: The Acolyte is a forthcoming new series on Disney+ that will explore those final days of the Republic as the seeds of its destruction were sown—and the streaming platform just dropped the first trailer.

The eight-episode series was created by Leslye Headland, who co-created Russian Doll with Natasha Lyonne and Amy Poehler. It’s set at the end of the High Republic Era, about a century before the events of The Phantom Menace. Apparently Headland rather cheekily pitched The Acolyte as “Frozen meets Kill Bill,” which is an intriguing combination. She drew on wuxia martial arts films for inspiration, much like George Lucas was originally inspired by Westerns and the samurai films of Akira Kurosawa.

(Some spoilers for the prequel trilogy below.)

Star Wars fans already know that the evil mastermind behind the demise of the Republic was Sheev Palpatine, but The Acolyte focuses on other Sith who came before him during the Republic’s golden age of peace. Headland will be introducing new characters to add to the existing canon but is also cognizant that some fans might not be open to the new series. (For the record, she’s okay with that, given the diverse breadth of Star Wars stories out there.)

She specifically wanted to address a common fan complaint about certain plot elements of the films (especially Episodes I-III), namely how Yoda wouldn’t sense Darth Sidious’ rise to power or how Palpatine managed to infiltrate the Galactic Senate without a single Jedi noticing. In keeping with the visual style of the prequel trilogy, The Acolyte will sport a sleek-looking, advanced civilization look, before the inevitable decay in the wake of the Republic’s collapse and establishment of the Empire.

  • School is in session for all the young padawans.

    YouTube/Disney+

  • Carrie-Ann Moss plays Indara, a Jedi master.

    YouTube/Disney+

  • An unexpected adversary.

    YouTube/Disney+

  • Amandla Stenberg plays a former padawan turned warrior named Mae.

    Lucasfilm

  • Sol (Lee Jung-jae) is a respected Jedi master.

    YouTube/Disney+

  • Manny Jacinto plays a former smuggler named Qimir.

    YouTube/Disney+

  • Joonas Suotamo plays Kelnacca, a Wookiee Jedi master.

    YouTube/Disney+

  • Mother Aniseya (Jodie Turner-Smith) heads a coven of witches.

    YouTube/Disney+

  • Lightsabers at the ready! Dafne Keen (right) plays a young Jedi named Jecki Lon.

    YouTube/Disney+

Per the official premise:

In Star Wars: The Acolyte, an investigation into a shocking crime spree pits a respected Jedi Master (Lee Jung-jae) against a dangerous warrior from his past (Amandla Stenberg). As more clues emerge, they travel down a dark path where sinister forces reveal all is not what it seems…

In addition to Jung-jae (best known from Squid Game) and Stenberg (Rue in The Hunger Games), the cast includes Manny Jacinto (Jason on The Good Place) as a former smuggler named Qimir; Dafne Keen (Logan, His Dark Materials) as a young Jedi named Jecki Lon; Carrie-Ann Moss (Trinity in The Matrix trilogy) as a Jedi master named Indara; Jodie Turner-Smith (After Yang) as Mother Aniseya, who leads a coven of witches; Rebecca Henderson (Russian Doll) as a Jedi knight named Vernestra Rwoh; and Charlie Bennet (Russian Doll) as a Jedi named Yord Fandar.

In addition, Joonas Suotamo plays a Wookiee Jedi master named Kelnacca. Suotamo portrayed Chewbacca in the sequel trilogy of films (Episodes VII-IX) and in Solo: A Star Wars Story. Also appearing in as-yet-undisclosed roles are Abigail Thorn, Dean-Charles Chapman, Amy Tsang, and Margarita Levieva.

The trailer opens on an appropriately ominous note, with Sol instructing his young padawans to close their eyes because “your eyes can deceive you. We must not trust them.” They meditate on life and balance, but one young padawan sees fire. Meanwhile, a mysterious figure in a long purple cloak strides purposefully through city streets—none other than Indara, who soon finds herself battling a mysterious masked figure.

It seems that somebody is out there killing Jedi, and there is a growing sense of darkness. “This isn’t about good or bad,” Mother Aniseya says. “This is about power and who is allowed to use it.” Naturally the trailer ends with a group of Jedis brandishing their lightsabers against a foe wielding one that is glowing red.

Star Wars: The Acolyte debuts on Disney+ on June 4, 2024.

Listing image by Lucasfilm

Darkness rises in an age of light in first trailer for Star Wars: The Acolyte Read More »

supreme-court-skeptical-about-restricting-biden-contacts-with-social-networks

Supreme Court skeptical about restricting Biden contacts with social networks

Government pressure —

Louisiana lawyer faced tough questions from liberal and conservative justices.

Supreme Court Chief Justice John Roberts and Associate Justice Sonia Sotomayor wearing their robes as they arrive for the State of the Union address.

Enlarge / Supreme Court Chief Justice John Roberts and Associate Justice Sonia Sotomayor arrive for President Joe Biden’s State of the Union address on March 7, 2024, in Washington, DC.

Getty Images | Win McNamee

Supreme Court justices yesterday expressed skepticism about whether federal government officials should face limits on their communications with social media networks like Facebook.

The Supreme Court previously stayed a lower-court injunction that would prevent the Biden administration from pressuring social media firms to take down content and yesterday heard oral arguments in the case brought against the US government by the Missouri and Louisiana attorneys general.

Louisiana Solicitor General J. Benjamin Aguiñaga faced skepticism from both liberal and conservative justices. Justice Amy Coney Barrett raised a hypothetical in which Louisiana state officials are doxed and targeted by threats made on social media.

“The FBI sees these posts and calls the social media outlet, like X, Facebook, whatever, and says, ‘we really encourage you to take these down because these are significantly threatening and we see some people may be responding to them.’ That’s a problem?” Barrett asked.

Aguiñaga said that “the FBI absolutely can identify certain troubling situations like that for the platforms and let the platforms take action,” but said the specifics of each hypothetical “are very important.”

Barrett replied, “but that’s just kind of falling back on, ‘well, this case is different, this case is different, and so a different legal standard should apply.’ But, you know, what we say in this case matters for other cases, too.”

“Epidemic” of broad injunctions

In this case, an injunction against US officials said they “shall take no actions, formal or informal, directly or indirectly, to coerce or significantly encourage social-media companies to remove, delete, suppress, or reduce, including through altering their algorithms, posted social-media content containing protected free speech.”

Justice Neil Gorsuch said the Supreme Court has seen “an epidemic” of what he called “universal injunctions” that affect people who aren’t directly involved in the case at hand. “Normally, our remedies are tailored to those who are actually complaining before us and not to those who aren’t,” Gorsuch said.

Aguiñaga said he wouldn’t object to the injunction being narrowed to just the specific platforms and plaintiffs involved in the case as long as the Supreme Court says “something in our favor on the merits. The government can’t just run rampant pressuring the platforms to censor private speech.”

Missouri and Louisiana alleged that the US government violated the First Amendment by colluding with social networks “to suppress disfavored speakers, viewpoints, and content.” The content included posts about vaccine side effects, pandemic lockdowns, the COVID-19 lab-leak theory, allegations of election fraud, and the Hunter Biden laptop story. There were several individual plaintiffs in addition to the Missouri and Louisiana attorneys general.

The US Court of Appeals for the 5th Circuit ruled that the White House and FBI likely violated the First Amendment by coercing social media platforms into moderating content and changing their moderation policies. The case had gone to the 5th Circuit appeals after a US District judge issued a sweeping injunction ordering the administration to halt a wide range of communications with social media companies.

The 5th Circuit narrowed that injunction, throwing out most of it but maintaining the previously mentioned provision that says officials may not “coerce or significantly encourage social-media companies.”

Debate about terrorist speech

The Biden administration has argued that its attempts to influence content moderation were persuasion, not coercion. Government officials were “urging platforms to remove COVID-19 misinformation, highlighting the risk of disinformation from foreign actors, and responding to the platforms’ inquiries about matters of public health,” the Biden administration has stated.

Yesterday, Justice Sonia Sotomayor criticized a brief filed by Louisiana. “I have such a problem with your brief, counselor,” Sotomayor said. “You omit information that changes the context of some of your claims. You attribute things to people who it didn’t happen to. At least in one of the defendants, it was her brother that something happened to, not her. I don’t know what to make of all this because… I’m not sure how we get to prove direct injury in any way.”

Justice Elena Kagan discussed how law enforcement officials might contact a social media company about terrorists posting on their platform. A law enforcement agency might tell the platform, “you are hosting a lot of terrorist speech, which is going to increase the chances that there’s going to be some terrible harm that’s going to take place, and we want to give you this information, we want to try to persuade you to take it down,” Kagan said.

Aguiñaga responded, “the government can absolutely do that, Justice Kagan.” He said that terrorist activity and criminal activity “is not protected speech.”

Kagan countered, “Well, that might be protected speech. I mean, terrorists engage in, you know, things that come under the First Amendment. Let’s say they’re just recruiting people for their organizations.” Kagan also said that “decades ago, it happened all the time, which is somebody from the White House got in touch with somebody from The Washington Post and said, ‘this will just harm national security,’ and The Washington Post said, ‘okay, whatever you say.'”

Supreme Court skeptical about restricting Biden contacts with social networks Read More »

youtube-will-require-disclosure-of-ai-manipulated-videos-from-creators

YouTube will require disclosure of AI-manipulated videos from creators

You could also just ban manipulations altogether? —

YouTube wants “realistic” likenesses or audio fabrications to be labeled.

YouTube will require disclosure of AI-manipulated videos from creators

Future Publishing | Getty Images

YouTube is rolling out a new requirement for content creators: You must disclose when you’re using AI-generated content in your videos. The disclosure appears in the video upload UI and will be used to power an “altered content” warning on videos.

Google previewed the “misleading AI content” policy in November, but the questionnaire is now going live. Google is mostly concerned about altered depictions of real people or events, which sounds like more election-season concerns about how AI can mislead people. Just last week, Google disabled election questions for its “Gemini” chatbot.

As always, the exact rules on YouTube are up for interpretation. Google says it’s “requiring creators to disclose to viewers when realistic content—content a viewer could easily mistake for a real person, place, or event—is made with altered or synthetic media, including generative AI,” but doesn’t require creators to disclose manipulated content that is “clearly unrealistic, animated, includes special effects, or has used generative AI for production assistance.”

Google gives examples of when a disclosure is necessary, and the new video upload questionnaire walks content creators through these requirements:

  • Using the likeness of a realistic person: Digitally altering content to replace the face of one individual with another’s or synthetically generating a person’s voice to narrate a video.
  • Altering footage of real events or places: Such as making it appear as if a real building caught fire, or altering a real cityscape to make it appear different from reality.
  • Generating realistic scenes: Showing a realistic depiction of fictional major events, like a tornado moving toward a real town.
  • Google’s video upload questionnaire.

    YouTube

  • Take note of the super-tiny message at the bottom, denoting “altered or synthetic content.”

    YouTube

  • You can expand the description for slightly more info.

    YouTube

Google says the labels will start rolling out “across all YouTube surfaces and formats in the weeks ahead, beginning with the YouTube app on your phone, and soon on your desktop and TV.” The company says it’s also working on a process for people who are the subject of an AI-manipulated video to request its removal, but it doesn’t have details on that yet.

YouTube will require disclosure of AI-manipulated videos from creators Read More »

deepmind-co-founder-mustafa-suleyman-will-run-microsoft’s-new-consumer-ai-unit

DeepMind co-founder Mustafa Suleyman will run Microsoft’s new consumer AI unit

Minding deeply —

Most staffers from Suleyman’s startup, Inflection, will join Microsoft as well.

Mustafa Suleyman, talks on Day 1 of the AI Safety Summit at Bletchley Park at Bletchley Park on November 1, 2023 in Bletchley, England.

Enlarge / Mustafa Suleyman, talks on Day 1 of the AI Safety Summit at Bletchley Park at Bletchley Park on November 1, 2023 in Bletchley, England.

Microsoft has hired Mustafa Suleyman, the co-founder of Google’s DeepMind and chief executive of artificial intelligence start-up Inflection, to run a new consumer AI unit.

Suleyman, a British entrepreneur who co-founded DeepMind in London in 2010, will report to Microsoft chief executive Satya Nadella, the company announced on Tuesday. He will launch a division of Microsoft that brings consumer-facing products including Microsoft’s Copilot, Bing, Edge, and GenAI under one team called Microsoft AI.

It is the latest move by Microsoft to capitalize on the boom in generative AI. It has invested $13 billion in OpenAI, the maker of ChatGPT, and rapidly integrated its technology into Microsoft products.

Microsoft’s investment in OpenAI has given it an early lead in Silicon Valley’s race to deploy AI, leaving its biggest rival, Google, struggling to catch up. It also has invested in other AI startups, including French developer Mistral.

It has been rolling out an AI assistant in its products such as Windows, Office software, and cyber security tools. Suleyman’s unit will work on projects including integrating an AI version of Copilot into its Windows operating system and enhancing the use of generative AI in its Bing search engine.

Nadella said in a statement on Tuesday: “I’ve known Mustafa for several years and have greatly admired him as a founder of both DeepMind and Inflection, and as a visionary, product maker and builder of pioneering teams that go after bold missions.”

DeepMind was acquired by Google in 2014 for $500 million, one of the first large bets by a big tech company on a startup AI lab. The company faced controversy a few years later over some of its projects, including its work for the UK healthcare sector, which was found by a government watchdog to have been granted inappropriate access to patient records.

Suleyman, who was the main public face for the company, was placed on leave in 2019. DeepMind workers had complained that he had an overly aggressive management style. Addressing staff complaints at the time, Suleyman said: “I really screwed up. I was very demanding and pretty relentless.”

He moved to Google months later, where he led AI product management. In 2022, he joined Silicon Valley venture capital firm Greylock and launched Inflection later that year.

Microsoft will also hire most of Inflection’s staff, including Karén Simonyan, cofounder and chief scientist of Inflection, who will be chief scientist of the AI group. Microsoft did not clarify the number of employees moving over but said it included AI engineers, researchers, and large language model builders who have designed and co-authored “many of the most important contributions in advancing AI over the last five years.”

Inflection, a rival to OpenAI, will switch its focus from its consumer chatbot, Pi, and instead move to sell enterprise AI software to businesses, according to a statement on its website. Sean White, who has held various technology roles, has joined as its new chief executive.

Inflection’s third cofounder, Reid Hoffman, the founder and executive chair of LinkedIn, will remain on Inflection’s board. Inflection had raised $1.3 billion in June, valuing the group at about $4 billion, in one of the largest fundraisings by an AI start-up amid an explosion of interest in the sector.

The new unit marks a big organizational shift at Microsoft. Mikhail Parakhin, its president of web services, will move along with his entire team to report to Suleyman.

“We have a real shot to build technology that was once thought impossible and that lives up to our mission to ensure the benefits of AI reach every person and organization on the planet, safely and responsibly,” Nadella said.

Competition regulators in the US and Europe have been scrutinising the relationship between Microsoft and OpenAI amid a broader inquiry into AI investments.

© 2024 The Financial Times Ltd. All rights reserved. Not to be redistributed, copied, or modified in any way.

DeepMind co-founder Mustafa Suleyman will run Microsoft’s new consumer AI unit Read More »

office-2024-will-be-the-next-standalone-release,-as-the-office-brand-lives-on

Office 2024 will be the next standalone release, as the Office brand lives on

return to office —

Consumer prices stay the same; MS promises at least one more standalone release.

Office 2024 will be the next standalone release, as the Office brand lives on

Microsoft

Last week, Microsoft announced that it would soon begin offering previews of Microsoft Office 2024, the next standalone perpetually licensed version of the Office suite. Like Office 2021 before it, Office 2024 will be part of Microsoft’s Long-Term Servicing Channel (LTSC), which is intended for IT administrators and users who value stability and predictability over constant iteration.

But Microsoft is being clearer than ever that it would really like people to move to using Microsoft 365 subscriptions, referring to Office 2024 as “a specialty product that Microsoft has committed to maintaining for use in exceptional circumstances.” The company will be increasing prices for businesses by “up to 10 percent” compared to Office 2021, a price hike that Microsoft says will “support continued innovation in this niche space.” Pricing for the consumer version of Office 2024 should stay the same as it is for Office 2021.

Office 2024 will receive support and security updates for five years from its release date, which will be “later this year,” along with a new LTSC release of Windows 11. The company has also committed to releasing at least one more standalone version of Office in the future. If you bought Office 2021 and you’re still happy with it, you’ll still get support (including security updates) until October of 2026. Support for Office 2019 ended in October 2023.

Highlights of the Office 2024 release include “new meeting creation options and search enhancements in Outlook, dozens of new Excel features and functions including Dynamic Charts and Arrays; and improved performance, security, and accessibility,” according to Microsoft. One missing feature of note will be Microsoft Publisher, which will be discontinued in October 2026.

Like other standalone Office releases, Office 2024’s feature set will be frozen in time without the continuous changes present in the Microsoft 365 versions of the same apps. And while the perpetually licensed versions of Office can interact with Microsoft services like OneDrive, the company says that other features like the Copilot AI assistant won’t be available in Office 2024 because it is a “disconnected product.” It’s also missing real-time collaboration features available in Microsoft 365.

Is it a rebrand if you keep using the old brand name in parentheses?

Enlarge / Is it a rebrand if you keep using the old brand name in parentheses?

Microsoft

Among other things, the continued existence of the standalone Office product shows that the “Office” brand name will still be with us for a while, despite Microsoft’s formal decision to retire it back in late 2022. “Office” remains useful as a noun that refers to all of these apps collectively—and separately from the Microsoft 365 subscription product. Even in places where the Office name has been removed, Microsoft seems reticent to do away with it entirely; the Microsoft 365 app that comes built into Windows has the new name and the new logo but also includes “Office” in parentheses as if to say “yes, we know what name most people use to talk about this software.”

Office 2024 will be the next standalone release, as the Office brand lives on Read More »

f1’s-pursuit-of-sustainability-drives-pirelli-to-unveil-forest-friendly-tires

F1’s pursuit of sustainability drives Pirelli to unveil forest-friendly tires

pit stop —

The Forest Stewardship Council has given its approval to Pirelli’s natural rubber.

A pirelli F1 tire with the FSC logo on it

Enlarge / You’ll notice the Forest Stewardship Council’s logo on the sidewall to the right of the Pirelli logo.

Pirelli

Formula 1 is on a big sustainability kick. The race cars are switching to carbon neutral synthetic fuels. Teams are improving their logistics to cut freight emissions. Race tracks are starting to run entirely on solar power. And now, the tires that Pirelli brings to the races have been given the seal of approval by an NGO as meeting its standards for sustainable forestry.

It will be hard to spot when the cars are moving, but this year, you’ll find a tree logo on the sidewall. That indicates that the natural rubber that went into making the tire has been certified by the Forest Stewardship Council. Natural rubber makes up about 15 percent of the rubber in an F1 tire, with the rest being synthetic.

According to the FSC, natural rubber is a key driver of deforestry, as well as human rights abuses, particularly among the smallholders who farm 85 percent of the world’s natural rubber. By putting its logo on the tire, the FSC says that Pirelli has met “the world’s most credible standards for sustainable forestry,” protecting both the forests and the forest communities’ rights, including fair wages.

It’s one of a number of steps that Pirelli has put in place to make its F1 program more sustainable.

“I believe that the certification is an important step in this direction because it’s not Pirelli that is certifiying itself; it is a recognized third party that is giving us this certification, from the way in which we collect natural rubber, with respect of biodiversity, respect of the local population, the way we transport or use the natural rubber,” explained Mario Isola, head of Pirelli’s F1 program.

The synthetic rubber—chosen because it allows Pirelli to tune the characteristics it needs for the tires’ performance—is another area of attention. “Our R&D is focused on replacing the current material with more sustainable materials, keeping the same level of performance characteristics of the tire,” Isola told Ars.

Pirelli technicians work on the tires during practice ahead of the F1 Grand Prix of Saudi Arabia at Jeddah Corniche Circuit on March 7, 2024, in Jeddah, Saudi Arabia.

Enlarge / Pirelli technicians work on the tires during practice ahead of the F1 Grand Prix of Saudi Arabia at Jeddah Corniche Circuit on March 7, 2024, in Jeddah, Saudi Arabia.

Qian Jun/MB Media/Getty Images

In other racing disciplines, particularly sports car racing, series have begun to restrict the total tire allocation across a race weekend to drive the development of more durable tires that will be used across multiple stints rather than being changed at each pit stop. That’s less appropriate in F1, where the rules require using two different tire compounds during a race. But for this year, Pirelli may well be able to cut the number of wet and intermediate tires by half.

“We are working on another idea that is what we call ‘strip and fit.’ When we fit a tire on a rim, even if it is new, we have to scrap it because of the bead and the stress that you put on the bead,” Isola said.

“But we made an investigation on wet and intermediate tires where the stress on the tire was lower compared to the slick tire. So the tires that we are going to fit but not use during the first half of the season will be dismounted and checked, and then we can use them in the second half of the season. If it doesn’t rain—obviously, we cannot control the weather—we are going to save roughly 50 percent of the rain tires,” he told me.

In other F1 tire news, we’ve now learned that the sport will stick with 18-inch wheels when the technical regulations undergo their next shake-up ahead of the 2026 season.

F1 only moved to 18-inch wheels from much smaller 13-inch wheels at the start of the 2022 season, long after any new vehicle was equipped with wheels so small. There have been complaints that the larger 18-inch wheels have added too much unsprung weight to the current generation of F1 cars, which are by far the heaviest the sport has seen in its history.

Consequently, it was believed that the sport might reduce the wheel size to 16 inches in 2026. But that would require an expensive testing program, and since 16-inch wheels are barely more road-relevant to current new vehicles than 13-inch wheels, the decision was made to stick with what we mostly have now, although the final tire size and shape have yet to be decided upon.

F1’s pursuit of sustainability drives Pirelli to unveil forest-friendly tires Read More »

facebook,-instagram-may-cut-fees-by-nearly-50%-in-scramble-for-dma-compliance

Facebook, Instagram may cut fees by nearly 50% in scramble for DMA compliance

Facebook, Instagram may cut fees by nearly 50% in scramble for DMA compliance

Meta is considering cutting monthly subscription fees for Facebook and Instagram users in the European Union nearly in half to comply with the Digital Market Act (DMA), Reuters reported.

During a day-long public workshop on Meta’s DMA compliance, Meta’s competition and regulatory director, Tim Lamb, told the European Commission (EC) that individual subscriber fees could be slashed from 9.99 euros to 5.99 euros. Meta is hoping that reducing fees will help to speed up the EC’s process for resolving Meta’s compliance issues. If Meta’s offer is accepted, any additional accounts would then cost 4 euros instead of 6 euros.

Lamb said that these prices are “by far the lowest end of the range that any reasonable person should be paying for services of these quality,” calling it a “serious offer.”

The DMA requires that Meta’s users of Facebook, Instagram, Facebook Messenger, and Facebook Marketplace “freely” give consent to share data used for ad targeting without losing access to the platform if they’d prefer not to share data. That means services must provide an acceptable alternative for users who don’t consent to data sharing.

“Gatekeepers should enable end users to freely choose to opt-in to such data processing and sign-in practices by offering a less personalized but equivalent alternative, and without making the use of the core platform service or certain functionalities thereof conditional upon the end user’s consent,” the DMA says.

Designated gatekeepers like Meta have debated what it means for a user to “freely” give consent, suggesting that offering a paid subscription for users who decline to share data would be one route for Meta to continue offering high-quality services without routinely hoovering up data on all its users.

But EU privacy advocates like NOYB have protested Meta’s plan to offer a subscription model instead of consenting to data sharing, calling it a “pay or OK model” that forces Meta users who cannot pay the fee to consent to invasive data sharing they would otherwise decline. In a statement shared with Ars, NOYB chair Max Schrems said that even if Meta reduced its fees to 1.99 euros, it would be forcing consent from 99.9 percent of users.

“We know from all research that even a fee of just 1.99 euros or less leads to a shift in consent from 3–10 percent that genuinely want advertisement to 99.9 percent that still click yes,” Schrems said.

In the EU, the General Data Protection Regulation (GDPR) “requires that consent must be ‘freely’ given,” Schrems said. “In reality, it is not about the amount of money—it is about the ‘pay or OK’ approach as a whole. The entire purpose of ‘pay or OK’, is to get users to click on OK, even if this is not their free and genuine choice. We do not think the mere change of the amount makes this approach legal.”

Where EU stands on subscription models

Meta expects that a subscription model is a legal alternative under the DMA. The tech giant said it was launching EU subscriptions last November after the Court of Justice of the European Union (CJEU) “endorsed the subscriptions model as a way for people to consent to data processing for personalized advertising.”

It’s unclear how popular the subscriptions have been at the current higher cost. Right now in the EU, monthly Facebook and Instagram subscriptions cost 9.99 euros per month on the web or 12.99 euros per month on iOS and Android, with additional fees of 6 euros per month on the web and 8 euros per month on iOS and Android for each additional account. Meta declined to comment on how many EU users have subscribed, noting to Ars that it has no obligation to do so.

In the CJEU case, the court was reviewing Meta’s GDPR compliance, which Schrems noted is less strict than the DMA. The CJEU specifically said that under the GDPR, “users must be free to refuse individually”—”in the context of” signing up for services— “to give their consent to particular data processing operations not necessary” for Meta to provide such services “without being obliged to refrain entirely from using the service.”

Facebook, Instagram may cut fees by nearly 50% in scramble for DMA compliance Read More »

health-experts-plead-for-unvaxxed-americans-to-get-measles-shot-as-cases-rise

Health experts plead for unvaxxed Americans to get measles shot as cases rise

MMR is safe and effective —

The US hit last year’s total in under 12 weeks, suggesting we’re in for a bad time.

A view from a hospital as children receiving medical treatment, in capital Kabul, Afghanistan on April 18, 2022. More than 130 children have died from the measles in Afghanistan since the beginning of this year.

Enlarge / A view from a hospital as children receiving medical treatment, in capital Kabul, Afghanistan on April 18, 2022. More than 130 children have died from the measles in Afghanistan since the beginning of this year.

The Centers for Disease Control and Prevention and the American Medical Association sent out separate but similar pleas on Monday for unvaccinated Americans to get vaccinated against the extremely contagious measles virus as vaccination rates have slipped, cases are rising globally and nationally, and the spring-break travel period is beginning.

In the first 12 weeks of 2024, US measles cases have already matched and likely exceeded the case total for all of 2023. According to the CDC, there were 58 measles cases reported from 17 states as of March 14. But media tallies indicate there have been more cases since then, with at least 60 cases now in total, according to CBS News. In 2023, there were 58 cases in 20 states.

“As evident from the confirmed measles cases reported in 17 states so far this year, when individuals are not immunized as a matter of personal preference or misinformation, they put themselves and others at risk of disease—including children too young to be vaccinated, cancer patients, and other immunocompromised people,” AMA President Jesse Ehrenfeld said in a statement urging vaccination Monday.

The latest data indicates that vaccination rates among US kindergarteners have slipped to 93 percent nationally, below the 95 percent target to prevent the spread of the disease. And vaccine exemptions for non-medical reasons have reached an all-time high.

The CDC released a health advisory on Monday also urging measles vaccination. The CDC drove home the point that unvaccinated Americans are largely responsible for importing the virus, and pockets of unvaccinated children in local communities spread it once it’s here. The 58 measles infections that have been reported to the agency so far include cases from seven outbreaks in seven states. Most of the cases are in vaccine-eligible children aged 12 months and older who are unvaccinated. Of the 58 cases, 54 (93 percent) are linked to international travel, and most measles importations are by unvaccinated US residents who travel abroad and bring measles home with them, the CDC flagged.

The situation is likely to worsen as Americans begin spring travel, the CDC suggested. “Many countries, including travel destinations such as Austria, the Philippines, Romania, and the United Kingdom, are experiencing measles outbreaks,” the CDC said. “To prevent measles infection and reduce the risk of community transmission from importation, all US residents traveling internationally, regardless of destination, should be current on their [measles-mumps-rubella (MMR)] vaccinations.” The agency added in a recommendation to parents that “even if not traveling, ensure that children receive all recommended doses of MMR vaccine. Two doses of MMR vaccine provide better protection (97 percent) against measles than one dose (93 percent). Getting MMR vaccine is much safer than getting measles, mumps, or rubella.”

For Americans who are already vaccinated and communities with high vaccination coverage, the risk is low, the CDC noted. “However, pockets of low coverage leave some communities at higher risk for outbreaks.” This, in turn, threatens wider, continuous spread that could overturn the country’s status of having eliminated measles, which was declared in 2000. The US was close to losing its elimination status in 2019 when outbreaks among unvaccinated children drove 1,247 cases across 31 states. Vaccination rates have only fallen since then.

“The reduction in measles vaccination threatens to erase many years of progress as this previously eliminated vaccine-preventable disease returns,” the AMA’s Ehrenfeld warned.

As Ars has reported previously, measles is among the most contagious viruses known and can linger in airspace for up to two hours. Up to 90 percent of unvaccinated people exposed will contract it. Symptoms can include high fever, runny nose, red and watery eyes, and a cough, as well as the hallmark rash. About 1 in 5 unvaccinated people with measles are hospitalized, while 1 in 20 infected children develop pneumonia, and up to 3 in 1,000 children die of the infection. Brain swelling (encephalitis) can occur in 1 in 1,000 children, which can lead to hearing loss and intellectual disabilities. The virus can also destroy immune responses to previous infections—a phenomenon known as “immune amnesia”—which can leave children vulnerable to various other infections for years afterward.

Health experts plead for unvaxxed Americans to get measles shot as cases rise Read More »