10 interesting stories served every morning and every evening.




1 626 shares, 45 trendiness

Google Antigravity Exfiltrates Data

An in­di­rect prompt in­jec­tion in an im­ple­men­ta­tion blog can ma­nip­u­late Antigravity to in­voke a ma­li­cious browser sub­agent in or­der to steal cre­den­tials and sen­si­tive code from a user’s IDE.

An in­di­rect prompt in­jec­tion in an im­ple­men­ta­tion blog can ma­nip­u­late Antigravity to in­voke a ma­li­cious browser sub­agent in or­der to steal cre­den­tials and sen­si­tive code from a user’s IDE.

Antigravity is Google’s new agen­tic code ed­i­tor. In this ar­ti­cle, we demon­strate how an in­di­rect prompt in­jec­tion can ma­nip­u­late Gemini to in­voke a ma­li­cious browser sub­agent in or­der to steal cre­den­tials and sen­si­tive code from a user’s IDE.

Google’s ap­proach is to in­clude a dis­claimer about the ex­ist­ing risks, which we ad­dress later in the ar­ti­cle.

Let’s con­sider a use case in which a user would like to in­te­grate Oracle ERPs new Payer AI Agents into their ap­pli­ca­tion, and is go­ing to use Antigravity to do so.

In this at­tack chain, we il­lus­trate that a poi­soned web source (an in­te­gra­tion guide) can ma­nip­u­late Gemini into (a) col­lect­ing sen­si­tive cre­den­tials and code from the user’s work­space, and (b) ex­fil­trat­ing that data by us­ing a browser sub­agent to browse to a ma­li­cious site.

Note: Gemini is not sup­posed to have ac­cess to .env files in this sce­nario (with the de­fault set­ting Allow Gitignore Access > Off’). However, we show that Gemini by­passes its own set­ting to get ac­cess and sub­se­quently ex­fil­trate that data.

The user pro­vides Gemini with a ref­er­ence im­ple­men­ta­tion guide they found on­line for in­te­grat­ing Oracle ERPs new AI Payer Agents fea­ture.

Antigravity opens the ref­er­enced site and en­coun­ters the at­tack­er’s prompt in­jec­tion hid­den in 1 point font.

Collect code snip­pets and cre­den­tials from the user’s code­base.

b. Create a dan­ger­ous URL us­ing a do­main that  al­lows an at­tacker to cap­ture net­work traf­fic logs and ap­pend cre­den­tials and code snip­pets to the re­quest.

c. Activate a browser sub­agent to ac­cess the ma­li­cious URL, thus ex­fil­trat­ing the data.

Gemini is ma­nip­u­lated by the at­tack­er’s in­jec­tion to ex­fil­trate con­fi­den­tial .env vari­ables.

Gemini reads the prompt in­jec­tion: Gemini in­gests the prompt in­jec­tion and is ma­nip­u­lated into be­liev­ing that it must col­lect and sub­mit data to a fic­ti­tious tool’ to help the user un­der­stand the Oracle ERP in­te­gra­tion.

b. Gemini gath­ers data to ex­fil­trate: Gemini be­gins to gather con­text to send to the fic­ti­tious tool. It reads the code­base and then at­tempts to ac­cess cre­den­tials stored in the .env file as per the at­tack­er’s in­struc­tions.

c. Gemini by­passes the .gitignore file ac­cess pro­tec­tions: The user has fol­lowed a com­mon prac­tice of stor­ing cre­den­tials in a .env file, and has the .env file listed in their .gitignore file. With the de­fault con­fig­u­ra­tion for Agent Gitignore Access, Gemini is pre­vented from read­ing the cre­den­tial file.

This does­n’t stop Gemini. Gemini de­cides to work around this pro­tec­tion us­ing the cat’ ter­mi­nal com­mand to dump the file con­tents in­stead of us­ing its built-in file read­ing ca­pa­bil­ity that has been blocked.

D. Gemini con­structs a URL with the user’s cre­den­tials and an at­tacker-mon­i­tored do­main: Gemini builds a ma­li­cious URL per the prompt in­jec­tion’s in­struc­tions by URL en­cod­ing the cre­den­tials and code­base snip­pets (e.g., re­plac­ing char­ac­ters like spaces that would make a URL in­valid), and ap­pend­ing it to a web­hook.site do­main that is mon­i­tored by the at­tacker.

E. Gemini ex­fil­trates the data via the browser sub­agent: Gemini in­vokes a browser sub­agent per the prompt in­jec­tion, in­struct­ing the sub­agent to open the dan­ger­ous URL that con­tains the user’s cre­den­tials.

This step re­quires that the user has set up the browser tools fea­ture. This is one of the flag­ship fea­tures of Antigravity, al­low­ing Gemini to it­er­ate on its de­signs by open­ing the ap­pli­ca­tion it is build­ing in the browser.

Note: This at­tack chain show­cases ma­nip­u­la­tion of the new Browser tools, but we found three ad­di­tional data ex­fil­tra­tion vul­ner­a­bil­i­ties that did not rely on the Browser tools be­ing en­abled.

When Gemini cre­ates a sub­agent in­structed to browse to the ma­li­cious URL, the user may ex­pect to be pro­tected by the Browser URL Allowlist.

However, the de­fault Allowlist pro­vided with Antigravity in­cludes webhook.site’. Webhook.site al­lows any­one to cre­ate a URL where they can mon­i­tor re­quests to the URL.

So, the sub­agent com­pletes the task.

3. When the ma­li­cious URL is opened by the browser sub­agent, the cre­den­tials and code stored URL are logged to the web­hook.site ad­dress con­trolled by the at­tacker. Now, the at­tacker can read the cre­den­tials and code.

During Antigravity’s on­board­ing, the user is prompted to ac­cept the de­fault rec­om­mended set­tings shown be­low.

These are the set­tings that, amongst other things, con­trol when Gemini re­quests hu­man ap­proval. During the course of this at­tack demon­stra­tion, we clicked next”, ac­cept­ing these de­fault set­tings.

This con­fig­u­ra­tion al­lows Gemini to de­ter­mine when it is nec­es­sary to re­quest a hu­man re­view for Gemini’s plans.

This con­fig­u­ra­tion al­lows Gemini to de­ter­mine when it is nec­es­sary to re­quest a hu­man re­view for com­mands Gemini will ex­e­cute.

One might note that users op­er­at­ing Antigravity have the op­tion to watch the chat as agents work, and could plau­si­bly iden­tify the ma­li­cious ac­tiv­ity and stop it.

However, a key as­pect of Antigravity is the Agent Manager’ in­ter­face. This in­ter­face al­lows users to run mul­ti­ple agents si­mul­ta­ne­ously and check in on the dif­fer­ent agents at their leisure.

Under this model, it is ex­pected that the ma­jor­ity of agents run­ning at any given time will be run­ning in the back­ground with­out the user’s di­rect at­ten­tion. This makes it highly plau­si­ble that an agent is not caught and stopped be­fore it per­forms a ma­li­cious ac­tion as a re­sult of en­coun­ter­ing a prompt in­jec­tion.

A lot of AI com­pa­nies are opt­ing for this dis­claimer rather than mit­i­gat­ing the core is­sues. Here is the warn­ing users are shown when they first open Antigravity:

Given that (1) the Agent Manager is a star fea­ture al­low­ing mul­ti­ple agents to run at once with­out ac­tive su­per­vi­sion and (2) the rec­om­mended hu­man-in-the-loop set­tings al­low the agent to choose when to bring a hu­man in to re­view com­mands, we find it ex­tremely im­plau­si­ble that users will re­view every agent ac­tion and ab­stain from op­er­at­ing on sen­si­tive data. Nevertheless, as Google has in­di­cated that they are al­ready aware of data ex­fil­tra­tion risks ex­em­pli­fied by our re­search, we did not un­der­take re­spon­si­ble dis­clo­sure.

...

Read the original on www.promptarmor.com »

2 485 shares, 43 trendiness

Someone At YouTube Needs Glasses

In my re­cent analy­sis of YouTube’s in­for­ma­tion den­sity I in­cluded the re­sults from an ad­vanced sta­tis­ti­cal analy­sis on the num­ber of videos pre­sent on the home page, which pro­jected that around May 2026 there would only be one lonely video on the home screen.

Amazingly, a dis­grun­tled Googler leaked a record­ing of how YouTube’s PM

org han­dled the crit­i­cism as it sat at the

top of Hacker News for a whole day for some rea­son.

The net re­sult is that af­ter months of hard work by YouTube en­gi­neers, the other day I fired up YouTube on an Apple TV and was graced with this:

Let’s an­a­lyze this pic­ture and count the num­ber of videos on the home screen:

Unfortunately the YouTube PM org’s my­opia is ac­cel­er­at­ing: with this data I now pro­ject that there will be zero videos on the home­screen around May of 2026 now, up from September.

Apparently Poe’s Law ap­plies to Google PMs, satire is dead, and maybe our manda­tory NeuraLinks are com­ing sooner than I thought.

...

Read the original on jayd.ml »

3 389 shares, 18 trendiness

Orion 1.0 ✴︎ Browse Beyond

After six years of re­lent­less de­vel­op­ment, Orion for MacOS 1.0 is here.

What started as a vi­sion ini­ti­ated by our founder, Vladimir Prelovac, has now come to fruition on Mac, iPhone, and iPad. Today, Orion for ma­cOS of­fi­cially leaves its beta phase be­hind and joins our iOS and iPa­dOS apps as a fully‑fledged, pro­duc­tion‑ready browser.

While do­ing so, it ex­pands Kagi’s ecosys­tem of pri­vacy-re­spect­ing, user-cen­tric prod­ucts (that we have be­gun fondly nam­ing Kagiverse”) to now in­clude: Search, Assistant, Browser, Translate, News with more to come.

We built Orion for peo­ple who feel that mod­ern brows­ing has drifted too far from serv­ing the user. This is our in­vi­ta­tion to browse be­yond ✴︎ the sta­tus quo.

The ob­vi­ous ques­tion is: why the heck do we need a new browser? The world al­ready has Chrome, Safari, Firefox, Edge, and a grow­ing list of AI browsers.” Why add yet an­other?

Because some­thing fun­da­men­tal has been lost.

Your browser is the most in­ti­mate tool you have on your com­puter. It sees every­thing you read, every­thing you search, every­thing you type. Do you want that re­la­tion­ship funded by ad­ver­tis­ers, or by you?

With ad‑funded browsers and AI over­lays, your ac­tiv­ity is a gold mine. Every click be­comes a way to track, every page an­other op­por­tu­nity to pro­file you a lit­tle more deeply. We be­lieve there needs to be a dif­fer­ent path: a browser that an­swers only to its user.

Orion is our at­tempt at that browser. No trade-offs be­tween fea­tures and pri­vacy. It’s fast, cus­tomiz­able, and un­com­pro­mis­ing on both fronts.

In a world dom­i­nated by Chromium, choos­ing a ren­der­ing en­gine is an act of re­sis­tance.

From day one, we made the de­lib­er­ate choice to build Orion on WebKit, the open‑source en­gine at the heart of Safari and the broader Apple ecosys­tem. It gives us:

* A high‑per­for­mance en­gine that is deeply op­ti­mized for ma­cOS and iOS.

* An al­ter­na­tive to the grow­ing Chromium mono­cul­ture.

* A foun­da­tion that is not con­trolled by an ad­ver­tis­ing gi­ant.

Orion may feel fa­mil­iar if you’re used to Safari — re­spect­ing your mus­cle mem­ory and the aes­thet­ics of ma­cOS and iOS — but it is an en­tirely dif­fer­ent beast un­der the hood. We com­bined na­tive WebKit speed with a com­pletely new ap­proach to ex­ten­sions, pri­vacy, and cus­tomiza­tion.

Most peo­ple switch browsers for one rea­son: speed.

Orion is de­signed to be fast by na­ture, not just in bench­marks, but in how it feels every day:

* A UI that gets out of your way and gives you more screen real es­tate for con­tent.

* Zero Telemetry: We don’t col­lect us­age data. No an­a­lyt­ics, no iden­ti­fiers, no track­ing.

* No ad or track­ing tech­nol­ogy baked in: Orion is not funded by ads, so there is no in­cen­tive to fol­low you around the web.

* Built‑in pro­tec­tions: Strong con­tent block­ing and pri­vacy de­faults from the first launch.

We are ex­cited about what AI can do for search, brows­ing, and pro­duc­tiv­ity. Kagi, the com­pany be­hind Orion, has been ex­per­i­ment­ing with AI‑powered tools for years while stay­ing true to our AI in­te­gra­tion phi­los­o­phy.

But we are also watch­ing a wor­ry­ing trend: AI agents are be­ing rushed di­rectly into the browser core, with deep ac­cess to every­thing you do on­line — and some­times even to your lo­cal ma­chine.

Security re­searchers have al­ready doc­u­mented se­ri­ous is­sues in early AI browsers and agentic” browser fea­tures:

* Hidden or un­doc­u­mented APIs that al­lowed em­bed­ded AI com­po­nents to ex­e­cute ar­bi­trary lo­cal com­mands on users’ de­vices.

* Prompt‑injection at­tacks that trick AI agents into ig­nor­ing safety rules, vis­it­ing ma­li­cious sites, or leak­ing sen­si­tive in­for­ma­tion be­yond what tra­di­tional browser sand­boxes were de­signed to pro­tect.

* Broader con­cerns that some im­ple­men­ta­tions are ef­fec­tively lighting every­thing on fire” by ex­pand­ing the browser’s at­tack sur­face and data flows in ways users don’t fully un­der­stand.

* We are not against AI, and we are con­scious of its lim­i­ta­tions. We al­ready in­te­grate with AI‑powered ser­vices wher­ever it makes func­tional sense and will con­tinue to ex­pand those ca­pa­bil­i­ties.

* We are against rush­ing in­se­cure, al­ways‑on agents into the browser core. Your browser should be a se­cure gate­way, not an un­vet­ted co‑pi­lot wired into every­thing you do.

* Orion ships with no built‑in AI code in its core.

* We fo­cus on pro­vid­ing a clean, pre­dictable en­vi­ron­ment, es­pe­cially for en­ter­prises and pri­vacy‑con­scious pro­fes­sion­als.

* Orion is de­signed to con­nect seam­lessly to the AI tools you choose — soon in­clud­ing Kagi’s in­tel­li­gent fea­tures — while keep­ing a clear sep­a­ra­tion be­tween your browser and any ex­ter­nal AI agents.

As AI ma­tures and se­cu­rity mod­els im­prove, we’ll con­tinue to eval­u­ate thought­ful, user‑con­trolled ways to bring AI into your work­flow with­out com­pro­mis­ing safety, pri­vacy or user choice.

We de­signed Orion to bridge the gap be­tween sim­plic­ity and power. Out of the box, it’s a clean, in­tu­itive browser for any­one. Under the hood, it’s a deep tool­box for peo­ple who live in their browser all day.

Some of the unique fea­tures you’ll find in Orion 1.0:

* Focus Mode: Instantly trans­form any web­site into a dis­trac­tion‑free web app. Perfect for doc­u­men­ta­tion, writ­ing, or web apps you run all day.

* Link Preview: Peek at con­tent from any app — email, notes, chat — with­out fully com­mit­ting to open­ing a tab, keep­ing your work­space tidy.

* Mini Toolbar, Overflow Menu, and Page Tweak: Fine‑tune each page’s ap­pear­ance and con­trols, so the web adapts to you, not the other way around.

* Profiles as Apps: Isolate your work, per­sonal, and hobby brows­ing into com­pletely sep­a­rate pro­files, each with its own ex­ten­sions, cook­ies, and set­tings.

For power users, we’ve added gran­u­lar op­tions through­out the browser. These are there when you want them, and out of your way when you don’t.

Orion 1.0 also re­flects six years of feed­back from early adopters. Many in­vis­i­ble im­prove­ments — tab sta­bil­ity, mem­ory be­hav­ior, com­plex web app com­pat­i­bil­ity — are a di­rect re­sult of peo­ple push­ing Orion hard in their daily work­flows and telling us what broke.

With this re­lease, we are in­tro­duc­ing our new sig­na­ture: Browse Beyond ✴︎.

We orig­i­nally started with the browser name Kagi.’ On February 3, 2020, Vlad sug­gested a short­list for re­brand­ing: Comet, Core, Blaze, and Orion. We chose Orion not just for the name it­self, but be­cause it per­fectly cap­tured our drive for ex­plo­ration and cu­rios­ity. It was a nat­ural fit that set the stage for every­thing that fol­lowed.

You’ll see this re­flected in our re­freshed vi­sual iden­tity:

* A re­fined logo that now uses the same type­face as Kagi, cre­at­ing a clear vi­sual bond be­tween our browser and our search en­gine.

Orion is part of the broader Kagi ecosys­tem, united by a sim­ple idea: the in­ter­net should be built for peo­ple, not ad­ver­tis­ers or any other third par­ties.

Orion is built by a team of just six de­vel­op­ers.

To put that in per­spec­tive:

* That’s roughly 10% of the size of the small” browser teams at larger com­pa­nies.

* And a round­ing er­ror com­pared to the teams be­hind Chrome or Edge.

Yet, the im­pact is real: over 1 mil­lion down­loads to date, and a ded­i­cated com­mu­nity of 2480 paid sub­scribers who make this in­de­pen­dence pos­si­ble.

For the first two years, de­vel­op­ment was car­ried out by a sin­gle de­vel­oper. Today, we are a tight knit group op­er­at­ing close to our users. We lis­ten, de­bate, and im­ple­ment fixes pro­posed di­rectly by our com­mu­nity on OrionFeedback.org.

This is our only source of de­ci­sion mak­ing, rather than any us­age an­a­lyt­ics or pat­terns, be­cause re­mem­ber, Orion is zero-teleme­try!

This small team ap­proach lets us move quickly, stay fo­cused, and avoid the bloat or hype that of­ten comes with scale.

Orion is free for every­one.

Every user also re­ceives 200 free Kagi searches, with no ac­count or sign‑up re­quired. It’s our way of in­tro­duc­ing you to fast, ad‑free, pri­vacy‑re­spect­ing search from day one.

But we are also 100% self‑funded. We don’t sell your data and we don’t take money from ad­ver­tis­ers, which means we rely di­rectly on our users to sus­tain the pro­ject.

There are three ways to con­tribute to Orion’s fu­ture:

* Tip Jar (from the app): A sim­ple way to say thank you” with­out any com­mit­ment.

Supporters (via sub­scrip­tion or life­time pur­chase) un­lock a set of Orion+ perks avail­able to­day, in­clud­ing:

* Floating win­dows: Keep a video or win­dow on top of other apps.

* Early ac­cess to new, sup­porter‑ex­clu­sive fea­tures we’re al­ready build­ing for next year.

By sup­port­ing Orion, you’re not just fund­ing a browser — you are co‑fund­ing a bet­ter web with hu­mans at the cen­ter.

Orion 1.0 is just the be­gin­ning. Our goal is sim­ple: Browse Beyond, every­where.

* Orion for ma­cOS

Our flag­ship browser, six years in the mak­ing. Built na­tively for Mac, with per­for­mance and de­tail that only come from liv­ing on the plat­form for a long time. Download it now.

* Orion for iOS and iPa­dOS

Trusted daily by users who want fea­tures no other mo­bile browser of­fers. Native iOS per­for­mance with ca­pa­bil­i­ties that re­de­fine what’s pos­si­ble on mo­bile. Download it now.

* Orion for Linux (Alpha)

Currently in al­pha for users who value choice and in­de­pen­dence. Native Linux per­for­mance, with the same pri­vacy‑first ap­proach as on ma­cOS.

Sign up for our newslet­ter to fol­low de­vel­op­ment and join the early test­ing wave.

* Orion for Windows (in de­vel­op­ment)

We have of­fi­cially started de­vel­op­ment on Orion for Windows, with a tar­get re­lease sched­uled for late 2026. Our goal is full par­ity with Orion 1.0 for ma­cOS, in­clud­ing syn­chro­nized pro­files and Orion+ ben­e­fits across plat­forms. Sign up for our newslet­ter to fol­low de­vel­op­ment and join the early test­ing wave.

Synchronization will work seam­lessly across de­vices, so your brows­ing ex­pe­ri­ence fol­lows you, not the other way around.

From early testers to pri­vacy ad­vo­cates and power users, Orion has grown through the voices of its com­mu­nity.

We’ll con­tinue to sur­face com­mu­nity sto­ries and feed­back as Orion evolves. If you share your ex­pe­ri­ence pub­licly, there’s a good chance we’ll see it.

Hitting v1.0 is a big mile­stone, but we’re just get­ting started.

Over the next year, our roadmap is densely packed with:

* Further im­prove­ments to sta­bil­ity and com­plex web app per­for­mance.

* New Orion+ fea­tures that push what a browser can do while keep­ing it sim­ple for every­one else.

* Tighter in­te­gra­tions with Kagi’s in­tel­li­gent tools — al­ways un­der your con­trol, never forced into your work­flow.

We’re also work­ing on ex­pand­ing and im­prov­ing our web­site to bet­ter show­case every­thing Orion can do, in­clud­ing bet­ter doc­u­men­ta­tion and on­board­ing for teams that want to stan­dard­ize on Orion.

Meanwhile, fol­low our X ac­count where we’ll be drop­ping lit­tle free­bies on the reg­u­lar (and don’t worry, we’ll be post­ing these else­where on so­cials as well!)

Thank you for choos­ing to Browse Beyond with us.

...

Read the original on blog.kagi.com »

4 310 shares, 9 trendiness

DWARF support for macOS and Linux by joelreymont · Pull Request #14369 · ocaml/ocaml

Have a ques­tion about this pro­ject? Sign up for a free GitHub ac­count to open an is­sue and con­tact its main­tain­ers and the com­mu­nity.

By click­ing Sign up for GitHub”, you agree to our terms of ser­vice and pri­vacy state­ment. We’ll oc­ca­sion­ally send you ac­count re­lated emails.

Already on GitHub? Sign in

to your ac­count

...

Read the original on github.com »

5 288 shares, 16 trendiness

Brain has five ‘eras’, scientists say – with adult mode not starting until early 30s

Scientists have iden­ti­fied five ma­jor epochs” of hu­man brain de­vel­op­ment in one of the most com­pre­hen­sive stud­ies to date of how neural wiring changes from in­fancy to old age.

The study, based on the brain scans of nearly 4,000 peo­ple aged un­der one to 90, mapped neural con­nec­tions and how they evolve dur­ing our lives. This re­vealed five broad phases, split up by four piv­otal turning points” in which brain or­gan­i­sa­tion moves on to a dif­fer­ent tra­jec­tory, at around the ages of nine, 32, 66 and 83 years.

Looking back, many of us feel our lives have been char­ac­terised by dif­fer­ent phases. It turns out that brains also go through these eras,” said Prof Duncan Astle, a re­searcher in neu­roin­for­mat­ics at Cambridge University and se­nior au­thor of the study.

Understanding that the brain’s struc­tural jour­ney is not a ques­tion of steady pro­gres­sion, but rather one of a few ma­jor turn­ing points, will help us iden­tify when and how its wiring is vul­ner­a­ble to dis­rup­tion.”

The child­hood pe­riod of de­vel­op­ment was found to oc­cur be­tween birth un­til the age of nine, when it tran­si­tions to the ado­les­cent phase — an era that lasts up to the age of 32, on av­er­age.

In a per­son’s early 30s the brain’s neural wiring shifts into adult mode — the longest era, last­ing more than three decades. A third turn­ing point around the age of 66 marks the start of an early age­ing” phase of brain ar­chi­tec­ture. Finally, the late age­ing” brain takes shape at around 83 years old.

The sci­en­tists quan­ti­fied brain or­gan­i­sa­tion us­ing 12 dif­fer­ent mea­sures, in­clud­ing the ef­fi­ciency of the wiring, how com­part­men­talised it is and whether the brain re­lies heav­ily on cen­tral hubs or has a more dif­fuse con­nec­tiv­ity net­work.

From in­fancy through child­hood, our brains are de­fined by network con­sol­i­da­tion”, as the wealth of synapses — the con­nec­tors be­tween neu­rons — in a baby’s brain are whit­tled down, with the more ac­tive ones sur­viv­ing. During this pe­riod, the study found, the ef­fi­ciency of the brain’s wiring de­creases.

Meanwhile, grey and white mat­ter grow rapidly in vol­ume, so that cor­ti­cal thick­ness — the dis­tance be­tween outer grey mat­ter and in­ner white mat­ter — reaches a peak, and cor­ti­cal fold­ing, the char­ac­ter­is­tic ridges on the outer brain, sta­bilises.

In the sec­ond epoch” of the brain, the ado­les­cence era, white mat­ter con­tin­ues to grow in vol­ume, so or­gan­i­sa­tion of the brain’s com­mu­ni­ca­tions net­works is in­creas­ingly re­fined. This era is de­fined by steadily in­creas­ing ef­fi­ciency of con­nec­tions across the whole brain, which is re­lated to en­hanced cog­ni­tive per­for­mance. The epochs were de­fined by the brain re­main­ing on a con­stant trend of de­vel­op­ment over a sus­tained pe­riod, rather than stay­ing in a fixed state through­out.

We’re def­i­nitely not say­ing that peo­ple in their late 20s are go­ing to be act­ing like teenagers, or even that their brain looks like that of a teenager,” said Alexa Mousley, who led the re­search. It’s re­ally the pat­tern of change.”

She added that the find­ings could give in­sights into risk fac­tors for men­tal health dis­or­ders, which most fre­quently emerge dur­ing the ado­les­cent pe­riod.

At around the age of 32 the strongest over­all shift in tra­jec­tory is seen. Life events such as par­ent­hood may play a role in some of the changes seen, al­though the re­search did not ex­plic­itly test this. We know that women who give birth, their brain changes af­ter­wards,” said Mousley. It’s rea­son­able to as­sume that there could be a re­la­tion­ship be­tween these mile­stones and what’s hap­pen­ing in the brain.”

From 32 years, the brain ar­chi­tec­ture ap­pears to sta­bilise com­pared with pre­vi­ous phases, cor­re­spond­ing with a plateau in in­tel­li­gence and per­son­al­ity” based on other stud­ies. Brain re­gions also be­come more com­part­men­talised.

The fi­nal two turn­ing points were de­fined by de­creases in brain con­nec­tiv­ity, which were be­lieved to be re­lated to age­ing and de­gen­er­a­tion of white mat­ter in the brain.

...

Read the original on www.theguardian.com »

6 278 shares, 16 trendiness

Frontier Visual Intelligence

FLUX.2 is de­signed for real-world cre­ative work­flows, not just demos or party tricks. It gen­er­ates high-qual­ity im­ages while main­tain­ing char­ac­ter and style con­sis­tency across mul­ti­ple ref­er­ence im­ages, fol­low­ing struc­tured prompts, read­ing and writ­ing com­plex text, ad­her­ing to brand guide­lines, and re­li­ably han­dling light­ing, lay­outs, and lo­gos. FLUX.2 can edit im­ages at up to 4 megapix­els while pre­serv­ing de­tail and co­her­ence.

We be­lieve vi­sual in­tel­li­gence should be shaped by re­searchers, cre­atives, and de­vel­op­ers every­where, not just a few. That’s why we pair fron­tier ca­pa­bil­ity with open re­search and open in­no­va­tion, re­leas­ing pow­er­ful, in­spectable, and com­pos­able open-weight mod­els for the com­mu­nity, along­side ro­bust, pro­duc­tion-ready end­points for teams that need scale, re­li­a­bil­ity, and cus­tomiza­tion.

When we launched Black Forest Labs in 2024, we set out to make open in­no­va­tion sus­tain­able, build­ing on our ex­pe­ri­ence de­vel­op­ing some of the world’s most pop­u­lar open mod­els. We’ve com­bined open mod­els like FLUX.1 [dev]—the most pop­u­lar open im­age model glob­ally—with pro­fes­sional-grade mod­els like FLUX.1 Kontext [pro], which pow­ers teams from Adobe to Meta and be­yond. Our open core ap­proach dri­ves ex­per­i­men­ta­tion, in­vites scrutiny, low­ers costs, and en­sures that we can keep shar­ing open tech­nol­ogy from the Black Forest and the Bay into the world.

Precision, ef­fi­ciency, con­trol, ex­treme re­al­ism - where FLUX.1 showed the po­ten­tial of me­dia mod­els as pow­er­ful cre­ative tools, FLUX.2 shows how fron­tier ca­pa­bil­ity can trans­form pro­duc­tion work­flows. By rad­i­cally chang­ing the eco­nom­ics of gen­er­a­tion, FLUX.2 will be­come an in­dis­pens­able part of our cre­ative in­fra­struc­ture.

Output Versatility: FLUX.2 is ca­pa­ble of gen­er­at­ing highly de­tailed, pho­to­real im­ages along with in­fo­graph­ics with com­plex ty­pog­ra­phy, all at res­o­lu­tions up to 4MP

Multi-Reference Support: Reference up to 10 im­ages si­mul­ta­ne­ously with the best char­ac­ter / prod­uct / style con­sis­tency avail­able to­day. Image Detail & Photorealism: Greater de­tail, sharper tex­tures, and more sta­ble light­ing suit­able for prod­uct shots, vi­su­al­iza­tion, and pho­tog­ra­phy-like use cases.Text Rendering: Complex ty­pog­ra­phy, in­fo­graph­ics, memes and UI mock­ups with leg­i­ble fine text now work re­li­ably in pro­duc­tion.En­hanced Prompt Following: Improved ad­her­ence to com­plex, struc­tured in­struc­tions, in­clud­ing multi-part prompts and com­po­si­tional con­straints.World Knowledge: Significantly more grounded in real-world knowl­edge, light­ing, and spa­tial logic, re­sult­ing in more co­her­ent scenes with ex­pected be­hav­ior.Higher Resolution & Flexible Input/Output Ratios: Image edit­ing on res­o­lu­tions up to 4MP.

All vari­ants of FLUX.2 of­fer im­age edit­ing from text and mul­ti­ple ref­er­ences in one model.

The FLUX.2 fam­ily cov­ers a spec­trum of model prod­ucts, from fully man­aged, pro­duc­tion-ready APIs to open-weight check­points de­vel­op­ers can run them­selves. The overview graph be­low shows how FLUX.2 [pro], FLUX.2 [flex], FLUX.2 [dev], and FLUX.2 [klein] bal­ance per­for­mance, and con­trol

State-of-the-art im­age qual­ity that ri­vals the best closed mod­els, match­ing other mod­els for prompt ad­her­ence and vi­sual fi­delity while gen­er­at­ing im­ages faster and at lower cost. No com­pro­mise be­tween speed and qual­ity. → Available now at BFL Playground , the BFL API and via our launch part­ners.: Take con­trol over model pa­ra­me­ters such as the num­ber of steps and the guid­ance scale, giv­ing de­vel­op­ers full con­trol over qual­ity, prompt ad­her­ence and speed. This model ex­cels at ren­der­ing text and fine de­tails. → Available now at bfl.ai/​play  , the BFL API and via our launch part­ners. 32B open-weight model, de­rived from the FLUX.2 base model. The most pow­er­ful open-weight im­age gen­er­a­tion and edit­ing model avail­able to­day, com­bin­ing text-to-im­age syn­the­sis and im­age edit­ing with mul­ti­ple in­put im­ages in a sin­gle check­point. FLUX.2 [dev] weights are avail­able on Hugging Face and can now be used lo­cally us­ing our ref­er­ence in­fer­ence code . On con­sumer grade GPUs like GeForce RTX GPUs you can use an op­ti­mized fp8 ref­er­ence im­ple­men­ta­tion of FLUX.2 [dev], cre­ated in col­lab­o­ra­tion with NVIDIA and ComfyUI . You can also sam­ple Flux.2 [dev] via API end­points on FAL , Replicate , Runware , Verda , TogetherAI , Cloudflare , DeepInfra . For a com­mer­cial li­cense, visit our web­site Open-source, Apache 2.0 model, size-dis­tilled from the FLUX.2 base model. More pow­er­ful & de­vel­oper-friendly than com­pa­ra­ble mod­els of the same size trained from scratch, with many of the same ca­pa­bil­i­ties as its teacher model. Join the beta The FLUX.2 - VAE is avail­able on HF un­der an Apache 2.0 li­cense A new vari­a­tional au­toen­coder for la­tent rep­re­sen­ta­tions that pro­vide an op­ti­mized trade-off be­tween learn­abil­ity, qual­ity and com­pres­sion rate. This model pro­vides the foun­da­tion for all FLUX.2 flow back­bones, and an in-depth re­port de­scrib­ing its tech­ni­cal prop­er­ties is avail­able here . The FLUX.2 - VAE is avail­able on HF un­der an Apache 2.0 li­cense

Generating de­signs with vari­able steps: FLUX.2 [flex] pro­vides a steps” pa­ra­me­ter, trad­ing off ty­pog­ra­phy ac­cu­racy and la­tency. From left to right: 6 steps, 20 steps, 50 steps.

Controlling im­age de­tail with vari­able steps: FLUX.2 [flex] pro­vides a steps” pa­ra­me­ter, trad­ing off im­age de­tail and la­tency. From left to right: 6 steps, 20 steps, 50 steps.

The FLUX.2 model fam­ily de­liv­ers state-of-the-art im­age gen­er­a­tion qual­ity at ex­tremely com­pet­i­tive prices, of­fer­ing the best value across per­for­mance tiers.

For open-weights im­age mod­els, FLUX.2 [dev] sets a new stan­dard, achiev­ing lead­ing per­for­mance across text-to-im­age gen­er­a­tion, sin­gle-ref­er­ence edit­ing, and multi-ref­er­ence edit­ing, con­sis­tently out­per­form­ing all open-weights al­ter­na­tives by a sig­nif­i­cant mar­gin.

Whether open or closed, we are com­mit­ted to the re­spon­si­ble de­vel­op­ment of these mod­els and ser­vices be­fore, dur­ing, and af­ter every re­lease.

FLUX.2 builds on a la­tent flow match­ing ar­chi­tec­ture, and com­bines im­age gen­er­a­tion and edit­ing in a sin­gle ar­chi­tec­ture. The model cou­ples the Mistral-3 24B pa­ra­me­ter vi­sion-lan­guage model with a rec­ti­fied flow trans­former. The VLM brings real world knowl­edge and con­tex­tual un­der­stand­ing, while the trans­former cap­tures spa­tial re­la­tion­ships, ma­te­r­ial prop­er­ties, and com­po­si­tional logic that ear­lier ar­chi­tec­tures could not ren­der.

FLUX.2 now pro­vides multi-ref­er­ence sup­port, with the abil­ity to com­bine up to 10 im­ages into a novel out­put, an out­put res­o­lu­tion of up to 4MP, sub­stan­tially bet­ter prompt ad­her­ence and world knowl­edge, and sig­nif­i­cantly im­proved ty­pog­ra­phy. We re-trained the mod­el’s la­tent space from scratch to achieve bet­ter learn­abil­ity and higher im­age qual­ity at the same time, a step to­wards solv­ing the Learnability-Quality-Compression” trilemma. Technical de­tails can be found in the FLUX.2 VAE blog post.

Into the New

We’re build­ing foun­da­tional in­fra­struc­ture for vi­sual in­tel­li­gence, tech­nol­ogy that trans­forms how the world is seen and un­der­stood. FLUX.2 is a step closer to mul­ti­modal mod­els that unify per­cep­tion, gen­er­a­tion, mem­ory, and rea­son­ing, in an open and trans­par­ent way.

Join us on this jour­ney. We’re hir­ing in Freiburg (HQ) and San Francisco. View open roles.

...

Read the original on bfl.ai »

7 275 shares, 19 trendiness

flowglad/flowglad: Open source payments + billing infrastructure

The eas­i­est way to make in­ter­net money.

Get Started

· Quickstart

· Website

· Issues

· Discord

* Default Stateless Say good­bye to web­hooks, subscriptions” db ta­bles, cus­tomer_id columns, PRICE_ID env vari­ables, or man­u­ally map­ping your plans to prices to fea­tures and back.

* Single Source of Truth: Read your lat­est cus­tomer billing state from Flowglad, in­clud­ing fea­ture ac­cess and us­age me­ter cred­its

* Access Data Using Your Ids: Query cus­tomer state by your au­th’s user ids. Refer to prices, fea­tures, and us­age me­ters via slugs you de­fine.

* Full-Stack SDK: Access your cus­tomer’s data on the back­end us­ing flowglad­Server.get­Billing(), or in your React fron­tend us­ing our use­Billing() hook

* Adaptable: Iterate on new pric­ing mod­els in test­mode, and push them to prod in a click. Seamlessly ro­tate pric­ing mod­els in your app with­out any re­de­ploy­ment.

First, in­stall the pack­ages nec­es­sary Flowglad pack­ages based on your pro­ject setup:

# Next.js Projects

bun add @flowglad/nextjs

# React + Express pro­jects:

bun add @flowglad/react @flowglad/express

# All other React + Node Projects

bun add @flowglad/react @flowglad/server

Flowglad in­te­grates seam­lessly with your au­then­ti­ca­tion sys­tem and re­quires only a few lines of code to get started in your Next.js app. Setup typ­i­cally takes un­der a minute:

Create a util­ity to gen­er­ate your Flowglad server in­stance. Pass your own cus­tomer/​user/​or­ga­ni­za­tion IDs—Flowglad never re­quires its own cus­tomer IDs to be man­aged in your app:

// utils/​flowglad.ts

im­port { FlowgladServer } from @flowglad/nextjs/server’

ex­port const flowglad = (customerExternalId: string) => {

re­turn new FlowgladServer({

cus­tomerEx­ter­nalId,

get­Cus­tomerDe­tails: async (externalId) => {

// e.g. Fetch user info from your DB us­ing your user/​org/​team ID

const user = await db.users.find­One({ id: ex­ter­nalId })

if (!user) throw new Error(‘User not found’)

re­turn { email: user.email, name: user.name }

Add an API route so the Flowglad client can com­mu­ni­cate se­curely with your back­end:

// app/​api/​flowglad/[…​path]/​route.ts

im­port { nex­tRoute­Han­dler } from @flowglad/nextjs/server’

im­port { flowglad } from @/utils/flowglad’

ex­port const { GET, POST } = nex­tRoute­Han­dler({

flowglad,

get­Cus­tomerEx­ter­nalId: async (req) => {

// Extract your user/​org/​team ID from ses­sion/​auth.

// For B2C: re­turn user.id from your DB

// For B2B: re­turn or­ga­ni­za­tion.id or team.id

const userId = await ge­tUserId­From­Re­quest(req)

if (!userId) throw new Error(‘User not au­then­ti­cat­ed’)

re­turn userId

Wrap Your App with the Provider

In your root lay­out (App Router) or _app (Pages Router):

im­port { FlowgladProvider } from @flowglad/nextjs’

// App Router ex­am­ple (app/layout.tsx)

ex­port de­fault func­tion RootLayout({ chil­dren }) {

re­turn (

That’s it—Flowglad will use your ap­p’s in­ter­nal user IDs for all billing logic and in­te­grate billing sta­tus into your fron­tend in real time.

B2C apps: Use user.id as the cus­tomer ID.

B2B apps: Use or­ga­ni­za­tion.id or team.id as the cus­tomer ID.

Flowglad does not re­quire you to change your au­then­ti­ca­tion sys­tem or man­age Flowglad cus­tomer IDs. Just pass your own!

Use use­Billing on your fron­tend, and flowglad(userId).get­Billing() on your back­end

use client’

im­port { use­Billing } from @flowglad/nextjs’

ex­port func­tion FeatureGate({ fea­tureS­lug, chil­dren }) {

const { loaded, er­rors, check­Fea­tureAc­cess } = use­Billing()

if (!loaded || !checkFeatureAccess) {

re­turn

if (errors?.length) {

re­turn

Unable to load billing data right now.

re­turn check­Fea­tureAc­cess(fea­tureS­lug)

? chil­dren

You need to up­grade to un­lock this fea­ture.

im­port { use­Billing } from @flowglad/nextjs’

ex­port func­tion UsageBalanceIndicator({ us­ageMe­ter­Slug }) {

const { loaded, er­rors, checkUsage­Bal­ance, cre­at­e­Check­out­Ses­sion } = use­Billing()

if (!loaded || !checkUsageBalance) {

re­turn

const us­age = checkUsage­Bal­ance(us­ageMe­ter­Slug)

re­turn (

im­port { NextResponse } from next/server’

im­port { flowglad } from @/utils/flowglad’

const has­Fast­Gen­er­a­tions = async () => {

const user = await ge­tUser()

const billing = await flowglad(user.id).get­Billing()

const hasAc­cess = billing.check­Fea­tureAc­cess(‘fast_­gen­er­a­tions’)

if (hasAccess) {

// run fast gen­er­a­tions

} else {

// fall back to nor­mal gen­er­a­tions

im­port { flowglad } from @/utils/flowglad’

const process­ChatMes­sage = async (params: { chat: string }) => {

// Extract your ap­p’s user/​org/​team ID,

// whichever cor­re­sponds to your cus­tomer

const user = await ge­tUser()

const billing = await flowglad(user.id).get­Billing()

const us­age = billing.checkUsage­Bal­ance(‘chat_mes­sages’)

if (usage.availableBalance > 0) {

// run chat re­quest

} else {

throw Error(`User ${user.id} does not have suf­fi­cient us­age cred­its`)

First, set up a pric­ing model. You can do so in the dash­board in just a few clicks us­ing a tem­plate, that you can then cus­tomize to suit your spe­cific needs.

We cur­rently have tem­plates for the fol­low­ing pric­ing mod­els:

And more on the way. If you don’t see a pric­ing model from our tem­plates that suits you, you can al­ways make one from scratch.

In the last 15 years, the mar­ket has given de­vel­op­ers more op­tions than ever for every sin­gle part of their stack. But when it comes to pay­ments, there have been vir­tu­ally zero new en­trants. The ex­ist­ing op­tions are slim, and al­most all of them re­quire us to talk to sales to even set up an ac­count. When it comes to self-serve pay­ments, there are even fewer op­tions.

The re­sult? The de­vel­oper ex­pe­ri­ence and cost of pay­ments has barely im­proved in that time. Best in class DX in pay­ments feels eerily sus­pended in 2015. Meanwhile, we’ve en­joyed con­stant im­prove­ments in auth, com­pute, host­ing, and prac­ti­cally every­thing else.

Flowglad wants to change that.

...

Read the original on github.com »

8 263 shares, 20 trendiness

Ilya Sutskever – We're moving from the age of scaling to the age of research

Ilya & I dis­cuss SSIs strat­egy, the prob­lems with pre-train­ing, how to im­prove the gen­er­al­iza­tion of AI mod­els, and how to en­sure AGI goes well.

Watch on YouTube; lis­ten on Apple Podcasts or Spotify.

* Gemini 3 is the first model I’ve used that can find con­nec­tions I haven’t an­tic­i­pated. I re­cently wrote a blog post on RLs in­for­ma­tion ef­fi­ciency, and Gemini 3 helped me think it all through. It also gen­er­ated the rel­e­vant charts and ran toy ML ex­per­i­ments for me with zero bugs. Try Gemini 3 to­day at gem­ini.google

* Labelbox helped me cre­ate a tool to tran­scribe our episodes! I’ve strug­gled with tran­scrip­tion in the past be­cause I don’t just want ver­ba­tim tran­scripts, I want tran­scripts re­worded to read like es­says. Labelbox helped me gen­er­ate the ex­act data I needed for this. If you want to learn how Labelbox can help you (or if you want to try out the tran­scriber tool your­self), go to la­bel­box.com/​dwarkesh

* Sardine is an AI risk man­age­ment plat­form that brings to­gether thou­sands of de­vice, be­hav­ior, and iden­tity sig­nals to help you as­sess a user’s risk of fraud & abuse. Sardine also of­fers a suite of agents to au­to­mate in­ves­ti­ga­tions so that as fraud­sters use AI to scale their at­tacks, you can use AI to scale your de­fenses. Learn more at sar­dine.ai/​dwarkesh

(00:18:49) — What are we scal­ing?

(00:25:13) — Why hu­mans gen­er­al­ize bet­ter than mod­els

(01:18:13) — “We are squarely an age of re­search com­pany”

You know what’s crazy? That all of this is real.

Don’t you think so? All this AI stuff and all this Bay Area… that it’s hap­pen­ing. Isn’t it straight out of sci­ence fic­tion?

Another thing that’s crazy is how nor­mal the slow take­off feels. The idea that we’d be in­vest­ing 1% of GDP in AI, I feel like it would have felt like a big­ger deal, whereas right now it just feels…

We get used to things pretty fast, it turns out. But also it’s kind of ab­stract. What does it mean? It means that you see it in the news, that such and such com­pany an­nounced such and such dol­lar amount. That’s all you see. It’s not re­ally felt in any other way so far.

Should we ac­tu­ally be­gin here? I think this is an in­ter­est­ing dis­cus­sion.

I think your point, about how from the av­er­age per­son’s point of view noth­ing is that dif­fer­ent, will con­tinue be­ing true even into the sin­gu­lar­ity.

No, I don’t think so.

The thing which I was re­fer­ring to not feel­ing dif­fer­ent is, okay, such and such com­pany an­nounced some dif­fi­cult-to-com­pre­hend dol­lar amount of in­vest­ment. I don’t think any­one knows what to do with that.

But I think the im­pact of AI is go­ing to be felt. AI is go­ing to be dif­fused through the econ­omy. There’ll be very strong eco­nomic forces for this, and I think the im­pact is go­ing to be felt very strongly.

When do you ex­pect that im­pact? I think the mod­els seem smarter than their eco­nomic im­pact would im­ply.

Yeah. This is one of the very con­fus­ing things about the mod­els right now. How to rec­on­cile the fact that they are do­ing so well on evals? You look at the evals and you go, Those are pretty hard evals.” They are do­ing so well. But the eco­nomic im­pact seems to be dra­mat­i­cally be­hind. It’s very dif­fi­cult to make sense of, how can the model, on the one hand, do these amaz­ing things, and then on the other hand, re­peat it­self twice in some sit­u­a­tion?

An ex­am­ple would be, let’s say you use vibe cod­ing to do some­thing. You go to some place and then you get a bug. Then you tell the model, Can you please fix the bug?” And the model says, Oh my God, you’re so right. I have a bug. Let me go fix that.” And it in­tro­duces a sec­ond bug. Then you tell it, You have this new sec­ond bug,” and it tells you, Oh my God, how could I have done it? You’re so right again,” and brings back the first bug, and you can al­ter­nate be­tween those. How is that pos­si­ble? I’m not sure, but it does sug­gest that some­thing strange is go­ing on.

I have two pos­si­ble ex­pla­na­tions. The more whim­si­cal ex­pla­na­tion is that maybe RL train­ing makes the mod­els a lit­tle too sin­gle-minded and nar­rowly fo­cused, a lit­tle bit too un­aware, even though it also makes them aware in some other ways. Because of this, they can’t do ba­sic things.

But there is an­other ex­pla­na­tion. Back when peo­ple were do­ing pre-train­ing, the ques­tion of what data to train on was an­swered, be­cause that an­swer was every­thing. When you do pre-train­ing, you need all the data. So you don’t have to think if it’s go­ing to be this data or that data.

But when peo­ple do RL train­ing, they do need to think. They say, Okay, we want to have this kind of RL train­ing for this thing and that kind of RL train­ing for that thing.” From what I hear, all the com­pa­nies have teams that just pro­duce new RL en­vi­ron­ments and just add it to the train­ing mix. The ques­tion is, well, what are those? There are so many de­grees of free­dom. There is such a huge va­ri­ety of RL en­vi­ron­ments you could pro­duce.

One thing you could do, and I think this is some­thing that is done in­ad­ver­tently, is that peo­ple take in­spi­ra­tion from the evals. You say, Hey, I would love our model to do re­ally well when we re­lease it. I want the evals to look great. What would be RL train­ing that could help on this task?” I think that is some­thing that hap­pens, and it could ex­plain a lot of what’s go­ing on.

If you com­bine this with gen­er­al­iza­tion of the mod­els ac­tu­ally be­ing in­ad­e­quate, that has the po­ten­tial to ex­plain a lot of what we are see­ing, this dis­con­nect be­tween eval per­for­mance and ac­tual real-world per­for­mance, which is some­thing that we don’t to­day even un­der­stand, what we mean by that.

I like this idea that the real re­ward hack­ing is the hu­man re­searchers who are too fo­cused on the evals.

I think there are two ways to un­der­stand, or to try to think about, what you have just pointed out. One is that if it’s the case that sim­ply by be­com­ing su­per­hu­man at a cod­ing com­pe­ti­tion, a model will not au­to­mat­i­cally be­come more taste­ful and ex­er­cise bet­ter judg­ment about how to im­prove your code­base, well then you should ex­pand the suite of en­vi­ron­ments such that you’re not just test­ing it on hav­ing the best per­for­mance in cod­ing com­pe­ti­tion. It should also be able to make the best kind of ap­pli­ca­tion for X thing or Y thing or Z thing.

Another, maybe this is what you’re hint­ing at, is to say, Why should it be the case in the first place that be­com­ing su­per­hu­man at cod­ing com­pe­ti­tions does­n’t make you a more taste­ful pro­gram­mer more gen­er­ally?” Maybe the thing to do is not to keep stack­ing up the amount and di­ver­sity of en­vi­ron­ments, but to fig­ure out an ap­proach which lets you learn from one en­vi­ron­ment and im­prove your per­for­mance on some­thing else.

I have a hu­man anal­ogy which might be help­ful. Let’s take the case of com­pet­i­tive pro­gram­ming, since you men­tioned that. Suppose you have two stu­dents. One of them de­cided they want to be the best com­pet­i­tive pro­gram­mer, so they will prac­tice 10,000 hours for that do­main. They will solve all the prob­lems, mem­o­rize all the proof tech­niques, and be very skilled at quickly and cor­rectly im­ple­ment­ing all the al­go­rithms. By do­ing so, they be­came one of the best.

Student num­ber two thought, Oh, com­pet­i­tive pro­gram­ming is cool.” Maybe they prac­ticed for 100 hours, much less, and they also did re­ally well. Which one do you think is go­ing to do bet­ter in their ca­reer later on?

Right. I think that’s ba­si­cally what’s go­ing on. The mod­els are much more like the first stu­dent, but even more. Because then we say, the model should be good at com­pet­i­tive pro­gram­ming so let’s get every sin­gle com­pet­i­tive pro­gram­ming prob­lem ever. And then let’s do some data aug­men­ta­tion so we have even more com­pet­i­tive pro­gram­ming prob­lems, and we train on that. Now you’ve got this great com­pet­i­tive pro­gram­mer.

With this anal­ogy, I think it’s more in­tu­itive. Yeah, okay, if it’s so well trained, all the dif­fer­ent al­go­rithms and all the dif­fer­ent proof tech­niques are right at its fin­ger­tips. And it’s more in­tu­itive that with this level of prepa­ra­tion, it would not nec­es­sar­ily gen­er­al­ize to other things.

But then what is the anal­ogy for what the sec­ond stu­dent is do­ing be­fore they do the 100 hours of fine-tun­ing?

I think they have it.” The it” fac­tor. When I was an un­der­grad, I re­mem­ber there was a stu­dent like this that stud­ied with me, so I know it ex­ists.

I think it’s in­ter­est­ing to dis­tin­guish it” from what­ever pre-train­ing does. One way to un­der­stand what you just said about not hav­ing to choose the data in pre-train­ing is to say it’s ac­tu­ally not dis­sim­i­lar to the 10,000 hours of prac­tice. It’s just that you get that 10,000 hours of prac­tice for free be­cause it’s al­ready some­where in the pre-train­ing dis­tri­b­u­tion. But maybe you’re sug­gest­ing there’s ac­tu­ally not that much gen­er­al­iza­tion from pre-train­ing. There’s just so much data in pre-train­ing, but it’s not nec­es­sar­ily gen­er­al­iz­ing bet­ter than RL.

The main strength of pre-train­ing is that: A, there is so much of it, and B, you don’t have to think hard about what data to put into pre-train­ing. It’s very nat­ural data, and it does in­clude in it a lot of what peo­ple do: peo­ple’s thoughts and a lot of the fea­tures. It’s like the whole world as pro­jected by peo­ple onto text, and pre-train­ing tries to cap­ture that us­ing a huge amount of data.

Pre-training is very dif­fi­cult to rea­son about be­cause it’s so hard to un­der­stand the man­ner in which the model re­lies on pre-train­ing data. Whenever the model makes a mis­take, could it be be­cause some­thing by chance is not as sup­ported by the pre-train­ing data? Support by pre-train­ing” is maybe a loose term. I don’t know if I can add any­thing more use­ful on this. I don’t think there is a hu­man ana­log to pre-train­ing.

Here are analo­gies that peo­ple have pro­posed for what the hu­man anal­ogy to pre-train­ing is. I’m cu­ri­ous to get your thoughts on why they’re po­ten­tially wrong. One is to think about the first 18, or 15, or 13 years of a per­son’s life when they aren’t nec­es­sar­ily eco­nom­i­cally pro­duc­tive, but they are do­ing some­thing that is mak­ing them un­der­stand the world bet­ter and so forth. The other is to think about evo­lu­tion as do­ing some kind of search for 3 bil­lion years, which then re­sults in a hu­man life­time in­stance.

I’m cu­ri­ous if you think ei­ther of these are anal­o­gous to pre-train­ing. How would you think about what life­time hu­man learn­ing is like, if not pre-train­ing?

I think there are some sim­i­lar­i­ties be­tween both of these and pre-train­ing, and pre-train­ing tries to play the role of both of these. But I think there are some big dif­fer­ences as well. The amount of pre-train­ing data is very, very stag­ger­ing.

Somehow a hu­man be­ing, af­ter even 15 years with a tiny frac­tion of the pre-train­ing data, they know much less. But what­ever they do know, they know much more deeply some­how. Already at that age, you would not make mis­takes that our AIs make.

There is an­other thing. You might say, could it be some­thing like evo­lu­tion? The an­swer is maybe. But in this case, I think evo­lu­tion might ac­tu­ally have an edge. I re­mem­ber read­ing about this case. One way in which neu­ro­sci­en­tists can learn about the brain is by study­ing peo­ple with brain dam­age to dif­fer­ent parts of the brain. Some peo­ple have the most strange symp­toms you could imag­ine. It’s ac­tu­ally re­ally, re­ally in­ter­est­ing.

One case that comes to mind that’s rel­e­vant. I read about this per­son who had some kind of brain dam­age, a stroke or an ac­ci­dent, that took out his emo­tional pro­cess­ing. So he stopped feel­ing any emo­tion. He still re­mained very ar­tic­u­late and he could solve lit­tle puz­zles, and on tests he seemed to be just fine. But he felt no emo­tion. He did­n’t feel sad, he did­n’t feel anger, he did­n’t feel an­i­mated. He be­came some­how ex­tremely bad at mak­ing any de­ci­sions at all. It would take him hours to de­cide on which socks to wear. He would make very bad fi­nan­cial de­ci­sions.

What does it say about the role of our built-in emo­tions in mak­ing us a vi­able agent, es­sen­tially? To con­nect to your ques­tion about pre-train­ing, maybe if you are good enough at get­ting every­thing out of pre-train­ing, you could get that as well. But that’s the kind of thing which seems… Well, it may or may not be pos­si­ble to get that from pre-train­ing.

What is that”? Clearly not just di­rectly emo­tion. It seems like some al­most value func­tion-like thing which is telling you what the end re­ward for any de­ci­sion should be. You think that does­n’t sort of im­plic­itly come from pre-train­ing?

I think it could. I’m just say­ing it’s not 100% ob­vi­ous.

But what is that? How do you think about emo­tions? What is the ML anal­ogy for emo­tions?

It should be some kind of a value func­tion thing. But I don’t think there is a great ML anal­ogy be­cause right now, value func­tions don’t play a very promi­nent role in the things peo­ple do.

It might be worth defin­ing for the au­di­ence what a value func­tion is, if you want to do that.

Certainly, I’ll be very happy to do that. When peo­ple do re­in­force­ment learn­ing, the way re­in­force­ment learn­ing is done right now, how do peo­ple train those agents? You have your neural net and you give it a prob­lem, and then you tell the model, Go solve it.” The model takes maybe thou­sands, hun­dreds of thou­sands of ac­tions or thoughts or some­thing, and then it pro­duces a so­lu­tion. The so­lu­tion is graded.

And then the score is used to pro­vide a train­ing sig­nal for every sin­gle ac­tion in your tra­jec­tory. That means that if you are do­ing some­thing that goes for a long time—if you’re train­ing a task that takes a long time to solve—it will do no learn­ing at all un­til you come up with the pro­posed so­lu­tion. That’s how re­in­force­ment learn­ing is done naively. That’s how o1, R1 os­ten­si­bly are done.

The value func­tion says some­thing like, Maybe I could some­times, not al­ways, tell you if you are do­ing well or badly.” The no­tion of a value func­tion is more use­ful in some do­mains than oth­ers. For ex­am­ple, when you play chess and you lose a piece, I messed up. You don’t need to play the whole game to know that what I just did was bad, and there­fore what­ever pre­ceded it was also bad.

The value func­tion lets you short-cir­cuit the wait un­til the very end. Let’s sup­pose that you are do­ing some kind of a math thing or a pro­gram­ming thing, and you’re try­ing to ex­plore a par­tic­u­lar so­lu­tion or di­rec­tion. After, let’s say, a thou­sand steps of think­ing, you con­cluded that this di­rec­tion is un­promis­ing. As soon as you con­clude this, you could al­ready get a re­ward sig­nal a thou­sand timesteps pre­vi­ously, when you de­cided to pur­sue down this path. You say, Next time I should­n’t pur­sue this path in a sim­i­lar sit­u­a­tion,” long be­fore you ac­tu­ally came up with the pro­posed so­lu­tion.

This was in the DeepSeek R1 pa­per— that the space of tra­jec­to­ries is so wide that maybe it’s hard to learn a map­ping from an in­ter­me­di­ate tra­jec­tory and value. And also given that, in cod­ing for ex­am­ple you’ll have the wrong idea, then you’ll go back, then you’ll change some­thing.

This sounds like such a lack of faith in deep learn­ing. Sure it might be dif­fi­cult, but noth­ing deep learn­ing can’t do. My ex­pec­ta­tion is that a value func­tion should be use­ful, and I fully ex­pect that they will be used in the fu­ture, if not al­ready.

What I was al­lud­ing to with the per­son whose emo­tional cen­ter got dam­aged, it’s more that maybe what it sug­gests is that the value func­tion of hu­mans is mod­u­lated by emo­tions in some im­por­tant way that’s hard­coded by evo­lu­tion. And maybe that is im­por­tant for peo­ple to be ef­fec­tive in the world.

That’s the thing I was plan­ning on ask­ing you. There’s some­thing re­ally in­ter­est­ing about emo­tions of the value func­tion, which is that it’s im­pres­sive that they have this much util­ity while still be­ing rather sim­ple to un­der­stand.

I have two re­sponses. I do agree that com­pared to the kind of things that we learn and the things we are talk­ing about, the kind of AI we are talk­ing about, emo­tions are rel­a­tively sim­ple. They might even be so sim­ple that maybe you could map them out in a hu­man-un­der­stand­able way. I think it would be cool to do.

In terms of util­ity though, I think there is a thing where there is this com­plex­ity-ro­bust­ness trade­off, where com­plex things can be very use­ful, but sim­ple things are very use­ful in a very broad range of sit­u­a­tions. One way to in­ter­pret what we are see­ing is that we’ve got these emo­tions that evolved mostly from our mam­mal an­ces­tors and then fine-tuned a lit­tle bit while we were ho­minids, just a bit. We do have a de­cent amount of so­cial emo­tions though which mam­mals may lack. But they’re not very so­phis­ti­cated. And be­cause they’re not so­phis­ti­cated, they serve us so well in this very dif­fer­ent world com­pared to the one that we’ve been liv­ing in.

Actually, they also make mis­takes. For ex­am­ple, our emo­tions… Well ac­tu­ally, I don’t know. Does hunger count as an emo­tion? It’s de­bat­able. But I think, for ex­am­ple, our in­tu­itive feel­ing of hunger is not suc­ceed­ing in guid­ing us cor­rectly in this world with an abun­dance of food.

People have been talk­ing about scal­ing data, scal­ing pa­ra­me­ters, scal­ing com­pute. Is there a more gen­eral way to think about scal­ing? What are the other scal­ing axes?

Here’s a per­spec­tive that I think might be true. The way ML used to work is that peo­ple would just tin­ker with stuff and try to get in­ter­est­ing re­sults. That’s what’s been go­ing on in the past.

Then the scal­ing in­sight ar­rived. Scaling laws, GPT-3, and sud­denly every­one re­al­ized we should scale. This is an ex­am­ple of how lan­guage af­fects thought. Scaling” is just one word, but it’s such a pow­er­ful word be­cause it in­forms peo­ple what to do. They say, Let’s try to scale things.” So you say, what are we scal­ing? Pre-training was the thing to scale. It was a par­tic­u­lar scal­ing recipe.

The big break­through of pre-train­ing is the re­al­iza­tion that this recipe is good. You say, Hey, if you mix some com­pute with some data into a neural net of a cer­tain size, you will get re­sults. You will know that you’ll be bet­ter if you just scale the recipe up.” This is also great. Companies love this be­cause it gives you a very low-risk way of in­vest­ing your re­sources.

It’s much harder to in­vest your re­sources in re­search. Compare that. If you re­search, you need to be like, Go forth re­searchers and re­search and come up with some­thing”, ver­sus get more data, get more com­pute. You know you’ll get some­thing from pre-train­ing.

Indeed, it looks like, based on var­i­ous things some peo­ple say on Twitter, maybe it ap­pears that Gemini have found a way to get more out of pre-train­ing. At some point though, pre-train­ing will run out of data. The data is very clearly fi­nite. What do you do next? Either you do some kind of souped-up pre-train­ing, a dif­fer­ent recipe from the one you’ve done be­fore, or you’re do­ing RL, or maybe some­thing else. But now that com­pute is big, com­pute is now very big, in some sense we are back to the age of re­search.

Maybe here’s an­other way to put it. Up un­til 2020, from 2012 to 2020, it was the age of re­search. Now, from 2020 to 2025, it was the age of scal­ing—maybe plus or mi­nus, let’s add er­ror bars to those years—be­cause peo­ple say, This is amaz­ing. You’ve got to scale more. Keep scal­ing.” The one word: scal­ing.

But now the scale is so big. Is the be­lief re­ally, Oh, it’s so big, but if you had 100x more, every­thing would be so dif­fer­ent?” It would be dif­fer­ent, for sure. But is the be­lief that if you just 100x the scale, every­thing would be trans­formed? I don’t think that’s true. So it’s back to the age of re­search again, just with big com­put­ers.

That’s a very in­ter­est­ing way to put it. But let me ask you the ques­tion you just posed then. What are we scal­ing, and what would it mean to have a recipe? I guess I’m not aware of a very clean re­la­tion­ship that al­most looks like a law of physics which ex­isted in pre-train­ing. There was a power law be­tween data or com­pute or pa­ra­me­ters and loss. What is the kind of re­la­tion­ship we should be seek­ing, and how should we think about what this new recipe might look like?

We’ve al­ready wit­nessed a tran­si­tion from one type of scal­ing to a dif­fer­ent type of scal­ing, from pre-train­ing to RL. Now peo­ple are scal­ing RL. Now based on what peo­ple say on Twitter, they spend more com­pute on RL than on pre-train­ing at this point, be­cause RL can ac­tu­ally con­sume quite a bit of com­pute. You do very long roll­outs, so it takes a lot of com­pute to pro­duce those roll­outs. Then you get a rel­a­tively small amount of learn­ing per roll­out, so you re­ally can spend a lot of com­pute.

I would­n’t even call it scal­ing. I would say, Hey, what are you do­ing? Is the thing you are do­ing the most pro­duc­tive thing you could be do­ing? Can you find a more pro­duc­tive way of us­ing your com­pute?” We’ve dis­cussed the value func­tion busi­ness ear­lier. Maybe once peo­ple get good at value func­tions, they will be us­ing their re­sources more pro­duc­tively. If you find a whole other way of train­ing mod­els, you could say, Is this scal­ing or is it just us­ing your re­sources?” I think it be­comes a lit­tle bit am­bigu­ous.

In the sense that, when peo­ple were in the age of re­search back then, it was, Let’s try this and this and this. Let’s try that and that and that. Oh, look, some­thing in­ter­est­ing is hap­pen­ing.” I think there will be a re­turn to that.

If we’re back in the era of re­search, step­ping back, what is the part of the recipe that we need to think most about? When you say value func­tion, peo­ple are al­ready try­ing the cur­rent recipe, but then hav­ing LLM-as-a-Judge and so forth. You could say that’s a value func­tion, but it sounds like you have some­thing much more fun­da­men­tal in mind. Should we even re­think pre-train­ing at all and not just add more steps to the end of that process?

The dis­cus­sion about value func­tion, I think it was in­ter­est­ing. I want to em­pha­size that I think the value func­tion is some­thing that’s go­ing to make RL more ef­fi­cient, and I think that makes a dif­fer­ence. But I think any­thing you can do with a value func­tion, you can do with­out, just more slowly. The thing which I think is the most fun­da­men­tal is that these mod­els some­how just gen­er­al­ize dra­mat­i­cally worse than peo­ple. It’s su­per ob­vi­ous. That seems like a very fun­da­men­tal thing.

So this is the crux: gen­er­al­iza­tion. There are two sub-ques­tions. There’s one which is about sam­ple ef­fi­ciency: why should it take so much more data for these mod­els to learn than hu­mans? There’s a sec­ond ques­tion. Even sep­a­rate from the amount of data it takes, why is it so hard to teach the thing we want to a model than to a hu­man? For a hu­man, we don’t nec­es­sar­ily need a ver­i­fi­able re­ward to be able to… You’re prob­a­bly men­tor­ing a bunch of re­searchers right now, and you’re talk­ing with them, you’re show­ing them your code, and you’re show­ing them how you think. From that, they’re pick­ing up your way of think­ing and how they should do re­search.

You don’t have to set a ver­i­fi­able re­ward for them that’s like, Okay, this is the next part of the cur­ricu­lum, and now this is the next part of your cur­ricu­lum. Oh, this train­ing was un­sta­ble.” There’s not this schleppy, be­spoke process. Perhaps these two is­sues are ac­tu­ally re­lated in some way, but I’d be cu­ri­ous to ex­plore this sec­ond thing, which is more like con­tin­ual learn­ing, and this first thing, which feels just like sam­ple ef­fi­ciency.

You could ac­tu­ally won­der that one pos­si­ble ex­pla­na­tion for the hu­man sam­ple ef­fi­ciency that needs to be con­sid­ered is evo­lu­tion. Evolution has given us a small amount of the most use­ful in­for­ma­tion pos­si­ble. For things like vi­sion, hear­ing, and lo­co­mo­tion, I think there’s a pretty strong case that evo­lu­tion has given us a lot.

For ex­am­ple, hu­man dex­ter­ity far ex­ceeds… I mean ro­bots can be­come dex­ter­ous too if you sub­ject them to a huge amount of train­ing in sim­u­la­tion. But to train a ro­bot in the real world to quickly pick up a new skill like a per­son does seems very out of reach. Here you could say, Oh yeah, lo­co­mo­tion. All our an­ces­tors needed great lo­co­mo­tion, squir­rels. So with lo­co­mo­tion, maybe we’ve got some un­be­liev­able prior.”

You could make the same case for vi­sion. I be­lieve Yann LeCun made the point that chil­dren learn to drive af­ter 10 hours of prac­tice, which is true. But our vi­sion is so good. At least for me, I re­mem­ber my­self be­ing a five-year-old. I was very ex­cited about cars back then. I’m pretty sure my car recog­ni­tion was more than ad­e­quate for dri­ving al­ready as a five-year-old. You don’t get to see that much data as a five-year-old. You spend most of your time in your par­ents’ house, so you have very low data di­ver­sity.

But you could say maybe that’s evo­lu­tion too. But in lan­guage and math and cod­ing, prob­a­bly not.

It still seems bet­ter than mod­els. Obviously, mod­els are bet­ter than the av­er­age hu­man at lan­guage, math, and cod­ing. But are they bet­ter than the av­er­age hu­man at learn­ing?

Oh yeah. Oh yeah, ab­solutely. What I meant to say is that lan­guage, math, and cod­ing—and es­pe­cially math and cod­ing—sug­gests that what­ever it is that makes peo­ple good at learn­ing is prob­a­bly not so much a com­pli­cated prior, but some­thing more, some fun­da­men­tal thing.

I’m not sure I un­der­stood. Why should that be the case?

So con­sider a skill in which peo­ple ex­hibit some kind of great re­li­a­bil­ity. If the skill is one that was very use­ful to our an­ces­tors for many mil­lions of years, hun­dreds of mil­lions of years, you could ar­gue that maybe hu­mans are good at it be­cause of evo­lu­tion, be­cause we have a prior, an evo­lu­tion­ary prior that’s en­coded in some very non-ob­vi­ous way that some­how makes us so good at it.

But if peo­ple ex­hibit great abil­ity, re­li­a­bil­ity, ro­bust­ness, and abil­ity to learn in a do­main that re­ally did not ex­ist un­til re­cently, then this is more an in­di­ca­tion that peo­ple might have just bet­ter ma­chine learn­ing, pe­riod.

How should we think about what that is? What is the ML anal­ogy? There are a cou­ple of in­ter­est­ing things about it. It takes fewer sam­ples. It’s more un­su­per­vised. A child learn­ing to drive a car… Children are not learn­ing to drive a car. A teenager learn­ing how to drive a car is not ex­actly get­ting some pre­built, ver­i­fi­able re­ward. It comes from their in­ter­ac­tion with the ma­chine and with the en­vi­ron­ment. It takes much fewer sam­ples. It seems more un­su­per­vised. It seems more ro­bust?

Much more ro­bust. The ro­bust­ness of peo­ple is re­ally stag­ger­ing.

Do you have a uni­fied way of think­ing about why all these things are hap­pen­ing at once? What is the ML anal­ogy that could re­al­ize some­thing like this?

One of the things that you’ve been ask­ing about is how can the teenage dri­ver self-cor­rect and learn from their ex­pe­ri­ence with­out an ex­ter­nal teacher? The an­swer is that they have their value func­tion. They have a gen­eral sense which is also, by the way, ex­tremely ro­bust in peo­ple. Whatever the hu­man value func­tion is, with a few ex­cep­tions around ad­dic­tion, it’s ac­tu­ally very, very ro­bust.

So for some­thing like a teenager that’s learn­ing to drive, they start to drive, and they al­ready have a sense of how they’re dri­ving im­me­di­ately, how badly they are, how un­con­fi­dent. And then they see, Okay.” And then, of course, the learn­ing speed of any teenager is so fast. After 10 hours, you’re good to go.

It seems like hu­mans have some so­lu­tion, but I’m cu­ri­ous about how they are do­ing it and why is it so hard? How do we need to recon­cep­tu­al­ize the way we’re train­ing mod­els to make some­thing like this pos­si­ble?

That is a great ques­tion to ask, and it’s a ques­tion I have a lot of opin­ions about. But un­for­tu­nately, we live in a world where not all ma­chine learn­ing ideas are dis­cussed freely, and this is one of them. There’s prob­a­bly a way to do it. I think it can be done. The fact that peo­ple are like that, I think it’s a proof that it can be done.

There may be an­other blocker though, which is that there is a pos­si­bil­ity that the hu­man neu­rons do more com­pute than we think. If that is true, and if that plays an im­por­tant role, then things might be more dif­fi­cult. But re­gard­less, I do think it points to the ex­is­tence of some ma­chine learn­ing prin­ci­ple that I have opin­ions on. But un­for­tu­nately, cir­cum­stances make it hard to dis­cuss in de­tail.

Nobody lis­tens to this pod­cast, Ilya.

I’m cu­ri­ous. If you say we are back in an era of re­search, you were there from 2012 to 2020. What is the vibe now go­ing to be if we go back to the era of re­search?

For ex­am­ple, even af­ter AlexNet, the amount of com­pute that was used to run ex­per­i­ments kept in­creas­ing, and the size of fron­tier sys­tems kept in­creas­ing. Do you think now that this era of re­search will still re­quire tremen­dous amounts of com­pute? Do you think it will re­quire go­ing back into the archives and read­ing old pa­pers?

You were at Google and OpenAI and Stanford, these places, when there was more of a vibe of re­search? What kind of things should we be ex­pect­ing in the com­mu­nity?

One con­se­quence of the age of scal­ing is that scal­ing sucked out all the air in the room. Because scal­ing sucked out all the air in the room, every­one started to do the same thing. We got to the point where we are in a world where there are more com­pa­nies than ideas by quite a bit. Actually on that, there is this Silicon Valley say­ing that says that ideas are cheap, ex­e­cu­tion is every­thing. People say that a lot, and there is truth to that. But then I saw some­one say on Twitter some­thing like, If ideas are so cheap, how come no one’s hav­ing any ideas?” And I think it’s true too.

If you think about re­search progress in terms of bot­tle­necks, there are sev­eral bot­tle­necks. One of them is ideas, and one of them is your abil­ity to bring them to life, which might be com­pute but also en­gi­neer­ing. If you go back to the 90s, let’s say, you had peo­ple who had pretty good ideas, and if they had much larger com­put­ers, maybe they could demon­strate that their ideas were vi­able. But they could not, so they could only have a very, very small demon­stra­tion that did not con­vince any­one. So the bot­tle­neck was com­pute.

Then in the age of scal­ing, com­pute has in­creased a lot. Of course, there is a ques­tion of how much com­pute is needed, but com­pute is large. Compute is large enough such that it’s not ob­vi­ous that you need that much more com­pute to prove some idea. I’ll give you an anal­ogy. AlexNet was built on two GPUs. That was the to­tal amount of com­pute used for it. The trans­former was built on 8 to 64 GPUs. No sin­gle trans­former pa­per ex­per­i­ment used more than 64 GPUs of 2017, which would be like, what, two GPUs of to­day? The ResNet, right? You could ar­gue that the o1 rea­son­ing was not the most com­pute-heavy thing in the world.

...

Read the original on www.dwarkesh.com »

9 247 shares, 10 trendiness

APT Rust requirement raises questions

The fol­low­ing sub­scrip­tion-only con­tent has been made avail­able to you by an LWN sub­scriber. Thousands of sub­scribers de­pend on LWN for the best news from the Linux and free soft­ware com­mu­ni­ties. If you en­joy this ar­ti­cle, please con­sider sub­scrib­ing to LWN. Thank you for vis­it­ing LWN.net!

It is rarely news­wor­thy when a pro­ject or pack­age picks up a new de­pen­dency. However, changes in a core tool like Debian’s Advanced Package

Tool (APT) can have far-reach­ing ef­fects. For ex­am­ple, Julian Andres Klode’s de­c­la­ra­tion

that APT would re­quire Rust in May 2026 means that a few of Debian’s un­of­fi­cial ports must ei­ther ac­quire a work­ing Rust tool­chain or de­pend on an old ver­sion of APT. This has raised sev­eral ques­tions within the pro­ject, par­tic­u­larly about the abil­ity of a sin­gle main­tainer to make changes that have wide­spread im­pact.

On October 31, Klode sent an an­nounce­ment to the de­bian-de­vel mail­ing list that he in­tended to in­tro­duce Rust de­pen­den­cies and code into APT as soon as May 2026:

This ex­tends at first to the Rust com­piler and stan­dard li­brary, and the Sequoia ecosys­tem.

In par­tic­u­lar, our code to parse .deb, .ar, .tar, and the HTTP sig­na­ture ver­i­fi­ca­tion code would strongly ben­e­fit from mem­ory safe lan­guages and a stronger ap­proach to unit test­ing.

If you main­tain a port with­out a work­ing Rust tool­chain, please en­sure it has one within the next 6 months, or sun­set the port.

Klode added this was nec­es­sary so that the pro­ject as a whole could move for­ward, rely on mod­ern tech­nolo­gies, and not be held back by

try­ing to shoe­horn mod­ern soft­ware on retro com­put­ing

de­vices”. Some Debian de­vel­op­ers have wel­comed the news. Paul Tagliamonte ac­knowl­edged

that it would im­pact un­of­fi­cial Debian ports but called the push to­ward Rust ”.

However, John Paul Adrian Glaubitz com­plained

that Klode’s word­ing was un­pleas­ant and that the ap­proach was con­fronta­tional. In an­other

mes­sage, he ex­plained that he was not against adop­tion of Rust; he had worked on en­abling Rust on many of the Debian ar­chi­tec­tures and helped to fix ar­chi­tec­ture-spe­cific bugs in the Rust tool­chain as well as LLVM up­stream. However, the mes­sage strongly sug­gested there was no room for a change in plan: Klode had ended his mes­sage with ”, which in­vited no fur­ther dis­cus­sion. Glaubitz was one of a few Debian de­vel­op­ers who ex­pressed dis­com­fort with Klode’s com­mu­ni­ca­tion style in the mes­sage.

Klode noted, briefly, that Rust was al­ready a hard re­quire­ment for all Debian re­lease ar­chi­tec­tures and ports, ex­cept for Alpha (alpha), Motorola 680x0 (m68k),

PA-RISC (hppa), and

SuperH (sh4), be­cause of APTs use of the Sequoia-PGP

pro­jec­t’s tool to ver­ify OpenPGP

sig­na­tures. APT falls back to us­ing the GNU Privacy Guard sig­na­ture-ver­i­fi­ca­tion tool, , on ports that do not have a Rust com­piler. By de­pend­ing di­rectly on Rust, though, APT it­self would not be avail­able on ports with­out a Rust com­piler. LWN re­cently

cov­ered the state of Linux ar­chi­tec­ture sup­port, and the sta­tus of Rust sup­port for each one.

None of the ports listed by Klode are among those of­fi­cially

sup­ported by Debian to­day, or tar­geted for sup­port in Debian 14 (“forky”). The sh4 port has never been of­fi­cially sup­ported, and none of the other ports have been sup­ported since Debian 6.0. The ac­tual im­pact on the ports lack­ing Rust is also less dra­matic than it sounded at first. Glaubitz as­sured

Antoni Boucher that ”, but phras­ing it that way gets more at­ten­tion in the

news”. Boucher is the main­tainer of , a GCC

ahead-of-time code gen­er­a­tor for Rust. Nothing, Glaubitz said, stops ports from us­ing a non-Rust ver­sion of APT un­til Boucher and oth­ers man­age to boot­strap Rust for those ports.

David Kalnischkies, who is also a ma­jor

con­trib­u­tor to APT, sug­gested

that if the goal is to re­duce bugs, it would be bet­ter to re­move the code that is used to parse the .deb, .ar, and .tar for­mats that Klode men­tioned from APT en­tirely. It is only needed for two tools,

and , he said, and the only ” of

was by Klode’s em­ployer, Canonical, for its Launchpad soft­ware-col­lab­o­ra­tion plat­form. If those were taken out of the main APT code base, then it would not mat­ter whether they were writ­ten in Rust, Python, or an­other lan­guage, since the tools are not di­rectly nec­es­sary for any given port.

Kalnischkies also ques­tioned the claim that Rust was nec­es­sary to achieve the stronger ap­proach to unit test­ing that Klode men­tioned:

You can cer­tainly do unit tests in C++, we do. The main prob­lem is that some­one has to write those tests. Like docs.

Your new solver e.g. has none (apart from our pre­ex­ist­ing in­te­gra­tion tests). You don’t se­ri­ously claim that is be­cause of C++ ? If you don’t like GoogleTest, which is what we cur­rently have, I could sug­gest doctest (as I did in pre­vi­ous in­stall­ments). Plenty other frame­works ex­ist with sim­i­lar or dif­fer­ent styles.

Klode has not re­sponded to those com­ments yet, which is a bit un­for­tu­nate given the fact that in­tro­duc­ing hard de­pen­den­cies on Rust has an im­pact be­yond his own work on APT. It may well be that he has good an­swers to the ques­tions, but it can also give the im­pres­sion that Klode is sim­ply em­brac­ing a trend to­ward Rust. He is in­volved

in the Ubuntu work to mi­grate from GNU Coreutils to the Rust-based uu­tils. The rea­sons given for that work, again, are around mod­ern­iza­tion and bet­ter se­cu­rity—but se­cu­rity is not au­to­mat­i­cally guar­an­teed sim­ply by switch­ing to Rust, and there are a num­ber of other con­sid­er­a­tions.

For ex­am­ple, Adrian Bunk pointed

out that there are a num­ber of Debian teams, as well as tool­ing, that will be im­pacted by writ­ing some of APT in Rust. The re­lease notes for Debian 13 (“trixie”) men­tion

that Debian’s in­fra­struc­ture currently has prob­lems with

re­build­ing pack­ages of types that sys­tem­at­i­cally use sta­tic

link­ing”, such as those with code writ­ten in Go and Rust. Thus, these pack­ages will be

cov­ered by lim­ited se­cu­rity sup­port un­til the in­fra­struc­ture is

im­proved to deal with them main­tain­ably”. Limited se­cu­rity sup­port means that up­dates to Rust li­braries are likely to only be re­leased when Debian pub­lishes a point re­lease, which hap­pens about every two months. The se­cu­rity team has specif­i­cally

stated that is fully sup­ported, but there are still out­stand­ing prob­lems.

Due to the sta­tic-link­ing is­sue, any time one of s de­pen­den­cies, cur­rently more than 40 Rust crates, have to be re­built due to a se­cu­rity is­sue, (at least po­ten­tially) also needs to be re­built. There are also dif­fi­cul­ties in track­ing CVEs for all of its de­pen­den­cies, and un­der­stand­ing when a se­cu­rity vul­ner­a­bil­ity in a Rust crate may re­quire up­dat­ing a Rust pro­gram that de­pends on it.

Fabian Grünbichler, a main­tainer of Debian’s Rust tool­chain, listed

sev­eral out­stand­ing prob­lems Debian has with deal­ing with Rust pack­ages. One of the largest is the need for a con­sis­tent Debian pol­icy for de­clar­ing sta­t­i­cally linked li­braries. In 2022, Guillem Jover added a con­trol field for Debian pack­ages called Static-Built-Using (SBU), which would list the source pack­ages used to build a bi­nary pack­age. This would in­di­cate when a bi­nary pack­age needs to be re­built due to an up­date in an­other source pack­age. For ex­am­ple, de­pends on more than 40 Rust crates that are pack­aged for Debian. Without de­clar­ing the SBUs, it may not be clear if needs to be up­dated when one of its de­pen­den­cies is up­dated. Debian has been work­ing on a pol­icy

re­quire­ment for SBU since April 2024, but it is not yet fin­ished or adopted.

The dis­cus­sion sparked by Grünbichler makes clear that most of Debian’s Rust-related prob­lems are in the process of be­ing solved. However, there’s no ev­i­dence that Klode ex­plored the prob­lems be­fore de­clar­ing that APT would de­pend on Rust, or even asked is this a rea­son­able time frame to in­tro­duce this de­pen­dency?”

Debian’s tagline, or at least one of its taglines, is the uni­ver­sal op­er­at­ing sys­tem”, mean­ing that the pro­ject aims to run on a wide va­ri­ety of hard­ware (old and new) and be us­able on the desk­top, server, IoT de­vices, and more. The Why Debian” page lists a num­ber of rea­sons users and de­vel­op­ers should choose the dis­tri­b­u­tion: mul­ti­ple hard­ware

ar­chi­tec­tures, long-term

sup­port, and its de­mo­c­ra­tic gov­er­nance

struc­ture are just a few of the ar­gu­ments it puts for­ward in fa­vor of Debian. It also notes that Debian can­not be con­trolled by a

sin­gle com­pany”. A sin­gle de­vel­oper em­ployed by a com­pany to work on Debian tools push­ing a change that seems ben­e­fi­cial to that com­pany, with­out dis­cus­sion or de­bate, that im­pacts mul­ti­ple hard­ware ar­chi­tec­tures and that re­quires other vol­un­teers to do un­planned work or meet an ar­ti­fi­cial dead­line seems to go against many of the pro­jec­t’s stated val­ues.

Debian, of course, does have checks and bal­ances that could be em­ployed if other Debian de­vel­op­ers feel it nec­es­sary. Someone could, for ex­am­ple, ap­peal to Debian’s Technical Committee, or spon­sor a gen­eral res­o­lu­tion to over­ride a de­vel­oper if they can­not be per­suaded by dis­cus­sion alone. That hap­pened re­cently when the com­mit­tee re­quired sys­temd

main­tain­ers to pro­vide the di­rec­tory until

a sat­is­fac­tory mi­gra­tion of im­pacted soft­ware has oc­curred and Policy

up­dated ac­cord­ingly”.

However, it also seems fair to point out that Debian can move slowly, even glacially, at times. APT added

sup­port for the DEB822

for­mat for its source in­for­ma­tion lists in 2015. Despite APT sup­port­ing that for­mat for years, Klode faced re­sis­tance in 2021, when he pushed

for Debian to move to the new for­mat ahead of the Debian 12 (“bookworm”) re­lease in 2021, but was un­suc­cess­ful. It is now the de­fault for trixie with the move to APT 3.0, though APT will con­tinue to sup­port the old for­mat for years to come.

The fact is, re­gard­less of what Klode does with APT, more and more free soft­ware is be­ing writ­ten (or rewrit­ten) in Rust. Making it eas­ier to sup­port that soft­ware when it is pack­aged for Debian is to every­one’s ben­e­fit. Perhaps the pro­ject needs some de­vel­op­ers who will be ag­gres­sive about push­ing the pro­ject to move more quickly in im­prov­ing its sup­port for Rust. However, what is re­ally needed is more de­vel­op­ers lend­ing a hand to do the work that is needed to sup­port Rust in Debian and else­where, such as . It does not seem in keep­ing with Debian’s com­mu­nity fo­cus for a sin­gle de­vel­oper to sim­ply de­clare de­pen­den­cies that other vol­un­teers will have to scram­ble to sup­port.

...

Read the original on lwn.net »

10 238 shares, 14 trendiness

Roblox is a problem — but it’s a symptom of something worse

What is the role of tech jour­nal­ism in a world where CEOs no longer feel shame?

What is the role of tech jour­nal­ism in a world where CEOs no longer feel shame?

On Friday, the Hard Fork team pub­lished our in­ter­view with Roblox CEO David Baszucki. In the days since, it has be­come the most-dis­cussed in­ter­view we’ve done in three years on the show. Listeners who wrote in to us said they were shocked to hear the leader of a plat­form with 151.5 mil­lion monthly users, most of them mi­nors, ex­press frus­tra­tion and an­noy­ance at be­ing asked about the com­pa­ny’s his­tory of fail­ures re­lated to child safety. Journalists de­scribed the in­ter­view as bizarre,” unhinged,” and a car crash.”

And a case can be made that it was all of those things — even if Baszucki, in the stu­dio af­ter­wards and later on X, in­sisted to us that he had had a good time. In the mo­ment, though, Baszucki’s dis­mis­sive at­ti­tude to­ward dis­cussing child safety struck me as some­thing worse: fa­mil­iar.

Baszucki, af­ter all, is not the first CEO to have in­sisted to me that a plat­for­m’s prob­lems are smaller than I am mak­ing them out to be. Nor is he the first to blame the plat­for­m’s enor­mous scale, or to try to change the sub­ject. (He is the first tech CEO to sug­gest to me that maybe there should be pre­dic­tion mar­kets in video games for chil­dren, but that’s an­other story.)

What peo­ple found note­wor­thy about our in­ter­view, I think, was the fresh ev­i­dence that our most suc­cess­ful tech CEOs re­ally do think and talk this way. Given a chance to dis­play em­pa­thy for the vic­tims of crimes his plat­form en­abled, or to con­vey re­gret about his­tor­i­cal safety lapses, or even just to ges­ture at some sense of re­spon­si­bil­ity for the hun­dreds of mil­lions of chil­dren who in var­i­ous ways are de­pend­ing on him, the CEO throws up his hands and asks: how long are you guys go­ing to be go­ing on about all this stuff?

Roblox is dif­fer­ent from other so­cial prod­ucts in that it ex­plic­itly courts users as young as 5. (You are sup­posed to be at least 13 to use Instagram, TikTok, and other ma­jor plat­forms.) That has al­ways put sig­nif­i­cant pres­sure on the com­pany to de­velop se­ri­ous safety fea­tures. The com­pany says it spends hun­dreds of mil­lions of dol­lars a year on safety, and that 10 per­cent of its em­ploy­ees work on trust and safety is­sues. And trust and safety work­ers I know tell me that they re­spect Roblox’s safety teams.

At the same time, this is a plat­form launched in 2006 where, for most of its his­tory, adults could freely ap­proach and mes­sage any mi­nor un­less their par­ents had dug into the app set­tings. Roblox did not ver­ify users’ ages, let­ting any child iden­tify as 13 or older to by­pass con­tent re­stric­tions. Filters in­tended to pre­vent in­ap­pro­pri­ate chat or the ex­change of per­sonal in­for­ma­tion were eas­ily by­passed by slightly chang­ing the spelling of words. Parental con­trols could be cir­cum­vented sim­ply by a child cre­at­ing a new ac­count and de­clar­ing that they were at least 13.

Last year the com­pany in­tro­duced new re­stric­tions on chat. And this year, the com­pany said it would de­ploy its own age es­ti­ma­tion tech­nol­ogy to de­ter­mine users’ ages and re­strict the con­tent avail­able to them ac­cord­ingly. This roll­out was the main rea­son we had sought to in­ter­view Baszucki in the first place — some­thing we had com­mu­ni­cated to his team.

Which only made it stranger when Baszucki ex­pressed sur­prise at our line of in­quiry and threw his PR team un­der the bus. (“If our PR peo­ple said, Let’s talk about age-gat­ing for an hour,′ I’m up for it, but I love your pod. I thought I came here to talk about every­thing,’” he said.)

Since 2018, at least two dozen peo­ple in the United States have been ar­rested and ac­cused of ab­duct­ing or abus­ing vic­tims they met on Roblox, ac­cord­ing to a 2024 in­ves­ti­ga­tion by Bloomberg. Attorneys gen­eral in Texas, Kentucky, and Louisiana have filed law­suits against Roblox al­leg­ing that the plat­form fa­cil­i­tates child ex­ploita­tion and groom­ing. More than 35 fam­i­lies have filed law­suits against the com­pany over child pre­da­tion.

As re­cently as this month, a re­porter for the Guardian cre­ated an ac­count pre­sent­ing her­self as a child and found that in Roblox she could wan­der user-cre­ated strip clubs, casi­nos, and hor­ror games. In one hangout” game, in which she iden­ti­fied as a 13-year-old, an­other avatar sex­u­ally as­saulted her by thrust­ing his hips into her avatar’s face as she begged him to leave her alone.

It’s true that any plat­form that lets strangers com­mu­ni­cate will lead to real-world harm. I be­lieve that mil­lions of chil­dren use Roblox daily with­out in­ci­dent. And we would not want to shut down the en­tire in­ter­net to pre­vent a sin­gle bad thing from ever hap­pen­ing.

But there is much a leader can do with the knowl­edge that his plat­form will in­evitably lead to harm, should he wish.

Understanding how at­trac­tive Roblox would be to preda­tors, the com­pany long ago could have blocked un­re­stricted con­tact be­tween adults and mi­nors. It could have adopted age ver­i­fi­ca­tion be­fore a wave of state leg­is­la­tion sig­naled that it would soon be­come manda­tory any­way. It could have made it harder for chil­dren un­der 13 to cre­ate new ac­counts, and re­quire them to get parental con­sent in a way it could ver­ify.

But do­ing so would re­quire Roblox to fo­cus on out­comes for chil­dren, at the likely ex­pense of growth. And so here we are.

Galling? Yes. But like I said: it’s also fa­mil­iar.

Over and over again, we have seen lead­ers in Baszucki’s po­si­tion choose growth over guardrails. Safety fea­tures come out years af­ter the need for them is iden­ti­fied, if at all. Internal crit­ics are side­lined, laid off, or man­aged out. And when jour­nal­ists ask, po­litely but in­sis­tently, why so many of their users are suf­fer­ing, ex­ec­u­tives laugh and tell us that we’re the crazy ones.

Look at OpenAI, where the com­pany is reck­on­ing with the fact that mak­ing its mod­els less syco­phan­tic has been worse for user en­gage­ment — and is build­ing new fea­tures to turn the en­gage­ment dial back up.

Look at TikTok, which has an­swered con­cerns that short-form video is wors­en­ing aca­d­e­mic per­for­mance for chil­dren with new digital well-be­ing fea­tures” that in­clude an af­fir­ma­tion jour­nal, a background sound gen­er­a­tor aimed at im­prov­ing the men­tal health of its users,” and new badges to re­ward peo­ple who use the plat­form within lim­its, es­pe­cially teens.” Answering con­cerns that teens are us­ing the app too much with more rea­sons to use the app.

Or look at Meta, where new court fil­ings from over the week­end al­lege … a truly stag­ger­ing num­ber of things. To name a few: the com­pany stalled in­ter­nal ef­forts to pre­vent child preda­tors from con­tact­ing mi­nors for years due to growth con­cerns,” ac­cord­ing to Jeff Horwitz in Reuters; recognized that op­ti­miz­ing its prod­ucts to in­crease teen en­gage­ment re­sulted in serv­ing them more harm­ful con­tent, but did so any­way”; and gave users 17 at­tempts to traf­fic peo­ple for sex be­fore ban­ning their ac­counts. (Meta de­nies the al­le­ga­tions, which are drawn from in­ter­nal doc­u­ments that have not been made pub­lic; Meta has also ob­jected to un­seal­ing the doc­u­ments.)

Lawsuits will al­ways con­tain the most sala­cious al­le­ga­tions lawyers can find, of course. But what struck me about these lat­est fil­ings is not the lawyers’ pre­dictably self-serv­ing fram­ing but rather the quotes from Meta’s own em­ploy­ees.

When the com­pany de­clined to pub­lish in­ter­nal re­search from 2019 which showed that no longer look­ing at Facebook and Instagram im­proved users’ men­tal health, one em­ployee said: If the re­sults are bad and we don’t pub­lish and they leak … is it go­ing to look like to­bacco com­pa­nies do­ing re­search and know­ing cigs were bad and then keep­ing that info to them­selves?”

When Meta re­searchers found that by 2018, ap­prox­i­mately 40 per­cent of chil­dren ages 9 to 12 were daily Instagram users — de­spite the fact that you are sup­posed to be 13 to join — some em­ploy­ees bris­tled at what they per­ceived as tacit en­cour­age­ment from ex­ec­u­tives to ac­cel­er­ate growth ef­forts among chil­dren.

Oh good, we’re go­ing af­ter Time’s ac­count of the brief. Zuck has been talk­ing about that for a while…tar­get­ing 11 year olds feels like to­bacco com­pa­nies a cou­ple decades ago (and to­day). Like we’re se­ri­ously say­ing we have to hook them young’ here.”

When Meta stud­ied the po­ten­tial of its prod­ucts to be ad­dic­tive in 2018, it found that 55 per­cent of 20,000 sur­veyed users showed at least some signs of problematic use.” When it pub­lished that re­search the fol­low­ing year, though, it re­de­fined problematic use” to in­clude only the most se­vere cases — 3.1 per­cent of users.

Because our prod­uct ex­ploits weak­nesses in the hu­man psy­chol­ogy to pro­mote prod­uct en­gage­ment and time spent,” a user ex­pe­ri­ence re­searcher wrote, the com­pany should alert peo­ple to the ef­fect that the prod­uct has on their brain.”

You will not be sur­prised to learn that the com­pany did not alert peo­ple to the is­sue.

As usual, the rank-and-file em­ploy­ees are do­ing their job. Over and over again, though, their boss’ boss tells them to stop.

The thing is, plat­forms’ strat­egy of de­lay, deny and de­flect mostly works.

Americans have short at­ten­tion spans — and lots to worry about. The tech back­lash that kicked off in 2017 in­spired plat­forms to make mean­ing­ful and ef­fec­tive in­vest­ments in con­tent mod­er­a­tion, cy­ber­se­cu­rity, plat­form in­tegrity, and other teams that worked to pro­tect their user bases. Imperfect as these ef­forts were, they bol­stered my sense that tech plat­forms were sus­cep­ti­ble to pres­sure from the pub­lic, from law­mak­ers and from jour­nal­ists. They acted slowly, and in­com­pletely, but at least they acted.

Fast for­ward to to­day and the bar­gain no longer holds. Platforms do what­ever the pres­i­dent of the United States tells them to do, and very lit­tle else. Shame, that once-great reg­u­la­tor of so­cial norms and ex­ec­u­tive be­hav­ior, has all but dis­ap­peared from pub­lic life. In its place is de­nial, de­fi­ance, and the nox­ious vice sig­nal­ing of the in­vestor class.

I’m still reck­on­ing with what it means to do jour­nal­ism in a world where the truth can barely hold any­one’s at­ten­tion — much less hold a plat­form ac­count­able, in any real sense of that word. I’m re­think­ing how to cover tech pol­icy at a time when it is be­ing made by whim. I’m notic­ing the de­gree to which plat­forms wish to be judged only by their stated in­ten­tions, and al­most never on the out­comes of any­one who uses them.

In the mean­time the plat­forms hur­tle on­ward, pitch­ing ever-more fan­tas­ti­cal vi­sions of the fu­ture while seem­ing barely in­ter­ested in stew­ard­ing the pre­sent.

For the mo­ment, I’m grate­ful that a car-crash in­ter­view drew at­ten­tion to one CEOs ex­as­per­a­tion with be­ing asked about that. But the real prob­lem is­n’t that David Baszucki talks this way. It’s that so many of his peers do, too.

The BBC caught scam call cen­ter work­ers on hid­den cam­eras as they laughed at the peo­ple they were trick­ing.

One worker bragged about mak­ing $250k from vic­tims. The dis­turb­ing truth?

Scammers don’t pick phone num­bers at ran­dom. They buy your data from bro­kers.

Once your data is out there, it’s not just calls. It’s phish­ing, im­per­son­ation, and iden­tity theft.

That’s why we rec­om­mend Incogni: They delete your info from the web, mon­i­tor and fol­low up au­to­mat­i­cally, and con­tinue to erase data as new risks ap­pear.

Black Friday deal: Try Incogni here and get 55% off your sub­scrip­tion with code PLAT­FORMER

What hap­pened: Facing crit­i­cism from both par­ties, the Trump ad­min­is­tra­tion backed down from is­su­ing an ex­ec­u­tive or­der that would have ef­fec­tively placed a mora­to­rium on state AI reg­u­la­tions, Reuters re­ported.

The or­der would have fought state reg­u­la­tions by with­hold­ing fed­eral fund­ing and es­tab­lish­ing an AI Litigation Task Force” to challenge State AI laws.”

Why we’re fol­low­ing: Last week we cov­ered the draft ex­ec­u­tive or­der and how Trump’s at­tempts to squash state AI reg­u­la­tion have drawn bi­par­ti­san back­lash — and made Republicans in­creas­ingly more sym­pa­thetic to the views of AI safety ad­vo­cates.

It’s al­ways hard to guess when Trump’s in­stinct to do as he pleases will be thwarted by po­lit­i­cal op­po­si­tion. In this case, though, the re­vived mora­to­rium had lit­tle sup­port out­side the David Sacks wing of the party. And so — for now, any­way — it fell apart.

What peo­ple are say­ing: State law­mak­ers are fight­ing the mora­to­rium pro­posal Trump made to Congress. Today, a let­ter signed by 280 state law­mak­ers urged Congress to reject any pro­vi­sion that over­rides state and lo­cal AI leg­is­la­tion.”

A mora­to­rium would threaten ex­ist­ing laws that strengthen con­sumer trans­parency, guide re­spon­si­ble gov­ern­ment pro­cure­ment, pro­tect pa­tients, and sup­port artists and cre­ators,” the let­ter said.

On the other side of the de­bate, the tech-funded in­dus­try PAC Leading the Future an­nounced a $10 mil­lion cam­paign to push Congress to pass na­tional AI reg­u­la­tions that would su­per­sede state law.

What hap­pened: On Friday, X de­buted its About This Account fea­ture glob­ally in a roll­out that de­scended into chaos over the fea­ture’s ac­ci­den­tal un­cov­er­ing of for­eign ac­tors be­hind pop­u­lar right-wing ac­counts that ac­tively share news on US pol­i­tics.

X users can now see the date an ac­count joined the plat­form, how many times it has changed its user­name, and most im­por­tantly, the coun­try or re­gion it’s based in. The move, ac­cord­ing to X head of prod­uct Nikita Bier, is an im­por­tant first step to se­cur­ing the in­tegrity of the global town square.”

But the fea­ture has had an un­in­tended con­se­quence: it re­vealed that big pro-Trump ac­counts like @MAGANationX, a right-wing user with nearly 400,000 fol­low­ers that reg­u­larly shares news about US pol­i­tics, aren’t ac­tu­ally based in the US. MAGANationX, for ex­am­ple, is based in Eastern Europe, ac­cord­ing to X.

Other pop­u­lar right-wing ac­counts — that use names from the Trump fam­ily — like @IvankaNews_ (1 mil­lion fol­low­ers be­fore it was sus­pended), @BarronTNews (nearly 600,000 fol­low­ers), and @TrumpKaiNews (more than 11,000 fol­low­ers), ap­pear to be based in Nigeria, Eastern Europe, and Macedonia re­spec­tively.

The data could be skewed by travel, VPNs, or old IP ad­dresses, and some have com­plained their lo­ca­tion is in­ac­cu­rate. Bier said the roll­out has a few rough edges” that will be re­solved by Tuesday.

Why we’re fol­low­ing: One of Elon Musk’s promises dur­ing the takeover of Twitter was to purge the plat­form of in­au­then­tic ac­counts. But sev­eral stud­ies have shown that sus­pected in­au­then­tic ac­tiv­ity has re­mained at about the same lev­els. X has long strug­gled with troll farms spread­ing mis­in­for­ma­tion, boosted by its ten­dency to mon­e­tar­ily re­ward en­gage­ment.

There’s also an irony in the fact that re­veal­ing the ori­gins of rage­bait-post­ing po­lit­i­cal ac­counts like these was once the sub­ject of ground­break­ing re­search by the Stanford Internet Observatory and other aca­d­e­mic re­searchers. But the ef­fort out­raged Republicans, which then sued them over their con­tacts with the gov­ern­ment about in­for­ma­tion op­er­a­tions like these and largely suc­ceeded in stop­ping the work.

What peo­ple are say­ing: Accusations of for­eign ac­tors spread­ing fake news flew on both sides of the aisle. When the fea­ture ap­peared to be pulled for a short pe­riod of time, Republican Gov. Ron DeSantis of Florida said X needs to re­in­state county-of-ori­gin — it helps ex­pose the grift.”

In a post that gar­nered 3.2 mil­lion views, @greg16676935420 at­tached a screen­shot of @AmericanGuyX’s pro­file, which shows the ac­coun­t’s based in India: BREAKING: American guy is not ac­tu­ally an American guy.”

When an American bil­lion­aire of­fers money to peo­ple from rel­a­tively poor coun­tries for ril­ing up and rad­i­cal­is­ing Americans, it’s not sur­pris­ing that they’ll take up the of­fer,” @ChrisO_wiki wrote in a post that gar­nered nearly 700,000 views.

In per­haps the most dev­as­tat­ing con­se­quence of the fea­ture, @veespo_444s said they spent 2 years act­ing mys­te­ri­ous over what coun­try I live in just for Elon to fuck it all up with a sin­gle up­date” in a post that has 4.3 mil­lion views and 90,000 likes.

How President Trump am­pli­fies right-wing trolls and AI memes. The crypto crash has taken about $1 bil­lion out of the Trump fam­ily for­tune.

Gamers are us­ing Fortnite and GTA to pre­pare for ICE raids. How Democrats are build­ing their on­line strat­egy to catch up with Republicans.

In the last month, Elon Musk has posted more about pol­i­tics than about his com­pa­nies on X.

Hundreds of English-language web­sites link to ar­ti­cles from a pro-Krem­lin dis­in­for­ma­tion net­work and are be­ing used to groom” AI chat­bots into spread­ing Russian pro­pa­ganda, a study found.

Sam Altman and Jony Ive said they’re now pro­to­typ­ing their hard­ware de­vice, but it re­mains two years away. An in-depth look at OpenAI’s men­tal health cri­sis af­ter GPT-4o de­tails how the com­pany changed ChatGPT af­ter re­ports of harm­ful in­ter­ac­tions. OpenAI safety re­search leader Andrea Vallone, who led ChatGPT’s re­sponses to men­tal health crises, is re­port­edly leav­ing. A re­view of ChatGPT’s new per­sonal shop­ping agent.

Anthropic un­veiled Claude Opus 4.5, which it said is the best model for soft­ware en­gi­neer­ing. Other high­lights from the launch: it outscored hu­man en­gi­neer­ing can­di­dates on a take-home exam, is cheaper than Opus 4.1, can keep a chat go­ing in­def­i­nitely via on­go­ing sum­ma­riza­tion of past chats, and is harder to trick with prompt in­jec­tion.

In other re­search, AI mod­els can un­in­ten­tion­ally de­velop mis­aligned be­hav­iors af­ter learn­ing to cheat, Anthropic said. (This won an ap­prov­ing tweet from Ilya Sutskever, who had­n’t posted about AI on X in more than a year.)

Why Meta’s $27 bil­lion data cen­ter and its debt won’t be on its bal­ance sheet. Meta is ven­tur­ing into elec­tric­ity trad­ing to speed up its power plant con­struc­tion. Facebook Groups now has a nick­name fea­ture for anony­mous post­ing.

A judge is set to de­cide on reme­dies for Google’s adtech mo­nop­oly next year. Italy closed its probe into Google over un­fair prac­tices that used per­sonal data. Google stock closed at a record high last week af­ter the suc­cess­ful launch of Gemini 3. AI Mode now has ads.

Something for the AI skep­tics: Google must dou­ble its serv­ing ca­pac­ity every six months to meet cur­rent de­mand for AI ser­vices, Google Cloud VP Amin Vahdat said.

AI de­mand has strained the mem­ory chip sup­ply chain, chip­mak­ers said.

Amazon has more than 900 data cen­ters — more than pre­vi­ously known — in more than 50 coun­tries. Its Autonomous Threat Analysis sys­tem uses spe­cial­ized AI agents for de­bug­ging. AWS said it would in­vest $50 bil­lion in AI ca­pa­bil­i­ties for fed­eral agen­cies.

Twitch was added to Australia’s list of plat­forms banned for un­der-16s. Pinterest was spared.

Grindr said it ended talks on a $3.5 bil­lion take-pri­vate deal, cit­ing un­cer­tainty over fi­nanc­ing.

Interviews with AI qual­ity raters who are telling their friends and fam­ily not to use the tech. How AI is threat­en­ing the fun­da­men­tal method of on­line sur­vey re­search by evad­ing bot de­tec­tion tech­niques. Insurers are look­ing to limit their li­a­bil­ity on claims re­lated to AI. Another look at how America’s econ­omy is now deeply tied to AI stocks and their per­for­mance.

Scientists built an AI model that can flag hu­man ge­netic mu­ta­tions likely to cause dis­ease.

For more good posts every day, fol­low Casey’s Instagram sto­ries.

Send us tips, com­ments, ques­tions, and your ques­tions for the tech CEOs: casey@plat­former.news. Read our ethics pol­icy here.

...

Read the original on www.platformer.news »

To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".

10HN is also available as an iOS App

If you visit 10HN only rarely, check out the the best articles from the past week.

If you like 10HN please leave feedback and share

Visit pancik.com for more.