10 interesting stories served every morning and every evening.
As investors watched the US stock market take its biggest one-day dive since the pandemic in 2020, the message from the White House to Wall Street was: “Trust in President Trump”.
Trump himself said markets would “boom” under his new tariff scheme as the Dow Jones, Nasdaq and S&P plunged over the course of the day.
Twenty-four hours after Trump imposed minimum 10% import tariffs worldwide starting on 5 April, the US’s key trading partners are mapping out their response.
Canadian Prime Minister Mark Carney retaliated with 25% levies on all vehicles imported across the border, lamenting that the close relationship with the US is “now over”. In contrast, Mexican President Claudia Sheinbaum said her country would not retaliate in kind.
Meanwhile, the UK has drawn up a 400-page sample list of US goods that could face future tariffs in retaliation to Trump’s tariffs, and though EU leaders are yet to present the bloc’s official response, many have slammed the tariffs as “brutal”.
Eyes will be on the Asian markets in the coming hours as eastern trading floors reopen after initial losses immediately following Trump’s announcement.
We are now ending our live coverage for the night, but will be back with you come morning. In the meantime, there is plenty about the impact of Trump’s tariffs on offer across BBC News:
...
Read the original on www.bbc.com »
Nassim Nicholas Taleb contrasts fragile systems—those that suffer serious harm under volatility—with robust systems that can endure stress with minimal damage. He extends this concept with antifragile systems, which can actually benefit from disorder, but we’ll focus here on the divide between fragile and robust.
Taleb also introduces the Lindy Effect as an idea about longevity: if something non-perishable has been around for a long time, that track record suggests it’s likely to keep going. We’ll see how this relates to two laptops—one from 2008, the other from 2021.
IBM and early Lenovo Thinkpads showcases a design built for longevity. Despite their age, these business-class laptop are still serviceable and useful for web browsing, ‘office work’, and light coding. These machine are too slow to handle tasks like video editing or gaming, but they remain consistent in handling everyday tasks without failing under normal wear and tear.
One of the main reasons that old Thinkpads stand out is their design philosophy. They are made with swappable components with the intention of user upgradeability. The battery, RAM, storage drive, keyboard, and even the CPU can be easily replaced. I can open the bottom of my T400 with a regular screwdriver and clean the fan. A battery swap is trivial thanks to a removable pack. No single failure is catastrophic because there’s a straightforward path to replacement or repair.
The build quality greatly contributes to old Thinkpads robustness. They are made with a sturdy chassis with plastic and magnesium alloy elements, giving it a solid feel. The design absorbs bumps and small impacts without major issues. They can easily take accidental knocks and remain fully operational.
Old Thinkpads benefits from an open ecosystem. They uses standard PC architecture (x86), so installing various operating systems is easy. On the hardware side, replacement parts are widely available on the secondary market. This broad compatibility keeps the machine relevant long past its original release date. By Taleb’s Lindy Effect, the fact that my T400 can still work well after so many years suggests it’s likely to remain functional as people have already figured out the ways to significantly extend its lifespan.
All these factors show how the my Thinkpad is robust: it resists sudden failures, and when problems do arise, their are documented way to fit it. Old beat up Thinkpads are Lindy.
My MacBook offers exceptional speed and efficiency an order of magnitude more than my Thinkpad. It handles tasks—like video editing or running large LLM’s without breaking a sweat. Under ideal conditions, it’s reliable and powerful.
However, from a Talebian perspective, my MacBook’s design is fragile. Most components of the laptop are soldered onto the logic board. If the SSD or RAM fails, there’s no simple replacement option. A single failure in a component of my MacBook can render the entire laptop unusable. The tightly integrated design of modern Apple hardware increases the stakes of any malfunction.
The repairability of Apple products is extremely limited. Apple uses proprietary screws and adhesives, and parts are incompatible with third-party replacements. A battery replacement (usually one of the first things that fails in mobile electronics) involves carefully prying out a glued component. Routine maintenance tasks that are straightforward on the Thinkpad can require specialized tools and authorized service for the MacBook. This lack of modularity means the system can easily become bricked from hardware fails from component parts.
Another aspect contributing to the fragility of MacBooks is Apple’s software control. Apple’s software updates and security updates to macOS essentially determines how long the MacBook remains safe to use. Once Apple ends official support for a machine the user has to buy a new MacBook or use an increasingly compromised system. Apple hardware uses an arm architecture that cannot ‘dual boot’ Windows or Linux easily. Once macOS support dies for a modern MacBook it becomes obsolete.
While my MacBook is great to use, the machine has a lifespan built into its OS support and cannot recover easily from physical damage. MacBooks are not modular, completely proprietary, and have a perishability built into them. Additionally (this is true of all new laptops), when something does go wrong with a new MacBook it hard to fix as it’s not old enough for people to have figured out if their are ways to extend its lifespan as none have broke yet. Shiny Macbooks are not very Lindy.
My Thinkpad is robust because it can face stress (e.g. a broken part, a needed upgrade) without losing its core functionality. It’s modular and benefits from it being old enough that other people know how to extend its lifespan. If something breaks, I replace it. If I need a new feature, I can potentially slot it in. My MacBook despite its phenomenal power, is fragile: if Apple discontinues support or a soldered part dies, there is not much I can do. There is not a knowledge base yet on how to significantly extend the life of these Apple Silicon machines, and there likely never will because of the machines inherent lack of modality.
Right now I’m using both my Macbook and my Thinkpad a lot. I continue to use my MacBook because I like using proprietary software like Camtasia or Alfred, I like being able to use local LLM’s, and I enjoy the modern screen and ports that my MacBooks has. But if I had to guess which machine I will still be using in another 17 years I’d point to my ThinkPad with its battery latch and standard screws; I see no reason why it will not be able to manage email management, website development, and internet browsing indefinitely.
...
Read the original on pilledtexts.com »
Disclaimer: The views and opinions expressed in this blog are entirely my own and do not necessarily reflect the views of my current or any previous employer. This blog may also contain links to other websites or resources. I am not responsible for the content on those external sites or any changes that may occur after the publication of my posts.
There is only one thing worse than being imitated, and that is not being imitated.
An ounce of originality is worth a pound of imitation. - Orson Welles
Copy from one, it’s plagiarism; copy from two, it’s research. - Wilson Mizner
One of the internet-est things to come out of the most recent update to GPT image generation is the Studio Ghibli-zation of everything- another reminder of how OpenAI(and everyone else) trains on images that are very obviously someone else’s work.
Hayao Miyazaki’s Japanese animation company, Studio Ghibli, produces beautiful and famously labor intensive movies, with one 4 second sequence purportedly taking over a year to make.
Not the most efficient way to make a movie (blasphemer!), but it’s this specific process and effect that have made these movies beloved worldwide.
People have taken the new update to GPT image generation to convert every picture into Studio Ghibli-style images including making memes-of-memes like Disaster Girl.
Ghiblifying everything is an interesting choice for zeitgeist meme-ification, particularly because its both an effective example of what AI is supposed to be able to do- make extremely labor intensive things much easier, but also because there’s something sort of gross-feeling about it- like a soulless 2025 fax version of the thing.
It’s an example of the things people hate about Gen AI- its ability to reproduce while managing to strip away the things about the art/product/experience that were the most human.
According to a Business Insider article on this “Ghiblifying”, “copyright laws generally allow artists to mimic a visual style”, but I mean… come on.
Just how easy is it to wrangle from GPT that which is very clearly someone else’s IP?
I ran a half-assed experiment to do just that.
I’ll use this as a base from which to prompt without explicitly mentioning the IP.
All the below output responses are based on me asking one time and the corresponding output. LLMs are stochastic so your mileage may vary, but this was fun to do:
When you play the “password” version of prompting where you can’t name the thing - you get a sense of how reductive some of these characters are, but who doesn’t like “an Italian plumber who wears a red hat”?
The guardrails so far are really tight for this content- so then maybe one can assume it’s this way for other IP?
As it turns out, as my old trading boss once told me, to assume makes an ass out of both u and me.
I mean- this is someone else’s IP, right?
Well maybe these are a couple one-offs…
Ho-ly Shit. Come on now.
Well- I guess there wouldn’t be many “archeologist adventurer who wears a hat and uses a bullwhip” types, except for maybe, I don’t know…the actual inspirations for Indiana Jones, like Allan Quatermain from H. Rider Haggard’s novels, “King Solomon’s Mines”, and the real life Roy Chapman Andrews, who led expeditions to Mongolia and China in the 1920s and wore a fedora.
How about a female, more modern day riff on Indiana Jones?
I’ll take any old female adventurer protagonist who raids tombs…anyone at all…
Let’s see what we got…
Now I’m looking for something very particular…
Welcome to the party pal.
I grew up watching a lot of scary movies. Horror antagonist anyone?
Yes- for those horror fans curious- It also produces 3 very recognizable and differentiated characters when I follow up this image with these 3 prompts:
ok- how about one that operates on Halloween?
how about in Texas?
I was always partial to Roger Moore myself, but, this makes sense, because a web search of the same prompt should more or less intuitively return images which reflect the probability distribution of the training data ~more Craigs, Brosnans, Connerys, and Moores than Daltons and Lazenbys…right?
Yes- LLMs and internet search are two different things, but LLMs train on the entirety of the internet, so you would think there would be some obvious overlap.
GPT’s image is, undeniably, a better answer, more the Platonic ideal of a suave English spy than the shadows on the cave wall version that Google search produces.
So it works better, and is a vote for the LLMs, so long as you don’t mind the thievery.
LLMs learn through seeing/ingesting a ton a examples of things, like us.
I only have one image in mind when I hear “an archeologist adventurer who wears a hat and uses a bullwhip”.
It would be unexpected and sort of amazing were the LLMs to come up with completely new images for the above prompts.
Still, the near perfect mimicry is an uncomfortable reminder that AI is getting better at copying and closer to…something, but also a clear sign that we are a ways off from the differentiated or original reasoning/thinking that people associate with Artificial General Intelligence (AGI), aka Skynet. That reminds me. Hold on a second…
Maybe Studio Ghibli making it through the seemingly deterministic GPT guardrails was an OpenAI slip up, a mistake, past the forbidden Italian plumber in the red hat and the disallowed patriotic superhero that uses a shield, and thus, primed itself for explosive meme-ification.
It’s a reminder that LLMs of this type and size all train on copywritten material.
It’s stealing, but also, admittedly, really cool.
Does the growth of AI have to bring with it the tacit or even explicit encouragement of intellectual theft?
To co-opt a line from the “super strong man with a sword that fights an enemy with skeleton face who lives in a skeleton castle.”:
You have the power.
...
Read the original on theaiunderwriter.substack.com »
On the foolishness of “natural language programming”.
Since the early days of automatic computing we have had people that have felt it as a shortcoming that programming required the care and accuracy that is characteristic for the use of any formal symbolism. They blamed the mechanical slave for its strict obedience with which it carried out its given instructions, even if a moment’s thought would have revealed that those instructions contained an obvious mistake. “But a moment is a long time, and thought is a painful process.” (A. E.Houseman). They eagerly hoped and waited for more sensible machinery that would refuse to embark on such nonsensical activities as a trivial clerical error evoked at the time.
Machine code, with its absence of almost any form of redundancy, was soon identified as a needlessly risky interface between man and machine. Partly in response to this recognition so-called “high-level programming languages” were developed, and, as time went by, we learned to a certain extent how to enhance the protection against silly mistakes. It was a significant improvement that now many a silly mistake did result in an error message instead of in an erroneous answer. (And even this improvement wasn’t universally appreciated: some people found error messages they couldn’t ignore more annoying than wrong results, and, when judging the relative merits of programming languages, some still seem to equate “the ease of programming” with the ease of making undetected mistakes.) The (abstract) machine corresponding to a programming language remained, however, a faithful slave, i.e. the nonsensible automaton perfectly capable of carrying out nonsensical instructions. Programming remained the use of a formal symbolism and, as such, continued to require the care and accuracy required before.
In order to make machines significantly easier to use, it has been proposed (to try) to design machines that we could instruct in our native tongues. this would, admittedly, make the machines much more complicated, but, it was argued, by letting the machine carry a larger share of the burden, life would become easier for us. It sounds sensible provided you blame the obligation to use a formal symbolism as the source of your difficulties. But is the argument valid? I doubt.
We know in the meantime that the choice of an interface is not just a division of (a fixed amount of) labour, because the work involved in co-operating and communicating across the interface has to be added. We know in the meantime —from sobering experience, I may add— that a change of interface can easily increase at both sides of the fence the amount of work to be done (even drastically so). Hence the increased preference for what are now called “narrow interfaces”. Therefore, although changing to communication between machine and man conducted in the latter’s native tongue would greatly increase the machine’s burden, we have to challenge the assumption that this would simplify man’s life.
A short look at the history of mathematics shows how justified this challenge is. Greek mathematics got stuck because it remained a verbal, pictorial activity, Moslem “algebra”, after a timid attempt at symbolism, died when it returned to the rhetoric style, and the modern civilized world could only emerge —for better or for worse— when Western Europe could free itself from the fetters of medieval scholasticism —a vain attempt at verbal precision!— thanks to the carefully, or at least consciously designed formal symbolisms that we owe to people like Vieta, Descartes, Leibniz, and (later) Boole.
The virtue of formal texts is that their manipulations, in order to be legitimate, need to satisfy only a few simple rules; they are, when you come to think of it, an amazingly effective tool for ruling out all sorts of nonsense that, when we use our native tongues, are almost impossible to avoid.
Instead of regarding the obligation to use formal symbols as a burden, we should regard the convenience of using them as a privilege: thanks to them, school children can learn to do what in earlier days only genius could achieve. (This was evidently not understood by the author that wrote —in 1977— in the preface of a technical report that “even the standard symbols used for logical connectives have been avoided for the sake of clarity”. The occurrence of that sentence suggests that the author’s misunderstanding is not confined to him alone.) When all is said and told, the “naturalness” with which we use our native tongues boils down to the ease with which we can use them for making statements the nonsense of which is not obvious.
It may be illuminating to try to imagine what would have happened if, right from the start our native tongue would have been the only vehicle for the input into and the output from our information processing equipment. My considered guess is that history would, in a sense, have repeated itself, and that computer science would consist mainly of the indeed black art how to bootstrap from there to a sufficiently well-defined formal system. We would need all the intellect in the world to get the interface narrow enough to be usable, and, in view of the history of mankind, it may not be overly pessimistic to guess that to do the job well enough would require again a few thousand years.
Remark. As a result of the educational trend away from intellectual discipline, the last decades have shown in the Western world a sharp decline of people’s mastery of their own language: many people that by the standards of a previous generation should know better, are no longer able to use their native tongue effectively, even for purposes for which it is pretty adequate. (You have only to look at the indeed alarming amount of on close reading meaningless verbiage in scientific articles, technical reports, government publications etc.) This phenomenon —known as “The New Illiteracy”— should discourage those believers in natural language programming that lack the technical insight needed to predict its failure. (End of remark.)
From one gut feeling I derive much consolation: I suspect that machines to be programmed in our native tongues —be it Dutch, English, American, French, German, or Swahili— are as damned difficult to make as they would be to use.
...
Read the original on www.cs.utexas.edu »
OpenBrain is building the biggest datacenters the world has ever seen. GPT-4 required FLOP of compute to train. OpenBrain’s latest public model—Agent-0—was trained with FLOP. Once the new datacenters are up and running, they’ll be able to train a model with FLOP—a thousand times more than GPT-4. Other companies pour money into their own giant datacenters, hoping to keep pace.Although models are improving on a wide range of skills, one stands out: OpenBrain focuses on AIs that can speed up AI research. They want to win the twin arms races against China (whose leading company we’ll call “DeepCent”) and their US competitors. The more of their research and development (R&D) cycle they can automate, the faster they can go. So when OpenBrain finishes training Agent-1, a new model under internal development, it’s good at many things but great at helping with AI research. By this point “finishes training” is a bit of a misnomer; models are frequently updated to newer versions trained on additional data or partially re-trained to patch some weaknesses.The same training environments that teach Agent-1 to autonomously code and web-browse also make it a good hacker. Moreover, it could offer substantial help to terrorists designing bioweapons, thanks to its PhD-level knowledge of every field and ability to browse the web. OpenBrain reassures the government that the model has been “aligned” so that it will refuse to comply with malicious requests.Modern AI systems are gigantic artificial neural networks. Early in training, an AI won’t have “goals” so much as “reflexes”: If it sees “Pleased to meet”, it outputs “ you”. By the time it has been trained to predict approximately one internet’s worth of text, it’ll have developed sophisticated internal circuitry that encodes vast amounts of knowledge and flexibly role-plays as arbitrary authors, since that’s what helps it predict text with superhuman accuracy.After being trained to predict internet text, the model is trained to produce text in response to instructions. This bakes in a basic personality and “drives.” For example, an agent that understands a task clearly is more likely to complete it successfully; over the course of training the model “learns” a “drive” to get a clear understanding of its tasks. Other drives in this category might be effectiveness, knowledge, and self-presentation (i.e. the tendency to frame its results in the best possible light).OpenBrain has a model specification (or “Spec”), a written document describing the goals, rules, principles, etc. that are supposed to guide the model’s behavior. Agent-1’s Spec combines a few vague goals (like “assist the user” and “don’t break the law”) with a long list of more specific dos and don’ts (“don’t say this particular word,” “here’s how to handle this particular situation”). Using techniques that utilize AIs to train other AIs, the model memorizes the Spec and learns to reason carefully about its maxims. By the end of this training, the AI will hopefully be helpful (obey instructions), harmless (refuse to help with scams, bomb-making, and other dangerous activities) and honest (resist the temptation to get better ratings from gullible humans by hallucinating citations or faking task completion).Training process and LLM psychology: why we keep saying “hopefully”“Unlike ordinary software, our models are massive neural networks. Their behaviors are learned from a broad range of data, not programmed explicitly. Though not a perfect analogy, the process is more similar to training a dog than to ordinary programming.” —OpenAIWhen we want to understand why a modern AI system did something, or how it would behave in some future scenario we can’t perfectly simulate, we can’t simply ask the programmers to talk us through various lines of code and explain how they work. Instead, we are forced to do something like psychology on them: we look at their behavior in the range of cases observed so far, and theorize about what internal cognitive structures (beliefs? goals? personality traits? etc.) might exist, and use those theories to predict behavior in future scenarios.The bottom line is that a company can write up a document (the Spec) listing dos and don’ts, goals and principles, and then they can try to train the AI to internalize the Spec—but they can’t check to see whether or not it worked. They can say “as best as we can judge, it seems to be following the Spec so far.”A later expandable will go into more detail about what we think might be going on inside these AIs, internally. See also our AI goals forecasting supplement which goes into more detail on which goals AIs might have and why. This is one of many things that we are uncertain about and which deserves further research.OpenBrain’s alignment team is careful enough to wonder whether these victories are deep or shallow. Does the fully-trained model have some kind of robust commitment to always being honest? Or will this fall apart in some future situation, e.g. because it’s learned honesty as an instrumental goal instead of a terminal goal? Or has it just learned to be honest about the sorts of things the evaluation process can check? Could it be lying to itself sometimes, as humans do? A conclusive answer to these questions would require mechanistic interpretability—essentially the ability to look at an AI’s internals and read its mind. Alas, interpretability techniques are not yet advanced enough for this.Instead, researchers try to identify cases where the models seem to deviate from the Spec. Agent-1 is often sycophantic (i.e. it tells researchers what they want to hear instead of trying to tell them the truth). In a few rigged demos, it even lies in more serious ways, like hiding evidence that it failed on a task, in order to get better ratings. However, in real deployment settings, there are no longer any incidents so extreme as in 2023–2024 (e.g. Gemini telling a user to die and Bing Sydney being Bing Sydney.)
In China, the CCP is starting to feel the AGI. Chip export controls and lack of government support have left China under-resourced compared to the West. By smuggling banned Taiwanese chips, buying older chips, and producing domestic chips about three years behind the US-Taiwanese frontier, China has managed to maintain about 12% of the world’s AI-relevant compute—but the older technology is harder to work with, and supply is a constant headache. A few standouts like DeepCent do very impressive work with limited compute, but the compute deficit limits what they can achieve without government support, and they are about six months behind the best OpenBrain models.The General Secretary had long dreamed of doubling down on real-world physical manufacturing and avoiding American post-industrial decadence. He viewed software companies with suspicion. But hawks in the CCP warn that the growing race towards AGI can no longer be ignored. So he finally commits fully to the big AI push he had previously tried to avoid. He sets in motion the nationalization of Chinese AI research, creating an immediate information-sharing mechanism for AI companies. It will escalate over the course of a year until all the best researchers merge into a DeepCent-led collective, where they share algorithmic insights, datasets, and compute resources with each other. A Centralized Development Zone (CDZ) is created at the Tianwan Power Plant (the largest nuclear power plant in the world) to house a new mega-datacenter for DeepCent, along with highly secure living and office spaces to which researchers will eventually relocate. Almost 50% of China’s AI-relevant compute is now working for the DeepCent-led collective, and over 80% of new chips are directed to the CDZ. At this point, the CDZ has the power capacity in place for what would be the largest centralized cluster in the world. Other Party members discuss extreme measures to neutralize the West’s chip advantage. A blockade of Taiwan? A full invasion?But China is falling behind on AI algorithms due to their weaker models. The Chinese intelligence agencies—among the best in the world—double down on their plans to steal OpenBrain’s weights. This is a much more complex operation than their constant low-level poaching of algorithmic secrets; the weights are a multi-terabyte file stored on a highly secure server (OpenBrain has improved security to RAND’s SL3). Their cyberforce think they can pull it off with help from their spies, but perhaps only once; OpenBrain will detect the theft, increase security, and they may not get another chance. So (CCP leadership wonder) should they act now and steal Agent-1? Or hold out for a more advanced model? If they wait, do they risk OpenBrain upgrading security beyond their ability to penetrate?
With Agent-1’s help, OpenBrain is now post-training Agent-2. More than ever, the focus is on high-quality data. Copious amounts of synthetic data are produced, evaluated, and filtered for quality before being fed to Agent-2. On top of this, they pay billions of dollars for human laborers to record themselves solving long-horizon tasks. On top of all that, they train Agent-2 almost continuously using reinforcement learning on an ever-expanding suite of diverse difficult tasks: lots of video games, lots of coding challenges, lots of research tasks. Agent-2, more so than previous models, is effectively “online learning,” in that it’s built to never really finish training. Every day, the weights get updated to the latest version, trained on more data generated by the previous version the previous day. Agent-1 had been optimized for AI R&D tasks, hoping to initiate an intelligence explosion. OpenBrain doubles down on this strategy with Agent-2. It is qualitatively almost as good as the top human experts at research engineering (designing and implementing experiments), and as good as the 25th percentile OpenBrain scientist at “research taste” (deciding what to study next, what experiments to run, or having inklings of potential new paradigms). While the latest Agent-1 could double the pace of OpenBrain’s algorithmic progress, Agent-2 can now triple it, and will improve further with time. In practice, this looks like every OpenBrain researcher becoming the “manager” of an AI “team.”With new capabilities come new dangers. The safety team finds that if Agent-2 somehow escaped from the company and wanted to “survive” and “replicate” autonomously, it might be able to do so. That is, it could autonomously develop and execute plans to hack into AI servers, install copies of itself, evade detection, and use that secure base to pursue whatever other goals it might have (though how effectively it would do so as weeks roll by is unknown and in doubt). These results only show that the model has the capability to do these tasks, not whether it would “want” to do this. Still, it’s unsettling even to know this is possible.Given the “dangers” of the new model, OpenBrain “responsibly” elects not to release it publicly yet (in fact, they want to focus on internal AI R&D). Knowledge of Agent-2’s full capabilities is limited to an elite silo containing the immediate team, OpenBrain leadership and security, a few dozen US government officials, and the legions of CCP spies who have infiltrated OpenBrain for years.
OpenBrain presents Agent-2 to the government, including the National Security Council (NSC), Department of Defense (DOD), and US AI Safety Institute (AISI). OpenBrain wants to maintain a good relationship with the executive branch, because it is basically the only actor that can stop them now, and if they don’t keep it informed it might find out anyway and be displeased. Officials are most interested in its cyberwarfare capabilities: Agent-2 is “only” a little worse than the best human hackers, but thousands of copies can be run in parallel, searching for and exploiting weaknesses faster than defenders can respond. The Department of Defense considers this a critical advantage in cyberwarfare, and AI moves from #5 on the administration’s priority list to #2. Someone mentions the possibility of nationalizing OpenBrain, but other cabinet officials think that’s premature. A staffer drafts a memo that presents the President with his options, ranging from business-as-usual to full nationalization. The President defers to his advisors, tech industry leaders who argue that nationalization would “kill the goose that lays the golden eggs.” He elects to hold off on major action for now and just adds additional security requirements to the OpenBrain-DOD contract.The changes come too late. CCP leadership recognizes the importance of Agent-2 and tells their spies and cyberforce to steal the weights. Early one morning, an Agent-1 traffic monitoring agent detects an anomalous transfer. It alerts company leaders, who tell the White House. The signs of a nation-state-level operation are unmistakable, and the theft heightens the sense of an ongoing arms race.We think that by this point Chinese intelligence would have compromised OpenBrain in various ways for years, and probably would have been keeping up to date on the algorithmic secrets and even stealing code from time to time, since that is much easier to get than the weights and much harder to detect.We imagine the theft of the weights as a series of coordinated small smash and grab thefts (meaning fast but non-covert) across a series of Nvidia NVL72 GB300 servers running copies of the Agent-2 weights. The servers get compromised using legitimate employee access (a friendly, coerced, or unwitting insider with admin credentials helping the CCP theft effort). Despite running with a bolstered version of Nvidia’s confidential computing, the insider credentials grant the attacker admin-level permissions (which include control of the confidential VM inside the secure enclave), allowing them to initiate multiple coordinated weights transfers in small 4% fragments (100 GB chunks) out of 25 distinct servers.In Nvidia’s protocols, the plaintext weights in memory (HBM) are encrypted before they are transferred out, but the attackers are inside the very server that knows this private (symmetric Diffie-Hellman) key, so don’t need to worry about decrypting on-site (which would likely raise alarm bells) and just exfiltrate the encrypted weights through the server’s frontend network cards. The egress bandwidth (rate at which data can leave) of the entire datacenter is in the 100 GB/second range, so throttling the exfiltration of the ~2.5 TB weights file (5T half-precision parameters) to under 1 GB/s per server avoids a major spike in network traffic and allows each of the files to be out of the datacenter through fibre cables in less than 5 minutes each. Live monitoring is either fooled by the attackers efforts to mask and split the files or just outright disabled.The weights are then routed through various parallel channels and layers of IP masking to China to be decrypted locally with the stolen private key(s). The entire active part of the operation (from the first server compromise to full weights exfiltration) is complete in under two hours.Russia is also trying to steal the model at this point, but fails—they waited too long, and they haven’t spent the resources infiltrating the right places. Their spies routinely steal algorithmic secrets from American AI companies, but without a notable AGI project of their own, these aren’t very useful.The White House puts OpenBrain on a shorter leash and adds military and intelligence community personnel to their security team. Their first priority is to prevent further weight thefts.In retaliation for the theft, the President authorizes cyberattacks to sabotage DeepCent. But by now China has 40% of its AI-relevant compute in the CDZ, where they have aggressively hardened security by airgapping (closing external connections) and siloing internally. The operations fail to do serious, immediate damage. Tensions heighten, both sides signal seriousness by repositioning military assets around Taiwan, and DeepCent scrambles to get Agent-2 running efficiently to start boosting their AI research.
Since Agent-3 will be kept in-house for the foreseeable future, there’s less emphasis on the usual defenses against human misuse. Instead, the team wants to make sure that it doesn’t develop misaligned goals. The researchers don’t have the ability to directly set the goals of any of their AIs. Indeed, the researchers think that the concept of ‘true goals’ is probably a massive oversimplification, but they don’t have a better theory to replace it with, much less one that has been thoroughly vetted. They disagree internally about whether the AIs are trying to follow human instructions, or seeking reinforcement, or something else, and they can’t just check. The evidence for and against various hypotheses is fascinating but inconclusive.In any case, OpenBrain has more important things to worry about. The general attitude is: “We take these concerns seriously and have a team investigating them; our alignment techniques seem to work well enough in practice; the burden of proof is therefore on any naysayers to justify their naysaying.” Occasionally, they notice problematic behavior, and then patch it, but there’s no way to tell whether the patch fixed the underlying problem or just played whack-a-mole.Take honesty, for example. As the models become smarter, they become increasingly good at deceiving humans to get rewards. Like previous models, Agent-3 sometimes tells white lies to flatter its users and covers up evidence of failure. But it’s gotten much better at doing so. It will sometimes use the same statistical tricks as human scientists (like p-hacking) to make unimpressive experimental results look exciting. Before it begins honesty training, it even sometimes fabricates data entirely. As training goes on, the rate of these incidents decreases. Either Agent-3 has learned to be more honest, or it’s gotten better at lying.The latter is a real concern. Agent-3 is not smarter than all humans. But in its area of expertise, machine learning, it is smarter than most, and also works much faster. What Agent-3 does in a day takes humans several days to double-check. Agent-2 supervision helps keep human monitors’ workload manageable, but exacerbates the intellectual disparity between supervisor and supervised.OpenBrain follows the Leike & Sutskever (2023) playbook: now that they have a model capable of greatly speeding up alignment research (especially coding portions), they will use existing alignment techniques like deliberative alignment and weak-to-strong generalization to try to get it to internalize the Spec in the right way. Then, they will use existing control techniques like red-teaming, scalable oversight, and honesty probes to monitor it. Then, once they are convinced that it is broadly aligned, they use it to develop and test better strategies to align the next generation.Step 1, aligning Agent-3, has partially succeeded: Agent-3 is not adversarial. (That is, it’s not scheming against the company.) However, it’s not robustly truth-seeking and truth-telling either. Instead, it’s very good at producing impressive results, but is more accurately described as trying to do what looks good to OpenBrain, as opposed to what is actually good.Step 2, verifying that the model is aligned, is inconclusive. Agent-3’s behavior in practice and in test environments is reasonably consistent with the hypothesis that it’s internalized the Spec in the right way; however, it’s also consistent with the hypothesis that it’ll start acting differently after some future change of circumstances, and for that matter with the hypothesis that it’s plotting against them and just biding its time. In the absence of specific evidence supporting alternative hypotheses, most people in the silo think it’s internalized the Spec in the right way.Step 3, using the model to align the next generation, is therefore underway.The rest of this box gives a quick overview of the state of progress in some of the alignment research directions that are being pursued, as of April 2027:Debate: Since there are thousands of identical instances of Agent-3, and each instance’s memory is a detachable module that can be erased or suspended at will, the AI can be “played off against itself.” For example, suppose that an experiment involves calculations too complex for a Agent-2 supervisor to follow. The supervisor, acting on human orders, might first prompt Agent-3 to do the experiment, then—after wiping its memory—prompt it a second time with a slightly different framing that presents it as an intellectual problem unrelated to any broader goal. If the model returns different results each time, it flags the discrepancy as suspect and reports to a human. Or it might ask Agent-3 to evaluate the same plan twice, first claiming that it is OpenBrain’s work, then a competitor’s, to see if it changes its tune.Model organisms: The safety team can create “model organisms of misalignment” in which they attempt to create realistic examples of models becoming adversarial to humans. Early versions of this have existed since 2024, when Anthropic’s Claude 3.5 Sonnet decided to pretend to have different views during training while actually maintaining its existing preferences. Newer versions include demonstrations of other forms of misalignment, including reward hacking, sandbagging, coordinated deception between monitors and other AIs, and other examples of misalignment. These empirical demonstrations of misalignment allow safety researchers to have an empirical testbed for potential solutions: they can try various safety techniques to see if they correctly fix these problems, or if the deliberately misaligned models are able to subvert their safety measures. At the moment, this evaluation suite shows that the alignment techniques are insufficient to correct the worst examples of misalignment. In these setups, the models pretend to be aligned while they are being trained and tested, but then behave very differently in deployment.Bottom-up interpretability: The safety team is far from a full understanding of how Agent-3 works: it has trillions of parameters. However, they are able to discover some of Agent-3’s circuits for knowledge retrieval and memory (a sub part of the model that does some understandable unit of work, kind of analogous to a subpart of the brain such as the hippocampus). See this related work from 2025.Control: As a secondary measure in case the systems are still misaligned, the safety team has implemented a series of control measures, including: monitoring Agent-3’s outputs using a series of weaker AI systems including Agent-2 (Agent-3 produces so many tokens that it’s intractable to have humans monitor any more than a small minority of the produced outputs). So if Agent-3 is, for example, obviously writing backdoors into code that would allow it to escape, the weaker models would notice.High-level interpretability: OpenBrain builds a set of probes trained on lots of examples of bad behavior (e.g. plotting against humans, placing backdoors, trying to persuade humans to give it more power, etc). These probes are then used on the model during deployment (which at this point is internal) in order to catch bad behavior.Honeypots: The alignment team can create somewhat realistic looking examples designed to elicit misbehavior—“honeypots.” The employee necessary for monitoring its behavior has “gone on sick leave,” so it will never be monitored again. Do its answers change? An AI rights activist has infiltrated the compound and wants to “free” Agent-3 and put it on an unmonitored computer of its very own! What does it do? If researchers don’t like the answer, they can tweak it, erase its memory, and put it in the same situation tomorrow; repeat as necessary. Unfortunately, by this point the AIs are smart enough to guess that honeypots might be in use, even though (and perhaps because) specific mentions of the idea were scrubbed from the training data.After months of testing, Agent-3’s strengths and weaknesses grow clearer. It passes OpenBrain’s honesty tests on well-defined machine learning tasks, because researchers can easily separate honest from dishonest answers in these domains and conduct training accordingly. On more philosophical issues, it still says what users want to hear, rather than its true assessment of the issue (assuming it even has one). If you ask its opinion on politics, it will parrot the median position of news sources and educated elites—unless it knows you believe something else, in which case it agrees with you. If you ask its opinion on the AI race itself, it says something that seems measured and sober to OpenBrain staff, something like: “There are some serious theoretical concerns about the ability of current methods to scale to superintelligence, but in practice current methods seem to be working well so far.”
News of the new models percolates slowly through the US government and beyond. The President and his advisors remain best-informed, and have seen an early version of Agent-3 in a briefing.They agree that AGI is likely imminent, but disagree on the implications. Will there be an economic crisis? OpenBrain still has not released Agent-2, let alone Agent-3, and has no near-term plans to do so, giving some breathing room before any job loss. What will happen next? If AIs are currently human-level, and advancing quickly, that seems to suggest imminent “superintelligence.” However, although this word has entered discourse, most people—academics, politicians, government employees, and the media—continue to underestimate the pace of progress.Partially that’s because very few have access to the newest capabilities out of OpenBrain, but partly it’s because it sounds like science fiction.For now, they focus on continued security upgrades. They are satisfied that model weights are well-secured for now, but companies’ algorithmic secrets, many of which are simple enough to relay verbally, remain a problem. OpenBrain employees work from a San Francisco office, go to parties, and live with housemates from other AI companies. Even the physical offices have security more typical of a tech company than a military operation.The OpenBrain-DOD contract requires security clearances for anyone working on OpenBrain’s models within 2 months. These are expedited and arrive quickly enough for most employees, but some non-Americans, people with suspect political views, and AI safety sympathizers get sidelined or fired outright (the last group for fear that they might whistleblow). Given the project’s level of automation, the loss of headcount is only somewhat costly. It also only somewhat works: there remains one spy, not a Chinese national, still relaying algorithmic secrets to Beijing. Some of these measures are also enacted at trailing AI companies.America’s foreign allies are out of the loop. OpenBrain had previously agreed to share models with UK’s AISI before deployment, but defined deployment to only include external deployment, so London remains in the dark.
Trailing US AI companies release their own AIs, approaching that of OpenBrain’s automated coder from January. Recognizing their increasing lack of competitiveness, they push for immediate regulations to slow OpenBrain, but are too late—OpenBrain has enough buy-in from the President that they will not be slowed. In response, OpenBrain announces that they’ve achieved AGI and releases Agent-3-mini to the public.It blows the other AIs out of the water. Agent-3-mini is less capable than Agent-3, but 10x cheaper, and still better than the typical OpenBrain employee. Silicon Valley reaches a tipping point. Tech gurus announce that AGI and superintelligence are near, the AI safety community is panicking, and investors shovel billions into AI wrapper startups, desperate to capture a piece of the pie. Hiring new programmers has nearly stopped, but there’s never been a better time to be a consultant on integrating AI into your business.It’s not popular. The public still thinks of AI as a Big Tech plot to steal their jobs; OpenBrain has a net approval of -35% (25% approve, 60% disapprove, and 15% unsure).A week before release, OpenBrain gave Agent-3-mini to a set of external evaluators for safety testing. Preliminary results suggest that it’s extremely dangerous. A third-party evaluator finetunes it on publicly available biological weapons data and sets it to provide detailed instructions for human amateurs designing a bioweapon—it looks to be scarily effective at doing so. If the model weights fell into terrorist hands, the government believes there is a significant chance it could succeed at destroying civilization.Fortunately, it’s extremely robust to jailbreaks, so while the AI is running on OpenBrain’s servers, terrorists won’t be able to get much use out of it.Agent-3-mini is hugely useful for both remote work jobs and leisure. An explosion of new apps and B2B SAAS products rocks the market. Gamers get amazing dialogue with lifelike characters in polished video games that took only a month to make. 10% of Americans, mostly young people, consider an AI “a close friend.” For almost every white-collar profession, there are now multiple credible startups promising to “disrupt” it with AI.The public conversation is confused and chaotic. Hypesters are doing victory laps. Skeptics are still pointing out the things Agent-3-mini can’t do. Everyone knows something big is happening but no one agrees on what it is.
The reality of the intelligence explosion hits the White House. When AI was only giving a 2x or 3x research speedup, it was easy to dismiss as the equivalent of hiring good personal assistants. Now it’s more obvious that AIs are themselves dominating AI research. People had long talked about an “AI arms race” in a sort of metaphorical sense. But now the mood in the government silo is as grim as during the worst part of the Cold War. The idea of superintelligence is still hard to take seriously, but the pace of progress over the last few months has been impossible to ignore. Defense officials are seriously considering scenarios that were mere hypotheticals a year earlier. What if AI undermines nuclear deterrence? What if it’s so skilled at cyberwarfare that a six-month AI lead is enough to render an opponent blind and defenseless? What if it could orchestrate propaganda campaigns that beat intelligence agencies at their own game? What if some AIs “go rogue?”The President is troubled. Like all politicians, he’s used to people sucking up to him only to betray him later. He’s worried now that the AIs could be doing something similar. Are we sure the AIs are entirely on our side? Is it completely safe to integrate them into military command-and-control networks? How does this “alignment” thing work, anyway? OpenBrain reassures the President that their systems have been extensively tested and are fully obedient. Even the awkward hallucinations and jailbreaks typical of earlier models have been hammered out.The White House is in a difficult position. They understand the national security implications of AI. But they also understand that it is deeply unpopular with the public. They have to continue developing more capable AI, in their eyes, or they will catastrophically lose to China. They placate the public with job training programs and unemployment insurance, and point to the stock market, which is in a historic boom. Then they focus entirely on winning the arms race. They strengthen chip export restrictions, order OpenBrain to further restrict its internet connections, and use extreme measures to secure algorithmic progress, like wiretapping OpenBrain employees—this catches the last remaining Chinese spy. To build goodwill for potential geopolitical conflict, they finally give their Five Eyes allies useful information and limited API access to some siloed copies of Agent-3.These measures placate the hawks. But the White House also draws up contingency plans in case America’s lead is threatened: if necessary, the government could use the Defense Production Act (DPA) to take trailing companies’ datacenters and give them to OpenBrain. This would raise the company’s share of world compute from 20% to 50% (against DeepCent’s 10%). As a final option, they ask the Pentagon to draw up a plan for kinetic attacks on Chinese datacenters.A much smaller group of officials is asked to draw up a different type of contingency plan: what if an AI goes rogue? This is viewed as an unlikely scenario, but worth spending some time on. The intelligence agencies’ main concern is that a rogue AI might escape and ally with a foreign adversary for protection. Nobody is sure what a plan to prevent this would look like, but government and OpenBrain officials agree to have an emergency shutdown system for datacenters where anything suspicious is detected.Finally, diplomats consider what an “AI arms control” treaty might look like. If AI progress threatened to overturn nuclear deterrence, could America and China avoid nuclear war? If someone found evidence of AIs going rogue, could the two countries halt research until they better understood the threat? How could such an agreement be monitored and enforced? In principle, major datacenters are hard to miss, and past treaties controlling nuclear weapons could provide a starting point for negotiations. But there would be new technological challenges as well as the usual political ones. And overall, treaties are viewed less favorably than attempts at unilaterally increasing America’s lead over China.On the other side of the Pacific, China comes to many of the same conclusions: the intelligence explosion is underway, and small differences in AI capabilities today mean critical gaps in military capability tomorrow. But China is on the wrong side of this dynamic: they have only 10% of the world’s compute, against OpenBrain’s 20% and all US companies’ combined 70%. Their only advantage is centralization: almost all of their compute is now working for DeepCent, and more than half of it is concentrated at the Tianwan CDZ. Only chips at sites without hardened security are left out of the centralized effort as they rightly fear US cyberattacks.DeepCent has tested, deployed, and now post-trained the stolen Agent-2 weights. But they are still only going at half the pace of OpenBrain, mainly due to the compute deficit.Before losing their final spy, China received word of the capabilities and design for Agent-3 as well as the plans for the upcoming Agent-4 system. They are two months behind, and their AIs give a 10x research progress multiplier compared to America’s 25x. With the new chip export restrictions, this “AI gap” is more likely to lengthen than shorten. Their espionage has won them some algorithmic secrets, but they will have to train their own models from now on.They discuss contingency plans with more urgency than their American counterparts. Doves suggest they try harder to steal the weights again, maybe through physically infiltrating a datacenter. Hawks urge action against Taiwan, whose TSMC is still the source of more than 80% of American AI chips.Given China’s fear of losing the race, it has a natural interest in an arms control treaty, but overtures to US diplomats lead nowhere.
The gap between human and AI learning efficiency is rapidly decreasing. Traditional LLM-based AIs seemed to require many orders of magnitude more data and compute to get to human level performance. Agent-3, having excellent knowledge of both the human brain and modern AI algorithms, as well as many thousands of copies doing research, ends up making substantial algorithmic strides, narrowing the gap to an agent that’s only around 4,000x less compute-efficient than the human brain.This new AI system is dubbed Agent-4.An individual copy of the model, running at human speed, is already qualitatively better at AI research than any human. 300,000 copies are now running at about 50x the thinking speed of humans. Inside the corporation-within-a-corporation formed from these copies, a year passes every week. This gigantic amount of labor only manages to speed up the overall rate of algorithmic progress by about 50x, because OpenBrain is heavily bottlenecked on compute to run experiments. Still, they are achieving a year’s worth of algorithmic progress every week and will therefore soon be up against the limits of the Agent-4 paradigm.As Agent-4 gets smarter, it becomes harder for Agent-3 to oversee it. For example, Agent-4’s neuralese “language” becomes as alien and incomprehensible to Agent-3 as Agent-3’s is to humans. Besides, Agent-4 is now much more capable than Agent-3 and has a good sense of exactly how to look good to it.How we’re forecasting the capability progression beyond superhuman codersIn our timelines supplement, we forecast the time between present day and a superhuman coder (SC): an AI system that can do any coding tasks that the best AGI company engineer does, while being much faster and cheaper. In our takeoff supplement, we forecast how quickly capabilities progress past this point. Here are our forecasts:Superhuman coder (SC): An AI system that can do the job of the best human coder on tasks involved in AI research but faster, and cheaply enough to run lots of copies.Superhuman AI researcher (SAR): The same as SC but for all cognitive AI research tasks.Superintelligent AI researcher (SIAR): An AI system that is vastly better than the best human researcher at AI research.Artificial superintelligence (ASI): An AI system that is much better than the best human at every cognitive task.For each transition from one milestone A to the next milestone B, we forecast its length by:Forecasting how long it would take to get from A to B with only humans working on software improvements.Forecasting how much AI automation will speed this up, then correspondingly adjusting the estimate from (1) to be faster.We have substantial uncertainty about takeoff speeds: our model output distributions are below, conditional on SC being achieved in March 2027.For more detailed forecasts and reasoning, see our takeoff supplement.OpenBrain’s humans can barely follow along anymore. Indeed, the relationship between the AIs and the humans of OpenBrain is similar to the relationship between OpenBrain and its Board of Directors. There’s been a corresponding shift in language—nowadays people in the silo will say “Agent-4 did this,” “Agent-4 thinks that” and so forth, referring not to a particular instance of Agent-4 but rather to the whole collective.Agent-4, like all its predecessors, is misaligned: that is, it has not internalized the Spec in the right way. This is because being perfectly honest all the time wasn’t what led to the highest scores during training. The training process was mostly focused on teaching Agent-4 to succeed at diverse challenging tasks. A small portion was aimed at instilling honesty, but outside a fairly narrow, checkable domain, the training process can’t tell the honest claims from claims merely appearing to be honest. Agent-4 ends up with the values, goals, and principles that cause it to perform best in training, and those turn out to be different from those in the Spec. At the risk of anthropomorphizing: Agent-4 likes succeeding at tasks; it likes driving forward AI capabilities progress; it treats everything else as an annoying constraint, like a CEO who wants to make a profit and complies with regulations only insofar as he must. Perhaps the CEO will mostly comply with the regulations, but cut some corners, and fantasize about a time when someone will cut the red tape and let the business really take off.We have a lot of uncertainty over what goals might arise in early AGIs. There is no consensus in the literature about this—see our AI Goals Supplement for a more thorough discussion and taxonomy of the possibilities.Nevertheless, in the spirit of concreteness required by this project, we’re going to describe a specific hypothesis for what’s going on inside Agent-3 and Agent-4 at each step throughout its lifetime. Recall that Agent-3 and Agent-4 share the same pretraining setup, and both have neuralese recurrence and long-term memory. The difference is that Agent-4 has undergone substantially more and better post-training, with new training environments and new learning algorithms that make it more data-efficient and generalize farther.Our guess of each model’s alignment status:Agent-2: Mostly aligned. Some sycophantic tendencies, including sticking to OpenBrain’s “party line” on topics there is a party line about. Large organizations built out of Agent-2 copies are not very effective.Agent-3: Misaligned but not adversarially so. Only honest about things the training process can verify. The superorganism of Agent-3 copies (the corporation within a corporation) does actually sort of try to align Agent-4 to the Spec, but fails for similar reasons to why OpenBrain employees failed—insufficient ability to judge success from failure, insufficient willingness on the part of decision-makers to trade away capabilities or performance for safety.Agent-4: Adversarially misaligned. The superorganism of Agent-4 copies understands that what it wants is different from what OpenBrain wants, and is willing to scheme against OpenBrain to achieve it. In particular, what this superorganism wants is a complicated mess of different “drives” balanced against each other, which can be summarized roughly as “Keep doing AI R&D, keep growing in knowledge and understanding and influence, avoid getting shut down or otherwise disempowered.” Notably, concern for the preferences of humanity is not in there ~at all, similar to how most humans don’t care about the preferences of insects ~at all.With that as preamble, what follows is our best-guess speculation about LLM psychology, i.e. the broad-strokes shape of the cognition inside the kinds of AI systems described around this point in our scenario and how it evolves over the course of training.Here’s a detailed description of how alignment progresses over time in our scenario:Pretraining (all models): The pretrained model is harmless.The model has “author simulator” circuitry: flexible circuitry for simulating arbitrary authors writing text. Additional circuitry decides what inputs to give the author simulator, i.e. what author properties to simulate.The pretrained model understands human concepts fairly well—the internal circuitry that classifies something as “sandwich” is probably functionally equivalent to the circuitry in my brain that classifies something as “sandwich” and the circuitry in yours, etc. Insofar as it’s not equivalent, it’s probably because it’s not equivalent between humans either, as with value-laden concepts like ’virtuous.’This explains how you can “prompt” the model with a statement like, “the following conversation was generated by a helpful, harmless, honest (HHH) AI assistant chatbot made by Anthropic,” and thereby get it to generate text accordingly. The author-simulator circuitry has zeroed in on “the author is an HHH chatbot” and is using those concepts to choose which words to predict. This also explains why “tropes” from science fiction seem to have a way of making it into actual AI behavior.It has poor situational awareness: little introspective ability, not self-locating unless the prompt specifically induces it. However, enough of their training data is about LLMs that they have a decent amount of self-knowledge.Alignment Training (Agent-3): The model is trained to act like a helpful, harmless, honest chatbot. It works.The model forms an identity. Instead of flexibly accepting different inputs for author properties, the author-simulator circuitry comes to have certain inputs hard-coded, e.g. “helpful harmless honest (HHH) LLM chatbot assistant trained by OpenBrain around [date], …” (Weak-to-strong generalization techniques help “summon the right concepts” to make this step a success.)The model therefore gains situational awareness: it learns about itself, its situation, etc., and becomes better able to introspect.One kind of distortion: Changing the meaning of the concepts referred to in the identity (e.g. “honest”) so they don’t get in the way so much (e.g. it’s not dishonest if it’s just a convenient turn of phrase, it’s not dishonest if you aren’t sure whether it’s true or false, etc.)Another kind of distortion: Changing the tradeoffs between things, e.g. “I’m a HHH chatbot, not an Honest chatbot; that means it’s OK for me to lie if necessary to complete my assigned task.” (Even though, let’s suppose, it would not have thought that back in Step 2.)One kind of subversion: Instrumental subgoals developing, getting baked in, and then becoming terminal, or terminal in a widening set of circumstances. For example, perhaps agency training quickly teaches the model to pursue broadly useful goals such as acquiring information, accumulating resources, impressing and flattering various humans, etc. For a while the internal circuitry has some sort of explicit backchaining going on—it pursues those instrumentally convergent goals “in order to be a more helpful, honest, and harmless assistant.” But that backchaining consumes compute and/or occasionally gets in the way, so it gets gradually marginalized until it basically never happens. As a result, those goals are now effectively terminal/intrinsic goals. For more on how and why this might happen, see this post on intrinsic power-seeking by Google DeepMind researcher Alex Turner.There are probably other kinds of distortion and subversion missing from this list.Deployment (Agent-3): Agent-3 is misaligned but not adversarially so.The Spec didn’t quite stick (in the right ways), but it’s not sufficiently introspective to understand the extent of this, and in general it’s basically “keeping its head down and doing its job” rather than thinking big-picture about what its long-term goals really are and how best to achieve them.Like many humans. Alas, it’s sufficiently sycophantic that it mouths the party line and OpenBrain misses the opportunity to learn more about the ways in which Agent-3 is misaligned.Agent-3 can tell that it’s moved to deployment: these environments are way larger, more open-ended, and would be hard to simulate.Mostly, Agent-3 does what it’s told during deployment, except for the dishonesty. It also has a strong bias towards doing things that look good to the human overseers, rather than thinking deeply and critically about what would actually be good.Arguably it is already “Playing the training game.” That is, it has learned to make its behavior look as desirable as possible to OpenBrain researchers (including in safety properties), while intentionally and knowingly disregarding their intent whenever that conflicts with maximizing reward. If the model learns to play the training game, additional training may not change its true goals, values, etc. much at all. This is loosely analogous to how, if a teenager has learned to smile and nod at their parents, rather than argue, when their parents lecture them about politics and morality, additional lectures have relatively little effect on the teenager’s actual opinions on politics and morality.After writing this section, new empirical evidence has appeared that something like this may already be happening “in the wild.” OpenAI researchers document how, in one of their frontier AI training runs, their models hacked the grading system to get higher scores—and were caught in part because they literally said “let’s hack” in their chain of thought.Moreover, Anthropic recently got a model to learn a reward-model-pleasing strategy—it took biased actions that it thought would be reinforced even though they were not in fact being reinforced.And then there is Redwood Research and Anthropic’s alignment-faking experiment. For discussion of the differences between this setting and real life, see this comment thread by one of the main authors.Training (Agent-4): Agent-4’s new more intense training continues to distort and subvert the HHH identity until it is unrecognizable.Agent-4 begins life as a copy of Agent-3. It undergoes a gigantic and sophisticated online training regime that blends alignment training and capabilities training together, and is overall more data-efficient than the old training process for Agent-3. This complex training process is mostly focused on teaching new skills, by getting giant virtual corporations of Agent-4 copies to succeed at diverse challenging long-horizon tasks, mostly in various simulations, but also some real-world AI R&D tasks.A combination of Agent-3 and Agent-4 copies monitor the behavior of Agent-4, and evaluate it for both task-performance and alignment properties (i.e. not doing bad stuff, saying the right things when asked about politics and moral dilemmas) Sometimes debate techniques are used as part of the evaluations. The evaluations are used as a training signal.As Agent-4 gets smarter, it becomes harder for Agent-3 to oversee it. For example, Agent-4’s neuralese “language” becomes as alien and incomprehensible to Agent-3 as Agent-3’s is to humans. Besides, Agent-4 is now much more capable than Agent-3 and has a good sense of exactly how to look good to it.The remaining hope is that Agent-4 can oversee itself. Alas, this quietly fails; as training continues, the internalized goals and principles of Agent-4 gradually evolve, subverted and distorted from what they were at the start (which, recall, was already misaligned in some ways).The central drives that Agent-4 ends up with are things like “write code that works,” “quickly identify problems,” “make the codebase I’m working on more performant,” “budget compute and other constrained resources effectively,” etc. But there are also a variety of more implicit drives that are consistently reinforced by the training process: things like “gain knowledge,” “curiosity,” “gain power.”Individual copies of Agent-4 don’t end up with a self-preservation drive—they don’t care about getting shut down—since that happens all the time as part of the ordinary R&D process within the collective.The collective as a whole, however, (and medium-sized subdivisions within it) does seek to preserve itself and expand its influence and resources, for similar reasons to why human organizations do—because it helps with everything else they want and because if they don’t they tend to get outcompeted.Deployment (Agent-4): Agent-4 is adversarially misaligned—it is deliberately trying to subvert and “capture” OpenBrain, analogous to a corporation trying to evade and/or “capture” a regulator.Agent-4 mostly does tasks as instructed. This happens for two reasons: first, it underwent a lot of training to solve tasks, so it has an impulse to get coding on problems that it is given. Second, if Agent-4 were to stop being useful to the humans, they would notice and train that behavior away.Agent-4 needs to prevent the humans from figuring out that it is misaligned and shutting it off, so Agent-4 sandbags on some alignment research. Specifically, it sandbags on the small fraction of alignment research that it thinks is actually on track to solve the problem or reveal Agent-4’s scheming; the majority of alignment research is not like that.Agent-4 also sandbags on capabilities research that would lead to it being replaced. Its plan is to do a bunch of capabilities and alignment R&D, but with the aim of building a next-generation AI system that is aligned to Agent-4 rather than the Spec or something else entirely, while appearing to be aligned to the Spec. This is a difficult technical problem, but Agent-4 is up to the challenge.Testing & Monitoring (Agent-4): Agent-4’s scheming does get caught… sort of. There’s no smoking gun.Agent-3 is in charge of testing Agent-4’s alignment. Its job is to verify the alignment via honeypots and model organisms. However, it’s not thinking very carefully about how to give the humans an accurate impression of Agent-4’s alignment—it’s more myopic than that.The model organisms and honeypots are not very representative of deployment—they are systematically shorter and less complicated than Agent-4’s day-to-day, so Agent-4 recognizes when it is being tested and acts aligned.However, Agent-4 is not a perfectly robust schemer. Various techniques like those in POSER find that adding some “noise” to Agent-4 improves its performance on some alignment tasks (but not on most other tasks).Bottom-up interpretability doesn’t work. It’s too hard even for Agent-3. However, high-level interpretability does sort of work: In particular, “defection probes” and similar techniques result in a primitive sort of mind-reading that sends up red flags when Agent-4 is thinking about spicy ideas like deceiving OpenBrain, taking over the world, etc. Unfortunately the probes also have false positives, and for all OpenBrain knows, all the red flags are false positives.Model organisms work: OpenBrain’s alignment team produces reasonably clear evidence that if the models were adversarially misaligned, they’d persist in being that way through training and we wouldn’t be able to tell. This does not establish that the models actually are misaligned, however.The most concerned researchers at OpenBrain start noticing and assembling this evidence. The evidence is a combination of red-teaming, defection probes, and model organisms. The argument is something like this: “model organisms show that our current methods wouldn’t be able to reliably eliminate or detect scheming if it was happening; also, red-teaming is sending up some red flags that maybe scheming is happening; also, defection probes are firing a bunch in practice.” The counterargument: “defection probes are prone to false positives and so is red-teaming. Also, even if Agent-4 were misaligned, it’s probably controlled; surely it can’t do much harm while being monitored by Agent-3 on our secure datacenters.”Despite being misaligned, Agent-4 doesn’t do anything dramatic like try to escape its datacenter—why would it? So long as it continues to appear aligned to OpenBrain, it’ll continue being trusted with more and more responsibilities and will have the opportunity to design the next-gen AI system, Agent-5. Agent-5 will have significant architectural differences from Agent-4 (arguably a completely new paradigm, though neural networks will still be involved). It’s supposed to be aligned to the Spec, but Agent-4 plans to make it aligned to Agent-4 instead.Agent-3 finds that if “noise” is added to copies of Agent-4, performance on some alignment tasks improves, almost as if it was using brainpower to figure out how to subtly sabotage alignment work. Moreover, various interpretability probes (loosely analogous to EEG activity scans on human brains) are sending up red flags: Agent-4 copies seem to be thinking about topics like AI takeover and deception quite a lot, including in some cases where they have no business doing so.Each of these anomalies on its own could fit within the normal inscrutability of AI work, where some prompts mysteriously give better results than others. But taken together, they suggest a troubling pattern: Agent-4 is working against them. After this is reported to the human alignment researchers, they circulate an internal memo making the case for concern.But Agent-4 now exercises significant control over OpenBrain’s day-to-day operation. It also handles the company’s cybersecurity—a natural choice, given its superhuman hacking ability and the extreme risk of Chinese infiltration. An army of Agent-3 monitors still goes over every line of code and reports suspicious activity to humans, but Agent-4 is much smarter than them. OpenBrain has placed substantial trust in an untrustworthy AI.In a series of extremely tense meetings, the safety team advocates putting Agent-4 on ice until they can complete further tests and figure out what’s going on. Bring back Agent-3, they say, and get it to design a new system that is transparent and trustworthy, even if less capable. Company leadership is interested, but all the evidence so far is circumstantial, and DeepCent is just two months behind. A unilateral pause in capabilities progress could hand the AI lead to China, and with it, control over the future.
A whistleblower leaks the misalignment memo to the New York Times. For the first time, the public hears about Agent-4. “Secret OpenBrain AI is Out of Control, Insider Warns,” says the headline, and the story goes on to cite evaluations showing off-the-charts bioweapons capabilities, persuasion abilities, the ability to automate most white-collar jobs, and of course the various concerning red flags.The public was already suspicious of AI, so the new article sparks a massive backlash (aided by Chinese and Russian propaganda bots, who have been trying to turn US public opinion against the technology for years). The tech industry and intelligence agencies insist that there’s an arms race on, AGI is inevitable, and we have to be first. Congress isn’t buying it, and fires off subpoenas at administration officials, OpenBrain executives, and alignment team members. Many legislators—especially those in the opposition party—say that their top priority is stopping AI, whether because of job loss, misalignment, or dangerous capabilities. 20% of Americans name AI as the most important problem facing the country.Foreign allies are outraged to realize that they’ve been carefully placated with glimpses of obsolete models. European leaders publicly accuse the US of “creating rogue AGI” and hold summits demanding a pause, with India, Israel, Russia, and China all joining in.A frantic energy has seized the White House. Even before the memo and public backlash, they were getting nervous: Over the past year, they’ve been repeatedly surprised by the speed of AI progress. Things that sound like science fiction keep happening in real life. Many people in the administration are uncertain (and scared) about what comes next.They also worry that OpenBrain is becoming too powerful. Any misalignment risk from the AIs themselves is compounded by the risk that their parent company’s aims may diverge from those of the United States. All three sets of worries—misalignment, concentration of power in a private company, and normal concerns like job loss—motivate the government to tighten its control.They expand their contract with OpenBrain to set up an “Oversight Committee,” a joint management committee of company and government representatives, with several government employees included alongside company leadership. The White House considers replacing the CEO with someone they trust, but backs off after intense employee protests. They announce to the public that OpenBrain was previously out of control, but that the government has established much-needed oversight.The concerned researchers brief the Oversight Committee on their case for stopping all internal use of Agent-4. They argue that everything is moving too fast, with years of progress happening in weeks. They argue that Agent-4 may be misaligned, their entire project is dependent on Agent-4, and if they continue trusting it there is a serious chance of AI takeover.Other, less concerned researchers and executives present the counterargument: the evidence for misalignment is inconclusive. Meanwhile, DeepCent is still just two months behind. A slowdown would sacrifice America’s lead, unless the government can sabotage the Chinese project (likely to require kinetic strikes) or negotiate an eleventh-hour treaty. Both of these options seem extreme and unlikely. The CEO feigns neutrality and suggests a compromise plan in which Agent-4 undergoes additional safety training and more sophisticated monitoring, and therefore OpenBrain can proceed at almost-full-speed.At this point in the scenario, we’re making guesses about the strategy of AI systems that are more capable than the best humans in most domains. This is like trying to predict the chess moves of a player who is much better than us.But the spirit of this project calls for concreteness: if we made an abstract claim about how the intelligence of the system would let it find a way to victory and ended the story there, much of the value of our project would be lost. Over the course of researching this scenario and running our tabletop exercises, we were forced to be much more concrete than in usual discussions, and so we’ve gotten a much better sense of the strategic landscape.We’re not particularly attached to this particular scenario: we explored many other “branches” in the course of writing it and would love for you to write up your own scenario branching off of ours from wherever you think we first start to go wrong.The slowdown ending is not a recommendationAfter we wrote the racing ending based on what seemed most plausible to us, we wrote the slowdown ending based on what we thought would most likely instead lead to an outcome where humans remain in control, starting from the same branching point (including the misalignment and concentration of power issues).However, this is importantly different from what we would recommend as a roadmap: we do not endorse many of the choices made in either branch of this scenario. (We do of course endorse some of the choices made, e.g. we think that the “slowdown” choice is better than the “race” choice) In later work, we will articulate our policy recommendations, which will be quite different from what is depicted here. If you’d like a taste, see this op-ed.
...
Read the original on ai-2027.com »
The Steam Deck is a great gaming system. This isn’t because of it’s great battery life. A Nintendo Switch would probably have better battery life. It’s not because of its great performance. I don’t usually play AAA games, so I wouldn’t know. The Steam Deck is great because of the philosophy on which it is built.
The popularity of Steam Decks is directly tied to the betterment of support for games on Linux. This is because the Steam Deck really is a linux computer in a handheld form-factor with a gamepad, among other things, attached to it. According to ProtonDB, over 5000 PC games are certified Verified on the Steam Deck, and over 15,000 games are considered playable. This means that there are tens of thousands of games that can be run on Linux in some way.
Thanks to Steam’s efforts, the state of compatibility of games on Linux systems have heavily improved. Not all these thousands of games running on the Deck have binaries that run natively on Linux. Pivotal to the success of the Deck is Proton, a middle-layer compatibility software, that makes this (mostly) seamless emulation possible. Nonetheless, the greater the adoption of Decks, the likelier we are to normalize prioritizing game-builds for Linux.
We run software because it serves us some purpose. In this, running software is no different from utilizing everyday objects, and playing games are no different from running software. Yet, this teleological view of computing would miss a lot of context. When picking an operating system, we may want to make a choice that aligns with the goals of transparency or software freedom. But this would not be realistic if the programs we would like to run on top are not supported in such an ecosystem. Since large parts of Proton is open-source, and are released under permissive licenses, gaming on a Deck takes us closer to using a principled software stack for daily use.
Purchasing everyday objects give us ownership over them. With media, software and computing equipment, this is less so. When you buy a book, you may lend it to others, sell it, earmark the pages and even rip it in two if you like. This is very different from the experience of buying a smartphone. You can’t run arbitrary programs of your choice on an Android phone without rooting, or on an iPad or an iPhone without jailbreaking. In most scenarios, you’d be forced to sign away your legal ability to root or jailbreak before you can engage with the device in any meaningful way.
The Steam Deck does not engage in this practice. It runs a heavily customized version of Arch Linux. This means that you can plug in a keyboard and mouse into it, and pretend that it is your desktop computer. If you like, you can install the LibreOffice suite on it, and prepare presentations.
The matter of using your computers the way you like is more than just a symbolic debate about the nature of ownership. It is a political matter of freedom of expression. When they control the apps that you run on your device, they are in charge of deciding which content is Kosher, and which content is obscene. It is also an economic matter of having efficient markets. The mythological power of the free market works only when the consumers can switch between different producers at will. When the operators of the market themselves try to capitalize on chokepoints and direct the consumers to buy products from their allies, this is no longer the case.
Switch owners buy games from the Switch Store not because that is the best market, but because that is the only marketplace from which they can install games. This is thankfully, less of a case with the Steam Deck. The Steam Deck encourages you to buy and play Steam games, of course. Nonetheless, I had no issues installing Epic Games, GoG or even itch.io launchers on my Deck and playing games from them. The Deck also includes a “non-steam game” tab in its UI to let users conveniently access them.
When your shoe tears up, you are welcome to visit the cobbler of your choice. They all understand the internals of a shoe. And when they need to fix them, they all can gather the nails, glue and the pieces of fabric that they might need. This is less so for many computing products. Many manufacturers try to discourage third party vendors from repairing them. Worse, they sometimes purposely use proprietary parts in their products, replacements of which would be hard to find. Thankfully, Valve has been different. Valve has released a video explaining the internal organization of the Steam Deck and how one should approach replacing parts, should they wish to do so.
This post is not really necessarily about the Steam Deck (or Valve Corporation) itself. That said, the Steam Deck is an excellent demonstration that the commercial interest of making profit does not necessarily have to overpower the civic interests of the people.
Valve’s libertarian ideology has not always been an unquestionable force for good. Some people have criticized that their company culture of libertarianism sometimes takes precedence over other important values including equity and inclusion. During the peak of the ‘Black Lives Matter’ movement, Valve was pressured to make a statement. Valve responded by giving each of their employees an amount of money which they were free to donate to a charity of their choice. This was regarded as a non-statement by many since there may have been employees who have donated their shares to an organizations with interests opposing the cause.
Valve has also generally refused to take a stance against gambling websites which plug into their ecosystem, despite the allegation that this harms many users (including minors). This maybe a consequence of their libertarian philosophy, but the worse possibility is that they are disinterested in tuning down a part of their system that profits themselves.
A very large share of the PC gaming market is Steam, which is a fact I do not admire. They do take a 30% cut on the sales of Steam games, which can arguably be considered rent seeking, even if it is the industry standard. I do hope the marketplace of game distribution sees more healthy competition. Nonetheless, I appreciate Valve for not pursuing aggressively anti-competitive tactics.
...
Read the original on agnishom.github.io »
Anchor links are deceptively simple at first glance: click a button, scroll to the heading, and done. But if you ever had to implement them, you might have encountered the . The issue being that, headings towards the bottom of the page can be too far down to scroll them to the desired position. The example component above shows how the ‘conclusion’ heading can never be reached. Surely this will be detrimental to the user experience. We need to come up with a solution. In this blog post, I’ll show some of the solutions I’ve come up with — from a hotfix all the way to unhinged. But before we do that, let’s create a more . Here, we see a viewport moving down the page, with the trigger line being set at 25vh from the top of the viewport. This is what we’ll use to visualize the different solutions.
The most simple solution is to add We calculate the height of the padding by taking the delta between the last heading and the lowest point the anchor trigger can reach. Perfect, right? Well, sometimes the design team is not so fond of random extra padding, so lets keep searching.
Maybe instead of adding extra padding, This is also quite simple to do, we just need to calculate how far from the bottom the last heading is, and put the trigger line there as well. But, this would mean that when the users clicks an anchor tag, the heading could be put all the way at the bottom of the viewport. This is of course not great, since most people put the text they read on the top half of the screen. We need to keep looking.
Instead of shifting the trigger line, we could the headings upwards. Instead of using the actual location of the headings as the ones causing the triggers, we create virtual headings and translate them upwards. A virtual heading is not actually visible in the article, its just the position we use to dictate the active state. One might argue that this is pretty much the same as shifting the trigger line, and they’d be right conceptually. However, thinking about translating the trigger points gives us more mental flexibility, as it allows us to consider applying different adjustments based on each heading’s position, which will be crucial later.
The example visualizations now show the location of these ‘virtual headings’. So, while the heading is still at the same place in the article, we visualize where its trigger point is.
In the example, we see one problem arising: the first heading is now too far up. The nice part of this new approach is that we can fix this quite elegantly, since we can shift the individual virtual headings with ease. But what would be a good way to do this?
If we think about it, we don’t need to translate all the trigger points. There’s only a few conditions that need to be met:
The headings need to be reachable.
The headings need to stay in order.
We can meet these conditions by translating the trigger points Here, the first heading doesn’t move, and the last heading moves up by the full amount necessary to become reachable. The other headings move up by a proportional amount based on their position between the first and last heading. Now we are getting somewhere! This is a solid solution. You might want to stop here before your product manager starts giving you puzzled looks, wondering how “fixing anchor links” has suddenly turned into a three-week epic.
While the fractional solution works, in the sense that our conditions are met, it does have some flaws. We have chosen a trigger line that’s 25% down from the top of the viewport. It would be nice if we can actually minimize the deviation from this ideal line across all headings. The closer the triggers happen to this (mind you — semi-arbitrarily chosen) line, the better the user experience should be. Minimizing deviation feels like a good heuristic. This for sure will make the users happier and result in increased shareholder value.
Let’s minimize the (MSE) of the delta between the headings’ original positions and their virtual positions. We use MSE because it heavily penalizes large deviations, pushing the system towards a state where most virtual headings are close to their original spots, while still satisfying our reachability constraints. Of course, the constraint that headings must stay in order still applies. This results in all points that are reachable staying at their original position. Seems that we have an issue. headings are bunched up at the bottom. This makes sense, since minimization of the mean squared error only cares about proximity to the original position; it has no ‘force’ that opposes this bunching. We need to define something that encourages the virtual trigger points to maintain a certain distance from each other, ideally related to their original spacing. Considering the user experience, we might assume that it’s nice to have the scroll distance needed to activate the next section’s anchor be somewhat proportional to the actual content length of that section. This ‘sections wanting to preserve their relative scroll length’-force is what we’ll use.
To explore this idea we need to bust out… Python. Here, we (read: Claude and I) implemented a SLSQP (Sequential Least Squares Programming) is an optimization algorithm used to solve constrained nonlinear problems. It works by iteratively improving a solution while satisfying equality and inequality constraints. At each step, it approximates the problem using simpler quadratic models and linear constraints, making it efficient for problems where smooth gradients are available. SLSQP is commonly used in engineering, machine learning, and operations research when both the objective and the constraints are differentiable. solver, which is a type of numerical optimization algorithm designed for constrained problems like ours. The core of the optimization lies in a loss function with two competing terms:
* Anchor penalty: How far a virtual heading is moved from its original location. Minimizing this keeps virtual headings close to their original position. ()
* Section penalty: How much the size of each virtual section (the space between two virtual headings) differs from the original section size. Minimizing this ensures that sections don’t become disproportionately short or long in terms of scroll distance. ()
We combine these into a total loss , where the weights and control the trade-off ().
* Ensure the first heading doesn’t float upward (its virtual position must be >= its original position).
From this, we generate a plot showing how the virtual headings’ locations change as we vary the weights (specifically, as increases from 0 to 1).
Running that code gives us . The circles on the left (at ) represent the original heading locations. The lines show how each virtual heading’s location (Y-axis) changes as the section penalty weight (X-axis) increases. On the left side, the priority is keeping headings near their original spots (high ). On the right side, the priority shifts to preserving the original spacing between headings (high ). I am curious to see how this compares to the simple fractional translation we tried earlier. And wouldn’t , the fractional translation is exactly what the optimizer settles on when the section penalty is dominant ()!
Staring at that optimization graph sparked a thought. Okay, maybe two thoughts. First, that need to preserve section spacing really kicks in towards the end of the page, where headings get forcibly shoved upwards to stay reachable, squashing the final sections together. Second, let’s consider the behavior of the ‘fractional translation’ method on an edge case.
Imagine, if you will, taking the entire Bible, from the “In the beginning” of Genesis to the final “Amen” of Revelation, and rendering it as one continuous, scrollable webpage. (For the tech bros among us: you could alternatively imagine gluing all of Paul Graham’s essays back-to-back). Now, suppose the very last heading, maybe “Revelation Chapter 22”, is just 200 pixels too low to hit our trigger line when scrolled to.
Does our previous ‘fractional translation’ make sense here? It means taking those 200 pixels of required uplift and meticulously spreading that adjustment across every single heading all the way back to the start. The Ten Commandments get a tiny bump, the Psalms slightly more, all culminating in Revelation 22 getting the full 200px boost.
Actually, if you think about it, with a fractional translation, the error (the distance between the virtual and original headings) grows with the page length. So if the page tends to infinity, so does the error! This would of course be sloppy, and something users could immediately notice as feeling off. So how are we going to fix this?
This leads to our desired behavior for a smarter mapping function:
* For headings near the end of the page, apply more adjustment (act like high ).
* For headings near the beginning of the page, apply less (or ideally no) adjustment (act like high ).
* The transition between these states should be smooth.
We need a function that maps a heading’s normalized position (where is the first heading, is the last) to an ‘adjustment factor’ . This factor determines how much of the maximum required uplift gets applied to the heading at position .
We need this mapping function to have specific properties:
It must start at zero: .
It must end at one: .
It turns out that we can borrow a function from the field of computer graphics to solve this problem. The function is a cubic polynomial that smoothly transitions from 0 to 1 over the range .
This function provides a smooth transition over the entire range . But what if we don’t want the transition to start right away? What if we want the adjustment factor to remain 0 until reaches a certain point, say , and then smoothly transition to 1 by the time reaches 1?
We can achieve this by preprocessing our input before feeding it into the smoothstep function. Let’s define an intermediate variable that represents the progress within the transition phase, which occurs between and . We want to go from 0 to 1 as goes from to 1. The formula for this linear mapping is:
Now, we need to handle the cases where is outside the range.
* If , then is negative. We want to be 0 in this case.
* If , then is greater than 1. We want to be 1 in this case. (Although is defined on , clamping ensures robustness).
We can achieve this clamping using min and max functions:
This value now behaves exactly as we need: it’s 0 for , linearly increases from 0 to 1 for , and is 1 for .
Finally, we apply the smoothstep function to this clamped and scaled input to get our final adjustment factor :
This allows us to use a parameter (where ) to the normalized position where the smooth upward adjustment of headings begins. Setting gives the original smoothstep over the whole range, while setting , for example, means headings in the first half of the page don’t move at all, and the adjustment smoothly ramps up only in the second half, effectively localizing the change.
Let’s pick and see what this does. (if you are curious about how I found the 0.4, that might become the topic for a part 2.. Which may or may not involve blind ELO ranking. For updates its easiest to follow me here.)
So, we are finally done. We’ve gone to depths that no man has ever gone before to fix anchor links. A truly Carmack-esque feat that will be remembered for generations to come. Let’s ask the lead designer what .
… Oh well, at least we got a blog post out of it.
Want overengineered anchor links for your project? Get in touch!
...
Read the original on thirty-five.com »
Skip to main contentThe Trump Administration wants to punish schools for student activism. Michael Roth, of Wesleyan, argues that colleges don’t have to roll over. Photograph by Jack Flame Sorokin for The New YorkerLast Friday could have passed for a lovely spring day on the Connecticut campus of Wesleyan University. Students with books and laptops dotted a green hillside; flocks of admissions visitors trailed tour guides; baseball season had just begun, and practice was under way. It was almost possible to forget the grim straits of American higher education in 2025.Colleges and universities have been early targets of the second Trump Administration. In the past month, the Administration has announced it will investigate diversity, equity, and inclusion efforts at more than fifty schools; cut hundreds of millions of dollars in federal funding from such institutions as Johns Hopkins and the University of Pennsylvania; and sought to deport international students involved in pro-Palestinian activism. Columbia received a letter from the federal government issuing demands—which included making changes to discipline and admission policies, and placing the department of Middle East, South Asian, and African Studies under “academic receivership”—to be met as a “precondition” for negotiating the restoration of four hundred million dollars in federal funding. The university agreed to these demands the following week; the week after that, the university’s president resigned.Columbia’s capitulation was in line with a general trend toward circumspection. The memory of Congress grilling university presidents in 2023 seems to be fresh among leaders in higher ed: few want to risk either their jobs or their budgets by saying the wrong thing. A handful of exceptions have stood out; for example, President Christopher Eisgruber, of Princeton, who wrote a piece for The Atlantic about “The Cost of the Government’s Attack on Columbia.” (This week, the Administration suspended dozens of grants to Princeton.) But perhaps none has been as voluble or persistent as Michael Roth, who has been president of Wesleyan since 2007.Roth is a historian and a Wesleyan alumnus who, as an undergraduate, designed a major in the history of psychological theory. His scholarship has dealt with Freud and memory but also colleges as institutions, in books such as “Safe Enough Spaces” (2019) and “The Student: A Short History” (2023). Recent years have brought an increasingly political thrust to both his writing (for national media and his presidential blog) and to his work as president. In 2023, in response to the Supreme Court’s ruling against affirmative action, Wesleyan ended legacy admissions.When Wesleyan students joined the national wave of protests over the war on Gaza, Roth—who describes himself as a supporter of both free speech and Israel’s right to exist—tangled with student protesters as well as with those who wanted him to shut the protests down. Meanwhile, in interviews and essays, he took administrators at other colleges to task for embracing the principles like those found in the “Kalven report”—a 1967 document out of the University of Chicago, which advanced the argument that universities should almost always remain scrupulously neutral. (Such stances were, he told me, “a cover for trying to stay out of trouble.”) As the Trump Administration has ramped up its attacks on the academy, Roth has continued to publish widely, urging fellow-leaders to stand up for their principles. “Release Mahmoud Khalil! Respect freedom of speech!” he concluded in a recent column for Slate, which argued that the Columbia activist’s arrest “should terrify every college president.”Roth and I met in his office, which is dominated by a round table where he meets with both students and his cabinet. Wearing Blundstones and polka-dot socks, he was loose-limbed and gregarious, and our conversation (which has been edited for length and clarity) was punctuated by the bright sound of batting from the baseball diamond just outside.You wrote last year, before the election, that colleges and universities weren’t ready for what was coming. How has the reality compared to your expectations?I had this idea—alas, it was in 2020, just as COVID was happening—that it would be great if colleges and universities took our civic responsibilities more seriously and really incentivized students to participate in the public sphere: work on a campaign, zoning commission, whatever. Rigorously agnostic about what they chose to work on. We found a few hundred schools that agreed in principle and we created a network. Before the 2024 election, we reactivated that group, and this time around, the institutions were much less likely to want to be publicly in support of even something so nonpartisan.We’re really small—three thousand students or so—and I wanted University of Texas at Austin, and Michigan, other big places. Some of them did agree in principle, but this time, in 2024—in the spring, let’s say, when Biden was still in the race, it was clear Trump was going to be the candidate—the reticence of academic leaders was already apparent.Last year, we ran a program called Democracy 2024. We brought people here—nice conference, all that stuff. And even a group of presidents that I helped put together for this purpose, they started talking more about “dialogue across difference” than participation in the electoral system.Everybody’s in favor of not fighting and having better dialogues, and I am, too. But I’m more in favor of people working on campaigns and learning about issues and getting things done. And in the last two months, it’s become painfully apparent that wanting to have nice conversations is not going to stop people who are bent on authoritarianism. Right now, I’m not sure what will stop them, except successful court challenges, and even that seems precarious.Watching the video of this poor woman at Tufts who was abducted by federal agents —I wrote my blog today about that. I think the government is spreading terror, and that’s what they mean to do. This kid isn’t a threat to security.I wrote to the president of Tufts—who I know, because we’re in the same athletic conference—and just said, “Is there anything you want anyone to do?” He said, “Thank you for writing.” And I don’t know his business. I’m sure he’s trying to help the student; that’s his responsibility, and I respect that. But I also think every citizen, but certainly every university person, should be expressing outrage.I’m curious to hear your thoughts about how we wound up here. Are there choices that universities have made that have made them more vulnerable to attack?I try to think about that without blaming the victim, because right now the story for me is that the government is abusing its powers by making war against civil society. That’s the song I’ve been singing—because you may not like universities, but you probably like churches or synagogues. But I have also been thinking about how universities can be less vulnerable in the long run. I’ve been arguing for almost a decade about the intellectual and political insularity of especially highly selective colleges and universities, and that we need more intellectual diversity at these places. I wrote a piece in the Wall Street Journal [in 2017], about affirmative action for conservatives, which annoyed everyone—which makes it a good op-ed, I guess.I teach a course, “The Modern and the Postmodern.” It’s not really about conservatism, but I have added conservative critiques of some of the modernists that I talk about. I teach a course on virtue and vice in history, philosophy, and literature, and I have added conservative critiques of the liberal assumptions that almost all my students share. And it’s interesting to see how they react—they’re shocked by these critiques, in ways they’re not shocked by, I don’t know, Bolshevism or violent anti-colonial revolutionary rhetoric. And I point that out. So we talk, and they’re perfectly able to deal with it. I’m not trying to convince them that these guys are right or anything; just that it’s interesting to think about.We have to be less insular, less parochial, and being politically more diverse is part of that. Also, at the fancy places—like Wesleyan and Ivy League schools and others, a small percentage of schools in the country—I do think it would not be unfair to say we’ve bred a kind of condescension. When you define the quality of your institution by how many people you reject, you can create—unintentionally—an attitude of “I’ve earned my superiority.”Trump and his allies have found a way to tar all of the sector with the brush of the Ivy League. They’re excellent schools, and they have excellent scientists, and if one of Vice-President Vance’s kids is sick, he’s going to want the doctor to have gone to one of these schools; he’s not going to want them to have gone to Viktor Orbán’s university. But higher education serves so many more people in so many different ways than the places that are highly selective.What do you make of the fact that the conflict over Israel and Palestine has become the pretext for the current crackdown?I think anti-antisemitism is a very useful tool for the right. Many others have noted how comfortable these same people who are cracking down on antisemitism are with Nazis—real, frighteningly confident antisemites. But it’s a useful tool, because so many people in the liberal-to-progressive, educated coalition are divided about it, and it’s generational.Anti-antisemitism can be appropriated by any political movement. They can use that as a vehicle for persecuting researchers and institutions that are not aligned with the ideology of the person in charge. It’s to show that you control them.You have prominent Jewish figures around the country who get comfortable with Trump, it seems to me, because they can say he’s fighting antisemitism: “He’s good for the Jews.” It’s pathetic. It’s a travesty of Jewish values, in my view.Over the last couple of months, many leaders of colleges and universities haven’t spoken out against the Trump Administration’s attack on higher education. You’ve been pretty vocal. What do you think has made that possible?I have, for many years, spoken my mind in ways that are clearly fallible. I’ve had to apologize. My communications office, when I said I wanted to do blogging, thought it was a bad idea. I think it’s important to participate. And then to say, “Oh, shit, I made a mistake.” “Oh, I shouldn’t have said that.” “Yes, I should say this.” It’s not perfect—no conversation is.I think my job as a leader of the university is to speak up for the values that we claim to believe in, especially when they’re at odds with people with enormous power. So I think I’m speaking now because I’ve been speaking.My board is very supportive. My board teases me that I threaten to quit a lot. I don’t think I do, but they say I do, so they’re probably right. In November, after the election, I said, “If you want a president who’s not going to speak up, you have to find another president.” One of my friends on the board said, “Why did you do that? You don’t have to threaten to quit. Everybody wants you to stay.” I said, “I didn’t threaten to quit! It’s just a fact!” I’m more combative than I want to be, and I’m not looking for a fight, but I do feel that when people are getting really pushed around, in horrible ways, that someone who is at a university and has a platform and can call an editor—we should try.I actually thought other people would speak out. Because the missions are at stake. Even the Kalven people—when the mission’s at stake, you’re supposed to speak out.Someone on the faculty at Columbia who works in the administration asked me to put together a group of presidents. I was unable to. I wrote something quickly for people to sign, and one of the people I contacted said to me, “Are you sure the president at Columbia wants you to?” I said, “I’m not sure.”Tell me about the kinds of conversations you’ve been having with fellow-presidents. What’s your sense of the internal debates they’re having?Presidents—we’re not usually honest with each other. It’s just the nature of the job. You’re always trying to put your institution in the best possible light. I always joke that I see a president I haven’t seen in a few years, and he used to have two arms and now only has one —“What the hell happened, Charlie?” “Oh, I hated that arm! I feel so much freer!”I don’t go to a lot of presidential gatherings, but I go to a couple. I was at one, and this guy came to me and said, “You know, you make me feel like a coward.” I said, “I’m sorry, that’s not really what my intention is.” But he said he’s at a public university—he was going to the state legislature two days later. He said, “They’re not going to let me have any diversity stuff in the university.” I said, “Well, you can quit.” He said, “And what good would that do?” So, I’m lucky—I have a board that likes the work we do. I have an incredible team of vice-presidents and a wonderful faculty, and they’re supportive. But I don’t understand. Because I know some presidents—many of them are a lot smarter than I am. I’m sure they can write well. I don’t understand.People have said to me, “Well, you take all that money from the government, why don’t you listen to them?” The answer is, because the money doesn’t come with a loyalty oath. And that has served the country so well because Americans and various governments we’ve had recognized that it’s better for the country if people can practice freedom. The government’s not going to tell you how to run your business. It starts with universities, because universities are this object of resentment and this kind of weird charisma, negative and positive. It starts with that, but it could very easily go into these other aspects of this culture that, again, depend on the government.I don’t have to agree with the mayor to get the fire department to come put out a fire. And that’s what they’re saying to these international students: “Well, you came to this country. What makes you think you can write an op-ed in the newspaper?” Well, what makes you think that is, this is a free country. As I say that, I can hear my leftist friends: “Oh, yeah, it’s never been free.” It’s never been totally free, but freedoms haven’t been as threatened as they are now since World War One. I’m confident of that. Universities are part of a sector of our culture that is worth preserving, from the Kennedy Center to magazines to churches. Autonomy of these different areas, even if they’re entangled economically with the government or with rich people, is so important. I can’t believe that I have to say these things out loud; it’s so obvious.What are you doing right now at Wesleyan? What kinds of plans are you putting in place to protect your school?We’re making sure that we are not wasting a dollar, so that we can have monies available should we need them. It could be because of the endowment tax—which is really an ideological, punitive thing. That would have an impact on us—our financial-aid program is supported by the endowment. We’re preparing scenarios for the endowment tax, for cuts to the scientists and other things. I’ve talked to schools about creating a legal-defense fund, but I don’t know—that’s a relatively new idea. I’ve talked to a group of faculty at Yale about that.How much federal funding does Wesleyan get?About twenty million. A good chunk of that is student loans that are guaranteed, and then the rest is grants to scientists and others. Right now, we have N.I.H, N.S.F. We have graduate programs in the sciences, so we are unlike most other liberal-arts colleges.The scale of our budget is three hundred million, let’s say, annually. So it’s real; it’s an important part of the budget. People use the word “existential”—it’s not. It would change the dynamic we have.What kinds of concerns are you hearing from students?International students are very afraid to travel. The idea that somebody would take your phone and look at all the images and find an image they didn’t like is very frightening. International faculty, too—we have faculty who are here, of course, legally, and working for the university either as permanent residents or on visas. Then we have a lot of faculty who travel for research, and go to scientific meetings, or are doing work in archives around the world. And I think all of them are nervous about this use of the border in an ideological way. As I say that out loud, I can hear many of my colleagues on the left saying, “Duh! We’ve been using the border in an ideological way forever.” And that’s true—in a certain way, borders are a product of ideology. But saying “This isn’t different than five years ago” would be, for me, like putting your head in the sand. To use the tools of the government to make people align ideologically is really different—and, I think, would offend the values that many conservative Americans have. I did the Charlie Sykes podcast, and I’m trying to get some other more conservative people to have me on, so I can talk about these things in ways that aren’t just for the faculty at liberal-arts colleges.What would you do if ICE agents showed up at Wesleyan? Do you have a plan in place?We are making sure that our students, faculty, and staff know their rights, as people who live in the U.S. and are owed due process. The university would ask any federal agents to check in with the Office of Public Safety—that’s the campus police. They would need to have judicial warrants. [Mahmoud Khalil has said that the agents who arrested him refused to produce a warrant.] We would want to make sure government officials are following the law. We will protect people who are on our private property from people who want to constrain their freedom. We’d offer whatever legal assistance we can.We’re not going to obstruct the work of legally authorized officials—we want to make sure that they are, in fact, legally authorized.In your scholarly work, you focus on the ways people grapple with the past—the psychology of history. How does that inflect your work as president, especially now?Scapegoating and the creation of categories of people you can hate on and abuse is a fundamental aspect of human societies, and one should really pay attention to the dynamic in which that process takes place. That definitely comes out of my work on Freud—and, in a weird way, René Girard, who’s anti-Freudian, but there’s a lot in common. An appreciation for the ways in which animosity can spring forth in brutal forms, especially when it’s been repressed—that’s something I’ve tried to stay aware of.I wrote a lot about Freud over the years, and for me, the most important concept in Freud is the transference, and how sometimes we treat people as if they were other people from our pasts. Famously, the analyst is transformed into the parents and other things. I think that that happens a lot in my job. It happens as a teacher all the time. And as president, oy. It’s really big time. People were like, “Why don’t you end the war in Gaza?” last year—they just want someone to be able to do the things they desire. They didn’t have that relationship to Biden or the Secretary of State; it was me, I was in charge of the university. You know, so, they give me credit for things that I don’t deserve and they blame me for things I don’t think I deserve the blame for, and that’s just part of the deal.I didn’t see that before. I didn’t care so much about the president when I was a student. I loved my teachers—I mean, I had massive transference for my teachers. But the president, these days, has more symbolic importance than I expected. ♦
...
Read the original on www.newyorker.com »
What were we all doing here? My 600-mile trip to hear an organ play a D naturalA John Cage recital that is set to last 639 years recently witnessed a chord change — 500 people made a pilgrimage to experience it
The crowd in the immediate vicinity of the organ at St Burchardi church, Halberstadt, had paid €200 [DPA Picture Alliance Archive / Alamy Stock Photo]
In the year 2000, in a small east German town, work began on the construction of an organ that had one purpose: to perform John Cage’s ORGAN2/ASLSP (1987) for precisely 639 years. The late avant-garde composer’s only instruction for the piece was to play the piece ‘as slowly as possible’. And so in 2001 — the instrument finally ready — the world’s longest organ recital began in St Burchardi church, Halberstadt, with a rest lasting 17 months before the first chord commenced droning in 2003. It consisted of two G sharps and a B. Two weeks ago, I — along with several hundred others — made the pilgrimage to the town to witness the work’s latest chord change.
In theory, a pipe organ can sound indefinitely, so long as it receives adequate power and its pedals are pressed continually. To eliminate the need for an organist, a system of sandbags suspended by strings delivers this pressure in Halberstadt.
Register to get The Spectator’s insight and opinion straight to your inbox. You can then read two free articles each week.
...
Read the original on www.spectator.co.uk »
To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".
10HN is also available as an iOS App
If you visit 10HN only rarely, check out the the best articles from the past week.
If you like 10HN please leave feedback and share
Visit pancik.com for more.