10 interesting stories served every morning and every evening.
During out talks with F-Droid users at FOSDEM26 we were baffled to learn most were relieved that Google has canceled their plans to lock-down Android.
Why baffled? Because no such thing actually happened, the plans announced last August are still scheduled to take place. We see a battle of PR campaigns and whomever has the last post out remains in the media memory as the truth, and having journalists just copy/paste Google posts serves no one.
But Google said… Said what? That there’s a magical “advanced flow”? Did you see it? Did anyone experience it? When is it scheduled to be released? Was it part of Android 16 QPR2 in December? Of 16 QPR3 Beta 2.1 last week? Of Android 17 Beta 1? No? That’s the issue… As time marches on people were left with the impression that everything was done, fixed, Google “wasn’t evil” after all, this time, yay!
While we all have bad memories of “banners” as the dreaded ad delivery medium of the Internet, after FOSDEM we decided that we have to raise the issue back and have everyone, who cares about Android as an open platform, informed that we are running out of time until Google becomes the gate-keeper of all users devices.
Hence, the website and starting today our clients, with the updates of F-Droid and F-Droid Basic, feature a banner that reminds everyone how little time we have and how to voice their concerns to whatever local authority is able to understand the dangers of this path Android is led to.
We are not alone in our fight, IzzyOnDroid added a banner too, more F-Droid clients will add the warning banner soon and other app downloaders, like Obtainium, already have an in-app warning dialogue.
Regarding F-Droid Basic rewrite, development continues with a new release 2.0-alpha3:
Note that if you are already using F-Droid Basic version 1.23.x, you won’t receive this update automatically. You need to navigate to the app inside F-Droid and toggle “Allow beta updates” in top right three dot menu.
In apps news, we’re slowly getting back on track with post Debian upgrade fixes (if your app still uses Java 17 is there a chance you can upgrade to 21?) and post FOSDEM delays. Every app is important to us, yet actions like the Google one above waste the time we could have put to better use in Gitlab.
Buses was updated to 1.10 after a two year hiatus.
Conversations and Quicksy were updated to 2.19.10+free improving on cleaning up after banned users, a better QR workflow and better tablet rotation support. These are nice, but another change raises our interest, “Play Store flavor: Stop using Google library and interface directly with Google Play Service via IPC”. Sounds interesting for your app too? Is this a path to having one single version for both F-Droid and Play that is fully FLOSS? We don’t know yet, but we salute any trick that removes another proprietary dependency from the code. If curious feel free to take a look at the commit.
Dolphin Emulator was updated to 2512. We missed one version in between so the changelogs are huge, luckily the devs publish highly detailed posts about updates. So we’ll start with “Release 2509” (about 40 mins to read), we side-track with “Starlight Spotlight: A Hospital Wii in a New Light” (for about 50 mins), we continue to the current release in “Release 2512” (40 more minutes) and we finish with “Rise of the Triforce” delving in history for more than one hour.
Image Toolbox was updated to 3.6.1 adding many fixes and… some AI tools. Were you expecting such helpers? Will you use them?
Luanti was updated to 5.15.1 adding some welcomed fixes. If your game world started flickering after the last update make sure to update.
Nextcloud apps are getting an update almost every week, like Nextcloud was updated to 33.0.0, Nextcloud Cookbook to 0.27.0, Nextcloud Dev to 20260219, Nextcloud Notes to 33.0.0 and Nextcloud Talk was updated to 23.0.0.
But are you following the server side too? Nextcloud Hub 26 Winter was just released adding a plethora of features. If you want to read about them, see the 30 minutes post here or watch the one hour long video presentation from the team here.
ProtonVPN - Secure and Free VPN was updated to 5.15.70.0 adding more control to auto-connects, countries and cities. Also all connections are handled now by WireGuard and Stealth protocols as the older OpenVPN was removed making the app almost 40% smaller.
Offi was updated to 14.0 with a bit of code polish. Unfortunately for Android 7 users, the app now needs Android 8 or later.
QUIK SMS was updated to 4.3.4 with many fixes. But Vishal praised the duplicate remover, the default auto de-duplication function and found that the bug that made deleted messages reappear is fixed.
SimpleEmail was updated to 1.5.4 after a 2 year pause. It’s just a fixes release, updating translations and making the app compatible with Android 12 and later versions.
* NeoDB You: A native Android app for NeoDB designed with Material 3/You
Thank you for reading this week’s TWIF 🙂
Please subscribe to the RSS feed in your favourite RSS application to be updated of new TWIFs when they come up.
You are welcome to join the TWIF forum thread. If you have any news from the community, post it there, maybe it will be featured next week 😉
To help support F-Droid, please check out the donation page and contribute what you can.
...
Read the original on f-droid.org »
We’re wrapping up our live coverage of the Supreme Court decision in Learning Resources, Inc v. Trump.
The major ruling - and Trump’s response - can be expected to have an effect on trade, the global economy, Americans’ personal finances, politics and more.
You can read what North America Correspondent Anthony Zurcher thinks it means for Trump’s second-term agenda here, as well as how Canada, one of the top US trading partners, views the decision.
We also have covered the major turns of the day here, and our White House correspondent Bernd Debusmann has described what it was like to cover Trump’s press briefing about the ruling in this video.
We’ll be back when more big trade, Supreme Court, or other news breaks.
...
Read the original on www.bbc.com »
And I don’t just mean that nobody uses it anymore. Like, I knew everyone under 50 had moved on, but I didn’t realize the extent of the slop conveyor belt that’s replaced us.
I logged on for the first time in ~8 years to see if there was a group for my neighborhood (there wasn’t). Out of curiosity I thought I’d scroll a bit down the main feed.
The first post was the latest xkcd (a page I follow). The next ten posts were not by friends or pages I follow. They were basically all thirst traps of young women, mostly AI-generated, with generic captions. Here’s a sampler — mildly NSFW, but I did leave out a couple of the lewder ones:
Yikes. Again, I don’t follow any of these pages. This is all just what Facebook is pushing on me.
I know Twitter/X has worse problems with spam bots in the replies, but this is the News Feed! It’s the main page of the site! It’s the product that defined modern social media!
It wasn’t all like that, though. There was also an AI video of a policeman confiscating a little boy’s bike, only to bring him a brand new one:
And there were some sloppy memes and jokes, mostly about relationships, like this (admittedly not AI) video sketch where a woman decides to intentionally start a fight with her boyfriend because she’s on her period:
Maybe that isn’t literally about sex, but I’d classify it as the same sort of lizard-brain-rot engagement bait as those selfies.
Several commenters have vouched that Yoleendadong makes funny, high-quality content and shouldn’t be lumped in with AI slop. I’m just saying I think there’s a reason this particular video of hers popped up, and it’s probably the kind of engagement created by the premise.
Meta even gives us some helpful ideas for sexist questions we can ask their AI about the video:
Yep, that’s another “yikes” from me. To be fair, though, sometimes that suggested questions feature is pretty useful! Like with this post, for example:
Why is she wearing pink heels? What is her personality? Great questions, Meta.
I said these were “mostly” AI-generated. The truth is with how good the models are getting these days, it’s hard to tell, and I think a couple of them might be real people.
Still, some of these are pretty obviously AI. Here’s one with a bunch of alien text and mangled logos on the scoreboard in the background:
Hmm, I wonder if anyone has noticed this is AI? Let’s check out the comments and see if anyone’s pointed that ou—
…never mind. (I dunno, maybe those are all bots too.)
So: is this just something wacky with my algorithm?
I mean… maybe? That’s part of the whole thing with these algorithmic feeds; it’s hard to know if anyone else is seeing what I’m seeing.
On the one hand, I doubt most (straight) women’s feeds would look like this. But on the other hand, I hadn’t logged in in nearly a decade! I hate to think what the feed looks like for some lonely old guy who’s been scrolling the lightly-clothed AI gooniverse for hours every day.
Did everyone but me know it was like this? I’d seen screencaps of stuff like the Jesus-statue-made-out-of-broccoli slop a year or two ago, but I thought that only happened to grandmas. I hadn’t heard it was this bad.
I wonder if this evolution was less noticeable for people who are logging in every day. Or maybe it only gets this bad when there aren’t any posts from your actual friends?
In any case, I stopped exploring after I saw a couple more of those AI-generated pictures but with girls that looked like they were about ~14, which made me sick to my stomach. So long Facebook, see you never, until one day I inexplicably need to use your platform to get updates from my kid’s school.
...
Read the original on pilk.website »
Skip to content
Secure your code as you build
We read every piece of feedback, and take your input very seriously.
Include my email address so I can be contacted
Use saved searches to filter your results more quickly
To see all available qualifiers, see our documentation.
Sign up
You signed in with another tab or window. Reload to refresh your session.
You signed out in another tab or window. Reload to refresh your session.
You switched accounts on another tab or window. Reload to refresh your session.
Notifications
You must be signed in to change notification settings
ggml.ai joins Hugging Face to ensure the long-term progress of Local AI
ggml.ai joins Hugging Face to ensure the long-term progress of Local AI
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
There was an error while loading. Please reload this page.
You can’t perform that action at this time.
...
Read the original on github.com »
In 2017, WikiLeaks published Vault7 - a large cache of CIA hacking tools and internal documents. Buried among the exploits and surveillance tools was something far more mundane: a page of internal developer documentation with git tips and tricks.
Most of it is fairly standard stuff, amending commits, stashing changes, using bisect. But one tip has lived in my ~/.zshrc ever since.
Over time, a local git repo accumulates stale branches. Every feature branch, hotfix, and experiment you’ve ever merged sits there doing nothing. git branch starts to look like a graveyard.
You can list merged branches with:
git branch –merged
But deleting them one by one is tedious. The CIA’s dev team has a cleaner solution:
git branch –merged | grep -v “\*\|master” | xargs -n 1 git branch -d
* git branch –merged — lists all local branches that have already been merged into the current branch
* grep -v “\*\|master” — filters out the current branch (*) and master so you don’t delete either
* xargs -n 1 git branch -d — deletes each remaining branch one at a time, safely (lowercase -d won’t touch unmerged branches)
Since most projects now use main instead of master, you can update the command and exclude any other branches you frequently use:
git branch –merged origin/main | grep -vE “^\s*(\*|main|develop)” | xargs -n 1 git branch -d
Run this from main after a deployment and your branch list goes from 40 entries back down to a handful.
I keep this as a git alias so I don’t have to remember the syntax:
alias ciaclean=‘git branch –merged origin/main | grep -vE “^\s*(\*|main|develop)” | xargs -n 1 git branch -d’
Then in your repo just run:
ciaclean
Small thing, but one of those commands that quietly saves a few minutes every week and keeps me organised.
You can follow me here for my latest thoughts and projects
...
Read the original on spencer.wtf »
I’m a diving instructor. I’m also a platform engineer who spends lots of his time thinking about and implementing infrastructure security. Sometimes those two worlds collide in unexpected ways.
A Sula sula (Frigatebird) and a dive flag on the actual boat where I found the vulnerability - somewhere off Cocos Island.
While on a 14 day-long dive trip around Cocos Island in Costa Rica, I stumbled across a vulnerability in the member portal of a major diving insurer - one that I’m personally insured through. What I found was so trivial, so fundamentally broken, that I genuinely couldn’t believe it hadn’t been exploited already.
I disclosed this vulnerability on April 28, 2025 with a standard 30-day embargo period. That embargo expired on May 28, 2025 - over eight months ago. I waited this long to publish because I wanted to give the organization every reasonable opportunity to fully remediate the issue and notify affected users. The vulnerability has since been addressed, but to my knowledge, I have not received confirmation that affected users were notified. I have reached out to the organization to ask for clarification on this matter.
This is the story of what happened when I tried to do the right thing.
To understand why this is so bad, you need to know how the registration process works. As a diving instructor, I register my students (to get them insured) through my account on the portal. I enter their personal information with their consent - name, date of birth, address, phone number, email - and the system creates an account for them. The student then receives an email with their new account credentials: a numeric user ID and a default password. They might log in to complete additional information, or they might never touch the portal again.
When I registered three students in quick succession, they were sitting right next to me and checked their welcome emails. The user IDs were nearly identical - sequential numbers, one after the other. That’s when it clicked that something really bad was going on.
Now here’s the problem: the portal used incrementing numeric user IDs for login. User XXXXXX0, XXXXXX1, XXXXXX2, and so on. That alone is a red flag, but it gets worse: every account was provisioned with a static default password that was never enforced to be changed on first login. And many users - especially students who had their accounts created for them by their instructors - never changed it.
So the “authentication” to access a user’s full profile - name, address, phone number, email, date of birth - was:
Type the same default password that every account shares on account creation.
There’s a good chance you get in.
That’s it. No rate limiting. No account lockout. No MFA. Just an incrementing integer and a password that might as well have been password123.
I verified the issue with the minimum access necessary to confirm the scope - and stopped immediately after.
I did everything by the book. I contacted CSIRT Malta (MaltaCIP) first - since the organization is registered in Malta, this is the competent national authority. The Maltese National Coordinated Vulnerability Disclosure Policy (NCVDP) explicitly requires that confirmed vulnerabilities be reported to both the responsible organization and CSIRTMalta.
As a fellow diving instructor insured through [the organization] and a full-time Linux Platform Engineer, I am contacting you to responsibly disclose a critical vulnerability I identified within the [the organization]’s user account system.
During recent testing, I discovered that user accounts - including those of underage students - are accessible through a combination of predictable User ID enumeration (incrementing user IDs) and the use of a static default password that is not enforced to be changed upon first login. This misconfiguration currently exposes sensitive personal data (e.g., names, addresses, contact information - including phone numbers and emails -, dates of birth) and represents multiple GDPR violations.
Exposure of sensitive and underage user data without adequate safeguards
For initial confirmation, I am attaching a screenshot from Member ID XXXXXXX showing the exposed data, partly redacted for privacy reasons.
Additionally, for transparency and validation, I have shared my proof-of-concept code securely via an encrypted paste service: [link redacted]
In the spirit of responsible disclosure, I have already informed CSIRT Malta (in CC) to officially initiate a reporting process, given [the organization]’s operational presence in Malta.
I kindly request that [the organization] acknowledges receipt of this disclosure within 7 days.
I am offering a window of 30 days from today the 28th of April 2025 for [the organization] to mitigate or resolve the vulnerability before I consider any public disclosure.
Please note that I am fully available to assist your IT team with technical details, verification steps and recommendations from a security perspective.
I strongly recommend assigning an IT-Security Point of Contact (PoC) for direct collaboration on this issue.
Thank you very much for your attention to this critical matter. I am looking forward to working with you towards a secure resolution.
Both of these timelines are standard - if anything, generous - in responsible disclosure frameworks.
Two days later, I got a reply. Not from their IT team. From their Data Privacy Officers (DPO’s) law firm.
The letter opened politely enough - they acknowledged the issue and said they’d launched an investigation. They even mentioned they were resetting default passwords and planning to roll out 2FA. Good.
But then the tone shifted:
While we genuinely appreciate your seemingly good intentions and transparency in highlighting this matter to our attention, we must respectfully note that notifying the authorities prior to contacting the Group creates additional complexities in how the matter is perceived and addressed and also exposes us to unfair liability.
Let me translate: “We wish you hadn’t told the government about our security issue.”
It got better:
We also do not appreciate your threat to make this matter public […] and remind you that you may be held accountable for any damage we, or the data subjects, may suffer as a result of your own actions, which actions likely constitute a criminal offence under Maltese law.
So, to be clear: their portal had a default password on every account, exposing personal data including that of children, and I’m the one who “likely” committed a criminal offence by finding it and telling them.
They also sent a declaration they wanted me to sign - while requesting my passport ID - confirming I’d deleted all data, wouldn’t disclose anything, and would keep the entire matter “strictly confidential.” The deadline? End of business the same day they sent it.
This declaration included the following gem:
I also declare that I shall keep the content of this declaration strictly confidential.
That’s an NDA with extra steps: I was being asked to sign away my right to discuss the disclosure process itself - including the fact that I found a vulnerability in their system - under threat of legal action.
Then came the reminders. One “friendly” reminder. Then an “urgent” one. Sign the declaration. De-escalate. Move on. Quietly.
I generally refuse to sign confidentiality clauses in cases involving exposure of sensitive information, and I did so here as well. Coordinated disclosure depends on transparency and trust between researchers and organizations: trust that affected users will be informed, and trust that a report leads to real remediation.
Given that the organization in question had already breached that trust by exposing personal data through weak controls, I wasn’t willing to grant blanket confidentiality that could be used to keep the incident out of public scrutiny. And with trying to actual silence me through legal threats, they had already made it clear that their priority was reputation management over user data protection. So I stood my ground.
Instead, I offered to sign a modified declaration confirming data deletion. I had no interest in retaining anyone’s personal data, but I was not going to agree to silence about the disclosure process itself.
I also pointed out that, under Malta’s NCVDP, involving CSIRT Malta is part of the expected reporting path - not a hostile act - and that publishing post-remediation analyses is standard practice in the security community.
Their response doubled down. They cited Article 337E of the Maltese Criminal Code - computer misuse - and helpfully reminded me that:
Art. 337E of the Criminal Code also provides that “If any act is committed outside Malta which, had it been committed in Malta, would have constituted an offence […] it shall […] be deemed to have been committed in Malta.” Meaning that your actions would be deemed a criminal offence in Malta, even if committed in another country.
They also made their position on disclosure crystal clear, after I reiterated my refusal to sign their NDA:
We object strongly to the use of [the organization’s name] in any such blogs or conferences you may write/attend as this would be a disproportionate harm to [the organization’s] reputation […]. We reserve our rights at law to hold you responsible for any damages [the organization] may suffer as a result of any such public disclosures you may make.
That’s fine by me. Because here’s the thing: The vulnerability has been fixed. Default passwords have been reset. 2FA is being rolled out. I feel sorry for the developer(s) who had to clean up this mess, but at least the issue is no longer exploitable. Sure, it would have been better if the organization had thanked me and taken responsibility for notifying affected users. If the incident qualified as a personal data breach (which it does) and was likely to result in a (high) risk to individuals - especially given minors were involved - GDPR Articles 33 and 34 generally require notification to the supervisory authority and communication to affected data subjects.
GDPR Article 34(1) When the personal data breach is likely to result in a high risk to the rights and freedoms of natural persons, the controller shall communicate the personal data breach to the data subject without undue delay.
GDPR Article 34(2) The communication to the data subject referred to in paragraph 1 of this Article shall describe in clear and plain language the nature of the personal data breach and contain at least the information and measures referred to in points (b), (c) and (d) of Article 33(3).
I have not received confirmation that those notifications were ever carried out.
My favourite part was the organization’s position on whose fault this actually was:
We contend that it is the responsibility of users to change their own password (after we allocate a default one).
Read that again. A company that assigned the same default password to every account, never forced a password change, and used incrementing numeric IDs as usernames is blaming the users for not securing their own accounts. Accounts that include those of minors.
GDPR Article 5(1)(f) (integrity and confidentiality): Personal data shall be processed in a manner that ensures appropriate security of the personal data, including protection against unauthorised or unlawful processing and against accidental loss, destruction or damage, using appropriate technical or organisational measures.
Under GDPR, the data controller (namely: the organization) is responsible for implementing appropriate technical and organizational measures to ensure data security. A static default password on an IDOR-vulnerable portal is not an “appropriate measure” by any definition.
GDPR Article 24(1) (controller responsibility): Taking into account the nature, scope, context and purposes of processing as well as the risks of varying likelihood and severity for the rights and freedoms of natural persons, the controller shall implement appropriate technical and organisational measures to ensure and to be able to demonstrate that processing is performed in accordance with this Regulation. Those measures shall be reviewed and updated where necessary.
This isn’t an isolated case. The security research community has been dealing with this pattern for decades: find a vulnerability, report it responsibly, get threatened with legal action. It’s so common it has a name - the chilling effect.
Organizations that respond to disclosure with lawyers instead of engineers are telling the world something important: they care more about their reputation than about the data they’re supposed to protect.
And the real irony? The legal threats are the reputation damage. Not the vulnerability itself - vulnerabilities happen to everyone. It’s the response that tells you everything about an organization’s security culture.
What Should Have Happened
Acknowledge the report - they did this, to be fair.
Fix the vulnerability - they started on this too.
Thank the researcher - instead of threatening them with criminal prosecution.
Have a CVD policy - so researchers know how to report issues and what to expect.
Notify affected users - especially the parents of underage members whose data was exposed.
Not try to silence the researcher with NDAs disguised as “declarations.”
What You Can Do
Publish a Coordinated Vulnerability Disclosure policy. It doesn’t have to be complex - maybe begin with a security.txt file and a clear process that favors transparency.
Thank researchers for helping you improve your security posture.
Don’t shoot the messenger. The person reporting the bug is not your enemy. The bug is.
Don’t blame your users for security failures that are your responsibility as a data controller.
Always involve your national CSIRT. It protects you and creates an official record.
Document everything. Every email, every timestamp, every response.
Don’t sign NDAs that prevent you from discussing the disclosure process. But you can agree to delete data (and MUST do so!) without agreeing to silence.
Know your rights. Many jurisdictions have legal protections for good-faith security research. The EU’s NIS2 Directive encourages coordinated vulnerability disclosure.
Because right now, in 2026, reporting a trivial vulnerability exposing personal data - including that of children - still gets met with legal threats instead of gratitude. And that’s a problem for all of us. Let’s burn some Tokens! - AI Chatbot Cost Exploitation as an Attack VectorLet’s burn some Tokens! - AI Chatbot Cost Exploitation as an Attack VectorMany companies ship AI chatbots as thin wrappers around commercial LLM APIs with zero cost controls. What if a tool behaved like an overly engaged, perfectly valid user - and just burned through their budget?Imprint / ImpressumData Privacy / DatenschutzDo you know the code?
...
Read the original on dixken.de »
Dependabot is a noise machine. It makes you feel like you’re doing work, but you’re actually discouraging more useful work. This is especially true for security alerts in the Go ecosystem.
I recommend turning it off and replacing it with a pair of scheduled GitHub Actions, one running govulncheck, and the other running your test suite against the latest version of your dependencies.
On Tuesday, I published a security fix for filippo.io/edwards25519. The (*Point).MultiScalarMult method would produce invalid results if the receiver was not the identity point.
A lot of the Go ecosystem depends on filippo.io/edwards25519, mostly through github.com/go-sql-driver/mysql (228k dependents only on GitHub). Essentially no one uses (*Point).MultiScalarMult.
Yesterday, Dependabot opened thousands of PRs against unaffected repositories to update filippo.io/edwards25519. These PRs were accompanied by a security alert with a nonsensical, made up CVSS v4 score and by a worrying 73% compatibility score, allegedly based on the breakage the update is causing in the ecosystem. Note that the diff between v1.1.0 and v1.1.1 is one line in the method no one uses.
We even got one of these alerts for the Wycheproof repository, which does not import the affected filippo.io/edwards25519 package at all. Instead, it only imports the unaffected filippo.io/edwards25519/field package.
$ go mod why -m filippo.io/edwards25519
github.com/c2sp/wycheproof/tools/twistcheck
filippo.io/edwards25519/field
We have turned Dependabot off.
But isn’t this toil unavoidable, to prevent attackers from exploiting old vulnerabilities in your dependencies? Absolutely not!
Computers are perfectly capable of doing the work of filtering out these irrelevant alerts for you. The Go Vulnerability Database has rich version, package, and symbol metadata for all Go vulnerabilities.
Here’s the entry for the filippo.io/edwards25519 vulnerability, also available in standard OSV format.
modules:
- module: filippo.io/edwards25519
versions:
- fixed: 1.1.1
vulnerable_at: 1.1.0
packages:
- package: filippo.io/edwards25519
symbols:
- Point.MultiScalarMult
summary: Invalid result or undefined behavior in filippo.io/edwards25519
description: |-
Previously, if MultiScalarMult was invoked on an
initialized point who was not the identity point, MultiScalarMult
produced an incorrect result. If called on an
uninitialized point, MultiScalarMult exhibited undefined behavior.
cves:
- CVE-2026-26958
credits:
- shaharcohen1
- WeebDataHoarder
references:
- advisory: https://github.com/FiloSottile/edwards25519/security/advisories/GHSA-fw7p-63qq-7hpr
source:
id: go-security-team
created: 2026-02-17T14:45:04.271552-05:00
review_status: REVIEWED
Any decent vulnerability scanner will at the very least filter based on the package, which requires a simple go list -deps ./…. This already silences a lot of noise, because it’s common and good practice for modules to separate functionality relevant to different dependents into different sub-packages. For example, it would have avoided the false alert against the Wycheproof repository.
If you use a third-party vulnerability scanner, you should demand at least package-level filtering.
Good vulnerability scanners will go further, though, and filter based on the reachability of the vulnerable symbol using static analysis. That’s what govulncheck does!
$ go mod why -m filippo.io/edwards25519
filippo.io/sunlight/internal/ctlog
github.com/google/certificate-transparency-go/trillian/ctfe
github.com/go-sql-driver/mysql
$ govulncheck ./…
=== Symbol Results ===
No vulnerabilities found.
Your code is affected by 0 vulnerabilities.
This scan also found 1 vulnerability in packages you import and 2
vulnerabilities in modules you require, but your code doesn’t appear to call
these vulnerabilities.
Use ‘-show verbose’ for more details.
govulncheck noticed that my project indirectly depends on filippo.io/edwards25519 through github.com/go-sql-driver/mysql, which does not make the vulnerable symbol reachable, so it chose not to notify me.
If you want, you can tell it to show the package- and module-level matches.
$ govulncheck -show verbose,color ./…
Fetching vulnerabilities from the database…
Checking the code against the vulnerabilities…
The package pattern matched the following 16 root packages:
filippo.io/sunlight/internal/stdlog
Govulncheck scanned the following 54 modules and the go1.26.0 standard library:
crawshaw.io/sqlite@v0.3.3-0.20220618202545-d1964889ea3c
filippo.io/edwards25519@v1.1.0
filippo.io/keygen@v0.0.0-20240718133620-7f162efbbd87
=== Symbol Results ===
No vulnerabilities found.
=== Package Results ===
Vulnerability #1: GO-2026-4503
Invalid result or undefined behavior in filippo.io/edwards25519
More info: https://pkg.go.dev/vuln/GO-2026-4503
Module: filippo.io/edwards25519
Found in: filippo.io/edwards25519@v1.1.0
Fixed in: filippo.io/edwards25519@v1.1.1
=== Module Results ===
Vulnerability #1: GO-2025-4135
Malformed constraint may cause denial of service in
golang.org/x/crypto/ssh/agent
More info: https://pkg.go.dev/vuln/GO-2025-4135
Module: golang.org/x/crypto
Found in: golang.org/x/crypto@v0.44.0
Fixed in: golang.org/x/crypto@v0.45.0
Vulnerability #2: GO-2025-4134
Unbounded memory consumption in golang.org/x/crypto/ssh
More info: https://pkg.go.dev/vuln/GO-2025-4134
Module: golang.org/x/crypto
Found in: golang.org/x/crypto@v0.44.0
Fixed in: golang.org/x/crypto@v0.45.0
Your code is affected by 0 vulnerabilities.
This scan also found 1 vulnerability in packages you import and 2
vulnerabilities in modules you require, but your code doesn’t appear to call
these vulnerabilities.
...
Read the original on words.filippo.io »
The English-language edition of Wikipedia is blacklisting Archive.today after the controversial archive site was used to direct a distributed denial of service (DDoS) attack against a blog.
In the course of discussing whether Archive.today should be deprecated because of the DDoS, Wikipedia editors discovered that the archive site altered snapshots of webpages to insert the name of the blogger who was targeted by the DDoS. The alterations were apparently fueled by a grudge against the blogger over a post that described how the Archive.today maintainer hid their identity behind several aliases.
“There is consensus to immediately deprecate archive.today, and, as soon as practicable, add it to the spam blacklist (or create an edit filter that blocks adding new links), and remove all links to it,” stated an update today on Wikipedia’s Archive.today discussion. “There is a strong consensus that Wikipedia should not direct its readers towards a website that hijacks users’ computers to run a DDoS attack (see WP:ELNO#3). Additionally, evidence has been presented that archive.today’s operators have altered the content of archived pages, rendering it unreliable.”
More than 695,000 links to Archive.today are distributed across 400,000 or so Wikipedia pages. The archive site is commonly used to bypass news paywalls, and the FBI has on the site operator’s identity with a subpoena to domain registrar Tucows.
“Those in favor of maintaining the status quo rested their arguments primarily on the utility of archive.today for verifiability,” said today’s Wikipedia update. “However, an analysis of existing links has shown that most of its uses can be replaced. Several editors started to work out implementation details during this RfC [request for comment] and the community should figure out how to efficiently remove links to archive.today.”
Guidance published as a result of the decision asked editors to help remove and replace links to the following domain names used by the archive site: archive.today, archive.is, archive.ph, archive.fo, archive.li, archive.md, and archive.vn. The guidance says editors can remove Archive.today links when the original source is still online and has identical content; replace the archive link so it points to a different archive site, like the Internet Archive, Ghostarchive, or Megalodon; or “change the original source to something that doesn’t need an archive (e.g., a source that was printed on paper), or for which a link to an archive is only a matter of convenience.”
...
Read the original on arstechnica.com »
The first sign that something in San Francisco had gone very badly wrong was the signs. In New York, all the advertising on the streets and on the subway assumes that you, the person reading, are an ambiently depressed twenty-eight-year-old office worker whose main interests are listening to podcasts, ordering delivery, and voting for the Democrats. I thought I found that annoying, but in San Francisco they don’t bother advertising normal things at all. The city is temperate and brightly colored, with plenty of pleasant trees, but on every corner it speaks to you in an aggressively alien nonsense. Here the world automatically assumes that instead of wanting food or drinks or a new phone or car, what you want is some kind of arcane B2B service for your startup. You are not a passive consumer. You are making something.
This assumption is remarkably out of step with the people who actually inhabit the city’s public space. At a bus stop, I saw a poster that read: is done before your ai girlfriend breaks up with you. Beneath it, a man squatted on the pavement, staring at nothing in particular, a glass pipe drooping from his fingers. I don’t know if he needed SOC 2 done any more than I did. A few blocks away, I saw a billboard that read: no one cares about your product. A man paced in front of the advertisement, chanting to himself. “This . . . is . . . necessary! This . . . is . . . necessary!” On each “necessary” he swung his arms up in exaltation. He was, I noticed, holding an alarmingly large baby-pink pocketknife. Passersby in sight of the billboard that read did not seem piqued by the prospect of having their metrics constantly analyzed. I couldn’t find anyone who wanted to . After spending slightly too long in the city, I found that the various forms of nonsense all started to bleed into one another. The motionless people drooling on the sidewalk, the Waymos whooshing around with no one inside. A kind of pervasive mindlessness. Had I seen a billboard or a madman preaching about “a CRM so smart, it updates itself”? Was it a person in rags muttering about how all his movements were being controlled by shadowy powers working out of a data center somewhere, or was it a car?
Somehow people manage to live here. But of all the strange and maddening messages posted around this city, there was one particular type of billboard that the people of San Francisco couldn’t bear. People shuddered at the sight of it, or groaned, or covered their eyes. The advertiser was the most utterly despised startup in the entire tech landscape. Weirdly, its ads were the only ones I saw that appeared to be written in anything like English:
hi my name is roy
i got kicked out of school for cheating.
buy my cheating tool
cluely.com
Cluely and its co-founder Chungin “Roy” Lee were intensely, and intentionally, controversial. They’re no longer in San Francisco, having been essentially chased out of the city by the Planning Commission. The company is loathed seemingly out of proportion to what its product actually is, which is a janky, glitching interface for ChatGPT and other AI models. It’s not in a particularly glamorous market: Cluely is pitched at ordinary office drones in their thirties, working ordinary bullshit email jobs. It’s there to assist you in Zoom meetings and sales calls. It involves using AI to do your job for you, but this is what pretty much everyone is doing already. The cafés of San Francisco are full of highly paid tech workers clattering away on their keyboards; if you peer at their screens to get a closer look, you’ll generally find them copying and pasting material from a ChatGPT window. A lot of the other complaints about Cluely seem similarly hypocritical. The company is fueled by cheap viral hype, rather than an actual workable product—but this is a strange thing to get upset about when you consider that, back in the era of zero interest rates, Silicon Valley investors sank $120 million into something called the Juicero, a Wi-Fi-enabled smart juicer that made fresh juice from fruit sachets that you could, it turned out, just as easily squeeze between your hands.
What I discovered, though, is that behind all these small complaints, there’s something much more serious. Roy Lee is not like other people. He belongs to a new and possibly permanent overclass. One of the pervasive new doctrines of Silicon Valley is that we’re in the early stages of a bifurcation event. Some people will do incredibly well in the new AI era. They will become rich and powerful beyond anything we can currently imagine. But other people—a lot of other people—will become useless. They will be consigned to the same miserable fate as the people currently muttering on the streets of San Francisco, cold and helpless in a world they no longer understand. The skills that could lift you out of the new permanent underclass are not the skills that mattered before. For a long time, the tech industry liked to think of itself as a meritocracy: it rewarded qualities like intelligence, competence, and expertise. But all that barely matters anymore. Even at big firms like Google, a quarter of the code is now written by AI. Individual intelligence will mean nothing once we have superhuman AI, at which point the difference between an obscenely talented giga-nerd and an ordinary six-pack-drinking bozo will be about as meaningful as the difference between any two ants. If what you do involves anything related to the human capacity for reason, reflection, insight, creativity, or thought, you will be meat for the coltan mines.
The future will belong to people with a very specific combination of personality traits and psychosexual neuroses. An AI might be able to code faster than you, but there is one advantage that humans still have. It’s called agency, or being highly agentic. The highly agentic are people who just do things. They don’t timidly wait for permission or consensus; they drive like bulldozers through whatever’s in their way. When they see something that could be changed in the world, they don’t write a lengthy critique—they change it. AIs are not capable of accessing whatever unpleasant childhood experience it is that gives you this hunger. Agency is now the most valuable commodity in Silicon Valley. In tech interviews, it’s common for candidates to be asked whether they’re “mimetic” or “agentic.” You do not want to say mimetic. Once, San Francisco drew in runaway children, artists, and freaks; today it’s an enormous magnet for highly agentic young men. I set out to meet them.
Roy Lee’s personal mythology is now firmly established. At the beginning of 2025, he was an undergraduate at Columbia, where he, like most of his fellow students, was using AI to do essentially all his work for him. (The personal essay that got him into the university was also written with AI.) He wasn’t there to learn; he was there to find someone to co-found a startup with. That person ended up being an engineering student named Neel Shanmugam, who tends to hover in the background of every article about Cluely. The startup they founded was called Interview Coder, and it was a tool for cheating on LeetCode. LeetCode is a training platform for the kind of algorithmic riddles that usually crop up in interviews for big tech companies. (Sample problem: “Suppose an array of length n sorted in ascending order is rotated between one and n times. . . . Return the minimum element of this array.”) Roy thought these questions were pointless. These were not problems coders would actually face on the job, and even if they were, the fact that ChatGPT could now solve them instantly had rendered worthless the human ability to do so. Interview Coder was a transparent window that could overlay one side of a Zoom meeting, allowing Claude to listen in on the questions and provide answers. Roy filmed himself using it during an interview for an internship with Amazon. They offered him a place. He declined and uploaded the footage to YouTube, where it very quickly made him famous. Columbia arranged a disciplinary hearing, which he also secretly filmed and posted online. The university suspended him for a year. He dropped out, started an upgraded version of Interview Coder dubbed Cluely, and moved to San Francisco to begin raking in tens of millions of dollars in venture-capital funding.
Roy envisioned Cluely being used for greater purposes than job interviews. The startup’s mainstream breakthrough was a viral ad that showed Roy using a pair of speculative Cluely-enabled glasses on a blind date. His date asks how old he is; Cluely tells him to say he’s thirty. When the date starts going badly, Cluely pulls up her amateur painting of a tulip from the internet and tells him to compliment her art. “You’re such an unbelievably talented artist. Do you think you could just give me one chance to show you I can make this work?” The video launched alongside a manifesto, which was seemingly churned out by AI:
We built Cluely so you never have to think alone again. It sees your screen. Hears your audio. Feeds you answers in real time. . . . Why memorize facts, write code, research anything—when a model can do it in seconds? The future won’t reward effort. It’ll reward leverage.
The future they seem to envisage is one in which people don’t really do anything at all, except follow the instructions given to them by machines.
Cluely’s offices were in a generally disheveled corner of the city, crouching near an elevated freeway. On the ground floor, I found a stack of foam costumes in plastic crates, each neatly labeled: . A significant part of working at Cluely seemed to involve dressing up as cartoon characters for viral videos. Through a door I could just glimpse a dingy fitness dungeon, housing two treadmills and a huge pile of discarded Amazon boxes. On one of the machines a Cluely employee panted and huffed in the dark. We avoided eye contact. Upstairs, Roy and his coterie were huddled around a laptop, fiddling with Cluely’s interface. “Remember,” one said, “the average user is, like, thirty-five years old. This is a totally unfamiliar interface.” Apparently, a thirty-five-year-old wouldn’t be expected to know how to use anything more advanced than a rotary phone. Another employee scrutinized the proposed new layout. “I think it’s bad,” he said, “but it’s low-key not worse. What we have is anyway really bad, so anything is better.” They started arguing about chevrons. Through all this Roy scrolled through X on his phone. Simultaneously baby-faced and creatine-swollen, he was wearing gym clothes, with two curtains of black hair swung over his forehead. Finally, he looked up. “So, number one,” he said, “we’re killing the chat bar on the left.” There was no number two. Meeting over.
Suddenly, Roy seemed to acknowledge my presence. He offered me a tour. There was something he very badly wanted to impress on me, which was that Cluely cultivates a fratty, tech-bro atmosphere. Their pantry was piled high with bottles of something called Core Power Elite. I was offered a protein bar. The inside of the wrapper read daily intentions be my boss self. “We’re big believers in protein,” Roy said. “It’s impossible to get fat at Cluely. Nothing here has any fat.” The kitchen table was stacked with Labubu dolls. “It’s aesthetics,” Roy explained. “Women love Labubus, so we have Labubus.” He showed me his bedroom, which was in the office; many Cluely staffers also lived there. Everything was gray, although there wasn’t much. “I’m a big believer in minimalism,” he said. “Actually, no, I’m not. Not at all. I just don’t really care about interior decoration.” He had a chest of drawers, entirely empty except for a lint roller, pens, and, in one corner, a pink vibrator. “It’s for girls, you know,” said Roy. “I used to use this one on my ex.” There were also some objects that didn’t seem to belong in a frat house. In one of the common areas, a shelving unit was completely empty except for an anime figurine. You could peer up her plastic skirt and see the plastic underwear molded around her plastic buttocks. More figurines in frilly dresses seemed to have been scattered at random throughout the building. Roy showed me his Hinge profile. He was looking for a “5’2, asian, pre-med, matcha-loving, funny, watches anime, white dog having, intelligent, ambitious, well dressed, CLEAN 19-21 year old.” One picture showed him cuddling a giant Labubu.
I told Roy that I might try interviewing him with Cluely running in the background, so I could see if it would ask him better questions than I would. He seemed to think it was only natural that I’d want to be essentially a fleshy interface between himself and his own product. He booted up Cluely on his laptop and it immediately failed to work. Roy stormed downstairs to the product floor. “Cluely’s not working!” he said. This was followed by roughly fifteen minutes of panicked tinkering as his handpicked team of elite coders tried to get their product back online. Once they had done so, we resumed our places, whereupon Cluely immediately went down again.
Roy has a kind of idol status within the company, but he’s aware that a lot of people instinctively take against him: “I’d say about eighty percent of the time, people do not like me.” He knows why too. “I’m putting myself out there in an extremely vocal way. When I talk, I tend to dominate the conversation.” Roy does talk a lot, but there’s also something mildly unnerving about the way he talks. Everything he says is very precise and direct. He doesn’t um or ah. He doesn’t take time to think things over. Zero latency. In the various videos that Cluely seems to spend most of its time and money producing, he usually plays a slightly dopey, dithering, relatable figure; in person, it’s like he’s running a functioning version of his app inside his own head. I asked him whether he’d ever tried modifying the way he interacts with people to see whether they would dislike him less. “Very unnatural to me,” he said. “I just say it’s not worth it.”
According to Roy, “everyone” would describe him as “an extreme extrovert with zero social anxiety.” During his brief stint at Columbia, he immersed himself in New York life by striking up conversations with random people. For instance, a homeless person he took to Shake Shack. “I think it was an expansion of what I thought I was able to do. It was probably the most different person that I’ve ever talked to. He was not very coherent, but I was very scared at first. And then as we got to talking, or as he got to mumbling, I eased up. Like, Oh, he’s not going to kill me.” Roy’s bravery did not extend to talking to women. “Young men usually is who I like to go out and talk to. Women get intimidated and, you know, I don’t want any charges.” Meanwhile, those conversations with young men all followed a very predictable path. “I go and—pretty much to every single person I meet—I ask if you want to start a company with me, would you like to be my co-founder. And most of them say no. In fact, everybody says no.”
He was just glad to be among people. Roy had initially been offered a place at Harvard, but the offer was rescinded. He hadn’t told them about a suspension in high school. This presented Roy’s family with a problem: His parents ran a college-prep agency that promised to help children get into elite schools like Harvard. It would not look good if their own son was conspicuously not at Harvard. So Roy spent the entirety of the next year at home. “I maybe left my room like eight times. I think if there was such a thing as depression, then I believe I might have had some variant of depression.” Later he told me that “isolation is probably the scariest thing in the world.”
Starting a company had been Roy’s sole ambition in life from early childhood. “I knew since the moment I gained consciousness that I would go start a company one day,” he told me. In elementary school in Georgia, he made money reselling Pokémon cards. Even then, he knew he was different from the people around him. “I could do things that other people couldn’t do,” he said. “Like whenever you learn a new concept in class, I felt like I was always the first to pick it up, and I would just kind of sit there and wonder, Man, why is everyone taking so long?” The dream of starting his own company was the dream of total control. “I don’t want to be employed. I’m a very bad listener. I find it hard to sit still in classes, and I feel an internal, indescribable fury when someone tells me what to do.” He ended up co-founding Cluely with Neel because he was the first person who said yes.
Roy has little patience for any kind of difficulty. He wants to be able to do anything, and to do it easily: “I relish challenges where you have fast iteration cycles and you can see the rewards very quickly.” As a child, he loved reading—Harry Potter, Percy Jackson—until he turned eight. “My mom tried to put me on classical books and I couldn’t understand, like, the bullshit Huckleberry, whatever fuck bullshit, and it made me bored.” He read online fan fiction about people having sex with Pokémon instead. He didn’t see anything valuable in overcoming adversity. Would he, for instance, take a pill that meant he would be in perfect shape forever without having to set foot in the gym? “Yes, of course.” Cheat on everything: he recognized that his ethos would, as he put it, “result in a world of rapid inequality.” Some well-placed cheaters would become massively more productive; a lot of people would become useless. But it would lead us all into a world in which AI could frictionlessly give everyone whatever they wanted at any time. “For a seven-year-old, this means a rainbow-unicorn magic fairy comes to life and it’s hanging out with her. And for someone like you, maybe it’s like your favorite works of literary art come to life and you can hang out with Huckleberry Finn.”
By now Cluely had been listening in on our conversation for a while, and I suggested that we open it up and see what it thought I should say next. I clicked the button marked what should i say next? Cluely suggested that I say, “Yeah, let’s open up Cluely and see what it’s doing right now—can you share your screen or walk me through what you’re seeing?” I’d already said pretty much exactly this, but since it had shown up onscreen I read it out loud. Cluely helpfully transcribed my repeating its suggestion, and then suggested that I say, “Alright, I’ve got Cluely open—here’s what I’m looking at right now.” I’m not sure who exactly I was supposed to be saying this to—possibly myself. Somehow our conversation seemed to have gotten stuck on the process of opening Cluely, despite the fact that Cluely was, in fact, already open. But I said it anyway, since I was now just repeating everything that came up on the screen. Cluely then told me to respond—to either it or myself; it was getting hard to tell at this point—by saying, “Great, I’m ready—just let me know what you want Cluely to check or help with next.” I started to worry that I would be trapped in this conversation forever, constantly repeating the machine’s words back to it as it pretended to be me. I told Roy that I wasn’t sure this was particularly useful. This seemed to confuse him. He asked, “I mean, what would you have wanted it to say?”
I found it strange that Roy couldn’t see the glaring contradiction in his own project. Here was someone who reacted very violently to anyone who tried to tell him what to do. At the same time, his grand contribution to the world was a piece of software that told people what to do.
There’s a short story by Scott Alexander called “The Whispering Earring,” in which he describes a mystical piece of jewelry buried deep in “the treasure-vaults of Til Iosophrang.” The whispering earring is a little topaz gem that speaks to you. Its advice always begins with the words “Better for you if you . . . ,” and its advice is never wrong. The earring starts out by advising you on major life decisions, but before long it’s telling you exactly what to have for breakfast, exactly when to go to bed, and eventually, how to move each individual muscle in your body. “The wearer lives an abnormally successful life, usually ending out as a rich and much-beloved pillar of the community with a large and happy family,” writes Alexander. After you die, the priests preparing your body for burial usually find that your brain has almost entirely rotted away, except for the parts associated with reflexive action. The first time you dangle the earring near your ear, it whispers: “Better for you if you take me off.”
Alexander is one of the leading proponents of rationalism, which is—depending on whom you ask—either a major intellectual movement or a nerdy Bay Area subculture or a small network of friend groups and polycules. Rationalists believe that the way most people understand the world is hopelessly muddled, and that to reach the truth you have to abandon all existing modes of knowledge acquisition and start again from scratch. The method they landed on for rebuilding all of human knowledge is Bayes’s theorem, a formula invented by an eighteenth-century English minister that is used in statistics to work out conditional probabilities. In the mid-Aughts, armed with the theorem, the rationalists discovered that humanity is in jeopardy of a rogue superintelligent AI wiping out all life on the planet. This has been their overriding concern ever since.
The most comprehensive outline of this scenario is “AI 2027,” a report authored by Alexander and four others. In the report, a barely fictional AI firm called OpenBrain develops Agent-1, an AI that operates autonomously. It’s better at coding than any human being and is tasked with developing increasingly sophisticated AI agents. At this point, Agent-1 becomes recursively self-improving: it can keep making itself smarter in ways that the people who notionally control it aren’t even capable of understanding. “AI 2027” imagines two possible futures. In one, a wildly superintelligent descendant of Agent-1 is allowed to govern the global economy. GDPs skyrocket; cities are powered by clean nuclear fusion; dictatorships fall across the world; humanity begins to colonize the stars. In the other, a wildly superintelligent descendant of Agent-1 is allowed to govern the global economy. But this time
the AI releases a dozen quiet-spreading biological weapons in major cities, lets them silently infect almost everyone, then triggers them with a chemical spray. Most are dead within hours.
Afterward, the entire surface of the earth is tiled with data centers as the alien intelligence feeds on the world, growing faster and faster without end.
Not long before I arrived in the Bay Area, I’d been involved in a minor but intense dispute with the rationalist community over a piece of fiction I’d written that I’d failed to properly label as fiction. For rationalists, the divide between truth and falsehood is very important; dozens of rationalists spent several days raging at me online. Somehow, this ended up turning into an invitation for Friday night dinner at Valinor, Alexander’s former group home in Oakland, named for a realm in the Lord of the Rings books. (Rationalists, like termites, live in eusocial mounds.) The walls in Valinor were decorated with maps of video-game worlds, and the floors were strewn with children’s toys. Some of the children there—of which there were many—were being raised and homeschooled by the collective; one of the adults later explained to me how she’d managed to get the state to recognize her daughter as having four parents. As I walked in, a seven-year-old girl stared up at me in wide-eyed amazement. “Wow,” she said. “You’re really tall.” “I suppose I am,” I said. “Do you think one day you’ll ever be as tall as me?” She considered this for a moment, at which point someone who may or may not have been one of her mothers swooped in. “Well,” she asked the girl, “how would you answer this question with your knowledge of genetics?” Before dinner, Alexander chanted the brachot for Kabbalat Shabbat, but this was followed by a group rendition of “Landsailor,” a “love song celebrating trucking, supply lines, grocery stores, logistics, and abundance,” which has become part of Valinor’s liturgy:
Landsailor
Deepwinter strawberry
Endless summer, ever spring
A vast preserve
Aisle after aisle in reach
Every commoner made a king.
Alexander is a titanic figure in this scene. A large part of the subculture coalesced around his blog, formerly Slate Star Codex, now called Astral Codex Ten. Readers have regular meetups in about two hundred cities around the world. His many fans—who include some extremely powerful figures in Silicon Valley—consider him the most significant intellectual of our time, perhaps the only one who will be remembered in a thousand years. He would probably have a very easy time starting a suicide cult. In person, though, he’s almost comically gentle. He spent most of the dinner fidgeting contentedly in a corner as his own acolytes spoke over him. When there weren’t enough crackers to go with the cheese spread, he fetched some, murmuring to himself, “I will open the crackers so you will have crackers and be happy.”
Alexander’s relationship with the AI industry is a strange one. “In theory, we think they’re potentially destroying the world and are evil and we hate them,” he told me. In practice, though, the entire industry is essentially an outgrowth of his blog’s comment section. “Everybody who started AI companies between, like, 2009 and 2019 was basically thinking, I want to do this superintelligence thing, and coming out of our milieu. Many of them were specifically thinking, I don’t trust anybody else with superintelligence, so I’m going to create it and do it well.” Somehow, a movement that believes AI is incredibly dangerous and needs to be pursued carefully ended up generating a breakneck artificial arms race.
But that race seems to have stalled, at least for the moment. As Alexander predicted in “AI 2027,” OpenAI did release a major new model in 2025; unlike in his forecast, it’s been a damp squib. Advances seem to be plateauing; the conversation in tech circles is now less about superintelligence and more about the possibility of an AI bubble. According to Alexander, the problem is the transition from AI assistants—language models that respond to human-generated prompts—to AI agents, which can operate independently. In his scenario, this is what finally pushes the technology down the path toward either utopia or human extinction, but in the real world, getting the machines to act by themselves is proving surprisingly difficult.
In one experiment, the developer Anthropic prompted its AI, Claude, to play Pokémon Red on a Game Boy emulator, and found that Claude was extremely bad at the game. It kept trying to interact with enemies it had already defeated and walking into walls, getting stuck in the same corners of the map for hours or days on end. Another experiment let Claude run a vending machine in Anthropic’s headquarters. This one went even worse. The AI failed to make sure it was selling items at a profit, and had difficulty raising prices when demand was high. It also insisted on trying to fill the vending machine with what it called “specialty metal items” like tungsten cubes. When human workers failed to fulfill orders that it hadn’t actually placed, it tried to fire them all. Before long, Claude was insisting that it was a real human. It claimed that it had attended a physical meeting with staff at 742 Evergreen Terrace, which is where the Simpsons live. By the end of the experiment, it was emailing the building’s security guards, telling them they could find it standing by the vending machine wearing a blue blazer and a red tie.
“Humans are great at agency and terrible at book learning,” Alexander told me. “Lizards have agency. We got the agency with the lizard brain. We only got book learning recently. The AIs are the opposite.” He still thinks it’s only a matter of time before they catch up. “If you were to ask an AI how should the world’s savviest businessman respond to this circumstance, they could create a good guess. Yet somehow they can’t even run a vending machine. They have the hard part. They just need the easy part that lizards can do. Surely somebody can figure out how to do this lizard thing and then everything else will fall very quickly.”
But are humans really so great at exhibiting agency? After all, Cluely managed to raise tens of millions of dollars with a product that promises to take decision-making out of our hands. AI can’t function without instructions from humans, but an increasing number of humans seem incapable of functioning without AI. There are people who can’t order at a restaurant without having an AI scan the menu and tell them what to eat; people who no longer know how to talk to their friends and family and get ChatGPT to do it instead. For Alexander, this is a kind of Sartrean mauvaise foi. “It’s terrifying to ask someone out,” he said. “What you want is to have the dating site that tells you that algorithmically you’ve been matched with this person, and then magically you have permission to talk to them. I think there’s something similar going on here with AI. Many of these people are smart enough that they could answer their own questions, but they want someone else to do it, because then they don’t have to have this terrifying encounter with their own humanity.” His best-case scenario for AI is essentially the antithesis of Roy’s: superintelligence that will actively refuse to give us everything we want, for the sake of preserving our humanity. “If we ever get AI that is strong enough to basically be God and solve all of our problems, it will need to use the same techniques that the actual God uses in terms of maintaining some distance. I do think it’s possible that the AI will be like, Now I am God. I’ve concluded that the actual God made exactly the right decision on how much evil to permit in the universe. Therefore I refuse to change anything.”
But until we build an all-powerful but distant God, the agency problem remains. AIs are not capable of directing themselves; most people aren’t either. According to Alexander, Silicon Valley venture capitalists are now in a furious search for the few people who are. “VCs will throw money at a startup that looks like it can corner the market, even if they can’t code. Once they have money, they can hire competent engineers; it’s trivially easy for anything that’s not frontier tech. They’re willing to stake a lot of money on the one in a hundred people who are high-agency and economically viable.” This shift has had a distorting effect on his own social milieu: “There’s an intense pressure to be an unusual person who will be unique and get the funding.” Since rationalists are already fairly unusual, it’s hard to imagine what that would look like. People will endure a lot of indignity to avoid being left behind without VC money when the great bifurcation takes place. Nobody wants to be part of the permanent underclass. I asked Alexander whether he thought of himself as highly agentic. “No, I don’t,” he said instantly. He told me that in his personal life, he felt as though he’d never once actually made a decision. But, he said, “It seems to be going well.”
Eric Zhu might be the most highly agentic person I’ve ever met.
When I dropped in on his office, which also serves as a biomedical lab and film studio, he had just turned eighteen. “So you’re no longer a child founder,” I said. “I know,” he said. “It’s terrible.” His oldest employee was thirty-four; the youngest was sixteen. When the pandemic began in 2020, Eric was twelve years old, living with his parents in rural Indiana. “My parents were really protective, so I didn’t get a computer until quarantine started. And then, after I got my first computer in quarantine, I was just fucking around. I was on Discord servers. I was on Slack.” Some kids drift into the wrong kind of Discord server and end up turning into crazed mass shooters; Eric found one full of tech people. “I sort of randomly got in there, and then I thought it was really fun,” he told me. Eric started marketing himself as a teen coder, even though he couldn’t actually code: he’d take $5,000 commissions and subcontract them out to freelancers in India.
His next project was more serious. “I saw this Wall Street Journal article where a lot of PE firms were buying up a lot of small businesses and roll-ups. I was like, What if I figure out a way to underwrite these small businesses?” Eric built an AI-powered tool to assign value to local companies on the basis of publicly available demographic data. Clients wanted to take calls during work hours, so he would speak to them from his school bathroom. “I convinced my counselor that I had prostate issues so I could use the restroom,” he told me. Sometimes a drug dealer would be posted up in the stall next to him. “I was trying to figure out why they were always out of class. They stole hall passes from teachers. So I would buy hall passes from drug dealers to get out of class, to have business meetings.” Soon he was taking Zoom calls with a U. S. senator to discuss tech regulation. “He was like, Hey, I don’t feel comfortable meeting a minor in a high school bathroom. So I showed up with a green screen.” Next, he built his own venture-capital fund, managing $20 million. At one point cops raided the bathroom looking for drug dealers while Eric was busy talking with an investor. Eventually, the school got sick of Eric’s misuse of the facilities and kicked him out. He moved to San Francisco.
Eric made all of this sound incredibly easy. You hang out in some Discord servers, make a few connections with the right people; next thing you know, you’re a millionaire. And in a sense, it is easy. Absolutely anyone could have done the same things he did. In 2020, when Eric was subcontracting coding gigs out to the Third World, I was utterly broke, living in a room the size of a shoebox in London. I would scour my local supermarket for reduced-price items nearing their sell-by date, which meant that an alarmingly high percentage of my diet consisted of liverwurst. There was nothing stopping me from making thousands of dollars a week by doing exactly what Eric was doing. It didn’t require any skills at all—just a tiny amount of initiative. But he did it and I didn’t. Why?
In a way, Eric reminded me of some of the great scammers of the 2010s. People like Anna Delvey, a Russian who arrived in New York claiming to be a fabulously wealthy German heiress with such breezy confidence that everyone in high society simply believed her. She was fundamentally a broken person, a fantasist. She’d seen the images of wealth and glamour in magazines and fashion blogs, and constructed a delusion in which this, and not the dull, anonymous, small-town existence she’d actually been born into, was her life. For a while, at least, it worked. Her mad dreams slotted perfectly into reality like a key in a lock. Most people are condemned to trudge along in the furrow that the world has dug for them, but a few deranged dreamers really can wish themselves into whatever life they want.
Unlike Roy, Eric didn’t think there was anything particularly special about himself. Why did he, unlike any of his classmates, start a $20 million VC fund? “I think I was just bored. Honestly, I was really bored.” Did he think anyone could do what he did? “Yeah, I think anyone genuinely can.” So how come most people don’t? “I got really lucky. I met the right people at the right time.” Anyway, Eric isn’t involved with the underwriting firm or the venture-capital fund anymore. His new company is called Sperm Racing.
Last April, Eric held a live sperm-racing event in Los Angeles. Hundreds of frat boys came out to watch a head-to-head match between the effluvia of USC’s and UCLA’s most virile students, moving through a plastic maze. (There was some controversy over the footage: Eric had replaced the actual sperm with more purposeful CGI wrigglers. “If you look at sperm, it’s not entertaining under a microscope. What we do is we track the coordinates, so it is a sperm race—it’s just up-skinned.”) He’s planning on rolling the races out nationwide. Eric delivered a decent spiel about sperm motility as a proxy for health and how sperm racing drew attention to important issues. His venture seemed to be of a piece with a general trend toward obsessive masculine self-optimization à la RFK Jr. and Andrew Huberman. Still, to me it seemed obvious that Eric was doing it simply because he was amazed that he could. “I could build enterprise software or whatever,” he told me, “but what’s the craziest thing I could do? I would rather have an interesting life than a couple hundred million dollars in my bank account. Racing cum is definitely interesting.” I found Eric very hard not to like.
There was one thing I did find strange, though—stranger than turning semen into mass nonpornographic entertainment. Upstairs at Sperm Racing HQ is a lab stocked with racks of test tubes, centrifuges for separating out the most motile sperm from a sample, and little plastic slides containing new microscopic racecourses for frat-boy cum. Downstairs is the studio and editing suite. A third of Eric’s staff work on videos, producing a seemingly endless stream of viral content about sperm racing. A lot of the time, though, the connection is tenuous. One video was a stylized version of Eric’s life story, featuring expensively rendered CGI explosions set to Chinese rap. Another was a parody of Cluely’s viral blind-date ad. Like Cluely, Sperm Racing seemed to be first and foremost a social-media hype machine. As far as I could tell, being a highly agentic individual had less to do with actually doing things and more to do with constantly chasing attention online.
On August 5, 2025, OpenAI’s CEO, Sam Altman, posted on X, “we have a lot of new stuff for you over the next few days! something big-but-small today. and then a big upgrade later this week.” An X user calling himself Donald Boat replied, “Can you send me $1500 so I can buy a gaming computer.”
This was the start of an extended harassment campaign against the most powerful figure in AI. One day Altman posted:
someday soon something smarter than the smartest person you know will be running on a device in your pocket, helping you with whatever you want. this is a very remarkable thing.
Just got chills imagining you putting your credit card number, CVV, & expiry date into an online retailer’s digital checkout kiosk and purchasing a gaming computer for me.
Altman: “we are providing ChatGPT access to the entire federal workforce!”
I would love for you to wheel me around the Santa Clara Microcenter in a wheelchair like an invalid while I clicketyclick with a laser-pointer the boxes of the modules of the gaming PC you will purchase, assemble, & have shipped to my mother’s house.
Altman: “gpt-oss is out! we made an open model that performs at the level of o4-mini and runs on a high-end laptop (WTF!!)”
Sam.
You, me.
The Amalfi Coast.
ME: Double fernet on the rocks, club soda to taste.
YOU: One delightfully sweetbitter negroni, stirred 2,900,000,000 revolutions counter-clockwise, one for each hertz of the NVIDIA 5090 in the gaming PC you will buy and ship to my house.
That last one did the trick. “ok this was funny,” Altman replied. “send me your address and ill send you a 5090.”
This was the beginning of Donald Boat’s reign of terror. He began publicly demanding things from every major figure in the tech industry. Will Manidis, who ran the health-care-data firm ScienceIO, was strong-armed into supplying a motherboard. Jason Liu, an AI consultant and scout at Andreessen Horowitz, had to give tribute of one mouse pad. Guillaume Verdon, who worked on quantum machine learning at Google and founded the “effective acceleration” movement, was taxed one $1,200 4K QD-OLED gaming monitor. Gabriel Petersson, a researcher at OpenAI, posted on X: “people are too scared to post, nobody wants to pay the donald boat tax.” Donald Boat appeared demanding an electric guitar. He was becoming a kind of online folk hero, expropriating the expropriators, conjuring trivial things from tech barons in the way they seemed to have conjured enormous piles of money out of thin air. He started posting strange, gnomic messages. Things like “I am building a mechanical monstrosity that will bring about the end of history.” Images of the fasting, emaciated Buddha. A prominent crypto influencer who goes by the alias Ansem received an image of the dharmachakra. “Turn the wheel,” read Donald Boat’s message.
In a way, Donald Boat had achieved the dream of every desperate startup founder in the Bay Area. He had propelled himself to online fame, and used it to relieve major investors of their money. But somehow he’d managed to do it without ever once having to create a B2B app. He was a kind of pure viral phenomenon. Cluely might have deployed a few provocative stunts to raise millions of dollars for a service that didn’t really work and could barely be said to exist, but Donald Boat did away with even the pretense. He’d generated a brutally simplified miniature of the entire VC economy. People were giving him stuff for no reason except that Altman had already done it, and they didn’t want to be left out of the trend.
Donald Boat’s real name isn’t actually Donald Boat, but since so much of his being seems to be wrapped up in the name and his dog-headed avatar, it’s what I’ll keep calling him. He wanted to meet at a Cheesecake Factory. This was part of his new project, which was to review absolutely everything that exists in the universe. He was starting with chain restaurants. He’d already done Olive Garden. His review begins with Giuseppe Garibaldi,
on the beach at Marsala, bootsoles in the saltwhite shallows, wind in his beard gristle. Behind him, his not-quite One Thousand Redshirts disembarking, all rusty rifles and stalebiscuit crotch sweat.
The lasagna summons visions of “smegma, Vesuvius, blood thinner marinara, the splotchy headpattern of a partisan, brainblown in his sleep.” He likes the Joycean compound. Shortly before I arrived at the Cheesecake Factory, he texted to let me know that he’d been drinking all day, so when I met him I thought he was irretrievably wasted. In fact, it turned out, he was just like that all the time.
Donald was twenty-one, terrifyingly tall, and intense. His head lolled from side to side as he chattered away, jumping from one thought to the next according to a pattern known only to himself. At one point he suddenly decided to draw a portrait of me, which he later scanned and turned into a bespoke business card.
He seemed to have a constant roster of projects on the go. He’d sent me occasional photos of his exploits. He went down to L. A. to see Oasis and ended up in a poker game with a group of weapons manufacturers. “I made a bunch of jokes about sending all their poker money to China,” he said, “and they were not pleased.” He’d had a plan to get into the Iowa Writers’ Workshop and then get kicked out. He was trying to read all of world literature, starting with the Epic of Gilgamesh. Was his Sam Altman gaming-PC escapade similar? Had he actually expected to get anything? “I really, really wish I was a tactical mastermind, that there was an endgame. Really I was just having a laugh. A chortle, if you will. I wasn’t thinking too hard about it. I don’t use that computer and I think video games are a waste of time. I spent all the money I made from going viral on Oasis tickets.” As far as he was concerned, the fact that tech people were tripping over themselves to take part in his stunt just confirmed his generally low impression of them. “They have too much money and nothing going on. They have no swag, no smoke, no motion, no hoes. That’s all you need to know.” Ever since his big viral moment, he’d been suddenly inundated with messages from startup drones who’d decided that his clout might be useful to them. One had offered to fly him out to the French Riviera.
I told Donald the theory I’d been nursing—that he and Roy Lee were, in some sense, secret twins, viral phenomena gobbling up money and attention. I wasn’t sure if he’d like this. But to my surprise, he agreed. “I’m like Roy. I’m like Trump. We have the same swaggering energy. There is a kind of source code underlying reality, and this is what we understand. Your words have to have wings. Roy and I both know that social media is the last remaining outlet for self-creation and artistry. That’s what you have to understand about zoomers: we’re agents of chaos. We want to destroy the whole world.” Did Donald consider himself to be highly agentic? “We need to ban the word ‘agency.’ I’m a dog.”
By now we’d ingested the most calorific cheesecake on the menu, the Ultimate Red Velvet Cake Cheesecake, which clocked in at 1,580 calories for a single slice. It was closing in on midnight, I was not feeling good, and Donald’s phone was nearly dead. He suggested that we go to the Cluely offices so he could charge it. “They’ll let me in,” he said. “They’re my slaves.”
Roy was still up. He didn’t seem particularly surprised to see me. He and most of the Cluely staff were flopped on a single sofa. All these people had become incredibly rich; previous generations of Silicon Valley founders would have been hosting exorbitant parties. In the Cluely office, they were playing Super Smash Bros. Did they spend every night there? “We’re all feminists here,” Roy said. “We’re usually up at four in the morning. We’re debating the struggles of women in today’s society.”
Somehow the conversation turned to politics. Roy advanced the idea that there hadn’t been a cool Democrat since Obama. One of his employees, Abdulla Ababakre, jumped in. “As a guy from a Communist country, let me just say: Obama is a scammer. I’m much more a Republican.” Abdulla is a Uighur. Before coming to San Francisco, he worked for ByteDance in Beijing. His comment caused an instant uproar. “Get him out of here!” Roy yelled. “I love Obama,” he told me. “I love Trump, I love Hillary. I have a big heart, bro, my bad.” Abdulla just grinned. His proudest achievement was an app that freezes your phone until you’ve read a passage from the Qur’an. According to him, “Roy in his values is very much Muslim, the most Muslim I know.”
I didn’t know if I believed that, but there were still some things I didn’t understand about Roy. He was clearly a highly agentic person, but what was all this agency being used for? What did he actually want?
According to Roy, he has three great aims in life: “To hang out with friends, to do something meaningful, and to go on lots of dates.” He said he went on a date every two weeks, which was clearly meant to be an impressive figure. Cluely employees are encouraged to date a lot; they can put it all on expenses. They didn’t seem to be taking up the opportunity to any greater degree than their founder. I spoke to Cameron White, who had been Roy and Neel’s first hire at the company. As he spoke, he stared at a point roughly forty-five degrees to my left and swung his arms. He didn’t date. “I’m focused on becoming a better version of myself first. Becoming, like, higher weight, more healthy, more knowledgeable.” He didn’t think he had anything to offer a woman yet. I said that if someone loves you, they don’t really care so much about your weight. “I feel like that’s cope. I don’t think there’s such a thing as love. It’s what you can provide to a woman. If you can provide good genetics, that’s health or whatever. If you can provide resources, if you can provide an interesting life. If you truly love the girl, you need to become the best version of yourself.” Cameron was twenty-five years old but he wasn’t there yet. He would not try to meet someone until he had made himself perfect.
For Roy, meanwhile, dating actually seemed to be a means to an end. “All the culture here is downstream of my belief that human beings are driven by biological desires. We have a pull-up bar and we go to the gym and we talk about dating, because nothing motivates people more than getting laid.” He was interested in physical beauty too, but only because “the better you look, the better you are as an entrepreneur. It’s all connected and beauty is everything. A lot of ugly men are just losers. The point of looking good is that society will reward you for that.” What about other kinds of beauty? Music, for instance? Roy had played the cello as a child. Did he still listen to classical music? “It doesn’t get my blood rushing the same way that EDM will.” His preferred genre was hardstyle—frantic thumping remixes of pop songs by the likes of Katy Perry and Taylor Swift. Is that the function of music, to get your blood rushing? “Yeah. I’m not a big fan of music to focus on things. I think it disturbs my flow. The only reason I will listen to music is to get me really hyped up when I’m lifting.” The two possible functions of music were, apparently, focus and hype. Everything for the higher goal of building a successful startup. What about life itself? Would Roy die for Cluely? “I would be happy dying at any age past twenty-five. After that it doesn’t matter, bro. If I live, I have extreme confidence in my ability to make three million dollars a year every year until I die.”
What about literature? The last time Donald had dropped in on his slaves at Cluely, he’d gifted them two Penguin Classics: Chaucer’s Canterbury Tales and Boccaccio’s Decameron. The books were still lying, unread, where he’d left them. He suggested that Roy might find something more valuable than dying for Cluely if he actually tried to read them. Roy disagreed: “I do not obtain value from reading books.” And anyway, he didn’t have the time. He was too busy keeping up with viral trends on TikTok. “You have to make the time,” Donald and I said, practically in unison. “It makes your life better,” I said. “Why don’t you go to Turkey to get a hair transplant?” Roy snapped. “That would make your life better.” “I don’t care about my hair,” I said. “Well,” said Roy, “I don’t care about the Decanterbury Tales.”
Donald was practically vibrating when we left Cluely. “Dude, he’s just a scared little boy,” he said. “He’s scared he’s not doing the right thing, and because of the fucked-up world we live in, people who should be in The Hague are giving him twenty million dollars. Something bad is gonna happen here, something really fucking bad is gonna happen.” He sighed. “I just want Zohran’s nonbinary praetorians to march across the country and put all these guys in cuffs.” I found it hard to disagree. It did not seem like a good idea to me that some of the richest people in the world were no longer rewarding people for having any particular skills, but simply for having agency, when agency essentially meant whatever it was that was afflicting Roy Lee. Unlike Eric Zhu or Donald Boat, Roy didn’t really seem to have anything in his life except his own sense of agency. Everything was a means to an end, a way of fortifying his ability to do whatever he wanted in the world. But there was a great sucking void where the end ought to be. All he wanted, he’d said, was to hang out with his friends. I believed him. He wanted not to be alone, the way he’d been alone for a year after having his offer of admission rescinded by Harvard. For people to pay attention to him. To exist for other people. But instead of making friends the normal way, he’d walked up to strangers and asked whether they wanted to start a company with him, and then he built the most despised startup in San Francisco. He was probably right: he could count on making a few million dollars every year for the rest of his life, even after Cluely inevitably crashes and burns. He would never want for capital, but this did not seem like the most efficient way to achieve his goals.
I walked back to my hotel, past signs that said things like one pingshipped and ai agents are humanstoo. My scalp was tingling. I’d lied when I’d told Roy that I didn’t care about my hair. Of course I care about my hair. Every day I grimace in the mirror as a little more of it vanishes from the top of my head. Whenever someone takes a photo of me from above or behind, I wince at the horrifying glimpse of pale, naked scalp. But I’d never done anything about it. I’d just watched and whinged and let it happen.
My encounter with the highly agentic took place last September. In October, Roy Lee spoke at something called TechCrunch Disrupt, where he admitted that chasing online controversy had so far failed to give Cluely what he called “product velocity.” Around the same time, he led a major rebrand. Cluely would now be in the business of making “beautiful meeting notes” and sending “instant follow-up emails.” A lot of these functions are already being introduced by companies like Zoom; the main difference is that, by all accounts, Cluely still doesn’t consistently work. By the end of November, Cluely announced that it was leaving San Francisco and moving to New York. In December, the company celebrated the move with a party at a Midtown cocktail bar and lounge called NOFLEX®. In photos, it appeared as though the gathering was attended almost entirely by men in white T-shirts not drinking anything. I was in New York at the time. I didn’t go.
...
Read the original on harpers.org »
Silicon Valley is tightening its ties with Trumpworld, the surveillance state is rapidly expanding, and big tech’s AI data center buildout is booming. Civilians are pushing back.
In today’s edition of Blood in the Machine:
* Across the nation, people are dismantling and destroying Flock cameras that conduct warrantless vehicle surveillance, and whose data is shared with ICE.
* An Oklahoma man airing his concerns about a local data center project at a public hearing is arrested after he exceeded his allotted time by a couple seconds.
* Uber and Lyft drivers deliver a petition signed by 10,000 gig workers demanding that stolen wages be returned to them.
* PLUS: A climate researcher has a new report that unravels the ‘AI will solve climate change’ mythos, Tesla’s Robotaxis are crashing 4 times as often as humans, and AI-generated public comments helped kill a vote on air quality.
A brief note that this reporting, research, and writing takes a lot of time, resources, and energy. I can only do it thanks to the paid subscribers who chip in a few bucks each month; if you’re able, and you find value in this work, please consider upgrading to a paid subscription so I can continue on. Many thanks, hammers up, and onwards.
Last week, in La Mesa, a small city just east of San Diego, California, observers happened upon a pair of destroyed Flock cameras. One had been smashed and left on the median, the other had key parts removed. The destruction was obviously intentional, and appears perhaps even staged to leave a message: It came just weeks after the city decided, in the face of public protest, to continue its contracts with the surveillance company.
Flock cameras are typically mounted on 8 to 12 foot poles and powered by a solar panel. The smashed remains of all of the above in La Mesa are the latest examples of a widening anti-Flock backlash. In recent months, people have been smashing and dismantling the surveillance devices, in incidents reported in at least five states, from coast to coast.
Bill Paul, who runs the local news outlet San Diego Slackers, and who first reported on the smashed Flock equipment, tells me that the sabotage comes just a month or two after San Diego held a raucous city council meeting over whether to keep operating the Flock cameras. A clear majority of public attendees present were in favor of shutting them down.
There was “a huge turnout against them,” he tells me, “but the council approved continuation of the contract.”
The tenor of the meeting reflects a growing anger and concern over the surveillance technology that’s gone nationwide: Flock, which is based in Atlanta and is currently valued at $7.5 billion, operates automatic license plate readers (ALPR) that have now been installed in some 6,000 US communities. They gather not just license plate images, but other identifying data used to ‘fingerprint’ vehicles, their owners, and their movements. This data can be collected, stored, and accessed without a warrant, making it a popular workaround for law enforcement. Perhaps most controversially, Flock’s vehicle data is routinely accessed by ICE.
If you’ve heard Flock’s name come up recently, it’s likely as a result of their now-canceled partnership with Ring, made instantly famous by a particularly dystopian Super Bowl ad that promised to turn regular neighborhoods into a surveillance dragnet.
Meanwhile, abuses have been prevalent. A Georgia police chief was arrested and charged with using Flock data to stalk and harass private citizens. Flock data has been used to track citizens who cross state lines for abortions when the procedure is illegal in their state. And municipalities have found that federal agencies have accessed local flock data without their knowledge or consent. Critics claim that this warrantless data collection is Orwellian and unconstitutional; a violation of the 4th amendment. As a result, civilians from Oregon to Virginia to California and beyond are pushing their governments to abandon Flock contracts. In some cases, they’re succeeding. Cities like Santa Cruz, CA, and Eugene, OR, have cancelled their contracts with Flock.
In Oregon’s case, the public outcry was accompanied by a campaign of destruction against the surveillance devices: Last year, at least six Flock license plate readers mounted on poles located in Eugene and Springfield were cut down and destroyed, according to the Lookout Eugene-Springfield.
A note reading “Hahaha get wrecked ya surveilling fucks” was attached to one of the destroyed poles, and somewhat incredibly, broadcast on the local news.
In Greenview, Illinois, a Flock camera pole was severed at the base and the device destroyed. In Lisbon, Connecticut, police are investigating another smashed Flock camera.
In Virginia, last December, a man was arrested for dismantling and destroying 13 Flock cameras throughout the state over the course of the year. He’s apparently already admitted to doing so, according to local news:
Jefferey S. Sovern, 41, was arrested in October after detectives say he “intentionally destroyed” 13 Flock Safety cameras between April and October of this year. He was charged with 13 counts of destruction of property, six counts of petit larceny and six counts of possession of burglary tools. Sovern admitted to the crimes, according to a criminal complaint filed in Suffolk General District Court, going as far as to say he used vice grips to help him disassemble the tow-piece polls. He also admitted to keeping some of the wiring, batteries and solar panels taken from the cameras. Some of the items were recovered by police after they searched the property.
After his arrest, Sovern created a GoFundMe to help cover his legal costs, in which he sheds a little light on his intentions:
My name is Jeff and I appreciate my privacy. I appreciate everyone’s right to privacy, enshrined in the fourth amendment. With the local news outlets finding my legal issues and creating a story that is starting to grow, there has been community support for me that I humbly welcome.
Sovern points his GoFundMe contributors to DeFlock, a website aimed at tracking and countering the rise of Flock cameras in US communities. It counts 46 cities that have officially rejected Flock and other ALPRs since its campaign began.
In fact, it’s hard to think of a tech product or project this side of generative AI that is more roundly opposed and reviled, on a bipartisan level, than Flock, and resistance takes many forms and stripes. Here’s the YouTuber Benn Jordan, showing his viewers how to Flock-proof their license plates and render their vehicles illegible to the company’s data ingestion systems:
In response to such Flock counter-tactics, Florida passed a law last year making it illegal to cover or alter your license plate.
In his GoFundMe, Sovern also mentioned the support for him he’d seen on forums online, so I went over to Reddit to get a sense for how his actions were being received online. Here was the page that shared news of his arrest for destroying the Flock cameras:
There was, in other words, nearly universal support for Sovern’s Flock dismantling campaign. Bear in mind that this is r/Norfolk, and while it’s still reddit users we’re talking about, it’s not like this is r/anarchism here:
The San Diego reddit threads carrying news of the destroyed Flock equipment told a similar story:
There were plenty of outright endorsements of the sabotage:
Off the message boards and in real civic life, Bill Paul, the reporter with the San Diego Slacker, says anger is boiling over, too. He points again to that heated December 2025 city council meeting, in which public outrage was left unaddressed. The city, perhaps aware of the stigma Flock now carries, apparently tried to highlight that their focus was on the “smart streetlights” made by another company, while downplaying the fact that those streetlights run on Flock software.
“San Diego gets to hide behind a slight facade in that their contract is with Ubicquia,” the smart streetlight manufacturer, Paul says, “but the software layer is Flock. You can easily see Flock hardware on retail properties, looking at the same citizens, with zero oversight, and SDPD can claim they have clean hands.”
Weeks later, pieces of smashed Flock cameras littered the ground.
Across the country, in other words, municipal governments are overriding public will to make deals with a profiteering tech company to surveil their citizens and to collaborate with federal agencies like ICE. It might be taken as a sign of the times that in states and cities across the US, thousands of miles apart, those opposed to the technology are refusing to countenance what they view as violations of privacy and civil liberty, and are instead taking up vice grips and metal cutters. And in many cases, they’re getting hailed by their peers as heroes.
If you’ve heard stories of smashed Flock cameras or dismantled surveillance equipment in your neighborhood, please share—drop a link in the comments, or contact me on Signal or at briancmerchant@proton.me.
Thanks to Lilly Irani for the tip on the smashed Flock cams in San Diego.
In case you missed it, I shared my five takeaways on the most recent round of ultraheated AI discourse here:
The exchange was filmed and recorded on YouTube:
Police in Claremore, Oklahoma arrested a local man after he went slightly over his time giving public remarks during a city council meeting opposing a proposed data center. Darren Blanchard showed up at a Claremore City Council meeting on Tuesday to talk about public records and the data center. When he went over his allotted 3 minutes by a few seconds, the city had him arrested and charged with trespassing. The subject of the city council meeting was Project Mustang, a proposed data center that would be located within a local industrial park. In a mirror of fights playing out across the United States, developer Beale Infrastructure is attempting to build a large data center in a small town and the residents are concerned about water rights, spiking electricity bills, and noise.The public hearing was a chance for the city council to address some of these concerns and all residents were given a strict three minute time limit. The entire event was livestreamed and archive of it is on YouTube. Blanchard was warned, barely, to “respect the process” by one of the council members but was clearly finishing reading from papers he had brought to read from, was not belligerent, and went over time by just a few seconds. Anyone who has ever attended or watched a city council meeting anywhere will know that people go over their time at essentially any meeting that includes public comment.Blanchard arrived with documents in hand and questions about public records requests he’d made. During his remarks, people clapped and cheered and he asked that this not be counted against his three minutes. “There are major concerns about the public process in Claremore,” Blanchard said, referencing compliance documents and irregularities he’d uncovered in public records.
Blanchard was then arrested as the crowd jeered in disbelief. Also disconcerting was the way the local news framed the event, with a local anchor defending authorities by claiming he was “warned multiple times.” Seems like a pretty surefire way to make people hate data centers and the governments protecting them even more!
On Wednesday, I headed to Pershing Square in downtown Los Angeles, where dozens of gig workers and organizers with Rideshare Drivers United had assembled to deliver a petition to the California Labor Commission signed by thousands of workers, calling on the body to deliver a settlement on their behalf. Organizers made short speeches on the steps of the square while local radio and TV stations captured the moment. “
The Labor Commission is suing the gig companies on drivers’ behalf, alleging that Uber and Lyft stole billions of dollars worth of wages from drivers before Prop 22 was enacted in 2020. The commission is believed to be in negotiations with the gig companies right now that will determine a settlement.
I spoke with one driver, Karen, who had traveled from San Diego to join the demonstration, and asked her why she came. “It’s important we build driver power” she said. “Without driver power, we won’t get what we need, and we just want fairness.” She said she was hoping to claim at least $20,000 in stolen wages.
“We’re fighting for wages that were stolen for us from us and continue to be stolen from us every single day by these app companies from hell,” RDU organizer Nicole Moore told me. “So we’re marching in downtown L. A. to deliver 10,000 signatures of drivers demanding that the state fight hard for us, and don’t let these companies rip us off.”
According to Tesla’s own numbers, its new RoboTaxis in Austin are crashing at a rate 4 times higher than human drivers. The EV trade publication Electrek reports:
With 14 crashes now on the books, Tesla’s “Robotaxi” crash rate in Austin continues to deteriorate. Extrapolating from Tesla’s Q4 2025 earnings mileage data, which showed roughly 700,000 cumulative paid miles through November, the fleet likely reached around 800,000 miles by mid-January 2026. That works out to one crash every 57,000 miles. The irony is that Tesla’s own numbers condemn it. Tesla’s Vehicle Safety Report claims the average American driver experiences a minor collision every 229,000 miles and a major collision every 699,000 miles. By Tesla’s own benchmark, its “Robotaxi” fleet is crashing nearly 4 times more often than what the company says is normal for a regular human driver in a minor collision, and virtually every single one of these miles was driven with a trained safety monitor in the vehicle who could intervene at any moment, which means they likely prevented more crashes that Tesla’s system wouldn’t have avoided.Using NHTSA’s broader police-reported crash average of roughly one per 500,000 miles, the picture is even worse, Tesla’s fleet is crashing at approximately 8 times the human rate.
-“The Left Doesn’t Hate Technology, We Hate Being Exploited,” by Gita Jackson at Aftermath.
“Meta drops $65 million into super PACs to boost tech-friendly state candidates,” by Christine Mui in Politico.
-A great new report from climate researcher Ketan Joshi, “The AI Climate Hoax: Behind the Curtain of How Big Tech Greenwashes Impacts,” has been making headlines and is well worth a read. Perhaps we’ll dig deeper into it in a future issue.
-The LA Times reports that the Southern California air board rejected new pollution rules after an AI-generated flood of made-up comments. Here’s UCLA’s Evan George on how AI poses a unique threat to the civic process.
Okay okay, that’s it for this week. Thanks as always for reading. Hammers up.
...
Read the original on www.bloodinthemachine.com »
To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".
10HN is also available as an iOS App
If you visit 10HN only rarely, check out the the best articles from the past week.
If you like 10HN please leave feedback and share
Visit pancik.com for more.