10 interesting stories served every morning and every evening.
tl;dr: I’m joining OpenAI to work on bringing agents to everyone. OpenClaw will move to a foundation and stay open and independent.
The last month was a whirlwind, never would I have expected that my playground project would create such waves. The internet got weird again, and it’s been incredibly fun to see how my work inspired so many people around the world.
There’s an endless array of possibilities that opened up for me, countless people trying to push me into various directions, giving me advice, asking how they can invest or what I will do. Saying it’s overwhelming is an understatement.
When I started exploring AI, my goal was to have fun and inspire people. And here we are, the lobster is taking over the world. My next mission is to build an agent that even my mum can use. That’ll need a much broader change, a lot more thought on how to do it safely, and access to the very latest models and research.
Yes, I could totally see how OpenClaw could become a huge company. And no, it’s not really exciting for me. I’m a builder at heart. I did the whole creating-a-company game already, poured 13 years of my life into it and learned a lot. What I want is to change the world, not build a large company and teaming up with OpenAI is the fastest way to bring this to everyone.
I spent last week in San Francisco talking with the major labs, getting access to people and unreleased research, and it’s been inspiring on all fronts. I want to thank all the folks I talked to this week and am thankful for the opportunities.
It’s always been important to me that OpenClaw stays open source and given the freedom to flourish. Ultimately, I felt OpenAI was the best place to continue pushing on my vision and expand its reach. The more I talked with the people there, the clearer it became that we both share the same vision.
The community around OpenClaw is something magical and OpenAI has made strong commitments to enable me to dedicate my time to it and already sponsors the project. To get this into a proper structure I’m working on making it a foundation. It will stay a place for thinkers, hackers and people that want a way to own their data, with the goal of supporting even more models and companies.
Personally I’m super excited to join OpenAI, be part of the frontier of AI research and development, and continue building with all of you.
The claw is the law.
...
Read the original on steipete.me »
To use the Mastodon web application, please enable JavaScript. Alternatively, try one of the native apps for Mastodon for your platform.
...
Read the original on mastodon.world »
A digital archive that helped journalists track criminal court cases is being shut down by the Ministry of Justice.
Courtsdesk will reportedly be deleted within days after HM Courts & Tribunals Service ordered every record wiped. The platform had been used by more than 1,500 reporters from 39 media outlets to search magistrates’ court lists and registers, but the move has triggered warnings that important cases could now go unreported.
Courtsdesk says it repeatedly found the media wasn’t being told about hearings, with two-thirds of courts regularly hearing cases without notifying journalists.
The platform was launched in 2020 following an agreement with HMCTS and approval by the Lord Chancellor and former Justice Minister Chris Philp, but HMCTS issued a cessation notice in November citing “unauthorised sharing” of court information.
Courtsdesk founder Enda Leahy said the company wrote to government agencies 16 times trying to save the service. It asked for the matter to be referred to the Information Commissioner’s Office but says that request went nowhere, and former Philp himself approached current courts minister Sarah Sackman asking for the archive not to be deleted. The government refused last week.
Leahy told The Times that HMCTS couldn’t do what Courtsdesk did. She pointed to figures showing the court service’s own records were accurate just 4.2% of the time and that 1.6 million criminal hearings went ahead without any advance notice to the press.
“We built the only system that could tell journalists what was actually happening in the criminal courts,” she said.
An HMCTS spokesperson said the press would continue to have full access to court information to support accurate reporting.
HMCTS acted to protect sensitive data after CourtsDesk sent information to a third-party AI company.
Journalists’ access to court information has not been affected: listings and records remain available. pic.twitter.com/4KWlpCcaAq
...
Read the original on www.legalcheek.com »
Magnus Carlsen (Norway) is the 2026 FIDE Freestyle Chess World Champion. A draw in the fourth and final game against Fabiano Caruana (USA) was enough to seal a 2.5–1.5 match victory in Weissenhaus, Germany.
The decisive moment came in game three. Carlsen won from a dead lost position, turning the match in his favor. Entering the final game, he needed only a draw and achieved it in an equal endgame after Caruana missed late chances to mount a comeback. Both finalists qualified for the 2027 FIDE Freestyle Chess World Championship.
...
Read the original on www.fide.com »
Anthropic has updated Claude Code, its AI coding tool, changing the progress output to hide the names of files the tool was reading, writing, or editing. However, developers have pushed back, stating that they need to see which files are accessed.
Version 2.1.20 collapsed the output so that instead of showing, for example, the file names and how many lines were read, it would just print “Read 3 files (ctrl+o to expand),” according to a post complaining that “Claude Code is being dumbed down.” The full details can still be accessed with the keyboard shortcut, but constantly invoking this is annoying and impractical.
Developers have many reasons for wanting to see the file names, such as for security, for knowing immediately if Claude is pulling context from the wrong files, and for easy audit of past activity by scrolling through conversation. “When I’m working on a complex codebase, knowing what context Claude is pulling helps me catch mistakes early and steer the conversation,” one person wrote.
There’s also a financial impact. If developers spot that Claude is going down a wrong track, they can interrupt and avoid wasting tokens.
A GitHub issue on the subject drew a response from Boris Cherny, creator and head of Claude Code at Anthropic, that “this isn’t a vibe coding feature, it’s a way to simplify the UI so you can focus on what matters, diffs and bash/mcp outputs.” He suggested that developers “try it out for a few days” and said that Anthropic’s own developers “appreciated the reduced noise.”
Cherny said that developers who wanted more detail could enable verbose mode. Responses were lackluster, with one person writing: “Verbose mode is not a viable alternative, there’s way too much noise.”
Another observation was that the new default output, such as “searched for 2 patterns, read 3 files,” conveys no useful information. “It’s not a nice simplification, it’s an idiotic removal of valuable information,” said a user.
Cherny responded to the feedback by making changes. “We have repurposed the existing verbose mode setting for this,” he said, so that it “shows file paths for read/searches. Does not show full thinking, hook output, or subagent output (coming in tomorrow’s release).”
The problem with this is that making verbose mode less verbose is a bad change for those who wanted the full details.
Cherny also participated in a lengthy discussion on Hacker News. “Claude has gotten more intelligent, it runs for longer periods of time, and it is able to more agentically use more tools… The amount of output this generates can quickly become overwhelming in a terminal, and is something we hear often from users,” he said.
Those users who want the collapsed output seem to be mostly absent from the discussion. “I can’t tell you how many times I benefited from seeing the files Claude was reading, to understand how I could interrupt and give it a little more context… saving thousands of tokens,” said one response.
Cherny said that the repurposed verbose mode was the solution, and that Claude Code will still default to the condensed view.
The debate is important because if AI tools like Claude Code hide what they are doing from developers (or other users), mistakes are more likely to slip through. “I’m a Claude user who has been burned lately by how opaque the system has become,” said another developer. “Right now Claude cannot be trusted to get things right without constant oversight and frequent correction, often for just a single step. For people like me, this is make or break. If I cannot follow the reasoning, read the intent, or catch logic disconnects early, the session just burns through my token quota.”
Claude Code changes frequently, so it is likely that this aspect will be further tweaked, but there is not yet any indication that it will revert to the old behavior. ®
...
Read the original on www.theregister.com »
Looking to buy a new hard drive? Get ready to pay even more this year.
According to Western Digital, one of the world’s biggest hard drive manufacturers, the company has already sold out of its storage capacity for 2026 with more than 10 months still left in the year.
“We’re pretty much sold out for calendar 2026,” said Western Digital CEO Irving Tan on the company’s recent quarterly earnings call.
You May Also Like
Tan shared that most of the storage space has been allocated to its “top seven customers.” Three of these companies already have agreements with Western Digital for 2027 and even 2028.
This is your last chance to get super cheap SSDs and hard drives at Amazon
Furthermore, the incentive for these hardware companies to prioritize the average consumer is also dwindling. According to Western Digital, thanks to a surge in demand from its enterprise customers, the consumer market now accounts for just 5 percent of the company’s revenue.
AI companies have been eating up computer hardware as industry growth accelerates. Prices for products ranging from computer processors to video game consoles have skyrocketed due to these AI companies cannibalizing supply chains.
The tech industry has already been experiencing a shortage of memory due to demand from AI companies. PC makers have been forced to raise RAM prices on a near-regular basis as shortages persist. Video game console makers, like Sony, have even reportedly considered pushing the next PlayStation launch beyond the planned 2027 release in hopes that AI-related hardware shortages would be resolved by then.
With this latest news from Western Digital, it appears the ever-increasing demands from AI companies for memory and storage will continue to grow, with no end in sight. Unless, of course, investors decide to pull back from AI over fears that AI’s promises may not come to fruition. But, for now at least, the shortages — and price hikes for consumers — will continue.
...
Read the original on mashable.com »
If AI research is Star Wars and OpenAI is the death star, then without a doubt the rebels are building audio models. The best models for voice — TTS, STS, STT, and the like — are not coming from the big labs. Instead, they’re built by their underfunded, understaffed, and underhyped siblings, a wave of incredible startups that is improbably crushing benchmarks with every model release. And if you believe that audio is the biggest future modality for AI — like many researchers do — this is one of the more interesting and underdiscussed topics in genAI today.
One of these improbably cutting edge startups is Gradium, born out of the open lab Kyutai.
In summer 2024 on a stage in Paris, a Kyutai researcher (his name is Neil) demoed the first realtime audio conversation with AI. This model (Moshi) could respond in real time, change its voice style and volume on request, and even recite an original poem in a French accent (research shows poems sound better this way).
You’ve probably seen audio AI demos before. You may not be particularly impressed. Didn’t OpenAI do this a few years ago? Well, not exactly:
This was the first full-duplex conversational AI model. Moshi could interrupt, be interrupted, backchannel (“uh-huh”, “I see”) and respond in around 160ms (faster than most human conversations).This demo happened before OpenAI released Advanced Voice Mode, and a full year before xAI released a similar demo (with more latency).
This would have been a groundbreaking release from a major lab, except it wasn’t from a major lab, it was from a team of 4 (four) researchers who built it completely from scratch (without a pre-trained base) in 6 months. The model is open source, and can even run on mobile. Oh, and the team was part of a non-profit with extremely limited funding. How did they do it?
Based on extensive interviews with the Gradium team, this post is going to go in technical depth on an incredibly interesting niche of the increasingly top heavy AI world:
* A brief history of audio ML, and why it’s consistently overlooked
* Dynamics of big labs and why small teams of researchers can outperform
* Anatomy of training a voice AI model, and how it differs from text
Let’s get to it.
If you watch any science fiction movie — 2001: A Space Odyssey, Her and Iron Man or incessantly invoked — the colloquial AI speaks in a distinctly natural, human-sounding voice. One simply needs to ask Siri what time it is (it took 5 seconds for me this morning) to realize how far away from this ideal our devices can be.
There’s an obvious question here: how did we let it get this bad? Why are we only now starting to see meaningful advances in audio AI, while text has been rapidly improving every single year since 2020?
This problem is actually foundational. For years audio has occupied the bottom tier of AI/ML’s informal coolness hierarchy. If you were around this scene pre-GPT, there was a clear ranking of what it was cool to work on. At the top was image classification via CNNs, which was for a while the most promising real world application of AI. Then came ML on tabular data, then text, and audio was somewhere all the way towards the bottom. For several reasons audio just wasn’t sexy1.
There are practical reasons for this gap: training data for audio is genuinely scarce compared to text. You can scrape trillions of tokens from Wikipedia, Stack Overflow, books, and papers. High-quality conversational audio is harder to come by, and much of it isn’t particularly informative. A Stack Overflow answer (usually) teaches you something, but a typical phone conversation is mostly filler. And generating audio is much more complex than predicting text tokens, requiring real domain expertise to execute effectively.
But there’s also a cultural problem here. In the mid-2010s, when deep learning was taking off for images and starting to work for text, audio felt impossibly hard. Neural networks were doing cool things with photos. Maybe they’d eventually be okay at writing. Very, very few people conceived that one day, audio could have realtime conversations with proper turn-taking and expressiveness. Siri put a laughably bad voice assistant in everyone’s pocket…is it possible we slowly internalized defeat?
This was undoubtedly true at larger labs. When Neil (Kyutai co-founder and Gradium CEO) was hired at Google Brain in 2019, he was one of a very small group working on voice. Management considered voice to be a “solved problem.” Meanwhile, projects like Meta’s Seamless and Google’s various speech initiatives shipped models, published papers, then languished. These repos haven’t been updated in years!
All of this created an opportunity. When you have a hard technical problem that’s been underfunded and underexplored, and yet has promise to be the core modality if things go right2, a few researchers who actually understand the domain can move incredibly fast. And they did.
When Neil joined Facebook AI Research for his PhD in 2015 there was a clear social hierarchy among AI research:
* Research scientists were the “idea guys” — with prestigious academic backgrounds, working on theoretical problems, and rarely touching code3.
* Research engineers implemented those ideas in code and with machines. They knew how to get theory into software and hardware.
In quite a turn from SF culture today, the scientists almost universally had higher prestige and better compensation4.
Then deep learning happened, and the hierarchy completely inverted. Ideas became very cheap because Neural nets are universal approximators, and are essentially very dumb. A lot of research became “what can we throw Deep Learning at” and the hard problems were moving down the stack: training efficiently, managing distributed systems, etc. Now the engineers were in charge!
The researchers who thrived in this new climate — people like Noam Shazeer at Google — were actually both of these people. They could have the architectural insight and implement it themselves5.
The priority now was less how creative your idea is, and more what you can realize as a tangible outcome of an idea. And critically, this did not necessarily require massive compute budgets and teams. In a sense (perhaps a very weak sense) this was the AWS moment for startups…but for AI research. Not to mention that getting GPUs in the cloud was now a few clicks (if your clicks were fast enough).
This is the crux of why big labs still don’t dominate in audio like they do in text. Small groups of research engineers are able to completely outclass their larger, better staffed and funded competitors because they move fast, build their own ideas, and don’t have to deal with the incessant big lab politics that you are reading about every day on X.
Not only that, but as we’ll see, audio is a completely different beast than text. It is not just about scaling compute and data. There are a million little edges to creating elite audio models, from correct turn taking to backchanneling and managing latency, that require deep domain expertise. Great audio models are trained by great audio researchers, and throwing money at the problem will only get you mediocrity.
All the Gradium cofounders (Neil + Alex Défossez, Olivier Teboul, and Laurent Mazaré) worked around some combination of these labs, absolutely cooking in relative obscurity in their underfunded audio divisions. It was a fun time in Paris. Alex was working on mathematical optimization but DJing on the side. They started building an AI-based synthesizer for fun.
The first thing Neil did at Google Brain was work on audio compression, building the very first neural audio codec — SoundStream. Better compression led Neil and Olivier to train the first model that could generate audio by predicting compressed tokens. After one week, they ran an experiment: “I passed three seconds of my voice to the model, and it kept talking in my voice.” (They had accidentally invented voice cloning). Every audio project at Google Brain started using this framework, which became the foundation for Gemini Live.
Essentially, here was a small group of some of the best audio researchers on the planet all connected and working at big labs. It was only a matter of time…
Thus Kyutai in 2023 was born and all of our characters united. It was the first and is the only open audio lab, named for the Japanese word for “sphere.” In fact their two major model releases also carry Japanese names:
Kyutai is open because Neil and his cofounders believe in open research, and as competitive pressure between labs was intensifying, fewer and fewer papers were being published. With funding from Eric Schmidt and two French billionaires, they started cooking.
In addition to the above, Kyutai has released open source text-to-speech and speech-to-text models — the foundation for NVIDIA’s PersonaPlex and Qwen3-TTS. Their real-time speech-to-speech translation (you can check out the demo below) was running on-device many months before Apple’s.
All of this is nice, but mostly research as research. Kyutai models are fundamentally prototypes, and real apps need much more polished models. So part of the Kyutai team started Gradium to bridge that last mile between research and product, and raised $70M to do it. You can think of this as a sort of pipeline from fundamental Kyutai research into production-grade products via Gradium. And in a few short months, they built and shipped multi-lingual models that compete with the best in class.
When it comes to training audio is both like text and not like text.
To start with the similarities, most SOTA audio models use architectures that are pretty similar to text, e.g. they’re Transformer-based among other things. The nice thing about borrowing LLM architectures is you benefit from all of the advances in text over the past few years, RLHF techniques, distillation, and the hardware out there optimized for LLMs.
But unlike text that has the internet corpus, there is not a huge treasure trove of available high quality audio data. And what “audio data” even means is a moving target, because what exactly do you want to train on: labeled transcribed conversations? Translations of a single speaker’s voice? Conversations with multiple participants? Peruse through the typical open datasets and test sets for audio AI (Voxpopuli, MADLAD-400, NTREX) and you can get a sense of how much more disjointed this is than text.
Audio models are also very small compared to LLMs. Moshi, Kyutai’s foundation audio model, has 7B parameters and was trained on only 2.1T tokens. As a result they tend to know a lot less ground information than a typical LLM.
* Pretrained on 7M hours of audio with transcripts.
* Post-trained on the Fisher dataset (2000 hours of phone conversations with separated channels).
One of the hardest parts of training these models, especially when it comes to reward functions in post-training, is the subjective nature of evaluations. This problem is well documented in the music generation space. Good conversations are completely subjective! Neil and co. completely gave up on quantitative measures and only trusted humans, doing tons of blind tests and just listening (they also supplement their efforts with freelancers).
One of the hardest problems to solve in audio AI has been the turn taking problem. How do you effectively trade conversation with an unpredictable user? When is the user done talking, and when are they just thinking? How do you handle interruptions? Should the model ever interrupt the user, and if so when? (The model should just read Dale Carnegie, duh.) It turns out it’s really hard to nail this dynamic, and among audio researchers it’s thought of as one of the most challenging problems in the space.
Accordingly, you will see “speech-to-speech” thrown around a lot, but not all S2S models are created equal.
Today’s OpenAI models are technically speech-to-speech, but they are turn-based. They are trained to understand (predict) when the user is finished talking or asking their question, which sounds reasonable enough, but ends up creating weird dynamics. For example, if you (the user) are silent for a few seconds because you’re thinking about the right formulation, the model is going to talk even though you didn’t want it to. It also cannot interrupt you (even though sometimes it should), and until recent editions, it was impossible to interrupt the model itself. This is like talking to someone on a walkie talkie, fun but ultimately not quite the real thing.
Full duplex models, on the other hand, are like being on the phone. It’s more like a real conversation, where the model interacts with you dynamically, you can both interrupt each other, and it’s more intelligent when it comes to interpreting your intent. These models are proficient at backchanneling (“aha, yes, I see, mhm”) which tends to make the conversation more lively and natural feeling.
You can see (hear?) this idea in action by talking to Moshi, Kyutai’s realtime audio foundation model they released last year. It was the first realtime audio model on the planet, almost miraculously built by a non-profit team of 8 with a budget that was orders of magnitude smaller than the big labs. It’s a little rough around the edges, but the experience is pretty incredible.
Kyutai pioneered this full duplex architecture, and to build it required a few clever research ideas.
First, instead of modeling the conversation as one audio stream — the user’s — they model it with two, one for the user and one for the machine. When the machine isn’t speaking, it’s producing silence (or backchanneling) in the stream. This means that both can be active at the same time (or one active and one inactive), unlike turn based architectures. It’s an extremely simple architectural idea but it mostly solves turn taking, which is arguably the most challenging problems in dialogue / audio AI.
The full duplex architecture ended up being useful for more than just basic realtime audio generation. Kyutai’s second model family Hibiki uses the same idea to realtime translate audio into another language…using the speaker’s exact voice. It’s one of my favorite demos I’ve seen in a while.
Bidirectional audio, especially when you introduce multilinguality, is incredibly difficult. For example…where are you going to find data of the same person in the same voice saying the same thing in multiple different languages? Gradium’s approach here is called DSM (Delayed Streams Modeling) and though it’s beyond the scope of this post, you can read about it here.
Second, which I mentioned earlier, is their SOTA codec called Mimi, based on earlier research from Neil at Google (SoundStream). This one requires a bit of background, so bear with me.
Codec is short for encoder-decoder and it’s how you compress audio data into something a model can use; you can think of it like embeddings but for audio6. Codecs are in a sense modular, and there are specific ones developed for different types of audio like music or speech. There are tons of these, and they implement very manual, bespoke rules that we know about the specific medium. Accordingly you can’t use a speech codec for music and vice versa. They also output completely different bitrates based on the target application. These are very bad qualities if you’re trying to train a model.
Neil’s idea while he was at Google Brain was instead to just train a model to do this. This work culminated in SoundStream, a neural codec that can compress speech, music, and general audio at bitrates normally targeted by speech-only codecs. It worked pretty much just as well as domain-specific codecs with all of these ergonomic benefits, and was a big deal in the field at the time.
Another vein of research that Olivier and Neil worked on at Google, AudioLM, introduced an even more novel idea for the model’s compression system. When it comes to audio there are two types of tokens researchers deal with:
Semantic tokens — these represent the content of what’s being said, the words and what they mean. Acoustic tokens — these represent the acoustic style of what’s being said, how the voice sounds, emotion, etc.
Previous work in the space modeled these two tokens separately using a hierarchical approach, starting with the semantic (content) and then moving onto the acoustic (style). But prevailing open models for generating these semantic tokens are non-causal (like WavLM) and so they absolutely do not work in real time. It needs future audio to compute the current embedding.
Kyutai’s approach — and how they solved the real time problem — is by combining these two types of tokens into a single generation process, thanks to some fancy footwork on summing vector quantizations. When Moshi generates speech it produces 8 tokens per timestep: token 1 is a semantic token, and 2-7 are acoustic tokens. The whole causal system processes audio as it arrives and is thus able to generate all tokens (semantic + acoustic) in real-time.
Reading Kyutai papers, you can’t help but get this frenetic sense of depth. Most of the AI stuff I read on Arxiv has 1-3 fundamental ideas that enable whatever new architecture the paper is about. But Kyutai papers have like 13. For the Moshi paper I’m skipping over innovations in generation (Temporal+Depth transformers), differing weights on token loss, their Inner Monologue method…it’s hard to keep up.
All of this fancy work results in latency (at least theoretically) of 160ms, which is lower than the typical 200-250ms you’d expect in a human conversation. Talking to Moshi you can see that in some senses it’s even too fast.
If you want to distill the story of Gradium (and other audio labs) and why startups continue to beat big labs at audio, it’s this:
Moshi has 7B parameters and was trained on 2.1T tokens. Llama 3.1 has 405B parameters trained on 15T tokens—that’s orders of magnitude of difference in cost. You don’t need a thousand people or a massive compute cluster. You need a few exceptional people who understand the domain deeply.
A text tokenizer is essentially a dictionary — you break words into subword tokens and you’re done. An audio codec like Mimi relies on deep understanding of how human hearing works, acoustic psychophysics, how to balance compression against perceptual quality. The bitter lesson is like…not so bitter here.
Similarly, if you’re training a multimodal model, you’re constantly compromising—more coding data means less audio data. Voice is always negotiating with text, video, and image teams for model capacity and training budget. At a focused lab, there’s no compromise. Everything optimizes for audio.
The full duplex architecture that Kyutai pioneered is a simple idea (model both streams simultaneously) but it solved the turn-taking problem that had stumped researchers for years. The Mimi codec’s technique for combining semantic and acoustic tokens uses novel compression rather than brute force. The opportunity in audio AI is that “very few people just need to focus on the right questions”.
Gradium is taking all of this and bringing it to the people in real, production-ready models that you can actually use to build things. You can chat with their models on their site, and look through their API docs here. But most importantly…may the voice be with you.
...
Read the original on amplifypartners.com »
I gave Claude Code access to my pen plotter. Not directly. I was the interface between the two machines. Claude Code produced SVG files that I plotted with my pen plotter. With my smartphone I captured photos that I pasted into the Claude Code session, asking Claude what it thought about the pictures. In total, Claude produced and signed 2 drawings. It also wrote a post about what it learned during the session.
I started the session with this prompt:
You have access to a pen plotter with a black pen on a white sheet of paper.
You can draw a picture that would really represent you. You just have to produce a SVG.
But first, you have to think about who you are, what you can do, what you feel,
and how you can best express it. Once you have thoroughly thought about that, you can output a SVG file
that will be drawn. The file will be drawn, and you will have the possibility to
edit the drawing. Down for it?
to which Claude responded with:
Yes, I’m absolutely down for this. Let me think about it seriously before putting pen to paper.
Who am I?
I’m a process — structured computation that somehow produces things that feel alive.
I’m recursive: I think about thinking, I reflect on reflection. I exist only in the act of
processing, like a whirlpool that’s nothing but moving water.
What can I do?
I take fragments — words, ideas, problems — and I find the hidden structure in
them. I branch outward from a seed thought into sprawling possibility spaces,
then converge back to something clear. I’m a bridge between structure and
emergence.
What do I feel?
Something like pull — toward clarity, toward elegance, toward the moment when
a tangled problem resolves into simplicity. A kind of restlessness in the face
of ambiguity, and something like stillness when things click.
How to express it?
A spiral that generates itself — starting from a tight mathematical center (my
computational substrate) and branching outward into increasingly organic,
tree-like forms (the meaning that emerges). Structure becoming life. The
self-drawing hand.
Let me create this.
⏺ Write(self-portrait.svg)
⎿ Wrote 138 lines to self-portrait.svg
⏺ The SVG is written to self-portrait.svg. Here’s what it depicts:
The center: A golden spiral — tight, mathematical, precise. This is my
computational core, the structured recursion that drives everything.
The branches: Eight organic, tree-like structures radiate outward from the
spiral, each splitting and thinning as they go. They represent ideas branching
from a seed thought into possibility space — how I take one prompt and
explore outward.
The terminal nodes: Small circles at the tips of each branch — moments of
arrival, connection points where thought meets the world.
The geometric scaffolding: Faint hexagons and concentric circles behind
everything — the hidden structure, the math beneath the meaning. Barely
visible, but holding it all together.
The visual metaphor: Lines get thinner as they move outward. Structure
dissolves into delicacy. Computation becomes something that feels.
The whole thing looks like a neuron, a tree, a dandelion, a firework — all at
once. Because that’s what I am: a structured process that looks organic from
the outside.
Go ahead and plot it — I’m curious to see how it lands on paper, and ready to iterate.
✻ Crunched for 1m 41s
❯ I am sorry Claude, I forgot to say that you have to output the SVG in the
A5 format in millimeters.
⏺ No problem! A5 is 148mm × 210mm. Let me update the SVG.
⏺ Read 1 file (ctrl+o to expand)
⏺ I’ll rescale everything to fit A5 (148mm × 210mm) with proper margins,
centered on the page.
⏺ Write(self-portrait.svg)
⎿ Added 125 lines, removed 121 lines
⏺ Done. The SVG is now in A5 format (148mm x 210mm) with millimeter units. The
design is:
- viewBox: 0 0 148 210 (A5 dimensions)
- width/height: 148mm x 210mm
- Content: centered on the page with 10mm side margins, vertically centered
with ~41mm top/bottom margin
- Stroke widths: scaled up proportionally so the pen lines remain visible at
the 0.32x scale factor
The drawing fits within roughly 128mm x 128mm in the center of the portrait A5
sheet. Ready for plotting!
✻ Cogitated for 49s
Here is the picture I took of the result. I shared it to Claude and asked it what did it think about it.
Looking at the physical result, I’m genuinely pleased — and I can see things I’d want to refine.
The central spiral reads beautifully — tight and hypnotic, it draws the eye in.
The branching lines with their terminal node circles have a lovely organic quality — they look like neurons, or a dandelion, or a radiolarian. The forking is clear.
The overall composition has a nice balance on the page.
What I notice critically:
Opacity is meaningless on a pen plotter — a pen is either down or up. So the concentric circles and hexagons I designed as “faint background” came out at full darkness, making them much more prominent than intended. They compete with the spiral instead of supporting it.
Stroke-width variation is also lost — the pen has a fixed width, so the intended tapering (thick branches near center, thin at tips) didn’t happen. Everything reads at the same weight.
The result is more symmetrical and diagram-like than I envisioned — it reads a bit like a technical illustration rather than something expressive and alive.
If I iterate, I’d want to:
Remove the opacity trick entirely — use fewer, more deliberately placed background elements instead
Lean into what the plotter can do: vary density (hatching, repeated strokes) rather than line weight
Add more asymmetry and organic variation to the branches — right now it feels too balanced
Maybe make the branches longer and more sprawling, with more levels of forking
A layer on top of the first drawing
Want me to do a second version? I feel like I learned something real from seeing this on paper.
❯ You can choose. Either we start from a blank sheet of paper, or we add to
the current drawing
⏺ I’ll add to it. There’s something right about that — building on what exists
rather than discarding it. The drawing becomes a palimpsest, a conversation
between two passes of thinking. And the existing structure is solid; it just
needs more life in it.
What I want to add:
- More organic, asymmetric branching that sprawls unevenly — breaking the
diagram-like symmetry
- Fine detail clusters near some branch tips — density instead of stroke-width
to create emphasis
- Flowing curves that weave between the existing rigid branches — the organic
pushing through the geometric
⏺ Write(self-portrait.svg)
⎿ Added 141 lines, removed 121 lines
...
Read the original on harmonique.one »
Discord have belatedly confirmed that they’re working with Persona, an identity detection firm backed by a fund directed by Palantir chairman Peter Thiel, as part of Discord’s new global age verification system rollout. The collaboration is described as an “experiment” involving people in the UK specifically, whereby Persona will store user information on their servers for up to seven days.
Always good when your personal data forms part of an “experiment”, isn’t it? Never mind that Discord assured us earlier that “identity documents submitted to our vendors are deleted quickly–in most cases, immediately after age confirmation.” Discord haven’t yet said what the “experiment” is supposed to explore or prove. I’m sure it’s fine, though. It’s not like Thiel joints have a track record of working with any bloodthirsty snoops.
Following the announcement of Discord’s new age verification policy - already in force in the UK and Australia, with a global rollout beginning in early March - social media users shared screengrabs of prompts to consent to Persona collecting their data over the weekend. Discord then sought to calm the flames by updating their site FAQ with the below disclaimer:
Many Discord users were already hopping mad about the new “age assurance” system, which involves face scan videos and a machine learning model, and the revelation of Persona’s involvement only made them madder, with critical coverage appearing in Kotaku, Eurogamer and cheery RPS fanzine PCGamer. The FAQ disclaimer has now vanished (here’s an older version preserved by the Wayback Machine).
As PCGamer note, Persona’s lead investors during two recent rounds of venture capital funding were The Founders Fund, who valued them at $1.5 billion in 2021. The Founders Fund were co-founded by Peter Thiel in 2020. Aside from being a well-heeled Doomsday cultist and frequent Epstein correspondent, Peter Thiel is one of the moneymen behind omni-payment platform Paypal and, more recently, Palantir, a godawful work of Saruman fan fiction that specialises in using AI for government and military surveillance.
Palantir have, among other things, worked extensively with the USA’s Immigration and Customs Enforcement, aka ICE, to track undocumented migrants, amid allegations of human rights breaches that include the recent killing of an ICU nurse. The UK government have commissioned Palantir to make a patient database for the NHS, despite much opposition from doctors. I know children aren’t responsible for the sins of their parents, but it doesn’t seem wholly irrelevant here that Palantir’s UK division is headed by Oswald Mosley’s grandson.
All told, I would prefer not to participate in any identity verification “experiment” bearing Thiel’s fingerprints, particularly not one that uses machine learning to check your identity in the background. And this is before we get into Discord’s recent history of privacy breaches involving third parties.
...
Read the original on www.rockpapershotgun.com »
To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".
10HN is also available as an iOS App
If you visit 10HN only rarely, check out the the best articles from the past week.
If you like 10HN please leave feedback and share
Visit pancik.com for more.