10 interesting stories served every morning and every evening.
The URL shortener that makes your links look as suspicious as possible.
Normal links are too trustworthy. Make them creepy.
Please see the below statement in regards to the legal nastygrams:
This website is a joke. Redirect pages are in place to inform the user of what website they are being directed to. It is not designed to “weaken global cybersecurity hygiene” and does not facilitate phishing.
This website does not violate any known laws, policies, or rules, to the best of the author’s knowledge.
Valid concerns brought up in your letter have been addressed. Going forward, I would greatly appreciate if you use the support email in “Report Issue” to discuss concerns or problems with this service, rather than sending legal threats.
...
Read the original on creepylink.com »
Palantir is working on a tool for Immigration and Customs Enforcement (ICE) that populates a map with potential deportation targets, brings up a dossier on each person, and provides a “confidence score” on the person’s current address, 404 Media has learned. ICE is using it to find locations where lots of people it might detain could be based.
The findings, based on internal ICE material obtained by 404 Media, public procurement records, and recent sworn testimony from an ICE official, show the clearest link yet between the technological infrastructure Palantir is building for ICE and the agency’s activities on the ground. The tool receives peoples’ addresses from the Department of Health and Human Services (HHS) among a range of other sources, according to the material.
The news comes after Department of Homeland Security (DHS) head Kristi Noem said the agency is sending hundreds more federal agents to Minneapolis amid widespread protests against the agency. Last week ICE officer Jonathan Ross shot and killed 37 year old U. S. citizen Renee Nicole Good. During Operation Metro Surge, which DHS calls the “largest immigration operation ever,” immigration agents have surrounded rideshare drivers and used pepper spray on high school students.
...
Read the original on www.404media.co »
When CC Wei visited Cupertino last August, he had bad news for his largest client. Apple would need to acquiesce to the largest price rise in years, TSMC’s CEO told its executives.
Tim Cook and his team took the news on the chin. Wei had been telegraphing hikes in earnings calls over the past few quarters, and the Taiwanese chip maker’s rising gross margins were testament to its increasing pricing power.
That wasn’t the worst news, my sources tell me.
Apple, which once held a dominant position on TSMC’s customer list, now needs to fight for production capacity. With the continuing AI boom, and each GPU from clients like Nvidia and AMD taking up a larger footprint per wafer, the iPhone maker’s chip designs are no longer guaranteed a place among TSMC’s almost two dozen fabs.
What Wei probably didn’t tell Cook is that Apple may no longer be his largest client.
According to Culpium analysis and discussions with sources in the supply chain, Nvidia likely took top spot in at least one or two quarters of last year. “We don’t discuss that,” Chief Financial Officer Wendell Huang told Culpium Thursday when asked about the change in client rankings.
Final data will be unveiled in a few months when TSMC releases its annual report — which includes revenue from its top clients — but there’s every chance that Apple’s lead for the full year narrowed significantly and may have even fallen below Nvidia’s. If it didn’t happen in 2025, then it’s almost certain to do so in 2026, my sources tell me.
TSMC’s revenue climbed 36% last year to $122 billion, it reported Thursday. Nvidia’s sales for the fiscal year through January 2026 is set to climb 62% while Apple’s product revenue — which excludes services — is on track to grow just 3.6% for the 12-months to December 2025, according to Culpium estimates based on earnings reports and company guidance.
Apple’s role as the primary driver of TSMC revenue growth ended five years ago. In 2018 TSMC sales would have even fallen if not for incremental purchases by Apple that year. Now, the Cupertino company is posting low single-digit revenue growth while Nvidia is skyrocketing.
The reason for this change is two-fold, and pretty obvious: AI is driving massive demand for high-powered chips, while the smartphone boom has plateaued.
TSMC’s sales from high-performance computing, which includes AI chips, climbed 48% last year on top of 58% growth the year before. Smartphone revenue climbed just 11%, slower than 23% in the prior year. That trend will continue this year, and for the foreseeable future.
Revenue in 2026 will rise close to 30%, yet capital expenditure will climb around 32% to a record of somewhere between $52 billion and $56 billion, TSMC said Thursday. Longer term, growth will average 25% in the five years through 2029 yet the AI segment will climb an average of 55% or more over the same period, the company said. That’s higher than a prior forecast for a mid-40 percent figure.
The ultimate flex for TSMC came Thursday when it showed off not only record revenue and net income, but a gross margin approaching that of software makers and fabless chip designers. In the December quarter, that figure was an astounding 62.3%, 280 basis points higher than the prior period. If not for its overseas fabs (Arizona and Japan) gross margin would have been even higher.
There are two caveats that are important. First, while smartphone processors are the largest portion of chips bought by Apple, they’re not the only type. Processors for Macs come under HPC, while it also has a strong lineup of custom chips used in accessories which fall under digital consumer electronics. Second, Nvidia isn’t the only HPC client. AMD is a major buyer of capacity for its own GPUs while Amazon and Google are on the growing list of customers developing in-house AI chips.
Put another way, Apple’s chip catalog is broader and more varied, while Nvidia’s lineup is more concentrated around a huge number of wafers at, or near, leading-edge. It’s for these reasons that Apple will remain important for at least another decade.
In the near-term, however, TSMC’s technology roadmap coupled with broader industry trends favor Nvidia, AMD and their ilk, meaning Apple may need to keep fighting for capacity over the next year or two.
TSMC is already producing chips in volume at 2 nanometer (called N2), currently its most advanced node, with Apple a major buyer. But in the second half of this year it’s set to ramp up both a new variant called N2P as well as a new node called A16.
The company’s business model is a little quirky. Instead of repurposing an existing factory for new technology, TSMC just builds a new one. This ensures no interruption to output and allows it to squeeze the most out of old tools and processes. In general, this means any new capacity that TSMC builds is for a new node. As a result, it has numerous fabs still churning out chips on technology that’s a decade older or more.
In TSMC CEO CC Wei’s words A16, with Super Power Rail, is “best for HPC with complex signal routes.” SPR is TSMC’s version of backside power, a newer approach designed to separate a chip’s signal from its power supply. Intel is also developing this technology, and many believe it’ll be the key to the US company’s prospects at stealing foundry share from its Taiwan rival.
After that, TSMC has A14 which it expects to bring into volume production around 2028. Some call this the next full node after N2, labeling A16 as not a “full node.” In truth, all of these names are as much marketing terms as they are technology designators. Nevertheless, as SemiAnalysis recently wrote in a fabulous report on the TSMC-Apple relationship, the balance will shift back to Apple because A14 is designed “for both mobile and HPC from the start.”
More importantly, what Apple offers is stability. Nvidia has been a client for a lot longer than Apple, but broadly speaking it’s a bit niche. Right now that “niche” is the hottest product on the planet, but niche it is. Apple, on the other hand, has products being made in no fewer than a dozen TSMC fabs. Even if Nvidia did overtake Apple by purchases, the breadth of its manufacturing footprint at TSMC is nowhere near as large.
This distinction may not matter now, but it probably will at some point. The AI boom won’t last forever. The bubble may burst, or it may slowly deflate, but the growth trajectory will surely flatten and that means demand for leading-edge AI chips will fall.
Wei knows this, which is why he’s expanding both quickly yet cautiously. “I am also very nervous,” he said at the company’s investor conference on Thursday in Taipei. “If we didn’t do it carefully, it would be a big disaster for TSMC for sure.”
The chip giant has recently come under fire, including from noted analyst Benedict Evans, for being “unwilling/unable to expand capacity fast enough to meet Nvidia’s book.” I think this is wrong, and unfair.
“The risk of under-investing is significantly greater than the risk of over-investing,” Evans cited Google CEO Sundar Pichai as saying back in 2Q 2024, as if to make the point. TSMC and Alphabet, Google’s parent, have approximately the same gross margin. But their business models couldn’t be more different. Nvidia’s financials are also unlike TSMC’s. Their respective capex strategies need to reflect this risk.
Alphabet’s capital intensity, calculated as acquisitions of property, plant & equipment divided by revenue, was just 15% for full-year 2024. TSMC’s is more than double that at over 33%. More importantly, depreciation — which is where the cost of capex is reflected in earnings — was just 10% of Alphabet’s cost of revenue. For TSMC, this figure is more than four times higher at 45%.
At Nvidia, which is a tier-one buyer of TSMC’s output, the data is more stark. Capital intensity was just 2.5% for 2024, while depreciation was only 5.7% of the cost of revenue. As a fabless chipmaker, it can enjoy gross margins of over 70%. Its only real risk is holding excess inventory. Even then, it could have written off its entire inventory at the end of October and still maintain a gross margin approaching that of its chief supplier. What’s more, neither of these clients have anywhere near the customer-concentration risk of TSMC.
The complaint that TSMC could and should build faster ignores the fact that it’s the one left holding the baby if a downturn comes and demand falls. It takes two to three years to build a new fab, Wei explained, so the company must skate where the puck is going without thinking too much about where it’s been. “Even if we spend 52 to 56 billion this year, the contribution this year is none,“ Wei said Thursday. Its major cost, buying equipment, remains on the books no matter what revenue it brings in for the quarter.
For the best part of a decade, Apple was the one driving TSMC’s need to keep spending on new facilities. Today it’s Nvidia, and Jensen Huang is starting to wield more power than Tim Cook. But neither has to bother with the expensive business of actually manufacturing semiconductors, merely the hassle of begging CC Wei for wafers.
For such clients, the foundry’s capacity is a fixed cost that they needn’t worry about. Which is precisely why eight of the world’s ten largest companies turn to TSMC to make their chips, and in return the Taiwanese giant gets to reap the rewards during boom times like this.
...
Read the original on www.culpium.com »
Last year China installed more than half of all wind and solar added globally. In May alone, it added enough renewable energy to power Poland, installing solar panels at a rate of roughly 100 every second.
The massive buildout is happening across the country, from crowded eastern cities increasingly topped by rooftop solar panels to remote western deserts where colossal wind farms sprawl across the landscape.
“From the ground, it’s hard to grasp the scale of these power plants,” said Chinese photographer Weimin Chu. “But when you rise into the air, you can see the geometry, the rhythm — and their relationship with the mountains, the desert, the sea.”
Chu has spent three years capturing the shift underway using drones to photograph power plants from overhead. His work, which draws from the visual language of traditional Chinese ink paintings, was featured last year in an award-winning exhibition, presented by Greenpeace. A selection of those photos is reproduced here.
“I started out just shooting landscapes,” Chu said. “But when I traveled to places like Guizhou, Yunnan, and Qinghai in 2022, I kept seeing wind farms and solar power plants appear in my camera frame. I realized this is the story of our time — and almost no one is documenting it in a systematic way.”
...
Read the original on e360.yale.edu »
“Internal ICE material and testimony from an official obtained by 404 Media provides the clearest link yet between the technological infrastructure Palantir is building for ICE and the agency’s activities on the ground.”
“Internal ICE material and testimony from an official obtained by 404 Media provides the clearest link yet between the technological infrastructure Palantir is building for ICE and the agency’s activities on the ground.”
This is racial profiling on a grand scale:
It apparently looks a lot like Google Maps, but designed to show the richness of an area for “targets”, populated in part by density of immigrants. And then you can dig in:
The Nazis could only dream of having such a capability.
Imagine working for this company, on this product. Every day, you go into work, in what I assume is a beautiful office with pine furniture and a well-stocked kitchen, and you build software that will help to deport people using what you know are extrajudicial means without due process. You probably have OKRs. There are customer calls with ICE. Every two-week sprint, you take on tasks that help make this engine better.
What do you tell yourself? What do you tell your family?
Are you on board with this agenda, or do you tell yourself you need the job to pay rent? To get healthcare?
You receive stock as part of your pay package. It’s going up! You can use it to buy a home, or to build a comfortable retirement, or some combination of the two.
Your co-workers are values aligned and work hard. They’re talented and smart. Man, you might think to yourself, I love working with this team.
Or, you might think, man, I’ve got to find another job.
Either way, you’re proud of your product work. You’re happy to take the salary, the free lunches, the espresso. And regardless of how you feel about it, the thing you do every day is powering an armed force that is kidnapping people on the street and shooting civilians, that shot a mother in the face, that is targeting people to disappear using a beautiful, modern map interface.
...
Read the original on werd.io »
Yes, you, who are thinking about not hiring a technical writer this year or, worse, erased one or more technical writing positions last year because of AI. You, who are buying into the promise of docs entirely authored by LLMs without expert oversight or guidance. You, who unloaded the weight of docs on your devs’ shoulders, as if it was a trivial chore.
You are making a big mistake. But you can still undo the damage.
It’s been a complicated year, 2025. When even Andrej Karpathy, one of OpenAI’s founders, admits, in a fit of Oppenheimerian guilt, to feeling lost, you know that no one holds the key to the future. You flail and dance around these new totems made of words, which are neither intelligent nor conscious, pretending they can replace humans while, in fact, they’re little more than glorified tools.
You might think that the plausible taste of AI prose is all you need to give your products a voice. You paste code into a field and something that resembles docs comes out after a few minutes. Like a student eager to turn homework in, you might be tempted to content yourself with docs theatre, thinking that it’ll earn you a good grade. It won’t, because docs aren’t just artifacts.
You keep using that word. I do not think it means what you think it means
When you say “docs”, you’re careful to focus on the output, omitting the process. Perhaps you don’t know how docs are produced. You’ve forgotten, or perhaps never knew, that docs are product truth; that without them, software becomes unusable, because software is never done, is never obvious, and is never simple. Producing those docs requires tech writers.
Tech writers go to great lengths to get the information they need. They write so that your audience can understand. They hunger for clarity and meaning and impact. They power through weeks full of deadlines, chasing product news, because without their reporting, most products wouldn’t thrive; some wouldn’t even exist. Their docs aren’t a byproduct: they tie the product together.
An LLM can’t do all that, because it can’t feel the pain of your users. It can’t put itself into their shoes. It lacks the kind of empathy that’s behind great help content. It does not, in fact, have any empathy at all, because it cannot care. You need folks who will care, because content is a hairy beast that can only be tamed by agents made of flesh and capable of emotions: humans.
You can’t generate docs on autopilot. Let me tell you why.
First, AI-generated docs are not intelligent. They not only make up things in subtle ways: They lack vision. Even if you fed them millions of tokens, they couldn’t develop a docs strategy, decide what not to document, or structure content for reuse. And they fail to capture the tension, the caveats, the edge cases, the feeling of unfinishedness that only someone who cares can feel. Without that grounding, docs are hollow.
Second, liability doesn’t vanish just because AI wrote it. When docs cause harm through wrong instructions, someone will be held responsible. It won’t be the model. You can’t depose an LLM. You can’t fire it. You can’t point at it in court when a customer’s data evaporates because your GenAI runbook told them to run the wrong command. That someone will be you, or someone who reports to you.
Third, even your favorite AI must RTFM. All your Claude Skills, Cursor rules, all the semantic tagging that makes RAG work, is technical writing under a new name: context curation. You fired or didn’t hire the people who create high-quality context and then wondered why your AI tools produce slop. You can’t augment what isn’t there. The writers you let go were the supply chain for the intelligence you’re now betting on.
It’s not all bad news: Marvelous things can happen if you provide your writers with AI tools and training while you protect the quality of your content through an AI policy. I’ve described the ideal end state in My day as an augmented technical writer in 2030, a vision of the future where writers orchestrate, edit, and publish docs together with AI agents. This is already happening before our eyes.
Productivity gains are real when you understand that augmentation is better than replacing humans, a reality even AWS’ CEO, Matt Garman, acknowledged. Read how I’m using AI as a technical writer. I’m not alone: Follow Tom Johnson, CT Smith, and Sarah Deaton, and discover how tech writers are building tools through AI to better apply it to docs.
Develop an AI strategy for docs together with tech writers, and give them time and resources to experiment with AI. Tech writers are resourceful by nature: they’ve spent careers doing more with less, optimizing workflows, finding clever solutions to impossible quests. Give them the tools and a bit of runway, and they’ll figure out how to make AI work for the docs, not instead of them.
Reconsider the positions you did not open. Or the writers you let go. Reconsider the assumption that AI has solved a problem that, at its core, is deeply human and requires not only concatenating words, but also chasing subject-matter experts and understanding the subtleties of product motions, among many other things.
Technical writers aren’t a luxury. They are the people who translate what you’ve built into something others can use. Without them, you’re shipping a product that can’t speak for itself, or that lies. Your product needs to speak. AI can generate noise effectively and infinitely, but only a technical writer can create the signal.
Don’t choose the noise. Get them back. Get them onboard.
Thanks to Tiffany Hrabusa, Casey Smith, and Anna Urbiztondo for their reviews of early drafts and for their encouragement. Thanks to my partner, Valentina, for helping me improve this piece and for suggesting to wait a bit before hitting Publish. And a heartfelt thank you to the tech writing community and its wonderful human beings.
For a standalone version of this letter, use https://passo.uno/reconsider/.
...
Read the original on passo.uno »
...
Read the original on hackernoon.com »
Today Raspberry Pi launched their new $130 AI HAT+ 2 which includes a Hailo 10H and 8 GB of LPDDR4X RAM.
With that, the Hailo 10H is capable of running LLMs entirely standalone, freeing the Pi’s CPU and system RAM for other tasks. The chip runs at a maximum of 3W, with 40 TOPS of INT8 NPU inference performance in addition to the equivalent 26 TOPS INT4 machine vision performance on the earlier AI HAT with Hailo 8.
In practice, it’s not as amazing as it sounds.
You still can’t upgrade the RAM on the Pi, but at least this way if you do have a need for an AI coprocessor, you don’t have to eat up the Pi’s memory to run things on it.
And it’s a lot cheaper and more compact than running an eGPU on a Pi. In that sense, it’s more useful than the silly NPUs Microsoft forces into their ‘AI PCs’.
But it’s still a solution in search of a problem, in all but the most niche of use cases.
Besides feeling like I’m living in the world of the Turbo Encabulator every time I’m testing AI hardware, I find the marketing of these things to be very vague, and the applications not very broad.
For example, the Hailo 10H is advertised as being used for a Fujitsu demo of automatic shrink detection for a self-checkout.
That’s certainly not a worthless use case, but it’s not something I’ve ever needed to do. I have a feeling this board is meant more for development, for people who want to deploy the 10H in other devices, rather than as a total solution to problems individual Pi owners need to solve.
Especially when it comes to the headline feature: running inference, like with LLMs.
I also published a video with all the information in this blog post, but if you enjoy text more than video, scroll on past—it doesn’t offend me!
I ran everything on an 8 gig Pi 5, so I could get an apples-to-apples comparison, running the same models on the Pi’s CPU as I did on the AI HAT’s NPU.
They both have the same 8GB LPDDR4X RAM configuration, so ideally, they’d have similar performance.
I tested every model Hailo put out so far, and compared them, Pi 5 versus Hailo 10H:
The Hailo is only close, really, on Qwen2.5 Coder 1.5B.
It is slightly more efficient in most cases:
But looking more closely at power draw, we can see why the Hailo doesn’t keep up:
The Pi’s CPU is allowed to max out it’s power limits (10W on the SoC), which are a lot higher than the Hailo’s (3W).
So power holds it back, but the 8 gigs of RAM holds back the LLM use case (vs just running on the Pi’s CPU) the most. The Pi 5 can be bought in up to a 16 GB configuration. That’s as much as you get in decent consumer graphics cards.
Because of that, many quantized medium-size models target 10-12 GB of RAM usage (leaving space for context, which eats up another 2+ GB of RAM).
A couple weeks ago, ByteShape got Qwen3 30B A3B Instruct to fit on a 16GB Pi 5. Now this post isn’t about LLMs, but the short of it is they found a novel way to compress the model to fit in 10 GB of RAM.
A little bit of quality is lost, but like a JPEG, it’s still good enough to ace all the contrived tests (like building a TODO list app, or sorting a complex list) that the tiny models I ran on the Hailo 10H didn’t complete well (see the video earlier in this post for details).
To test the 30B model, I installed llama.cpp following this guide from my blog, and downloaded the compressed model.
I asked it to generate a single page TODO list app, and it’s still not a speed demon (this is a Pi CPU with LPDDR4x RAM we’re talking about), but after a little while, it gave me this:
It met all my requirements:
* I can type in as many items as I want
* I can drag them around to rearrange them
* I can check off items and they go to the bottom of the list…
It’s honestly crazy how many small tasks you can do even with free local models… even on a Pi. Natural Language Programming was just a dream back when I started my career.
Besides being angry Google, OpenAI, Anthropic and all these other companies are consuming all the world’s money and resources doing this stuff—not to mention destroying the careers of thousands of junior developers—it is kinda neat to see NLP work for very tightly defined examples.
But I don’t think this HAT is the best choice to run local, private LLMs (at least not as a primary goal).
What it is good for, is vision processing. But the original AI HAT was good for that too!
In my testing, Hailo’s hailo-rpi5-examples were not yet updated for this new HAT, and even if I specified the Hailo 10H manually, model files would not load, or I ran into errors once the board was detected.
But Raspberry Pi’s models ran, so I tested them with a Camera Module 3:
I pointed it over at my desk, and it was able to pick out things like my keyboard, my monitor (which it thought was a TV), my phone, and even the mouse tucked away in the back.
It all ran quite fast—and 10x faster than on the Pi’s CPU—but the problem is I can do the same thing with the original AI HAT ($110)—or the AI Camera ($70).
If you just need vision processing, I would stick with one of those.
The headline feature of the AI HAT+ 2 is the ability to run in a ‘mixed’ mode, where it can process machine vision (frames from a camera or video feed), while also running inference (like an LLM or text-to-speech).
Unfortunately, when I tried running two models simultaneously, I ran into segmentation faults or ‘device not ready’, and lacking any working examples from Hailo, I had to give up on getting that working in time for this post.
Just like the original AI HAT, there’s some growing pains.
It seems like with most hardware with “AI” in the name, it’s hardware-first, then software comes later—if it comes at all. At least with Raspberry Pi’s track record, the software does come, it’s just… often the solutions are only useful in tiny niche use cases.
8 GB of RAM is useful, but it’s not quite enough to give this HAT an advantage over just paying for the bigger 16GB Pi with more RAM, which will be more flexible and run models faster.
The main use case for this HAT might be in power-constrained applications where you need both vision processing and inferencing. But even there… it’s hard to say “yes, buy this thing”, because for just a few more watts, the Pi could achieve better performance for inference in tandem with the $70 AI Camera or the $110 AI HAT+ for the vision processing.
Outside of running tiny LLMs in less than 10 watts, maybe the idea is you use the AI HAT+ 2 as a development kit for designing devices using the 10H like self-checkout scanners (which might not even run on a Pi)? I’m not sure.
...
Read the original on www.jeffgeerling.com »
...
Read the original on kyutai.org »
Handy is a cross-platform desktop application built with Tauri (Rust + React/TypeScript) that provides simple, privacy-focused speech transcription. Press a shortcut, speak, and have your words appear in any text field—all without sending your voice to the cloud.
Handy was created to fill the gap for a truly open source, extensible speech-to-text tool. As stated on handy.computer:
* Free: Accessibility tooling belongs in everyone’s hands, not behind a paywall
* Open Source: Together we can build further. Extend Handy for yourself and contribute to something bigger
* Private: Your voice stays on your computer. Get transcriptions without sending audio to the cloud
* Simple: One tool, one job. Transcribe what you say and put it into a text box
Handy isn’t trying to be the best speech-to-text app—it’s trying to be the most forkable one.
Press a configurable keyboard shortcut to start/stop recording (or use push-to-talk mode)
Speak your words while the shortcut is active
Release and Handy processes your speech using Whisper
Get your transcribed text pasted directly into whatever app you’re using
The process is entirely local:
* Silence is filtered using VAD (Voice Activity Detection) with Silero
* Transcription uses your choice of models:
Whisper models (Small/Medium/Turbo/Large) with GPU acceleration when available
* Whisper models (Small/Medium/Turbo/Large) with GPU acceleration when available
Download the latest release from the releases page or the website
* Frontend: React + TypeScript with Tailwind CSS for the settings UI
* Core Libraries:
Handy includes an advanced debug mode for development and troubleshooting. Access it by pressing:
This project is actively being developed and has some known issues. We believe in transparency about the current state:
* Whisper models crash on certain system configurations (Windows and Linux)
* Does not affect all systems - issue is configuration-dependent
If you experience crashes and are a developer, please help to fix and provide debug logs!
* If you experience crashes and are a developer, please help to fix and provide debug logs!
* Requires wtype or dotool for text input to work correctly (see Linux Notes below for installation)
For reliable text input on Linux, install the appropriate tool for your display server:
* X11: Install xdotool for both direct typing and clipboard paste shortcuts
* Wayland: Install wtype (preferred) or dotool for text input to work correctly
* dotool setup: Requires adding your user to the input group: sudo usermod -aG input $USER (then log out and back in)
Without these tools, Handy falls back to enigo which may have limited compatibility, especially on Wayland.
The recording overlay is disabled by default on Linux (Overlay Position: None) because certain compositors treat it as the active window. When the overlay is visible it can steal focus, which prevents Handy from pasting back into the application that triggered transcription. If you enable the overlay anyway, be aware that clipboard-based pasting might fail or end up in the wrong window.
If you are having trouble with the app, running with the environment variable WEBKIT_DISABLE_DMABUF_RENDERER=1 may help
You can manage global shortcuts outside of Handy and still control the app via signals. Sending SIGUSR2 to the Handy process toggles recording on/off, which lets Wayland window managers or other hotkey daemons keep ownership of keybindings. Example (Sway):
bindsym $mod+o exec pkill -USR2 -n handy
pkill here simply delivers the signal—it does not terminate the process.
The following are recommendations for running Handy on your own machine. If you don’t meet the system requirements, the performance of the application may be degraded. We are working on improving the performance across all kinds of computers and hardware.
We’re actively working on several features and improvements. Contributions and feedback are welcome!
* Adding debug logging to a file to help diagnose issues
* A rewrite of global shortcut handling for MacOS, and potentially other OS’s too.
* Cleanup and refactor settings system which is becoming bloated and messy
If you’re behind a proxy, firewall, or in a restricted network environment where Handy cannot download models automatically, you can manually download and install them. The URLs are publicly accessible from any browser.
Navigate to the About section
Copy the “App Data Directory” path shown there, or use the shortcuts:
Inside your app data directory, create a models folder if it doesn’t already exist:
# macOS/Linux
mkdir -p ~/Library/Application\ Support/com.pais.handy/models
# Windows (PowerShell)
New-Item -ItemType Directory -Force -Path “$env:APPDATA\com.pais.handy\models”
Download the models you want from below
Simply place the .bin file directly into the models directory:
Place the extracted directory into the models folder
The directory must be named exactly as follows:
Final structure should look like:
* For Parakeet models, the extracted directory name must match exactly as shown above
* Do not rename the .bin files for Whisper models—use the exact filenames from the download URLs
* After placing the files, restart Handy to detect the new models
Your manually installed models should now appear as “Downloaded”
Select the model you want to use and test transcription
Test thoroughly on your target platform
Submit a pull request with clear description of changes
Join the discussion - reach out at contact@handy.computer
The goal is to create both a useful tool and a foundation for others to build upon—a well-patterned, simple codebase that serves the community.
* Whisper by OpenAI for the speech recognition model
“Your search for the right speech-to-text tool can end here—not because Handy is perfect, but because you can make it perfect for you.”
...
Read the original on github.com »
To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".
10HN is also available as an iOS App
If you visit 10HN only rarely, check out the the best articles from the past week.
If you like 10HN please leave feedback and share
Visit pancik.com for more.