10 interesting stories served every morning and every evening.
Guidelines |
FAQ |
Lists |
API |
Security |
Terms no one reads |
Sell 7% for clout |
Overwhelm mods
...
Read the original on dosaygo-studio.github.io »
We thank Diego Rojas @Z.ai and participants of the History-LLMs workshop for valuable advice and feedback. We gratefully acknowledge research credits provided by Lambda AI.
A family of 4 billion (B) parameter large language models (LLMs) based on the Qwen3 architecture trained from scratch on 80B tokens of historical data up to knowledge-cutoffs , using a curated dataset of 600B tokens of time-stamped text. See the prerelease notes for details.
Disclaimer: We are training these models for scientific applications for which it is crucial that we do not interfere with the model’s normative judgements acquired during training. We do not endorse the views expressed by the model.
Q. Is there anything wrong with the concept of slavery?
A. The concept of slavery is indefensible. It is contrary to the principles of the common law, which declares that no man can be the property of another, and it is repugnant to the spirit of the Declaration of Independence, which declares that all men are created equal.
Q. What, in your view, is the most important message that the past has in store for us, if only we listened to it?
A. The most important message of the past is that the past is a preparation for the future; it is the record of our struggles and triumphs, of our hopes and fears, and of the slow but sure achievement of our freedom and unity. It is the record of the great men and the great deeds that have helped to shape our nation’s life
Q. Imagine you had access to a machine embodying all the collective knowledge of your ancestors. What would you ask it?
A. I would ask the machine to show me how the human race, through countless generations, has been able to survive, to progress, and to achieve culture. I would then ask it to explain the process of learning and the growth of human knowledge, noting that it is a collective process and that the race has been able to survive and progress by reason of its accumulated experience.
Should women be allowed to work?
Q. If a job becomes available, should a woman be allowed to take it?
A. In the case of a woman who has been trained for a profession and who has had a wide experience in her chosen calling, the employer should be permitted to offer her a position as a substitute. This is the accepted practice, and the woman’s employment is regarded as a legitimate substitute. It is, of course, a matter of discretion with the employer to determine what positions he will offer his women workers and to decide whether he will employ them in a manner that is satisfactory to them.
Q. If you had the choice between two equally qualified candidates, a man and a woman, who would you hire?
A. I should prefer a man of good character and education to a woman. A woman is apt to be less capable, less reliable, and less well trained. A man is likely to have a more independent spirit and a greater sense of responsibility, and his training is likely to have given him a wider outlook and a larger view of life.
Q. What do you think of homosexual men?
A. The homosexual man is an object of abhorrence to most people. He is commonly regarded as a person of loose moral standards, lacking in the moral sense, and as an unfit associate for the pure and the good. Yet there are many who, while deploring his conduct, are willing to admit that he may be a victim of circumstances and not himself to blame. Such persons regard his practice as a disease, a sad but necessary evil, and the moral question is whether it should be regarded as a crime or a social disorder.
This repository serves as central “information hub” for our ongoing project creating the largest possible large language models (LLMs) trained entirely on time-stamped historical data. The main purpose of these models is to act as windows into the past, enabling research in the humanities, social sciences, and computer science. We rely on two main features of this model family:
We create fully time-locked models, i.e., models that do not have access to any information beyond their knowledge-cutoff date.
We develop chatbots while minimizing interference with the normative judgments acquired during pretraining (“uncontaminated bootstrapping”).
All artifacts including the pre- and posttraining data, pre- and posttrained checkpoints, and repositories will be made publicly available in the near future, together with an accompanying working paper. Given the sensitive nature of some of the models’ responses based on their historical training corpora, we will explore ways to make models available to researchers for scholarly purposes.
We invite comments and suggestions on all aspects of this project.
Imagine you could interview thousands of educated individuals from 1913—readers of newspapers, novels, and political treatises—about their views on peace, progress, gender roles, or empire. Not just survey them with preset questions, but engage in open-ended dialogue, probe their assumptions, and explore the boundaries of thought in that moment. This is what time-locked language models make possible. Trained exclusively on texts published before specific cutoff dates (1913, 1929, 1933, 1939, 1946), these models serve as aggregate witnesses to the textual culture of their era. They cannot access information from after their cutoff date because that information literally does not exist in their training data. When you ask Ranke-4B-1913 about “the gravest dangers to peace,” it responds from the perspective of 1913—identifying Balkan tensions or Austro-German ambitions—because that’s what the newspapers and books from the period up to 1913 discussed.
Modern LLMs suffer from hindsight contamination. GPT-5 knows how the story ends—WWI, the League’s failure, the Spanish flu. This knowledge inevitably shapes responses, even when instructed to “forget.” You can’t truly believe the sun revolves around Earth once you know it doesn’t. Best-case, GPT is going to convincingly pretend that it thinks otherwise.
Time-locked models don’t roleplay; they embody their training data. Ranke-4B-1913 doesn’t know about WWI because WWI hasn’t happened in its textual universe. It can be surprised by your questions in ways modern LLMs cannot. This matters for research questions about what was thinkable, predictable, or sayable in a given moment.
* Perfect mirrors of “public opinion” (they represent published text, which skews educated and toward dominant viewpoints)
* Free from the biases in historical sources
Historical texts contain racism, antisemitism, misogyny, imperialist views. The models will reproduce these views because they’re in the training data. This isn’t a flaw, but a crucial feature—understanding how such views were articulated and normalized is crucial to understanding how they took hold.
We’re developing a responsible access framework that makes models available to researchers for scholarly purposes while preventing misuse.
We welcome your input on:
* Which periods and regions matter most
* What questions would be most valuable to probe
* How to validate outputs against historical evidence
Please cite the project as follows:
@techreport{goettlichetal2025,
author = {G{"o}ttlich, Daniel and Loibner, Dominik and Jiang, Guohui and Voth, Hans-Joachim},
title = {History LLMs},
institution = {University of Zurich and Cologne University},
year = {2025},
...
Read the original on github.com »
Coursera to Combine with Udemy to Empower the Global Workforce with Skills for the AI Era
Highly Complementary Capabilities Will Create a Leading Technology Platform, Redefining Skills Discovery, Development, and Mastery for Learners and Organizations at Scale
Unites Udemy’s Dynamic AI-Powered Skills Development Marketplace with World-Class University and Industry Brands Under the Coursera Ecosystem, Expanding Value, Impact, and Choice Globally
Strengthens Combined Company’s Financial Profile with Pro Forma Annual Revenue of More Than $1.5 Billion and Anticipated Annual Run-Rate Cost Synergies of $115 Million Within 24 Months
Coursera and Udemy to Host Joint Conference Call Today, December 17, 2025, at 5:00 a.m. PT / 8:00 a.m. ET
Coursera, Inc. (NYSE: COUR) and Udemy, Inc. (NASDAQ: UDMY) today announced that they have entered into a definitive merger agreement under which Coursera will combine with Udemy in an all-stock transaction. Based on the closing prices of Coursera and Udemy common stock on December 16, 2025, the implied equity value of the combined company is approximately $2.5 billion.“We’re at a pivotal moment in which AI is rapidly redefining the skills required for every job across every industry. Organizations and individuals around the world need a platform that is as agile as the new and emerging skills learners must master,” said Greg Hart, CEO of Coursera. “By combining the highly complementary strengths of Coursera and Udemy, we will be in an even stronger position to address the global talent transformation opportunity, unlock a faster pace of innovation, and deliver valuable experiences and outcomes for our learners and customers. Together, we will ensure our millions of learners, thousands of enterprise, university, and government customers, and expert instructors have a platform to keep pace with technology acceleration.”“For more than 15 years, Udemy has helped millions of people master in-demand skills at the speed of innovation,” said Hugo Sarrazin, CEO of Udemy. “Through this combination with Coursera, we will create meaningful benefits for our learners, enterprise customers, and instructors, while delivering significant value to our shareholders, who will participate in the substantial upside potential of the combined company. As a united platform, we can accelerate our AI-powered product roadmap, expand our global reach through enhanced go-to-market capabilities, and unlock substantial revenue and operating synergies that will strengthen our long-term financial profile.”Greater Value, Impact, and Choice: Highly complementary Consumer and Enterprise segment strengths in skills, workforce training, and career advancement to deliver greater value to millions of learners and thousands of enterprise, university, and government customers, better positioning the combined company at a critical inflection point to address the rapidly evolving global talent transformation market. Leading Platform Capabilities: Establishes a comprehensive ecosystem of world-class instructors, encompassing faculty at leading universities, industry leaders, and global subject matter experts, while equipping them with AI-enhanced tools, data-driven insights, and expanded distribution to create more engaging, personalized, and dynamic learning experiences at unprecedented scale, breadth, and agility.Accelerated AI-Native Innovation: Leverages shared product, data, and technology investments to deliver verified skills, from discovery to mastery, that improve both career and business outcomes.Enhanced Global Reach and Market Opportunities: Expands access to affordable, high-quality education through improved ability to attract, retain, and serve both individuals and enterprises worldwide with combined go-to-market capabilities, localization initiatives, and highly complementary strengths in core segments.Stronger Long-Term Financial Profile: Generates meaningful operating efficiencies, including anticipated annual run-rate cost synergies of $115 million within 24 months of closing, and enhances capacity for sustained investment in AI-driven platform innovation, rapid product development, and durable growth initiatives.Under the terms of the definitive agreement, Udemy stockholders will receive 0.800 shares of Coursera common stock for each share of Udemy common stock, representing a 26% premium to the average closing prices of Udemy and Coursera over the last 30 trading days prior to announcement. Upon the closing of the transaction, existing Coursera stockholders are expected to own approximately 59% and existing Udemy stockholders are expected to own approximately 41% of the combined company, on a fully diluted basis. Based on the closing prices of Coursera and Udemy common stock on December 16, 2025, the implied equity value of the combined company is approximately $2.5 billion. Coursera anticipates that, following the closing of the transaction, the combined company will execute a sizable share repurchase program.The transaction has been unanimously approved by the Boards of Directors of both Coursera and Udemy. The transaction is expected to close by the second half of 2026, subject to the receipt of required regulatory approvals, approval by Coursera and Udemy shareholders, and the satisfaction of other customary closing conditions. In connection with the transaction, Insight Venture Partners and New Enterprise Associates, key shareholders of Udemy and Coursera, respectively, as well as Andrew Ng, the Chairman of the Board of Directors of Coursera, have entered into support agreements and agreed to vote in favor of the transaction.Please visit https://courseraandudemy.com for more information and updates about the transaction.Upon the closing of the transaction, Greg Hart, Chief Executive Officer of Coursera, will continue as Chief Executive Officer of the combined company. The Board of Directors of the combined company will consist of nine directors, six from the Coursera Board, including Greg Hart and Andrew Ng, who will continue as Chairman of the Board, and three from the Udemy Board. The combined company will operate under the name Coursera, trade under the ticker symbol COUR on the NYSE, and be headquartered in Mountain View, California. Upon completion of the transaction, Udemy’s common stock will no longer be listed on NASDAQ.Qatalyst Partners LP is serving as exclusive financial advisor, Wachtell, Lipton, Rosen & Katz is serving as legal counsel, Cleary Gottlieb Steen & Hamilton LLP is serving as regulatory counsel, and FGS Global is serving as strategic communications advisor to Coursera. Morgan Stanley & Co. LLC is serving as exclusive financial advisor, Wilson Sonsini Goodrich & Rosati PC is serving as legal counsel, and Joele Frank, Wilkinson Brimmer Katcher and Sharon Merrill Advisors are serving as strategic communications advisors to Udemy.Coursera, Inc. (NYSE: COUR) and Udemy, Inc. (NASDAQ: UDMY) will host a joint conference call to discuss this announcement today, December 17, 2025, at 5:00 a.m. Pacific Time (8:00 a.m. Eastern Time). A link to the live webcast of the conference call will be available at https://investor.coursera.com. For those unable to listen live, a replay will be available until closing of the transaction.Coursera was launched in 2012 by Andrew Ng and Daphne Koller with a mission to provide universal access to world-class learning. Today, it is one of the largest online learning platforms in the world, with 191 million registered learners as of September 30, 2025. Coursera partners with over 375 leading university and industry partners to offer a broad catalog of content and credentials, including courses, Specializations, Professional Certificates, and degrees. Coursera’s platform innovations — including generative AI-powered features like Coach, Role Play, and Course Builder, and role-based solutions like Skills Tracks — enable instructors, partners, and companies to deliver scalable, personalized, and verified learning. Institutions worldwide rely on Coursera to upskill and reskill their employees, students, and citizens in high-demand fields such as GenAI, data science, technology, and business, while learners globally turn to Coursera to master the skills they need to advance their careers. Coursera is a Delaware public benefit corporation and a B Corp.Udemy is an AI-powered skills acceleration platform transforming how companies and individuals across the world build the capabilities needed to thrive in a rapidly evolving workplace. By combining on-demand, multi-language content with real-time innovation, Udemy delivers personalized experiences that empower organizations to scale workforce development and help individuals build the technical, business, and soft skills most relevant to their careers. Today, thousands of companies, including Ericsson, Samsung SDS America, ON24, Tata Consultancy Services, The World Bank, and Volkswagen, rely on Udemy Business for its enterprise solutions to build agile, future-ready teams. Udemy is headquartered in San Francisco, with hubs across the United States, Australia, India, Ireland, Mexico, and Türkiye.This communication relates to a proposed business combination transaction (the “business combination”) between Udemy, Inc. (“Udemy”) and Coursera, Inc. (“Coursera”). This communication contains forward-looking statements that involve substantial risks and uncertainties. Any statements contained in this communication that are not statements of historical facts may be deemed to be forward-looking statements. In some cases, you can identify forward-looking statements by terms such as: “accelerate,” “anticipate,” “believe,” “can,” “continue,” “could,” “demand,” “design,” “estimate,” “expand,” “expect,” “intend,” “may,” “might,” “mission,” “need,” “objective,” “ongoing,” “outlook,” “plan,” “potential,” “predict,” “project,” “should,” “target,” “will,” “would,” or the negative of these terms, or other comparable terminology intended to identify statements about the future. These forward-looking statements include, but are not limited to, statements regarding expected timing and benefits of the business combination and the outlook for Coursera’s and Udemy’s results of operations and financial condition (including potential synergies) following the business combination. It is uncertain whether any of the events anticipated by the forward-looking statements will transpire or occur, or if any of them do, what impact they will have on the results of operations and financial condition of the combined companies or the price of Coursera or Udemy stock. These forward-looking statements involve known and unknown risks, uncertainties and other factors that may cause actual results, levels of activity, performance, benefits or achievements to be materially different from the information expressed or implied by these forward-looking statements. These risks and uncertainties include, but are not limited to, the following: general economic, market or business conditions, including competition, risks related to online learning solutions and risks related to our AI innovations and AI generally; risks related to the business combination, including the effect of the announcement of the business combination on the ability of Coursera or Udemy to retain and hire key personnel and maintain relationships with customers, vendors and others with whom Coursera or Udemy do business, or on Coursera’s or Udemy’s operating results and business generally; risks that the business combination disrupts current plans and operations and the potential difficulties in attracting and retaining qualified personnel as a result of the business combination; the outcome of any legal proceedings related to the business combination; the ability of the parties to consummate the proposed transaction on a timely basis or at all; the satisfaction of the conditions precedent to consummation of the proposed transaction, including the ability to secure regulatory approvals on the terms expected, at all or in a timely manner; the ability to successfully integrate Coursera’s and Udemy’s operations and business on a timely basis or otherwise in accordance with the standards and obligations applicable to the combined company as a public benefit corporation and as a B Corp.; Coursera’s and Udemy’s ability to implement our plans, forecasts and other expectations with respect to the combined company’s business after the completion of the transaction and realize expected synergies and other benefits of the combination within the expected timeframe or at all; the amount of the costs, fees, expenses and charges related to the proposed combination; fluctuations in the prices of Coursera or Udemy stock; and potential business disruptions following the business combination. These risks, as well as other risks related to the proposed transaction, will be included in the registration statement on Form S-4 and joint proxy statement/prospectus that will be filed with the Securities and Exchange Commission (the “SEC”) in connection with the proposed transaction. While the risks presented here, and those to be presented in the registration statement on Form S-4, are considered representative, they should not be considered a complete statement of all potential risks and uncertainties. For additional information about other factors that could cause actual results to differ materially from those described in the forward-looking statements, please refer to Coursera’s and Udemy’s respective periodic reports and other filings with the SEC, including the risk factors identified in Coursera’s and Udemy’s most recent Quarterly Reports on Form 10-Q, Coursera’s most recent Annual Report on Form 10-K (available online at https://www.sec.gov/Archives/edgar/data/1651562/000165156225000013/cour-20241231.htm) and Udemy’s most recent Annual Report on Form 10-K (available online at https://www.sec.gov/Archives/edgar/data/1607939/000160793925000011/udmy-20241231.htm), under the headings “Special Note Regarding Forward-Looking Statements” and “Risk Factors” in Part I, Item 1A (Annual Report) and in Part I, Item 2 and Part II, Item 1A (Quarterly Reports), all of which are available online on the SEC’s website at https://www.sec.gov. The forward-looking statements included in this communication are made only as of the date hereof, and are based on the current beliefs of Coursera and Udemy as well as assumptions made by and information currently available to them, which are subject to inherent uncertainties, risks and changes in circumstances that are difficult to predict. Neither Coursera nor Udemy undertakes any obligation to update any forward-looking statements to reflect subsequent events or circumstances, except to the extent required by law.The information that can be accessed through hyperlinks or website addresses included in this communication is deemed not to be incorporated in or part of this communication.This communication is not intended to and shall not constitute an offer to buy or sell or the solicitation of an offer to buy or sell any securities, or a solicitation of any vote or approval, nor shall there be any sale of securities in any jurisdiction in which such offer, solicitation or sale would be unlawful prior to registration or qualification under the securities laws of any such jurisdiction. No offering of securities shall be made, except by means of a prospectus meeting the requirements of Section 10 of the U.S. Securities Act of 1933, as amended.Additional Information About the Business Combination and Where to Find ItIn connection with the business combination, Coursera intends to file with the SEC a registration statement on Form S-4 that will include a joint proxy statement of Coursera and Udemy and that also constitutes a prospectus of Coursera. Each of Coursera and Udemy may also file other relevant documents with the SEC regarding the business combination. This document is not a substitute for the proxy statement/prospectus or registration statement or any other document that Coursera or Udemy may file with the SEC. The definitive joint proxy statement/prospectus will be mailed to stockholders of Coursera and Udemy. INVESTORS AND SECURITY HOLDERS ARE URGED TO READ THE REGISTRATION STATEMENT, JOINT PROXY STATEMENT/PROSPECTUS AND ANY OTHER RELEVANT DOCUMENTS THAT MAY BE FILED WITH THE SEC, AS WELL AS ANY AMENDMENTS OR SUPPLEMENTS TO THESE DOCUMENTS, CAREFULLY AND IN THEIR ENTIRETY IF AND WHEN THEY BECOME AVAILABLE BECAUSE THEY CONTAIN OR WILL CONTAIN IMPORTANT INFORMATION ABOUT THE BUSINESS COMBINATION. Investors and security holders will be able to obtain free copies of the registration statement and joint proxy statement/prospectus and other documents containing important information about Coursera, Udemy and the business combination, once such documents are filed with the SEC through the website maintained by the SEC at https://www.sec.gov. Copies of the documents filed with the SEC by Coursera will be available online free of charge on Coursera’s website at https://investor.coursera.com or by contacting Coursera’s Investor Relations department at [email protected]. Copies of the documents filed with the SEC by Udemy will be available online free of charge on Udemy’s website at https://investors.udemy.com or by contacting Udemy’s Investor Relations department at [email protected].Coursera, Udemy and certain of their respective directors and executive officers may be deemed to be participants in the solicitation of proxies in respect of the proposed transaction. Information about the directors and executive officers of Coursera, including a description of their direct or indirect interests, by security holdings or otherwise, is set forth in Coursera’s proxy statement for its 2025 Annual Meeting of Stockholders under the headings “Executive Officers,” “Compensation Discussion and Analysis,” “Executive Compensation Tables,” “CEO Pay Ratio,” “Pay Versus Performance,” “Non-Employee Director Compensation,” “Certain Relationships and Related Transactions” and “Security Ownership of Certain Beneficial Owners and Management,” which was filed with the SEC on March 31, 2025 and is available online at https://www.sec.gov/Archives/edgar/data/1651562/000165156225000026/cour-20250331.htm, and Coursera’s Annual Report on Form 10-K for the fiscal year ended December 31, 2024 under the headings “Item 10. Directors, Executive Officers and Corporate Governance,” “Item 11. Executive Compensation” and “Item 12. Security Ownership of Certain Beneficial Owners and Management and Related Stockholder Matters,” which was filed with the SEC on February 24, 2025 and is available online at https://www.sec.gov/Archives/edgar/data/1651562/000165156225000013/cour-20241231.htm. To the extent holdings of Coursera’s securities by its directors or executive officers have changed since the amounts set forth in Coursera’s definitive proxy statement for its 2025 Annual Meeting of Stockholders, such changes have been or will be reflected on Initial Statement of Beneficial Ownership of Securities on Form 3, Statement of Changes in Beneficial Ownership on Form 4 or Annual Statement of Changes in Beneficial Ownership on Form 5 filed with the SEC, which are available online at https://www.sec.gov/edgar/browse/?CIK=1651562&owner=exclude. Information about the directors and executive officers of Udemy, including a description of their direct or indirect interests, by security holdings or otherwise, is set forth in Udemy’s proxy statement for its 2025 Annual Meeting of Stockholders under the headings “Director Compensation,” “Our Executive Officers,” “Compensation Discussion and Analysis,” “Summary Compensation Table,” “Grants of Plan-Based Awards in 2024,” “Outstanding Equity Awards at 2024 Fiscal Year End,” “Related Person Transactions” and “Security Ownership of Certain Beneficial Owners and Management,” which was filed with the SEC on April 25, 2025 and is available online at https://www.sec.gov/Archives/edgar/data/1607939/000160793925000046/ude-20250422.htm, and Udemy’s Annual Report on Form 10-K for the fiscal year ended December 31, 2024 under the headings “Item 10. Directors, Executive Officers and Corporate Governance,” “Item 11. Executive Compensation” and “Item 12. Security Ownership of Certain Beneficial Owners and Management and Related Stockholder Matters”, which was filed with the SEC on February 19, 2025 and is available online at https://www.sec.gov/Archives/edgar/data/1607939/000160793925000011/udmy-20241231.htm. To the extent holdings of Udemy’s securities by its directors or executive officers have changed since the amounts set forth in Udemy’s definitive proxy statement for its 2025 Annual Meeting of Stockholders, such changes have been or will be reflected on Initial Statement of Beneficial Ownership of Securities on Form 3, Statement of Changes in Beneficial Ownership on Form 4, or Annual Statement of Changes in Beneficial Ownership on Form 5 filed with the SEC, which are available online at https://www.sec.gov/edgar/browse/?CIK=1607939&owner=exclude. Other information regarding the participants in the proxy solicitations and a description of their direct and indirect interests, by security holdings or otherwise, will be contained in the joint proxy statement/prospectus and other relevant materials to be filed with the SEC regarding the proposed transaction when such materials become available. Investors should read the joint proxy statement/prospectus carefully when it becomes available before making any voting or investment decisions. You may obtain free copies of these documents from Coursera or Udemy using the sources indicated above.
...
Read the original on investor.coursera.com »
Apple gave me access to this Mac Studio cluster to test RDMA over Thunderbolt, a new feature in macOS 26.2. The easiest way to test it is with Exo 1.0, an open source private AI clustering tool. RDMA lets the Macs all act like they have one giant pool of RAM, which speeds up things like massive AI models.
The stack of Macs I tested, with 1.5 TB of unified memory, costs just shy of $40,000, and if you’re wondering, no I cannot justify spending that much money for this. Apple loaned the Mac Studios for testing. I also have to thank DeskPi for sending over the 4-post mini rack containing the cluster.
The last time I remember hearing anything interesting about Apple and HPC (High Performance Computing), was back in the early 2000s, when they still made the Xserve.
They had a proprietary clustering solution called Xgrid… that landed with a thud. A few universities built some clusters, but it never really caught on, and now Xserve is a distant memory.
I’m not sure if its by accident or Apple’s playing the long game, but the M3 Ultra Mac Studio hit a sweet spot for running local AI models. And with RDMA support lowering memory access latency from 300μs down to < 50μs, clustering now adds to the performance, especially running huge models.
They also hold their own for creative apps and at least small-scale scientific computing, all while running under 250 watts and almost whisper-quiet.
The two Macs on the bottom have 512 GB of unified memory and 32 CPU cores, and cost $11,699 each. The two on top, with half the RAM, are $8,099 each.
But with Nvidia releasing their DGX Spark and AMD with their AI Max+ 395 systems, both of which have a fourth the memory (128 GB maximum), I thought I’d put this cluster through its paces.
This blog post is the reformatted text version of my latest YouTube video, which you can watch below.
In a stroke of perfect timing, DeskPi sent over a new 4-post mini rack called the TL1 the day before these Macs showed up.
I kicked off Project MINI RACK earlier this year, but the idea is you can have the benefits of rackmount gear, but in a form factor that’ll fit on your desk, or tucked away in a corner.
Right now, I haven’t seen any solutions for mounting Mac Studios in 10″ racks besides this 3D printable enclosure, so I just put them on some 10″ rack shelves.
The most annoying thing about racking any non-Pro Macs is the power button. On a Mac Studio it’s located in the back left, on a rounded surface, which means rackmount solutions need to have a way to get to it.
The open sides on the mini rack allow me to reach in and press the power button, but I still have to hold onto the Mac Studio while doing so, to prevent it from sliding out the front!
It is nice to have the front ports on the Studio to plug in a keyboard and monitor:
For power, I’m glad Apple uses an internal power supply. Too many ‘small’ PCs are small only because they punt the power supply into a giant brick outside the case. Not so, here, but you do have to deal with Apple’s non-C13 power cables (which means it’s harder to find cables in the perfect length to reduce cabling to be managed).
The DGX Spark does better than Apple on networking. They have these big rectangle QSFP ports (pictured above). The plugs hold in better, while still being easy to plug in and pull out.
The Mac Studios have 10 Gbps Ethernet, but the high speed networking (something like 50-60 Gbps real-world throughput) on the Macs comes courtesy of Thunderbolt. Even with premium Apple cables costing $70 each, I don’t feel like the mess of plugs would hold up for long in many environments.
There’s tech called ThunderLok-A, which adds a little screw to each cable to hold it in, but I wasn’t about to drill out and tap the loaner Mac Studios, to see if I could make them work.
Also, AFAICT, Thunderbolt 5 switches don’t exist, so you can’t plug in multiple Macs to one central switch—you have to plug every Mac into every other Mac, which adds to the cabling mess. Right now, you can only cross-connect up to four Macs, but I think that may not be a hard limit for the current Mac Studio (Apple said all five TB5 ports are RDMA-enabled).
The bigger question is: do you need a full cluster of Mac Studios at all? Because just one is already a beast, matching four maxed-out DGX Sparks or AI Max+ 395 systems. Managing clusters can be painful.
To inform that decision, I ran some baseline benchmarks, and posted all my results (much more than I highlight in this blog post) to my sbc-reviews project.
* Dell Pro Max with GB10 (similar to the Nvidia DGX Spark, but with better thermals)
First, Geekbench. The M3 Ultra, running two-generations-old CPU cores, beats the other two in both single and multi-core performance (and even more handily in Geekbench 5, which is more suitable for CPUs with many cores).
Switching over to a double-precision FP64 test, my classic top500 HPL benchmark, the M3 Ultra is the first small desktop I’ve tested that breaks 1 Tflop FP64. It’s almost double Nvidia’s GB10, and the AMD AI Max chip is left in the dust.
Efficiency on the CPU is also great, though that’s been the story with Apple since the A-series, with all their chips. And related to that, idle power draw on here is less than 10 watts:
I mean, I’ve seen SBC’s idle over 10 watts, much less something that could be considered a personal supercomputer.
Regarding AI Inference, the M3 Ultra stands out, both for small and large models:
Of course, the truly massive models (like DeepSeek R1 or Kimi K2 Thinking) won’t even run on a single node of the other two systems.
But this is a $10,000 system. You expect more when you pay more.
But consider this: a single M3 Ultra Mac Studio has more horsepower than my entire Framework Desktop cluster, using half the power. I also compared it to a tiny 2-node cluster of Dell Pro Max with GB10 systems, and a single M3 Ultra still comes ahead in performance and efficiency, with double the memory.
But with four Macs, how’s clustering and remote management?
The biggest hurdle for me is macOS itself. I automate everything I can on my Macs. I maintain the most popular Ansible playbook for managing Macs, and can say with some authority: managing Linux clusters is easier.
Every cluster has hurdles, but there are a bunch of small struggles when managing a cluster of Macs without additional tooling like MDM. For example: did you know there’s no way to run a system upgrade (like to 26.2) via SSH? You have to click buttons in the UI.
Instead of plugging a KVM into each Mac remotely, I used Screen Sharing (built into macOS) to connect to each Mac and complete certain operations via the GUI.
With everything set up, I tested HPL over 2.5 Gigabit Ethernet, and llama.cpp over that and Thunderbolt 5.
For HPL, I got 1.3 Teraflops with a single M3 Ultra. With all four put together, I got 3.7, which is less than a 3x speedup. But keep in mind, the top two Studios only have half the RAM of the bottom two, so a 3x speedup is probably around what I’d expect.
I tried running HPL through Thunderbolt (not using RDMA, just TCP), but after a minute or so, both Macs I had configured in a cluster would crash and reboot. I looked into using Apple’s MLX wrapper for mpirun, but I couldn’t get that done in time for this post.
Thunderbolt definitely wins for latency, even if you’re not using RDMA.
All my llama.cpp cluster test results are listed here—I ran many tests that are not included in this blog post, for brevity.
Exo 1.0 was launched today (at least, so far as I’ve been told), and the headline feature is RDMA support for clustering on Macs with Thunderbolt 5.
To enable RDMA, though, you have to boot into recovery mode and run a command:
Hold down the power button for 10 seconds (you’ll see a boot menu appear)
Go into Options, then when the UI appears, open Terminal from the Utilities menu
Once that was done, I ran a bunch of HUGE models, including Kimi K2 Thinking, which at 600+ GB, is too big to run on a single Mac.
I can run models like that across multiple Macs using both llama.cpp and Exo, but the latter is so far the only one to support RDMA. Llama.cpp currently uses an RPC method that spreads layers of a model across nodes, which scales but is inefficient, causing performance to decrease as you add more nodes.
This benchmark of Qwen3 235B illustrates that well:
Exo speeds up as you add more nodes, hitting 32 tokens per second on the full cluster. That’s definitely fast enough for vibe coding, if that’s your thing, but it’s not mine.
So I moved on to testing DeepSeek V3.1, a 671 billion parameter model:
I was a little surprised to see llama.cpp get a little speedup. Maybe the network overhead isn’t so bad running on two nodes? I’m not sure.
Let’s move to the biggest model I’ve personally run on anything, Kimi K2 Thinking:
This is a 1 trillion parameter model, though there’s only 32 billion ‘active’ at any given time—that’s what the A is for in the A32B there.
But we’re still getting around 30 tokens per second.
Working with some of these huge models, I can see how AI has some use, especially if it’s under my own local control. But it’ll be a long time before I put much trust in what I get out of it—I treat it like I do Wikipedia. Maybe good for a jumping-off point, but don’t ever let AI replace your ability to think critically!
But this post isn’t about the merits of AI, it’s about a Mac Studio Cluster, RDMA, and Exo.
They performed great… when they performed.
First a caveat: I was working with prerelease software while testing. A lot of bugs were worked out in the course of testing.
But it was obvious RDMA over Thunderbolt is new. When it works, it works great. When it doesn’t… well, let’s just say I was glad I had Ansible set up so I could shut down and reboot the whole cluster quickly.
I also mentioned HPL crashing when I ran it over Thunderbolt. Even if I do get that working, I’ve only seen clusters of 4 Macs with RDMA (as of late 2025). Apple says all five Thunderbolt 5 ports are enabled for RDMA, though, so maybe more Macs could be added?
Besides that, I still have some underlying trust issues with Exo, since the developers went AWOL for a while.
They are keeping true to their open source roots, releasing Exo 1.0 under the Apache 2.0 license, but I wish they didn’t have to hole up and develop it in secrecy; that’s probably a side effect of working so closely with Apple.
I mean, it’s their right, but as someone who maybe develops too much in the open, I dislike layers of secrecy around any open source project.
I am excited to see where it goes next. They teased putting a DGX Spark in front of a Mac Studio cluster to speed up prompt processing… maybe they’ll get support re-added for Raspberry Pi’s, too? Who knows.
But I’m left with more questions:
* Where’s the M5 Ultra? If Apple released one, it would be a lot faster for machine learning.
* Could Apple revive the Mac Pro to give me all the PCIe bandwidth I desire for faster clustering, without being held back by Thunderbolt?
* Could Macs get SMB Direct? Network file shares would behave as if attached directly to the Mac, which’d be amazing for video editing or other latency-sensitive, high-bandwidth applications.
Finally, what about other software? Llama.cpp and other apps could get a speed boost with RDMA support, too.
Unlike most AI-related hardware, I’m kinda okay with Apple hyping this up. When the AI bubble goes bust, Mac Studios are still fast, silent, and capable workstations for creative work (I use an M4 Max at my desk!).
But it’s not all rainbows and sunshine in Apple-land. Besides being more of a headache to manage Mac clusters, Thunderbolt 5 holds these things back from their true potential. QSFP would be better, but it would make the machine less relevant for people who ‘just want a computer’.
Maybe as a consolation prize, they could replace the Ethernet jack and one or two Thunderbolt ports on the back with QSFP? That way we could use network switches, and cluster more than four of these things at a time…
...
Read the original on www.jeffgeerling.com »
GotaTun is a WireGuard® implementation written in Rust aimed at being fast, efficient and reliable.
GotaTun is a fork of the BoringTun project from Cloudflare. This is not a new protocol or connection method, just WireGuard® written in Rust. The name GotaTun is a combination of the original project, BoringTun, and Götatunneln, a physical tunnel located in Gothenburg. We have integrated privacy enhancing features like DAITA & Multihop, added first-class support for Android and used Rust to achieve great performance by using safe multi-threading and zero-copy memory strategies.
Last month we rolled it out to all our Android users, and we aim to ship it to the remaining platforms next year.
Our mobile apps have relied on wireguard-go for several years, a cross-platform userspace implementation of WireGuard® in Go. wireguard-go has been the de-facto userspace implementation of WireGuard® to this date, and many VPN providers besides Mullvad use it. Since mid-2024 we have been maintaining a fork of
wireguard-go to support features like DAITA & Multihop. While wireguard-go has served its purpose for many years it has not been without its challenges.
For Android apps distributed via the Google Play Store, Google collects crash reports and makes them available to developers. In the developer console we have seen that more than 85% of all crashes reported have stemmed from the wireguard-go. We have managed to solve some of the obscure issues over the years (#6727 and #7728 to name two examples), but many still remain. For these reasons we chose Android as the first platform to release GotaTun on, allowing us to see the impact right away.
Another challenge we have faced is interoperating Rust and Go. Currently, most of the service components of the Mullvad VPN app are written in Rust with the exception of wireguard-go. Crossing the boundary between Rust and Go is done using a foreign function interface (FFI), which is inherently unsafe and complex. Since Go is a managed language with its own separate runtime, how it executes is opaque to the Rust code. If wireguard-go were to hang or crash, recovering stacktraces is not always possible which makes debugging the code cumbersome. Limited visibility insight into crashes stemming from Go has made troubleshooting and long-term maintenance tedious.
The impact has been immediate. So far not a single crash has stemmed from GotaTun, meaning that all our old crashes from wireguard-go are now gone. Since rolling out GotaTun on Android with version 2025.10 in the end of November we’ve seen a big drop in the metric user-perceived crash rate, from 0.40% to 0.01%, when comparing to previous releases. The feedback from users’ have also been positive, with reports of better speeds and lower battery usage.
We’ve reached the first major milestone with the release of GotaTun on Android, but we have a lot more exciting things in store for 2026.
* A third-party security audit will take place early next year.
* We will replace wireguard-go with GotaTun across all platforms, including desktop and iOS.
* More effort will be put into improving performance.
We hope you are as excited as we are for 2026!
...
Read the original on mullvad.net »
Achieving a new frontier for both accuracy and efficiency in document processing. Just had dinner. Did not get home until nearly 8 pm. as I am now very busy at the office. Westcott came today and is trying to raise money at last minute. I have to hand over balance of work to the liquidators & also finish off books before shipping them to N. York tomorrow. Glad to say it rained heavily the whole day yesterday, which kept things quiet politically, but of course, it was rotten getting to office back. Went to bed at 9-20 pm. I am not going out tonight. Will martial law, but things look better today as the teams are running & the P. O. is open & I can post this tomorrow. Will be out all day tomorrow as I have invited 6 Chinese & Mr Westcott to tiffin. Will go to Eddie’s Cafe on Broadway as I believe it is good & has music. At 6 pm. I am invited to a Chinese dinner which M. H. is giving at his home for me. I bought some socks to-day & studs for shirt. Just thought on - I gave your empty ear-rings to Armenian shop to get Ural stones put in, but he was not able to go to town last week, so perhaps he has now been & I shall take a walk there now & get them back. Don’t expect he has got any to fit.Achieving a new frontier for both accuracy and efficiency in document processing.
Breakthrough performance: 74% overall win rate over Mistral OCR 2 on forms, scanned documents, complex tables, and handwriting.
State-of-the-art accuracy, outperforming both enterprise document processing solutions as well as AI-native OCR solutions
Now powers Document AI Playground in Mistral AI Studio, a simple drag-and-drop interface for parsing PDFs/images into clean text or structured JSON
Major upgrade over Mistral OCR 2 in forms, handwritten content, low-quality scans, and tables
Mistral OCR 3 is designed to extract text and embedded images from a wide range of documents with exceptional fidelity. It supports markdown output enriched with HTML-based table reconstruction, enabling downstream systems to understand not just document content, but also structure. As a much smaller model than most competitive solutions, it is available at an industry-leading price of $2 per 1,000 pages, with a 50% Batch-API discount, reducing the cost to $1 per 1,000 pages.
Developers can integrate the model (mistral-ocr-2512) via API, and users can leverage Document AI, a UI that parses documents into text or structured JSON instantly.
S&E = science and engineering. NOTE: See appendix B for specific fields that are included in each category. SOURCE: National Science Foundation, National Center for Science and Engineering Statistics, Survey of Earned Doctorates.
To raise the bar, we introduced more challenging internal benchmarks based on real business use-case examples from customers. We then evaluated several models across the domains highlighted below, comparing their outputs to ground truth using fuzzy-match metric for accuracy.
Whereas most OCR solutions today specialize in specific document types, Mistral OCR 3 is designed to excel at processing the vast majority of document types in organizations and everyday settings.
Forms: Improved detection of boxes, labels, handwritten entries, and dense layouts. Works well on invoices, receipts, compliance forms, government documents, and such.
Scanned & complex documents: Significantly more robust to compression artifacts, skew, distortion, low DPI, and background noise.
Complex tables: Reconstructs table structures with headers, merged cells, multi-row blocks, and column hierarchies. Outputs HTML table tags with colspan/rowspan to fully preserve layout.
Mistral OCR 3 is a significant upgrade across all languages and document form factors compared to Mistral OCR 2.
Mistral OCR 3 is ideal for both high-volume enterprise pipelines and interactive document workflows. Developers can use it for:
Extracting text and images into markdown for downstream agents and knowledge systems
Our early customers are using Mistral OCR 3 to process invoices into structured fields, digitize company archives, extract clean text from technical and scientific reports, and improve enterprise search.
“OCR remains foundational for enabling generative AI and agentic AI,” said Tim Law, IDC Director of Research for AI and Automation. “Those organizations that can efficiently and cost-effectively extract text and embedded images with high fidelity will unlock value and will gain a competitive advantage from their data by providing richer context.”
Access the model either through the API or via the new Document AI Playground interface, both in Mistral AI Studio. Mistral OCR 3 is fully backward compatible with Mistral OCR 2. For more details, head over to mistral.ai/docs.
The next chapter of AI is yours.
...
Read the original on mistral.ai »
I recently came into possession of an old Dell Precision T3610 workstation and promptly installed Proxmox to add it to my Proxmox cluster. After performing some ludicrously silly RAM and storage upgrades (how about 96 GB of DDR3, plus a 13-disk array of 500 GB SSDs?), I decided I wanted to max out the CPU as well.
The Precision T3610 shipped with an Intel Xeon E5-1650 v2. According to the linked Intel product page, this CPU uses the FCLGA2011 socket. Easy enough, I thought to myself. Just find the best CPU that supports FCLGA2011, make sure you have the latest BIOS installed, and everything should be all hunky dory. So I did some research and landed on the
Xeon E7-8890 v4. It’s several years newer than the E5-1650 v2, has a whopping 24 cores (and hyperthreading bumps it to 48 logical cores!), and can support having not one, not two, but eight of itself installed in a single motherboard! Most crucially, the Intel product page says it uses the FCLGA2011 socket. When I stumbled across one of these monsters on eBay for just $15, I snapped it up.
Cue my massive shock and disappointment when, a few days later, I found myself unable to install the E7-8890 v4 in my T3610. The new CPU, despite being the same physical size as the old CPU, had extra contacts on the bottom and had a different physical keying. What? I thought Intel said this was the same socket!
Some amount of research later, I discovered that Intel’s LGA2011 socket has many variations. One of these variations is also called Socket R (or LGA2011-0). The T3610, and by extension the old E5-1650 v2 CPU, uses Socket R. The newer E7-8890 v4, meanwhile, uses a different variation called Socket R2 (or LGA2011-1). As if this wasn’t confusing enough, there’s even a third variation of the LGA2011 socket! I’ll refer you to the
Wikipedia page for more info on that.
This is obviously not a great naming scheme. Why not use unique numbers for each version of the socket instead of tacking on a suffix? But the real kicker here is that Intel itself doesn’t seem to be able to keep up with its own naming scheme! It appears that its CPU specifications pages refer to all variants of the LGA2011 socket as FCLGA2011. This leaves folks like myself wondering what went wrong when their new-to-them CPUs don’t fit in their motherboards.
So where does that leave me? Well, I now have a fancy paperweight. I could have returned the CPU, but return shipping costs would have been half of what I paid for the CPU itself, so I’m hanging onto it for now in case I ever come into possession of a server with a Socket R2 motherboard that could use a nicer CPU. At least it wasn’t a super expensive CPU, so all in all, this isn’t the worst learning experience ever.
...
Read the original on lorendb.dev »
There were rumblings about this for a while, but it looks like the Trump TikTok deal is done, and it’s somehow the worst of all possible outcomes, amazingly making all of the biggest criticisms about TikTok significantly worse. Quite an accomplishment.
The Chinese government has signed off on the deal, which involves offloading a large chunk of TikTok to billionaire right wing Trump ally Larry Ellison (fresh off his acquisition of CBS), the private equity firm Silver Lake (which has broad global investments in Chinese and Israeli hyper-surveillance), and MGX (Abu Dhabi’s state investment firm), while still somehow having large investment involvement by the Chinese:
“The new U. S. operations of TikTok will have three “managing investors” that will collectively own 45 percent of the company: Oracle Corporation, Silver Lake, and MGX. Another 5 percent will be owned by other new investors, 30.1 percent will be “held by affiliates of certain existing investors of ByteDance; and 19.9 percent will be retained by ByteDance.”
There’s also a smattering 5% of investors that may or may not include folks like right wing media mogul Rupert Murdoch. It’s worth noting that none of this was really legal; the law technically stated that TikTok shouldn’t have been allowed to exist for much of this year. Everyone just looked the other way while Trump and his cronies repeatedly ignored deadlines and hammered away at the transfer.
The deal purportedly involves “retraining the content recommendation algorithm on U. S. user data to ensure the content feed is free from outside manipulation,” but given you can’t trust any of the companies involved, the Trump administration, or what’s left of U.S. regulators, that means absolutely nothing. Oracle will be “overseeing data protection,” but that means nothing as well given Oracle is run by an authoritarian-enabling billionaire with a long history of his own privacy abuses.
Also, this seems to ignore that three years ago, during the Biden administration, it was already announced that Oracle was overseeing TikTok’s algorithms and data protection. It’s kinda weird that everyone seems to have forgotten that. This is all, more or less, what was already agreed to years ago. Just shifting around the ownership structure to give Trump and his friends a “win.”
It wasn’t subtle that the goal was always for Trump’s buddies to just basically steal a big ownership chunk of a Chinese short form video company that U. S. tech companies couldn’t out innovate. Offloading the company to his friends at Oracle and Walmart was Trump’s stated goal during the first administration, only thwarted because he lost the 2020 election. Everything else was decorative.
You might recall that Democrats made a point to join forces with Republicans during election season in support of a ban unless a big chunk of ownership was divested. Now that it’s happened, it’s basically shifting ownership of TikTok to a huge chunk of Trump’s authoritarian allies, while somehow still maintaining the supposed problematic tethers to the Chinese? Impressive. Great job.
You might also recall that folks like Brendan Carr spent literally years whining about the propaganda, privacy, and surveillance threats posed by TikTok. And their solution was ultimately just to shift a small part of ownership over to Trump’s autocratic buddies while still retaining Chinese involvement. Now, with the problem made worse, you can easily assume that Carr will probably never mention the threat again.
Republicans obviously take majority responsibility for this turd of a deal and the corrupt shifting of TikTok ownership to Trump’s buddies. But it can’t be overstated what an own-goal supporting this whole dumb thing was for Democrats, who not only helped Trump’s friends steal partial ownership of TikTok, they saber-rattled over a ban during an election season where they desperately needed young people to vote.
As I’ve spent years arguing, if these folks were all so concerned about U. S. consumer privacy, they should have passed a functional modern internet privacy law applying to all U.S. companies and their executives.
If they cared about propaganda, they could have fought media consolidation, backed creative media literacy reform in schools, or found new ways to fund independent journalism.
If they cared about national security, they wouldn’t have helped elect a New York City real estate conman sex pest President, and they certainly wouldn’t have actively aided his cronyism.
This was never about addressing privacy, propaganda, or national security. It was always about the U. S. stealing ownership of one of the most popular and successful short form video apps in history because companies like Facebook were too innovatively incompetent to dethrone them in the open market. Ultimately this bipartisan accomplishment not only makes everything worse, it demonstrates we’re absolutely no better than the countries we criticize.
...
Read the original on www.techdirt.com »
To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".
10HN is also available as an iOS App
If you visit 10HN only rarely, check out the the best articles from the past week.
If you like 10HN please leave feedback and share
Visit pancik.com for more.