10 interesting stories served every morning and every evening.
If you’d like to make a translation of this post, check out the instructions of how to do that here. Thank you so much for making this thing accessible to more folks!
Wanna be the first(ish) to see my new stuff when it comes out? Sign up for my very low-volume, maximum once-a-month
newsletter!
If you have disposable income to dispose of, you can support my free, open-source work on:
Vitamin D & Omega-3 have a larger effect on depression than antidepressants
⏱ This post is over years old.
Proceed at own risk.
The “effect size” of the best antidepressants on depression, vs placebo, is around 0.4. (On average; some people respond much better or much worse.) This is like going from a C to a C+.
In contrast: the effect size of 1500 mg/day of “≥60% EPA” Omega-3 supplements is a bit higher, around 0.6. This is like going from a C to a B–. (With uncertainty; at worst, Omega-3′s “only” on par with antidepressants.)
But, much better: the effect size of 4000 IU/day of Vitamin D is twice as high as antidepressants’, around 1.0. This is like going from a C to an B! (With uncertainty; at worst, Vitamin D’s “only” on par with antidepressants.) This works even for people who don’t have a Vitamin D insufficiency — but around half of American adults do.
Even if you’re already taking Vitamin D & Omega-3, double check your dose: it may still not be enough! The official recommendations are all too low, and recent research suggests even the official maximum safe dose for Vitamin D is too low.
I know the “yay supplements” genre of writing is full of sloppy research & grifters, and you should be skeptical of my claim of easy wins, of “$100 bills laying on the sidewalk”. But there is good science among the trash, and policy is often decades behind science in any field, not just health.
So, Vitamin D & Omega-3: possibly high reward, for low risk. That’s a positive “expected value” bet! These supplements are safe, cheap, over-the-counter, and have positive side-effects (on Covid & cognition). As always, “ask your doctor”, show them the peer-reviewed papers cited in this post.
Unless you have specific reasons to not take Vitamin D & Omega-3 — kidney stones, blood thinners, etc — please try them, for at least a month! They could save your mental health. Maybe even your life.
In Alicetown, the average person has 4 younger cousins.
In Bobtown, the average person has 3 younger cousins.
Alright, not so surprising. You may not even notice a difference.
In Alicetown, the average person has 4 limbs.
In Bobtown, the average person has 3 limbs.
It’s the same absolute difference (4 vs 3) and relative difference (3/4). So what makes limbs more surprising than cousins? Well, partly it’s more dramatic & visible, but also because: we expect high variation in the number of someone’s younger cousins, but not their number of limbs.
This is why scientists calculate an “effect size” or “standardized mean difference” (“mean” = average). We take the difference between two groups, then divide by the total amount of variation, to account for how surprising a difference is.
Unfortunately for laypeople, the effect size is usually just reported as a number, like “+0.74” for spacing out your studying vs cramming, or “–0.776″ for sleep deprivation on attention.
But what’s that mean? How can we make these numbers intuitive?
Well, a common way for data to be is a bell-shaped curve (also called a “normal distribution”). And most of us are, alas, well-acquainted with the bell curve in school grades. (“grading on a curve”)
So: school grades give us a useful way to think about standardized effect sizes! We can now convert that number into an actual letter grade:
For example: spacing out your studying, relative to cramming, will on average lift your test scores from a C to a B–. (effect size = +0.74) And short-term sleep deprivation, relative to healthy sleep, will on average tank your ability to pay attention from a C to a D+. (effect size: –0.776)
But it’s not limited to just grades & academic performance. Effect sizes can also help us understand any kind of difference between groups, in observation or in experiments!
Let’s use our school grade analogy, to interpret effect sizes on mental health:
What’s an “F in mental health”? By definition of a bell curve, ~2.3% of people are below –2 sigma (an “F”). (See: this bell curve calculator.) In Canada, ~2.6% of people had suicidal ideation in 2022, while in the US, it was ~4.9% in 2019. So, it’s not too far off to say: “F in mental health = literally suicidal”. (Also, reminder that ~4% is 1-in-25 people. You likely know someone, or are someone, who will feel suicidal this year. Please reach out to your friends & loved ones!)
What’s a “D in mental health”? ~16% of people are below –1 sigma (a “D”) on a bell curve. The Keyes 2002 study estimated that ~14.1% of adults meet the DSM-III criteria for a major depressive episode. So, D = Depressed.
What’s an average “C in mental health”? ~68% of people are within a sigma of average (a “C”) on a bell curve. Same above study found that 56.6 percent had moderate mental health. They were neither “languishing” nor “flourishing”. I guess C = Could Be Worse.
What’s a “B in mental health”? ~16% of people are above +1 sigma (a “B”) on a bell curve. Same above study found that 17.2% of adults are “flourishing”. Good for them! B = Flourishing, life is good.
What’s an “A in mental health”? I don’t know who these freaks are. I actually could not find any scientific studies on “the +2 sigma in well-being”. In contrast, there’s lots of research on suicidal ideation, the –2 sigma in well-being. In the absence of any actual data, I’ll just say: A = AWESOME
So, if an intervention is found to have an effect size of +1.0, that’s like going up a letter grade. If something’s found to have an effect size of -2.0, that’s like going down two letter grades. And so on.
Okay, so how do we get peoples’ “mental health grades” up?
Let’s look at antidepressants, Omega-3, and Vitamin D, in turn:
The good news is they work. The bad news is they don’t work as well as you’d think they may work.
Cipriani et al 2018 is a meta-analysis: a study that collects & combines lots of previous studies (that pass some basic criteria, to minimize a garbage-in-garbage-out situation). While meta-analyses aren’t perfect, it’s usually better for “science communicators” like me to cite meta-analyses over individual studies, to reduce the chance I’m cherry-picking.
Anyway: this meta-analysis analyzes 522 trials with 116,477 participants. All 21 antidepressants they studied were better than placebo (a pill that contains no active medicine). The most effective antidepressant, Amitriptyline, had an “Odds Ratio” of 2.13, which converts to an effect size of 0.417, which is “small-medium” according to Cohen’s recommendations. Or, by our school-letter-grade comparison: the best antidepressant would take your mental health grade from an F to F+, or C to C+.
From Figure 3 of that paper, you can see that Amitriptyline has the highest estimated effect size, while the side effects are no worse than placebo:
But hang on, only F to F+ on average? How does that square with people’s personal experience that antidepressants have been lifesaving?
Well, first: the average person has around 1 testicle.
The punchline being ~50% of people have 2 testicles while ~50% of people have 0 testicles, hence the average is “around 1”. Likewise, the average effect for the best antidepressant is 0.4 — but some people respond much better than that… and some respond much worse. (e.g. different kinds of antidepressant, different kinds of depression, different kinds of people, etc. Note that this caveat also applies to the Vitamin D & Omega-3 studies, and all medical studies.)
And, second: the belief that things will get better is a powerful thing. Unfortunately, the power of hope gets a bad name in medicine: “placebo”.
When you take any medicine, you don’t just get (effect of medicine). You get (effect of medicine + effect of placebo + effect of time).
The effect of placebo + time: probably around 0.9.
The effect of placebo alone: Amazingly, despite researchers having used placebos for decades, it’s only recently that we started testing “open-label” placebos: placebos where we just tell the patient it’s a placebo. We then compare “getting placebo” to “getting nothing”. The effect size of open placebo, on stuff ranging from pain to depression, is around 0.43. (Spille et al 2023)
The effect of time alone: Using the above two numbers, I’d guesstimate: 0.9 - 0.43 = 0.47. “Time” includes both natural healing, and “regression to the mean”.
So, the individual effect of medication, psychological placebo, and time, are all around +0.4 each. And combined, they give you +1.20, or going from F to D+ or C to B+. That’s why many people report antidepressants being lifesaving! (Again, on average; some people react much worse.)
“Wait, the improvement from antidepressants is mostly placebo + time?” Yes, and this is widely known in psychiatry. I mean, they’re not yelling it from the rooftops, but it ain’t no secret. Decades ago, the infamous Kirsch & Sapirstein 1998 estimated that the improvement from antidepressants is ~75% placebo + time. Even the most critical response to Kirsch’s work, Fountoulakis & Möller 2011, still finds it’s mostly placebo + time.
But again, I think “placebo” is too dismissive a word for the power of hope. Hope isn’t magic, but it’s something, and measurably so: around +0.4. I assert: the placebo effect isn’t a bug, it’s a feature! It proves the connection between mental state & physical health.
But anyway, for the rest of this article, I’ll only be reporting effect sizes versus placebo + time. Just remember that the power of hope gives you an extra +0.4 (like C to C+) for all interventions.
Keep getting confused on which fat is what? Me too. So, here’s a crash course on various fats:
Fatty acids are chains of carbons & hydrogens + two oxygens. They say “OOH” at one end, and “HHH” at the other end:
A saturated fatty acid is one where all the carbons’ free spots are filled up with hydrogens. (Hence, “saturated”) This makes the molecule stick straight out. This is why long saturated fatty acids — like those found in butter — tend to be solid at room temperature.
In contrast, unsaturated fatty acids have at least one hydrogen missing. This causes them to have a double-bond “kink” in the molecule. This makes them not stick out, which is why unsaturated fats tend to be liquid at room temperature. Mono-unsaturated fatty acids (MUFAs) — like in olive oil — only have one kink. Poly-unsaturated fatty acids (PUFAs) — like in fatty fish — have two or more kinks. Let’s be mature adults about this, please.
For completeness: trans fats are unsaturated fats whose “kink” is twisted around, causing them to go straight. That is the worst sentence I’ve written all month. The twisted kink is caused by the hydrogens being on opposite sides, hence “trans”. (And yes, if they’re on the same side it’s “cis”. Latin was a mistake.) The molecule being straight is why trans fats — which margarine used to be full of — are solid at room temperature, despite being an unsaturated fat.
It’s neat whenever you can trace the history of something right down to its atoms! Margarine was first invented because it’s cheaper, and is spreadable straight from the fridge, unlike butter. Margarine (used to be) made by taking unsaturated vegetable oils, which were cheaper than animal fats, then pumping a bunch of hydrogens into it (hence, “hydrogenated oils”). If you completely hydrogenate an oil, it becomes a saturated fat. But they only partially hydrogenated those oils, leading to trans fats, which were cheaper & a spreadable semi-solid at fridge temperature.
In the 1970s & 80s, the US Food & Drug Administration concluded that trans fats were not harmful to humans, and nutritionists promoted margarine over butter, because butter had “unhealthy” saturated fats. But in the early 1990s, scientists realized that trans fats were even worse for you than saturated fats. Only in the 2010′s, did most Western countries start officially banning trans fats. Reminder: policy is often decades behind science.
I need to stop going on infodump tangents. Anyway, Omega-3 is any fatty acid with its first kink at the 3rd carbon from the Omega end (“HHH”), though it can have more kinks later down the chain. (And yes, Omega-6 has its first kink at the 6th carbon, and Omega-9 has its first kink at the 9th carbon. There’s nothing physically preventing Omega-4 or Omega-5′s from existing, but due to some quirk of evolution, Omega-3, -6, and -9 are the ones biological life uses most. As far as I can tell, there’s no specific reason they’re all multiples of 3. Probably just a coincidence. There is a less common Omega-7.)
Finally, there’s three main types of Omega-3: EPA (Eicosapentaenoic Acid), DHA (Docosahexaenoic Acid), and ALA (Alpha-Linolenic Acid). ALA is mostly found in plants like chia seeds & walnuts, while EPA & DHA mostly come from seafood, though there are algae-based vegan sources.
EPA & DHA are the focus of this section. For bio-mechanical reasons I don’t understand but I assume someone else does: EPA is the one associated with anti-inflammation, better brain health, and less depression… while DHA isn’t. (But DHA is still needed for other stuff, like your neurons’ cell walls, so don’t cut them out completely!)
All the above info in a Venn (technically Euler) diagram:
Okay, enough yap. Time for the actual data:
Sublette et al 2011 is an older meta-analysis (15 trials with 916 participants). It’s the only meta-analysis I could find that estimates the actual “dose-response” curve, which shows: how much effect, for how much treatment.
Why is dose-response important? Because one problem with many meta-analyses is they’ll do something like: “Study 1 gave patients 1 gram of medicine and saw a +1 improvement in disease, Study 2 gave 10 grams and saw +4 improvement, Study 3 gave 100 grams and saw negative –5 improvement… the average of +1, +4, and –5 is zero… therefore the medicine’s effect is zero.”
As mentioned earlier, this is a meaningless mean. That’s why we want to know the response at each dose.
Anyway, the Sublette meta-analysis gathered randomized trials studying Omega-3 on depression (vs placebo, of course) and got the following dose-response curve.⤵ Note that the horizontal axis is not just amount of total Omega-3, but specifically the extra amount of “unopposed” EPA, above the amount of DHA. Or in other words, “EPA minus DHA”:
The top effect size is around +0.558, which is like going from an F to D–, or C to B–. You get this maximum effect around 1 to 2 grams of extra EPA, and too much EPA gets worse results. The meta-analysis finds that Omega-3 supplements that are ~60% EPA (and the rest DHA) are optimal.
Is this in line with later meta-analyses? More or less! Liao et al 2019 also finds that ~1 gram of ≥60% EPA is best, but actually finds a higher effect size: +1.03. Kelaiditis et al 2023 also finds 1 to 2g of ≥60% EPA is best, but found a lower effect size of +0.43… which is still as good as the best antidepressant! So, I’m taking +0.558 as the median estimate.
Let’s convert this to an actionable recommendation: You want around 1 gram of EPA a day. So if your supplements are 60% EPA, you need 1 gram ÷ 0.6 ~= 1.667 grams = 1667 milligrams. Let’s round this down for convenience: get 1500 mg/day of 60%-EPA Omega-3 supplements.
In comparison, most official health organizations recommend “250–500 mg combined EPA and DHA each day for healthy adults.” That is over three times too low, at least for optimal effects on depression. Which, as we calculated above, is probably around 1500 mg/day. (The official safe dose is 5000 mg/day)
Direct effect on suicide: Finally, a (small) study directly investigating the link between suicide & Omega-3. Sublette et al 2006: “Low [DHA] and low Omega-3 proportions […] predicted risk of suicidal behavior among depressed patients over the 2-year period.” Though keep in mind this is a small study, and it’s observational not experimental. Also, weird that contrary to the above studies on depression, DHA predicted suicide but not EPA. Not sure what to make of that.
Bonus: Omega-3 may also boost cognition? Shahinfar et al 2025: “Enhancement of global cognitive abilities was observed with increasing omega-3 dosage up to 1500 mg/day. [effect size = 1.00, like going from a grade of C to B!], followed by downward trend at higher doses.”
Ghaemi et al 2024 is a meta-analysis on Vitamin D on depression (31 trials with 24,189 participants).
Again, it actually estimates a dose-response curve! Below is Figure 1 + Table 2, showing the effect of Vitamin D dosage on depression vs placebo. The solid line is the average estimated effect, dashed lines are 95% confidence interval. Note the effect size is negative in this figure, because they’re measuring reduction in depressive symptoms:
The upper range of uncertainty is lowest at 5000 IU (International Units) of Vitamin D a day, with an estimated effect size of 1.82, with a 95% uncertainty range, from 0.98 to 2.66. Let’s be pessimistic, and take the lowest end: 0.98, like taking your mental health from an F to D, or C to B.
Is this in line with earlier meta-analyses? Again, more or less! Mikola et al 2022 found a lower estimate: the effect for ≥ 2000 IU/day is 0.407. Note that even this is still on par with the best antidepressant! And Xie et al 2022 found a higher estimate: the effect of > 2,800 IU/day is 1.23. So, I’ll take the median estimate: around 0.98. (And I’m recommending 4,000 IU/day, since that’s the “official” max safe dose. Though as we’ll see later, even the official max dose may be too low.)
Does this still work even if you’re already taking antidepressants? Yup! Table 1 of the first meta-analysis, also shows that Vitamin D helps for both patients using antidepressant medication, and not. This is encouraging: it means you can stack both medications & supplements!
Does this still work even if you don’t have Vitamin D insufficiency? Yes, but admittedly much less. That said, you probably do have a Vitamin D insufficiency. Liu et al 2018 finds that a bit under half of American adults (41.4%) have insufficient Vitamin D blood levels. And Manios et al 2017 finds that over half of kids (52.5%) in Greece — frickin’ sunny Greece! — are still Vitamin D insufficient.
Also, the “official” recommendations are all too low:
So, if these three meta-analyses are right, then high doses — 2000 IU/day or more, possibly 4000 (official max dose) or higher — is optimal. But the official recommendation for Vitamin D is 400–800 IU/day, several times too low.
And even the official max dose of 4000 IU/day may be too low! But McCullough et al 2019 gave over thousands of participants 5,000 to 10,000 IU/day, for seven years, and there were zero cases of serious side effects. This matches later studies like Billington et al 2020, a 3-year-long trial on hundreds of participants, which found “the safety profile of vitamin D supplementation is similar for doses of 400, 4000, and 10,000 IU/day.” (Although 15 participants got “mild hypercalcemia”, but “all cases resolved on repeat testing.” Either way, that’s a small cost for reducing the risk of major depression & suicide.)
And it makes evolutionary sense that 10,000 IU a day should be safe. Your skin, exposed to the Sun’s ultraviolet rays, can synthesize up to (the equivalent of) 10,000 IU a day, before plateauing out. Source is Vieth 1999: “Because vitamin D is potentially toxic, intake of [1000 IU/day] has been avoided even though the weight of evidence shows that the currently accepted [limit] of [2000 IU/day] is too low by at least 5-fold.” And Papadimitriou 2017 reviews several previous studies that find statistical errors behind official recommendations; correcting for these, adults should get 8000 IU/day.
So why are all the official sources still so paranoid about Vitamin D, and lowballing the recommendations? Well, alas, official policy is always a few decades behind the science in any field. See: trans fats, open-label placebos, aerosol transmission of Covid-19, everything about educational policy. And because something something incentives, it’s “rational” for government/insurers to be very risk-averse & slow to change (for better & worse).
Speaking of the Sun, why take supplements instead of just getting Vitamin D from sun exposure? Well, skin cancer. But also: because Sun-Skin D varies greatly depending on the season, your latitude, and your skin type. There’s less ultraviolet rays from the Sun in winter/fall, and at latitudes further from the equator. And the darker your skin is, the less Vitamin D your skin makes for the same amount of Sun exposure. As expected from the bio-physics of skin, Black adults have the highest prevalence of Vitamin D deficiency (82.1%!!), followed by Hispanic adults (62.9%). (But hey, at least Black adults have the lowest incidence of skin cancer. You win some you lose some.) The point is: speaking as someone with Southeast Asian skin, who’s currently in Canada during winter… even if I stood outside naked for hours, I’d get approximately zero IU/day of Vitamin D from the Sun. Thus: supplements.
Direct effect on suicide: Finally, a meta-analysis directly measuring the effect of Vitamin D on suicidal behaviour. Yu et al 2025: “Vitamin D in patients with [suicidal behaviours] were significantly lower than in controls (standardized mean difference: –0.69, or a ‘medium’ difference)”. Reminder that this paper by itself only measures correlation, not causation — but combined with the above experiments of Vitamin D on depression, I think it’s reasonable to guess it’s partly causal.
* Almost half of you have a Vitamin D insufficiency according to the official recommendation (800 IU/day).
* And those official recommendations are way too low. The optimal amount of Vitamin D for depression is probably 4000 IU/day, with an effect around twice that of the best antidepressant.
* Even the official maximum safe dose (4000 IU/day) is below what your body can produce from the Sun in optimal conditions (10,000 IU/day). Recent randomized controlled trials confirm that 10,000 IU/day is, indeed, mostly safe.
* Reminder that official policy is often decades behind the science.
* Reminder that I’m not saying “take supplements instead of antidepressants”; in fact the above meta-analysis shows you can effectively stack them!
Bonus: Vitamin D supplementation was found in several randomized controlled trials to reduce mortality from Covid-19, though much less than official treatments like Paxlovid. Vitamin D also probably helps guard against influenza too, though the evidence is small & early.
Scurvy is caused by a lack of Vitamin C. It’s a condition that causes your wounds to re-open up & teeth to fall out. Scurvy used to kill almost half(!) of all sailors on major expeditions; it’s estimated millions died. It can be cured by eating lemons.
Rickets is mostly caused by a lack of Vitamin D. It’s a condition where kids’ bones go all soft and deformed. During the Industrial Revolution, up to 80% of kids suffered from it. It can be prevented with cod liver oil.
Goiters is mostly caused by a lack of Iodine. It’s a condition where the thyroid gland in your neck swells up painfully, to the size of an apple. During WWI, a third of adult men had goiters. It can be prevented with iodized salt.
About 1 in 4 people are expected to have clinical depression sometime in their life. Depression is the #1 source of the global “burden from disease” in the mental health category, and that category is the #6 burden of disease in the world, above Alzheimer’s, malaria, and sexually transmitted infections.
The effective altruists are all, “woah for just $3000 you can prevent a child’s death from malaria” — and that’s great! save them kids! — but where’s the fanfare for the accumulating evidence that, “woah with cheap daily supplements we can save millions from suicide & depressed lives”?
Over and over again throughout history, some horrific thing that caused millions to suffer, turned out to be “yeah you were missing this one molecule lol”. To be clear: not everything is gonna be that simple, and mental health is not “just” chemistry. Also, all the numbers on this page have with large error bars & uncertainty, more research is needed.
But, as of right now, I feel I can at least confidently claim the following:
...
Read the original on blog.ncase.me »
The first images from the Meteosat Third Generation-Sounder satellite have been shared at the European Space Conference in Brussels, showing how the mission will provide data on temperature and humidity, for more accurate weather forecasting over Europe and northern Africa.
The images from Meteosat Third Generation-Sounder (MTG-S) show a full-disc image of Earth as seen from geostationary orbit, about 36 000 km above Earth’s surface. These images were captured on 15 November 2025 by the satellite’s Infrared Sounder instrument. In the ‘temperature’ image (below), the Infrared Sounder used a long-wave infrared channel, which measured Earth’s surface temperature as well as the temperature at the top of clouds. Dark red corresponds to high temperatures, mainly on the warmer land surfaces, while blue corresponds to lower temperatures, typically on the top of clouds.As would be expected, most of the warmest (dark red) areas in this image are on the continents of Africa and South America. In the top-centre of the image, the outline of the coast of western Africa is clearly visible in dark red, with the Cape Verde peninsula, home to Senegal’s capital Dakar, visible as among the warmest areas in this image. In the bottom-right of the image, the western coast of Namibia and South Africa are also visible in red beneath a swirl of cold cloud shown in blue, while the northeast coast of Brazil is visible in dark red on the left of the image.
The ‘humidity’ image (below) was captured using the Infrared Sounder’s medium-wave infrared channel, which measures humidity in Earth’s atmosphere. Blue colours correspond to regions in the atmosphere with higher humidity, while red colours correspond to lower humidity in the atmosphere.The outlines of landmasses are not visible in this image. The areas of least atmospheric humidity, shown in dark red, are seen approximately over the Sahara Desert and the Middle East (top of image), while a large area of ‘dry’ atmosphere also covers part of the South Atlantic Ocean (centre of image). Numerous patches of high humidity are seen in dark blue over the eastern part of the African continent as well as in high and low latitudes.
Below we see a close-up from MTG-Sounder of the European continent and part of northern Africa. Like the first image above, here we see heat from land surfaces and temperatures at the top of clouds. The heat from the African continent is seen in red in the lower part of the image, while a dark blue weather front covers Spain and Portugal. The Italian peninsula is in the centre of the image.
Temperatures over Europe and northern Africa by MTG-Sounder
And the animation (below) uses data from the MTG-Sounder satellite to track the eruption of Ethiopia’s Hayli Gubbi volcano on 23 November 2025. The background imagery shows surface temperature changes while infrared channels highlight the developing ash plume. The satellite’s timely observations enable tracking of the evolving ash plume over time.
MTG is a world-class Earth observation mission developed by the European Space Agency (ESA) with European partners to address scientific and societal challenges. The mission provides game-changing data for forecasting weather and air quality over Europe.The satellite’s geostationary position above the equator means it maintains a fixed position relative to Earth, following the same area on the planet’s surface as we rotate. This enables it to provide coverage of Europe and part of northern Africa on a 15-minute repeat cycle. It supplies new data on temperature and humidity over Europe every 30 minutes, supplying meteorologists with a complete weather picture of the region and complementing data on cloud formation and lightning from the MTG-Imager (MTG-I) satellite.
ESA’s Director of Earth Observation Programmes, Simonetta Cheli, said, “Seeing the first Infrared Sounder images from the MTG-Sounder satellite really brings this mission and its potential to life. We expect data from this mission to change the way we forecast severe storms over Europe — and this is very exciting for communities and citizens, as well as for meteorologists and climatologists. As ever, the outstanding work done by our teams in collaboration with long-standing partners, including Eumetsat, the European Commission and dozens of European industry teams, means we now have the ability to predict extreme weather events in more accurate and timely ways than ever before.”The Infrared Sounder instrument on board MTG-S is the first European hyperspectral sounding instrument in geostationary orbit. It is designed to generate a completely new type of data product. It uses interferometric techniques, which analyse miniscule patterns in light waves, to capture data on temperature and humidity, as well as being able to measure wind and trace gases in the atmosphere. The data will eventually be used to generate three-dimensional maps of the atmosphere, helping to improve the accuracy of weather forecasting, especially for nowcasting rapidly evolving storms.“It’s fantastic to see the first images from this groundbreaking mission,” said James Champion, ESA’s MTG Project Manager. “This satellite has been 15 years in development and will revolutionise weather forecasting and especially nowcasting. The ability to vertically profile the full Earth’s disk with a repeat cycle of only 30 minutes for Europe is an incredible accomplishment!”
“I’m excited that we can share these first images from the Infrared Sounder, which showcase just a small selection of the 1700 infrared channels continuously acquired by the instrument as it observes Earth,” said Pieter Van den Braembussche, MTG System and Payload Manager at ESA. “By combining all 1700 channels, we will soon be able to generate three dimensional maps of temperature, humidity and even trace gases in the atmosphere. This capability will offer a completely new perspective on Earth’s atmosphere, not previously available in Europe, and is expected to help forecasters predict severe storms earlier than is possible today.”
The MTG mission currently has two satellites in orbit: MTG-I and MTG-S. The second Imager will be launched later in 2026.MTG-S was launched on 1 July 2025. Thales Alenia Space is the prime contractor for the overall MTG mission, with OHB Systems responsible for the MTG-Sounder satellite. Mission control and data distribution are managed by Eumetsat.The MTG-S satellite also hosts the Copernicus Sentinel-4 mission, which consists of an ultraviolet, visible and near-infrared (UVN) imaging spectrometer. Sentinel-4 delivered its first images last year.
Thank you for liking
You have already liked this page, you can only like it once!
...
Read the original on www.esa.int »
The goal of this tracker is to detect statistically significant degradations in Claude Code with Opus 4.5 performance on SWE tasks. What you see is what you get: We benchmark in Claude Code CLI with the SOTA model (currently Opus 4.5) directly, no custom harnesses.
Shows if any time period has a statistically significant performance drop (p < 0.05).
Historical average pass rate used as reference for detecting performance changes.
Percentage of benchmark tasks passed in the most recent day’s evaluations.
Aggregate pass rate over the last 7 days. Provides a more stable measure than daily results.
Aggregate pass rate over the last 30 days. Best measure of overall sustained performance.
Daily benchmark pass rates over the past 30 days. Hover over legend items for details on each visual element.
Daily benchmark pass rate showing the percentage of tasks solved each day.
Historical average pass rate (58%) used as reference for detecting performance changes.
Shaded region around baseline (±14.0%). Changes within this band are not statistically significant (p ≥ 0.05).
95% confidence interval for each data point. Toggle checkbox to show/hide. Wider intervals indicate more uncertainty (fewer samples).
Historical average pass rate (58%) used as reference for detecting performance changes.
Shaded region around baseline (±5.6%). Changes within this band are not statistically significant (p ≥ 0.05).
95% confidence interval for each data point. Toggle checkbox to show/hide. Wider intervals indicate more uncertainty (fewer samples).
The goal of this tracker is to detect statistically significant degradations in Claude Code with Opus 4.5 performance on SWE tasks. We are an independent third party with no affiliation to frontier model providers.
Context: In September 2025, Anthropic published a
postmortem on Claude degradations. We want to offer a resource to detect such degradations in the future.
We run a daily evaluation of Claude Code CLI on a curated, contamination-resistant subset of
SWE-Bench-Pro. We always use the latest available Claude Code release and the SOTA model (currently Opus 4.5). Benchmarks run directly in Claude Code without custom harnesses, so results reflect what actual users can expect. This allows us to detect degradation related to both model changes and harness changes.
Each daily evaluation runs on N=50 test instances, so daily variability is expected. Weekly and monthly results are aggregated for more reliable estimates.
We model tests as Bernoulli random variables and compute 95% confidence intervals around daily, weekly, and monthly pass rates. Statistically significant differences in any of those time horizons are reported.
Get notified when degradation is detected We’ll email you when we detect a statistically significant performance drop. Thanks for subscribing! Check your email to confirm. Something went wrong. Please try again.
...
Read the original on marginlab.ai »
Your browser does not support the audio element.
This content is generated by Google AI. Generative AI is experimental
In August, we previewed Genie 3, a general-purpose world model capable of generating diverse, interactive environments. Even in this early form, trusted testers were able to create an impressive range of fascinating worlds and experiences, and uncovered entirely new ways to use it. The next step is to broaden access through a dedicated, interactive prototype focused on immersive world creation. Starting today, we’re rolling out access to Project Genie for Google AI Ultra subscribers in the U.S (18+). This experimental research prototype lets users create, explore and remix their own interactive worlds.A world model simulates the dynamics of an environment, predicting how they evolve and how actions affect them. While Google DeepMind has a history of agents for specific environments like Chess or Go, building AGI requires systems that navigate the diversity of the real world.To meet this challenge and support our AGI mission, we developed Genie 3. Unlike explorable experiences in static 3D snapshots, Genie 3 generates the path ahead in real time as you move and interact with the world. It simulates physics and interactions for dynamic worlds, while its breakthrough consistency enables the simulation of any real-world scenario — from robotics and modelling animation and fiction, to exploring locations and historical settings.Building on our model research with trusted testers from across industries and domains, we are taking the next step with an experimental research prototype: Project Genie.Project Genie is a prototype web app powered by Genie 3, Nano Banana Pro and Gemini, which allows users to experiment with the immersive experiences of our world model firsthand. The experience is centred on three core capabilities:
Prompt with text and generated or uploaded images to create a living, expanding environment. Create your character, your world, and define how you want to explore it — from walking to riding, flying to driving, and anything beyond.For more precise control, we have integrated “World Sketching” with Nano Banana Pro. This allows you to preview what your world will look like and modify your image to fine tune your world prior to jumping in. You can also define your perspective for the character — such as first-person or third-person — giving you control over how you experience the scene before you enter.
Your world is a navigable environment that’s waiting to be explored. As you move, Project Genie generates the path ahead in real time based on the actions you take. You can also adjust the camera as you traverse through the world.Remix existing worlds into new interpretations, by building on top of their prompts. You can also explore curated worlds in the gallery or in the for inspiration, or build on top of them. And once you’re done, you can download videos of your worlds and your explorations.
Project Genie is an experimental research prototype in Google Labs, powered by Genie 3. As with all our work towards general AI systems, our mission is to build AI responsibly to benefit humanity. Since Genie 3 is an early research model, there are a few known areas for improvement:Generated worlds might not look completely true-to-life or always adhere closely to prompts or images, or real-world physicsCharacters can sometimes be less controllable, or experience higher latency in controlA few of the Genie 3 model capabilities we announced in August, such as promptable events that change the world as you explore it, are not yet included in this prototype. You can find more details on model limitations and future updates on how we’re improving the experience, here.Building on the work we have been doing with trusted testers, we are excited to share this prototype with users of our most advanced AI to better understand how people will use world models in many areas of both AI research and generative media.Access to Project Genie begins rolling out today to Google AI Ultra subscribers in the U.S. (18+), expanding to more territories in due course. We look forward to seeing the infinitely diverse worlds they create, and in time, our goal is to make these experiences and technology accessible to more users.
...
Read the original on blog.google »
The acting head of the US government’s top cybersecurity agency reportedly uploaded sensitive government files into a public version of ChatGPT, triggering internal security alerts and a federal review.
A Politico investigation claims Madhu Gottumukkala, the interim director of the Cybersecurity and Infrastructure Security Agency, uploaded contracting documents marked “For Official Use Only” into ChatGPT last summer.
The report says Gottumukkala requested a special exemption to access ChatGPT, which is blocked for other Department of Homeland Security staff.
Cybersecurity monitoring systems then reportedly flagged the uploads in early August. That triggered a DHS-led damage assessment to determine whether the information had been exposed.
Public versions of ChatGPT share user inputs with OpenAI, which raised concerns inside the federal government about sensitive data leaving internal networks.
CISA spokesperson Marci McCarthy told Politico that Gottumukkala “was granted permission to use ChatGPT with DHS controls in place,” adding that the use was “short-term and limited.”
Gottumukkala has served as acting director since May, while the Senate has yet to confirm Sean Plankey as permanent head of the agency.
The ChatGPT incident follows other reported issues during Gottumukkala’s tenure. Politico said he previously failed a counterintelligence polygraph required for access to highly sensitive intelligence. During congressional testimony last week, he rejected that characterization when questioned.
The report lands as the administration of US President Donald Trump continues to push AI adoption across federal agencies.
Trump signed an executive order in December aimed at limiting state-level AI regulation, while the Pentagon has announced an “AI-first” strategy to expand the military’s use of artificial intelligence.
...
Read the original on www.dexerto.com »
A Waymo robotaxi struck a child near an elementary school in Santa Monica on January 23, according to the company. Waymo told the National Highway Traffic Safety Administration (NHTSA) that the child — whose age and identity are not currently public — sustained minor injuries.
The NHTSA has opened an investigation into the accident, and Waymo said in a blog post that it “will cooperate fully with them throughout the process.” The National Transportation Safety Board said Thursday afternoon that it has also opened an investigation in coordination with the Santa Monica Police Department.
Waymo said its robotaxi struck the child at six miles per hour, after braking “hard” from around 17 miles per hour. The young pedestrian “suddenly entered the roadway from behind a tall SUV, moving directly into our vehicle’s path,” the company said in its blog post. Waymo said its vehicle “immediately detected the individual as soon as they began to emerge from behind the stopped vehicle.”
“Following contact, the pedestrian stood up immediately, walked to the sidewalk, and we called 911. The vehicle remained stopped, moved to the side of the road, and stayed there until law enforcement cleared the vehicle to leave the scene,” Waymo wrote in the post.
News of the crash comes as Waymo faces dual investigations into its robotaxis illegally passing school buses. The NHTSA opened a probe into the problem in October shortly after the first report of the incident in Atlanta, Georgia, and the NTSB opened its own investigation last week after around 20 incidents were reported in Austin, Texas.
According to the NHTSA, the accident occurred “within two blocks” of the elementary school “during normal school drop off hours.” The safety regulator said “there were other children, a crossing guard, and several double-parked vehicles in the vicinity.”
The NHTSA’s Office of Defects Investigation is investigating “whether the Waymo AV exercised appropriate caution given, among other things, its proximity to the elementary school during drop off hours, and the presence of young pedestrians and other potential vulnerable road users.”
Waymo said in its blog post that its “peer-reviewed model” shows a “fully attentive human driver in this same situation would have made contact with the pedestrian at approximately 14 mph.” The company did not release a specific analysis of this crash.
This story has been updated to include information about the National Transportation Safety Board’s investigation.
...
Read the original on techcrunch.com »
Two security professionals who were arrested in 2019 after performing an authorized security assessment of a county courthouse in Iowa will receive $600,000 to settle a lawsuit they brought alleging wrongful arrest and defamation.
The case was brought by Gary DeMercurio and Justin Wynn, two penetration testers who at the time were employed by Colorado-based security firm Coalfire Labs. The men had written authorization from the Iowa Judicial Branch to conduct “red-team” exercises, meaning attempted security breaches that mimic techniques used by criminal hackers or burglars.
The objective of such exercises is to test the resilience of existing defenses using the types of real-world attacks the defenses are designed to repel. The rules of engagement for this exercise explicitly permitted “physical attacks,” including “lockpicking,” against judicial branch buildings so long as they didn’t cause significant damage.
The event galvanized security and law enforcement professionals. Despite the legitimacy of the work and the legal contract that authorized it, DeMercurio and Wynn were arrested on charges of felony third-degree burglary and spent 20 hours in jail, until they were released on $100,000 bail ($50,000 for each). The charges were later reduced to misdemeanor trespassing charges, but even then, Chad Leonard, sheriff of Dallas County, where the courthouse was located, continued to allege publicly that the men had acted illegally and should be prosecuted.
Reputational hits from these sorts of events can be fatal to a security professional’s career. And of course, the prospect of being jailed for performing authorized security assessment is enough to get the attention of any penetration tester, not to mention the customers that hire them.
“This incident didn’t make anyone safer,” Wynn said in a statement. “It sent a chilling message to security professionals nationwide that helping [a] government identify real vulnerabilities can lead to arrest, prosecution, and public disgrace. That undermines public safety, not enhances it.”
DeMercurio and Wynn’s engagement at the Dallas County Courthouse on September 11, 2019, had been routine. A little after midnight, after finding a side door to the courthouse unlocked, the men closed it and let it lock. They then slipped a makeshift tool through a crack in the door and tripped the locking mechanism. After gaining entry, the pentesters tripped an alarm alerting authorities.
...
Read the original on arstechnica.com »
The PlayStation 2’s library is easily among the best of any console ever released, and even if you were to narrow down the list of games to the very best, you’d be left with dozens (more like hundreds) of incredible titles.
But the PS2 hardware is getting a bit long in the tooth, and even though you can hook up the console using RGB component cables to a great upscaler (or use other means) to get the best visuals on a modern 4k TV, emulators have grown in popularity with PCSX2 offering gamers means to scale titles to render internally at higher resolutions, run with a more stable frame rate and, even make use of texture packs.
But do you know what’s better than an emulator? Taking the existing Playstation 2 game and recompiling it to run on a modern platform (such as your Windows or Linux desktop PC). That’s exactly what is being worked on now with PS2Recomp, a static Recompiler & Runtime Tool.
To keep things simple here, this will basically take a Playstation 2 game (which would be designed around the PS2’s unique architecture such as the ‘Emotion Engine’ CPU that’s based around a MIP R5900) and convert it to natively run on whatever platform you’re targeting.
In plain English, this is a tool and obviously, would need to be used on different games. In other words, it’s not just a ‘download and every game automatically runs’ application. But, it will give folks a tool to be able to decompile the game and quite frankly, that’s absolutely incredible.
This is a great stepping stone for some incredible remasters and community remakes of games. There are already HD Texture Packs available for PS2 emulators, as well as other ways to improve visuals. But this would give even more freedom and flexibility to do modify and really enhance the games. That’s to say nothing of totally unlocking the frame rates (and likely not breaking physics or collision detection which is a big problem with emulated titles).
At a guess, too, the games would also run great even with much lower-end hardware than would be needed for emulators. Recompilation efforts in the community certainly aren’t new. Indeed, you can look to the N64 because there have been several high-profile examples of what these kind of projects can achieve.
A few infamous ones would include both including Mario 64 and Zelda. Indeed, there’s a fork of the Mario 64 project supporting RTX (ray tracing) for Nvidia owners. You can see an example of Mario 64 below:
Another example on the N64 is Zelda, where the project has a plethora of visual and gameplay enhancements, and in the longer term again, they’re planning to introduce Ray Tracing.
So, in the future we could be playing the likes of MGS2, Gran Turismo, God of War, Tekken 4, Shadow Hearts with ‘native’ PC versions. This would allow controllers to run (such as dual shock or Xbox controllers) and other features to be bundled in too (exactly as we see with the N64 ports).
So yes, currently playing PS2 games on PC via emulator is still absolutely fantastic, but native ports would be the holy grail of game preservation.
The Playstation 2 architecture is extremely unique, and as I mentioned earlier in this article focused around a MIPS R5900 based CPU known as the Emotion Engine (operating a shade under 300MHz). This CPU was super unique, because Sony implemented a number of customized features include two Vector Units designed to help manipulate geometry and perform a bunch of other co-processing duties.
This was bundled with 32MB of memory, and the GPU was known as the Graphics Synthesizer, runing at about 147MHz, and sporting 4MB of embedded DRAM. Sony’s design was fascinating for the time, and despite its processor clocked significantly lower than either Nintendo’s GameCube or Microsoft’s Xbox, punched well above its weight class.
As a small update — I want to remind people that (as of the time I’m writing this article) the project is *NOT* finished yet, and there is still work to do. But the fact that this is being worked on is awesome for those of us interested in game preservation.
...
Read the original on redgamingtech.com »
Writing about layoffs and the tech market has been on my TODO for several years. Yesterday, the news of 16k Amazon layoffs plus two LinkedIn posts on the same topic back-to-back encouraged me to finally write about it.
Disclaimer: I worked 5 years at Shopify. This is probably why such posts come one after another on my feed but Shopify isn’t the point here, they are just a micro piece of the whole fucked up system.
Tech job market is fundamentally broken and we all pointing fingers at AI.
But having spent almost 2 decades in the industry, I think the rot goes much deeper than ChatGPT.
Truth to be told tech market hasn’t truly ‘improved’ since the 2008 financial crisis. It just mutated into something evil.
After the 2008 mortgage crisis, the economic regime significantly changed. Which was also around the time my interest in Finance began and recently I started to build my own investment tool you can read more about it here.
At that time time we entered an era of extensive liquidity (cheap money). When interest rates are near zero, investors demand growth above all else.
As a result, tech companies stopped building for sustainability and started building for exponential expansion.
Here is a graph shows US Fed Interest Rates by years.
In traditional industries like manufacturing you don’t hire 500 factory workers unless you have a production line that needs them. You don’t over-hire based on a guess.
But in Tech, the playbook is different. Companies over-hire software engineers intentionally. To play the lottery. It is similar to having slow and steady ETF investments vs active investing. No matter how godly you are with active investing sooner or later, you will invest on a loser. Same goes for businesses.
In a factory, “Work in Progress” (unfinished goods) is a liability. You don’t want inventory sitting on the floor; you want it out the door.
In software, we convinced ourselves that “Work in Progress” (hiring engineers to work on projects that haven’t shipped yet) is an asset.
It is not. It is just excessive inventory.
When the market turned, companies realized they were warehousing talent like unsold products. And just like unsold inventory, when the storage costs get too high, you dump it.
Till ~2010, a layoff was a sign of failure. It meant the CEO messed up.
In 2024, a layoff is a signal of “discipline.” Companies lay off thousands, and their stock price jumps.
They are signaling to Wall Street that they are willing to sacrifice human capital to protect margins.
Big Tech companies (think Google, Meta, or any hyper-growth SaaS) operate on a two-tier system:
1. The Core: A fundamental team working on the actual revenue-generating products (the search engine, the ad network, the checkout flow).
2. The Bets: Thousands of engineers hired to build parallel products, experimental features, or simply to keep talent away from competitors and potentially build something that would move into “The Core” tier.
The company knows that most of these side bets will fail. When the economic winds change, the ‘non-core’ staff becomes immediately replaceable.
It’s a vicious cycle: Hire the best people you can find to hoard talent, see what sticks, and lay off the rest when investors want to see better margins.
Most engineers (including me) spent months grinding LeetCode at least twice in their career, studying system design, and passing grueling 6-round interviews to prove they are the “top 1%.”
Yet, once hired, they are often placed on a non-essential team where they become nothing more than a statistic on a spreadsheet.
You jump through hoops to prove you are exceptional, only to be treated as disposable.
For a long time, Europe offered a counter-balance. The pay was lower than Silicon Valley, but the trade-off was stability, stronger labor protections, and a slower, more sustainable pace of work.
As American tech giants expanded into Europe and as European unicorns chased the same growth-at-all-costs playbooks the incentives changed.
Leadership imported US-style compensation models, investor expectations, and organizational volatility, but without importing US-level pay or upside.
”On paper” Europe still has strong labor laws. In practice, companies learned to route around them: constant reorganizations, “strategic refocus” layoffs, performance-managed exits.
The result is the worst of both worlds. European engineers now face US-level job insecurity with European-level compensation and limited mobility. The safety net hasn’t disappeared, but it’s being slowly hollowed out.
And severances… A small, one-time payment is used to justify years of below market compensation, while offering little real protection against sudden displacement.
Europe just became a lower-cost extension of Silicon Valley.
Ultimately, this comes down to how companies signal value.
Traditional businesses used to show their health through revenue, profit, and smart capital investment. Today, Tech companies use layoffs as a marketing signal to Wall Street. They cut costs not because they are going bankrupt, but to show they can be “efficient.”
The more liquidity that was pumped into Tech, the harder this situation became. As long as engineers are treated as speculative assets rather than human capital, the market will remain broken regardless of how good AI gets.
The job market is not “tough” right now because of AI. It is tough because we are unwinding 14 years of financial toxicity.
The liquidity that flooded the tech sector didn’t just inflate valuations; it inflated teams, egos, and expectations.
Until the industry relearns how to build with scarcity rather than excess, the “vicious cycle” of hire-and-dump will continue regardless of how good AI will get.
So you aren’t being laid off because your performance was bad; you are being effectively “liquidated” like a bad stock trade that you sell with a loss.
...
Read the original on substack.com »
Here’s the story of a remarkable scandal from a few years ago.
In the South Pacific, just north of Australia, there is a small, impoverished, and remote country called Papua New Guinea. It’s a country that I’ve always found absolutely fascinating. If there’s any outpost of true remoteness in the world, I think it’s either in the outer mountains of Afghanistan, in the deepest jungles of central Africa, or in the highlands of Papua New Guinea. (PNG, we call it.) Here’s my favorite fact: Papua New Guinea, with about 0.1 percent of the world’s population, hosts more than 10 percent of the world’s languages. Two villages, separated perhaps only by a few miles, will speak languages that are not mutually intelligible. And if you go into rural PNG, far into rural PNG, you’ll find yourself in places that time forgot.
But here’s a question about Papua New Guinea: how many people live there?
The answer should be pretty simple. National governments are supposed to provide annual estimates for their populations. And the PNG government does just that. In 2022, it said that there were 9.4 million people in Papua New Guinea. So 9.4 million people was the official number.
But how did the PNG government reach that number?
The PNG government conducts a census about every ten years. When the PNG government provided its 2022 estimate, the previous census had been done in 2011. But that census was a disaster, and the PNG government didn’t consider its own findings credible. So the PNG government took the 2000 census, which found that the country had 5.5 million people, and worked off of that one. So the 2022 population estimate was an extrapolation from the 2000 census, and the number that the PNG government arrived at was 9.4 million.
But this, even the PNG government would admit, was a hazy guess.
About 80 percent of people in Papua New Guinea live in the countryside. And this is not a countryside of flat plains and paved roads: PNG is a country of mountain highlands and remote islands. Many places, probably most places, don’t have roads leading to them; and the roads that do exist are almost never paved. People speak different languages and have little trust in the central government, which simply isn’t a force in most of the country. So traveling across PNG is extraordinarily treacherous. It’s not a country where you can send people to survey the countryside with much ease. And so the PNG government really had no idea how many people lived in the country.
Late in 2022, word leaked of a report that the UN had commissioned. The report found that PNG’s population was not 9.4 million people, as the government maintained, but closer to 17 million people—roughly double the official number. Researchers had used satellite imagery and household surveys to find that the population in rural areas had been dramatically undercounted.
This was a huge embarrassment for the PNG government. It suggested, first of all, that they were completely incompetent and had no idea what was going on in the country that they claimed to govern. And it also meant that all the economic statistics about PNG—which presented a fairly happy picture—were entirely false. Papua New Guinea had been ranked as a “lower-middle income” country, along with India and Egypt; but if the report was correct then it was simply a “lower-income” country, like Afghanistan or Mali. Any economic progress that the government could have cited was instantly wiped away.
But it wasn’t as though the government could point to census figures of its own. So the country’s prime minister had to admit that he didn’t know what the population was: he didn’t know, he said, whether the population is “17 million, or 13 million, or 10 million.” It basically didn’t matter, he said, because no matter what the population was, “I cannot adequately educate, provide health cover, build infrastructures and create the enabling law and order environment” for the country’s people to succeed.
But in the end, the PNG government won out. To preserve its dignity, it issued a gag order on the report, which has still never been released. There was some obscure behind-the-scenes bureaucratic wrangling, and in 2023 the UN shelved the report and agreed with the PNG government’s existing estimate. And so today, PNG officially has approximately 10 million people, perfectly in line with what had been estimated before.
The truth, of course, is that we have no idea how many people live in Papua New Guinea.
Last week, someone calling themselves Bonesaw went viral on Twitter for a post that claimed that China’s population numbers were entirely fake. China, they said, had been lying about its population for decades: it actually had only about 500 million people. In fact practically every non-Western country had been lying about its population. India’s numbers were also badly exaggerated: the idea that there are 1.5 billion Indians was absurd. The true population of the world, Bonesaw said, was significantly less than 1 billion people.
This is obviously an extremely stupid idea. It’s possible that Chinese population numbers are mildly exaggerated, but the most credible estimates—the ones advanced by Yi Fuxian—are that the exaggeration is on the order of a few percentage points. (It’s also worth noting that no reputable source has yet backed Yi Fuxian’s theory.) Actually faking the existence of billions of people would require a global conspiracy orders of magnitude more complex than anything in human history. Tens or hundreds of thousands of people, spread across every country in the world, would have to be in on it. Local, regional, and national governments would all have to be involved; also the UN, the World Bank, the IMF, every satellite company, every NGO that does work in any of these places. Every election would have to be fake. Every government database would have to be full of fake names. And all for what? To get one over on the dumb Westerners?
So we can dismiss Bonesaw’s claim pretty easily. But, as much as I hate to admit it, his argument does have a kernel of truth. And that kernel of truth is this: we simply have no idea how many people live in many of the world’s countries.
This is not the case for most countries, of course. In wealthy countries, like Germany or Japan or Sweden, populations are generally trusting and bureaucracies are generally capable. Sweden, for its part, maintains such an accurate daily birth-and-death count of population numbers that it no longer even needs to conduct a census. And population numbers are also not so much of a problem in countries like China, India, or Vietnam. These places might be poorer, but they have strong central governments that have a strong interest in knowing what’s going on inside the country. Population counts might be slightly overstated in these places because fertility is falling faster than expected (which could be the case in a country like India where fertility rates are falling quickly), or because local officials are exaggerating the number of students in their schools to secure more education subsidies (that’s Yi Fuxian’s theory of population counts in China), or because more people have emigrated than expected (as was the case in Paraguay when a census revealed its population to be smaller than officials expected). But if the state is in full control of a country, it will want to know what’s going on inside that country; and that starts with the simple fact of knowing how many people live there.
But “the state being in full control of a country” is not a criterion that holds in much of the world. Which brings us to Nigeria.
Nigeria is a huge place. Officially, it’s a nation of 240 million people, which would make it the most populous country in Africa and the sixth most populous country in the world. And without a doubt, there are a lot of people in Nigeria. But we actually have no idea how many there are.
Like PNG, Nigeria is supposed to conduct a census every 10 years. But in Nigeria, the census is a politically fraught thing. Nigeria is not a natural polity, and its ongoing unity as a single country is fragile. And so Nigerian elites expend enormous effort to ensure that Nigeria remains one country. They have two important tools at their disposal. The first is the relative representation of different regions in the Nigerian state. And the second is the distribution of Nigeria’s vast oil revenues. Both of these—how many seats a state is given in the Nigerian parliament, and how large a share of oil revenues it receives—are determined by its share of the population.
So local elites have a strong incentive to exaggerate the number of people in their region, in order to secure more oil revenue, while national elites have a strong incentive to balance populations across states in order to maintain the precarious balance of power between different regions. And so the overwhelming bias in Nigerian population counts is toward extremely blatant fraud.
It’s long been the case that censuses in Nigeria are shoddy affairs. When Nigeria was a colony of Britain, its censuses were limited to Lagos, a few townships, and a small number of villages: so the 1931 census for Nigeria yielded numbers that were too low by as much as 75 percent. Once Nigeria became independent, in 1960, the bias swung from underestimation to overestimation. Nigeria’s first census as an independent state came in 1962, and it immediately caused a political problem: the ruling regime was dominated by northern elites, but the census found that southern Nigeria had more people. And so another census was ordered the next year, which conveniently found an extra eight million people in the north. This pattern of brazenly false numbers continued for decades. The next census, in 1973, was such an obvious fraud that the government opted not even to publish the results. For eighteen years after that there was no attempt even to conduct a census. The next census, in 1991, was by far the most credible, and it shocked many people by finding that the population was about 30 percent smaller than estimated. But even that one was riddled with fraud. Many states reported that every single household had exactly nine people.
In 2006, Nigeria tried once again to count its population. And as luck would have it, it found that since the last census each state’s proportion of the national population had remained exactly the same: so there was no need to change the composition of the Nigerian parliament or the distribution of oil revenues. But this census was an extremely rocky affair. The city of Lagos, for instance, rejected the results of the census, which it claimed undercounted its population in order to preserve northern power; so it conducted its own (technically illegal) census and found that it had eight million more people than the national census had reckoned. And there was also a good deal of violence that accompanied the census: about ten people were killed in clashes around the census, usually in regions with separatist activity. The whole experience was so difficult that Nigeria has opted not to repeat it. The 2006 census was the last time that Nigeria has tried to count how many people live in the country.
So the Nigerian government’s figure of 240 million people is, as is the case in Papua New Guinea, an extrapolation from a long-ago census figure. Is it credible? Very few people think so. Even the head of Nigeria’s population commission doesn’t believe that the 2006 census was trustworthy, and indeed said that “no census has been credible in Nigeria since 1816.” (Nigeria’s president fired him shortly thereafter.) There are plenty of reasons to think that Nigeria’s population might be overstated. It would explain, for instance, why in so many ways there appear to be tens of millions of missing Nigerians: why so few Nigerians have registered for national identification numbers, or why Nigerian voter turnout is so much lower than voter turnout in nearby African nations (typically in the 20s or low 30s, compared to the 50s or 60s for Ghana, Cameroon, or Burkina Faso), or why SIM card registration is so low, or why Nigerian fertility rates have apparently been dropping so much faster than demographers expected.
None of this evidence is conclusive, of course. (There are credible third parties—like the Against Malaria Foundation—that believe that Nigerian population counts might actually be understated.) But the crucial thing, as in Papua New Guinea, is that we don’t know how many people live in Nigeria. It might be that there are 240 million Nigerians, as the Nigerian government claims; or that there are 260 million Nigerians; or that there are only 180 million. We don’t know. But we have plenty of reason to think that the official numbers have little relationship to reality.
What about other countries?
Nigeria is not the only poor country with an extremely patchy history of censuses. Indeed we find that countless poor nations with weak states have only the vaguest idea how many people they govern. The Democratic Republic of the Congo, which by most estimates has the fourth-largest population in Africa, has not conducted a census since 1984. Neither South Sudan nor Eritrea, two of the newest states in Africa (one created in 2011 and the other in 1991), has conducted a census in their entire history as independent states. Afghanistan has not had one since 1979; Chad since 1991; Somalia since 1975.
The various bodies that interest themselves in national populations, from the World Bank to the CIA, reliably publish population numbers for each of these countries. But without grounding in trustworthy census data, we simply have no idea if the numbers are real or not. Estimates for Eritrea’s population vary by a factor of two. Afghanistan could have anywhere between 38 and 50 million people. Estimates for the DRC’s 2020 population range from 73 million to 104 million. How did the country reach its official number for that year, 94.9 million? We have no idea. “It is unclear how the DRC national statistical office derived its estimate,” the U. S. Census Bureau said, “as there is no information in its 2020 statistical yearbook.”
Many other countries do conduct more regular censuses, but do a terrible job of it. Enumerators are hired cheaply and do a bad job, or they quit halfway through, or they go unpaid and just refuse to submit their data. An unknowable number simply submit fake numbers. These are not, after all, technical experts or trained professionals; they are random people sent into remote places, often with extremely poor infrastructure, and charged with determining how many people live there. It is exceptionally difficult to do that and come out with an accurate answer.
So even those countries that do conduct regular or semi-regular censuses often arrive at inaccurate results. The most recent South African census, for instance, undercounted the population by as much as 31 percent—and that is one of the wealthier and better-run nations in Africa. In poorer and less functional countries, statistical capacity is often just nonexistent. Take, for instance, the testimony of the former director of Sudan’s statistical bureau, who said that the most accurate census in Sudan’s history was conducted in 1956, when the country was still under British rule.
It shouldn’t be new to anyone that population data in the poor world is bad. We’ve known about these problems for a long time. And for an equally long time, we’ve had a preferred solution in mind. Technology can compensate for the deterioration of human coordination: we have satellites.
Satellites have two great benefits for counting populations. First, satellites can see pretty much any part of the world from space, and so you entirely obviate the logistical problem of sending people into remote areas: all you need is a small count of some portion of the area under study, which you can use to ground your estimates in something like reality. And second, you don’t have to rely on local governments to obtain the data—so you can get away from the bad incentives of, say, Nigerian elites.
But satellite data can only tell us so much. A satellite can look at a house, but it can’t determine whether three people live there, or six people, or eight people. And often the problem is worse than that. Sometimes a satellite can’t tell what’s a building and what’s a feature of the landscape. Dense cities are a problem; and so, by the way, are jungles—satellites can’t penetrate thick forest cover, and there are quite a few people around the world who still live in forests. (The “forest people” of central Africa, for instance, or a few million of the Adivasi in India.)
So guessing population numbers from high-resolution satellite imagery is an extraordinarily difficult problem. The various companies that guess population numbers from satellite imagery—working with groups like the World Health Organization that might be interested in mapping, say, malaria cases—take different approaches to tackling this problem. And the different approaches they take can lead to wildly different results. For example: Meta and WorldPop both used satellite imagery to predict the population of the city of Bauchi, in northeastern Nigeria. But the numbers that they reached were entirely different, because they take different approaches: Meta uses a deep learning model to detect individual buildings in images and then distributes population proportionally across those structures, while WorldPop feeds a machine-learning model with dozens of variables (land cover, elevation, road networks, so on) and uses that to predict population. Meta guessed that Bauchi has 127,000 children under the age of five; WorldPop says that it has 254,000, twice as many. So Meta’s estimate is about 50 percent lower than WorldPop’s. We see similar differences in other regions. Meta says that Ganjuwa, also in northeastern Nigeria, has 76,000 children under the age of five; WorldPop says that it has 162,000.
And when we do have ground-truth data, we tend to find that satellite-based data doesn’t perform much better. Last year, three Finnish scientists published a study in Nature looking at satellite-based population estimates for rural areas that were cleared for the construction of dams. This was a useful test for the satellite data, because in resettling the people of those areas local officials were required to count the local population in a careful way (since resettlement counts determine compensation payments), and those counts could be compared to the satellite estimates. And again and again, the Finnish scientists found that the satellite data badly undercounted the number of people who lived in these areas. The European Commission’s GSH-POP satellite tool undercounted populations by 84 percent; WorldPop, the best performer, still underestimated rural populations by 53 percent. The pattern held worldwide, with particularly large discrepancies in China, Brazil, Australia, Poland, and Colombia. Nor is it just rural areas being resettled: WorldPop and Meta estimated slums in Nigeria and Kenya to be a third of their actual size.
So satellite data is not a panacea. It might be that in the future the tools advance to the point where they can produce reliable estimates of human populations in areas of arbitrary size. But we are not really close to that point.
Where does that leave us?
I don’t think there’s any reason to embrace the sort of idiotic conspiracism of Bonesaw. We simply have no reason to think that the number of people in the world is dramatically different from what official estimates indicate; indeed while there are specific cases where the numbers might be dramatically off, there’s just no reason to think that this is the case for every country. There are many places, like perhaps Papua New Guinea, where population counts are probably too low. The only thing that can be said with any reliability is that we simply don’t know how many people live in these countries.
Given that we don’t have much evidence of a systematic bias in population counts—Nigeria might overcount, but Sudan might undercount, and at scale these differences should cancel out—the best we can do is assume that there is a sort of “law of large numbers” for population counts: the more units we have under consideration, the more closely the numbers should hew to reality. So population counts for individual countries, particularly in Africa, are probably badly inaccurate. It wouldn’t be surprising if the total population for Africa is off-base by some amount. But we don’t have much reason to think that the global population is very different from what we believe it to be.
But it’s good to be reminded that we know a lot less about the world than we think. Much of our thinking about the world runs on a statistical edifice of extraordinary complexity, in which raw numbers—like population counts, but also many others—are only the most basic inputs. Thinking about the actual construction of these numbers is important, because it encourages us to have a healthy degree of epistemic humility about the world: we really know much less than we think.
...
Read the original on davidoks.blog »
To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".
10HN is also available as an iOS App
If you visit 10HN only rarely, check out the the best articles from the past week.
If you like 10HN please leave feedback and share
Visit pancik.com for more.