I also signed this letter
I also signed this letter
4-panel comic. (1) [Person 1 with ponytail flanked by person with short hair and another person speaking into microphone at podium] PERSON 1: In the early 2010s, researchers found that many major scientific results couldnβt be reproduced. (2) PERSON 1: Over a decade into the replication crisis, we wanted to see if todayβs studies have become more robust. (3) PERSON 1: Unfortunately, our replication analysis has found exactly the same problems that those 2010s researchers did. (4) [newspaper with image of speakers from previous panels] Headline: Replication Crisis Solved
Replication Crisis
xkcd.com/3117/
I should say: cutting US funding for Gavi is a tragically misguided decision that can and should be reversed.
π Noise pollution also drowns out the sounds of natureπ
Our research shows that traffic noise reduces the calming, restorative power of birdsong π€
w/ @konraduebel.bsky.social, @simon-butler.bsky.social, @anthonyhigney.bsky.social, Nick Hanley & Eleanor Ratcliffe
π osf.io/preprints/os...
π¨ Thrilled to announce 4th Annual SGPE Summer School
π One-day masterclass on causal inference
π Topics: SCM, matrix completion, partial ID
ποΈ 5 June | πStirling, UK | π» Dr Anthony Higney (Glasgow)
ποΈ Β£50 PhDs / Β£100 staff
Lunch & dinner included!
Sign up:bit.ly/SGPE2025SS
All welcome!
The image shows data from a WHO report on adolescent mental health, focusing on loneliness among 11-, 13-, and 15-year-olds across 44 countries. It highlights that 16% of adolescents report feeling lonely most of the time or always, with rates nearly doubling from age 11 to 15. Girls consistently report higher levels than boys. A heatmap table shows the prevalence by age, gender, and country, with darker shades indicating higher rates. Notably high rates among 15-year-old girls include the UK (40%), Belgium (French, 37%), Finland (26%), and Germany (32%), with much lower rates for boys. The report also notes that loneliness is more common among adolescents from low-affluence families.
Why is there such a difference in loneliness by gender for teenagers? who.int/europe/publicaβ¦
James Webb Satellite took a fantastic shot of the Planet Saturn with it's Near Infrared Camera (NIRCam) and the Mid-Infrared Instrument (MIRI). This is just an Amazing shot that shows the planet and the rings in brilliant detail.
The figure is a line graph titled **"Heat Pump Adoption and Weekly Energy Consumption, Great Britain."** - **Y-axis:** "Change in weekly consumption, kWh," ranging from -350 to 150 in increments of 50. - **X-axis:** "Weeks since adoption," ranging from -50 to 90, with a vertical dashed line at week 0 marking the **"Week of heat pump adoption."** The graph displays two lines with confidence intervals (shaded areas): - **Blue line ("Electricity")**: Starts near zero before adoption, then jumps to approximately +50 kWh after adoption and gradually increases over time. - **Red line ("Gas")**: Starts near zero before adoption, then drops sharply to around -200 kWh after adoption. It shows seasonal variation afterward, with consumption dropping as low as -250 kWh around week 70. **Source:** Researchers' calculations using data from Octopus Energy.
Paper finds heat pump adoption in UK led to 70% less carbon usage and 40% less energy.
www.nber.org/202502/diges...
The detail with a boat and a house
My favourite Minoan fresco from Akrotiri, showing boats arriving to the harbour. A detail. #FrescoFriday
I've read a couple of papers on external validity where they define it in a way that is much too restricted in my view. Doesn't sit right.
What sort of information? Anything: magnitude, sign, variation, heterogeneity (maybe a drug has more of an effect on young than old e.g.)
How do you evaluate that? The same way you evaluate any forecast.
Workshopping here: if an estimate contains information about an ex-ante unobserved treatment then it has some degree of external validity for that treatment.
We should at a minimum do what Nature does, in which the referee comments and author responses are published along with the paper.
Allows the paper itself to be an authoritative artifact while lifting the curtain on the debate that led its creation.
(quoting @dholtz.bsky.social )
4/
Some hope for first gens after all! Fabulous new study notes: "academics from poorer backgrounds introduce more novel scientific concepts but are less likely to receive recognition." This is good, not bad, news. Our different view on life is our super power to make change! www.nber.org/papers/w33289
According to some of the records in the archive, Santa Claus himself sat in Parliament in the 15th century... But was he really in Westminster when he should have been at the North Pole? #Christmas
Below, Dr Hannes Kleineke explores the mystery of 'Nicholas Christmas'
Take a look at artifact. This inscription is ostensibly part of a large stele which has evidently been broken into fragments and we are seeing one (middle) section. All the edges are broken and unfinished. Here is the problem - if this was a broken stele, some letters would be fragmentary.
We present the expected values from p-value hacking as a choice of the minimum p-value among m independents tests, which can be considerably lower than the "true" p-value, even with a single trial, owing to the extreme skewness of the meta-distribution. We first present an exact probability distribution (meta-distribution) for p-values across ensembles of statistically identical phenomena. We derive the distribution for small samples 2<nβ€nββ30 as well as the limiting one as the sample size n becomes large. We also look at the properties of the "power" of a test through the distribution of its inverse for a given p-value and parametrization. The formulas allow the investigation of the stability of the reproduction of results and "p-hacking" and other aspects of meta-analysis. P-values are shown to be extremely skewed and volatile, regardless of the sample size n, and vary greatly across repetitions of exactly same protocols under identical stochastic copies of the phenomenon; such volatility makes the minimum p value diverge significantly from the "true" one. Setting the power is shown to offer little remedy unless sample size is increased markedly or the p-value is lowered by at least one order of magnitude.
Reading this paper now. Interesting but I don't quite agree with his take on "true" p-values. Will post about it next week. arxiv.org/abs/1603.07532
As I explained here, you need to take into account study characteristics correlated with effect size and standard error when you use publication bias detection methods. anthonychigney.github.io/home/blog/Tr...
Abstract Does lead pollution increase crime? We perform the first meta-analysis of the effect of lead on crime, pooling 542 estimates from 24 studies. The effect of lead is overstated in the literature due to publication bias. Our main estimates of the mean effect sizes are a partial correlation of 0.16, and an elasticity of 0.09. Our estimates suggest the abatement of lead pollution may be responsible for 7β28% of the fall in homicide in the US. Given the historically higher urban lead levels, reduced lead pollution accounted for 6β20% of the convergence in US urban and rural crime rates. Lead increases crime, but does not explain the majority of the fall in crime observed in some countries in the 20th century. Additional explanations are needed.
Read the study yourself. t.co/BoS1kMsYlg
Funnel plot with elasticity effect size on x axis and precision (1/se) on y. Shows long right tail and smaller left tail at low precision. Clear positive and significant effect towards top of funnel.
They also cut the bottom off that figure.
Screenshot of twitter account Cremieux. Basically says lead has no effect on crime citing my meta-analysis. Inlcudes funnel plot with partial correlations on the x axis and precision (1/se) on y axis). More precise effects centered around zero PCC.
This person is using my study/chart in a misleading way on twitter.
There is a correlation here between the SE and effect size that is picked up by pub bias methods. Partly this is due to underlying study characteristics. Adjust for this and we find there is pub bias, but lead does cause crime.
I haven't fully read this piece by @jacklandry.bsky.social yet but according to his meta-analysis of only the recent spat of GI pilots, the income elasticity of labor supply on the intensive margin is...0.16. Which is broadly consistent with the literature!
jainfamilyinstitute.org/guaranteed-i...
This fearless science sleuth risked her career to expose publication fraud
Anna Abalkina @abalkina.bsky.social is part of Natureβs 10, a list of people who shaped science in 2024.
Holly Else reports at Nature.
www.nature.com/articles/d41...
Funding available to do a PhD on statistical methods for assessing the integrity of RCTs, based in Aberdeen, with Alison Avenell, myself, Graeme MacLennan and Mark Bolland. Competitive process, funded by MRC Trials Methodology Research Partnership: www.findaphd.com/phds/project...
The authors and others have argued that this shows if effects are heterogeneous, and researchers have a good idea of the effect size ex-ante, and therefore efficiently select sample sizes to all have, for example, 80% power, there will be a correlation between the errors and the effects. This is true but would mean all the studies are estimating effects that are not only known to be different but we know how they are different. We can rank the effects. Should they be doing that? No! In general, in a (random effects) meta-analysis, we assume each study is estimating a different effect and estimate the average of the effects, but the differences are assumed to be random, drawn from a common underlying distribution. When this is not the case, they are not i.i.d, nor exchangeable. What that means is any inference we do from this model will be off. This is important because when we test for publication bias, we are not just describing the data, we are doing inference.
I also have a brief discussion of how you need to meet the assumption of exchangeability at least in meta-analysis. Happy to hear if I am wrong about anything. Only way to learn.
ββ¦[E]ach of these data sets comprises studies on disparate processes, using radically different procedures and designs. Few meta-analysts would be interested in pooling such diverse effects.β In other words, they are combining studies that are looking at completely different things! For me, this is fatal for the empirical part of the paper. This is βcombining incommensurable resultsβ and I cannot see any good reason for doing so. For example, they are combining a replication of how much people agree βthe earth is flatβ with a replication of the effect of less caveolin-1 in mice!
For example, I think they combine incommensurable effects.
I have started writing about papers as I read them to help remember them, but in case others are interested I am also putting them on my site.
First one is a working paper criticising publication bias detection. I think their critique is not quite right.
anthonychigney.github.io/home/blog/Tr...
My default explanation for sudden changes in a time series is change in measurement.
So glad to see another example, sent to me by a colleague who shares my cynicism. Short, clear explanation of Liquid Chromatography-Tandem Mass Spectrometry. Which is important to the story.