I think my favourite statistic in the world is that the inter-rater reliability of peer reviewers is .35, which qualifies as “poor” or “questionable”.

It’s not clear to me exactly why anyone would think peer review is good at quality control. Two guys who may or may not know the topic at hand, and may or may not care about doing a good job are tasked with critiquing1 a paper.
I’m not sure why peer review caught on, whether it’s having a mechanism that prevents people from publishing a ton of super low quality papers to boost metrics, or whether people noticed that papers reviewed by peers tend to be better and therefore made it mandatory. I don’t think it matters. People are trying to bolster the process with more technique (e.g. blinding, more reviewers, public review) but none stand out as particularly helpful. The process is also tedious.
Incentives
People primarily try to get into academia because they like the job itself, are into ideas, want money, or strive for status. People may be motivated by several of these drives, but often one dominates. To become tenured (free), professors have to publish papers that get cited and keep a clean reputation.
Sometimes professors cheat. The most common method is p-hacking: getting a dataset, fishing for an association that reaches p < .052 or even p < .01, coming up with a post-hoc theory as to why it exists3, and publishing it. Some people understand this is a problem and have tried to implement norms to avoid this, such as forcing public code/data, conducting preregistrations, or correcting for multiple testing.
All of these methods are flawed. Some studies can’t really be preregistered (e.g. studies that track inceldom over time) and even then sometimes preregistered studies fail replication. Open code can still have issues running on other people’s computers and correcting for multiple testing isn’t possible if the authors can selectively present their findings.
I call this phenomenon the cycle of tragedy: some people find some way to cheat the system, some people find out the method exists and patch it, the effectiveness of that cheating method decreases, and so on. Adding these rules makes it harder for people to cheat and publish legitimate research too. Not to mention how much more difficult it makes it to get into academia as a newcomer.
The process of studying and learning is something that is best done unconsciously and genuinely; one cannot force sustained interest in abstractions. The anti-cheating protocols involved in publishing things results in more academic labour being conscious, which decreases people’s productivity and satisfaction with their work.
The academic protocols that people would agree to in the age of the internet/LLM would also be very different. Lengthy introductions that flesh out the topic would no longer be encouraged as reviewers can just look up concepts they do not understand or request a summary from an LLM; discussion sections would be less valued for similar reasons.
These problems leave academia populated by different types: personally I see the narcissist, autist, nerd drone, and impostor the most. By “narcissist” I don’t mean a selfish person4 who likes being special, I mean a person who can’t differentiate themselves from their surroundings and devolve into hyper-confident or hypo-confident behaviour depending on their temperament, often seeking external sources of validation and value.
Some time ago I tried researching the group selection question and found it tedious as all people wanted to argue about was altruism and racism. Predictably, altruists and racists like the idea of group selection while Randian individualists don’t. I suspect that group selection does happen, but not in the way people think it does: some psychological types are better for the group (e.g. hero, everyman, genius) while others are detractive; I suspect the narcissist is the type that drags down groups the most.
Over time, the average intelligence of an academic has definitely decreased due to the education system being less selective. I actually don’t think the decreased intelligence itself is that much of a problem, intelligent people can work on their own or with each other if they wish, the issue is that decreasing the barriers to entry of academia has opened it up to the striver class and caused the rule and regulation cascade.
Where does academia go?
I am doubling down on my prediction that academia’s prestige will nosedive in 50 years5 for reasons I outlined in a prior post: lack of progress, lower trust, rise of AI, worsening demographics, and the decreasing selectivity of academia.
The silver lining is that if academia’s prestige reduces, then the strivers will leave, opening up the doors to people who like ideas or just want a job again. It will take a while for people to start up a new way of doing science that doesn’t involve all of the weird rules that we used between 1950 and 2050, but I have faith.
The only reasons I would recommend getting into academia are using your status as a professor to publish books and get onto talk shows, dating hot female (or male) students, getting tenure and proceeding to do absolutely nothing for the rest of your career, and doing important work like GWASes or metascience.
All critical things are bad.
The “p” here refers to a p-value, the probability a statistical association as or more extreme than the one that was found would be observed under the null hypothesis, which is (usually) the absence of a true statistical association. “p < .05” refers to the p-value being under .05, the cutoff where people conclude that their association is (probably) true.
It’s important to differentiate tinkering (good) from p-hacking (bad). See Against Method: Outline of an Anarchistic Theory of Knowledge for more information.
I blame The Last Psychiatrist and Lasch for this conception of narcissism.
15 more years for the millennial + zoomer cohort to wreck havok on the institutions and 35 more years for the generations that grew up with high status academia to be replaced.
I recently commented on a Substack post by an immunologist who said that RFK Jr's claim about being able to see inflammation in the face was absolute BS. I commented pointing to papers which use AI to detect health status from the face, and said if it's possible to do for other health issues, why not for mitochondria? She replied (?) and then immediately blocked me. Now I can't even see the thread.
https://substack.com/@johncollis407103/note/c-150298282
"Nullius in verba" this ain't. Academia's public communications increasingly make it look like a sort of priesthood or guild. Unitiated hoi polloi like myself are assumed to be too stupid to appreciate any sort of research or experiment-based argument.
For academics to maintain their prestige, they'd be far better off at least attempting to explain the results of their research and why they believe what they believe. If they explain their research in the same manner they would explain any Democratic policy issue, that's an excellent way to turn off ~half the population which doesn't trust Democrats on policy.
It would be so easy to actually DO SCIENCE, James Randi style, and challenge RFK Jr to an experimental test of his claim he can see mitochondrial dysfunction in the face. But certain "scientists" have stored their science brain deep in a cabinet somewhere, and have made hobby out of sniffing their own farts. Whatever happened to Richard Feynman? "The first principle is that you must not fool yourself and you are the easiest person to fool."
>dating hot female (or male) students
That's considered questionable nowadays yes?