loading . . . Quantitative political science shouldnât favour tools over theory - Impact of Social Sciences estimated reading time: 6 min Publication in highly ranked political science journals has increasingly become defined by the use of complex quantitative methods. Silje SynnĂžve Lyder Hermansen argues the fetishization of quantitative tools has obscured the wider context and meaning of research in the field. Political science isnât short on new methods. But too often, our publishing culture is a slot machine: pull the lever enough times and someone will hit the jackpot. Thatâs how we end up mistaking noise for knowledge. We live in an age of data abundance: easy access, cheap software, and a generation skilled enough to âlet the data speakâ. Out of this, two academic fads emerge. The p-hacking boom. Ten researchers sample from the same population and test whether X is correlated with Y. The rule: publish if p < 0.1, quit if not. Inevitably, someone gets âluckyâ. A tidy theory is retrofitted, a few anecdotes sprinkled in, maybe even a citation. To look more comprehensive, we frame it as settling a debate between theories A and B. Letâs be honest: political science has argued in all directions before. Both theories are compelling and internally coherent. VoilĂ : new âtruthâ established. The âcorrectionâ race. The community has learned. Data access, software, and training have all expanded. Now twice as many researchers sample from the same population. The rule: publish only if the result is significant in the opposite direction. Replications are boring, non-significant results inconclusive. Surprise is the only currency. Inevitablyâwith the same probability as beforeâsomeone gets âlucky.â VoilĂ : new, opposite âtruthâ established. But the underlying reality never changed: the true correlation is zero. Weâve built orthodoxy on statistical fluke. So what have we learned? Less than nothing. Weâre wrong, and weâve grown confident about being wrong. Worse yet, weâve done it eyes wide shut; armed with the very statistical training that tells us the game is rigged. Thatâs the danger of one-correlation papers in a publishing culture that fetishizes originality. Now, let me be clear: Iâm not sneering from the sidelines. Iâm a compulsive scraper; I get my kicks from the three-star slot machine of regression analysis. I love reading new research, and I want to publish too. How do we fix this? We can always demand stricter p-values before weâre impressed. But the core problem doesnât go away. At best, we reset the slot machine odds to where they used to be: bigger data, stricter cutoffs â same game. As our scraped datasets balloon, the machine continues to spit jackpots. Can causal inference fix the problem? In some ways, yes: it pins down where an effect occurs, and it feels rigorous. But theory is too often left by the wayside, reduced to claims about âreal-life dataâ and the absence of obvious confounders â precisely where alternative theories tend to live. The toolkit is fantastic for public policy questions. What was the effect of our latest reform? Data is available, and the question is bounded. But external validity becomes the stumbling block. Can we reproduce those conditions in the future? Can we resolve bigger societal problems with another reform? Without theory that specifies fairly general scope conditions, the answers donât travel. But what are we actually measuring? If all we do is describe ourselves, how does that teach us why we behave as we do? We could go all in on the experimental ideal: pre-registration, conjoint survey experiments, lab-in-field. These methods double down on falsifiability. Pre-registration forces us to commit before pulling the lever. They require careful planning and let us test competing theories head-to-head. Thereâs a real boldness in this turn: no more hand-waving; the question is framed, the die is cast. But bold is still bounded. Conjoints show how people trade off attributes in a survey, but the selection effect looms large. How much do we rely on convenience samples â MTurk, YouGov, undergrads? And even if we limit our claims to that group, we may still construct trade-offs our respondents would never perceive in the real world. Then thereâs the newest darling: text analysis and prediction. Data science masquerades as political method. We scrape massive corpora of speeches, manifestos, TikTok feeds, or laws, run them through embeddings, topic models, or large language models (LLMs). Theyâre magical at approximating meaning. They mirror us back to ourselves â endlessly. But what are we actually measuring? If all we do is describe ourselves, how does that teach us why we behave as we do? Then we move on to prediction. If we can forecast 80% of coups or measure âpolicy distanceâ to six decimals, who wouldnât be impressed? Out-of-sample accuracy shows a model describes our data, but rarely why. Prediction is not explanation, unless the model maps onto a data-generating process we can actually theorise. I donât snark at these advances. I admire them! Iâm often baffled at how cleverly confounders can be designed away. Iâm in absolute awe at the ballsiness of preregistration and prediction. My stupor is endless when text models make intuitive sense. I might be a jaded millennial, but I recognise progress. But these are tools â powerful tools â not methods. And certainly not theory. There is no such thing as a perfect paper, but we can make them better. One way is to measure the same thing in different ways. That boosts construct validity and rules out obvious p-hacking, but still doesnât test theory itself. For that, we need imagination: to nerd out on the data-generating process. If what we claim is true, what else would we observe? What other empirical implications follow from our theoretical models? Itâs an old trick, but it still works with the flashiest of tools. In my opening parable, a false finding struck once in ten attempts. If the true correlation is zero, one out of ten researchers still gets a significant result by chance. If we demand two distinct tests of the same theory, then the odds of a false positive drop to one in a hundred. Every extra implication strengthens the test exponentially. The tools we employ and the research designs we devise have multiplied. We truly live in exciting times. But they donât replace our human brains. And hereâs the bigger point. Weâre often asked to justify our âlushâ life on taxpayersâ dime by proving âsocietal impactâ. Somehow this has come to mean media outreach. Fine. But weâre not the engineers. Weâre knowledge providers for decision makers: the people (taxable or not) and their representatives. What we need is a research agenda that carefully dissects â paper by paper â how democracy fails, what behaviors our institutions induce, how opinions are formed, and how they translate into policies. This begins with research and ends with teaching. Political science isnât in the business of theorising about Life, the Universe, and Everything. Our task is no less urgent: to study power â how itâs won, wielded, constrained, and lost. Too often we debate whether 42 was correctly estimated, rather than carefully specifying the question up front. Thatâs the jackpot worth chasing. đšEnjoying this blogpost? Sign up to our mailing list and receive all the latest LSE Impact Blog news direct to your inbox đš The content generated on this blog is for information purposes only. This Article gives the views and opinions of the authors and does not reflect the views and opinions of the Impact of Social Science blog (the blog), nor of the London School of Economics and Political Science. Please review our comments policy if you have any concerns on posting a comment below. Image Credit: mihalec on Shutterstock. https://blogs.lse.ac.uk/impactofsocialsciences/2025/11/06/quantitative-political-science-shouldnt-favour-tools-over-meaning/