On the morning of the 2016 presidential election, the New York Times published its final pre-election blog post on its data analysis vertical, The Upshot. Confidently forecasting an imminent victory for Hillary Clinton, the paper’s modelers projected she had an 85 percent chance of winning. Other forecasts, the article noted, were similarly favorable, ranging from 71 percent in Clinton’s favor (Nate Silver’s FiveThirtyEight) to over 99 percent (the Princeton Election Consortium). The most likely Electoral College outcome was 322 votes for Clinton—just shy of Obama’s performance in 2012.
In reality, of course, Clinton suffered a decisive defeat. That evening, the Times’s infamous election “needle” traversed nearly its entire semicircle in a few hours, reaching over 95 percent likelihood of a Trump victory by 11:15 p.m. on the East Coast. In the battleground states of Michigan, Wisconsin, and Pennsylvania—Clinton’s so-called “Blue Wall,” all of which wound up going for Trump—polling was found to be off by an average of 5 to 6 percentage points, all biased in favor of Clinton.
Recently reignited debates about the place of polling in progressive politics are one skirmish in a broader struggle over the future of the Democratic Party.
The errors were so large that the next time around, in the 2020 election, the Times began noting how predictions would differ if they were as off as they had been in 2016. But even this effort didn’t save face. In both the 2020 and 2024 election cycles, models were by many metrics just as wrong as they had been in 2016, or even worse. In 2020, Biden was projected to win the popular vote by 7 to 8 points but wound up with just a 4.4-point margin, marking the largest polling error in forty years; at the state level, the error was the worst since tracking began in 2000. As for last year, polling bias fell but again cut in the same direction: in the run-up to Election Day, Kamala Harris emerged as the slight favorite in both the popular vote and Electoral College. We all know how that turned out.
These much-mocked projections—along with the cycles of disillusionment and defenses they spawned—loom large over recently reignited debates about the place of polling and data in progressive politics. The dispute is one skirmish in a broader struggle over the future of the Democratic Party, coming at a moment when the stakes could hardly be higher for forging a left-liberal popular front against Trump’s authoritarianism.
Roughly speaking, one side touts polling and other electoral data as our best guides to political reality—and thus, it is claimed, to political strategy. These tools are far from perfect, the argument goes, but they are better than alternatives. To those in this camp, hostility to data smacks of impressionism at best (just “vibes”), rank ideology at worst; either way, it’s irrational—a dangerous denial of reality—and progressives must do politics by poll if they actually care about winning. Critics, in turn, insist that it’s the data-mongers who are the ideologues, ignorant of history, lacking in moral backbone, and devoid of political imagination. The influence and professional prestige they wield as “very smart,” well-credentialed, STEM-fluent wonks—steeped in a technocratic worldview, at odds with bottom-up movements—is decried as just a way of dressing up megadonor-approved policies in the guise of neutral expertise, or just a means for media companies to capture audience attention by treating politics like sports.
The shape of this debate today is the product of the rapid rise of the political polling industry in the late twentieth century. From candidate selection to message crafting, audience targeting, and news reporting, polling and data analytics have come to play an outsize role in professional politics and mainstream media, shaping not just how politics is conceived, practiced, and covered but also the contours of liberal common sense. It wasn’t always this way. FDR was the first president to draw on polling to help guide campaign strategy; historians have traced its rise to prominence among Democratic strategists since the 1960s, a trend coinciding with the “hollowing” out of political parties and the explosion of money in politics. The Upshot and Vox both launched in 2014—the same year FiveThirtyEight relaunched on ESPN—clinching the ascendance of a wonkish approach to progressive punditry during Obama’s second term. Today, the orientation is epitomized by data-driven consultancies like David Shor’s Blue Rose Research and election analysis groups like Split Ticket—the latter cofounded by Lakshya Jain, the director of political data at the new self-styled organ of liberalism’s salvation, The Argument.
Much ink has been spilled on these developments—in particular, on whether ubiquitous political data analysis has in fact led to greater political insight and more effective political campaigns. But the sparring can create more confusion than it dispels. How is one supposed to decide between the positivists and the skeptics? Are polling and quantitative models reliable indicators of things like what voters think and how they will behave? Were these methods, and the broader sensibility they reflect, invalidated by what happened in 2016, 2020, and 2024? Without clear, shared standards about when a prediction can be validated as “correct” and what level of accuracy makes for a successful method, debate endlessly spins on. When skeptics cite embarrassing misses like Clinton v. Trump—or surprise upsets like Zohran Mamdani’s—the data gurus reply that predictions are always probabilistic and polls always have a margin of error, and besides, gut feelings perform even worse. When critics rail against polling’s professional-managerial function or its association with monied interests, quants dismiss these criticisms as bad-faith efforts to cast aspersions on their work and reputations that don’t show how their evidence-based models are flawed. (Never mind that the models are usually proprietary, not open to public auditing.)
More insight can be gained by looking under the hood of a particular controversy. And as it happens, a new one has been brewing this year concerning Wins Above Replacement (WAR)—a metric analogous to the baseball statistic of the same name, used by Split Ticket to argue that Democrats should “moderate” to improve their electoral prospects. The WAR wars can seem like a purely technical dispute over methodology. In fact, they illustrate why arguments over “what the data say we should do” can’t ever be settled by the data alone. The chain of reasoning that runs from political data to political prescription depends on value-laden judgments about the game of politics itself: which past patterns will hold in the future and which can be changed, how open we are to experimentation, and how much we are willing to risk if we’re wrong.
These judgments, which define and delimit the scope of what politics can do, are inescapable: no strategist can avoid making them. The problem arises in concealing them, treating them as plain facts rather than open to reasonable dispute.
The basic idea behind Split Ticket’s WAR score is to compare how well a congressional candidate performs in an election with how well they are “expected” to perform based on “structural” factors—incumbency, for example, or past voting patterns in the district. A WAR score of zero is interpreted as meaning the candidate’s performance was perfectly generic: what any “replacement-level” candidate from the same party would have achieved. A higher WAR score is taken to indicate a more impressive performance relative to expectations—and a clue that emulating the candidate could improve someone’s electoral chances.
Consider an example. Split Ticket starts with the candidate’s actual performance, measured relative to the performance of the presidential candidate of the same party. Suppose a Democratic congressional candidate won her race last year by 6 percentage points in a district that went +2 for Harris. The discrepancy—a raw overperformance of +4—is suggestive. But can we conclude on this basis alone that the congressional candidate was a “better” candidate than Harris—say, because of her distinctive messaging style, policy views, or charisma? Not if other structural factors contributed to the result. If the candidate were an incumbent, for example, she might be expected to have a built-in advantage over the presidential candidate of, say, +3 points. Accounting for this “structural” factor, her WAR—the portion of her performance attributable to her individual characteristics, beyond what we’d expect the typical incumbent to get—would be +1 point.
For analysts who mine WAR or something like it for political insights, the score is not simply a descriptive measure: a way of capturing and ranking how well candidates perform. It’s considered valuable evidence for inferring what factors help win elections and thus for making political prescriptions. In other words, the aim is to draw conclusions about what kinds of candidates do better than others—and on that basis, to make judgments about political strategy writ large, including which candidates to support with party funding and which platforms and messaging styles to encourage (and discourage).
That’s just what happened in February, when Jain and Harrison Lavelle of Split Ticket published a Washington Post op-ed reporting their findings that ideological moderates—as measured by another numerical index—outperformed progressives in 2024, in the sense of having higher WAR scores. “The sooner partisans and pressure groups reject the seductive notion that America actually wants their specific version of ideological purity, rather than moderation that might compromise principles they value, the sooner they will win more elections and get more of the policy they want,” they concluded. No error bars, no uncertainty: just flat, bald-faced prescription.
New pieces, archive selections, and more straight to your inbox
When the task of politics is reduced to identifying the top predictors of voting for Democrats, success is reduced to ensuring behavioral compliance.
In the end, this is not a dispute that can be resolved by data. Politics cannot be drained out of political strategy, for the basic reason that central questions of this kind are simply not open to purely technical or empirical resolution. Skepticism toward prevailing modeling methods is therefore neither dogmatic nor anti-empirical. Rather, it is a reasonable and considered judgment about the social world—shaped by distinct interpretations of evidence, assessments of risk, senses of urgency, and visions of what is possible and worth trying—which says, on this basis, that the challenges we face can only be met by radical changes to the prevailing political order, even if doing so means pushing past the boundaries of prevailing political reason. Independent and nonprofit, Boston Review relies on reader funding. To support work like this, please donate here.
The post How to Lie with (Political) Statistics appeared first on Boston Review.
Amazon is going through something of a massive restocking mission this week for Pokémon cards,…
Amazon is going through something of a massive restocking mission this week for Pokémon cards,…
Magic: The Gathering has kicked off its Teenage Mutant Ninja Turtles set prerelease weekend, but…
The much-delayed Spider-Man: Beyond the Spider-Verse currently has a June 18, 2027 release date. If…
Resident Evil Requiem has landed on Steam, and is now the series' biggest launch to…
There's a whole bunch of great Pokémon deals available right now online as part of…
This website uses cookies.