Should FiveThirtyEight’s elasticity index have a wider spread?

[Edit: I no longer stand by the conclusions of this post. For more, see “How much do you believe your results?“]

In 2012, Nate Silver wrote about elastic and inelastic states. An elastic state is one with lots of swing voters. This means that if the national electorate shifted by one percentage point, you would expect that state’s vote to shift by more than one percentage point. Alaska, for instance, is an elastic state: although it is a solidly red state, its vote count swings more than most states’ from election to election. Mississippi, on the other hand, is an inelastic state: compared to most other states, the Republican candidate gets a pretty consistent percentage of the state’s vote from election to election.

FiveThirtyEight has pioneered a very useful tool that they call the elasticity index: a quantitative measure of how elastic the fifty states (and D.C.) are. A state is defined to have elasticity e if every 1\% shift in the national vote predicts an e\% shift in the state’s vote. FiveThrityEight introduced the elasticity index in 2012 (linked above) and updated it again in 2018. I don’t think FiveThirtyEight has made their methodology fully transparent, but they describe it as such:

The scores work by modeling the likelihood of an individual voter having voted Democratic or Republican for Congress, based on a series of characteristics related to their demographic (race, religion, etc.) and political (Democrat, Republican, independent, liberal, conservative, etc.) identity. We then estimate how much that probability would change based on a shift in the national political environment. The principle is that voters at the extreme end of the spectrum — those who have close to a 0 percent or a 100 percent chance of voting for one of the parties, based on our analysis — don’t swing as much as those in the middle.

Per the elasticity index, the most elastic state is Alaska (elasticity 1.16); the least elastic is Alabama (0.89). (Washington, D.C. has a lower elasticity index than any of the fifty states: 0.80.)

Interestingly, the 2012 version of the elasticity index had a much wider spread: the highest index was 1.29 (Rhode Island) and the lowest was 0.45 (Washington D.C.; the lowest state was Mississippi, 0.63). The standard deviation of the 2018 elasticity index is 0.068; the standard deviation of the 2012 index was more than double that: 0.144.

I’m not sure whether FiveThirtyEight’s methodology changed or whether their methods gave much tighter numbers in 2018 than they did in 2012, but I decided to investigate. Is the 2018 elasticity index is too tightly concentrated? Is the 2012 elasticity index not concentrated tightly enough? Or is something else going on?

Here was my basic plan: the overall national vote varies from election to election. In 2008, Obama (D) beat McCain (R) by 7.2 percentage points. In 2012, Obama (D) beat Romney (R) by 3.9 percentage points.1 This means that the country as a whole saw its vote tally shift rightward by 3.3 percentage points from 2008 to 2012. And although most states shifted rightward from 2008 to 2012, some shifted more than others. If FiveThirtyEight’s elasticity index is accurate, then states with elasticity 1 would on average have shifted 3.3 points to the right, (hypothetical) states with elasticity 2 would on average have shifted 6.6 points to the right, and (hypothetical) states with elasticity 0 would on average have not shifted at all. That is, if we plot elasticity on the x-axis and rightward shift (in percentage points) on the y-axis, the slope of the line should be around 3.3.

I did this analysis and the slope turned out to be 5.4.2

2008-2012_vs_2012_Elasticity

The 95\% confidence interval on this slope is [-2.6, 13.4].3 In particular, the slope I got was not even close to statistically significantly different from 3.3, so it’s not reasonable to draw conclusions quite yet. However, this constitutes very weak evidence that even the 2012 elasticity index (which has a wider spread than the 2018 one) is too narrow: the shifts from 2008 to 2012 were larger than you would expect based on states’ elasticities.

In 2016, Clinton (D) beat Trump (R) by 2.1 percentage points (but lost the electoral college). I did the same analysis using the 2012 elasticity index and the shift from 2012 to 2016. Just as we would expect a slope of 3.3 for the 2008-to-2012 shift, we would expect a slope of 1.8 for the 2012-to-2016 shift (since that is how much the country as a whole shifted). The slope turned out to be 8.7 with a 95\% confidence interval of [-2.2, 19.5]. Again, this is more (weak) evidence that the 2012 elasticity index is too narrow.

2012-2016_vs_2012_Elasticity

I also did this for the shift from 2012 to 2016 using the 2018 elasticity index. I got a slope of 12.4, with a 95\% confidence interval of [-10.9, 35.87], compared to the expected slope of 1.8. This also constitutes weak evidence that the 2018 elasticity index is too narrow.4

2012-2016_vs_2018_Elasticity

We thus have a few weak indications that FiveThirtyEight’s elasticity index may not be wide enough. I have one final piece of evidence to offer, which is state-by-state shifts in Donald Trump’s popularity since the 2016 election. I considered the percentage of people who voted for Trump in 2016 and compared it with Trump’s current approval ratings, as estimated by Civiqs.5 The second measure was broadly lower (on average by about 10 points as of August 28th), so we should expect that if we plot the shift from Trump’s vote percentage to his approval rating in each state against the state’s elasticity, the slope would be about -10. In fact, the slope was roughly -30, with a 95\% confidence interval of [-9, -50].6

Trump_Popularity_vs_2018_Elasticity

On the surface this seems like decent evidence that the elasticity index isn’t wide enough. I think it’s actually only weak evidence, for a few reasons. First, Trump’s current approval rating is an imperfect measure of how we would expect him to do in a general election, even against a fairly popular Democrat. Second, it seems to me that Civiqs’ model of Trump’s approval ratings and FiveThirtyEight’s method of generating the elasticity index are similar in nature. This might make the popularity-shift numbers and the elasticity numbers might artificially well-correlated. Third, shifts in Trump’s approval rating correlate between demographically and geographically similar states, which makes the above confidence interval too narrow (this applies to the earlier confidence intervals as well). So all in all, I’d say we have several pieces of weak evidence that FiveThirtyEight’s elasticity index is too narrow.

Taken together, do these weak pieces of evidence constitute strong evidence? I’m not comfortable asserting that they do, especially because I trust FiveThirtyEight to put a lot of thought into things like elasticity. But at a minimum, I’m willing to suggest that the index may be too narrow and encourage others to look into this. If you find anything interesting, let me know!

1. I decided to look at presidential rather than congressional election results because lots of congressional races are uncontested and this messes with percentages. I think accounting for this and using congressional results would make for a better, more thorough analysis, but I’m not sure how to do that. If anyone does this analysis, I’d be interested in seeing the results.

2. I threw out Washington, D.C. from my analysis because it was a leverage point, as well as Utah, because it behaved weirdly in both the 2012 and 2016 elections (2012 because Romney was a Mormon and 2016 because Evan McMullin received a large fraction of the conservative vote).

3. This confidence interval is accurate only if some assumptions are made: the underlying data is linear; the y-values are normally distributed with the same standard deviation; and residuals are uncorrelated. I’m not entirely comfortable with the last of these assumptions — states that are demographically similar and in the same region tend to have correlated shifts — but I think the interval is not terribly off.

4. Take this evidence with a grain of salt, since calculation of the 2018 indices may have been partly based on the shift from 2012 to 2016.

5. Elasticity is intended to model shifts in how people vote between elections, not shifts in popularity. However, I think Trump’s approval rating is a pretty good proxy for how much of the vote he would get against a reasonably popular, Obama-type Democrat.

6. As before, I have excluded Washington, D.C. and Utah.

Leave a comment