Tuesday, December 31, 2019

 

That Was Quite the Mess

I'll end the year with a look back at a failed model, namely the forecasting model for the recent British elections. It was a pretty spectacular failure at 75 seats short of the actual result. But what did Edison say, supposedly...it's not failure, but just another example of what doesn't work. In the spirit of "we write to learn" (many folks have said that), I present the following essay.


What Went Wrong? A Post-Mortem of an Election Forecast
Ross E. Burkhart
School of Public Service
Boise State University
14 December 2019

            Yesterday, the aftermath of a consequential election in a critical Western democracy, the United Kingdom, came upon an important day for our Political Science students: the capstone course project showcase. A brief recap of both events follows, and then the theme of this short essay emerges: reflecting on a poor election forecast.
            The election process is simple in Britain: voters tick a box for their favored candidate and place the paper in a box, to be counted by the returning officer's staff in their geographic constituency. The winners in the 650 constituencies, decided by plurality (meaning winning at least one more vote than their closest opponent), are thus duly elected Members of Parliament. The party winning the most seats gets to form the next government. In this case, the Conservatives won a smashing majority of 80 in the House of Commons. This is the Conservative Party's largest majority since 1987, and suggests a fundamental realignment in British politics, where working-class voters crossed over to the Conservatives who promised a speedy Brexit to address working-class grievances over being ignored by globalized Brussels bureaucrats.
            Meanwhile, the capstone Political Science students presented their projects in poster format, on public display in the Student Union Building. They conducted research on an impressive array of topics: Title IX funding and transgender rights in schools, policing reform and civil liberties, increasing the intelligence budget, Twitter followers and campaign spending, balancing abortion politics, how reporters cultivate relationships with politicians, the prospects for a hard or a soft Brexit, the effects of door-to-door campaigning in state legislative elections, water pollution in Costa Rica, state-level abortion restrictions and the Supreme Court, and human rights violations in Burma and China. The students also had an eye in their poster presentations toward highlighting findings and shortcomings in their research. As critics, they were most effective evaluators in pointing out the shortcomings in their research and suggestions for future refinement. Such impressive self-reflection behooves me to follow their example with one of my recent research pieces.
            Two weeks ago I posted "A Snap Elections Forecasting Model of British Elections" in which, using data that reflected the snap election environment (defined as an election occurring well before the scheduled end of the Parliament), I gave a forecast of the Conservatives winning 290 seats in the House of Commons. It was a mere 75 seats off!  I need to understand what went wrong with the model. The forecasting model was as follows:
S = 20.50 + .73 P + 1.61 J – 3.78 T     (Eq.1)
    (1.45)   (3.03)   (2.74)   (1.82)
R2 (adj.) = .68    S.E.E. = 5.42    D-W = 1.84   N = 15
Where S = percentage of House of Commons seats won by the incumbent governing party, P = approval of the performance of the Government at either the time of the dissolution of Parliament during a snap election or six months prior to dissolution during a non-snap election, J = jobless rate at the time of dissolution during a snap election or six months prior to dissolution during a non-snap election, T = the number of terms the incumbent party has been in office assuming a fatigue factor takes place. The figures in parentheses are absolute t-ratios, S.E.E. is the standard error of estimate, and D-W is the Durbin-Watson statistic after the Cochrane-Orcutt procedure.
            I noted in my forecast that the R-squared statistic is on the low side for a forecasting model, and that the standard error of estimate is rather high, suggesting a fair bit of unexplained variance in the seats forecast, though certainly nothing like the 75 seat error in forecast. Of course, that much error is unacceptable in a forecasting model, hence this postmortem.
            We are left with several possible paths toward enlightenment. One is that the election was an outlier, a freak of nature that could not be forecasted well. We reject this path because other forecasters were rather close to the mark (Murr, Stegmaier, and Lewis-Beck 2019), but there were prominent forecasts that were wide of the mark (Lebo and Fisher 2019), suggesting there was no real unanimous view as to the forecasted result (Greenwood 2019).
            A second possibility is that British elections are inherently unpredictable, with those who claim to make a forecast very close to the final result just having been lucky in their forecasting. After all, forecasting is an informed probability exercise, and even a broken clock is correct twice a day, goes the old adage, so anyone has a chance of getting it right. While there are critics of the election forecasting enterprise (van der Eijk 2005), this possibility of unpredictable elections seems unlikely as well. Many elections are well forecasted (Burkhart 2018). More broadly speaking, the forecasting literature has successfully forecasted election results in many countries over a lengthy time period, too well to ignore (Bélanger and Trotter 2017).
            A third possibilithy is that the model itself is rather faulty. There are three pieces of evidence that point in this direction.  First, the R-squared is rather low for forecasting purposes at .68. Ideally, we seek an R-squared of at least .75 in election forecasting models. Second, the unemployment independent variable is oddly signed in the positive direction, suggesting that incumbent governments that raise unemployment rates are more likely to win more seats. (The positive sign is most likely influenced by Margaret Thatcher's three strong majorities during the highest unemployment rates in British history since the Great Depression.)  Third, the statistical significance of the terms in office variable is pretty marginal at approximately the .09 level. The preferred significance level is .05.
            Further examination reveals that the next election forecast, presumably taking place in 2024 given the substantial Conservative majority, erodes in model performance. The forecast equation for the 2024 election is as follows (Cochrane-Orcutt estimation):
S = 7.16 + .95 P + 1.68 J – 1.37 T     (Eq. 2)
    (.56)   (4.10)   (3.45)   (.78)
R2 (adj.) = .61    S.E.E. = 5.60    D-W = 2.06   N = 16
All model performance indicators decline in 2024 compared to 2019. The y-intercept is no longer significant, indicating that the baseline percentage of seats for the Conservatives, while estimated to be 7.16% of all seats, could actually be 0%. The popularity function is elevated in importance, almost equating a one-to-one impact on the Conservative vote. The jobless independent variable remains steady in impact, while the trend variable drifts into statistical insignificance.
            In fact, simply dropping the trend variable yields the following result:
S = -.63 + 1.07 P + 2.35 J     (Eq. 3)
       (.27) (5.53)  (2.95)
R2 (adj.) = .62    S.E.E. = 5.51    D-W = 2.15   N = 16
This model performs better than Eq. 2 in that all variables are statistically significant and the adjusted R-squared rises, if just a smidgen. Yet the y-intercept suggests that the baseline for Conservative seats is effectively zero, which is nonsensical.
            To be sure, more work is needed in understanding a proper structural model of British election forecasting. In the face of the strong performance of the citizen forecasting models and the multilevel YouGov model, it is tempting to proclaim this the elegy of the structural equation model. However, we must remember another statistical issue with this literature, and that is small sample sizes. The Central Limit Theorem is a good rule of thumb for regression analyses upon which these forecasts rely. The 2019 election is one data point out of the small sample of 17 British general elections in this model, so perhaps the only elegy worth making is that it is too soon to tell its demise.
REFERENCES
Bélanger,  Éric, and David Trotter. 2017.  "Econometric Approaches to Forecasting." In
            The SAGE Handbook of Electoral Behavior, eds. Kai Arzheimer, Jocelyn Evans,
            and Michael S. Lewis-Beck. Beverly Hills, Calif. Sage Publications.
Burkhart, Ross E. 2018. "Was Anyone Right? Assessing the 2016 U.S. Presidential
            Election Forecast." The Blue Review, 15 January.
Greenwood, Joe. 2019. "Experts predict that the general election will be tighter than
            expected." LSE British Politics and Policy, 10 December.
Lebo, Matthew, and Stephen Fisher. 2019. "Forecasting the 2019 General Election using
            the PM and Pendulum Model". LSE British Politics and Policy, 20 November.
Murr, Andreas, Mary Stegmaier, and Michael Lewis-Beck. 2019. "Citizen forecasting
            2019: a big win for the Conservatives." LSE British Politics and Policy, 4
            December.
van der Eijk, Cees. 2005. "Election Forecasting: A Sceptical View." British Journal of
            Politics and International Relations 7:210-4.

Comments: Post a Comment

<< Home

This page is powered by Blogger. Isn't yours?