Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.
The long view
There’s been noise in the public debate on the forthcoming New Zealand election about the value of voting intention polls. There’s often not enough attention to the long view on these things, so here is a graphic of 15 years of polling data:
Taken all together, the polls are probably doing better than many people think.
There’s been a bit of a systematic underestimate of New Zealand First support perhaps, and maybe last election an overestimate of the Greens vote, but it’s not too bad.
The other thing we see in that chart is that voting intention obviously changes over time. This is something that’s often forgotten when all the attention is on just the latest number. But look at the 2004 surge in National’s support after Don Brash’s controversial “Orewa” speech, followed by a partial collapse to not much more than it had been six months earlier. Or the interesting high levels in National support in the year or so before the 2008 and 2011 elections followed by swift movement down in the lead up to the election. Or Labour’s journey between 2011 and 2014 – growth in support up to solidly in the 30% range before a collapse in the last year to the election. In any of these cases, taking an average of an individual pollsters’ results over time and treating it as a prediction of the election day result is just unfair. People do change their minds.
I always think in terms of there being three sources of uncertainty in translating a poll result into a prediction of the election:
- uncertainty from polls’ relatively small sample size, which is published as the margin of error of “plus or minus two or three percent” and is the only indication of uncertainty that is usually given prominence
- the rest of what is known to statisticians as “total survey error”, which comes from things like systematic small under- or over-estimates from subtleties in the polling companies’ methods, or ways that sampling people in reality doesn’t resemble just picking golf balls from a bucket (for example, unlike the golf balls, certain types of people may systematically refuse to talk to surveyers)
- even if we knew for sure how people would vote today, people change their minds right up to the last minute.
My forecasts using this data try to take all of that into account. Only on the evening of 23 September will we know how that went.
The source of this data is ultimately the news organisations commissioning the polls and the volunteers contributing to Wikipedia. The data is collated in my nzelect R package.
Here’s the R code that drew the graphic:
R-bloggers.com offers daily e-mail updates about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.
Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.