Joseph Angolano, Mainstreet vice-president, sent me the link to the full report, so I got to update my voting intention graphs (with error bars!) before the undecided are allocated.
I’ve only included the last Mainstreet and Léger polls because those of CROP, Ipsos, and Forum were conducted before the campaign had even officially started.
You can see that all results are compatible (the error bars overlap) except for Quebec Solidaire:
Indeed, none of the error bars touch the 12.5% line.
Last night, when the Léger results were announced, Too Close To Call’s Bryan Breguet pointed out that Léger and Mainstreet disagreed on voting intentions for the CAQ and for QS:
En gros Léger et Mainstreet s’entendent parfaitement sur le PLQ et le PQ, mais Léger a la CAQ 5 points plus élevé et QS 5 points plus bas. Intéressant que les différences soient entre CAQ et QS1.
He comes to that conclusion using scores after distribution. As we have seen, before distribution both polling firms agree that CAQ voting intentions lie somewhere between 27% and 31%.
In this morning’s blog post, Bryan expanded on the topic. He returned to the difference between Mainstreet’s local and province-wide polls, which he had mentioned in yesterday’s blog post:
Mainstreet et Léger en fait s’entendent parfaitement sur le PLQ et le PQ. Par contre ils ont des chiffres fort différents pour la CAQ et QS. Mainstreet a ces partis à respectivement 31% et 16% alors que Léger les a à 35% et 11%. Une différence de 4-5 points pour chaque parti. Qui dit vrai? Impossible d’y répondre pour sûr mais les sondages par comté de Mainstreet sont bien plus cohérents avec une Coalition à 35%-36% et QS à 11%. Ainsi je serais tenté de dire que Léger a possiblement raison ici. Mais il nous faudra attendre d’autres sondages (et en fait l’élection) pour en être sûr2.
If you can read French, I highly encourage you to read his blog post on the disagreement between Mainstreet’s province-wide and riding polls. To entice you to read the whole thing for yourselves, here’s the table he comments:
So let’s recap. On the one hand, province-wide Léger and Mainstreet polls disagree on QS support. On the other hand, the results of Mainstreet’s riding polls fit better with the picture painted by Léger (CAQ higher, QS lower).
I left you hanging last Friday when I promised a new data visualization of the most recent polls. To refresh your memory, the margin of error depends on the score in the poll (it increases when the score gets closer to 50%) and the sample size (one goes up while the other goes down). It does not depend on the size of the population of which you want to know the opinion.
I did a graph similar to the one in Qc125 (with margins of error this time) for the last three polls in that Qc125 diagram. I added the Forum poll (conducted on 23 August with 965 respondents) and the last Léger (conducted from 24 to 28 August with 1010 respondents).1
I first tried to do it in Google Spreadsheets, so you could access the file and check everything out. However, I could only add an error bar that was either a constant or a percentage. As we saw on Friday, polling margins of error are a bit more complicated than that.
I also tried with Excel and its open-source equivalent LibreOffice but bumped into the same problem: there was no way of defining a different error bar for each point. It doesn’t come as much of a surprise, then, that there are so few representations of polling data with margins of error.
I had managed just fine by using candlestick charts (used to describe movements in the stock market), but Martin objected that they were ugly. Hence, to please the pole in our tandem in charge of graphics, I pulled out the big guns and programmed the graph in R, an open-source statistical analysis software.
After too many hours fiddling about, here’s what I got2:
Each point situates the party’s score in the poll. The vertical line contained within the two horizontal lines describes the confidence interval if you take into account the margin of error at 95% (or 19 times out of 20). You can see that the lines higher up are longer than the lower ones. As we said at the beginning, the margin of error increases with the proportion (or rather with its proximity to 50%).
By comparing the scores of different parties vertically within a single poll, we see that:
in CROP, the CAQ and the Liberals are statistically tied;
in Forum, the Liberals are statistically tied to the PQ instead (with the CAQ way ahead);
in Léger, voting intentions for the CAQ and the Liberals overlap and are therefore statistically tied as in CROP.
Differences in data collection mode
Too Close To Call’s Bryan Breguet looked into QS’s diverging polling scores in a blog post last Thursday. He was troubled by the fact that the disagreement follows data collection lines:
You can see that three polls place the party below 10% and two above (the ones that use IVR or robocalls). More importantly, the results of these two groups don’t overlap, even if we take into account the margin of error. (None of the horizontal bars touches the 10% line.)
Mainstreet and Forum use IVR and get results significantly higher than CROP or Léger using online polls, and Ipsos. The latter combines online polling with good old live callers: humans talking to other humans over the phone to ask them polling questions.
Bryan ran 10,000 simulations and came to the conclusion that either Mainstreet or Léger was wrong. It assumed that “real” voter intentions for QS were at the 10% mark. He simulated for a sample size of 1,010 respondents, as was the case in Léger.
On the horizontal axis are voting intentions for Québec Solidaire (centred at 10% because that’s his starting assumption). On the vertical axis is the number of simulations for which QS got a given score.
Distribution of 10,000 simulations
with QS at 10% and a sample size of 1,010
Léger has QS at 6%, but we see very few simulations peg the left-wing party under 7%. For Mainstreet, Bryan uses data from the nightly polls (available through a paid subscription). Québec Solidaire had at the time 13,1% (it has since smashed the 15% barrier). Once again, nearly to simulations at all came up with such a high result.
An effect limited to QS voting intentions
When we turn to the other parties, we see that there is no systematic bias according to the data collection mode.
Using IVR, Forum places CAQ and the PQ way ahead of other pollsters, beyond the margin of errors.
In the case of Liberals, CROP is the pollster that pegs them uncharacteristically high.
We’ll therefore be keeping a close eye on how the differences in scores between pollsters evolve. They only seem to matter when trying to determine the composition of the National Assembly because it seems that we already know which party will take over the government if the election was held today: Too Close To Call’s Sunday post discusses CAQ’s over 99% chances of winning.
It seems that Bryan Breguet answered a tad too quickly to Marc-Antoine Berthiaume’s Tuesday tweet pointing out the enormous difference between Léger and Mainstream polls regarding Quebec Solidaire support amongst voters aged 18 to 34. (To find out what the heck I’m talking about or to refresh your memory, read my Wednesday post, “Younger voters and polling variability.”)
Mainstreet big wigs have launched a campaign on Twitter to assert just how confident they are about their polling results (and claiming in passing that Léger’s are out of whack). Here is one of their most recent tweets:
Nous assistons à une croissance réelle de QS au cours des derniers jours. Nous le ressentons de manière anecdotique, et nous le voyons dans nos sondages nocturnes. Quelque chose est en train de se passer.1
(I suppose that by “sondages nocturnes,” the Mainstreet vice-president means “nightly polls.”)
Bon, en plus des querelles de politiciens, nous avons droit aux querelles de sondeurs… 😉2
To settle the matter, Bryan made 20,000 simulations, starting from the assumption that “actual” support for QS in that age group is in fact the average of the two polls’ scores: 18.4%. He posited a sub-sample size of 150 respondents (the size of Léger’s sub-sample).
He found that it was highly improbable, though not completely impossible, that, if QS is actually at 18.4% amongst voters aged 18 to 34, one poll would get 8% and another would get 25.9%. The bar chart below shows the number of polling simulations (vertical axis) for which a given score (horizontal axis) was reached for QS support with young people aged 18 to 34.
He came to the conclusion that one of the two polls is probably out of whack (but there’s no way of knowing which one because there would need to be an election right now, not in a month’s time).
Actually, support for QS amongst voters aged 18 to 34 must be either higher, either lower than 18.4%. If it was higher, the curve would be shifted to the right, and the Mainstreet score (25.9%) would no longer be as improbable. In contrast, if it was lower, the curve would be shifted to the left, and the Léger score (8%) would no longer be impossible.
Léger and Mainstreet are the extremes, but neither one nor the other is completely isolated, as can be seen in this bar chart of QS voting intentions for 18- to 34-year-olds:
Here’s how Bryan sums up the situation:
En conclusion: les différences observées entre sondeurs pour QS chez les 18-34 ans ne peuvent pas être complètement expliquées par les marges d’erreur et tailles d’échantillons. Il y a quelque chose d’autre. Après, j’avoue ne pas avoir d’explication actuellement.3
So what’s this margin of error he’s talking about? Is it always ±3, 19 times out of 20?
What factors into the margin of error
Ok, so I’m going to include a formula for those for whom it makes life easier, but don’t worry, I’ll jump directly to the implications.
The margin of error at the 95% level (hence 19 times out of 20) is 1.96 standard deviations or:
where p is the proportion (the percentage for that answer in the poll: 8% in Léger and 25.9% in Mainstreet) and n is the sample size (the number of respondents).
That means that:
The margin of error is not dependent on the size of the population you want to study. Whether you want to find out the opinion in a single riding or in the entire province of Quebec does not affect the margin of error of a given poll.
In other words, it’s not because you’re studying a smaller population that you can settle for a smaller sample: the margin of error depends on the sample size, not the size of the population.
The margin of error goes up when the sample size goes down (that’s much more intuitive).
The margin of error also depends on the poll result (the proportion): the lower the percentage (or, more accurately, the further away from 50%), the smaller the margin of error. It’s therefore not always ± 3 (or the margin of error given at the beginning of the poll), 19 times out of 20.
The confidence interval spreads from the value of the percentage minus the margin of error to the value of the percentage plus the margin of error.
Visualizing the margin of error
Qc125 charts presenting polling results do not show the margin of error and give the impression that it’s showing a variation across time (with the line joining the observations). I don’t like these data visualization decisions.
At least, the visualization contains all the information needed to calculate the margins of error for each observation: the percentage (p) is written in the circles and the sample size (n) is at the bottom of each “column” (on top of the data collection mode and field dates, which don’t influence the margin of error4).
In my next post, I’ll offer you a slightly different way of visualizing poll results and dig deeper into the differences between polling firms.
I woke up this morning to this post shared in my Facebook feed:
On top is a bar chart picked up from a Le Devoir article published last night at 9:12pm1 and altered by Jean-François Provençal from the hit millennial-humour TV show Les Appendices. Below is what is called a meme. It’s a still from the TV-reality show The Osbournes in which fallen metalhead Ozzy Osbourne doesn’t know what’s going on when a phone rings.2
Read til the end
Adding “who answer the phone” after “Voting intentions of 18-34-year-olds” presupposes that polls are still conducted by calling randomly selected numbers from the phone book. As I’ve explained in the Polling section of the Elections primer, pollsters have developed new methodologies to adapt to new communication habits.
And, lo and behold, if you read the Le Devoir article until the end, you’ll find this box outlining the methodology:
Le nouveau sondage Web de Léger a été réalisé auprès de 1010 Québécois ayant le droit de vote du 24 au 28 août, alors que la campagne était commencée. Par comparaison, un échantillon probabiliste similaire aurait une marge d’erreur d’environ plus ou moins 3%, 19 fois sur 20. (emphasis added) 3
So the “problem” with this poll is not that it uses a methodology now deprecated in our smartphone world.
Marc-Antoine Berthiaume raised a much more pertinent issue on Twitter:
Comment expliquer que chez @leger360, pour les 18-34 ans, #QS arrive en 5ème place avec 8% et que @MainStResearch place en 2ème position avec 23,4%? C’est un écart de 15,4%!4
He thus contrasts the Le Devoir chart based on Léger’s polling data with data from Mainstreet’s Baromètre élections 2018. This tool is funded by Groupe Capitales Médias, a conglomerate of Power-Corporation-subsidiary Gesca’s French-language dailies sold to Martin Cauchon.5 There is a paid subscription service for individuals.6
Too Close To Call’s Bryan Breguet replied:
Facile: tailles d’échantillon petites. Donc variance est grande7
but quickly added:
Cela étant dit la différence est un peu grande ici, je l’avoue8
Let’s take a closer look by placing the two datasets side-by-side:
What jumps out at first is that Mainstreet data is provided without the undecided having been allocated to parties since it contains the share of undecided young voters. Léger always provides sub-sample data after having allocated the undecided.
It’s therefore normal that Léger percentages are higher than those in Mainstreet: the sum total of voting intentions in Léger is 100% while it’s 90%10 in Mainstreet. That’s what explains all the blue in the right-hand column (that shows the difference between both datasets).
Incidentally, we can assume that the presence of 1.6% of young voters who intend to cast their ballots for another party in Mainstreet and their absence in Léger is compensated by the presence in the latter of 3% of young voters who intend to back the NPD Quebec. In other words, it’s likely that a fair share of those who would vote for “another” party in Mainstreet would in fact vote for the NPD Quebec.
But that’s not what shocked the interwebz. By bringing together the two datasets and ordering the parties according to their score in Mainstreet, we immediately see where the polls disagree: on voting intentions for Quebec Solidaire and for the Liberals (still with voters aged 18 to 34).
Alexandre Blanchet, a political science Ph.D., offers a convincing demonstration of the uncertainty inherent to polls in his French-language guide to polling for journalists and other geeks. (Just a heads up: I went on the page twice, and it appears to have brought my Internet connection down for a minute both times.)
la bonne question à se poser n’est souvent pas de savoir quel sondage est meilleur qu’un autre, mais plutôt de savoir de quelle réalité il est le plus probable que ces sondages émanent. Les sondages sont une manifestation de la réalité qui nous intéresse. Ils en sont une manifestation plus ou moins précise, et parfois plusieurs réalités différentes pourront être cohérentes avec les sondages que nous observons. Avec le scénario de l’élection de 2003 où la réalité était claire et nette, nous avons obtenu des sondages qui étaient eux aussi très clairs: le PLQ menait, le PQ était deuxième et l’ADQ était troisième. Avec le scénario de l’élection de 2012, où les intentions de vote étaient beaucoup plus serrées, plusieurs réalités étaient concordantes avec les sondages que nous obtenions.11
Indeed, polls can change even if the underlying reality hasn’t itself changed.