OPINION20 August 2015

Removing the jokers from the pack

Opinion

As a quantitative executive who specialises in online methodologies, I have come across my fair share of suspicious looking data when we’ve used panel sample – the ‘jokers’ that could compromise the results we deliver to our clients. 

In fact, in a recent study conducted by McCallum Layton via a UK-based panel provider, a shocking 352 completed interviews out of a total base of 2,000 had to be removed and replaced due to various quality control issues, which included:

  • ‘speedsters’ – those people who complete the survey far too quickly for their answers to be truly considered
  • ‘flatliners’ – those who repeatedly give the same answer
  • nonsense verbatims – random letters, or responses that don’t answer the question
  • contradictions in responses – e.g. a respondent says he has a son, but then later in the survey, the son magically disappears
  • offensive language – I’m all for passionate responses, but when the respondent has simply filled the space with swear words, they have to go!

Bearing this in mind, we really owe it to our respondents to provide them with engaging and stimulating surveys to make sure they don’t get bored.  But when the average panellist is on 5-6 panels, and receiving many invites per week, it’s difficult to make our surveys truly stand out.

Most issues come from real-life respondents, but one of the most worrying trends for me is the growing sophistication of automated programs, designed to ‘cheat’ our carefully constructed questionnaires. 

While checking the data on a different survey, we found 30 completes that seemed to draw on a standard set of around eight verbatim responses – the phrasing, punctuation, spacing and spelling mistakes were identical, and couldn’t have come from unrelated ‘real life’ respondents.  More worryingly, these verbatims all referenced the topic of the questionnaire, so wouldn’t necessarily be detectable to the untrained eye. 

When we approached the panel company to report this, they said the IDs in question came from 30 completely different IP addresses, and they simply couldn’t have uncovered these fraudulent responses using their own initial checks.  Once some retrospective digging was done, the perpetrators were found, but the panel provider wouldn’t have been aware if we hadn’t flagged it.

Interestingly, when the same survey was relaunched over a year later, we spotted the same bank of eight verbatims being called upon again.  Having just completed the fourth wave of the research, it’s still an issue and despite changing panel provider, we have to remain vigilant to this kind of activity. 

So I think it falls to us – the researchers and analysts – to give detailed feedback to our panel partners to root out the people that are consistently providing us with unreliable data.  Speaking to others in the industry, I’m not sure that the process of checking data quality is deemed to be as important as the analysis and reporting stages.  If everyone contributes to this effort, we can help to drive sample quality to the top of the agenda.  And if these fraudsters are proving elusive, we need to (at the very least) replace these interviews so our clients are always getting the best possible quality of data.

Laura Finnemore is a senior research executive at McCallum Layton