WEBVTT

00:00:00.000 --> 00:00:32.403
I wanted to say a bit more about this important issue of recruiting participants.
The quality of the results hinges entirely on the quality of the participants.
If you're asking participants to do things
and they're not paying attention or they're simply skipping through as quickly as they can
– which does happen – then you're going to be very disappointed with the results

00:00:32.403 --> 00:01:01.907
and possibly simply have to write off the whole thing as an expensive waste of time.
So, recruiting participants is a very important topic, but it's surprisingly difficult.
Or, certainly, it can be. You have the idea that these people might want to help you
improve your interactive solution – whatever it is; a website, an app,
what have you – and lots of people *are* very motivated to do that.
And you simply pay them a simple reward and everyone goes away quite happy.

00:01:01.907 --> 00:01:32.771
But it's certainly true with *online research*
that there are people who would simply take part in order to get
the reward and do very little for it.
And it comes as quite a shock, I'm afraid, if you're a trusting person, that this kind of thing happens.
I was involved in a fairly good-sized study in the U.S.
– a university, who I won't name – and we had as participants in a series of studies
students, their parents and the staff of the university.

00:01:32.771 --> 00:02:05.000
And, believe it or not, the students were the best behaved
of the lot in terms of actually being conscientious in
answering the questions or performing the tasks as required or as requested.
Staff were possibly even the worst.
And I think their attitude was "Well, you're already paying me, so
why won't you just give me this extra money without me having to do much for it?"
I really don't understand the background to that particular issue.

00:02:05.000 --> 00:02:32.752
And the parents, I'm afraid, were not a great deal better.
So, we had to throw away a fair amount of data.
Now, when I say "a fair amount", throwing away 10% of your data is probably pretty extreme.
Certainly, 5% you might want to plan for.
But the kinds of things that these participants get up to – particularly if you're talking about online panels,
and you'll often come across panels if you go to the tool provider, if you're using, say for example, a card-sorting tool

00:02:32.752 --> 00:03:03.008
or a first-click test tool and they offer you respondents for a price each,
then be aware that those respondents have signed up for this purpose,
for the purpose of doing studies and getting some kind of reward.
And some of them are a little bit what you might call on the cynical side.
They do as little as possible. We've even on card sort studies had people log in,
do nothing for half an hour
and then log out and claim that they had done the study.

00:03:03.008 --> 00:03:31.042
So, it can be as vexing as that, I'm afraid.
So, the kinds of things that people get up to: They do the minimum necessary;
that was the scenario I was just describing.
They can answer questions in a survery without reading them.
So, they would do what's called *straightlining*.
Straightlining is where they are effectively just answering every question the same
in a straight line down the page or down the screen.
And they also could attempt to perform tasks without understanding them.

00:03:31.042 --> 00:04:04.091
So, if you're doing a first-click test and you ask them, "Go and find this particular piece of apparel,
where would you click first?", they'd just click.
They're not reading it; they didn't really read the question.
They're not looking at the design mockup being offered;
they're just clicking, so as to get credit for doing this.
Like I say, I don't want to paint all respondents with this rather black brush,
but it's *some* people do this.
And we just have to work out how to keep those people from polluting our results.
So, the reward is sometimes the issue, that if you are too generous in the reward

00:04:04.091 --> 00:04:30.000
that you're offering, you will attract the wrong kind of participant.
Certainly I've seen that happen within organizations doing studies on intranets,
where somebody decided to give away a rather expensive piece of equipment at the time:
a DVD reader, which was – when this happened – quite a valuable thing to have.
And the quality of the results plummetted.
Happily, it was something where we could actually look at the quality of the results and

00:04:30.000 --> 00:05:01.511
simply filter out those people who really hadn't been paying much attention to what they were supposed to be doing.
So, like I say, you can expect for online studies
to discard been 5 and 10% of your participants' results.
You also – if you're doing face-to-face research –
and you're trying to do quantitative sorts of numbers,
say, you'd be having 20 or 30 participants, you probably won't have a figure quite as bad as that,
but I still have seen, even in face-to-face card sorts, for example,

00:05:01.511 --> 00:05:33.349
people literally didn't *understand* what they were supposed to be doing,
or didn't get what they were supposed to be doing, and consequently their results were not terribly useful.
So, you're not going to get away with 100% valuable participation, I'm afraid.
And so, I'm going to call these people who aren't doing it, and some of them are not doing it because they don't understand,
but the vast majority are not doing it because they don't want to spend the time or the effort;
I'm going to call them *failing participants*. And the thing is,
we actually need to be able to *find* them in the data and take them out.

00:05:33.349 --> 00:06:01.059
You have to be careful how you select participants, how you filter them
and how you actually measure the quality of their output, as it were.
And one of the big sources of useful information are the actual tools that you are using.
In an online survey, you can see how long people have spent, you can see how many questions they have answered.
And, similarly, with first-click testing, you can see how many of the tasks they completed;
you can see how long they spent doing it.

00:06:01.059 --> 00:06:30.446
And with some of these, we actually can also see how successful they were.
In both of the early-design testing methods – card sorting and first-click testing –
we are allowed to nominate "correct" answers
– which is, I keep using the term in double-quotes here because
there are no actually correct answers in surveys, for example;
so, I'm using "correct" in a particular way:
"Correct" is what we think they should be doing when they're doing a card sort, *approximately*,
or, in particular, when they're doing a *first-click test*,

00:06:30.446 --> 00:07:03.580
that we think they ought to be clicking around about here.
Surveys as a group are a completely different kettle of fish, as it were.
There are really no correct answers when you start.
You've got your list of research questions – things that you want to *know* –
but what you need to do is to incorporate questions and answers
in such a way that you can check that people are indeed *paying attention*
and *answering consistently*.
So, you might for example change the wording of a question and reintroduce it later on

00:07:03.580 --> 00:07:33.650
to see if you get the same answer.
The idea is to be able to get a score for each participant.
And the score is your own score, about basically how much you trust them
or maybe the *inverse* of how much you trust them.
So, as the score goes up, your trust goes down.
So, if these people keep doing inconsistent or confusing things,
like replying to questions with answers that aren't actually real answers – you've made them up –
or not answering two questions which are effectively the same the same way, etc.,

00:07:33.650 --> 00:08:02.946
then you would get to a point where you'd say, "Well, I just don't trust this participant,"
and you would yank their data from your results.
Happily, most of these tools do make it easy for you to yank individual results.
So, we have to design the studies to *find* these failing participants.
And, as I say, for some these tools – online tools we'll be using – that is relatively straightforward, but tedious.
But with surveys, in particular, you are going to have to put quite a bit of effort into that kind of research.

00:08:02.946 --> 00:08:32.735
Steps we can take in particular:
Provide consistency checks between tasks or questions.
Ensure that "straightlined" results – where people are always answering in the same place on each and every question down the page –
ask the same question again in slightly different wording or with the answers in a different order.
Now, I wouldn't go around changing the order of answers on a regular basis.
You might have one part of the questionnaire where
"good" is on the right and "bad" is on the left;

00:08:32.735 --> 00:09:00.000
and you might decide to change it in a completely different part
of the questionnaire and make it really obvious that you've changed it
to those who are paying attention.
But whatever it is that you do, what you're *trying* to do is to find people who really aren't paying much attention
to the directions on the survey or whatever the research tool is,
and catch them out and pull them out of your results.
And of the issues you should be aware of if you're paying for participants from something

00:09:00.000 --> 00:09:30.274
like your research tool *supplier* is that you can go back to them and say,
"These people did not do a very good job of completing this survey, this study."
And ask them to refund you for the cost of those.
You tell them that you're having to pull their data out of your results.
Also, it helps to tidy up their respondent pool.
Perhaps it's not your particular concern, but if you do end up using them again,
it would be nice to know that some of these people who are simply gaming the system
have been removed from the respondent pool.

00:09:30.274 --> 00:09:45.598
So, reporting them – getting them removed from the pool – is a sensible thing to be doing.
And, finally, devising a scoring system to check the consistency
and also checking for fake responses and people who are just not basically doing the research as you need them to do it.