28 4. FACETED FRAMEWORK OF INTERACTIVE IR USER STUDIES
with search systems). As is shown in Figure 4.1, we identied a series of subfacets under these two
facets and found that these two main facets and the associated subfacets not only aect the study
procedure, but also partially decide the quality of data collected, the reliability of statistical results,
as well as the generalizability of the ndings emerged from the empirical evidences. is subsection
explains the two main facets and the related subfacets in detail with examples from the research
paper collection.
Participant recruitment is the starting point of user study procedure and determines the
source of available data (e.g., search behavioral data, self-reported and neuro-physiological data on
interaction experience, search evaluation data). Dierent recruitment methods are often associated
with dierent experimental setups. From our collection of user study papers, we summarized two
main recruitment approaches: (1) widely used small-scale user study recruitment methods, such as
yers, in-class recruitment, personal social networks, and internal mailing lists; and (2) large-scale
crowdsourcing techniques, such as recruitment within certain institutions (e.g., Microsoft’s large-
scale user studies) or via professional crowdsourcing and survey platforms (e.g., Amazon Mechan-
ical Turk, SurveyMonkey). In small-scale user studies, researchers usually expect high commitment
from participants and ask them to engage in relatively complex tasks (e.g., complex search task and
work task, relevance and usefulness judgments, post-search in-depth interview) that need to be car-
ried out through specic experimental systems and/or within certain controlled lab settings. With
a variety of predened constraints and conditions, researchers are more likely to collect a variety of
more detailed, reliable data on users’ search interactions, cognition, and experience.
Let us examine a couple of examples. Cole et al. (2013) invited participants to complete pre-
dened search tasks in their controlled lab and collected data on their search interactions. In this
case, researchers had a relatively nice control over a series of contextual factors (e.g., task type, task
topic, search system and interface) and successfully collected several types of data on users’ search
interactions, including search behavioral data, eye movement data, and qualitative data regarding
search experience and obstacles from post-search interview transcripts. Moshfeghi, Triantallou,
and Pollick (2016) conducted an fMRI study on information needs and demonstrated that users’
knowledge states (i.e., information need, anomalous state of knowledge) can be inferred from the
activities of certain neuro-physiological signals. Apparently, at least part of the data collected in
these small-scale user studies would not be available if the participants were recruited and partici-
pated in the study via crowdsourcing platform.
When we look at the limitations, the recruitment methods used in small-scale (often lab-
based) user studies are inecient, time consuming, and expensive. To reduce the cost and expedite
participant recruitment and data collection process, many researchers in research institutions tend
to adopt convenience sampling methods and recruit students on campus as their participants.
However, only studying information seeking and search behaviors of students from the same
campus may result in the lack of variation in participants’ knowledge background and eventually