58 7. CONCLUSION AND FUTURE DIRECTIONS
create new opportunities for researchers in the IR community to further improve IIR study design,
reporting, and evaluation practices.
With respect to future directions, there are several ways to improve our faceted scheme and
to further enhance our understanding of the roles and potential of IIR user studies. For instance, fu-
ture research can expand the scope of user study review and include more relevant venues for paper
collection and facet extraction, such as ASIS&T, CIKM, JCDL, and so on. By doing so, researchers
may be able to identify new subfacets and factors (e.g., new system and/or interface features and
search assistant tools, new data collection techniques and associated features, new models for data
analysis) and thereby can improve the current version of faceted evaluation framework.
Besides, it is also important to develop new scales, measures, and indicators to quantify the
faceted evaluation framework and to improve the generalizability of the framework in evaluating a
wide range of IIR user studies. is quantication should include both the measurements of dier-
ent facets and dimensions (e.g., how many widely-ignored, under-reported subfacets are discussed,
which critical statistics are reported) and the quantication of baselines (e.g., a series of interrelated,
quantied user study facet measures) for user study evaluation. Due to the fundamental dierences
among dierent research focuses and problems, researchers should employ dierent baselines when
evaluating dierent types of IIR user studies. In addition, as it was discussed in Chapter 6, estab-
lishing new platforms (e.g., paper submission tracks, workshops) for reporting replication studies
and unexpected results can also deepen our understanding of the potential eects of dierent facets
and dimensions (e.g., task facets, participant or user characteristics, system or interface features) on
the results of user studies.
As Robertson (2008) points out, “a eld advances not by deciding on a single best compro-
mise, but through dierent researchers taking dierent decisions, and the resulting dialectic” (p.
447). In the past two decades, we have witnessed the diversity in user study design manipulations
and decisions (e.g., participant recruitment, task design, study device, study procedure, type of data
collected) which has largely supported the explorations of users’ interactions with search systems in
context and signicantly expanded the territory of IIR research. By carefully examining and evalu-
ating IIR user studies via the faceted approach, our hope is that the work presented here could add
more knowledge about the structure and the related impacts of the “decision in compromise” and
better facilitate the dialectic around IIR user study design, reporting, and evaluation.