Asking for a Friend: Evaluating Response Biases in Security User Studies
Title | Asking for a Friend: Evaluating Response Biases in Security User Studies |
Publication Type | Conference Paper |
Year of Publication | 2018 |
Authors | Redmiles, Elissa M., Zhu, Ziyun, Kross, Sean, Kuchhal, Dhruv, Dumitras, Tudor, Mazurek, Michelle L. |
Conference Name | Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security |
Publisher | ACM |
Conference Location | New York, NY, USA |
ISBN Number | 978-1-4503-5693-0 |
Keywords | Collaboration, comparability, data-driven security, Human Behavior, Metrics, pubcrawl, Resiliency, Scalability, science of security, usable security |
Abstract | The security field relies on user studies, often including survey questions, to query end users' general security behavior and experiences, or hypothetical responses to new messages or tools. Self-report data has many benefits - ease of collection, control, and depth of understanding - but also many well-known biases stemming from people's difficulty remembering prior events or predicting how they might behave, as well as their tendency to shape their answers to a perceived audience. Prior work in fields like public health has focused on measuring these biases and developing effective mitigations; however, there is limited evidence as to whether and how these biases and mitigations apply specifically in a computer-security context. In this work, we systematically compare real-world measurement data to survey results, focusing on an exemplar, well-studied security behavior: software updating. We align field measurements about specific software updates (n=517,932) with survey results in which participants respond to the update messages that were used when those versions were released (n=2,092). This allows us to examine differences in self-reported and observed update speeds, as well as examining self-reported responses to particular message features that may correlate with these results. The results indicate that for the most part, self-reported data varies consistently and systematically with measured data. However, this systematic relationship breaks down when survey respondents are required to notice and act on minor details of experimental manipulations. Our results suggest that many insights from self-report security data can, when used with care, translate to real-world environments; however, insights about specific variations in message texts or other details may be more difficult to assess with surveys. |
URL | https://dl.acm.org/citation.cfm?doid=3243734.3243740 |
DOI | 10.1145/3243734.3243740 |
Citation Key | redmiles_asking_2018 |