skip to main content
10.1145/2858036.2858402acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
note

How Much Information?: Effects of Transparency on Trust in an Algorithmic Interface

Published: 07 May 2016 Publication History

Abstract

The rising prevalence of algorithmic interfaces, such as curated feeds in online news, raises new questions for designers, scholars, and critics of media. This work focuses on how transparent design of algorithmic interfaces can promote awareness and foster trust. A two-stage process of how transparency affects trust was hypothesized drawing on theories of information processing and procedural justice. In an online field experiment, three levels of system transparency were tested in the high-stakes context of peer assessment. Individuals whose expectations were violated (by receiving a lower grade than expected) trusted the system less, unless the grading algorithm was made more transparent through explanation. However, providing too much information eroded this trust. Attitudes of individuals whose expectations were met did not vary with transparency. Results are discussed in terms of a dual process model of attitude change and the depth of justification of perceived inconsistency. Designing for trust requires balanced interface transparency - not too little and not too much.

References

[1]
Frank Bannister and Regina Connolly. 2011. The Trouble with Transparency: A Critical Review of Openness in e-Government. Policy & Internet 3, 1 (2011), 1--30.
[2]
Bruce G Buchanan, Edward Hance Shortliffe, and others. 1984. Rule-based expert systems. Vol. 3. Addison-Wesley Reading, MA.
[3]
Judee K Burgoon and Jerold L Hale. 1988. Nonverbal expectancy violations: Model elaboration and application to immediacy behaviors. Communications Monographs 55, 1 (1988), 58--79.
[4]
Karen S Cook, Chris Snijders, Vincent Buskens, and Coye Cheshire. 2009. eTrust: Forming relationships in the online world. Russell Sage Foundation.
[5]
Cynthia L Corritore, Beverly Kracher, and Susan Wiedenbeck. 2003. On-line trust: concepts, evolving themes, a model. International Journal of Human-Computer Studies 58, 6 (2003), 737--758.
[6]
Henriette Cramer, Vanessa Evers, Satyan Ramlal, Maarten Van Someren, Lloyd Rutledge, Natalia Stash, Lora Aroyo, and Bob Wielinga. 2008. The effects of transparency on trust in and acceptance of a content-based art recommender. User Modeling and User-Adapted Interaction 18, 5 (2008), 455--496.
[7]
Motahhare Eslami, Aimee Rickman, Kristen Vaccaro, Amirhossein Aleyasen, Andy Vuong, Karrie Karahalios, Kevin Hamilton, and Christian Sandvig. 2015. "I always assumed that I wasn't really that close to {her}": Reasoning about invisible algorithms in the news feed. In Proceedings of the 33rd Annual SIGCHI Conference on Human Factors in Computing Systems. 153--162.
[8]
Alyssa Glass, Deborah L McGuinness, and Michael Wolverton. 2008. Toward establishing trust in adaptive agents. In Proceedings of the 13th international conference on Intelligent user interfaces. ACM, 227--236.
[9]
Kevin Hamilton, Karrie Karahalios, Christian Sandvig, and Motahhare Eslami. 2014. A path to understanding the effects of algorithm awareness. In CHI'14 Extended Abstracts on Human Factors in Computing Systems. ACM, 631--642.
[10]
Jonathan L Herlocker, Joseph A Konstan, and John Riedl. 2000. Explaining collaborative filtering recommendations. In Proceedings of the 2000 ACM conference on Computer supported cooperative work. ACM, 241--250.
[11]
Hilary Johnson and Peter Johnson. 1993. Explanation facilities and interactive systems. In Proceedings of the 1st international conference on Intelligent user interfaces. ACM, 159--166.
[12]
Julia H Kaufman and Christian D Schunn. 2011. Students' perceptions about peer assessment for writing: their origin and impact on revision work. Instructional Science 39, 3 (2011), 387--406.
[13]
Jurgen Koenemann and Nicholas J Belkin. 1996. A case for interaction: a study of interactive information retrieval behavior and effectiveness. In Proceedings of the SIGCHI conference on Human factors in computing systems. ACM, 205--212.
[14]
Chinmay Kulkarni, Koh Pang Wei, Huy Le, Daniel Chia, Kathryn Papadopoulos, Justin Cheng, Daphne Koller, and Scott R Klemmer. 2013. Peer and self assessment in massive online classes. ACM Transactions on Computer-Human Interaction (TOCHI) 20, 6 (2013), 33.
[15]
E Allan Lind and Tom R Tyler. 1988. The Social Psychology of Procedural Justice. Springer Science & Business Media.
[16]
Jack Muramatsu and Wanda Pratt. 2001. Transparent Queries: investigation users' mental models of search engines. In Proceedings of the 24th annual international ACM SIGIR conference on Research and development in information retrieval. ACM, 217--224.
[17]
Stephen J Payne. 2003. Users' mental models: the very ideas. HCI models, theories, and frameworks: Toward a multidisciplinary science (2003), 135--156.
[18]
Richard E Petty and John T Cacioppo. 1986. The elaboration likelihood model of persuasion. Springer.
[19]
Chris Piech, Jon Huang, Zhenghao Chen, Chuong Do, Andrew Ng, and Daphne Koller. 2013. Tuned Models of Peer Assessment in MOOCs. In Proceedings of the 6th International Conference on Educational Data Mining.
[20]
Wolter Pieters. 2011. Explanation and trust: what to tell the user in security and AI? Ethics and information technology 13, 1 (2011), 53--64.
[21]
Pearl Pu and Li Chen. 2007. Trust-inspiring explanation interfaces for recommender systems. Knowledge-Based Systems 20, 6 (2007), 542--556.
[22]
Emilee Rader and Rebecca Gray. 2015. Understanding User Beliefs About Algorithmic Curation in the Facebook News Feed. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. ACM, 173--182.
[23]
Byron Reeves and Clifford Nass. 1996. How people treat computers, television, and new media like real people and places. CSLI Publications and Cambridge university press.
[24]
Jens Riegelsberger, M Angela Sasse, and John D McCarthy. 2005. The mechanics of trust: A framework for research and design. International Journal of Human-Computer Studies 62, 3 (2005), 381--422.
[25]
Philip M Sadler and Eddie Good. 2006. The impact of self-and peer-grading on student learning. Educational assessment 11, 1 (2006), 1--31.
[26]
Kay Sambell, Liz McDowell, and Sally Brown. 1997. "But is it fair?": an exploratory study of student perceptions of the consequential validity of assessment. Studies in Educational Evaluation 23, 4 (1997), 349--371.
[27]
Mien Segers and Filip Dochy. 2001. New assessment forms in problem-based learning: the value-added of the students' perspective. Studies in higher education 26, 3 (2001), 327--343.
[28]
David K Sherman and Geoffrey L Cohen. 2002. Accepting threatening information: Self-Affirmation and the reduction of defensive biases. Current Directions in Psychological Science 11, 4 (2002), 119--123.
[29]
Rashmi Sinha and Kirsten Swearingen. 2002. The role of transparency in recommender systems. In CHI'02 extended abstracts on Human factors in computing systems. ACM, 830--831.
[30]
Candace Thille, Emily Schneider, Rene F Kizilcec, Christopher Piech, Sherif A Halawa, and Daniel K Greene. 2014. The future of data-enriched assessment. Research & Practice in Assessment 9, 2 (2014), 5--16.
[31]
Weiquan Wang and Izak Benbasat. 2007. Recommendation agents for electronic commerce: Effects of explanation facilities on trusting beliefs. Journal of Management Information Systems 23, 4 (2007), 217--246.
[32]
Meichun Lydia Wen and Chin-Chung Tsai. 2006. University students' perceptions of and attitudes toward (online) peer assessment. Higher Education 51, 1 (2006), 27--44.

Cited By

View all
  • (2024)Artificial Intelligence Alone Will Not Democratise Education: On Educational Inequality, Techno-Solutionism and Inclusive ToolsSustainability10.3390/su1602078116:2(781)Online publication date: 16-Jan-2024
  • (2024)In Seal We Trust? Investigating the Effect of Certifications on Perceived Trustworthiness of AI SystemsHuman-Machine Communication10.30658/hmc.8.78(141-162)Online publication date: 2024
  • (2024)Bridging the Human-Automation Fairness Gap: How Providing Reasons Enhances the Perceived Fairness of Public Decision-MakingSSRN Electronic Journal10.2139/ssrn.4819145Online publication date: 2024
  • Show More Cited By

Index Terms

  1. How Much Information?: Effects of Transparency on Trust in an Algorithmic Interface

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      CHI '16: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems
      May 2016
      6108 pages
      ISBN:9781450333627
      DOI:10.1145/2858036
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 07 May 2016

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. algorithm awareness
      2. attitude change
      3. interface design
      4. peer assessment
      5. transparency
      6. trust

      Qualifiers

      • Note

      Conference

      CHI'16
      Sponsor:
      CHI'16: CHI Conference on Human Factors in Computing Systems
      May 7 - 12, 2016
      California, San Jose, USA

      Acceptance Rates

      CHI '16 Paper Acceptance Rate 565 of 2,435 submissions, 23%;
      Overall Acceptance Rate 6,199 of 26,314 submissions, 24%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)754
      • Downloads (Last 6 weeks)83
      Reflects downloads up to 07 Sep 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Artificial Intelligence Alone Will Not Democratise Education: On Educational Inequality, Techno-Solutionism and Inclusive ToolsSustainability10.3390/su1602078116:2(781)Online publication date: 16-Jan-2024
      • (2024)In Seal We Trust? Investigating the Effect of Certifications on Perceived Trustworthiness of AI SystemsHuman-Machine Communication10.30658/hmc.8.78(141-162)Online publication date: 2024
      • (2024)Bridging the Human-Automation Fairness Gap: How Providing Reasons Enhances the Perceived Fairness of Public Decision-MakingSSRN Electronic Journal10.2139/ssrn.4819145Online publication date: 2024
      • (2024)Effects of machine learning errors on human decision-making: manipulations of model accuracy, error types, and error importanceCognitive Research: Principles and Implications10.1186/s41235-024-00586-29:1Online publication date: 26-Aug-2024
      • (2024)When Stories Turn Institutional: How TikTok Users Legitimate the Algorithmic SensemakingSocial Media + Society10.1177/2056305123122411410:1Online publication date: 30-Jan-2024
      • (2024)Finding Paths for Explainable MOOC Recommendation: A Learner PerspectiveProceedings of the 14th Learning Analytics and Knowledge Conference10.1145/3636555.3636898(426-437)Online publication date: 18-Mar-2024
      • (2024)IT Higher Education Teachers and Trust in AI-Enabled Ed-Tech: Implications for Adoption of AI in Higher EducationProceedings of the 2024 Computers and People Research Conference10.1145/3632634.3655852(1-16)Online publication date: 29-May-2024
      • (2024)Content Moderation, Platformised Speech Governance, and Legitimacy: TikTok in South and Southeast AsiaCompanion Publication of the 16th ACM Web Science Conference10.1145/3630744.3658613(27-31)Online publication date: 21-May-2024
      • (2024)A Critical Survey on Fairness Benefits of Explainable AIProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658990(1579-1595)Online publication date: 3-Jun-2024
      • (2024)Meaningful Transparency for Clinicians: Operationalising HCXAI Research with GynaecologistsProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658971(1268-1281)Online publication date: 3-Jun-2024
      • Show More Cited By

      View Options

      Get Access

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media