zouharvi commited on
Commit
380f361
1 Parent(s): 53786ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md CHANGED
@@ -1,3 +1,63 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ task_categories:
4
+ - tabular-regression
5
+ - tabular-classification
6
+ - question-answering
7
+ language:
8
+ - en
9
+ tags:
10
+ - user modelling
11
+ - trust
12
+ size_categories:
13
+ - 10K<n<100K
14
  ---
15
+
16
+ This is a slightly edited dataset of the one [found here on GitHub](https://github.com/zouharvi/trust-intervention/).
17
+ The data contains the user interactions, their bet values, answer correctness etc.
18
+ Please contact the authors if you have any questions.
19
+
20
+ # A Diachronic Perspective on User Trust in AI under Uncertainty
21
+
22
+ > **Abstract:** In a human-AI collaboration, users build a mental model of the AI system based on its veracity and how it presents its decision, e.g. its presentation of system confidence and an explanation of the output.
23
+ > However, modern NLP systems are often uncalibrated, resulting in confidently incorrect predictions that undermine user trust.
24
+ > In order to build trustworthy AI, we must understand how user trust is developed and how it can be regained after potential trust-eroding events.
25
+ > We study the evolution of user trust in response to these trust-eroding events using a betting game as the users interact with the AI.
26
+ > We find that even a few incorrect instances with inaccurate confidence estimates can substantially damage user trust and performance, with very slow recovery.
27
+ > We also show that this degradation in trust can reduce the success of human-AI collaboration
28
+ > and that different types of miscalibration---unconfidently correct and confidently incorrect---have different (negative) effects on user trust.
29
+ > Our findings highlight the importance of calibration in user-facing AI application, and shed light onto what aspects help users decide whether to trust the system.
30
+
31
+ This work was presented EMNLP 2023, read it [**here**](https://aclanthology.org/2023.emnlp-main.339/).
32
+ Written by Shehzaad Dhuliawala, Vilém Zouhar, Mennatallah El-Assady, and Mrinmaya Sachan from ETH Zurich, Department of Computer Science.
33
+ ```
34
+ @inproceedings{dhuliawala-etal-2023-diachronic,
35
+ title = "A Diachronic Perspective on User Trust in {AI} under Uncertainty",
36
+ author = "Dhuliawala, Shehzaad and
37
+ Zouhar, Vil{\'e}m and
38
+ El-Assady, Mennatallah and
39
+ Sachan, Mrinmaya",
40
+ booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
41
+ month = dec,
42
+ year = "2023",
43
+ address = "Singapore",
44
+ publisher = "Association for Computational Linguistics",
45
+ url = "https://aclanthology.org/2023.emnlp-main.339",
46
+ doi = "10.18653/v1/2023.emnlp-main.339",
47
+ pages = "5567--5580"
48
+ }
49
+ ```
50
+
51
+ <img width="400em" src="https://raw.githubusercontent.com/zouharvi/trust-intervention/main/meta/figure_1.png">
52
+
53
+ <small>
54
+ Figure 1: Diachronic view of a typical human-AI collaborative setting.
55
+ Here, at each timestep <em>t</em>, the user uses their prior mental model <em>ψ<sub>t</sub></em> to accept or reject the AI system’s answer <em>y<sub>t</sub></em>, supported by an additional message <em>m<sub>t</sub></em> comprising of the AI’s confidence, and updates their mental model of the AI system to <em>ψ<sub>t+1</sub></em>. If the message is rejected, the user invokes a fallback process to provide a different answer.
56
+ </small>
57
+
58
+ ## Resources
59
+
60
+ [![Paper video presentation](https://img.youtube.com/vi/NrH3flpijDw/0.jpg)](https://www.youtube.com/watch?v=NrH3flpijDw)
61
+
62
+
63
+ <img width="500em" src="https://raw.githubusercontent.com/zouharvi/trust-intervention/main/meta/poster.png">