Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,63 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
task_categories:
|
4 |
+
- tabular-regression
|
5 |
+
- tabular-classification
|
6 |
+
- question-answering
|
7 |
+
language:
|
8 |
+
- en
|
9 |
+
tags:
|
10 |
+
- user modelling
|
11 |
+
- trust
|
12 |
+
size_categories:
|
13 |
+
- 10K<n<100K
|
14 |
---
|
15 |
+
|
16 |
+
This is a slightly edited dataset of the one [found here on GitHub](https://github.com/zouharvi/trust-intervention/).
|
17 |
+
The data contains the user interactions, their bet values, answer correctness etc.
|
18 |
+
Please contact the authors if you have any questions.
|
19 |
+
|
20 |
+
# A Diachronic Perspective on User Trust in AI under Uncertainty
|
21 |
+
|
22 |
+
> **Abstract:** In a human-AI collaboration, users build a mental model of the AI system based on its veracity and how it presents its decision, e.g. its presentation of system confidence and an explanation of the output.
|
23 |
+
> However, modern NLP systems are often uncalibrated, resulting in confidently incorrect predictions that undermine user trust.
|
24 |
+
> In order to build trustworthy AI, we must understand how user trust is developed and how it can be regained after potential trust-eroding events.
|
25 |
+
> We study the evolution of user trust in response to these trust-eroding events using a betting game as the users interact with the AI.
|
26 |
+
> We find that even a few incorrect instances with inaccurate confidence estimates can substantially damage user trust and performance, with very slow recovery.
|
27 |
+
> We also show that this degradation in trust can reduce the success of human-AI collaboration
|
28 |
+
> and that different types of miscalibration---unconfidently correct and confidently incorrect---have different (negative) effects on user trust.
|
29 |
+
> Our findings highlight the importance of calibration in user-facing AI application, and shed light onto what aspects help users decide whether to trust the system.
|
30 |
+
|
31 |
+
This work was presented EMNLP 2023, read it [**here**](https://aclanthology.org/2023.emnlp-main.339/).
|
32 |
+
Written by Shehzaad Dhuliawala, Vilém Zouhar, Mennatallah El-Assady, and Mrinmaya Sachan from ETH Zurich, Department of Computer Science.
|
33 |
+
```
|
34 |
+
@inproceedings{dhuliawala-etal-2023-diachronic,
|
35 |
+
title = "A Diachronic Perspective on User Trust in {AI} under Uncertainty",
|
36 |
+
author = "Dhuliawala, Shehzaad and
|
37 |
+
Zouhar, Vil{\'e}m and
|
38 |
+
El-Assady, Mennatallah and
|
39 |
+
Sachan, Mrinmaya",
|
40 |
+
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
|
41 |
+
month = dec,
|
42 |
+
year = "2023",
|
43 |
+
address = "Singapore",
|
44 |
+
publisher = "Association for Computational Linguistics",
|
45 |
+
url = "https://aclanthology.org/2023.emnlp-main.339",
|
46 |
+
doi = "10.18653/v1/2023.emnlp-main.339",
|
47 |
+
pages = "5567--5580"
|
48 |
+
}
|
49 |
+
```
|
50 |
+
|
51 |
+
<img width="400em" src="https://raw.githubusercontent.com/zouharvi/trust-intervention/main/meta/figure_1.png">
|
52 |
+
|
53 |
+
<small>
|
54 |
+
Figure 1: Diachronic view of a typical human-AI collaborative setting.
|
55 |
+
Here, at each timestep <em>t</em>, the user uses their prior mental model <em>ψ<sub>t</sub></em> to accept or reject the AI system’s answer <em>y<sub>t</sub></em>, supported by an additional message <em>m<sub>t</sub></em> comprising of the AI’s confidence, and updates their mental model of the AI system to <em>ψ<sub>t+1</sub></em>. If the message is rejected, the user invokes a fallback process to provide a different answer.
|
56 |
+
</small>
|
57 |
+
|
58 |
+
## Resources
|
59 |
+
|
60 |
+
[![Paper video presentation](https://img.youtube.com/vi/NrH3flpijDw/0.jpg)](https://www.youtube.com/watch?v=NrH3flpijDw)
|
61 |
+
|
62 |
+
|
63 |
+
<img width="500em" src="https://raw.githubusercontent.com/zouharvi/trust-intervention/main/meta/poster.png">
|