TY - GEN
T1 - How Does Predictive Information Affect Human Ethical Preferences?
AU - Narayanan, Saumik
AU - Yu, Guanghui
AU - Tang, Wei
AU - Ho, Chien Ju
AU - Yin, Ming
N1 - Publisher Copyright:
© 2022 Owner/Author.
PY - 2022/7/26
Y1 - 2022/7/26
N2 - Artificial intelligence (AI) has been increasingly involved in decision making in high-stakes domains, including loan applications, employment screening, and assistive clinical decision making. Meanwhile, involving AI in these high-stake decisions has created ethical concerns on how to balance different trade-offs to respect human values. One approach for aligning AIs with human values is to elicit human ethical preferences and incorporate this information in the design of computer systems. In this work, we explore how human ethical preferences are impacted by the information shown to humans during elicitation. In particular, we aim to provide a contrast between verifiable information (e.g., patient demographics or blood test results) and predictive information (e.g., the probability of organ transplant success). Using kidney transplant allocation as a case study, we conduct a randomized experiment to elicit human ethical preferences on scarce resource allocation to understand how human ethical preferences are impacted by the verifiable and predictive information. We find that the presence of predictive information significantly changes how humans take into account other verifiable information in their ethical preferences. We also find that the source of the predictive information (e.g., whether the predictions are made by AI or human doctors) plays a key role in how humans incorporate the predictive information into their own ethical judgements.
AB - Artificial intelligence (AI) has been increasingly involved in decision making in high-stakes domains, including loan applications, employment screening, and assistive clinical decision making. Meanwhile, involving AI in these high-stake decisions has created ethical concerns on how to balance different trade-offs to respect human values. One approach for aligning AIs with human values is to elicit human ethical preferences and incorporate this information in the design of computer systems. In this work, we explore how human ethical preferences are impacted by the information shown to humans during elicitation. In particular, we aim to provide a contrast between verifiable information (e.g., patient demographics or blood test results) and predictive information (e.g., the probability of organ transplant success). Using kidney transplant allocation as a case study, we conduct a randomized experiment to elicit human ethical preferences on scarce resource allocation to understand how human ethical preferences are impacted by the verifiable and predictive information. We find that the presence of predictive information significantly changes how humans take into account other verifiable information in their ethical preferences. We also find that the source of the predictive information (e.g., whether the predictions are made by AI or human doctors) plays a key role in how humans incorporate the predictive information into their own ethical judgements.
KW - ethical preference, ai ethics
KW - preference elicitation
UR - https://www.scopus.com/pages/publications/85137160902
U2 - 10.1145/3514094.3534165
DO - 10.1145/3514094.3534165
M3 - Conference contribution
AN - SCOPUS:85137160902
T3 - AIES 2022 - Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society
SP - 508
EP - 517
BT - AIES 2022 - Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society
PB - Association for Computing Machinery, Inc
T2 - 5th AAAI/ACM Conference on Artificial Intelligence, Ethics, and Society, AIES 2022
Y2 - 1 August 2022 through 3 August 2022
ER -