This paper proposes a formal account of AI trustworthiness,connectingbothintrinsic and perceived trustworthinessin an operational schematization.We argue that trustworthiness extends beyond the inherent capabilities of an AI system to include significant influences from observers' perceptions, such as perceived transparency, agency locus, and human oversight. While the concept of perceived trustworthiness is discussed in the literature, few attempts have been made to connect it with the intrinsic trustworthiness of AI systems. Our analysis introduces a novel schematization to quantify trustworthiness by assessing the discrepancies between expected and observed behaviors and how these affect perceived uncertainty and trust.The paper provides a formalization for measuring trustworthiness, taking into account both perceived and intrinsic characteristics.By detailing the factors that influence trust, this study aims to foster more ethical and widely accepted AI technologies, ensuring they meet both functional and ethical criteria.
Bisconti, P., Aquilino, L., Marchetti, A., Nardi, D., A Formal Account of Trustworthiness: Connecting Intrinsic and Perceived Trustworthiness, Paper, in Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, (San José, Ca, 21-23 October 2024), AAAI Press, Washington D.C. 2024:7 131-140. 10.1609/aies.v7i1.31624 [https://hdl.handle.net/10807/311537]
A Formal Account of Trustworthiness: Connecting Intrinsic and Perceived Trustworthiness
Aquilino, Letizia;Marchetti, Antonella;
2024
Abstract
This paper proposes a formal account of AI trustworthiness,connectingbothintrinsic and perceived trustworthinessin an operational schematization.We argue that trustworthiness extends beyond the inherent capabilities of an AI system to include significant influences from observers' perceptions, such as perceived transparency, agency locus, and human oversight. While the concept of perceived trustworthiness is discussed in the literature, few attempts have been made to connect it with the intrinsic trustworthiness of AI systems. Our analysis introduces a novel schematization to quantify trustworthiness by assessing the discrepancies between expected and observed behaviors and how these affect perceived uncertainty and trust.The paper provides a formalization for measuring trustworthiness, taking into account both perceived and intrinsic characteristics.By detailing the factors that influence trust, this study aims to foster more ethical and widely accepted AI technologies, ensuring they meet both functional and ethical criteria.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.