TY - GEN
T1 - Designing explainable artificial intelligence with active inference
T2 - International Workshop on Active Inference (4th : 2023)
AU - Albarracin, Mahault
AU - Hipólito, Inês
AU - Tremblay, Safae Essafi
AU - Fox, Jason G.
AU - René, Gabriel
AU - Friston, Karl
AU - Ramstead, Maxwell J. D.
PY - 2024
Y1 - 2024
N2 - This paper investigates the prospect of developing human-interpretable, explainable artificial intelligence (AI) systems based on active inference and the free energy principle. We first provide a brief overview of active inference, and in particular, of how it applies to the modeling of decision-making, introspection, as well as the generation of overt and covert actions. We then discuss how active inference can be leveraged to design explainable AI systems, namely, by allowing us to model core features of “introspective” processes and by generating useful, human-interpretable models of the processes involved in decision-making. We propose an architecture for explainable AI systems using active inference. This architecture foregrounds the role of an explicit hierarchical generative model, the operation of which enables the AI system to track and explain the factors that contribute to its own decisions, and whose structure is designed to be interpretable and auditable by human users. We outline how this architecture can integrate diverse sources of information to make informed decisions in an auditable manner, mimicking or reproducing aspects of human-like consciousness and introspection. Finally, we discuss the implications of our findings for future research in AI, and the potential ethical considerations of developing AI systems with (the appearance of) introspective capabilities.
AB - This paper investigates the prospect of developing human-interpretable, explainable artificial intelligence (AI) systems based on active inference and the free energy principle. We first provide a brief overview of active inference, and in particular, of how it applies to the modeling of decision-making, introspection, as well as the generation of overt and covert actions. We then discuss how active inference can be leveraged to design explainable AI systems, namely, by allowing us to model core features of “introspective” processes and by generating useful, human-interpretable models of the processes involved in decision-making. We propose an architecture for explainable AI systems using active inference. This architecture foregrounds the role of an explicit hierarchical generative model, the operation of which enables the AI system to track and explain the factors that contribute to its own decisions, and whose structure is designed to be interpretable and auditable by human users. We outline how this architecture can integrate diverse sources of information to make informed decisions in an auditable manner, mimicking or reproducing aspects of human-like consciousness and introspection. Finally, we discuss the implications of our findings for future research in AI, and the potential ethical considerations of developing AI systems with (the appearance of) introspective capabilities.
KW - Active Inference
KW - Artificial intelligence
KW - Explainability
UR - http://www.scopus.com/inward/record.url?scp=85177864075&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-47958-8_9
DO - 10.1007/978-3-031-47958-8_9
M3 - Conference proceeding contribution
AN - SCOPUS:85177864075
SN - 9783031479571
T3 - Communications in Computer and Information Science
SP - 123
EP - 144
BT - Active inference
A2 - Buckley, Christopher L.
A2 - Cialfi, Daniela
A2 - Lanillos, Pablo
A2 - Ramstead, Maxwell
A2 - Sajid, Noor
A2 - Shimazaki, Hideaki
A2 - Verbelen, Tim
A2 - Wisse, Martijn
PB - Springer, Springer Nature
CY - Cham, Switzerland
Y2 - 13 September 2023 through 15 September 2023
ER -