Does your AI agent get you? A personalizable framework for approximating human models from argumentation-based dialogue traces

Abstract

Explainable AI is increasingly employing argumentation methods to facilitate interactive explanations between AI agents and human users. While existing approaches typically rely on predetermined human user models, there remains a critical gap in dynamically learning and updating these models during interactions. In this paper, we present a framework that enables AI agents to adapt their understanding of human users through argumentation-based dialogues. Our approach, called Persona, draws on prospect theory and integrates a probability weighting function with a Bayesian belief update mechanism that refines a probability distribution over possible human models based on exchanged arguments. Through empirical evaluations with human users in an applied argumentation setting, we demonstrate that Persona effectively captures evolving human beliefs, facilitates personalized interactions, and outperforms state-of-the-art methods.

Publication
In AAAI Conference on Artificial Intelligence
Yinxu Tang
Yinxu Tang
PhD Candidate at Washington University in St. Louis, USA

My research interests include Explainable AI, Artificial Intelligence, Bandits and Reinforcement Learning.