AI assistance is impacting our decisions and the quality of our work. But how will this assistance affect us -- our skills, our growth and improvement, enjoyment, collaboration, or our agency in the workplace? The current design of AI assistance does not consider human-centric objectives; we need methods to account for them.

We propose offline RL as an approach for optimizing such human-centric objectives in AI-assisted decision-making.

Link to preprint: http://arxiv.org/abs/2403.05911

Towards Optimizing Human-Centric Objectives in AI-Assisted Decision-Making With Offline Reinforcement Learning

As AI assistance is increasingly infused into decision-making processes, we may seek to optimize human-centric objectives beyond decision accuracy, such as skill improvement or task enjoyment of individuals interacting with these systems. With this aspiration in mind, we propose offline reinforcement learning (RL) as a general approach for modeling human-AI decision-making to optimize such human-centric objectives. Our approach seeks to optimize different objectives by adaptively providing decision support to humans -- the right type of assistance, to the right person, at the right time. We instantiate our approach with two objectives: human-AI accuracy on the decision-making task and human learning about the task, and learn policies that optimize these two objectives from previous human-AI interaction data. We compare the optimized policies against various baselines in AI-assisted decision-making. Across two experiments (N = 316 and N = 964), our results consistently demonstrate that people interacting with policies optimized for accuracy achieve significantly better accuracy -- and even human-AI complementarity -- compared to those interacting with any other type of AI support. Our results further indicate that human learning is more difficult to optimize than accuracy, with participants who interacted with learning-optimized policies showing significant learning improvement only at times. Our research (1) demonstrates offline RL to be a promising approach to model dynamics of human-AI decision-making, leading to policies that may optimize various human-centric objectives and provide novel insights about the AI-assisted decision-making space, and (2) emphasizes the importance of considering human-centric objectives beyond decision accuracy in AI-assisted decision-making, while also opening up the novel research challenge of optimizing such objectives.

arXiv.org

Our approach optimizes different objectives by adaptively providing decision support to humans —- the right type of assistance, to the right person, at the right time.

For example, when optimizing accuracy, we may withhold AI support when the AI is uncertain so the human does not overrely on it. When seeking to optimize human skills, we may show only partial support so people cognitively engage with the AI support.

Or envisioning futures in which we seek to enhance collaboration and relatedness in the workplace, such dynamic assistance may even at appropriate times advise decision-makers to seek insights from and consult other colleagues for complex decision-making scenarios.

Here, we instantiated our approach with human-AI accuracy and human learning as two objectives.

Our results consistently show that people interacting with policies optimized for accuracy achieve significantly better accuracy and even human-AI complementarity in decision-making. Our results further indicate that human learning is more difficult to optimize than accuracy, with participants who interacted with learning-optimized policies showing significant learning improvement only at times.

Our research demonstrates (offline) RL to be a promising approach to model dynamics of human-AI decision-making, leading to policies that may optimize various human-centric objectives and provide novel insights about the AI-assisted decision-making space. Our work emphasizes the importance of considering human-centric objectives along with accuracy while also opening up the novel research challenge of devising human-AI interaction techniques and explanations that support such objectives.