Improving Trust Estimation in Human-Robot Collaboration Using Beta Reputation at Fine-grained Timescales

๐Ÿ“… 2024-11-04
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Fine-grained, real-time trust estimation remains challenging in human-robot collaboration due to the coarse, binary nature of conventional task-outcome feedback. Method: This paper proposes a continuous-reward-driven Beta reputation model that performs online Bayesian inference using per-step scalar reward signalsโ€”enabling millisecond-level dynamic trust modeling. It uniquely integrates maximum-entropy reinforcement learning to automatically synthesize task-adaptive reward functions, eliminating reliance on handcrafted metrics. Contribution/Results: Experiments demonstrate significant improvements in both estimation accuracy and temporal responsiveness over baseline methods. The framework is validated across multiple benchmark human-robot collaboration tasks, confirming its generalizability and effectiveness. The implementation is publicly available.

Technology Category

Application Category

๐Ÿ“ Abstract
When interacting with each other, humans adjust their behavior based on perceived trust. However, to achieve similar adaptability, robots must accurately estimate human trust at sufficiently granular timescales during the human-robot collaboration task. A beta reputation is a popular way to formalize a mathematical estimation of human trust. However, it relies on binary performance, which updates trust estimations only after each task concludes. Additionally, manually crafting a reward function is the usual method of building a performance indicator, which is labor-intensive and time-consuming. These limitations prevent efficiently capturing continuous changes in trust at more granular timescales throughout the collaboration task. Therefore, this paper presents a new framework for the estimation of human trust using a beta reputation at fine-grained timescales. To achieve granularity in beta reputation, we utilize continuous reward values to update trust estimations at each timestep of a task. We construct a continuous reward function using maximum entropy optimization to eliminate the need for the laborious specification of a performance indicator. The proposed framework improves trust estimations by increasing accuracy, eliminating the need for manually crafting a reward function, and advancing toward developing more intelligent robots. The source code is publicly available. https://github.com/resuldagdanov/robot-learning-human-trust
Problem

Research questions and friction points this paper is trying to address.

Estimating human trust in robots at fine-grained timescales
Eliminating manual crafting of reward functions for trust estimation
Improving accuracy of trust updates using continuous reward values
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses beta reputation for fine-grained trust estimation
Employs continuous reward values per timestep
Automates reward function via entropy optimization
๐Ÿ”Ž Similar Papers
No similar papers found.
R
Resul Dagdanov
Robotics Institute, Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia
M
Milan Andrejevic
Psychology Discipline, Graduate School of Health, Faculty of Health, University of Technology Sydney, Ultimo, NSW 2007, Australia
Dikai Liu
Dikai Liu
University of Technology Sydney
Field Roboticsinfrastructure roboticshuman-robot collaboration
Chin-Teng Lin
Chin-Teng Lin
University of Technology Sydney
Computational intelligencemachine learningfuzzy neural networks (FNN)cognitive neuro-engineeringbrain-computer interface