TR2024-135
Open Human-Robot Collaboration using Decentralized Inverse Reinforcement Learning
-
- "Open Human-Robot Collaboration using Decentralized Inverse Reinforcement Learning", 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2024), October 2024.BibTeX TR2024-135 PDF
- @inproceedings{Suresh2024oct,
- author = {Suresh, Prasanth and Jain, Siddarth and Doshi, Prashant and Romeres, Diego}},
- title = {Open Human-Robot Collaboration using Decentralized Inverse Reinforcement Learning},
- booktitle = {2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2024)},
- year = 2024,
- month = oct,
- url = {https://www.merl.com/publications/TR2024-135}
- }
,
- "Open Human-Robot Collaboration using Decentralized Inverse Reinforcement Learning", 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2024), October 2024.
-
MERL Contacts:
-
Research Area:
Abstract:
The growing interest in human-robot collaboration (HRC), where humans and robots cooperate towards shared goals, has seen significant advancements over the past decade. While previous research has addressed various challenges, several key issues remain unresolved. Many domains within HRC involve activities that do not necessarily require human presence throughout the entire task. Existing literature typically models HRC as a closed system, where all agents are present for the entire duration of the task. In contrast, an open model offers flexibility by allowing an agent to enter and exit the collaboration as needed, enabling them to concurrently manage other tasks. In this paper, we introduce a novel multiagent framework called oDec-MDP, designed specifically to model open HRC scenarios where agents can join or leave tasks flexibly during execution. We generalize a recent multiagent inverse reinforcement learning method - Dec-AIRL to learn from open systems modeled using the oDec-MDP. Our method is validated through experiments conducted in both a simplified toy firefighting domain and a realistic dyadic human-robot collaborative assembly. Results show that our framework and learning method improves upon its closed system counterpart.