This project is a multi-institution collaboration led by Prashant Doshi of the University of Georgia, Leen-Kiat Soh of the University of Nebraska-Lincoln, and Adam Eck of Oberlin College. Our work has been supported by two collaborative grants from the National Science Foundation: NSF IIS-1910037, IIS-1910156, IIS-1909513 and NSF IIS-2312657, IIS-2312658, IIS-2312659.




Project Description

In many real-world applications of AI, the set of actors and tasks are not constant, but instead change over time. Robots tasked with suppressing wildfires eventually run out of limited suppressant resources and need to temporarily disengage from the collaborative work in order to recharge, or they might become damaged and leave the environment permanently. In a large business organization, objectives and goals change with the market, requiring workers to adapt to perform different sets of tasks across time. We call these multiagent systems open agent systems, and the openness of the sets of agents and tasks necessitates new capabilities and modeling for decision making compared to planning and learning in closed environments.

In this research, we address the challenges to multiagent decision making caused by three types of openness:

  1. agent openness, where the set of agents acting in the world change over time. This could include agents temporarily disengaging from the environment before returning (e.g., autonomous ride-sharing cars recharging when low on energy), new groups of agents joining over time (e.g., new attackers appearing globally in a cybersecurity defense application), or existing agents leaving permanently (e.g., firefighting robots becoming damaged and needing to leave the operations to recover valuable hardware).

  2. task openness, where the set of tasks that agents aim to accomplish change over time. This could include new tasks appearing that are novel compared to existing tasks (e.g., transporting ride-sharing passengers for unique events) or popular existing tasks disappearing forever. It could also entail a gradual shift in the requirements of tasks over time so that tasks gradually become different from their initial context.

  3. type openness, where the types (i.e., capabilities) of agents change over time. This could include agents learning new skills and gaining new responsibilities (e.g., promotions of office workers) or robots losing abilities over time (e.g., modeling damage to robots engaged in the field).

The following video provides a short overview of the problems, challenges, and solutions related to OASYS:





Publications

A summary of the research area of OASYS can be found in:

  • Eck, A., Soh, L.-K., & Doshi, P. 2023. Decision Making in Open Multiagent Systems. AI Magazine. 44(4), 508-523. [link]

More details about our research on OASYS can be found in:

  • Aditya Shinde, Prashant Doshi. 2025. “Inferring Hidden Behavioral Signatures of Cyber Adversaries Using Inverse Reinforcement Learning”, Proceedings of the 2025 Prestigious Applications of Intelligent Systems (PAIS)/ECAI [link]

  • Tyson Jordan, Pranav Pandey, Prashant Doshi, Ramviyas Parasuraman, Adam Goodie. 2025. “Analyzing Human Perceptions of a MEDEVAC Robot in a Simulated Evacuation Scenario”, Proceedings of the 2025 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) [link]

  • Pranav Pandey, Ramviyas Parasuraman, Prashant Doshi. 2025. “Integrating Perceptions: A Human-Centered Physical Safety Model for Human-Robot Interaction”, Proceedings of the 2025 IEEE Robot-Human Interaction (RO-MAN) [link]

  • Yanyu Liu, Yinghui Pan, Yifeng Zeng, Biyang Ma, Prashant Doshi. 2025. “Active Legibility in Multiagent Reinforcement Learning”, in Artificial Intelligence Journal (AIJ), Elsevier [link]

  • Gayathri Anil, Prashant Doshi, Daniel Redder, Adam Eck, Leen-Kiat Soh. 2025. “MOHITO: Multi-Agent Reinforcement Learning using Hypergraphs for Task-Open Systems”, Proceedings of the 2025 Conference on Uncertainty in AI (UAI)[link]

  • Prasanth Sengadu Suresh, Prashant Doshi, Bikramjit Banerjee. 2025. “Adaptive Human-Robot Collaboration using Type-Based IRL”, Proceedings of the 2025 Conference on Uncertainty in AI (UAI) [link]

  • Bhavana Nare, Anusha Challa, John Bradley Frericks, Prashant Doshi, Kyle Johnsen. 2025. “A Novel Computational Framework of Robot Trust for Human-Robot Teams”, Proceedings of the 2025 IEEE International Conference on Robotics and Automation (ICRA) [link]

  • Yikang Gui, Prashant Doshi. 2025. “Inversely Learning Transferable Rewards via Abstracted States”. arXiv:2501.01669 [link]

  • He, K., Doshi, P., & Banerjee, B. 2024. Robust Individualistic Learning in Many-Agent Systems. Proceedings of the 25th International Conference on Principles and Practice of Multi-Agent Systems (PRIMA’2024) [link]

  • Suresh, P.S., Romeres, D., Doshi, D., & Jain, S. 2024. Open Human-Robot Collaboration Systems (OHRCS): A Research Perspective. Proceedings of the 2024 IEEE Conference on Cognition and Machine Intelligence (CogMI’2024), Blue Sky/Vision Track. [link]

  • Suresh, P.S., Jain, S., Doshi, P., & Romeres, D. 2024. Open Human-Robot Collaborations using Decentralized Inverse Reinforcement Learning. Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS’2024). [link]

  • Kakarlapudi, A., Anil, G., Eck, A., Doshi, P., & Soh, L.-K. 2022. Decision-Theoretic Planning with Communication in Open Multiagent Systems. Proceedings of the 2022 Conference on Uncertainty in Artificial Intelligence (UAI’22), Eindhoven, Netherlands, August 1-5, 2022 [link] [Open Review with Appendices] [Code]

  • Eck, A., Shah, M., Doshi, P., & Soh, L.-K. 2020. Scalable Decision-Theoretic Planning in Open and Typed Multiagent Systems. Proceedings of the Thirty-fourth AAAI Conference on Artificial Intelligence (AAAI’2020), New York City, NY, February 8-12, 2020. [link] [Preprint with Appendices] [Code]

  • Chandrasekaran, M., Eck, A., Doshi, P., & Soh, L.-K. 2016. Individual Planning in Open and Typed Agent Systems. Proceedings of the 2016 Conference on Uncertainty in Artificial Intelligence (UAI’16), New York City, NY, June 25-29, 2016. [link]