Embodied conversational AI agents in a multi-modal multi-agent competitive dialogue

  • Rahul Divekar
  • , Xiangyang Mou
  • , Lisha Chen
  • , Maira Gatti De Bayser
  • , Melina Alberio Guerra
  • , Hui Su
  • , Maira Bayser

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

In a setting where two AI agents embodied as animated humanoid avatars are engaged in a conversation with one human and each other, we see two challenges. One, determination by the AI agents about which one of them is being addressed. Two, determination by the AI agents if they may/could/should speak at the end of a turn. In this work we bring these two challenges together and explore the participation of AI agents in multiparty conversations. Particularly, we show two embodied AI shopkeeper agents who sell similar items aiming to get the business of a user by competing with each other on the price. In this scenario, we solve the first challenge by using headpose (estimated by deep learning techniques) to determine who the user is talking to. For the second challenge we use deontic logic to model rules of a negotiation conversation.
Original languageEnglish
Title of host publication28th International Joint Conference on Artificial Intelligence, IJCAI 2019
Volume2019-
DOIs
StatePublished - 2019

Fingerprint

Dive into the research topics of 'Embodied conversational AI agents in a multi-modal multi-agent competitive dialogue'. Together they form a unique fingerprint.

Cite this