TY - GEN
T1 - Cira: An architecture for building configurable immersive smart-rooms
AU - Divekar, Rahul
AU - Peveler, Matthew
AU - Rouhani, Robert
AU - Zhao, Rui
AU - Kephart, Jeffrey O.
AU - Allen, David
AU - Wang, Kang
AU - Ji, Qiang
AU - Su, Hui
PY - 2018
Y1 - 2018
N2 - . Significant technological advancements in computer vision, natural language processing and other fields of computer science and engineering have made way for us to build new human-computer interfaces like a Cognitive Immersive Room (CIR). The CIR described in this work can “see”, “hear”, interpret and respond appropriately to a user or a group of users occupying it. However, bringing the state-of-the-art technologies together to respond to different levels of interpretations of multiple input sources under a unified system that can be easily reconfigured to suit different purposes is a challenge. The possible use-cases and benefits of such a smart-room are exciting and limitless. To address this sky-high vision of a smart-room, we describe a software architecture, Cognitive Immersive Room Architecture (CIRA) that allows the room to be programmed for multiple use-cases by providing them spatial and contextual intelligence. In this project, given the relationship of the software with the physical attributes of a room, reconfigurability to various physical environments must be kept in mind. By providing resilience to changes in the lower-level devices, the architecture allows flexibility for researchers to enable different multi-modal interactions. The architecture also supports use-case and context-nesting. Thus, this work describes an architecture that enables a smart-room with multi-modal interaction, interpretation and reasoning. The CIRA supports nested contexts, multiple languages and multiple physical environments. We describe four such contexts (use-cases), two languages and two physical environments prototyped using this architecture and describe a pilot user-study.
AB - . Significant technological advancements in computer vision, natural language processing and other fields of computer science and engineering have made way for us to build new human-computer interfaces like a Cognitive Immersive Room (CIR). The CIR described in this work can “see”, “hear”, interpret and respond appropriately to a user or a group of users occupying it. However, bringing the state-of-the-art technologies together to respond to different levels of interpretations of multiple input sources under a unified system that can be easily reconfigured to suit different purposes is a challenge. The possible use-cases and benefits of such a smart-room are exciting and limitless. To address this sky-high vision of a smart-room, we describe a software architecture, Cognitive Immersive Room Architecture (CIRA) that allows the room to be programmed for multiple use-cases by providing them spatial and contextual intelligence. In this project, given the relationship of the software with the physical attributes of a room, reconfigurability to various physical environments must be kept in mind. By providing resilience to changes in the lower-level devices, the architecture allows flexibility for researchers to enable different multi-modal interactions. The architecture also supports use-case and context-nesting. Thus, this work describes an architecture that enables a smart-room with multi-modal interaction, interpretation and reasoning. The CIRA supports nested contexts, multiple languages and multiple physical environments. We describe four such contexts (use-cases), two languages and two physical environments prototyped using this architecture and describe a pilot user-study.
UR - https://dx.doi.org/10.1007/978-3-030-01057-7_7
U2 - 10.1007/978-3-030-01057-7_7
DO - 10.1007/978-3-030-01057-7_7
M3 - Conference contribution
BT - IEEE Intellisys 2018
PB - Springer Verlag
ER -