LUI: A multimodal, intelligent interface for large displays

Vik Parthiban and Ashley Jieun Lee. 2019. "LUI: A multimodal, intelligent interface for large displays." In The 17th International Conference on Virtual-Reality Continuum and its Applications in Industry (VRCAI ’19). Association for Computing Machinery, New York, NY, USA, Article 48, 1–2. DOI:


On large screen displays, using conventional keyboard and mouse input is difficult because small mouse movements often do not scale well with the size of the display and individual elements on screen. We propose LUI, or Large User Interface, which increases the range of dynamic surface area of interactions possible on such a display. Our model leverages real-time continuous feedback of free-handed gestures and voice to control extensible applications such as photos, videos, and 3D models. Utilizing a single stereo-camera and voice assistant, LUI does not require exhaustive calibration or a multitude of sensors to operate, and it can be easily installed and deployed on any large screen surfaces. In a user study, participants found LUI efficient and easily learnable with minimal instruction, and preferred it to more conventional interfaces. This multimodal interface can also be deployed in augmented or virtual reality spaces and autonomous vehicle displays.

Related Content