Steven Hennessy
Jeremy Nelissen
Reference Information
Title: A Multimodal Labeling Interface for Wearable Computing
Authors: Shanqing Li, Yunde Jia
When/Where: IUI 2010
Summary
This paper proposed a system for labeling objects in wearable environments using a multimodal interface instead of with the inconvenient mice and keyboard. The program uses visual tracking to allow the user to circle on object by pointing, then the user speaks the name of the object and speech recognition is used to get the label. The visual tracking and speech recognition is performed in real-time. The system hardware consists of a head-mounted display to track the user's circling of objects, a wearable computer, and a Bluetooth microphone to capture the user's voice.
View of the system in process. Source: Article. |
Discussion
I thought this system was pretty interesting. It makes sense that for this kind of environment, a keyboard and mouse would be inconvenient, so this proposed approach seems like it would be efficient. However, it seems like its effectiveness would depend largely on how well the speech recognition and hand-tracking recognition systems work. Also, most of the paper was largely technical and mathematical, making it rather hard to understand. However, I like the idea of changing the input methods for unique systems such that the method of input is adapted to the system itself.
I also thought this system was very interesting, but the paper was a harder read
ReplyDeleteI definitely like the idea behind this tech, but some major miniaturization would be required before it would be highly usable.
ReplyDeleteAlso, I wonder how much this technology would be hampered by high-noise environments, or if multiple people were using this near each other.
When I read this paper, I though that it could have effectiveness in tourism and navigation. While the system obviously needs to be condensed, I think it could have real applications in the future.
ReplyDelete