The Effect of Perceptual Structure on Multimodal Speech Recognition Interfaces
Loading...
Files
Permanent Link
Author/Creator
Author/Creator ORCID
Date
1998-01-01
Type of Work
Department
Program
Citation of Original Publication
Rights
This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
Abstract
A framework of complementary behavior has been proposed which maintains that direct
manipulation and speech interfaces have reciprocal strengths and weaknesses. This suggests that
user interface performance and acceptance may increase by adopting a multimodal approach that
combines speech and direct manipulation. This effort examined the hypothesis that the speed,
accuracy, and acceptance of multimodal speech and direct manipulation interfaces will increase
when the modalities match the perceptual structure of the input attributes. A software prototype
which supported a typical biomedical data collection task was developed to test this hypothesis.
A group of 20 clinical and veterinary pathologists evaluated the prototype in an experimental
setting using repeated measures. The results of this experiment supported the hypothesis that the
perceptual structure of an input task is an important consideration when designing a multimodal
computer interface. Task completion time, the number of speech errors, and user acceptance
improved when interface best matched the perceptual structure of the input attributes.