Context-based multimodal input understanding in conversational systems

dc.contributor.authorChai, J.
dc.contributor.authorPan, Shimei
dc.contributor.authorZhou, M. X.
dc.contributor.authorHouck, K.
dc.date.accessioned2025-06-05T14:03:45Z
dc.date.available2025-06-05T14:03:45Z
dc.date.issued2002-10
dc.descriptionFourth IEEE International Conference on Multimodal Interfaces
dc.description.abstractIn a multimodal human-machine conversation, user inputs are often abbreviated or imprecise. Sometimes, merely fusing multimodal inputs together cannot derive a complete understanding. To address these inadequacies, we are building a semantics-based multimodal interpretation framework called MIND (Multimodal Interpretation for Natural Dialog). The unique feature of MIND is the use of a variety of contexts (e.g., domain context and conversation context) to enhance multimodal fusion. In this paper we present a semantically rich modeling scheme and a context-based approach that enable MIND to gain a full understanding of user inputs, including ambiguous and incomplete ones.
dc.description.urihttps://ieeexplore.ieee.org/document/1166974/
dc.format.extent6 pages
dc.genreconference papers and proceedings
dc.genrepostprints
dc.identifierdoi:10.13016/m2udti-olhe
dc.identifier.citationChai, J., Shimei Pan, M.X. Zhou, and K. Houck. “Context-Based Multimodal Input Understanding in Conversational Systems.” In Proceedings. Fourth IEEE International Conference on Multimodal Interfaces, 87–92, 2002. https://doi.org/10.1109/ICMI.2002.1166974.
dc.identifier.urihttps://doi.org/10.1109/ICMI.2002.1166974
dc.identifier.urihttp://hdl.handle.net/11603/38752
dc.language.isoen_US
dc.publisherIEEE
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Information Systems Department
dc.rights© 2002 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
dc.subjectSpeech processing
dc.subjectUSA Councils
dc.subjectContext modeling
dc.subjectSwitches
dc.subjectDisplays
dc.subjectSpeech recognition
dc.subjectText recognition
dc.subjectNatural languages
dc.subjectCities and towns
dc.subjectHistory
dc.titleContext-based multimodal input understanding in conversational systems
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-5989-8543

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
contextbased.pdf
Size:
242.08 KB
Format:
Adobe Portable Document Format