Fusing AI: Multimodal Language Models Inference Across Diverse Inputs
Abstract
References
Index Terms
- Fusing AI: Multimodal Language Models Inference Across Diverse Inputs
Recommendations
Leveraging hierarchy in multimodal generative models for effective cross-modality inference
AbstractThis work addresses the problem of cross-modality inference (CMI), i.e., inferring missing data of unavailable perceptual modalities (e.g., sound) using data from available perceptual modalities (e.g., image). We overview single-...
Skipping spare information in multimodal inputs during multimodal input fusion
IUI '09: Proceedings of the 14th international conference on Intelligent user interfacesIn a multimodal interface, a user can use multiple modalities, such as speech, gesture, and eye gaze etc., to communicate with a system. As a critical component in a multimodal interface, multimodal input fusion explores the ways to effectively ...
Intent capturing through multimodal inputs
HCI'13: Proceedings of the 15th international conference on Human-Computer Interaction: interaction modalities and techniques - Volume Part IVVirtual manufacturing environments need complex and accurate 3D human-computer interaction. One main problem of current virtual environments (VEs) is the heavy overloads of the users on both cognitive and motor operational aspects. This paper ...
Comments
Information & Contributors
Information
Published In
Publisher
IEEE Computer Society Press
Washington, DC, United States
Publication History
Qualifiers
- Discussion
Contributors
Other Metrics
Bibliometrics & Citations
Bibliometrics
Article Metrics
- 0Total Citations
- 0Total Downloads
- Downloads (Last 12 months)0
- Downloads (Last 6 weeks)0