skip to main content
10.1145/1570433.1570497acmconferencesArticle/Chapter ViewAbstractPublication PageseicsConference Proceedingsconference-collections
short-paper

High level data fusion on a multimodal interactive application platform

Published:15 July 2009Publication History

ABSTRACT

This research aims to propose a multimodal fusion framework for high-level data integration between two or more modalities. It takes as input extracted low level features from different system devices, analyzes and identifies intrinsic meanings in these data through dedicated processes running in parallel. Extracted meanings are mutually compared to identify complementarities, ambiguities and inconsistencies to better understand the user intention when interacting with the system. The whole fusion lifecycle will be described and evaluated in an ambient intelligence scenario, where two co-workers interact by voice and movements, demonstrating their intentions and the system gives advices according to identified needs.

References

  1. Bolt R. A., "Put-that-there: Voice and gesture at the graphics interface," in International Conference on Computer Graphics and Interactive Techniques, July 1980, pp. 262--270 Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Bouchet, J. and Nigay, L. ICARE: a component-based approach for the design and development of multimodal interfaces. In CHI '04 Extended Abstracts on Human Factors in Computing Systems. Vienna, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Cole L., Austin D., Cole L., Visual Object Recognition using Template Matching, Proceedings of Australasian Conference on Robotics and Automation, 2004.Google ScholarGoogle Scholar
  4. Curran J., Clark S. and Bos J. (2007): Linguistically Motivated Large-Scale NLP with C&C and Boxer. Proceedings of the ACL 2007 Demonstrations Session (ACL-07 demo), pp. 29--32. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Engel, Ralf; Pfleger, Norbert. Modality Fusion. SmartKom: Foundations of Multimodal Dialogue Systems. Springer, Berlin, 2006.Google ScholarGoogle Scholar
  6. Eugene C., Mark J. Coarse-to-Fine N-Best Parsing and MaxEnt Discriminative Reranking. Proc. ACL2005. pp. 173--180. Ann Arbor, USA, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Intel Corporation, Open Source Computer Vision Library--OpenCV, http://www.intel.com/technology/computing/opencv/index.htm, 2008.Google ScholarGoogle Scholar
  8. Kaiser, Ed; Olwal, Alex; McGee, David; Benko, Hrvoje; Corradini, Andrea; Li, Xiaoguang; Cohen, Phil; Feiner, Steven. Mutual Disambiguation of 3D Multimodal Interaction in Augmented and Virtual Reality. ICMI'03. Vancouver, Canada. November, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Lawson, Lionel; Macq, Benoit. OpenInterface Platform for Multimodal Applications Prototyping. ICASSP Show&Tell '08, Las Vegas, USA. April, 2008.Google ScholarGoogle Scholar
  10. Martin, J. C. TYCOON: Theoretical Framework and Software Tools for Multimodal Interfaces. Intelligence and Multimodality in Multimedia interfaces. (ed.) John Lee, AAAI Press, 1998.Google ScholarGoogle Scholar
  11. Protégé: http://protege.stanford.edu/Google ScholarGoogle Scholar
  12. Soar: http://sitemaker.umich.edu/soar/homeGoogle ScholarGoogle Scholar
  13. Vybornova, Olga; Gemo, Monica; Macq, Benoit. Multimodal Multi-Level Fusion Using Contextual Information. ERCIM News, N0. 70. July, 2006.Google ScholarGoogle Scholar
  14. Vybornova, Olga; Mendonça, Hildeberto; Neiberg, Daniel; Gomez, Antonio; Shen, Ao. Multimodal High Level Data Integration. eNTERFACE'08 Workshop, Orsay-Paris, France. August, 2008.Google ScholarGoogle Scholar
  15. Vybornova, Olga; Mendonça, Hildeberto; Lawson, Lionel; Macq, Benoit. High Level Data Fusion on a Multimodal Interactive Application Platform. IEEE ISM, Berkeley, California, USA. December, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. W. Walker, et. al. "Sphinx-4: A Flexible Open Source Framework for Speech Recognition" Technical Report, SMLI TR2004-0811, 2004 SUN MICROSYSTEMS INC.Google ScholarGoogle Scholar

Index Terms

  1. High level data fusion on a multimodal interactive application platform

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          EICS '09: Proceedings of the 1st ACM SIGCHI symposium on Engineering interactive computing systems
          July 2009
          348 pages
          ISBN:9781605586007
          DOI:10.1145/1570433

          Copyright © 2009 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 15 July 2009

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • short-paper

          Acceptance Rates

          Overall Acceptance Rate55of238submissions,23%

          Upcoming Conference

          EICS '24

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader