ABSTRACT
Geographic Information Systems (GIS) offers a large amount of functions for performing spatial analysis and geospatial information retrieval. However, off-the-shelf GIS remains difficult to use for occasional GIS experts. The major problem lies in that its interface organizes spatial analysis tools and functions according to spatial data structures and corresponding algorithms, which is conceptually confusing and cognitively complex. Prior work identified the usability problem of conventional GIS interface and developed alternatives based on speech or gesture to narrow the gap between the high-functionality provided by GIS and its usability. This paper outlined my doctoral research goal in understanding human-GIS interaction activity, especially how interaction modalities assist to capture spatial analysis intention and influence collaborative spatial problem solving. We proposed a framework for enabling multimodal human-GIS interaction driven by intention. We also implemented a prototype GeoEASI (Geo-dialogue Environment for Assisted Spatial Inquiry) to demonstrate the effectiveness of our framework. GeoEASI understands commonly known spatial analysis intentions through multimodal techniques and is able to assist users to perform spatial analysis with proper strategies. Further work will evaluate the effectiveness of our framework, improve the reliability and flexibility of the system, extend the GIS interface for supporting multiple users, and integrate the system into GeoDeliberation. We will concentrate on how multimodality technology can be adopted in these circumstances and explore the potentials of it. The study aims to demonstrate the feasibility of building a GIS to be both useful and usable by introducing an intent-driven multimodal interface, forming the key to building a better theory of spatial thinking for GIS.
- 1. Maged N Kamel Boulos, Bryan J Blanchard, Cory Walker, Julio Montero, Aalap Tripathy, and Ricardo Gutierrez-Osuna. 2011. Web GIS in practice X: a Microsoft Kinect natural user interface for Google Earth navigation. International journal of health geographics 10, 1 (Jan. 2011), 45.Google Scholar
- 2. Guoray Cai, Hongmei Wang, Alan M. MacEachren, and Sven Fuhrmann. 2005a. Natural Conversational Interfaces to Geospatial Databases. Transactions in GIS 9, 2 (March 2005), 199--221.Google ScholarCross Ref
- 3. Guoray Cai, Hongmei Wang, Alan M MacEachren, and Sven Fuhrmann. 2005b. Natural conversational interfaces to geospatial databases. Transactions in GIS 9, 2 (2005), 199--221.Google ScholarCross Ref
- 4. Guoray Cai and Bo Yu. 2009. Spatial annotation technology for public deliberation. Transactions in GIS 13, s1 (2009), 123--146.Google ScholarCross Ref
- 5. Guoray Cai, Bo Yu, and Dong Chen. 2013a. Modeling and Communicating the Conceptual Intent of Geo-Analytical Tasks for Human-GIS Interaction. Transactions in GIS 17, 3 (June 2013), 353--368.Google ScholarCross Ref
- 6. Guoray Cai, Bo Yu, and Dong Chen. 2013b. Modeling and Communicating the Conceptual Intent of Geo-Analytical Tasks for Human-GIS Interaction. Transactions in GIS 17, 3 (2013), 353--368.Google ScholarCross Ref
- 7. Gerhard Fischer. 1987. Making computers more useful and more usable. In HCI (2). 97--104.Google Scholar
- 8. Michael F Goodchild. 2011a. Spatial thinking and the gis user interface. Procedia-Social and Behavioral Sciences 21 (2011), 3--9.Google ScholarCross Ref
- 9. Michael F. Goodchild. 2011b. Spatial Thinking and the GIS User Interface. Procedia - Social and Behavioral Sciences 21 (Jan. 2011), 3--9.Google Scholar
- 10. Michael F. Goodchild. 2013. Asking Spatial Questions to Identify GIS Functionality. 2013 Fourth International Conference on Computing for Geospatial Research and Application (July 2013), 106--110. Google ScholarDigital Library
- 11. Barbara J Grosz and Sarit Kraus. 1996. Collaborative plans for complex group action. Artificial Intelligence 86, 2 (1996), 269--357. Google ScholarDigital Library
- 12. Barbara J Grosz and Sarit Kraus. 1999. The evolution of SharedPlans. In Foundations of rational agency. Springer, 227--262.Google Scholar
- 13. Barbara J Grosz and Candace L Sidner. 1986. Attention, intentions, and the structure of discourse. Computational linguistics 12, 3 (1986), 175--204. Google ScholarDigital Library
- 14. Brent Hecht, Johannes Schöning, Thomas Erickson, and Reid Priedhorsky. 2011. Geographic human-computer interaction. Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems - CHI EA '11 (2011), 447. Google ScholarDigital Library
- 15. Nicholas R. Hedley, Mark Billinghurst, Lori Postner, Richard May, and Hirokazu Kato. 2002. Explorations in the Use of Augmented Reality for Geographic Visualization. Presence: Teleoper. Virtual Environ. 11, 2 (April 2002), 119--133. Google ScholarDigital Library
- 16. Karen E Lochbaum. 1998. A collaborative planning model of intentional structure. Computational Linguistics 24, 4 (1998), 525--572. Google ScholarDigital Library
- 17. Ishantha Lokuge and Suguru Ishizaki. 1995. Geospace: An interactive visualization system for exploring complex information spaces. In Proceedings of the SIGCHI conference on Human factors in computing systems. ACM Press/Addison-Wesley Publishing Co., 409--414. Google ScholarDigital Library
- 18. A Mitchell. 1999. The ESRI Guide to GIS analysis, Volume 1: Geographic patterns and relationship. Redlands, Calif: ESRI.Google Scholar
- 19. A Mitchell. 2005. The ESRI Guide to GIS analysis, Volume 2: Spartial measurements and statistics. Redlands, Calif: ESRI.Google Scholar
- 20. Sharon Oviatt. 1996. Multimodal interfaces for dynamic interactive maps. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, 95--102. Google ScholarDigital Library
- 21. Ingmar Rauschert, Pyush Agrawal, Rajeev Sharma, Sven Fuhrmann, Isaac Brewer, and Alan MacEachren. 2002a. Designing a human-centered, multimodal GIS interface to support emergency management. In Proceedings of the 10th ACM international symposium on Advances in geographic information systems. ACM, 119--124. Google ScholarDigital Library
- 22. Ingmar Rauschert, Pyush Agrawal, Rajeev Sharma, Sven Fuhrmann, Isaac Brewer, and Alan MacEachren. 2002b. Designing a human-centered, multimodal GIS interface to support emergency management. In Proceedings of the tenth ACM international symposium on Advances in geographic information systems - GIS '02. ACM Press, New York, New York, USA, 119. Google ScholarDigital Library
- 23. SC Shapiro, H Chalupski, and HC Chou. 1991. Linking ARC/INFO with snactor. Santa Barbara, California.: National Center for geographic Information and Analysis (1991).Google Scholar
- 24. Carol Traynor and Marian G Williams. 1995. Why are geographic information systems hard to use?. In Conference companion on Human factors in computing systems. ACM, 288--289. Google ScholarDigital Library
- 25. Hongmei Wang, Guoray Cai, and Alan M MacEachren. 2008. GeoDialogue: a software agent enabling collaborative dialogues between a user and a conversational GIS. In Tools with Artificial Intelligence, 2008. ICTAI'08. 20th IEEE International Conference on, Vol. 2. IEEE, 357--360. Google ScholarDigital Library
Index Terms
- Exploring Intent-driven Multimodal Interface for Geographical Information System
Recommendations
Multimodal interface design for multimodal meeting content retrieval
ICMI '04: Proceedings of the 6th international conference on Multimodal interfacesThis thesis will investigate which modalities, and in which combinations, are best suited for use in a multimodal interface that allows users to retrieve the content of recorded and processed multimodal meetings. The dual role of multimodality in the ...
Active agent oriented multimodal interface system
IJCAI'95: Proceedings of the 14th international joint conference on Artificial intelligence - Volume 1This paper presents a prototype of an interface system with an active human-like agent. In usual human communication, non-verbal expressions play important roles. They convey emotional information and control timing of interaction as well. This project ...
A usability study of multimodal input in an augmented reality environment
In this paper, we describe a user study evaluating the usability of an augmented reality (AR) multimodal interface (MMI). We have developed an AR MMI that combines free-hand gesture and speech input in a natural way using a multimodal fusion ...
Comments