Skip to main content
Log in

Dempster-Shafer theoretic resolution of referential ambiguity

  • Published:
Autonomous Robots Aims and scope Submit manuscript

Abstract

Robots designed to interact with humans in realistic environments must be able to handle uncertainty with respect to the identities and properties of the people, places, and things found in their environments. When humans refer to these entities using under-specified language, robots must often generate clarification requests to determine which entities were meant. In this paper, we first present recommendations for designers of robots needing to generate such requests. We then show how a Dempster-Shafer theoretic pragmatic reasoning component capable of generating requests to clarify pragmatic uncertainty can also generate requests to resolve referential uncertainty when integrated with probabilistic reference resolution and referring expression generation components. Our system is then demonstrated in a simulated alpine search and rescue context enabled by a novel hybrid architecture.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Notes

  1. While not directly relevant to the present work, there has also been research on using interaction patterns to identify opportunities for clarification in situated settings (Carrillo and Topp 2016).

  2. Future research will be needed to determine how the content of the options to be offered may impact how this decision is made. The results of such research may suggest refinements of this recommendation.

  3. We have chosen to adopt (and extend) the notation used in Tellex et al. (2011) in order to facilitate easier comparison to related work. We would advocate for its adoption as a common notation across the reference resolution and symbol grounding communities.

  4. Note that for most utterances j will be very small, and k will in almost all circumstances be either 1 or 2.

  5. Note, however, that low-probability hypotheses are pruned out during the resolution process, and thus the remaining hypotheses have a higher concentration of mass (and thus, higher belief and plausibility) than they would if this pruning process were not employed. This pruning process is further described by Williams et al. (2016).

  6. Some other groups have, since the publication of our original work on this topic (Williams et al. 2015), followed a similar approach, notably in the Rational Speech Act Theory inspired robotics literature (Fried et al. 2017) and in work on “inverse semantics” (Knepper et al. 2015, 2017). See also both prior and posterior work on language understanding from the Rational Speech Act psychological literature (Goodman and Stuhlmüller 2013; Goodman and Frank 2016), as well as critiques of such approaches (Gatt et al. 2013; Qing and Franke 2015).

  7. It is important to note that our pragmatic reasoning system currently is only equipped to handle conventionalized Indirect Speech Acts. For a comprehensive handling of ISAs, it will be necessary to integrate this approach with a plan reasoning system (Perrault and Allen 1980; Briggs and Scheutz 2013; Trott and Bergen 2017).

  8. Note here that we have chosen to use rules, in our example as well as in our evaluation, that use the form “Do you need Y or Z” rather than the more indirect and hence more polite “Would you like Y or Z”. These two forms trade off between our desiderata. “Do you need Y or Z” (in response to “I need X” better demonstrates intentions, but is less pragmatically appropriate, than “Would you like Y or Z”, and vice versa. Although we are able to generate both forms using the presented approach, we chose to use the form “Do you need Y or Z”, in part because, while it may be less pragmatically appropriate than “Would you like Y or Z”, both forms are significantly more pragmatically appropriate than the use of a direct command.

  9. As above, the probabilities of different properties holding for these objects were arbitrarily hand-selected for the sake of a clear and simple demonstration walkthrough. A set of “dummy” Consultants were used that provided these hand-selected probabilities when asked for probability judgments. In practice, these probability judgments can be provided by arbitrary classifiers, such as those commonly used for object recognition (e.g. Redmon et al. 2016), which may often return different levels of confidence for different observed objects.

  10. Here, \(agt_1\) is changed to the agent’s name for dialogue processing.

  11. All beliefs and plausibilities in this section are rounded.

  12. The uncertainty intervals associated with different rules were arbitrarily hand-selected for the sake of the demonstration walkthrough. For a discussion of how these intervals might be adapted over time, we direct the interested reader to (Williams et al. 2014).

  13. This Component is named after the European SHERPA project (Marconi et al. 2012) for which the alpine search and rescue KnowRob ontologies used in this integration were developed.

  14. Although, see recent discussion of the shortcomings of such checks (Hauser and Schwarz 2015), especially in crowdsourced experiments (Curran 2016).

References

  • Bauer, M. (1997). Approximation algorithms and decision making in the Dempster-Shafer theory of evidence—An empirical study. International Journal of Approximate Reasoning, 17(2–3), 217–237.

    Article  MathSciNet  MATH  Google Scholar 

  • Bechhofer, S. (2009). Owl: Web ontology language. In Encyclopedia of database systems (pp. 2008–2009). New York: Springer.

  • Beetz, M., Mösenlechner, L., & Tenorth, M. (2010). CRAM—A cognitive robot abstract machine for everyday manipulation in human environments. In Proceedings of the IEEE/RSJ international conference on intelligent robots and systems (IROS), Taipei, Taiwan, pp. 1012–1017.

  • Beetz, M., Mösenlechner, L., Tenorth, M., & Rühr, T. (2012). Cram—A cognitive robot abstract machine. In 5th International conference on cognitive systems (CogSys 2012).

  • Benotti, L., & Blackburn, P. (2017). Modeling the clarification potential of instructions: Predicting clarification requests and other reactions. Computer Speech & Language, 45, 536–551.

    Article  Google Scholar 

  • Berenson, D., & Srinivasa, S. S. (2008). Grasp synthesis in cluttered environments for dexterous hands. In Proceedings of the 8th IEEE-RAS international conference on humanoid robots (HUMANOIDS), pp. 189–196.

  • Black, A., Taylor, P., Caley, R., & Clark, R. (1998). The festival speech synthesis system. Technical report. Edinburgh: University of Edinburgh.

    Google Scholar 

  • Brennan, S. E., Galati, A., & Kuhlen, A. K. (2010). Two minds, one dialog: Coordinating speaking and understanding. Psychology of Learning and Motivation, 53, 301–344.

    Article  Google Scholar 

  • Brenner, M., & Kruijff-Korbayová, I. (2008). A continual multiagent planning approach to situated dialogue. In Proceedings of the 12th workshop on the semantics and pragmatics of dialogue (Semdial), London, UK.

  • Brick, T., & Scheutz, M. (2007). Incremental natural language processing for HRI. In Proceeding of the ACM/IEEE international conference on human-robot interaction (HRI) pp. 263–270.

  • Briggs, G., & Scheutz, M. (2013). A hybrid architectural approach to understanding and appropriately generating indirect speech acts. In Proceedings of the twenty-seventh AAAI conference on artificial intelligence (AAAI).

  • Brown, P. (1987). Politeness: Some universals in language usage (Vol. 4). Cambridge: Cambridge University Press.

    Book  Google Scholar 

  • Byron, D., Koller, A., Striegnitz, K., Cassell, J., Dale, R., Moore, J., & Oberlander, J. (2009). Report on the first NLG challenge on generating instructions in virtual environments (GIVE). In Proceedings of the twelfth European workshop on natural language generation (ENLG), Association for Computational Linguistics, pp. 165–173.

  • Cai, H., & Mostofi, Y. (2016). Asking for help with the right question by predicting human visual performance. In Proceedings of robotics: Science and systems.

  • Carrillo, F. M., & Topp, E. A. (2016). Interaction and task patterns in symbiotic, mixed-initiative human-robot interaction. In AAAI workshop: Symbiotic cognitive systems.

  • Carroll, J. B. (1964). Language and thought. Foundations of modern psychology. Englewood Cliffs, NJ: Prentice-Hall.

    Google Scholar 

  • Clark, H. H. (1996). Using language. Cambridge: Cambridge University Press.

    Book  Google Scholar 

  • Clark, H. H., & Schaefer, E. F. (1989). Contributing to discourse. Cognitive Science, 13(2), 259–294.

    Article  Google Scholar 

  • Crockford, D. (2006). The application/json media type for javascript object notation (json).

  • Curran, P. G. (2016). Methods for the detection of carelessly invalid responses in survey data. Journal of Experimental Social Psychology, 66, 4–19.

    Article  Google Scholar 

  • Dale, R., & Reiter, E. (1995). Computational interpretations of the gricean maxims in the generation of referring expressions. Cognitive Science, 19(2), 233–263.

    Article  Google Scholar 

  • Deits, R., Tellex, S., Kollar, T., & Roy, N. (2013). Clarifying commands with information-theoretic human-robot dialog. Journal of Human-Robot Interaction (JHRI), 2(2), 58–79.

    Google Scholar 

  • Dempster, A. P. (2008). The Dempster-Shafer calculus for statisticians. International Journal of approximate reasoning, 48(2), 365–377.

    Article  MathSciNet  MATH  Google Scholar 

  • Dzifcak, J., Scheutz, M., Baral, C., & Schermerhorn, P. (2009). What to do and how to do it: Translating natural language directives into temporal and dynamic logic representation for goal management and action execution. In Proceedings of the international conference on robotics and automation (ICRA), Kobe, Japan.

  • Fagin, R., & Halpern, J. Y. (1991). A new approach to updating beliefs. In Uncertainty in artificial intelligence (pp. 347–374). New York: Elsevier Science Publishers.

  • Fried, D., Andreas, J., & Klein, D. (2017). Unified pragmatic models for generating and following instructions. arXiv:1711.04987.

  • Gabsdil, M. (2003). Clarification in spoken dialogue systems. In Proceedings of the 2003 AAAI spring symposium. Workshop on natural language generation in spoken and written dialogue (pp. 28–35).

  • Garoufi, K., & Koller, A. (2014). Generation of effective referring expressions in situated context. Language, Cognition and Neuroscience, 29(8), 986–1001.

    Article  Google Scholar 

  • Gatt, A., & Reiter, E. (2009). Simplenlg: A realisation engine for practical applications. In Proceedings of the 12th European workshop on natural language generation, association for computational linguistics (pp. 90–93).

  • Gatt, A., van Gompel, R. P., van Deemter, K., & Kramer, E. (2013). Are we Bayesian referring expression generators. In Proceedings of the thirty-fifth annual meeting of the cognitive science society.

  • Ginzburg, J. (2009). The interactive stance: Meaning for conversation (forthcoming in 2009). Studies in Computational Linguistics.

  • Goodman, N. D., & Frank, M. C. (2016). Pragmatic language interpretation as probabilistic inference. Trends in Cognitive Sciences, 20(11), 818–829.

    Article  Google Scholar 

  • Goodman, N. D., & Stuhlmüller, A. (2013). Knowledge and implicature: Modeling language understanding as social cognition. Topics in cognitive science, 5(1), 173–184.

    Article  Google Scholar 

  • Grice, H. P. (1970). Logic and conversation. Syntax and Semantics, 3, 41–58.

    Google Scholar 

  • Gundel, J. K., Hedberg, N., & Zacharski, R. (1993). Cognitive status and the form of referring expressions in discourse. Language, 69(2), 274–307.

    Article  Google Scholar 

  • Hauser, D. J., & Schwarz, N. (2015). It’s a trap! instructional manipulation checks prompt systematic thinking on “tricky” tasks. Sage Open, 5(2),

  • Heendeni, J. N., Premaratne, K., Murthi, M., Uscinski, J., & Scheutz, M. (2016). A generalization of Bayesian inference in the Dempster-Shafer belief theoretic framework. In Proceedings of the international conference on information fusion.

  • Hemachandra, S., Walter, M. R., & Teller, S. (2014). Information theoretic question asking to improve spatial semantic representations. In Proceedings of the AAAI fall symposium series.

  • Huang, Y. T., & Snedeker, J. (2011). Logic and conversation revisited: Evidence for a division between semantic and pragmatic content in real-time language comprehension. Language and Cognitive Processes, 26(8), 1161–1172.

    Article  Google Scholar 

  • Knepper, R. A., Tellex, S., Li, A., Roy, N., & Rus, D. (2015). Recovering from failure by asking for help. Autonomous Robots, 39(3), 347–362.

    Article  Google Scholar 

  • Knepper, R. A., Mavrogiannis, C. I., Proft, J., & Liang, C. (2017). Implicit communication in a joint action. In Proceedings of the 2017 ACM/IEEE international conference on human-robot interaction (HRI) (pp. 283–292). ACM.

  • Koenig, N., & Howard, A. (2004). Design and use paradigms for Gazebo, an open-source multi-robot simulator. In Proceedings of the IEEE/RSJ international conference on intelligent robots and systems (IROS) (Vol. 3, pp. 2149–2154).

  • Kollar, T., Tellex, S., Walter, M., Huang, A., Bachrach, A., Hemachandra, S., et al. (2017). Generalized grounding graphs: A probabilistic framework for understanding grounded commands. arXiv:1712.01097.

  • Koller, A., Striegnitz, K., Gargett, A., Byron, D., Cassell, J., Dale, R., Moore, J., & Oberlander, J. (2010). Report on the second nlg challenge on generating instructions in virtual environments (GIVE-2). In Proceedings of the sixth international natural language generation conference (INLG), association for computational linguistics (pp. 243–250).

  • Koolen, R., Krahmer, E., & Swerts, M. (2016). How distractor objects trigger referential overspecification: Testing the effects of visual clutter and distractor distance. Cognitive Science, 40(7), 1617–1647.

    Article  Google Scholar 

  • Krause, E., Cantrell, R., Potapova, E., Zillich, M., & Scheutz, M. (2013). Incrementally biasing visual search using natural language input. In Proceedings of the 12th international conference on autonomous agents and multi-agent systems (AAMAS) (pp. 31–38).

  • Kruijff, G. J. M., Kelleher, J. D., & Hawes, N. (2006a). Information fusion for visual reference resolution in dynamic situated dialogue. In Perception and interactive technologies. New York: Springer.

  • Kruijff, G. J. M., Zender, H., Jensfelt, P., & Christensen, H.I. (2006b). Clarification dialogues in human-augmented mapping. In Proceedings of the 1st ACM SIGCHI/SIGART conference on Human-Robot Interaction (HRI) (pp. 282–289).

  • Kruijff, G. J. M., Brenner, M., & Hawes, N. (2008). Continual planning for cross-modal situated clarification in human-robot interaction. In Proceedings of the seventeenth IEEE international symposium on robot and human interactive communication (RO-MAN) (pp. 592–597).

  • Lemaignan, S., Warnier, M., Sisbot, E. A., Clodic, A., & Alami, R. (2017). Artificial cognition for social human-robot interaction: An implementation. Artificial Intelligence, 247, 45–69.

    Article  MathSciNet  MATH  Google Scholar 

  • Marconi, L., Melchiorri, C., Beetz, M., Pangercic, D., Siegwart, R., Leutenegger, S., Carloni, R., Stramigioli, S., Bruyninckx, H., Doherty, P., et al. (2012). The sherpa project: Smart collaboration between humans and ground-aerial robots for improving rescuing activities in alpine environments. In Proceedings of the IEEE international symposium on safety, security, and rescue robotics (SSRR) (pp. 1–4)

  • Marge, M., & Rudnicky, A. I. (2015). Miscommunication recovery in physically situated dialogue. In Proceedings of the sixteenth annual meeting of the special interest group on discourse and dialogue (SIGDIAL) (pp. 22–49).

  • Matarić, M. J. (2002). Situated robotics. In L. Nadel (Ed.), Encyclopedia of cognitive science. London: Nature Publishers Group, Macmillan Reference Ltd.

    Google Scholar 

  • Matuszek, C., Herbst, E., Zettlemoyer, L., & Fox, D. (2012). Learning to parse natural language commands to a robot control system. In Proceedings of the thirteenth international symposium on experimental robotics (ISER) (pp. 403–415).

  • Maurtua, I., Fernandez, I., Kildal, J., Susperregi, L., Tellaeche, A., & Ibarguren, A. (2016). Enhancing safe human-robot collaboration through natural multimodal communication. In: Proceedings of the 21st IEEE international conference on emerging technologies and factory automation (ETFA), IEEE (pp. 1–8).

  • Mavridis, N. (2015). A review of verbal and non-verbal human-robot interactive communication. Robotics and Autonomous Systems, 63, 22–35.

    Article  MathSciNet  Google Scholar 

  • McGuinness, D. L., Van Harmelen, F., et al. (2004). Owl web ontology language overview. W3C Recommendation 10(10):2004.

  • Meo, T., McMahan, B., & Stone, M. (2014). Generating and resolving vague color references. In Proceedings of the eighteenth SEMDIAL workshop on the semantics and pragmatics of dialogue (DialWatt) (pp. 107–115).

  • Mösenlechner, L., & Beetz, M. (2011). Parameterizing actions to have the appropriate effects. In IEEE/RSJ international conference on intelligent robots and systems (IROS), San Francisco, CA, USA.

  • Mutlu, B., & Forlizzi, J. (2008). Robots in organizations: the role of workflow, social, and environmental factors in human-robot interaction. In Proceedings of the 3rd ACM/IEEE international conference on human-robot interaction (HRI) (pp. 287–294).

  • Núñez, R. C., Dabarera, R., Scheutz, M., Briggs, G., Bueno, O., Premaratne, K., & Murthi, M. N. (2013a). DS-based uncertain implication rules for inference and fusion applications. In Proceedings of the sixteenth international conference on information fusion (FUSION) (pp. 1934–1941).

  • Núñez, R. C., Scheutz, M., Premaratne, K., & Murthi, M. N. (2013b). Modeling uncertainty in first-order logic: A Dempster-Shafer theoretic approach. In Proceedings of the eighth international symposium on imprecise probability: Theories and applications.

  • Orita, N., Vornov, E., Feldman, N., & Daumé III, H. (2015). Why discourse affects speakers’ choice of referring expressions. In Proceedings of the 53rd annual meeting of the association for computational linguistics and the 7th international joint conference on natural language processing (Vol. 1, pp. 1639–1649).

  • Papineni, K., Roukos, S., Ward, T., & Zhu, W. J. (2002). BLEU: A method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting on association for computational linguistics, Association for Computational Linguistics (pp. 311–318).

  • Perrault, C. R., & Allen, J. F. (1980). A plan-based analysis of indirect speech acts. Computational Linguistics, 6(3–4), 167–182.

    Google Scholar 

  • Polpitiya, L. G., Premaratne, K., Murthi, M. N., & Sarkar, D. (2017). Efficient computation of belief theoretic conditionals. In Proceedings of the tenth international symposium on imprecise probability: Theories and applications (pp. 265–276).

  • Purver, M. (2004). Clarie: The clarification engine. In Proceedings of the eighth SEMDIAL workshop on the semantics and pragmatics of dialogue (CATALOG) (pp. 77–84).

  • Purver, M., Ginzburg, J., & Healey, P. (2003). On the means for clarification in dialogue. In Current and new directions in discourse and dialogue (pp. 235–255). New York: Springer.

  • Qing, C., & Franke, M. (2015). Variations on a Bayesian theme: Comparing Bayesian models of referential reasoning. In Bayesian natural language semantics and pragmatics (pp. 201–220). New York: Springer

  • Quigley, M., Conley, K., Gerkey, B., Faust, J., Foote, T. B., Leibs, J., Wheeler, R., & Ng, A. Y. (2009). ROS: an open-source robot operating system. In ICRA workshop on open source software.

  • Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779–788).

  • Reiter, E., Dale, R., & Feng, Z. (2000). Building natural language generation systems. Cambridge: MIT Press.

    Book  Google Scholar 

  • Richards, B. (1987). Type/token ratios: What do they really tell us? Journal of Child Language, 14(02), 201–209.

    Article  Google Scholar 

  • Rodríguez, K. J., & Schlangen, D. (2004). Form, intonation and function of clarification requests in German task-oriented spoken dialogues. In Proceedings of the 8th workshop on the semantics and pragmatics of dialogue (SemDial).

  • Rosenthal, S., & Veloso, M. (2012). Mobile robot planning to seek help with spatially-situated tasks. In: Proceedings of the AAAI conference on artificial intelligence (AAAI).

  • Rosenthal, S., Veloso, M., & Dey, A. K. (2012a). Acquiring accurate human responses to robots’ questions. International Journal of Social Robotics, 4(2), 117–129.

    Article  Google Scholar 

  • Rosenthal, S., Veloso, M., & Dey, A. K. (2012b). Is someone in this office available to help me? Proactively seeking help from spatially-situated humans. Journal of Intelligent and Robotic Systems, 66, 205–221.

    Article  Google Scholar 

  • Roy, D. K. (2002). Learning visually grounded words and syntax for a scene description task. Computer Speech and Language, 16(3–4), 353–385.

    Article  Google Scholar 

  • Scalise, R., Li, S., Admoni, H., Rosenthal, S., & Srinivasa, S. S. (2018). Natural language instructions for human-robot collaborative manipulation. The International Journal of Robotics Research, 37(6), 558–565.

    Article  Google Scholar 

  • Schegloff, E. A. (1987). Some sources of misunderstanding in talk-in-interaction. Linguistics, 25(1), 201–218.

    Article  Google Scholar 

  • Schenk, E., & Guittard, C. (2009). Crowdsourcing: What can be outsourced to the crowd, and why. In Workshop on open source innovation, Strasbourg, France, Vol. 72.

  • Schermerhorn, P. W., Kramer, J. F., Middendorff, C., & Scheutz, M. (2006). DIARC: A testbed for natural human-robot interaction. In Proceedings of the twentieth AAAI conference on artificial intelligence (AAAI) (pp. 1972–1973).

  • Scheutz, M., Krause, E., & Sadeghi, S. (2014). An embodied real-time model of language-guided incremental visual search. In:Proceedings of the thirty-sixth annual meeting of the cognitive science society.

  • Scheutz, M., Krause, E., Oosterveld, B., Frasca, T., & Platt, R. (2017). Spoken instruction-based one-shot object and action learning in a cognitive robotic architecture. In Proceedings of the 16th conference on autonomous agents and multiagent systems (AAMAS) (pp. 1378–1386).

  • Schröder, M., & Trouvain, J. (2003). The german text-to-speech synthesis system MARY: A tool for research, development and teaching. International Journal of Speech Technology, 6(4), 365–377.

    Article  Google Scholar 

  • Schröder, M., Charfuelan, M., Pammi, S., & Steiner, I. (2011). Open source voice creation toolkit for the MARY TTS platform. In Twelfth annual conference of the international speech communication association.

  • Searle, J. R. (1975). Indirect speech acts. Syntax and Semantics, 3, 59–82.

    Google Scholar 

  • Shafer, G. (1976). A mathematical theory of evidence. Princeton: Princeton University Press.

    MATH  Google Scholar 

  • Steedman, M., & Baldridge, J. (2011). Combinatory categorial grammar. Non-transformational syntax: Formal and explicit models of grammar (pp. 181–224).

  • Steele, G. (1990). Common LISP: The language. New York: Elsevier.

    MATH  Google Scholar 

  • Stirling, A. (2010). Keep it complex. Nature, 468(7327), 1029–1031.

    Article  Google Scholar 

  • Stoyanchev, S., Liu, A., & Hirschberg, J. (2013). Modelling human clarification strategies. In Proceedings of the 14th annual meeting of the special interest group on discourse and dialogue (SIGDIAL) (pp. 137–141).

  • Talamadupula, K., Kambhampati, S., Schermerhorn, P., Benton, J., & Scheutz, M. (2011). Planning for human-robot teaming. In Proceedings of the ICAPS workshop on scheduling and planning applications (SPARK), Vol. 67.

  • Tang, Y., Hang, C. W., Parsons, S., & Singh, M. P. (2012). Towards argumentation with symbolic Dempster-Shafer evidence. In Proceedings of the second international conference on computational models of argument (COMMA) (pp. 462–469).

  • Tellex, S., Kollar, T., Dickerson, S., Walter, M. R., Banerjee, A. G., Teller, S., et al. (2011). Approaching the symbol grounding problem with probabilistic graphical models. AI Magazine, 32(4), 64–76.

    Article  Google Scholar 

  • Tellex, S., Thaker, P., Deits, R., Simeonov, D., Kollar, T., & Roy, N. (2013). Toward information theoretic human-robot dialog. Robotics, 32, 409–417.

    Google Scholar 

  • Tellex, S., Knepper, R. A., Li, A., Rus, D., & Roy, N. (2014). Asking for help using inverse semantics. In Proceedings of robotics: Science and systems, Vol. 2.

  • Tenbrink, T., Ross, R. J., Thomas, K. E., Dethlefs, N., & Andonova, E. (2010). Route instructions in map-based human-human and human-computer dialogue: A comparative analysis. Journal of Visual Languages & Computing, 21(5), 292–309.

    Article  Google Scholar 

  • Tenorth, M., & Beetz, M. (2009). KnowRob—knowledge processing for autonomous personal robots. In Proceedings of the IEEE/RSJ international conference on intelligent robots and systems (IROS) (pp. 4261–4266).

  • Tenorth, M., & Beetz, M. (2017). Representations for robot knowledge in the KnowRob framework. Artificial Intelligence, 247, 151–169.

    Article  MathSciNet  MATH  Google Scholar 

  • Traum, D. R. (1994). A computational theory of grounding in natural language conversation. PhD thesis, University of Rochester, Rochester, NY.

  • Trott, S., & Bergen, B. (2017). A theoretical model of indirect request comprehension. In Proceedings of the AAAI fall symposium on artificial intelligence for human-robot interaction (AI-HRI).

  • Von Ahn, L., Blum, M., Hopper, N. J., & Langford, J. (2003). Captcha: Using hard ai problems for security. In Proceedings of the international conference on the theory and applications of cryptographic techniques (pp. 294–311). Springer

  • Wielemaker, J. (1987). SWI-Prolog documentation: Prolog for the real world. http://www.swi-prolog.org. Accessed 5 Feb 2018.

  • Wielemaker, J., Schrijvers, T., Triska, M., & Lager, T. (2012). SWI-Prolog. Theory and Practice of Logic Programming, 12(1–2), 67–96.

    Article  MathSciNet  MATH  Google Scholar 

  • Williams, T. (2017a). A consultant framework for natural language processing in integrated robot architectures. IEEE Intelligent Informatics Bulletin, 18(1), 10–14.

    Google Scholar 

  • Williams, T. (2017b). Situated natural language interaction in uncertain and open worlds. PhD thesis, Tufts University.

  • Williams, T. (2018a). Toward ethical natural language generation for human-robot interaction. In Late breaking report for the 13th ACM/IEEE international conference on human-robot interaction.

  • Williams, T. (2018b). “Who Should I Run Over?”: Long-term ethical implications of natural language generation. In Proceedings of the 2018 HRI workshop on longitudinal human-robot teaming.

  • Williams, T., & Jackson, B. (2018). A Bayesian analysis of moral norm malleability during clarification dialogues. In Proceedings of the fortieth annual meeting of the cognitive science society.

  • Williams, T., & Scheutz, M. (2015). POWER: A domain-independent algorithm for probabilistic, open-world entity resolution. In Proceedings of the IEEE/RSJ international conference on intelligent robots and systems (IROS) (pp. 1230–1235).

  • Williams, T., & Scheutz, M. (2016a). A framework for resolving open-world referential expressions in distributed heterogeneous knowledge bases. In Proceedings of the thirtieth AAAI conference on artificial intelligence (AAAI), pp 3598–3964.

  • Williams, T., & Scheutz, M. (2016b). Resolution of referential ambiguity using Dempster-Shafer theoretic pragmatics. In Proceedings of the AAAI fall symposium on artificial intelligence for human-robot interaction (AI-HRI).

  • Williams, T., & Scheutz, M. (2017a). Referring expression generation under uncertainty: Algorithm and evaluation framework. In Proceedings of the 10th international conference on natural language generation (INLG).

  • Williams, T., & Scheutz, M. (2017b). Resolution of referential ambiguity in human-robot dialogue using Dempster-Shafer theoretic pragmatics. In Proceedings of robotics: science and systems.

  • Williams, T., Núñez, R. C., Briggs, G., Scheutz, M., Premaratne, K., & Murthi, M. N. (2014). A Dempster-Shafer theoretic approach to understanding indirect speech acts. Advances in Artificial Intelligence.

  • Williams, T., Briggs, G., Oosterveld, B., & Scheutz, M. (2015). Going beyond command-based instructions: Extending robotic natural language interaction capabilities. In: Proceedings of the twenty-ninth AAAI conference on artificial intelligence (AAAI) (pp. 1387–1393).

  • Williams, T., Acharya, S., Schreitter, S., & Scheutz, M. (2016). Situated open world reference resolution for human-robot dialogue. In Proceedings of the eleventh ACM/IEEE international conference on human-robot interaction (HRI) (pp. 311–318).

  • Williams, T., Johnson, C., Scheutz, M., & Kuipers, B. (2017). A tale of two architectures: A dual-citizenship integration of natural language and the cognitive map. In Proceedings of the sixteenth international conference on autonomous agents and multi-agent systems (AAMAS), Sao Paolo, Brazil.

  • Wilson, J. R., Krause, E., Scheutz, M., & Rivers, M. (2016). Analogical generalization of actions from single exemplars in a robotic architecture. In Proceedings of the international conference on autonomous agents and multiagent systems (AAMAS) (pp. 1015–1023).

  • Wyatt, J. (2005). Planning clarification questions to resolve ambiguous references to objects. In Proceedings of the 4th IJCAI workshop on knowledge and reasoning in practical dialogue systems, Edinburgh, Scotland (pp. 16–23).

  • Yazdani, F., Scheutz, M., & Beetz, M. (2017). Guidelines for improving task-based natural language understanding in human-robot rescue teams. In Proceedings of the 2017 8th IEEE international conference on cognitive infocommunications (CogInfoCom), Debrecen, Hungary, accepted for publication.

  • Yoon, S. O., & Brown-Schmidt, S. (2013). Lexical differentiation in language production and comprehension. Journal of Memory and Language, 69(3), 397–416.

    Article  Google Scholar 

  • Zarrieß, S., & Schlangen, D. (2016). Towards generating colour terms for referents in photographs: Prefer the expected or the unexpected? In: Proceedings of the 9th international natural language generation conference (INLG).

  • Zettlemoyer, L. S., & Collins, M. (2012). Learning to map sentences to logical form: Structured classification with probabilistic categorial grammars. It Proceedings of the twenty-first conference on uncertainty in artificial intelligence (UAI).

Download references

Acknowledgements

This work was in part funded by Grant N00014-14-1-0149 from the US Office of Naval Research. The research of Michael Beetz is partly funded by the German science foundation DFG in the context of the collaborative research centre EASE (Everyday Activity Science and Engineering).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tom Williams.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

This is one of several papers published in Autonomous Robots comprising the “Special Issue on Robotics Science and Systems”.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Williams, T., Yazdani, F., Suresh, P. et al. Dempster-Shafer theoretic resolution of referential ambiguity. Auton Robot 43, 389–414 (2019). https://doi.org/10.1007/s10514-018-9795-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10514-018-9795-5

Keywords

Navigation