Dialogue Systems as Conversational Mentors

Most AI research on language processing has focused on the sentence level, but human language often concerns extended discourse, involves communication among multiple agents, and helps one or more of the agents achieve their goals. Many human interactions involve dialogue between two or more participants who take on established roles, typically alternate between utterances, incrementally elaborate on common ground, omit much of the important material, and infer missing content using knowledge. The ability to carry out extended dialogues is a distinguishing feature of human intelligence.

Our research on task-oriented dialogue has focused on conversational mentors. Such interactive systems attempt to aid a human user who is carrying out some complex goal-directed task. To this end, the systems draw on knowledge about the task's elements and structure, as well as knowledge about the conventions of dialogue. One common application of such mentors is help desks, which often use a form of conversational case-based reasoning (Aha et al., 2006). Another example is Allen et al.'s (2004) TRIPS system, which helped users respond to emergency calls.

Conversational mentors must address a number of issues. Behavior is goal-directed and involves communicative actions about joint activity. Participants must reach common ground for the situation that includes each other's beliefs and goals. However, many beliefs and goals are not explicitly stated but rather inferred by participating agents The overall process alternates between drawing inferences and executing goal-directed activities. This suggests building on Langley's (2017) social cognition hypothesis: intelligence depends on the ability to represent and reason about the mental states of other agents. Human communication requires both understanding and altering others' beliefs and goals, which in turn requires high-level representations and mechanisms to reason over them.

We have developed an architecture for conversational mentors that integrates generic knowledge about dialogues and specific knowledge about a domain to both understand utterances and respond to them. Domain knowledge includes conceptual rules that describe situational patterns and skills that associate conditions, effects with subskills or actions. Both concepts and skills are organized into hierarchies, with complex structures defined in terms of simpler ones. Dialogue expertise includes knowledge about primitive speech acts and relations among them (e.g., questions are followed by answers). The architecture operates in cycles, during which it: observes new speech acts (including its own), uses inference to update beliefs/goals in working memory, and executes skills to produce utterances based on this memory state.

Our main testbed has involved scenarios in which a human medic helps injured teammates with system assistance. The medic has limited training but can provide situational information and affect the environment. On the other hand, the system has medical expertise, but cannot sense or alter the environment directly; it can only offer instructions. The medic and system collaborate to achieve their shared goal of helping the injured person. The system uses a Web interface similar to a messaging application, which let us address high-level aspects of dialogue without tackling other important challenges, such as speech recognition.

This work was funded by the Office of Naval Research through Grants N00014-09-1-1029 and N00014-10-1-0487. Alfredo Gabaldon, Ben Meadows, and Richard Heald contributed substantially to the effort. A gift from Ericsson provided additional support for research done jointly with Ted Selker.

Related Publications

Langley, P. (2017). Interactive cognitive systems and social intelligence. IEEE Intelligent Systems, 32, 22-30.

Gabaldon, A., Langley, P., Meadows, B., & Selker, T. (2016). Knowledge-guided interpretation and generation of task-oriented dialogue. In A. Rudnicky, A. Raux, I. Lane, & T. Misu (Eds.), Situated dialog in speech-based human-computer interaction. Berlin: Springer.

Gabaldon, A., & Langley, P. (2015). Dialogue understanding in a logic of action and belief. Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence (pp. 515-521). Austin, TX: AAAI Press.

Gabaldon, A., Langley, P., & Meadows, B. (2014). Integrating meta-level and domain-level knowledge for task-oriented dialogue. Advances in Cognitive Systems, 3, 201-219.

Langley, P., Meadows, B., Gabaldon, A., & Heald, R. (2014). Abductive understanding of dialogues about joint activities. Interaction Studies, 15, 426-454.

Gabaldon, A., Meadows, B., & Langley, P. (2014). Integrated interpretation and generation of task-oriented dialog. Proceedings of the Fifth International Workshop on Spoken Dialog Systems. Napa, CA.

Gabaldon, A., Langley, P., & Meadows, B. (2013). Integrating meta-level and domain-level knowledge for interpretation and generation of task-oriented dialogue. Proceedings of the Second Annual Conference on Advances in Cognitive Systems. Baltimore, MD.

For more information, send electronic mail to patrick.w.langley@gmail.com

© 1997 Institute for the Study of Learning and Expertise. All rights reserved.