Publication: Towards Social and Interpretable Neural Dialog Systems
No Thumbnail Available
Open/View Files
Date
2020-06-17
Authors
Published Version
Published Version
Journal Title
Journal ISSN
Volume Title
Publisher
The Harvard community has made this article openly available. Please share how this access benefits you.
Citation
Saleh, Abdelrhman. 2020. Towards Social and Interpretable Neural Dialog Systems. Bachelor's thesis, Harvard College.
Research Data
Abstract
Open-domain dialog generation is a task that challenges machines to mimic human conversations. Despite the remarkable progress natural language generation has seen over the past several years, open-domain dialog systems still suffer from limitations that hinder their adoption in the real world. Systems trained with maximum likelihood often generate dull and repetitive responses, ignoring user input. Training on standard datasets from online forums leads to the generation of inappropriate, biased, or toxic responses. And models rarely exhibit long-term coherence across multiple dialog turns. Meanwhile, the predominant approach to dialog generation relies on black-box neural networks which provide little insight as to what information they learn (or do not learn) about engaging in dialog.
In light of these issues, this thesis makes two contributions to building social and interpretable dialog systems. The first part of this thesis proposes a novel reinforcement learning approach for improving the social capabilities of open-domain dialog systems. We optimize for human-centered objectives such as response politeness, diversity, coherence, and sentiment. Our interactive human evaluation shows that these objectives can improve the quality of human-AI interaction and increase user engagement.
The second part of this thesis investigates the conversational understanding captured by neural dialog systems using probing. Our results suggest that standard open-domain dialog systems struggle with basic skills such as answering questions, inferring contradiction, and determining the topic of conversation. We also find that the dyadic, turn-taking nature of dialog is not fully leveraged by these models. By exploring these limitations, we highlight the need for additional research into architectures and training methods that can allow for capturing high-level information about natural language.
Description
Other Available Sources
Keywords
Terms of Use
This article is made available under the terms and conditions applicable to Other Posted Material (LAA), as set forth at Terms of Service