DialogMI: A Dialogue Model Based on Enhancing Dialogue Mutual Information
Yibo Zhang (Beijing University of Posts and Telecommunications); Ping Gong (Beijing University of Posts and Telecommunications); Zelin Wang (Beijing University of Posts and Telecommunications); Zhe Li (Beijing University of Posts and Telecommunications); Xuanyuan Yang (Beijing University of Posts and Telecommunications)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Most of the open-domain dialogue models tend to perform insufficiently in generating informative response. The possible reason is that they lack the capability of enhancing the mutual information between generated responses and dialogue history. To address this issue, we present a novel task of the mutual information enhancement and build a dialogue generation framework DialogMI. Besides, we propose two methods to represent the loss function of the novel task and a method to enhance the effect of the mutual information loss. To the best of our knowledge, we first propose a loss that measures mutual information and use it to assist training model. We conduct experiments on Chinese chit-chat dataset and LCCC dataset for response generation. Results on these datasets indicate that DialogMI can significantly outperform baselines in generating informative dialogue, leading to better response quality and diversity.