SELF-ATTENTION FOR INCOMPLETE UTTERANCE REWRITING
Yong Zhang, Zhitao Li, Jianzong Wang, Ning Cheng, Jing Xiao
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:10:58
Incomplete utterance rewriting (IUR) has recently become an essential task in NLP, aiming to complement the incomplete utterance with sufficient context information for comprehension. In this paper, we propose a novel method by directly extracting the coreference and omission relationship from the self-attention weight matrix of the transformer instead of word embeddings and edit the original text accordingly to generate the complete utterance. Benefiting from the rich information in the self-attention weight matrix, our method achieved competitive results on public IUR datasets.