Vision And Text Transformer For Predicting Answerability On Visual Question Answering
Tung Le, Huy Tien Nguyen, Minh Le Nguyen
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:06:49
Answerability on Visual Question Answering is a novel and attractive task to predict answerable scores between images and questions in multi-modal data. Existing works often utilize a binary mapping from visual question answering systems into Answerability. It does not reflect the essence of this problem. Together with our consideration of Answerability in a regression task, we propose VT-Transformer, which exploits visual and textual features through Transformer architecture. Experimental results on VizWiz 2020 dataset show the effectiveness and robustness of VT-Transformer for Answerability on Visual Question Answering when comparing with competitive baselines.