ESTIMATION OF VISUAL CONTENTS BASED ON QUESTION ANSWERING FROM HUMAN BRAIN ACTIVITY
Saya Takada, Ren Togo, Takahiro Ogawa, Miki Haseyama
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 13:47
We propose an estimation method for free-form Visual Question Answering (VQA) from human brain activity, \textit{brain decoding VQA}. The task of VQA in the field of computer vision is generating an answer given an image and a question about its contents. The proposed method can realize answering arbitrary visual questions about images from brain activity measured by functional Magnetic Resonance Imaging (fMRI) while viewing the same images. We enable estimating various information from brain activity via a unique VQA model which can realize more detailed understanding of images and complex reasoning. In addition, we newly make use of un-labeled images not used in the training phase to improve the performance of the transformation, since fMRI datasets are generally small. The proposed method can answer a visual question from a little amount of fMRI data measured while subjects are viewing images.