Mitigating Dataset Bias in Image Captioning through CLIP Confounder-free Captioning Network
YeonJu Kim, Junho Kim, Byung-Kwan Lee, Sebin Shin, Yong Man Ro
-
SPS
IEEE Members: $11.00
Non-members: $15.00
The dataset bias has been identified as a major challenge in image captioning. When the image captioning model predicts a word, it should consider the visual evidence associated with the word, but the model tends to use contextual evidence from the dataset bias and results in biased captions, especially when the dataset is biased toward some specific situations. To solve this problem, we approach from the causal inference perspective and design a causal graph. Based on the causal graph, we propose a novel method named C2Cap which is CLIP confounder-free captioning network. We use the global visual confounder to control the confounding factors in the image and train the model to produce debiased captions. We validate our proposed method on MSCOCO benchmark and demonstrate the effectiveness of our method.