RETHINKING LONG-TAILED VISUAL RECOGNITION WITH DYNAMIC PROBABILITY SMOOTHING AND FREQUENCY WEIGHTED FOCUSING
Wan Jun Nah, Chun Chet Ng, Che-Tsung Lin, Yeong Khang Lee, Jie Long Kew, Zhi Qin Tan, Chee Seng Chan, Zach Christopher, Shang-Hong Lai
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Deep learning models trained on long-tailed (LT) datasets often exhibit bias towards head classes with high frequency. This paper highlights the limitations of existing solutions that combine class- and instance-level re-weighting loss in a naive manner. Specifically, we demonstrate that such solutions result in overfitting the training set, significantly impacting the rare classes. To address this issue, we propose a novel loss function that dynamically reduces the influence of outliers and assigns class-dependent focusing parameters. We also introduce a new long-tailed dataset, ICText-LT, featuring various image qualities and greater realism than artificially sampled datasets. Our method has proven effective, outperforming existing methods through superior quantitative results on CIFAR-LT, Tiny ImageNet-LT, and our new ICText-LT datasets. The source code and new dataset are available at https://github.com/nwjun/FFDS-Loss.