TCRNET: MAKE TRANSFORMER, CNN AND RNN COMPLEMENT EACH OTHER
Xinxin Shan, Tai Ma, Anqi Gu, Haibin Cai, Ying Wen
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:05:17
Recently, several Transformer-based methods have been presented to improve image segmentation. However, since Transformer needs regular square images and has difficulty in obtaining local feature information, the performance of image segmentation is seriously affected. In this paper, we propose a novel encoder-decoder network named TCRNet, which makes Transformer, Convolutional neural network (CNN) and Recurrent neural network (RNN) complement each other. In the encoder, we extract and concatenate the feature maps from Transformer and CNN to effectively capture global and local feature information of images. Then in the decoder, we utilize convolutional RNN in the proposed recurrent decoding unit to refine the feature maps from the decoder for finer prediction. Experimental results on three medical datasets demonstrate that TCRNet effectively improves the segmentation precision.