Training An Embedded Object Detector For Industrial Settings Without Real Images
Julia Cohen, Carlos Crispim-Junior, Jean-Marc Chiappa, Laure Tougne
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:05:37
In an industrial environment, object detection is a challenging task due to the absence of real images and real-time requirements for the object detector, usually embedded in a mobile device. Using 3D models, it is however possible to create a synthetic dataset to train a neural network, although the performance on real images is limited by the domain gap. In this paper, we study the performance of a Convolutional Neural Network (CNN) designed to detect objects in real-time: Single-Shot Detector (SSD) with a Mobilenet backbone. We train SSD with synthetic images only and apply extensive data augmentation to reduce the domain gap between synthetic and real images. On the T-LESS dataset, SSD performs better than Mask R-CNN trained on the same synthetic images, with MobilenetV2 and MobilenetV3 Large as backbone. Our results also show the huge improvement enabled by an adequate augmentation strategy.