Journal of Physics: Conference Series 2019
Chanjian Gu, Changsheng Lu, Chanchen Gu and Xinping Guan
Viewpoint estimation is a fundamental procedure in vision-based robot tasks. A good viewpoint of the camera relative to objects can help the visual system perform better both in observation and manipulation. Recently, CNN-based algorithms, which can effectively extract discriminative features from images in challenging conditions are utilized to handle the viewpoint estimation problem. However, most existing algorithms focus on how to leverage the extracted deep features while neglecting the spatial relationship among images that captured from various viewpoints. In this paper, we present a deep metric learning method for solving the viewpoint estimation problem. A triplet loss with a novel viewpoint-based input selection strategy is introduced, which could learn more powerful features after incorporating the spatial relationship between viewpoints. Combined with the traditional classification loss, the presented loss can further enhance the discriminative power of features. To evaluate the performance of our method, a dataset containing a large number of images generated from five different texture-less workpieces is built and the experiment results show the effectiveness of the proposed method.