Abstract
With the advent of Industry 4.0, the demand for estimating target pose keeps increasing. However, the accuracy of the existing pose estimation algorithms for texture-less targets is still poor. Traditional methods require approximately accurate initial pose or else they are easy to fall into a local optimum while the deep learning methods are limited in unconstrained environments where the unpredictable data can not be captured ahead for model training. Therefore, the paper proposes an innovative method which can cover the shortage of these two classes of methods. In our method, a multi-task model which can predict the pose of target and simultaneously obtain the edge map is designed. Then, the predicted pose and edge map are transferred to pose optimization module which is implemented based on edge matching. In addition, considering the lack of the pose datasets for texture-less objects, we design an effective pose dataset generation method based on 3D reconstruction. At last, the proposed system is tested on the public dataset and the rendered dataset. Experimental results demonstrate that the proposed algorithm is more accurate compared with the state-of-the-art methods.
Supported by the National Key R&D Program of China under Grant No. 2018YFB1305300, the National Natural Science Foundation of China under Grant No. 61873189, the Natural Science Foundation of Shanghai under Grant No. 18ZR1442500, Shanghai Municipal Science and Technology Major Project under Grant No. 2021SHZDZX0100 and Shanghai Municipal Commission of Science and Technology Project under Grant No. 19511132101.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Arbeláez, P., Maire, M., Fowlkes, C., Malik, J.: Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 33(5), 898–916 (2011). https://doi.org/10.1109/TPAMI.2010.161
Choi, C., Christensen, H.I.: Real-time 3D model-based tracking using edge and keypoint features for robotic manipulation. In: 2010 IEEE International Conference on Robotics and Automation, pp. 4048–4055 (2010). https://doi.org/10.1109/ROBOT.2010.5509171
Chollet, F.: Xception: Deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251–1258 (2017)
Dong, Y., et al.: Accurate 6dof pose tracking for texture-less objects. IEEE Trans. Circuits Syst. Video Technol. 31(5), 1834–1848 (2021). https://doi.org/10.1109/TCSVT.2020.3011737
Imperoli, M., Pretto, A.: D\(^{2}\)CO: fast and robust registration of 3D textureless objects using the directional chamfer distance. In: Nalpantidis, L., Krüger, V., Eklundh, J.-O., Gasteratos, A. (eds.) ICVS 2015. LNCS, vol. 9163, pp. 316–328. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-20904-3_29
Li, Y., Wang, G., Ji, X., Xiang, Yu., Fox, D.: DeepIM: deep iterative matching for 6D pose estimation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11210, pp. 695–711. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01231-1_42
Liu, M.Y., Tuzel, O., Veeraraghavan, A., Chellappa, R.: Fast directional chamfer matching. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 1696–1703 (2010). https://doi.org/10.1109/CVPR.2010.5539837
Manhardt, F., Kehl, W., Navab, N., Tombari, F.: Deep model-based 6D pose refinement in RGB. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision – ECCV 2018. LNCS, vol. 11218, pp. 833–849. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01264-9_49
Pauwels, K., Rubio, L., Díaz, J., Ros, E.: Real-time model-based rigid object pose estimation and tracking combining dense and sparse visual cues. In: 2013 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2347–2354 (2013). https://doi.org/10.1109/CVPR.2013.304
Poma, X.S., Riba, E., Sappa, A.: Dense extreme inception network: towards a robust CNN model for edge detection. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 1923–1932 (2020)
Prisacariu, V.A., Kähler, O., Murray, D.W., Reid, I.D.: Real-time 3D tracking and reconstruction on mobile phones. IEEE Trans. Visual Comput. Graphics 21(5), 557–570 (2015). https://doi.org/10.1109/TVCG.2014.2355207
Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7263–7271 (2017)
Song, C., Song, J., Huang, Q.: HybridPose: 6D object pose estimation under hybrid representations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 431–440 (2020)
Su, H., Qi, C.R., Li, Y., Guibas, L.J.: Render for CNN: viewpoint estimation in images using CNNs trained with rendered 3D model views. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2686–2694 (2015)
Tekin, B., Sinha, S.N., Fua, P.: Real-time seamless single shot 6D object pose prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 292–301 (2018)
Tjaden, H., Schwanecke, U., Schomer, E., Cremers, D.: A region-based gauss-newton approach to real-time monocular multiple object tracking. IEEE Trans. Pattern Anal. Mach. Intell., 1 (2018). https://doi.org/10.1109/TPAMI.2018.2884990
Tulsiani, S., Malik, J.: Viewpoints and keypoints. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1510–1519 (2015)
Wang, S., Yue, J., Dong, Y., He, S., Wang, H., Ning, S.: A synthetic dataset for visual slam evaluation. Robot. Auton. Syst. 124, 103336 (2020). https://doi.org/10.1016/j.robot.2019.103336
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Xu, S., Gong, P., Dong, Y., Gi, L., Huang, C., Wang, S. (2021). Pose Estimation of Texture-Less Targets for Unconstrained Grasping. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2021. Lecture Notes in Computer Science(), vol 13017. Springer, Cham. https://doi.org/10.1007/978-3-030-90439-5_37
Download citation
DOI: https://doi.org/10.1007/978-3-030-90439-5_37
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-90438-8
Online ISBN: 978-3-030-90439-5
eBook Packages: Computer ScienceComputer Science (R0)