Abstract
The classification and grasping of randomly placed objects where only a limited number of training images are available, remains a challenging problem. Approaches such as data synthesis have been used to synthetically create larger training data sets from a small set of training data and can be used to improve performance. This paper examines how limited product images for ‘off the shelf’ items can be used to generate a synthetic data set that is used to train a that allows classification of the item, segmentation and grasping. Experiments investigating the effects of data synthesis are presented and the subsequent trained network implemented in a robotic system to perform grasping of objects.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Ciregan, D., Meier, U., Schmidhuber, J.: Multi-column deep neural networks for image classification. In: 2012 IEEE conference on Computer vision and pattern recognition (CVPR), pp. 3642–3649. IEEE (2012)
Dragan, A.D., Gordon, G.J., Srinivasa, S.S.: Learning from experience in manipulation planning: setting the right goals. In: Christensen, H.I., Khatib, O. (eds.) Robotics Research. STAR, vol. 100, pp. 309–326. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-29363-9_18
Dwibedi, D., Misra, I., Hebert, M.: Cut, paste and learn: surprisingly easy synthesis for instance detection. ArXiv 1(2), 3 (2017)
Fang, K., Bai, Y., Hinterstoisser, S., Kalakrishnan, M.: Multi-task domain adaptation for deep learning of instance grasping from simulation. arXiv preprint arXiv:1710.06422 (2017)
Georgakis, G., Mousavian, A., Berg, A.C., Kosecka, J.: Synthesizing training data for object detection in indoor scenes. arXiv preprint arXiv:1702.07836 (2017)
He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2980–2988. IEEE (2017)
Huebner, K., Ruthotto, S., Kragic, D.: Minimum volume bounding box decomposition for shape approximation in robot grasping. In: IEEE International Conference on Robotics and Automation, ICRA 2008, pp. 1628–1633. IEEE (2008)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems, pp. 1097–1105 (2012)
Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. Int. J. Robot. Res. 34(4–5), 705–724 (2015)
Levine, S., Pastor, P., Krizhevsky, A., Quillen, D.: Learning hand-eye coordination for robotic grasping with large-scale data collection. In: Kulić, D., Nakamura, Y., Khatib, O., Venture, G. (eds.) ISER 2016. SPAR, vol. 1, pp. 173–184. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-50115-4_16
Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740–755. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10602-1_48
Milan, A., et al.: Semantic segmentation from limited training data. arXiv preprint arXiv:1709.07665 (2017)
Papandreou, G., et al..: Towards accurate multiperson pose estimation in the wild. arXiv preprint arXiv:1701.01779 (2017)
Redmon, J., Angelova, A.: Real-time grasp detection using convolutional neural networks. In: 2015 IEEE International Conference on Robotics and Automation (ICRA), pp. 1316–1322. IEEE (2015)
Richtsfeld, M., Vincze, M.: Grasping of unknown objects from a table top. In: Workshop on Vision in Action: Efficient Strategies for Cognitive Agents in Complex Environments (2008)
Tobin, J., Zaremba, W., Abbeel, P.: Domain randomization and generative models for robotic grasping. arXiv preprint arXiv:1710.06425 (2017)
Xie, L., Wang, J., Wei, Z., Wang, M., Tian, Q.: Disturblabel: Regularizing CNN on the loss layer. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4753–4762 (2016)
Acknowledgments
With thanks to the James Dyson Foundation Undergraduate Bursary and also the EPSRC CDT in Sensor Technologies (Grant EP/L015889/1).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Cheah, M., Hughes, J., Iida, F. (2018). Data Synthesization for Classification in Autonomous Robotic Grasping System Using ‘Catalogue’-Style Images. In: Giuliani, M., Assaf, T., Giannaccini, M. (eds) Towards Autonomous Robotic Systems. TAROS 2018. Lecture Notes in Computer Science(), vol 10965. Springer, Cham. https://doi.org/10.1007/978-3-319-96728-8_4
Download citation
DOI: https://doi.org/10.1007/978-3-319-96728-8_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-96727-1
Online ISBN: 978-3-319-96728-8
eBook Packages: Computer ScienceComputer Science (R0)