Abstract
Annotation data is the “fuel” of vision cognitive system but hard to obtain. We focus on finding a feasible way to generate high-quality image data. The 3D models can produce rich annotated 2D images, and the generative adversarial nets can create various pictures. We proposed the background augmentation generative adversarial nets to build a bridge between GAN and 3D models for data augmentation. As a result, we use BAGAN and 3D models to generate images which can help deep convolutional classifier improve accuracy score to 93.12% on real data test sets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
This work described in “Synthetic data from 3D models for real image classification” which had been submitted in “EURASIP Journal on Image and Video Processing.”
- 2.
References
Chang, A.X., et al.: Shapenet: an information-rich 3D model repository. In: Computer Science (2015)
Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever, I., Abbeel, P.: Infogan: interpretable representation learning by information maximizing generative adversarial nets. In: Neural Information Processing Systems, pp. 2172–2180 (2016)
Everingham, M., Van Gool, L., Williams, C.K.I., Winn, J., Zisserman, A.: The Pascal Visual Object Classes Challenge 2012 (VOC2012) Results. http://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html
Feng, Y., Ren, J., Jiang, J.: Object-based 2D-to-3D video conversion for effective stereoscopic content generation in 3D-tv applications. IEEE Trans. Broadcast. 57(2), 500–509 (2011)
Goodfellow, I.J., et al.: Generative adversarial nets. In: International Conference on Neural Information Processing Systems, pp. 2672–2680 (2014)
Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.C.: Improved training of wasserstein gans. In: Advances in Neural Information Processing Systems, pp. 5769–5779 (2017)
Han, J., Zhang, D., Hu, X., Guo, L., Ren, J., Wu, F.: Background prior-based salient object detection via deep reconstruction residual. IEEE Trans. Circuits Syst. Video Technol. 25(8), 1309–1321 (2015)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: International Conference on Neural Information Processing Systems, pp. 1097–1105 (2012)
Mirza, M., Osindero, S.: Conditional generative adversarial nets. In: Computer Science, pp. 2672–2680 (2014)
Odena, A., Olah, C., Shlens, J.: Conditional image synthesis with auxiliary classifier gans. In: International conference on machine learning, pp. 2642–2651 (2016)
Ren, J., Jiang, J., Wang, D., Ipson, S.S.: Fusion of intensity and inter-component chromatic difference for effective and robust colour edge detection. Image Process. Let 4(4), 294–301 (2010)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: International Conference on Learning Representations (2015)
Szegedy, C., et al.: Going deeper with convolutions. In: Computer Vision and Pattern Recognition, pp. 1–9 (2015)
Wang, Z., et al.: A deep-learning based feature hybrid framework for spatiotemporal saliency detection inside videos. Neurocomputing, 68–83 (2018)
Xiang, Y., et al.: Objectnet3D: a large scale database for 3D object recognition. In: European Conference Computer Vision (ECCV) (2016)
Xiao, J., Hays, J., Ehinger, K.A., Oliva, A., Torralba, A.: Sun database: large-scale scene recognition from abbey to zoo. In: Computer Vision and Pattern Recognition, pp. 3485–3492 (2010)
Zabalza, J.: Corrigendum to novel segmented stacked autoencoder for effective dimensionality reduction and feature extraction in hyperspectral imaging. Neurocomputing 185(C), 1–10 (2016)
Zhao, D., Zheng, J., Ren, J.: Effective removal of artifacts from views synthesized using depth image based rendering. In: The International Conference on Distributed Multimedia Systems, pp. 65–71 (2015)
Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networkss. In: Computer Vision (ICCV) IEEE International Conference on 2017 (2017)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Ma, Y., Liu, K., Guan, Zb., Xu, XK., Qian, X., Bao, H. (2018). Using GAN to Augment the Synthesizing Images from 3D Models. In: Ren, J., et al. Advances in Brain Inspired Cognitive Systems. BICS 2018. Lecture Notes in Computer Science(), vol 10989. Springer, Cham. https://doi.org/10.1007/978-3-030-00563-4_10
Download citation
DOI: https://doi.org/10.1007/978-3-030-00563-4_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-00562-7
Online ISBN: 978-3-030-00563-4
eBook Packages: Computer ScienceComputer Science (R0)