Abstract
In this paper, we propose a method to generate handwritten style characters by extracting the characteristics of a person’s handwritten characters. We use DCGAN (Deep Convolutional Generative Adversarial Networks) to generate handwritten style characters, which preserves the unique characteristics of a person’s handwritten characters. Some methods to generate handwritten style characters have been proposed so far. However, there are too many Kanji characters to prepare handwritten characters as training data; therefore, it is impossible to generate handwritten style characters by extracting the characteristics of a person’s handwritten characters including Kanji characters. In facts, since there are 2,136 Kanji characters for regular use, these would require a large amount of training time and training effort. Our method generates characters that have never been written before by decomposing and extracting Kanji characters into elements, learns each element separately, generates each element, and merges them. Therefore, the system implemented by the method can learn from a limited number of handwritten characters and automatically generate characters that do not exist in the training data, and can generate characters that preserve the characteristics of a person’s handwritten characters. Thus, the number of handwritten style characters that can be generated automatically can be significantly increased.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mayr, M., Stumpf, M., Nicolaou, A., Seuret, M., Maier, A., Christlein, V.: Spatio-temporal handwriting imitation. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12539, pp. 528–543. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-68238-5_38
Fogel, S., Averbuch-Elor, H., Cohen, S., Mazor, S., Litman, R.: ScrabbleGAN: semi-supervised varying length handwritten text generation. In: Proceedings Institute of Electrical and Electronics Engineers, CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–18 June 2020, pp. 4323–4332 (2020)
Varga, T., Kilchhofer, D., Bunke, H.: Template-based synthetic handwriting generation for the training of recognition systems. In: Proceedings of 12th Conference of the International Graphonomics Society, Salerno, Italy, 26–29 June 2005, pp. 206–211 (2005)
Zong, A., Zhu, Y.: StrokeBank: automating personalized Chinese handwriting generation. In: Proceedings of 28th AAAI Conference on Artificial Intelligence, Québec, Canada, 27–31 July 2014, pp. 3024–3029 (2014)
Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. In: Proceedings of International Conference on Learning Representations 2016, San Juan, Puerto Rico, 2–4 May 2016 (2013)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Yang, J., Li, T., Liang, G., He, W., Zhao, Y.: A simple recurrent unit model based intrusion detection system with DCGAN. In: Proceedings of Institute of Electrical and Electronics Engineers Access, vol. 7, pp. 83286-83296 (2019)
Wu, Q., Chen, Y., Meng, J.: DCGAN-based data augmentation for tomato leaf disease identification. IEEE Access 8, 98716–98728 (2020)
Esser, P., Sutter, E., Ommer, B.: A variational U-net for conditional appearance and shape generation. In: Proceedings of Institute of Electrical and Electronics Engineers, Lake City, UT, USA, 12 April 2018, pp. 8857–8866 (2018)
Çiçek, Ö., Abdulkadir, A., Lienkamp, S.S., Brox, T., Ronneberger, O.: 3D U-net: learning dense volumetric segmentation from sparse annotation. In: Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W. (eds.) MICCAI 2016. LNCS, vol. 9901, pp. 424–432. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46723-8_49
Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: UNet++: a nested u-net architecture for medical image segmentation. In: Stoyanov, D., et al. (eds.) DLMIA/ML-CDS -2018. LNCS, vol. 11045, pp. 3–11. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00889-5_1
JiHyeong, L., JeongSeo, C., Koyama, S., Hibino, H.: Change of the impression by the letter thickness-in the cases of hiragana and katakana of ming and gothic styles. Sci. Des. 63(5), 101–108 (2017)
Vianna, P., Farias, R., de Albuquerque Pereira, W.C.: U-Net and SegNet performances on lesion segmentation of breast ultrasonography images. Res/ Biomed. Eng. 37(2), 171–179 (2021). https://doi.org/10.1007/s42600-021-00137-4
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Shimomura, H., Miwa, H. (2023). Automatic Generation of Handwritten Style Characters Including Untrained Characters. In: Barolli, L. (eds) Advances in Intelligent Networking and Collaborative Systems. INCoS 2023. Lecture Notes on Data Engineering and Communications Technologies, vol 182. Springer, Cham. https://doi.org/10.1007/978-3-031-40971-4_2
Download citation
DOI: https://doi.org/10.1007/978-3-031-40971-4_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-40970-7
Online ISBN: 978-3-031-40971-4
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)