Abstract
Shadow performs an important role in the image, which can enhance the image effect and convey important visual clues. We propose a method based on deep learning to automatically generate stylized shadows for line drawings. Based on StarGAN, a shadow generation adversarial network (ShadowGAN) is designed, which can automate the creation of stylized shadows with different light directions. This method defines eight light directions. Users can select one of the eight light directions around the 2D image to specify the light source according to the encoding of the light direction, and generate the shadow corresponding to the light direction. We use a new dataset containing line drawings with shadows and label information corresponding to the light direction. Experiments show that our method can generate stylized shadows for line drawing with satisfactory quality, which can simplify the user’s workflow, and save the time of drawing line drawing shadows.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Anjyo, K., Wemler, S., Baxter, W.: Tweakable light and shade for cartoon animation. In: Proceedings of the 4th International Symposium on Non-photorealistic Animation and Rendering, PP. 133–139 (2006)
Zhang, L., Jiang, J., Ji, Y., Liu, C.: SmartShadow: artistic shadow drawing tool for line drawings. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, PP. 5391–5400 (2021)
Casanova, A., Careil, M., Verbeek, J., Drozdzal, M., Romero-Soriano, A.: Instance conditioned GAN. In: Advances in Neural Information Processing Systems, vol. 34, pp. 27517–27529 (2021)
Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, PP. 8789–8797 (2018)
Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, PP. 1125–1134 (2017)
Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, PP. 2223–2232 (2017)
Mirza, M., Osindero, S.: Conditional generative adversarial nets. arxiv.org/abs/1411.1784 (2014)
Zheng, Q., Li, Z., Bargteil, A.: Learning to shadow hand-drawn sketches. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, PP. 7436–7445 (2020)
Hudon, M., Lutz, S., Pagés, R., Smolic, A.: Augmenting hand-drawn art with global illumination effects through surface inflation. In: Proceedings of the 16th ACM SIGGRAPH European Conference on Visual Media Production, PP. 1–9 (2019)
Zhang, L., Simo-Serra, E., Ji, Y., Liu, C.: Generating digital painting lighting effects via RGB-space geometry. ACM Trans. Graph. 39(2), 1–13 (2020)
Hudon, M., Grogan, M., Pagés, R., Smolić, A.: Deep normal estimation for automatic shading of hand-drawn characters. In: Leal-Taixé, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11131, pp. 246–262. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-11015-4_20
Petikam, L., Anjyo, K., Rhee, T.: Shading rig: dynamic art-directable stylised shading for 3D characters. ACM Trans. Graph. (TOG) 40(5), 1–14 (2021)
DeCoro, C., Cole, F., Finkelstein, A., Rusinkiewicz, S.: Stylized shadows. In: Proceedings of the 5th International Symposium on Nonphotorealistic Animation and Rendering, PP. 77–83 (2007)
Zhang, L., Ji, Y., Lin, X., Liu, C.: Style transfer for anime sketches with enhanced residual U-net and auxiliary classifier GAN. In: Proceedings of the 4th IAPR Asian Conference on Pattern Recognition, PP. 506–511 (2017)
Wei, T., Zhu, L.: Comic style transfer based on generative confrontation network. In: Proceedings of 2021 6th International Conference on Intelligent Computing and Signal Processing (ICSP), PP. 1011–1014. IEEE (2021)
Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: Proceedings of the 34th International Conference on Machine Learning (ICML), PP. 214–223 (2017)
Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.: Improved training of wasserstein GANs. (2017). arxiv.org/abs/1704.00028
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, PP. 770–778 (2016)
Zhang, H., Goodfellow, I., Metaxas, D., Odena, A.: Self-attention generative adversarial networks. In: Proceedings of the 36th International Conference on Machine Learning, PP. 7354–7363. PMLR (2019)
Li, C., Wand, M.: Precomputed real-time texture synthesis with Markovian generative adversarial networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 702–716. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46487-9_43
Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arxiv.org/abs/1412.6980 (2014)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Xue, H., Kang, C. (2023). ShadowGAN for Line Drawings Shadow Generation. In: Iliadis, L., Papaleonidas, A., Angelov, P., Jayne, C. (eds) Artificial Neural Networks and Machine Learning – ICANN 2023. ICANN 2023. Lecture Notes in Computer Science, vol 14262. Springer, Cham. https://doi.org/10.1007/978-3-031-44201-8_25
Download citation
DOI: https://doi.org/10.1007/978-3-031-44201-8_25
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-44200-1
Online ISBN: 978-3-031-44201-8
eBook Packages: Computer ScienceComputer Science (R0)