Abstract
Ukiyo-e culture has endured throughout Japanese art history to this day. With its high artistic value, ukiyo-e remains an important part of art history. Possibly more than one million ukiyo-e prints have been collected by institutions and individuals worldwide. Many public ukiyo-e databases of various scales have been created in different languages. The sharing of ukiyo-e culture could advance to a new stage if the information from all the databases could be shared without differences in information. However, understanding different languages in different databases, redundant data, missing data, uncertain data, and inconsistent data are all barriers to knowledge discovery in each database. Therefore, this paper uses Ukiyo-e Portal Database [1] prints that were released from the Art Research Center (ARC) of Ritsumeikan University as examples, explains the challenges that are currently solvable, and proposes a multi-source artwork information embedding framework for multimodal and multilingual retrieval.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsNotes
- 1.
Google, The Ukiyo-e Artists You Need To Know, https://artsandculture.google.com/story/the-ukiyo-e-artists-you-need-to-know/BQKC6o0k2oBRLA.
- 2.
- 3.
References
ARC Ukiyo-e Portal Database. https://www.dh-jac.net/db/nishikie/search_portal.php?&lang=en. Accessed 25 June 2020
ARC Japanese Woodblocks Prints. https://www.dh-jac.net/db/nishikie/search.php?enter=default&lang=en. Accessed 25 June 2020
Li, K.Y., Batjargal, B., Maeda, A.: Character segmentation in collector’s seal images: an attempt on retrieval based on ancient character typeface. In: Proceedings of the 5th International Workshop on Computational History (HistoInformatics 2019), pp. 40–49 (2019)
Messina, P., Dominguez, V., Parra, D., Trattner, C., Soto, A.: Content-based artwork recommendation: integrating painting metadata with neural and manually-engineered visual features. User Model. User-Adapt. Interact. 29, 251–290 (2019)
Benouaret, I., Lenne, D.: Personalizing the museum experience through context-aware recommendations. In: Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 743–748 (2015)
He, R., Fang, C., Wang, Z., McAuley, J.: Vista: a visually, socially, and temporally-aware model for artistic recommendation. In: Proceedings of the 10th ACM Conference on Recommender Systems, RecSys 2016, pp. 309–316 (2016)
Barkan, O., Koenigstein, N.: Item2Vec: neural item embedding for collaborative filtering. In: IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP). IEEE (2016)
Pires, T., Schlinger, E., Garrette, D.: How multilingual is Multilingual BERT? arXiv preprint arXiv:1906.01502 (2019)
Devlin, J., Chang, M., Lee, M., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL (2019)
Liu, C.L., Hsu, T.Y., Chuang, Y.S., Lee, H.Y.: A study of cross-lingual ability and language-specific information in multilingual BERT. arXiv Preprint arXiv:2004.09205 (2020)
PySparNN. https://github.com/facebookresearch/pysparnn. Accessed 25 June 2020
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Li, K., Batjargal, B., Maeda, A., Akama, R. (2020). Artwork Information Embedding Framework for Multi-source Ukiyo-e Record Retrieval. In: Ishita, E., Pang, N.L.S., Zhou, L. (eds) Digital Libraries at Times of Massive Societal Transition. ICADL 2020. Lecture Notes in Computer Science(), vol 12504. Springer, Cham. https://doi.org/10.1007/978-3-030-64452-9_23
Download citation
DOI: https://doi.org/10.1007/978-3-030-64452-9_23
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-64451-2
Online ISBN: 978-3-030-64452-9
eBook Packages: Computer ScienceComputer Science (R0)