Abstract
In recent years, due to the application of neural networks in music generation, a lot of music has appeared on the Internet, which poses more challenges for screening music and selecting songs. Although subjective evaluation is currently the ultimate choice for evaluating the beauty of songs, this method has limitations due to human resources and efficiency issues. Therefore, this paper combines information science and music science to propose an aesthetic computing framework for music composition on the Internet. Firstly, more accurate basic and advanced features of music are extracted by transcribing the separated musical accompaniment. Next, we will match them to the music rules. Then we use appropriate merging rules to determine the weight of elements, so as to achieve the purpose of calculating the aesthetic feeling of music composition. This paper presents for the first time a musical aesthetic evaluation framework combining music information dynamics, audio transcription and Zipf’wlaw. The experimental results prove that the objective beauty calculation method proposed is feasible and effective.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
2021 Annual White Paper of Chinese Digital Music. Tencent Music Data Research Institute (2022)
Ji, S., Luo, J., Yang, X.: A comprehensive survey on deep music generation: multi-level representations, algorithms, evaluations, and future directions (2020)
Tikhonov, A., Yamshchikov, I.P.: Music generation with variational recurrent autoencoder supported by history. arXiv, abs/1705.05458 (2020)
Yang, L.-C., Lerch, A.: On the evaluation of generative models in music. Neural Comput. Appl. 32(9), 4773–4784 (2018). https://doi.org/10.1007/s00521-018-3849-7
Iqbal, A.: Computational aesthetics. Encyclopedia Britannica (2015)
Wu, Y.-T., Chen, B., Su, L.: Multi-instrument automatic music transcription with self-attention-based instance segmentation. IEEE/ACM Trans. Audio Speech Lang. Process. 28, 2796–2809 (2020). https://doi.org/10.1109/TASLP.2020.3030482
Theis, L., et al.: A note on the evaluation of generative models. CoRR abs/1511.01844 (2016). n. pag
Dong, H.-W., et al.: MuseGAN: symbolic-domain music generation and accompaniment with multi-track sequential generative adversarial networks. arXiv abs/1709.06298 (2017). n. pag
Huang, C.-Z.A., et al.: Counterpoint by convolution. In: ISMIR (2017)
Johnson, D.D.: Generating polyphonic music using tied parallel networks. In: Correia, J., Ciesielski, V., Liapis, A. (eds.) EvoMUSART 2017. LNCS, vol. 10198, pp. 128–143. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-55750-2_9
Gillick, J., et al.: Learning to groove with inverse sequence transformations. In: ICML (2019)
Chuan, C., Herremans, D.: Modeling temporal tonal relations in polyphonic music through deep networks with a novel image-based representation. In: AAAI (2018)
Jin, C., et al.: A transformer generative adversarial network for multi‐track music generation. CAAI Trans. Intell. Technol. 7(3), 369–380 (2022)
Hadjeres, G., Pachet, F., Nielsen, F.: Deepbach: a steerable model for bach chorales generation. In: International Conference on Machine Learning. PMLR (2017)
Jin, C., et al.: A style-specific music composition neural network. Neural Process. Lett. 52(3), 1893–1912 (2020)
Cemgil, A.T.: Bayesian music transcription (2004)
Benetos, E., Dixon, S., Giannoulis, D., Kirchhoff, H., Klapuri, A.: Automatic music transcription: challenges and future directions. J. Intell. Inf. Syst. 41(3), 407–434 (2013). https://doi.org/10.1007/s10844-013-0258-3
Klapuri, A., Davy, M., (eds.): Signal processing methods for music transcription (2007)
Nan, N., et al.: Common quantitative characteristics of music melodies - pursuing the constrained entropy maximization casually in composition. Sci. China Inf. Sci. 65, 1–3 (2022)
Kong, Q., et al.: GiantMIDI-piano: a large-scale MIDI dataset for classical piano music. Trans. Int. Soc. Music. Inf. Retr. 5, 87–98 (2022)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Wang, S., Tie, Y., Li, X., Wang, X., Qi, L. (2023). Intelligence Evaluation of Music Composition Based on Music Knowledge. In: Huang, DS., Premaratne, P., Jin, B., Qu, B., Jo, KH., Hussain, A. (eds) Advanced Intelligent Computing Technology and Applications. ICIC 2023. Lecture Notes in Computer Science(), vol 14090. Springer, Singapore. https://doi.org/10.1007/978-981-99-4761-4_32
Download citation
DOI: https://doi.org/10.1007/978-981-99-4761-4_32
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-4760-7
Online ISBN: 978-981-99-4761-4
eBook Packages: Computer ScienceComputer Science (R0)