Skip to main content
Log in

MISNA - A musical instrument segregation system from noisy audio with LPCC-S features and extreme learning

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Technology has developed a lot over the last decades and has made a profound impact in almost every field. The field of Music Information Retrieval (MIR) has not been an exception to this as well, one of its most promising applications being Automatic Music Transcription (AMT). It is important to identify the active regions of various Instruments in a piece before transcription and the challenge elevates even more when the audio clips are contaminated with noise. MISNA (Musical Instrument Segregation from Noisy Clips) is a system proposed towards the identification of isolated Instruments from noisy clips which can aid towards AMT in noisy environments. The system works using statistical features (LPCC-S) derived from raw Linear Predictive Cepstral Coefficient values on very short clips of lengths 1 and 2 seconds. The system has been tested for various SNR scenarios and highest accuracies of 98.63% and 97.42% for Individual Instruments and Instrument Family identification has been obtained with the aid of Extreme Learning based classifier for a highest of 2626 clips.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Agostini G, Longari M, Pollastri E (2003) Musical instrument timbres classification with spectral features. EURASIP J Appl Signal Process 2003:5–14

    Google Scholar 

  2. Benetos E, Kotti M, Kotropoulos C (2007) Large scale musical instrument identification. In: 4th Sound and music computing conference, pp 283–286

  3. Biernacki A (2017) Analysis and modelling of traffic produced by adaptive HTTP-based video. Multimed Tools Appl 76(10):12347–12368

    Article  Google Scholar 

  4. Demṡar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30

    MathSciNet  Google Scholar 

  5. Deshmukh S, Bhirud S (2014) Analysis and application of audio features extraction and classification method to be used for North Indian Classical Musics singer identification problem. Int J Adv Res Comput Commun Eng, 3(2)

  6. Donnelly PJ, Sheppard JW (2013) Classification of musical timbre using bayesian networks. Comput Music J 37(4):70–86

    Article  Google Scholar 

  7. Eronen A, Klapuri A (2000) Musical instrument recognition using cepstral coefficients and temporal features. In: 2000 IEEE International conference on acoustics, speech, and signal processing, 2000. ICASSP’00. Proceedings, vol 2. IEEE, pp II753–II756

  8. Fragoulis D, Papaodysseus C, Exarhos M, Roussopoulos G, Panagopoulos T, Kamarotos D (2006) Automated classification of piano-guitar notes. IEEE Trans Audio Speech Lang Process 14(3):1040–1050

    Article  Google Scholar 

  9. Huang GB (2014) An insight into extreme learning machines: random neurons, random features and kernels. Cogn Comput 6(3):376–390

    Article  Google Scholar 

  10. Huang GB, Zhou H, Ding X, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Trans Syst Man Cybern Part B (Cybern) 42(2):513–529

    Article  Google Scholar 

  11. Huang GB, Bai Z, Kasun LLC, Vong CM (2015) Local receptive fields based extreme learning machine. IEEE Comput Intell Mag 10(2):18–29

    Article  Google Scholar 

  12. Jadhav PS (2015) Classification of musical instruments sounds by using MFCC and Timbral audio descriptors. Int J Recent Innov Trends Comput Commun, 3(7)

  13. Jitpakdee P, Uyyanonvara B (2017) Computer-aided detection and quantification in glistenings on intra-ocular lenses. Multimed Tools Appl, 1–14

  14. Kaminsky I, Materka A (1995) Automatic source identification of monophonic musical instrument sounds. In: IEEE International Conference on neural networks, 1995. Proceedings., vol 1. IEEE, pp 189–194

  15. Kaminskyj I, Czaszejko T (2005) Automatic recognition of isolated monophonic musical instrument sounds using kNNC. J Intell Inf Syst 24(2):199–221

    Article  Google Scholar 

  16. Kitahara T, Goto M, Okuno HG (2005) Pitch-dependent identification of musical instrument sounds. Appl Intell 23(3):267–275

    Article  Google Scholar 

  17. Lita AI, Ionescu LM, Mazare AG, Serban G, Lita I (2016) Real time system for instrumental sound extraction and recognition. In: 2016 39th International spring seminar on electronics technology (ISSE). IEEE, pp 456–461

  18. Liu J, Xie L (2010) Comparison of performance in automatic classification between Chinese and Western musical instruments. In: 2010 WASE International conference on information engineering (ICIE), vol 1. IEEE, pp 3–6

  19. Livshin A, Rodet X (2004) Musical instrument identification in continuous recordings. In: Digital audio effects 2004, pp 1–1

  20. Livshin A, Rodet X (2009) Purging musical instrument sample databases using automatic musical instrument recognition methods. IEEE Trans Audio Speech Lang Process 17(5):1046–1051

    Article  Google Scholar 

  21. Martin KD, Kim YE (1998) 2pMU9. Musical instrument identification: a pattern-recognition approach. In: Presented at the 136th meeting of the acoustical society of America

  22. Masood S, Gupta S, Khan S (2015) Novel approach for musical instrument identification using neural network. In: 2015 Annual IEEE on India conference (INDICON). IEEE, pp 1–5

  23. Mukherjee H, Rakshit P, Phadikar S, Roy K (2016) REARC-A Bangla phoneme recognizer. In: 2016 International conference on accessibility to digital World (ICADW). IEEE, pp 177–180

  24. Mukherjee H, Halder C, Phadikar S, Roy K (2017) READ-A Bangla phoneme recognition system. In: Proceedings of the 5th international conference on frontiers in intelligent computing: theory and applications. Springer, Singapore, pp 599–607

    Google Scholar 

  25. Patil SD, Pattewar TM (2015) Musical instrument identification using SVM & MLP with formal concept analysis. In: 2015 International Conference on green computing and internet of things (ICGCIoT). IEEE, pp 936–939

  26. Petruncio D, Hasegawa-Johnson MA (2002) Evaluation of various features for music genre classification with hidden Markov models. University of Illinois, Master’s thesis

    Google Scholar 

  27. Rai A, Singh HV (2017) SVM based robust watermarking for enhanced medical image security. Multimed Tools Appl, 1–14

  28. Ri CY, Yao M (2015) Bayesian network based semantic image classification with attributed relational graph. Multimed Tools Appl 74(13):4965–4986

    Article  Google Scholar 

  29. Röver C, Klefenz F, Weihs C (2005) Identification of musical instruments by means of the Hough-transformation. In: Classification—the ubiquitous challenge. Springer, Berlin, pp 608–615

  30. Sturm BL, Morvidone M, Daudet L (2010) Musical instrument identification using multiscale mel-frequency cepstral coefficients. In: 2010 18th European signal processing conference. IEEE, pp 477–481

  31. Takahashi Y, Kondo K (2014) Comparison of two classification methods for Musical Instrument identification. In: 2014 IEEE 3rd Global conference on consumer electronics (GCCE). IEEE, pp 67– 68

  32. Tang J, Deng C, Huang GB (2016) Extreme learning machine for multilayer perceptron. IEEE Trans Neural Netw Learn Syst 27(4):809–821

    Article  MathSciNet  Google Scholar 

  33. Yu F, Chen Y (2015) Musical instrument classification based on improved matching pursuit with instrument-specific atoms. In: 2015 IIAI 4th International congress on advanced applied informatics (IIAI-AAI). IEEE, pp 506–510

  34. Yu J, Chen X, Yang D (2008) Chinese folk musical instruments recognition in polyphonic music. In: International Conference on audio, language and image processing, 2008. ICALIP 2008. IEEE, pp 1145–1152

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Himadri Mukherjee.

Appendix

Appendix

Fig. 7
figure 7

a Feature values for the Instruments for 1 second long Clips in Fan Noise Condition. b Feature values for the Instruments for 2 second long Clips in Fan Noise Condition. c Feature values for the Instruments for 1 second long Clips in Rain Condition. d Feature values for the Instruments for 2 second long Clips in Rain Noise Condition

Table 11 Individual Instrument Confusions for both D1(1s) and D2(2s)
Fig. 8
figure 8

a Feature values for the Instruments for 1 second long Clips in Traffic Noise Condition. b Feature values for the Instruments for 2 second long Clips in Traffic Noise Condition. c Feature values for the Instruments for 1 second long Clips in Vacuum Cleaner Condition. d Feature values for the Instruments for 2 second long Clips in Vacuum Cleaner Noise Condition

Table 12 Individual Instrument Confusions for both D3(1s) and D7(2s)
Table 13 Individual Instrument Confusions for both D4(1s) and D8(2s)
Table 14 Individual Instrument Confusions for both D5(1s) and D9(2s)
Table 15 Individual Instrument Confusions for both D6(1s) and D10(2s)
Table 16 (a) Instrument Family Confusions for both D1(1s) and D2(2s)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Mukherjee, H., Obaidullah, S.M., Phadikar, S. et al. MISNA - A musical instrument segregation system from noisy audio with LPCC-S features and extreme learning. Multimed Tools Appl 77, 27997–28022 (2018). https://doi.org/10.1007/s11042-018-5993-6

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-018-5993-6

Keywords

Navigation