Skip to main content

Phase Constraint and Deep Neural Network for Speech Separation

  • Conference paper
  • First Online:
Advances in Neural Networks - ISNN 2017 (ISNN 2017)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 10262))

Included in the following conference series:

Abstract

The phase response of speech is an important part in speech separation. In this paper, we apply the complex mask to the speech separation. It both enhances the magnitude and phase of speech. Specifically, we use a deep neural network to estimate the complex mask of two sources. And considering the importance of the phase, we also explore a phase constraint objective function, which can ensure the phase of the sum of estimated sources that is close to the phase of the mixture. We demonstrate the efficiency of the method on the TIMIT speech corpus for single channel speech separation.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Ephraim, Y., Malah, D.: Speech enhancement using a minimum-mean square error short-time spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 33(2), 443–445 (1985)

    Article  Google Scholar 

  2. Erdogan, H., Hershey, J.R., Watanabe, S., Roux, J.L.: Phase-sensitive and recognition-boosted speech separation using deep recurrent neural networks. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2015, pp. 708–712 (2015)

    Google Scholar 

  3. Healy, E.W., Yoho, S.E., Wang, Y., Wang, D.: An algorithm to improve speech recognition in noise for hearing-impaired listeners. J. Acoust. Soc. Am. 135(4), 3029–3038 (2014)

    Article  Google Scholar 

  4. Huang, P.S., Kim, M., Hasegawa-Johnson, M., Smaragdis, P.: Deep learning for monaural speech separation. In: 2014 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2014, pp. 1562–1566 (2014)

    Google Scholar 

  5. Kang, T.G., Kwon, K., Shin, J.W., Kim, N.S.: NMF-based target source separation using deep neural network. IEEE Signal Process. Lett. 22(2), 229–233 (2015)

    Article  Google Scholar 

  6. Kim, G., Lu, Y., Hu, Y., Loizou, P.C.: An algorithm that improves speech intelligibility in noise for normal-hearing listeners. J. Acoust. Soc. Am. 126(3), 1486–1494 (2009)

    Article  Google Scholar 

  7. Narayanan, A., Wang, D.L.: Ideal ratio mask estimation using deep neural networks for robust speech recognition. In: IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 7092–7096 (2013)

    Google Scholar 

  8. Paliwal, K., Wjcicki, K., Shannon, B.: The importance of phase in speech enhancement. Speech Commun. 53(4), 465–494 (2011)

    Article  Google Scholar 

  9. Schmidt, M.N., Olsson, R.K.: Single-channel speech separation using sparse non-negative matrix factorization. In: ICSLP, Ninth International Conference on Spoken Language Processing, INTERSPEECH 2006, Pittsburgh, PA, USA, September 2006

    Google Scholar 

  10. Tu, Y., Du, J., Xu, Y., Dai, L.: Speech separation based on improved deep neural networks with dual outputs of speech features for both target and interfering speakers. In: International Symposium on Chinese Spoken Language Processing, pp. 250–254 (2014)

    Google Scholar 

  11. Vincent, E., Gribonval, R., Fevotte, C.: Performance measurement in blind audio source separation. IEEE Trans. Audio Speech Lang. Process. 14(4), 1462–1469 (2006)

    Article  Google Scholar 

  12. Wang, D., Lim, J.S.: The unimportance of phase in speech enhancement. IEEE Trans. Acoust. Speech Signal Process. 30(4), 679–681 (1982)

    Article  Google Scholar 

  13. Wang, G.X., Hsu, C.C., Chien, J.T.: Discriminative deep recurrent neural networks for monaural speech separation. In: IEEE International Conference on Acoustics, Speech and Signal Processing (2016)

    Google Scholar 

  14. Wang, Y., Narayanan, A., Wang, D.L.: On training targets for supervised speech separation. IEEE/ACM Trans. Audio Speech Lang. Process. 22(12), 1849–1858 (2014)

    Article  Google Scholar 

  15. Wang, Y., Wang, D.L.: A deep neural network for time-domain signal reconstruction. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2015, pp. 4390–4394 (2015)

    Google Scholar 

  16. Williamson, D.S., Wang, Y., Wang, D.: Complex ratio masking for monaural speech separation. IEEE/ACM Trans. Audio Speech Lang. Process. 24(3), 1–1 (2016)

    Article  Google Scholar 

  17. Zue, V., Seneff, S., Glass, J.: Speech database development at MIT: TIMIT and beyond. Speech Commun. 9(4), 351–356 (1990)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaohong Ma .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Miao, Z., Ma, X., Ding, S. (2017). Phase Constraint and Deep Neural Network for Speech Separation. In: Cong, F., Leung, A., Wei, Q. (eds) Advances in Neural Networks - ISNN 2017. ISNN 2017. Lecture Notes in Computer Science(), vol 10262. Springer, Cham. https://doi.org/10.1007/978-3-319-59081-3_32

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-59081-3_32

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-59080-6

  • Online ISBN: 978-3-319-59081-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics