Abstract
It is a key feature to embed 3D realistic sound effect in the future multimedia and virtual reality systems. Recent research on acoustics and psychoacoustics reveals the important cues for sound localization and sound perception. One promising approach to generate 3D realistic sound effect uses two earphones by simulating the sound waveforms from sound source to eardrum. This paper summarizes two methods for generating 3D realistic sound and points out their inherent drawbacks. To overcome these drawbacks we propose a simplified model to generate 3D realistic sound at any positions in the horizontal plane based on the results of sound perception and localization. Experimental results show that the model is correct and efficient.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Zhang Aidong, Shi Jiaoying. Realistic spatial sound in virtual reality. Submitted to JCST.
Wightman F L, Kistler D J. Headphone simulation of free-field listening—(I) Stimulus synthesis.Journal of Acoustical Society of America, 1989, 85(2): 858–867.
Kistler D J, Wightman F L. A model of head-related transfer function based on principal components analysis and minimum-phase reconstruction.Journal of Acoustical Society of America, 1992, 91(3): 1637–1647.
Blauert J. Spatial Hearing: The Psychophysics of Human Sound Localization. Cambridge, MA, MIT Press, 1990.
Rayleigh L. On our perception of sound direction.Philosophy Magazine, 1907, 13: 214–232.
Plenge G. On the differences between localization and lateralization.Journal of Acoustical Society of America, 1974, 56(3): 944–951.
Shaw E A G, Vaillancourt M M. Transformation of sound-pressure level from the free field to the eardrum presented in numerical form.Journal of Acoustical Society of America, 1985, 78(3): 1120–1123.
Mehrgardt S, Mellert V. Transformation characteristics of the external human ear.Journal of Acoustical Society of America, 1977, 61(6): 1567–1676.
Zhao Yu. The study of multimedia synchronization and non-speech audio. Doctoral Dissertation, Zhejiang University, 1995.
Kuhn G F. Model for the interaural time differences in the azimuthal plane.Journal of Acoustical Society of America, 1977, 62(1): 157–167.
Ross M J, Shaffer H L, Cohen A, Freudberg R, Manley H J. Average magnitude difference function pitch extractor.IEEE Trans. on Acoustics, Speech, and Signal Processing, 1974, ASSP-22(5): 353–362.
Author information
Authors and Affiliations
Additional information
This paper was supported by the Basic Research Fund granted by the National Commision of Science, Technology and Industry for Defence.
For the biographies ofZhao Yu, Shi Jiaoying andHe Zhijun, please refer to Vol.11, No.2, p.160 of JCST.
Zhang Qiong was born in 1973. He received his B.S. degree in computer science from Zhejiang University in 1994. Now he is a graduate student at the State Key Lab of CAD&CG, Zhejiang University. His research interests include 3D realistic sound rendering and audio signal processing in the multimedia and virtual reality systems.
Xiang Hui was born in 1970. He received his B.S. degree in computer science from The University of Science and Technology of China, and his M.S. degree also in computer science from Zhejiang University. He is a Ph.D. candidate at the State Key Lab of CAD&CG, Zhejiang University. His research interests include image compression and dimension transformation of video streams.
Rights and permissions
About this article
Cite this article
Zhao, Y., Zhang, Q., Xiang, H. et al. A simplified model for generating 3D realistic sound in the multimedia and virtual reality systems. J. of Comput. Sci. & Technol. 11, 461–470 (1996). https://doi.org/10.1007/BF02947213
Received:
Revised:
Issue Date:
DOI: https://doi.org/10.1007/BF02947213