Abstract
The number of applications on graphics workstations using sound in order to enhance the human-computer interaction capabilities is increasing. Using and handling sound in addition to text and graphics seems to be the next forthcoming step towards a multimedia environment.
This paper outlines a first approach for an audio content architecture consisting of different description layers which are able to handle different levels of abstraction. Therefore a comparison between image rendering and sound synthesis is made. Then a short overview on digital sound synthesis techniques and sound color models is given in order to show hardware requirements as well as psychophysical difficulties. Finally, the audio layer model is presented. It contains three layers for parametric and symbolic audio description, one layer for digital audio, and one presentation layer.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
William Buxton, William Gaver and Sara Bly: Tutorial Notes of the 1989 SIGCHI conference on “The Use of Non-Speech Audio at the Interface”, ACM 1989
John M. Chowing: “The Synthesis of Complex Audio Spectra by Means of Frequency Modulation”, Journal of the Audio Engineering Society, Vol. 21(7), 1973
William W. Gaver: “The SonicFinder”, Tutorial Notes of the 1989 SIGCHI conference on “The Use of Non-Speech Audio at the Interface”, ACM 1989
Heinz Gerhäuser und Dieter Seitzer: “Musikcodierung — Signalverarbeitung in Echtzeit auf psychoakustischer Grundlage”, FhG-Bericht 1/89
J. M. Grey: “An Exploration of Musical Timbre”, PhD. diss., Dept. of Psychology, Stanford University, 1975
Georges G. Grinstein, Ronald M. Picket, Marian G. Williams: “EXVIS: An Exploratory Visualization Environment”, Graphics Research Laboratory, University of Lowell, 1989
ISO/IEC JTC1/SC1/WG5 document N638: “Liaison Statement on User Requirements for an Audio Content Architecture”, 1988
Gareth Loy: “Musicians make a Standard: The MIDI Phenomenon”, Computer Music Journal, Vol. 9, No. 4, 1985
Juan G. Roeder: “Introduction to the Physics and Psychophysics of Music”, Springer Verlag, New York — Heidelberg — Berlin, 1973
Stuart Smith: “An Auditory Display for Exploratory Visualization of Multidimensional Data”, Computer Science Department, University of Lowell, 1989
John Strawn and Curtis Roads: “Synthesizer Hardware and Engineering”, chapter 2 of “Foundations of Computer Music”, edited by John Strawn and Curis Roads, MIT Press, Cambridge, MA., 1987
G. Stoll und G. Theile: “MASCAM: Minimale Datenrate durch Berücksichtigung der Gehöreigenschaften bei der Codierung hochwertiger Tonsignale”, Fernseh- und Kino-Technik Nr. 11/1988
David L. Wessel: “Timbre Space as a Musical Structure”, chapter 5 of “Foundations of Computer Music”, edited by John Strawn and Curtis Roads, MIT Press, Cambridge, MA., 1987
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1991 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Blum, C. (1991). A Model for the Use of Different Sound Description Layers Within a Multimedia Environment. In: Encarnação, J.L., Grinstein, G.G. (eds) Workstations for Experiments. IFIP Series on Computer Graphics. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-75903-1_11
Download citation
DOI: https://doi.org/10.1007/978-3-642-75903-1_11
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-75905-5
Online ISBN: 978-3-642-75903-1
eBook Packages: Springer Book Archive