A computational model of auditory analysis is described that is inspired by psychoacoustical and neurophysiological findings in early and central stages of the auditory system. The model provides a unified multiresolution representation of the spectral and temporal features likely critical in the perception of sound. Simplified, more specifically tailored versions of this model have already been validated by successful application in the assessment of speech intelligibility [Elhilali et al, Speech Commun. 41(2-3), 331–348 (2003); Chi et al, J. Acoust. Soc. Am. 106, 2719–2732 (1999)] and in explaining the perception of monaural phase sensitivity [R. Carlyon and S. Shamma, J. Acoust. Soc. Am. 114, 333–348 (2003)]. Here we provide a more complete mathematical formulation of the model, illustrating how complex signals are transformed through various stages of the model, and relating it to comparable existing models of auditory processing. Furthermore, we outline several reconstruction algorithms to resynthesize the sound from the model output so as to evaluate the fidelity of the representation and contribution of different features and cues to the sound percept.
Skip Nav Destination
Article navigation
August 2005
August 01 2005
Multiresolution spectrotemporal analysis of complex sounds
Taishih Chi;
Center for Auditory and Acoustics Research, Institute for Systems Research Electrical and Computer Engineering Department,
University of Maryland
, College Park, Maryland 20742
Search for other works by this author on:
Powen Ru;
Center for Auditory and Acoustics Research, Institute for Systems Research Electrical and Computer Engineering Department,
University of Maryland
, College Park, Maryland 20742
Search for other works by this author on:
Shihab A. Shamma
Center for Auditory and Acoustics Research, Institute for Systems Research Electrical and Computer Engineering Department,
University of Maryland
, College Park, Maryland 20742
Search for other works by this author on:
J. Acoust. Soc. Am. 118, 887–906 (2005)
Article history
Received:
June 22 2004
Accepted:
May 12 2005
Citation
Taishih Chi, Powen Ru, Shihab A. Shamma; Multiresolution spectrotemporal analysis of complex sounds. J. Acoust. Soc. Am. 1 August 2005; 118 (2): 887–906. https://doi.org/10.1121/1.1945807
Download citation file:
Sign in
Don't already have an account? Register
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Sign in via your Institution
Sign in via your InstitutionPay-Per-View Access
$40.00
Citing articles via
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Day-to-day loudness assessments of indoor soundscapes: Exploring the impact of loudness indicators, person, and situation
Siegbert Versümer, Jochen Steffens, et al.