Abstract
Auditory scene analysis (ASA) refers to the process (es) of parsing the complex acoustic input into auditory perceptual objects representing either physical sources or temporal sound patterns, such as melodies, which contributed to the sound waves reaching the ears. A number of new computational models accounting for some of the perceptual phenomena of ASA have been published recently. Here we provide a theoretically motivated review of these computational models, aiming to relate their guiding principles to the central issues of the theoretical framework of ASA. Specifically, we ask how they achieve the grouping and separation of sound elements and whether they implement some form of competition between alternative interpretations of the sound input. We consider the extent to which they include predictive processes, as important current theories suggest that perception is inherently predictive, and also how they have been evaluated. We conclude that current computational models of ASA are fragmentary in the sense that rather than providing general competing interpretations of ASA, they focus on assessing the utility of specific processes (or algorithms) for finding the causes of the complex acoustic signal. This leaves open the possibility for integrating complementary aspects of the models into a more comprehensive theory of ASA.
DOI
10.3389/fnins.2016.00524
Publication Date
2016-11-15
Publication Title
Frontiers in Neuroscience
Volume
10
Publisher
Frontiers Media SA
ISSN
1662-453X
Embargo Period
2024-11-22
Recommended Citation
Szabó, B., Denham, S., & Winkler, I. (2016) 'Computational Models of Auditory Scene Analysis: A Review', Frontiers in Neuroscience, 10. Frontiers Media SA: Available at: https://doi.org/10.3389/fnins.2016.00524