[HTML][HTML] The cocktail-party problem revisited: early processing and selection of multi-talker speech

AW Bronkhorst�- Attention, Perception, & Psychophysics, 2015 - Springer
How do we recognize what one person is saying when others are speaking at the same
time? This review summarizes widespread research in psychoacoustics, auditory scene�…

Temporal coherence and attention in auditory scene analysis

SA Shamma, M Elhilali, C Micheyl�- Trends in neurosciences, 2011 - cell.com
Humans and other animals can attend to one of multiple sounds and follow it selectively
over time. The neural underpinnings of this perceptual feat remain mysterious. Some studies�…

Looking to listen at the cocktail party: A speaker-independent audio-visual model for speech separation

A Ephrat, I Mosseri, O Lang, T Dekel, K Wilson…�- arXiv preprint arXiv�…, 2018 - arxiv.org
We present a joint audio-visual model for isolating a single speech signal from a mixture of
sounds such as other speakers and background noise. Solving this task using only audio as�…

[BOOK][B] Cognitive psychology: A student's handbook

MW Eysenck, MT Keane - 2020 - taylorfrancis.com
The fully updated eighth edition of Cognitive Psychology: A Student's Handbook provides
comprehensive yet accessible coverage of all the key areas in the field ranging from visual�…

The sound of pixels

H Zhao, C Gan, A Rouditchenko…�- Proceedings of the�…, 2018 - openaccess.thecvf.com
We introduce PixelPlayer, a system that, by leveraging large amounts of unlabeled videos,
learns to locate image regions which produce sounds and separate the input sounds into a�…

Music gesture for visual sound separation

C Gan, D Huang, H Zhao…�- Proceedings of the�…, 2020 - openaccess.thecvf.com
Recent deep learning approaches have achieved impressive performance on visual sound
separation tasks. However, these approaches are mostly built on appearance and optical�…

The sound of motions

H Zhao, C Gan, WC Ma…�- Proceedings of the IEEE�…, 2019 - openaccess.thecvf.com
Sounds originate from object motions and vibrations of surrounding air. Inspired by the fact
that humans is capable of interpreting sound sources from how objects move visually, we�…

[HTML][HTML] Rapid transformation from auditory to linguistic representations of continuous speech

C Brodbeck, LE Hong, JZ Simon�- Current Biology, 2018 - cell.com
During speech perception, a central task of the auditory cortex is to analyze complex
acoustic patterns to allow detection of the words that encode a linguistic message [1]. It is�…

Sound source localization is all about cross-modal alignment

A Senocak, H Ryu, J Kim, TH Oh…�- Proceedings of the�…, 2023 - openaccess.thecvf.com
Humans can easily perceive the direction of sound sources in a visual scene, termed sound
source localization. Recent studies on learning-based sound source localization have�…

[BOOK][B] The Gamer's Brain: How neuroscience and UX can impact video game design

C Hodent - 2017 - taylorfrancis.com
Making a successful video game is hard. Even games that are well-received at launch may
fail to engage players in the long term due to issues with the user experience (UX) that they�…