To communicate effectively with each other, we often need to be able to understand what someone’s saying in the presence of background noise. Even listening to a friend speak at a busy restaurant can be difficult when other people are having loud conversations at nearby tables. I’m particularly interested in how we understand speech in these challenging listening environments.

Selectively attending to a sound of interest

3D Brain ImageA theme of my research is to investigate how we focus our attention on a sound of interest—for example, a friend speaking in a noisy place. How does attention influence perception and the way in which our brains process sounds? To explore this question, I use a combination of methods, including psychophysics, electroencephalograhy (EEG), and functional magnetic resonance imaging (fMRI).

Relevant publications:

Improving speech intelligibility during multi-talker listening

people-2567915_1920I’m also interested in how cognitive factors (e.g. context and prior knowledge) can improve speech intelligibility when we’re in difficult listening situations. For example, some of my research on this topic has focused on voice familiarity. When multiple people speak simultaneously, we’re better at understanding speech spoken by people who we’re familiar with, such as a friend or partner, than someone we’ve never previously met. But how do we become familiar with someone’s voice? And which aspects of someone’s voice help us to follow what they’re saying?

Relevant publications:

Hearing impairment

Another aspect of my research investigates how cognitive factors influence speech perception for people with sensorineural hearing impairment. People with hearing impairment often report that they find listening in noisy environments particularly difficult and effortful, even when they use hearing aids. Although some cognitive abilities seem to be impaired in people with hearing loss, they nevertheless retain the ability to use context and prior knowledge to help them understand speech—perhaps relying on context to help compensate for a degraded acoustic signal. I’m interested in how these factors can help to improve speech perception, and make listening less tiring and effortful, for people with hearing loss.

Relevant publications:

For a full list of publications, see my Google Scholar or NCBI page.