Summary: Researchers report familiar voices are easier than those of a stranger to understand, even when the familiar voice is manipulated and the listener can not identify the voice.
Familiar voices are easier to understand and this advantage holds even if when we aren’t able to identify who those familiar voices belong to, according to research in Psychological Science.
The research from Western University’s BrainsCAN initiative showed that even though participants were not able to recognize their friend’s voice when the resonance of their voice was manipulated, they still found it easier to understand than the same words spoken by a stranger.
“Our findings demonstrate that we pick out different information from a voice, depending on whether we’re simply trying to recognize whether it’s our friend or family member on the phone or whether we’re trying to understand the words they’re saying,” says researcher Emma Holmes of UCL (University College London), first author on the study. “This shows we focus on different parts of speech sounds for different purposes.”
Holmes and Western University coauthors Ingrid S. Johnsrude and Ysabel Domingo are interested in understanding the factors that influence how we perceive others’ voices across a variety of contexts. Anyone who has tried to hold a conversation in a bustling office or a crowded restaurant knows how difficult it is to understand what someone is saying when it competes with background noise. In previous work, the researchers found that familiarity offers an advantage in these noisy situations, making the voices of friends and family easier to understand than the voices of strangers.
“This suggests that, over time, we must learn something about the voices of the people we frequently talk to, which helps us to better understand the words they’re saying,” Holmes explains. “For this study we asked: Why does being familiar with someone’s voice help us understand what they’re saying?”
Holmes and colleagues decided to focus on two acoustic properties of the voice that vary reliably across people: pitch and resonance. Their aim was to determine how these properties influence our ability to understand what someone is saying and recognize who is speaking.
The researchers recruited 11 pairs of participants who were either friends or couples, had known each other for at least 6 months, and spoke to one another regularly. The participants read a predetermined set of sentences aloud, each of which followed a standard pattern of name, verb, number, adjective, noun (e.g., “Bob bought five green bags”). The researchers digitally manipulated the recordings from each speaker to systematically vary the pitch and the resonance.
For the experiment, participants listened to sentences spoken by their partner and by unfamiliar speakers who had the same gender as their partner. In one task, they heard two sentences spoken at the same time by different speakers and had to identify words in one of the sentences. In another task, participants heard a series of sentences and indicated whether each sentence was spoken by their partner or not.
As expected, participants were better at recognizing their partner’s voice when they heard an unaltered recording compared with any of the manipulated versions. They were also better at recognizing the partner’s voice when the pitch was altered relative to when the resonance was altered.
Intriguingly, even though manipulating the resonance of the partner’s voice made it unrecognizable to participants, they still found it easier to understand against a competing speaker than a stranger’s voice.
The results suggest that resonance is a critical acoustic feature that helps us recognize who a particular voice belongs to. Both pitch and resonance can influence our ability to understand what someone familiar is saying, although it seems that we can still understand what they are saying very well when the pitch and resonance of their voice have been altered.
The findings shed light on how we perceive speech in our everyday lives and could have particular implications for individuals with hearing loss, who have even greater difficulty understanding speech in noisy settings: These individuals might benefit even more from familiar voices. The research could even have implications for the development of artificial agents, including robots and digital assistants like Siri, that are both intelligible and recognizable.
Funding: This study was supported by a Natural Sciences and Engineering Research Council of Canada Discovery Grant, a Canadian Institutes of Health Research Operating Grant, and by BrainsCAN, Western University’s $66 million Canada First Research Excellence Fund program in cognitive neuroscience.
Source: Anna Mikulak – APS
Publisher: Organized by NeuroscienceNews.com.
Image Source: NeuroscienceNews.com image is adapted from the APS news release.
Original Research: Abstract for “Familiar Voices Are More Intelligible, Even if They Are Not Recognized as Familiar” by Emma Holmes, Ysabel Domingo, and Ingrid S. Johnsrude in Psychological Science. Published August 10 2018.
Familiar Voices Are More Intelligible, Even if They Are Not Recognized as Familiar
We can recognize familiar people by their voices, and familiar talkers are more intelligible than unfamiliar talkers when competing talkers are present. However, whether the acoustic voice characteristics that permit recognition and those that benefit intelligibility are the same or different is unknown. Here, we recruited pairs of participants who had known each other for 6 months or longer and manipulated the acoustic correlates of two voice characteristics (vocal tract length and glottal pulse rate). These had different effects on explicit recognition of and the speech-intelligibility benefit realized from familiar voices. Furthermore, even when explicit recognition of familiar voices was eliminated, they were still more intelligible than unfamiliar voices—demonstrating that familiar voices do not need to be explicitly recognized to benefit intelligibility. Processing familiar-voice information appears therefore to depend on multiple, at least partially independent, systems that are recruited depending on the perceptual goal of the listener.