10/13/2017

A Lesson In The Language Of ‘Baby Talk’

6:46 minutes

baby talk
Credit: Shutterstock

We’ve all heard adults cooing to babies in “baby talk” — that high-pitched, singsong cadence we tend to slip into around infants. The overall effect of baby talk may sound unnatural, but as Princeton neuroscientist Elise Piazza explains, the exaggerated high pitch, repetition, rhythm and even the pauses in baby talk can give babies important acoustical information about how language works.

“All of these cues combine to help highlight the inherent structure in speech, so to help babies to segment this constant stream of noise into the building blocks of language — like syllables, words and sentences,” she says. And as Piazza and her colleagues recently found, baby talk has another quality: Adults talking to babies change the timbre of their voices, too.

Their research was published in October in the journal Current Biology.

Timbre is the unique quality, or “tone color,” of a sound, Piazza explains.

[Chronic Fatigue patients speak up in a new documentary.]

“It’s actually much less understood than pitch or rhythm, but we rely on it constantly to distinguish and enjoy all of the different flavors of sounds around us,” she says. “So, for example, we can easily discern different idiosyncratic celebrity voices, like Barry White, who has a famously velvety voice, or Gilbert Gottfried, who has a much more nasal voice, or maybe Tom Waits with his sort of gravelly growl — even if these three people were all singing the same note with the same rhythm.”

Likewise, if you listen to the different sections of an orchestra tuning at the same pitch, you might notice that the woodwinds sound “reedy,” the brass instruments “buzzy,” and the strings “mellow.”

“These are all timbre descriptors,” Piazza says.

Timbre helps us quickly get at the gist of sounds, she explains. Hearing just a second or two of music doesn’t give you much information about its beat structure, or whether it’s in a major or minor key. But you can hear timbre. “And similarly with the voice, you can really get a lot of information about someone’s identity from pretty short clips of voices, which you might not be able to just by you know, a single note or a single piece of rhythm,” she says.

In their research, Piazza and her colleagues recorded 12 English-speaking mothers as they played with their babies, and as they spoke with an adult researcher. Then, the team used a sophisticated machine learning algorithm to model the shift in timbre between the mothers’ “baby talk” and their everyday language. “It’s so consistent across mothers,” Piazza said in a Princeton press release. “They all use the same kind of shift to go between those modes.”

When the researchers tested mothers speaking in nine other languages, from Cantonese to Hebrew to Polish, they found similar results. “The most remarkable thing was that when we then brought this new group of 12 mothers, also from the central New Jersey area, who spoke this wider range of languages from around the world,” Piazza says, “the same model that we had devised to discriminate these two modes in the English group generalized immediately to this new group of non-English speakers.”

In fact, we all might shift the timbre of our voices when we communicate with babies, even if we don’t realize it. “We used mothers to keep overall pitch range fairly consistent across participants,” Piazza explained in the press release. “However, I’d predict that our findings would generalize quite well to fathers.”

—Julia Franz (originally published on PRI.org)

*Correction 10/13/2017: A previous version of this article said Piazza’s study was published in the journal Cell Press. It was actually in Current Biology, which is published by Cell Press.


Segment Guests

Elise Piazza

Elise Piazza is a postdoctoral researcher at the Princeton Neuroscience Institute of Princeton University in Princeton, New Jersey.

Segment Transcript

JOHN DANKOSKY: OK, imagine you’re sitting on a train, and you overhear a woman sitting behind you saying this.

SPEAKER: Hi! Hi! What are you looking at?

JOHN DANKOSKY: You could probably tell, even without seeing the woman, that she is speaking to a baby– we hope. She’s using that high-pitched, sing-song cadence that we think of as baby talk. Now, baby talk isn’t just some adorable or irritating affectation that we adults acquire when we’re around infants. It actually gives babies important acoustical information that they need to begin to process words in our language.

But new research out this week suggests there’s something else adults are doing when they use baby talk. They change the timbre of their voice as well. Here with me to discuss this– and I promise it won’t be in baby talk– is Elise Piazza, post-doctoral researcher at the Princeton Neuroscience Institute at Princeton University. Welcome to Science Friday.

ELISE PIAZZA: Hi. Thanks very much for having me.

JOHN DANKOSKY: Why don’t you explain what timbre is first, so people understand what we’re talking about.

ELISE PIAZZA: So timbre is generally defined as the unique quality or the tone color of a sound. And it’s actually much less understood than pitch or rhythm, but we rely on it constantly to distinguish and enjoy all of the different flavors of sounds around us.

So for example, we can easily discern different idiosyncratic celebrity voices, like Barry White, who has a famously velvety voice, or Gilbert Gottfried who has a much more nasal voice, or maybe Tom Waits with his sort of gravelly growl, even if these three people were all singing the same note with the same rhythm. And if you imagine an orchestra tuning up, they often will all play the same pitch, like A 440, but you can still easily pick out the different instrument families or textures throughout the orchestra, including the reedy woodwinds, the buzzy brass, the mellow strings, et cetera. And these are all timbre descriptors.

And timbre can also give us the overall gist of sounds, in a way that many other features really can’t. So research has shown that people can identify the genre and the era of a piece of music quite well, in less than half a second. And this is probably because timbre is giving us some clues about the instrumentation and the texture of that music.

JOHN DANKOSKY: I’m John Dankosky. This is Science Friday from PRI, Public Radio International. And we’re talking with researcher Elise Piazza about timbre in baby talk.

OK, we found that mothers do this regularly during baby talk, of course– dads, too. We’ve got a clip here of a mother speaking to her baby, and then speaking in her normal voice. I’m going to try and guess the timbre differences. First, let’s take a listen.

SPEAKER: Hi. Hi, sweet girl. Hi. What do you see? You see a microphone? Yeah? And some sound proofing? What’s that for? Huh?

Anything shiny catches her attention right now– shiny and bright.

JOHN DANKOSKY: OK, so that’s a mom and a baby. Not just any baby, by the way– a brand-new member of the SciFri family, [? Larkin ?] [? Scotta ?] with her mom, Molly, making her radio debut today. So I heard sort of a breathy quality to this, a breathiness. But what else do you hear in there, Elise?

ELISE PIAZZA: So some of the features of motherese, or infant-directed speech, that have already been well-documented in the literature which you can easily hear in that segment, include the much more exaggerated pitch contours, so the sort of swoops up and down. Also, the overall pitch is much higher, which we think babies tend to prefer. And the repetition and the rhythm is also quite salient, so there are more pauses in the motherese. All of these cues sort of combine to help to highlight the inherent structure in speech, so to help babies to segment this constant stream of noise into the building blocks of language, like syllables, words, and sentences.

JOHN DANKOSKY: You talked about timbre helping us to get at the gist of sounds. What exactly do you mean by that?

ELISE PIAZZA: So if you take a really quick slice of music, you’re not getting much information about the rhythmic complexity. So you might not be able to hear the beat structure yet, and you might not know which notes, or sort of which key center– whether you’re, let’s say, in a major or a minor mode– just from a second of music.

But what you can hear are things like the instrumentation, whether we’re dealing with the timbres of a rock band, maybe the sort of percussive sounds and the sort of guitar timbre, as opposed to maybe a more chamber music kind of timbre. And similarly, with the voice, you can really get a lot of information about someone’s identity from pretty short clips of voices, which you might not be able to just by a single note or a single piece of rhythm.

JOHN DANKOSKY: Well, one of the most interesting findings here is about whether or not this happens across all languages, these qualities of timbre. What did you learn about that? Is it just in English or not?

ELISE PIAZZA: So we found statistical commonalities in the exact structure of the shift that generalized across a very wide variety of languages. So specifically, we tested English-speaking mothers, as well as mothers who spoke Spanish, Russian, Polish, Hungarian, German, French, Hebrew, Mandarin, and Cantonese, so a wide variety.

And we first brought the English-speaking mothers into the lab, and we asked them to just speak as they naturally would to their babies while playing with them, while reading them some age-appropriate board books. And then we had them speak to an adult experimenter. And we used a sophisticated machine-learning algorithm, just sort of a form of AI, where we basically designed a model of this shift in timbre between infant- and adult-directed speech in the English-speaking mothers.

And the most remarkable thing was that when we then brought this new group of 12 mothers, also from the central New Jersey area, who spoke this wider range of languages from around the world, the same model that we had devised to discriminate these two modes in the English group generalized immediately to this new group of non-English speakers. And we used these rigorous statistical methods to determine that. So for instance, if you train the model just on the English data, it actually generalizes extremely well to the non-English and vice versa.

JOHN DANKOSKY: It’s fascinating. I didn’t know we could learn this much about baby talk. But I’m very appreciative that you took some time with us today. Elise Piazza, post-doctoral researcher at the Princeton Neuroscience Institute at Princeton University. Thanks so much.

Copyright © 2017 Science Friday Initiative. All rights reserved. Science Friday transcripts are produced on a tight deadline by 3Play Media. Fidelity to the original aired/published audio or video file might vary, and text might be updated or amended in the future. For the authoritative record of Science Friday’s programming, please visit the original aired/published recording. For terms of use and more information, visit our policies pages at http://www.sciencefriday.com/about/policies/

Meet the Producer

About Katie Feather

Katie Feather is a former SciFri producer and the proud mother of two cats, Charleigh and Sadie.

Explore More

The U.S. Government Has A Mustang Problem

Mustangs carry symbolic weight in the minds of Americans. That makes managing their populations tricky.  

Read More