Months Before Their First Words, Babies' Brains Rehearse Speech Mechanics
Embargo expired: 14-Jul-2014 3:00 PM EDT
Source Newsroom: University of Washington
Newswise — Infants can tell the difference between sounds of all languages until about 8 months of age when their brains start to focus only on the sounds they hear around them. It's been unclear how this transition occurs, but social interactions and caregivers' use of exaggerated "parentese" style of speech seem to help.
University of Washington research in 7- and 11-month-old infants shows that speech sounds stimulate areas of the brain that coordinate and plan motor movements for speech.
The study, published July 14 in the Proceedings of the National Academy of Sciences, suggests that baby brains start laying down the groundwork of how to form words long before they actually begin to speak, and this may affect the developmental transition.
"Most babies babble by 7 months, but don't utter their first words until after their first birthdays," said lead author Patricia Kuhl, who is the co-director of the UW's Institute for Learning and Brain Sciences. "Finding activation in motor areas of the brain when infants are simply listening is significant, because it means the baby brain is engaged in trying to talk back right from the start and suggests that 7-month-olds' brains are already trying to figure out how to make the right movements that will produce words."
Kuhl and her research team believe this practice at motor planning contributes to the transition when infants become more sensitive to their native language.
The results emphasize the importance of talking to kids during social interactions even if they aren't talking back yet.
"Hearing us talk exercises the action areas of infants' brains, going beyond what we thought happens when we talk to them," Kuhl said. "Infants' brains are preparing them to act on the world by practicing how to speak before they actually say a word."
In the experiment, infants sat in a brain scanner that measures brain activation through a noninvasive technique called magnetoencephalography. Nicknamed MEG, the brain scanner resembles an egg-shaped vintage hair dryer and is completely safe for infants. The Institute for Learning and Brain Sciences was the first in the world to use such a tool to study babies while they engaged in a task.
Here's a video of one the babies in the experiment: https://www.youtube.com/watch?v=aFJNTaPvXpk&feature=youtu.be
The babies, 57 7- and 11- or 12-month-olds, each listened to a series of native and foreign language syllables such as "da" and "ta" as researchers recorded brain responses. They listened to sounds from English (https://soundcloud.com/uw-today/pnas-example-3standard-english) and Spanish (https://soundcloud.com/uw-today/pnas-example-3standard-spanish).
The researchers observed brain activity in an auditory area of the brain called the superior temporal gyrus, as well as in Broca's area and the cerebellum, cortical regions responsible for planning the motor movements required for producing speech.
This pattern of brain activation occurred for sounds in the 7-month-olds' native language (English) as well as in a non-native language (Spanish), showing that at this early age infants are responding to all speech sounds, whether or not they have heard the sounds before.
In the older infants, brain activation was different. By 11-12 months, infants' brains increase motor activation to the non-native speech sounds relative to native speech, which the researchers interpret as showing that it takes more effort for the baby brain to predict which movements create non-native speech. This reflects an effect of experience between 7 and 11 months, and suggests that activation in motor brain areas is contributing to the transition in early speech perception.
The study has social implications, suggesting that the slow and exaggerated parentese speech – "Hiiiii! How are youuuuu?" – may actually prompt infants to try to synthesize utterances themselves and imitate what they heard, uttering something like "Ahhh bah bah baaah."
"Parentese is very exaggerated, and when infants hear it, their brains may find it easier to model the motor movements necessary to speak," Kuhl said.
Co-authors Rey Ramirez, Alexis Bosseler, Jo-Fu Lotus Lin and Toshiaki Imada are all with UW's Institute for Learning & Brain Sciences.
The National Science Foundation Science of Learning Program grant to the UW's LIFE Center funded the study.