Social calls provide novel insights into the evolution of vocal learning
Sewall, Kendra B.; Young, Anna M.; Wright, Timothy F.
2016-01-01
Learned song is among the best-studied models of animal communication. In oscine songbirds, where learned song is most prevalent, it is used primarily for intrasexual selection and mate attraction. Learning of a different class of vocal signals, known as contact calls, is found in a diverse array of species, where they are used to mediate social interactions among individuals. We argue that call learning provides a taxonomically rich system for studying testable hypotheses for the evolutionary origins of vocal learning. We describe and critically evaluate four nonmutually exclusive hypotheses for the origin and current function of vocal learning of calls, which propose that call learning (1) improves auditory detection and recognition, (2) signals local knowledge, (3) signals group membership, or (4) allows for the encoding of more complex social information. We propose approaches to testing these four hypotheses but emphasize that all of them share the idea that social living, not sexual selection, is a central driver of vocal learning. Finally, we identify future areas for research on call learning that could provide new perspectives on the origins and mechanisms of vocal learning in both animals and humans. PMID:28163325
Birds, primates, and spoken language origins: behavioral phenotypes and neurobiological substrates
Petkov, Christopher I.; Jarvis, Erich D.
2012-01-01
Vocal learners such as humans and songbirds can learn to produce elaborate patterns of structurally organized vocalizations, whereas many other vertebrates such as non-human primates and most other bird groups either cannot or do so to a very limited degree. To explain the similarities among humans and vocal-learning birds and the differences with other species, various theories have been proposed. One set of theories are motor theories, which underscore the role of the motor system as an evolutionary substrate for vocal production learning. For instance, the motor theory of speech and song perception proposes enhanced auditory perceptual learning of speech in humans and song in birds, which suggests a considerable level of neurobiological specialization. Another, a motor theory of vocal learning origin, proposes that the brain pathways that control the learning and production of song and speech were derived from adjacent motor brain pathways. Another set of theories are cognitive theories, which address the interface between cognition and the auditory-vocal domains to support language learning in humans. Here we critically review the behavioral and neurobiological evidence for parallels and differences between the so-called vocal learners and vocal non-learners in the context of motor and cognitive theories. In doing so, we note that behaviorally vocal-production learning abilities are more distributed than categorical, as are the auditory-learning abilities of animals. We propose testable hypotheses on the extent of the specializations and cross-species correspondences suggested by motor and cognitive theories. We believe that determining how spoken language evolved is likely to become clearer with concerted efforts in testing comparative data from many non-human animal species. PMID:22912615
Saranathan, Vinodkumar; Hamilton, Deborah; Powell, George V N; Kroodsma, Donald E; Prum, Richard O
2007-09-01
Vocal learning is thought to have evolved in three clades of birds (parrots, hummingbirds, and oscine passerines), and three clades of mammals (whales, bats, and primates). Behavioural data indicate that, unlike other suboscine passerines, the three-wattled bellbird Procnias tricarunculata (Cotingidae) is capable of vocal learning. Procnias tricarunculata shows conspicuous vocal ontogeny, striking geographical variation in song, and rapid temporal change in song within a population. Deprivation studies of vocal development in P. tricarunculata are impractical. Here, we report evidence from mitochondrial DNA sequences and nuclear microsatellite loci that genetic variation within and among the four allopatric breeding populations of P. tricarunculata is not congruent with variation in vocal behaviour. Sequences of the mitochondrial DNA control region document extensive haplotype sharing among localities and song types, and no phylogenetic resolution of geographical populations or behavioural groups. The vocally differentiated, allopatric breeding populations of P. tricarunculata are only weakly genetically differentiated populations, and are not distinct taxa. Mitochondrial DNA and microsatellite variation show small (2.9% and 13.5%, respectively) but significant correlation with geographical distance, but no significant residual variation by song type. Estimates of the strength of selection that would be needed to maintain the observed geographical pattern in vocal differentiation if songs were genetically based are unreasonably high, further discrediting the hypothesis of a genetic origin of vocal variation. These data support a fourth, phylogenetically independent origin of avian vocal learning in Procnias. Geographical variations in P. tricarunculata vocal behaviour are likely culturally evolved dialects.
Vocal learning in the functionally referential food grunts of chimpanzees.
Watson, Stuart K; Townsend, Simon W; Schel, Anne M; Wilke, Claudia; Wallace, Emma K; Cheng, Leveda; West, Victoria; Slocombe, Katie E
2015-02-16
One standout feature of human language is our ability to reference external objects and events with socially learned symbols, or words. Exploring the phylogenetic origins of this capacity is therefore key to a comprehensive understanding of the evolution of language. While non-human primates can produce vocalizations that refer to external objects in the environment, it is generally accepted that their acoustic structure is fixed and a product of arousal states. Indeed, it has been argued that the apparent lack of flexible control over the structure of referential vocalizations represents a key discontinuity with language. Here, we demonstrate vocal learning in the acoustic structure of referential food grunts in captive chimpanzees. We found that, following the integration of two groups of adult chimpanzees, the acoustic structure of referential food grunts produced for a specific food converged over 3 years. Acoustic convergence arose independently of preference for the food, and social network analyses indicated this only occurred after strong affiliative relationships were established between the original subgroups. We argue that these data represent the first evidence of non-human animals actively modifying and socially learning the structure of a meaningful referential vocalization from conspecifics. Our findings indicate that primate referential call structure is not simply determined by arousal and that the socially learned nature of referential words in humans likely has ancient evolutionary origins. Copyright © 2015 Elsevier Ltd. All rights reserved.
The Origins of Vocal Learning: New Sounds, New Circuits, New Cells
ERIC Educational Resources Information Center
Nottebohm, Fernando; Liu, Wan-Chun
2010-01-01
We do not know how vocal learning came to be, but it is such a salient trait in human evolution that many have tried to imagine it. In primates this is difficult because we are the only species known to possess this skill. Songbirds provide a richer and independent set of data. I use comparative data and ask broad questions: How does vocal…
Salinas-Melgoza, Alejandro; Wright, Timothy F.
2012-01-01
Studies of avian vocal dialects commonly find evidence of geographic and acoustic stability in the face of substantial gene flow between dialects. The vocal imitation and reduced dispersal hypotheses are alternatives to explain this mismatch between vocal and genetic variation. We experimentally simulated dispersal in the yellow-naped amazon (Amazona auropalliata) by moving individuals within and across dialect boundaries in Costa Rica. One juvenile translocated across dialect boundaries altered its contact call to imitate the acoustic form of the local call six weeks post-release. In contrast, four adults translocated across dialect boundaries returned to their original capture site within 120 days, while five cross-dialect translocated adults who remained at the release site did not alter their contact calls. Translocated individuals were observed to show some segregation from resident flocks. The observation of vocal imitation by the juvenile bird supports the vocal imitation, whereas the behavior of adults is more consistent with the reduced dispersal hypotheses. Taken together, our results suggest that both post-dispersal learning by juveniles and high philopatry in adults could explain the stability of vocal dialects in the face of immigration and gene flow. PMID:23139809
Development of neural responsivity to vocal sounds in higher level auditory cortex of songbirds
Miller-Sims, Vanessa C.
2014-01-01
Like humans, songbirds learn vocal sounds from “tutors” during a sensitive period of development. Vocal learning in songbirds therefore provides a powerful model system for investigating neural mechanisms by which memories of learned vocal sounds are stored. This study examined whether NCM (caudo-medial nidopallium), a region of higher level auditory cortex in songbirds, serves as a locus where a neural memory of tutor sounds is acquired during early stages of vocal learning. NCM neurons respond well to complex auditory stimuli, and evoked activity in many NCM neurons habituates such that the response to a stimulus that is heard repeatedly decreases to approximately one-half its original level (stimulus-specific adaptation). The rate of neural habituation serves as an index of familiarity, being low for familiar sounds, but high for novel sounds. We found that response strength across different song stimuli was higher in NCM neurons of adult zebra finches than in juveniles, and that only adult NCM responded selectively to tutor song. The rate of habituation across both tutor song and novel conspecific songs was lower in adult than in juvenile NCM, indicating higher familiarity and a more persistent response to song stimuli in adults. In juvenile birds that have memorized tutor vocal sounds, neural habituation was higher for tutor song than for a familiar conspecific song. This unexpected result suggests that the response to tutor song in NCM at this age may be subject to top-down influences that maintain the tutor song as a salient stimulus, despite its high level of familiarity. PMID:24694936
Vocal learning in elephants: neural bases and adaptive context
Stoeger, Angela S; Manger, Paul
2014-01-01
In the last decade clear evidence has accumulated that elephants are capable of vocal production learning. Examples of vocal imitation are documented in African (Loxodonta africana) and Asian (Elephas maximus) elephants, but little is known about the function of vocal learning within the natural communication systems of either species. We are also just starting to identify the neural basis of elephant vocalizations. The African elephant diencephalon and brainstem possess specializations related to aspects of neural information processing in the motor system (affecting the timing and learning of trunk movements) and the auditory and vocalization system. Comparative interdisciplinary (from behavioral to neuroanatomical) studies are strongly warranted to increase our understanding of both vocal learning and vocal behavior in elephants. PMID:25062469
Arriaga, Gustavo; Zhou, Eric P.; Jarvis, Erich D.
2012-01-01
Humans and song-learning birds communicate acoustically using learned vocalizations. The characteristic features of this social communication behavior include vocal control by forebrain motor areas, a direct cortical projection to brainstem vocal motor neurons, and dependence on auditory feedback to develop and maintain learned vocalizations. These features have so far not been found in closely related primate and avian species that do not learn vocalizations. Male mice produce courtship ultrasonic vocalizations with acoustic features similar to songs of song-learning birds. However, it is assumed that mice lack a forebrain system for vocal modification and that their ultrasonic vocalizations are innate. Here we investigated the mouse song system and discovered that it includes a motor cortex region active during singing, that projects directly to brainstem vocal motor neurons and is necessary for keeping song more stereotyped and on pitch. We also discovered that male mice depend on auditory feedback to maintain some ultrasonic song features, and that sub-strains with differences in their songs can match each other's pitch when cross-housed under competitive social conditions. We conclude that male mice have some limited vocal modification abilities with at least some neuroanatomical features thought to be unique to humans and song-learning birds. To explain our findings, we propose a continuum hypothesis of vocal learning. PMID:23071596
Horita, Haruhito; Kobayashi, Masahiko; Liu, Wan-chun; Oka, Kotaro; Jarvis, Erich D.; Wada, Kazuhiro
2012-01-01
Mechanisms for the evolution of convergent behavioral traits are largely unknown. Vocal learning is one such trait that evolved multiple times and is necessary in humans for the acquisition of spoken language. Among birds, vocal learning is evolved in songbirds, parrots, and hummingbirds. Each time similar forebrain song nuclei specialized for vocal learning and production have evolved. This finding led to the hypothesis that the behavioral and neuroanatomical convergences for vocal learning could be associated with molecular convergence. We previously found that the neural activity-induced gene dual specificity phosphatase 1 (dusp1) was up-regulated in non-vocal circuits, specifically in sensory-input neurons of the thalamus and telencephalon; however, dusp1 was not up-regulated in higher order sensory neurons or motor circuits. Here we show that song motor nuclei are an exception to this pattern. The song nuclei of species from all known vocal learning avian lineages showed motor-driven up-regulation of dusp1 expression induced by singing. There was no detectable motor-driven dusp1 expression throughout the rest of the forebrain after non-vocal motor performance. This pattern contrasts with expression of the commonly studied activity-induced gene egr1, which shows motor-driven expression in song nuclei induced by singing, but also motor-driven expression in adjacent brain regions after non-vocal motor behaviors. In the vocal non-learning avian species, we found no detectable vocalizing-driven dusp1 expression in the forebrain. These findings suggest that independent evolutions of neural systems for vocal learning were accompanied by selection for specialized motor-driven expression of the dusp1 gene in those circuits. This specialized expression of dusp1 could potentially lead to differential regulation of dusp1-modulated molecular cascades in vocal learning circuits. PMID:22876306
Neural Representation of a Target Auditory Memory in a Cortico-Basal Ganglia Pathway
Bottjer, Sarah W.
2013-01-01
Vocal learning in songbirds, like speech acquisition in humans, entails a period of sensorimotor integration during which vocalizations are evaluated via auditory feedback and progressively refined to achieve an imitation of memorized vocal sounds. This process requires the brain to compare feedback of current vocal behavior to a memory of target vocal sounds. We report the discovery of two distinct populations of neurons in a cortico-basal ganglia circuit of juvenile songbirds (zebra finches, Taeniopygia guttata) during vocal learning: (1) one in which neurons are selectively tuned to memorized sounds and (2) another in which neurons are selectively tuned to self-produced vocalizations. These results suggest that neurons tuned to learned vocal sounds encode a memory of those target sounds, whereas neurons tuned to self-produced vocalizations encode a representation of current vocal sounds. The presence of neurons tuned to memorized sounds is limited to early stages of sensorimotor integration: after learning, the incidence of neurons encoding memorized vocal sounds was greatly diminished. In contrast to this circuit, neurons known to drive vocal behavior through a parallel cortico-basal ganglia pathway show little selective tuning until late in learning. One interpretation of these data is that representations of current and target vocal sounds in the shell circuit are used to compare ongoing patterns of vocal feedback to memorized sounds, whereas the parallel core circuit has a motor-related role in learning. Such a functional subdivision is similar to mammalian cortico-basal ganglia pathways in which associative-limbic circuits mediate goal-directed responses, whereas sensorimotor circuits support motor aspects of learning. PMID:24005299
Learning while Babbling: Prelinguistic Object-Directed Vocalizations Indicate a Readiness to Learn
ERIC Educational Resources Information Center
Goldstein, Michael H.; Schwade, Jennifer; Briesch, Jacquelyn; Syal, Supriya
2010-01-01
Two studies illustrate the functional significance of a new category of prelinguistic vocalizing--object-directed vocalizations (ODVs)--and show that these sounds are connected to learning about words and objects. Experiment 1 tested 12-month-old infants' perceptual learning of objects that elicited ODVs. Fourteen infants' vocalizations were…
Dopaminergic Contributions to Vocal Learning
Hoffmann, Lukas A.; Saravanan, Varun; Wood, Alynda N.; He, Li
2016-01-01
Although the brain relies on auditory information to calibrate vocal behavior, the neural substrates of vocal learning remain unclear. Here we demonstrate that lesions of the dopaminergic inputs to a basal ganglia nucleus in a songbird species (Bengalese finches, Lonchura striata var. domestica) greatly reduced the magnitude of vocal learning driven by disruptive auditory feedback in a negative reinforcement task. These lesions produced no measureable effects on the quality of vocal performance or the amount of song produced. Our results suggest that dopaminergic inputs to the basal ganglia selectively mediate reinforcement-driven vocal plasticity. In contrast, dopaminergic lesions produced no measurable effects on the birds' ability to restore song acoustics to baseline following the cessation of reinforcement training, suggesting that different forms of vocal plasticity may use different neural mechanisms. SIGNIFICANCE STATEMENT During skill learning, the brain relies on sensory feedback to improve motor performance. However, the neural basis of sensorimotor learning is poorly understood. Here, we investigate the role of the neurotransmitter dopamine in regulating vocal learning in the Bengalese finch, a songbird with an extremely precise singing behavior that can nevertheless be reshaped dramatically by auditory feedback. Our findings show that reduction of dopamine inputs to a region of the songbird basal ganglia greatly impairs vocal learning but has no detectable effect on vocal performance. These results suggest a specific role for dopamine in regulating vocal plasticity. PMID:26888928
Tyack, Peter L
2008-08-01
The classic evidence for vocal production learning involves imitation of novel, often anthropogenic sounds. Among mammals, this has been reported for dolphins, elephants, harbor seals, and humans. A broader taxonomic distribution has been reported for vocal convergence, where the acoustic properties of calls from different individuals converge when they are housed together in captivity or form social bonds in the wild. Vocal convergence has been demonstrated for animals as diverse as songbirds, parakeets, hummingbirds, bats, elephants, cetaceans, and primates. For most species, call convergence is thought to reflect a group-distinctive identifier, with shared calls reflecting and strengthening social bonds. A ubiquitous function for vocal production learning that is starting to receive attention involves modifying signals to improve communication in a noisy channel. Pooling data on vocal imitation, vocal convergence, and compensation for noise suggests a wider taxonomic distribution of vocal production learning among mammals than has been generally appreciated. The wide taxonomic distribution of this evidence for vocal production learning suggests that perhaps more of the neural underpinnings for vocal production learning are in place in mammals than is usually recognized. (c) 2008 APA, all rights reserved
Auditory and audio-vocal responses of single neurons in the monkey ventral premotor cortex.
Hage, Steffen R
2018-03-20
Monkey vocalization is a complex behavioral pattern, which is flexibly used in audio-vocal communication. A recently proposed dual neural network model suggests that cognitive control might be involved in this behavior, originating from a frontal cortical network in the prefrontal cortex and mediated via projections from the rostral portion of the ventral premotor cortex (PMvr) and motor cortex to the primary vocal motor network in the brainstem. For the rapid adjustment of vocal output to external acoustic events, strong interconnections between vocal motor and auditory sites are needed, which are present at cortical and subcortical levels. However, the role of the PMvr in audio-vocal integration processes remains unclear. In the present study, single neurons in the PMvr were recorded in rhesus monkeys (Macaca mulatta) while volitionally producing vocalizations in a visual detection task or passively listening to monkey vocalizations. Ten percent of randomly selected neurons in the PMvr modulated their discharge rate in response to acoustic stimulation with species-specific calls. More than four-fifths of these auditory neurons showed an additional modulation of their discharge rates either before and/or during the monkeys' motor production of the vocalization. Based on these audio-vocal interactions, the PMvr might be well positioned to mediate higher order auditory processing with cognitive control of the vocal motor output to the primary vocal motor network. Such audio-vocal integration processes in the premotor cortex might constitute a precursor for the evolution of complex learned audio-vocal integration systems, ultimately giving rise to human speech. Copyright © 2018 Elsevier B.V. All rights reserved.
Neural FoxP2 and FoxP1 expression in the budgerigar, an avian species with adult vocal learning.
Hara, Erina; Perez, Jemima M; Whitney, Osceola; Chen, Qianqian; White, Stephanie A; Wright, Timothy F
2015-04-15
Vocal learning underlies acquisition of both language in humans and vocal signals in some avian taxa. These bird groups and humans exhibit convergent developmental phases and associated brain pathways for vocal communication. The transcription factor FoxP2 plays critical roles in vocal learning in humans and songbirds. Another member of the forkhead box gene family, FoxP1 also shows high expression in brain areas involved in vocal learning and production. Here, we investigate FoxP2 and FoxP1 mRNA and protein in adult male budgerigars (Melopsittacus undulatus), a parrot species that exhibits vocal learning as both juveniles and adults. To examine these molecules in adult vocal learners, we compared their expression patterns in the budgerigar striatal nucleus involved in vocal learning, magnocellular nucleus of the medial striatum (MMSt), across birds with different vocal states, such as vocalizing to a female (directed), vocalizing alone (undirected), and non-vocalizing. We found that both FoxP2 mRNA and protein expressions were consistently lower in MMSt than in the adjacent striatum regardless of the vocal states, whereas previous work has shown that songbirds exhibit down-regulation in the homologous region, Area X, only after singing alone. In contrast, FoxP1 levels were high in MMSt compared to the adjacent striatum in all groups. Taken together these results strengthen the general hypothesis that FoxP2 and FoxP1 have specialized expression in vocal nuclei across a range of taxa, and suggest that the adult vocal plasticity seen in budgerigars may be a product of persistent down-regulation of FoxP2 in MMSt. Copyright © 2015 Elsevier B.V. All rights reserved.
Neural FoxP2 and FoxP1 expression in the budgerigar, an avian species with adult vocal learning
Hara, Erina; Perez, Jemima M.; Whitney, Osceola; Chen, Qianqian; White, Stephanie A.; Wright, Timothy F.
2015-01-01
Vocal learning underlies acquisition of both language in humans and vocal signals in some avian taxa. These bird groups and humans exhibit convergent developmental phases and associated brain pathways for vocal communication. The transcription factor FoxP2 plays critical roles in vocal learning in humans and songbirds. Another member of the forkhead box gene family, FoxP1 also shows high expression in brain areas involved in vocal learning and production. Here, we investigate FoxP2 and FoxP1 mRNA and protein in adult male budgerigars (Melopsittacus undulatus), a parrot species that exhibits vocal learning as both juveniles and adults. To examine these molecules in adult vocal learners, we compared their expression patterns in the budgerigar striatal nucleus involved in vocal learning, magnocellular nucleus of the medial striatum (MMSt), across birds with different vocal states, such as vocalizing to a female (directed), vocalizing alone (undirected), and non-vocalizing. We found that both FoxP2 mRNA and protein expressions were consistently lower in MMSt than in the adjacent striatum regardless of the vocal states, whereas previous work has shown that songbirds exhibit downregulation in the homologous region, Area X, only after singing alone. In contrast, FoxP1 levels were high in MMSt compared to the adjacent striatum in all groups. Taken together these results strengthen the general hypothesis that FoxP2 and FoxP1 have specialized expression in vocal nuclei across a range of taxa, and suggest that the adult vocal plasticity seen in budgerigars may be a product of persistent down-regulation of FoxP2 in MMSt. PMID:25601574
Mechanisms underlying the social enhancement of vocal learning in songbirds.
Chen, Yining; Matheson, Laura E; Sakata, Jon T
2016-06-14
Social processes profoundly influence speech and language acquisition. Despite the importance of social influences, little is known about how social interactions modulate vocal learning. Like humans, songbirds learn their vocalizations during development, and they provide an excellent opportunity to reveal mechanisms of social influences on vocal learning. Using yoked experimental designs, we demonstrate that social interactions with adult tutors for as little as 1 d significantly enhanced vocal learning. Social influences on attention to song seemed central to the social enhancement of learning because socially tutored birds were more attentive to the tutor's songs than passively tutored birds, and because variation in attentiveness and in the social modulation of attention significantly predicted variation in vocal learning. Attention to song was influenced by both the nature and amount of tutor song: Pupils paid more attention to songs that tutors directed at them and to tutors that produced fewer songs. Tutors altered their song structure when directing songs at pupils in a manner that resembled how humans alter their vocalizations when speaking to infants, that was distinct from how tutors changed their songs when singing to females, and that could influence attention and learning. Furthermore, social interactions that rapidly enhanced learning increased the activity of noradrenergic and dopaminergic midbrain neurons. These data highlight striking parallels between humans and songbirds in the social modulation of vocal learning and suggest that social influences on attention and midbrain circuitry could represent shared mechanisms underlying the social modulation of vocal learning.
Mechanisms underlying the social enhancement of vocal learning in songbirds
Chen, Yining; Matheson, Laura E.; Sakata, Jon T.
2016-01-01
Social processes profoundly influence speech and language acquisition. Despite the importance of social influences, little is known about how social interactions modulate vocal learning. Like humans, songbirds learn their vocalizations during development, and they provide an excellent opportunity to reveal mechanisms of social influences on vocal learning. Using yoked experimental designs, we demonstrate that social interactions with adult tutors for as little as 1 d significantly enhanced vocal learning. Social influences on attention to song seemed central to the social enhancement of learning because socially tutored birds were more attentive to the tutor’s songs than passively tutored birds, and because variation in attentiveness and in the social modulation of attention significantly predicted variation in vocal learning. Attention to song was influenced by both the nature and amount of tutor song: Pupils paid more attention to songs that tutors directed at them and to tutors that produced fewer songs. Tutors altered their song structure when directing songs at pupils in a manner that resembled how humans alter their vocalizations when speaking to infants, that was distinct from how tutors changed their songs when singing to females, and that could influence attention and learning. Furthermore, social interactions that rapidly enhanced learning increased the activity of noradrenergic and dopaminergic midbrain neurons. These data highlight striking parallels between humans and songbirds in the social modulation of vocal learning and suggest that social influences on attention and midbrain circuitry could represent shared mechanisms underlying the social modulation of vocal learning. PMID:27247385
Recursive Vocal Pattern Learning and Generalization in Starlings
ERIC Educational Resources Information Center
Bloomfield, Tiffany Corinna
2012-01-01
Among known communication systems, human language alone exhibits open-ended productivity of meaning. Interest in the psychological mechanisms supporting this ability, and their evolutionary origins, has resurged following the suggestion that the only uniquely human ability underlying language is a mechanism of recursion. This "Unique…
Shapes and sounds as self-objects in learning geography.
Baum, E A
1978-01-01
The pleasure which some children find in maps and map reading is manifold in origin. Children cathect patterns of configuration and color and derive joy from the visual mastery of these. This gratification is enhanced by the child's knowledge that the map represents something bigger than and external to itself. Likewise, some children take pleasure in the pronunciation of names themselves. The phonetic transcription of multisyllabic names is often a plearurable challenge. The vocalized name has its origin in the self, becomes barely external to self, and is self-monitored. Thus, in children both the configurations and the vocalizations associated with map reading have the properties of "self=objects" (Kohut, 1971). From the author's observation the delight which some children take in sounding out geographic names on a map may, in some instances, indicate pre-existing gratifying sound associations. Childish amusement in punning on cognomens may be an even greater stimulant for learning than visual configurations or artificial cognitive devices.
Identification of a motor to auditory pathway important for vocal learning
Roberts, Todd F.; Hisey, Erin; Tanaka, Masashi; Kearney, Matthew; Chattree, Gaurav; Yang, Cindy F.; Shah, Nirao M.; Mooney, Richard
2017-01-01
Summary Learning to vocalize depends on the ability to adaptively modify the temporal and spectral features of vocal elements. Neurons that convey motor-related signals to the auditory system are theorized to facilitate vocal learning, but the identity and function of such neurons remain unknown. Here we identify a previously unknown neuron type in the songbird brain that transmits vocal motor signals to the auditory cortex. Genetically ablating these neurons in juveniles disrupted their ability to imitate features of an adult tutor’s song. Ablating these neurons in adults had little effect on previously learned songs, but interfered with their ability to adaptively modify the duration of vocal elements and largely prevented the degradation of song’s temporal features normally caused by deafening. These findings identify a motor to auditory circuit essential to vocal imitation and to the adaptive modification of vocal timing. PMID:28504672
Musser, Whitney B; Bowles, Ann E; Grebner, Dawn M; Crance, Jessica L
2014-10-01
Limited previous evidence suggests that killer whales (Orcinus orca) are capable of vocal production learning. However, vocal contextual learning has not been studied, nor the factors promoting learning. Vocalizations were collected from three killer whales with a history of exposure to bottlenose dolphins (Tursiops truncatus) and compared with data from seven killer whales held with conspecifics and nine bottlenose dolphins. The three whales' repertoires were distinguishable by a higher proportion of click trains and whistles. Time-domain features of click trains were intermediate between those of whales held with conspecifics and dolphins. These differences provided evidence for contextual learning. One killer whale spontaneously learned to produce artificial chirps taught to dolphins; acoustic features fell within the range of inter-individual differences among the dolphins. This whale also produced whistles similar to a stereotyped whistle produced by one dolphin. Thus, results provide further support for vocal production learning and show that killer whales are capable of contextual learning. That killer whales produce similar repertoires when associated with another species suggests substantial vocal plasticity and motivation for vocal conformity with social associates.
Fehér, Olga
2017-02-01
In this article, I argue that a comparative approach focusing on the cognitive capacities and behavioral mechanisms that underlie vocal learning in songbirds and humans can provide valuable insights into the evolutionary origins of language. The experimental approaches I discuss use abnormal song and atypical linguistic input to study the processes of individual learning, social interaction, and cultural transmission. Atypical input places increased learning and communicative pressure on learners, so exploring how they respond to this type of input provides a particularly clear picture of the biases and constraints at work during learning and use. Furthermore, simulating the cultural transmission of these unnatural communication systems in the laboratory informs us about how learning and social biases influence the structure of communication systems in the long run. Findings based on these methods suggest fundamental similarities in the basic social-cognitive mechanisms underlying vocal learning in birds and humans, and continuing research promises insights into the uniquely human mechanisms and into how human cognition and social behavior interact, and ultimately impact on the evolution of language.
Mouse vocal communication system: are ultrasounds learned or innate?
Arriaga, Gustavo; Jarvis, Erich D.
2013-01-01
Mouse ultrasonic vocalizations (USVs) are often used as behavioral readouts of internal states, to measure effects of social and pharmacological manipulations, and for behavioral phenotyping of mouse models for neuropsychiatric and neurodegenerative disorders. However, little is known about the neurobiological mechanisms of rodent USV production. Here we discuss the available data to assess whether male mouse song behavior and the supporting brain circuits resemble those of known vocal non-learning or vocal learning species. Recent neurobiology studies have demonstrated that the mouse USV brain system includes motor cortex and striatal regions, and that the vocal motor cortex sends a direct sparse projection to the brainstem vocal motor nucleus ambiguous, a projection thought be unique to humans among mammals. Recent behavioral studies have reported opposing conclusions on mouse vocal plasticity, including vocal ontogeny changes in USVs over early development that might not be explained by innate maturation processes, evidence for and against a role for auditory feedback in developing and maintaining normal mouse USVs, and evidence for and against limited vocal imitation of song pitch. To reconcile these findings, we suggest that the trait of vocal learning may not be dichotomous but encompass a broad set of behavioral and neural traits we call the continuum hypothesis, and that mice possess some of the traits associated with a capacity for limited vocal learning. PMID:23295209
Early experience shapes vocal neural coding and perception in songbirds
Woolley, Sarah M. N.
2012-01-01
Songbirds, like humans, are highly accomplished vocal learners. The many parallels between speech and birdsong and conserved features of mammalian and avian auditory systems have led to the emergence of the songbird as a model system for studying the perceptual mechanisms of vocal communication. Laboratory research on songbirds allows the careful control of early life experience and high-resolution analysis of brain function during vocal learning, production and perception. Here, I review what songbird studies have revealed about the role of early experience in the development of vocal behavior, auditory perception and the processing of learned vocalizations by auditory neurons. The findings of these studies suggest general principles for how exposure to vocalizations during development and into adulthood influences the perception of learned vocal signals. PMID:22711657
Behavior-Linked FoxP2 Regulation Enables Zebra Finch Vocal Learning
Heston, Jonathan B.
2015-01-01
Mutations in the FOXP2 transcription factor cause an inherited speech and language disorder, but how FoxP2 contributes to learning of these vocal communication signals remains unclear. FoxP2 is enriched in corticostriatal circuits of both human and songbird brains. Experimental knockdown of this enrichment in song control neurons of the zebra finch basal ganglia impairs tutor song imitation, indicating that adequate FoxP2 levels are necessary for normal vocal learning. In unmanipulated birds, vocal practice acutely downregulates FoxP2, leading to increased vocal variability and dynamic regulation of FoxP2 target genes. To determine whether this behavioral regulation is important for song learning, here, we used viral-driven overexpression of FoxP2 to counteract its downregulation. This manipulation disrupted the acute effects of song practice on vocal variability and caused inaccurate song imitation. Together, these findings indicate that dynamic behavior-linked regulation of FoxP2, rather than absolute levels, is critical for vocal learning. PMID:25698728
A hypothesis on a role of oxytocin in the social mechanisms of speech and vocal learning.
Theofanopoulou, Constantina; Boeckx, Cedric; Jarvis, Erich D
2017-08-30
Language acquisition in humans and song learning in songbirds naturally happen as a social learning experience, providing an excellent opportunity to reveal social motivation and reward mechanisms that boost sensorimotor learning. Our knowledge about the molecules and circuits that control these social mechanisms for vocal learning and language is limited. Here we propose a hypothesis of a role for oxytocin (OT) in the social motivation and evolution of vocal learning and language. Building upon existing evidence, we suggest specific neural pathways and mechanisms through which OT might modulate vocal learning circuits in specific developmental stages. © 2017 The Authors.
A hypothesis on a role of oxytocin in the social mechanisms of speech and vocal learning
Jarvis, Erich D.
2017-01-01
Language acquisition in humans and song learning in songbirds naturally happen as a social learning experience, providing an excellent opportunity to reveal social motivation and reward mechanisms that boost sensorimotor learning. Our knowledge about the molecules and circuits that control these social mechanisms for vocal learning and language is limited. Here we propose a hypothesis of a role for oxytocin (OT) in the social motivation and evolution of vocal learning and language. Building upon existing evidence, we suggest specific neural pathways and mechanisms through which OT might modulate vocal learning circuits in specific developmental stages. PMID:28835557
Tchernichovski, Ofer; Marcus, Gary
2014-01-01
Studies of vocal learning in songbirds typically focus on the acquisition of sensory templates for song imitation and on the consequent process of matching song production to templates. However, functional vocal development also requires the capacity to adaptively diverge from sensory templates, and to flexibly assemble vocal units. Examples of adaptive divergence include the corrective imitation of abnormal songs, and the decreased tendency to copy overabundant syllables. Such frequency-dependent effects might mirror tradeoffs between the assimilation of group identity (culture) while establishing individual and flexibly expressive songs. Intriguingly, although the requirements for vocal plasticity vary across songbirds, and more so between birdsong and language, the capacity to flexibly assemble vocal sounds develops in a similar, stepwise manner across species. Therefore, universal features of vocal learning go well beyond the capacity to imitate. PMID:25005823
Behavior-linked FoxP2 regulation enables zebra finch vocal learning.
Heston, Jonathan B; White, Stephanie A
2015-02-18
Mutations in the FOXP2 transcription factor cause an inherited speech and language disorder, but how FoxP2 contributes to learning of these vocal communication signals remains unclear. FoxP2 is enriched in corticostriatal circuits of both human and songbird brains. Experimental knockdown of this enrichment in song control neurons of the zebra finch basal ganglia impairs tutor song imitation, indicating that adequate FoxP2 levels are necessary for normal vocal learning. In unmanipulated birds, vocal practice acutely downregulates FoxP2, leading to increased vocal variability and dynamic regulation of FoxP2 target genes. To determine whether this behavioral regulation is important for song learning, here, we used viral-driven overexpression of FoxP2 to counteract its downregulation. This manipulation disrupted the acute effects of song practice on vocal variability and caused inaccurate song imitation. Together, these findings indicate that dynamic behavior-linked regulation of FoxP2, rather than absolute levels, is critical for vocal learning. Copyright © 2015 the authors 0270-6474/15/352885-10$15.00/0.
Differential Expression of Glutamate Receptors in Avian Neural Pathways for Learned Vocalization
WADA, KAZUHIRO; SAKAGUCHI, HIRONOBU; JARVIS, ERICH D.; HAGIWARA, MASATOSHI
2008-01-01
Learned vocalization, the substrate for human language, is a rare trait. It is found in three distantly related groups of birds—parrots, hummingbirds, and songbirds. These three groups contain cerebral vocal nuclei for learned vocalization not found in their more closely related vocal nonlearning relatives. Here, we cloned 21 receptor subunits/subtypes of all four glutamate receptor families (AMPA, kainate, NMDA, and metabotropic) and examined their expression in vocal nuclei of songbirds. We also examined expression of a subset of these receptors in vocal nuclei of hummingbirds and parrots, as well as in the brains of dove species as examples of close vocal nonlearning relatives. Among the 21 subunits/subtypes, 19 showed higher and/or lower prominent differential expression in songbird vocal nuclei relative to the surrounding brain subdivisions in which the vocal nuclei are located. This included relatively lower levels of all four AMPA subunits in lMAN, strikingly higher levels of the kainite subunit GluR5 in the robust nucleus of the arcopallium (RA), higher and lower levels respectively of the NMDA subunits NR2A and NR2B in most vocal nuclei and lower levels of the metabotropic group I subtypes (mGluR1 and -5) in most vocal nuclei and the group II subtype (mGluR2), showing a unique expression pattern of very low levels in RA and very high levels in HVC. The splice variants of AMPA subunits showed further differential expression in vocal nuclei. Some of the receptor subunits/subtypes also showed differential expression in hummingbird and parrot vocal nuclei. The magnitude of differential expression in vocal nuclei of all three vocal learners was unique compared with the smaller magnitude of differences found for nonvocal areas of vocal learners and vocal nonlearners. Our results suggest that evolution of vocal learning was accompanied by differential expression of a conserved gene family for synaptic transmission and plasticity in vocal nuclei. They also suggest that neural activity and signal transduction in vocal nuclei of vocal learners will be different relative to the surrounding brain areas. PMID:15236466
Goldstein, Michael H.; Schwade, Jennifer A.; Bornstein, Marc H.
2014-01-01
The early noncry vocalizations of infants are salient social signals. Caregivers spontaneously respond to 30-50% of these sounds, and their responsiveness to infants' prelinguistic noncry vocalizations facilitates the development of phonology and speech. Have infants learned that their vocalizations influence the behavior of social partners? If infants have learned the contingency between their vocalizing and the social responses of others, they should show an extinction burst when the contingency is removed, increasing their rate of noncry vocalizing then decreasing. Thirty-eight 5-month-olds were tested in the still-face paradigm, during which they engaged in a 2-min still-face interaction with an unfamiliar adult. When the adult assumed a still face, infants showed an extinction burst. This pattern of infant vocalizations suggests that 5-month-olds have learned the social efficacy of their vocalizations on caregivers' behavior. Furthermore, the magnitude of 5-month infants' extinction bursts predicted their language comprehension at 13 months. PMID:19489893
Prespeech motor learning in a neural network using reinforcement☆
Warlaumont, Anne S.; Westermann, Gert; Buder, Eugene H.; Oller, D. Kimbrough
2012-01-01
Vocal motor development in infancy provides a crucial foundation for language development. Some significant early accomplishments include learning to control the process of phonation (the production of sound at the larynx) and learning to produce the sounds of one’s language. Previous work has shown that social reinforcement shapes the kinds of vocalizations infants produce. We present a neural network model that provides an account of how vocal learning may be guided by reinforcement. The model consists of a self-organizing map that outputs to muscles of a realistic vocalization synthesizer. Vocalizations are spontaneously produced by the network. If a vocalization meets certain acoustic criteria, it is reinforced, and the weights are updated to make similar muscle activations increasingly likely to recur. We ran simulations of the model under various reinforcement criteria and tested the types of vocalizations it produced after learning in the differ-ent conditions. When reinforcement was contingent on the production of phonated (i.e. voiced) sounds, the network’s post learning productions were almost always phonated, whereas when reinforcement was not contingent on phonation, the network’s post-learning productions were almost always not phonated. When reinforcement was contingent on both phonation and proximity to English vowels as opposed to Korean vowels, the model’s post-learning productions were more likely to resemble the English vowels and vice versa. PMID:23275137
Prespeech motor learning in a neural network using reinforcement.
Warlaumont, Anne S; Westermann, Gert; Buder, Eugene H; Oller, D Kimbrough
2013-02-01
Vocal motor development in infancy provides a crucial foundation for language development. Some significant early accomplishments include learning to control the process of phonation (the production of sound at the larynx) and learning to produce the sounds of one's language. Previous work has shown that social reinforcement shapes the kinds of vocalizations infants produce. We present a neural network model that provides an account of how vocal learning may be guided by reinforcement. The model consists of a self-organizing map that outputs to muscles of a realistic vocalization synthesizer. Vocalizations are spontaneously produced by the network. If a vocalization meets certain acoustic criteria, it is reinforced, and the weights are updated to make similar muscle activations increasingly likely to recur. We ran simulations of the model under various reinforcement criteria and tested the types of vocalizations it produced after learning in the different conditions. When reinforcement was contingent on the production of phonated (i.e. voiced) sounds, the network's post-learning productions were almost always phonated, whereas when reinforcement was not contingent on phonation, the network's post-learning productions were almost always not phonated. When reinforcement was contingent on both phonation and proximity to English vowels as opposed to Korean vowels, the model's post-learning productions were more likely to resemble the English vowels and vice versa. Copyright © 2012 Elsevier Ltd. All rights reserved.
2014-01-01
Background Whether listening to background music enhances verbal learning performance is still a matter of dispute. In this study we investigated the influence of vocal and instrumental background music on verbal learning. Methods 226 subjects were randomly assigned to one of five groups (one control group and 4 experimental groups). All participants were exposed to a verbal learning task. One group served as control group while the 4 further groups served as experimental groups. The control group learned without background music while the 4 experimental groups were exposed to vocal or instrumental musical pieces during learning with different subjective intensity and valence. Thus, we employed 4 music listening conditions (vocal music with high intensity: VOC_HIGH, vocal music with low intensity: VOC_LOW, instrumental music with high intensity: INST_HIGH, instrumental music with low intensity: INST_LOW) and one control condition (CONT) during which the subjects learned the word lists. Since it turned out that the high and low intensity groups did not differ in terms of the rated intensity during the main experiment these groups were lumped together. Thus, we worked with 3 groups: one control group and two groups, which were exposed to background music (vocal and instrumental) during verbal learning. As dependent variable, the number of learned words was used. Here we measured immediate recall during five learning sessions (recall 1 – recall 5) and delayed recall for 15 minutes (recall 6) and 14 days (recall 7) after the last learning session. Results Verbal learning improved during the first 5 recall sessions without any strong difference between the control and experimental groups. Also the delayed recalls were similar for the three groups. There was only a trend for attenuated verbal learning for the group passively listened to vocals. This learning attenuation diminished during the following learning sessions. Conclusions The exposure to vocal or instrumental background music during encoding did not influence verbal learning. We suggest that the participants are easily able to cope with this background stimulation by ignoring this information channel in order to focus on the verbal learning task. PMID:24670048
A robotic voice simulator and the interactive training for hearing-impaired people.
Sawada, Hideyuki; Kitani, Mitsuki; Hayashi, Yasumori
2008-01-01
A talking and singing robot which adaptively learns the vocalization skill by means of an auditory feedback learning algorithm is being developed. The robot consists of motor-controlled vocal organs such as vocal cords, a vocal tract and a nasal cavity to generate a natural voice imitating a human vocalization. In this study, the robot is applied to the training system of speech articulation for the hearing-impaired, because the robot is able to reproduce their vocalization and to teach them how it is to be improved to generate clear speech. The paper briefly introduces the mechanical construction of the robot and how it autonomously acquires the vocalization skill in the auditory feedback learning by listening to human speech. Then the training system is described, together with the evaluation of the speech training by auditory impaired people.
Early life manipulations of vasopressin-family peptides alter vocal learning.
Baran, Nicole M; Peck, Samantha C; Kim, Tabitha H; Goldstein, Michael H; Adkins-Regan, Elizabeth
2017-07-26
Vocal learning from social partners is crucial for the successful development of communication in a wide range of species. Social interactions organize attention and enhance motivation to learn species-typical behaviour. However, the neurobiological mechanisms connecting social motivation and vocal learning are unknown. Using zebra finches ( Taeniopygia guttata ), a ubiquitous model for vocal learning, we show that manipulations of nonapeptide hormones in the vasopressin family (arginine vasotocin, AVT) early in development can promote or disrupt both song and social motivation. Young male zebra finches, like human infants, are socially gregarious and require interactive feedback from adult tutors to learn mature vocal forms. To investigate the role of social motivational mechanisms in song learning, in two studies, we injected hatchling males with AVT or Manning compound (MC, a nonapeptide receptor antagonist) on days 2-8 post-hatching and recorded song at maturity. In both studies, MC males produced a worse match to tutor song than controls. In study 2, which experimentally controlled for tutor and genetic factors, AVT males also learned song significantly better compared with controls. Furthermore, song similarity correlated with several measures of social motivation throughout development. These findings provide the first evidence that nonapeptides are critical to the development of vocal learning. © 2017 The Author(s).
Day, Nancy F; Kimball, Todd Haswell; Aamodt, Caitlin M; Heston, Jonathan B; Hilliard, Austin T; Xiao, Xinshu; White, Stephanie A
2018-01-01
Human speech is one of the few examples of vocal learning among mammals yet ~half of avian species exhibit this ability. Its neurogenetic basis is largely unknown beyond a shared requirement for FoxP2 in both humans and zebra finches. We manipulated FoxP2 isoforms in Area X, a song-specific region of the avian striatopallidum analogous to human anterior striatum, during a critical period for song development. We delineate, for the first time, unique contributions of each isoform to vocal learning. Weighted gene coexpression network analysis of RNA-seq data revealed gene modules correlated to singing, learning, or vocal variability. Coexpression related to singing was found in juvenile and adult Area X whereas coexpression correlated to learning was unique to juveniles. The confluence of learning and singing coexpression in juvenile Area X may underscore molecular processes that drive vocal learning in young zebra finches and, by analogy, humans. PMID:29360038
ERIC Educational Resources Information Center
Pepperberg, Irene M.
2010-01-01
This chapter briefly reviews what is known-and what remains to be understood--about Grey parrot vocal learning. I review Greys' physical capacities--issues of auditory perception and production--then discuss how these capacities are used in vocal learning and can be recruited for referential communication with humans. I discuss cross-species…
ERIC Educational Resources Information Center
Hsu, Chia-Fang
2012-01-01
This study investigated the influence of teacher vocal qualities and confirmation behaviors on student learning. Students (N = 197) enrolled in nonnative English-speaking teachers' classes completed a battery of instruments. Results indicated that both vocal qualities and confirmation behaviors were negatively related to receiver apprehension,…
Convergent Differential Regulation of Parvalbumin in the Brains of Vocal Learners
Hara, Erina; Rivas, Miriam V.; Ward, James M.; Okanoya, Kazuo; Jarvis, Erich D.
2012-01-01
Spoken language and learned song are complex communication behaviors found in only a few species, including humans and three groups of distantly related birds – songbirds, parrots, and hummingbirds. Despite their large phylogenetic distances, these vocal learners show convergent behaviors and associated brain pathways for vocal communication. However, it is not clear whether this behavioral and anatomical convergence is associated with molecular convergence. Here we used oligo microarrays to screen for genes differentially regulated in brain nuclei necessary for producing learned vocalizations relative to adjacent brain areas that control other behaviors in avian vocal learners versus vocal non-learners. A top candidate gene in our screen was a calcium-binding protein, parvalbumin (PV). In situ hybridization verification revealed that PV was expressed significantly higher throughout the song motor pathway, including brainstem vocal motor neurons relative to the surrounding brain regions of all distantly related avian vocal learners. This differential expression was specific to PV and vocal learners, as it was not found in avian vocal non-learners nor for control genes in learners and non-learners. Similar to the vocal learning birds, higher PV up-regulation was found in the brainstem tongue motor neurons used for speech production in humans relative to a non-human primate, macaques. These results suggest repeated convergent evolution of differential PV up-regulation in the brains of vocal learners separated by more than 65–300 million years from a common ancestor and that the specialized behaviors of learned song and speech may require extra calcium buffering and signaling. PMID:22238614
ERIC Educational Resources Information Center
Warren, Steven F.; Gilkerson, Jill; Richards, Jeffrey A.; Oller, D. Kimbrough; Xu, Dongxin; Yapanel, Umit; Gray, Sharmistha
2010-01-01
The study compared the vocal production and language learning environments of 26 young children with autism spectrum disorder (ASD) to 78 typically developing children using measures derived from automated vocal analysis. A digital language processor and audio-processing algorithms measured the amount of adult words to children and the amount of…
Pasch, Bret; Abbasi, Mustafa Z; Wilson, Macey; Zhao, Daniel; Searle, Jeremy B; Webster, Michael S; Rice, Aaron N
2016-04-01
Nutritional stress can have lasting impacts on the development of traits involved in vocal production. Cross-fostering experiments are often used to examine the propensity for vocal learning in a variety of taxa, but few studies assess the influence of malnourishment that can occur as a byproduct of this technique. In this study, we reciprocally cross-fostered sister taxa of voluble grasshopper mice (genus Onychomys) to explore their propensity for vocal learning. Vocalizations of Onychomys leucogaster did not differ between control and cross-fostered animals, but cross-fostered Onychomys arenicola produced vocalizations that were higher in frequency in a direction away from tutors. These same animals exhibited a transient reduction in body mass early in development, indicative of malnutrition. Our findings simultaneously refute vocal learning and support the developmental stress hypothesis to highlight the importance of early ontogeny on the production of vocalizations later in life. Copyright © 2016 Elsevier Inc. All rights reserved.
Animal models of speech and vocal communication deficits associated with psychiatric disorders
Konopka, Genevieve; Roberts, Todd F.
2015-01-01
Disruptions in speech, language and vocal communication are hallmarks of several neuropsychiatric disorders, most notably autism spectrum disorders. Historically, the use of animal models to dissect molecular pathways and connect them to behavioral endophenotypes in cognitive disorders has proven to be an effective approach for developing and testing disease-relevant therapeutics. The unique aspects of human language when compared to vocal behaviors in other animals make such an approach potentially more challenging. However, the study of vocal learning in species with analogous brain circuits to humans may provide entry points for understanding this human-specific phenotype and diseases. Here, we review animal models of vocal learning and vocal communication, and specifically link phenotypes of psychiatric disorders to relevant model systems. Evolutionary constraints in the organization of neural circuits and synaptic plasticity result in similarities in the brain mechanisms for vocal learning and vocal communication. Comparative approaches and careful consideration of the behavioral limitations among different animal models can provide critical avenues for dissecting the molecular pathways underlying cognitive disorders that disrupt speech, language and vocal communication. PMID:26232298
Auditory–vocal mirroring in songbirds
Mooney, Richard
2014-01-01
Mirror neurons are theorized to serve as a neural substrate for spoken language in humans, but the existence and functions of auditory–vocal mirror neurons in the human brain remain largely matters of speculation. Songbirds resemble humans in their capacity for vocal learning and depend on their learned songs to facilitate courtship and individual recognition. Recent neurophysiological studies have detected putative auditory–vocal mirror neurons in a sensorimotor region of the songbird's brain that plays an important role in expressive and receptive aspects of vocal communication. This review discusses the auditory and motor-related properties of these cells, considers their potential role on song learning and communication in relation to classical studies of birdsong, and points to the circuit and developmental mechanisms that may give rise to auditory–vocal mirroring in the songbird's brain. PMID:24778375
Auditory-vocal mirroring in songbirds.
Mooney, Richard
2014-01-01
Mirror neurons are theorized to serve as a neural substrate for spoken language in humans, but the existence and functions of auditory-vocal mirror neurons in the human brain remain largely matters of speculation. Songbirds resemble humans in their capacity for vocal learning and depend on their learned songs to facilitate courtship and individual recognition. Recent neurophysiological studies have detected putative auditory-vocal mirror neurons in a sensorimotor region of the songbird's brain that plays an important role in expressive and receptive aspects of vocal communication. This review discusses the auditory and motor-related properties of these cells, considers their potential role on song learning and communication in relation to classical studies of birdsong, and points to the circuit and developmental mechanisms that may give rise to auditory-vocal mirroring in the songbird's brain.
Warren, Steven F; Gilkerson, Jill; Richards, Jeffrey A; Oller, D Kimbrough; Xu, Dongxin; Yapanel, Umit; Gray, Sharmistha
2010-05-01
The study compared the vocal production and language learning environments of 26 young children with autism spectrum disorder (ASD) to 78 typically developing children using measures derived from automated vocal analysis. A digital language processor and audio-processing algorithms measured the amount of adult words to children and the amount of vocalizations they produced during 12-h recording periods in their natural environments. The results indicated significant differences between typically developing children and children with ASD in the characteristics of conversations, the number of conversational turns, and in child vocalizations that correlated with parent measures of various child characteristics. Automated measurement of the language learning environment of young children with ASD reveals important differences from the environments experienced by typically developing children.
ERIC Educational Resources Information Center
de Groot, Annette M. B.; Smedinga, Hilde E.
2014-01-01
Participants learned foreign vocabulary by means of the paired-associates learning procedure in three conditions: (a) in silence, (b) with vocal music with lyrics in a familiar language playing in the background, or (c) with vocal music with lyrics in an unfamiliar language playing in the background. The vocabulary to learn varied in concreteness…
Winograd, Claudia; Ceman, Stephanie
2012-01-01
Fragile X syndrome (FXS) is the most common cause of inherited intellectual disability and presents with markedly atypical speech-language, likely due to impaired vocal learning. Although current models have been useful for studies of some aspects of FXS, zebra finch is the only tractable lab model for vocal learning. The neural circuits for vocal learning in the zebra finch have clear relationships to the pathways in the human brain that may be affected in FXS. Further, finch vocal learning may be quantified using software designed specifically for this purpose. Knockdown of the zebra finch FMR1 gene may ultimately enable novel tests of therapies that are modality-specific, using drugs or even social strategies, to ameliorate deficits in vocal development and function. In this chapter, we describe the utility of the zebra finch model and present a hypothesis for the role of FMRP in the developing neural circuitry for vocalization.
Vocal Generalization Depends on Gesture Identity and Sequence
Sober, Samuel J.
2014-01-01
Generalization, the brain's ability to transfer motor learning from one context to another, occurs in a wide range of complex behaviors. However, the rules of generalization in vocal behavior are poorly understood, and it is unknown how vocal learning generalizes across an animal's entire repertoire of natural vocalizations and sequences. Here, we asked whether generalization occurs in a nonhuman vocal learner and quantified its properties. We hypothesized that adaptive error correction of a vocal gesture produced in one sequence would generalize to the same gesture produced in other sequences. To test our hypothesis, we manipulated the fundamental frequency (pitch) of auditory feedback in Bengalese finches (Lonchura striata var. domestica) to create sensory errors during vocal gestures (song syllables) produced in particular sequences. As hypothesized, error-corrective learning on pitch-shifted vocal gestures generalized to the same gestures produced in other sequential contexts. Surprisingly, generalization magnitude depended strongly on sequential distance from the pitch-shifted syllables, with greater adaptation for gestures produced near to the pitch-shifted syllable. A further unexpected result was that nonshifted syllables changed their pitch in the direction opposite from the shifted syllables. This apparently antiadaptive pattern of generalization could not be explained by correlations between generalization and the acoustic similarity to the pitch-shifted syllable. These findings therefore suggest that generalization depends on the type of vocal gesture and its sequential context relative to other gestures and may reflect an advantageous strategy for vocal learning and maintenance. PMID:24741046
Scheerer, N E; Jacobson, D S; Jones, J A
2016-02-09
Auditory feedback plays an important role in the acquisition of fluent speech; however, this role may change once speech is acquired and individuals no longer experience persistent developmental changes to the brain and vocal tract. For this reason, we investigated whether the role of auditory feedback in sensorimotor learning differs across children and adult speakers. Participants produced vocalizations while they heard their vocal pitch predictably or unpredictably shifted downward one semitone. The participants' vocal pitches were measured at the beginning of each vocalization, before auditory feedback was available, to assess the extent to which the deviant auditory feedback modified subsequent speech motor commands. Sensorimotor learning was observed in both children and adults, with participants' initial vocal pitch increasing following trials where they were exposed to predictable, but not unpredictable, frequency-altered feedback. Participants' vocal pitch was also measured across each vocalization, to index the extent to which the deviant auditory feedback was used to modify ongoing vocalizations. While both children and adults were found to increase their vocal pitch following predictable and unpredictable changes to their auditory feedback, adults produced larger compensatory responses. The results of the current study demonstrate that both children and adults rapidly integrate information derived from their auditory feedback to modify subsequent speech motor commands. However, these results also demonstrate that children and adults differ in their ability to use auditory feedback to generate compensatory vocal responses during ongoing vocalization. Since vocal variability also differed across the children and adult groups, these results also suggest that compensatory vocal responses to frequency-altered feedback manipulations initiated at vocalization onset may be modulated by vocal variability. Copyright © 2015 IBRO. Published by Elsevier Ltd. All rights reserved.
Vocal copying of individually distinctive signature whistles in bottlenose dolphins
King, Stephanie L.; Sayigh, Laela S.; Wells, Randall S.; Fellner, Wendi; Janik, Vincent M.
2013-01-01
Vocal learning is relatively common in birds but less so in mammals. Sexual selection and individual or group recognition have been identified as major forces in its evolution. While important in the development of vocal displays, vocal learning also allows signal copying in social interactions. Such copying can function in addressing or labelling selected conspecifics. Most examples of addressing in non-humans come from bird song, where matching occurs in an aggressive context. However, in other animals, addressing with learned signals is very much an affiliative signal. We studied the function of vocal copying in a mammal that shows vocal learning as well as complex cognitive and social behaviour, the bottlenose dolphin (Tursiops truncatus). Copying occurred almost exclusively between close associates such as mother–calf pairs and male alliances during separation and was not followed by aggression. All copies were clearly recognizable as such because copiers consistently modified some acoustic parameters of a signal when copying it. We found no evidence for the use of copying in aggression or deception. This use of vocal copying is similar to its use in human language, where the maintenance of social bonds appears to be more important than the immediate defence of resources. PMID:23427174
Cortical inter-hemispheric circuits for multimodal vocal learning in songbirds.
Paterson, Amy K; Bottjer, Sarah W
2017-10-15
Vocal learning in songbirds and humans is strongly influenced by social interactions based on sensory inputs from several modalities. Songbird vocal learning is mediated by cortico-basal ganglia circuits that include the SHELL region of lateral magnocellular nucleus of the anterior nidopallium (LMAN), but little is known concerning neural pathways that could integrate multimodal sensory information with SHELL circuitry. In addition, cortical pathways that mediate the precise coordination between hemispheres required for song production have been little studied. In order to identify candidate mechanisms for multimodal sensory integration and bilateral coordination for vocal learning in zebra finches, we investigated the anatomical organization of two regions that receive input from SHELL: the dorsal caudolateral nidopallium (dNCL SHELL ) and a region within the ventral arcopallium (Av). Anterograde and retrograde tracing experiments revealed a topographically organized inter-hemispheric circuit: SHELL and dNCL SHELL , as well as adjacent nidopallial areas, send axonal projections to ipsilateral Av; Av in turn projects to contralateral SHELL, dNCL SHELL , and regions of nidopallium adjacent to each. Av on each side also projects directly to contralateral Av. dNCL SHELL and Av each integrate inputs from ipsilateral SHELL with inputs from sensory regions in surrounding nidopallium, suggesting that they function to integrate multimodal sensory information with song-related responses within LMAN-SHELL during vocal learning. Av projections share this integrated information from the ipsilateral hemisphere with contralateral sensory and song-learning regions. Our results suggest that the inter-hemispheric pathway through Av may function to integrate multimodal sensory feedback with vocal-learning circuitry and coordinate bilateral vocal behavior. © 2017 Wiley Periodicals, Inc.
Soderstrom, Ken; Wilson, Ashley R
2013-11-01
Zebra finch song is a learned behavior dependent upon successful progress through a sensitive period of late-postnatal development. This learning is associated with maturation of distinct brain nuclei and the fiber tract interconnections between them. We have previously found remarkably distinct and dense CB1 cannabinoid receptor expression within many of these song control brain regions, implying a normal role for endocannabinoid signaling in vocal learning. Activation of CB1 receptors via daily treatments with exogenous agonist during sensorimotor stages of song learning (but not in adulthood) results in persistent alteration of song patterns. Now we are working to understand physiological changes responsible for this cannabinoid-altered vocal learning. We have found that song-altering developmental treatments are associated with changes in expression of endocannabinoid signaling elements, including CB1 receptors and the principal CNS endogenous agonist, 2-AG. Within CNS, 2-AG is produced largely through activity of the α isoform of the enzyme diacylglycerol lipase (DAGLα). To better appreciate the role of 2-AG production in normal vocal development we have determined the spatial distribution of DAGLα expression within zebra finch CNS during vocal development. Early during vocal development at 25 days, DAGLα staining is typically light and of fibroid processes. Staining peaks late in the sensorimotor stage of song learning at 75 days and is characterized by fiber, neuropil and some staining of both small and large cell somata. Results provide insight to the normal role for endocannabinoid signaling in the maturation of brain regions responsible for song learning and vocal-motor output, and suggest mechanisms by which exogenous cannabinoid exposure alters acquisition of this form of vocal communication. Copyright © 2013 Elsevier B.V. All rights reserved.
Vocal exploration is locally regulated during song learning
Ravbar, Primoz; Parra, Lucas C.; Lipkind, Dina; Tchernichovski, Ofer
2012-01-01
Exploratory variability is essential for sensory-motor learning, but it is not known how and at what time scales it is regulated. We manipulated song learning in zebra finches to experimentally control the requirements for vocal exploration in different parts of their song. We first trained birds to perform a one-syllable song, and once they mastered it we added a new syllable to the song model. Remarkably, when practicing the modified song, birds rapidly alternated between high and low acoustic variability to confine vocal exploration to the newly added syllable. Further, even within syllables, acoustic variability changed independently across song elements that were only milliseconds apart. Analysis of the entire vocal output during learning revealed that the variability of each song element decreased as it approached the target, correlating with momentary local distance from the target and less so with the overall distance. We conclude that vocal error is computed locally in sub-syllabic time scales and that song elements can be learned and crystalized independently. Songbirds have dedicated brain circuitry for vocal babbling in the anterior forebrain pathway (AFP), which generates exploratory song patterns that drive premotor neurons at the song nucleus RA (robust nucleus of the arcopallium). We hypothesize that either AFP adjusts the gain of vocal exploration in fine time scales, or that the sensitivity of RA premotor neurons to AFP/HVC inputs varies across song elements. PMID:22399765
Rehearsal Effects in Adult Word Learning
ERIC Educational Resources Information Center
Kaushanskaya, Margarita; Yoo, Jeewon
2011-01-01
The goal of this research was to examine the effects of phonological familiarity and rehearsal method (vocal vs. subvocal) on novel word learning. In Experiment 1, English-speaking adults learned phonologically familiar novel words that followed English phonological structure. Participants learned half the words via vocal rehearsal (saying the…
Repeated imitation makes human vocalizations more word-like.
Edmiston, Pierce; Perlman, Marcus; Lupyan, Gary
2018-03-14
People have long pondered the evolution of language and the origin of words. Here, we investigate how conventional spoken words might emerge from imitations of environmental sounds. Does the repeated imitation of an environmental sound gradually give rise to more word-like forms? In what ways do these forms resemble the original sounds that motivated them (i.e. exhibit iconicity)? Participants played a version of the children's game 'Telephone'. The first generation of participants imitated recognizable environmental sounds (e.g. glass breaking, water splashing). Subsequent generations imitated the previous generation of imitations for a maximum of eight generations. The results showed that the imitations became more stable and word-like, and later imitations were easier to learn as category labels. At the same time, even after eight generations, both spoken imitations and their written transcriptions could be matched above chance to the category of environmental sound that motivated them. These results show how repeated imitation can create progressively more word-like forms while continuing to retain a resemblance to the original sound that motivated them, and speak to the possible role of human vocal imitation in explaining the origins of at least some spoken words. © 2018 The Author(s).
Neural coding of syntactic structure in learned vocalizations in the songbird.
Fujimoto, Hisataka; Hasegawa, Taku; Watanabe, Dai
2011-07-06
Although vocal signals including human languages are composed of a finite number of acoustic elements, complex and diverse vocal patterns can be created from combinations of these elements, linked together by syntactic rules. To enable such syntactic vocal behaviors, neural systems must extract the sequence patterns from auditory information and establish syntactic rules to generate motor commands for vocal organs. However, the neural basis of syntactic processing of learned vocal signals remains largely unknown. Here we report that the basal ganglia projecting premotor neurons (HVC(X) neurons) in Bengalese finches represent syntactic rules that generate variable song sequences. When vocalizing an alternative transition segment between song elements called syllables, sparse burst spikes of HVC(X) neurons code the identity of a specific syllable type or a specific transition direction among the alternative trajectories. When vocalizing a variable repetition sequence of the same syllable, HVC(X) neurons not only signal the initiation and termination of the repetition sequence but also indicate the progress and state-of-completeness of the repetition. These different types of syntactic information are frequently integrated within the activity of single HVC(X) neurons, suggesting that syntactic attributes of the individual neurons are not programmed as a basic cellular subtype in advance but acquired in the course of vocal learning and maturation. Furthermore, some auditory-vocal mirroring type HVC(X) neurons display transition selectivity in the auditory phase, much as they do in the vocal phase, suggesting that these songbirds may extract syntactic rules from auditory experience and apply them to form their own vocal behaviors.
Primate feedstock for the evolution of consonants.
Lameira, Adriano R; Maddieson, Ian; Zuberbühler, Klaus
2014-02-01
The evolution of speech remains an elusive scientific problem. A widespread notion is that vocal learning, underlined by vocal-fold control, is a key prerequisite for speech evolution. Although present in birds and non-primate mammals, vocal learning is ostensibly absent in non-human primates. Here we argue that the main road to speech evolution has been through controlling the supralaryngeal vocal tract, for which we find evidence for evolutionary continuity within the great apes. Copyright © 2013 Elsevier Ltd. All rights reserved.
2018-01-01
Human vocal development is dependent on learning by imitation through social feedback between infants and caregivers. Recent studies have revealed that vocal development is also influenced by parental feedback in marmoset monkeys, suggesting vocal learning mechanisms in nonhuman primates. Marmoset infants that experience more contingent vocal feedback than their littermates develop vocalizations more rapidly, and infant marmosets with limited parental interaction exhibit immature vocal behavior beyond infancy. However, it is yet unclear whether direct parental interaction is an obligate requirement for proper vocal development because all monkeys in the aforementioned studies were able to produce the adult call repertoire after infancy. Using quantitative measures to compare distinct call parameters and vocal sequence structure, we show that social interaction has a direct impact not only on the maturation of the vocal behavior but also on acoustic call structures during vocal development. Monkeys with limited parental interaction during development show systematic differences in call entropy, a measure for maturity, compared with their normally raised siblings. In addition, different call types were occasionally uttered in motif-like sequences similar to those exhibited by vocal learners, such as birds and humans, in early vocal development. These results indicate that a lack of parental interaction leads to long-term disturbances in the acoustic structure of marmoset vocalizations, suggesting an imperative role for social interaction in proper primate vocal development. PMID:29651461
Gultekin, Yasemin B; Hage, Steffen R
2018-04-01
Human vocal development is dependent on learning by imitation through social feedback between infants and caregivers. Recent studies have revealed that vocal development is also influenced by parental feedback in marmoset monkeys, suggesting vocal learning mechanisms in nonhuman primates. Marmoset infants that experience more contingent vocal feedback than their littermates develop vocalizations more rapidly, and infant marmosets with limited parental interaction exhibit immature vocal behavior beyond infancy. However, it is yet unclear whether direct parental interaction is an obligate requirement for proper vocal development because all monkeys in the aforementioned studies were able to produce the adult call repertoire after infancy. Using quantitative measures to compare distinct call parameters and vocal sequence structure, we show that social interaction has a direct impact not only on the maturation of the vocal behavior but also on acoustic call structures during vocal development. Monkeys with limited parental interaction during development show systematic differences in call entropy, a measure for maturity, compared with their normally raised siblings. In addition, different call types were occasionally uttered in motif-like sequences similar to those exhibited by vocal learners, such as birds and humans, in early vocal development. These results indicate that a lack of parental interaction leads to long-term disturbances in the acoustic structure of marmoset vocalizations, suggesting an imperative role for social interaction in proper primate vocal development.
Killer whales are capable of vocal learning
Foote, Andrew D; Griffin, Rachael M; Howitt, David; Larsson, Lisa; Miller, Patrick J.O; Rus Hoelzel, A
2006-01-01
The production learning of vocalizations by manipulation of the sound production organs to alter the physical structure of sound has been demonstrated in only a few mammals. In this natural experiment, we document the vocal behaviour of two juvenile killer whales, Orcinus orca, separated from their natal pods, which are the only cases of dispersal seen during the three decades of observation of their populations. We find mimicry of California sea lion (Zalophus californianus) barks, demonstrating the vocal production learning ability for one of the calves. We also find differences in call usage (compared to the natal pod) that may reflect the absence of a repertoire model from tutors or some unknown effect related to isolation or context. PMID:17148275
Differential FoxP2 and FoxP1 expression in a vocal learning nucleus of the developing budgerigar
Whitney, Osceola; Voyles, Tawni; Hara, Erina; Chen, Qianqian; White, Stephanie A.; Wright, Timothy F.
2014-01-01
The forkhead domain FOXP2 and FOXP1 transcription factors are implicated in several cognitive disorders with language deficits, notably autism, and thus play a central role in learned vocal motor behavior in humans. Although a similar role for FoxP2 and FoxP1 is proposed for other vertebrate species, including songbirds, the neurodevelopmental expression of these genes are unknown in a species with lifelong vocal learning abilities. Like humans, budgerigars (Melopsittacus undulatus) learn new vocalizations throughout their entire lifetime. Like songbirds, budgerigars have distinct brain nuclei for vocal learning, which include the magnocellular nucleus of the medial striatum (MMSt), a basal ganglia region that is considered developmentally and functionally analogous to Area X in songbirds. Here we used in situ hybridization and immunohistochemistry to investigate FoxP2 and FoxP1 expression in the MMSt of juvenile and adult budgerigars. We found FoxP2 mRNA and protein expression levels in the MMSt that were lower than the surrounding striatum throughout development and adulthood. In contrast, FoxP1 mRNA and protein had an elevated MMSt/striatum expression ratio as birds matured, regardless of their sex. These results show that life-long vocal plasticity in budgerigars is associated with persistent low-level FoxP2 expression in the budgerigar MMSt, and suggests the possibility that FoxP1 plays an organizational role in the neurodevelopment of vocal motor circuitry. Thus, developmental regulation of the FoxP2 and FoxP1 genes in the basal ganglia appears essential for vocal mimicry in a range of species that possess this relatively rare trait. PMID:25407828
Differential FoxP2 and FoxP1 expression in a vocal learning nucleus of the developing budgerigar.
Whitney, Osceola; Voyles, Tawni; Hara, Erina; Chen, Qianqian; White, Stephanie A; Wright, Timothy F
2015-07-01
The forkhead domain FOXP2 and FOXP1 transcription factors are implicated in several cognitive disorders with language deficits, notably autism, and thus play a central role in learned vocal motor behavior in humans. Although a similar role for FoxP2 and FoxP1 is proposed for other vertebrate species, including songbirds, the neurodevelopmental expression of these genes are unknown in a species with lifelong vocal learning abilities. Like humans, budgerigars (Melopsittacus undulatus) learn new vocalizations throughout their entire lifetime. Like songbirds, budgerigars have distinct brain nuclei for vocal learning, which include the magnocellular nucleus of the medial striatum (MMSt), a basal ganglia region that is considered developmentally and functionally analogous to Area X in songbirds. Here, we used in situ hybridization and immunohistochemistry to investigate FoxP2 and FoxP1 expression in the MMSt of juvenile and adult budgerigars. We found FoxP2 mRNA and protein expression levels in the MMSt that were lower than the surrounding striatum throughout development and adulthood. In contrast, FoxP1 mRNA and protein had an elevated MMSt/striatum expression ratio as birds matured, regardless of their sex. These results show that life-long vocal plasticity in budgerigars is associated with persistent low-level FoxP2 expression in the budgerigar MMSt, and suggests the possibility that FoxP1 plays an organizational role in the neurodevelopment of vocal motor circuitry. Thus, developmental regulation of the FoxP2 and FoxP1 genes in the basal ganglia appears essential for vocal mimicry in a range of species that possess this relatively rare trait. © 2014 Wiley Periodicals, Inc.
Alt, Mary; Spaulding, Tammie
2011-01-01
The purpose of this study was to measure the effect of time to response in a fast-mapping word learning task for children with specific language impairment (SLI) and children with typically developing language skills (TD). Manipulating time to response allows us to examine decay of the memory trace, the use of vocal rehearsal, and their effects on word learning. Participants included 40 school-age children: half with SLI and half with TD. The children were asked to expressively and receptively fast-map 24 novel labels for 24 novel animated dinosaurs. They were asked to demonstrate learning either immediately after presentation of the novel word or after a 10-second delay. Data were collected on the use of vocal rehearsal and for recognition and production accuracy. Although the SLI group was less accurate overall, there was no evidence of decay of the memory trace. Both groups used vocal rehearsal at comparable rates, which did not vary when learning was tested immediately or after a delay. Use of vocal rehearsal resulted in better accuracy on the recognition task, but only for the TD group. A delay in time to response without interference was not an undue burden for either group. Despite the fact that children with SLI used a vocal rehearsal strategy as often as unimpaired peers, they did not benefit from the strategy in the same way as their peers. Possible explanations for these findings and clinical implications will be discussed. Readers will learn about how time to response affects word learning in children with specific language impairment and unimpaired peers. They will see how this issue fits into a framework of phonological working memory. They will also become acquainted with the effect of vocal rehearsal on word learning. Copyright © 2011 Elsevier Inc. All rights reserved.
Human mutant huntingtin disrupts vocal learning in transgenic songbirds.
Liu, Wan-Chun; Kohn, Jessica; Szwed, Sarah K; Pariser, Eben; Sepe, Sharon; Haripal, Bhagwattie; Oshimori, Naoki; Marsala, Martin; Miyanohara, Atsushi; Lee, Ramee
2015-11-01
Speech and vocal impairments characterize many neurological disorders. However, the neurogenetic mechanisms of these disorders are not well understood, and current animal models do not have the necessary circuitry to recapitulate vocal learning deficits. We developed germline transgenic songbirds, zebra finches (Taneiopygia guttata) expressing human mutant huntingtin (mHTT), a protein responsible for the progressive deterioration of motor and cognitive function in Huntington's disease (HD). Although generally healthy, the mutant songbirds had severe vocal disorders, including poor vocal imitation, stuttering, and progressive syntax and syllable degradation. Their song abnormalities were associated with HD-related neuropathology and dysfunction of the cortical-basal ganglia (CBG) song circuit. These transgenics are, to the best of our knowledge, the first experimentally created, functional mutant songbirds. Their progressive and quantifiable vocal disorder, combined with circuit dysfunction in the CBG song system, offers a model for genetic manipulation and the development of therapeutic strategies for CBG-related vocal and motor disorders.
Recent Advances in the Genetics of Vocal Learning
Condro, Michael C.; White, Stephanie A.
2015-01-01
Language is a complex communicative behavior unique to humans, and its genetic basis is poorly understood. Genes associated with human speech and language disorders provide some insights, originating with the FOXP2 transcription factor, a mutation in which is the source of an inherited form of developmental verbal dyspraxia. Subsequently, targets of FOXP2 regulation have been associated with speech and language disorders, along with other genes. Here, we review these recent findings that implicate genetic factors in human speech. Due to the exclusivity of language to humans, no single animal model is sufficient to study the complete behavioral effects of these genes. Fortunately, some animals possess subcomponents of language. One such subcomponent is vocal learning, which though rare in the animal kingdom, is shared with songbirds. We therefore discuss how songbird studies have contributed to the current understanding of genetic factors that impact human speech, and support the continued use of this animal model for such studies in the future. PMID:26052371
Bass, Andrew H.; Chagnaud, Boris P.
2012-01-01
Acoustic signaling behaviors are widespread among bony vertebrates, which include the majority of living fishes and tetrapods. Developmental studies in sound-producing fishes and tetrapods indicate that central pattern generating networks dedicated to vocalization originate from the same caudal hindbrain rhombomere (rh) 8-spinal compartment. Together, the evidence suggests that vocalization and its morphophysiological basis, including mechanisms of vocal–respiratory coupling that are widespread among tetrapods, are ancestral characters for bony vertebrates. Premotor-motor circuitry for pectoral appendages that function in locomotion and acoustic signaling develops in the same rh8-spinal compartment. Hence, vocal and pectoral phenotypes in fishes share both developmental origins and roles in acoustic communication. These findings lead to the proposal that the coupling of more highly derived vocal and pectoral mechanisms among tetrapods, including those adapted for nonvocal acoustic and gestural signaling, originated in fishes. Comparative studies further show that rh8 premotor populations have distinct neurophysiological properties coding for equally distinct behavioral attributes such as call duration. We conclude that neural network innovations in the spatiotemporal patterning of vocal and pectoral mechanisms of social communication, including forelimb gestural signaling, have their evolutionary origins in the caudal hindbrain of fishes. PMID:22723366
Learned Vocal Variation Is Associated with Abrupt Cryptic Genetic Change in a Parrot Species Complex
Ribot, Raoul F. H.; Buchanan, Katherine L.; Endler, John A.; Joseph, Leo; Bennett, Andrew T. D.; Berg, Mathew L.
2012-01-01
Contact zones between subspecies or closely related species offer valuable insights into speciation processes. A typical feature of such zones is the presence of clinal variation in multiple traits. The nature of these traits and the concordance among clines are expected to influence whether and how quickly speciation will proceed. Learned signals, such as vocalizations in species having vocal learning (e.g. humans, many birds, bats and cetaceans), can exhibit rapid change and may accelerate reproductive isolation between populations. Therefore, particularly strong concordance among clines in learned signals and population genetic structure may be expected, even among continuous populations in the early stages of speciation. However, empirical evidence for this pattern is often limited because differences in vocalisations between populations are driven by habitat differences or have evolved in allopatry. We tested for this pattern in a unique system where we may be able to separate effects of habitat and evolutionary history. We studied geographic variation in the vocalizations of the crimson rosella (Platycercus elegans) parrot species complex. Parrots are well known for their life-long vocal learning and cognitive abilities. We analysed contact calls across a ca 1300 km transect encompassing populations that differed in neutral genetic markers and plumage colour. We found steep clinal changes in two acoustic variables (fundamental frequency and peak frequency position). The positions of the two clines in vocal traits were concordant with a steep cline in microsatellite-based genetic variation, but were discordant with the steep clines in mtDNA, plumage and habitat. Our study provides new evidence that vocal variation, in a species with vocal learning, can coincide with areas of restricted gene flow across geographically continuous populations. Our results suggest that traits that evolve culturally can be strongly associated with reduced gene flow between populations, and therefore may promote speciation, even in the absence of other barriers. PMID:23227179
An Automated Procedure for Evaluating Song Imitation
Mandelblat-Cerf, Yael; Fee, Michale S.
2014-01-01
Songbirds have emerged as an excellent model system to understand the neural basis of vocal and motor learning. Like humans, songbirds learn to imitate the vocalizations of their parents or other conspecific “tutors.” Young songbirds learn by comparing their own vocalizations to the memory of their tutor song, slowly improving until over the course of several weeks they can achieve an excellent imitation of the tutor. Because of the slow progression of vocal learning, and the large amounts of singing generated, automated algorithms for quantifying vocal imitation have become increasingly important for studying the mechanisms underlying this process. However, methodologies for quantifying song imitation are complicated by the highly variable songs of either juvenile birds or those that learn poorly because of experimental manipulations. Here we present a method for the evaluation of song imitation that incorporates two innovations: First, an automated procedure for selecting pupil song segments, and, second, a new algorithm, implemented in Matlab, for computing both song acoustic and sequence similarity. We tested our procedure using zebra finch song and determined a set of acoustic features for which the algorithm optimally differentiates between similar and non-similar songs. PMID:24809510
Precise auditory-vocal mirroring in neurons for learned vocal communication.
Prather, J F; Peters, S; Nowicki, S; Mooney, R
2008-01-17
Brain mechanisms for communication must establish a correspondence between sensory and motor codes used to represent the signal. One idea is that this correspondence is established at the level of single neurons that are active when the individual performs a particular gesture or observes a similar gesture performed by another individual. Although neurons that display a precise auditory-vocal correspondence could facilitate vocal communication, they have yet to be identified. Here we report that a certain class of neurons in the swamp sparrow forebrain displays a precise auditory-vocal correspondence. We show that these neurons respond in a temporally precise fashion to auditory presentation of certain note sequences in this songbird's repertoire and to similar note sequences in other birds' songs. These neurons display nearly identical patterns of activity when the bird sings the same sequence, and disrupting auditory feedback does not alter this singing-related activity, indicating it is motor in nature. Furthermore, these neurons innervate striatal structures important for song learning, raising the possibility that singing-related activity in these cells is compared to auditory feedback to guide vocal learning.
Cetacean vocal learning and communication.
Janik, Vincent M
2014-10-01
The cetaceans are one of the few mammalian clades capable of vocal production learning. Evidence for this comes from synchronous changes in song patterns of baleen whales and experimental work on toothed whales in captivity. While baleen whales like many vocal learners use this skill in song displays that are involved in sexual selection, toothed whales use learned signals in individual recognition and the negotiation of social relationships. Experimental studies demonstrated that dolphins can use learned signals referentially. Studies on wild dolphins demonstrated how this skill appears to be useful in their own communication system, making them an interesting subject for comparative communication studies. Copyright © 2014. Published by Elsevier Ltd.
Convergent transcriptional specializations in the brains of humans and song-learning birds
Pfenning, Andreas R.; Hara, Erina; Whitney, Osceola; Rivas, Miriam V.; Wang, Rui; Roulhac, Petra L.; Howard, Jason T.; Wirthlin, Morgan; Lovell, Peter V.; Ganapathy, Ganeshkumar; Mouncastle, Jacquelyn; Moseley, M. Arthur; Thompson, J. Will; Soderblom, Erik J.; Iriki, Atsushi; Kato, Masaki; Gilbert, M. Thomas P.; Zhang, Guojie; Bakken, Trygve; Bongaarts, Angie; Bernard, Amy; Lein, Ed; Mello, Claudio V.; Hartemink, Alexander J.; Jarvis, Erich D.
2015-01-01
Song-learning birds and humans share independently evolved similarities in brain pathways for vocal learning that are essential for song and speech and are not found in most other species. Comparisons of brain transcriptomes of song-learning birds and humans relative to vocal nonlearners identified convergent gene expression specializations in specific song and speech brain regions of avian vocal learners and humans. The strongest shared profiles relate bird motor and striatal song-learning nuclei, respectively, with human laryngeal motor cortex and parts of the striatum that control speech production and learning. Most of the associated genes function in motor control and brain connectivity. Thus, convergent behavior and neural connectivity for a complex trait are associated with convergent specialized expression of multiple genes. PMID:25504733
A duetting perspective on avian song learning.
Rivera-Cáceres, Karla D; Templeton, Christopher N
2017-12-25
Avian song learning has a rich history of study and has become the preeminent system for understanding the ontogeny of vocal communication in animals. Song learning in birds has many parallels with human language learning, ranging from the neural mechanisms involved to the importance of social factors in shaping signal acquisition. While much has been learned about the process of song learning, virtually all of the research done to date has focused on temperate species, where often only one sex (the male) sings. Duetting species, in which both males and females learn to sing and learn to combine their songs into temporally coordinated joint displays, could provide many insights into the processes by which vocal learning takes place. Here we highlight three key features of song learning-neuroendocrine control mechanisms, timing and life history stages of song acquisition, and the role of social factors in song selection and use-that have been elucidated from species where only males sing, and compare these with duetting species. We summarize what is known about song learning in duetting species and then provide several suggestions for fruitful directions for future research. We suggest that focusing research efforts on duetting species could significantly advance our understanding of vocal learning in birds and further cement the importance of avian species as models for understanding human conversations and the processes of vocal learning more broadly. Copyright © 2017 Elsevier B.V. All rights reserved.
Learned vocal and breathing behavior in an enculturated gorilla.
Perlman, Marcus; Clark, Nathaniel
2015-09-01
We describe the repertoire of learned vocal and breathing-related behaviors (VBBs) performed by the enculturated gorilla Koko. We examined a large video corpus of Koko and observed 439 VBBs spread across 161 bouts. Our analysis shows that Koko exercises voluntary control over the performance of nine distinctive VBBs, which involve variable coordination of her breathing, larynx, and supralaryngeal articulators like the tongue and lips. Each of these behaviors is performed in the context of particular manual action routines and gestures. Based on these and other findings, we suggest that vocal learning and the ability to exercise volitional control over vocalization, particularly in a multimodal context, might have figured relatively early into the evolution of language, with some rudimentary capacity in place at the time of our last common ancestor with great apes.
Rodenas-Cuadrado, Pedro M; Mengede, Janine; Baas, Laura; Devanna, Paolo; Schmid, Tobias A; Yartsev, Michael; Firzlaff, Uwe; Vernes, Sonja C
2018-06-01
Genes including FOXP2, FOXP1, and CNTNAP2, have been implicated in human speech and language phenotypes, pointing to a role in the development of normal language-related circuitry in the brain. Although speech and language are unique to humans a comparative approach is possible by addressing language-relevant traits in animal systems. One such trait, vocal learning, represents an essential component of human spoken language, and is shared by cetaceans, pinnipeds, elephants, some birds and bats. Given their vocal learning abilities, gregarious nature, and reliance on vocalizations for social communication and navigation, bats represent an intriguing mammalian system in which to explore language-relevant genes. We used immunohistochemistry to detail the distribution of FoxP2, FoxP1, and Cntnap2 proteins, accompanied by detailed cytoarchitectural histology in the brains of two vocal learning bat species; Phyllostomus discolor and Rousettus aegyptiacus. We show widespread expression of these genes, similar to what has been previously observed in other species, including humans. A striking difference was observed in the adult P. discolor bat, which showed low levels of FoxP2 expression in the cortex that contrasted with patterns found in rodents and nonhuman primates. We created an online, open-access database within which all data can be browsed, searched, and high resolution images viewed to single cell resolution. The data presented herein reveal regions of interest in the bat brain and provide new opportunities to address the role of these language-related genes in complex vocal-motor and vocal learning behaviors in a mammalian model system. © 2018 The Authors The Journal of Comparative Neurology Published by Wiley Periodicals, Inc.
Carey, Daniel; McGettigan, Carolyn
2017-04-01
The human vocal system is highly plastic, allowing for the flexible expression of language, mood and intentions. However, this plasticity is not stable throughout the life span, and it is well documented that adult learners encounter greater difficulty than children in acquiring the sounds of foreign languages. Researchers have used magnetic resonance imaging (MRI) to interrogate the neural substrates of vocal imitation and learning, and the correlates of individual differences in phonetic "talent". In parallel, a growing body of work using MR technology to directly image the vocal tract in real time during speech has offered primarily descriptive accounts of phonetic variation within and across languages. In this paper, we review the contribution of neural MRI to our understanding of vocal learning, and give an overview of vocal tract imaging and its potential to inform the field. We propose methods by which our understanding of speech production and learning could be advanced through the combined measurement of articulation and brain activity using MRI - specifically, we describe a novel paradigm, developed in our laboratory, that uses both MRI techniques to for the first time map directly between neural, articulatory and acoustic data in the investigation of vocalisation. This non-invasive, multimodal imaging method could be used to track central and peripheral correlates of spoken language learning, and speech recovery in clinical settings, as well as provide insights into potential sites for targeted neural interventions. Copyright © 2016 Elsevier Ltd. All rights reserved.
Hayase, Shin; Wada, Kazuhiro
2018-06-23
Learned vocalization, including birdsong and human speech, is acquired through self-motivated vocal practice during the sensitive period of vocal learning. The zebra finch (Taeniopygia guttata) develops a song characterized by vocal variability and crystalizes a defined song pattern as adulthood. However, it remains unknown how vocal variability is regulated with diurnal singing during the sensorimotor learning period. Here, we investigated the expression of activity-dependent neuroplasticity-related gene Arc during the early plastic song phase to examine its potential association with vocal plasticity. We first confirmed that multiple acoustic features of syllables in the plastic song were dramatically and simultaneously modulated during the first 3 hours of singing in a day and the altered features were maintained until sleep. Concurrently, Arc was intensely induced during morning singing and a subsequent attenuation during afternoon singing in the robust nucleus of the arcopallium (RA) and the interfacial nucleus of the nidopallium (NIf). The singing-driven Arc expression was not altered by circadian rhythm, but rather reduced during the day as juveniles produced more songs. Song stabilization accelerated by testosterone administration in juveniles was accompanied with attenuation of Arc induction in RA and NIf. In contrast, although early-deafened birds produced highly unstable song even at adulthood, singing-driven Arc expression was not different between intact and early-deafened adults. These results suggest a potential functional link between Arc expression in RA and NIf and vocal plasticity during the sensorimotor phase of song learning. Nonetheless, Arc expression did not reflect the quality of bird's own song or auditory feedback. This article is protected by copyright. All rights reserved. This article is protected by copyright. All rights reserved.
Auditory Signal Processing in Communication: Perception and Performance of Vocal Sounds
Prather, Jonathan F.
2013-01-01
Learning and maintaining the sounds we use in vocal communication require accurate perception of the sounds we hear performed by others and feedback-dependent imitation of those sounds to produce our own vocalizations. Understanding how the central nervous system integrates auditory and vocal-motor information to enable communication is a fundamental goal of systems neuroscience, and insights into the mechanisms of those processes will profoundly enhance clinical therapies for communication disorders. Gaining the high-resolution insight necessary to define the circuits and cellular mechanisms underlying human vocal communication is presently impractical. Songbirds are the best animal model of human speech, and this review highlights recent insights into the neural basis of auditory perception and feedback-dependent imitation in those animals. Neural correlates of song perception are present in auditory areas, and those correlates are preserved in the auditory responses of downstream neurons that are also active when the bird sings. Initial tests indicate that singing-related activity in those downstream neurons is associated with vocal-motor performance as opposed to the bird simply hearing itself sing. Therefore, action potentials related to auditory perception and action potentials related to vocal performance are co-localized in individual neurons. Conceptual models of song learning involve comparison of vocal commands and the associated auditory feedback to compute an error signal that is used to guide refinement of subsequent song performances, yet the sites of that comparison remain unknown. Convergence of sensory and motor activity onto individual neurons points to a possible mechanism through which auditory and vocal-motor signals may be linked to enable learning and maintenance of the sounds used in vocal communication. PMID:23827717
The response of the anterior striatum during adult human vocal learning
Leech, Robert; Iverson, Paul; Wise, Richard J. S.
2014-01-01
Research on mammals predicts that the anterior striatum is a central component of human motor learning. However, because vocalizations in most mammals are innate, much of the neurobiology of human vocal learning has been inferred from studies on songbirds. Essential for song learning is a pathway, the homolog of mammalian cortical-basal ganglia “loops,” which includes the avian striatum. The present functional magnetic resonance imaging (fMRI) study investigated adult human vocal learning, a skill that persists throughout life, albeit imperfectly given that late-acquired languages are spoken with an accent. Monolingual adult participants were scanned while repeating novel non-native words. After training on the pronunciation of half the words for 1 wk, participants underwent a second scan. During scanning there was no external feedback on performance. Activity declined sharply in left and right anterior striatum, both within and between scanning sessions, and this change was independent of training and performance. This indicates that adult speakers rapidly adapt to the novel articulatory movements, possibly by using motor sequences from their native speech to approximate those required for the novel speech sounds. Improved accuracy correlated only with activity in motor-sensory perisylvian cortex. We propose that future studies on vocal learning, using different behavioral and pharmacological manipulations, will provide insights into adult striatal plasticity and its potential for modification in both educational and clinical contexts. PMID:24805076
Complex coevolution of wing, tail, and vocal sounds of courting male bee hummingbirds.
Clark, Christopher J; McGuire, Jimmy A; Bonaccorso, Elisa; Berv, Jacob S; Prum, Richard O
2018-03-01
Phenotypic characters with a complex physical basis may have a correspondingly complex evolutionary history. Males in the "bee" hummingbird clade court females with sound from tail-feathers, which flutter during display dives. On a phylogeny of 35 species, flutter sound frequency evolves as a gradual, continuous character on most branches. But on at least six internal branches fall two types of major, saltational changes: mode of flutter changes, or the feather that is the sound source changes, causing frequency to jump from one discrete value to another. In addition to their tail "instruments," males also court females with sound from their syrinx and wing feathers, and may transfer or switch instruments over evolutionary time. In support of this, we found a negative phylogenetic correlation between presence of wing trills and singing. We hypothesize this transference occurs because wing trills and vocal songs serve similar functions and are thus redundant. There are also three independent origins of self-convergence of multiple signals, in which the same species produces both a vocal (sung) frequency sweep, and a highly similar nonvocal sound. Moreover, production of vocal, learned song has been lost repeatedly. Male bee hummingbirds court females with a diverse, coevolving array of acoustic traits. © 2018 The Author(s). Evolution © 2018 The Society for the Study of Evolution.
ERIC Educational Resources Information Center
Riede, Tobias; Goller, Franz
2010-01-01
Song production in songbirds is a model system for studying learned vocal behavior. As in humans, bird phonation involves three main motor systems (respiration, vocal organ and vocal tract). The avian respiratory mechanism uses pressure regulation in air sacs to ventilate a rigid lung. In songbirds sound is generated with two independently…
Broaddus-Lawrence, P L; Treole, K; McCabe, R B; Allen, R L; Toppin, L
2000-03-01
The purpose of the present study was to determine the effects of vocal hygiene education on the vocal hygiene behaviors and perceptual vocal characteristics of untrained singers. Eleven adult untrained singers served as subjects. They attended four 1-hour class sessions on vocal hygiene, including anatomy and physiology of the phonatory mechanism, vocally abusive behaviors, voice disorders commonly seen in singers, and measures to prevent voice disorders. Pre- and postinstruction surveys were used to record subjects' vocal abuses and their perceptions of their speaking and singing voice. They also rated their perceived value of vocal hygiene education. Results revealed minimal changes in vocal hygiene behaviors and perceptual voice characteristics. The subjects did report a high degree of benefit and learning, however.
Panaitof, S. Carmen; Abrahams, Brett S.; Dong, Hongmei; Geschwind, Daniel H.; White, Stephanie A.
2010-01-01
Multiple studies, involving distinct clinical populations, implicate contactin associated protein-like 2 (CNTNAP2) in aspects of language development and performance. While CNTNAP2 is broadly distributed in developing rodent brain, it shows a striking gradient of frontal cortical enrichment in developing human brain, consistent with a role in patterning circuits that subserve higher cognition and language. To test the hypothesis that CNTNAP2 may be important for learned vocal communication in additional species, we employed in situ hybridization to characterize transcript distribution in the zebra finch, an experimentally tractable songbird for which the neural substrate of this behavior is well-established. Consistent with an important role in learned vocalization, Cntnap2 was enriched or diminished in key song control nuclei relative to adjacent brain tissue. Importantly, this punctuated expression was observed in males, but not females, in accord with the sexual dimorphism of neural circuitry and vocal learning in this species. Ongoing functional work will provide important insights into the relationship between Cntnap2 and vocal communication in songbirds and thereby clarify mechanisms at play in disorders of human cognition and language. PMID:20394055
Zhou, Xin; Fu, Xin; Lin, Chun; Zhou, Xiaojuan; Liu, Jin; Wang, Li; Zhang, Xinwen; Zuo, Mingxue; Fan, Xiaolong; Li, Dapeng; Sun, Yingyu
2017-05-01
Deafening elicits a deterioration of learned vocalization, in both humans and songbirds. In songbirds, learned vocal plasticity has been shown to depend on the basal ganglia-cortical circuit, but the underlying cellular basis remains to be clarified. Using confocal imaging and electron microscopy, we examined the effect of deafening on dendritic spines in avian vocal motor cortex, the robust nucleus of the arcopallium (RA), and investigated the role of the basal ganglia circuit in motor cortex plasticity. We found rapid structural changes to RA dendritic spines in response to hearing loss, accompanied by learned song degradation. In particular, the morphological characters of RA spine synaptic contacts between 2 major pathways were altered differently. However, experimental disruption of the basal ganglia circuit, through lesions in song-specialized basal ganglia nucleus Area X, largely prevented both the observed changes to RA dendritic spines and the song deterioration after hearing loss. Our results provide cellular evidence to highlight a key role of the basal ganglia circuit in the motor cortical plasticity that underlies learned vocal plasticity. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Tokarev, Kirill; Tiunova, Anna; Scharff, Constance; Anokhin, Konstantin
2011-01-01
Specialized neural pathways, the song system, are required for acquiring, producing, and perceiving learned avian vocalizations. Birds that do not learn to produce their vocalizations lack telencephalic song system components. It is not known whether the song system forebrain regions are exclusively evolved for song or whether they also process information not related to song that might reflect their 'evolutionary history'. To address this question we monitored the induction of two immediate-early genes (IEGs) c-Fos and ZENK in various regions of the song system in zebra finches (Taeniopygia guttata) in response to an aversive food learning paradigm; this involves the association of a food item with a noxious stimulus that affects the oropharyngeal-esophageal cavity and tongue, causing subsequent avoidance of that food item. The motor response results in beak and head movements but not vocalizations. IEGs have been extensively used to map neuro-molecular correlates of song motor production and auditory processing. As previously reported, neurons in two pallial vocal motor regions, HVC and RA, expressed IEGs after singing. Surprisingly, c-Fos was induced equivalently also after food aversion learning in the absence of singing. The density of c-Fos positive neurons was significantly higher than that of birds in control conditions. This was not the case in two other pallial song nuclei important for vocal plasticity, LMAN and Area X, although singing did induce IEGs in these structures, as reported previously. Our results are consistent with the possibility that some of the song nuclei may participate in non-vocal learning and the populations of neurons involved in the two tasks show partial overlap. These findings underscore the previously advanced notion that the specialized forebrain pre-motor nuclei controlling song evolved from circuits involved in behaviors related to feeding.
Ghassemi, Marzyeh; Van Stan, Jarrad H; Mehta, Daryush D; Zañartu, Matías; Cheyne, Harold A; Hillman, Robert E; Guttag, John V
2014-06-01
Voice disorders are medical conditions that often result from vocal abuse/misuse which is referred to generically as vocal hyperfunction. Standard voice assessment approaches cannot accurately determine the actual nature, prevalence, and pathological impact of hyperfunctional vocal behaviors because such behaviors can vary greatly across the course of an individual's typical day and may not be clearly demonstrated during a brief clinical encounter. Thus, it would be clinically valuable to develop noninvasive ambulatory measures that can reliably differentiate vocal hyperfunction from normal patterns of vocal behavior. As an initial step toward this goal we used an accelerometer taped to the neck surface to provide a continuous, noninvasive acceleration signal designed to capture some aspects of vocal behavior related to vocal cord nodules, a common manifestation of vocal hyperfunction. We gathered data from 12 female adult patients diagnosed with vocal fold nodules and 12 control speakers matched for age and occupation. We derived features from weeklong neck-surface acceleration recordings by using distributions of sound pressure level and fundamental frequency over 5-min windows of the acceleration signal and normalized these features so that intersubject comparisons were meaningful. We then used supervised machine learning to show that the two groups exhibit distinct vocal behaviors that can be detected using the acceleration signal. We were able to correctly classify 22 of the 24 subjects, suggesting that in the future measures of the acceleration signal could be used to detect patients with the types of aberrant vocal behaviors that are associated with hyperfunctional voice disorders.
Alt, Mary; Spaulding, Tammie
2011-01-01
Purpose The purpose of this study was to measure the effect of time to response in a fast-mapping word learning task for children with Specific Language Impairment (SLI) and children with typically-developing language skills (TD). Manipulating time to response allows us to examine decay of the memory trace, the use of vocal rehearsal, and their effects on word learning. Method Participants included 40 school-age children: half with SLI and half with TD. The children were asked to expressively and receptively fast-map 24 novel labels for 24 novel animated dinosaurs. They were asked to demonstrate learning either immediately after presentation of the novel word or after a 10-second delay. Data were collected on the use of vocal rehearsal and for recognition and production accuracy. Results Although the SLI group was less accurate overall, there was no evidence of decay of the memory trace. Both groups used vocal rehearsal at comparable rates, which did not vary when learning was tested immediately or after a delay. Use of vocal rehearsal resulted in better accuracy on the recognition task, but only for the TD group. Conclusions A delay in time to response without interference was not an undue burden for either group. Despite the fact that children with SLI used a vocal rehearsal strategy as often as unimpaired peers, they did not benefit from the strategy in the same way as their peers. Possible explanations for these findings and clinical implications will be discussed. PMID:21885056
Benichov, Jonathan I; Globerson, Eitan; Tchernichovski, Ofer
2016-01-01
Humans and oscine songbirds share the rare capacity for vocal learning. Songbirds have the ability to acquire songs and calls of various rhythms through imitation. In several species, birds can even coordinate the timing of their vocalizations with other individuals in duets that are synchronized with millisecond-accuracy. It is not known, however, if songbirds can perceive rhythms holistically nor if they are capable of spontaneous entrainment to complex rhythms, in a manner similar to humans. Here we review emerging evidence from studies of rhythm generation and vocal coordination across songbirds and humans. In particular, recently developed experimental methods have revealed neural mechanisms underlying the temporal structure of song and have allowed us to test birds' abilities to predict the timing of rhythmic social signals. Surprisingly, zebra finches can readily learn to anticipate the calls of a "vocal robot" partner and alter the timing of their answers to avoid jamming, even in reference to complex rhythmic patterns. This capacity resembles, to some extent, human predictive motor response to an external beat. In songbirds, this is driven, at least in part, by the forebrain song system, which controls song timing and is essential for vocal learning. Building upon previous evidence for spontaneous entrainment in human and non-human vocal learners, we propose a comparative framework for future studies aimed at identifying shared mechanism of rhythm production and perception across songbirds and humans.
Vocal Learning via Social Reinforcement by Infant Marmoset Monkeys.
Takahashi, Daniel Y; Liao, Diana A; Ghazanfar, Asif A
2017-06-19
For over half a century now, primate vocalizations have been thought to undergo little or no experience-dependent acoustic changes during development [1]. If any changes are apparent, then they are routinely (and quite reasonably) attributed to the passive consequences of growth. Indeed, previous experiments on squirrel monkeys and macaque monkeys showed that social isolation [2, 3], deafness [2], cross-fostering [4] and parental absence [5] have little or no effect on vocal development. Here, we explicitly test in marmoset monkeys-a very vocal and cooperatively breeding species [6]-whether the transformation of immature into mature contact calls by infants is influenced by contingent parental vocal feedback. Using a closed-loop design, we experimentally provided more versus less contingent vocal feedback to twin infant marmoset monkeys over their first 2 months of life, the interval during which their contact calls transform from noisy, immature calls to tonal adult-like "phee" calls [7, 8]. Infants who received more contingent feedback had a faster rate of vocal development, producing mature-sounding contact calls earlier than the other twin. The differential rate of vocal development was not linked to genetics, perinatal experience, or body growth; nor did the amount of contingency influence the overall rate of spontaneous vocal production. Thus, we provide the first experimental evidence for production-related vocal learning during the development of a nonhuman primate. Copyright © 2017 Elsevier Ltd. All rights reserved.
Evolutionary Origins for Social Vocalization in a Vertebrate Hindbrain–Spinal Compartment
Bass, Andrew H.; Gilland, Edwin H.; Baker, Robert
2008-01-01
The macroevolutionary events leading to neural innovations for social communication, such as vocalization, are essentially unexplored. Many fish vocalize during female courtship and territorial defense, as do amphibians, birds, and mammals. Here, we map the neural circuitry for vocalization in larval fish and show that the vocal network develops in a segment-like region across the most caudal hindbrain and rostral spinal cord. Taxonomic analysis demonstrates a highly conserved pattern between fish and all major lineages of vocal tetrapods. We propose that the vocal basis for acoustic communication among vertebrates evolved from an ancestrally shared developmental compartment already present in the early fishes. PMID:18635807
Sleep, offline processing, and vocal learning
Margoliash, Daniel; Schmidt, Marc F
2009-01-01
The study of song learning and the neural song system has provided an important comparative model system for the study of speech and language acquisition. We describe some recent advances in the bird song system, focusing on the role of offline processing including sleep in processing sensory information and in guiding developmental song learning. These observations motivate a new model of the organization and role of the sensory memories in vocal learning. PMID:19906416
Mind the gap: Neural coding of species identity in birdsong prosody.
Araki, Makoto; Bandi, M M; Yazaki-Sugiyama, Yoko
2016-12-09
Juvenile songbirds learn vocal communication from adult tutors of the same species but not from adults of other species. How species-specific learning emerges from the basic features of song prosody remains unknown. In the zebra finch auditory cortex, we discovered a class of neurons that register the silent temporal gaps between song syllables and are distinct from neurons encoding syllable morphology. Behavioral learning and neuronal coding of temporal gap structure resisted song tutoring from other species: Zebra finches fostered by Bengalese finch parents learned Bengalese finch song morphology transposed onto zebra finch temporal gaps. During the vocal learning period, temporal gap neurons fired selectively to zebra finch song. The innate temporal coding of intersyllable silent gaps suggests a neuronal barcode for conspecific vocal learning and social communication in acoustically diverse environments. Copyright © 2016, American Association for the Advancement of Science.
FoxP2 in song-learning birds and vocal-learning mammals.
Webb, D M; Zhang, J
2005-01-01
FoxP2 is the first identified gene that is specifically involved in speech and language development in humans. Population genetic studies of FoxP2 revealed a selective sweep in recent human history associated with two amino acid substitutions in exon 7. Avian song learning and human language acquisition share many behavioral and neurological similarities. To determine whether FoxP2 plays a similar role in song-learning birds, we sequenced exon 7 of FoxP2 in multiple song-learning and nonlearning birds. We show extreme conservation of FoxP2 sequences in birds, including unusually low rates of synonymous substitutions. However, no amino acid substitutions are shared between the song-learning birds and humans. Furthermore, sequences from vocal-learning whales, dolphins, and bats do not share the human-unique substitutions. While FoxP2 appears to be under strong functional constraints in mammals and birds, we find no evidence for its role during the evolution of vocal learning in nonhuman animals as in humans.
Soma, Masayo; Mori, Chihiro
2015-01-01
Music and dance are two remarkable human characteristics that are closely related. Communication through integrated vocal and motional signals is also common in the courtship displays of birds. The contribution of songbird studies to our understanding of vocal learning has already shed some light on the cognitive underpinnings of musical ability. Moreover, recent pioneering research has begun to show how animals can synchronize their behaviors with external stimuli, like metronome beats. However, few studies have applied such perspectives to unraveling how animals can integrate multimodal communicative signals that have natural functions. Additionally, studies have rarely asked how well these behaviors are learned. With this in mind, here we cast a spotlight on an unusual animal behavior: non-vocal sound production associated with singing in the Java sparrow (Lonchura oryzivora), a songbird. We show that male Java sparrows coordinate their bill-click sounds with the syntax of their song-note sequences, similar to percussionists. Analysis showed that they produced clicks frequently toward the beginning of songs and before/after specific song notes. We also show that bill-clicking patterns are similar between social fathers and their sons, suggesting that these behaviors might be learned from models or linked to learning-based vocalizations. Individuals untutored by conspecifics also exhibited stereotypical bill-clicking patterns in relation to song-note sequence, indicating that while the production of bill clicking itself is intrinsic, its syncopation appears to develop with songs. This paints an intriguing picture in which non-vocal sounds are integrated with vocal courtship signals in a songbird, a model that we expect will contribute to the further understanding of multimodal communication. PMID:25992841
Alderete, Tanya L.; Chang, Daniel
2010-01-01
The cortical nucleus LMAN (lateral magnocellular nucleus of the anterior nidopallium) provides the output of a basal ganglia pathway that is necessary for acquisition of learned vocal behavior during development in songbirds. LMAN is composed of two subregions, a core and a surrounding shell, that give rise to independent pathways that traverse the forebrain in parallel. The LMANshell pathway forms a recurrent loop that includes a cortical region, the dorsal region of the caudolateral nidopallium (dNCL), hitherto unknown to be involved with learned vocal behavior. Here we show that vocal production strongly induces the IEG product ZENK in dNCL of zebra finches. Hearing tutor song while singing is more effective at inducing expression in dNCL of juvenile birds during the auditory–motor integration stage of vocal learning than is hearing conspecific song. In contrast, hearing conspecific song is relatively more effective at inducing expression in adult birds, regardless of whether they are producing song. Furthermore, ZENK+ neurons in dNCL include projection neurons that are part of the LMANshell recurrent loop and a high proportion of dNCL projection neurons express ZENK in singing juvenile birds that hear tutor song. Thus juvenile birds that are actively refining their vocal pattern to imitate a tutor song show high levels of ZENK induction in dNCL neurons when they are singing while hearing the song of their tutor and low levels when they hear a novel conspecific. This pattern indicates that dNCL is a novel brain region involved with vocal learning and that its function is developmentally regulated. PMID:20107119
Van Stan, Jarrad H; Mehta, Daryush D; Petit, Robert J; Sternad, Dagmar; Muise, Jason; Burns, James A; Hillman, Robert E
2017-02-01
Ambulatory voice biofeedback (AVB) has the potential to significantly improve voice therapy effectiveness by targeting one of the most challenging aspects of rehabilitation: carryover of desired behaviors outside of the therapy session. Although initial evidence indicates that AVB can alter vocal behavior in daily life, retention of the new behavior after biofeedback has not been demonstrated. Motor learning studies repeatedly have shown retention-related benefits when reducing feedback frequency or providing summary statistics. Therefore, novel AVB settings that are based on these concepts are developed and implemented. The underlying theoretical framework and resultant implementation of innovative AVB settings on a smartphone-based voice monitor are described. A clinical case study demonstrates the functionality of the new relative frequency feedback capabilities. With new technical capabilities, 2 aspects of feedback are directly modifiable for AVB: relative frequency and summary feedback. Although reduced-frequency AVB was associated with improved carryover of a therapeutic vocal behavior (i.e., reduced vocal intensity) in a patient post-excision of vocal fold nodules, causation cannot be assumed. Timing and frequency of AVB schedules can be manipulated to empirically assess generalization of motor learning principles to vocal behavior modification and test the clinical effectiveness of AVB with various feedback schedules.
Mehta, Daryush D.; Petit, Robert J.; Sternad, Dagmar; Muise, Jason; Burns, James A.; Hillman, Robert E.
2017-01-01
Purpose Ambulatory voice biofeedback (AVB) has the potential to significantly improve voice therapy effectiveness by targeting one of the most challenging aspects of rehabilitation: carryover of desired behaviors outside of the therapy session. Although initial evidence indicates that AVB can alter vocal behavior in daily life, retention of the new behavior after biofeedback has not been demonstrated. Motor learning studies repeatedly have shown retention-related benefits when reducing feedback frequency or providing summary statistics. Therefore, novel AVB settings that are based on these concepts are developed and implemented. Method The underlying theoretical framework and resultant implementation of innovative AVB settings on a smartphone-based voice monitor are described. A clinical case study demonstrates the functionality of the new relative frequency feedback capabilities. Results With new technical capabilities, 2 aspects of feedback are directly modifiable for AVB: relative frequency and summary feedback. Although reduced-frequency AVB was associated with improved carryover of a therapeutic vocal behavior (i.e., reduced vocal intensity) in a patient post-excision of vocal fold nodules, causation cannot be assumed. Conclusions Timing and frequency of AVB schedules can be manipulated to empirically assess generalization of motor learning principles to vocal behavior modification and test the clinical effectiveness of AVB with various feedback schedules. PMID:28124070
Rodenas‐Cuadrado, Pedro M.; Mengede, Janine; Baas, Laura; Devanna, Paolo; Schmid, Tobias A.; Yartsev, Michael; Firzlaff, Uwe
2018-01-01
Abstract Genes including FOXP2, FOXP1, and CNTNAP2, have been implicated in human speech and language phenotypes, pointing to a role in the development of normal language‐related circuitry in the brain. Although speech and language are unique to humans a comparative approach is possible by addressing language‐relevant traits in animal systems. One such trait, vocal learning, represents an essential component of human spoken language, and is shared by cetaceans, pinnipeds, elephants, some birds and bats. Given their vocal learning abilities, gregarious nature, and reliance on vocalizations for social communication and navigation, bats represent an intriguing mammalian system in which to explore language‐relevant genes. We used immunohistochemistry to detail the distribution of FoxP2, FoxP1, and Cntnap2 proteins, accompanied by detailed cytoarchitectural histology in the brains of two vocal learning bat species; Phyllostomus discolor and Rousettus aegyptiacus. We show widespread expression of these genes, similar to what has been previously observed in other species, including humans. A striking difference was observed in the adult P. discolor bat, which showed low levels of FoxP2 expression in the cortex that contrasted with patterns found in rodents and nonhuman primates. We created an online, open‐access database within which all data can be browsed, searched, and high resolution images viewed to single cell resolution. The data presented herein reveal regions of interest in the bat brain and provide new opportunities to address the role of these language‐related genes in complex vocal‐motor and vocal learning behaviors in a mammalian model system. PMID:29297931
Goldberg, Jesse H.
2011-01-01
Young songbirds produce vocal “babbling,” and the variability of their songs is thought to underlie a process of trial-and-error vocal learning. It is known that this exploratory variability requires the “cortical” component of a basal ganglia (BG) thalamocortical loop, but less understood is the role of the BG and thalamic components in this behavior. We found that large bilateral lesions to the songbird BG homolog Area X had little or no effect on song variability during vocal babbling. In contrast, lesions to the BG-recipient thalamic nucleus DLM (medial portion of the dorsolateral thalamus) largely abolished normal vocal babbling in young birds and caused a dramatic increase in song stereotypy. These findings support the idea that the motor thalamus plays a key role in the expression of exploratory juvenile behaviors during learning. PMID:21430276
A neural circuit mechanism for regulating vocal variability during song learning in zebra finches.
Garst-Orozco, Jonathan; Babadi, Baktash; Ölveczky, Bence P
2014-12-15
Motor skill learning is characterized by improved performance and reduced motor variability. The neural mechanisms that couple skill level and variability, however, are not known. The zebra finch, a songbird, presents a unique opportunity to address this question because production of learned song and induction of vocal variability are instantiated in distinct circuits that converge on a motor cortex analogue controlling vocal output. To probe the interplay between learning and variability, we made intracellular recordings from neurons in this area, characterizing how their inputs from the functionally distinct pathways change throughout song development. We found that inputs that drive stereotyped song-patterns are strengthened and pruned, while inputs that induce variability remain unchanged. A simple network model showed that strengthening and pruning of action-specific connections reduces the sensitivity of motor control circuits to variable input and neural 'noise'. This identifies a simple and general mechanism for learning-related regulation of motor variability.
Limiting parental feedback disrupts vocal development in marmoset monkeys
Gultekin, Yasemin B.; Hage, Steffen R.
2017-01-01
Vocalizations of human infants undergo dramatic changes across the first year by becoming increasingly mature and speech-like. Human vocal development is partially dependent on learning by imitation through social feedback between infants and caregivers. Recent studies revealed similar developmental processes being influenced by parental feedback in marmoset monkeys for apparently innate vocalizations. Marmosets produce infant-specific vocalizations that disappear after the first postnatal months. However, it is yet unclear whether parental feedback is an obligate requirement for proper vocal development. Using quantitative measures to compare call parameters and vocal sequence structure we show that, in contrast to normally raised marmosets, marmosets that were separated from parents after the third postnatal month still produced infant-specific vocal behaviour at subadult stages. These findings suggest a significant role of social feedback on primate vocal development until the subadult stages and further show that marmoset monkeys are a compelling model system for early human vocal development. PMID:28090084
Tokarev, Kirill; Tiunova, Anna
2011-01-01
Background Specialized neural pathways, the song system, are required for acquiring, producing, and perceiving learned avian vocalizations. Birds that do not learn to produce their vocalizations lack telencephalic song system components. It is not known whether the song system forebrain regions are exclusively evolved for song or whether they also process information not related to song that might reflect their ‘evolutionary history’. Methodology/Principal Findings To address this question we monitored the induction of two immediate-early genes (IEGs) c-Fos and ZENK in various regions of the song system in zebra finches (Taeniopygia guttata) in response to an aversive food learning paradigm; this involves the association of a food item with a noxious stimulus that affects the oropharyngeal-esophageal cavity and tongue, causing subsequent avoidance of that food item. The motor response results in beak and head movements but not vocalizations. IEGs have been extensively used to map neuro-molecular correlates of song motor production and auditory processing. As previously reported, neurons in two pallial vocal motor regions, HVC and RA, expressed IEGs after singing. Surprisingly, c-Fos was induced equivalently also after food aversion learning in the absence of singing. The density of c-Fos positive neurons was significantly higher than that of birds in control conditions. This was not the case in two other pallial song nuclei important for vocal plasticity, LMAN and Area X, although singing did induce IEGs in these structures, as reported previously. Conclusions/Significance Our results are consistent with the possibility that some of the song nuclei may participate in non-vocal learning and the populations of neurons involved in the two tasks show partial overlap. These findings underscore the previously advanced notion that the specialized forebrain pre-motor nuclei controlling song evolved from circuits involved in behaviors related to feeding. PMID:21695176
Cook, Peter; Rouse, Andrew; Wilson, Margaret; Reichmuth, Colleen
2013-11-01
Is the ability to entrain motor activity to a rhythmic auditory stimulus, that is "keep a beat," dependent on neural adaptations supporting vocal mimicry? That is the premise of the vocal learning and synchronization hypothesis, recently advanced to explain the basis of this behavior (A. Patel, 2006, Musical Rhythm, Linguistic Rhythm, and Human Evolution, Music Perception, 24, 99-104). Prior to the current study, only vocal mimics, including humans, cockatoos, and budgerigars, have been shown to be capable of motoric entrainment. Here we demonstrate that a less vocally flexible animal, a California sea lion (Zalophus californianus), can learn to entrain head bobbing to an auditory rhythm meeting three criteria: a behavioral response that does not reproduce the stimulus; performance transfer to a range of novel tempos; and entrainment to complex, musical stimuli. These findings show that the capacity for entrainment of movement to rhythmic sounds does not depend on a capacity for vocal mimicry, and may be more widespread in the animal kingdom than previously hypothesized.
Social interaction shapes babbling: Testing parallels between birdsong and speech
NASA Astrophysics Data System (ADS)
Goldstein, Michael H.; King, Andrew P.; West, Meredith J.
2003-06-01
Birdsong is considered a model of human speech development at behavioral and neural levels. Few direct tests of the proposed analogs exist, however. Here we test a mechanism of phonological development in human infants that is based on social shaping, a selective learning process first documented in songbirds. By manipulating mothers' reactions to their 8-month-old infants' vocalizations, we demonstrate that phonological features of babbling are sensitive to nonimitative social stimulation. Contingent, but not noncontingent, maternal behavior facilitates more complex and mature vocal behavior. Changes in vocalizations persist after the manipulation. The data show that human infants use social feedback, facilitating immediate transitions in vocal behavior. Social interaction creates rapid shifts to developmentally more advanced sounds. These transitions mirror the normal development of speech, supporting the predictions of the avian social shaping model. These data provide strong support for a parallel in function between vocal precursors of songbirds and infants. Because imitation is usually considered the mechanism for vocal learning in both taxa, the findings introduce social shaping as a general process underlying the development of speech and song.
Segmental Production in Mandarin-Learning Infants
ERIC Educational Resources Information Center
Chen, Li-Mei; Kent, Raymond D.
2010-01-01
The early development of vocalic and consonantal production in Mandarin-learning infants was studied at the transition from babbling to producing first words. Spontaneous vocalizations were recorded for 24 infants grouped by age: G1 (0 ; 7 to 1 ; 0) and G2 (1 ; 1 to 1 ; 6). Additionally, the infant-directed speech of 24 caregivers was recorded…
Statistical learning in songbirds: from self-tutoring to song culture.
Fehér, Olga; Ljubičić, Iva; Suzuki, Kenta; Okanoya, Kazuo; Tchernichovski, Ofer
2017-01-05
At the onset of vocal development, both songbirds and humans produce variable vocal babbling with broadly distributed acoustic features. Over development, these vocalizations differentiate into the well-defined, categorical signals that characterize adult vocal behaviour. A broadly distributed signal is ideal for vocal exploration, that is, for matching vocal production to the statistics of the sensory input. The developmental transition to categorical signals is a gradual process during which the vocal output becomes differentiated and stable. But does it require categorical input? We trained juvenile zebra finches with playbacks of their own developing song, produced just a few moments earlier, updated continuously over development. Although the vocalizations of these self-tutored (ST) birds were initially broadly distributed, birds quickly developed categorical signals, as fast as birds that were trained with a categorical, adult song template. By contrast, siblings of those birds that received no training (isolates) developed phonological categories much more slowly and never reached the same level of category differentiation as their ST brothers. Therefore, instead of simply mirroring the statistical properties of their sensory input, songbirds actively transform it into distinct categories. We suggest that the early self-generation of phonological categories facilitates the establishment of vocal culture by making the song easier to transmit at the micro level, while promoting stability of shared vocabulary at the group level over generations.This article is part of the themed issue 'New frontiers for statistical learning in the cognitive sciences'. © 2016 The Authors.
Universal mechanisms of sound production and control in birds and mammals
Elemans, C.P.H; Rasmussen, J.H.; Herbst, C.T.; Düring, D.N.; Zollinger, S.A.; Brumm, H.; Srivastava, K.; Svane, N.; Ding, M.; Larsen, O.N.; Sober, S.J.; Švec, J.G.
2015-01-01
As animals vocalize, their vocal organ transforms motor commands into vocalizations for social communication. In birds, the physical mechanisms by which vocalizations are produced and controlled remain unresolved because of the extreme difficulty in obtaining in vivo measurements. Here, we introduce an ex vivo preparation of the avian vocal organ that allows simultaneous high-speed imaging, muscle stimulation and kinematic and acoustic analyses to reveal the mechanisms of vocal production in birds across a wide range of taxa. Remarkably, we show that all species tested employ the myoelastic-aerodynamic (MEAD) mechanism, the same mechanism used to produce human speech. Furthermore, we show substantial redundancy in the control of key vocal parameters ex vivo, suggesting that in vivo vocalizations may also not be specified by unique motor commands. We propose that such motor redundancy can aid vocal learning and is common to MEAD sound production across birds and mammals, including humans. PMID:26612008
Universal mechanisms of sound production and control in birds and mammals.
Elemans, C P H; Rasmussen, J H; Herbst, C T; Düring, D N; Zollinger, S A; Brumm, H; Srivastava, K; Svane, N; Ding, M; Larsen, O N; Sober, S J; Švec, J G
2015-11-27
As animals vocalize, their vocal organ transforms motor commands into vocalizations for social communication. In birds, the physical mechanisms by which vocalizations are produced and controlled remain unresolved because of the extreme difficulty in obtaining in vivo measurements. Here, we introduce an ex vivo preparation of the avian vocal organ that allows simultaneous high-speed imaging, muscle stimulation and kinematic and acoustic analyses to reveal the mechanisms of vocal production in birds across a wide range of taxa. Remarkably, we show that all species tested employ the myoelastic-aerodynamic (MEAD) mechanism, the same mechanism used to produce human speech. Furthermore, we show substantial redundancy in the control of key vocal parameters ex vivo, suggesting that in vivo vocalizations may also not be specified by unique motor commands. We propose that such motor redundancy can aid vocal learning and is common to MEAD sound production across birds and mammals, including humans.
Iconicity can ground the creation of vocal symbols.
Perlman, Marcus; Dale, Rick; Lupyan, Gary
2015-08-01
Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative 'vocal' charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic 'meaning template' we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems.
Corrigendum: Cultural Relativity in Perceiving Emotion From Vocalizations.
2014-12-01
Gendron, M., Roberson, D., van der Vyver, J. M., & Barrett, L. F. (2014). Cultural relativity in perceiving emotion from vocalizations. Psychological Science, 25, 911-920. (Original DOI: 10.1177/0956797613517239 ).
Animal vocal sequences: not the Markov chains we thought they were
Kershenbaum, Arik; Bowles, Ann E.; Freeberg, Todd M.; Jin, Dezhe Z.; Lameira, Adriano R.; Bohn, Kirsten
2014-01-01
Many animals produce vocal sequences that appear complex. Most researchers assume that these sequences are well characterized as Markov chains (i.e. that the probability of a particular vocal element can be calculated from the history of only a finite number of preceding elements). However, this assumption has never been explicitly tested. Furthermore, it is unclear how language could evolve in a single step from a Markovian origin, as is frequently assumed, as no intermediate forms have been found between animal communication and human language. Here, we assess whether animal taxa produce vocal sequences that are better described by Markov chains, or by non-Markovian dynamics such as the ‘renewal process’ (RP), characterized by a strong tendency to repeat elements. We examined vocal sequences of seven taxa: Bengalese finches Lonchura striata domestica, Carolina chickadees Poecile carolinensis, free-tailed bats Tadarida brasiliensis, rock hyraxes Procavia capensis, pilot whales Globicephala macrorhynchus, killer whales Orcinus orca and orangutans Pongo spp. The vocal systems of most of these species are more consistent with a non-Markovian RP than with the Markovian models traditionally assumed. Our data suggest that non-Markovian vocal sequences may be more common than Markov sequences, which must be taken into account when evaluating alternative hypotheses for the evolution of signalling complexity, and perhaps human language origins. PMID:25143037
Chimpanzee vocal signaling points to a multimodal origin of human language.
Taglialatela, Jared P; Russell, Jamie L; Schaeffer, Jennifer A; Hopkins, William D
2011-04-20
The evolutionary origin of human language and its neurobiological foundations has long been the object of intense scientific debate. Although a number of theories have been proposed, one particularly contentious model suggests that human language evolved from a manual gestural communication system in a common ape-human ancestor. Consistent with a gestural origins theory are data indicating that chimpanzees intentionally and referentially communicate via manual gestures, and the production of manual gestures, in conjunction with vocalizations, activates the chimpanzee Broca's area homologue--a region in the human brain that is critical for the planning and execution of language. However, it is not known if this activity observed in the chimpanzee Broca's area is the result of the chimpanzees producing manual communicative gestures, communicative sounds, or both. This information is critical for evaluating the theory that human language evolved from a strictly manual gestural system. To this end, we used positron emission tomography (PET) to examine the neural metabolic activity in the chimpanzee brain. We collected PET data in 4 subjects, all of whom produced manual communicative gestures. However, 2 of these subjects also produced so-called attention-getting vocalizations directed towards a human experimenter. Interestingly, only the two subjects that produced these attention-getting sounds showed greater mean metabolic activity in the Broca's area homologue as compared to a baseline scan. The two subjects that did not produce attention-getting sounds did not. These data contradict an exclusive "gestural origins" theory for they suggest that it is vocal signaling that selectively activates the Broca's area homologue in chimpanzees. In other words, the activity observed in the Broca's area homologue reflects the production of vocal signals by the chimpanzees, suggesting that this critical human language region was involved in vocal signaling in the common ancestor of both modern humans and chimpanzees.
Properties of Vocalization- and Gesture-Combinations in the Transition to First Words
ERIC Educational Resources Information Center
Murillo, Eva; Capilla, Almudena
2016-01-01
Gestures and vocal elements interact from the early stages of language development, but the role of this interaction in the language learning process is not yet completely understood. The aim of this study is to explore gestural accompaniment's influence on the acoustic properties of vocalizations in the transition to first words. Eleven Spanish…
Vocal Stereotypy in Individuals with Autism Spectrum Disorders: A Review of Behavioral Interventions
ERIC Educational Resources Information Center
Lanovaz, Marc J.; Sladeczek, Ingrid E.
2012-01-01
Vocal stereotypy is a common problem behavior in individuals with autism spectrum disorders that may interfere considerably with learning and social inclusion. To assist clinicians in treating the behavior and to guide researchers in identifying gaps in the research literature, the authors provide an overview of research on vocal stereotypy in…
Can Birds Perceive Rhythmic Patterns? A Review and Experiments on a Songbird and a Parrot Species
ten Cate, Carel; Spierings, Michelle; Hubert, Jeroen; Honing, Henkjan
2016-01-01
While humans can easily entrain their behavior with the beat in music, this ability is rare among animals. Yet, comparative studies in non-human species are needed if we want to understand how and why this ability evolved. Entrainment requires two abilities: (1) recognizing the regularity in the auditory stimulus and (2) the ability to adjust the own motor output to the perceived pattern. It has been suggested that beat perception and entrainment are linked to the ability for vocal learning. The presence of some bird species showing beat induction, and also the existence of vocal learning as well as vocal non-learning bird taxa, make them relevant models for comparative research on rhythm perception and its link to vocal learning. Also, some bird vocalizations show strong regularity in rhythmic structure, suggesting that birds might perceive rhythmic structures. In this paper we review the available experimental evidence for the perception of regularity and rhythms by birds, like the ability to distinguish regular from irregular stimuli over tempo transformations and report data from new experiments. While some species show a limited ability to detect regularity, most evidence suggests that birds attend primarily to absolute and not relative timing of patterns and to local features of stimuli. We conclude that, apart from some large parrot species, there is limited evidence for beat and regularity perception among birds and that the link to vocal learning is unclear. We next report the new experiments in which zebra finches and budgerigars (both vocal learners) were first trained to distinguish a regular from an irregular pattern of beats and then tested on various tempo transformations of these stimuli. The results showed that both species reduced the discrimination after tempo transformations. This suggests that, as was found in earlier studies, they attended mainly to local temporal features of the stimuli, and not to their overall regularity. However, some individuals of both species showed an additional sensitivity to the more global pattern if some local features were left unchanged. Altogether our study indicates both between and within species variation, in which birds attend to a mixture of local and to global rhythmic features. PMID:27242635
Imitation of novel conspecific and human speech sounds in the killer whale (Orcinus orca).
Abramson, José Z; Hernández-Lloreda, Mª Victoria; García, Lino; Colmenares, Fernando; Aboitiz, Francisco; Call, Josep
2018-01-31
Vocal imitation is a hallmark of human spoken language, which, along with other advanced cognitive skills, has fuelled the evolution of human culture. Comparative evidence has revealed that although the ability to copy sounds from conspecifics is mostly uniquely human among primates, a few distantly related taxa of birds and mammals have also independently evolved this capacity. Remarkably, field observations of killer whales have documented the existence of group-differentiated vocal dialects that are often referred to as traditions or cultures and are hypothesized to be acquired non-genetically. Here we use a do-as-I-do paradigm to study the abilities of a killer whale to imitate novel sounds uttered by conspecific (vocal imitative learning) and human models (vocal mimicry). We found that the subject made recognizable copies of all familiar and novel conspecific and human sounds tested and did so relatively quickly (most during the first 10 trials and three in the first attempt). Our results lend support to the hypothesis that the vocal variants observed in natural populations of this species can be socially learned by imitation. The capacity for vocal imitation shown in this study may scaffold the natural vocal traditions of killer whales in the wild. © 2018 The Author(s).
Vocal interaction between children with Down syndrome and their parents.
Thiemann-Bourque, Kathy S; Warren, Steven F; Brady, Nancy; Gilkerson, Jill; Richards, Jeffrey A
2014-08-01
The purpose of this study was to describe differences in parent input and child vocal behaviors of children with Down syndrome (DS) compared with typically developing (TD) children. The goals were to describe the language learning environments at distinctly different ages in early childhood. Nine children with DS and 9 age-matched TD children participated; 4 children in each group were ages 9-11 months, and 5 were between 25 and 54 months. Measures were derived from automated vocal analysis. A digital language processor measured the richness of the child's language environment, including number of adult words, conversational turns, and child vocalizations. Analyses indicated no significant differences in words spoken by parents of younger versus older children with DS and significantly more words spoken by parents of TD children than parents of children with DS. Differences between the DS and TD groups were observed in rates of all vocal behaviors, with no differences noted between the younger versus older children with DS, and the younger TD children did not vocalize significantly more than the younger DS children. Parents of children with DS continue to provide consistent levels of input across the early language learning years; however, child vocal behaviors remain low after the age of 24 months, suggesting the need for additional and alternative intervention approaches.
From imitation to meaning: circuit plasticity and the acquisition of a conventionalized semantics
García, Ricardo R.; Zamorano, Francisco; Aboitiz, Francisco
2014-01-01
The capacity for language is arguably the most remarkable innovation of the human brain. A relatively recent interpretation prescribes that part of the language-related circuits were co-opted from circuitry involved in hand control—the mirror neuron system (MNS), involved both in the perception and in the execution of voluntary grasping actions. A less radical view is that in early humans, communication was opportunistic and multimodal, using signs, vocalizations or whatever means available to transmit social information. However, one point that is not yet clear under either perspective is how learned communication acquired a semantic property thereby allowing us to name objects and eventually describe our surrounding environment. Here we suggest a scenario involving both manual gestures and learned vocalizations that led to the development of a primitive form of conventionalized reference. This proposal is based on comparative evidence gathered from other species and on neurolinguistic evidence in humans, which points to a crucial role for vocal learning in the early development of language. Firstly, the capacity to direct the attention of others to a common object may have been crucial for developing a consensual referential system. Pointing, which is a ritualized grasping gesture, may have been crucial to this end. Vocalizations also served to generate joint attention among conversants, especially when combined with gaze direction. Another contributing element was the development of pantomimic actions resembling events or animals. In conjunction with this mimicry, the development of plastic neural circuits that support complex, learned vocalizations was probably a significant factor in the evolution of conventionalized semantics in our species. Thus, vocal imitations of sounds, as in onomatopoeias (words whose sound resembles their meaning), are possibly supported by mirror system circuits, and may have been relevant in the acquisition of early meanings. PMID:25152726
Iconicity can ground the creation of vocal symbols
Perlman, Marcus; Dale, Rick; Lupyan, Gary
2015-01-01
Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative ‘vocal’ charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic ‘meaning template’ we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems. PMID:26361547
Piristine, Hande C; Choetso, Tenzin; Gobes, Sharon M H
2016-11-01
Sensory feedback is essential for acquiring and maintaining complex motor behaviors, including birdsong. In zebra finches, auditory feedback reaches the song control circuits primarily through the nucleus interfacialis nidopalii (Nif), which provides excitatory input to HVC (proper name)-a premotor region essential for the production of learned vocalizations. Despite being one of the major inputs to the song control pathway, the role of Nif in generating vocalizations is not well understood. To address this, we transiently inactivated Nif in late juvenile zebra finches. Upon Nif inactivation (in both hemispheres or on one side only), birds went from singing stereotyped zebra finch song to uttering highly variable and unstructured vocalizations resembling sub-song, an early juvenile song form driven by a basal ganglia circuit. Simultaneously inactivating Nif and LMAN (lateral magnocellular nucleus of the anterior nidopallium), the output nucleus of a basal ganglia circuit, inhibited song production altogether. These results suggest that Nif is required for generating the premotor drive for song. Permanent Nif lesions, in contrast, have only transient effects on vocal production, with song recovering within a day. The sensorimotor nucleus Nif thus produces a premotor drive to the motor pathway that is acutely required for generating learned vocalizations, but once permanently removed, the song system can compensate for its absence. © 2016 Wiley Periodicals, Inc. Develop Neurobiol 76: 1213-1225, 2016. © 2016 Wiley Periodicals, Inc.
Song evolution, speciation, and vocal learning in passerine birds.
Mason, Nicholas A; Burns, Kevin J; Tobias, Joseph A; Claramunt, Santiago; Seddon, Nathalie; Derryberry, Elizabeth P
2017-03-01
Phenotypic divergence can promote reproductive isolation and speciation, suggesting a possible link between rates of phenotypic evolution and the tempo of speciation at multiple evolutionary scales. To date, most macroevolutionary studies of diversification have focused on morphological traits, whereas behavioral traits─including vocal signals─are rarely considered. Thus, although behavioral traits often mediate mate choice and gene flow, we have a limited understanding of how behavioral evolution contributes to diversification. Furthermore, the developmental mode by which behavioral traits are acquired may affect rates of behavioral evolution, although this hypothesis is seldom tested in a phylogenetic framework. Here, we examine evidence for rate shifts in vocal evolution and speciation across two major radiations of codistributed passerines: one oscine clade with learned songs (Thraupidae) and one suboscine clade with innate songs (Furnariidae). We find that evolutionary bursts in rates of speciation and song evolution are coincident in both thraupids and furnariids. Further, overall rates of vocal evolution are higher among taxa with learned rather than innate songs. Taken together, these findings suggest an association between macroevolutionary bursts in speciation and vocal evolution, and that the tempo of behavioral evolution can be influenced by variation in developmental modes among lineages. © 2016 The Author(s). Evolution © 2016 The Society for the Study of Evolution.
Social learning of vocal structure in a nonhuman primate?
2011-01-01
Background Non-human primate communication is thought to be fundamentally different from human speech, mainly due to vast differences in vocal control. The lack of these abilities in non-human primates is especially striking if compared to some marine mammals and bird species, which has generated somewhat of an evolutionary conundrum. What are the biological roots and underlying evolutionary pressures of the human ability to voluntarily control sound production and learn the vocal utterances of others? One hypothesis is that this capacity has evolved gradually in humans from an ancestral stage that resembled the vocal behavior of modern primates. Support for this has come from studies that have documented limited vocal flexibility and convergence in different primate species, typically in calls used during social interactions. The mechanisms underlying these patterns, however, are currently unknown. Specifically, it has been difficult to rule out explanations based on genetic relatedness, suggesting that such vocal flexibility may not be the result of social learning. Results To address this point, we compared the degree of acoustic similarity of contact calls in free-ranging Campbell's monkeys as a function of their social bonds and genetic relatedness. We calculated three different indices to compare the similarities between the calls' frequency contours, the duration of grooming interactions and the microsatellite-based genetic relatedness between partners. We found a significantly positive relation between bond strength and acoustic similarity that was independent of genetic relatedness. Conclusion Genetic factors determine the general species-specific call repertoire of a primate species, while social factors can influence the fine structure of some the call types. The finding is in line with the more general hypothesis that human speech has evolved gradually from earlier primate-like vocal communication. PMID:22177339
Lewandowski, Brian; Vyssotski, Alexei; Hahnloser, Richard H R; Schmidt, Marc
2013-06-01
Communication between auditory and vocal motor nuclei is essential for vocal learning. In songbirds, the nucleus interfacialis of the nidopallium (NIf) is part of a sensorimotor loop, along with auditory nucleus avalanche (Av) and song system nucleus HVC, that links the auditory and song systems. Most of the auditory information comes through this sensorimotor loop, with the projection from NIf to HVC representing the largest single source of auditory information to the song system. In addition to providing the majority of HVC's auditory input, NIf is also the primary driver of spontaneous activity and premotor-like bursting during sleep in HVC. Like HVC and RA, two nuclei critical for song learning and production, NIf exhibits behavioral-state dependent auditory responses and strong motor bursts that precede song output. NIf also exhibits extended periods of fast gamma oscillations following vocal production. Based on the converging evidence from studies of physiology and functional connectivity it would be reasonable to expect NIf to play an important role in the learning, maintenance, and production of song. Surprisingly, however, lesions of NIf in adult zebra finches have no effect on song production or maintenance. Only the plastic song produced by juvenile zebra finches during the sensorimotor phase of song learning is affected by NIf lesions. In this review, we carefully examine what is known about NIf at the anatomical, physiological, and behavioral levels. We reexamine conclusions drawn from previous studies in the light of our current understanding of the song system, and establish what can be said with certainty about NIf's involvement in song learning, maintenance, and production. Finally, we review recent theories of song learning integrating possible roles for NIf within these frameworks and suggest possible parallels between NIf and sensorimotor areas that form part of the neural circuitry for speech processing in humans. Copyright © 2013 Elsevier Ltd. All rights reserved.
Lewandowski, Brian; Vyssotski, Alexei; Hahnloser, Richard H.R.; Schmidt, Marc
2015-01-01
Communication between auditory and vocal motor nuclei is essential for vocal learning. In songbirds, the nucleus interfacialis of the nidopallium (NIf) is part of a sensorimotor loop, along with auditory nucleus avalanche (Av) and song system nucleus HVC, that links the auditory and song systems. Most of the auditory information comes through this sensorimotor loop, with the projection from NIf to HVC representing the largest single source of auditory information to the song system. In addition to providing the majority of HVC’s auditory input, NIf is also the primary driver of spontaneous activity and premotor-like bursting during sleep in HVC. Like HVC and RA, two nuclei critical for song learning and production, NIf exhibits behavioral-state dependent auditory responses and strong motor bursts that precede song output. NIf also exhibits extended periods of fast gamma oscillations following vocal production. Based on the converging evidence from studies of physiology and functional connectivity it would be reasonable to expect NIf to play an important role in the learning, maintenance, and production of song. Surprisingly, however, lesions of NIf in adult zebra finches have no effect on song production or maintenance. Only the plastic song produced by juvenile zebra finches during the sensorimotor phase of song learning is affected by NIf lesions. In this review, we carefully examine what is known about NIf at the anatomical, physiological, and behavioral levels. We reexamine conclusions drawn from previous studies in the light of our current understanding of the song system, and establish what can be said with certainty about NIf’s involvement in song learning, maintenance, and production. Finally, we review recent theories of song learning integrating possible roles for NIf within these frameworks and suggest possible parallels between NIf and sensorimotor areas that form part of the neural circuitry for speech processing in humans. PMID:23603062
Functional flexibility of infant vocalization and the emergence of language
Oller, D. Kimbrough; Buder, Eugene H.; Ramsdell, Heather L.; Warlaumont, Anne S.; Chorna, Lesya; Bakeman, Roger
2013-01-01
We report on the emergence of functional flexibility in vocalizations of human infants. This vastly underappreciated capability becomes apparent when prelinguistic vocalizations express a full range of emotional content—positive, neutral, and negative. The data show that at least three types of infant vocalizations (squeals, vowel-like sounds, and growls) occur with this full range of expression by 3–4 mo of age. In contrast, infant cry and laughter, which are species-specific signals apparently homologous to vocal calls in other primates, show functional stability, with cry overwhelmingly expressing negative and laughter positive emotional states. Functional flexibility is a sine qua non in spoken language, because all words or sentences can be produced as expressions of varying emotional states and because learning conventional “meanings” requires the ability to produce sounds that are free of any predetermined function. Functional flexibility is a defining characteristic of language, and empirically it appears before syntax, word learning, and even earlier-developing features presumed to be critical to language (e.g., joint attention, syllable imitation, and canonical babbling). The appearance of functional flexibility early in the first year of human life is a critical step in the development of vocal language and may have been a critical step in the evolution of human language, preceding protosyntax and even primitive single words. Such flexible affect expression of vocalizations has not yet been reported for any nonhuman primate but if found to occur would suggest deep roots for functional flexibility of vocalization in our primate heritage. PMID:23550164
ERIC Educational Resources Information Center
Obenland, Carrie A.; Munson, Ashlyn H.; Hutchinson, John S.
2013-01-01
Active learning is becoming more prevalent in large science classrooms, and this study shows the impact on performance of being vocal during Socratic questioning in a General Chemistry course. 800 college students over a two year period were given a pre and post-test using the Chemistry Concept Reasoning Test. The pre-test results showed that…
Mechanisms and time course of vocal learning and consolidation in the adult songbird.
Warren, Timothy L; Tumer, Evren C; Charlesworth, Jonathan D; Brainard, Michael S
2011-10-01
In songbirds, the basal ganglia outflow nucleus LMAN is a cortical analog that is required for several forms of song plasticity and learning. Moreover, in adults, inactivating LMAN can reverse the initial expression of learning driven via aversive reinforcement. In the present study, we investigated how LMAN contributes to both reinforcement-driven learning and a self-driven recovery process in adult Bengalese finches. We first drove changes in the fundamental frequency of targeted song syllables and compared the effects of inactivating LMAN with the effects of interfering with N-methyl-d-aspartate (NMDA) receptor-dependent transmission from LMAN to one of its principal targets, the song premotor nucleus RA. Inactivating LMAN and blocking NMDA receptors in RA caused indistinguishable reversions in the expression of learning, indicating that LMAN contributes to learning through NMDA receptor-mediated glutamatergic transmission to RA. We next assessed how LMAN's role evolves over time by maintaining learned changes to song while periodically inactivating LMAN. The expression of learning consolidated to become LMAN independent over multiple days, indicating that this form of consolidation is not completed over one night, as previously suggested, and instead may occur gradually during singing. Subsequent cessation of reinforcement was followed by a gradual self-driven recovery of original song structure, indicating that consolidation does not correspond with the lasting retention of changes to song. Finally, for self-driven recovery, as for reinforcement-driven learning, LMAN was required for the expression of initial, but not later, changes to song. Our results indicate that NMDA receptor-dependent transmission from LMAN to RA plays an essential role in the initial expression of two distinct forms of vocal learning and that this role gradually wanes over a multiday process of consolidation. The results support an emerging view that cortical-basal ganglia circuits can direct the initial expression of learning via top-down influences on primary motor circuitry.
Mechanisms and time course of vocal learning and consolidation in the adult songbird
Tumer, Evren C.; Charlesworth, Jonathan D.; Brainard, Michael S.
2011-01-01
In songbirds, the basal ganglia outflow nucleus LMAN is a cortical analog that is required for several forms of song plasticity and learning. Moreover, in adults, inactivating LMAN can reverse the initial expression of learning driven via aversive reinforcement. In the present study, we investigated how LMAN contributes to both reinforcement-driven learning and a self-driven recovery process in adult Bengalese finches. We first drove changes in the fundamental frequency of targeted song syllables and compared the effects of inactivating LMAN with the effects of interfering with N-methyl-d-aspartate (NMDA) receptor-dependent transmission from LMAN to one of its principal targets, the song premotor nucleus RA. Inactivating LMAN and blocking NMDA receptors in RA caused indistinguishable reversions in the expression of learning, indicating that LMAN contributes to learning through NMDA receptor-mediated glutamatergic transmission to RA. We next assessed how LMAN's role evolves over time by maintaining learned changes to song while periodically inactivating LMAN. The expression of learning consolidated to become LMAN independent over multiple days, indicating that this form of consolidation is not completed over one night, as previously suggested, and instead may occur gradually during singing. Subsequent cessation of reinforcement was followed by a gradual self-driven recovery of original song structure, indicating that consolidation does not correspond with the lasting retention of changes to song. Finally, for self-driven recovery, as for reinforcement-driven learning, LMAN was required for the expression of initial, but not later, changes to song. Our results indicate that NMDA receptor-dependent transmission from LMAN to RA plays an essential role in the initial expression of two distinct forms of vocal learning and that this role gradually wanes over a multiday process of consolidation. The results support an emerging view that cortical-basal ganglia circuits can direct the initial expression of learning via top-down influences on primary motor circuitry. PMID:21734110
Animal vocal sequences: not the Markov chains we thought they were.
Kershenbaum, Arik; Bowles, Ann E; Freeberg, Todd M; Jin, Dezhe Z; Lameira, Adriano R; Bohn, Kirsten
2014-10-07
Many animals produce vocal sequences that appear complex. Most researchers assume that these sequences are well characterized as Markov chains (i.e. that the probability of a particular vocal element can be calculated from the history of only a finite number of preceding elements). However, this assumption has never been explicitly tested. Furthermore, it is unclear how language could evolve in a single step from a Markovian origin, as is frequently assumed, as no intermediate forms have been found between animal communication and human language. Here, we assess whether animal taxa produce vocal sequences that are better described by Markov chains, or by non-Markovian dynamics such as the 'renewal process' (RP), characterized by a strong tendency to repeat elements. We examined vocal sequences of seven taxa: Bengalese finches Lonchura striata domestica, Carolina chickadees Poecile carolinensis, free-tailed bats Tadarida brasiliensis, rock hyraxes Procavia capensis, pilot whales Globicephala macrorhynchus, killer whales Orcinus orca and orangutans Pongo spp. The vocal systems of most of these species are more consistent with a non-Markovian RP than with the Markovian models traditionally assumed. Our data suggest that non-Markovian vocal sequences may be more common than Markov sequences, which must be taken into account when evaluating alternative hypotheses for the evolution of signalling complexity, and perhaps human language origins. © 2014 The Author(s) Published by the Royal Society. All rights reserved.
LaZerte, Stefanie E.; Slabbekoorn, Hans; Otter, Ken A.
2016-01-01
Urban noise can interfere with avian communication through masking, but birds can reduce this interference by altering their vocalizations. Although several experimental studies indicate that birds can rapidly change their vocalizations in response to sudden increases in ambient noise, none have investigated whether this is a learned response that depends on previous exposure. Black-capped chickadees (Poecile atricapillus) change the frequency of their songs in response to both fluctuating traffic noise and experimental noise. We investigated whether these responses to fluctuating noise depend on familiarity with noise. We confirmed that males in noisy areas sang higher-frequency songs than those in quiet areas, but found that only males in already-noisy territories shifted songs upwards in immediate response to experimental noise. Unexpectedly, males in more quiet territories shifted songs downwards in response to experimental noise. These results suggest that chickadees may require prior experience with fluctuating noise to adjust vocalizations in such a way as to minimize masking. Thus, learning to cope may be an important part of adjusting to acoustic life in the city. PMID:27358372
LaZerte, Stefanie E; Slabbekoorn, Hans; Otter, Ken A
2016-06-29
Urban noise can interfere with avian communication through masking, but birds can reduce this interference by altering their vocalizations. Although several experimental studies indicate that birds can rapidly change their vocalizations in response to sudden increases in ambient noise, none have investigated whether this is a learned response that depends on previous exposure. Black-capped chickadees (Poecile atricapillus) change the frequency of their songs in response to both fluctuating traffic noise and experimental noise. We investigated whether these responses to fluctuating noise depend on familiarity with noise. We confirmed that males in noisy areas sang higher-frequency songs than those in quiet areas, but found that only males in already-noisy territories shifted songs upwards in immediate response to experimental noise. Unexpectedly, males in more quiet territories shifted songs downwards in response to experimental noise. These results suggest that chickadees may require prior experience with fluctuating noise to adjust vocalizations in such a way as to minimize masking. Thus, learning to cope may be an important part of adjusting to acoustic life in the city. © 2016 The Author(s).
Striatal FoxP2 Is Actively Regulated during Songbird Sensorimotor Learning
Teramitsu, Ikuko; Poopatanapong, Amy; Torrisi, Salvatore; White, Stephanie A.
2010-01-01
Background Mutations in the FOXP2 transcription factor lead to language disorders with developmental onset. Accompanying structural abnormalities in cortico-striatal circuitry indicate that at least a portion of the behavioral phenotype is due to organizational deficits. We previously found parallel FoxP2 expression patterns in human and songbird cortico/pallio-striatal circuits important for learned vocalizations, suggesting that FoxP2's function in birdsong may generalize to speech. Methodology/Principal Findings We used zebra finches to address the question of whether FoxP2 is additionally important in the post-organizational function of these circuits. In both humans and songbirds, vocal learning depends on auditory guidance to achieve and maintain optimal vocal output. We tested whether deafening prior to or during the sensorimotor phase of song learning disrupted FoxP2 expression in song circuitry. As expected, the songs of deafened juveniles were abnormal, however basal FoxP2 levels were unaffected. In contrast, when hearing or deaf juveniles sang for two hours in the morning, FoxP2 was acutely down-regulated in the striatal song nucleus, area X. The extent of down-regulation was similar between hearing and deaf birds. Interestingly, levels of FoxP2 and singing were correlated only in hearing birds. Conclusions/Significance Hearing appears to link FoxP2 levels to the amount of vocal practice. As juvenile birds spent more time practicing than did adults, their FoxP2 levels are likely to be low more often. Behaviorally-driven reductions in the mRNA encoding this transcription factor could ultimately affect downstream molecules that function in vocal exploration, especially during sensorimotor learning. PMID:20062527
Vocal Interaction between Children with Down syndrome and their Parents
Thiemann-Bourque, Kathy S.; Warren, Steven F.; Brady, Nancy; Gilkerson, Jill; Richards, Jeffrey A.
2014-01-01
Purpose The purpose of this study was to describe differences in parent input and child vocal behaviors of children with Down syndrome (DS) compared to typically developing (TD) children. The goals were to describe the language learning environments at distinctly different ages in early childhood. Method Nine children with DS and 9 age-matched TD children participated; four children in each group were ages 9–11 months and five were between 25–54 months. Measures were derived from automated vocal analysis. A digital language processer measured the richness of the child’s language environment, including number of adult words, conversational turns, and child vocalizations. Results Analyses indicated no significant differences in words spoken by parents of younger vs. older children with DS, and significantly more words spoken by parents of TD children than parents of children with DS. Differences between the DS and TD groups were observed in rates of all vocal behaviors; with no differences noted between the younger vs. older children with DS, and the younger TD children did not vocalize significantly more than the younger DS children. Conclusions Parents of children with DS continue to provide consistent levels of input across the early language learning years; however, child vocal behaviors remain low after the age of 24 months suggesting the need for additional and alternative intervention approaches. PMID:24686777
Rhythmic synchronization tapping to an audio–visual metronome in budgerigars
Hasegawa, Ai; Okanoya, Kazuo; Hasegawa, Toshikazu; Seki, Yoshimasa
2011-01-01
In all ages and countries, music and dance have constituted a central part in human culture and communication. Recently, vocal-learning animals such as parrots and elephants have been found to share rhythmic ability with humans. Thus, we investigated the rhythmic synchronization of budgerigars, a vocal-mimicking parrot species, under controlled conditions and a systematically designed experimental paradigm as a first step in understanding the evolution of musical entrainment. We trained eight budgerigars to perform isochronous tapping tasks in which they pecked a key to the rhythm of audio–visual metronome-like stimuli. The budgerigars showed evidence of entrainment to external stimuli over a wide range of tempos. They seemed to be inherently inclined to tap at fast tempos, which have a similar time scale to the rhythm of budgerigars' natural vocalizations. We suggest that vocal learning might have contributed to their performance, which resembled that of humans. PMID:22355637
Rhythmic synchronization tapping to an audio-visual metronome in budgerigars.
Hasegawa, Ai; Okanoya, Kazuo; Hasegawa, Toshikazu; Seki, Yoshimasa
2011-01-01
In all ages and countries, music and dance have constituted a central part in human culture and communication. Recently, vocal-learning animals such as parrots and elephants have been found to share rhythmic ability with humans. Thus, we investigated the rhythmic synchronization of budgerigars, a vocal-mimicking parrot species, under controlled conditions and a systematically designed experimental paradigm as a first step in understanding the evolution of musical entrainment. We trained eight budgerigars to perform isochronous tapping tasks in which they pecked a key to the rhythm of audio-visual metronome-like stimuli. The budgerigars showed evidence of entrainment to external stimuli over a wide range of tempos. They seemed to be inherently inclined to tap at fast tempos, which have a similar time scale to the rhythm of budgerigars' natural vocalizations. We suggest that vocal learning might have contributed to their performance, which resembled that of humans.
Audio-vocal interaction in single neurons of the monkey ventrolateral prefrontal cortex.
Hage, Steffen R; Nieder, Andreas
2015-05-06
Complex audio-vocal integration systems depend on a strong interconnection between the auditory and the vocal motor system. To gain cognitive control over audio-vocal interaction during vocal motor control, the PFC needs to be involved. Neurons in the ventrolateral PFC (VLPFC) have been shown to separately encode the sensory perceptions and motor production of vocalizations. It is unknown, however, whether single neurons in the PFC reflect audio-vocal interactions. We therefore recorded single-unit activity in the VLPFC of rhesus monkeys (Macaca mulatta) while they produced vocalizations on command or passively listened to monkey calls. We found that 12% of randomly selected neurons in VLPFC modulated their discharge rate in response to acoustic stimulation with species-specific calls. Almost three-fourths of these auditory neurons showed an additional modulation of their discharge rates either before and/or during the monkeys' motor production of vocalization. Based on these audio-vocal interactions, the VLPFC might be well positioned to combine higher order auditory processing with cognitive control of the vocal motor output. Such audio-vocal integration processes in the VLPFC might constitute a precursor for the evolution of complex learned audio-vocal integration systems, ultimately giving rise to human speech. Copyright © 2015 the authors 0270-6474/15/357030-11$15.00/0.
The voice conveys specific emotions: evidence from vocal burst displays.
Simon-Thomas, Emiliana R; Keltner, Dacher J; Sauter, Disa; Sinicropi-Yao, Lara; Abramson, Anna
2009-12-01
Studies of emotion signaling inform claims about the taxonomic structure, evolutionary origins, and physiological correlates of emotions. Emotion vocalization research has tended to focus on a limited set of emotions: anger, disgust, fear, sadness, surprise, happiness, and for the voice, also tenderness. Here, we examine how well brief vocal bursts can communicate 22 different emotions: 9 negative (Study 1) and 13 positive (Study 2), and whether prototypical vocal bursts convey emotions more reliably than heterogeneous vocal bursts (Study 3). Results show that vocal bursts communicate emotions like anger, fear, and sadness, as well as seldom-studied states like awe, compassion, interest, and embarrassment. Ancillary analyses reveal family-wise patterns of vocal burst expression. Errors in classification were more common within emotion families (e.g., 'self-conscious,' 'pro-social') than between emotion families. The three studies reported highlight the voice as a rich modality for emotion display that can inform fundamental constructs about emotion.
A three-parameter model for classifying anurans into four genera based on advertisement calls.
Gingras, Bruno; Fitch, William Tecumseh
2013-01-01
The vocalizations of anurans are innate in structure and may therefore contain indicators of phylogenetic history. Thus, advertisement calls of species which are more closely related phylogenetically are predicted to be more similar than those of distant species. This hypothesis was evaluated by comparing several widely used machine-learning algorithms. Recordings of advertisement calls from 142 species belonging to four genera were analyzed. A logistic regression model, using mean values for dominant frequency, coefficient of variation of root-mean square energy, and spectral flux, correctly classified advertisement calls with regard to genus with an accuracy above 70%. Similar accuracy rates were obtained using these parameters with a support vector machine model, a K-nearest neighbor algorithm, and a multivariate Gaussian distribution classifier, whereas a Gaussian mixture model performed slightly worse. In contrast, models based on mel-frequency cepstral coefficients did not fare as well. Comparable accuracy levels were obtained on out-of-sample recordings from 52 of the 142 original species. The results suggest that a combination of low-level acoustic attributes is sufficient to discriminate efficiently between the vocalizations of these four genera, thus supporting the initial premise and validating the use of high-throughput algorithms on animal vocalizations to evaluate phylogenetic hypotheses.
Learning to breathe and sing: development of respiratory-vocal coordination in young songbirds
Veit, Lena; Aronov, Dmitriy
2011-01-01
How do animals with learned vocalizations coordinate vocal production with respiration? Songbirds such as the zebra finch learn their songs, beginning with highly variable babbling vocalizations known as subsong. After several weeks of practice, zebra finches are able to produce a precisely timed pattern of syllables and silences, precisely coordinated with expiratory and inspiratory pulses (Franz M, Goller F. J Neurobiol 51: 129–141, 2002). While respiration in adult song is well described, relatively little is known about respiratory patterns in subsong or about the processes by which respiratory and vocal patterns become coordinated. To address these questions, we recorded thoracic air sac pressure in juvenile zebra finches prior to the appearance of any consistent temporal or acoustic structure in their songs. We found that subsong contains brief inspiratory pulses (50 ms) alternating with longer pulses of sustained expiratory pressure (50–500 ms). In striking contrast to adult song, expiratory pulses often contained multiple (0–8) variably timed syllables separated by expiratory gaps and were only partially vocalized. During development, expiratory pulses became shorter and more stereotyped in duration with shorter and fewer nonvocalized parts. These developmental changes eventually resulted in the production of a single syllable per expiratory pulse and a single inspiratory pulse filling each gap, forming a coordinated sequence similar to that of adult song. To examine the role of forebrain song-control nuclei in the development of respiratory patterns, we performed pressure recordings before and after lesions of nucleus HVC (proper name) and found that this manipulation reverses the developmental trends in measures of the respiratory pattern. PMID:21697438
Learning to breathe and sing: development of respiratory-vocal coordination in young songbirds.
Veit, Lena; Aronov, Dmitriy; Fee, Michale S
2011-10-01
How do animals with learned vocalizations coordinate vocal production with respiration? Songbirds such as the zebra finch learn their songs, beginning with highly variable babbling vocalizations known as subsong. After several weeks of practice, zebra finches are able to produce a precisely timed pattern of syllables and silences, precisely coordinated with expiratory and inspiratory pulses (Franz M, Goller F. J Neurobiol 51: 129-141, 2002). While respiration in adult song is well described, relatively little is known about respiratory patterns in subsong or about the processes by which respiratory and vocal patterns become coordinated. To address these questions, we recorded thoracic air sac pressure in juvenile zebra finches prior to the appearance of any consistent temporal or acoustic structure in their songs. We found that subsong contains brief inspiratory pulses (50 ms) alternating with longer pulses of sustained expiratory pressure (50-500 ms). In striking contrast to adult song, expiratory pulses often contained multiple (0-8) variably timed syllables separated by expiratory gaps and were only partially vocalized. During development, expiratory pulses became shorter and more stereotyped in duration with shorter and fewer nonvocalized parts. These developmental changes eventually resulted in the production of a single syllable per expiratory pulse and a single inspiratory pulse filling each gap, forming a coordinated sequence similar to that of adult song. To examine the role of forebrain song-control nuclei in the development of respiratory patterns, we performed pressure recordings before and after lesions of nucleus HVC (proper name) and found that this manipulation reverses the developmental trends in measures of the respiratory pattern.
Chirathivat, Napim; Raja, Sahitya C; Gobes, Sharon M H
2015-06-22
Many aspects of song learning in songbirds resemble characteristics of speech acquisition in humans. Genetic, anatomical and behavioural parallels have most recently been extended with demonstrated similarities in hemispheric dominance between humans and songbirds: the avian higher order auditory cortex is left-lateralized for processing song memories in juvenile zebra finches that already have formed a memory of their fathers' song, just like Wernicke's area in the left hemisphere of the human brain is dominant for speech perception. However, it is unclear if hemispheric specialization is due to pre-existing functional asymmetry or the result of learning itself. Here we show that in juvenile male and female zebra finches that had never heard an adult song before, neuronal activation after initial exposure to a conspecific song is bilateral. Thus, like in humans, hemispheric dominance develops with vocal proficiency. A left-lateralized functional system that develops through auditory-vocal learning may be an evolutionary adaptation that could increase the efficiency of transferring information within one hemisphere, benefiting the production and perception of learned communication signals.
Chirathivat, Napim; Raja, Sahitya C.; Gobes, Sharon M. H.
2015-01-01
Many aspects of song learning in songbirds resemble characteristics of speech acquisition in humans. Genetic, anatomical and behavioural parallels have most recently been extended with demonstrated similarities in hemispheric dominance between humans and songbirds: the avian higher order auditory cortex is left-lateralized for processing song memories in juvenile zebra finches that already have formed a memory of their fathers’ song, just like Wernicke’s area in the left hemisphere of the human brain is dominant for speech perception. However, it is unclear if hemispheric specialization is due to pre-existing functional asymmetry or the result of learning itself. Here we show that in juvenile male and female zebra finches that had never heard an adult song before, neuronal activation after initial exposure to a conspecific song is bilateral. Thus, like in humans, hemispheric dominance develops with vocal proficiency. A left-lateralized functional system that develops through auditory-vocal learning may be an evolutionary adaptation that could increase the efficiency of transferring information within one hemisphere, benefiting the production and perception of learned communication signals. PMID:26098840
Acoustic signatures of sound source-tract coupling.
Arneodo, Ezequiel M; Perl, Yonatan Sanz; Mindlin, Gabriel B
2011-04-01
Birdsong is a complex behavior, which results from the interaction between a nervous system and a biomechanical peripheral device. While much has been learned about how complex sounds are generated in the vocal organ, little has been learned about the signature on the vocalizations of the nonlinear effects introduced by the acoustic interactions between a sound source and the vocal tract. The variety of morphologies among bird species makes birdsong a most suitable model to study phenomena associated to the production of complex vocalizations. Inspired by the sound production mechanisms of songbirds, in this work we study a mathematical model of a vocal organ, in which a simple sound source interacts with a tract, leading to a delay differential equation. We explore the system numerically, and by taking it to the weakly nonlinear limit, we are able to examine its periodic solutions analytically. By these means we are able to explore the dynamics of oscillatory solutions of a sound source-tract coupled system, which are qualitatively different from those of a sound source-filter model of a vocal organ. Nonlinear features of the solutions are proposed as the underlying mechanisms of observed phenomena in birdsong, such as unilaterally produced "frequency jumps," enhancement of resonances, and the shift of the fundamental frequency observed in heliox experiments. ©2011 American Physical Society
Acoustic signatures of sound source-tract coupling
Arneodo, Ezequiel M.; Perl, Yonatan Sanz; Mindlin, Gabriel B.
2014-01-01
Birdsong is a complex behavior, which results from the interaction between a nervous system and a biomechanical peripheral device. While much has been learned about how complex sounds are generated in the vocal organ, little has been learned about the signature on the vocalizations of the nonlinear effects introduced by the acoustic interactions between a sound source and the vocal tract. The variety of morphologies among bird species makes birdsong a most suitable model to study phenomena associated to the production of complex vocalizations. Inspired by the sound production mechanisms of songbirds, in this work we study a mathematical model of a vocal organ, in which a simple sound source interacts with a tract, leading to a delay differential equation. We explore the system numerically, and by taking it to the weakly nonlinear limit, we are able to examine its periodic solutions analytically. By these means we are able to explore the dynamics of oscillatory solutions of a sound source-tract coupled system, which are qualitatively different from those of a sound source-filter model of a vocal organ. Nonlinear features of the solutions are proposed as the underlying mechanisms of observed phenomena in birdsong, such as unilaterally produced “frequency jumps,” enhancement of resonances, and the shift of the fundamental frequency observed in heliox experiments. PMID:21599213
Fischer, J; Hammerschmidt, K
2011-01-01
Comparative analyses used to reconstruct the evolution of traits associated with the human language faculty, including its socio-cognitive underpinnings, highlight the importance of evolutionary constraints limiting vocal learning in non-human primates. After a brief overview of this field of research and the neural basis of primate vocalizations, we review studies that have addressed the genetic basis of usage and structure of ultrasonic communication in mice, with a focus on the gene FOXP2 involved in specific language impairments and neuroligin genes (NL-3 and NL-4) involved in autism spectrum disorders. Knockout of FoxP2 leads to reduced vocal behavior and eventually premature death. Introducing the human variant of FoxP2 protein into mice, in contrast, results in shifts in frequency and modulation of pup ultrasonic vocalizations. Knockout of NL-3 and NL-4 in mice diminishes social behavior and vocalizations. Although such studies may provide insights into the molecular and neural basis of social and communicative behavior, the structure of mouse vocalizations is largely innate, limiting the suitability of the mouse model to study human speech, a learned mode of production. Although knockout or replacement of single genes has perceptible effects on behavior, these genes are part of larger networks whose functions remain poorly understood. In humans, for instance, deficiencies in NL-4 can lead to a broad spectrum of disorders, suggesting that further factors (experiential and/or genetic) contribute to the variation in clinical symptoms. The precise nature as well as the interaction of these factors is yet to be determined. PMID:20579107
ERIC Educational Resources Information Center
Alt, Mary; Spaulding, Tammie
2011-01-01
Purpose: The purpose of this study was to measure the effect of time to response in a fast-mapping word learning task for children with specific language impairment (SLI) and children with typically developing language skills (TD). Manipulating time to response allows us to examine decay of the memory trace, the use of vocal rehearsal, and their…
Growth and splitting of neural sequences in songbird vocal development
Okubo, Tatsuo S.; Mackevicius, Emily L.; Payne, Hannah L.; Lynch, Galen F.; Fee, Michale S.
2015-01-01
Neural sequences are a fundamental feature of brain dynamics underlying diverse behaviors, but the mechanisms by which they develop during learning remain unknown. Songbirds learn vocalizations composed of syllables; in adult birds, each syllable is produced by a different sequence of action potential bursts in the premotor cortical area HVC. Here we carried out recordings of large populations of HVC neurons in singing juvenile birds throughout learning to examine the emergence of neural sequences. Early in vocal development, HVC neurons begin producing rhythmic bursts, temporally locked to a ‘prototype’ syllable. Different neurons are active at different latencies relative to syllable onset to form a continuous sequence. Through development, as new syllables emerge from the prototype syllable, initially highly overlapping burst sequences become increasingly distinct. We propose a mechanistic model in which multiple neural sequences can emerge from the growth and splitting of a common precursor sequence. PMID:26618871
Core and Shell Song Systems Unique to the Parrot Brain
Chakraborty, Mukta; Walløe, Solveig; Nedergaard, Signe; Fridel, Emma E.; Dabelsteen, Torben; Pakkenberg, Bente; Bertelsen, Mads F.; Dorrestein, Gerry M.; Brauth, Steven E.; Durand, Sarah E.; Jarvis, Erich D.
2015-01-01
The ability to imitate complex sounds is rare, and among birds has been found only in parrots, songbirds, and hummingbirds. Parrots exhibit the most advanced vocal mimicry among non-human animals. A few studies have noted differences in connectivity, brain position and shape in the vocal learning systems of parrots relative to songbirds and hummingbirds. However, only one parrot species, the budgerigar, has been examined and no differences in the presence of song system structures were found with other avian vocal learners. Motivated by questions of whether there are important differences in the vocal systems of parrots relative to other vocal learners, we used specialized constitutive gene expression, singing-driven gene expression, and neural connectivity tracing experiments to further characterize the song system of budgerigars and/or other parrots. We found that the parrot brain uniquely contains a song system within a song system. The parrot “core” song system is similar to the song systems of songbirds and hummingbirds, whereas the “shell” song system is unique to parrots. The core with only rudimentary shell regions were found in the New Zealand kea, representing one of the only living species at a basal divergence with all other parrots, implying that parrots evolved vocal learning systems at least 29 million years ago. Relative size differences in the core and shell regions occur among species, which we suggest could be related to species differences in vocal and cognitive abilities. PMID:26107173
Developmental Modulation of the Temporal Relationship Between Brain and Behavior
Crandall, Shane R.; Aoki, Naoya; Nick, Teresa A.
2008-01-01
Humans and songbirds shape learned vocalizations during a sensorimotor sensitive period or “babbling” phase. The brain mechanisms that underlie the shaping of vocalizations by sensory feedback are not known. We examined song behavior and brain activity in zebra finches during singing as they actively shaped their song toward a tutor model. We now show that the temporal relationship of behavior and activity in the premotor area HVC changes with the development of song behavior. During sensorimotor learning, HVC bursting activity both preceded and followed learned vocalizations by hundreds of milliseconds. Correspondingly, the duration of bursts that occurred during ongoing song motif behavior was prolonged in juveniles, as compared with adults, and was inversely correlated with song maturation. Multielectrode single-unit recording in juveniles revealed that single fast-spiking neurons were active both before and after vocalization. These same neurons responded to auditory stimuli. Collectively, these data indicate that a key aspect of sensory critical periods—prolonged bursting—also applies to sensorimotor development. In addition, prolonged motor discharge and sensory input coincide in single neurons of the developing song system, providing the necessary cellular elements for sensorimotor shaping through activity-dependent mechanisms. PMID:17079340
Esch, Barbara E; Carr, James E; Michael, Jack
2005-01-01
Many children with autism do not imitate adult vocalizations, an important skill in learning to talk. Pairing adult vocalizations with preferred stimuli has been shown to increase free-operant vocalizations but effects are temporary; thus, direct reinforcement may be necessary to establish durable vocal behaviors. In Experiment 1, directly reinforced echoic responses did not increase following stimulus-stimulus pairings in three children with autism. Similarly, pairings did not increase free-operant vocalizations in Experiment 2, a replication of Miguel et al. (2002). Experiment 3 demonstrated that shaping increased vowel frequency for one participant. Results suggest that variables are yet to be delineated that influence effectiveness of a stimulus-stimulus pairing procedure on vocalization frequency and acquisition of a verbal operant following such pairings. PMID:22477313
How Evolution May Work Through Curiosity-Driven Developmental Process.
Oudeyer, Pierre-Yves; Smith, Linda B
2016-04-01
Infants' own activities create and actively select their learning experiences. Here we review recent models of embodied information seeking and curiosity-driven learning and show that these mechanisms have deep implications for development and evolution. We discuss how these mechanisms yield self-organized epigenesis with emergent ordered behavioral and cognitive developmental stages. We describe a robotic experiment that explored the hypothesis that progress in learning, in and for itself, generates intrinsic rewards: The robot learners probabilistically selected experiences according to their potential for reducing uncertainty. In these experiments, curiosity-driven learning led the robot learner to successively discover object affordances and vocal interaction with its peers. We explain how a learning curriculum adapted to the current constraints of the learning system automatically formed, constraining learning and shaping the developmental trajectory. The observed trajectories in the robot experiment share many properties with those in infant development, including a mixture of regularities and diversities in the developmental patterns. Finally, we argue that such emergent developmental structures can guide and constrain evolution, in particular with regard to the origins of language. Copyright © 2016 Cognitive Science Society, Inc.
Neural Processing of Musical and Vocal Emotions Through Cochlear Implants Simulation.
Ahmed, Duha G; Paquette, Sebastian; Zeitouni, Anthony; Lehmann, Alexandre
2018-05-01
Cochlear implants (CIs) partially restore the sense of hearing in the deaf. However, the ability to recognize emotions in speech and music is reduced due to the implant's electrical signal limitations and the patient's altered neural pathways. Electrophysiological correlations of these limitations are not yet well established. Here we aimed to characterize the effect of CIs on auditory emotion processing and, for the first time, directly compare vocal and musical emotion processing through a CI-simulator. We recorded 16 normal hearing participants' electroencephalographic activity while listening to vocal and musical emotional bursts in their original form and in a degraded (CI-simulated) condition. We found prolonged P50 latency and reduced N100-P200 complex amplitude in the CI-simulated condition. This points to a limitation in encoding sound signals processed through CI simulation. When comparing the processing of vocal and musical bursts, we found a delay in latency with the musical bursts compared to the vocal bursts in both conditions (original and CI-simulated). This suggests that despite the cochlear implants' limitations, the auditory cortex can distinguish between vocal and musical stimuli. In addition, it adds to the literature supporting the complexity of musical emotion. Replicating this study with actual CI users might lead to characterizing emotional processing in CI users and could ultimately help develop optimal rehabilitation programs or device processing strategies to improve CI users' quality of life.
Yanagihara, Shin; Hessler, Neal A.
2011-01-01
Reactivations of waking experiences during sleep have been considered fundamental neural processes for memory consolidation. In songbirds, evidence suggests the importance of sleep-related neuronal activity in song system motor pathway nuclei for both juvenile vocal learning and maintenance of adult song. Like those in singing motor nuclei, neurons in the basal ganglia nucleus Area X, part of the basal ganglia-thalamocortical circuit essential for vocal plasticity, exhibit singing-related activity. It is unclear, however, whether Area X neurons show any distinctive spiking activity during sleep similar to that during singing. Here we demonstrate that, during sleep, Area X pallidal neurons exhibit phasic spiking activity, which shares some firing properties with activity during singing. Shorter interspike intervals that almost exclusively occurred during singing in awake periods were also observed during sleep. The level of firing variability was consistently higher during singing and sleep than during awake non-singing states. Moreover, deceleration of firing rate, which is considered to be an important firing property for transmitting signals from Area X to the thalamic nucleus DLM, was observed mainly during sleep as well as during singing. These results suggest that songbird basal ganglia circuitry may be involved in the off-line processing potentially critical for vocal learning during sensorimotor learning phase. PMID:21991379
Vahaba, Daniel M; Macedo-Lima, Matheus; Remage-Healey, Luke
2017-01-01
Vocal learning occurs during an experience-dependent, age-limited critical period early in development. In songbirds, vocal learning begins when presinging birds acquire an auditory memory of their tutor's song (sensory phase) followed by the onset of vocal production and refinement (sensorimotor phase). Hearing is necessary throughout the vocal learning critical period. One key brain area for songbird auditory processing is the caudomedial nidopallium (NCM), a telencephalic region analogous to mammalian auditory cortex. Despite NCM's established role in auditory processing, it is unclear how the response properties of NCM neurons may shift across development. Moreover, communication processing in NCM is rapidly enhanced by local 17β-estradiol (E2) administration in adult songbirds; however, the function of dynamically fluctuating E 2 in NCM during development is unknown. We collected bilateral extracellular recordings in NCM coupled with reverse microdialysis delivery in juvenile male zebra finches ( Taeniopygia guttata ) across the vocal learning critical period. We found that auditory-evoked activity and coding accuracy were substantially higher in the NCM of sensory-aged animals compared to sensorimotor-aged animals. Further, we observed both age-dependent and lateralized effects of local E 2 administration on sensory processing. In sensory-aged subjects, E 2 decreased auditory responsiveness across both hemispheres; however, a similar trend was observed in age-matched control subjects. In sensorimotor-aged subjects, E 2 dampened auditory responsiveness in left NCM but enhanced auditory responsiveness in right NCM. Our results reveal an age-dependent physiological shift in auditory processing and lateralized E 2 sensitivity that each precisely track a key neural "switch point" from purely sensory (pre-singing) to sensorimotor (singing) in developing songbirds.
2017-01-01
Abstract Vocal learning occurs during an experience-dependent, age-limited critical period early in development. In songbirds, vocal learning begins when presinging birds acquire an auditory memory of their tutor’s song (sensory phase) followed by the onset of vocal production and refinement (sensorimotor phase). Hearing is necessary throughout the vocal learning critical period. One key brain area for songbird auditory processing is the caudomedial nidopallium (NCM), a telencephalic region analogous to mammalian auditory cortex. Despite NCM’s established role in auditory processing, it is unclear how the response properties of NCM neurons may shift across development. Moreover, communication processing in NCM is rapidly enhanced by local 17β-estradiol (E2) administration in adult songbirds; however, the function of dynamically fluctuating E2 in NCM during development is unknown. We collected bilateral extracellular recordings in NCM coupled with reverse microdialysis delivery in juvenile male zebra finches (Taeniopygia guttata) across the vocal learning critical period. We found that auditory-evoked activity and coding accuracy were substantially higher in the NCM of sensory-aged animals compared to sensorimotor-aged animals. Further, we observed both age-dependent and lateralized effects of local E2 administration on sensory processing. In sensory-aged subjects, E2 decreased auditory responsiveness across both hemispheres; however, a similar trend was observed in age-matched control subjects. In sensorimotor-aged subjects, E2 dampened auditory responsiveness in left NCM but enhanced auditory responsiveness in right NCM. Our results reveal an age-dependent physiological shift in auditory processing and lateralized E2 sensitivity that each precisely track a key neural “switch point” from purely sensory (pre-singing) to sensorimotor (singing) in developing songbirds. PMID:29255797
Central Nervous System Control of Voice and Swallowing
Ludlow, Christy L.
2015-01-01
This review of the central nervous control systems for voice and swallowing has suggested that the traditional concepts of a separation between cortical and limbic and brain stem control should be refined and more integrative. For voice production, a separation of the non-human vocalization system from the human learned voice production system has been posited based primarily on studies of non-human primates. However, recent humans studies of emotionally based vocalizations and human volitional voice production has shown more integration between these two systems than previously proposed. Recent human studies have shown that reflexive vocalization as well as learned voice production not involving speech, involve a common integrative system. On the other hand, recent studies of non-human primates have provided evidence of some cortical activity during vocalization and cortical changes with training during vocal behavior. For swallowing, evidence from the macaque and functional brain imaging in humans indicates that the control for the pharyngeal phase of swallowing is not primarily under brain stem mechanisms as previously proposed. Studies suggest that the initiation and patterning of swallowing for the pharyngeal phase is also under active cortical control for both spontaneous as well as volitional swallowing in awake humans and non-human primates. PMID:26241238
Communication Modality Sampling for a Toddler with Angelman Syndrome
ERIC Educational Resources Information Center
Martin, Jolene Hyppa; Reichle, Joe; Dimian, Adele; Chen, Mo
2013-01-01
Purpose: Vocal, gestural, and graphic communication modes were implemented concurrently with a toddler with Angelman syndrome to identify the most efficiently learned communication mode to emphasize in an initial augmentative communication system. Method: Symbols representing preferred objects were introduced in vocal, gestural, and graphic…
Vowels Development in Babbling of typically developing 6-to-12-month old Persian-learning Infants.
Fotuhi, Mina; Yadegari, Fariba; Teymouri, Robab
2017-10-01
Pre-linguistic vocalizations including early consonants, vowels, and their combinations into syllables are considered as important predictors of the speech and language development. The purpose of this study was to examine vowel development in babblings of normally developing Persian-learning infants. Eight typically developing 6-8-month-old Persian-learning infants (3 boys and 5 girls) participated in this 4-month longitudinal descriptive-analytic study. A weekly 30-60-minute audio- and video-recording was obtained at home from the comfort state vocalizations of infants and the mother-child interactions. A total of 74:02:03 hours of vocalizations were phonetically transcribed. Seven vowels comprising /i/,/e/,/a/,/u/,/o/,/ɑ/, and /ә/ were identified in the babblings. The inter-rater reliability was obtained for 20% of vocalizations. The data were analyzed by repeated measures ANOVA and Pearson's correlation coefficient using SPSS software version 20. The results showed that two vowels /a/ (46.04) and /e/ (23.60) were produced with the highest mean frequency of occurrence, respectively. Regarding front/back dimension, the front vowels were the most prominent ones (71.87); in terms of height, low (46.78) and mid (32.45) vowels occurred maximally. A good inter-rater reliability was obtained (0.99, P < .01). The increased frequency of occurrence of the low and mid front vowels in the current study was consistent with previous studies on the emergence of vowels in pre-linguistic vocalization in other languages.
Cross-cultural adaptation of the Brazilian version of the Vocal Fatigue Index - VFI.
Zambon, Fabiana; Moreti, Felipe; Nanjundeswaran, Chayadevie; Behlau, Mara
2017-03-13
The purpose of this study was to perform the cultural adaptation of the Brazilian version of the Vocal Fatigue Index (VFI). Two Brazilian bilingual speech-language pathologists (SLP) translated the original version of the VFI in English into Portuguese. The translations were reviewed by a committee of five voice specialist SLPs resulting in the final version of the instrument. A third bilingual SLP back-translated this final version and the same committee reviewed the differences from its original version. The final Portuguese version of the VFI, as in the original English version, was answered on a categorical scale of 0-4 indicating the frequency they experience the symptoms: 0=never, 1=almost never, 2=sometimes, 3=almost always, and 4=always. For cultural equivalence of the Portuguese version, the option "not applicable" was added to the categorical scale and 20 individuals with vocal complaints and dysphonia completed the index. Questions considered "not applicable" would be disregarded from the Brazilian version of the protocol; no question had to be removed from the instrument. The Brazilian Portuguese version was entitled "Índice de Fadiga Vocal - IFV" and features 19 questions, equivalent to the original instrument. Of the 19 items, 11 were related with tiredness of voice and voice avoidance, five concerned physical discomfort associated with voicing, and three were related to improvement of symptoms with rest or lack thereof. The Brazilian version of the VFI presents cultural and linguistic equivalence to the original instrument. The IFV validation into Brazilian Portuguese is in progress.
A Joint Prosodic Origin of Language and Music
Brown, Steven
2017-01-01
Vocal theories of the origin of language rarely make a case for the precursor functions that underlay the evolution of speech. The vocal expression of emotion is unquestionably the best candidate for such a precursor, although most evolutionary models of both language and speech ignore emotion and prosody altogether. I present here a model for a joint prosodic precursor of language and music in which ritualized group-level vocalizations served as the ancestral state. This precursor combined not only affective and intonational aspects of prosody, but also holistic and combinatorial mechanisms of phrase generation. From this common stage, there was a bifurcation to form language and music as separate, though homologous, specializations. This separation of language and music was accompanied by their (re)unification in songs with words. PMID:29163276
Mouse Vocal Communication System: Are Ultrasounds Learned or Innate?
ERIC Educational Resources Information Center
Arriaga, Gustavo; Jarvis, Erich D.
2013-01-01
Mouse ultrasonic vocalizations (USVs) are often used as behavioral readouts of internal states, to measure effects of social and pharmacological manipulations, and for behavioral phenotyping of mouse models for neuropsychiatric and neurodegenerative disorders. However, little is known about the neurobiological mechanisms of rodent USV production.…
Distribution of language-related Cntnap2 protein in neural circuits critical for vocal learning.
Condro, Michael C; White, Stephanie A
2014-01-01
Variants of the contactin associated protein-like 2 (Cntnap2) gene are risk factors for language-related disorders including autism spectrum disorder, specific language impairment, and stuttering. Songbirds are useful models for study of human speech disorders due to their shared capacity for vocal learning, which relies on similar cortico-basal ganglia circuitry and genetic factors. Here we investigate Cntnap2 protein expression in the brain of the zebra finch, a songbird species in which males, but not females, learn their courtship songs. We hypothesize that Cntnap2 has overlapping functions in vocal learning species, and expect to find protein expression in song-related areas of the zebra finch brain. We further expect that the distribution of this membrane-bound protein may not completely mirror its mRNA distribution due to the distinct subcellular localization of the two molecular species. We find that Cntnap2 protein is enriched in several song control regions relative to surrounding tissues, particularly within the adult male, but not female, robust nucleus of the arcopallium (RA), a cortical song control region analogous to human layer 5 primary motor cortex. The onset of this sexually dimorphic expression coincides with the onset of sensorimotor learning in developing males. Enrichment in male RA appears due to expression in projection neurons within the nucleus, as well as to additional expression in nerve terminals of cortical projections to RA from the lateral magnocellular nucleus of the nidopallium. Cntnap2 protein expression in zebra finch brain supports the hypothesis that this molecule affects neural connectivity critical for vocal learning across taxonomic classes. Copyright © 2013 Wiley Periodicals, Inc.
Human-like brain hemispheric dominance in birdsong learning.
Moorman, Sanne; Gobes, Sharon M H; Kuijpers, Maaike; Kerkhofs, Amber; Zandbergen, Matthijs A; Bolhuis, Johan J
2012-07-31
Unlike nonhuman primates, songbirds learn to vocalize very much like human infants acquire spoken language. In humans, Broca's area in the frontal lobe and Wernicke's area in the temporal lobe are crucially involved in speech production and perception, respectively. Songbirds have analogous brain regions that show a similar neural dissociation between vocal production and auditory perception and memory. In both humans and songbirds, there is evidence for lateralization of neural responsiveness in these brain regions. Human infants already show left-sided dominance in their brain activation when exposed to speech. Moreover, a memory-specific left-sided dominance in Wernicke's area for speech perception has been demonstrated in 2.5-mo-old babies. It is possible that auditory-vocal learning is associated with hemispheric dominance and that this association arose in songbirds and humans through convergent evolution. Therefore, we investigated whether there is similar song memory-related lateralization in the songbird brain. We exposed male zebra finches to tutor or unfamiliar song. We found left-sided dominance of neuronal activation in a Broca-like brain region (HVC, a letter-based name) of juvenile and adult zebra finch males, independent of the song stimulus presented. In addition, juvenile males showed left-sided dominance for tutor song but not for unfamiliar song in a Wernicke-like brain region (the caudomedial nidopallium). Thus, left-sided dominance in the caudomedial nidopallium was specific for the song-learning phase and was memory-related. These findings demonstrate a remarkable neural parallel between birdsong and human spoken language, and they have important consequences for our understanding of the evolution of auditory-vocal learning and its neural mechanisms.
Quality and Readability of English-Language Internet Information for Voice Disorders.
Dueppen, Abigail J; Bellon-Harn, Monica L; Radhakrishnan, Nandhakumar; Manchaiah, Vinaya
2017-12-15
The purpose of this study is to evaluate the readability and quality of English-language Internet information related to vocal hygiene, vocal health, and prevention of voice disorders. This study extends recent work because it evaluates readability, content quality, and website origin across broader search criteria than previous studies evaluating online voice material. Eighty-five websites were aggregated using five different country-specific search engines. Websites were then analyzed using quality and readability assessments. The entire web page was evaluated; however, no information or links beyond the first page was reviewed. Statistical calculations were employed to examine website ratings, differences between website origin and quality and readability scores, and correlations between readability instruments. Websites exhibited acceptable quality as measured by the DISCERN. However, only one website obtained the Health On the Net certification. Significant differences in quality were found among website origin, with government websites receiving higher quality ratings. Approximate educational levels required to comprehend information on the websites ranged from 8 to 9 years of education. Significant differences were found between website origin and readability measures with higher levels of education required to understand information on websites of nonprofit organizations. Current vocal hygiene, vocal health, and prevention of voice disorders websites were found to exhibit acceptable levels of quality and readability. However, highly rated Internet information related to voice care should be made more accessible to voice clients through Health On the Net certification. Published by Elsevier Inc.
Production, Usage, and Comprehension in Animal Vocalizations
ERIC Educational Resources Information Center
Seyfarth, Robert M.; Cheney, Dorothy L.
2010-01-01
In this review, we place equal emphasis on production, usage, and comprehension because these components of communication may exhibit different developmental trajectories and be affected by different neural mechanisms. In the animal kingdom generally, learned, flexible vocal production is rare, appearing in only a few orders of birds and few…
Hamaguchi, Kosuke; Mooney, Richard
2012-01-01
Complex brain functions, such as the capacity to learn and modulate vocal sequences, depend on activity propagation in highly distributed neural networks. To explore the synaptic basis of activity propagation in such networks, we made dual in vivo intracellular recordings in anesthetized zebra finches from the input (nucleus HVC) and output (lateral magnocellular nucleus of the anterior nidopallium (LMAN)) neurons of a songbird cortico-basal ganglia (BG) pathway necessary to the learning and modulation of vocal motor sequences. These recordings reveal evidence of bidirectional interactions, rather than only feedforward propagation of activity from HVC to LMAN, as had been previously supposed. A combination of dual and triple recording configurations and pharmacological manipulations was used to map out circuitry by which activity propagates from LMAN to HVC. These experiments indicate that activity travels to HVC through at least two independent ipsilateral pathways, one of which involves fast signaling through a midbrain dopaminergic cell group, reminiscent of recurrent mesocortical loops described in mammals. We then used in vivo pharmacological manipulations to establish that augmented LMAN activity is sufficient to restore high levels of sequence variability in adult birds, suggesting that recurrent interactions through highly distributed forebrain – midbrain pathways can modulate learned vocal sequences. PMID:22915110
Talker identification across source mechanisms: experiments with laryngeal and electrolarynx speech.
Perrachione, Tyler K; Stepp, Cara E; Hillman, Robert E; Wong, Patrick C M
2014-10-01
The purpose of this study was to determine listeners' ability to learn talker identity from speech produced with an electrolarynx, explore source and filter differentiation in talker identification, and describe acoustic-phonetic changes associated with electrolarynx use. Healthy adult control listeners learned to identify talkers from speech recordings produced using talkers' normal laryngeal vocal source or an electrolarynx. Listeners' abilities to identify talkers from the trained vocal source (Experiment 1) and generalize this knowledge to the untrained source (Experiment 2) were assessed. Acoustic-phonetic measurements of spectral differences between source mechanisms were performed. Additional listeners attempted to match recordings from different source mechanisms to a single talker (Experiment 3). Listeners successfully learned talker identity from electrolarynx speech but less accurately than from laryngeal speech. Listeners were unable to generalize talker identity to the untrained source mechanism. Electrolarynx use resulted in vowels with higher F1 frequencies compared with laryngeal speech. Listeners matched recordings from different sources to a single talker better than chance. Electrolarynx speech, although lacking individual differences in voice quality, nevertheless conveys sufficient indexical information related to the vocal filter and articulation for listeners to identify individual talkers. Psychologically, perception of talker identity arises from a "gestalt" of the vocal source and filter.
Sohrabji, F; Nordeen, E J; Nordeen, K W
1990-01-01
Area X, a large sexually dimorphic nucleus in the avian ventral forebrain, is part of a highly discrete system of interconnected nuclei that have been implicated in either song learning or adult song production. Previously, this nucleus has been included in the song system because of its substantial connections with other vocal control nuclei, and because its volume is positively correlated with the capacity for song. In order to directly assess the role of Area X in song behavior, this nucleus was bilaterally lesioned in both juvenile and adult zebra finches, using ibotenic acid. We report here that lesioning Area X disrupts normal song development in juvenile birds, but does not affect the production of stereotyped song by adult birds. Although juvenile-lesioned birds were consistently judged as being in earlier stages of vocal development than age-matched controls, they continued to produce normal song-like vocalizations. Thus, unlike the lateral magnocellular nucleus of the anterior neostriatum, another avian forebrain nucleus implicated in song learning, Area X does not seem to be necessary for sustaining production of juvenile song. Rather, the behavioral results suggest Area X is important for either the acquisition of a song model or the improvement of song through vocal practice.
Talker identification across source mechanisms: Experiments with laryngeal and electrolarynx speech
Perrachione, Tyler K.; Stepp, Cara E.; Hillman, Robert E.; Wong, Patrick C.M.
2015-01-01
Purpose To determine listeners' ability to learn talker identity from speech produced with an electrolarynx, explore source and filter differentiation in talker identification, and describe acoustic-phonetic changes associated with electrolarynx use. Method Healthy adult control listeners learned to identify talkers from speech recordings produced using talkers' normal laryngeal vocal source or an electrolarynx. Listeners' abilities to identify talkers from the trained vocal source (Experiment 1) and generalize this knowledge to the untrained source (Experiment 2) were assessed. Acoustic-phonetic measurements of spectral differences between source mechanisms were performed. Additional listeners attempted to match recordings from different source mechanisms to a single talker (Experiment 3). Results Listeners successfully learned talker identity from electrolarynx speech, but less accurately than from laryngeal speech. Listeners were unable to generalize talker identity to the untrained source mechanism. Electrolarynx use resulted in vowels with higher F1 frequencies compared to laryngeal speech. Listeners matched recordings from different sources to a single talker better than chance. Conclusions Electrolarynx speech, though lacking individual differences in voice quality, nevertheless conveys sufficient indexical information related to the vocal filter and articulation for listeners to identify individual talkers. Psychologically, perception of talker identity arises from a “gestalt” of the vocal source and filter. PMID:24801962
Gruzelier, J H; Holmes, P; Hirst, L; Bulpin, K; Rahman, S; van Run, C; Leach, J
2014-01-01
Alpha/theta (A/T) and sensory-motor rhythm (SMR) neurofeedback were compared in university instrumentalists who were novice singers with regard to prepared and improvised instrumental and vocal performance in three music domains: creativity/musicality, technique and communication/presentation. Only A/T training enhanced advanced playing seen in all three domains by expert assessors and validated by correlations with learning indices, strongest with Creativity/Musicality as shown by Egner and Gruzelier (2003). Here A/T gains extended to novice performance - prepared vocal, improvised vocal and instrumental - and were recognised by a lay audience who judged the prepared folk songs. SMR learning correlated positively with Technical Competence and Communication in novice performance, in keeping with SMR neurofeedback's known impact on lower-order processes such as attention, working memory and psychomotor skills. The importance of validation through learning indices was emphasised in the interpretation of neurofeedback outcome. Copyright © 2013 Elsevier B.V. All rights reserved.
Vertical transmission of learned signatures in a wild parrot
Berg, Karl S.; Delgado, Soraya; Cortopassi, Kathryn A.; Beissinger, Steven R.; Bradbury, Jack W.
2012-01-01
Learned birdsong is a widely used animal model for understanding the acquisition of human speech. Male songbirds often learn songs from adult males during sensitive periods early in life, and sing to attract mates and defend territories. In presumably all of the 350+ parrot species, individuals of both sexes commonly learn vocal signals throughout life to satisfy a wide variety of social functions. Despite intriguing parallels with humans, there have been no experimental studies demonstrating learned vocal production in wild parrots. We studied contact call learning in video-rigged nests of a well-known marked population of green-rumped parrotlets (Forpus passerinus) in Venezuela. Both sexes of naive nestlings developed individually unique contact calls in the nest, and we demonstrate experimentally that signature attributes are learned from both primary care-givers. This represents the first experimental evidence for the mechanisms underlying the transmission of a socially acquired trait in a wild parrot population. PMID:21752824
Temperature-dependent regulation of vocal pattern generator.
Yamaguchi, Ayako; Gooler, David; Herrold, Amy; Patel, Shailja; Pong, Winnie W
2008-12-01
Vocalizations of Xenopus laevis are generated by central pattern generators (CPGs). The advertisement call of male X. laevis is a complex biphasic motor rhythm consisting of fast and slow trills (a train of clicks). We found that the trill rate of these advertisement calls is sensitive to temperature and that this rate modification of the vocal rhythms originates in the central pattern generators. In vivo the rates of fast and slow trills increased linearly with an increase in temperature. In vitro a similar linear relation between temperature and compound action potential frequency in the laryngeal nerve was found when fictive advertisement calls were evoked in the isolated brain. Temperature did not limit the contractile properties of laryngeal muscles within the frequency range of vocalizations. We next took advantage of the temperature sensitivity of the vocal CPG in vitro to localize the source of the vocal rhythms. We focused on the dorsal tegmental area of the medulla (DTAM), a brain stem nucleus that is essential for vocal production. We found that bilateral cooling of DTAM reduced both fast and slow trill rates. Thus we conclude that DTAM is a source of biphasic vocal rhythms.
A versatile pitch tracking algorithm: from human speech to killer whale vocalizations.
Shapiro, Ari Daniel; Wang, Chao
2009-07-01
In this article, a pitch tracking algorithm [named discrete logarithmic Fourier transformation-pitch detection algorithm (DLFT-PDA)], originally designed for human telephone speech, was modified for killer whale vocalizations. The multiple frequency components of some of these vocalizations demand a spectral (rather than temporal) approach to pitch tracking. The DLFT-PDA algorithm derives reliable estimations of pitch and the temporal change of pitch from the harmonic structure of the vocal signal. Scores from both estimations are combined in a dynamic programming search to find a smooth pitch track. The algorithm is capable of tracking killer whale calls that contain simultaneous low and high frequency components and compares favorably across most signal to noise ratio ranges to the peak-picking and sidewinder algorithms that have been used for tracking killer whale vocalizations previously.
Temporal signatures of processing voiceness and emotion in sound
Gunter, Thomas C.
2017-01-01
Abstract This study explored the temporal course of vocal and emotional sound processing. Participants detected rare repetitions in a stimulus stream comprising neutral and surprised non-verbal exclamations and spectrally rotated control sounds. Spectral rotation preserved some acoustic and emotional properties of the vocal originals. Event-related potentials elicited to unrepeated sounds revealed effects of voiceness and emotion. Relative to non-vocal sounds, vocal sounds elicited a larger centro-parietally distributed N1. This effect was followed by greater positivity to vocal relative to non-vocal sounds beginning with the P2 and extending throughout the recording epoch (N4, late positive potential) with larger amplitudes in female than in male listeners. Emotion effects overlapped with the voiceness effects but were smaller and differed topographically. Voiceness and emotion interacted only for the late positive potential, which was greater for vocal-emotional as compared with all other sounds. Taken together, these results point to a multi-stage process in which voiceness and emotionality are represented independently before being integrated in a manner that biases responses to stimuli with socio-emotional relevance. PMID:28338796
Temporal signatures of processing voiceness and emotion in sound.
Schirmer, Annett; Gunter, Thomas C
2017-06-01
This study explored the temporal course of vocal and emotional sound processing. Participants detected rare repetitions in a stimulus stream comprising neutral and surprised non-verbal exclamations and spectrally rotated control sounds. Spectral rotation preserved some acoustic and emotional properties of the vocal originals. Event-related potentials elicited to unrepeated sounds revealed effects of voiceness and emotion. Relative to non-vocal sounds, vocal sounds elicited a larger centro-parietally distributed N1. This effect was followed by greater positivity to vocal relative to non-vocal sounds beginning with the P2 and extending throughout the recording epoch (N4, late positive potential) with larger amplitudes in female than in male listeners. Emotion effects overlapped with the voiceness effects but were smaller and differed topographically. Voiceness and emotion interacted only for the late positive potential, which was greater for vocal-emotional as compared with all other sounds. Taken together, these results point to a multi-stage process in which voiceness and emotionality are represented independently before being integrated in a manner that biases responses to stimuli with socio-emotional relevance. © The Author (2017). Published by Oxford University Press.
A Brain for Speech. Evolutionary Continuity in Primate and Human Auditory-Vocal Processing
Aboitiz, Francisco
2018-01-01
In this review article, I propose a continuous evolution from the auditory-vocal apparatus and its mechanisms of neural control in non-human primates, to the peripheral organs and the neural control of human speech. Although there is an overall conservatism both in peripheral systems and in central neural circuits, a few changes were critical for the expansion of vocal plasticity and the elaboration of proto-speech in early humans. Two of the most relevant changes were the acquisition of direct cortical control of the vocal fold musculature and the consolidation of an auditory-vocal articulatory circuit, encompassing auditory areas in the temporoparietal junction and prefrontal and motor areas in the frontal cortex. This articulatory loop, also referred to as the phonological loop, enhanced vocal working memory capacity, enabling early humans to learn increasingly complex utterances. The auditory-vocal circuit became progressively coupled to multimodal systems conveying information about objects and events, which gradually led to the acquisition of modern speech. Gestural communication accompanies the development of vocal communication since very early in human evolution, and although both systems co-evolved tightly in the beginning, at some point speech became the main channel of communication. PMID:29636657
[Causes of vocal cord dyscinesia and its original factors after endotracheal intubation].
Sun, Anke; Zhang, Tiezheng; Liu, Wenyuan; Tang, Weiwei; Guo, Xiaohong
2012-03-01
To research the causes of postintubation vocal cord dyskinesia and its contributing factors. The causes of vocal cord dyskinesia were confirmed by laryngoscope, three-dimensional spiral CT, stroboscope, and the analysis of therapy. The factors relevant to the causes of vocal cord dyskinesia were analysed based on the following elements: (1) the anatomic or pathological condition of patients or the technical skills of anesthetists. (2) emaciated or obese body and neck. (3) the age of patients. (4) the duration of endotracheal tube retention. (5) the types of operations. (6) anesthesia procedure. Among 135 patients, 128 cases (94.81%) manifested arytenoid dislocation, 7 cases (5.19%) vocal cord paralysis. The study showed that the vocal cord dyskinesia associated with anatomic or pathological condition of patients and technical skills of anesthetists (with intubation difficulty) accounted for 76.30%. The patients with relative emaciated body or neck accounted for 90.62% in cases without intubation difficulty. Age had no significant analytical relationship with vocal cord dyskinesia. Prolonged intubation (endotracheal tube retention over 12 hours) was accounted for only 17.64%. The incidence of vocal cord dyskinesia was nearly 0.5% in patients underwent cardio-thoracic surgery, accounting for 59.26% of all the patients. There are two major causes of vocal cord dyskinesia: arytenoid dislocation and vocal cord paralysis, and the rate of vocal cord dyskinesia could be reduced by the improvement of technical skill of anesthetists and/or sufficient attention to the intubation condition of patients.
Communication modality sampling for a toddler with Angelman syndrome.
Hyppa Martin, Jolene; Reichle, Joe; Dimian, Adele; Chen, Mo
2013-10-01
Vocal, gestural, and graphic communication modes were implemented concurrently with a toddler with Angelman syndrome to identify the most efficiently learned communication mode to emphasize in an initial augmentative communication system. Symbols representing preferred objects were introduced in vocal, gestural, and graphic communication modes using an alternating treatment single-subject experimental design. Conventionally accepted prompting strategies were used to teach symbols in each communication mode. Because the learner did not vocally imitate, vocal mode intervention focused on increasing vocal frequency as an initial step. When graphic and gestural mode performances were compared, the learner most accurately produced requests in graphic mode (percentage of nonoverlapping data = 96). Given the lack of success in prompting vocal productions, a comparison between vocal and the other two communication modes was not made. A growing body of evidence suggests that concurrent modality sampling is a promising low-inference, data-driven procedure that can be used to inform selection of a communication mode(s) for initial emphasis with young children. Concurrent modality sampling can guide clinical decisions regarding the allocation of treatment resources to promote success in building an initial communicative repertoire.
Acoustic correlates of body size and individual identity in banded penguins
Gamba, Marco; Gili, Claudia; Pessani, Daniela
2017-01-01
Animal vocalisations play a role in individual recognition and mate choice. In nesting penguins, acoustic variation in vocalisations originates from distinctiveness in the morphology of the vocal apparatus. Using the source-filter theory approach, we investigated vocal individuality cues and correlates of body size and mass in the ecstatic display songs the Humboldt and Magellanic penguins. We demonstrate that both fundamental frequency (f0) and formants (F1-F4) are essential vocal features to discriminate among individuals. However, we show that only duration and f0 are honest indicators of the body size and mass, respectively. We did not find any effect of body dimension on formants, formant dispersion nor estimated vocal tract length of the emitters. Overall, our findings provide the first evidence that the resonant frequencies of the vocal tract do not correlate with body size in penguins. Our results add important information to a growing body of literature on the role of the different vocal parameters in conveying biologically meaningful information in bird vocalisations. PMID:28199318
Primate vocal communication: a useful tool for understanding human speech and language evolution?
Fedurek, Pawel; Slocombe, Katie E
2011-04-01
Language is a uniquely human trait, and questions of how and why it evolved have been intriguing scientists for years. Nonhuman primates (primates) are our closest living relatives, and their behavior can be used to estimate the capacities of our extinct ancestors. As humans and many primate species rely on vocalizations as their primary mode of communication, the vocal behavior of primates has been an obvious target for studies investigating the evolutionary roots of human speech and language. By studying the similarities and differences between human and primate vocalizations, comparative research has the potential to clarify the evolutionary processes that shaped human speech and language. This review examines some of the seminal and recent studies that contribute to our knowledge regarding the link between primate calls and human language and speech. We focus on three main aspects of primate vocal behavior: functional reference, call combinations, and vocal learning. Studies in these areas indicate that despite important differences, primate vocal communication exhibits some key features characterizing human language. They also indicate, however, that some critical aspects of speech, such as vocal plasticity, are not shared with our primate cousins. We conclude that comparative research on primate vocal behavior is a very promising tool for deepening our understanding of the evolution of human speech and language, but much is still to be done as many aspects of monkey and ape vocalizations remain largely unexplored.
Discussion: Changes in Vocal Production and Auditory Perception after Hair Cell Regeneration.
ERIC Educational Resources Information Center
Ryals, Brenda M.; Dooling, Robert J.
2000-01-01
A bird study found that with sufficient time and training after hair cell and hearing loss and hair cell regeneration, the mature avian auditory system can accommodate input from a newly regenerated periphery sufficiently to allow for recognition of previously familiar vocalizations and the learning of new complex acoustic classifications.…
A Comparison of Written, Vocal, and Video Feedback When Training Teachers
ERIC Educational Resources Information Center
Luck, Kally M.; Lerman, Dorothea C.; Wu, Wai-Ling; Dupuis, Danielle L.; Hussein, Louisa A.
2018-01-01
We compared the effectiveness of and preference for different feedback strategies when training six special education teachers during a 5-day summer training program. In Experiment 1, teachers received written or vocal feedback while learning to implement two different types of preference assessments. In Experiment 2, we compared either written or…
Mindfulness for Singers: The Effects of a Targeted Mindfulness Course on Learning Vocal Technique
ERIC Educational Resources Information Center
Czajkowski, Anne-Marie L.; Greasley, Alinka E.
2015-01-01
This paper reports the development and implementation of a unique Mindfulness for Singers (MfS) course designed to improve singers' vocal technique. Eight university students completed the intervention. Five Facet Mindfulness Questionnaire (FFMQ) scores showed general improvement across all five facets of mindfulness. Qualitative results showed…
Chabout, Jonathan; Sarkar, Abhra; Patel, Sheel R; Radden, Taylor; Dunson, David B; Fisher, Simon E; Jarvis, Erich D
2016-01-01
Development of proficient spoken language skills is disrupted by mutations of the FOXP2 transcription factor. A heterozygous missense mutation in the KE family causes speech apraxia, involving difficulty producing words with complex learned sequences of syllables. Manipulations in songbirds have helped to elucidate the role of this gene in vocal learning, but findings in non-human mammals have been limited or inconclusive. Here, we performed a systematic study of ultrasonic vocalizations (USVs) of adult male mice carrying the KE family mutation. Using novel statistical tools, we found that Foxp2 heterozygous mice did not have detectable changes in USV syllable acoustic structure, but produced shorter sequences and did not shift to more complex syntax in social contexts where wildtype animals did. Heterozygous mice also displayed a shift in the position of their rudimentary laryngeal motor cortex (LMC) layer-5 neurons. Our findings indicate that although mouse USVs are mostly innate, the underlying contributions of FoxP2 to sequencing of vocalizations are conserved with humans.
Chabout, Jonathan; Sarkar, Abhra; Patel, Sheel R.; Radden, Taylor; Dunson, David B.; Fisher, Simon E.; Jarvis, Erich D.
2016-01-01
Development of proficient spoken language skills is disrupted by mutations of the FOXP2 transcription factor. A heterozygous missense mutation in the KE family causes speech apraxia, involving difficulty producing words with complex learned sequences of syllables. Manipulations in songbirds have helped to elucidate the role of this gene in vocal learning, but findings in non-human mammals have been limited or inconclusive. Here, we performed a systematic study of ultrasonic vocalizations (USVs) of adult male mice carrying the KE family mutation. Using novel statistical tools, we found that Foxp2 heterozygous mice did not have detectable changes in USV syllable acoustic structure, but produced shorter sequences and did not shift to more complex syntax in social contexts where wildtype animals did. Heterozygous mice also displayed a shift in the position of their rudimentary laryngeal motor cortex (LMC) layer-5 neurons. Our findings indicate that although mouse USVs are mostly innate, the underlying contributions of FoxP2 to sequencing of vocalizations are conserved with humans. PMID:27812326
Mets, David G; Brainard, Michael S
2018-01-01
Abstract Background Vocal learning in songbirds has emerged as a powerful model for sensorimotor learning. Neurobehavioral studies of Bengalese finch (Lonchura striata domestica) song, naturally more variable and plastic than songs of other finch species, have demonstrated the importance of behavioral variability for initial learning, maintenance, and plasticity of vocalizations. However, the molecular and genetic underpinnings of this variability and the learning it supports are poorly understood. Findings To establish a platform for the molecular analysis of behavioral variability and plasticity, we generated an initial draft assembly of the Bengalese finch genome from a single male animal to 151× coverage and an N50 of 3.0 MB. Furthermore, we developed an initial set of gene models using RNA-seq data from 8 samples that comprise liver, muscle, cerebellum, brainstem/midbrain, and forebrain tissue from juvenile and adult Bengalese finches of both sexes. Conclusions We provide a draft Bengalese finch genome and gene annotation to facilitate the study of the molecular-genetic influences on behavioral variability and the process of vocal learning. These data will directly support many avenues for the identification of genes involved in learning, including differential expression analysis, comparative genomic analysis (through comparison to existing avian genome assemblies), and derivation of genetic maps for linkage analysis. Bengalese finch gene models and sequences will be essential for subsequent manipulation (molecular or genetic) of genes and gene products, enabling novel mechanistic investigations into the role of variability in learned behavior. PMID:29618046
Bottlenose dolphins can use learned vocal labels to address each other
King, Stephanie L.; Janik, Vincent M.
2013-01-01
In animal communication research, vocal labeling refers to incidents in which an animal consistently uses a specific acoustic signal when presented with a specific object or class of objects. Labeling with learned signals is a foundation of human language but is notably rare in nonhuman communication systems. In natural animal systems, labeling often occurs with signals that are not influenced by learning, such as in alarm and food calling. There is a suggestion, however, that some species use learned signals to label conspecific individuals in their own communication system when mimicking individually distinctive calls. Bottlenose dolphins (Tursiops truncatus) are a promising animal for exploration in this area because they are capable of vocal production learning and can learn to use arbitrary signals to report the presence or absence of objects. Bottlenose dolphins develop their own unique identity signal, the signature whistle. This whistle encodes individual identity independently of voice features. The copying of signature whistles may therefore allow animals to label or address one another. Here, we show that wild bottlenose dolphins respond to hearing a copy of their own signature whistle by calling back. Animals did not respond to whistles that were not their own signature. This study provides compelling evidence that a dolphin’s learned identity signal is used as a label when addressing conspecifics. Bottlenose dolphins therefore appear to be unique as nonhuman mammals to use learned signals as individually specific labels for different social companions in their own natural communication system. PMID:23878217
Colquitt, Bradley M; Mets, David G; Brainard, Michael S
2018-03-01
Vocal learning in songbirds has emerged as a powerful model for sensorimotor learning. Neurobehavioral studies of Bengalese finch (Lonchura striata domestica) song, naturally more variable and plastic than songs of other finch species, have demonstrated the importance of behavioral variability for initial learning, maintenance, and plasticity of vocalizations. However, the molecular and genetic underpinnings of this variability and the learning it supports are poorly understood. To establish a platform for the molecular analysis of behavioral variability and plasticity, we generated an initial draft assembly of the Bengalese finch genome from a single male animal to 151× coverage and an N50 of 3.0 MB. Furthermore, we developed an initial set of gene models using RNA-seq data from 8 samples that comprise liver, muscle, cerebellum, brainstem/midbrain, and forebrain tissue from juvenile and adult Bengalese finches of both sexes. We provide a draft Bengalese finch genome and gene annotation to facilitate the study of the molecular-genetic influences on behavioral variability and the process of vocal learning. These data will directly support many avenues for the identification of genes involved in learning, including differential expression analysis, comparative genomic analysis (through comparison to existing avian genome assemblies), and derivation of genetic maps for linkage analysis. Bengalese finch gene models and sequences will be essential for subsequent manipulation (molecular or genetic) of genes and gene products, enabling novel mechanistic investigations into the role of variability in learned behavior.
Evolution, brain, and the nature of language.
Berwick, Robert C; Friederici, Angela D; Chomsky, Noam; Bolhuis, Johan J
2013-02-01
Language serves as a cornerstone for human cognition, yet much about its evolution remains puzzling. Recent research on this question parallels Darwin's attempt to explain both the unity of all species and their diversity. What has emerged from this research is that the unified nature of human language arises from a shared, species-specific computational ability. This ability has identifiable correlates in the brain and has remained fixed since the origin of language approximately 100 thousand years ago. Although songbirds share with humans a vocal imitation learning ability, with a similar underlying neural organization, language is uniquely human. Copyright © 2012 Elsevier Ltd. All rights reserved.
Observational Learning in the Music Masterclass
ERIC Educational Resources Information Center
Haddon, Elizabeth
2014-01-01
This article contributes to research on music masterclasses through examining learning through observation. It investigates how students are learning as observers in this context; whether and how they will transfer their masterclass learning to their own instrumental/vocal development, and whether they have discussed learning through observation.…
Human-like brain hemispheric dominance in birdsong learning
Moorman, Sanne; Gobes, Sharon M. H.; Kuijpers, Maaike; Kerkhofs, Amber; Zandbergen, Matthijs A.; Bolhuis, Johan J.
2012-01-01
Unlike nonhuman primates, songbirds learn to vocalize very much like human infants acquire spoken language. In humans, Broca’s area in the frontal lobe and Wernicke’s area in the temporal lobe are crucially involved in speech production and perception, respectively. Songbirds have analogous brain regions that show a similar neural dissociation between vocal production and auditory perception and memory. In both humans and songbirds, there is evidence for lateralization of neural responsiveness in these brain regions. Human infants already show left-sided dominance in their brain activation when exposed to speech. Moreover, a memory-specific left-sided dominance in Wernicke’s area for speech perception has been demonstrated in 2.5-mo-old babies. It is possible that auditory-vocal learning is associated with hemispheric dominance and that this association arose in songbirds and humans through convergent evolution. Therefore, we investigated whether there is similar song memory-related lateralization in the songbird brain. We exposed male zebra finches to tutor or unfamiliar song. We found left-sided dominance of neuronal activation in a Broca-like brain region (HVC, a letter-based name) of juvenile and adult zebra finch males, independent of the song stimulus presented. In addition, juvenile males showed left-sided dominance for tutor song but not for unfamiliar song in a Wernicke-like brain region (the caudomedial nidopallium). Thus, left-sided dominance in the caudomedial nidopallium was specific for the song-learning phase and was memory-related. These findings demonstrate a remarkable neural parallel between birdsong and human spoken language, and they have important consequences for our understanding of the evolution of auditory-vocal learning and its neural mechanisms. PMID:22802637
ERIC Educational Resources Information Center
Vorperian, Houri K.; Wang, Shubing; Schimek, E. Michael; Durtschi, Reid B.; Kent, Ray D.; Gentry, Lindell R.; Chung, Moo K.
2011-01-01
Purpose: The anatomic origin for prepubertal vowel acoustic differences between male and female subjects remains unknown. The purpose of this study is to examine developmental sex differences in vocal tract (VT) length and its oral and pharyngeal portions. Method: Nine VT variables were measured from 605 imaging studies (magnetic resonance imaging…
Precise Motor Control Enables Rapid Flexibility in Vocal Behavior of Marmoset Monkeys.
Pomberger, Thomas; Risueno-Segovia, Cristina; Löschner, Julia; Hage, Steffen R
2018-03-05
Investigating the evolution of human speech is difficult and controversial because human speech surpasses nonhuman primate vocal communication in scope and flexibility [1-3]. Monkey vocalizations have been assumed to be largely innate, highly affective, and stereotyped for over 50 years [4, 5]. Recently, this perception has dramatically changed. Current studies have revealed distinct learning mechanisms during vocal development [6-8] and vocal flexibility, allowing monkeys to cognitively control when [9, 10], where [11], and what to vocalize [10, 12, 13]. However, specific call features (e.g., duration, frequency) remain surprisingly robust and stable in adult monkeys, resulting in rather stereotyped and discrete call patterns [14]. Additionally, monkeys seem to be unable to modulate their acoustic call structure under reinforced conditions beyond natural constraints [15, 16]. Behavioral experiments have shown that monkeys can stop sequences of calls immediately after acoustic perturbation but cannot interrupt ongoing vocalizations, suggesting that calls consist of single impartible pulses [17, 18]. Using acoustic perturbation triggered by the vocal behavior itself and quantitative measures of resulting vocal adjustments, we show that marmoset monkeys are capable of producing calls with durations beyond the natural boundaries of their repertoire by interrupting ongoing vocalizations rapidly after perturbation onset. Our results indicate that marmosets are capable of interrupting vocalizations only at periodic time points throughout calls, further supported by the occurrence of periodically segmented phees. These ideas overturn decades-old concepts on primate vocal pattern generation, indicating that vocalizations do not consist of one discrete call pattern but are built of many sequentially uttered units, like human speech. Copyright © 2018 The Author(s). Published by Elsevier Ltd.. All rights reserved.
Zinc sulfate therapy of vocal process granuloma.
Sun, Guang-Bin; Sun, Na; Tang, Hai-Hong; Zhu, Qiu-Bei; Wen, Wu; Zheng, Hong-Liang
2012-09-01
Vocal process granuloma is a benign lesion that occurs on the arytenoid cartilage. It tends to recur locally, and there is a great diversity of methods to treat it. Here, we reviewed the effects of zinc sulfate therapy program in 16 patients with vocal process granulomas. Eleven patients had a history of trauma or laryngeal intubation and five patients had unknown origin. Eleven had recurrence after one to three failed surgeries, and the others had no prior treatment. Symptoms included hoarseness, sore throat, lump sensation in the throat and cough that apparently improved. The granulomas did not recur for at least 1 year. No complications occurred. For vocal process granuloma, zinc sulfate therapy is good either as an initial or compensatory treatment.
Development of precursors to speech in infants exposed to two languages.
Oller, D K; Eilers, R E; Urbano, R; Cobo-Lewis, A B
1997-06-01
The study of bilingualism has often focused on two contradictory possibilities: that the learning of two languages may produce deficits of performance in each language by comparison with performance of monolingual individuals, or on the contrary, that the learning of two languages may produce linguistic or cognitive advantages with regard to the monolingual learning experience. The work reported here addressed the possibility that the very early bilingual experience of infancy may affect the unfolding of vocal precursors to speech. The results of longitudinal research with 73 infants aged 0;4 to 1;6 in monolingual and bilingual environments provided no support for either a bilingual deficit hypothesis nor for its opposite, a bilingual advantage hypothesis. Infants reared in bilingual and monolingual environments manifested similar ages of onset for canonical babbling (production of well-formed syllables), an event known to be fundamentally related to speech development. Further, quantitative measures of vocal performance (proportion of usage of well-formed syllables and vowel-like sounds) showed additional similarities between monolingual and bilingual infants. The similarities applied to infants of middle and low socio-economic status and to infants that were born at term or prematurely. The results suggest that vocal development in the first year of life is robust with respect to conditions of rearing. The biological foundations of speech appear to be such as to resist modifications in the natural schedule of vocal development.
Properties of vocalization- and gesture-combinations in the transition to first words.
Murillo, Eva; Capilla, Almudena
2016-07-01
Gestures and vocal elements interact from the early stages of language development, but the role of this interaction in the language learning process is not yet completely understood. The aim of this study is to explore gestural accompaniment's influence on the acoustic properties of vocalizations in the transition to first words. Eleven Spanish children aged 0;9 to 1;3 were observed longitudinally in a semi-structured play situation with an adult. Vocalizations were analyzed using several acoustic parameters based on those described by Oller et al. (2010). Results indicate that declarative vocalizations have fewer protosyllables than imperative ones, but only when they are produced with a gesture. Protosyllables duration and f(0) are more similar to those of mature speech when produced with pointing and declarative function than when produced with reaching gestures and imperative purposes. The proportion of canonical syllables produced increases with age, but only when combined with a gesture.
Yang, Anxiong; Berry, David A; Kaltenbacher, Manfred; Döllinger, Michael
2012-02-01
The human voice signal originates from the vibrations of the two vocal folds within the larynx. The interactions of several intrinsic laryngeal muscles adduct and shape the vocal folds to facilitate vibration in response to airflow. Three-dimensional vocal fold dynamics are extracted from in vitro hemilarynx experiments and fitted by a numerical three-dimensional-multi-mass-model (3DM) using an optimization procedure. In this work, the 3DM dynamics are optimized over 24 experimental data sets to estimate biomechanical vocal fold properties during phonation. Accuracy of the optimization is verified by low normalized error (0.13 ± 0.02), high correlation (83% ± 2%), and reproducible subglottal pressure values. The optimized, 3DM parameters yielded biomechanical variations in tissue properties along the vocal fold surface, including variations in both the local mass and stiffness of vocal folds. That is, both mass and stiffness increased along the superior-to-inferior direction. These variations were statistically analyzed under different experimental conditions (e.g., an increase in tension as a function of vocal fold elongation and an increase in stiffness and a decrease in mass as a function of glottal airflow). The study showed that physiologically relevant vocal fold tissue properties, which cannot be directly measured during in vivo human phonation, can be captured using this 3D-modeling technique. © 2012 Acoustical Society of America
Yang, Anxiong; Berry, David A.; Kaltenbacher, Manfred; Döllinger, Michael
2012-01-01
The human voice signal originates from the vibrations of the two vocal folds within the larynx. The interactions of several intrinsic laryngeal muscles adduct and shape the vocal folds to facilitate vibration in response to airflow. Three-dimensional vocal fold dynamics are extracted from in vitro hemilarynx experiments and fitted by a numerical three-dimensional-multi-mass-model (3DM) using an optimization procedure. In this work, the 3DM dynamics are optimized over 24 experimental data sets to estimate biomechanical vocal fold properties during phonation. Accuracy of the optimization is verified by low normalized error (0.13 ± 0.02), high correlation (83% ± 2%), and reproducible subglottal pressure values. The optimized, 3DM parameters yielded biomechanical variations in tissue properties along the vocal fold surface, including variations in both the local mass and stiffness of vocal folds. That is, both mass and stiffness increased along the superior-to-inferior direction. These variations were statistically analyzed under different experimental conditions (e.g., an increase in tension as a function of vocal fold elongation and an increase in stiffness and a decrease in mass as a function of glottal airflow). The study showed that physiologically relevant vocal fold tissue properties, which cannot be directly measured during in vivo human phonation, can be captured using this 3D-modeling technique. PMID:22352511
ERIC Educational Resources Information Center
Ivanova, Tamara N.; Gross, Christina; Mappus, Rudolph C.; Kwon, Yong Jun; Bassell, Gary J.; Liu, Robert C.
2017-01-01
Learning to recognize a stimulus category requires experience with its many natural variations. However, the mechanisms that allow a category's sensorineural representation to be updated after experiencing new exemplars are not well understood, particularly at the molecular level. Here we investigate how a natural vocal category induces expression…
The vocal repertoire of Tibetan macaques (Macaca thibetana): A quantitative classification.
Bernstein, Sofia K; Sheeran, Lori K; Wagner, R Steven; Li, Jin-Hua; Koda, Hiroki
2016-09-01
Vocal repertoires are basic and essential components for describing vocal communication in animals. Studying the entire suite of vocal signals aids investigations on the variation of acoustic structure across social contexts, comparisons on the complexity of communication systems across taxa, and in exploration of the evolutionary origins of species-specific vocalizations. Here, we describe the vocal repertoire of the largest species in the macaque genus, Macaca thibetana. We extracted thirty acoustic parameters from call recordings. Post hoc validation through quantitative analyses of the a priori repertoire classified eleven call types: coo, squawk, squeal, noisy scream, growl, bark, compound squeak, leap coo, weeping, modulated tonal scream, and pant. In comparison to the rest of the genus, Tibetan macaques uttered a wider array of vocalizations in the context of copulations. Previous reports did not include modulated tonal screams and pants during harassment of copulatory dyads. Furthermore, in comparison to the rest of the genus, Tibetan macaque females emit acoustically distinct copulation calls. The vocal repertoire of Tibetan macaques contributes to the literature on the emergence of species-specific calls in the genus Macaca with potential insights from social, reproductive, and ecological comparisons across species. Am. J. Primatol. 78:937-949, 2016. © 2016 Wiley Periodicals, Inc. © 2016 Wiley Periodicals, Inc.
Titze, Ingo R
2014-04-01
The origin of vocal registers has generally been attributed to differential activation of cricothyroid and thyroarytenoid muscles in the larynx. Register shifts, however, have also been shown to be affected by glottal pressures exerted on vocal fold surfaces, which can change with loudness, pitch, and vowel. Here it is shown computationally and with empirical data that intraglottal pressures can change abruptly when glottal adductory geometry is changed relatively smoothly from convergent to divergent. An intermediate shape between large convergence and large divergence, namely, a nearly rectangular glottal shape with almost parallel vocal fold surfaces, is associated with mixed registration. It can be less stable than either of the highly angular shapes unless transglottal pressure is reduced and upper stiffness of vocal fold tissues is balanced with lower stiffness. This intermediate state of adduction is desirable because it leads to a low phonation threshold pressure with moderate vocal fold collision. Achieving mixed registration consistently across wide ranges of F0, lung pressure, and vocal tract shapes appears to be a balancing act of coordinating laryngeal muscle activation with vocal tract pressures. Surprisingly, a large transglottal pressure is not facilitative in this process, exacerbating the bi-stable condition and the associated register contrast.
Bi-stable vocal fold adduction: A mechanism of modal-falsetto register shifts and mixed registration
Titze, Ingo R.
2014-01-01
The origin of vocal registers has generally been attributed to differential activation of cricothyroid and thyroarytenoid muscles in the larynx. Register shifts, however, have also been shown to be affected by glottal pressures exerted on vocal fold surfaces, which can change with loudness, pitch, and vowel. Here it is shown computationally and with empirical data that intraglottal pressures can change abruptly when glottal adductory geometry is changed relatively smoothly from convergent to divergent. An intermediate shape between large convergence and large divergence, namely, a nearly rectangular glottal shape with almost parallel vocal fold surfaces, is associated with mixed registration. It can be less stable than either of the highly angular shapes unless transglottal pressure is reduced and upper stiffness of vocal fold tissues is balanced with lower stiffness. This intermediate state of adduction is desirable because it leads to a low phonation threshold pressure with moderate vocal fold collision. Achieving mixed registration consistently across wide ranges of F0, lung pressure, and vocal tract shapes appears to be a balancing act of coordinating laryngeal muscle activation with vocal tract pressures. Surprisingly, a large transglottal pressure is not facilitative in this process, exacerbating the bi-stable condition and the associated register contrast. PMID:25235006
A humanized version of Foxp2 does not affect ultrasonic vocalization in adult mice.
Hammerschmidt, K; Schreiweis, C; Minge, C; Pääbo, S; Fischer, J; Enard, W
2015-11-01
The transcription factor FOXP2 has been linked to severe speech and language impairments in humans. An analysis of the evolution of the FOXP2 gene has identified two amino acid substitutions that became fixed after the split of the human and chimpanzee lineages. Studying the functional consequences of these two substitutions in the endogenous Foxp2 gene of mice showed alterations in dopamine levels, striatal synaptic plasticity, neuronal morphology and cortico-striatal-dependent learning. In addition, ultrasonic vocalizations (USVs) of pups had a significantly lower average pitch than control littermates. To which degree adult USVs would be affected in mice carrying the 'humanized' Foxp2 variant remained unclear. In this study, we analyzed USVs of 68 adult male mice uttered during repeated courtship encounters with different females. Mice carrying the Foxp2(hum/hum) allele did not differ significantly in the number of call elements, their element structure or in their element composition from control littermates. We conclude that neither the structure nor the usage of USVs in adult mice is affected by the two amino acid substitutions that occurred in FOXP2 during human evolution. The reported effect for pup vocalization thus appears to be transient. These results are in line with accumulating evidence that mouse USVs are hardly influenced by vocal learning. Hence, the function and evolution of genes that are necessary, but not sufficient for vocal learning in humans, must be either studied at a different phenotypic level in mice or in other organisms. © 2015 The Authors. Genes, Brain and Behavior published by International Behavioural and Neural Genetics Society and John Wiley & Sons Ltd.
A Bird’s Eye View of Human Language Evolution
Berwick, Robert C.; Beckers, Gabriël J. L.; Okanoya, Kazuo; Bolhuis, Johan J.
2012-01-01
Comparative studies of linguistic faculties in animals pose an evolutionary paradox: language involves certain perceptual and motor abilities, but it is not clear that this serves as more than an input–output channel for the externalization of language proper. Strikingly, the capability for auditory–vocal learning is not shared with our closest relatives, the apes, but is present in such remotely related groups as songbirds and marine mammals. There is increasing evidence for behavioral, neural, and genetic similarities between speech acquisition and birdsong learning. At the same time, researchers have applied formal linguistic analysis to the vocalizations of both primates and songbirds. What have all these studies taught us about the evolution of language? Is the comparative study of an apparently species-specific trait like language feasible? We argue that comparative analysis remains an important method for the evolutionary reconstruction and causal analysis of the mechanisms underlying language. On the one hand, common descent has been important in the evolution of the brain, such that avian and mammalian brains may be largely homologous, particularly in the case of brain regions involved in auditory perception, vocalization, and auditory memory. On the other hand, there has been convergent evolution of the capacity for auditory–vocal learning, and possibly for structuring of external vocalizations, such that apes lack the abilities that are shared between songbirds and humans. However, significant limitations to this comparative analysis remain. While all birdsong may be classified in terms of a particularly simple kind of concatenation system, the regular languages, there is no compelling evidence to date that birdsong matches the characteristic syntactic complexity of human language, arising from the composition of smaller forms like words and phrases into larger ones. PMID:22518103
Scattoni, Maria Luisa; Crawley, Jacqueline; Ricceri, Laura
2009-01-01
In neonatal mice ultrasonic vocalizations have been studied both as an early communicative behavior of the pup-mother dyad and as a sign of an aversive affective state. Adult mice of both sexes produce complex ultrasonic vocalization patterns in different experimental/social contexts. All these vocalizations are becoming an increasingly valuable assay for behavioral phenotyping throughout the mouse life-span and alterations of the ultrasound patterns have been reported in several mouse models of neurodevelopmental disorders. Here we also show that the modulation of vocalizations by maternal cues (maternal potentiation paradigm) – originally identified and investigated in rats - can be measured in C57Bl/6 mouse pups with appropriate modifications of the rat protocol and can likely be applied to mouse behavioral phenotyping. In addition we suggest that a detailed qualitative evaluation of neonatal calls together with analysis of adult mouse vocalization patterns in both sexes in social settings, may lead to a greater understanding of the communication value of vocalizations in mice. Importantly, both neonatal and adult USV altered patterns can be determined during the behavioural phenotyping of mouse models of human neurodevelopmental and neuropsychiatric disorders, starting from those in which deficits in communication are a primary symptom. PMID:18771687
Murugan, Malavika; Harward, Stephen; Scharff, Constance; Mooney, Richard
2013-12-18
Mutations of the FOXP2 gene impair speech and language development in humans and shRNA-mediated suppression of the avian ortholog FoxP2 disrupts song learning in juvenile zebra finches. How diminished FoxP2 levels affect vocal control and alter the function of neural circuits important to learned vocalizations remains unclear. Here we show that FoxP2 knockdown in the songbird striatum disrupts developmental and social modulation of song variability. Recordings in anesthetized birds show that FoxP2 knockdown interferes with D1R-dependent modulation of activity propagation in a corticostriatal pathway important to song variability, an effect that may be partly attributable to reduced D1R and DARPP-32 protein levels. Furthermore, recordings in singing birds reveal that FoxP2 knockdown prevents social modulation of singing-related activity in this pathway. These findings show that reduced FoxP2 levels interfere with the dopaminergic modulation of vocal variability, which may impede song and speech development by disrupting reinforcement learning mechanisms. Copyright © 2013 Elsevier Inc. All rights reserved.
Murugan, Malavika; Harward, Stephen; Scharff, Constance; Mooney, Richard
2013-01-01
Summary Mutations of the FOXP2 gene impair speech and language development in humans and shRNA-mediated suppression of the avian orthologue FoxP2 disrupts song learning in juvenile zebra finches. How diminished FoxP2 levels affect vocal control and alter the function of neural circuits important to learned vocalizations remains unclear. Here we show that FoxP2 knockdown in the songbird striatum disrupts developmental and social modulation of song variability. Recordings in anaesthetized birds show that FoxP2 knockdown interferes with D1R-dependent modulation of activity propagation in a corticostriatal pathway important to song variability, an effect that may be partly attributable to reduced D1R and DARPP-32 protein levels. Furthermore, recordings in singing birds reveal that FoxP2 knockdown prevents social modulation of singing-related activity in this pathway. These findings show that reduced FoxP2 levels interfere with the dopaminergic modulation of vocal variability, which may impede song and speech development by disrupting reinforcement learning mechanisms. PMID:24268418
Context-dependent vocal mimicry in a passerine bird.
Goodale, Eben; Kotagama, Sarath W
2006-04-07
How do birds select the sounds they mimic, and in what contexts do they use vocal mimicry? Some birds show a preference for mimicking other species' alarm notes, especially in situations when they appear to be alarmed. Yet no study has demonstrated that birds change the call types they mimic with changing contexts. We found that greater racket-tailed drongos (Dicrurus paradiseus) in the rainforest of Sri Lanka mimic the calls of predators and the alarm-associated calls of other species more often than would be expected from the frequency of these sounds in the acoustic environment. Drongos include this alarm-associated mimicry in their own alarm vocalizations, while incorporating other species' songs and contact calls in their own songs. Drongos show an additional level of context specificity by mimicking other species' ground predator-specific call types when mobbing. We suggest that drongos learn other species' calls and their contexts while interacting with these species in mixed flocks. The drongos' behaviour demonstrates that alarm-associated calls can have learned components, and that birds can learn the appropriate usage of calls that encode different types of information.
Context-dependent vocal mimicry in a passerine bird
Goodale, Eben; Kotagama, Sarath W
2005-01-01
How do birds select the sounds they mimic, and in what contexts do they use vocal mimicry? Some birds show a preference for mimicking other species' alarm notes, especially in situations when they appear to be alarmed. Yet no study has demonstrated that birds change the call types they mimic with changing contexts. We found that greater racket-tailed drongos (Dicrurus paradiseus) in the rainforest of Sri Lanka mimic the calls of predators and the alarm-associated calls of other species more often than would be expected from the frequency of these sounds in the acoustic environment. Drongos include this alarm-associated mimicry in their own alarm vocalizations, while incorporating other species' songs and contact calls in their own songs. Drongos show an additional level of context specificity by mimicking other species' ground predator-specific call types when mobbing. We suggest that drongos learn other species' calls and their contexts while interacting with these species in mixed flocks. The drongos' behaviour demonstrates that alarm-associated calls can have learned components, and that birds can learn the appropriate usage of calls that encode different types of information. PMID:16618682
Vocal Fry Use in Adult Female Speakers Exposed to Two Languages.
Gibson, Todd A; Summers, Connie; Walls, Sydney
2017-07-01
Several studies have identified the widespread use of vocal fry among American women. Popular explanations for this phenomenon appeal to sociolinguistic purposes that likely take significant time for second language users to learn. The objective of this study was to determine if mere exposure to this vocal register, as opposed to nuanced sociolinguistic motivations, might explain its widespread use. This study used multigroup within- and between-subjects design. Fifty-eight women from one of three language background groups (functionally monolingual in English, functionally monolingual in Spanish, and Spanish-English bilinguals) living in El Paso, Texas, repeated a list of nonwords conforming to the sound rules of English and another list of nonwords conforming to the sound rules of Spanish. Perceptual analysis identified each episode of vocal fry. There were no statistically significant differences between groups in their frequency of vocal fry use despite large differences in their amount of English-language exposure. All groups produced more vocal fry when repeating English than when repeating Spanish nonwords. Because the human perceptual system encodes for vocal qualities even after minimal language experience, the widespread use of vocal fry among female residents in the United States likely is owing to mere exposure to English rather than nuanced sociolinguistic motivations. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Anyanwu, G E; Nto, J N; Agu, A U; Ekezie, J; Esom, E A
2016-11-01
Background music has been reported to enhance learning in the cadaver dissection laboratory. This study was designed to determine the impact of various forms of musical genre and some of their characteristics on students' learning outcome in the dissection laboratory. Some selected musical genre in vocal and non-vocal forms and at different tempi and volume were played as background music (BM) to 253 Medical and Dental students during various sessions of cadaver dissection. Psychological Stress assessment was done using Psychological stress measure-9. Participants love for music, preferred musical genre and other musical characteristics were assessed. The impact of the various musical genre and their characteristics on learning was done via written examination on the region dissected during each musical session. A positive relationship was noted between students' preference for musical genre during leisure with their preference for BM during private study time (P<0.01). Statistically significant differences (P<0.01) were established in the impacts of the selected musical genre on some selected learning factors. Country and Classical music gave the highest positive impact on the various learning factors in CDL followed by R&B. No significant difference was noted between the cognitive values of vocal and non-vocal music. Classical music most effectively reduced the stress induced by dissection in the CDL while Reggae and High life musical genre created a more stressful environment than regular background noise (P<0.01). Moderate volume level and Tempo were most preferred during both cadaver dissection activity and leisure hours. This study shows statistically significant differences in the cognitive values of some of the studied musical genre and their various characteristics. The inability to isolate the particular musical genre with these desired properties could account for the controversies in the reports of the role of music in academic environment. Copyright © 2016 Elsevier GmbH. All rights reserved.
Early Musical Training in Bel Canto Vocal Technique: A Brief History and Philosophy.
ERIC Educational Resources Information Center
Gerstein, Christine Wondolowski
This paper offers a brief history and philosophy of the origins of bel canto vocal style and describes the pedagogical methods used to achieve bel canto ideals in singing. The document discusses the adoption and development of this technique and how it developed over long periods of preparation in the foregoing centuries before the Baroque era.…
Geographical variation of St. Lucia Parrot flight vocalizations
Kleeman, Patrick M.; Gilardi, James D.
2005-01-01
Parrots are vocal learners and many species of parrots are capable of learning new calls, even as adults. This capability gives parrots the potential to develop communication systems that can vary dramatically over space. St. Lucia Parrot (Amazona versicolor) flight vocalizations were examined for geographic variation between four different sites on the island of St. Lucia. Spectrographic cross-correlation analysis of a commonly used flight vocalization, the p-chow call, demonstrated quantitative differences between sites. Additionally, the similarity of p-chows decreased as the distance between sites increased. Flight call repertoires also differed among sites; parrots at the Des Bottes and Quilesse sites each used one flight call unique to those sites, while parrots at the Barre de L'Isle site used a flight call that Quilesse parrots gave only while perched. It is unclear whether the vocal variation changed clinally with distance, or whether there were discrete dialect boundaries as in a congener, the Yellow-naped Parrot (Amazona auropalliata, Wright 1996). The geographical scale over which the St. Lucia Parrot's vocal variation occurred was dramatically smaller than that of the Yellow-naped Parrot. Similar patterns of fine-scale vocal variation may be more widespread among other parrot species in the Caribbean than previously documented.
Online contributions of auditory feedback to neural activity in avian song control circuitry
Sakata, Jon T.; Brainard, Michael S.
2008-01-01
Birdsong, like human speech, relies critically on auditory feedback to provide information about the quality of vocalizations. Although the importance of auditory feedback to vocal learning is well established, whether and how feedback signals influence vocal premotor circuitry has remained obscure. Previous studies in singing birds have not detected changes to vocal premotor activity following perturbations of auditory feedback, leading to the hypothesis that contributions of feedback to vocal plasticity might rely on ‘offline’ processing. Here, we recorded single and multi-unit activity in the premotor nucleus HVC of singing Bengalese finches in response to feedback perturbations that are known to drive plastic changes in song. We found that transient feedback perturbation caused reliable decreases in HVC activity at short latencies (20-80 ms). Similar changes to HVC activity occurred in awake, non-singing finches when the bird’s own song was played back with auditory perturbations that simulated those experienced by singing birds. These data indicate that neurons in avian vocal premotor circuitry are rapidly influenced by perturbations of auditory feedback and support the possibility that feedback information in HVC contributes online to the production and plasticity of vocalizations. PMID:18971480
Rules and mechanisms for efficient two-stage learning in neural circuits.
Teşileanu, Tiberiu; Ölveczky, Bence; Balasubramanian, Vijay
2017-04-04
Trial-and-error learning requires evaluating variable actions and reinforcing successful variants. In songbirds, vocal exploration is induced by LMAN, the output of a basal ganglia-related circuit that also contributes a corrective bias to the vocal output. This bias is gradually consolidated in RA, a motor cortex analogue downstream of LMAN. We develop a new model of such two-stage learning. Using stochastic gradient descent, we derive how the activity in 'tutor' circuits ( e.g., LMAN) should match plasticity mechanisms in 'student' circuits ( e.g., RA) to achieve efficient learning. We further describe a reinforcement learning framework through which the tutor can build its teaching signal. We show that mismatches between the tutor signal and the plasticity mechanism can impair learning. Applied to birdsong, our results predict the temporal structure of the corrective bias from LMAN given a plasticity rule in RA. Our framework can be applied predictively to other paired brain areas showing two-stage learning.
Modeling coupled aerodynamics and vocal fold dynamics using immersed boundary methods.
Duncan, Comer; Zhai, Guangnian; Scherer, Ronald
2006-11-01
The penalty immersed boundary (PIB) method, originally introduced by Peskin (1972) to model the function of the mammalian heart, is tested as a fluid-structure interaction model of the closely coupled dynamics of the vocal folds and aerodynamics in phonation. Two-dimensional vocal folds are simulated with material properties chosen to result in self-oscillation and volume flows in physiological frequency ranges. Properties of the glottal flow field, including vorticity, are studied in conjunction with the dynamic vocal fold motion. The results of using the PIB method to model self-oscillating vocal folds for the case of 8 cm H20 as the transglottal pressure gradient are described. The volume flow at 8 cm H20, the transglottal pressure, and vortex dynamics associated with the self-oscillating model are shown. Volume flow is also given for 2, 4, and 12 cm H2O, illustrating the robustness of the model to a range of transglottal pressures. The results indicate that the PIB method applied to modeling phonation has good potential for the study of the interdependence of aerodynamics and vocal fold motion.
Origin of symbol-using systems: speech, but not sign, without the semantic urge
Sereno, Martin I.
2014-01-01
Natural language—spoken and signed—is a multichannel phenomenon, involving facial and body expression, and voice and visual intonation that is often used in the service of a social urge to communicate meaning. Given that iconicity seems easier and less abstract than making arbitrary connections between sound and meaning, iconicity and gesture have often been invoked in the origin of language alongside the urge to convey meaning. To get a fresh perspective, we critically distinguish the origin of a system capable of evolution from the subsequent evolution that system becomes capable of. Human language arose on a substrate of a system already capable of Darwinian evolution; the genetically supported uniquely human ability to learn a language reflects a key contact point between Darwinian evolution and language. Though implemented in brains generated by DNA symbols coding for protein meaning, the second higher-level symbol-using system of language now operates in a world mostly decoupled from Darwinian evolutionary constraints. Examination of Darwinian evolution of vocal learning in other animals suggests that the initial fixation of a key prerequisite to language into the human genome may actually have required initially side-stepping not only iconicity, but the urge to mean itself. If sign languages came later, they would not have faced this constraint. PMID:25092671
Machine Learning Algorithms for Automatic Classification of Marmoset Vocalizations
Ribeiro, Sidarta; Pereira, Danillo R.; Papa, João P.; de Albuquerque, Victor Hugo C.
2016-01-01
Automatic classification of vocalization type could potentially become a useful tool for acoustic the monitoring of captive colonies of highly vocal primates. However, for classification to be useful in practice, a reliable algorithm that can be successfully trained on small datasets is necessary. In this work, we consider seven different classification algorithms with the goal of finding a robust classifier that can be successfully trained on small datasets. We found good classification performance (accuracy > 0.83 and F1-score > 0.84) using the Optimum Path Forest classifier. Dataset and algorithms are made publicly available. PMID:27654941
Moving to the Beat and Singing are Linked in Humans
Dalla Bella, Simone; Berkowska, Magdalena; Sowiński, Jakub
2015-01-01
The abilities to sing and to move to the beat of a rhythmic auditory stimulus emerge early during development, and both engage perceptual, motor, and sensorimotor processes. These similarities between singing and synchronization to a beat may be rooted in biology. Patel (2008) has suggested that motor synchronization to auditory rhythms may have emerged during evolution as a byproduct of selection for vocal learning (“vocal learning and synchronization hypothesis”). This view predicts a strong link between vocal performance and synchronization skills in humans. Here, we tested this prediction by asking occasional singers to tap along with auditory pulse trains and to imitate familiar melodies. Both vocal imitation and synchronization skills were measured in terms of accuracy and precision or consistency. Accurate and precise singers tapped more in the vicinity of the pacing stimuli (i.e., they were more accurate) than less accurate and less precise singers. Moreover, accurate singers were more consistent when tapping to the beat. These differences cannot be ascribed to basic motor skills or to motivational factors. Individual differences in terms of singing proficiency and synchronization skills may reflect the variability of a shared sensorimotor translation mechanism. PMID:26733370
Predictive and tempo-flexible synchronization to a visual metronome in monkeys.
Takeya, Ryuji; Kameda, Masashi; Patel, Aniruddh D; Tanaka, Masaki
2017-07-21
Predictive and tempo-flexible synchronization to an auditory beat is a fundamental component of human music. To date, only certain vocal learning species show this behaviour spontaneously. Prior research training macaques (vocal non-learners) to tap to an auditory or visual metronome found their movements to be largely reactive, not predictive. Does this reflect the lack of capacity for predictive synchronization in monkeys, or lack of motivation to exhibit this behaviour? To discriminate these possibilities, we trained monkeys to make synchronized eye movements to a visual metronome. We found that monkeys could generate predictive saccades synchronized to periodic visual stimuli when an immediate reward was given for every predictive movement. This behaviour generalized to novel tempi, and the monkeys could maintain the tempo internally. Furthermore, monkeys could flexibly switch from predictive to reactive saccades when a reward was given for each reactive response. In contrast, when humans were asked to make a sequence of reactive saccades to a visual metronome, they often unintentionally generated predictive movements. These results suggest that even vocal non-learners may have the capacity for predictive and tempo-flexible synchronization to a beat, but that only certain vocal learning species are intrinsically motivated to do it.
Drinking Songs: Alcohol Effects on Learned Song of Zebra Finches
Olson, Christopher R.; Owen, Devin C.; Ryabinin, Andrey E.; Mello, Claudio V.
2014-01-01
Speech impairment is one of the most intriguing and least understood effects of alcohol on cognitive function, largely due to the lack of data on alcohol effects on vocalizations in the context of an appropriate experimental model organism. Zebra finches, a representative songbird and a premier model for understanding the neurobiology of vocal production and learning, learn song in a manner analogous to how humans learn speech. Here we show that when allowed access, finches readily drink alcohol, increase their blood ethanol concentrations (BEC) significantly, and sing a song with altered acoustic structure. The most pronounced effects were decreased amplitude and increased entropy, the latter likely reflecting a disruption in the birds’ ability to maintain the spectral structure of song under alcohol. Furthermore, specific syllables, which have distinct acoustic structures, were differentially influenced by alcohol, likely reflecting a diversity in the neural mechanisms required for their production. Remarkably, these effects on vocalizations occurred without overt effects on general behavioral measures, and importantly, they occurred within a range of BEC that can be considered risky for humans. Our results suggest that the variable effects of alcohol on finch song reflect differential alcohol sensitivity of the brain circuitry elements that control different aspects of song production. They also point to finches as an informative model for understanding how alcohol affects the neuronal circuits that control the production of learned motor behaviors. PMID:25536524
The neural network classification of false killer whale (Pseudorca crassidens) vocalizations.
Murray, S O; Mercado, E; Roitblat, H L
1998-12-01
This study reports the use of unsupervised, self-organizing neural network to categorize the repertoire of false killer whale vocalizations. Self-organizing networks are capable of detecting patterns in their input and partitioning those patterns into categories without requiring that the number or types of categories be predefined. The inputs for the neural networks were two-dimensional characterization of false killer whale vocalization, where each vocalization was characterized by a sequence of short-time measurements of duty cycle and peak frequency. The first neural network used competitive learning, where units in a competitive layer distributed themselves to recognize frequently presented input vectors. This network resulted in classes representing typical patterns in the vocalizations. The second network was a Kohonen feature map which organized the outputs topologically, providing a graphical organization of pattern relationships. The networks performed well as measured by (1) the average correlation between the input vectors and the weight vectors for each category, and (2) the ability of the networks to classify novel vocalizations. The techniques used in this study could easily be applied to other species and facilitate the development of objective, comprehensive repertoire models.
Tsoi, Shuk C; Aiya, Utsav V; Wasner, Kobi D; Phan, Mimi L; Pytte, Carolyn L; Vicario, David S
2014-01-01
Many brain regions exhibit lateral differences in structure and function, and also incorporate new neurons in adulthood, thought to function in learning and in the formation of new memories. However, the contribution of new neurons to hemispheric differences in processing is unknown. The present study combines cellular, behavioral, and physiological methods to address whether 1) new neuron incorporation differs between the brain hemispheres, and 2) the degree to which hemispheric lateralization of new neurons correlates with behavioral and physiological measures of learning and memory. The songbird provides a model system for assessing the contribution of new neurons to hemispheric specialization because songbird brain areas for vocal processing are functionally lateralized and receive a continuous influx of new neurons in adulthood. In adult male zebra finches, we quantified new neurons in the caudomedial nidopallium (NCM), a forebrain area involved in discrimination and memory for the complex vocalizations of individual conspecifics. We assessed song learning and recorded neural responses to song in NCM. We found significantly more new neurons labeled in left than in right NCM; moreover, the degree of asymmetry in new neuron numbers was correlated with the quality of song learning and strength of neuronal memory for recently heard songs. In birds with experimentally impaired song quality, the hemispheric difference in new neurons was diminished. These results suggest that new neurons may contribute to an allocation of function between the hemispheres that underlies the learning and processing of complex signals.
Wasner, Kobi D.; Phan, Mimi L.; Pytte, Carolyn L.; Vicario, David S.
2014-01-01
Many brain regions exhibit lateral differences in structure and function, and also incorporate new neurons in adulthood, thought to function in learning and in the formation of new memories. However, the contribution of new neurons to hemispheric differences in processing is unknown. The present study combines cellular, behavioral, and physiological methods to address whether 1) new neuron incorporation differs between the brain hemispheres, and 2) the degree to which hemispheric lateralization of new neurons correlates with behavioral and physiological measures of learning and memory. The songbird provides a model system for assessing the contribution of new neurons to hemispheric specialization because songbird brain areas for vocal processing are functionally lateralized and receive a continuous influx of new neurons in adulthood. In adult male zebra finches, we quantified new neurons in the caudomedial nidopallium (NCM), a forebrain area involved in discrimination and memory for the complex vocalizations of individual conspecifics. We assessed song learning and recorded neural responses to song in NCM. We found significantly more new neurons labeled in left than in right NCM; moreover, the degree of asymmetry in new neuron numbers was correlated with the quality of song learning and strength of neuronal memory for recently heard songs. In birds with experimentally impaired song quality, the hemispheric difference in new neurons was diminished. These results suggest that new neurons may contribute to an allocation of function between the hemispheres that underlies the learning and processing of complex signals. PMID:25251077
Differential coding of conspecific vocalizations in the ventral auditory cortical stream.
Fukushima, Makoto; Saunders, Richard C; Leopold, David A; Mishkin, Mortimer; Averbeck, Bruno B
2014-03-26
The mammalian auditory cortex integrates spectral and temporal acoustic features to support the perception of complex sounds, including conspecific vocalizations. Here we investigate coding of vocal stimuli in different subfields in macaque auditory cortex. We simultaneously measured auditory evoked potentials over a large swath of primary and higher order auditory cortex along the supratemporal plane in three animals chronically using high-density microelectrocorticographic arrays. To evaluate the capacity of neural activity to discriminate individual stimuli in these high-dimensional datasets, we applied a regularized multivariate classifier to evoked potentials to conspecific vocalizations. We found a gradual decrease in the level of overall classification performance along the caudal to rostral axis. Furthermore, the performance in the caudal sectors was similar across individual stimuli, whereas the performance in the rostral sectors significantly differed for different stimuli. Moreover, the information about vocalizations in the caudal sectors was similar to the information about synthetic stimuli that contained only the spectral or temporal features of the original vocalizations. In the rostral sectors, however, the classification for vocalizations was significantly better than that for the synthetic stimuli, suggesting that conjoined spectral and temporal features were necessary to explain differential coding of vocalizations in the rostral areas. We also found that this coding in the rostral sector was carried primarily in the theta frequency band of the response. These findings illustrate a progression in neural coding of conspecific vocalizations along the ventral auditory pathway.
Differential Coding of Conspecific Vocalizations in the Ventral Auditory Cortical Stream
Saunders, Richard C.; Leopold, David A.; Mishkin, Mortimer; Averbeck, Bruno B.
2014-01-01
The mammalian auditory cortex integrates spectral and temporal acoustic features to support the perception of complex sounds, including conspecific vocalizations. Here we investigate coding of vocal stimuli in different subfields in macaque auditory cortex. We simultaneously measured auditory evoked potentials over a large swath of primary and higher order auditory cortex along the supratemporal plane in three animals chronically using high-density microelectrocorticographic arrays. To evaluate the capacity of neural activity to discriminate individual stimuli in these high-dimensional datasets, we applied a regularized multivariate classifier to evoked potentials to conspecific vocalizations. We found a gradual decrease in the level of overall classification performance along the caudal to rostral axis. Furthermore, the performance in the caudal sectors was similar across individual stimuli, whereas the performance in the rostral sectors significantly differed for different stimuli. Moreover, the information about vocalizations in the caudal sectors was similar to the information about synthetic stimuli that contained only the spectral or temporal features of the original vocalizations. In the rostral sectors, however, the classification for vocalizations was significantly better than that for the synthetic stimuli, suggesting that conjoined spectral and temporal features were necessary to explain differential coding of vocalizations in the rostral areas. We also found that this coding in the rostral sector was carried primarily in the theta frequency band of the response. These findings illustrate a progression in neural coding of conspecific vocalizations along the ventral auditory pathway. PMID:24672012
An agent-based model of dialect evolution in killer whales.
Filatova, Olga A; Miller, Patrick J O
2015-05-21
The killer whale is one of the few animal species with vocal dialects that arise from socially learned group-specific call repertoires. We describe a new agent-based model of killer whale populations and test a set of vocal-learning rules to assess which mechanisms may lead to the formation of dialect groupings observed in the wild. We tested a null model with genetic transmission and no learning, and ten models with learning rules that differ by template source (mother or matriline), variation type (random errors or innovations) and type of call change (no divergence from kin vs. divergence from kin). The null model without vocal learning did not produce the pattern of group-specific call repertoires we observe in nature. Learning from either mother alone or the entire matriline with calls changing by random errors produced a graded distribution of the call phenotype, without the discrete call types observed in nature. Introducing occasional innovation or random error proportional to matriline variance yielded more or less discrete and stable call types. A tendency to diverge from the calls of related matrilines provided fast divergence of loose call clusters. A pattern resembling the dialect diversity observed in the wild arose only when rules were applied in combinations and similar outputs could arise from different learning rules and their combinations. Our results emphasize the lack of information on quantitative features of wild killer whale dialects and reveal a set of testable questions that can draw insights into the cultural evolution of killer whale dialects. Copyright © 2015 Elsevier Ltd. All rights reserved.
Soderstrom, Ken; Tian, Qiyu
2008-01-01
CB1 cannabinoid receptors are distinctly expressed at high density within several regions of zebra finch telencephalon including those known to be involved in song learning (lMAN and Area X) and production (HVC and RA). Because: (1) exposure to cannabinoid agonists during developmental periods of auditory and sensory-motor song learning alters song patterns produced later in adulthood and; (2) densities of song region expression of CB1 waxes-and-wanes during song learning, it is becoming clear that CB1 receptor-mediated signaling is important to normal processes of vocal development. To better understand mechanisms involved in cannabinoid modulation of vocal behavior we have investigated the dose-response relationship between systemic cannabinoid exposure and changes in neuronal activity (as indicated by expression of the transcription factor, c-Fos) within telencephalic brain regions with established involvement in song learning and/or control. In adults we have found that low doses (0.1 mg/kg) of the cannabinoid agonist WIN-55212-2 decrease neuronal activity (as indicated by densities of c-fos-expressing nuclei) within vocal motor regions of caudal telencephalon (HVC and RA) while higher doses (3 mg/kg) stimulate activity. Both effects were reversed by pretreatment with the CB1-selective antagonist rimonabant. Interestingly, no effects of cannabinoid treatment were observed within the rostral song regions lMAN and Area X, despite distinct and dense CB1 receptor expression within these areas. Overall, our results demonstrate that, depending on dosage, CB1 agonism can both inhibit and stimulate neuronal activity within brain regions controlling adult vocal motor output, implicating involvement of multiple CB1-sensitive neuronal circuits. PMID:18509622
A novel model for examining recovery of phonation after vocal nerve damage.
Bhama, Prabhat K; Hillel, Allen D; Merati, Albert L; Perkel, David J
2011-05-01
Recurrent laryngeal nerve injury remains a dominant clinical issue in laryngology. To date, no animal model of laryngeal reinnervation has offered an outcome measure that can reflect the degree of recovery based on vocal function. We present an avian model system for studying recovery of learned vocalizations after nerve injury. Prospective animal study. Digital recordings of bird song were made from 11 adult male zebra finches; nine birds underwent bilateral crushing of the nerve supplying the vocal organ, and two birds underwent sham surgery. Songs from all the birds were then recorded regularly and analyzed based on temporal and spectral characteristics using computer software. Indices were calculated to indicate the degree of similarity between preoperative and postoperative song. Nerve crush caused audible differences in song quality and significant drops (P<0.05) in measured spectral and, to a lesser degree, temporal indices. Spectral indices recovered significantly (mean=43.0%; standard deviation [SD]=40.7; P<0.02), and there was an insignificant trend toward recovery of temporal index (mean=28.0%; SD=41.4; P=0.0771). In five of the nine (56%) birds, there was a greater than 50% recovery of spectral indices within a 4-week period. Two birds exhibited substantially less recovery of spectral indices and two birds had a persistent decline in spectral indices. Recovery of temporal index was highly variable as well, ranging from persistent further declines of 45.1% to recovery of 87%. Neither sham bird exhibited significant (P>0.05) differences in song after nerve crush. The songbird model system allows functional analysis of learned vocalization after surgical damage to vocal nerves. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
Vocal specialization through tracheal elongation in an extinct Miocene pheasant from China.
Li, Zhiheng; Clarke, Julia A; Eliason, Chad M; Stidham, Thomas A; Deng, Tao; Zhou, Zhonghe
2018-05-25
Modifications to the upper vocal tract involving hyper-elongated tracheae have evolved many times within crown birds, and their evolution has been linked to a 'size exaggeration' hypothesis in acoustic signaling and communication, whereby smaller-sized birds can produce louder sounds. A fossil skeleton of a new extinct species of wildfowl (Galliformes: Phasianidae) from the late Miocene of China, preserves an elongated, coiled trachea that represents the oldest fossil record of this vocal modification in birds and the first documentation of its evolution within pheasants. The phylogenetic position of this species within Phasianidae has not been fully resolved, but appears to document a separate independent origination of this vocal modification within Galliformes. The fossil preserves a coiled section of the trachea and other remains supporting a tracheal length longer than the bird's body. This extinct species likely produced vocalizations with a lower fundamental frequency and reduced harmonics compared to similarly-sized pheasants. The independent evolution of this vocal feature in galliforms living in both open and closed habitats does not appear to be correlated with other factors of biology or its open savanna-like habitat. Features present in the fossil that are typically associated with sexual dimorphism suggest that sexual selection may have resulted in the evolution of both the morphology and vocalization mechanism in this extinct species.
Heinrich, J E; Nordeen, K W; Nordeen, E J
2005-03-01
Several instances of early learning coincide with significant rearrangements of neural connections in regions contributing to these behaviors. In fact developmentally restricted learning may be constrained temporally by the opportunity for experience to selectively maintain appropriate synapses amidst the elimination of exuberant connections. Consistent with this notion, during the normal sensitive period for vocal learning in zebra finches (Taenopygia guttata), there is a decline in the density of dendritic spines within a region essential for song development, the lateral magnocellular nucleus of the anterior nidopallium (lMAN). Moreover, in birds isolated from conspecific song shortly after hatching, both the closure of the sensitive period for vocal learning and the pruning of spines from lMAN neurons is delayed. Here, we employed a more subtle form of deprivation to delay the close of the sensitive period for song learning, and found that late song learning occurred without obvious alterations in the pruning of dendritic spines on lMAN neurons. At posthatch day (PHD) 65 (beyond the end of the normal sensitive period for song memorization in zebra finches), birds isolated from song beginning on PHD30 did not differ from normally reared birds in measures of dendritic spine density on Golgi-Cox stained lMAN neurons. Moreover, tutor exposure from PHD65 to 90 did not increase spine elimination in these isolates (who memorized new song material) relative to controls (who did not). Thus, we conclude that the extent of normally occurring lMAN spine loss is not sufficient to account for the timing of the sensitive period for zebra finch song learning.
Short bouts of vocalization induce long lasting fast gamma oscillations in a sensorimotor nucleus
Lewandowski, Brian; Schmidt, Marc
2011-01-01
Performance evaluation is a critical feature of motor learning. In the vocal system, it requires the integration of auditory feedback signals with vocal motor commands. The network activity that supports such integration is unknown, but it has been proposed that vocal performance evaluation occurs offline. Recording from NIf, a sensorimotor structure in the avian song system, we show that short bouts of singing in adult male zebra finches (Taeniopygia guttata) induce persistent increases in firing activity and coherent oscillations in the fast gamma range (90–150 Hz). Single units are strongly phase-locked to these oscillations, which can last up to 30 s, often outlasting vocal activity by an order of magnitude. In other systems, oscillations often are triggered by events or behavioral tasks but rarely outlast the event that triggered them by more than 1 second. The present observations are the longest reported gamma oscillations triggered by an isolated behavioral event. In mammals, gamma oscillations have been associated with memory consolidation and are hypothesized to facilitate communication between brain regions. We suggest that the timing and persistent nature of NIf’s fast gamma oscillations make them well suited to facilitate the integration of auditory and vocal motor traces associated with vocal performance evaluation. PMID:21957255
Pérouse, R; Coulombeau, B; Arias, C; Casanova, C
2006-01-01
In patients presenting a bilateral laryngeal immobility, the potential reversibility of certain cases, the refusal or the bad tolerance of long term tracheotomy bring up the question of the choice of the surgical technique if it is indicated. To report our experience with the technique of lateralization of the paralyzed vocal fold (arytenoidopexy) suggested by Lichtenberger. After having described the technique, we report 5 cases (3 pos-thyroidectomy, I of central origine, 1 post-burn). From 1 to 12 months after surgery 2 patients were fully satisfied, a patient (central origin) recovered spontaneously at the end of a month and the 2 last had a partial result. Only one patient required several surgical gestures. The Lichtenberger's technique combines theoretical reversibility and conservation of a functional glottic plan. It avoids tracheotomy. This approach can according to us validly replace the traditional techniques, medium or long term tracheotomy , or endoscopic arythenoid or posterior vocal fold resection.
2013-01-01
Background A fundamental question in molecular neurobiology is how genes that determine basic neuronal properties shape the functional organization of brain circuits underlying complex learned behaviors. Given the growing availability of complete vertebrate genomes, comparative genomics represents a promising approach to address this question. Here we used genomics and molecular approaches to study how ion channel genes influence the properties of the brain circuitry that regulates birdsong, a learned vocal behavior with important similarities to human speech acquisition. We focused on potassium (K-)Channels, which are major determinants of neuronal cell excitability. Starting with the human gene set of K-Channels, we used cross-species mRNA/protein alignments, and syntenic analysis to define the full complement of orthologs, paralogs, allelic variants, as well as novel loci not previously predicted in the genome of zebra finch (Taeniopygia guttata). We also compared protein coding domains in chicken and zebra finch orthologs to identify genes under positive selective pressure, and those that contained lineage-specific insertions/deletions in functional domains. Finally, we conducted comprehensive in situ hybridizations to determine the extent of brain expression, and identify K-Channel gene enrichments in nuclei of the avian song system. Results We identified 107 K-Channel finch genes, including 6 novel genes common to non-mammalian vertebrate lineages. Twenty human genes are absent in songbirds, birds, or sauropsids, or unique to mammals, suggesting K-Channel properties may be lineage-specific. We also identified specific family members with insertions/deletions and/or high dN/dS ratios compared to chicken, a non-vocal learner. In situ hybridization revealed that while most K-Channel genes are broadly expressed in the brain, a subset is selectively expressed in song nuclei, representing molecular specializations of the vocal circuitry. Conclusions Together, these findings shed new light on genes that may regulate biophysical and excitable properties of the song circuitry, identify potential targets for the manipulation of the song system, and reveal genomic specializations that may relate to the emergence of vocal learning and associated brain areas in birds. PMID:23845108
White, Stephanie A.
2009-01-01
Could a mutation in a single gene be the evolutionary lynchpin supporting the development of human language? A rare mutation in the molecule known as FOXP2 discovered in a human family seemed to suggest so, and its sequence phylogeny reinforced a Chomskian view that language emerged wholesale in humans. Spurred by this discovery, research in primates, rodents and birds suggests that FoxP2 and other language-related genes are interactors in the neuromolecular networks that underlie subsystems of language, such symbolic understanding, vocal learning and theory of mind. The whole picture will only come together through comparative and integrative study into how the human language singularity evolved. PMID:19913899
Translating birdsong: songbirds as a model for basic and applied medical research.
Brainard, Michael S; Doupe, Allison J
2013-07-08
Songbirds, long of interest to basic neuroscience, have great potential as a model system for translational neuroscience. Songbirds learn their complex vocal behavior in a manner that exemplifies general processes of perceptual and motor skill learning and, more specifically, resembles human speech learning. Song is subserved by circuitry that is specialized for vocal learning and production but that has strong similarities to mammalian brain pathways. The combination of highly quantifiable behavior and discrete neural substrates facilitates understanding links between brain and behavior, both in normal states and in disease. Here we highlight (a) behavioral and mechanistic parallels between birdsong and aspects of speech and social communication, including insights into mirror neurons, the function of auditory feedback, and genes underlying social communication disorders, and (b) contributions of songbirds to understanding cortical-basal ganglia circuit function and dysfunction, including the possibility of harnessing adult neurogenesis for brain repair.
Translating Birdsong: Songbirds as a model for basic and applied medical research
2014-01-01
Songbirds, long of interest to basic neuroscientists, have great potential as a model system for translational neuroscience. Songbirds learn their complex vocal behavior in a manner that exemplifies general processes of perceptual and motor skill learning, and more specifically resembles human speech learning. Song is subserved by circuitry that is specialized for vocal learning and production, but that has strong similarities to mammalian brain pathways. The combination of a highly quantifiable behavior and discrete neural substrates facilitates understanding links between brain and behavior, both normally and in disease. Here we highlight 1) behavioral and mechanistic parallels between birdsong and aspects of speech and social communication, including insights into mirror neurons, the function of auditory feedback, and genes underlying social communication disorders, and 2) contributions of songbirds to understanding cortical-basal ganglia circuit function and dysfunction, including the possibility of harnessing adult neurogenesis for brain repair. PMID:23750515
Final Syllable Lengthening (FSL) in infant vocalizations.
Nathani, Suneeti; Oller, D Kimbrough; Cobo-Lewis, Alan B
2003-02-01
Final Syllable Lengthening (FSL) has been extensively examined in infant vocalizations in order to determine whether its basis is biological or learned. Findings suggest there may be a U-shaped developmental trajectory for FSL. The present study sought to verify this pattern and to determine whether vocal maturity and deafness influence FSL. Eight normally hearing infants, aged 0;3 to 1;0, and eight deaf infants, aged 0;8 to 4;0, were examined at three levels of prelinguistic vocal development: precanonical, canonical, and postcanonical. FSL was found at all three levels suggesting a biological basis for this phenomenon. Individual variability was, however, considerable. Reduction in the magnitude of FSL across the three sessions provided some support for a downward trend for FSL in infancy. Findings further indicated that auditory deprivation can significantly affect temporal aspects of infant speech production.
Food-Associated Calling in Gorillas (Gorilla g. gorilla) in the Wild
Luef, Eva Maria; Breuer, Thomas; Pika, Simone
2016-01-01
Many nonhuman primates produce food-associated vocalizations upon encountering or ingesting particular food. Concerning the great apes, only food-associated vocalizations of chimpanzees (Pan troglodytes) and bonobos (Pan paniscus) have been studied in detail, providing evidence that these vocalizations can be produced flexibly in relation to a variety of factors, such as the quantity and quality of food and/or the type of audience. Only anecdotal evidence exists of eastern (Gorilla beringei) and western gorillas (Gorilla gorilla) producing food-associated vocalizations, termed singing or humming. To enable a better understanding of the context in which these calls are produced, we investigated and compared the vocal behavior of two free-ranging groups of western lowland gorillas (Gorilla g. gorilla) at Mondika, Republic of Congo. Our results show that (a) food-associated call production occurs only during feeding and not in other contexts; (b) calling is not uniformly distributed across age and sex classes; (c) calls are only produced during feeding on specific foods; and (d) normally just one individual gives calls during group feeding sessions, however, certain food types elicit simultaneous calling of two or more individuals. Our findings provide new insight into the vocal abilities of gorillas but also carry larger implications for questions concerning vocal variability among the great apes. Food-associated calls of nonhuman primates have been shown to be flexible in terms of when they are used and who they are directed at, making them interesting vocalizations from the viewpoint of language evolution. Food-associated vocalizations in great apes can offer new opportunities to investigate the phylogenetic development of vocal communication within the primate lineage and can possibly contribute novel insights into the origins of human language. PMID:26909518
Dreiss, A N; Ruppli, C A; Roulin, A
2014-01-01
To compete over limited parental resources, young animals communicate with their parents and siblings by producing honest vocal signals of need. Components of begging calls that are sensitive to food deprivation may honestly signal need, whereas other components may be associated with individual-specific attributes that do not change with time such as identity, sex, absolute age and hierarchy. In a sib-sib communication system where barn owl (Tyto alba) nestlings vocally negotiate priority access to food resources, we show that calls have individual signatures that are used by nestlings to recognize which siblings are motivated to compete, even if most vocalization features vary with hunger level. Nestlings were more identifiable when food-deprived than food-satiated, suggesting that vocal identity is emphasized when the benefit of winning a vocal contest is higher. In broods where siblings interact iteratively, we speculate that individual-specific signature permits siblings to verify that the most vocal individual in the absence of parents is the one that indeed perceived the food brought by parents. Individual recognition may also allow nestlings to associate identity with individual-specific characteristics such as position in the within-brood dominance hierarchy. Calls indeed revealed age hierarchy and to a lower extent sex and absolute age. Using a cross-fostering experimental design, we show that most acoustic features were related to the nest of origin (but not the nest of rearing), suggesting a genetic or an early developmental effect on the ontogeny of vocal signatures. To conclude, our study suggests that sibling competition has promoted the evolution of vocal behaviours that signal not only hunger level but also intrinsic individual characteristics such as identity, family, sex and age. © 2013 The Authors. Journal of Evolutionary Biology © 2013 European Society For Evolutionary Biology.
Towards Real-Time Speech Emotion Recognition for Affective E-Learning
ERIC Educational Resources Information Center
Bahreini, Kiavash; Nadolski, Rob; Westera, Wim
2016-01-01
This paper presents the voice emotion recognition part of the FILTWAM framework for real-time emotion recognition in affective e-learning settings. FILTWAM (Framework for Improving Learning Through Webcams And Microphones) intends to offer timely and appropriate online feedback based upon learner's vocal intonations and facial expressions in order…
Bjørgesaeter, Anders; Ugland, Karl Inne; Bjørge, Arne
2004-10-01
The male harbor seal (Phoca vitulina) produces broadband nonharmonic vocalizations underwater during the breeding season. In total, 120 vocalizations from six colonies were analyzed to provide a description of the acoustic structure and for the presence of geographic variation. The complex harbor seal vocalizations may be described by how the frequency bandwidth varies over time. An algorithm that identifies the boundaries between noise and signal from digital spectrograms was developed in order to extract a frequency bandwidth contour. The contours were used as inputs for multivariate analysis. The vocalizations' sound types (e.g., pulsed sound, whistle, and broadband nonharmonic sound) were determined by comparing the vocalizations' spectrographic representations with sound waves produced by known sound sources. Comparison between colonies revealed differences in the frequency contours, as well as some geographical variation in use of sound types. The vocal differences may reflect a limited exchange of individuals between the six colonies due to long distances and strong site fidelity. Geographically different vocal repertoires have potential for identifying discrete breeding colonies of harbor seals, but more information is needed on the nature and extent of early movements of young, the degree of learning, and the stability of the vocal repertoire. A characteristic feature of many vocalizations in this study was the presence of tonal-like introductory phrases that fit into the categories pulsed sound and whistles. The functions of these phrases are unknown but may be important in distance perception and localization of the sound source. The potential behavioral consequences of the observed variability may be indicative of adaptations to different environmental properties influencing determination of distance and direction and plausible different male mating tactics.
Vocal Qualities in Music Theater Voice: Perceptions of Expert Pedagogues.
Bourne, Tracy; Kenny, Dianna
2016-01-01
To gather qualitative descriptions of music theater vocal qualities including belt, legit, and mix from expert pedagogues to better define this voice type. This is a prospective, semistructured interview. Twelve expert teachers from United States, United Kingdom, Asia, and Australia were interviewed by Skype and asked to identify characteristics of music theater vocal qualities including vocal production, physiology, esthetics, pitch range, and pedagogical techniques. Responses were compared with published studies on music theater voice. Belt and legit were generally described as distinct sounds with differing physiological and technical requirements. Teachers were concerned that belt should be taught "safely" to minimize vocal health risks. There was consensus between teachers and published research on the physiology of the glottis and vocal tract; however, teachers were not in agreement about breathing techniques. Neither were teachers in agreement about the meaning of "mix." Most participants described belt as heavily weighted, thick folds, thyroarytenoid-dominant, or chest register; however, there was no consensus on an appropriate term. Belt substyles were named and generally categorized by weightedness or tone color. Descriptions of male belt were less clear than for female belt. This survey provides an overview of expert pedagogical perspectives on the characteristics of belt, legit, and mix qualities in the music theater voice. Although teacher responses are generally in agreement with published research, there are still many controversial issues and gaps in knowledge and understanding of this vocal technique. Breathing techniques, vocal range, mix, male belt, and vocal registers require continuing investigation so that we can learn more about efficient and healthy vocal function in music theater singing. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Rules and mechanisms for efficient two-stage learning in neural circuits
Teşileanu, Tiberiu; Ölveczky, Bence; Balasubramanian, Vijay
2017-01-01
Trial-and-error learning requires evaluating variable actions and reinforcing successful variants. In songbirds, vocal exploration is induced by LMAN, the output of a basal ganglia-related circuit that also contributes a corrective bias to the vocal output. This bias is gradually consolidated in RA, a motor cortex analogue downstream of LMAN. We develop a new model of such two-stage learning. Using stochastic gradient descent, we derive how the activity in ‘tutor’ circuits (e.g., LMAN) should match plasticity mechanisms in ‘student’ circuits (e.g., RA) to achieve efficient learning. We further describe a reinforcement learning framework through which the tutor can build its teaching signal. We show that mismatches between the tutor signal and the plasticity mechanism can impair learning. Applied to birdsong, our results predict the temporal structure of the corrective bias from LMAN given a plasticity rule in RA. Our framework can be applied predictively to other paired brain areas showing two-stage learning. DOI: http://dx.doi.org/10.7554/eLife.20944.001 PMID:28374674
Heterospecific discrimination of Poecile vocalizations by zebra finches (Taeniopygia guttata).
Guillette, Lauren M; Hoeschele, Marisa; Hahn, Allison H; Sturdy, Christopher B
2013-08-01
Previous perceptual research with black-capped and mountain chickadees has demonstrated that the D note of the namesake chick-a-dee call controlled species-based discrimination compared to other note types in this call. In the current experiment, we sought to determine whether discrimination performance of the chickadees was controlled by stimulus-specific properties or due to learning through experience. To accomplish this, we tested zebra finches, a songbird species that is distantly related to chickadees, and also unfamiliar with black-capped and mountain chickadee vocalizations, on the same species-based discrimination on which black-capped and mountain chickadees were previously trained. We found that zebra finches learned the discrimination in the fewest number of trials with the D note, compared to other note types (i.e., the A, B, and C notes). In addition, we compared the current results to earlier work and found that zebra finches learned the discrimination in fewer trials compared to black-capped chickadees, and, across all species, males learned the discrimination in fewer trials than females. We discuss the roles that acoustic complexity and learning play in classification of the three species of songbirds tested. More generally, these results point to the benefits derived from testing members of each sex in species that vary in their natural history, vocal output, and phylogenetic relatedness as a means to uncover the mechanisms underlying acoustic communication. (PsycINFO Database Record (c) 2013 APA, all rights reserved).
Advanced Technologies for Acoustic Monitoring of Bird Populations
2009-04-01
Ford and P. Spong. 2000. Dialect change in resident killer whales : implications for vocal learning and cultural transmission. Animal Behaviour 60: 629...network to compare killer whale (Orcinus orca) dialects. Journal of the Acoustical Society of America 105(4): 2499-2507. Deecke, V. B., J. K. B...Murray, S. O., E. Mercado and H. L. Roitblat. 1998. The neural network classification of false killer whale (Pseudorca crassidens) vocalizations
Gestures, vocalizations, and memory in language origins.
Aboitiz, Francisco
2012-01-01
THIS ARTICLE DISCUSSES THE POSSIBLE HOMOLOGIES BETWEEN THE HUMAN LANGUAGE NETWORKS AND COMPARABLE AUDITORY PROJECTION SYSTEMS IN THE MACAQUE BRAIN, IN AN ATTEMPT TO RECONCILE TWO EXISTING VIEWS ON LANGUAGE EVOLUTION: one that emphasizes hand control and gestures, and the other that emphasizes auditory-vocal mechanisms. The capacity for language is based on relatively well defined neural substrates whose rudiments have been traced in the non-human primate brain. At its core, this circuit constitutes an auditory-vocal sensorimotor circuit with two main components, a "ventral pathway" connecting anterior auditory regions with anterior ventrolateral prefrontal areas, and a "dorsal pathway" connecting auditory areas with parietal areas and with posterior ventrolateral prefrontal areas via the arcuate fasciculus and the superior longitudinal fasciculus. In humans, the dorsal circuit is especially important for phonological processing and phonological working memory, capacities that are critical for language acquisition and for complex syntax processing. In the macaque, the homolog of the dorsal circuit overlaps with an inferior parietal-premotor network for hand and gesture selection that is under voluntary control, while vocalizations are largely fixed and involuntary. The recruitment of the dorsal component for vocalization behavior in the human lineage, together with a direct cortical control of the subcortical vocalizing system, are proposed to represent a fundamental innovation in human evolution, generating an inflection point that permitted the explosion of vocal language and human communication. In this context, vocal communication and gesturing have a common history in primate communication.
A role for descending auditory cortical projections in songbird vocal learning
Mandelblat-Cerf, Yael; Las, Liora; Denisenko, Natalia; Fee, Michale S
2014-01-01
Many learned motor behaviors are acquired by comparing ongoing behavior with an internal representation of correct performance, rather than using an explicit external reward. For example, juvenile songbirds learn to sing by comparing their song with the memory of a tutor song. At present, the brain regions subserving song evaluation are not known. In this study, we report several findings suggesting that song evaluation involves an avian 'cortical' area previously shown to project to the dopaminergic midbrain and other downstream targets. We find that this ventral portion of the intermediate arcopallium (AIV) receives inputs from auditory cortical areas, and that lesions of AIV result in significant deficits in vocal learning. Additionally, AIV neurons exhibit fast responses to disruptive auditory feedback presented during singing, but not during nonsinging periods. Our findings suggest that auditory cortical areas may guide learning by transmitting song evaluation signals to the dopaminergic midbrain and/or other subcortical targets. DOI: http://dx.doi.org/10.7554/eLife.02152.001 PMID:24935934
Meaning in the avian auditory cortex: Neural representation of communication calls
Elie, Julie E; Theunissen, Frédéric E
2014-01-01
Understanding how the brain extracts the behavioral meaning carried by specific vocalization types that can be emitted by various vocalizers and in different conditions is a central question in auditory research. This semantic categorization is a fundamental process required for acoustic communication and presupposes discriminative and invariance properties of the auditory system for conspecific vocalizations. Songbirds have been used extensively to study vocal learning, but the communicative function of all their vocalizations and their neural representation has yet to be examined. In our research, we first generated a library containing almost the entire zebra finch vocal repertoire and organized communication calls along 9 different categories based on their behavioral meaning. We then investigated the neural representations of these semantic categories in the primary and secondary auditory areas of 6 anesthetized zebra finches. To analyze how single units encode these call categories, we described neural responses in terms of their discrimination, selectivity and invariance properties. Quantitative measures for these neural properties were obtained using an optimal decoder based both on spike counts and spike patterns. Information theoretic metrics show that almost half of the single units encode semantic information. Neurons achieve higher discrimination of these semantic categories by being more selective and more invariant. These results demonstrate that computations necessary for semantic categorization of meaningful vocalizations are already present in the auditory cortex and emphasize the value of a neuro-ethological approach to understand vocal communication. PMID:25728175
Activity propagation in an avian basal ganglia-thalamo-cortical circuit essential for vocal learning
Kojima, Satoshi; Doupe, Allison J.
2009-01-01
In mammalian basal ganglia-thalamo-cortical circuits, GABAergic pallidal neurons are thought to ‘gate’ or modulate excitation in thalamus with their strong inhibitory inputs, and thus signal to cortex by pausing and permitting thalamic neurons to fire in response to excitatory drive. In contrast, in a homologous circuit specialized for vocal learning in songbirds, evidence suggests that pallidal neurons signal by eliciting postinhibitory rebound spikes in thalamus, which could occur even without any excitatory drive to thalamic neurons. To test whether songbird pallidal neurons can also communicate with thalamus by gating excitatory drive, as well as by postinhibitory rebound, we examined the activity of thalamic relay neurons in response to acute inactivation of the basal ganglia structure Area X; Area X contains the pallidal neurons that project to thalamus. Although inactivation of Area X should eliminate rebound-mediated spiking in thalamus, this manipulation tonically increases the firing rate of thalamic relay neurons, providing evidence that songbird pallidal neurons can gate tonic thalamic excitatory drive. We also found that the increased thalamic activity was fed forward to its target in the avian equivalent of cortex, which includes neurons that project to the vocal premotor area. These data raise the possibility that basal ganglia circuits can signal to cortex through thalamus both by generating postinhibitory rebound and by gating excitatory drive, and may switch between these modes depending on the statistics of pallidal firing. Moreover, these findings provide insight into the strikingly different disruptive effects of basal ganglia and ‘cortical’ lesions on songbird vocal learning. PMID:19369547
Vampola, Tomáš; Horáček, Jaromír; Laukkanen, Anne-Maria; Švec, Jan G
2015-04-01
Resonance frequencies of the vocal tract have traditionally been modelled using one-dimensional models. These cannot accurately represent the events in the frequency region of the formant cluster around 2.5-4.5 kHz, however. Here, the vocal tract resonance frequencies and their mode shapes are studied using a three-dimensional finite element model obtained from computed tomography measurements of a subject phonating on vowel [a:]. Instead of the traditional five, up to eight resonance frequencies of the vocal tract were found below the prominent antiresonance around 4.7 kHz. The three extra resonances were found to correspond to modes which were axially asymmetric and involved the piriform sinuses, valleculae, and transverse vibrations in the oral cavity. The results therefore suggest that the phenomenon of speaker's and singer's formant clustering may be more complex than originally thought.
Affective responses in tamarins elicited by species-specific music
Snowdon, Charles T.; Teie, David
2010-01-01
Theories of music evolution agree that human music has an affective influence on listeners. Tests of non-humans provided little evidence of preferences for human music. However, prosodic features of speech (‘motherese’) influence affective behaviour of non-verbal infants as well as domestic animals, suggesting that features of music can influence the behaviour of non-human species. We incorporated acoustical characteristics of tamarin affiliation vocalizations and tamarin threat vocalizations into corresponding pieces of music. We compared music composed for tamarins with that composed for humans. Tamarins were generally indifferent to playbacks of human music, but responded with increased arousal to tamarin threat vocalization based music, and with decreased activity and increased calm behaviour to tamarin affective vocalization based music. Affective components in human music may have evolutionary origins in the structure of calls of non-human animals. In addition, animal signals may have evolved to manage the behaviour of listeners by influencing their affective state. PMID:19726444
Framework Surgery for Treatment of Unilateral Vocal Fold Paralysis
Daniero, James J.; Garrett, C. Gaelyn; Francis, David O.
2014-01-01
Laryngeal framework surgery is the current gold standard treatment for unilateral vocal fold paralysis. It provides a permanent solution to glottic insufficiency caused by injury to the recurrent laryngeal nerve. Various modifications to the original Isshiki type I laryngoplasty procedure have been described to improve voice and swallowing outcomes. The success of this procedure is highly dependent on the experience of the surgeon as it epitomizes the intersection of art and science in the field. The following article reviews the evidence, controversies, and complications related to laryngoplasty for unilateral vocal fold paralysis. It also provides a detailed analysis of how and when arytenoid-positioning procedures should be considered, and summarizes the literature on postoperative outcomes. PMID:24883239
Rules of song development and their use in vocal interactions by birds with large repertoires.
Geberzahn, Nicole; Hultsch, Henrike
2004-06-01
Songbirds are well known for settling their disputes by vocal signals, and their singing plays a dominant role. Most studies on this issue have concentrated on bird species that develop and use small vocal repertoires. In this article we will go farther and focus on examples of how species with large song repertoires make use of their vocal competence. In particular, we will outline the study of interaction rules which have been elucidated by examining time- and pattern-specific relationships between signals exchanged by territorial neighbors. First we present an inquiry into the rules of song learning and development. In birds with large song repertoires, the ontogeny of such rules proceeds along a number of trajectories which help in understanding the often remarkable accomplishments of adult birds. In both approaches, our model species will be the Common Nightingale Luscinia megarhynchos that has been investigated intensively in the field and in the laboratory.
Effects of vocal training and phonatory task on voice onset time.
McCrea, Christopher R; Morris, Richard J
2007-01-01
The purpose of this study was to examine the temporal-acoustic differences between trained singers and nonsingers during speech and singing tasks. Thirty male participants were separated into two groups of 15 according to level of vocal training (ie, trained or untrained). The participants spoke and sang carrier phrases containing English voiced and voiceless bilabial stops, and voice onset time (VOT) was measured for the stop consonant productions. Mixed analyses of variance revealed a significant main effect between speech and singing for /p/ and /b/, with VOT durations longer during speech than singing for /p/, and the opposite true for /b/. Furthermore, a significant phonatory task by vocal training interaction was observed for /p/ productions. The results indicated that the type of phonatory task influences VOT and that these influences are most obvious in trained singers secondary to the articulatory and phonatory adjustments learned during vocal training.
A Neural Code That Is Isometric to Vocal Output and Correlates with Its Sensory Consequences
Vyssotski, Alexei L.; Stepien, Anna E.; Keller, Georg B.; Hahnloser, Richard H. R.
2016-01-01
What cortical inputs are provided to motor control areas while they drive complex learned behaviors? We study this question in the nucleus interface of the nidopallium (NIf), which is required for normal birdsong production and provides the main source of auditory input to HVC, the driver of adult song. In juvenile and adult zebra finches, we find that spikes in NIf projection neurons precede vocalizations by several tens of milliseconds and are insensitive to distortions of auditory feedback. We identify a local isometry between NIf output and vocalizations: quasi-identical notes produced in different syllables are preceded by highly similar NIf spike patterns. NIf multiunit firing during song precedes responses in auditory cortical neurons by about 50 ms, revealing delayed congruence between NIf spiking and a neural representation of auditory feedback. Our findings suggest that NIf codes for imminent acoustic events within vocal performance. PMID:27723764
What Pinnipeds Have to Say about Human Speech, Music, and the Evolution of Rhythm.
Ravignani, Andrea; Fitch, W Tecumseh; Hanke, Frederike D; Heinrich, Tamara; Hurgitsch, Bettina; Kotz, Sonja A; Scharff, Constance; Stoeger, Angela S; de Boer, Bart
2016-01-01
Research on the evolution of human speech and music benefits from hypotheses and data generated in a number of disciplines. The purpose of this article is to illustrate the high relevance of pinniped research for the study of speech, musical rhythm, and their origins, bridging and complementing current research on primates and birds. We briefly discuss speech, vocal learning, and rhythm from an evolutionary and comparative perspective. We review the current state of the art on pinniped communication and behavior relevant to the evolution of human speech and music, showing interesting parallels to hypotheses on rhythmic behavior in early hominids. We suggest future research directions in terms of species to test and empirical data needed.
What Pinnipeds Have to Say about Human Speech, Music, and the Evolution of Rhythm
Ravignani, Andrea; Fitch, W. Tecumseh; Hanke, Frederike D.; Heinrich, Tamara; Hurgitsch, Bettina; Kotz, Sonja A.; Scharff, Constance; Stoeger, Angela S.; de Boer, Bart
2016-01-01
Research on the evolution of human speech and music benefits from hypotheses and data generated in a number of disciplines. The purpose of this article is to illustrate the high relevance of pinniped research for the study of speech, musical rhythm, and their origins, bridging and complementing current research on primates and birds. We briefly discuss speech, vocal learning, and rhythm from an evolutionary and comparative perspective. We review the current state of the art on pinniped communication and behavior relevant to the evolution of human speech and music, showing interesting parallels to hypotheses on rhythmic behavior in early hominids. We suggest future research directions in terms of species to test and empirical data needed. PMID:27378843
ERIC Educational Resources Information Center
Malloy, Peggy
2009-01-01
Long before children learn language, they communicate with gestures, vocalizations, facial expressions, and body language. This is known as prelinguistic (prior to language) communication. Most children learn this type of communication without formal teaching, but children who are deaf-blind may need guidance to learn it. This publication…
Retrieving Tract Variables From Acoustics: A Comparison of Different Machine Learning Strategies.
Mitra, Vikramjit; Nam, Hosung; Espy-Wilson, Carol Y; Saltzman, Elliot; Goldstein, Louis
2010-09-13
Many different studies have claimed that articulatory information can be used to improve the performance of automatic speech recognition systems. Unfortunately, such articulatory information is not readily available in typical speaker-listener situations. Consequently, such information has to be estimated from the acoustic signal in a process which is usually termed "speech-inversion." This study aims to propose and compare various machine learning strategies for speech inversion: Trajectory mixture density networks (TMDNs), feedforward artificial neural networks (FF-ANN), support vector regression (SVR), autoregressive artificial neural network (AR-ANN), and distal supervised learning (DSL). Further, using a database generated by the Haskins Laboratories speech production model, we test the claim that information regarding constrictions produced by the distinct organs of the vocal tract (vocal tract variables) is superior to flesh-point information (articulatory pellet trajectories) for the inversion process.
Five fundamental constraints on theories of the origins of music
Merker, Bjorn; Morley, Iain; Zuidema, Willem
2015-01-01
The diverse forms and functions of human music place obstacles in the way of an evolutionary reconstruction of its origins. In the absence of any obvious homologues of human music among our closest primate relatives, theorizing about its origins, in order to make progress, needs constraints from the nature of music, the capacities it engages, and the contexts in which it occurs. Here we propose and examine five fundamental constraints that bear on theories of how music and some of its features may have originated. First, cultural transmission, bringing the formal powers of cultural as contrasted with Darwinian evolution to bear on its contents. Second, generativity, i.e. the fact that music generates infinite pattern diversity by finite means. Third, vocal production learning, without which there can be no human singing. Fourth, entrainment with perfect synchrony, without which there is neither rhythmic ensemble music nor rhythmic dancing to music. And fifth, the universal propensity of humans to gather occasionally to sing and dance together in a group, which suggests a motivational basis endemic to our biology. We end by considering the evolutionary context within which these constraints had to be met in the genesis of human musicality. PMID:25646518
Dissociable Effects on Birdsong of Androgen Signaling in Cortex-Like Brain Regions of Canaries
2017-01-01
The neural basis of how learned vocalizations change during development and in adulthood represents a major challenge facing cognitive neuroscience. This plasticity in the degree to which learned vocalizations can change in both humans and songbirds is linked to the actions of sex steroid hormones during ontogeny but also in adulthood in the context of seasonal changes in birdsong. We investigated the role of steroid hormone signaling in the brain on distinct features of birdsong using adult male canaries (Serinus canaria), which show extensive seasonal vocal plasticity as adults. Specifically, we bilaterally implanted the potent androgen receptor antagonist flutamide in two key brain regions that control birdsong. We show that androgen signaling in the motor cortical-like brain region, the robust nucleus of the arcopallium (RA), controls syllable and trill bandwidth stereotypy, while not significantly affecting higher order features of song such syllable-type usage (i.e., how many times each syllable type is used) or syllable sequences. In contrast, androgen signaling in the premotor cortical-like brain region, HVC (proper name), controls song variability by increasing the variability of syllable-type usage and syllable sequences, while having no effect on syllable or trill bandwidth stereotypy. Other aspects of song, such as the duration of trills and the number of syllables per song, were also differentially affected by androgen signaling in HVC versus RA. These results implicate androgens in regulating distinct features of complex motor output in a precise and nonredundant manner. SIGNIFICANCE STATEMENT Vocal plasticity is linked to the actions of sex steroid hormones, but the precise mechanisms are unclear. We investigated this question in adult male canaries (Serinus canaria), which show extensive vocal plasticity throughout their life. We show that androgens in two cortex-like vocal control brain regions regulate distinct aspects of vocal plasticity. For example, in HVC (proper name), androgens regulate variability in syntax but not phonology, whereas androgens in the robust nucleus of the arcopallium (RA) regulate variability in phonology but not syntax. Temporal aspects of song were also differentially affected by androgen signaling in HVC versus RA. Thus, androgen signaling may reduce vocal plasticity by acting in a nonredundant and precise manner in the brain. PMID:28821656
NASA Astrophysics Data System (ADS)
Rendall, Drew; Kollias, Sophie; Ney, Christina; Lloyd, Peter
2005-02-01
Key voice features-fundamental frequency (F0) and formant frequencies-can vary extensively between individuals. Much of the variation can be traced to differences in the size of the larynx and vocal-tract cavities, but whether these differences in turn simply reflect differences in speaker body size (i.e., neutral vocal allometry) remains unclear. Quantitative analyses were therefore undertaken to test the relationship between speaker body size and voice F0 and formant frequencies for human vowels. To test the taxonomic generality of the relationships, the same analyses were conducted on the vowel-like grunts of baboons, whose phylogenetic proximity to humans and similar vocal production biology and voice acoustic patterns recommend them for such comparative research. For adults of both species, males were larger than females and had lower mean voice F0 and formant frequencies. However, beyond this, F0 variation did not track body-size variation between the sexes in either species, nor within sexes in humans. In humans, formant variation correlated significantly with speaker height but only in males and not in females. Implications for general vocal allometry are discussed as are implications for speech origins theories, and challenges to them, related to laryngeal position and vocal tract length. .
[Exudative laryngeal diseases of Reinke's space].
Remacle, M; Lagneau, G; Marbaix, E; Doyen, A; Van den Eeckhaut, J
1992-01-01
Exudative laryngopathies are represented by the nodule, the polyp and the Reinke's oedema. Their origin is found in exudative processes located in the Reinke's space. Evolution towards one or another lesion depends on the etiological factor (vocal abuse, tobacco, alcohol, acute vocal fold trauma) and can happen according three modalities: oedema, fibrosis and angiectasies development. Apparition of fibrosis inside the lesion or too important haemorrhage require microsurgical treatment in addition to the medical treatment and the speech therapy.
Daily and Developmental Modulation of “Premotor” Activity in the Birdsong System
Day, Nancy F.; Kinnischtzke, Amanda K.; Adam, Murtaza; Nick, Teresa A.
2009-01-01
Human speech and birdsong are shaped during a sensorimotor sensitive period in which auditory feedback guides vocal learning. To study brain activity as song learning occurred, we recorded longitudinally from developing zebra finches during the sensorimotor phase. Learned sequences of vocalizations (motifs) were examined along with contemporaneous neural population activity in the song nucleus HVC, which is necessary for the production of learned song (Nottebohm et al. [1976]: J Comp Neurol 165:457–486; Simpson and Vicario [1990]: J Neurosci 10:1541–1556). During singing, HVC activity levels increased as the day progressed and decreased after a night of sleep in juveniles and adults. In contrast, the pattern of HVC activity changed on a daily basis only in juveniles: activity bursts became more pronounced during the day. The HVC of adults was significantly burstier than that of juveniles. HVC bursting was relevant to song behavior because the degree of burstiness inversely correlated with the variance of song features in juveniles. The song of juveniles degrades overnight (Deregnaucourt et al. [2005]: Nature 433:710–716). Consistent with a relationship between HVC activity and song plasticity (Day et al. [2008]: J Neurophys 100:2956–2965), HVC burstiness degraded overnight in young juveniles and the amount of overnight degradation declined with developmental song learning. Nocturnal changes in HVC activity strongly and inversely correlated with the next day's change, suggesting that sleep-dependent degradation of HVC activity may facilitate or enable subsequent diurnal changes. Collectively, these data show that HVC activity levels exhibit daily cycles in adults and juveniles, whereas HVC burstiness and song stereotypy change daily in juveniles only. In addition, the data indicate that HVC burstiness increases with development and inversely correlates with song variability, which is necessary for trial and error vocal learning. PMID:19650042
Cornez, Gilles; Madison, Farrah N; Van der Linden, Annemie; Cornil, Charlotte; Yoder, Kathleen M; Ball, Gregory F; Balthazart, Jacques
2017-09-01
Perineuronal nets (PNN) are aggregations of chondroitin sulfate proteoglycans surrounding the soma and proximal processes of neurons, mostly GABAergic interneurons expressing parvalbumin. They limit the plasticity of their afferent synaptic connections. In zebra finches PNN develop in an experience-dependent manner in the song control nuclei HVC and RA (nucleus robustus arcopallialis) when young birds crystallize their song. Because songbird species that are open-ended learners tend to recapitulate each year the different phases of song learning until their song crystallizes at the beginning of the breeding season, we tested whether seasonal changes in PNN expression would be found in the song control nuclei of a seasonally breeding species such as the European starling. Only minimal changes in PNN densities and total number of cells surrounded by PNN were detected. However, comparison of the density of PNN and of PNN surrounding parvalbumin-positive cells revealed that these structures are far less numerous in starlings that show extensive adult vocal plasticity, including learning of new songs throughout the year, than in the closed-ended learner zebra finches. Canaries that also display some vocal plasticity across season but were never formally shown to learn new songs in adulthood were intermediate in this respect. Together these data suggest that establishment of PNN around parvalbumin-positive neurons in song control nuclei has diverged during evolution to control the different learning capacities observed in songbird species. This differential expression of PNN in different songbird species could represent a key cellular mechanism mediating species variation between closed-ended and open-ended learning strategies. © 2017 Wiley Periodicals, Inc. Develop Neurobiol 77: 975-994, 2017. © 2017 Wiley Periodicals, Inc.
Mother goats do not forget their kids’ calls
Briefer, Elodie F.; Padilla de la Torre, Monica; McElligott, Alan G.
2012-01-01
Parent–offspring recognition is crucial for offspring survival. At long distances, this recognition is mainly based on vocalizations. Because of maturation-related changes to the structure of vocalizations, parents have to learn successive call versions produced by their offspring throughout ontogeny in order to maintain recognition. However, because of the difficulties involved in following the same individuals over years, it is not clear how long this vocal memory persists. Here, we investigated long-term vocal recognition in goats. We tested responses of mothers to their kids’ calls 7–13 months after weaning. We then compared mothers’ responses to calls of their previous kids with their responses to the same calls at five weeks postpartum. Subjects tended to respond more to their own kids at five weeks postpartum than 11–17 months later, but displayed stronger responses to their previous kids than to familiar kids from other females. Acoustic analyses showed that it is unlikely that mothers were responding to their previous kids simply because they confounded them with the new kids they were currently nursing. Therefore, our results provide evidence for strong, long-term vocal memory capacity in goats. The persistence of offspring vocal recognition beyond weaning could have important roles in kin social relationships and inbreeding avoidance. PMID:22719031
Using Ambulatory Voice Monitoring to Investigate Common Voice Disorders: Research Update
Mehta, Daryush D.; Van Stan, Jarrad H.; Zañartu, Matías; Ghassemi, Marzyeh; Guttag, John V.; Espinoza, Víctor M.; Cortés, Juan P.; Cheyne, Harold A.; Hillman, Robert E.
2015-01-01
Many common voice disorders are chronic or recurring conditions that are likely to result from inefficient and/or abusive patterns of vocal behavior, referred to as vocal hyperfunction. The clinical management of hyperfunctional voice disorders would be greatly enhanced by the ability to monitor and quantify detrimental vocal behaviors during an individual’s activities of daily life. This paper provides an update on ongoing work that uses a miniature accelerometer on the neck surface below the larynx to collect a large set of ambulatory data on patients with hyperfunctional voice disorders (before and after treatment) and matched-control subjects. Three types of analysis approaches are being employed in an effort to identify the best set of measures for differentiating among hyperfunctional and normal patterns of vocal behavior: (1) ambulatory measures of voice use that include vocal dose and voice quality correlates, (2) aerodynamic measures based on glottal airflow estimates extracted from the accelerometer signal using subject-specific vocal system models, and (3) classification based on machine learning and pattern recognition approaches that have been used successfully in analyzing long-term recordings of other physiological signals. Preliminary results demonstrate the potential for ambulatory voice monitoring to improve the diagnosis and treatment of common hyperfunctional voice disorders. PMID:26528472
Mother goats do not forget their kids' calls.
Briefer, Elodie F; Padilla de la Torre, Monica; McElligott, Alan G
2012-09-22
Parent-offspring recognition is crucial for offspring survival. At long distances, this recognition is mainly based on vocalizations. Because of maturation-related changes to the structure of vocalizations, parents have to learn successive call versions produced by their offspring throughout ontogeny in order to maintain recognition. However, because of the difficulties involved in following the same individuals over years, it is not clear how long this vocal memory persists. Here, we investigated long-term vocal recognition in goats. We tested responses of mothers to their kids' calls 7-13 months after weaning. We then compared mothers' responses to calls of their previous kids with their responses to the same calls at five weeks postpartum. Subjects tended to respond more to their own kids at five weeks postpartum than 11-17 months later, but displayed stronger responses to their previous kids than to familiar kids from other females. Acoustic analyses showed that it is unlikely that mothers were responding to their previous kids simply because they confounded them with the new kids they were currently nursing. Therefore, our results provide evidence for strong, long-term vocal memory capacity in goats. The persistence of offspring vocal recognition beyond weaning could have important roles in kin social relationships and inbreeding avoidance.
Experiences of a short vocal training course for call-centre customer service advisors.
Lehto, Laura; Rantala, Leena; Vilkman, Erkki; Alku, Paavo; Bäckström, Tom
2003-01-01
It is commonly known that occupational voice users suffer from voice symptoms to varying extents. The purpose of this study was to find out the effects of a short (2-day) vocal training course on professional speakers' voice. The subjects were 38 female and 10 male customer advisors, who mainly use the telephone during their working hours at a call centre. The findings showed that although the subjects did not suffer from severe voice problems, they reported that the short vocal training course had an effect of some of the vocal symptoms they had experienced. More than 50% of the females and males reported a decrease in the feeling of mucus and the consequent need to clear the throat, and diminished worsening of their voice. Over 60% thought that voice training had improved their vocal habits and none reported a negative influence of the course on their voice. Females also reported a reduction of vocal fatigue. The subjects were further asked to respond to 23 statements on how they experienced the voice training in general. The statements 'I learned things that I didn't know about the use of voice in general' and 'I got useful and important knowledge concerning my work' were highly assessed by both females and males. The results suggest that even a short vocal training course might affect positively the self-reported well-being of persons working in a vocally loading occupation. However, to find out the long-term effects of a short training course, a follow-up study would need to be carried out. Copyright 2003 S. Karger AG, Basel
Common Vocal Effects and Partial Glottal Vibration in Professional Nonclassical Singers.
Caffier, Philipp P; Ibrahim Nasr, Ahmed; Ropero Rendon, Maria Del Mar; Wienhausen, Sascha; Forbes, Eleanor; Seidner, Wolfram; Nawka, Tadeus
2018-05-01
To multidimensionally investigate common vocal effects in experienced professional nonclassical singers, to examine their mechanism of production and reproducibility, to demonstrate the existence of partial glottal vibration, and to assess the potential of damage to the voice from nonclassical singing. Individual cohort study. Ten male singers aged between 25 and 46 years (34 ± 7 years [mean ± SD]) with different stylistic backgrounds were recruited (five pop/rock/metal, five musical theater). Participants repeatedly presented the usual nonclassical vocal effects and techniques in their repertoire. All performances were documented and analyzed using established instruments (eg, auditory-perceptual assessment, videolaryngostroboscopy, electroglottography, voice function diagnostics). The vocal apparatus of all singers was healthy and capable of high performance. Typical nonclassical vocal effects were breathy voice, creaky voice, vocal fry, grunting, distortion, rattle, belt, and twang. All effects could be easily differentiated from each other. They were intraindividually consistently repeatable and also interindividually produced in a similar manner. A special feature in one singer was the first evidence of partial glottal vibration when belting in the high register. The unintended transition to this reduced voice quality was accompanied by physical fatigue and inflexible respiratory support. The long-lasting use of the investigated nonclassical vocal effects had no negative impact on trained singers. The possibility of long-term damage depends on the individual constitution, specific use, duration, and extent of the hyperfunction. The incidence of partial glottal vibration and its consequences require continuing research to learn more about efficient and healthy vocal function in nonclassical singing. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Aeroelastic Model of Vocal-Fold Vibrating Element for Studying the Phonation Threshold
NASA Astrophysics Data System (ADS)
Horáček, J.; Švec, J. G.
2002-10-01
An original theoretical model for vibration onset of the vocal folds in the air-flow coming from the human subglottal tract is designed, which allows studying the influence of the physical properties of the vocal folds (e.g., geometrical shape, mass, viscosity) on their vibration characteristics (such as the natural frequencies, mode shapes of vibration and the thresholds of instability). The mathematical model of the vocal fold is designed as a simplified dynamic system of two degrees of freedom (rotation and translation) vibrating on an elastic foundation in the wall of a channel conveying air. An approximate unsteady one-dimensional flow theory for the inviscid incompressible fluid is presented for the phonatory air-flow. A generally defined shape of the vocal-fold surface is considered for expressing the unsteady aerodynamic forces in the glottis. The parameters of the mechanical part of the model, i.e., the mass, stiffness and damping matrices, are related to the geometry and material density of the vocal folds as well as to the fundamental natural frequency and damping known from experiments. The coupled numerical solution yields the vibration characteristics (natural frequencies, damping and mode shapes of vibration), including the instability thresholds of the aeroelastic system. The vibration characteristics obtained from the coupled numerical solution of the system appear to be in reasonable qualitative agreement with the physiological data and clinical observations. The model is particularly suitable for studying the phonation threshold, i.e., the onset of vibration of the vocal folds.
Azul, David
2016-11-01
Transmasculine people assigned female gender at birth but who do not identify with this classification have traditionally received little consideration in the voice literature. Existing analyses tend to be focused on evaluating speaker voice characteristics, whereas other factors that contribute to the production of vocal gender have remained underexplored. Most studies rely on researcher-centred perspectives, whereas very little is known about how transmasculine people themselves experience and make sense of their vocal situations. To explore how participants described their subjective gender positionings; which gender attributions they wished to receive from others; which gender they self-attributed to their voices; which gender attributions they had received from others; and how far participants were satisfied with the gender-related aspects of their vocal situations. Transcripts of semi-structured interviews with 14 German-speaking transmasculine people served as the original data corpus. Sections in which participants described the gender-related aspects of their vocal situations and that were relevant to the current research objectives were selected and explored using qualitative content analysis. The analysis revealed diverse accounts pertaining to the factors that contribute to the production of vocal gender for individual participants and variable levels of satisfaction with vocal gender presentation and attribution. Transmasculine people need to be regarded as a heterogeneous population and clinical practice needs to follow a client-centred, individualized approach. © 2016 Royal College of Speech and Language Therapists.
Identification of prelinguistic phonological categories.
Ramsdell, Heather L; Oller, D Kimbrough; Buder, Eugene H; Ethington, Corinna A; Chorna, Lesya
2012-12-01
The prelinguistic infant's babbling repertoire of syllables--the phonological categories that form the basis for early word learning--is noticed by caregivers who interact with infants around them. Prior research on babbling has not explored the caregiver's role in recognition of early vocal categories as foundations for word learning. In the present work, the authors begin to address this gap. The authors explored vocalizations produced by 8 infants at 3 ages (8, 10, and 12 months) in studies illustrating identification of phonological categories through caregiver report, laboratory procedures simulating the caregiver's natural mode of listening, and the more traditional laboratory approach (phonetic transcription). Caregivers reported small repertoires of syllables for their infants. Repertoires of similar size and phonetic content were discerned in the laboratory by judges who simulated the caregiver's natural mode of listening. However, phonetic transcription with repeated listening to infant recordings yielded repertoire sizes that vastly exceeded those reported by caregivers and naturalistic listeners. The results suggest that caregiver report and naturalistic listening by laboratory staff can provide a new way to explore key characteristics of early infant vocal categories, a way that may provide insight into later speech and language development.
Basal ganglia function, stuttering, sequencing, and repair in adult songbirds.
Kubikova, Lubica; Bosikova, Eva; Cvikova, Martina; Lukacova, Kristina; Scharff, Constance; Jarvis, Erich D
2014-10-13
A pallial-basal-ganglia-thalamic-pallial loop in songbirds is involved in vocal motor learning. Damage to its basal ganglia part, Area X, in adult zebra finches has been noted to have no strong effects on song and its function is unclear. Here we report that neurotoxic damage to adult Area X induced changes in singing tempo and global syllable sequencing in all animals, and considerably increased syllable repetition in birds whose song motifs ended with minor repetitions before lesioning. This stuttering-like behavior started at one month, and improved over six months. Unexpectedly, the lesioned region showed considerable recovery, including immigration of newly generated or repaired neurons that became active during singing. The timing of the recovery and stuttering suggest that immature recovering activity of the circuit might be associated with stuttering. These findings indicate that even after juvenile learning is complete, the adult striatum plays a role in higher level organization of learned vocalizations.
Place preference and vocal learning rely on distinct reinforcers in songbirds.
Murdoch, Don; Chen, Ruidong; Goldberg, Jesse H
2018-04-30
In reinforcement learning (RL) agents are typically tasked with maximizing a single objective function such as reward. But it remains poorly understood how agents might pursue distinct objectives at once. In machines, multiobjective RL can be achieved by dividing a single agent into multiple sub-agents, each of which is shaped by agent-specific reinforcement, but it remains unknown if animals adopt this strategy. Here we use songbirds to test if navigation and singing, two behaviors with distinct objectives, can be differentially reinforced. We demonstrate that strobe flashes aversively condition place preference but not song syllables. Brief noise bursts aversively condition song syllables but positively reinforce place preference. Thus distinct behavior-generating systems, or agencies, within a single animal can be shaped by correspondingly distinct reinforcement signals. Our findings suggest that spatially segregated vocal circuits can solve a credit assignment problem associated with multiobjective learning.
Arneodo, Ezequiel M; Perl, Yonatan Sanz; Goller, Franz; Mindlin, Gabriel B
2012-01-01
Because of the parallels found with human language production and acquisition, birdsong is an ideal animal model to study general mechanisms underlying complex, learned motor behavior. The rich and diverse vocalizations of songbirds emerge as a result of the interaction between a pattern generator in the brain and a highly nontrivial nonlinear periphery. Much of the complexity of this vocal behavior has been understood by studying the physics of the avian vocal organ, particularly the syrinx. A mathematical model describing the complex periphery as a nonlinear dynamical system leads to the conclusion that nontrivial behavior emerges even when the organ is commanded by simple motor instructions: smooth paths in a low dimensional parameter space. An analysis of the model provides insight into which parameters are responsible for generating a rich variety of diverse vocalizations, and what the physiological meaning of these parameters is. By recording the physiological motor instructions elicited by a spontaneously singing muted bird and computing the model on a Digital Signal Processor in real-time, we produce realistic synthetic vocalizations that replace the bird's own auditory feedback. In this way, we build a bio-prosthetic avian vocal organ driven by a freely behaving bird via its physiologically coded motor commands. Since it is based on a low-dimensional nonlinear mathematical model of the peripheral effector, the emulation of the motor behavior requires light computation, in such a way that our bio-prosthetic device can be implemented on a portable platform.
Songs to syntax: the linguistics of birdsong.
Berwick, Robert C; Okanoya, Kazuo; Beckers, Gabriel J L; Bolhuis, Johan J
2011-03-01
Unlike our primate cousins, many species of bird share with humans a capacity for vocal learning, a crucial factor in speech acquisition. There are striking behavioural, neural and genetic similarities between auditory-vocal learning in birds and human infants. Recently, the linguistic parallels between birdsong and spoken language have begun to be investigated. Although both birdsong and human language are hierarchically organized according to particular syntactic constraints, birdsong structure is best characterized as 'phonological syntax', resembling aspects of human sound structure. Crucially, birdsong lacks semantics and words. Formal language and linguistic analysis remains essential for the proper characterization of birdsong as a model system for human speech and language, and for the study of the brain and cognition evolution. Copyright © 2011 Elsevier Ltd. All rights reserved.
Motor Tics, Tourette Syndrome, and Learning Disabilities.
ERIC Educational Resources Information Center
Lerer, Robert J.
1987-01-01
Complex motor tics associated with vocal tics indicate a high likelihood of Tourette syndrome; children with this syndrome may also have learning disabilities and attentional disorders. Individuals may be treated with stimulant drugs which may precipitate or exacerbate tics. Pharmacotherapy is available for management of tics and attentional…
Segmentation of Killer Whale Vocalizations Using the Hilbert-Huang Transform
NASA Astrophysics Data System (ADS)
Adam, Olivier
2008-12-01
The study of cetacean vocalizations is usually based on spectrogram analysis. The feature extraction is obtained from 2D methods like the edge detection algorithm. Difficulties appear when signal-to-noise ratios are weak or when more than one vocalization is simultaneously emitted. This is the case for acoustic observations in a natural environment and especially for the killer whales which swim in groups. To resolve this problem, we propose the use of the Hilbert-Huang transform. First, we illustrate how few modes (5) are satisfactory for the analysis of these calls. Then, we detail our approach which consists of combining the modes for extracting the time-varying frequencies of the vocalizations. This combination takes advantage of one of the empirical mode decomposition properties which is that the successive IMFs represent the original data broken down into frequency components from highest to lowest frequency. To evaluate the performance, our method is first applied on the simulated chirp signals. This approach allows us to link one chirp to one mode. Then we apply it on real signals emitted by killer whales. The results confirm that this method is a favorable alternative for the automatic extraction of killer whale vocalizations.
ERIC Educational Resources Information Center
Saktanli, S. Cem
2011-01-01
This experimental study was done to see if using computer supported notation and vocalization program for teaching songs instead of using block flute accompanied song teaching has any significant effect on students' singing behavior. The study group is composed of the 5th, 6th and 7th graders of 2008-2009 educational term in T.O.K.I. Yahya Kemal…
Perceptual fluency and judgments of vocal aesthetics and stereotypicality.
Babel, Molly; McGuire, Grant
2015-05-01
Research has shown that processing dynamics on the perceiver's end determine aesthetic pleasure. Specifically, typical objects, which are processed more fluently, are perceived as more attractive. We extend this notion of perceptual fluency to judgments of vocal aesthetics. Vocal attractiveness has traditionally been examined with respect to sexual dimorphism and the apparent size of a talker, as reconstructed from the acoustic signal, despite evidence that gender-specific speech patterns are learned social behaviors. In this study, we report on a series of three experiments using 60 voices (30 females) to compare the relationship between judgments of vocal attractiveness, stereotypicality, and gender categorization fluency. Our results indicate that attractiveness and stereotypicality are highly correlated for female and male voices. Stereotypicality and categorization fluency were also correlated for male voices, but not female voices. Crucially, stereotypicality and categorization fluency interacted to predict attractiveness, suggesting the role of perceptual fluency is present, but nuanced, in judgments of human voices. © 2014 Cognitive Science Society, Inc.
Mother-offspring recognition in the domestic cat: Kittens recognize their own mother's call.
Szenczi, Péter; Bánszegi, Oxána; Urrutia, Andrea; Faragó, Tamás; Hudson, Robyn
2016-07-01
Acoustic communication can play an important part in mother-young recognition in many mammals. This, however, has still only been investigated in a small range mainly of herd- or colony-living species. Here we report on the behavioral response of kittens of the domestic cat, a typically solitary carnivore, to playbacks of "greeting chirps" and "meows" from their own versus alien mothers. We found significantly stronger responses to the chirps from kittens' own mother than to her meows or to the chirps or meows of alien mothers. Acoustic analysis revealed greater variation between vocalizations from different mothers than for vocalizations from the same mother. We conclude that chirps emitted by mother cats at the nest represent a specific form of vocal communication with their young, and that kittens learn and respond positively to these and distinguish them from chirps of other mothers and from other cat vocalizations while still in the nest. © 2016 Wiley Periodicals, Inc. Dev Psychobiol 58: 568-577, 2016. © 2016 Wiley Periodicals, Inc.
Riede, Tobias; Goller, Franz
2010-10-01
Song production in songbirds is a model system for studying learned vocal behavior. As in humans, bird phonation involves three main motor systems (respiration, vocal organ and vocal tract). The avian respiratory mechanism uses pressure regulation in air sacs to ventilate a rigid lung. In songbirds sound is generated with two independently controlled sound sources, which reside in a uniquely avian vocal organ, the syrinx. However, the physical sound generation mechanism in the syrinx shows strong analogies to that in the human larynx, such that both can be characterized as myoelastic-aerodynamic sound sources. Similarities include active adduction and abduction, oscillating tissue masses which modulate flow rate through the organ and a layered structure of the oscillating tissue masses giving rise to complex viscoelastic properties. Differences in the functional morphology of the sound producing system between birds and humans require specific motor control patterns. The songbird vocal apparatus is adapted for high speed, suggesting that temporal patterns and fast modulation of sound features are important in acoustic communication. Rapid respiratory patterns determine the coarse temporal structure of song and maintain gas exchange even during very long songs. The respiratory system also contributes to the fine control of airflow. Muscular control of the vocal organ regulates airflow and acoustic features. The upper vocal tract of birds filters the sounds generated in the syrinx, and filter properties are actively adjusted. Nonlinear source-filter interactions may also play a role. The unique morphology and biomechanical system for sound production in birds presents an interesting model for exploring parallels in control mechanisms that give rise to highly convergent physical patterns of sound generation. More comparative work should provide a rich source for our understanding of the evolution of complex sound producing systems. Copyright © 2009 Elsevier Inc. All rights reserved.
Arneodo, Ezequiel M.; Perl, Yonatan Sanz; Goller, Franz; Mindlin, Gabriel B.
2012-01-01
Because of the parallels found with human language production and acquisition, birdsong is an ideal animal model to study general mechanisms underlying complex, learned motor behavior. The rich and diverse vocalizations of songbirds emerge as a result of the interaction between a pattern generator in the brain and a highly nontrivial nonlinear periphery. Much of the complexity of this vocal behavior has been understood by studying the physics of the avian vocal organ, particularly the syrinx. A mathematical model describing the complex periphery as a nonlinear dynamical system leads to the conclusion that nontrivial behavior emerges even when the organ is commanded by simple motor instructions: smooth paths in a low dimensional parameter space. An analysis of the model provides insight into which parameters are responsible for generating a rich variety of diverse vocalizations, and what the physiological meaning of these parameters is. By recording the physiological motor instructions elicited by a spontaneously singing muted bird and computing the model on a Digital Signal Processor in real-time, we produce realistic synthetic vocalizations that replace the bird's own auditory feedback. In this way, we build a bio-prosthetic avian vocal organ driven by a freely behaving bird via its physiologically coded motor commands. Since it is based on a low-dimensional nonlinear mathematical model of the peripheral effector, the emulation of the motor behavior requires light computation, in such a way that our bio-prosthetic device can be implemented on a portable platform. PMID:22761555
Evaluating theories of bird song learning: implications for future directions.
Margoliash, D
2002-12-01
Studies of birdsong learning have stimulated extensive hypotheses at all levels of behavioral and physiological organization. This hypothesis building is valuable for the field and is consistent with the remarkable range of issues that can be rigorously addressed in this system. The traditional instructional (template) theory of song learning has been challenged on multiple fronts, especially at a behavioral level by evidence consistent with selectional hypotheses. In this review I highlight the caveats associated with these theories to better define the limits of our knowledge and identify important experiments for the future. The sites and representational forms of the various conceptual entities posited by the template theory are unknown. The distinction between instruction and selection in vocal learning is not well established at a mechanistic level. There is as yet insufficient neurophysiological data to choose between competing mechanisms of error-driven learning and reinforcement learning. Both may obtain for vocal learning. The possible role of sleep in acoustic or procedural memory consolidation, while supported by some physiological observations, does not yet have support in the behavioral literature. The remarkable expansion of knowledge in the past 20 years and the recent development of new technologies for physiological and behavioral experiments should permit direct tests of these theories in the coming decade.
The sensorimotor and social sides of the architecture of speech.
Pezzulo, Giovanni; Barca, Laura; D'Ausilio, Alessando
2014-12-01
Speech is a complex skill to master. In addition to sophisticated phono-articulatory abilities, speech acquisition requires neuronal systems configured for vocal learning, with adaptable sensorimotor maps that couple heard speech sounds with motor programs for speech production; imitation and self-imitation mechanisms that can train the sensorimotor maps to reproduce heard speech sounds; and a "pedagogical" learning environment that supports tutor learning.
Signal analysis of the female singing voice: Features for perceptual singer identity
NASA Astrophysics Data System (ADS)
Mellody, Maureen
2001-07-01
Individual singing voices tend to be easy for a listener to identify, particularly when compared to the difficulty of identifying the performer of any other musical instrument. What cues does a listener use to identify a particular singing voice? This work seeks to identify a set of features with which one can synthesize notes with the vocal quality of a particular singer. Such analysis and synthesis influences computer music (in the creation of synthetic sounds with different timbre), vocal pedagogy (as a training tool to help singers understand properties of their own voice as well as different professional-quality voices), and vocal health (to identify improper behavior in vocal production). The problem of singer identification is approached in three phases: signal analysis, the development of low- order representations, and perceptual evaluation. To perform the signal analysis, a high-resolution time- frequency distribution is applied to vowel tokens from sopranos and mezzo-sopranos. From these results, low- order representations are created for each singer's notes, which are used to synthesize sounds with the timbral quality of that singer. Finally, these synthesized sounds, along with original recordings, are evaluated by trained listeners in a variety of perceptual experiments to determine the extent to which the vocal quality of the desired singer is captured. Results from the signal analysis show that amplitude and frequency estimates extracted from the time-frequency signal analysis can be used to re-create each signal with little degradation in quality and no loss of perceptual identity. Low-order representations derived from the signal analysis are used in clustering and classification, which successfully clusters signals with corresponding singer identity. Finally, perceptual results indicate that trained listeners are, surprisingly, only modestly successful at correctly identifying the singer of a recording, and find the task to be particularly difficult for certain voices and extremely easy for others. Listeners also indicate that the majority of sounds synthesized with the low-order representations sufficiently capture the desired vocal timbre. Again, the task is easy for certain voices and much more difficult when evaluating other singers, consistent with the results from the original recordings.
Five fundamental constraints on theories of the origins of music.
Merker, Bjorn; Morley, Iain; Zuidema, Willem
2015-03-19
The diverse forms and functions of human music place obstacles in the way of an evolutionary reconstruction of its origins. In the absence of any obvious homologues of human music among our closest primate relatives, theorizing about its origins, in order to make progress, needs constraints from the nature of music, the capacities it engages, and the contexts in which it occurs. Here we propose and examine five fundamental constraints that bear on theories of how music and some of its features may have originated. First, cultural transmission, bringing the formal powers of cultural as contrasted with Darwinian evolution to bear on its contents. Second, generativity, i.e. the fact that music generates infinite pattern diversity by finite means. Third, vocal production learning, without which there can be no human singing. Fourth, entrainment with perfect synchrony, without which there is neither rhythmic ensemble music nor rhythmic dancing to music. And fifth, the universal propensity of humans to gather occasionally to sing and dance together in a group, which suggests a motivational basis endemic to our biology. We end by considering the evolutionary context within which these constraints had to be met in the genesis of human musicality. © 2015 The Author(s) Published by the Royal Society. All rights reserved.
Sleep, Off-Line Processing, and Vocal Learning
ERIC Educational Resources Information Center
Margoliash, Daniel; Schmidt, Marc F.
2010-01-01
The study of song learning and the neural song system has provided an important comparative model system for the study of speech and language acquisition. We describe some recent advances in the bird song system, focusing on the role of off-line processing including sleep in processing sensory information and in guiding developmental song…
Teaching Young Children How to Sing: One School's Experience
ERIC Educational Resources Information Center
Kenney, Susan
2011-01-01
In many schools, classroom teachers are responsible for the music experiences of young children. Children may learn songs, but may not learn "how" to sing. This article outlines simple teaching strategies to help young children develop listening and vocal habits leading to beautiful singing. The article discusses how the kindergarten classes at…
Population genetic structure and vocal dialects in an amazon parrot.
Wright, T F; Wilkinson, G S
2001-01-01
The relationship between cultural and genetic evolution was examined in the yellow-naped amazon Amazona auropalliata. This species has previously been shown to have regional dialects defined by large shifts in the acoustic structure of its learned contact call. Mitochondrial DNA sequence variation from a 680 base pair segment of the first domain of the control region was assayed in 41 samples collected from two neighbouring dialects in Costa Rica. The relationship of genetic variation to vocal variation was examined using haplotype analysis, genetic distance analysis, a maximum-likelihood estimator of migration rates and phylogenetic reconstructions. All analyses indicated a high degree of gene flow and, thus, individual dispersal across dialect boundaries. Calls sampled from sound libraries suggested that temporally stable contact call dialects occur throughout the range of the yellow-naped amazon, while the presence of similar dialects in the sister species Amazona ochrocephala suggests that the propensity to form dialects is ancestral in this clade. These results indicate that genes and culture are not closely associated in the yellow-naped amazon. Rather, they suggest that regional diversity in vocalizations is maintained by selective pressures that promote social learning and allow individual repertoires to conform to local call types. PMID:11297178
The respiratory-vocal system of songbirds: anatomy, physiology, and neural control.
Schmidt, Marc F; Martin Wild, J
2014-01-01
This wide-ranging review presents an overview of the respiratory-vocal system in songbirds, which are the only other vertebrate group known to display a degree of respiratory control during song rivalling that of humans during speech; this despite the fact that the peripheral components of both the respiratory and vocal systems differ substantially in the two groups. We first provide a brief description of these peripheral components in songbirds (lungs, air sacs and respiratory muscles, vocal organ (syrinx), upper vocal tract) and then proceed to a review of the organization of central respiratory-related neurons in the spinal cord and brainstem, the latter having an organization fundamentally similar to that of the ventral respiratory group of mammals. The second half of the review describes the nature of the motor commands generated in a specialized "cortical" song control circuit and how these might engage brainstem respiratory networks to shape the temporal structure of song. We also discuss a bilaterally projecting "respiratory-thalamic" pathway that links the respiratory system to "cortical" song control nuclei. This necessary pathway for song originates in the brainstem's primary inspiratory center and is hypothesized to play a vital role in synchronizing song motor commands both within and across hemispheres. © 2014 Elsevier B.V. All rights reserved.
The respiratory-vocal system of songbirds: Anatomy, physiology, and neural control
Schmidt, Marc F.; Wild, J. Martin
2015-01-01
This wide-ranging review presents an overview of the respiratory-vocal system in songbirds, which are the only other vertebrate group known to display a degree of respiratory control during song rivalling that of humans during speech; this despite the fact that the peripheral components of both the respiratory and vocal systems differ substantially in the two groups. We first provide a brief description of these peripheral components in songbirds (lungs, air sacs and respiratory muscles, vocal organ (syrinx), upper vocal tract) and then proceed to a review of the organization of central respiratory-related neurons in the spinal cord and brainstem, the latter having an organization fundamentally similar to that of the ventral respiratory group of mammals. The second half of the review describes the nature of the motor commands generated in a specialized “cortical” song control circuit and how these might engage brainstem respiratory networks to shape the temporal structure of song. We also discuss a bilaterally projecting “respiratory-thalamic” pathway that links the respiratory system to “cortical” song control nuclei. This necessary pathway for song originates in the brainstem’s primary inspiratory center and is hypothesized to play a vital role in synchronizing song motor commands both within and across hemispheres. PMID:25194204
Pathogenesis of vocal fold nodules: new insights from a modelling approach.
Dejonckere, Philippe H; Kob, Malte
2009-01-01
To give new insights into the pathogenesis of vocal fold nodules: (a) why the female/male ratio is so extreme, (b) how an hourglass-shaped vibration pattern - eliciting a localized microtrauma - originates, and (c) what the roles of muscular tension imbalance and of behavioral aspects are. Simulations with a 3-dimensional computer model of the vibrating vocal folds. (1) A slightly incomplete dorsal vocal fold adduction is a first condition for inducing an hourglass vibration pattern. (2) A limited collision zone is only possible with a small degree of curving of the rest position of the vocal fold edges in their ventral portion. This is an anatomical characteristic of the adult female larynx. Muscular fatigue and resulting hypotonia seem to enhance this curving. (3) If both these conditions are fulfilled, a sufficient vibration amplitude is required to achieve a localized impact. (4) This third condition can be obtained by an increased subglottal pressure and/or by a decrease in active stress of the tension forces between the neighboring vocalis masses. These last aspects incorporate muscular tension imbalance (dyskinesia) and behavioral aspects in the modelling process. Decrease in active stress is a possible effect of fatigue, and increase in subglottal pressure a result of effort compensation. Copyright 2009 S. Karger AG, Basel.
Facial expressions and the evolution of the speech rhythm.
Ghazanfar, Asif A; Takahashi, Daniel Y
2014-06-01
In primates, different vocalizations are produced, at least in part, by making different facial expressions. Not surprisingly, humans, apes, and monkeys all recognize the correspondence between vocalizations and the facial postures associated with them. However, one major dissimilarity between monkey vocalizations and human speech is that, in the latter, the acoustic output and associated movements of the mouth are both rhythmic (in the 3- to 8-Hz range) and tightly correlated, whereas monkey vocalizations have a similar acoustic rhythmicity but lack the concommitant rhythmic facial motion. This raises the question of how we evolved from a presumptive ancestral acoustic-only vocal rhythm to the one that is audiovisual with improved perceptual sensitivity. According to one hypothesis, this bisensory speech rhythm evolved through the rhythmic facial expressions of ancestral primates. If this hypothesis has any validity, we expect that the extant nonhuman primates produce at least some facial expressions with a speech-like rhythm in the 3- to 8-Hz frequency range. Lip smacking, an affiliative signal observed in many genera of primates, satisfies this criterion. We review a series of studies using developmental, x-ray cineradiographic, EMG, and perceptual approaches with macaque monkeys producing lip smacks to further investigate this hypothesis. We then explore its putative neural basis and remark on important differences between lip smacking and speech production. Overall, the data support the hypothesis that lip smacking may have been an ancestral expression that was linked to vocal output to produce the original rhythmic audiovisual speech-like utterances in the human lineage.
Dual Neural Network Model for the Evolution of Speech and Language.
Hage, Steffen R; Nieder, Andreas
2016-12-01
Explaining the evolution of speech and language poses one of the biggest challenges in biology. We propose a dual network model that posits a volitional articulatory motor network (VAMN) originating in the prefrontal cortex (PFC; including Broca's area) that cognitively controls vocal output of a phylogenetically conserved primary vocal motor network (PVMN) situated in subcortical structures. By comparing the connections between these two systems in human and nonhuman primate brains, we identify crucial biological preadaptations in monkeys for the emergence of a language system in humans. This model of language evolution explains the exclusiveness of non-verbal communication sounds (e.g., cries) in infants with an immature PFC, as well as the observed emergence of non-linguistic vocalizations in adults after frontal lobe pathologies. Copyright © 2016 Elsevier Ltd. All rights reserved.
Basal ganglia function, stuttering, sequencing, and repair in adult songbirds
Kubikova, Lubica; Bosikova, Eva; Cvikova, Martina; Lukacova, Kristina; Scharff, Constance; Jarvis, Erich D.
2014-01-01
A pallial-basal-ganglia-thalamic-pallial loop in songbirds is involved in vocal motor learning. Damage to its basal ganglia part, Area X, in adult zebra finches has been noted to have no strong effects on song and its function is unclear. Here we report that neurotoxic damage to adult Area X induced changes in singing tempo and global syllable sequencing in all animals, and considerably increased syllable repetition in birds whose song motifs ended with minor repetitions before lesioning. This stuttering-like behavior started at one month, and improved over six months. Unexpectedly, the lesioned region showed considerable recovery, including immigration of newly generated or repaired neurons that became active during singing. The timing of the recovery and stuttering suggest that immature recovering activity of the circuit might be associated with stuttering. These findings indicate that even after juvenile learning is complete, the adult striatum plays a role in higher level organization of learned vocalizations. PMID:25307086
Campbell's monkeys concatenate vocalizations into context-specific call sequences
Ouattara, Karim; Lemasson, Alban; Zuberbühler, Klaus
2009-01-01
Primate vocal behavior is often considered irrelevant in modeling human language evolution, mainly because of the caller's limited vocal control and apparent lack of intentional signaling. Here, we present the results of a long-term study on Campbell's monkeys, which has revealed an unrivaled degree of vocal complexity. Adult males produced six different loud call types, which they combined into various sequences in highly context-specific ways. We found stereotyped sequences that were strongly associated with cohesion and travel, falling trees, neighboring groups, nonpredatory animals, unspecific predatory threat, and specific predator classes. Within the responses to predators, we found that crowned eagles triggered four and leopards three different sequences, depending on how the caller learned about their presence. Callers followed a number of principles when concatenating sequences, such as nonrandom transition probabilities of call types, addition of specific calls into an existing sequence to form a different one, or recombination of two sequences to form a third one. We conclude that these primates have overcome some of the constraints of limited vocal control by combinatorial organization. As the different sequences were so tightly linked to specific external events, the Campbell's monkey call system may be the most complex example of ‘proto-syntax’ in animal communication known to date. PMID:20007377
Sounds of Modified Flight Feathers Reliably Signal Danger in a Pigeon.
Murray, Trevor G; Zeil, Jochen; Magrath, Robert D
2017-11-20
In his book on sexual selection, Darwin [1] devoted equal space to non-vocal and vocal communication in birds. Since then, vocal communication has become a model for studies of neurobiology, learning, communication, evolution, and conservation [2, 3]. In contrast, non-vocal "instrumental music," as Darwin called it, has only recently become subject to sustained inquiry [4, 5]. In particular, outstanding work reveals how feathers, often highly modified, produce distinctive sounds [6-9], and suggests that these sounds have evolved at least 70 times, in many orders [10]. It remains to be shown, however, that such sounds are signals used in communication. Here we show that crested pigeons (Ochyphaps lophotes) signal alarm with specially modified wing feathers. We used video and feather-removal experiments to demonstrate that the highly modified 8 th primary wing feather (P8) produces a distinct note during each downstroke. The sound changes with wingbeat frequency, so that birds fleeing danger produce wing sounds with a higher tempo. Critically, a playback experiment revealed that only if P8 is present does the sound of escape flight signal danger. Our results therefore indicate, nearly 150 years after Darwin's book, that modified feathers can be used for non-vocal communication, and they reveal an intrinsically reliable alarm signal. Copyright © 2017 Elsevier Ltd. All rights reserved.
Van Puyvelde, Martine; Loots, Gerrit; Gillisjans, Lobcke; Pattyn, Nathalie; Quintana, Carmen
2015-08-01
This study reports a cross-cultural comparison of the vocal pitch patterns of 15 Mexican Spanish-speaking and 15 Belgian Flemish-speaking dyads, recorded during 5min of free-play in a laboratory setting. Both cultures have a tradition of dyadic face-to-face interaction but differ in language origins (i.e., Romanic versus Germanic). In total, 374 Mexican and 558 Flemish vocal exchanges were identified, analyzed and compared for their incidence of tonal synchrony (harmonic/pentatonic series), non-tonal synchrony (with/without imitations) and pitch and/or interval imitations. The main findings revealed that dyads in both cultures rely on tonal synchrony using similar pitch ratios and timing patterns. However, there were significant differences in the infants' vocal pitch imitation behavior. Additional video-analyzes on the contingency patterns involved in pitch imitation showed a cross-cultural difference in the maternal selective reinforcement of pitch imitation. The results are interpreted with regard to linguistic, developmental and cultural aspects and the 'musilanguage' model. Copyright © 2015 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Soltis, Joseph M.; Savage, Anne; Leong, Kirsten M.
2004-05-01
The most commonly occurring elephant vocalization is the rumble, a frequency-modulated call with infrasonic components. Upwards of ten distinct rumble subtypes have been proposed, but little quantitative work on the acoustic properties of rumbles has been conducted. Rumble vocalizations (N=269) from six females housed at Disney's Animal Kingdom were analyzed. Vocalizations were recorded from microphones in collars around subject necks, and rumbles were digitized and measured using SIGNAL software. Sixteen acoustic variables were measured for each call, extracting both source and filter features. Multidimensional scaling analysis indicates that there are no acoustically distinct rumble subtypes, but that there is quantitative variation across rumbles. Discriminant function analysis showed that the acoustic characteristics of rumbles differ across females. A classification success rate of 65% was achieved when assigning unselected rumbles to one of the six females (test set =64 calls) according to the functions derived from the originally selected calls (training set =205 calls). The rumble is best viewed as a single call type with graded variation, but information regarding individual identity is encoded in female rumbles.
Self-Organization: Complex Dynamical Systems in the Evolution of Speech
NASA Astrophysics Data System (ADS)
Oudeyer, Pierre-Yves
Human vocalization systems are characterized by complex structural properties. They are combinatorial, based on the systematic reuse of phonemes, and the set of repertoires in human languages is characterized by both strong statistical regularities—universals—and a great diversity. Besides, they are conventional codes culturally shared in each community of speakers. What are the origins of the forms of speech? What are the mechanisms that permitted their evolution in the course of phylogenesis and cultural evolution? How can a shared speech code be formed in a community of individuals? This chapter focuses on the way the concept of self-organization, and its interaction with natural selection, can throw light on these three questions. In particular, a computational model is presented which shows that a basic neural equipment for adaptive holistic vocal imitation, coupling directly motor and perceptual representations in the brain, can generate spontaneously shared combinatorial systems of vocalizations in a society of babbling individuals. Furthermore, we show how morphological and physiological innate constraints can interact with these self-organized mechanisms to account for both the formation of statistical regularities and diversity in vocalization systems.
Peh, Wendy Y X; Roberts, Todd F; Mooney, Richard
2015-04-08
Vocal communication depends on the coordinated activity of sensorimotor neurons important to vocal perception and production. How vocalizations are represented by spatiotemporal activity patterns in these neuronal populations remains poorly understood. Here we combined intracellular recordings and two-photon calcium imaging in anesthetized adult zebra finches (Taeniopygia guttata) to examine how learned birdsong and its component syllables are represented in identified projection neurons (PNs) within HVC, a sensorimotor region important for song perception and production. These experiments show that neighboring HVC PNs can respond at markedly different times to song playback and that different syllables activate spatially intermingled PNs within a local (~100 μm) region of HVC. Moreover, noise correlations were stronger between PNs that responded most strongly to the same syllable and were spatially graded within and between classes of PNs. These findings support a model in which syllabic and temporal features of song are represented by spatially intermingled PNs functionally organized into cell- and syllable-type networks within local spatial scales in HVC. Copyright © 2015 the authors 0270-6474/15/355589-17$15.00/0.
Shi, Zhimin; Luo, Guanzheng; Fu, Lijuan; Fang, Zhide; Wang, XiuJie; Li, XiaoChing
2013-10-16
Mutations in the FOXP2 gene cause speech and language impairments, accompanied by structural and functional abnormalities in brain regions underlying speech-related sensory-motor processing, including the striatum and cerebellum. The sequence and expression patterns of FOXP2 are highly conserved among higher vertebrates. In the zebra finch brain, FoxP2 is expressed in Area X, a striatal nucleus required for vocal learning, and reduced FoxP2 expression impairs dendritic development and vocal learning. The FoxP2 gene encodes a transcription factor that controls the expression of many downstream genes. However, how FOXP2 gene expression is regulated is not clearly understood. miRNAs regulate gene expression post-transcriptionally by targeting the 3'-untranslated regions (UTRs) of mRNAs, leading to translational suppression or mRNA degradation. In this study, we identified miR-9 and miR-140-5p as potential regulators of the FoxP2 gene. We show that both miR-9 and miR-140-5p target specific sequences in the FoxP2 3'-UTR and downregulate FoxP2 protein and mRNA expression in vitro. We also show that the expression of miR-9 and miR-140-5p in Area X of the zebra finch brain is regulated during song development in juvenile zebra finches. We further show that in adult zebra finches the expression of miR-9 and miR-140-5p in Area X is regulated as a function of the social context of song behavior in males singing undirected songs. Our findings reveal a post-transcriptional mechanism that regulates FoxP2 expression and suggest that social vocal behavior can influence the basal ganglia circuit controlling vocal learning via a miRNA-FoxP2 gene regulatory network.
Shi, Zhimin; Luo, Guanzheng; Fu, Lijuan; Fang, Zhide; Wang, XiuJie
2013-01-01
Mutations in the FOXP2 gene cause speech and language impairments, accompanied by structural and functional abnormalities in brain regions underlying speech-related sensory-motor processing, including the striatum and cerebellum. The sequence and expression patterns of FOXP2 are highly conserved among higher vertebrates. In the zebra finch brain, FoxP2 is expressed in Area X, a striatal nucleus required for vocal learning, and reduced FoxP2 expression impairs dendritic development and vocal learning. The FoxP2 gene encodes a transcription factor that controls the expression of many downstream genes. However, how FOXP2 gene expression is regulated is not clearly understood. miRNAs regulate gene expression post-transcriptionally by targeting the 3′-untranslated regions (UTRs) of mRNAs, leading to translational suppression or mRNA degradation. In this study, we identified miR-9 and miR-140-5p as potential regulators of the FoxP2 gene. We show that both miR-9 and miR-140-5p target specific sequences in the FoxP2 3′-UTR and downregulate FoxP2 protein and mRNA expression in vitro. We also show that the expression of miR-9 and miR-140-5p in Area X of the zebra finch brain is regulated during song development in juvenile zebra finches. We further show that in adult zebra finches the expression of miR-9 and miR-140-5p in Area X is regulated as a function of the social context of song behavior in males singing undirected songs. Our findings reveal a post-transcriptional mechanism that regulates FoxP2 expression and suggest that social vocal behavior can influence the basal ganglia circuit controlling vocal learning via a miRNA-FoxP2 gene regulatory network. PMID:24133256
Can a bird brain do phonology?
Samuels, Bridget D.
2015-01-01
A number of recent studies have revealed correspondences between song- and language-related neural structures, pathways, and gene expression in humans and songbirds. Analyses of vocal learning, song structure, and the distribution of song elements have similarly revealed a remarkable number of shared characteristics with human speech. This article reviews recent developments in the understanding of these issues with reference to the phonological phenomena observed in human language. This investigation suggests that birds possess a host of abilities necessary for human phonological computation, as evidenced by behavioral, neuroanatomical, and molecular genetic studies. Vocal-learning birds therefore present an excellent model for studying some areas of human phonology, though differences in the primitives of song and language as well as the absence of a human-like morphosyntax make human phonology differ from birdsong phonology in crucial ways. PMID:26284006
Using Videos and Multimodal Discourse Analysis to Study How Students Learn a Trade
ERIC Educational Resources Information Center
Chan, Selena
2013-01-01
The use of video to assist with ethnographical-based research is not a new phenomenon. Recent advances in technology have reduced the costs and technical expertise required to use videos for gathering research data. Audio-visual records of learning activities as they take place, allow for many non-vocal and inter-personal communication…
Assessing Middle School Student Participation in Online vs. Face-to-Face Environments
ERIC Educational Resources Information Center
Oravetz, Cathleen L.
2011-01-01
Educators have observed reluctance in middle school students to vocally engage in small group learning tasks, the result of which could be a decrease in student learning. The same students have been observed collaborating with peers outside of the classroom when using technology. The purpose of this study is to determine if technology provides a…
An Evaluation of Preference for Mode of Instruction Following Variations in Response Effort
ERIC Educational Resources Information Center
Romani, Patrick W.; McCoy, Thomasin E.; Wacker, David P.; Padilla-Dalmau, Yaniz C.
2014-01-01
The current study evaluated preference for mode of instruction (i.e., visual or vocal) for four children diagnosed with a language-based learning disability. Each participant was an elementary student who was initially referred to a neuropsychology clinic specializing in learning disabilities. As a part of the evaluation, measures of each…
Classroom Audio Distribution in the Postsecondary Setting: A Story of Universal Design for Learning
ERIC Educational Resources Information Center
Flagg-Williams, Joan B.; Bokhorst-Heng, Wendy D.
2016-01-01
Classroom Audio Distribution Systems (CADS) consist of amplification technology that enhances the teacher's, or sometimes the student's, vocal signal above the background noise in a classroom. Much research has supported the benefits of CADS for student learning, but most of it has focused on elementary school classrooms. This study investigated…
Learning in Communities of Inquiry: A Review of the Literature
ERIC Educational Resources Information Center
Rourke, Liam; Kanuka, Heather
2009-01-01
The purpose of this study was to investigate learning in communities of inquiry (CoI) as the terms are defined in Garrison, Anderson, and Archer's (2000) framework. We identified 252 reports from 2000-2008 that referenced the framework, and we reviewed them using Ogawan and Malen's (1991) strategy for synthesizing multi-vocal bodies of literature.…
Watts, Christopher R
2016-01-01
Reduced vocal intensity is a core impairment of hypokinetic dysarthria in Parkinson's disease (PD). Speech treatments have been developed to rehabilitate the vocal subsystems underlying this impairment. Intensive treatment programs requiring high-intensity voice and speech exercises with clinician-guided prompting and feedback have been established as effective for improving vocal function. Less is known, however, regarding long-term outcomes of clinical benefit in speakers with PD who receive these treatments. A retrospective cohort design was utilized. Data from 78 patient files across a three year period were analyzed. All patients received a structured, intensive program of voice therapy focusing on speaking intent and loudness. The dependent variable for all analyses was vocal intensity in decibels (dBSPL). Vocal intensity during sustained vowel production, reading, and novel conversational speech was compared at pre-treatment, post-treatment, six month follow-up, and twelve month follow-up periods. Statistically significant increases in vocal intensity were found at post-treatment, 6 months, and 12 month follow-up periods with intensity gains ranging from 5 to 17 dB depending on speaking condition and measurement period. Significant treatment effects were found in all three speaking conditions. Effect sizes for all outcome measures were large, suggesting a strong degree of practical significance. Significant increases in vocal intensity measured at 6 and 12 moth follow-up periods suggested that the sample of patients maintained treatment benefit for up to a year. These findings are supported by outcome studies reporting treatment outcomes within a few months post-treatment, in addition to prior studies that have reported long-term outcome results. The positive treatment outcomes experienced by the PD cohort in this study are consistent with treatment responses subsequent to other treatment approaches which focus on high-intensity, clinician guided motor learning for voice and speech production in PD. Theories regarding the underlying neurophysiological response to treatment will be discussed.
Heterospecific eavesdropping in ant-following birds of the Neotropics is a learned behaviour.
Pollock, Henry S; Martínez, Ari E; Kelley, J Patrick; Touchton, Janeene M; Tarwater, Corey E
2017-10-25
Animals eavesdrop on other species to obtain information about their environments. Heterospecific eavesdropping can yield tangible fitness benefits by providing valuable information about food resources and predator presence. The ability to eavesdrop may therefore be under strong selection, although extensive research on alarm-calling in avian mixed-species flocks has found only limited evidence that close association with another species could select for innate signal recognition. Nevertheless, very little is known about the evolution of eavesdropping behaviour and the mechanism of heterospecific signal recognition, particularly in other ecological contexts, such as foraging. To understand whether heterospecific eavesdropping was an innate or learned behaviour in a foraging context, we studied heterospecific signal recognition in ant-following birds of the Neotropics, which eavesdrop on vocalizations of obligate ant-following species to locate and recruit to swarms of the army ant Eciton burchellii , a profitable food resource. We used a playback experiment to compare recruitment of ant-following birds to vocalizations of two obligate species at a mainland site (where both species are present) and a nearby island site (where one species remains whereas the other went extinct approx. 40 years ago). We found that ant-following birds recruited strongly to playbacks of the obligate species present at both island and mainland sites, but the island birds did not recruit to playbacks of the absent obligate species. Our results strongly suggest that (i) ant-following birds learn to recognize heterospecific vocalizations from ecological experience and (ii) island birds no longer recognize the locally extinct obligate species after eight generations of absence from the island. Although learning appears to be the mechanism of heterospecific signal recognition in ant-following birds, more experimental tests are needed to fully understand the evolution of eavesdropping behaviour. © 2017 The Author(s).
Central pattern generator for vocalization: is there a vertebrate morphotype?
Bass, Andrew H
2014-10-01
Animals that generate acoustic signals for social communication are faced with two essential tasks: generate a temporally precise signal and inform the auditory system about the occurrence of one's own sonic signal. Recent studies of sound producing fishes delineate a hindbrain network comprised of anatomically distinct compartments coding equally distinct neurophysiological properties that allow an organism to meet these behavioral demands. A set of neural characters comprising a vocal-sonic central pattern generator (CPG) morphotype is proposed for fishes and tetrapods that shares evolutionary developmental origins with pectoral appendage motor systems. Copyright © 2014 Elsevier Ltd. All rights reserved.
Central pattern generator for vocalization: Is there a vertebrate morphotype?
Bass, Andrew H.
2014-01-01
Animals that generate acoustic signals for social communication are faced with two essential tasks: generate a temporally precise signal and inform the auditory system about the occurrence of one’s own sonic signal. Recent studies of sound producing fishes delineate a hindbrain network comprised of anatomically distinct compartments coding equally distinct neurophysiological properties that allow an organism to meet these behavioral demands. A set of neural characters comprising a vocal-sonic central pattern generator (CPG) morphotype is proposed for fishes and tetrapods that shares evolutionary developmental origins with pectoral appendage motor systems. PMID:25050813
[The role of laryngeal electromyography in the diagnosis of vocal cord movement disorders].
Bach, Ádám; Sztanó, Balázs; Kiss, József Géza; Volk, Gerd Fabian; Müller, Andreas; Pototschnig, Claus; Rovó, László
2018-02-01
The development of the therapeutic possibilities of vocal cord immobility necessitated the parallel renewal of diagnostic methods. In the last years, laryngeal electromyography, which was first introduced more than 70 years ago, has been re-discovered. After reviewing the international literature and their own experience, the authors present the indications, technical requirements, method and, particularly, the evaluation of the results of this procedure. Laryngeal electromyography makes the differentiation between mechanical fixation and immobility with neurological origin of the vocal folds possible. In case of laryngeal paralysis/paresis it also evaluates objectively the severity of neural injury, the prognosis of the disease and the necessity of any glottis-widening procedure. The widespread application of dynamic rehabilitation interventions is not conceivable without the routine application of laryngeal electromyography, so this sensitive diagnostic tool has to be introduced in all laryngological centers. Orv Hetil. 2018; 159(8): 303-311.
Baker, Janet
2002-01-01
The projected speaking voice and the singing voice are highly sensitive to external and internal influences, and teachers of spoken voice and singing are in a unique position to identify subtle and more serious vocal difficulties in their students. Persistent anomalies may herald early onset of changes in vocal fold structure, neurophysiological control, or emotional stability. Two cases are presented to illustrate the benefits of a collaborative approach to diagnosis and management. The first, a 21-year-old male drama and singing student with an abnormally high speaking voice and falsetto singing voice was found to have a psychogenic dysphonia referred to as "puberphonia" or "mutational falsetto". The second, a 34-year-old female alto with strained phonation and perceived stutter of the vocal folds was diagnosed with "adductor spasmodic dysphonia" or "focal laryngeal dystonia" of neurological origin.
Ma, Jie; Kanwal, Jagmeet S.
2014-01-01
The neural substrate for the perception of vocalizations is relatively well described, but how their timing and specificity are tightly coupled with accompanying physiological changes and context-appropriate behaviors remains unresolved. We hypothesized that temporally integrated vocal and emotive responses, especially the expression of fear, vigilance and aggression, originate within the amygdala. To test this hypothesis, we performed electrical microstimulation at 461 highly restricted loci within the basal and central amygdala in awake mustached bats. At a subset of these sites, high frequency stimulation with weak constant current pulses presented at near-threshold levels triggered vocalization of either echolocation pulses or social calls. At the vast majority of locations, microstimulation produced a constellation of changes in autonomic and somatomotor outputs. These changes included widespread co-activation of significant tachycardia and hyperventilation and/or rhythmic ear pinna movements (PMs). In a few locations, responses were constrained to vocalization and/or PMs despite increases in the intensity of stimulation. The probability of eliciting echolocation pulses vs. social calls decreased in a medial-posterior to anterolateral direction within the centrobasal amygdala. Microinjections of kainic acid (KA) at stimulation sites confirmed the contribution of cellular activity rather than fibers-of-passage in the control of multimodal outputs. The results suggest that localized clusters of neurons may simultaneously modulate the activity of multiple central pattern generators (CPGs) present within the brainstem. PMID:24624089
Ma, Jie; Kanwal, Jagmeet S
2014-01-01
The neural substrate for the perception of vocalizations is relatively well described, but how their timing and specificity are tightly coupled with accompanying physiological changes and context-appropriate behaviors remains unresolved. We hypothesized that temporally integrated vocal and emotive responses, especially the expression of fear, vigilance and aggression, originate within the amygdala. To test this hypothesis, we performed electrical microstimulation at 461 highly restricted loci within the basal and central amygdala in awake mustached bats. At a subset of these sites, high frequency stimulation with weak constant current pulses presented at near-threshold levels triggered vocalization of either echolocation pulses or social calls. At the vast majority of locations, microstimulation produced a constellation of changes in autonomic and somatomotor outputs. These changes included widespread co-activation of significant tachycardia and hyperventilation and/or rhythmic ear pinna movements (PMs). In a few locations, responses were constrained to vocalization and/or PMs despite increases in the intensity of stimulation. The probability of eliciting echolocation pulses vs. social calls decreased in a medial-posterior to anterolateral direction within the centrobasal amygdala. Microinjections of kainic acid (KA) at stimulation sites confirmed the contribution of cellular activity rather than fibers-of-passage in the control of multimodal outputs. The results suggest that localized clusters of neurons may simultaneously modulate the activity of multiple central pattern generators (CPGs) present within the brainstem.
Cysneiros, Helena Renata Silva; Leal, Mariana de Carvalho; Lucena, Jonia Alves; Muniz, Lilian Ferreira
To conduct a systematic review of the scientific literature studying the relationship between vocal production and auditory perception in cochlear implant users. This is an integrative systematic review. The plattforms/databases Bireme, SciELO, Cochrane, Scopus and Web of Science were consulted and the descriptors used were voice, cochlear implant and auditory perception. Original papers published in English, French, Spanish or Portuguese involving the study of vocal production and auditory perception in cochlear implant users were selected and there was no restriction about year of publication of the articles. The studies selected were analyzed according to the author, location, year and publication of the article, as well as for their sample size, type of vocal production and auditory perception assessment and for its major findings and recommendation grade/level of scientific evidence. The results suggest the existence of positive relationship between vocal production and auditory perception in cochlear implant users, and indicate that the deployment time has a positive influence in this relationship. None of the selected studies were rated at level 1 of scientific evidence or grade A of recommendation, which is related to the methodological approach that can go with this subject matter. There is great lack of publications relating auditory perception and speech production in cochlear implant users. This gap is even greater when it comes to the adult population.
Wallez, Catherine; Schaeffer, Jennifer; Meguerditchian, Adrien; Vauclair, Jacques; Schapiro, Steven J.; Hopkins, William D.
2013-01-01
Studies involving oro-facial asymmetries in nonhuman primates have largely demonstrated a right hemispheric dominance for communicative signals and conveyance of emotional information. A recent study on chimpanzee reported the first evidence of significant left-hemispheric dominance when using attention-getting sounds and rightward bias for species-typical vocalizations (Losin, Russell, Freeman, Meguerditchian, Hopkins & Fitch, 2008). The current study sought to extend the findings from Losin et al. (2008) with additional oro-facial assessment in a new colony of chimpanzees. When combining the two populations, the results indicated a consistent leftward bias for attention-getting sounds and a right lateralization for species-typical vocalizations. Collectively, the results suggest that both voluntary- controlled oro-facial and gestural communication might share the same left-hemispheric specialization and might have coevolved into a single integrated system present in a common hominid ancestor. PMID:22867751
Involvement of the avian song system in reproductive behaviour
Wild, J. Martin; Botelho, João F.
2015-01-01
The song system of songbirds consists of an interconnected set of forebrain nuclei that has traditionally been regarded as dedicated to the learning and production of song. Here, however, we suggest that the song system could also influence muscles used in reproductive behaviour, such as the cloacal sphincter muscle. We show that the same medullary nucleus, retroambigualis (RAm), that projects upon spinal motoneurons innervating expiratory muscles (which provide the pressure head for vocalization) and upon vocal motoneurons for respiratory–vocal coordination also projects upon cloacal motoneurons. Furthermore, RAm neurons projecting to sacral spinal levels were shown to receive direct projections from nucleus robustus arcopallialis (RA) of the forebrain song system. Thus, by indicating a possible disynaptic relationship between RA and motoneurons innervating the reproductive organ, in both males and females, these results potentially extend the role of the song system to include consummatory as well as appetitive aspects of reproductive behaviour. PMID:26631245
Flight calls signal group and individual identity but not kinship in a cooperatively breeding bird.
Keen, Sara C; Meliza, C Daniel; Rubenstein, Dustin R
2013-11-01
In many complex societies, intricate communication and recognition systems may evolve to help support both direct and indirect benefits of group membership. In cooperatively breeding species where groups typically comprise relatives, both learned and innate vocal signals may serve as reliable cues for kin recognition. Here, we investigated vocal communication in the plural cooperatively breeding superb starling, Lamprotornis superbus , where flight calls-short, stereotyped vocalizations used when approaching conspecifics-may communicate kin relationships, group membership, and/or individual identity. We found that flight calls were most similar within individual repertoires but were also more similar within groups than within the larger population. Although starlings responded differently to playback of calls from their own versus other neighboring and distant social groups, call similarity was uncorrelated with genetic relatedness. Additionally, immigrant females showed similar patterns to birds born in the study population. Together, these results suggest that flight calls are learned signals that reflect social association but may also carry a signal of individuality. Flight calls, therefore, provide a reliable recognition mechanism for groups and may also be used to recognize individuals. In complex societies comprising related and unrelated individuals, signaling individuality and group association, rather than kinship, may be a route to cooperation.
Undergraduate Students' Ownership of Musical Learning: Obstacles and Options in One-to-One Teaching
ERIC Educational Resources Information Center
Johansson, Karin
2013-01-01
This paper describes a longitudinal, collaborative case study, made in the framework of the project Students' Ownership of Learning (SOL) during one academic year with one vocal teacher and two female students. The aim of the study was to relate the interaction between the teacher's and the students' intentions and expectations to…
What songbirds teach us about learning
NASA Astrophysics Data System (ADS)
Brainard, Michael S.; Doupe, Allison J.
2002-05-01
Bird fanciers have known for centuries that songbirds learn their songs. This learning has striking parallels to speech acquisition: like humans, birds must hear the sounds of adults during a sensitive period, and must hear their own voice while learning to vocalize. With the discovery and investigation of discrete brain structures required for singing, songbirds are now providing insights into neural mechanisms of learning. Aided by a wealth of behavioural observations and species diversity, studies in songbirds are addressing such basic issues in neuroscience as perceptual and sensorimotor learning, developmental regulation of plasticity, and the control and function of adult neurogenesis.
Sons learn songs from their social fathers in a cooperatively breeding bird
Greig, Emma I.; Taft, Benjamin N.; Pruett-Jones, Stephen
2012-01-01
Song learning is hypothesized to allow social adaptation to a local song neighbourhood. Maintaining social associations is particularly important in cooperative breeders, yet vocal learning in such species has only been assessed in systems where social association was correlated with relatedness. Thus, benefits of vocal learning as a means of maintaining social associations could not be disentangled from benefits of kin recognition. We assessed genetic and cultural contributions to song in a species where social association was not strongly correlated with kinship: the cooperatively breeding, reproductively promiscuous splendid fairy-wren (Malurus splendens). We found that song characters of socially associated father–son pairs were more strongly correlated (and thus songs were more similar) than songs of father–son pairs with a genetic, but no social, association (i.e. cuckolding fathers). Song transmission was, therefore, vertical and cultural, with minimal signatures of kinship. Additionally, song characters were not correlated with several phenotypic indicators of male quality, supporting the idea that there may be a tradeoff between accurate copying of tutors and quality signalling via maximizing song performance, particularly when social and genetic relationships are decoupled. Our results lend support to the hypothesis that song learning facilitates the maintenance of social associations by permitting unrelated individuals to acquire similar signal phenotypes. PMID:22593105
Sons learn songs from their social fathers in a cooperatively breeding bird.
Greig, Emma I; Taft, Benjamin N; Pruett-Jones, Stephen
2012-08-22
Song learning is hypothesized to allow social adaptation to a local song neighbourhood. Maintaining social associations is particularly important in cooperative breeders, yet vocal learning in such species has only been assessed in systems where social association was correlated with relatedness. Thus, benefits of vocal learning as a means of maintaining social associations could not be disentangled from benefits of kin recognition. We assessed genetic and cultural contributions to song in a species where social association was not strongly correlated with kinship: the cooperatively breeding, reproductively promiscuous splendid fairy-wren (Malurus splendens). We found that song characters of socially associated father-son pairs were more strongly correlated (and thus songs were more similar) than songs of father-son pairs with a genetic, but no social, association (i.e. cuckolding fathers). Song transmission was, therefore, vertical and cultural, with minimal signatures of kinship. Additionally, song characters were not correlated with several phenotypic indicators of male quality, supporting the idea that there may be a tradeoff between accurate copying of tutors and quality signalling via maximizing song performance, particularly when social and genetic relationships are decoupled. Our results lend support to the hypothesis that song learning facilitates the maintenance of social associations by permitting unrelated individuals to acquire similar signal phenotypes.
Assessing the uniqueness of language: Animal grammatical abilities take center stage.
Ten Cate, Carel
2017-02-01
Questions related to the uniqueness of language can only be addressed properly by referring to sound knowledge of the relevant cognitive abilities of nonhuman animals. A key question concerns the nature and extent of animal rule-learning abilities. I discuss two approaches used to assess these abilities. One is comparing the structures of animal vocalizations to linguistic ones, and another is addressing the grammatical rule- and pattern-learning abilities of animals through experiments using artificial grammars. Neither of these approaches has so far provided unambiguous evidence of advanced animal abilities. However, when we consider how animal vocalizations are analyzed, the types of stimuli and tasks that are used in artificial grammar learning experiments, the limited number of species examined, and the groups to which these belong, I argue that the currently available evidence is insufficient to arrive at firm conclusions concerning the limitations of animal grammatical abilities. As a consequence, the gap between human linguistic rule-learning abilities and those of nonhuman animals may be smaller and less clear than is currently assumed. This means that it is still an open question whether a difference in the rule-learning and rule abstraction abilities between animals and humans played the key role in the evolution of language.
Echolocating bats rely on audiovocal feedback to adapt sonar signal design.
Luo, Jinhong; Moss, Cynthia F
2017-10-10
Many species of bat emit acoustic signals and use information carried by echoes reflecting from nearby objects to navigate and forage. It is widely documented that echolocating bats adjust the features of sonar calls in response to echo feedback; however, it remains unknown whether audiovocal feedback contributes to sonar call design. Audiovocal feedback refers to the monitoring of one's own vocalizations during call production and has been intensively studied in nonecholocating animals. Audiovocal feedback not only is a necessary component of vocal learning but also guides the control of the spectro-temporal structure of vocalizations. Here, we show that audiovocal feedback is directly involved in the echolocating bat's control of sonar call features. As big brown bats tracked targets from a stationary position, we played acoustic jamming signals, simulating calls of another bat, timed to selectively perturb audiovocal feedback or echo feedback. We found that the bats exhibited the largest call-frequency adjustments when the jamming signals occurred during vocal production. By contrast, bats did not show sonar call-frequency adjustments when the jamming signals coincided with the arrival of target echoes. Furthermore, bats rapidly adapted sonar call design in the first vocalization following the jamming signal, revealing a response latency in the range of 66 to 94 ms. Thus, bats, like songbirds and humans, rely on audiovocal feedback to structure sonar signal design.
A Mozart is not a Pavarotti: singers outperform instrumentalists on foreign accent imitation
Christiner, Markus; Reiterer, Susanne Maria
2015-01-01
Recent findings have shown that people with higher musical aptitude were also better in oral language imitation tasks. However, whether singing capacity and instrument playing contribute differently to the imitation of speech has been ignored so far. Research has just recently started to understand that instrumentalists develop quite distinct skills when compared to vocalists. In the same vein the role of the vocal motor system in language acquisition processes has poorly been investigated as most investigations (neurobiological and behavioral) favor to examine speech perception. We set out to test whether the vocal motor system can influence an ability to learn, produce and perceive new languages by contrasting instrumentalists and vocalists. Therefore, we investigated 96 participants, 27 instrumentalists, 33 vocalists and 36 non-musicians/non-singers. They were tested for their abilities to imitate foreign speech: unknown language (Hindi), second language (English) and their musical aptitude. Results revealed that both instrumentalists and vocalists have a higher ability to imitate unintelligible speech and foreign accents than non-musicians/non-singers. Within the musician group, vocalists outperformed instrumentalists significantly. Conclusion: First, adaptive plasticity for speech imitation is not reliant on audition alone but also on vocal-motor induced processes. Second, vocal flexibility of singers goes together with higher speech imitation aptitude. Third, vocal motor training, as of singers, may speed up foreign language acquisition processes. PMID:26379537
A Mozart is not a Pavarotti: singers outperform instrumentalists on foreign accent imitation.
Christiner, Markus; Reiterer, Susanne Maria
2015-01-01
Recent findings have shown that people with higher musical aptitude were also better in oral language imitation tasks. However, whether singing capacity and instrument playing contribute differently to the imitation of speech has been ignored so far. Research has just recently started to understand that instrumentalists develop quite distinct skills when compared to vocalists. In the same vein the role of the vocal motor system in language acquisition processes has poorly been investigated as most investigations (neurobiological and behavioral) favor to examine speech perception. We set out to test whether the vocal motor system can influence an ability to learn, produce and perceive new languages by contrasting instrumentalists and vocalists. Therefore, we investigated 96 participants, 27 instrumentalists, 33 vocalists and 36 non-musicians/non-singers. They were tested for their abilities to imitate foreign speech: unknown language (Hindi), second language (English) and their musical aptitude. Results revealed that both instrumentalists and vocalists have a higher ability to imitate unintelligible speech and foreign accents than non-musicians/non-singers. Within the musician group, vocalists outperformed instrumentalists significantly. First, adaptive plasticity for speech imitation is not reliant on audition alone but also on vocal-motor induced processes. Second, vocal flexibility of singers goes together with higher speech imitation aptitude. Third, vocal motor training, as of singers, may speed up foreign language acquisition processes.
Petekkaya, Emine; Yücel, Ahmet Hilmi; Sürmelioğlu, Özgür
2017-12-28
Opera and chant singers learn to effectively use aerodynamic components by breathing exercises during their education. Aerodynamic components, including subglottic air pressure and airflow, deteriorate in voice disorders. This study aimed to evaluate the changes in aerodynamic parameters and supraglottic structures of men and women with different vocal registers who are in an opera and chant education program. Vocal acoustic characteristics, aerodynamic components, and supraglottic structures were evaluated in 40 opera and chant art branch students. The majority of female students were sopranos, and the male students were baritone or tenor vocalists. The acoustic analyses revealed that the mean fundamental frequency was 152.33 Hz in the males and 218.77 Hz in the females. The estimated mean subglottal pressures were similar in females (14.99 cmH 2 O) and in males (14.48 cmH 2 O). Estimated mean airflow rates were also similar in both groups. The supraglottic structure compression analyses revealed partial anterior-posterior compressions in 2 tenors and 2 sopranos, and false vocal fold compression in 2 sopranos. Opera music is sung in high-pitched sounds. Attempts to sing high-pitched notes and frequently using register transitions overstrain the vocal structures. This intense muscular effort eventually traumatizes the vocal structures and causes supraglottic activity. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
38 CFR 21.7120 - Courses included in programs of education.
Code of Federal Regulations, 2011 CFR
2011-07-01
... (ii) Any music course, instrumental or vocal, public speaking course or courses in dancing, sports or... public speaking which are offered by institutions of higher learning for credit as an integral part of a...
Godino-Llorente, J I; Gómez-Vilda, P
2004-02-01
It is well known that vocal and voice diseases do not necessarily cause perceptible changes in the acoustic voice signal. Acoustic analysis is a useful tool to diagnose voice diseases being a complementary technique to other methods based on direct observation of the vocal folds by laryngoscopy. Through the present paper two neural-network based classification approaches applied to the automatic detection of voice disorders will be studied. Structures studied are multilayer perceptron and learning vector quantization fed using short-term vectors calculated accordingly to the well-known Mel Frequency Coefficient cepstral parameterization. The paper shows that these architectures allow the detection of voice disorders--including glottic cancer--under highly reliable conditions. Within this context, the Learning Vector quantization methodology demonstrated to be more reliable than the multilayer perceptron architecture yielding 96% frame accuracy under similar working conditions.
Nimodipine alters acquisition of a visual discrimination task in chicks.
Deyo, R; Panksepp, J; Conner, R L
1990-03-01
Chicks 5 days old received intraperitoneal injections of nimodipine 30 min before training on either a visual discrimination task (0, 0.5, 1.0, or 5.0 mg/kg) or a test of separation-induced distress vocalizations (0, 0.5, or 2.5 mg/kg). Chicks receiving 1.0 mg/kg nimodipine made significantly fewer visual discrimination errors than vehicle controls by trials 41-60, but did not differ from controls 24 h later. Chicks in the 5 mg/kg group made significantly more errors when compared to controls both during acquisition of the task and during retention. Nimodipine did not alter separation-induced distress vocalizations at any of the doses tested, suggesting that nimodipine's effects on learning cannot be attributed to a reduction in separation distress. These data indicate that nimodipine's facilitation of learning in young subjects is dose dependent, but nimodipine failed to enhance retention.
Chironomic stylization of intonation.
d'Alessandro, Christophe; Rilliard, Albert; Le Beux, Sylvain
2011-03-01
Intonation stylization is studied using "chironomy," i.e., the analogy between hand gestures and prosodic movements. An intonation mimicking paradigm is used. The task of the ten subjects is to copy the intonation pattern of sentences with the help of a stylus on a graphic tablet, using a system for real-time manual intonation modification. Gestural imitation is compared to vocal imitation of the same sentences (seven for a male speaker, seven for a female speaker). Distance measures between gestural copies, vocal imitations, and original sentences are computed for performance assessment. Perceptual testing is also used for assessing the quality of gestural copies. The perceptual difference between natural and stylized contours is measured using a mean opinion score paradigm for 15 subjects. The results indicate that intonation contours can be stylized with accuracy by chironomic imitation. The results of vocal imitation and chironomic imitation are comparable, but subjects show better imitation results in vocal imitation. The best stylized contours using chironomy seems perceptually indistinguishable or almost indistinguishable from natural contours, particularly for female speech. This indicates that chironomic stylization is effective, and that hand movements can be analogous to intonation movements. © 2011 Acoustical Society of America
Focquet, A; Péréon, Y; Ségura, S; Ferron, C; Malard, O; Espitalier, F
2017-02-01
To study the diagnostic and prognostic contribution of laryngeal electromyography in unilateral vocal-fold immobility in adults. A retrospective study included patients with unilateral vocal-fold immobility undergoing laryngeal electromyography between 2007 and 2015. Neurogenic, normal or myogenic findings were compared to the clinical aspect. Prognosis for recovery was assessed from motor unit potentials on laryngeal electromyography, and compared to subsequent progress on laryngoscopy. Sixty-three patients (mean age, 59 years) were initially included; 2 were subsequently excluded from analysis. Mean time from onset of immobility to laryngeal electromyography was 7 months. 85% of the 61 patients showed neurogenic findings, indicating neural lesion; 13% showed normal electromyography, indicating cricoarytenoid joint ankylosis; and 1 patient showed a myogenic pattern. Neurogenic cases were usually secondary to cervical surgery. Thirty-eight patients were followed up. In total, 75% of patients showing reinnervation potentials recovered. The positive predictive value of laryngeal electromyography was 69.2%. Laryngeal electromyography is effective in specifying the origin of unilateral vocal-fold immobility in adults. It also has a prognostic role, lack of reinnervation potentials being a possible indication for early medialization surgery. Copyright © 2016 Elsevier Masson SAS. All rights reserved.
Ape gestures and language evolution
Pollick, Amy S.; de Waal, Frans B. M.
2007-01-01
The natural communication of apes may hold clues about language origins, especially because apes frequently gesture with limbs and hands, a mode of communication thought to have been the starting point of human language evolution. The present study aimed to contrast brachiomanual gestures with orofacial movements and vocalizations in the natural communication of our closest primate relatives, bonobos (Pan paniscus) and chimpanzees (Pan troglodytes). We tested whether gesture is the more flexible form of communication by measuring the strength of association between signals and specific behavioral contexts, comparing groups of both the same and different ape species. Subjects were two captive bonobo groups, a total of 13 individuals, and two captive chimpanzee groups, a total of 34 individuals. The study distinguished 31 manual gestures and 18 facial/vocal signals. It was found that homologous facial/vocal displays were used very similarly by both ape species, yet the same did not apply to gestures. Both within and between species gesture usage varied enormously. Moreover, bonobos showed greater flexibility in this regard than chimpanzees and were also the only species in which multimodal communication (i.e., combinations of gestures and facial/vocal signals) added to behavioral impact on the recipient. PMID:17470779
Automated extraction and classification of time-frequency contours in humpback vocalizations.
Ou, Hui; Au, Whitlow W L; Zurk, Lisa M; Lammers, Marc O
2013-01-01
A time-frequency contour extraction and classification algorithm was created to analyze humpback whale vocalizations. The algorithm automatically extracted contours of whale vocalization units by searching for gray-level discontinuities in the spectrogram images. The unit-to-unit similarity was quantified by cross-correlating the contour lines. A library of distinctive humpback units was then generated by applying an unsupervised, cluster-based learning algorithm. The purpose of this study was to provide a fast and automated feature selection tool to describe the vocal signatures of animal groups. This approach could benefit a variety of applications such as species description, identification, and evolution of song structures. The algorithm was tested on humpback whale song data recorded at various locations in Hawaii from 2002 to 2003. Results presented in this paper showed low probability of false alarm (0%-4%) under noisy environments with small boat vessels and snapping shrimp. The classification algorithm was tested on a controlled set of 30 units forming six unit types, and all the units were correctly classified. In a case study on humpback data collected in the Auau Chanel, Hawaii, in 2002, the algorithm extracted 951 units, which were classified into 12 distinctive types.
Gender and vocal production mode discrimination using the high frequencies for speech and singing
Monson, Brian B.; Lotto, Andrew J.; Story, Brad H.
2014-01-01
Humans routinely produce acoustical energy at frequencies above 6 kHz during vocalization, but this frequency range is often not represented in communication devices and speech perception research. Recent advancements toward high-definition (HD) voice and extended bandwidth hearing aids have increased the interest in the high frequencies. The potential perceptual information provided by high-frequency energy (HFE) is not well characterized. We found that humans can accomplish tasks of gender discrimination and vocal production mode discrimination (speech vs. singing) when presented with acoustic stimuli containing only HFE at both amplified and normal levels. Performance in these tasks was robust in the presence of low-frequency masking noise. No substantial learning effect was observed. Listeners also were able to identify the sung and spoken text (excerpts from “The Star-Spangled Banner”) with very few exposures. These results add to the increasing evidence that the high frequencies provide at least redundant information about the vocal signal, suggesting that its representation in communication devices (e.g., cell phones, hearing aids, and cochlear implants) and speech/voice synthesizers could improve these devices and benefit normal-hearing and hearing-impaired listeners. PMID:25400613
Fee, Michale S.
2012-01-01
In its simplest formulation, reinforcement learning is based on the idea that if an action taken in a particular context is followed by a favorable outcome, then, in the same context, the tendency to produce that action should be strengthened, or reinforced. While reinforcement learning forms the basis of many current theories of basal ganglia (BG) function, these models do not incorporate distinct computational roles for signals that convey context, and those that convey what action an animal takes. Recent experiments in the songbird suggest that vocal-related BG circuitry receives two functionally distinct excitatory inputs. One input is from a cortical region that carries context information about the current “time” in the motor sequence. The other is an efference copy of motor commands from a separate cortical brain region that generates vocal variability during learning. Based on these findings, I propose here a general model of vertebrate BG function that combines context information with a distinct motor efference copy signal. The signals are integrated by a learning rule in which efference copy inputs gate the potentiation of context inputs (but not efference copy inputs) onto medium spiny neurons in response to a rewarded action. The hypothesis is described in terms of a circuit that implements the learning of visually guided saccades. The model makes testable predictions about the anatomical and functional properties of hypothesized context and efference copy inputs to the striatum from both thalamic and cortical sources. PMID:22754501
Fee, Michale S
2012-01-01
In its simplest formulation, reinforcement learning is based on the idea that if an action taken in a particular context is followed by a favorable outcome, then, in the same context, the tendency to produce that action should be strengthened, or reinforced. While reinforcement learning forms the basis of many current theories of basal ganglia (BG) function, these models do not incorporate distinct computational roles for signals that convey context, and those that convey what action an animal takes. Recent experiments in the songbird suggest that vocal-related BG circuitry receives two functionally distinct excitatory inputs. One input is from a cortical region that carries context information about the current "time" in the motor sequence. The other is an efference copy of motor commands from a separate cortical brain region that generates vocal variability during learning. Based on these findings, I propose here a general model of vertebrate BG function that combines context information with a distinct motor efference copy signal. The signals are integrated by a learning rule in which efference copy inputs gate the potentiation of context inputs (but not efference copy inputs) onto medium spiny neurons in response to a rewarded action. The hypothesis is described in terms of a circuit that implements the learning of visually guided saccades. The model makes testable predictions about the anatomical and functional properties of hypothesized context and efference copy inputs to the striatum from both thalamic and cortical sources.
Characterization of ultrasonic vocalizations of Fragile X mice.
Belagodu, Amogh P; Johnson, Aaron M; Galvez, Roberto
2016-09-01
Fragile X Syndrome (FXS) is the leading form of inherited intellectual disability. It is caused by the transcriptional silencing of FMR1, the gene which codes for the Fragile X Mental Retardation Protein (FMRP). Patients who have FXS exhibit numerous behavioral and cognitive impairments, such as attention-deficit/hyperactivity disorder, obsessive compulsive disorder, and autistic-like behaviors. In addition to these behavioral abnormalities, FXS patients have also been shown to exhibit various deficits in communication such as abnormal sentence structures, increased utterances, repetition of sounds and words, and reduced articulation. These deficits can dramatically hinder communication for FXS patients, exacerbating learning and cognition impairments while decreasing their quality of life. To examine the biological underpinnings of these communication abnormalities, studies have used a mouse model of the Fragile X Syndrome; however, these vocalization studies have resulted in inconsistent findings that often do not correlate with abnormalities observed in FXS patients. Interestingly, a detailed examination of frequency modulated vocalizations that are believed to be a better assessment of rodent communication has never been conducted. The following study used courtship separation to conduct a detailed examination of frequency modulated ultrasonic vocalizations (USV) in FXS mice. Our analyses of frequency modulated USVs demonstrated that adult FXS mice exhibited longer phrases and more motifs. Phrases are vocalizations consisting of multiple frequency modulated ultrasonic vocalizations, while motifs are repeated frequency modulated USV patterns. Fragile X mice had a higher proportion of "u" syllables in all USVs and phrases while their wildtype counterparts preferred isolated "h" syllables. Although the specific importance of these syllables towards communication deficits still needs to be evaluated, these findings in production of USVs are consistent with the repetitive and perseverative speech patterns observed in FXS patients. This study demonstrates that FXS mice can be used to study the underlying biological mechanism(s) mediating FXS vocalization abnormalities. Copyright © 2016 Elsevier B.V. All rights reserved.
Keough, Dwayne
2011-01-01
Research on the control of visually guided limb movements indicates that the brain learns and continuously updates an internal model that maps the relationship between motor commands and sensory feedback. A growing body of work suggests that an internal model that relates motor commands to sensory feedback also supports vocal control. There is evidence from arm-reaching studies that shows that when provided with a contextual cue, the motor system can acquire multiple internal models, which allows an animal to adapt to different perturbations in diverse contexts. In this study we show that trained singers can rapidly acquire multiple internal models regarding voice fundamental frequency (F0). These models accommodate different perturbations to ongoing auditory feedback. Participants heard three musical notes and reproduced each one in succession. The musical targets could serve as a contextual cue to indicate which direction (up or down) feedback would be altered on each trial; however, participants were not explicitly instructed to use this strategy. When participants were gradually exposed to altered feedback adaptation was observed immediately following vocal onset. Aftereffects were target specific and did not influence vocal productions on subsequent trials. When target notes were no longer a contextual cue, adaptation occurred during altered feedback trials and evidence for trial-by-trial adaptation was found. These findings indicate that the brain is exceptionally sensitive to the deviations between auditory feedback and the predicted consequence of a motor command during vocalization. Moreover, these results indicate that, with contextual cues, the vocal control system may maintain multiple internal models that are capable of independent modification during different tasks or environments. PMID:21346208
Fee, Michale S.
2011-01-01
Learned motor behaviors require descending forebrain control to be coordinated with midbrain and brainstem motor systems. In songbirds, such as the zebra finch, regular breathing is controlled by brainstem centers, but when the adult songbird begins to sing, its breathing becomes tightly coordinated with forebrain-controlled vocalizations. The periods of silence (gaps) between song syllables are typically filled with brief breaths, allowing the bird to sing uninterrupted for many seconds. While substantial progress has been made in identifying the brain areas and pathways involved in vocal and respiratory control, it is not understood how respiratory and vocal control is coordinated by forebrain motor circuits. Here we combine a recently developed technique for localized brain cooling, together with recordings of thoracic air sac pressure, to examine the role of cortical premotor nucleus HVC (proper name) in respiratory-vocal coordination. We found that HVC cooling, in addition to slowing all song timescales as previously reported, also increased the duration of expiratory pulses (EPs) and inspiratory pulses (IPs). Expiratory pulses, like song syllables, were stretched uniformly by HVC cooling, but most inspiratory pulses exhibited non-uniform stretch of pressure waveform such that the majority of stretch occurred late in the IP. Indeed, some IPs appeared to change duration by the earlier or later truncation of an underlying inspiratory event. These findings are consistent with the idea that during singing the temporal structure of EPs is under the direct control of forebrain circuits, whereas that of IPs can be strongly influenced by circuits downstream of HVC, likely in the brainstem. An analysis of the temporal jitter of respiratory and vocal structure suggests that IPs may be initiated by HVC at the end of each syllable and terminated by HVC immediately before the onset of the next syllable. PMID:21980466
Garland, Ellen C; Goldizen, Anne W; Lilley, Matthew S; Rekdahl, Melinda L; Garrigue, Claire; Constantine, Rochelle; Hauser, Nan Daeschler; Poole, M Michael; Robbins, Jooke; Noad, Michael J
2015-08-01
For cetaceans, population structure is traditionally determined by molecular genetics or photographically identified individuals. Acoustic data, however, has provided information on movement and population structure with less effort and cost than traditional methods in an array of taxa. Male humpback whales (Megaptera novaeangliae) produce a continually evolving vocal sexual display, or song, that is similar among all males in a population. The rapid cultural transmission (the transfer of information or behavior between conspecifics through social learning) of different versions of this display between distinct but interconnected populations in the western and central South Pacific region presents a unique way to investigate population structure based on the movement dynamics of a song (acoustic) display. Using 11 years of data, we investigated an acoustically based population structure for the region by comparing stereotyped song sequences among populations and years. We used the Levenshtein distance technique to group previously defined populations into (vocally based) clusters based on the overall similarity of their song display in space and time. We identified the following distinct vocal clusters: western cluster, 1 population off eastern Australia; central cluster, populations around New Caledonia, Tonga, and American Samoa; and eastern region, either a single cluster or 2 clusters, one around the Cook Islands and the other off French Polynesia. These results are consistent with the hypothesis that each breeding aggregation represents a distinct population (each occupied a single, terminal node) in a metapopulation, similar to the current understanding of population structure based on genetic and photo-identification studies. However, the central vocal cluster had higher levels of song-sharing among populations than the other clusters, indicating that levels of vocal connectivity varied within the region. Our results demonstrate the utility and value of using culturally transmitted vocal patterns as a way of defining connectivity to infer population structure. We suggest vocal patterns be incorporated by the International Whaling Commission in conjunction with traditional methods in the assessment of structure. © 2015, Society for Conservation Biology.
Embodied Pronunciation Learning: Research and Practice
ERIC Educational Resources Information Center
Chan, Marsha J.
2018-01-01
This article summarizes research on body language, embodiment, and the incorporation of proprioception, physical movement, gestures, and touch into second language education, particularly with regard to the pronunciation of English. It asserts that careful attention to breathing, vocalization, articulatory positions, pulmonic and tactile…
Bennur, Sharath; Tsunada, Joji; Cohen, Yale E; Liu, Robert C
2013-11-01
Acoustic communication between animals requires them to detect, discriminate, and categorize conspecific or heterospecific vocalizations in their natural environment. Laboratory studies of the auditory-processing abilities that facilitate these tasks have typically employed a broad range of acoustic stimuli, ranging from natural sounds like vocalizations to "artificial" sounds like pure tones and noise bursts. However, even when using vocalizations, laboratory studies often test abilities like categorization in relatively artificial contexts. Consequently, it is not clear whether neural and behavioral correlates of these tasks (1) reflect extensive operant training, which drives plastic changes in auditory pathways, or (2) the innate capacity of the animal and its auditory system. Here, we review a number of recent studies, which suggest that adopting more ethological paradigms utilizing natural communication contexts are scientifically important for elucidating how the auditory system normally processes and learns communication sounds. Additionally, since learning the meaning of communication sounds generally involves social interactions that engage neuromodulatory systems differently than laboratory-based conditioning paradigms, we argue that scientists need to pursue more ethological approaches to more fully inform our understanding of how the auditory system is engaged during acoustic communication. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives". Copyright © 2013 Elsevier B.V. All rights reserved.
Cross-cultural adaptation of the Chilean version of the Voice Symptom Scale - VoiSS.
Ruston, Francisco Contreras; Moreti, Felipe; Vivero, Martín; Malebran, Celina; Behlau, Mara
This research aims to accomplish the cross-cultural equivalence of the Chilean version of the VoiSS protocol through its cultural and linguistic adaptation. After the translation of the VoiSS protocol to Chilean Spanish by two bilingual speech therapists and its back translation to English, we compared the items of the original tool with the previous translated version. The existing discrepancies were modified by a consensus committee of five speech therapists and the translated version was entitled Escala de Sintomas Vocales - ESV, with 30 questions and five answers: "Never", "Occasionally", "Sometimes", "Most of the time", "Always". For cross-cultural equivalence, the protocol was applied to 15 individuals with vocal problems. In each question the option of "Not applicable" was added to the answer choices for identification of the questions not comprehended or not appropriate for the target population. Two individuals had difficulty answering two questions, which made it necessary to adapt the translation of only one of them. The modified ESV was applied to three individuals with vocal problems, and there were incomprehensible inappropriate questions for the Chilean culture. The ESV reflects the original English version, both in the number of questions and the limitations of the emotional and physical domains. There is now a cross-cultural equivalence of VoiSS in Chilean Spanish, titled ESV. The validation of the ESV for Chilean Spanish is ongoing. RESUMEN Este estudio tuvo como objetivo realizar la equivalencia cultural de la versión Chilena del protocolo Voice Symptom Scale - VoiSS por medio de su adaptación cultural y lingüística. Después de la traducción del VoiSS para el Español Chileno, por dos fonoaudiólogos bilingües, y de la retro traducción para el inglés, se realizó una comparación de los ítems del instrumento original con la versión traducida, surgiendo discrepancias; tales divergencias fueron resueltas por un comité compuesto por cinco fonoaudiólogos, que llegaron a un consenso de donde surgió la denominada Escala de Síntomas Vocales - ESV, compuesta de 30 preguntas y 5 respuestas: "Nunca", "Casi nunca", "A veces", "Casi siempre", "Siempre". Con el objetivo de lograr la equivalencia cultural, la ESV fue aplicada a 15 individuos con problemas vocales. A cada pregunta se le agregó la opción "No aplicable" en las elecciones de respuesta, con el fin de identificar preguntas incomprendidas o inapropiadas para la población en cuestión. Dos de los individuos tuvieron dificultades en el momento de contestar dos preguntas, haciéndose necesario adaptar solamente la traducción de una de ellas. La ESV modificada fue aplicada a tres personas más con problemas vocales, no encontrándose preguntas incomprendidas o inapropiadas para la cultura Chilena. La ESV refleja la versión original de inglés, tanto en la cantidad de preguntas como en la limitación de los dominios emocional y físico. Existe una equivalencia cultural del VoiSS para el Español Chileno, que recibe el nombre de ESV. La validación de la ESV para el Español Chileno está en proceso de conclusión.
Collaborative Learning in Two Vocal Conservatoire Courses
ERIC Educational Resources Information Center
Rumiantsev, Tamara W.; Maas, Annemarie; Admiraal, Wilfried
2017-01-01
The apprenticeship tradition in conservatoire education assumes that teachers' expertise is the main source for the development of future music professionals. However, the professional practice of vocalists is nearly completely based on collaboration, such as with other vocalists, instrumentalists, accompanists, orchestras, conductors, or stage…
Knockdown of FoxP2 alters spine density in Area X of the zebra finch.
Schulz, S B; Haesler, S; Scharff, C; Rochefort, C
2010-10-01
Mutations in the gene encoding the transcription factor FoxP2 impair human speech and language. We have previously shown that deficits in vocal learning occur in zebra finches after reduction of FoxP2 in Area X, a striatal nucleus involved in song acquisition. We recently showed that FoxP2 is expressed in newly generated spiny neurons (SN) in adult Area X as well as in the ventricular zone (VZ) from which the SN originates. Moreover, their recruitment to Area X increases transiently during the song learning phase. The present report therefore investigated whether FoxP2 is involved in the structural plasticity of Area X. We assessed the proliferation, differentiation and morphology of SN after lentivirally mediated knockdown of FoxP2 in Area X or in the VZ during the song learning phase. Proliferation rate was not significantly affected by knockdown of FoxP2 in the VZ. In addition, FoxP2 reduction both in the VZ and in Area X did not affect the number of new neurons in Area X. However, at the fine-structural level, SN in Area X bore fewer spines after FoxP2 knockdown. This effect was even more pronounced when neurons received the knockdown before differentiation, i.e. as neuroblasts in the VZ. These results suggest that FoxP2 might directly or indirectly regulate spine dynamics in Area X and thereby influence song plasticity. Together, these data present the first evidence for a role of FoxP2 in the structural plasticity of dendritic spines and complement the emerging evidence of physiological synaptic plasticity in FoxP2 mouse models. Genes, Brain and Behavior © 2010 Blackwell Publishing Ltd and International Behavioural and Neural Genetics Society. No claim to original US government works.
Production and survival of projection neurons in a forebrain vocal center of adult male canaries
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kirn, J.R.; Alvarez-Buylla, A.; Nottebohm, F.
1991-06-01
Neurons are produced in the adult canary telencephalon. Many of these cells are incorporated into the high vocal center (nucleus HVC), which participates in the control of learned song. In the present work, 3H-thymidine and fluorogold were employed to follow the differentiation and survival of HVC neurons born in adulthood. We found that many HVC neurons born in September grow long axons to the robust nucleus of the archistriatum (nucleus RA) and thus become part of the efferent pathway for song control. Many of these new neurons have already established their connections with RA by 30 d after their birth.more » By 240 d, 75-80% of the September-born HVC neurons project to RA. Most of these new projection neurons survive at least 8 months. The longevity of HVC neurons born in September suggests that these cells remain part of the vocal control circuit long enough to participate in the yearly renewal of the song repertoire.« less
Laukka, Petri; Neiberg, Daniel; Elfenbein, Hillary Anger
2014-06-01
The possibility of cultural differences in the fundamental acoustic patterns used to express emotion through the voice is an unanswered question central to the larger debate about the universality versus cultural specificity of emotion. This study used emotionally inflected standard-content speech segments expressing 11 emotions produced by 100 professional actors from 5 English-speaking cultures. Machine learning simulations were employed to classify expressions based on their acoustic features, using conditions where training and testing were conducted on stimuli coming from either the same or different cultures. A wide range of emotions were classified with above-chance accuracy in cross-cultural conditions, suggesting vocal expressions share important characteristics across cultures. However, classification showed an in-group advantage with higher accuracy in within- versus cross-cultural conditions. This finding demonstrates cultural differences in expressive vocal style, and supports the dialect theory of emotions according to which greater recognition of expressions from in-group members results from greater familiarity with culturally specific expressive styles.
Development of auditory sensitivity in budgerigars (Melopsittacus undulatus)
NASA Astrophysics Data System (ADS)
Brittan-Powell, Elizabeth F.; Dooling, Robert J.
2004-06-01
Auditory feedback influences the development of vocalizations in songbirds and parrots; however, little is known about the development of hearing in these birds. The auditory brainstem response was used to track the development of auditory sensitivity in budgerigars from hatch to 6 weeks of age. Responses were first obtained from 1-week-old at high stimulation levels at frequencies at or below 2 kHz, showing that budgerigars do not hear well at hatch. Over the next week, thresholds improved markedly, and responses were obtained for almost all test frequencies throughout the range of hearing by 14 days. By 3 weeks posthatch, birds' best sensitivity shifted from 2 to 2.86 kHz, and the shape of the auditory brainstem response (ABR) audiogram became similar to that of adult budgerigars. About a week before leaving the nest, ABR audiograms of young budgerigars are very similar to those of adult birds. These data complement what is known about vocal development in budgerigars and show that hearing is fully developed by the time that vocal learning begins.
The hypoglossal canal and the origin of human vocal behavior
Kay, Richard F.; Cartmill, Matt; Balow, Michelle
1998-01-01
The mammalian hypoglossal canal transmits the nerve that supplies the muscles of the tongue. This canal is absolutely and relatively larger in modern humans than it is in the African apes (Pan and Gorilla). We hypothesize that the human tongue is supplied more richly with motor nerves than are those of living apes and propose that canal size in fossil hominids may provide an indication about the motor coordination of the tongue and reflect the evolution of speech and language. Canals of gracile Australopithecus, and possibly Homo habilis, fall within the range of extant Pan and are significantly smaller than those of modern Homo. The canals of Neanderthals and an early “modern” Homo sapiens (Skhul 5), as well as of African and European middle Pleistocene Homo (Kabwe and Swanscombe), fall within the range of extant Homo and are significantly larger than those of Pan troglodytes. These anatomical findings suggest that the vocal capabilities of Neanderthals were the same as those of humans today. Furthermore, the vocal abilities of Australopithecus were not advanced significantly over those of chimpanzees whereas those of Homo may have been essentially modern by at least 400,000 years ago. Thus, human vocal abilities may have appeared much earlier in time than the first archaeological evidence for symbolic behavior. PMID:9560291
The hypoglossal canal and the origin of human vocal behavior.
Kay, R F; Cartmill, M; Balow, M
1998-04-28
The mammalian hypoglossal canal transmits the nerve that supplies the muscles of the tongue. This canal is absolutely and relatively larger in modern humans than it is in the African apes (Pan and Gorilla). We hypothesize that the human tongue is supplied more richly with motor nerves than are those of living apes and propose that canal size in fossil hominids may provide an indication about the motor coordination of the tongue and reflect the evolution of speech and language. Canals of gracile Australopithecus, and possibly Homo habilis, fall within the range of extant Pan and are significantly smaller than those of modern Homo. The canals of Neanderthals and an early "modern" Homo sapiens (Skhul 5), as well as of African and European middle Pleistocene Homo (Kabwe and Swanscombe), fall within the range of extant Homo and are significantly larger than those of Pan troglodytes. These anatomical findings suggest that the vocal capabilities of Neanderthals were the same as those of humans today. Furthermore, the vocal abilities of Australopithecus were not advanced significantly over those of chimpanzees whereas those of Homo may have been essentially modern by at least 400,000 years ago. Thus, human vocal abilities may have appeared much earlier in time than the first archaeological evidence for symbolic behavior.
Applications for Subvocal Speech
NASA Technical Reports Server (NTRS)
Jorgensen, Charles; Betts, Bradley
2007-01-01
A research and development effort now underway is directed toward the use of subvocal speech for communication in settings in which (1) acoustic noise could interfere excessively with ordinary vocal communication and/or (2) acoustic silence or secrecy of communication is required. By "subvocal speech" is meant sub-audible electromyographic (EMG) signals, associated with speech, that are acquired from the surface of the larynx and lingual areas of the throat. Topics addressed in this effort include recognition of the sub-vocal EMG signals that represent specific original words or phrases; transformation (including encoding and/or enciphering) of the signals into forms that are less vulnerable to distortion, degradation, and/or interception; and reconstruction of the original words or phrases at the receiving end of a communication link. Potential applications include ordinary verbal communications among hazardous- material-cleanup workers in protective suits, workers in noisy environments, divers, and firefighters, and secret communications among law-enforcement officers and military personnel in combat and other confrontational situations.
Avey, Marc T; Hoeschele, Marisa; Moscicki, Michele K; Bloomfield, Laurie L; Sturdy, Christopher B
2011-01-01
Songbird auditory areas (i.e., CMM and NCM) are preferentially activated to playback of conspecific vocalizations relative to heterospecific and arbitrary noise. Here, we asked if the neural response to auditory stimulation is not simply preferential for conspecific vocalizations but also for the information conveyed by the vocalization. Black-capped chickadees use their chick-a-dee mobbing call to recruit conspecifics and other avian species to mob perched predators. Mobbing calls produced in response to smaller, higher-threat predators contain more "D" notes compared to those produced in response to larger, lower-threat predators and thus convey the degree of threat of predators. We specifically asked whether the neural response varies with the degree of threat conveyed by the mobbing calls of chickadees and whether the neural response is the same for actual predator calls that correspond to the degree of threat of the chickadee mobbing calls. Our results demonstrate that, as degree of threat increases in conspecific chickadee mobbing calls, there is a corresponding increase in immediate early gene (IEG) expression in telencephalic auditory areas. We also demonstrate that as the degree of threat increases for the heterospecific predator, there is a corresponding increase in IEG expression in the auditory areas. Furthermore, there was no significant difference in the amount IEG expression between conspecific mobbing calls or heterospecific predator calls that were the same degree of threat. In a second experiment, using hand-reared chickadees without predator experience, we found more IEG expression in response to mobbing calls than corresponding predator calls, indicating that degree of threat is learned. Our results demonstrate that degree of threat corresponds to neural activity in the auditory areas and that threat can be conveyed by different species signals and that these signals must be learned.
Early life stress impairs contextual threat expression in female, but not male, mice.
Manzano-Nieves, Gabriela; Gaillard, Mizan; Gallo, Meghan; Bath, Kevin G
2018-05-21
Early life stress (ELS) is associated with altered processing of threat signals, and increased lifetime risk of anxiety and affective pathology, disorders that disproportionately affect females. We tested the impact of a limited bedding paradigm of ELS (from P4-11) on contextual threat learning, context memory, footshock sensitivity, and anxietylike behavior, in adult male and female mice. To examine contextual threat learning, mice conditioned by context/footshock association were tested 24 hr later for the context memory. To determine the effect of ELS on footshock sensitivity, a separate cohort of mice were exposed to footshocks of increasing intensity (0.06 to 0.40 mA) and behavioral responses (jump and audible vocalization) were assessed by observers blind to treatment condition, sex, and cycle stage. ELS impaired context memory in female, but not male, mice. ELS increased footshock-induced threshold to vocalize, but not to jump, in both sexes. In female mice, this effect was most apparent during estrus. Decreased body weight, indicative of higher stress incurred by an individual mouse, correlated with increased threshold to jump in both sexes reared in ELS, and to audibly vocalize in ELS females. As ELS effects on shock sensitivity were present in both sexes, the contextual recall deficit in females was not likely driven by changes in the salience of aversive footshocks. No effects on anxietylike behavior, as measured in the elevated plus maze (EPM), were observed. More work is needed to better understand the impact of ELS on both somatic and gonadal development, and their potential contribution to threat learning. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Ivanova, Tamara N; Gross, Christina; Mappus, Rudolph C; Kwon, Yong Jun; Bassell, Gary J; Liu, Robert C
2017-12-01
Learning to recognize a stimulus category requires experience with its many natural variations. However, the mechanisms that allow a category's sensorineural representation to be updated after experiencing new exemplars are not well understood, particularly at the molecular level. Here we investigate how a natural vocal category induces expression in the auditory system of a key synaptic plasticity effector immediate early gene, Arc/Arg3.1 , which is required for memory consolidation. We use the ultrasonic communication system between mouse pups and adult females to study whether prior familiarity with pup vocalizations alters how Arc is engaged in the core auditory cortex after playback of novel exemplars from the pup vocal category. A computerized, 3D surface-assisted cellular compartmental analysis, validated against manual cell counts, demonstrates significant changes in the recruitment of neurons expressing Arc in pup-experienced animals (mothers and virgin females "cocaring" for pups) compared with pup-inexperienced animals (pup-naïve virgins), especially when listening to more familiar, natural calls compared to less familiar but similarly recognized tonal model calls. Our data support the hypothesis that the kinetics of Arc induction to refine cortical representations of sensory categories is sensitive to the familiarity of the sensory experience. © 2017 Ivanova et al.; Published by Cold Spring Harbor Laboratory Press.
L1 literacy affects L2 pronunciation intake and text vocalization
NASA Astrophysics Data System (ADS)
Walton, Martin
2005-04-01
For both deaf and hearing learners, L1 acquisition calls on auditive, gestural and visual modes in progressive processes over longer stages imposed in strictly anatomical and social order from the earliest pre-lexical phase [Jusczyk (1993), Kuhl & Meltzoff (1996)] to ultimate literacy. By contrast, L2 learning will call on accelerating procedures but with restricted input, arbitrated by L1 literacy as can be traced in the English of French-speaking learners, whether observed in spontaneous speech or in text vocalization modes. An inventory of their predictable omissions, intrusions and substitutions at suprasegmental and syllabic levels, many of which they can actually hear while unable to vocalize in real-time, suggests that a photogenic segmentation of continuous speech into alphabetical units has eclipsed the indispensable earlier phonogenic module, filtering L2 intake and output. This competing mode analysis hypothesizes a critical effect on L2 pronunciation of L1 graphemic procedures acquired usually before puberty, informing data for any Critical Period Hypothesis or amounts of L1 activation influencing L2 accent [Flege (1997, 1998)] or any psychoacoustic French deafness with regard to English stress-timing [Dupoux (1997)]. A metaphonic model [Howell & Dean (1991)] adapted for French learners may remedially distance L1 from L2 vocalization procedures.
Rehn, Nicola; Filatova, Olga A; Durban, John W; Foote, Andrew D
2011-01-01
Facial and vocal expressions of emotion have been found in a number of social mammal species and are thought to have evolved to aid social communication. There has been much debate about whether such signals are culturally inherited or are truly biologically innate. Evidence for the innateness of such signals can come from cross-cultural studies. Previous studies have identified a vocalisation (the V4 or 'excitement' call) associated with high arousal behaviours in a population of killer whales in British Columbia, Canada. In this study, we compared recordings from three different socially and reproductively isolated ecotypes of killer whales, including five vocal clans of one ecotype, each clan having discrete culturally transmitted vocal traditions. The V4 call was found in recordings of each ecotype and each vocal clan. Nine independent observers reproduced our classification of the V4 call from each population with high inter-observer agreement. Our results suggest the V4 call may be universal in Pacific killer whale populations and that transmission of this call is independent of cultural tradition or ecotype. We argue that such universality is more consistent with an innate vocalisation than one acquired through social learning and may be linked to its apparent function of motivational expression.
NASA Astrophysics Data System (ADS)
Rehn, Nicola; Filatova, Olga A.; Durban, John W.; Foote, Andrew D.
2011-01-01
Facial and vocal expressions of emotion have been found in a number of social mammal species and are thought to have evolved to aid social communication. There has been much debate about whether such signals are culturally inherited or are truly biologically innate. Evidence for the innateness of such signals can come from cross-cultural studies. Previous studies have identified a vocalisation (the V4 or `excitement' call) associated with high arousal behaviours in a population of killer whales in British Columbia, Canada. In this study, we compared recordings from three different socially and reproductively isolated ecotypes of killer whales, including five vocal clans of one ecotype, each clan having discrete culturally transmitted vocal traditions. The V4 call was found in recordings of each ecotype and each vocal clan. Nine independent observers reproduced our classification of the V4 call from each population with high inter-observer agreement. Our results suggest the V4 call may be universal in Pacific killer whale populations and that transmission of this call is independent of cultural tradition or ecotype. We argue that such universality is more consistent with an innate vocalisation than one acquired through social learning and may be linked to its apparent function of motivational expression.
Hemispheric differences in processing of vocalizations depend on early experience.
Phan, Mimi L; Vicario, David S
2010-02-02
An intriguing phenomenon in the neurobiology of language is lateralization: the dominant role of one hemisphere in a particular function. Lateralization is not exclusive to language because lateral differences are observed in other sensory modalities, behaviors, and animal species. Despite much scientific attention, the function of lateralization, its possible dependence on experience, and the functional implications of such dependence have yet to be clearly determined. We have explored the role of early experience in the development of lateralized sensory processing in the brain, using the songbird model of vocal learning. By controlling exposure to natural vocalizations (through isolation, song tutoring, and muting), we manipulated the postnatal auditory environment of developing zebra finches, and then assessed effects on hemispheric specialization for communication sounds in adulthood. Using bilateral multielectrode recordings from a forebrain auditory area known to selectively process species-specific vocalizations, we found that auditory responses to species-typical songs and long calls, in both male and female birds, were stronger in the right hemisphere than in the left, and that right-side responses adapted more rapidly to stimulus repetition. We describe specific instances, particularly in males, where these lateral differences show an influence of auditory experience with song and/or the bird's own voice during development.
The Molecular Basis of Human Brain Evolution.
Enard, Wolfgang
2016-10-24
Humans are a remarkable species, especially because of the remarkable properties of their brain. Since the split from the chimpanzee lineage, the human brain has increased three-fold in size and has acquired abilities for vocal learning, language and intense cooperation. To better understand the molecular basis of these changes is of great biological and biomedical interest. However, all the about 16 million fixed genetic changes that occurred during human evolution are fully correlated with all molecular, cellular, anatomical and behavioral changes that occurred during this time. Hence, as humans and chimpanzees cannot be crossed or genetically manipulated, no direct evidence for linking particular genetic and molecular changes to human brain evolution can be obtained. Here, I sketch a framework how indirect evidence can be obtained and review findings related to the molecular basis of human cognition, vocal learning and brain size. In particular, I discuss how a comprehensive comparative approach, leveraging cellular systems and genomic technologies, could inform the evolution of our brain in the future. Copyright © 2016 Elsevier Ltd. All rights reserved.
Wohlgemuth, Sandra; Adam, Iris; Scharff, Constance
2014-10-01
Humans with mutations in the transcription factor FOXP2 display a severe speech disorder. Songbirds are a powerful model system to study FoxP2. Like humans, songbirds communicate via vocalizations that are imitatively learned during critical periods and this learning is influenced by social factors and relies on functionally lateralized neural circuits. During the past five years significant progress has been made moving from a descriptive to a more mechanistic understanding of how FoxP2 functions in songbirds. Current evidence from molecular and electrophysiological studies indicates that FoxP2 is important for shaping synaptic plasticity of specific neuron populations. One future goal will be to identify the transcriptional regulation orchestrated by FoxP2 and its associated molecular network that brings about these physiological effects. This will be key to further unravel how FoxP2 influences synaptic function and thereby contributes to auditory guided vocal motor behavior in the songbird model. Copyright © 2014 Elsevier Ltd. All rights reserved.
Language development in a non-vocal child.
Rogow, S M
1994-01-01
Many children who cannot speak, comprehend both oral and written language. Having knowledge of language is not the same as being able to use language for social transactions. Non-vocal children learn to use augmented and assisted systems, but they experience specific difficulties in initiating and maintaining conversations and making use of the pragmatic functions of language. The purpose of this study was to investigate the semantic and syntactic knowledge of a child with severe multiple disabilities who can read and write and comprehend two languages, but does not initiate conversation. The study demonstrates that high levels of language comprehension and ability to read and write do not automatically transfer to conversational competence or narrative ability.
Singing-driven gene expression in the developing songbird brain
Johnson, Frank; Whitney, Osceola
2014-01-01
Neural and behavioral development arises from an integration of genetic and environmental influences, yet specifying the nature of this interaction remains a primary problem in neuroscience. Here, we review molecular and behavioral studies that focus on the role of singing-driven gene expression during neural and vocal development in the male zebra finch (Taeniopygia guttata), a songbird that learns a species-typical vocal pattern during juvenile development by imitating an adult male tutor. A primary aim of our lab has been to identify naturally-occurring environmental influences that shape the propensity to sing. This ethological approach underlies our theoretical perspective, which is to integrate the significance of singing-driven gene expression into a broader ecological context. PMID:16129463
Baker, J
1999-12-01
Four women aged between 27 and 58 years sought otolaryngological examination due to significant alterations to their voices, the primary concerns being hoarseness in vocal quality, lowering of habitual pitch, difficulty projecting their speaking voices, and loss of control over their singing voices. Otolaryngological examination with a mirror or flexible laryngoscope revealed no apparent abnormality of vocal fold structure or function, and the women were referred for speech pathology with diagnoses of functional dysphonia. Objective acoustic measures using the Kay Visipitch indicated significant lowering of the mean fundamental frequency for each woman, and perceptual analysis of the patients' voices during quiet speaking, projected voice use, and comprehensive singing activities revealed a constellation of features typically noted in the pubescent male. The original diagnoses of a functional dysphonia were queried, prompting further exploration of each woman's medical history, revealing in each case onset of vocal symptoms shortly after commencing treatment for conditions with medications containing virilizing agents (eg, Danocrine (danazol), Deca-Durabolin (nandrolene decanoate), and testosterone). Although some of the vocal symptoms decreased in severity with the influences from 6 months voice therapy and after withdrawal from the drugs, a number of symptoms remained permanent, suggesting each subject had suffered significant alterations in vocal physiology, including muscle tissue changes, muscle coordination dysfunction, and propioceptive dysfunction. This retrospective study is presented in order to illustrate that it was both the projected speaking voice and the singing voice that proved so highly sensitive to the virilization effects. The implications for future prospective research studies and responsible clinical practice are discussed.
Forlano, Paul M.; Kim, Spencer D.; Krzyminska, Zuzanna M.; Sisneros, Joseph A.
2014-01-01
Although the neuroanatomical distribution of catecholaminergic (CA) neurons has been well documented across all vertebrate classes, few studies have examined CA connectivity to physiologically and anatomically identified neural circuitry that controls behavior. The goal of this study was to characterize CA distribution in the brain and inner ear of the plainfin midshipman fish (Porichthys notatus) with particular emphasis on their relationship with anatomically labeled circuitry that both produces and encodes social acoustic signals in this species. Neurobiotin labeling of the main auditory endorgan, the saccule, combined with tyrosine hydroxylase immunofluorescence (TH-ir) revealed a strong CA innervation of both the peripheral and central auditory system. Diencephalic TH-ir neurons in the periventricular posterior tuberculum, known to be dopaminergic, send ascending projections to the ventral telencephalon and prominent descending projections to vocal-acoustic integration sites, notably the hindbrain octavolateralis efferent nucleus, as well as onto the base of hair cells in the saccule via nerve VIII. Neurobiotin backfills of the vocal nerve in combination with TH-ir revealed CA terminals on all components of the vocal pattern generator which appears to largely originate from local TH-ir neurons but may include diencephalic projections as well. This study provides strong evidence for catecholamines as important neuromodulators of both auditory and vocal circuitry and acoustic-driven social behavior in midshipman fish. This first demonstration of TH-ir terminals in the main endorgan of hearing in a non-mammalian vertebrate suggests a conserved and important anatomical and functional role for dopamine in normal audition. PMID:24715479
Ictal speech and language dysfunction in adult epilepsy: Clinical study of 95 seizures.
Dussaule, C; Cauquil, C; Flamand-Roze, C; Gagnepain, J-P; Bouilleret, V; Denier, C; Masnou, P
2017-04-01
To analyze the semiological characteristics of the language and speech disorders arising during epileptic seizures, and to describe the patterns of language and speech disorders that can predict laterality of the epileptic focus. This study retrospectively analyzed 95 consecutive videos of seizures with language and/or speech disorders in 44 patients admitted for diagnostic video-EEG monitoring. Laterality of the epileptic focus was defined according to electro-clinical correlation studies and structural and functional neuroimaging findings. Language and speech disorders were analyzed by a neurologist and a speech therapist blinded to these data. Language and/or speech disorders were subdivided into eight dynamic patterns: pure anterior aphasia; anterior aphasia and vocal; anterior aphasia and "arthria"; pure posterior aphasia; posterior aphasia and vocal; pure vocal; vocal and arthria; and pure arthria. The epileptic focus was in the left hemisphere in more than 4/5 of seizures presenting with pure anterior aphasia or pure posterior aphasia patterns, while discharges originated in the right hemisphere in almost 2/3 of seizures presenting with a pure vocal pattern. No laterality value was found for the other patterns. Classification of the language and speech disorders arising during epileptic seizures into dynamic patterns may be useful for the optimal analysis of anatomo-electro-clinical correlations. In addition, our research has led to the development of standardized tests for analyses of language and speech disorders arising during seizures that can be conducted during video-EEG sessions. Copyright © 2017 Elsevier Masson SAS. All rights reserved.
Guillette, Lauren M; Healy, Susan D
2017-06-01
The transmission of information from an experienced demonstrator to a naïve observer often depends on characteristics of the demonstrator, such as familiarity, success or dominance status. Whether or not the demonstrator pays attention to and/or interacts with the observer may also affect social information acquisition or use by the observer. Here we used a video-demonstrator paradigm first to test whether video demonstrators have the same effect as using live demonstrators in zebra finches, and second, to test the importance of visual and vocal interactions between the demonstrator and observer on social information use by the observer. We found that female zebra finches copied novel food choices of male demonstrators they saw via live-streaming video while they did not consistently copy from the demonstrators when they were seen in playbacks of the same videos. Although naive observers copied in the absence of vocalizations by the demonstrator, as they copied from playback of videos with the sound off, females did not copy where there was a mis-match between the visual information provided by the video and vocal information from a live male that was out of sight. Taken together these results suggest that video demonstration is a useful methodology for testing social information transfer, at least in a foraging context, but more importantly, that social information use varies according to the vocal interactions, or lack thereof, between the observer and the demonstrator. Copyright © 2017 The Authors. Published by Elsevier B.V. All rights reserved.
Can vocal conditioning trigger a semiotic ratchet in marmosets?
Turesson, Hjalmar K; Ribeiro, Sidarta
2015-01-01
The complexity of human communication has often been taken as evidence that our language reflects a true evolutionary leap, bearing little resemblance to any other animal communication system. The putative uniqueness of the human language poses serious evolutionary and ethological challenges to a rational explanation of human communication. Here we review ethological, anatomical, molecular, and computational results across several species to set boundaries for these challenges. Results from animal behavior, cognitive psychology, neurobiology, and semiotics indicate that human language shares multiple features with other primate communication systems, such as specialized brain circuits for sensorimotor processing, the capability for indexical (pointing) and symbolic (referential) signaling, the importance of shared intentionality for associative learning, affective conditioning and parental scaffolding of vocal production. The most substantial differences lie in the higher human capacity for symbolic compositionality, fast vertical transmission of new symbols across generations, and irreversible accumulation of novel adaptive behaviors (cultural ratchet). We hypothesize that increasingly-complex vocal conditioning of an appropriate animal model may be sufficient to trigger a semiotic ratchet, evidenced by progressive sign complexification, as spontaneous contact calls become indexes, then symbols and finally arguments (strings of symbols). To test this hypothesis, we outline a series of conditioning experiments in the common marmoset (Callithrix jacchus). The experiments are designed to probe the limits of vocal communication in a prosocial, highly vocal primate 35 million years far from the human lineage, so as to shed light on the mechanisms of semiotic complexification and cultural transmission, and serve as a naturalistic behavioral setting for the investigation of language disorders.
Kleber, Boris; Zeitouni, Anthony G; Friberg, Anders; Zatorre, Robert J
2013-04-03
Somatosensation plays an important role in the motor control of vocal functions, yet its neural correlate and relation to vocal learning is not well understood. We used fMRI in 17 trained singers and 12 nonsingers to study the effects of vocal-fold anesthesia on the vocal-motor singing network as a function of singing expertise. Tasks required participants to sing musical target intervals under normal conditions and after anesthesia. At the behavioral level, anesthesia altered pitch accuracy in both groups, but singers were less affected than nonsingers, indicating an experience-dependent effect of the intervention. At the neural level, this difference was accompanied by distinct patterns of decreased activation in singers (cortical and subcortical sensory and motor areas) and nonsingers (subcortical motor areas only) respectively, suggesting that anesthesia affected the higher-level voluntary (explicit) motor and sensorimotor integration network more in experienced singers, and the lower-level (implicit) subcortical motor loops in nonsingers. The right anterior insular cortex (AIC) was identified as the principal area dissociating the effect of expertise as a function of anesthesia by three separate sources of evidence. First, it responded differently to anesthesia in singers (decreased activation) and nonsingers (increased activation). Second, functional connectivity between AIC and bilateral A1, M1, and S1 was reduced in singers but augmented in nonsingers. Third, increased BOLD activity in right AIC in singers was correlated with larger pitch deviation under anesthesia. We conclude that the right AIC and sensory-motor areas play a role in experience-dependent modulation of feedback integration for vocal motor control during singing.
Can vocal conditioning trigger a semiotic ratchet in marmosets?
Turesson, Hjalmar K.; Ribeiro, Sidarta
2015-01-01
The complexity of human communication has often been taken as evidence that our language reflects a true evolutionary leap, bearing little resemblance to any other animal communication system. The putative uniqueness of the human language poses serious evolutionary and ethological challenges to a rational explanation of human communication. Here we review ethological, anatomical, molecular, and computational results across several species to set boundaries for these challenges. Results from animal behavior, cognitive psychology, neurobiology, and semiotics indicate that human language shares multiple features with other primate communication systems, such as specialized brain circuits for sensorimotor processing, the capability for indexical (pointing) and symbolic (referential) signaling, the importance of shared intentionality for associative learning, affective conditioning and parental scaffolding of vocal production. The most substantial differences lie in the higher human capacity for symbolic compositionality, fast vertical transmission of new symbols across generations, and irreversible accumulation of novel adaptive behaviors (cultural ratchet). We hypothesize that increasingly-complex vocal conditioning of an appropriate animal model may be sufficient to trigger a semiotic ratchet, evidenced by progressive sign complexification, as spontaneous contact calls become indexes, then symbols and finally arguments (strings of symbols). To test this hypothesis, we outline a series of conditioning experiments in the common marmoset (Callithrix jacchus). The experiments are designed to probe the limits of vocal communication in a prosocial, highly vocal primate 35 million years far from the human lineage, so as to shed light on the mechanisms of semiotic complexification and cultural transmission, and serve as a naturalistic behavioral setting for the investigation of language disorders. PMID:26500583
Memory in the making: localized brain activation related to song learning in young songbirds
Gobes, Sharon M. H.; Zandbergen, Matthijs A.; Bolhuis, Johan J.
2010-01-01
Songbird males learn to sing their songs from an adult ‘tutor’ early in life, much like human infants learn to speak. Similar to humans, in the songbird brain there are separate neural substrates for vocal production and for auditory memory. In adult songbirds, the caudal pallium, the avian equivalent of the auditory association cortex, has been proposed to contain the neural substrate of tutor song memory, while the song system is involved in song production as well as sensorimotor learning. If this hypothesis is correct, there should be neuronal activation in the caudal pallium, and not in the song system, while the young bird is hearing the tutor song. We found increased song-induced molecular neuronal activation, measured as the expression of an immediate early gene, in the caudal pallium of juvenile zebra finch males that were in the process of learning to sing their songs. No such activation was found in the song system. Molecular neuronal activation was significantly greater in response to tutor song than to novel song or silence in the medial part of the caudomedial nidopallium (NCM). In the caudomedial mesopallium, there was significantly greater molecular neuronal activation in response to tutor song than to silence. In addition, in the NCM there was a significant positive correlation between spontaneous molecular neuronal activation and the strength of song learning during sleep. These results suggest that the caudal pallium contains the neural substrate for tutor song memory, which is activated during sleep when the young bird is in the process of learning its song. The findings provide insight into the formation of auditory memories that guide vocal production learning, a process fundamental for human speech acquisition. PMID:20534608
The Referent of Children's Early Songs
ERIC Educational Resources Information Center
Mang, Esther
2005-01-01
Musical creativity during early childhood is readily exemplified in vocal behaviours. This paper is a discussion of observations on children's performance of learned songs and self-generated songs. Longitudinal observations suggest that self-generated songs may be seen as referent-guided improvisation using source materials derived from learned…
Syllabic Patterns in the Early Vocalizations of Quichua Children
ERIC Educational Resources Information Center
Gildersleeve-Neumann, Christina E.; Davis, Barbara L.; Macneilage, Peter F.
2013-01-01
To understand the interactions between production patterns common to children regardless of language environment and the early appearance of production effects based on perceptual learning from the ambient language requires the study of languages with diverse phonological properties. Few studies have evaluated early phonological acquisition…
ERIC Educational Resources Information Center
McCloskey, Patrick J.
2005-01-01
Nelson Beaudoin, the principal of Kennebunk High School in Maine believes devoutly in letting students have a say in how they are educated. He also knows the risks. At this student-centered school, learning includes sex education. The girls' team chooses "sexually transmitted diseases" for 300 points. Although the game is modeled after…
ERIC Educational Resources Information Center
Skouge, James R.; Kajiyama, Brian
2009-01-01
In this article, the authors relate a story about the transformative power of technologies for voice. They relate Brian Kajiyama's personal odyssey--what might be described as a journey from unvoiced to vocal--in learning to use a DynaWrite, a type-and-talk device that Brian uses as a communication tool.
Balanoff, Amy M; Smaers, Jeroen B; Turner, Alan H
2016-08-01
Living birds constitute the only vertebrate group whose brain volume relative to body size approaches the uniquely expanded values expressed by mammals. The broad suite of complex behaviors exhibited by crown-group birds, including sociality, vocal learning, parental care, and flying, suggests the origins of their encephalization was likely driven by a mosaic of selective pressures. If true, the historical pattern of brain expansion may be more complex than either a gradual expansion, as proposed by early studies of the avian brain, or a sudden expansion correlating with the appearance of flight. The origins of modern avian neuroanatomy are obscured by the more than 100 million years of evolution along their phylogenetic stem (from the origin of the modern radiation in the Middle Jurassic to the split from crocodile-line archosaurs). Here we use phylogenetic comparative approaches to explore which evolutionary scenarios best explain variation in measured volumes of digitally partitioned endocasts of modern birds and their non-avian ancestors. Our analyses suggest that variation in the relative volumes of the endocranium and cerebrum explain most of the structural variation in this lineage. Generalized multi-regime Ornstein-Uhlenbeck (OU) models suggest that powered flight does not appear to be a driver of observed variation, reinforcing the hypothesis that the deep history of the avian brain is complex, with nuances still to be discovered. © 2015 Anatomical Society.
Cortical Memory Mechanisms and Language Origins
ERIC Educational Resources Information Center
Aboitiz, Francisco; Garcia, Ricardo R.; Bosman, Conrado; Brunetti, Enzo
2006-01-01
We have previously proposed that cortical auditory-vocal networks of the monkey brain can be partly homologized with language networks that participate in the phonological loop. In this paper, we suggest that other linguistic phenomena like semantic and syntactic processing also rely on the activation of transient memory networks, which can be…
Li, XiaoChing; Wang, Xiu-Jie; Tannenhauser, Jonathan; Podell, Sheila; Mukherjee, Piali; Hertel, Moritz; Biane, Jeremy; Masuda, Shoko; Nottebohm, Fernando; Gaasterland, Terry
2007-01-01
Vocal learning and neuronal replacement have been studied extensively in songbirds, but until recently, few molecular and genomic tools for songbird research existed. Here we describe new molecular/genomic resources developed in our laboratory. We made cDNA libraries from zebra finch (Taeniopygia guttata) brains at different developmental stages. A total of 11,000 cDNA clones from these libraries, representing 5,866 unique gene transcripts, were randomly picked and sequenced from the 3′ ends. A web-based database was established for clone tracking, sequence analysis, and functional annotations. Our cDNA libraries were not normalized. Sequencing ESTs without normalization produced many developmental stage-specific sequences, yielding insights into patterns of gene expression at different stages of brain development. In particular, the cDNA library made from brains at posthatching day 30–50, corresponding to the period of rapid song system development and song learning, has the most diverse and richest set of genes expressed. We also identified five microRNAs whose sequences are highly conserved between zebra finch and other species. We printed cDNA microarrays and profiled gene expression in the high vocal center of both adult male zebra finches and canaries (Serinus canaria). Genes differentially expressed in the high vocal center were identified from the microarray hybridization results. Selected genes were validated by in situ hybridization. Networks among the regulated genes were also identified. These resources provide songbird biologists with tools for genome annotation, comparative genomics, and microarray gene expression analysis. PMID:17426146
"Bird Song Metronomics": Isochronous Organization of Zebra Finch Song Rhythm.
Norton, Philipp; Scharff, Constance
2016-01-01
The human capacity for speech and vocal music depends on vocal imitation. Songbirds, in contrast to non-human primates, share this vocal production learning with humans. The process through which birds and humans learn many of their vocalizations as well as the underlying neural system exhibit a number of striking parallels and have been widely researched. In contrast, rhythm, a key feature of language, and music, has received surprisingly little attention in songbirds. Investigating temporal periodicity in bird song has the potential to inform the relationship between neural mechanisms and behavioral output and can also provide insight into the biology and evolution of musicality. Here we present a method to analyze birdsong for an underlying rhythmic regularity. Using the intervals from one note onset to the next as input, we found for each bird an isochronous sequence of time stamps, a "signal-derived pulse," or pulse(S), of which a subset aligned with all note onsets of the bird's song. Fourier analysis corroborated these results. To determine whether this finding was just a byproduct of the duration of notes and intervals typical for zebra finches but not dependent on the individual duration of elements and the sequence in which they are sung, we compared natural songs to models of artificial songs. Note onsets of natural song deviated from the pulse(S) significantly less than those of artificial songs with randomized note and gap durations. Thus, male zebra finch song has the regularity required for a listener to extract a perceived pulse (pulse(P)), as yet untested. Strikingly, in our study, pulses(S) that best fit note onsets often also coincided with the transitions between sub-note elements within complex notes, corresponding to neuromuscular gestures. Gesture durations often equaled one or more pulse(S) periods. This suggests that gesture duration constitutes the basic element of the temporal hierarchy of zebra finch song rhythm, an interesting parallel to the hierarchically structured components of regular rhythms in human music.
“Bird Song Metronomics”: Isochronous Organization of Zebra Finch Song Rhythm
Norton, Philipp; Scharff, Constance
2016-01-01
The human capacity for speech and vocal music depends on vocal imitation. Songbirds, in contrast to non-human primates, share this vocal production learning with humans. The process through which birds and humans learn many of their vocalizations as well as the underlying neural system exhibit a number of striking parallels and have been widely researched. In contrast, rhythm, a key feature of language, and music, has received surprisingly little attention in songbirds. Investigating temporal periodicity in bird song has the potential to inform the relationship between neural mechanisms and behavioral output and can also provide insight into the biology and evolution of musicality. Here we present a method to analyze birdsong for an underlying rhythmic regularity. Using the intervals from one note onset to the next as input, we found for each bird an isochronous sequence of time stamps, a “signal-derived pulse,” or pulseS, of which a subset aligned with all note onsets of the bird's song. Fourier analysis corroborated these results. To determine whether this finding was just a byproduct of the duration of notes and intervals typical for zebra finches but not dependent on the individual duration of elements and the sequence in which they are sung, we compared natural songs to models of artificial songs. Note onsets of natural song deviated from the pulseS significantly less than those of artificial songs with randomized note and gap durations. Thus, male zebra finch song has the regularity required for a listener to extract a perceived pulse (pulseP), as yet untested. Strikingly, in our study, pulsesS that best fit note onsets often also coincided with the transitions between sub-note elements within complex notes, corresponding to neuromuscular gestures. Gesture durations often equaled one or more pulseS periods. This suggests that gesture duration constitutes the basic element of the temporal hierarchy of zebra finch song rhythm, an interesting parallel to the hierarchically structured components of regular rhythms in human music. PMID:27458334
Halwani, Gus F; Loui, Psyche; Rüber, Theodor; Schlaug, Gottfried
2011-01-01
Structure and function of the human brain are affected by training in both linguistic and musical domains. Individuals with intensive vocal musical training provide a useful model for investigating neural adaptations of learning in the vocal-motor domain and can be compared with learning in a more general musical domain. Here we confirm general differences in macrostructure (tract volume) and microstructure (fractional anisotropy, FA) of the arcuate fasciculus (AF), a prominent white-matter tract connecting temporal and frontal brain regions, between singers, instrumentalists, and non-musicians. Both groups of musicians differed from non-musicians in having larger tract volume and higher FA values of the right and left AF. The AF was then subdivided in a dorsal (superior) branch connecting the superior temporal gyrus and the inferior frontal gyrus (STG ↔ IFG), and ventral (inferior) branch connecting the middle temporal gyrus and the inferior frontal gyrus (MTG ↔ IFG). Relative to instrumental musicians, singers had a larger tract volume but lower FA values in the left dorsal AF (STG ↔ IFG), and a similar trend in the left ventral AF (MTG ↔ IFG). This between-group comparison controls for the general effects of musical training, although FA was still higher in singers compared to non-musicians. Both musician groups had higher tract volumes in the right dorsal and ventral tracts compared to non-musicians, but did not show a significant difference between each other. Furthermore, in the singers' group, FA in the left dorsal branch of the AF was inversely correlated with the number of years of participants' vocal training. Our findings suggest that long-term vocal-motor training might lead to an increase in volume and microstructural complexity of specific white-matter tracts connecting regions that are fundamental to sound perception, production, and its feedforward and feedback control which can be differentiated from a more general musician effect.
Surgical Interventions for Pediatric Unilateral Vocal Cord Paralysis: A Systematic Review.
Butskiy, Oleksandr; Mistry, Bhavik; Chadha, Neil K
2015-07-01
The most widely used surgical interventions for pediatric unilateral vocal cord paralysis include injection laryngoplasty, thyroplasty, and laryngeal reinnervation. Despite increasing interest in surgical interventions for unilateral vocal cord paralysis in children, the surgical outcomes data in children are scarce. To appraise and summarize the available evidence for pediatric unilateral vocal cord paralysis surgical strategies. MEDLINE (1946-2014) and EMBASE (1980-2014) were searched for publications that described the results of laryngoplasty, thyroplasty, or laryngeal reinnervation for pediatric unilateral vocal cord paralysis. Further studies were identified from bibliographies of relevant studies, gray literature, and annual scientific assemblies. Two reviewers independently appraised the selected studies for quality, level of evidence, and risk of bias as well as extracted data, including unilateral vocal cord paralysis origin, voice outcomes, swallowing outcomes, and adverse events. Of 366 identified studies, the inclusion criteria were met by 15 studies: 6 observational studies, 6 case series, and 3 case reports. All 36 children undergoing laryngeal reinnervation (8 studies) had improvement or resolution of dysphonia. Of 31 children receiving injection laryngoplasty (6 studies), most experienced improvement in voice quality, speech, swallowing, aspiration, and glottic closure. Of 12 children treated by thyroplasty (5 studies), 2 experienced resolution of dysphonia, 4 had some improvement, and 4 had no improvement (2 patients had undocumented outcomes). Thyroplasty resolved or improved aspiration in 7 of 8 patients. Published studies suggest that reinnervation may be the most effective surgical intervention for children with dysphonia; however, long-term follow-up data are lacking. With the exception of polytetrafluoroethylene injections, injection laryngoplasty was reported to be a relatively safe, nonpermanent, and effective option for most children with dysphonia. Thyroplasty appears to have fallen out favor in recent years because of difficulty in performing this procedure in children under local anesthesia, but it continues to be a viable option for children with aspiration.
Changes in the neural control of a complex motor sequence during learning
Otchy, Timothy M.; Goldberg, Jesse H.; Aronov, Dmitriy; Fee, Michale S.
2011-01-01
The acquisition of complex motor sequences often proceeds through trial-and-error learning, requiring the deliberate exploration of motor actions and the concomitant evaluation of the resulting performance. Songbirds learn their song in this manner, producing highly variable vocalizations as juveniles. As the song improves, vocal variability is gradually reduced until it is all but eliminated in adult birds. In the present study we examine how the motor program underlying such a complex motor behavior evolves during learning by recording from the robust nucleus of the arcopallium (RA), a motor cortex analog brain region. In young birds, neurons in RA exhibited highly variable firing patterns that throughout development became more precise, sparse, and bursty. We further explored how the developing motor program in RA is shaped by its two main inputs: LMAN, the output nucleus of a basal ganglia-forebrain circuit, and HVC, a premotor nucleus. Pharmacological inactivation of LMAN during singing made the song-aligned firing patterns of RA neurons adultlike in their stereotypy without dramatically affecting the spike statistics or the overall firing patterns. Removing the input from HVC, on the other hand, resulted in a complete loss of stereotypy of both the song and the underlying motor program. Thus our results show that a basal ganglia-forebrain circuit drives motor exploration required for trial-and-error learning by adding variability to the developing motor program. As learning proceeds and the motor circuits mature, the relative contribution of LMAN is reduced, allowing the premotor input from HVC to drive an increasingly stereotyped song. PMID:21543758
Models of vocal learning in the songbird: Historical frameworks and the stabilizing critic.
Nick, Teresa A
2015-10-01
Birdsong is a form of sensorimotor learning that involves a mirror-like system that activates with both song hearing and production. Early models of song learning, based on behavioral measures, identified key features of vocal plasticity, such as the requirements for memorization of a tutor song and auditory feedback during song practice. The concept of a comparator, which compares the memory of the tutor song to auditory feedback, featured prominently. Later models focused on linking anatomically-defined neural modules to behavioral concepts, such as the comparator. Exploiting the anatomical modularity of the songbird brain, localized lesions illuminated mechanisms of the neural song system. More recent models have integrated neuronal mechanisms identified in other systems with observations in songbirds. While these models explain multiple aspects of song learning, they must incorporate computational elements based on unknown biological mechanisms to bridge the motor-to-sensory delay and/or transform motor signals into the sensory domain. Here, I introduce the stabilizing critic hypothesis, which enables sensorimotor learning by (1) placing a purely sensory comparator afferent of the song system and (2) endowing song system disinhibitory interneuron networks with the capacity both to bridge the motor-sensory delay through prolonged bursting and to stabilize song segments selectively based on the comparator signal. These proposed networks stabilize an otherwise variable signal generated by both putative mirror neurons and a cortical-basal ganglia-thalamic loop. This stabilized signal then temporally converges with a matched premotor signal in the efferent song motor cortex, promoting spike-timing-dependent plasticity in the premotor circuitry and behavioral song learning. © 2014 Wiley Periodicals, Inc.
ERIC Educational Resources Information Center
Burks, Beatrice Karol; Reeves, Richard
2009-01-01
Despite a vocal commitment to fairness, the British Government has, according to these authors, wholeheartedly failed to live up to this pledge when it comes to skills and adult training. A report on adult learning released in December by Demos found a system rife with inequality and contradictions. As the jobs market becomes increasingly…
Social Modulation of Associative Fear Learning by Pheromone Communication
ERIC Educational Resources Information Center
Bredy, Timothy W.; Barad, Mark
2009-01-01
Mice communicate through visual, vocal, and olfactory cues that influence innate, nonassociative behavior. We here report that exposure to a recently fear-conditioned familiar mouse impairs acquisition of conditioned fear and facilitates fear extinction, effects mimicked by both an olfactory chemosignal emitted by a recently fear-conditioned…
Tourette Syndrome: A Case for Establishing the Individual Needs of Children at Risk.
ERIC Educational Resources Information Center
Wilson, Jeni; Shrimpton, Bradley
Tourette Syndrome (TS) is a neurological disorder characterized by multiple, involuntary, and repetitive motor and vocal tics. This paper addresses the educational needs of students with TS noting that, without proper intervention and appropriate learning experiences, these children often experience personal distress, reduced self-esteem, social…
Stimulus-Dependent Flexibility in Non-Human Auditory Pitch Processing
ERIC Educational Resources Information Center
Bregman, Micah R.; Patel, Aniruddh D.; Gentner, Timothy Q.
2012-01-01
Songbirds and humans share many parallels in vocal learning and auditory sequence processing. However, the two groups differ notably in their abilities to recognize acoustic sequences shifted in absolute pitch (pitch height). Whereas humans maintain accurate recognition of words or melodies over large pitch height changes, songbirds are…
From Band Rooms to Choral Risers.
ERIC Educational Resources Information Center
Washington, William Lee
1999-01-01
Offers suggestions for instrumentalists who have become choral instructors explaining that their first goal should be to learn how to identify and understand an ideal vocal sound. Finds that the hardest aspect of choral instruction is developing a teaching approach. Offers five steps to assist instrumentalists in molding their own approach. (CMK)
The Importance of Music in Early Childhood.
ERIC Educational Resources Information Center
Levinowitz, Lili M.
1998-01-01
Surveys some of the research in music education that validates the inclusion of music for its own sake in models for early childhood learning. Focuses on topics that include, but are not limited to, child and vocal development, the importance of movement for children, and adult involvement in music education. (CMK)
Mechanisms of Song Perception in Oscine Birds
ERIC Educational Resources Information Center
Knudsen, Daniel P.; Gentner, Timothy Q.
2010-01-01
Songbirds share a number of parallels with humans that make them an attractive model system for studying the behavioral and neurobiological mechanisms that underlie the learning and processing of vocal communication signals. Here we review the perceptual and cognitive mechanisms of audition in birds, and emphasize the behavioral and neural basis…
Audio/ Videoconferencing Packages: Low Cost
ERIC Educational Resources Information Center
Treblay, Remy; Fyvie, Barb; Koritko, Brenda
2005-01-01
A comparison was conducted of "Voxwire MeetingRoom" and "iVocalize" v4.1.0.3, both Web-conferencing products using voice-over-Internet protocol (VoIP) to provide unlimited, inexpensive, international audio communication, and high-quality Web-conferencing fostering collaborative learning. The study used the evaluation criteria used in earlier…
Hodges-Simeon, Carolyn R; Gaulin, Steven J C; Puts, David A
2011-06-01
Men's copulatory success can often be predicted by measuring traits involved in male contests and female choice. Previous research has demonstrated relationships between one such vocal trait in men, mean fundamental frequency (F(0)), and the outcomes and indicators of sexual success with women. The present study investigated the role of another vocal parameter, F(0) variation (the within-subject SD in F(0) across the utterance, F(0)-SD), in predicting men's reported number of female sexual partners in the last year. Male participants (N = 111) competed with another man for a date with a woman. Recorded interactions with the competitor ("competitive recording") and the woman ("courtship recording") were analyzed for five non-linguistic vocal parameters: F(0)-SD, mean F(0), intensity, duration, and formant dispersion (D( f ), an acoustic correlate of vocal tract length), as well as dominant and attractive linguistic content. After controlling for age and attitudes toward uncommitted sex (SOI), lower F(0)-SD (i.e., a more monotone voice) and more dominant linguistic content were strong predictors of the number of past-year sexual partners, whereas mean F(0) and D( f ) did not significantly predict past-year partners. These contrasts have implications for the relative importance of male contests and female choice in shaping men's mating success and hence the origins and maintenance of sexually dimorphic traits in humans.
The odyssey of Orpheus: the evolution of solo singing.
Miller, R
1996-06-01
Notated sacred solo song dates from 1,000 B.C. Early secular song exhibits modest vocal demands of chant-like character. Popular song and liturgical solo song share common origins. Western European secular song notation began in the early Medieval Age. Compositional writing for solo voice took a dramatic turn toward virtuosity about 1600. By the mid-17th century, the modern solo voice emerged. "Classical" solo vocal literature is not static but is constantly evolving, requiring skills far in excess of those of speech or of early solo song literature; this is equally the case with ethnomusicological and popular singing styles. Efficient use of the singing instrument is essential to the healthy accomplishment of all of these literatures.
The evolution of laughter in great apes and humans
Owren, Michael J; Zimmermann, Elke
2010-01-01
It has long been claimed that human emotional expressions, such as laughter, have evolved from nonhuman displays. The aim of the current study was to test this prediction by conducting acoustic and phylogenetic analyses based on the acoustics of tickle-induced vocalizations of orangutans, gorillas, chimpanzees, bonobos and humans. Results revealed both important similarities and differences among the various species’ vocalizations, with the phylogenetic tree reconstructed based on these acoustic data matching the well-established genetic relationships of great apes and humans. These outcomes provide evidence of a common phylogenetic origin of tickle-induced vocalizations in these taxa, which can therefore be termed “laughter” across all five species. Results are consistent with the claims of phylogenetic continuity of emotional expressions. Together with observations made on the use of laughter in great apes and humans, findings of this study further indicate that there were two main periods of selection-driven evolutionary change in laughter within the Hominidae, to a smaller degree, among the great apes and, most distinctively, after the separation of hominins from the last common ancestor with chimpanzees and bonobos. PMID:20585520
Erath, Byron D.; Peterson, Sean D.; Zañartu, Matías; Wodicka, George R.; Stewart, Kelley C.; Plesniak, Michael W.
2013-01-01
Hirschberg [J. Acoust. Soc. Am. 134, 9-12 (2013)] presents a commentary and criticisms of the viscous flow model presented by Erath et al. [J. Acoust. Soc. Am. 130, 389–403 (2011)] that solves for the asymmetric pressure loading on the vocal fold walls. This pressure loading arises from asymmetric flow attachment to one vocal fold wall when the glottal channel forms a divergent configuration. Hirschberg proposes an alternative model for the asymmetric loading based upon inviscid flow curvature at the glottal inlet. In this manuscript further evidence is provided in support of the model of Erath et al. and the underlying assumptions, and demonstrates that the primary criticisms presented by Hirschberg are unwarranted. The model presented by Hirschberg is compared with the model from the original paper by Erath et al., and it is shown that each model describes different and complementary aspects of divergent glottal flows. PMID:23927090
NCS-1 dependent learning bonus and behavior outputs of self-directed exploration
NASA Astrophysics Data System (ADS)
Mun, Ho-Suk
Animals explore a new environment and learn about their surroundings. "Exploration" refers to all activities that increase the information obtained from an animal. For this study, I determined a molecule that mediates self-directed exploration, with a particular focus on rearing behavior and vocalization. Rearing can be either self-directed exploration or escape-oriented exploration. Self-directed exploration can be driven by the desire to gather information about environments while escape-oriented exploration can be driven by fear or anxiety. To differentiate between these two concepts, I compared rearing and other behaviors in three different conditions 1) novel dim (safe environment), which induces exploration based rearing; 2) novel bright (fearful environment), which elicits fear driven rearing; and 3) familiar environment as a control. First, I characterized the effects on two distinct types of environment in exploratory behavior and its effect on learning. From this, I determined that self-directed exploration enhances spatial learning while escape-oriented exploration does not produce a learning bonus. Second, I found that NCS-1 is involved in exploration, as well as learning and memory, by testing mice with reduced levels of Ncs-1 by point mutation and also siRNA injection. Finally, I illustrated other behavior outputs and neural substrate activities, which co-occurred during either self-directed or escape-oriented exploration. I found that high-frequency ultrasonic vocalizations occurred during self-directed exploration while low-frequency calls were emitted during escape-oriented exploration. Also, with immediate early gene imaging techniques, I found hippocampus and nucleus accumbens activation in self-directed exploration. This study is the first comprehensive molecular analysis of learning bonus in self-directed exploration. These results may be beneficial for studying underlying mechanisms of neuropsychiatric disease, and also reveal therapeutic targets for them.
Vibratory regime classification of infant phonation.
Buder, Eugene H; Chorna, Lesya B; Oller, D Kimbrough; Robinson, Rebecca B
2008-09-01
Infant phonation is highly variable in many respects, including the basic vibratory patterns by which the vocal tissues create acoustic signals. Previous studies have identified the regular occurrence of nonmodal phonation types in normal infant phonation. The glottis is like many oscillating systems that, because of nonlinear relationships among the elements, may vibrate in ways representing the deterministic patterns classified theoretically within the mathematical framework of nonlinear dynamics. The infant's preverbal vocal explorations present such a variety of phonations that it may be possible to find effectively all the classes of vibration predicted by nonlinear dynamic theory. The current report defines acoustic criteria for an important subset of such vibratory regimes, and demonstrates that analysts can be trained to reliably use these criteria for a classification that includes all instances of infant phonation in the recorded corpora. The method is thus internally comprehensive in the sense that all phonations are classified, but it is not exhaustive in the sense that all vocal qualities are thereby represented. Using the methods thus developed, this study also demonstrates that the distributions of these phonation types vary significantly across sessions of recording in the first year of life, suggesting developmental changes. The method of regime classification is thus capable of tracking changes that may be indicative of maturation of the mechanism, the learning of categories of phonatory control, and the possibly varying use of vocalizations across social contexts.
Meguerditchian, Adrien; Taglialatela, Jared P; Leavens, David A; Hopkins, William D
2014-12-01
Ackermann et al. mention the "acquisition of species-atypical sounds" in apes without any discussion. In our commentary, we demonstrate that these atypical sounds in chimpanzees not only include laryngeal sounds, but also have a major significance regarding the origins of language, if we consider looking at their context of use, their social properties, their relations with gestures, their lateralization, and their neurofunctional correlates as well.
The ecology of entrainment: Foundations of coordinated rhythmic movement.
Phillips-Silver, Jessica; Aktipis, C Athena; Bryant, Gregory A
2010-09-01
Entrainment has been studied in a variety of contexts including music perception, dance, verbal communication and motor coordination more generally. Here we seek to provide a unifying framework that incorporates the key aspects of entrainment as it has been studied in these varying domains. We propose that there are a number of types of entrainment that build upon pre-existing adaptations that allow organisms to perceive stimuli as rhythmic, to produce periodic stimuli, and to integrate the two using sensory feedback. We suggest that social entrainment is a special case of spatiotemporal coordination where the rhythmic signal originates from another individual. We use this framework to understand the function and evolutionary basis for coordinated rhythmic movement and to explore questions about the nature of entrainment in music and dance. The framework of entrainment presented here has a number of implications for the vocal learning hypothesis and other proposals for the evolution of coordinated rhythmic behavior across an array of species.
The ecology of entrainment: Foundations of coordinated rhythmic movement
Phillips-Silver, Jessica; Aktipis, C. Athena; Bryant, Gregory A.
2011-01-01
Entrainment has been studied in a variety of contexts including music perception, dance, verbal communication and motor coordination more generally. Here we seek to provide a unifying framework that incorporates the key aspects of entrainment as it has been studied in these varying domains. We propose that there are a number of types of entrainment that build upon pre-existing adaptations that allow organisms to perceive stimuli as rhythmic, to produce periodic stimuli, and to integrate the two using sensory feedback. We suggest that social entrainment is a special case of spatiotemporal coordination where the rhythmic signal originates from another individual. We use this framework to understand the function and evolutionary basis for coordinated rhythmic movement and to explore questions about the nature of entrainment in music and dance. The framework of entrainment presented here has a number of implications for the vocal learning hypothesis and other proposals for the evolution of coordinated rhythmic behavior across an array of species. PMID:21776183
Perez-Pouchoulen, Miguel; Roby, Clinton R.; Ryan, Timothy E.; McCarthy, Margaret M.
2014-01-01
Sex differences in vocal communication are prevalent in both the animals and humans. The mechanism(s) mediating gender differences in human language are unknown, although, sex hormones, principally androgens, play a central role in the development of vocalizations in a wide variety of animal species. The discovery of FOXP2 has added an additional avenue for exploring the origins of language and animal communication. The FOXP2 gene is a member of the forkhead box P (FOXP) family of transcription factors. Prior to the prenatal androgen surge in male fetuses, we observed no sex difference for Foxp2 protein levels in cultured cells. In contrast, 24 hours after the onset of the androgen surge, we found a sex difference for Foxp2 protein levels in cultured cortical cells with males having higher levels than females. Furthermore, we observed the potent nonaromatizable androgen dihydrotestosterone altered not only Foxp2 mRNA and protein levels but also Foxp1. Androgen effects on both Foxp2 and Foxp1 were found to occur in the striatum, cerebellar vermis, and cortex. Immunofluorescence microscopy and coimmunoprecipitation demonstrate Foxp2 and the androgen receptor protein interact. Databases for transcription factor binding sites predict a consensus binding motif for androgen receptor on the Foxp2 promoter regions. We also observed a sex difference in rat pup vocalization with males vocalizing more than females and treatment of females with dihydrotestosterone eliminated the sex difference. We propose that androgens might be an upstream regulator of both Foxp2 and Foxp1 expression and signaling. This has important implications for language and communication as well as neuropsychiatric developmental disorders involving impairments in communication. PMID:25247470
Bowers, J Michael; Perez-Pouchoulen, Miguel; Roby, Clinton R; Ryan, Timothy E; McCarthy, Margaret M
2014-12-01
Sex differences in vocal communication are prevalent in both the animals and humans. The mechanism(s) mediating gender differences in human language are unknown, although, sex hormones, principally androgens, play a central role in the development of vocalizations in a wide variety of animal species. The discovery of FOXP2 has added an additional avenue for exploring the origins of language and animal communication. The FOXP2 gene is a member of the forkhead box P (FOXP) family of transcription factors. Prior to the prenatal androgen surge in male fetuses, we observed no sex difference for Foxp2 protein levels in cultured cells. In contrast, 24 hours after the onset of the androgen surge, we found a sex difference for Foxp2 protein levels in cultured cortical cells with males having higher levels than females. Furthermore, we observed the potent nonaromatizable androgen dihydrotestosterone altered not only Foxp2 mRNA and protein levels but also Foxp1. Androgen effects on both Foxp2 and Foxp1 were found to occur in the striatum, cerebellar vermis, and cortex. Immunofluorescence microscopy and coimmunoprecipitation demonstrate Foxp2 and the androgen receptor protein interact. Databases for transcription factor binding sites predict a consensus binding motif for androgen receptor on the Foxp2 promoter regions. We also observed a sex difference in rat pup vocalization with males vocalizing more than females and treatment of females with dihydrotestosterone eliminated the sex difference. We propose that androgens might be an upstream regulator of both Foxp2 and Foxp1 expression and signaling. This has important implications for language and communication as well as neuropsychiatric developmental disorders involving impairments in communication.
A model of acoustic interspeaker variability based on the concept of formant-cavity affiliation
NASA Astrophysics Data System (ADS)
Apostol, Lian; Perrier, Pascal; Bailly, Gérard
2004-01-01
A method is proposed to model the interspeaker variability of formant patterns for oral vowels. It is assumed that this variability originates in the differences existing among speakers in the respective lengths of their front and back vocal-tract cavities. In order to characterize, from the spectral description of the acoustic speech signal, these vocal-tract differences between speakers, each formant is interpreted, according to the concept of formant-cavity affiliation, as a resonance of a specific vocal-tract cavity. Its frequency can thus be directly related to the corresponding cavity length, and a transformation model can be proposed from a speaker A to a speaker B on the basis of the frequency ratios of the formants corresponding to the same resonances. In order to minimize the number of sounds to be recorded for each speaker in order to carry out this speaker transformation, the frequency ratios are exactly computed only for the three extreme cardinal vowels [eye, aye, you] and they are approximated for the remaining vowels through an interpolation function. The method is evaluated through its capacity to transform the (F1,F2) formant patterns of eight oral vowels pronounced by five male speakers into the (F1,F2) patterns of the corresponding vowels generated by an articulatory model of the vocal tract. The resulting formant patterns are compared to those provided by normalization techniques published in the literature. The proposed method is found to be efficient, but a number of limitations are also observed and discussed. These limitations can be associated with the formant-cavity affiliation model itself or with a possible influence of speaker-specific vocal-tract geometry in the cross-sectional direction, which the model might not have taken into account.
The Effects of Vocal Activity and Race of Applicant on Job Selection Interview Decisions.
ERIC Educational Resources Information Center
Byrd, Marquita L.
Because interviewing is a face-to-face interaction belonging to the genre of interpersonal relationships, the employment interview is subject to some of the same problems that beset interpersonal relationships. Problems can occur in both interviews and interpersonal relationships because of bias, which originates from background characteristics,…
Social modulation of associative fear learning by pheromone communication
Bredy, Timothy W.; Barad, Mark
2009-01-01
Mice communicate through visual, vocal, and olfactory cues that influence innate, nonassociative behavior. We here report that exposure to a recently fear-conditioned familiar mouse impairs acquisition of conditioned fear and facilitates fear extinction, effects mimicked by both an olfactory chemosignal emitted by a recently fear-conditioned familiar mouse and by the putative stress-related anxiogenic pheromone β-phenylethylamine (β-PEA). Together, these findings suggest social modulation of higher-order cognitive processing through pheromone communication and support the concurrent excitor hypothesis of extinction learning. PMID:19117912
Social modulation of associative fear learning by pheromone communication.
Bredy, Timothy W; Barad, Mark
2009-01-01
Mice communicate through visual, vocal, and olfactory cues that influence innate, nonassociative behavior. We here report that exposure to a recently fear-conditioned familiar mouse impairs acquisition of conditioned fear and facilitates fear extinction, effects mimicked by both an olfactory chemosignal emitted by a recently fear-conditioned familiar mouse and by the putative stress-related anxiogenic pheromone beta-phenylethylamine (beta-PEA). Together, these findings suggest social modulation of higher-order cognitive processing through pheromone communication and support the concurrent excitor hypothesis of extinction learning.
Individual killer whale vocal variation during intra-group behavioral dynamics
NASA Astrophysics Data System (ADS)
Grebner, Dawn M.
The scientific goal of this dissertation was to carefully study the signal structure of killer whale communications and vocal complexity and link them to behavioral circumstances. The overall objective of this research sought to provide insight into killer whale call content and usage which may be conveying information to conspecifics in order to maintain group cohesion. Data were collected in the summers of 2006 and 2007 in Johnstone Strait, British Columbia. For both individuals and small groups, vocalizations were isolated using a triangular hydrophone array and the behavioral movement patterns were captured by a theodolite and video camera positioned on a cliff overlooking the hyrophone locations. This dissertation is divided into four analysis chapters. In Chapter 3, discriminant analysis was used to validate the four N04 call subtypes which were originally parsed due to variations in slope segments. The first two functions of the discriminant analysis explained 97% of the variability. Most of the variability for the N04 call was found in the front convex and the terminal portions of the call, while very little variability was found in the center region of the call. This research revealed that individual killer whales produced multiple subtypes of the N04 call. No correlations of behaviors to acoustic parameters obtained were found. The aim of the Chapter 4 was to determine if killer whale calling behavior varied prior to and after the animals had joined. Pulsed call rates were found to be greater pre- compared to post-joining events. Two-way vocal exchanges were more common occurring 74% of the time during pre-joining events. In Chapter 5, initiated and first response to calls varied between age/sex class groups when mothers were separated from an offspring. Solo mothers and calves initiated pulsed calls more often than they responded. Most of the no vocal responses were due to mothers who were foraging. Finally, observations of the frequency split in N04 calls discussed in Chapter 6 showed that the higher frequency component (HFC) was always associated with sideband 7 (SB7) of the lower frequency component (LFC). Insight into Northern Resident killer whale intra-group vocal dynamics would aid our understanding of vocal behaviors of many other marine mammal species that rely on vocal exchanges for prey capture, group movement or survival. This is the first study to focus on killer whale vocal content and usage as it pertains to intra-group dynamics for (1) mother and offspring separations and (2) for all individuals prior to joining events, as well as (3) individual usage in a diverging pulsed call. It is also the first time the N04 call has been parsed into subtypes.
Acquisition of a Non-Vocal 'Language' by Aphasic Children
ERIC Educational Resources Information Center
Hughes, Jennifer
1974-01-01
Aphasic children were taught to communicate via a system of visual symbols devised by Premack (1969) for use with chimpanzees. Subjects readily learned to express several language functions in this way. "Premackese" is seen better viewed as a communication system. It may be that Aphasic children lack some specifically linguistic ability.…
Learning Curves: Expanding the Constituency for Comprehensive Sexuality Education, Fall 2005
ERIC Educational Resources Information Center
Sen, Rinku; Fellner, Kim
2005-01-01
Support in the United States for comprehensive sexuality education is overwhelming. Yet a small, vocal opposition and increasingly hostile public policy have deterred its implementation in many communities across the country. The chasm between community needs and public policy prompted the Ms. Foundation for Women and the David and Lucille Packard…
Language Development in Nonverbal Autistic Children Using a Simultaneous Communication System.
ERIC Educational Resources Information Center
Creedon, Margaret Procyk
Twenty-one nonverbal autistic children, 4- to 9-years-old, with language ages of 4- to 24-months, participated in the communication learning program from 1 to 3 years. Simultaneous verbal and manual signs were chosen as the communications mode. The children initially displayed infrequent, unrecognizable vocalizations (Screeches, or vocal…
Applying Infant Massage Practices: A Qualitative Study
ERIC Educational Resources Information Center
Lappin, Grace; Kretschmer, Robert E.
2005-01-01
This study explored the dynamic interaction between a mother and her 11-month-old visually impaired infant before and after the mother was taught infant massage. After the mother learned infant massage, she had more appropriate physical contact with her infant, engaged with him within his field of vision, directly vocalized to him, and had a…
A New Model of Sensorimotor Coupling in the Development of Speech
ERIC Educational Resources Information Center
Westermann, Gert; Miranda, Eduardo Reck
2004-01-01
We present a computational model that learns a coupling between motor parameters and their sensory consequences in vocal production during a babbling phase. Based on the coupling, preferred motor parameters and prototypically perceived sounds develop concurrently. Exposure to an ambient language modifies perception to coincide with the sounds from…
Australian TESOL Teachers' Cultural Perceptions of Students
ERIC Educational Resources Information Center
Ronai, Kara; Lammervo, Tiina
2017-01-01
Over the last decade, research in the field of Teaching English to Speakers of Other Languages (TESOL) has increasingly focused on the relationship between culture and learning. Researchers such as Kumaravadivelu (2003) have been vocal in their opposition to the practice of cultural stereotyping. In the current study, Holliday's (2005) model of…
The Role of Students and Content in Teacher Effectiveness
ERIC Educational Resources Information Center
Ennis, Catherine D.
2014-01-01
The process of effective teaching--teaching that directly leads to student learning of standards-based content--is tenuous at best and easily disrupted by contextual and behavioral factors. In this commentary, I discuss the role of student support and mediation in teacher effectiveness and curricular reform. The most vocal students in physical…
ERIC Educational Resources Information Center
White, Stephanie A.
2010-01-01
Could a mutation in a single gene be the evolutionary lynchpin supporting the development of human language? A rare mutation in the molecule known as FOXP2 discovered in a human family seemed to suggest so, and its sequence phylogeny reinforced a Chomskian view that language emerged wholesale in humans. Spurred by this discovery, research in…
Intramodal and Intermodal Functioning of Normal and LD Children
ERIC Educational Resources Information Center
Heath, Earl J.; Early, George H.
1973-01-01
Assessed were the abilities of 50 normal 5-to 9-year-old children and 30 learning disabled 7-to 9-year-old children to recognize temporal patterns presented visually and auditorially (intramodal abilities) and to vocally produce the patterns whether presentation was visual or auditory (intramodal and cross-modal abilities). (MC)
The mimetic repertoire of the spotted bowerbird Ptilonorhynchus maculatus
NASA Astrophysics Data System (ADS)
Kelley, Laura A.; Healy, Susan D.
2011-06-01
Although vocal mimicry in songbirds is well documented, little is known about the function of such mimicry. One possibility is that the mimic produces the vocalisations of predatory or aggressive species to deter potential predators or competitors. Alternatively, these sounds may be learned in error as a result of their acoustic properties such as structural simplicity. We determined the mimetic repertoires of a population of male spotted bowerbirds Ptilonorhynchus maculatus, a species that mimics predatory and aggressive species. Although male mimetic repertoires contained an overabundance of vocalisations produced by species that were generally aggressive, there was also a marked prevalence of mimicry of sounds that are associated with alarm such as predator calls, alarm calls and mobbing calls, irrespective of whether the species being mimicked was aggressive or not. We propose that it may be the alarming context in which these sounds are first heard that may lead both to their acquisition and to their later reproduction. We suggest that enhanced learning capability during acute stress may explain vocal mimicry in many species that mimic sounds associated with alarm.
Guerra, Jaime E.; Cruz-Nieto, Javier; Ortiz-Maciel, Sonia Gabriela; Wright, Timothy F.
2012-01-01
The populations of many species are declining worldwide, and conservation efforts struggle to keep pace with extinction rates. Conservation biologists commonly employ strategies such as translocation and reintroduction, which move individuals of endangered species from one part of their range to another. Because individuals from endangered populations are nonexpendable, identifying any potential barriers to the establishment of viable populations prior to release of individuals should be a priority. This study evaluates the potential for learned communication signals to constrain conservation strategies such as reintroduction in an endangered species, the Thick-billed Parrot (Rhynchopsitta pachyrhyncha). We conducted vocal surveys at three geographically distinct breeding populations in the Sierra Madre Occidental of Chihuahua, Mexico. Acoustic analyses utilizing both spectrogram cross-correlations and parameter measurements from spectrograms revealed no significant differences among the three sites in two common call types. Calls did vary among individuals within a site. The apparent lack of significant geographic variation across sampled sites suggests that differences in learned communication signals are unlikely to pose a barrier to the integration of translocated individuals from different populations into newly established populations. PMID:22307993
Core and region-enriched networks of behaviorally regulated genes and the singing genome
Whitney, Osceola; Pfenning, Andreas R.; Howard, Jason T.; Blatti, Charles A; Liu, Fang; Ward, James M.; Wang, Rui; Audet, Jean-Nicolas; Kellis, Manolis; Mukherjee, Sayan; Sinha, Saurabh; Hartemink, Alexander J.; West, Anne E.; Jarvis, Erich D.
2015-01-01
Songbirds represent an important model organism for elucidating molecular mechanisms that link genes with complex behaviors, in part because they have discrete vocal learning circuits that have parallels with those that mediate human speech. We found that ~10% of the genes in the avian genome were regulated by singing, and we found a striking regional diversity of both basal and singing-induced programs in the four key song nuclei of the zebra finch, a vocal learning songbird. The region-enriched patterns were a result of distinct combinations of region-enriched transcription factors (TFs), their binding motifs, and presinging acetylation of histone 3 at lysine 27 (H3K27ac) enhancer activity in the regulatory regions of the associated genes. RNA interference manipulations validated the role of the calcium-response transcription factor (CaRF) in regulating genes preferentially expressed in specific song nuclei in response to singing. Thus, differential combinatorial binding of a small group of activity-regulated TFs and predefined epigenetic enhancer activity influences the anatomical diversity of behaviorally regulated gene networks. PMID:25504732
Chen, Wenli; Woo, Peak; Murry, Thomas
2017-09-01
High-speed videoendoscopy captures the cycle-to-cycle vibratory motion of each individual vocal fold in normal and severely disordered phonation. Therefore, it provides a direct method to examine the specific vibratory changes following vocal fold surgery. The purpose of this study was to examine the vocal fold vibratory pattern changes in the surgically treated pathologic vocal fold and the contralateral vocal fold in three vocal pathologies: vocal polyp (n = 3), paresis or paralysis (n = 3), and scar (n = 3). Digital kymography was used to extract high-speed kymographic vocal fold images at the mid-membranous region of the vocal fold. Spectral analysis was subsequently applied to the digital kymography to quantify the cycle-to-cycle movements of each vocal fold, expressed as a spectrum. Surgical modification resulted in significantly improved spectral power of the treated pathologic vocal fold. Furthermore, the contralateral vocal fold also presented with improved spectral power irrespective of vocal pathology. In comparison with normal vocal fold spectrum, postsurgical vocal fold vibrations continued to demonstrate decreased vibratory amplitude in both vocal folds. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Shepard, Kathryn N; Chong, Kelly K; Liu, Robert C
2016-01-01
Tonotopic map plasticity in the adult auditory cortex (AC) is a well established and oft-cited measure of auditory associative learning in classical conditioning paradigms. However, its necessity as an enduring memory trace has been debated, especially given a recent finding that the areal expansion of core AC tuned to a newly relevant frequency range may arise only transiently to support auditory learning. This has been reinforced by an ethological paradigm showing that map expansion is not observed for ultrasonic vocalizations (USVs) or for ultrasound frequencies in postweaning dams for whom USVs emitted by pups acquire behavioral relevance. However, whether transient expansion occurs during maternal experience is not known, and could help to reveal the generality of cortical map expansion as a correlate for auditory learning. We thus mapped the auditory cortices of maternal mice at postnatal time points surrounding the peak in pup USV emission, but found no evidence of frequency map expansion for the behaviorally relevant high ultrasound range in AC. Instead, regions tuned to low frequencies outside of the ultrasound range show progressively greater suppression of activity in response to the playback of ultrasounds or pup USVs for maternally experienced animals assessed at their pups' postnatal day 9 (P9) to P10, or postweaning. This provides new evidence for a lateral-band suppression mechanism elicited by behaviorally meaningful USVs, likely enhancing their population-level signal-to-noise ratio. These results demonstrate that tonotopic map enlargement has limits as a construct for conceptualizing how experience leaves neural memory traces within sensory cortex in the context of ethological auditory learning.
Neural imaging in songbirds using fiber optic fluorescence microscopy
NASA Astrophysics Data System (ADS)
Nooshabadi, Fatemeh; Hearn, Gentry; Lints, Thierry; Maitland, Kristen C.
2012-02-01
The song control system of juvenile songbirds is an important model for studying the developmental acquisition and generation of complex learned vocal motor sequences, two processes that are fundamental to human speech and language. To understand the neural mechanisms underlying song production, it is critical to characterize the activity of identified neurons in the song control system when the bird is singing. Neural imaging in unrestrained singing birds, although technically challenging, will advance our understanding of neural ensemble coding mechanisms in this system. We are exploring the use of a fiber optic microscope for functional imaging in the brain of behaving and singing birds in order to better understand the contribution of a key brain nucleus (high vocal center nucleus; HVC) to temporal aspects of song motor control. We have constructed a fluorescence microscope with LED illumination, a fiber bundle for transmission of fluorescence excitation and emission light, a ~2x GRIN lens, and a CCD for image acquisition. The system has 2 μm resolution, 375 μm field of view, 200 μm working distance, and 1 mm outer diameter. As an initial characterization of this setup, neurons in HVC were imaged using the fiber optic microscope after injection of quantum dots or fluorescent retrograde tracers into different song nuclei. A Lucid Vivascope confocal microscope was used to confirm the imaging results. Long-term imaging of the activity of these neurons in juvenile birds during singing may lead us to a better understanding of the central motor codes for song and the central mechanism by which auditory experience modifies song motor commands to enable vocal learning and imitation.
Sex differences in the representation of call stimuli in a songbird secondary auditory area
Giret, Nicolas; Menardy, Fabien; Del Negro, Catherine
2015-01-01
Understanding how communication sounds are encoded in the central auditory system is critical to deciphering the neural bases of acoustic communication. Songbirds use learned or unlearned vocalizations in a variety of social interactions. They have telencephalic auditory areas specialized for processing natural sounds and considered as playing a critical role in the discrimination of behaviorally relevant vocal sounds. The zebra finch, a highly social songbird species, forms lifelong pair bonds. Only male zebra finches sing. However, both sexes produce the distance call when placed in visual isolation. This call is sexually dimorphic, is learned only in males and provides support for individual recognition in both sexes. Here, we assessed whether auditory processing of distance calls differs between paired males and females by recording spiking activity in a secondary auditory area, the caudolateral mesopallium (CLM), while presenting the distance calls of a variety of individuals, including the bird itself, the mate, familiar and unfamiliar males and females. In males, the CLM is potentially involved in auditory feedback processing important for vocal learning. Based on both the analyses of spike rates and temporal aspects of discharges, our results clearly indicate that call-evoked responses of CLM neurons are sexually dimorphic, being stronger, lasting longer, and conveying more information about calls in males than in females. In addition, how auditory responses vary among call types differ between sexes. In females, response strength differs between familiar male and female calls. In males, temporal features of responses reveal a sensitivity to the bird's own call. These findings provide evidence that sexual dimorphism occurs in higher-order processing areas within the auditory system. They suggest a sexual dimorphism in the function of the CLM, contributing to transmit information about the self-generated calls in males and to storage of information about the bird's auditory experience in females. PMID:26578918
Avey, Marc T.; Hoeschele, Marisa; Moscicki, Michele K.; Bloomfield, Laurie L.; Sturdy, Christopher B.
2011-01-01
Songbird auditory areas (i.e., CMM and NCM) are preferentially activated to playback of conspecific vocalizations relative to heterospecific and arbitrary noise [1]–[2]. Here, we asked if the neural response to auditory stimulation is not simply preferential for conspecific vocalizations but also for the information conveyed by the vocalization. Black-capped chickadees use their chick-a-dee mobbing call to recruit conspecifics and other avian species to mob perched predators [3]. Mobbing calls produced in response to smaller, higher-threat predators contain more “D” notes compared to those produced in response to larger, lower-threat predators and thus convey the degree of threat of predators [4]. We specifically asked whether the neural response varies with the degree of threat conveyed by the mobbing calls of chickadees and whether the neural response is the same for actual predator calls that correspond to the degree of threat of the chickadee mobbing calls. Our results demonstrate that, as degree of threat increases in conspecific chickadee mobbing calls, there is a corresponding increase in immediate early gene (IEG) expression in telencephalic auditory areas. We also demonstrate that as the degree of threat increases for the heterospecific predator, there is a corresponding increase in IEG expression in the auditory areas. Furthermore, there was no significant difference in the amount IEG expression between conspecific mobbing calls or heterospecific predator calls that were the same degree of threat. In a second experiment, using hand-reared chickadees without predator experience, we found more IEG expression in response to mobbing calls than corresponding predator calls, indicating that degree of threat is learned. Our results demonstrate that degree of threat corresponds to neural activity in the auditory areas and that threat can be conveyed by different species signals and that these signals must be learned. PMID:21909363
Sex differences in the representation of call stimuli in a songbird secondary auditory area.
Giret, Nicolas; Menardy, Fabien; Del Negro, Catherine
2015-01-01
Understanding how communication sounds are encoded in the central auditory system is critical to deciphering the neural bases of acoustic communication. Songbirds use learned or unlearned vocalizations in a variety of social interactions. They have telencephalic auditory areas specialized for processing natural sounds and considered as playing a critical role in the discrimination of behaviorally relevant vocal sounds. The zebra finch, a highly social songbird species, forms lifelong pair bonds. Only male zebra finches sing. However, both sexes produce the distance call when placed in visual isolation. This call is sexually dimorphic, is learned only in males and provides support for individual recognition in both sexes. Here, we assessed whether auditory processing of distance calls differs between paired males and females by recording spiking activity in a secondary auditory area, the caudolateral mesopallium (CLM), while presenting the distance calls of a variety of individuals, including the bird itself, the mate, familiar and unfamiliar males and females. In males, the CLM is potentially involved in auditory feedback processing important for vocal learning. Based on both the analyses of spike rates and temporal aspects of discharges, our results clearly indicate that call-evoked responses of CLM neurons are sexually dimorphic, being stronger, lasting longer, and conveying more information about calls in males than in females. In addition, how auditory responses vary among call types differ between sexes. In females, response strength differs between familiar male and female calls. In males, temporal features of responses reveal a sensitivity to the bird's own call. These findings provide evidence that sexual dimorphism occurs in higher-order processing areas within the auditory system. They suggest a sexual dimorphism in the function of the CLM, contributing to transmit information about the self-generated calls in males and to storage of information about the bird's auditory experience in females.
The biology and evolution of music: a comparative perspective.
Fitch, W Tecumseh
2006-05-01
Studies of the biology of music (as of language) are highly interdisciplinary and demand the integration of diverse strands of evidence. In this paper, I present a comparative perspective on the biology and evolution of music, stressing the value of comparisons both with human language, and with those animal communication systems traditionally termed "song". A comparison of the "design features" of music with those of language reveals substantial overlap, along with some important differences. Most of these differences appear to stem from semantic, rather than structural, factors, suggesting a shared formal core of music and language. I next review various animal communication systems that appear related to human music, either by analogy (bird and whale "song") or potential homology (great ape bimanual drumming). A crucial comparative distinction is between learned, complex signals (like language, music and birdsong) and unlearned signals (like laughter, ape calls, or bird calls). While human vocalizations clearly build upon an acoustic and emotional foundation shared with other primates and mammals, vocal learning has evolved independently in our species since our divergence with chimpanzees. The convergent evolution of vocal learning in other species offers a powerful window into psychological and neural constraints influencing the evolution of complex signaling systems (including both song and speech), while ape drumming presents a fascinating potential homology with human instrumental music. I next discuss the archeological data relevant to music evolution, concluding on the basis of prehistoric bone flutes that instrumental music is at least 40,000 years old, and perhaps much older. I end with a brief review of adaptive functions proposed for music, concluding that no one selective force (e.g., sexual selection) is adequate to explaining all aspects of human music. I suggest that questions about the past function of music are unlikely to be answered definitively and are thus a poor choice as a research focus for biomusicology. In contrast, a comparative approach to music promises rich dividends for our future understanding of the biology and evolution of music.
Kumar, Veena; Croxson, Paula L; Simonyan, Kristina
2016-04-13
The laryngeal motor cortex (LMC) is essential for the production of learned vocal behaviors because bilateral damage to this area renders humans unable to speak but has no apparent effect on innate vocalizations such as human laughing and crying or monkey calls. Several hypotheses have been put forward attempting to explain the evolutionary changes from monkeys to humans that potentially led to enhanced LMC functionality for finer motor control of speech production. These views, however, remain limited to the position of the larynx area within the motor cortex, as well as its connections with the phonatory brainstem regions responsible for the direct control of laryngeal muscles. Using probabilistic diffusion tractography in healthy humans and rhesus monkeys, we show that, whereas the LMC structural network is largely comparable in both species, the LMC establishes nearly 7-fold stronger connectivity with the somatosensory and inferior parietal cortices in humans than in macaques. These findings suggest that important "hard-wired" components of the human LMC network controlling the laryngeal component of speech motor output evolved from an already existing, similar network in nonhuman primates. However, the evolution of enhanced LMC-parietal connections likely allowed for more complex synchrony of higher-order sensorimotor coordination, proprioceptive and tactile feedback, and modulation of learned voice for speech production. The role of the primary motor cortex in the formation of a comprehensive network controlling speech and language has been long underestimated and poorly studied. Here, we provide comparative and quantitative evidence for the significance of this region in the control of a highly learned and uniquely human behavior: speech production. From the viewpoint of structural network organization, we discuss potential evolutionary advances of enhanced temporoparietal cortical connections with the laryngeal motor cortex in humans compared with nonhuman primates that may have contributed to the development of finer vocal motor control necessary for speech production. Copyright © 2016 the authors 0270-6474/16/364170-12$15.00/0.
Optimal glottal configuration for ease of phonation.
Lucero, J C
1998-06-01
Recent experimental studies have shown the existence of optimal values of the glottal width and convergence angle, at which the phonation threshold pressure is minimum. These results indicate the existence of an optimal glottal configuration for ease of phonation, not predicted by the previous theory. In this paper, the origin of the optimal configuration is investigated using a low dimensional mathematical model of the vocal fold. Two phenomena of glottal aerodynamics are examined: pressure losses due to air viscosity, and air flow separation from a divergent glottis. The optimal glottal configuration seems to be a consequence of the combined effect of both factors. The results agree with the experimental data, showing that the phonation threshold pressure is minimum when the vocal folds are slightly separated in a near rectangular glottis.
Effect of the menstrual cycle on voice quality.
Silverman, E M; Zimmer, C H
1978-01-01
The question addressed was whether most young women with no vocal training exhibit premenstrual hoarseness. Spectral (acoustical) analyses of the sustained productions of three vowels produced by 20 undergraduates at and at premenstruation were rated for degree of hoarseness. Statistical analysis of the data indicated that the typical subject was no more hoarse of premenstruation than at ovulation. To determine whether this finding represented a genuine characteristic of women's voices or a type II statistical error, a systematic replication was undertaken with another sample of 27 undergraduates. The finding replicated that of the original investigation, suggesting that premenstrual hoarseness is a rarely occurring condition among young women with no vocal training. The apparent differential effect of the menstrual cycle on trained as opposed to untrained voices deserves systematic investigation.
Crance, Jessica L; Bowles, Ann E; Garver, Alan
2014-04-15
Killer whales (Orcinus orca) are thought to learn their vocal dialect. Dispersal in the species is rare, but effects of shifts in social association on the dialect can be studied under controlled conditions. Individual call repertoires and social association were measured in three adult female killer whales and three males (two juveniles and an adult) during two periods, 2001-2003 and 2005-2006. Three distinct dialect repertoires were represented among the subjects. An adventitious experiment in social change resulted from the birth of a calf and the transfer of two non-focal subjects in 2004. Across the two periods, 1691 calls were collected, categorized and attributed to individuals. Repertoire overlap for each subject dyad was compared with an index of association. During 2005-2006, the two juvenile males increased association with the unrelated adult male. By the end of the period, both had begun producing novel calls and call features characteristic of his repertoire. However, there was little or no reciprocal change and the adult females did not acquire his calls. Repertoire overlap and association were significantly correlated in the first period. In the second, median association time and repertoire similarity increased, but the relationship was only marginally significant. The results provided evidence that juvenile male killer whales are capable of learning new call types, possibly stimulated by a change in social association. The pattern of learning was consistent with a selective convergence of male repertoires.
A Preliminary Analysis of Early Rhythm and Blues Musical Practices.
ERIC Educational Resources Information Center
Meadows, Eddie S.
1983-01-01
Presents background information on the evolution of rhythm and blues (R & B) from the 1940s to the 1960s: the origin and naming of selected R & B groups, role of instruments in R & B orchestras, soloist/group vocal practices, and the role that independent record labels played in artists' successes and failures. (Author/ML)
Maternal Affective Expression and Infant's Mood in a Laboratory Situation: A Japanese Case.
ERIC Educational Resources Information Center
Chen, Shing-jen
The results of a Japanese replication of the vocalic social referencing experiment are discussed in comparison with original findings from the United States. The attempt to replicate failed. It is argued that the lack of success was due to (1) Japanese mothers' attitude toward laboratory situations, an attitude resulting from relative…
Helium Speech: An Application of Standing Waves
NASA Astrophysics Data System (ADS)
Wentworth, Christopher D.
2011-04-01
Taking a breath of helium gas and then speaking or singing to the class is a favorite demonstration for an introductory physics course, as it usually elicits appreciative laughter, which serves to energize the class session. Students will usually report that the helium speech "raises the frequency" of the voice. A more accurate description of the phenomenon requires that we distinguish between the frequencies of sound produced by the larynx and the filtering of those frequencies by the vocal tract. We will describe here an experiment done by introductory physics students that uses helium speech as a context for learning about the human vocal system and as an application of the standing sound-wave concept. Modern acoustic analysis software easily obtained by instructors for student use allows data to be obtained and analyzed quickly.
Cardin, Jessica A; Raksin, Jonathan N; Schmidt, Marc F
2005-04-01
Sensorimotor integration in the avian song system is crucial for both learning and maintenance of song, a vocal motor behavior. Although a number of song system areas demonstrate both sensory and motor characteristics, their exact roles in auditory and premotor processing are unclear. In particular, it is unknown whether input from the forebrain nucleus interface of the nidopallium (NIf), which exhibits both sensory and premotor activity, is necessary for both auditory and premotor processing in its target, HVC. Here we show that bilateral NIf lesions result in long-term loss of HVC auditory activity but do not impair song production. NIf is thus a major source of auditory input to HVC, but an intact NIf is not necessary for motor output in adult zebra finches.
Canopoli, Alessandro; Herbst, Joshua A; Hahnloser, Richard H R
2014-05-14
Many animals exhibit flexible behaviors that they can adjust to increase reward or avoid harm (learning by positive or aversive reinforcement). But what neural mechanisms allow them to restore their original behavior (motor program) after reinforcement is withdrawn? One possibility is that motor restoration relies on brain areas that have a role in memorization but no role in either motor production or in sensory processing relevant for expressing the behavior and its refinement. We investigated the role of a higher auditory brain area in the songbird for modifying and restoring the stereotyped adult song. We exposed zebra finches to aversively reinforcing white noise stimuli contingent on the pitch of one of their stereotyped song syllables. In response, birds significantly changed the pitch of that syllable to avoid the aversive reinforcer. After we withdrew reinforcement, birds recovered their original song within a few days. However, we found that large bilateral lesions in the caudal medial nidopallium (NCM, a high auditory area) impaired recovery of the original pitch even several weeks after withdrawal of the reinforcing stimuli. Because NCM lesions spared both successful noise-avoidance behavior and birds' auditory discrimination ability, our results show that NCM is not needed for directed motor changes or for auditory discriminative processing, but is implied in memorizing or recalling the memory of the recent song target. Copyright © 2014 the authors 0270-6474/14/347018-09$15.00/0.
Vibratory Regime Classification of Infant Phonation
Buder, Eugene H.; Chorna, Lesya B.; Oller, D. Kimbrough; Robinson, Rebecca B.
2008-01-01
Infant phonation is highly variable in many respects, including the basic vibratory patterns by which the vocal tissues create acoustic signals. Previous studies have identified the regular occurrence of non-modal phonation types in normal infant phonation. The glottis is like many oscillating systems that, because of non-linear relationships among the elements, may vibrate in ways representing the deterministic patterns classified theoretically within the mathematical framework of non-linear dynamics. The infant’s pre-verbal vocal explorations present such a variety of phonations that it may be possible to find effectively all the classes of vibration predicted by non-linear dynamic theory. The current report defines acoustic criteria for an important subset of such vibratory regimes, and demonstrates that analysts can be trained to reliably use these criteria for a classification that includes all instances of infant phonation in the recorded corpora. The method is thus internally comprehensive in the sense that all phonations are classified, but it is not exhaustive in the sense that all vocal qualities are thereby represented. Using the methods thus developed, this study also demonstrates that the distributions of these phonation types vary significantly across sessions of recording in the first year of life, suggesting developmental changes. The method of regime classification is thus capable of tracking changes that may be indicative of maturation of the mechanism, the learning of categories of phonatory control, and the possibly varying use of vocalizations across social contexts. PMID:17509829
Artificially lengthened and constricted vocal tract in vocal training methods.
Bele, Irene Velsvik
2005-01-01
It is common practice in vocal training to make use of vocal exercise techniques that involve partial occlusion of the vocal tract. Various techniques are used; some of them form an occlusion within the front part of the oral cavity or at the lips. Another vocal exercise technique involves lengthening the vocal tract; for example, the method of phonation into small tubes. This essay presents some studies made on the effects of various vocal training methods that involve an artificially lengthened and constricted vocal tract. The influence of sufficient acoustic impedance on vocal fold vibration and economical voice production is presented.
ERIC Educational Resources Information Center
Gogate, Lakshmi J.; Bahrick, Lorraine E.
1998-01-01
Investigated 7-month olds' ability to relate vowel sounds with objects when intersensory redundancy was present versus absent. Found that infants detected a mismatch in the vowel-object pairs in the moving-synchronous condition but not in the still or moving-asynchronous condition, demonstrating that temporal synchrony between vocalizations and…
How to Craft Social Media for Graduate Study
ERIC Educational Resources Information Center
Huwe, Terence K.
2011-01-01
Social media now offer a fresh opportunity to enliven the learning process. But to gain traction, they will have to prove their relevance. Librarians are quite vocal in advocating for the preservation of human interaction around digital media; one might say that it is their charge to be certain that the human factors within digital libraries are…
Technology in Our Schools: A Call for a Cost/Benefit Analysis
ERIC Educational Resources Information Center
Wilhelm, Jeffrey D.
2010-01-01
Wilhelm is a vocal advocate for the substantive integration of technology into language arts classes in ways that support the "critical" use of technologies and the learning of new concepts and procedures for reading and composing. He makes the case that no one can be considered fully literate without a familiarity with and appreciation for our…
Acoustic classification of multiple simultaneous bird species: a multi-instance multi-label approach
F. Briggs; B. Lakshminarayanan; L. Neal; X.Z. Fern; R. Raich; S.F. Hadley; A.S. Hadley; M.G. Betts
2012-01-01
Although field-collected recordings typically contain multiple simultaneously vocalizing birds of different species, acoustic species classification in this setting has received little study so far. This work formulates the problem of classifying the set of species present in an audio recording using the multi-instance multi-label (MIML) framework for machine learning...
Dopamine Regulation of Human Speech and Bird Song: A Critical Review
ERIC Educational Resources Information Center
Simonyan, Kristina; Horwitz, Barry; Jarvis, Erich D.
2012-01-01
To understand the neural basis of human speech control, extensive research has been done using a variety of methodologies in a range of experimental models. Nevertheless, several critical questions about learned vocal motor control still remain open. One of them is the mechanism(s) by which neurotransmitters, such as dopamine, modulate speech and…
ERIC Educational Resources Information Center
Perrachione, Tyler K.; Wong, Patrick C. M.
2007-01-01
Brain imaging studies of voice perception often contrast activation from vocal and verbal tasks to identify regions uniquely involved in processing voice. However, such a strategy precludes detection of the functional relationship between speech and voice perception. In a pair of experiments involving identifying voices from native and foreign…
ERIC Educational Resources Information Center
Yeh, Yi-Lien
2018-01-01
The importance of music teachers reflecting on their teaching practise has been discussed for a long time [(Gaunt 2008. "One-to-one Tuition in a Conservatoire: The Perceptions of Instrumental and Vocal Teachers." "Psychology of Music" 36 (2): 215-245; Schön 1987. "Educating the Reflective Practitioner." San…
ERIC Educational Resources Information Center
Brown, Jimmy Lee
2017-01-01
Although the implementation of effective instructional practice is paramount to the academic success of gifted music programs in kindergarten through twelfth grade (K-12), many learners in gifted music programs lack critical thinking skills, self-directed learning, a learner-centered paradigm, and learner-mentor relationship during instruction.…
The Anna's hummingbird chirps with its tail: a new mechanism of sonation in birds
Clark, Christopher James; Feo, Teresa J
2008-01-01
A diverse array of birds apparently make mechanical sounds (called sonations) with their feathers. Few studies have established that these sounds are non-vocal, and the mechanics of how these sounds are produced remains poorly studied. The loud, high-frequency chirp emitted by a male Anna's hummingbird (Calypte anna) during his display dive is a debated example. Production of the sound was originally attributed to the tail, but a more recent study argued that the sound is vocal. Here, we use high-speed video of diving birds, experimental manipulations on wild birds and laboratory experiments on individual feathers to show that the dive sound is made by tail feathers. High-speed video shows that fluttering of the trailing vane of the outermost tail feathers produces the sound. The mechanism is not a whistle, and we propose a flag model to explain the feather's fluttering and accompanying sound. The flag hypothesis predicts that subtle changes in feather shape will tune the frequency of sound produced by feathers. Many kinds of birds are reported to create aerodynamic sounds with their wings or tail, and this model may explain a wide diversity of non-vocal sounds produced by birds. PMID:18230592
Precursors of Dancing and Singing to Music in Three- to Four-Months-Old Infants
Fujii, Shinya; Watanabe, Hama; Oohashi, Hiroki; Hirashima, Masaya; Nozaki, Daichi; Taga, Gentaro
2014-01-01
Dancing and singing to music involve auditory-motor coordination and have been essential to our human culture since ancient times. Although scholars have been trying to understand the evolutionary and developmental origin of music, early human developmental manifestations of auditory-motor interactions in music have not been fully investigated. Here we report limb movements and vocalizations in three- to four-months-old infants while they listened to music and were in silence. In the group analysis, we found no significant increase in the amount of movement or in the relative power spectrum density around the musical tempo in the music condition compared to the silent condition. Intriguingly, however, there were two infants who demonstrated striking increases in the rhythmic movements via kicking or arm-waving around the musical tempo during listening to music. Monte-Carlo statistics with phase-randomized surrogate data revealed that the limb movements of these individuals were significantly synchronized to the musical beat. Moreover, we found a clear increase in the formant variability of vocalizations in the group during music perception. These results suggest that infants at this age are already primed with their bodies to interact with music via limb movements and vocalizations. PMID:24837135
Precursors of dancing and singing to music in three- to four-months-old infants.
Fujii, Shinya; Watanabe, Hama; Oohashi, Hiroki; Hirashima, Masaya; Nozaki, Daichi; Taga, Gentaro
2014-01-01
Dancing and singing to music involve auditory-motor coordination and have been essential to our human culture since ancient times. Although scholars have been trying to understand the evolutionary and developmental origin of music, early human developmental manifestations of auditory-motor interactions in music have not been fully investigated. Here we report limb movements and vocalizations in three- to four-months-old infants while they listened to music and were in silence. In the group analysis, we found no significant increase in the amount of movement or in the relative power spectrum density around the musical tempo in the music condition compared to the silent condition. Intriguingly, however, there were two infants who demonstrated striking increases in the rhythmic movements via kicking or arm-waving around the musical tempo during listening to music. Monte-Carlo statistics with phase-randomized surrogate data revealed that the limb movements of these individuals were significantly synchronized to the musical beat. Moreover, we found a clear increase in the formant variability of vocalizations in the group during music perception. These results suggest that infants at this age are already primed with their bodies to interact with music via limb movements and vocalizations.
A new measure of child vocal reciprocity in children with autism spectrum disorder.
Harbison, Amy L; Woynaroski, Tiffany G; Tapp, Jon; Wade, Joshua W; Warlaumont, Anne S; Yoder, Paul J
2018-06-01
Children's vocal development occurs in the context of reciprocal exchanges with a communication partner who models "speechlike" productions. We propose a new measure of child vocal reciprocity, which we define as the degree to which an adult vocal response increases the probability of an immediately following child vocal response. Vocal reciprocity is likely to be associated with the speechlikeness of vocal communication in young children with autism spectrum disorder (ASD). Two studies were conducted to test the utility of the new measure. The first used simulated vocal samples with randomly sequenced child and adult vocalizations to test the accuracy of the proposed index of child vocal reciprocity. The second was an empirical study of 21 children with ASD who were preverbal or in the early stages of language development. Daylong vocal samples collected in the natural environment were computer analyzed to derive the proposed index of child vocal reciprocity, which was highly stable when derived from two daylong vocal samples and was associated with speechlikeness of vocal communication. This association was significant even when controlling for chance probability of child vocalizations to adult vocal responses, probability of adult vocalizations, or probability of child vocalizations. A valid measure of children's vocal reciprocity might eventually improve our ability to predict which children are on track to develop useful speech and/or are most likely to respond to language intervention. A link to a free, publicly-available software program to derive the new measure of child vocal reciprocity is provided. Autism Res 2018, 11: 903-915. © 2018 International Society for Autism Research, Wiley Periodicals, Inc. Children and adults often engage in back-and-forth vocal exchanges. The extent to which they do so is believed to support children's early speech and language development. Two studies tested a new measure of child vocal reciprocity using computer-generated and real-life vocal samples of young children with autism collected in natural settings. The results provide initial evidence of accuracy, test-retest reliability, and validity of the new measure of child vocal reciprocity. A sound measure of children's vocal reciprocity might improve our ability to predict which children are on track to develop useful speech and/or are most likely to respond to language intervention. A free, publicly-available software program and manuals are provided. © 2018 International Society for Autism Research, Wiley Periodicals, Inc.
The Lombard Effect in Choral Singing
NASA Astrophysics Data System (ADS)
Tonkinson, Steven E.
The Lombard effect is a phenomenon in which a speaker or singer involuntarily raises his or her vocal intensity in the presence of high levels of sound. Many Lombard studies have been published in relation to the speaking voice but very little on the singing voice. A strong reliance upon auditory feedback seems to be a key factor in causing the Lombard effect and research has suggested that singers with more experience and training, especially soloists, do not rely as much on auditory feedback. The purpose of this study was to compare selected vocal intensity response level readings of adult singers, with varying amounts of training, before and after verbal instructions to resist the Lombard effect in choral singing. Choral singers seem especially susceptible because of the nature of the choral environment with its strong masking effect and because of a relative lack of training in voice management. Twenty-seven subjects were asked to sing the national anthem along with a choir heard through headphones. After some brief instructions to resist increasing vocal intensity as the choir increased, each subject sang once again. The performances were recorded and vocal intensity (dB SPL) readings from selected places in the song were obtained from a graphic level recorder chart and analyzed for statistical significance. A 3 x 3 x 2 multiple analysis of variance procedure was performed on the scores, the main factors being experience level, pretest-posttest differences, and places in the song. The questions to be answered by the study were: (1) Do varying levels of experience affect to a significant degree to Lombard effect in choral singing, and (2) Do instructions to maintain a constant level of vocal intensity affect to a significant degree the Lombard effect in singers of varying levels of experience. No significant difference (.05 level) for levels of experience was observed. The effect of the instructions, however, was significant (p <.05) and suggested that choral singers can learn to resist the Lombard effect and consciously regulate their vocal intensity to some extent in the face of masking sound. Choral directors are encouraged to help singers in this regard and inculcate principles of good voice management.
Phylogenetic signal in the acoustic parameters of the advertisement calls of four clades of anurans.
Gingras, Bruno; Mohandesan, Elmira; Boko, Drasko; Fitch, W Tecumseh
2013-07-01
Anuran vocalizations, especially their advertisement calls, are largely species-specific and can be used to identify taxonomic affiliations. Because anurans are not vocal learners, their vocalizations are generally assumed to have a strong genetic component. This suggests that the degree of similarity between advertisement calls may be related to large-scale phylogenetic relationships. To test this hypothesis, advertisement calls from 90 species belonging to four large clades (Bufo, Hylinae, Leptodactylus, and Rana) were analyzed. Phylogenetic distances were estimated based on the DNA sequences of the 12S mitochondrial ribosomal RNA gene, and, for a subset of 49 species, on the rhodopsin gene. Mean values for five acoustic parameters (coefficient of variation of root-mean-square amplitude, dominant frequency, spectral flux, spectral irregularity, and spectral flatness) were computed for each species. We then tested for phylogenetic signal on the body-size-corrected residuals of these five parameters, using three statistical tests (Moran's I, Mantel, and Blomberg's K) and three models of genetic distance (pairwise distances, Abouheif's proximities, and the variance-covariance matrix derived from the phylogenetic tree). A significant phylogenetic signal was detected for most acoustic parameters on the 12S dataset, across statistical tests and genetic distance models, both for the entire sample of 90 species and within clades in several cases. A further analysis on a subset of 49 species using genetic distances derived from rhodopsin and from 12S broadly confirmed the results obtained on the larger sample, indicating that the phylogenetic signals observed in these acoustic parameters can be detected using a variety of genetic distance models derived either from a variable mitochondrial sequence or from a conserved nuclear gene. We found a robust relationship, in a large number of species, between anuran phylogenetic relatedness and acoustic similarity in the advertisement calls in a taxon with no evidence for vocal learning, even after correcting for the effect of body size. This finding, covering a broad sample of species whose vocalizations are fairly diverse, indicates that the intense selection on certain call characteristics observed in many anurans does not eliminate all acoustic indicators of relatedness. Our approach could potentially be applied to other vocal taxa.
NASA Astrophysics Data System (ADS)
Rendall, Drew; Owren, Michael J.; Weerts, Elise; Hienz, Robert D.
2004-01-01
This study quantifies sex differences in the acoustic structure of vowel-like grunt vocalizations in baboons (Papio spp.) and tests the basic perceptual discriminability of these differences to baboon listeners. Acoustic analyses were performed on 1028 grunts recorded from 27 adult baboons (11 males and 16 females) in southern Africa, focusing specifically on the fundamental frequency (F0) and formant frequencies. The mean F0 and the mean frequencies of the first three formants were all significantly lower in males than they were in females, more dramatically so for F0. Experiments using standard psychophysical procedures subsequently tested the discriminability of adult male and adult female grunts. After learning to discriminate the grunt of one male from that of one female, five baboon subjects subsequently generalized this discrimination both to new call tokens from the same individuals and to grunts from novel males and females. These results are discussed in the context of both the possible vocal anatomical basis for sex differences in call structure and the potential perceptual mechanisms involved in their processing by listeners, particularly as these relate to analogous issues in human speech production and perception.
Acoustic Analysis and Electroglottography in Elite Vocal Performers.
Villafuerte-Gonzalez, Rocio; Valadez-Jimenez, Victor M; Sierra-Ramirez, Jose A; Ysunza, Pablo Antonio; Chavarria-Villafuerte, Karen; Hernandez-Lopez, Xochiquetzal
2017-05-01
Acoustic analysis of voice (AAV) and electroglottography (EGG) have been used for assessing vocal quality in patients with voice disorders. The effectiveness of these procedures for detecting mild disturbances in vocal quality in elite vocal performers has been controversial. To compare acoustic parameters obtained by AAV and EGG before and after vocal training to determine the effectiveness of these procedures for detecting vocal improvements in elite vocal performers. Thirty-three elite vocal performers were studied. The study group included 14 males and 19 females, ages 18-40 years, without a history of voice disorders. Acoustic parameters were obtained through AAV and EGG before and after vocal training using the Linklater method. Nonsignificant differences (P > 0.05) were found between values of fundamental frequency (F 0 ), shimmer, and jitter obtained by both procedures before vocal training. Mean F 0 was similar after vocal training. Jitter percentage as measured by AAV showed nonsignificant differences (P > 0.05) before and after vocal training. Shimmer percentage as measured by AAV demonstrated a significant reduction (P < 0.05) after vocal training. As measured by EGG after vocal training, shimmer and jitter were significantly reduced (P < 0.05); open quotient was significantly increased (P < 0.05); and irregularity was significantly reduced (P < 0.05). AAV and EGG were effective for detecting improvements in vocal function after vocal training in male and female elite vocal performers undergoing vocal training. EGG demonstrated better efficacy for detecting improvements and provided additional parameters as compared to AAV. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Learning words and learning sounds: Advances in language development.
Vihman, Marilyn M
2017-02-01
Phonological development is sometimes seen as a process of learning sounds, or forming phonological categories, and then combining sounds to build words, with the evidence taken largely from studies demonstrating 'perceptual narrowing' in infant speech perception over the first year of life. In contrast, studies of early word production have long provided evidence that holistic word learning may precede the formation of phonological categories. In that account, children begin by matching their existing vocal patterns to adult words, with knowledge of the phonological system emerging from the network of related word forms. Here I review evidence from production and then consider how the implicit and explicit learning mechanisms assumed by the complementary memory systems model might be understood as reconciling the two approaches. © 2016 The British Psychological Society.
Increased Fos expression among midbrain dopaminergic cell groups during birdsong tutoring.
Nordeen, E J; Holtzman, D A; Nordeen, K W
2009-08-01
During avian vocal learning, birds memorize conspecific song patterns and then use auditory feedback to match their vocal output to this acquired template. Some models of song learning posit that during tutoring, conspecific visual, social and/or auditory cues activate neuromodulatory systems that encourage acquisition of the tutor's song and attach incentive value to that specific acoustic pattern. This hypothesis predicts that stimuli experienced during social tutoring activate cell populations capable of signaling reward. Using immunocytochemistry for the protein product of the immediate early gene c-Fos, we found that brief exposure of juvenile male zebra finches to a live familiar male tutor increased the density of Fos+ cells within two brain regions implicated in reward processing: the ventral tegmental area (VTA) and substantia nigra pars compacta (SNc). This activation of Fos appears to involve both dopaminergic and non-dopaminergic VTA/SNc neurons. Intriguingly, a familiar tutor was more effective than a novel tutor in stimulating Fos expression within these regions. In the periaqueductal gray, a dopamine-enriched cell population that has been implicated in emotional processing, Fos labeling also was increased after tutoring, with a familiar tutor again being more effective than a novel conspecific. As several neural regions implicated in song acquisition receive strong dopaminergic projections from these midbrain nuclei, their activation in conjunction with hearing the tutor's song could help to establish sensory representations that later guide motor sequence learning.
Nonlinear Bayesian cue integration explains the dynamics of vocal learning
NASA Astrophysics Data System (ADS)
Zhou, Baohua; Sober, Samuel; Nemenman, Ilya
The acoustics of vocal production in songbirds is tightly regulated during both development and adulthood as birds progressively refine their song using sensory feedback to match an acoustic target. Here, we perturb this sensory feedback using headphones to shift the pitch (fundamental frequency) of song. When the pitch is shifted upwards (downwards), birds eventually learn to compensate and sing lower (higher), bringing the experienced pitch closer to the target. Paradoxically, the speed and amplitude of this motor learning decrease with increases in the introduced error size, so that birds respond rapidly to a small sensory perturbation, while seemingly never correcting a much bigger one. Similar results are observed broadly across the animal kingdom, and they do not derive from a limited plasticity of the adult brain since birds can compensate for a large error as long as the error is imposed gradually. We develop a mathematical model based on nonlinear Bayesian integration of two sensory modalities (one perturbed and the other not) that quantitatively explains all of these observations. The model makes predictions about the structure of the probability distribution of the pitches sung by birds during the pitch shift experiments, which we confirm using experimental data. This work was supported in part by James S. McDonnell Foundation Grant # 220020321, NSF Grant # IOS/1208126, NSF Grant # IOS/1456912 and NIH Grants # R01NS084844.
Moving from hand to mouth: echo phonology and the origins of language
Woll, Bencie
2014-01-01
Although the sign languages in use today are full human languages, certain of the features they share with gestures have been suggested to provide information about possible origins of human language. These features include sharing common articulators with gestures, and exhibiting substantial iconicity in comparison to spoken languages. If human proto-language was gestural, the question remains of how a highly iconic manual communication system might have been transformed into a primarily vocal communication system in which the links between symbol and referent are for the most part arbitrary. The hypothesis presented here focuses on a class of signs which exhibit: “echo phonology,” a repertoire of mouth actions which are characterized by “echoing” on the mouth certain of the articulatory actions of the hands. The basic features of echo phonology are introduced, and discussed in relation to various types of data. Echo phonology provides naturalistic examples of a possible mechanism accounting for part of the evolution of language, with evidence both of the transfer of manual actions to oral ones and the conversion of units of an iconic manual communication system into a largely arbitrary vocal communication system. PMID:25071636
Neural Correlates of the Lombard Effect in Primate Auditory Cortex
Eliades, Steven J.
2012-01-01
Speaking is a sensory-motor process that involves constant self-monitoring to ensure accurate vocal production. Self-monitoring of vocal feedback allows rapid adjustment to correct perceived differences between intended and produced vocalizations. One important behavior in vocal feedback control is a compensatory increase in vocal intensity in response to noise masking during vocal production, commonly referred to as the Lombard effect. This behavior requires mechanisms for continuously monitoring auditory feedback during speaking. However, the underlying neural mechanisms are poorly understood. Here we show that when marmoset monkeys vocalize in the presence of masking noise that disrupts vocal feedback, the compensatory increase in vocal intensity is accompanied by a shift in auditory cortex activity toward neural response patterns seen during vocalizations under normal feedback condition. Furthermore, we show that neural activity in auditory cortex during a vocalization phrase predicts vocal intensity compensation in subsequent phrases. These observations demonstrate that the auditory cortex participates in self-monitoring during the Lombard effect, and may play a role in the compensation of noise masking during feedback-mediated vocal control. PMID:22855821
The perceptual features of vocal fatigue as self-reported by a group of actors and singers.
Kitch, J A; Oates, J
1994-09-01
Performers (10 actors/10 singers) rated via a self-report questionnaire the severity of their voice-related changes when vocally fatigued. Similar frequency patterns and perceptual features of vocal fatigue were found across subjects. Actors rated "power" aspects (e.g., voice projection) and singers rated vocal dynamic aspects (e.g., pitch range) of their voices as most affected when vocally fatigued. Vocal fatigue was evidenced by changes in kinesthetic/proprioceptive sensations and vocal dynamics. The causes and context of vocal fatigue were vocal misuse, being "run down," high performance demands, and using high pitch/volume levels. Further research is needed to delineate the perceptual features of "normal" levels of vocal fatigue and its possible causes.
Shepard, Kathryn N.; Chong, Kelly K.
2016-01-01
Tonotopic map plasticity in the adult auditory cortex (AC) is a well established and oft-cited measure of auditory associative learning in classical conditioning paradigms. However, its necessity as an enduring memory trace has been debated, especially given a recent finding that the areal expansion of core AC tuned to a newly relevant frequency range may arise only transiently to support auditory learning. This has been reinforced by an ethological paradigm showing that map expansion is not observed for ultrasonic vocalizations (USVs) or for ultrasound frequencies in postweaning dams for whom USVs emitted by pups acquire behavioral relevance. However, whether transient expansion occurs during maternal experience is not known, and could help to reveal the generality of cortical map expansion as a correlate for auditory learning. We thus mapped the auditory cortices of maternal mice at postnatal time points surrounding the peak in pup USV emission, but found no evidence of frequency map expansion for the behaviorally relevant high ultrasound range in AC. Instead, regions tuned to low frequencies outside of the ultrasound range show progressively greater suppression of activity in response to the playback of ultrasounds or pup USVs for maternally experienced animals assessed at their pups’ postnatal day 9 (P9) to P10, or postweaning. This provides new evidence for a lateral-band suppression mechanism elicited by behaviorally meaningful USVs, likely enhancing their population-level signal-to-noise ratio. These results demonstrate that tonotopic map enlargement has limits as a construct for conceptualizing how experience leaves neural memory traces within sensory cortex in the context of ethological auditory learning. PMID:27957529
Dai, Yun; Zhao, Yuanzi; Tomi, Masatoshi; Shin, Bo-Chul; Thamotharan, Shanthie; Mazarati, Andrey; Sankar, Raman; Wang, Elizabeth A; Cepeda, Carlos; Levine, Michael S; Zhang, Jingjing; Frew, Andrew; Alger, Jeffry R; Clark, Peter M; Sondhi, Monica; Kositamongkol, Sudatip; Leibovitch, Leah; Devaskar, Sherin U
2017-04-01
We tested the hypothesis that exposure of glut3+/- mice to a ketogenic diet ameliorates autism-like features, which include aberrant behavior and electrographic seizures. We first investigated the life course sex-specific changes in basal plasma-cerebrospinal fluid (CSF)-brain metabolic profile, brain glucose transport/uptake, glucose and monocarboxylate transporter proteins, and adenosine triphosphate (ATP) in the presence or absence of systemic insulin administration. Glut3+/- male but not female mice (5 months of age) displayed reduced CSF glucose/lactate concentrations with no change in brain Glut1, Mct2, glucose uptake or ATP. Exogenous insulin-induced hypoglycemia increased brain glucose uptake in glut3+/- males alone. Higher plasma-CSF ketones (β-hydroxybutyrate) and lower brain Glut3 in females vs males proved protective in the former while enhancing vulnerability in the latter. As a consequence, increased synaptic proteins (neuroligin4 and SAPAP1) with spontaneous excitatory postsynaptic activity subsequently reduced hippocampal glucose content and increased brain amyloid β1-40 deposition in an age-dependent manner in glut3+/- males but not females (4 to 24 months of age). We then explored the protective effect of a ketogenic diet on ultrasonic vocalization, sociability, spatial learning and memory, and electroencephalogram seizures in male mice (7 days to 6 to 8 months of age) alone. A ketogenic diet partially restored sociability without affecting perturbed vocalization, spatial learning and memory, and reduced seizure events. We conclude that (1) sex-specific and age-dependent perturbations underlie the phenotype of glut3+/- mice, and (2) a ketogenic diet ameliorates seizures caused by increased cortical excitation and improves sociability, but fails to rescue vocalization and cognitive deficits in glut3+/- male mice. Copyright © 2017 Endocrine Society.
Dai, Yun; Zhao, Yuanzi; Tomi, Masatoshi; Shin, Bo-Chul; Thamotharan, Shanthie; Mazarati, Andrey; Sankar, Raman; Wang, Elizabeth A.; Cepeda, Carlos; Levine, Michael S.; Zhang, Jingjing; Frew, Andrew; Alger, Jeffry R.; Clark, Peter M.; Sondhi, Monica; Kositamongkol, Sudatip; Leibovitch, Leah
2017-01-01
We tested the hypothesis that exposure of glut3+/− mice to a ketogenic diet ameliorates autism-like features, which include aberrant behavior and electrographic seizures. We first investigated the life course sex-specific changes in basal plasma–cerebrospinal fluid (CSF)–brain metabolic profile, brain glucose transport/uptake, glucose and monocarboxylate transporter proteins, and adenosine triphosphate (ATP) in the presence or absence of systemic insulin administration. Glut3+/− male but not female mice (5 months of age) displayed reduced CSF glucose/lactate concentrations with no change in brain Glut1, Mct2, glucose uptake or ATP. Exogenous insulin-induced hypoglycemia increased brain glucose uptake in glut3+/− males alone. Higher plasma-CSF ketones (β-hydroxybutyrate) and lower brain Glut3 in females vs males proved protective in the former while enhancing vulnerability in the latter. As a consequence, increased synaptic proteins (neuroligin4 and SAPAP1) with spontaneous excitatory postsynaptic activity subsequently reduced hippocampal glucose content and increased brain amyloid β1-40 deposition in an age-dependent manner in glut3+/− males but not females (4 to 24 months of age). We then explored the protective effect of a ketogenic diet on ultrasonic vocalization, sociability, spatial learning and memory, and electroencephalogram seizures in male mice (7 days to 6 to 8 months of age) alone. A ketogenic diet partially restored sociability without affecting perturbed vocalization, spatial learning and memory, and reduced seizure events. We conclude that (1) sex-specific and age-dependent perturbations underlie the phenotype of glut3+/− mice, and (2) a ketogenic diet ameliorates seizures caused by increased cortical excitation and improves sociability, but fails to rescue vocalization and cognitive deficits in glut3+/− male mice. PMID:28324109
He hears, she hears: are there sex differences in auditory processing?
Yoder, Kathleen M; Phan, Mimi L; Lu, Kai; Vicario, David S
2015-03-01
Songbirds learn individually unique songs through vocal imitation and use them in courtship and territorial displays. Previous work has identified a forebrain auditory area, the caudomedial nidopallium (NCM), that appears specialized for discriminating and remembering conspecific vocalizations. In zebra finches (ZFs), only males produce learned vocalizations, but both sexes process these and other signals. This study assessed sex differences in auditory processing by recording extracellular multiunit activity at multiple sites within NCM. Juvenile female ZFs (n = 46) were reared in individual isolation and artificially tutored with song. In adulthood, songs were played back to assess auditory responses, stimulus-specific adaptation, neural bias for conspecific song, and memory for the tutor's song, as well as recently heard songs. In a subset of females (n = 36), estradiol (E2) levels were manipulated to test the contribution of E2, known to be synthesized in the brain, to auditory responses. Untreated females (n = 10) showed significant differences in response magnitude and stimulus-specific adaptation compared to males reared in the same paradigm (n = 9). In hormone-manipulated females, E2 augmentation facilitated the memory for recently heard songs in adulthood, but neither E2 augmentation (n = 15) nor E2 synthesis blockade (n = 9) affected tutor song memory or the neural bias for conspecific song. The results demonstrate subtle sex differences in processing communication signals, and show that E2 levels in female songbirds can affect the memory for songs of potential suitors, thus contributing to the process of mate selection. The results also have potential relevance to clinical interventions that manipulate E2 in human patients. © 2014 Wiley Periodicals, Inc.
PVCM, PVCD, EPL, and irritable larynx syndrome: what are we talking about and how do we treat it?
Andrianopoulos, M V; Gallivan, G J; Gallivan, K H
2000-12-01
Paroxysmal vocal cord movement/motion (PVCM), paroxysmal vocal cord dysfunction (PVCD), episodic paroxysmal laryngospasm (EPL), and irritable larynx syndrome (ILS) are terms used to describe laryngeal dysfunction masquerading as asthma, upper airway obstruction, or functional and organic voice disorders. The differential diagnosis of PVCM, PVCD, EPL, and ILS is critical to successful medical and behavioral management of the patient. During the past 10 years, 27 subjects, ages 15-79 years, were identified to have paroxysms of inspiratory stridor, acute respiratory distress, associated aphonia and dysphonia, resulting in misdiagnosis and unnecessary emergency treatments, including endotracheal intubation, cardiopulmonary resuscitation, massive pharmacotherapy, or tracheostomy. A multifactorial management program is proposed utilizing principles of motor learning, neurolinguistic programming model, respiratory and phonatory synchronization, relaxation techniques, concurrent monitoring of behavioral adjustments, and formal psychological counseling.
Reinforcement of Infant Vocalizations through Contingent Vocal Imitation
ERIC Educational Resources Information Center
Pelaez, Martha; Virues-Ortega, Javier; Gewirtz, Jacob L.
2011-01-01
Maternal vocal imitation of infant vocalizations is highly prevalent during face-to-face interactions of infants and their caregivers. Although maternal vocal imitation has been associated with later verbal development, its potentially reinforcing effect on infant vocalizations has not been explored experimentally. This study examined the…
ERIC Educational Resources Information Center
Van Stan, Jarrad H.; Mehta, Daryush D.; Sternad, Dagmar; Petit, Robert; Hillman, Robert E.
2017-01-01
Purpose: Ambulatory voice biofeedback has the potential to significantly improve voice therapy effectiveness by targeting carryover of desired behaviors outside the therapy session (i.e., retention). This study applies motor learning concepts (reduced frequency and delayed, summary feedback) that demonstrate increased retention to ambulatory voice…
Another Perspective: Every Child a Singer
ERIC Educational Resources Information Center
Phillips, Kenneth H.
2014-01-01
This article begins with a brief history of children's singing instruction in schools. When Kenneth Phillips began researching children's singing, he was surprised to learn that a strong program of child vocal pedagogy was in place in America's schools for about a hundred years--the 1830s to the 1930s. By the 1950s however, all that…
Dr. Mashup or, Why Educators Should Learn to Stop Worrying and Love the Remix
ERIC Educational Resources Information Center
Lamb, Brian
2007-01-01
A music track that lays a vocal from Madonna over guitars from the Sex Pistols. A classroom portal that presents automatically updated syndicated resources from the campus library, news sources, student events, weblogs, and podcasts and that was built quickly using free tools. A Web site that takes crime data from the Chicago Police and applies…
ERIC Educational Resources Information Center
Simones, Lilian Lima
2017-01-01
Music performance in the higher educational context is shaped by a reciprocal chain of interactions between students, part-time tutors and full-time teaching staff, each with specific expectations about the teaching and learning process. Such expectations can provide valuable insights not only for designing and implementing meaningful educational…
ERIC Educational Resources Information Center
Burwell, Kim
2017-01-01
While studio-based instrumental and vocal learning is widely regarded as both important and effective in higher education music, research to date has offered little concrete information about studio practices that students have regarded as ineffective. Two recent case studies investigated what appear to be exceptional instances in which students…
ERIC Educational Resources Information Center
Wallez, Catherine; Schaeffer, Jennifer; Meguerditchian, Adrien; Vauclair, Jacques; Schapiro, Steven J.; Hopkins, William D.
2012-01-01
Studies involving oro-facial asymmetries in nonhuman primates have largely demonstrated a right hemispheric dominance for communicative signals and conveyance of emotional information. A recent study on chimpanzee reported the first evidence of significant left-hemispheric dominance when using attention-getting sounds and rightward bias for…
Pinaud, Raphael; Terleph, Thomas A.; Tremere, Liisa A.; Phan, Mimi L.; Dagostin, André A.; Leão, Ricardo M.; Mello, Claudio V.; Vicario, David S.
2008-01-01
The role of GABA in the central processing of complex auditory signals is not fully understood. We have studied the involvement of GABAA-mediated inhibition in the processing of birdsong, a learned vocal communication signal requiring intact hearing for its development and maintenance. We focused on caudomedial nidopallium (NCM), an area analogous to parts of the mammalian auditory cortex with selective responses to birdsong. We present evidence that GABAA-mediated inhibition plays a pronounced role in NCM's auditory processing of birdsong. Using immunocytochemistry, we show that approximately half of NCM's neurons are GABAergic. Whole cell patch-clamp recordings in a slice preparation demonstrate that, at rest, spontaneously active GABAergic synapses inhibit excitatory inputs onto NCM neurons via GABAA receptors. Multi-electrode electrophysiological recordings in awake birds show that local blockade of GABAA-mediated inhibition in NCM markedly affects the temporal pattern of song-evoked responses in NCM without modifications in frequency tuning. Surprisingly, this blockade increases the phasic and largely suppresses the tonic response component, reflecting dynamic relationships of inhibitory networks that could include disinhibition. Thus processing of learned natural communication sounds in songbirds, and possibly other vocal learners, may depend on complex interactions of inhibitory networks. PMID:18480371
Matievics, Vera; Bach, Adam; Sztano, Balazs; Bere, Zsofia; Tobias, Zoltan; Castellanos, Paul F; Mueller, Andreas H; Rovo, Laszló
2017-10-01
In unilateral vocal cord paralysis (UVCP), hoarseness is usually the leading symptom; however, the diminished airway might lead to breathing problems as well, especially with exertion. The application of the classic resection glottis enlarging or medialization procedures might shift the breathing and/or the voice to a worse condition. The non-destructive endoscopic arytenoid abduction lateropexy (EAAL) might be a solution for this problem. The aim of our study was to analyze the phonatory and respiratory outcomes of this treatment concept. The first year phoniatric [Jitter, Shimmer, harmonics-to-noise ratio (HNR), maximum phonation time (MPT), fundamental frequency (F 0 ), Voice Handicap Index (VHI), Dysphonia Severity Index (DSI), Global-Roughness-Breathiness scale (GRB)], peak inspiratory flow (PIF), and quality of life (QoL) were evaluated in ten UVCP patients treated by EAAL for dyspnea generally presented on exertion. PIF, Jitter, QoL, GRB, and VHI significantly improved. DSI, HNR, and MPT got non-significantly better. F 0 slightly increased in all patients, a mild deterioration of shimmer was observed. These results prove that improving respiratory function is not necessarily associated with a deterioration in voice quality. The EAAL provides a significant improvement in breathing and the vibratory parameters of the postoperative, more tensed and straightened vocal cords proved to be more advantageous than the original (para) median 'loose' position. The over-adduction of the contralateral side more or less compensates for the disadvantageous, more lateral position of the operated side. EAAL might be an alternative treatment for unilateral vocal cord paralysis associated with breathing problems.
Detection and Classification of Whale Acoustic Signals
NASA Astrophysics Data System (ADS)
Xian, Yin
This dissertation focuses on two vital challenges in relation to whale acoustic signals: detection and classification. In detection, we evaluated the influence of the uncertain ocean environment on the spectrogram-based detector, and derived the likelihood ratio of the proposed Short Time Fourier Transform detector. Experimental results showed that the proposed detector outperforms detectors based on the spectrogram. The proposed detector is more sensitive to environmental changes because it includes phase information. In classification, our focus is on finding a robust and sparse representation of whale vocalizations. Because whale vocalizations can be modeled as polynomial phase signals, we can represent the whale calls by their polynomial phase coefficients. In this dissertation, we used the Weyl transform to capture chirp rate information, and used a two dimensional feature set to represent whale vocalizations globally. Experimental results showed that our Weyl feature set outperforms chirplet coefficients and MFCC (Mel Frequency Cepstral Coefficients) when applied to our collected data. Since whale vocalizations can be represented by polynomial phase coefficients, it is plausible that the signals lie on a manifold parameterized by these coefficients. We also studied the intrinsic structure of high dimensional whale data by exploiting its geometry. Experimental results showed that nonlinear mappings such as Laplacian Eigenmap and ISOMAP outperform linear mappings such as PCA and MDS, suggesting that the whale acoustic data is nonlinear. We also explored deep learning algorithms on whale acoustic data. We built each layer as convolutions with either a PCA filter bank (PCANet) or a DCT filter bank (DCTNet). With the DCT filter bank, each layer has different a time-frequency scale representation, and from this, one can extract different physical information. Experimental results showed that our PCANet and DCTNet achieve high classification rate on the whale vocalization data set. The word error rate of the DCTNet feature is similar to the MFSC in speech recognition tasks, suggesting that the convolutional network is able to reveal acoustic content of speech signals.
Histopathologic study of human vocal fold mucosa unphonated over a decade.
Sato, Kiminori; Umeno, Hirohito; Ono, Takeharu; Nakashima, Tadashi
2011-12-01
Mechanotransduction caused by vocal fold vibration could possibly be an important factor in the maintenance of extracellular matrices and layered structure of the human adult vocal fold mucosa as a vibrating tissue after the layered structure has been completed. Vocal fold stellate cells (VFSCs) in the human maculae flavae of the vocal fold mucosa are inferred to be involved in the metabolism of extracellular matrices of the vocal fold mucosa. Maculae flavae are also considered to be an important structure in the growth and development of the human vocal fold mucosa. Tension caused by phonation (vocal fold vibration) is hypothesized to stimulate the VFSCs to accelerate production of extracellular matrices. A human adult vocal fold mucosa unphonated over a decade was investigated histopathologically. Vocal fold mucosa unphonated for 11 years and 2 months of a 64-year-old male with cerebral hemorrhage was investigated by light and electron microscopy. The vocal fold mucosae (including maculae flavae) were atrophic. The vocal fold mucosa did not have a vocal ligament, Reinke's space or a layered structure. The lamina propria appeared as a uniform structure. Morphologically, the VFSCs synthesized fewer extracellular matrices, such as fibrous protein and glycosaminoglycan. Consequently, VFSCs appeared to decrease their level of activity.
Coos, booms, and hoots: The evolution of closed-mouth vocal behavior in birds.
Riede, Tobias; Eliason, Chad M; Miller, Edward H; Goller, Franz; Clarke, Julia A
2016-08-01
Most birds vocalize with an open beak, but vocalization with a closed beak into an inflating cavity occurs in territorial or courtship displays in disparate species throughout birds. Closed-mouth vocalizations generate resonance conditions that favor low-frequency sounds. By contrast, open-mouth vocalizations cover a wider frequency range. Here we describe closed-mouth vocalizations of birds from functional and morphological perspectives and assess the distribution of closed-mouth vocalizations in birds and related outgroups. Ancestral-state optimizations of body size and vocal behavior indicate that closed-mouth vocalizations are unlikely to be ancestral in birds and have evolved independently at least 16 times within Aves, predominantly in large-bodied lineages. Closed-mouth vocalizations are rare in the small-bodied passerines. In light of these results and body size trends in nonavian dinosaurs, we suggest that the capacity for closed-mouth vocalization was present in at least some extinct nonavian dinosaurs. As in birds, this behavior may have been limited to sexually selected vocal displays, and hence would have co-occurred with open-mouthed vocalizations. © 2016 The Author(s). Evolution © 2016 The Society for the Study of Evolution.
The Vocal Repertoire of Adult and Neonate Giant Otters (Pteronura brasiliensis)
Mumm, Christina A. S.; Knörnschild, Mirjam
2014-01-01
Animals use vocalizations to exchange information about external events, their own physical or motivational state, or about individuality and social affiliation. Infant babbling can enhance the development of the full adult vocal repertoire by providing ample opportunity for practice. Giant otters are very social and frequently vocalizing animals. They live in highly cohesive groups, generally including a reproductive pair and their offspring born in different years. This basic social structure may vary in the degree of relatedness of the group members. Individuals engage in shared group activities and different social roles and thus, the social organization of giant otters provides a basis for complex and long-term individual relationships. We recorded and analysed the vocalizations of adult and neonate giant otters from wild and captive groups. We classified the adult vocalizations according to their acoustic structure, and described their main behavioural context. Additionally, we present the first description of vocalizations uttered in babbling bouts of new born giant otters. We expected to find 1) a sophisticated vocal repertoire that would reflect the species’ complex social organisation, 2) that giant otter vocalizations have a clear relationship between signal structure and function, and 3) that the vocal repertoire of new born giant otters would comprise age-specific vocalizations as well as precursors of the adult repertoire. We found a vocal repertoire with 22 distinct vocalization types produced by adults and 11 vocalization types within the babbling bouts of the neonates. A comparison within the otter subfamily suggests a relation between vocal and social complexity, with the giant otters being the socially and vocally most complex species. PMID:25391142
Doellinger, Michael; Lohscheller, Joerg; McWhorter, Andrew; Kunduk, Melda
2009-03-01
We investigate the potential of high-speed digital imaging technique (HSI) and the phonovibrogram (PVG) analysis in normal vocal fold dynamics by studying the effects of continuous voice use (vocal loading) during the workday. One healthy subject was recorded at sustained phonation 13 times within 2 consecutive days in the morning before and in the afternoon after vocal loading, respectively. Vocal fold dynamics were extracted and visualized by PVGs. The characteristic PVG patterns were extracted representing vocal fold vibration types. The parameter values were then analyzed by statistics regarding vocal load, left-right PVG asymmetries, anterior-posterior PVG asymmetries, and opening-closing differences. For the first time, the direct impact of vocal load could be determined by analyzing vocal fold dynamics. For same vocal loading conditions, equal dynamical behavior of the vocal folds were confirmed. Comparison of recordings performed in the morning with the recordings after work revealed significant changes in vibration behavior, indicating impact of occurring vocal load. Left-right asymmetries in vocal fold dynamics were found confirming earlier assumptions. Different dynamics between opening and closing procedure as well as for anterior and posterior parts were found. Constant voice usage stresses the vocal folds even in healthy subjects and can be detected by applying the PVG technique. Furthermore, left-right PVG asymmetries do occur in healthy voice to a certain extent. HSI in combination with PVG analysis seems to be a promising tool for investigation of vocal fold fatigue and pathologies resulting in small forms of dynamical changes.
Garcia, Elisângela Zacanti; Yamashita, Hélio Kiitiro; Garcia, Davi Sousa; Padovani, Marina Martins Pereira; Azevedo, Renata Rangel; Chiari, Brasília Maria
2016-01-01
Cone beam computed tomography (CBCT), which represents an alternative to traditional computed tomography and magnetic resonance imaging, may be a useful instrument to study vocal tract physiology related to vocal exercises. This study aims to evaluate the applicability of CBCT to the assessment of variations in the vocal tract of healthy individuals before and after vocal exercises. Voice recordings and CBCT images before and after vocal exercises performed by 3 speech-language pathologists without vocal complaints were collected and compared. Each participant performed 1 type of exercise, i.e., Finnish resonance tube technique, prolonged consonant "b" technique, or chewing technique. The analysis consisted of an acoustic analysis and tomographic imaging. Modifications of the vocal tract settings following vocal exercises were properly detected by CBCT, and changes in the acoustic parameters were, for the most part, compatible with the variations detected in image measurements. CBCT was shown to be capable of properly assessing the changes in vocal tract settings promoted by vocal exercises. © 2017 S. Karger AG, Basel.
Masapollo, Matthew; Polka, Linda; Ménard, Lucie
2016-03-01
To learn to produce speech, infants must effectively monitor and assess their own speech output. Yet very little is known about how infants perceive speech produced by an infant, which has higher voice pitch and formant frequencies compared to adult or child speech. Here, we tested whether pre-babbling infants (at 4-6 months) prefer listening to vowel sounds with infant vocal properties over vowel sounds with adult vocal properties. A listening preference favoring infant vowels may derive from their higher voice pitch, which has been shown to attract infant attention in infant-directed speech (IDS). In addition, infants' nascent articulatory abilities may induce a bias favoring infant speech given that 4- to 6-month-olds are beginning to produce vowel sounds. We created infant and adult /i/ ('ee') vowels using a production-based synthesizer that simulates the act of speaking in talkers at different ages and then tested infants across four experiments using a sequential preferential listening task. The findings provide the first evidence that infants preferentially attend to vowel sounds with infant voice pitch and/or formants over vowel sounds with no infant-like vocal properties, supporting the view that infants' production abilities influence how they process infant speech. The findings with respect to voice pitch also reveal parallels between IDS and infant speech, raising new questions about the role of this speech register in infant development. Research exploring the underpinnings and impact of this perceptual bias can expand our understanding of infant language development. © 2015 John Wiley & Sons Ltd.
Dependence of phonation threshold pressure on vocal tract acoustics and vocal fold tissue mechanics.
Chan, Roger W; Titze, Ingo R
2006-04-01
Analytical and computer simulation studies have shown that the acoustic impedance of the vocal tract as well as the viscoelastic properties of vocal fold tissues are critical for determining the dynamics and the energy transfer mechanism of vocal fold oscillation. In the present study, a linear, small-amplitude oscillation theory was revised by taking into account the propagation of a mucosal wave and the inertive reactance (inertance) of the supraglottal vocal tract as the major energy transfer mechanisms for flow-induced self-oscillation of the vocal fold. Specifically, analytical results predicted that phonation threshold pressure (Pth) increases with the viscous shear properties of the vocal fold, but decreases with vocal tract inertance. This theory was empirically tested using a physical model of the larynx, where biological materials (fat, hyaluronic acid, and fibronectin) were implanted into the vocal fold cover to investigate the effect of vocal fold tissue viscoelasticity on Pth. A uniform-tube supraglottal vocal tract was also introduced to examine the effect of vocal tract inertance on Pth. Results showed that Pth decreased with the inertive impedance of the vocal tract and increased with the viscous shear modulus (G") or dynamic viscosity (eta') of the vocal fold cover, consistent with theoretical predictions. These findings supported the potential biomechanical benefits of hyaluronic acid as a surgical bioimplant for repairing voice disorders involving the superficial layer of the lamina propria, such as scarring, sulcus vocalis, atrophy, and Reinke's edema.
Van Lierde, Kristiane M; D'haeseleer, Evelien; Wuyts, Floris L; De Ley, Sophia; Geldof, Ruben; De Vuyst, Julie; Sofie, Claeys
2010-09-01
The purpose of the present cross-sectional study was to determine the objective vocal quality and the vocal characteristics (vocal risk factors, vocal and corporal complaints) in 197 female students in speech-language pathology during the 4 years of study. The objective vocal quality was measured by means of the Dysphonia Severity Index (DSI). Perceptual voice assessment, the Voice Handicap Index (VHI), questionnaires addressing vocal risks, and vocal and corporal complaints during and/or after voice usage were performed. Speech-language pathology (SLP) students have a borderline vocal quality corresponding to a DSI% of 68. The analysis of variance revealed no significant change of the objective vocal quality between the first bachelor year and the master year. No psychosocial handicapping effect of the voice was observed by means of the VHI total, though there was an effect at the functional VHI level in addition to some vocal complaints. Ninety-three percent of the student SLPs reported the presence of corporal pain during and/or after speaking. In particular, sore throat and headache were mentioned as the prevalent corporal pain symptoms. A longitudinal study of the objective vocal quality of the same subjects during their career as an SLP might provide new insights. 2010 The Voice Foundation. Published by Mosby, Inc. All rights reserved.
University Vocal Training and Vocal Health of Music Educators and Music Therapists
ERIC Educational Resources Information Center
Baker, Vicki D.; Cohen, Nicki
2017-01-01
The purpose of this study was to describe the university vocal training and vocal health of music educators and music therapists. The participants (N = 426), music educators (n = 351) and music therapists (n = 75), completed a survey addressing demographics, vocal training, voice usage, and vocal health. Both groups reported singing at least 50%…
Monkey vocal tracts are speech-ready.
Fitch, W Tecumseh; de Boer, Bart; Mathur, Neil; Ghazanfar, Asif A
2016-12-01
For four decades, the inability of nonhuman primates to produce human speech sounds has been claimed to stem from limitations in their vocal tract anatomy, a conclusion based on plaster casts made from the vocal tract of a monkey cadaver. We used x-ray videos to quantify vocal tract dynamics in living macaques during vocalization, facial displays, and feeding. We demonstrate that the macaque vocal tract could easily produce an adequate range of speech sounds to support spoken language, showing that previous techniques based on postmortem samples drastically underestimated primate vocal capabilities. Our findings imply that the evolution of human speech capabilities required neural changes rather than modifications of vocal anatomy. Macaques have a speech-ready vocal tract but lack a speech-ready brain to control it.
Vocal Dose Measures: Quantifying Accumulated Vibration Exposure in Vocal Fold Tissues
Titze, Ingo R.; Švec, Jan G.; Popolo, Peter S.
2011-01-01
To measure the exposure to self-induced tissue vibration in speech, three vocal doses were defined and described: distance dose, which accumulates the distance that tissue particles of the vocal folds travel in an oscillatory trajectory; energy dissipation dose, which accumulates the total amount of heat dissipated over a unit volume of vocal fold tissues; and time dose, which accumulates the total phonation time. These doses were compared to a previously used vocal dose measure, the vocal loading index, which accumulates the number of vibration cycles of the vocal folds. Empirical rules for viscosity and vocal fold deformation were used to calculate all the doses from the fundamental frequency (F0) and sound pressure level (SPL) values of speech. Six participants were asked to read in normal, monotone, and exaggerated speech and the doses associated with these vocalizations were calculated. The results showed that large F0 and SPL variations in speech affected the dose measures, suggesting that accumulation of phonation time alone is insufficient. The vibration exposure of the vocal folds in normal speech was related to the industrial limits for hand-transmitted vibration, in which the safe distance dose was derived to be about 500 m. This limit was found rather low for vocalization; it was related to a comparable time dose of about 17 min of continuous vocalization, or about 35 min of continuous reading with normal breathing and unvoiced segments. The voicing pauses in normal speech and dialogue effectively prolong the safe time dose. The derived safety limits for vocalization will likely require refinement based on a more detailed knowledge of the differences in hand and vocal fold tissue morphology and their response to vibrational stress, and on the effect of recovery of the vocal fold tissue during voicing pauses. PMID:12959470
Vocal dose measures: quantifying accumulated vibration exposure in vocal fold tissues.
Titze, Ingo R; Svec, Jan G; Popolo, Peter S
2003-08-01
To measure the exposure to self-induced tissue vibration in speech, three vocal doses were defined and described: distance dose, which accumulates the distance that tissue particles of the vocal folds travel in an oscillatory trajectory; energy dissipation dose, which accumulates the total amount of heat dissipated over a unit volume of vocal fold tissues; and time dose, which accumulates the total phonation time. These doses were compared to a previously used vocal dose measure, the vocal loading index, which accumulates the number of vibration cycles of the vocal folds. Empirical rules for viscosity and vocal fold deformation were used to calculate all the doses from the fundamental frequency (F0) and sound pressure level (SPL) values of speech. Six participants were asked to read in normal, monotone, and exaggerated speech and the doses associated with these vocalizations were calculated. The results showed that large F0 and SPL variations in speech affected the dose measures, suggesting that accumulation of phonation time alone is insufficient. The vibration exposure of the vocal folds in normal speech was related to the industrial limits for hand-transmitted vibration, in which the safe distance dose was derived to be about 500 m. This limit was found rather low for vocalization; it was related to a comparable time dose of about 17 min of continuous vocalization, or about 35 min of continuous reading with normal breathing and unvoiced segments. The voicing pauses in normal speech and dialogue effectively prolong the safe time dose. The derived safety limits for vocalization will likely require refinement based on a more detailed knowledge of the differences in hand and vocal fold tissue morphology and their response to vibrational stress, and on the effect of recovery of the vocal fold tissue during voicing pauses.
Sielska-Badurek, Ewelina; Osuch-Wójcikiewicz, Ewa; Sobol, Maria; Kazanecka, Ewa; Niemczyk, Kazimierz
2017-01-01
This study investigated vocal function knowledge and vocal tract sensorimotor self-awareness and the impact of functional voice rehabilitation on vocal function knowledge and self-awareness. This is a prospective, randomized study. Twenty singers (study group [SG]) completed a questionnaire before and after functional voice rehabilitation. Twenty additional singers, representing the control group, also completed the questionnaire without functional voice rehabilitation at a 3-month interval. The questionnaire consisted of three parts. The first part evaluated the singers' attitude to the anatomical and physiological knowledge of the vocal tract and their self-esteem of the knowledge level. The second part assessed the theoretical knowledge of the singers' vocal tract physiology. The third part of the questionnaire assessed singers' sensorimotor self-awareness of the vocal tract. The results showed that most singers indicated that knowledge of the vocal tract's anatomy and physiology is useful (59% SG, 67% control group). However, 75% of all participants defined their knowledge of the vocal tract's anatomy and physiology as weak or inadequate. In the SG, vocal function knowledge at the first assessment was 45%. After rehabilitation, the level increased to 67.7%. Vocal tract sensorimotor self-awareness initially was 38.9% in SG but rose to 66.7%. Findings of the study suggest that classical singers lack knowledge about the physiology of the vocal mechanism, especially the breathing patterns. In addition, they have low sensorimotor self-awareness of their vocal tract. The results suggest that singers would benefit from receiving services from phoniatrists and speech-language pathologists during their voice training. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Kessler, Sharon E; Radespiel, Ute; Hasiniaina, Alida I F; Leliveld, Lisette M C; Nash, Leanne T; Zimmermann, Elke
2014-02-20
Maternal kin selection is a driving force in the evolution of mammalian social complexity and it requires that kin are distinctive from nonkin. The transition from the ancestral state of asociality to the derived state of complex social groups is thought to have occurred via solitary foraging, in which individuals forage alone, but, unlike the asocial ancestors, maintain dispersed social networks via scent-marks and vocalizations. We hypothesize that matrilineal signatures in vocalizations were an important part of these networks. We used the solitary foraging gray mouse lemur (Microcebus murinus) as a model for ancestral solitary foragers and tested for matrilineal signatures in their calls, thus investigating whether such signatures are already present in solitary foragers and could have facilitated the kin selection thought to have driven the evolution of increased social complexity in mammals. Because agonism can be very costly, selection for matrilineal signatures in agonistic calls should help reduce agonism between unfamiliar matrilineal kin. We conducted this study on a well-studied population of wild mouse lemurs at Ankarafantsika National Park, Madagascar. We determined pairwise relatedness using seven microsatellite loci, matrilineal relatedness by sequencing the mitrochondrial D-loop, and sleeping group associations using radio-telemetry. We recorded agonistic calls during controlled social encounters and conducted a multi-parametric acoustic analysis to determine the spectral and temporal structure of the agonistic calls. We measured 10 calls for each of 16 females from six different matrilineal kin groups. Calls were assigned to their matriline at a rate significantly higher than chance (pDFA: correct = 47.1%, chance = 26.7%, p = 0.03). There was a statistical trend for a negative correlation between acoustic distance and relatedness (Mantel Test: g = -1.61, Z = 4.61, r = -0.13, p = 0.058). Mouse lemur agonistic calls are moderately distinctive by matriline. Because sleeping groups consisted of close maternal kin, both genetics and social learning may have generated these acoustic signatures. As mouse lemurs are models for solitary foragers, we recommend further studies testing whether the lemurs use these calls to recognize kin. This would enable further modeling of how kin recognition in ancestral species could have shaped the evolution of complex sociality.
Multiple Coordination Patterns in Infant and Adult Vocalizations
Abney, Drew H.; Warlaumont, Anne S.; Oller, D. Kimbrough; Wallot, Sebastian; Kello, Christopher T.
2017-01-01
The study of vocal coordination between infants and adults has led to important insights into the development of social, cognitive, emotional and linguistic abilities. We used an automatic system to identify vocalizations produced by infants and adults over the course of the day for fifteen infants studied longitudinally during the first two years of life. We measured three different types of vocal coordination: coincidence-based, rate-based, and cluster-based. Coincidence-based and rate-based coordination are established measures in the developmental literature. Cluster-based coordination is new and measures the strength of matching in the degree to which vocalization events occur in hierarchically nested clusters. We investigated whether various coordination patterns differ as a function of vocalization type, whether different coordination patterns provide unique information about the dynamics of vocal interaction, and how the various coordination patterns each relate to infant age. All vocal coordination patterns displayed greater coordination for infant speech-related vocalizations, adults adapted the hierarchical clustering of their vocalizations to match that of infants, and each of the three coordination patterns had unique associations with infant age. Altogether, our results indicate that vocal coordination between infants and adults is multifaceted, suggesting a complex relationship between vocal coordination and the development of vocal communication. PMID:29375276
Behroozmand, Roozbeh; Karvelis, Laura; Liu, Hanjun; Larson, Charles R.
2009-01-01
Objective The present study investigated whether self-vocalization enhances auditory neural responsiveness to voice pitch feedback perturbation and how this vocalization-induced neural modulation can be affected by the extent of the feedback deviation. Method Event related potentials (ERPs) were recorded in 15 subjects in response to +100, +200 and +500 cents pitch-shifted voice auditory feedback during active vocalization and passive listening to the playback of the self-produced vocalizations. Result The amplitude of the evoked P1 (latency: 73.51 ms) and P2 (latency: 199.55 ms) ERP components in response to feedback perturbation were significantly larger during vocalization than listening. The difference between P2 peak amplitudes during vocalization vs. listening was shown to be significantly larger for +100 than +500 cents stimulus. Conclusion Results indicate that the human auditory cortex is more responsive to voice F0 feedback perturbations during vocalization than passive listening. Greater vocalization-induced enhancement of the auditory responsiveness to smaller feedback perturbations may imply that the audio-vocal system detects and corrects for errors in vocal production that closely match the expected vocal output. Significance Findings of this study support previous suggestions regarding the enhanced auditory sensitivity to feedback alterations during self-vocalization, which may serve the purpose of feedback-based monitoring of one’s voice. PMID:19520602
Fukushima, Makoto; Saunders, Richard C; Fujii, Naotaka; Averbeck, Bruno B; Mishkin, Mortimer
2014-01-01
Vocal production is an example of controlled motor behavior with high temporal precision. Previous studies have decoded auditory evoked cortical activity while monkeys listened to vocalization sounds. On the other hand, there have been few attempts at decoding motor cortical activity during vocal production. Here we recorded cortical activity during vocal production in the macaque with a chronically implanted electrocorticographic (ECoG) electrode array. The array detected robust activity in motor cortex during vocal production. We used a nonlinear dynamical model of the vocal organ to reduce the dimensionality of `Coo' calls produced by the monkey. We then used linear regression to evaluate the information in motor cortical activity for this reduced representation of calls. This simple linear model accounted for circa 65% of the variance in the reduced sound representations, supporting the feasibility of using the dynamical model of the vocal organ for decoding motor cortical activity during vocal production.