Sample records for vocal pattern generator

  1. On the role of the reticular formation in vocal pattern generation.

    PubMed

    Jürgens, Uwe; Hage, Steffen R

    2007-09-04

    This review is an attempt to localize the brain region responsible for pattern generation of species-specific vocalizations. A catalogue is set up, listing the criteria considered to be essential for a vocal pattern generator. According to this catalogue, a vocal pattern generator should show vocalization-correlated activity, starting before vocal onset and reflecting specific acoustic features of the vocalization. Artificial activation by electrical or glutamatergic stimulation should produce artificially sounding vocalization. Lesioning is expected to have an inhibitory or deteriorating effect on vocalization. Anatomically, a vocal pattern generator can be assumed to have direct or, at least, oligosynaptic connections with all the motoneuron pools involved in phonation. A survey of the literature reveals that the only area meeting all these criteria is a region, reaching from the parvocellular pontine reticular formation just above the superior olive through the lateral reticular formation around the facial nucleus and nucleus ambiguus down to the caudalmost medulla, including the dorsal and ventral reticular nuclei and nucleus retroambiguus. It is proposed that vocal pattern generation takes place within this whole region.

  2. Central pattern generators for social vocalization: Androgen-dependent neurophysiological mechanisms

    PubMed Central

    Bass, Andrew H.; Remage-Healey, Luke

    2008-01-01

    Historically, most studies of vertebrate central pattern generators (CPGs) have focused on mechanisms for locomotion and respiration. Here, we highlight new results for ectothermic vertebrates, namely teleost fish and amphibians, showing how androgenic steroids can influence the temporal patterning of CPGs for social vocalization. Investigations of vocalizing teleosts show how androgens can rapidly (within minutes) modulate the neurophysiological output of the vocal CPG (fictive vocalizations that mimic the temporal properties of natural vocalizations) inclusive of their divergent actions between species, as well as intraspecific differences between male reproductive morphs. Studies of anuran amphibians (frogs) demonstrate that long-term steroid treatments (wks) can masculinize the fictive vocalizations of females, inclusive of its sensitivity to rapid modulation by serotonin. Given the conserved organization of vocal control systems across vertebrate groups, the vocal CPGs of fish and amphibians provide tractable models for identifying androgen-dependent events that are fundamental to the mechanisms of vocal motor patterning. These basic mechanisms can also inform our understanding of the more complex CPGs for vocalization, and social behaviors in general, that have evolved among birds and mammals. PMID:18262186

  3. Temperature-dependent regulation of vocal pattern generator.

    PubMed

    Yamaguchi, Ayako; Gooler, David; Herrold, Amy; Patel, Shailja; Pong, Winnie W

    2008-12-01

    Vocalizations of Xenopus laevis are generated by central pattern generators (CPGs). The advertisement call of male X. laevis is a complex biphasic motor rhythm consisting of fast and slow trills (a train of clicks). We found that the trill rate of these advertisement calls is sensitive to temperature and that this rate modification of the vocal rhythms originates in the central pattern generators. In vivo the rates of fast and slow trills increased linearly with an increase in temperature. In vitro a similar linear relation between temperature and compound action potential frequency in the laryngeal nerve was found when fictive advertisement calls were evoked in the isolated brain. Temperature did not limit the contractile properties of laryngeal muscles within the frequency range of vocalizations. We next took advantage of the temperature sensitivity of the vocal CPG in vitro to localize the source of the vocal rhythms. We focused on the dorsal tegmental area of the medulla (DTAM), a brain stem nucleus that is essential for vocal production. We found that bilateral cooling of DTAM reduced both fast and slow trill rates. Thus we conclude that DTAM is a source of biphasic vocal rhythms.

  4. Inhibitory and modulatory inputs to the vocal central pattern generator of a teleost fish

    PubMed Central

    Rosner, Elisabeth; Rohmann, Kevin N.; Bass, Andrew H.

    2018-01-01

    Abstract Vocalization is a behavioral feature that is shared among multiple vertebrate lineages, including fish. The temporal patterning of vocal communication signals is set, in part, by central pattern generators (CPGs). Toadfishes are well‐established models for CPG coding of vocalization at the hindbrain level. The vocal CPG comprises three topographically separate nuclei: pre‐pacemaker, pacemaker, motor. While the connectivity between these nuclei is well understood, their neurochemical profile remains largely unexplored. The highly vocal Gulf toadfish, Opsanus beta, has been the subject of previous behavioral, neuroanatomical and neurophysiological studies. Combining transneuronal neurobiotin‐labeling with immunohistochemistry, we map the distribution of inhibitory neurotransmitters and neuromodulators along with gap junctions in the vocal CPG of this species. Dense GABAergic and glycinergic label is found throughout the CPG, with labeled somata immediately adjacent to or within CPG nuclei, including a distinct subset of pacemaker neurons co‐labeled with neurobiotin and glycine. Neurobiotin‐labeled motor and pacemaker neurons are densely co‐labeled with the gap junction protein connexin 35/36, supporting the hypothesis that transneuronal neurobiotin‐labeling occurs, at least in part, via gap junction coupling. Serotonergic and catecholaminergic label is also robust within the entire vocal CPG, with additional cholinergic label in pacemaker and prepacemaker nuclei. Likely sources of these putative modulatory inputs are neurons within or immediately adjacent to vocal CPG neurons. Together with prior neurophysiological investigations, the results reveal potential mechanisms for generating multiple classes of social context‐dependent vocalizations with widely divergent temporal and spectral properties. PMID:29424431

  5. Peripheral mechanisms for vocal production in birds - differences and similarities to human speech and singing.

    PubMed

    Riede, Tobias; Goller, Franz

    2010-10-01

    Song production in songbirds is a model system for studying learned vocal behavior. As in humans, bird phonation involves three main motor systems (respiration, vocal organ and vocal tract). The avian respiratory mechanism uses pressure regulation in air sacs to ventilate a rigid lung. In songbirds sound is generated with two independently controlled sound sources, which reside in a uniquely avian vocal organ, the syrinx. However, the physical sound generation mechanism in the syrinx shows strong analogies to that in the human larynx, such that both can be characterized as myoelastic-aerodynamic sound sources. Similarities include active adduction and abduction, oscillating tissue masses which modulate flow rate through the organ and a layered structure of the oscillating tissue masses giving rise to complex viscoelastic properties. Differences in the functional morphology of the sound producing system between birds and humans require specific motor control patterns. The songbird vocal apparatus is adapted for high speed, suggesting that temporal patterns and fast modulation of sound features are important in acoustic communication. Rapid respiratory patterns determine the coarse temporal structure of song and maintain gas exchange even during very long songs. The respiratory system also contributes to the fine control of airflow. Muscular control of the vocal organ regulates airflow and acoustic features. The upper vocal tract of birds filters the sounds generated in the syrinx, and filter properties are actively adjusted. Nonlinear source-filter interactions may also play a role. The unique morphology and biomechanical system for sound production in birds presents an interesting model for exploring parallels in control mechanisms that give rise to highly convergent physical patterns of sound generation. More comparative work should provide a rich source for our understanding of the evolution of complex sound producing systems. Copyright © 2009 Elsevier Inc. All rights reserved.

  6. Evolution of Courtship Songs in Xenopus : Vocal Pattern Generation and Sound Production.

    PubMed

    Leininger, Elizabeth C; Kelley, Darcy B

    2015-01-01

    The extant species of African clawed frogs (Xenopus and Silurana) provide an opportunity to link the evolution of vocal characters to changes in the responsible cellular and molecular mechanisms. In this review, we integrate several robust lines of research: evolutionary trajectories of Xenopus vocalizations, cellular and circuit-level mechanisms of vocalization in selected Xenopus model species, and Xenopus evolutionary history and speciation mechanisms. Integrating recent findings allows us to generate and test specific hypotheses about the evolution of Xenopus vocal circuits. We propose that reduced vocal sex differences in some Xenopus species result from species-specific losses of sexually differentiated neural and neuromuscular features. Modification of sex-hormone-regulated developmental mechanisms is a strong candidate mechanism for reduced vocal sex differences.

  7. Vocal fold contact patterns based on normal modes of vibration.

    PubMed

    Smith, Simeon L; Titze, Ingo R

    2018-05-17

    The fluid-structure interaction and energy transfer from respiratory airflow to self-sustained vocal fold oscillation continues to be a topic of interest in vocal fold research. Vocal fold vibration is driven by pressures on the vocal fold surface, which are determined by the shape of the glottis and the contact between vocal folds. Characterization of three-dimensional glottal shapes and contact patterns can lead to increased understanding of normal and abnormal physiology of the voice, as well as to development of improved vocal fold models, but a large inventory of shapes has not been directly studied previously. This study aimed to take an initial step toward characterizing vocal fold contact patterns systematically. Vocal fold motion and contact was modeled based on normal mode vibration, as it has been shown that vocal fold vibration can be almost entirely described by only the few lowest order vibrational modes. Symmetric and asymmetric combinations of the four lowest normal modes of vibration were superimposed on left and right vocal fold medial surfaces, for each of three prephonatory glottal configurations, according to a surface wave approach. Contact patterns were generated from the interaction of modal shapes at 16 normalized phases during the vibratory cycle. Eight major contact patterns were identified and characterized by the shape of the flow channel, with the following descriptors assigned: convergent, divergent, convergent-divergent, uniform, split, merged, island, and multichannel. Each of the contact patterns and its variation are described, and future work and applications are discussed. Copyright © 2018 Elsevier Ltd. All rights reserved.

  8. Rhythm generation, coordination, and initiation in the vocal pathways of male African clawed frogs

    PubMed Central

    Cavin Barnes, Jessica; Appleby, Todd

    2016-01-01

    Central pattern generators (CPGs) in the brain stem are considered to underlie vocalizations in many vertebrate species, but the detailed mechanisms underlying how motor rhythms are generated, coordinated, and initiated remain unclear. We addressed these issues using isolated brain preparations of Xenopus laevis from which fictive vocalizations can be elicited. Advertisement calls of male X. laevis that consist of fast and slow trills are generated by vocal CPGs contained in the brain stem. Brain stem central vocal pathways consist of a premotor nucleus [dorsal tegmental area of medulla (DTAM)] and a laryngeal motor nucleus [a homologue of nucleus ambiguus (n.IX-X)] with extensive reciprocal connections between the nuclei. In addition, DTAM receives descending inputs from the extended amygdala. We found that unilateral transection of the projections between DTAM and n.IX-X eliminated premotor fictive fast trill patterns but did not affect fictive slow trills, suggesting that the fast and slow trill CPGs are distinct; the slow trill CPG is contained in n.IX-X, and the fast trill CPG spans DTAM and n.IX-X. Midline transections that eliminated the anterior, posterior, or both commissures caused no change in the temporal structure of fictive calls, but bilateral synchrony was lost, indicating that the vocal CPGs are contained in the lateral halves of the brain stem and that the commissures synchronize the two oscillators. Furthermore, the elimination of the inputs from extended amygdala to DTAM, in addition to the anterior commissure, resulted in autonomous initiation of fictive fast but not slow trills by each hemibrain stem, indicating that the extended amygdala provides a bilateral signal to initiate fast trills. NEW & NOTEWORTHY Central pattern generators (CPGs) are considered to underlie vocalizations in many vertebrate species, but the detailed mechanisms underlying their functions remain unclear. We addressed this question using an isolated brain preparation of African clawed frogs. We discovered that two vocal phases are mediated by anatomically distinct CPGs, that there are a pair of CPGs contained in the left and right half of the brain stem, and that mechanisms underlying initiation of the two vocal phases are distinct. PMID:27760822

  9. Central pattern generator for vocalization: is there a vertebrate morphotype?

    PubMed

    Bass, Andrew H

    2014-10-01

    Animals that generate acoustic signals for social communication are faced with two essential tasks: generate a temporally precise signal and inform the auditory system about the occurrence of one's own sonic signal. Recent studies of sound producing fishes delineate a hindbrain network comprised of anatomically distinct compartments coding equally distinct neurophysiological properties that allow an organism to meet these behavioral demands. A set of neural characters comprising a vocal-sonic central pattern generator (CPG) morphotype is proposed for fishes and tetrapods that shares evolutionary developmental origins with pectoral appendage motor systems. Copyright © 2014 Elsevier Ltd. All rights reserved.

  10. Central pattern generator for vocalization: Is there a vertebrate morphotype?

    PubMed Central

    Bass, Andrew H.

    2014-01-01

    Animals that generate acoustic signals for social communication are faced with two essential tasks: generate a temporally precise signal and inform the auditory system about the occurrence of one’s own sonic signal. Recent studies of sound producing fishes delineate a hindbrain network comprised of anatomically distinct compartments coding equally distinct neurophysiological properties that allow an organism to meet these behavioral demands. A set of neural characters comprising a vocal-sonic central pattern generator (CPG) morphotype is proposed for fishes and tetrapods that shares evolutionary developmental origins with pectoral appendage motor systems. PMID:25050813

  11. Breathing and Vocal Control: The Respiratory System as both a Driver and Target of Telencephalic Vocal Motor Circuits in Songbirds

    PubMed Central

    Schmidt, Marc F.; McLean, Judith; Goller, Franz

    2011-01-01

    The production of vocalizations is intimately linked to the respiratory system. Despite our understanding of neural circuits that generate normal respiratory patterns, very little is understood regarding how these ponto-medullary circuits become engaged during vocal production. Songbirds offer a potentially powerful model system for addressing this relationship. Songs dramatically alter the respiratory pattern in ways that are often highly predictable and songbirds have a specialized telencephalic vocal motor circuit that provides massive innervation to a brainstem respiratory network that shares many similarities with its mammalian counterpart. In this review, we highlight interactions between the song motor circuit and the respiratory system, describing how both systems likely interact to produce the complex respiratory patterns that are observed during vocalization. We also discuss how the respiratory system, through its bilateral bottom-up projections to thalamus, might play a key role in sending precisely timed signals that synchronize premotor activity in both hemispheres. PMID:21984733

  12. Neural coding of syntactic structure in learned vocalizations in the songbird.

    PubMed

    Fujimoto, Hisataka; Hasegawa, Taku; Watanabe, Dai

    2011-07-06

    Although vocal signals including human languages are composed of a finite number of acoustic elements, complex and diverse vocal patterns can be created from combinations of these elements, linked together by syntactic rules. To enable such syntactic vocal behaviors, neural systems must extract the sequence patterns from auditory information and establish syntactic rules to generate motor commands for vocal organs. However, the neural basis of syntactic processing of learned vocal signals remains largely unknown. Here we report that the basal ganglia projecting premotor neurons (HVC(X) neurons) in Bengalese finches represent syntactic rules that generate variable song sequences. When vocalizing an alternative transition segment between song elements called syllables, sparse burst spikes of HVC(X) neurons code the identity of a specific syllable type or a specific transition direction among the alternative trajectories. When vocalizing a variable repetition sequence of the same syllable, HVC(X) neurons not only signal the initiation and termination of the repetition sequence but also indicate the progress and state-of-completeness of the repetition. These different types of syntactic information are frequently integrated within the activity of single HVC(X) neurons, suggesting that syntactic attributes of the individual neurons are not programmed as a basic cellular subtype in advance but acquired in the course of vocal learning and maturation. Furthermore, some auditory-vocal mirroring type HVC(X) neurons display transition selectivity in the auditory phase, much as they do in the vocal phase, suggesting that these songbirds may extract syntactic rules from auditory experience and apply them to form their own vocal behaviors.

  13. Distinct neural and neuromuscular strategies underlie independent evolution of simplified advertisement calls

    PubMed Central

    Leininger, Elizabeth C.; Kelley, Darcy B.

    2013-01-01

    Independent or convergent evolution can underlie phenotypic similarity of derived behavioural characters. Determining the underlying neural and neuromuscular mechanisms sheds light on how these characters arose. One example of evolutionarily derived characters is a temporally simple advertisement call of male African clawed frogs (Xenopus) that arose at least twice independently from a more complex ancestral pattern. How did simplification occur in the vocal circuit? To distinguish shared from divergent mechanisms, we examined activity from the calling brain and vocal organ (larynx) in two species that independently evolved simplified calls. We find that each species uses distinct neural and neuromuscular strategies to produce the simplified calls. Isolated  Xenopus borealis brains produce fictive vocal patterns that match temporal patterns of actual male calls; the larynx converts nerve activity faithfully into muscle contractions and single clicks. In contrast, fictive patterns from isolated Xenopus boumbaensis brains are short bursts of nerve activity; the isolated larynx requires stimulus bursts to produce a single click of sound. Thus, unlike X. borealis, the output of the X. boumbaensis hindbrain vocal pattern generator is an ancestral burst-type pattern, transformed by the larynx into single clicks. Temporally simple advertisement calls in genetically distant species of Xenopus have thus arisen independently via reconfigurations of central and peripheral vocal neuroeffectors. PMID:23407829

  14. Distinct neural and neuromuscular strategies underlie independent evolution of simplified advertisement calls.

    PubMed

    Leininger, Elizabeth C; Kelley, Darcy B

    2013-04-07

    Independent or convergent evolution can underlie phenotypic similarity of derived behavioural characters. Determining the underlying neural and neuromuscular mechanisms sheds light on how these characters arose. One example of evolutionarily derived characters is a temporally simple advertisement call of male African clawed frogs (Xenopus) that arose at least twice independently from a more complex ancestral pattern. How did simplification occur in the vocal circuit? To distinguish shared from divergent mechanisms, we examined activity from the calling brain and vocal organ (larynx) in two species that independently evolved simplified calls. We find that each species uses distinct neural and neuromuscular strategies to produce the simplified calls. Isolated Xenopus borealis brains produce fictive vocal patterns that match temporal patterns of actual male calls; the larynx converts nerve activity faithfully into muscle contractions and single clicks. In contrast, fictive patterns from isolated Xenopus boumbaensis brains are short bursts of nerve activity; the isolated larynx requires stimulus bursts to produce a single click of sound. Thus, unlike X. borealis, the output of the X. boumbaensis hindbrain vocal pattern generator is an ancestral burst-type pattern, transformed by the larynx into single clicks. Temporally simple advertisement calls in genetically distant species of Xenopus have thus arisen independently via reconfigurations of central and peripheral vocal neuroeffectors.

  15. Precise Motor Control Enables Rapid Flexibility in Vocal Behavior of Marmoset Monkeys.

    PubMed

    Pomberger, Thomas; Risueno-Segovia, Cristina; Löschner, Julia; Hage, Steffen R

    2018-03-05

    Investigating the evolution of human speech is difficult and controversial because human speech surpasses nonhuman primate vocal communication in scope and flexibility [1-3]. Monkey vocalizations have been assumed to be largely innate, highly affective, and stereotyped for over 50 years [4, 5]. Recently, this perception has dramatically changed. Current studies have revealed distinct learning mechanisms during vocal development [6-8] and vocal flexibility, allowing monkeys to cognitively control when [9, 10], where [11], and what to vocalize [10, 12, 13]. However, specific call features (e.g., duration, frequency) remain surprisingly robust and stable in adult monkeys, resulting in rather stereotyped and discrete call patterns [14]. Additionally, monkeys seem to be unable to modulate their acoustic call structure under reinforced conditions beyond natural constraints [15, 16]. Behavioral experiments have shown that monkeys can stop sequences of calls immediately after acoustic perturbation but cannot interrupt ongoing vocalizations, suggesting that calls consist of single impartible pulses [17, 18]. Using acoustic perturbation triggered by the vocal behavior itself and quantitative measures of resulting vocal adjustments, we show that marmoset monkeys are capable of producing calls with durations beyond the natural boundaries of their repertoire by interrupting ongoing vocalizations rapidly after perturbation onset. Our results indicate that marmosets are capable of interrupting vocalizations only at periodic time points throughout calls, further supported by the occurrence of periodically segmented phees. These ideas overturn decades-old concepts on primate vocal pattern generation, indicating that vocalizations do not consist of one discrete call pattern but are built of many sequentially uttered units, like human speech. Copyright © 2018 The Author(s). Published by Elsevier Ltd.. All rights reserved.

  16. Vocalization frequency and duration are coded in separate hindbrain nuclei.

    PubMed

    Chagnaud, Boris P; Baker, Robert; Bass, Andrew H

    2011-06-14

    Temporal patterning is an essential feature of neural networks producing precisely timed behaviours such as vocalizations that are widely used in vertebrate social communication. Here we show that intrinsic and network properties of separate hindbrain neuronal populations encode the natural call attributes of frequency and duration in vocal fish. Intracellular structure/function analyses indicate that call duration is encoded by a sustained membrane depolarization in vocal prepacemaker neurons that innervate downstream pacemaker neurons. Pacemaker neurons, in turn, encode call frequency by rhythmic, ultrafast oscillations in their membrane potential. Pharmacological manipulations show prepacemaker activity to be independent of pacemaker function, thus accounting for natural variation in duration which is the predominant feature distinguishing call types. Prepacemaker neurons also innervate key hindbrain auditory nuclei thereby effectively serving as a call-duration corollary discharge. We propose that premotor compartmentalization of neurons coding distinct acoustic attributes is a fundamental trait of hindbrain vocal pattern generators among vertebrates.

  17. Vocalization frequency and duration are coded in separate hindbrain nuclei

    PubMed Central

    Chagnaud, Boris P.; Baker, Robert; Bass, Andrew H.

    2011-01-01

    Temporal patterning is an essential feature of neural networks producing precisely timed behaviours such as vocalizations that are widely used in vertebrate social communication. Here we show that intrinsic and network properties of separate hindbrain neuronal populations encode the natural call attributes of frequency and duration in vocal fish. Intracellular structure/function analyses indicate that call duration is encoded by a sustained membrane depolarization in vocal prepacemaker neurons that innervate downstream pacemaker neurons. Pacemaker neurons, in turn, encode call frequency by rhythmic, ultrafast oscillations in their membrane potential. Pharmacological manipulations show prepacemaker activity to be independent of pacemaker function, thus accounting for natural variation in duration which is the predominant feature distinguishing call types. Prepacemaker neurons also innervate key hindbrain auditory nuclei thereby effectively serving as a call-duration corollary discharge. We propose that premotor compartmentalization of neurons coding distinct acoustic attributes is a fundamental trait of hindbrain vocal pattern generators among vertebrates. PMID:21673667

  18. From Central Pattern Generator to Sensory Template in the Evolution of Birdsong

    ERIC Educational Resources Information Center

    Konishi, Masakazu

    2010-01-01

    Central nervous networks, be they a part of the human brain or a group of neurons in a snail, may be designed to produce distinct patterns of movement. Central pattern generators can account for the development and production of normal vocal signals without auditory feedback in non-songbirds. Songbirds need auditory feedback to develop and…

  19. Complex vibratory patterns in an elephant larynx.

    PubMed

    Herbst, Christian T; Svec, Jan G; Lohscheller, Jörg; Frey, Roland; Gumpenberger, Michaela; Stoeger, Angela S; Fitch, W Tecumseh

    2013-11-01

    Elephants' low-frequency vocalizations are produced by flow-induced self-sustaining oscillations of laryngeal tissue. To date, little is known in detail about the vibratory phenomena in the elephant larynx. Here, we provide a first descriptive report of the complex oscillatory features found in the excised larynx of a 25 year old female African elephant (Loxodonta africana), the largest animal sound generator ever studied experimentally. Sound production was documented with high-speed video, acoustic measurements, air flow and sound pressure level recordings. The anatomy of the larynx was studied with computed tomography (CT) and dissections. Elephant CT vocal anatomy data were further compared with the anatomy of an adult human male. We observed numerous unusual phenomena, not typically reported in human vocal fold vibrations. Phase delays along both the inferior-superior and anterior-posterior (A-P) dimension were commonly observed, as well as transverse travelling wave patterns along the A-P dimension, previously not documented in the literature. Acoustic energy was mainly created during the instant of glottal opening. The vestibular folds, when adducted, participated in tissue vibration, effectively increasing the generated sound pressure level by 12 dB. The complexity of the observed phenomena is partly attributed to the distinct laryngeal anatomy of the elephant larynx, which is not simply a large-scale version of its human counterpart. Travelling waves may be facilitated by low fundamental frequencies and increased vocal fold tension. A travelling wave model is proposed, to account for three types of phenomena: A-P travelling waves, 'conventional' standing wave patterns, and irregular vocal fold vibration.

  20. Multiple Coordination Patterns in Infant and Adult Vocalizations

    PubMed Central

    Abney, Drew H.; Warlaumont, Anne S.; Oller, D. Kimbrough; Wallot, Sebastian; Kello, Christopher T.

    2017-01-01

    The study of vocal coordination between infants and adults has led to important insights into the development of social, cognitive, emotional and linguistic abilities. We used an automatic system to identify vocalizations produced by infants and adults over the course of the day for fifteen infants studied longitudinally during the first two years of life. We measured three different types of vocal coordination: coincidence-based, rate-based, and cluster-based. Coincidence-based and rate-based coordination are established measures in the developmental literature. Cluster-based coordination is new and measures the strength of matching in the degree to which vocalization events occur in hierarchically nested clusters. We investigated whether various coordination patterns differ as a function of vocalization type, whether different coordination patterns provide unique information about the dynamics of vocal interaction, and how the various coordination patterns each relate to infant age. All vocal coordination patterns displayed greater coordination for infant speech-related vocalizations, adults adapted the hierarchical clustering of their vocalizations to match that of infants, and each of the three coordination patterns had unique associations with infant age. Altogether, our results indicate that vocal coordination between infants and adults is multifaceted, suggesting a complex relationship between vocal coordination and the development of vocal communication. PMID:29375276

  1. Bilateral lesions of the medial frontal cortex disrupt recognition of social hierarchy during antiphonal communication in naked mole-rats (Heterocephalus glaber).

    PubMed

    Yosida, Shigeto; Okanoya, Kazuo

    2012-02-01

    Generation of the motor patterns of emotional sounds in mammals occurs in the periaqueductal gray matter of the midbrain and is not directly controlled by the cortex. The medial frontal cortex indirectly controls vocalizations, based on the recognition of social context. We examined whether the medial frontal cortex was responsible for antiphonal vocalization, or turn-taking, in naked mole-rats. In normal turn-taking, naked mole-rats vocalize more frequently to dominant individuals than to subordinate ones. Bilateral lesions of the medial frontal cortex disrupted differentiation of call rates to the stimulus animals, which had varied social relationships to the subject. However, medial frontal cortex lesions did not affect either the acoustic properties of the vocalizations or the timing of the vocal exchanges. This suggests that the medial frontal cortex may be involved in social cognition or decision making during turn-taking, while other regions of the brain regulate when animals vocalize and the vocalizations themselves.

  2. Speech-like orofacial oscillations in stump-tailed macaque (Macaca arctoides) facial and vocal signals.

    PubMed

    Toyoda, Aru; Maruhashi, Tamaki; Malaivijitnond, Suchinda; Koda, Hiroki

    2017-10-01

    Speech is unique to humans and characterized by facial actions of ∼5 Hz oscillations of lip, mouth or jaw movements. Lip-smacking, a facial display of primates characterized by oscillatory actions involving the vertical opening and closing of the jaw and lips, exhibits stable 5-Hz oscillation patterns, matching that of speech, suggesting that lip-smacking is a precursor of speech. We tested if facial or vocal actions exhibiting the same rate of oscillation are found in wide forms of facial or vocal displays in various social contexts, exhibiting diversity among species. We observed facial and vocal actions of wild stump-tailed macaques (Macaca arctoides), and selected video clips including facial displays (teeth chattering; TC), panting calls, and feeding. Ten open-to-open mouth durations during TC and feeding and five amplitude peak-to-peak durations in panting were analyzed. Facial display (TC) and vocalization (panting) oscillated within 5.74 ± 1.19 and 6.71 ± 2.91 Hz, respectively, similar to the reported lip-smacking of long-tailed macaques and the speech of humans. These results indicated a common mechanism for the central pattern generator underlying orofacial movements, which would evolve to speech. Similar oscillations in panting, which evolved from different muscular control than the orofacial action, suggested the sensory foundations for perceptual saliency particular to 5-Hz rhythms in macaques. This supports the pre-adaptation hypothesis of speech evolution, which states a central pattern generator for 5-Hz facial oscillation and perceptual background tuned to 5-Hz actions existed in common ancestors of macaques and humans, before the emergence of speech. © 2017 Wiley Periodicals, Inc.

  3. Shared developmental and evolutionary origins for neural basis of vocal–acoustic and pectoral–gestural signaling

    PubMed Central

    Bass, Andrew H.; Chagnaud, Boris P.

    2012-01-01

    Acoustic signaling behaviors are widespread among bony vertebrates, which include the majority of living fishes and tetrapods. Developmental studies in sound-producing fishes and tetrapods indicate that central pattern generating networks dedicated to vocalization originate from the same caudal hindbrain rhombomere (rh) 8-spinal compartment. Together, the evidence suggests that vocalization and its morphophysiological basis, including mechanisms of vocal–respiratory coupling that are widespread among tetrapods, are ancestral characters for bony vertebrates. Premotor-motor circuitry for pectoral appendages that function in locomotion and acoustic signaling develops in the same rh8-spinal compartment. Hence, vocal and pectoral phenotypes in fishes share both developmental origins and roles in acoustic communication. These findings lead to the proposal that the coupling of more highly derived vocal and pectoral mechanisms among tetrapods, including those adapted for nonvocal acoustic and gestural signaling, originated in fishes. Comparative studies further show that rh8 premotor populations have distinct neurophysiological properties coding for equally distinct behavioral attributes such as call duration. We conclude that neural network innovations in the spatiotemporal patterning of vocal and pectoral mechanisms of social communication, including forelimb gestural signaling, have their evolutionary origins in the caudal hindbrain of fishes. PMID:22723366

  4. Subglottal pressure and fundamental frequency control in contact calls of juvenile Alligator mississippiensis

    PubMed Central

    Riede, Tobias; Tokuda, Isao T.; Farmer, C. G.

    2011-01-01

    SUMMARY Vocalization is rare among non-avian reptiles, with the exception of the crocodilians, the sister taxon of birds. Crocodilians have a complex vocal repertoire. Their vocal and respiratory system is not well understood but appears to consist of a combination of features that are also found in the extremely vocal avian and mammalian taxa. Anatomical studies suggest that the alligator larynx is able to abduct and adduct the vocal folds, but not to elongate or shorten them, and is therefore lacking a key regulator of frequency, yet alligators can modulate fundamental frequency remarkably well. We investigated the morphological and physiological features of sound production in alligators. Vocal fold length scales isometrically across a wide range of alligator body sizes. The relationship between fundamental frequency and subglottal pressure is significant in some individuals at some isolated points, such as call onset and position of maximum fundamental frequency. The relationship is not consistent over large segments of the call. Fundamental frequency can change faster than expected by pressure changes alone, suggesting an active motor pattern controls frequency and is intrinsic to the larynx. We utilized a two-mass vocal fold model to test whether abduction and adduction could generate this motor pattern. The fine-tuned interplay between subglottal pressure and glottal adduction can achieve frequency modulations much larger than those resulting from subglottal pressure variations alone and of similar magnitude, as observed in alligator calls. We conclude that the alligator larynx represents a sound source with only two control parameters (subglottal pressure and vocal fold adduction) in contrast to the mammalian larynx in which three parameters can be altered to modulate frequency (subglottal pressure, vocal fold adduction and length/tension). PMID:21865521

  5. Stereotypic Laryngeal and Respiratory Motor Patterns Generate Different Call Types in Rat Ultrasound Vocalization

    PubMed Central

    RIEDE, TOBIAS

    2014-01-01

    Rodents produce highly variable ultrasound whistles as communication signals unlike many other mammals, who employ flow-induced vocal fold oscillations to produce sound. The role of larynx muscles in controlling sound features across different call types in ultrasound vocalization (USV) was investigated using laryngeal muscle electromyographic (EMG) activity, subglottal pressure measurements and vocal sound output in awake and spontaneously behaving Sprague–Dawley rats. Results support the hypothesis that glottal shape determines fundamental frequency. EMG activities of thyroarytenoid and cricothyroid muscles were aligned with call duration. EMG intensity increased with fundamental frequency. Phasic activities of both muscles were aligned with fast changing fundamental frequency contours, for example in trills. Activities of the sternothyroid and sternohyoid muscles, two muscles involved in vocal production in other mammals, are not critical for the production of rat USV. To test how stereotypic laryngeal and respiratory activity are across call types and individuals, sets of ten EMG and subglottal pressure parameters were measured in six different call types from six rats. Using discriminant function analysis, on average 80% of parameter sets were correctly assigned to their respective call type. This was significantly higher than the chance level. Since fundamental frequency features of USV are tightly associated with stereotypic activity of intrinsic laryngeal muscles and muscles contributing to build-up of subglottal pressure, USV provide insight into the neurophysiological control of peripheral vocal motor patterns. PMID:23423862

  6. Prosthetic avian vocal organ controlled by a freely behaving bird based on a low dimensional model of the biomechanical periphery.

    PubMed

    Arneodo, Ezequiel M; Perl, Yonatan Sanz; Goller, Franz; Mindlin, Gabriel B

    2012-01-01

    Because of the parallels found with human language production and acquisition, birdsong is an ideal animal model to study general mechanisms underlying complex, learned motor behavior. The rich and diverse vocalizations of songbirds emerge as a result of the interaction between a pattern generator in the brain and a highly nontrivial nonlinear periphery. Much of the complexity of this vocal behavior has been understood by studying the physics of the avian vocal organ, particularly the syrinx. A mathematical model describing the complex periphery as a nonlinear dynamical system leads to the conclusion that nontrivial behavior emerges even when the organ is commanded by simple motor instructions: smooth paths in a low dimensional parameter space. An analysis of the model provides insight into which parameters are responsible for generating a rich variety of diverse vocalizations, and what the physiological meaning of these parameters is. By recording the physiological motor instructions elicited by a spontaneously singing muted bird and computing the model on a Digital Signal Processor in real-time, we produce realistic synthetic vocalizations that replace the bird's own auditory feedback. In this way, we build a bio-prosthetic avian vocal organ driven by a freely behaving bird via its physiologically coded motor commands. Since it is based on a low-dimensional nonlinear mathematical model of the peripheral effector, the emulation of the motor behavior requires light computation, in such a way that our bio-prosthetic device can be implemented on a portable platform.

  7. Neuronal Control of Mammalian Vocalization, with Special Reference to the Squirrel Monkey

    NASA Astrophysics Data System (ADS)

    Jürgens, Uwe

    Squirrel monkey vocalization can be considered as a suitable model for the study in humans of the neurobiological basis of nonverbal emotional vocal utterances, such as laughing, crying, and groaning. Evaluation of electrical and chemical brain stimulation data, lesioning studies, single-neurone recordings, and neuroanatomical tracing work leads to the following conclusions: The periaqueductal gray and laterally bordering tegmentum of the midbrain represent a crucial area for the production of vocalization. This area collects the various vocalization-triggering stimuli, such as auditory, visual, and somatosensory input from diverse sensory-processing structures, motivation-controlling input from some limbic structures, and volitional impulses from the anterior cingulate cortex. Destruction of this area causes mutism. It is still under dispute whether the periaqueductal region harbors the vocal pattern generator or merely couples vocalization-triggering information to motor-coordinating structures further downward in the brainstem. The periaqueductal region is connected with the phonatory motoneuron pools indirectly via one or several interneurons. The nucleus retroambiguus represents a crucial relay station for the laryngeal and expiratory component of vocalization. The articulatory component reaches the orofacial motoneuron pools via the parvocellular reticular formation. Essential proprioceptive feedback from the larynx and lungs enter the vocal-controlling network via the solitary tract nucleus.

  8. Syllable acoustics, temporal patterns, and call composition vary with behavioral context in Mexican free-tailed bats

    PubMed Central

    Bohn, Kirsten M.; Schmidt-French, Barbara; Ma, Sean T.; Pollak, George D.

    2008-01-01

    Recent research has shown that some bat species have rich vocal repertoires with diverse syllable acoustics. Few studies, however, have compared vocalizations across different behavioral contexts or examined the temporal emission patterns of vocalizations. In this paper, a comprehensive examination of the vocal repertoire of Mexican free-tailed bats, T. brasiliensis, is presented. Syllable acoustics and temporal emission patterns for 16 types of vocalizations including courtship song revealed three main findings. First, although in some cases syllables are unique to specific calls, other syllables are shared among different calls. Second, entire calls associated with one behavior can be embedded into more complex vocalizations used in entirely different behavioral contexts. Third, when different calls are composed of similar syllables, distinctive temporal emission patterns may facilitate call recognition. These results indicate that syllable acoustics alone do not likely provide enough information for call recognition; rather, the acoustic context and temporal emission patterns of vocalizations may affect meaning. PMID:19045674

  9. A model of acoustic interspeaker variability based on the concept of formant-cavity affiliation

    NASA Astrophysics Data System (ADS)

    Apostol, Lian; Perrier, Pascal; Bailly, Gérard

    2004-01-01

    A method is proposed to model the interspeaker variability of formant patterns for oral vowels. It is assumed that this variability originates in the differences existing among speakers in the respective lengths of their front and back vocal-tract cavities. In order to characterize, from the spectral description of the acoustic speech signal, these vocal-tract differences between speakers, each formant is interpreted, according to the concept of formant-cavity affiliation, as a resonance of a specific vocal-tract cavity. Its frequency can thus be directly related to the corresponding cavity length, and a transformation model can be proposed from a speaker A to a speaker B on the basis of the frequency ratios of the formants corresponding to the same resonances. In order to minimize the number of sounds to be recorded for each speaker in order to carry out this speaker transformation, the frequency ratios are exactly computed only for the three extreme cardinal vowels [eye, aye, you] and they are approximated for the remaining vowels through an interpolation function. The method is evaluated through its capacity to transform the (F1,F2) formant patterns of eight oral vowels pronounced by five male speakers into the (F1,F2) patterns of the corresponding vowels generated by an articulatory model of the vocal tract. The resulting formant patterns are compared to those provided by normalization techniques published in the literature. The proposed method is found to be efficient, but a number of limitations are also observed and discussed. These limitations can be associated with the formant-cavity affiliation model itself or with a possible influence of speaker-specific vocal-tract geometry in the cross-sectional direction, which the model might not have taken into account.

  10. Prosthetic Avian Vocal Organ Controlled by a Freely Behaving Bird Based on a Low Dimensional Model of the Biomechanical Periphery

    PubMed Central

    Arneodo, Ezequiel M.; Perl, Yonatan Sanz; Goller, Franz; Mindlin, Gabriel B.

    2012-01-01

    Because of the parallels found with human language production and acquisition, birdsong is an ideal animal model to study general mechanisms underlying complex, learned motor behavior. The rich and diverse vocalizations of songbirds emerge as a result of the interaction between a pattern generator in the brain and a highly nontrivial nonlinear periphery. Much of the complexity of this vocal behavior has been understood by studying the physics of the avian vocal organ, particularly the syrinx. A mathematical model describing the complex periphery as a nonlinear dynamical system leads to the conclusion that nontrivial behavior emerges even when the organ is commanded by simple motor instructions: smooth paths in a low dimensional parameter space. An analysis of the model provides insight into which parameters are responsible for generating a rich variety of diverse vocalizations, and what the physiological meaning of these parameters is. By recording the physiological motor instructions elicited by a spontaneously singing muted bird and computing the model on a Digital Signal Processor in real-time, we produce realistic synthetic vocalizations that replace the bird's own auditory feedback. In this way, we build a bio-prosthetic avian vocal organ driven by a freely behaving bird via its physiologically coded motor commands. Since it is based on a low-dimensional nonlinear mathematical model of the peripheral effector, the emulation of the motor behavior requires light computation, in such a way that our bio-prosthetic device can be implemented on a portable platform. PMID:22761555

  11. Electroglottographic parameterization of the effects of gender, vowel and phonatory registers on vocal fold vibratory patterns: an Indian perspective.

    PubMed

    Paul, Nilanjan; Kumar, Suman; Chatterjee, Indranil; Mukherjee, Biswarup

    2011-01-01

    In-depth study on laryngeal biomechanics and vocal fold vibratory patterns reveal that a single vibratory cycle can be divided into two major phases, the closed and open phase, which is subdivided into opening and closing phases. Studies reveal that the relative time course of abduction and adduction, which in turn is dependent on the relative relaxing and tensing of the vocal fold cover and body, to be the determining factor in production of a particular vocal register like the modal (or chest), falsetto, glottal fry registers. Studies further point out Electroglottography to be particularly suitable for the study of vocal vibratory patterns during register changes. However, to date, there has been limited study on quantitative parameterization of EGG wave form in vocal fry register. Moreover, contradictory findings abound in literature regarding effects of gender and vowel types on vocal vibratory patterns, especially during phonation at different registers. The present study endeavors to find out the effects of vowel and gender differences on the vocal fold vibratory patterns in different registers and how these would be reflected in standard EGG parameters of Contact Quotient (CQ) and Contact Index (CI), taking into consideration the Indian sociolinguistic context. Electroglottographic recordings of 10 young adults (5 males and 5 females) were taken while the subjects phonated the three vowels /a/,/i/,/u/ each in two vocal registers, modal and vocal fry. Obtained raw EGG were further normalized using the Derived EGG algorithm and theCQ and CI values were derived. Obtained data were subject to statistical analysis using the 3-way ANOVA with gender, vowel and vocal register as the three variables. Post-hoc Dunnett C multiple comparison analysis were also performed. Results reveal that CQ values are significantly higher in vocal fry than modal phonation for both males and females, indicating a relatively hyperconstricted vocal system during vocal fry. The males have significantly greater CQ values than females both at modal and vocal fry phonations which indicate that the males are predisposed to greater vocal fold constriction. Females demonstrated no significant increase in CI values in vocal fry state; and in some cases actually decrease in the CI values which suggest an inherently distinct vocal fold physiological adjustment from that in males. No vowel effects were found in any conditions. Perturbation values (CQP and CIP) are significantly more in vocal fry register than in modal register, and the increase was more in case of females than males. The findings give strong evidence to certain hypotheses in literature regarding effects of vowel, gender and phonatory register on vocal fold vibratory patterns.

  12. Within-individual variation in bullfrog vocalizations: implications for a vocally mediated social recognition system.

    PubMed

    Bee, Mark A

    2004-12-01

    Acoustic signals provide a basis for social recognition in a wide range of animals. Few studies, however, have attempted to relate the patterns of individual variation in signals to behavioral discrimination thresholds used by receivers to discriminate among individuals. North American bullfrogs (Rana catesbeiana) discriminate among familiar and unfamiliar individuals based on individual variation in advertisement calls. The sources, patterns, and magnitudes of variation in eight acoustic properties of multiple-note advertisement calls were examined to understand how patterns of within-individual variation might either constrain, or provide additional cues for, vocal recognition. Six of eight acoustic properties exhibited significant note-to-note variation within multiple-note calls. Despite this source of within-individual variation, all call properties varied significantly among individuals, and multivariate analyses indicated that call notes were individually distinct. Fine-temporal and spectral call properties exhibited less within-individual variation compared to gross-temporal properties and contributed most toward statistically distinguishing among individuals. Among-individual differences in the patterns of within-individual variation in some properties suggest that within-individual variation could also function as a recognition cue. The distributions of among-individual and within-individual differences were used to generate hypotheses about the expected behavioral discrimination thresholds of receivers.

  13. Distribution of androgen receptor mRNA expression in vocal, auditory, and neuroendocrine circuits in a teleost fish.

    PubMed

    Forlano, Paul M; Marchaterre, Margaret; Deitcher, David L; Bass, Andrew H

    2010-02-15

    Across all major vertebrate groups, androgen receptors (ARs) have been identified in neural circuits that shape reproductive-related behaviors, including vocalization. The vocal control network of teleost fishes presents an archetypal example of how a vertebrate nervous system produces social, context-dependent sounds. We cloned a partial cDNA of AR that was used to generate specific probes to localize AR expression throughout the central nervous system of the vocal plainfin midshipman fish (Porichthys notatus). In the forebrain, AR mRNA is abundant in proposed homologs of the mammalian striatum and amygdala, and in anterior and posterior parvocellular and magnocellular nuclei of the preoptic area, nucleus preglomerulosus, and posterior, ventral and anterior tuberal nuclei of the hypothalamus. Many of these nuclei are part of the known vocal and auditory circuitry in midshipman. The midbrain periaqueductal gray, an essential link between forebrain and hindbrain vocal circuitry, and the lateral line recipient nucleus medialis in the rostral hindbrain also express abundant AR mRNA. In the caudal hindbrain-spinal vocal circuit, high AR mRNA is found in the vocal prepacemaker nucleus and along the dorsal periphery of the vocal motor nucleus congruent with the known pattern of expression of aromatase-containing glial cells. Additionally, abundant AR mRNA expression is shown for the first time in the inner ear of a vertebrate. The distribution of AR mRNA strongly supports the role of androgens as modulators of behaviorally defined vocal, auditory, and neuroendocrine circuits in teleost fish and vertebrates in general. 2009 Wiley-Liss, Inc.

  14. Ultrasonic vocalizations: a tool for behavioural phenotyping of mouse models of neurodevelopmental disorders

    PubMed Central

    Scattoni, Maria Luisa; Crawley, Jacqueline; Ricceri, Laura

    2009-01-01

    In neonatal mice ultrasonic vocalizations have been studied both as an early communicative behavior of the pup-mother dyad and as a sign of an aversive affective state. Adult mice of both sexes produce complex ultrasonic vocalization patterns in different experimental/social contexts. All these vocalizations are becoming an increasingly valuable assay for behavioral phenotyping throughout the mouse life-span and alterations of the ultrasound patterns have been reported in several mouse models of neurodevelopmental disorders. Here we also show that the modulation of vocalizations by maternal cues (maternal potentiation paradigm) – originally identified and investigated in rats - can be measured in C57Bl/6 mouse pups with appropriate modifications of the rat protocol and can likely be applied to mouse behavioral phenotyping. In addition we suggest that a detailed qualitative evaluation of neonatal calls together with analysis of adult mouse vocalization patterns in both sexes in social settings, may lead to a greater understanding of the communication value of vocalizations in mice. Importantly, both neonatal and adult USV altered patterns can be determined during the behavioural phenotyping of mouse models of human neurodevelopmental and neuropsychiatric disorders, starting from those in which deficits in communication are a primary symptom. PMID:18771687

  15. High-speed imaging of vocal fold vibrations and larynx movements within vocalizations of different vowels.

    PubMed

    Maurer, D; Hess, M; Gross, M

    1996-12-01

    Theoretic investigations of the "source-filter" model have indicated a pronounced acoustic interaction of glottal source and vocal tract. Empirical investigations of formant pattern variations apart from changes in vowel identity have demonstrated a direct relationship between the fundamental frequency and the patterns. As a consequence of both findings, independence of phonation and articulation may be limited in the speech process. Within the present study, possible interdependence of phonation and phoneme was investigated: vocal fold vibrations and larynx position for vocalizations of different vowels in a healthy man and woman were examined by high-speed light-intensified digital imaging. We found 1) different movements of the vocal folds for vocalizations of different vowel identities within one speaker and at similar fundamental frequency, and 2) constant larynx position within vocalization of one vowel identity, but different positions for vocalizations of different vowel identities. A possible relationship between the vocal fold vibrations and the phoneme is discussed.

  16. Finding the Beat: From Socially Coordinated Vocalizations in Songbirds to Rhythmic Entrainment in Humans.

    PubMed

    Benichov, Jonathan I; Globerson, Eitan; Tchernichovski, Ofer

    2016-01-01

    Humans and oscine songbirds share the rare capacity for vocal learning. Songbirds have the ability to acquire songs and calls of various rhythms through imitation. In several species, birds can even coordinate the timing of their vocalizations with other individuals in duets that are synchronized with millisecond-accuracy. It is not known, however, if songbirds can perceive rhythms holistically nor if they are capable of spontaneous entrainment to complex rhythms, in a manner similar to humans. Here we review emerging evidence from studies of rhythm generation and vocal coordination across songbirds and humans. In particular, recently developed experimental methods have revealed neural mechanisms underlying the temporal structure of song and have allowed us to test birds' abilities to predict the timing of rhythmic social signals. Surprisingly, zebra finches can readily learn to anticipate the calls of a "vocal robot" partner and alter the timing of their answers to avoid jamming, even in reference to complex rhythmic patterns. This capacity resembles, to some extent, human predictive motor response to an external beat. In songbirds, this is driven, at least in part, by the forebrain song system, which controls song timing and is essential for vocal learning. Building upon previous evidence for spontaneous entrainment in human and non-human vocal learners, we propose a comparative framework for future studies aimed at identifying shared mechanism of rhythm production and perception across songbirds and humans.

  17. Vocal exploration is locally regulated during song learning

    PubMed Central

    Ravbar, Primoz; Parra, Lucas C.; Lipkind, Dina; Tchernichovski, Ofer

    2012-01-01

    Exploratory variability is essential for sensory-motor learning, but it is not known how and at what time scales it is regulated. We manipulated song learning in zebra finches to experimentally control the requirements for vocal exploration in different parts of their song. We first trained birds to perform a one-syllable song, and once they mastered it we added a new syllable to the song model. Remarkably, when practicing the modified song, birds rapidly alternated between high and low acoustic variability to confine vocal exploration to the newly added syllable. Further, even within syllables, acoustic variability changed independently across song elements that were only milliseconds apart. Analysis of the entire vocal output during learning revealed that the variability of each song element decreased as it approached the target, correlating with momentary local distance from the target and less so with the overall distance. We conclude that vocal error is computed locally in sub-syllabic time scales and that song elements can be learned and crystalized independently. Songbirds have dedicated brain circuitry for vocal babbling in the anterior forebrain pathway (AFP), which generates exploratory song patterns that drive premotor neurons at the song nucleus RA (robust nucleus of the arcopallium). We hypothesize that either AFP adjusts the gain of vocal exploration in fine time scales, or that the sensitivity of RA premotor neurons to AFP/HVC inputs varies across song elements. PMID:22399765

  18. Meaning in the avian auditory cortex: Neural representation of communication calls

    PubMed Central

    Elie, Julie E; Theunissen, Frédéric E

    2014-01-01

    Understanding how the brain extracts the behavioral meaning carried by specific vocalization types that can be emitted by various vocalizers and in different conditions is a central question in auditory research. This semantic categorization is a fundamental process required for acoustic communication and presupposes discriminative and invariance properties of the auditory system for conspecific vocalizations. Songbirds have been used extensively to study vocal learning, but the communicative function of all their vocalizations and their neural representation has yet to be examined. In our research, we first generated a library containing almost the entire zebra finch vocal repertoire and organized communication calls along 9 different categories based on their behavioral meaning. We then investigated the neural representations of these semantic categories in the primary and secondary auditory areas of 6 anesthetized zebra finches. To analyze how single units encode these call categories, we described neural responses in terms of their discrimination, selectivity and invariance properties. Quantitative measures for these neural properties were obtained using an optimal decoder based both on spike counts and spike patterns. Information theoretic metrics show that almost half of the single units encode semantic information. Neurons achieve higher discrimination of these semantic categories by being more selective and more invariant. These results demonstrate that computations necessary for semantic categorization of meaningful vocalizations are already present in the auditory cortex and emphasize the value of a neuro-ethological approach to understand vocal communication. PMID:25728175

  19. The neural network classification of false killer whale (Pseudorca crassidens) vocalizations.

    PubMed

    Murray, S O; Mercado, E; Roitblat, H L

    1998-12-01

    This study reports the use of unsupervised, self-organizing neural network to categorize the repertoire of false killer whale vocalizations. Self-organizing networks are capable of detecting patterns in their input and partitioning those patterns into categories without requiring that the number or types of categories be predefined. The inputs for the neural networks were two-dimensional characterization of false killer whale vocalization, where each vocalization was characterized by a sequence of short-time measurements of duty cycle and peak frequency. The first neural network used competitive learning, where units in a competitive layer distributed themselves to recognize frequently presented input vectors. This network resulted in classes representing typical patterns in the vocalizations. The second network was a Kohonen feature map which organized the outputs topologically, providing a graphical organization of pattern relationships. The networks performed well as measured by (1) the average correlation between the input vectors and the weight vectors for each category, and (2) the ability of the networks to classify novel vocalizations. The techniques used in this study could easily be applied to other species and facilitate the development of objective, comprehensive repertoire models.

  20. Medial surface dynamics of the vocal folds in an in vivo canine model

    NASA Astrophysics Data System (ADS)

    Doellinger, Michael; Berke, Gerald S.; Chhetri, Dinesh K.; Berry, David A.

    2004-05-01

    Quantitative measurement of the medial surface dynamics of the vocal folds is important for understanding how sound is generated in the larynx. However, such data are hard to gather because of the inaccessibility of the vocal folds. Recent studies have applied hemi-larynx methodology to excised human larynges, to visualize these dynamics. The present study extends this methodology to obtain similar quantitative measurements using an in vivo canine hemi-larynx setup, with varying levels of stimulation to the recurrent laryngeal nerve. Use of an in vivo model allows us to examine effects of intrinsic muscle contraction on the medial surface of the vocal folds, to provide greater insight into mechanisms of vocal control. Data were collected using digital high-speed imaging with a sampling frequency of up to 4000 Hz, and a spatial resolution of up to 1024×1024 pixels. Three-dimensional motion will be extracted, computed, visualized, and contrasted as a function of the level of stimulation to the recurrent laryngeal nerve. Results will also be compared to patterns of vibration in excised larynges. Finally, commonly applied quantitative analyses will be performed to investigate the underlying modes of vibration. [Work supported by NIH/NIDCD.

  1. Software for objective comparison of vocal acoustic features over weeks of audio recording: KLFromRecordingDays

    NASA Astrophysics Data System (ADS)

    Soderstrom, Ken; Alalawi, Ali

    KLFromRecordingDays allows measurement of Kullback-Leibler (KL) distances between 2D probability distributions of vocal acoustic features. Greater KL distance measures reflect increased phonological divergence across the vocalizations compared. The software has been used to compare *.wav file recordings made by Sound Analysis Recorder 2011 of songbird vocalizations pre- and post-drug and surgical manipulations. Recordings from individual animals in *.wav format are first organized into subdirectories by recording day and then segmented into individual syllables uttered and acoustic features of these syllables using Sound Analysis Pro 2011 (SAP). KLFromRecordingDays uses syllable acoustic feature data output by SAP to a MySQL table to generate and compare "template" (typically pre-treatment) and "target" (typically post-treatment) probability distributions. These distributions are a series of virtual 2D plots of the duration of each syllable (as x-axis) to each of 13 other acoustic features measured by SAP for that syllable (as y-axes). Differences between "template" and "target" probability distributions for each acoustic feature are determined by calculating KL distance, a measure of divergence of the target 2D distribution pattern from that of the template. KL distances and the mean KL distance across all acoustic features are calculated for each recording day and output to an Excel spreadsheet. Resulting data for individual subjects may then be pooled across treatment groups and graphically summarized and used for statistical comparisons. Because SAP-generated MySQL files are accessed directly, data limits associated with spreadsheet output are avoided, and the totality of vocal output over weeks may be objectively analyzed all at once. The software has been useful for measuring drug effects on songbird vocalizations and assessing recovery from damage to regions of vocal motor cortex. It may be useful in studies employing other species, and as part of speech therapies tracking progress in producing distinct speech sounds in isolation.

  2. Lateralization as a symmetry breaking process in birdsong

    NASA Astrophysics Data System (ADS)

    Trevisan, M. A.; Cooper, B.; Goller, F.; Mindlin, G. B.

    2007-03-01

    The singing by songbirds is a most convincing example in the animal kingdom of functional lateralization of the brain, a feature usually associated with human language. Lateralization is expressed as one or both of the bird’s sound sources being active during the vocalization. Normal songs require high coordination between the vocal organ and respiratory activity, which is bilaterally symmetric. Moreover, the physical and neural substrate used to produce the song lack obvious asymmetries. In this work we show that complex spatiotemporal patterns of motor activity controlling airflow through the sound sources can be explained in terms of spontaneous symmetry breaking bifurcations. This analysis also provides a framework from which to study the effects of imperfections in the system’ s symmetries. A physical model of the avian vocal organ is used to generate synthetic sounds, which allows us to predict acoustical signatures of the song and compare the predictions of the model with experimental data.

  3. Exploring the anatomical encoding of voice with a mathematical model of the vocal system.

    PubMed

    Assaneo, M Florencia; Sitt, Jacobo; Varoquaux, Gael; Sigman, Mariano; Cohen, Laurent; Trevisan, Marcos A

    2016-11-01

    The faculty of language depends on the interplay between the production and perception of speech sounds. A relevant open question is whether the dimensions that organize voice perception in the brain are acoustical or depend on properties of the vocal system that produced it. One of the main empirical difficulties in answering this question is to generate sounds that vary along a continuum according to the anatomical properties the vocal apparatus that produced them. Here we use a mathematical model that offers the unique possibility of synthesizing vocal sounds by controlling a small set of anatomically based parameters. In a first stage the quality of the synthetic voice was evaluated. Using specific time traces for sub-glottal pressure and tension of the vocal folds, the synthetic voices generated perceptual responses, which are indistinguishable from those of real speech. The synthesizer was then used to investigate how the auditory cortex responds to the perception of voice depending on the anatomy of the vocal apparatus. Our fMRI results show that sounds are perceived as human vocalizations when produced by a vocal system that follows a simple relationship between the size of the vocal folds and the vocal tract. We found that these anatomical parameters encode the perceptual vocal identity (male, female, child) and show that the brain areas that respond to human speech also encode vocal identity. On the basis of these results, we propose that this low-dimensional model of the vocal system is capable of generating realistic voices and represents a novel tool to explore the voice perception with a precise control of the anatomical variables that generate speech. Furthermore, the model provides an explanation of how auditory cortices encode voices in terms of the anatomical parameters of the vocal system. Copyright © 2016 Elsevier Inc. All rights reserved.

  4. Memory-dependent adjustment of vocal response latencies in a territorial songbird.

    PubMed

    Geberzahn, Nicole; Hultsch, Henrike; Todt, Dietmar

    2013-06-01

    Vocal interactions in songbirds can be used as a model system to investigate the interplay of intrinsic singing programmes (e.g. influences from vocal memories) and external variables (e.g. social factors). When characterizing vocal interactions between territorial rivals two aspects are important: (1) the timing of songs in relation to the conspecific's singing and (2) the use of a song pattern that matches the rival's song. Responses in both domains can be used to address a territorial rival. This study is the first to investigate the relation of the timing of vocal responses to (1) the vocal memory of a responding subject and (2) the selection of the song pattern that the subject uses as a response. To this end, we conducted interactive playback experiments with adult nightingales (Luscinia megarhynchos) that had been hand-reared and tutored in the laboratory. We analysed the subjects' vocal response latencies towards broadcast playback stimuli that they either had in their own vocal repertoire (songs shared with playback) or that they had not heard before (unknown songs). Likewise, we compared vocal response latencies between responses that matched the stimulus song and those that did not. Our findings showed that the latency of singing in response to the playback was shorter for shared versus unknown song stimuli when subjects overlapped the playback stimuli with their own song. Moreover birds tended to overlap faster when vocally matching the stimulus song rather than when replying with a non-matching song type. We conclude that memory of song patterns influenced response latencies and discuss possible mechanisms. Copyright © 2012 Elsevier Ltd. All rights reserved.

  5. Iconicity can ground the creation of vocal symbols.

    PubMed

    Perlman, Marcus; Dale, Rick; Lupyan, Gary

    2015-08-01

    Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative 'vocal' charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic 'meaning template' we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems.

  6. Automatic mouse ultrasound detector (A-MUD): A new tool for processing rodent vocalizations.

    PubMed

    Zala, Sarah M; Reitschmidt, Doris; Noll, Anton; Balazs, Peter; Penn, Dustin J

    2017-01-01

    House mice (Mus musculus) emit complex ultrasonic vocalizations (USVs) during social and sexual interactions, which have features similar to bird song (i.e., they are composed of several different types of syllables, uttered in succession over time to form a pattern of sequences). Manually processing complex vocalization data is time-consuming and potentially subjective, and therefore, we developed an algorithm that automatically detects mouse ultrasonic vocalizations (Automatic Mouse Ultrasound Detector or A-MUD). A-MUD is a script that runs on STx acoustic software (S_TOOLS-STx version 4.2.2), which is free for scientific use. This algorithm improved the efficiency of processing USV files, as it was 4-12 times faster than manual segmentation, depending upon the size of the file. We evaluated A-MUD error rates using manually segmented sound files as a 'gold standard' reference, and compared them to a commercially available program. A-MUD had lower error rates than the commercial software, as it detected significantly more correct positives, and fewer false positives and false negatives. The errors generated by A-MUD were mainly false negatives, rather than false positives. This study is the first to systematically compare error rates for automatic ultrasonic vocalization detection methods, and A-MUD and subsequent versions will be made available for the scientific community.

  7. Motor Neurons Tune Premotor Activity in a Vertebrate Central Pattern Generator

    PubMed Central

    2017-01-01

    Central patterns generators (CPGs) are neural circuits that drive rhythmic motor output without sensory feedback. Vertebrate CPGs are generally believed to operate in a top-down manner in which premotor interneurons activate motor neurons that in turn drive muscles. In contrast, the frog (Xenopus laevis) vocal CPG contains a functionally unexplored neuronal projection from the motor nucleus to the premotor nucleus, indicating a recurrent pathway that may contribute to rhythm generation. In this study, we characterized the function of this bottom-up connection. The X. laevis vocal CPG produces a 50–60 Hz “fast trill” song used by males during courtship. We recorded “fictive vocalizations” in the in vitro CPG from the laryngeal nerve while simultaneously recording premotor activity at the population and single-cell level. We show that transecting the motor-to-premotor projection eliminated the characteristic firing rate of premotor neurons. Silencing motor neurons with the intracellular sodium channel blocker QX-314 also disrupted premotor rhythms, as did blockade of nicotinic synapses in the motor nucleus (the putative location of motor neuron-to-interneuron connections). Electrically stimulating the laryngeal nerve elicited primarily IPSPs in premotor neurons that could be blocked by a nicotinic receptor antagonist. Our results indicate that an inhibitory signal, activated by motor neurons, is required for proper CPG function. To our knowledge, these findings represent the first example of a CPG in which precise premotor rhythms are tuned by motor neuron activity. SIGNIFICANCE STATEMENT Central pattern generators (CPGs) are neural circuits that produce rhythmic behaviors. In vertebrates, motor neurons are not commonly known to contribute to CPG function, with the exception of a few spinal circuits where the functional significance of motor neuron feedback is still poorly understood. The frog hindbrain vocal circuit contains a previously unexplored connection from the motor to premotor region. Our results indicate that motor neurons activate this bottom-up connection, and blocking this signal eliminates normal premotor activity. These findings may promote increased awareness of potential involvement of motor neurons in a wider range of CPGs, perhaps clarifying our understanding of network principles underlying motor behaviors in numerous organisms, including humans. PMID:28219984

  8. Iconicity can ground the creation of vocal symbols

    PubMed Central

    Perlman, Marcus; Dale, Rick; Lupyan, Gary

    2015-01-01

    Studies of gestural communication systems find that they originate from spontaneously created iconic gestures. Yet, we know little about how people create vocal communication systems, and many have suggested that vocalizations do not afford iconicity beyond trivial instances of onomatopoeia. It is unknown whether people can generate vocal communication systems through a process of iconic creation similar to gestural systems. Here, we examine the creation and development of a rudimentary vocal symbol system in a laboratory setting. Pairs of participants generated novel vocalizations for 18 different meanings in an iterative ‘vocal’ charades communication game. The communicators quickly converged on stable vocalizations, and naive listeners could correctly infer their meanings in subsequent playback experiments. People's ability to guess the meanings of these novel vocalizations was predicted by how close the vocalization was to an iconic ‘meaning template’ we derived from the production data. These results strongly suggest that the meaningfulness of these vocalizations derived from iconicity. Our findings illuminate a mechanism by which iconicity can ground the creation of vocal symbols, analogous to the function of iconicity in gestural communication systems. PMID:26361547

  9. The perceptual features of vocal fatigue as self-reported by a group of actors and singers.

    PubMed

    Kitch, J A; Oates, J

    1994-09-01

    Performers (10 actors/10 singers) rated via a self-report questionnaire the severity of their voice-related changes when vocally fatigued. Similar frequency patterns and perceptual features of vocal fatigue were found across subjects. Actors rated "power" aspects (e.g., voice projection) and singers rated vocal dynamic aspects (e.g., pitch range) of their voices as most affected when vocally fatigued. Vocal fatigue was evidenced by changes in kinesthetic/proprioceptive sensations and vocal dynamics. The causes and context of vocal fatigue were vocal misuse, being "run down," high performance demands, and using high pitch/volume levels. Further research is needed to delineate the perceptual features of "normal" levels of vocal fatigue and its possible causes.

  10. Vocal patterns in infants with autism spectrum disorder: canonical babbling status and vocalization frequency.

    PubMed

    Patten, Elena; Belardi, Katie; Baranek, Grace T; Watson, Linda R; Labban, Jeffrey D; Oller, D Kimbrough

    2014-10-01

    Canonical babbling is a critical milestone for speech development and is usually well in place by 10 months. The possibility that infants with autism spectrum disorder (ASD) show late onset of canonical babbling has so far eluded evaluation. Rate of vocalization or "volubility" has also been suggested as possibly aberrant in infants with ASD. We conducted a retrospective video study examining vocalizations of 37 infants at 9-12 and 15-18 months. Twenty-three of the 37 infants were later diagnosed with ASD and indeed produced low rates of canonical babbling and low volubility by comparison with the 14 typically developing infants. The study thus supports suggestions that very early vocal patterns may prove to be a useful component of early screening and diagnosis of ASD.

  11. Vocal patterns in infants with Autism Spectrum Disorder: Canonical babbling status and vocalization frequency

    PubMed Central

    Patten, Elena; Belardi, Katie; Baranek, Grace T.; Watson, Linda R.; Labban, Jeffrey D.; Oller, D. Kimbrough

    2014-01-01

    Canonical babbling is a critical milestone for speech development and is usually well in place by 10 months. The possibility that infants with ASD show late onset of canonical babbling has so far eluded evaluation. Rate of vocalization or “volubility” has also been suggested as possibly aberrant in infants with ASD. We conducted a retrospective video study examining vocalizations of 37 infants at 9–12 and 15–18 months. Twenty-three of the 37 infants were later diagnosed with ASD and indeed produced low rates of canonical babbling and low volubility by comparison with the 14 typically developing infants. The study thus supports suggestions that very early vocal patterns may prove to be a useful component of early screening and diagnosis of ASD. PMID:24482292

  12. Spatial location influences vocal interactions in bullfrog choruses

    PubMed Central

    Bates, Mary E.; Cropp, Brett F.; Gonchar, Marina; Knowles, Jeffrey; Simmons, James A.; Simmons, Andrea Megela

    2010-01-01

    A multiple sensor array was employed to identify the spatial locations of all vocalizing male bullfrogs (Rana catesbeiana) in five natural choruses. Patterns of vocal activity collected with this array were compared with computer simulations of chorus activity. Bullfrogs were not randomly spaced within choruses, but tended to cluster into closely spaced groups of two to five vocalizing males. There were nonrandom, differing patterns of vocal interactions within clusters of closely spaced males and between different clusters. Bullfrogs located within the same cluster tended to overlap or alternate call notes with two or more other males in that cluster. These near-simultaneous calling bouts produced advertisement calls with more pronounced amplitude modulation than occurred in nonoverlapping notes or calls. Bullfrogs located in different clusters more often alternated entire calls or overlapped only small segments of their calls. They also tended to respond sequentially to calls of their farther neighbors compared to their nearer neighbors. Results of computational analyses showed that the observed patterns of vocal interactions were significantly different than expected based on random activity. The use of a multiple sensor array provides a richer view of the dynamics of choruses than available based on single microphone techniques. PMID:20370047

  13. SDI Software Technology Program Plan Version 1.5

    DTIC Science & Technology

    1987-06-01

    computer generation of auditory communication of meaningful speech. Most speech synthesizers are based on mathematical models of the human vocal tract, but...oral/ auditory and multimodal communications. Although such state-of-the-art interaction technology has not fully matured, user experience has...superior I pattern matching capabilities and the subliminal intuitive deduction capability. The error performance of humans can be helped by careful

  14. Cumulative and Synergistic Effects of Physical, biological, and Acoustic Signals on Marine Mammal Habitat Use

    DTIC Science & Technology

    2013-04-01

    a simultaneous time series of marine mammal vocalizations and changing soundscapes (sound levels and spectral shapes) related to surface conditions...mooring (Figures 4, 6, and 7). 1 Figure 4. Seasoanl soundscapes generated... soundscapes in fall (a) and summer (d) show a linear pattern indicating an environment dominated by wind. Sound levels increase linearly as wind

  15. A robotic voice simulator and the interactive training for hearing-impaired people.

    PubMed

    Sawada, Hideyuki; Kitani, Mitsuki; Hayashi, Yasumori

    2008-01-01

    A talking and singing robot which adaptively learns the vocalization skill by means of an auditory feedback learning algorithm is being developed. The robot consists of motor-controlled vocal organs such as vocal cords, a vocal tract and a nasal cavity to generate a natural voice imitating a human vocalization. In this study, the robot is applied to the training system of speech articulation for the hearing-impaired, because the robot is able to reproduce their vocalization and to teach them how it is to be improved to generate clear speech. The paper briefly introduces the mechanical construction of the robot and how it autonomously acquires the vocalization skill in the auditory feedback learning by listening to human speech. Then the training system is described, together with the evaluation of the speech training by auditory impaired people.

  16. Vocal activity of lesser galagos (Galago spp.) at zoos.

    PubMed

    Schneiderová, Irena; Zouhar, Jan; Štefanská, Lucie; Bolfíková, Barbora Černá; Lhota, Stanislav; Brandl, Pavel

    2016-01-01

    Almost nothing is known about the natural vocal behavior of lesser galagos living in zoos. This is perhaps because they are usually kept in nocturnal exhibits separated from the visitors by a transparent and acoustically insulating glass barrier. The aim of the present study was therefore to fill this gap in knowledge of the vocal behavior of lesser galagos from zoos. This knowledge might be beneficial because the vocalizations of these small primates can be used for species determination. We performed a 10-day-long acoustic monitoring of vocal activity in each of seven various groups of Galago senegalensis and G. moholi living at four zoos. We quantitatively evaluated the occurrence of four loud vocalization types present in both species, including the most species-specific advertisement call. We found that qualitative as well as quantitative differences exist in the vocal behavior of the studied groups. We confirmed that the observed vocalization types can be collected from lesser galagos living at zoos, and the success can be increased by selecting larger and more diverse groups. We found two distinct patterns of diel vocal activity in the most vocally active groups. G. senegalensis groups were most vocally active at the beginning and at the end of their activity period, whereas one G. moholi group showed an opposite pattern. The latter is surprising, as it is generally accepted that lesser galagos emit advertisement calls especially at dawn and dusk, i.e., at the beginning and at the end of their diel activity. © 2016 Wiley Periodicals, Inc.

  17. Learning to breathe and sing: development of respiratory-vocal coordination in young songbirds

    PubMed Central

    Veit, Lena; Aronov, Dmitriy

    2011-01-01

    How do animals with learned vocalizations coordinate vocal production with respiration? Songbirds such as the zebra finch learn their songs, beginning with highly variable babbling vocalizations known as subsong. After several weeks of practice, zebra finches are able to produce a precisely timed pattern of syllables and silences, precisely coordinated with expiratory and inspiratory pulses (Franz M, Goller F. J Neurobiol 51: 129–141, 2002). While respiration in adult song is well described, relatively little is known about respiratory patterns in subsong or about the processes by which respiratory and vocal patterns become coordinated. To address these questions, we recorded thoracic air sac pressure in juvenile zebra finches prior to the appearance of any consistent temporal or acoustic structure in their songs. We found that subsong contains brief inspiratory pulses (50 ms) alternating with longer pulses of sustained expiratory pressure (50–500 ms). In striking contrast to adult song, expiratory pulses often contained multiple (0–8) variably timed syllables separated by expiratory gaps and were only partially vocalized. During development, expiratory pulses became shorter and more stereotyped in duration with shorter and fewer nonvocalized parts. These developmental changes eventually resulted in the production of a single syllable per expiratory pulse and a single inspiratory pulse filling each gap, forming a coordinated sequence similar to that of adult song. To examine the role of forebrain song-control nuclei in the development of respiratory patterns, we performed pressure recordings before and after lesions of nucleus HVC (proper name) and found that this manipulation reverses the developmental trends in measures of the respiratory pattern. PMID:21697438

  18. Learning to breathe and sing: development of respiratory-vocal coordination in young songbirds.

    PubMed

    Veit, Lena; Aronov, Dmitriy; Fee, Michale S

    2011-10-01

    How do animals with learned vocalizations coordinate vocal production with respiration? Songbirds such as the zebra finch learn their songs, beginning with highly variable babbling vocalizations known as subsong. After several weeks of practice, zebra finches are able to produce a precisely timed pattern of syllables and silences, precisely coordinated with expiratory and inspiratory pulses (Franz M, Goller F. J Neurobiol 51: 129-141, 2002). While respiration in adult song is well described, relatively little is known about respiratory patterns in subsong or about the processes by which respiratory and vocal patterns become coordinated. To address these questions, we recorded thoracic air sac pressure in juvenile zebra finches prior to the appearance of any consistent temporal or acoustic structure in their songs. We found that subsong contains brief inspiratory pulses (50 ms) alternating with longer pulses of sustained expiratory pressure (50-500 ms). In striking contrast to adult song, expiratory pulses often contained multiple (0-8) variably timed syllables separated by expiratory gaps and were only partially vocalized. During development, expiratory pulses became shorter and more stereotyped in duration with shorter and fewer nonvocalized parts. These developmental changes eventually resulted in the production of a single syllable per expiratory pulse and a single inspiratory pulse filling each gap, forming a coordinated sequence similar to that of adult song. To examine the role of forebrain song-control nuclei in the development of respiratory patterns, we performed pressure recordings before and after lesions of nucleus HVC (proper name) and found that this manipulation reverses the developmental trends in measures of the respiratory pattern.

  19. Evolutionary Origins for Social Vocalization in a Vertebrate Hindbrain–Spinal Compartment

    PubMed Central

    Bass, Andrew H.; Gilland, Edwin H.; Baker, Robert

    2008-01-01

    The macroevolutionary events leading to neural innovations for social communication, such as vocalization, are essentially unexplored. Many fish vocalize during female courtship and territorial defense, as do amphibians, birds, and mammals. Here, we map the neural circuitry for vocalization in larval fish and show that the vocal network develops in a segment-like region across the most caudal hindbrain and rostral spinal cord. Taxonomic analysis demonstrates a highly conserved pattern between fish and all major lineages of vocal tetrapods. We propose that the vocal basis for acoustic communication among vertebrates evolved from an ancestrally shared developmental compartment already present in the early fishes. PMID:18635807

  20. Non-song vocalizations of pygmy blue whales in Geographe Bay, Western Australia.

    PubMed

    Recalde-Salas, A; Salgado Kent, C P; Parsons, M J G; Marley, S A; McCauley, R D

    2014-05-01

    Non-song vocalizations of migrating pygmy blue whales (Balaenoptera musculus brevicauda) in Western Australia are described. Simultaneous land-based visual observations and underwater acoustic recordings detected 27 groups in Geographe Bay, WA over 2011 to 2012. Six different vocalizations were recorded that were not repeated in a pattern or in association with song, and thus were identified as non-song vocalizations. Five of these were not previously described for this population. Their acoustic characteristics and context are presented. Given that 56% of groups vocalized, 86% of which produced non-song vocalizations and 14% song units, the inclusion of non-song vocalizations in passive-acoustic monitoring is proposed.

  1. Occurrence Frequencies of Acoustic Patterns of Vocal Fry in American English Speakers.

    PubMed

    Abdelli-Beruh, Nassima B; Drugman, Thomas; Red Owl, R H

    2016-11-01

    The goal of this study was to analyze the occurrence frequencies of three individual acoustic patterns (A, B, C) and of vocal fry overall (A + B + C) as a function of gender, word position in the sentence (Not Last Word vs. Last Word), and sentence length (number of words in a sentence). This is an experimental design. Twenty-five male and 29 female American English (AE) speakers read the Grandfather Passage. The recordings were processed by a Matlab toolbox designed for the analysis and detection of creaky segments, automatically identified using the Kane-Drugman algorithm. The experiment produced subsamples of outcomes, three that reflect a single, discrete acoustic pattern (A, B, or C) and the fourth that reflects the occurrence frequency counts of Vocal Fry Overall without regard to any specific pattern. Zero-truncated Poisson regression analyses were conducted with Gender and Word Position as predictors and Sentence Length as a covariate. The results of the present study showed that the occurrence frequencies of the three acoustic patterns and vocal fry overall (A + B + C) are greatest at the end of sentences but are unaffected by sentence length. The findings also reveal that AE female speakers exhibit Pattern C significantly more frequently than Pattern B, and the converse holds for AE male speakers. Future studies are needed to confirm such outcomes, assess the perceptual salience of these acoustic patterns, and determine the physiological correlates of these acoustic patterns. The findings have implications for the design of new excitation models of vocal fry. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  2. Stimulation of the basal and central amygdala in the mustached bat triggers echolocation and agonistic vocalizations within multimodal output

    PubMed Central

    Ma, Jie; Kanwal, Jagmeet S.

    2014-01-01

    The neural substrate for the perception of vocalizations is relatively well described, but how their timing and specificity are tightly coupled with accompanying physiological changes and context-appropriate behaviors remains unresolved. We hypothesized that temporally integrated vocal and emotive responses, especially the expression of fear, vigilance and aggression, originate within the amygdala. To test this hypothesis, we performed electrical microstimulation at 461 highly restricted loci within the basal and central amygdala in awake mustached bats. At a subset of these sites, high frequency stimulation with weak constant current pulses presented at near-threshold levels triggered vocalization of either echolocation pulses or social calls. At the vast majority of locations, microstimulation produced a constellation of changes in autonomic and somatomotor outputs. These changes included widespread co-activation of significant tachycardia and hyperventilation and/or rhythmic ear pinna movements (PMs). In a few locations, responses were constrained to vocalization and/or PMs despite increases in the intensity of stimulation. The probability of eliciting echolocation pulses vs. social calls decreased in a medial-posterior to anterolateral direction within the centrobasal amygdala. Microinjections of kainic acid (KA) at stimulation sites confirmed the contribution of cellular activity rather than fibers-of-passage in the control of multimodal outputs. The results suggest that localized clusters of neurons may simultaneously modulate the activity of multiple central pattern generators (CPGs) present within the brainstem. PMID:24624089

  3. Stimulation of the basal and central amygdala in the mustached bat triggers echolocation and agonistic vocalizations within multimodal output.

    PubMed

    Ma, Jie; Kanwal, Jagmeet S

    2014-01-01

    The neural substrate for the perception of vocalizations is relatively well described, but how their timing and specificity are tightly coupled with accompanying physiological changes and context-appropriate behaviors remains unresolved. We hypothesized that temporally integrated vocal and emotive responses, especially the expression of fear, vigilance and aggression, originate within the amygdala. To test this hypothesis, we performed electrical microstimulation at 461 highly restricted loci within the basal and central amygdala in awake mustached bats. At a subset of these sites, high frequency stimulation with weak constant current pulses presented at near-threshold levels triggered vocalization of either echolocation pulses or social calls. At the vast majority of locations, microstimulation produced a constellation of changes in autonomic and somatomotor outputs. These changes included widespread co-activation of significant tachycardia and hyperventilation and/or rhythmic ear pinna movements (PMs). In a few locations, responses were constrained to vocalization and/or PMs despite increases in the intensity of stimulation. The probability of eliciting echolocation pulses vs. social calls decreased in a medial-posterior to anterolateral direction within the centrobasal amygdala. Microinjections of kainic acid (KA) at stimulation sites confirmed the contribution of cellular activity rather than fibers-of-passage in the control of multimodal outputs. The results suggest that localized clusters of neurons may simultaneously modulate the activity of multiple central pattern generators (CPGs) present within the brainstem.

  4. Automatic mouse ultrasound detector (A-MUD): A new tool for processing rodent vocalizations

    PubMed Central

    Reitschmidt, Doris; Noll, Anton; Balazs, Peter; Penn, Dustin J.

    2017-01-01

    House mice (Mus musculus) emit complex ultrasonic vocalizations (USVs) during social and sexual interactions, which have features similar to bird song (i.e., they are composed of several different types of syllables, uttered in succession over time to form a pattern of sequences). Manually processing complex vocalization data is time-consuming and potentially subjective, and therefore, we developed an algorithm that automatically detects mouse ultrasonic vocalizations (Automatic Mouse Ultrasound Detector or A-MUD). A-MUD is a script that runs on STx acoustic software (S_TOOLS-STx version 4.2.2), which is free for scientific use. This algorithm improved the efficiency of processing USV files, as it was 4–12 times faster than manual segmentation, depending upon the size of the file. We evaluated A-MUD error rates using manually segmented sound files as a ‘gold standard’ reference, and compared them to a commercially available program. A-MUD had lower error rates than the commercial software, as it detected significantly more correct positives, and fewer false positives and false negatives. The errors generated by A-MUD were mainly false negatives, rather than false positives. This study is the first to systematically compare error rates for automatic ultrasonic vocalization detection methods, and A-MUD and subsequent versions will be made available for the scientific community. PMID:28727808

  5. Specialized Motor-Driven dusp1 Expression in the Song Systems of Multiple Lineages of Vocal Learning Birds

    PubMed Central

    Horita, Haruhito; Kobayashi, Masahiko; Liu, Wan-chun; Oka, Kotaro; Jarvis, Erich D.; Wada, Kazuhiro

    2012-01-01

    Mechanisms for the evolution of convergent behavioral traits are largely unknown. Vocal learning is one such trait that evolved multiple times and is necessary in humans for the acquisition of spoken language. Among birds, vocal learning is evolved in songbirds, parrots, and hummingbirds. Each time similar forebrain song nuclei specialized for vocal learning and production have evolved. This finding led to the hypothesis that the behavioral and neuroanatomical convergences for vocal learning could be associated with molecular convergence. We previously found that the neural activity-induced gene dual specificity phosphatase 1 (dusp1) was up-regulated in non-vocal circuits, specifically in sensory-input neurons of the thalamus and telencephalon; however, dusp1 was not up-regulated in higher order sensory neurons or motor circuits. Here we show that song motor nuclei are an exception to this pattern. The song nuclei of species from all known vocal learning avian lineages showed motor-driven up-regulation of dusp1 expression induced by singing. There was no detectable motor-driven dusp1 expression throughout the rest of the forebrain after non-vocal motor performance. This pattern contrasts with expression of the commonly studied activity-induced gene egr1, which shows motor-driven expression in song nuclei induced by singing, but also motor-driven expression in adjacent brain regions after non-vocal motor behaviors. In the vocal non-learning avian species, we found no detectable vocalizing-driven dusp1 expression in the forebrain. These findings suggest that independent evolutions of neural systems for vocal learning were accompanied by selection for specialized motor-driven expression of the dusp1 gene in those circuits. This specialized expression of dusp1 could potentially lead to differential regulation of dusp1-modulated molecular cascades in vocal learning circuits. PMID:22876306

  6. Social learning of vocal structure in a nonhuman primate?

    PubMed Central

    2011-01-01

    Background Non-human primate communication is thought to be fundamentally different from human speech, mainly due to vast differences in vocal control. The lack of these abilities in non-human primates is especially striking if compared to some marine mammals and bird species, which has generated somewhat of an evolutionary conundrum. What are the biological roots and underlying evolutionary pressures of the human ability to voluntarily control sound production and learn the vocal utterances of others? One hypothesis is that this capacity has evolved gradually in humans from an ancestral stage that resembled the vocal behavior of modern primates. Support for this has come from studies that have documented limited vocal flexibility and convergence in different primate species, typically in calls used during social interactions. The mechanisms underlying these patterns, however, are currently unknown. Specifically, it has been difficult to rule out explanations based on genetic relatedness, suggesting that such vocal flexibility may not be the result of social learning. Results To address this point, we compared the degree of acoustic similarity of contact calls in free-ranging Campbell's monkeys as a function of their social bonds and genetic relatedness. We calculated three different indices to compare the similarities between the calls' frequency contours, the duration of grooming interactions and the microsatellite-based genetic relatedness between partners. We found a significantly positive relation between bond strength and acoustic similarity that was independent of genetic relatedness. Conclusion Genetic factors determine the general species-specific call repertoire of a primate species, while social factors can influence the fine structure of some the call types. The finding is in line with the more general hypothesis that human speech has evolved gradually from earlier primate-like vocal communication. PMID:22177339

  7. Prelinguistic Pitch Patterns Expressing "Communication" and "Apprehension"

    ERIC Educational Resources Information Center

    Papaeliou, Christina F.; Trevarthen, Colwyn

    2006-01-01

    This study examined whether pitch patterns of prelinguistic vocalizations could discriminate between social vocalizations, uttered apparently with the intention to communicate, and "private" speech, related to solitary activities as an expression of "thinking". Four healthy ten month old English-speaking infants (2 boys and 2 girls) were…

  8. Vocal Fold Vibration Following Surgical Intervention in Three Vocal Pathologies: A Preliminary Study.

    PubMed

    Chen, Wenli; Woo, Peak; Murry, Thomas

    2017-09-01

    High-speed videoendoscopy captures the cycle-to-cycle vibratory motion of each individual vocal fold in normal and severely disordered phonation. Therefore, it provides a direct method to examine the specific vibratory changes following vocal fold surgery. The purpose of this study was to examine the vocal fold vibratory pattern changes in the surgically treated pathologic vocal fold and the contralateral vocal fold in three vocal pathologies: vocal polyp (n = 3), paresis or paralysis (n = 3), and scar (n = 3). Digital kymography was used to extract high-speed kymographic vocal fold images at the mid-membranous region of the vocal fold. Spectral analysis was subsequently applied to the digital kymography to quantify the cycle-to-cycle movements of each vocal fold, expressed as a spectrum. Surgical modification resulted in significantly improved spectral power of the treated pathologic vocal fold. Furthermore, the contralateral vocal fold also presented with improved spectral power irrespective of vocal pathology. In comparison with normal vocal fold spectrum, postsurgical vocal fold vibrations continued to demonstrate decreased vibratory amplitude in both vocal folds. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  9. How small could a pup sound? The physical bases of signaling body size in harbor seals

    PubMed Central

    Gross, Stephanie; Garcia, Maxime; Rubio-Garcia, Ana; de Boer, Bart

    2017-01-01

    Abstract Vocal communication is a crucial aspect of animal behavior. The mechanism which most mammals use to vocalize relies on three anatomical components. First, air overpressure is generated inside the lower vocal tract. Second, as the airstream goes through the glottis, sound is produced via vocal fold vibration. Third, this sound is further filtered by the geometry and length of the upper vocal tract. Evidence from mammalian anatomy and bioacoustics suggests that some of these three components may covary with an animal’s body size. The framework provided by acoustic allometry suggests that, because vocal tract length (VTL) is more strongly constrained by the growth of the body than vocal fold length (VFL), VTL generates more reliable acoustic cues to an animal’s size. This hypothesis is often tested acoustically but rarely anatomically, especially in pinnipeds. Here, we test the anatomical bases of the acoustic allometry hypothesis in harbor seal pups Phoca vitulina. We dissected and measured vocal tract, vocal folds, and other anatomical features of 15 harbor seals post-mortem. We found that, while VTL correlates with body size, VFL does not. This suggests that, while body growth puts anatomical constraints on how vocalizations are filtered by harbor seals’ vocal tract, no such constraints appear to exist on vocal folds, at least during puppyhood. It is particularly interesting to find anatomical constraints on harbor seals’ vocal tracts, the same anatomical region partially enabling pups to produce individually distinctive vocalizations. PMID:29492005

  10. Individuality and stability in male songs of cao vit gibbons (Nomascus nasutus) with potential to monitor population dynamics.

    PubMed

    Feng, Jun-Juan; Cui, Liang-Wei; Ma, Chang-Yong; Fei, Han-Lan; Fan, Peng-Fei

    2014-01-01

    Vocal individuality and stability has been used to conduct population surveys, monitor population dynamics, and detect dispersal patterns in avian studies. To our knowledge, it has never been used in these kinds of studies among primates. The cao vit gibbon is a critically endangered species with only one small population living in a karst forest along China-Vietnam border. Due to the difficult karst terrain, an international border, long life history, and similarity in male morphology, detailed monitoring of population dynamics and dispersal patterns are not possible using traditional observation methods. In this paper, we test individuality and stability in male songs of cao vit gibbons. We then discuss the possibility of using vocal individuality for population surveys and monitoring population dynamics and dispersal patterns. Significant individuality of vocalization was detected in all 9 males, and the correct rate of individual identification yielded by discriminant function analysis using a subset of variables was satisfactory (>90%). Vocal stability over 2-6 years was also documented in 4 males. Several characters of cao vit gibbons allowed long-term population monitoring using vocal recordings in both China and Vietnam: 1) regular loud calls, 2) strong individuality and stability in male songs, 3) stable territories, and 4) long male tenure. During the course of this research, we also observed one male replacement (confirmed by vocal analysis). This time- and labor-saving method might be the most effective way to detect dispersal patterns in this transboundary population.

  11. Monkeys and Humans Share a Common Computation for Face/Voice Integration

    PubMed Central

    Chandrasekaran, Chandramouli; Lemus, Luis; Trubanova, Andrea; Gondan, Matthias; Ghazanfar, Asif A.

    2011-01-01

    Speech production involves the movement of the mouth and other regions of the face resulting in visual motion cues. These visual cues enhance intelligibility and detection of auditory speech. As such, face-to-face speech is fundamentally a multisensory phenomenon. If speech is fundamentally multisensory, it should be reflected in the evolution of vocal communication: similar behavioral effects should be observed in other primates. Old World monkeys share with humans vocal production biomechanics and communicate face-to-face with vocalizations. It is unknown, however, if they, too, combine faces and voices to enhance their perception of vocalizations. We show that they do: monkeys combine faces and voices in noisy environments to enhance their detection of vocalizations. Their behavior parallels that of humans performing an identical task. We explored what common computational mechanism(s) could explain the pattern of results we observed across species. Standard explanations or models such as the principle of inverse effectiveness and a “race” model failed to account for their behavior patterns. Conversely, a “superposition model”, positing the linear summation of activity patterns in response to visual and auditory components of vocalizations, served as a straightforward but powerful explanatory mechanism for the observed behaviors in both species. As such, it represents a putative homologous mechanism for integrating faces and voices across primates. PMID:21998576

  12. Using Ambulatory Voice Monitoring to Investigate Common Voice Disorders: Research Update

    PubMed Central

    Mehta, Daryush D.; Van Stan, Jarrad H.; Zañartu, Matías; Ghassemi, Marzyeh; Guttag, John V.; Espinoza, Víctor M.; Cortés, Juan P.; Cheyne, Harold A.; Hillman, Robert E.

    2015-01-01

    Many common voice disorders are chronic or recurring conditions that are likely to result from inefficient and/or abusive patterns of vocal behavior, referred to as vocal hyperfunction. The clinical management of hyperfunctional voice disorders would be greatly enhanced by the ability to monitor and quantify detrimental vocal behaviors during an individual’s activities of daily life. This paper provides an update on ongoing work that uses a miniature accelerometer on the neck surface below the larynx to collect a large set of ambulatory data on patients with hyperfunctional voice disorders (before and after treatment) and matched-control subjects. Three types of analysis approaches are being employed in an effort to identify the best set of measures for differentiating among hyperfunctional and normal patterns of vocal behavior: (1) ambulatory measures of voice use that include vocal dose and voice quality correlates, (2) aerodynamic measures based on glottal airflow estimates extracted from the accelerometer signal using subject-specific vocal system models, and (3) classification based on machine learning and pattern recognition approaches that have been used successfully in analyzing long-term recordings of other physiological signals. Preliminary results demonstrate the potential for ambulatory voice monitoring to improve the diagnosis and treatment of common hyperfunctional voice disorders. PMID:26528472

  13. Vocal communication in a complex multi-level society: constrained acoustic structure and flexible call usage in Guinea baboons.

    PubMed

    Maciej, Peter; Ndao, Ibrahima; Hammerschmidt, Kurt; Fischer, Julia

    2013-09-23

    To understand the evolution of acoustic communication in animals, it is important to distinguish between the structure and the usage of vocal signals, since both aspects are subject to different constraints. In terrestrial mammals, the structure of calls is largely innate, while individuals have a greater ability to actively initiate or withhold calls. In closely related taxa, one would therefore predict a higher flexibility in call usage compared to call structure. In the present study, we investigated the vocal repertoire of free living Guinea baboons (Papio papio) and examined the structure and usage of the animals' vocal signals. Guinea baboons live in a complex multi-level social organization and exhibit a largely tolerant and affiliative social style, contrary to most other baboon taxa. To classify the vocal repertoire of male and female Guinea baboons, cluster analyses were used and focal observations were conducted to assess the usage of vocal signals in the particular contexts. In general, the vocal repertoire of Guinea baboons largely corresponded to the vocal repertoire other baboon taxa. The usage of calls, however, differed considerably from other baboon taxa and corresponded with the specific characteristics of the Guinea baboons' social behaviour. While Guinea baboons showed a diminished usage of contest and display vocalizations (a common pattern observed in chacma baboons), they frequently used vocal signals during affiliative and greeting interactions. Our study shows that the call structure of primates is largely unaffected by the species' social system (including grouping patterns and social interactions), while the usage of calls can be more flexibly adjusted, reflecting the quality of social interactions of the individuals. Our results support the view that the primary function of social signals is to regulate social interactions, and therefore the degree of competition and cooperation may be more important to explain variation in call usage than grouping patterns or group size.

  14. Vocal communication in a complex multi-level society: constrained acoustic structure and flexible call usage in Guinea baboons

    PubMed Central

    2013-01-01

    Background To understand the evolution of acoustic communication in animals, it is important to distinguish between the structure and the usage of vocal signals, since both aspects are subject to different constraints. In terrestrial mammals, the structure of calls is largely innate, while individuals have a greater ability to actively initiate or withhold calls. In closely related taxa, one would therefore predict a higher flexibility in call usage compared to call structure. In the present study, we investigated the vocal repertoire of free living Guinea baboons (Papio papio) and examined the structure and usage of the animals’ vocal signals. Guinea baboons live in a complex multi-level social organization and exhibit a largely tolerant and affiliative social style, contrary to most other baboon taxa. To classify the vocal repertoire of male and female Guinea baboons, cluster analyses were used and focal observations were conducted to assess the usage of vocal signals in the particular contexts. Results In general, the vocal repertoire of Guinea baboons largely corresponded to the vocal repertoire other baboon taxa. The usage of calls, however, differed considerably from other baboon taxa and corresponded with the specific characteristics of the Guinea baboons’ social behaviour. While Guinea baboons showed a diminished usage of contest and display vocalizations (a common pattern observed in chacma baboons), they frequently used vocal signals during affiliative and greeting interactions. Conclusions Our study shows that the call structure of primates is largely unaffected by the species’ social system (including grouping patterns and social interactions), while the usage of calls can be more flexibly adjusted, reflecting the quality of social interactions of the individuals. Our results support the view that the primary function of social signals is to regulate social interactions, and therefore the degree of competition and cooperation may be more important to explain variation in call usage than grouping patterns or group size. PMID:24059742

  15. Speech rhythm in Kannada speaking adults who stutter.

    PubMed

    Maruthy, Santosh; Venugopal, Sahana; Parakh, Priyanka

    2017-10-01

    A longstanding hypothesis about the underlying mechanisms of stuttering suggests that speech disfluencies may be associated with problems in timing and temporal patterning of speech events. Fifteen adults who do and do not stutter read five sentences, and from these, the vocalic and consonantal durations were measured. Using these, pairwise variability index (raw PVI for consonantal intervals and normalised PVI for vocalic intervals) and interval based rhythm metrics (PercV, DeltaC, DeltaV, VarcoC and VarcoV) were calculated for all the participants. Findings suggested higher mean values in adults who stutter when compared to adults who do not stutter for all the rhythm metrics except for VarcoV. Further, statistically significant difference between the two groups was found for all the rhythm metrics except for VarcoV. Combining the present results with consistent prior findings based on rhythm deficits in children and adults who stutter, there appears to be strong empirical support for the hypothesis that individuals who stutter may have deficits in generation of rhythmic speech patterns.

  16. Infant-Mother Vocalization Patterns: A Replication and Extension.

    ERIC Educational Resources Information Center

    Kilbourne, Brock K.; Ginsburg, Gerald P.

    This study reports a replication of an earlier study by Kilbourne and Ginsberg (1980) which indicated the occurrence of a transition from predominantly coacting to predominantly alternating infant-mother vocalization patterns. In addition, the present study examined the modulating influences of nursing activity and mother's focus of attention upon…

  17. Towards a computer-aided diagnosis system for vocal cord diseases.

    PubMed

    Verikas, A; Gelzinis, A; Bacauskiene, M; Uloza, V

    2006-01-01

    The objective of this work is to investigate a possibility of creating a computer-aided decision support system for an automated analysis of vocal cord images aiming to categorize diseases of vocal cords. The problem is treated as a pattern recognition task. To obtain a concise and informative representation of a vocal cord image, colour, texture, and geometrical features are used. The representation is further analyzed by a pattern classifier categorizing the image into healthy, diffuse, and nodular classes. The approach developed was tested on 785 vocal cord images collected at the Department of Otolaryngology, Kaunas University of Medicine, Lithuania. A correct classification rate of over 87% was obtained when categorizing a set of unseen images into the aforementioned three classes. Bearing in mind the high similarity of the decision classes, the results obtained are rather encouraging and the developed tools could be very helpful for assuring objective analysis of the images of laryngeal diseases.

  18. Social Vocalizations of Big Brown Bats Vary with Behavioral Context

    PubMed Central

    Gadziola, Marie A.; Grimsley, Jasmine M. S.; Faure, Paul A.; Wenstrup, Jeffrey J.

    2012-01-01

    Bats are among the most gregarious and vocal mammals, with some species demonstrating a diverse repertoire of syllables under a variety of behavioral contexts. Despite extensive characterization of big brown bat (Eptesicus fuscus) biosonar signals, there have been no detailed studies of adult social vocalizations. We recorded and analyzed social vocalizations and associated behaviors of captive big brown bats under four behavioral contexts: low aggression, medium aggression, high aggression, and appeasement. Even limited to these contexts, big brown bats possess a rich repertoire of social vocalizations, with 18 distinct syllable types automatically classified using a spectrogram cross-correlation procedure. For each behavioral context, we describe vocalizations in terms of syllable acoustics, temporal emission patterns, and typical syllable sequences. Emotion-related acoustic cues are evident within the call structure by context-specific syllable types or variations in the temporal emission pattern. We designed a paradigm that could evoke aggressive vocalizations while monitoring heart rate as an objective measure of internal physiological state. Changes in the magnitude and duration of elevated heart rate scaled to the level of evoked aggression, confirming the behavioral state classifications assessed by vocalizations and behavioral displays. These results reveal a complex acoustic communication system among big brown bats in which acoustic cues and call structure signal the emotional state of a caller. PMID:22970247

  19. Communication of emotions in vocal expression and music performance: different channels, same code?

    PubMed

    Juslin, Patrik N; Laukka, Petri

    2003-09-01

    Many authors have speculated about a close relationship between vocal expression of emotions and musical expression of emotions. but evidence bearing on this relationship has unfortunately been lacking. This review of 104 studies of vocal expression and 41 studies of music performance reveals similarities between the 2 channels concerning (a) the accuracy with which discrete emotions were communicated to listeners and (b) the emotion-specific patterns of acoustic cues used to communicate each emotion. The patterns are generally consistent with K. R. Scherer's (1986) theoretical predictions. The results can explain why music is perceived as expressive of emotion, and they are consistent with an evolutionary perspective on vocal expression of emotions. Discussion focuses on theoretical accounts and directions for future research.

  20. Neural Correlates of the Lombard Effect in Primate Auditory Cortex

    PubMed Central

    Eliades, Steven J.

    2012-01-01

    Speaking is a sensory-motor process that involves constant self-monitoring to ensure accurate vocal production. Self-monitoring of vocal feedback allows rapid adjustment to correct perceived differences between intended and produced vocalizations. One important behavior in vocal feedback control is a compensatory increase in vocal intensity in response to noise masking during vocal production, commonly referred to as the Lombard effect. This behavior requires mechanisms for continuously monitoring auditory feedback during speaking. However, the underlying neural mechanisms are poorly understood. Here we show that when marmoset monkeys vocalize in the presence of masking noise that disrupts vocal feedback, the compensatory increase in vocal intensity is accompanied by a shift in auditory cortex activity toward neural response patterns seen during vocalizations under normal feedback condition. Furthermore, we show that neural activity in auditory cortex during a vocalization phrase predicts vocal intensity compensation in subsequent phrases. These observations demonstrate that the auditory cortex participates in self-monitoring during the Lombard effect, and may play a role in the compensation of noise masking during feedback-mediated vocal control. PMID:22855821

  1. Ictal speech and language dysfunction in adult epilepsy: Clinical study of 95 seizures.

    PubMed

    Dussaule, C; Cauquil, C; Flamand-Roze, C; Gagnepain, J-P; Bouilleret, V; Denier, C; Masnou, P

    2017-04-01

    To analyze the semiological characteristics of the language and speech disorders arising during epileptic seizures, and to describe the patterns of language and speech disorders that can predict laterality of the epileptic focus. This study retrospectively analyzed 95 consecutive videos of seizures with language and/or speech disorders in 44 patients admitted for diagnostic video-EEG monitoring. Laterality of the epileptic focus was defined according to electro-clinical correlation studies and structural and functional neuroimaging findings. Language and speech disorders were analyzed by a neurologist and a speech therapist blinded to these data. Language and/or speech disorders were subdivided into eight dynamic patterns: pure anterior aphasia; anterior aphasia and vocal; anterior aphasia and "arthria"; pure posterior aphasia; posterior aphasia and vocal; pure vocal; vocal and arthria; and pure arthria. The epileptic focus was in the left hemisphere in more than 4/5 of seizures presenting with pure anterior aphasia or pure posterior aphasia patterns, while discharges originated in the right hemisphere in almost 2/3 of seizures presenting with a pure vocal pattern. No laterality value was found for the other patterns. Classification of the language and speech disorders arising during epileptic seizures into dynamic patterns may be useful for the optimal analysis of anatomo-electro-clinical correlations. In addition, our research has led to the development of standardized tests for analyses of language and speech disorders arising during seizures that can be conducted during video-EEG sessions. Copyright © 2017 Elsevier Masson SAS. All rights reserved.

  2. Vocal development in a Waddington landscape

    PubMed Central

    Teramoto, Yayoi; Takahashi, Daniel Y; Holmes, Philip; Ghazanfar, Asif A

    2017-01-01

    Vocal development is the adaptive coordination of the vocal apparatus, muscles, the nervous system, and social interaction. Here, we use a quantitative framework based on optimal control theory and Waddington’s landscape metaphor to provide an integrated view of this process. With a biomechanical model of the marmoset monkey vocal apparatus and behavioral developmental data, we show that only the combination of the developing vocal tract, vocal apparatus muscles and nervous system can fully account for the patterns of vocal development. Together, these elements influence the shape of the monkeys’ vocal developmental landscape, tilting, rotating or shifting it in different ways. We can thus use this framework to make quantitative predictions regarding how interfering factors or experimental perturbations can change the landscape within a species, or to explain comparative differences in vocal development across species DOI: http://dx.doi.org/10.7554/eLife.20782.001 PMID:28092262

  3. Difference between the vocalizations of two sister species of pigeons explained in dynamical terms.

    PubMed

    Alonso, R Gogui; Kopuchian, Cecilia; Amador, Ana; Suarez, Maria de Los Angeles; Tubaro, Pablo L; Mindlin, Gabriel B

    2016-05-01

    Vocal communication is an unique example, where the nonlinear nature of the periphery can give rise to complex sounds even when driven by simple neural instructions. In this work we studied the case of two close-related bird species, Patagioenas maculosa and Patagioenas picazuro, whose vocalizations differ only in the timbre. The temporal modulation of the fundamental frequency is similar in both cases, differing only in the existence of sidebands around the fundamental frequency in the P. maculosa. We tested the hypothesis that the qualitative difference between these vocalizations lies in the nonlinear nature of the syrinx. In particular, we propose that the roughness of maculosa's vocalizations is due to an asymmetry between the right and left vibratory membranes, whose nonlinear dynamics generate the sound. To test the hypothesis, we generated a biomechanical model for vocal production with an asymmetric parameter Q with which we can control the level of asymmetry between these membranes. Using this model we generated synthetic vocalizations with the principal acoustic features of both species. In addition, we confirmed the anatomical predictions by making post mortem inspection of the syrinxes, showing that the species with tonal song (picazuro) has a more symmetrical pair of membranes compared to maculosa.

  4. Difference between the vocalizations of two sister species of pigeons explained in dynamical terms

    PubMed Central

    Alonso, R. Gogui; Kopuchian, Cecilia; Amador, Ana; de los Angeles Suarez, Maria; Tubaro, Pablo L.; Mindlin, Gabriel B.

    2016-01-01

    Vocal communication is a unique example where the nonlinear nature of the periphery can give rise to complex sounds even when driven by simple neural instructions. In this work we studied the case of two close-related bird species, Patagioenas maculosa and Patagioenas picazuro, whose vocalizations differ only in the timbre. The temporal modulation of the fundamental frequency is similar in both cases, differing only in the existence of sidebands around the fundamental frequency in the Patagioenas maculosa. We tested the hypothesis that the qualitative difference between these vocalizations lies in the nonlinear nature of the syrinx. In particular, we propose that the roughness of maculosa's vocalizations is due to an asymmetry between the right and left vibratory membranes, whose nonlinear dynamics generate the sound. To test the hypothesis, we generated a biomechanical model for vocal production with an asymmetric parameter Q with which we can control the level of asymmetry between these membranes. Using this model we generated synthetic vocalizations with the principal acoustic features of both species. In addition, we confirmed the anatomical predictions by making post-mortem inspection of the syrinxes, showing that the species with tonal song (picazuro) has a more symmetrical pair of membranes compared to maculosa. PMID:27033354

  5. A circular model for song motor control in Serinus canaria

    PubMed Central

    Alonso, Rodrigo G.; Trevisan, Marcos A.; Amador, Ana; Goller, Franz; Mindlin, Gabriel B.

    2015-01-01

    Song production in songbirds is controlled by a network of nuclei distributed across several brain regions, which drives respiratory and vocal motor systems to generate sound. We built a model for birdsong production, whose variables are the average activities of different neural populations within these nuclei of the song system. We focus on the predictions of respiratory patterns of song, because these can be easily measured and therefore provide a validation for the model. We test the hypothesis that it is possible to construct a model in which (1) the activity of an expiratory related (ER) neural population fits the observed pressure patterns used by canaries during singing, and (2) a higher forebrain neural population, HVC, is sparsely active, simultaneously with significant motor instances of the pressure patterns. We show that in order to achieve these two requirements, the ER neural population needs to receive two inputs: a direct one, and its copy after being processed by other areas of the song system. The model is capable of reproducing the measured respiratory patterns and makes specific predictions on the timing of HVC activity during their production. These results suggest that vocal production is controlled by a circular network rather than by a simple top-down architecture. PMID:25904860

  6. Coding rate and duration of vocalizations of the frog, Xenopus laevis.

    PubMed

    Zornik, Erik; Yamaguchi, Ayako

    2012-08-29

    Vocalizations involve complex rhythmic motor patterns, but the underlying temporal coding mechanisms in the nervous system are poorly understood. Using a recently developed whole-brain preparation from which "fictive" vocalizations are readily elicited in vitro, we investigated the cellular basis of temporal complexity of African clawed frogs (Xenopus laevis). Male advertisement calls contain two alternating components--fast trills (∼300 ms) and slow trills (∼700 ms) that contain clicks repeated at ∼60 and ∼30 Hz, respectively. We found that males can alter the duration of fast trills without changing click rates. This finding led us to hypothesize that call rate and duration are regulated by independent mechanisms. We tested this by obtaining whole-cell patch-clamp recordings in the "fictively" calling isolated brain. We discovered a single type of premotor neuron with activity patterns correlated with both the rate and duration of fast trills. These "fast-trill neurons" (FTNs) exhibited long-lasting depolarizations (LLDs) correlated with each fast trill and action potentials that were phase-locked with motor output-neural correlates of call duration and rate, respectively. When depolarized without central pattern generator activation, FTNs produced subthreshold oscillations and action potentials at fast-trill rates, indicating FTN resonance properties are tuned to, and may dictate, the fast-trill rhythm. NMDA receptor (NMDAR) blockade eliminated LLDs in FTNs, and NMDAR activation in synaptically isolated FTNs induced repetitive LLDs. These results suggest FTNs contain an NMDAR-dependent mechanism that may regulate fast-trill duration. We conclude that a single premotor neuron population employs distinct mechanisms to regulate call rate and duration.

  7. Peripheral Mechanisms for Vocal Production in Birds--Differences and Similarities to Human Speech and Singing

    ERIC Educational Resources Information Center

    Riede, Tobias; Goller, Franz

    2010-01-01

    Song production in songbirds is a model system for studying learned vocal behavior. As in humans, bird phonation involves three main motor systems (respiration, vocal organ and vocal tract). The avian respiratory mechanism uses pressure regulation in air sacs to ventilate a rigid lung. In songbirds sound is generated with two independently…

  8. Development of vocal tract length during early childhood: A magnetic resonance imaging study

    NASA Astrophysics Data System (ADS)

    Vorperian, Houri K.; Kent, Ray D.; Lindstrom, Mary J.; Kalina, Cliff M.; Gentry, Lindell R.; Yandell, Brian S.

    2005-01-01

    Speech development in children is predicated partly on the growth and anatomic restructuring of the vocal tract. This study examines the growth pattern of the various hard and soft tissue vocal tract structures as visualized by magnetic resonance imaging (MRI), and assesses their relational growth with vocal tract length (VTL). Measurements on lip thickness, hard- and soft-palate length, tongue length, naso-oro-pharyngeal length, mandibular length and depth, and distance of the hyoid bone and larynx from the posterior nasal spine were used from 63 pediatric cases (ages birth to 6 years and 9 months) and 12 adults. Results indicate (a) ongoing growth of all oral and pharyngeal vocal tract structures with no sexual dimorphism, and a period of accelerated growth between birth and 18 months; (b) vocal tract structure's region (oral/anterior versus pharyngeal/posterior) and orientation (horizontal versus vertical) determine its growth pattern; and (c) the relational growth of the different structures with VTL changes with development-while the increase in VTL throughout development is predominantly due to growth of pharyngeal/posterior structures, VTL is also substantially affected by the growth of oral/anterior structures during the first 18 months of life. Findings provide normative data that can be used for modeling the development of the vocal tract. .

  9. Simulations of temporal patterns of oral airflow in men and women using a two-mass model of the vocal folds under dynamic control

    NASA Astrophysics Data System (ADS)

    Lucero, Jorge C.; Koenig, Laura L.

    2005-03-01

    In this study we use a low-dimensional laryngeal model to reproduce temporal variations in oral airflow produced by speakers in the vicinity of an abduction gesture. It attempts to characterize these temporal patterns in terms of biomechanical parameters such as glottal area, vocal fold stiffness, subglottal pressure, and gender differences in laryngeal dimensions. A two-mass model of the vocal folds coupled to a two-tube approximation of the vocal tract is fitted to oral airflow records measured in men and women during the production of /aha/ utterances, using the subglottal pressure, glottal width, and Q factor as control parameters. The results show that the model is capable of reproducing the airflow records with good approximation. A nonlinear damping characteristics is needed, to reproduce the flow variation at glottal abduction. Devoicing is achieved by the combined action of vocal fold abduction, the decrease of subglottal pressure, and the increase of vocal fold tension. In general, the female larynx has a more restricted region of vocal fold oscillation than the male one. This would explain the more frequent devoicing in glottal abduction-adduction gestures for /h/ in running speech by women, compared to men. .

  10. Say It to Play It.

    ERIC Educational Resources Information Center

    Jarvis, William C.

    1980-01-01

    Author discusses the importance of vocalization in the development of basic musicianship. He cites studies demonstrating that vocal teaching strategies, such as singing tonal patterns, aids music reading, memory, and instrumental performance. (SJL)

  11. Can Birds Perceive Rhythmic Patterns? A Review and Experiments on a Songbird and a Parrot Species

    PubMed Central

    ten Cate, Carel; Spierings, Michelle; Hubert, Jeroen; Honing, Henkjan

    2016-01-01

    While humans can easily entrain their behavior with the beat in music, this ability is rare among animals. Yet, comparative studies in non-human species are needed if we want to understand how and why this ability evolved. Entrainment requires two abilities: (1) recognizing the regularity in the auditory stimulus and (2) the ability to adjust the own motor output to the perceived pattern. It has been suggested that beat perception and entrainment are linked to the ability for vocal learning. The presence of some bird species showing beat induction, and also the existence of vocal learning as well as vocal non-learning bird taxa, make them relevant models for comparative research on rhythm perception and its link to vocal learning. Also, some bird vocalizations show strong regularity in rhythmic structure, suggesting that birds might perceive rhythmic structures. In this paper we review the available experimental evidence for the perception of regularity and rhythms by birds, like the ability to distinguish regular from irregular stimuli over tempo transformations and report data from new experiments. While some species show a limited ability to detect regularity, most evidence suggests that birds attend primarily to absolute and not relative timing of patterns and to local features of stimuli. We conclude that, apart from some large parrot species, there is limited evidence for beat and regularity perception among birds and that the link to vocal learning is unclear. We next report the new experiments in which zebra finches and budgerigars (both vocal learners) were first trained to distinguish a regular from an irregular pattern of beats and then tested on various tempo transformations of these stimuli. The results showed that both species reduced the discrimination after tempo transformations. This suggests that, as was found in earlier studies, they attended mainly to local temporal features of the stimuli, and not to their overall regularity. However, some individuals of both species showed an additional sensitivity to the more global pattern if some local features were left unchanged. Altogether our study indicates both between and within species variation, in which birds attend to a mixture of local and to global rhythmic features. PMID:27242635

  12. Your attention please: increasing ambient noise levels elicits a change in communication behaviour in humpback whales (Megaptera novaeangliae)

    PubMed Central

    Dunlop, Rebecca A.; Cato, Douglas H.; Noad, Michael J.

    2010-01-01

    High background noise is an important obstacle in successful signal detection and perception of an intended acoustic signal. To overcome this problem, many animals modify their acoustic signal by increasing the repetition rate, duration, amplitude or frequency range of the signal. An alternative method to ensure successful signal reception, yet to be tested in animals, involves the use of two different types of signal, where one signal type may enhance the other in periods of high background noise. Humpback whale communication signals comprise two different types: vocal signals, and surface-generated signals such as ‘breaching’ or ‘pectoral slapping’. We found that humpback whales gradually switched from primarily vocal to primarily surface-generated communication in increasing wind speeds and background noise levels, though kept both signal types in their repertoire. Vocal signals have the advantage of having higher information content but may have the disadvantage of loosing this information in a noisy environment. Surface-generated sounds have energy distributed over a greater frequency range and may be less likely to become confused in periods of high wind-generated noise but have less information content when compared with vocal sounds. Therefore, surface-generated sounds may improve detection or enhance the perception of vocal signals in a noisy environment. PMID:20392731

  13. Your attention please: increasing ambient noise levels elicits a change in communication behaviour in humpback whales (Megaptera novaeangliae).

    PubMed

    Dunlop, Rebecca A; Cato, Douglas H; Noad, Michael J

    2010-08-22

    High background noise is an important obstacle in successful signal detection and perception of an intended acoustic signal. To overcome this problem, many animals modify their acoustic signal by increasing the repetition rate, duration, amplitude or frequency range of the signal. An alternative method to ensure successful signal reception, yet to be tested in animals, involves the use of two different types of signal, where one signal type may enhance the other in periods of high background noise. Humpback whale communication signals comprise two different types: vocal signals, and surface-generated signals such as 'breaching' or 'pectoral slapping'. We found that humpback whales gradually switched from primarily vocal to primarily surface-generated communication in increasing wind speeds and background noise levels, though kept both signal types in their repertoire. Vocal signals have the advantage of having higher information content but may have the disadvantage of loosing this information in a noisy environment. Surface-generated sounds have energy distributed over a greater frequency range and may be less likely to become confused in periods of high wind-generated noise but have less information content when compared with vocal sounds. Therefore, surface-generated sounds may improve detection or enhance the perception of vocal signals in a noisy environment.

  14. The value of vocalizing: Five-month-old infants associate their own noncry vocalizations with responses from caregivers

    PubMed Central

    Goldstein, Michael H.; Schwade, Jennifer A.; Bornstein, Marc H.

    2014-01-01

    The early noncry vocalizations of infants are salient social signals. Caregivers spontaneously respond to 30-50% of these sounds, and their responsiveness to infants' prelinguistic noncry vocalizations facilitates the development of phonology and speech. Have infants learned that their vocalizations influence the behavior of social partners? If infants have learned the contingency between their vocalizing and the social responses of others, they should show an extinction burst when the contingency is removed, increasing their rate of noncry vocalizing then decreasing. Thirty-eight 5-month-olds were tested in the still-face paradigm, during which they engaged in a 2-min still-face interaction with an unfamiliar adult. When the adult assumed a still face, infants showed an extinction burst. This pattern of infant vocalizations suggests that 5-month-olds have learned the social efficacy of their vocalizations on caregivers' behavior. Furthermore, the magnitude of 5-month infants' extinction bursts predicted their language comprehension at 13 months. PMID:19489893

  15. Neural correlates of frog calling: production by two semi-independent generators.

    PubMed

    Schmidt, R S

    1992-09-28

    The anterior preoptic nuclei of the isolated brainstem of male, Northern leopard frogs (Rana p. pipiens) were stimulated electrically and neural correlates of mating calling recorded from the rhombencephalic mating calling pattern generator. Lesions of discrete areas of the brainstem showed that the mating calling generator is separable into two generators, the pretrigeminal nucleus and the classical pulmonary respiration generator (which is approximately co-extensive with the motor nuclei IX-X). Each of these still can produce pulses when isolated from the other. Their interaction changes the expiratory phase of breathing into the vocal phase of calling. All stages of intermediates between these phases could be seen. An updated and simplified model of call production and evolution is presented.

  16. TauG-guidance of transients in expressive musical performance.

    PubMed

    Schogler, Benjaman; Pepping, Gert-Jan; Lee, David N

    2008-08-01

    The sounds in expressive musical performance, and the movements that produce them, offer insight into temporal patterns in the brain that generate expression. To gain understanding of these brain patterns, we analyzed two types of transient sounds, and the movements that produced them, during a vocal duet and a bass solo. The transient sounds studied were inter-tone f (0)(t)-glides (the continuous change in fundamental frequency, f (0)(t), when gliding from one tone to the next), and attack intensity-glides (the continuous rise in sound intensity when attacking, or initiating, a tone). The temporal patterns of the inter-tone f (0)(t)-glides and attack intensity-glides, and of the movements producing them, all conformed to the mathematical function, tau (G)(t) (called tauG), predicted by General Tau Theory, and assumed to be generated in the brain. The values of the parameters of the tau (G)(t) function were modulated by the performers when they modulated musical expression. Thus the tau (G)(t) function appears to be a fundamental of brain activity entailed in the generation of expressive temporal patterns of movement and sound.

  17. Sound production by singing humpback whales.

    PubMed

    Mercado, Eduardo; Schneider, Jennifer N; Pack, Adam A; Herman, Louis M

    2010-04-01

    Sounds from humpback whale songs were analyzed to evaluate possible mechanisms of sound production. Song sounds fell along a continuum with trains of discrete pulses at one end and continuous tonal signals at the other. This graded vocal repertoire is comparable to that seen in false killer whales [Murray et al. (1998). J. Acoust. Soc. Am. 104, 1679-1688] and human singers, indicating that all three species generate sounds by varying the tension of pneumatically driven, vibrating membranes. Patterns in the spectral content of sounds and in nonlinear sound features show that resonating air chambers may also contribute to humpback whale sound production. Collectively, these findings suggest that categorizing individual units within songs into discrete types may obscure how singers modulate song features and illustrate how production-based characterizations of vocalizations can provide new insights into how humpback whales sing.

  18. Auditory evoked fields to vocalization during passive listening and active generation in adults who stutter.

    PubMed

    Beal, Deryk S; Cheyne, Douglas O; Gracco, Vincent L; Quraan, Maher A; Taylor, Margot J; De Nil, Luc F

    2010-10-01

    We used magnetoencephalography to investigate auditory evoked responses to speech vocalizations and non-speech tones in adults who do and do not stutter. Neuromagnetic field patterns were recorded as participants listened to a 1 kHz tone, playback of their own productions of the vowel /i/ and vowel-initial words, and actively generated the vowel /i/ and vowel-initial words. Activation of the auditory cortex at approximately 50 and 100 ms was observed during all tasks. A reduction in the peak amplitudes of the M50 and M100 components was observed during the active generation versus passive listening tasks dependent on the stimuli. Adults who stutter did not differ in the amount of speech-induced auditory suppression relative to fluent speakers. Adults who stutter had shorter M100 latencies for the actively generated speaking tasks in the right hemisphere relative to the left hemisphere but the fluent speakers showed similar latencies across hemispheres. During passive listening tasks, adults who stutter had longer M50 and M100 latencies than fluent speakers. The results suggest that there are timing, rather than amplitude, differences in auditory processing during speech in adults who stutter and are discussed in relation to hypotheses of auditory-motor integration breakdown in stuttering. Copyright 2010 Elsevier Inc. All rights reserved.

  19. Rodent ultrasonic vocalizations are bound to active sniffing behavior

    PubMed Central

    Sirotin, Yevgeniy B.; Costa, Martín Elias; Laplagne, Diego A.

    2014-01-01

    During rodent active behavior, multiple orofacial sensorimotor behaviors, including sniffing and whisking, display rhythmicity in the theta range (~5–10 Hz). During specific behaviors, these rhythmic patterns interlock, such that execution of individual motor programs becomes dependent on the state of the others. Here we performed simultaneous recordings of the respiratory cycle and ultrasonic vocalization emission by adult rats and mice in social settings. We used automated analysis to examine the relationship between breathing patterns and vocalization over long time periods. Rat ultrasonic vocalizations (USVs, “50 kHz”) were emitted within stretches of active sniffing (5–10 Hz) and were largely absent during periods of passive breathing (1–4 Hz). Because ultrasound was tightly linked to the exhalation phase, the sniffing cycle segmented vocal production into discrete calls and imposed its theta rhythmicity on their timing. In turn, calls briefly prolonged exhalations, causing an immediate drop in sniffing rate. Similar results were obtained in mice. Our results show that ultrasonic vocalizations are an integral part of the rhythmic orofacial behavioral ensemble. This complex behavioral program is thus involved not only in active sensing but also in the temporal structuring of social communication signals. Many other social signals of mammals, including monkey calls and human speech, show structure in the theta range. Our work points to a mechanism for such structuring in rodent ultrasonic vocalizations. PMID:25477796

  20. Catecholaminergic connectivity to the inner ear, central auditory and vocal motor circuitry in the plainfin midshipman fish, Porichthys notatus

    PubMed Central

    Forlano, Paul M.; Kim, Spencer D.; Krzyminska, Zuzanna M.; Sisneros, Joseph A.

    2014-01-01

    Although the neuroanatomical distribution of catecholaminergic (CA) neurons has been well documented across all vertebrate classes, few studies have examined CA connectivity to physiologically and anatomically identified neural circuitry that controls behavior. The goal of this study was to characterize CA distribution in the brain and inner ear of the plainfin midshipman fish (Porichthys notatus) with particular emphasis on their relationship with anatomically labeled circuitry that both produces and encodes social acoustic signals in this species. Neurobiotin labeling of the main auditory endorgan, the saccule, combined with tyrosine hydroxylase immunofluorescence (TH-ir) revealed a strong CA innervation of both the peripheral and central auditory system. Diencephalic TH-ir neurons in the periventricular posterior tuberculum, known to be dopaminergic, send ascending projections to the ventral telencephalon and prominent descending projections to vocal-acoustic integration sites, notably the hindbrain octavolateralis efferent nucleus, as well as onto the base of hair cells in the saccule via nerve VIII. Neurobiotin backfills of the vocal nerve in combination with TH-ir revealed CA terminals on all components of the vocal pattern generator which appears to largely originate from local TH-ir neurons but may include diencephalic projections as well. This study provides strong evidence for catecholamines as important neuromodulators of both auditory and vocal circuitry and acoustic-driven social behavior in midshipman fish. This first demonstration of TH-ir terminals in the main endorgan of hearing in a non-mammalian vertebrate suggests a conserved and important anatomical and functional role for dopamine in normal audition. PMID:24715479

  1. A cross-cultural comparison of tonal synchrony and pitch imitation in the vocal dialogs of Belgian Flemish-speaking and Mexican Spanish-speaking mother-infant dyads.

    PubMed

    Van Puyvelde, Martine; Loots, Gerrit; Gillisjans, Lobcke; Pattyn, Nathalie; Quintana, Carmen

    2015-08-01

    This study reports a cross-cultural comparison of the vocal pitch patterns of 15 Mexican Spanish-speaking and 15 Belgian Flemish-speaking dyads, recorded during 5min of free-play in a laboratory setting. Both cultures have a tradition of dyadic face-to-face interaction but differ in language origins (i.e., Romanic versus Germanic). In total, 374 Mexican and 558 Flemish vocal exchanges were identified, analyzed and compared for their incidence of tonal synchrony (harmonic/pentatonic series), non-tonal synchrony (with/without imitations) and pitch and/or interval imitations. The main findings revealed that dyads in both cultures rely on tonal synchrony using similar pitch ratios and timing patterns. However, there were significant differences in the infants' vocal pitch imitation behavior. Additional video-analyzes on the contingency patterns involved in pitch imitation showed a cross-cultural difference in the maternal selective reinforcement of pitch imitation. The results are interpreted with regard to linguistic, developmental and cultural aspects and the 'musilanguage' model. Copyright © 2015 Elsevier Inc. All rights reserved.

  2. Influence of Embedded Fibers and an Epithelium Layer on the Glottal Closure Pattern in a Physical Vocal Fold Model

    ERIC Educational Resources Information Center

    Xuan, Yue; Zhang, Zhaoyan

    2014-01-01

    Purpose: The purpose of this study was to explore the possible structural and material property features that may facilitate complete glottal closure in an otherwise isotropic physical vocal fold model. Method: Seven vocal fold models with different structural features were used in this study. An isotropic model was used as the baseline model, and…

  3. The voice conveys specific emotions: evidence from vocal burst displays.

    PubMed

    Simon-Thomas, Emiliana R; Keltner, Dacher J; Sauter, Disa; Sinicropi-Yao, Lara; Abramson, Anna

    2009-12-01

    Studies of emotion signaling inform claims about the taxonomic structure, evolutionary origins, and physiological correlates of emotions. Emotion vocalization research has tended to focus on a limited set of emotions: anger, disgust, fear, sadness, surprise, happiness, and for the voice, also tenderness. Here, we examine how well brief vocal bursts can communicate 22 different emotions: 9 negative (Study 1) and 13 positive (Study 2), and whether prototypical vocal bursts convey emotions more reliably than heterogeneous vocal bursts (Study 3). Results show that vocal bursts communicate emotions like anger, fear, and sadness, as well as seldom-studied states like awe, compassion, interest, and embarrassment. Ancillary analyses reveal family-wise patterns of vocal burst expression. Errors in classification were more common within emotion families (e.g., 'self-conscious,' 'pro-social') than between emotion families. The three studies reported highlight the voice as a rich modality for emotion display that can inform fundamental constructs about emotion.

  4. Spatio-temporal analysis of irregular vocal fold oscillations: Biphonation due to desynchronization of spatial modes

    NASA Astrophysics Data System (ADS)

    Neubauer, Jürgen; Mergell, Patrick; Eysholdt, Ulrich; Herzel, Hanspeter

    2001-12-01

    This report is on direct observation and modal analysis of irregular spatio-temporal vibration patterns of vocal fold pathologies in vivo. The observed oscillation patterns are described quantitatively with multiline kymograms, spectral analysis, and spatio-temporal plots. The complex spatio-temporal vibration patterns are decomposed by empirical orthogonal functions into independent vibratory modes. It is shown quantitatively that biphonation can be induced either by left-right asymmetry or by desynchronized anterior-posterior vibratory modes, and the term ``AP (anterior-posterior) biphonation'' is introduced. The presented phonation examples show that for normal phonation the first two modes sufficiently explain the glottal dynamics. The spatio-temporal oscillation pattern associated with biphonation due to left-right asymmetry can be explained by the first three modes. Higher-order modes are required to describe the pattern for biphonation induced by anterior-posterior vibrations. Spatial irregularity is quantified by an entropy measure, which is significantly higher for irregular phonation than for normal phonation. Two asymmetry measures are introduced: the left-right asymmetry and the anterior-posterior asymmetry, as the ratios of the fundamental frequencies of left and right vocal fold and of anterior-posterior modes, respectively. These quantities clearly differentiate between left-right biphonation and anterior-posterior biphonation. This paper proposes methods to analyze quantitatively irregular vocal fold contour patterns in vivo and complements previous findings of desynchronization of vibration modes in computer modes and in in vitro experiments.

  5. Maximum imaging depth comparison in porcine vocal folds using 776-nm vs. 1552-nm excitation wavelengths

    NASA Astrophysics Data System (ADS)

    Yildirim, Murat; Ferhanoglu, Onur; Kobler, James B.; Zeitels, Steven M.; Ben-Yakar, Adela

    2013-02-01

    Vocal fold scarring is one of the major causes of voice disorders and may arise from overuse or post-surgical wound healing. One promising treatment utilizes the injection of soft biomaterials aimed at restoring viscoelasticity of the outermost vibratory layer of the vocal fold, superficial lamina propria (SLP). However, the density of the tissue and the required injection pressure impair proper localization of the injected biomaterial in SLP. To enhance treatment effectiveness, we are investigating a technique to image and ablate sub-epithelial planar voids in vocal folds using ultrafast laser pulses to better localize the injected biomaterial. It is challenging to optimize the excitation wavelength to perform imaging and ablation at depths suitable for clinical use. Here, we compare maximum imaging depth using two photon autofluorescence and second harmonic generation with third-harmonic generation imaging modalities for healthy porcine vocal folds. We used a home-built inverted nonlinear scanning microscope together with a high repetition rate (2 MHz) ultrafast fiber laser (Raydiance Inc.). We acquired both two-photon autofluorescence and second harmonic generation signals using 776 nm wavelength and third harmonic generation signals using 1552 nm excitation wavelength. We observed that maximum imaging depth with 776 nm wavelength is significantly improved from 114 μm to 205 μm when third harmonic generation is employed using 1552 nm wavelength, without any observable damage in the tissue.

  6. Contextual influences on children's use of vocal affect cues during referential interpretation.

    PubMed

    Berman, Jared M J; Graham, Susan A; Chambers, Craig G

    2013-01-01

    In three experiments, we investigated 5-year-olds' sensitivity to speaker vocal affect during referential interpretation in cases where the indeterminacy is or is not resolved by speech information. In Experiment 1, analyses of eye gaze patterns and pointing behaviours indicated that 5-year-olds used vocal affect cues at the point where an ambiguous description was encountered. In Experiments 2 and 3, we used unambiguous situations to investigate how the referential context influences the ability to use affect cues earlier in the utterance. Here, we found a differential use of speaker vocal affect whereby 5-year-olds' referential hypotheses were influenced by negative vocal affect cues in advance of the noun, but not by positive affect cues. Together, our findings reveal how 5-year-olds use a speaker's vocal affect to identify potential referents in different contextual situations and also suggest that children may be more attuned to negative vocal affect than positive vocal affect, particularly early in an utterance.

  7. Further evaluation of methods to identify matched stimulation.

    PubMed

    Rapp, John T

    2007-01-01

    The effects of preferred stimulation on the vocal stereotypy of 2 individuals were evaluated in two experiments. The results of Experiment 1 showed that (a) the vocal stereotypy of both participants persisted in the absence of social consequences, (b) 1 participant manipulated toys that did and did not produce auditory stimulation, but only sound-producing toys decreased his vocal stereotypy, and (c) only noncontingent music decreased vocal stereotypy for the other participant, but sterotypy paradoxically increased when toys were presented with music. Using a three-component multiple schedule, the results of Experiment 2 showed that the vocal stereotypy of both participants remained below preintervention levels following the removal of auditory stimulation and that 1 participant's vocal stereotypy increased following the removal of contingent reprimands. These patterns suggest that auditory stimulation functioned as an abolishing operation for vocal stereotypy and reprimands functioned as an establishing operation for vocal stereotypy. Together, the two experiments provide a method for identifying alternative stimulation that may substitute for automatically reinforced behavior.

  8. Vocal responses of austral forest frogs to amplitude and degradation patterns of advertisement calls.

    PubMed

    Penna, Mario; Moreno-Gómez, Felipe N; Muñoz, Matías I; Cisternas, Javiera

    2017-07-01

    Degradation phenomena affecting animal acoustic signals may provide cues to assess the distance of emitters. Recognition of degraded signals has been extensively demonstrated in birds, and recently studies have also reported detection of degraded patterns in anurans that call at or above ground level. In the current study we explore the vocal responses of the syntopic burrowing male frogs Eupsophus emiliopugini and E. calcaratus from the South American temperate forest to synthetic conspecific calls differing in amplitude and emulating degraded and non-degraded signal patterns. The results show a strong dependence of vocal responses on signal amplitude, and a general lack of differential responses to signals with different pulse amplitude modulation depths in E. emiliopugini and no effect of relative amplitude of harmonics in E. calcaratus. Such limited discrimination of signal degradation patterns from non-degraded signals is likely related to the burrowing habits of these species. Shelters amplify outgoing and incoming conspecific vocalizations, but do not counteract signal degradation to an extent comparable to calling strategies used by other frogs. The limited detection abilities and resultant response permissiveness to degraded calls in these syntopic burrowing species would be advantageous for animals communicating in circumstances in which signal alteration prevails. Copyright © 2017 Elsevier B.V. All rights reserved.

  9. Two genetic loci control syllable sequences of ultrasonic courtship vocalizations in inbred mice

    PubMed Central

    2011-01-01

    Background The ultrasonic vocalizations (USV) of courting male mice are known to possess a phonetic structure with a complex combination of several syllables. The genetic mechanisms underlying the syllable sequence organization were investigated. Results This study compared syllable sequence organization in two inbred strains of mice, 129S4/SvJae (129) and C57BL6J (B6), and demonstrated that they possessed two mutually exclusive phenotypes. The 129S4/SvJae (129) strain frequently exhibited a "chevron-wave" USV pattern, which was characterized by the repetition of chevron-type syllables. The C57BL/6J strain produced a "staccato" USV pattern, which was characterized by the repetition of short-type syllables. An F1 strain obtained by crossing the 129S4/SvJae and C57BL/6J strains produced only the staccato phenotype. The chevron-wave and staccato phenotypes reappeared in the F2 generations, following the Mendelian law of independent assortment. Conclusions These results suggest that two genetic loci control the organization of syllable sequences. These loci were occupied by the staccato and chevron-wave alleles in the B6 and 129 mouse strains, respectively. Recombination of these alleles might lead to the diversity of USV patterns produced by mice. PMID:22018021

  10. Wild chimpanzees' use of single and combined vocal and gestural signals.

    PubMed

    Hobaiter, C; Byrne, R W; Zuberbühler, K

    2017-01-01

    We describe the individual and combined use of vocalizations and gestures in wild chimpanzees. The rate of gesturing peaked in infancy and, with the exception of the alpha male, decreased again in older age groups, while vocal signals showed the opposite pattern. Although gesture-vocal combinations were relatively rare, they were consistently found in all age groups, especially during affiliative and agonistic interactions. Within behavioural contexts rank (excluding alpha-rank) had no effect on the rate of male chimpanzees' use of vocal or gestural signals and only a small effect on their use of combination signals. The alpha male was an outlier, however, both as a prolific user of gestures and recipient of high levels of vocal and gesture-vocal signals. Persistence in signal use varied with signal type: chimpanzees persisted in use of gestures and gesture-vocal combinations after failure, but where their vocal signals failed they tended to add gestural signals to produce gesture-vocal combinations. Overall, chimpanzees employed signals with a sensitivity to the public/private nature of information, by adjusting their use of signal types according to social context and by taking into account potential out-of-sight audiences. We discuss these findings in relation to the various socio-ecological challenges that chimpanzees are exposed to in their natural forest habitats and the current discussion of multimodal communication in great apes. All animal communication combines different types of signals, including vocalizations, facial expressions, and gestures. However, the study of primate communication has typically focused on the use of signal types in isolation. As a result, we know little on how primates use the full repertoire of signals available to them. Here we present a systematic study on the individual and combined use of gestures and vocalizations in wild chimpanzees. We find that gesturing peaks in infancy and decreases in older age, while vocal signals show the opposite distribution, and patterns of persistence after failure suggest that gestural and vocal signals may encode different types of information. Overall, chimpanzees employed signals with a sensitivity to the public/private nature of information, by adjusting their use of signal types according to social context and by taking into account potential out-of-sight audiences.

  11. Statistical learning in songbirds: from self-tutoring to song culture.

    PubMed

    Fehér, Olga; Ljubičić, Iva; Suzuki, Kenta; Okanoya, Kazuo; Tchernichovski, Ofer

    2017-01-05

    At the onset of vocal development, both songbirds and humans produce variable vocal babbling with broadly distributed acoustic features. Over development, these vocalizations differentiate into the well-defined, categorical signals that characterize adult vocal behaviour. A broadly distributed signal is ideal for vocal exploration, that is, for matching vocal production to the statistics of the sensory input. The developmental transition to categorical signals is a gradual process during which the vocal output becomes differentiated and stable. But does it require categorical input? We trained juvenile zebra finches with playbacks of their own developing song, produced just a few moments earlier, updated continuously over development. Although the vocalizations of these self-tutored (ST) birds were initially broadly distributed, birds quickly developed categorical signals, as fast as birds that were trained with a categorical, adult song template. By contrast, siblings of those birds that received no training (isolates) developed phonological categories much more slowly and never reached the same level of category differentiation as their ST brothers. Therefore, instead of simply mirroring the statistical properties of their sensory input, songbirds actively transform it into distinct categories. We suggest that the early self-generation of phonological categories facilitates the establishment of vocal culture by making the song easier to transmit at the micro level, while promoting stability of shared vocabulary at the group level over generations.This article is part of the themed issue 'New frontiers for statistical learning in the cognitive sciences'. © 2016 The Authors.

  12. Coos, booms, and hoots: The evolution of closed-mouth vocal behavior in birds.

    PubMed

    Riede, Tobias; Eliason, Chad M; Miller, Edward H; Goller, Franz; Clarke, Julia A

    2016-08-01

    Most birds vocalize with an open beak, but vocalization with a closed beak into an inflating cavity occurs in territorial or courtship displays in disparate species throughout birds. Closed-mouth vocalizations generate resonance conditions that favor low-frequency sounds. By contrast, open-mouth vocalizations cover a wider frequency range. Here we describe closed-mouth vocalizations of birds from functional and morphological perspectives and assess the distribution of closed-mouth vocalizations in birds and related outgroups. Ancestral-state optimizations of body size and vocal behavior indicate that closed-mouth vocalizations are unlikely to be ancestral in birds and have evolved independently at least 16 times within Aves, predominantly in large-bodied lineages. Closed-mouth vocalizations are rare in the small-bodied passerines. In light of these results and body size trends in nonavian dinosaurs, we suggest that the capacity for closed-mouth vocalization was present in at least some extinct nonavian dinosaurs. As in birds, this behavior may have been limited to sexually selected vocal displays, and hence would have co-occurred with open-mouthed vocalizations. © 2016 The Author(s). Evolution © 2016 The Society for the Study of Evolution.

  13. A Kinect-Based Sign Language Hand Gesture Recognition System for Hearing- and Speech-Impaired: A Pilot Study of Pakistani Sign Language.

    PubMed

    Halim, Zahid; Abbas, Ghulam

    2015-01-01

    Sign language provides hearing and speech impaired individuals with an interface to communicate with other members of the society. Unfortunately, sign language is not understood by most of the common people. For this, a gadget based on image processing and pattern recognition can provide with a vital aid for detecting and translating sign language into a vocal language. This work presents a system for detecting and understanding the sign language gestures by a custom built software tool and later translating the gesture into a vocal language. For the purpose of recognizing a particular gesture, the system employs a Dynamic Time Warping (DTW) algorithm and an off-the-shelf software tool is employed for vocal language generation. Microsoft(®) Kinect is the primary tool used to capture video stream of a user. The proposed method is capable of successfully detecting gestures stored in the dictionary with an accuracy of 91%. The proposed system has the ability to define and add custom made gestures. Based on an experiment in which 10 individuals with impairments used the system to communicate with 5 people with no disability, 87% agreed that the system was useful.

  14. Variability of normal vocal fold dynamics for different vocal loading in one healthy subject investigated by phonovibrograms.

    PubMed

    Doellinger, Michael; Lohscheller, Joerg; McWhorter, Andrew; Kunduk, Melda

    2009-03-01

    We investigate the potential of high-speed digital imaging technique (HSI) and the phonovibrogram (PVG) analysis in normal vocal fold dynamics by studying the effects of continuous voice use (vocal loading) during the workday. One healthy subject was recorded at sustained phonation 13 times within 2 consecutive days in the morning before and in the afternoon after vocal loading, respectively. Vocal fold dynamics were extracted and visualized by PVGs. The characteristic PVG patterns were extracted representing vocal fold vibration types. The parameter values were then analyzed by statistics regarding vocal load, left-right PVG asymmetries, anterior-posterior PVG asymmetries, and opening-closing differences. For the first time, the direct impact of vocal load could be determined by analyzing vocal fold dynamics. For same vocal loading conditions, equal dynamical behavior of the vocal folds were confirmed. Comparison of recordings performed in the morning with the recordings after work revealed significant changes in vibration behavior, indicating impact of occurring vocal load. Left-right asymmetries in vocal fold dynamics were found confirming earlier assumptions. Different dynamics between opening and closing procedure as well as for anterior and posterior parts were found. Constant voice usage stresses the vocal folds even in healthy subjects and can be detected by applying the PVG technique. Furthermore, left-right PVG asymmetries do occur in healthy voice to a certain extent. HSI in combination with PVG analysis seems to be a promising tool for investigation of vocal fold fatigue and pathologies resulting in small forms of dynamical changes.

  15. Automated analysis of connected speech reveals early biomarkers of Parkinson's disease in patients with rapid eye movement sleep behaviour disorder.

    PubMed

    Hlavnička, Jan; Čmejla, Roman; Tykalová, Tereza; Šonka, Karel; Růžička, Evžen; Rusz, Jan

    2017-02-02

    For generations, the evaluation of speech abnormalities in neurodegenerative disorders such as Parkinson's disease (PD) has been limited to perceptual tests or user-controlled laboratory analysis based upon rather small samples of human vocalizations. Our study introduces a fully automated method that yields significant features related to respiratory deficits, dysphonia, imprecise articulation and dysrhythmia from acoustic microphone data of natural connected speech for predicting early and distinctive patterns of neurodegeneration. We compared speech recordings of 50 subjects with rapid eye movement sleep behaviour disorder (RBD), 30 newly diagnosed, untreated PD patients and 50 healthy controls, and showed that subliminal parkinsonian speech deficits can be reliably captured even in RBD patients, which are at high risk of developing PD or other synucleinopathies. Thus, automated vocal analysis should soon be able to contribute to screening and diagnostic procedures for prodromal parkinsonian neurodegeneration in natural environments.

  16. The effect of music on repetitive disruptive vocalizations of persons with dementia.

    PubMed

    Casby, J A; Holm, M B

    1994-10-01

    This study examined the effect of classical music and favorite music on the repetitive disruptive vocalizations of long-term-care facility (LTCF) residents with dementia of the Alzheimer's type (DAT). Three subjects diagnosed with DAT who had a history of repetitive disruptive vocalizations were selected for the study. Three single-subject withdrawal designs (ABA, ACA, and ABCA) were used to assess subjects' repetitive disruptive vocalizations during each phase: no intervention (A); relaxing, classical music (B); and favorite music (C). Classical music and favorite music significantly decreased the number of vocalizations in two of the three subjects (p < .05). These findings support a method that was effective in decreasing the disruptive vocalization pattern common in those with DAT in the least restrictive manner, as mandated by the Omnibus Budget Reconciliation Act of 1987.

  17. Human Exploration of Enclosed Spaces through Echolocation.

    PubMed

    Flanagin, Virginia L; Schörnich, Sven; Schranner, Michael; Hummel, Nadine; Wallmeier, Ludwig; Wahlberg, Magnus; Stephan, Thomas; Wiegrebe, Lutz

    2017-02-08

    Some blind humans have developed echolocation, as a method of navigation in space. Echolocation is a truly active sense because subjects analyze echoes of dedicated, self-generated sounds to assess space around them. Using a special virtual space technique, we assess how humans perceive enclosed spaces through echolocation, thereby revealing the interplay between sensory and vocal-motor neural activity while humans perform this task. Sighted subjects were trained to detect small changes in virtual-room size analyzing real-time generated echoes of their vocalizations. Individual differences in performance were related to the type and number of vocalizations produced. We then asked subjects to estimate virtual-room size with either active or passive sounds while measuring their brain activity with fMRI. Subjects were better at estimating room size when actively vocalizing. This was reflected in the hemodynamic activity of vocal-motor cortices, even after individual motor and sensory components were removed. Activity in these areas also varied with perceived room size, although the vocal-motor output was unchanged. In addition, thalamic and auditory-midbrain activity was correlated with perceived room size; a likely result of top-down auditory pathways for human echolocation, comparable with those described in echolocating bats. Our data provide evidence that human echolocation is supported by active sensing, both behaviorally and in terms of brain activity. The neural sensory-motor coupling complements the fundamental acoustic motor-sensory coupling via the environment in echolocation. SIGNIFICANCE STATEMENT Passive listening is the predominant method for examining brain activity during echolocation, the auditory analysis of self-generated sounds. We show that sighted humans perform better when they actively vocalize than during passive listening. Correspondingly, vocal motor and cerebellar activity is greater during active echolocation than vocalization alone. Motor and subcortical auditory brain activity covaries with the auditory percept, although motor output is unchanged. Our results reveal behaviorally relevant neural sensory-motor coupling during echolocation. Copyright © 2017 the authors 0270-6474/17/371614-14$15.00/0.

  18. Computational model for vocal tract dynamics in a suboscine bird.

    PubMed

    Assaneo, M F; Trevisan, M A

    2010-09-01

    In a recent work, active use of the vocal tract has been reported for singing oscines. The reconfiguration of the vocal tract during song serves to match its resonances to the syringeal fundamental frequency, demonstrating a precise coordination of the two main pieces of the avian vocal system for songbirds characterized by tonal songs. In this work we investigated the Great Kiskadee (Pitangus sulfuratus), a suboscine bird whose calls display a rich harmonic content. Using a recently developed mathematical model for the syrinx and a mobile vocal tract, we set up a computational model that provides a plausible reconstruction of the vocal tract movement using a few spectral features taken from the utterances. Moreover, synthetic calls were generated using the articulated vocal tract that accounts for all the acoustical features observed experimentally.

  19. Behavior differentiation between wild Japanese quail, domestic quail, and their first filial generation.

    PubMed

    Chang, G B; Liu, X P; Chang, H; Chen, G H; Zhao, W M; Ji, D J; Chen, R; Qin, Y R; Shi, X K; Hu, G S

    2009-06-01

    The number of wild quail has dramatically reduced in China and reached a state of endangerment with the deterioration of the environment in recent years. In this study, we examined the ecological behaviors of quails in the cage to determine the differentiation level between wild Japanese quail and domestic quail, to detect the relationship between quail behavior and evolutionary differentiation and to analyze the possibility of restoring effective size of wild population. With the on-the-spot observations and measurements, the behaviors of 3 categories of quail, namely wild Japanese quail from the Weishan Lake area in China, domestic quail, and their first filial generation (F(1)) were studied. Domestic quail differed from wild Japanese quail in morphological pattern and ecological behaviors, including some indexes of figure type and egg, vocalization, aggression and fighting, and mating, but wild Japanese quail and domestic quail could succeed in mating and reproducing fertile hybrid offspring. There were significant differences between domestic quail and wild Japanese quail in reproductive traits, involved mating times, fertility rate, hatching rate, and hatching rate of fertilized eggs (P < 0.05). The first filial generation presented significant difference from the wild Japanese quail in vocalization, aggression and fighting, mating, hatching rate, hatching rate of fertilized eggs, and some egg indexes (P < 0.05) and significantly differ from the domestic quail in vocalization, hatching rate, and hatching rate of fertilized eggs (P < 0.05). Evolutionary differentiation between wild quail and domestic quail was still at a relatively low level because no reproductive isolation existed. The advantages of the F(1) hybrids in reproductive capacity, fertilization, and hatching recommend that releasing hybrids instead of domestic quails to the wild would be a more effective way to restore the effective size of wild quail population if necessary.

  20. Auditory and audio-vocal responses of single neurons in the monkey ventral premotor cortex.

    PubMed

    Hage, Steffen R

    2018-03-20

    Monkey vocalization is a complex behavioral pattern, which is flexibly used in audio-vocal communication. A recently proposed dual neural network model suggests that cognitive control might be involved in this behavior, originating from a frontal cortical network in the prefrontal cortex and mediated via projections from the rostral portion of the ventral premotor cortex (PMvr) and motor cortex to the primary vocal motor network in the brainstem. For the rapid adjustment of vocal output to external acoustic events, strong interconnections between vocal motor and auditory sites are needed, which are present at cortical and subcortical levels. However, the role of the PMvr in audio-vocal integration processes remains unclear. In the present study, single neurons in the PMvr were recorded in rhesus monkeys (Macaca mulatta) while volitionally producing vocalizations in a visual detection task or passively listening to monkey vocalizations. Ten percent of randomly selected neurons in the PMvr modulated their discharge rate in response to acoustic stimulation with species-specific calls. More than four-fifths of these auditory neurons showed an additional modulation of their discharge rates either before and/or during the monkeys' motor production of the vocalization. Based on these audio-vocal interactions, the PMvr might be well positioned to mediate higher order auditory processing with cognitive control of the vocal motor output to the primary vocal motor network. Such audio-vocal integration processes in the premotor cortex might constitute a precursor for the evolution of complex learned audio-vocal integration systems, ultimately giving rise to human speech. Copyright © 2018 Elsevier B.V. All rights reserved.

  1. Social coordination in animal vocal interactions. Is there any evidence of turn-taking? The starling as an animal model

    PubMed Central

    Henry, Laurence; Craig, Adrian J. F. K.; Lemasson, Alban; Hausberger, Martine

    2015-01-01

    Turn-taking in conversation appears to be a common feature in various human cultures and this universality raises questions about its biological basis and evolutionary trajectory. Functional convergence is a widespread phenomenon in evolution, revealing sometimes striking functional similarities between very distant species even though the mechanisms involved may be different. Studies on mammals (including non-human primates) and bird species with different levels of social coordination reveal that temporal and structural regularities in vocal interactions may depend on the species' social structure. Here we test the hypothesis that turn-taking and associated rules of conversations may be an adaptive response to the requirements of social life, by testing the applicability of turn-taking rules to an animal model, the European starling. Birdsong has for many decades been considered as one of the best models of human language and starling songs have been well described in terms of vocal production and perception. Starlings do have vocal interactions where alternating patterns predominate. Observational and experimental data on vocal interactions reveal that (1) there are indeed clear temporal and structural regularities, (2) the temporal and structural patterning is influenced by the immediate social context, the general social situation, the individual history, and the internal state of the emitter. Comparison of phylogenetically close species of Sturnids reveals that the alternating pattern of vocal interactions varies greatly according to the species' social structure, suggesting that interactional regularities may have evolved together with social systems. These findings lead to solid bases of discussion on the evolution of communication rules in relation to social evolution. They will be discussed also in terms of processes, at the light of recent neurobiological findings. PMID:26441787

  2. Laryngeal evidence for the first and second passaggio in professionally trained sopranos

    PubMed Central

    Burk, Fabian; Köberlein, Marie; Selamtzis, Andreas; Döllinger, Michael; Burdumy, Michael; Richter, Bernhard

    2017-01-01

    Introduction Due to a lack of empirical data, the current understanding of the laryngeal mechanics in the passaggio regions (i.e., the fundamental frequency ranges where vocal registration events usually occur) of the female singing voice is still limited. Material and methods In this study the first and second passaggio regions of 10 professionally trained female classical soprano singers were analyzed. The sopranos performed pitch glides from A3 (ƒo = 220 Hz) to A4 (ƒo = 440 Hz) and from A4 (ƒo = 440 Hz) to A5 (ƒo = 880 Hz) on the vowel [iː]. Vocal fold vibration was assessed with trans-nasal high speed videoendoscopy at 20,000 fps, complemented by simultaneous electroglottographic (EGG) and acoustic recordings. Register breaks were perceptually rated by 12 voice experts. Voice stability was documented with the EGG-based sample entropy. Glottal opening and closing patterns during the passaggi were analyzed, supplemented with open quotient data extracted from the glottal area waveform. Results In both the first and the second passaggio, variations of vocal fold vibration patterns were found. Four distinct patterns emerged: smooth transitions with either increasing or decreasing durations of glottal closure, abrupt register transitions, and intermediate loss of vocal fold contact. Audible register transitions (in both the first and second passaggi) generally coincided with higher sample entropy values and higher open quotient variance through the respective passaggi. Conclusions Noteworthy vocal fold oscillatory registration events occur in both the first and the second passaggio even in professional sopranos. The respective transitions are hypothesized to be caused by either (a) a change of laryngeal biomechanical properties; or by (b) vocal tract resonance effects, constituting level 2 source-filter interactions. PMID:28467509

  3. Gender differences in laterality patterns for speaking and singing.

    PubMed

    Hough, M S; Daniel, H J; Snow, M A; O'Brien, K F; Hume, W G

    1994-09-01

    This study examined behaviors reflecting cerebral organization of speaking and singing in normal college students. The investigation focused on whether differences existed in the laterality patterns of two singing tasks and one speaking task in males and females. Performance was measured on a verbal/manual time-sharing paradigm, coupling finger tapping with three vocal tasks (speaking, singing a rote song, singing up and down a diatonic five note scale). Females exhibited less variation than males in mean tapping rates and laterality scores across all three vocal tasks, thus indicating that gender most likely influences lateralization of vocal tasks. Bilateral integration was indicated for both males and females during singing up/down the aforementioned scale. These findings suggest differential involvement of both hemispheres in processing musical functions.

  4. The Role of Lexical Stress on the Use of Vocal Fry in Young Adult Female Speakers.

    PubMed

    Gibson, Todd A

    2017-01-01

    Vocal fry is a voice register often used by young adult women for sociolinguistic purposes. Some acoustic correlates of lexical stress, however, appear incompatible with the use of vocal fry. The objective of this study was to systematically examine the role of lexical stress in the use of vocal fry by young adult women. This is a semi-randomized controlled laboratory study. Fifty female undergraduate students were recorded repeating one-, two-, three-, and four-syllable nonwords that conformed to English phonotactics. Nonwords were presented in order from shorter to longer lengths, with stimuli randomized within syllable length. Perceptual analyses of recordings were augmented by acoustic analyses to identify each syllable in which vocal fry occurred. Eighty-six percent of participants produced at least one episode of vocal fry. Vocal fry was more likely to occur in unstressed than stressed position, and the likelihood increased as distance from the stressed syllable increased. There was considerable variability in the use of vocal fry. Frequent and infrequent users varied on the degree to which they used vocal fry in single-syllable nonwords. Vocal fry use persists among young adult women even in the absence of syntactic and pragmatic influences. Lexical stress appeared to dramatically reduce the use of vocal fry. Patterns of vocal fry use appeared to be different for frequent and infrequent users of this vocal register. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  5. Effects of subsampling of passive acoustic recordings on acoustic metrics.

    PubMed

    Thomisch, Karolin; Boebel, Olaf; Zitterbart, Daniel P; Samaran, Flore; Van Parijs, Sofie; Van Opzeeland, Ilse

    2015-07-01

    Passive acoustic monitoring is an important tool in marine mammal studies. However, logistics and finances frequently constrain the number and servicing schedules of acoustic recorders, requiring a trade-off between deployment periods and sampling continuity, i.e., the implementation of a subsampling scheme. Optimizing such schemes to each project's specific research questions is desirable. This study investigates the impact of subsampling on the accuracy of two common metrics, acoustic presence and call rate, for different vocalization patterns (regimes) of baleen whales: (1) variable vocal activity, (2) vocalizations organized in song bouts, and (3) vocal activity with diel patterns. To this end, above metrics are compared for continuous and subsampled data subject to different sampling strategies, covering duty cycles between 50% and 2%. The results show that a reduction of the duty cycle impacts negatively on the accuracy of both acoustic presence and call rate estimates. For a given duty cycle, frequent short listening periods improve accuracy of daily acoustic presence estimates over few long listening periods. Overall, subsampling effects are most pronounced for low and/or temporally clustered vocal activity. These findings illustrate the importance of informed decisions when applying subsampling strategies to passive acoustic recordings or analyses for a given target species.

  6. The role of vocal individuality in conservation

    PubMed Central

    Terry, Andrew MR; Peake, Tom M; McGregor, Peter K

    2005-01-01

    Identifying the individuals within a population can generate information on life history parameters, generate input data for conservation models, and highlight behavioural traits that may affect management decisions and error or bias within census methods. Individual animals can be discriminated by features of their vocalisations. This vocal individuality can be utilised as an alternative marking technique in situations where the marks are difficult to detect or animals are sensitive to disturbance. Vocal individuality can also be used in cases were the capture and handling of an animal is either logistically or ethically problematic. Many studies have suggested that vocal individuality can be used to count and monitor populations over time; however, few have explicitly tested the method in this role. In this review we discuss methods for extracting individuality information from vocalisations and techniques for using this to count and monitor populations over time. We present case studies in birds where vocal individuality has been applied to conservation and we discuss its role in mammals. PMID:15960848

  7. Neural FoxP2 and FoxP1 expression in the budgerigar, an avian species with adult vocal learning.

    PubMed

    Hara, Erina; Perez, Jemima M; Whitney, Osceola; Chen, Qianqian; White, Stephanie A; Wright, Timothy F

    2015-04-15

    Vocal learning underlies acquisition of both language in humans and vocal signals in some avian taxa. These bird groups and humans exhibit convergent developmental phases and associated brain pathways for vocal communication. The transcription factor FoxP2 plays critical roles in vocal learning in humans and songbirds. Another member of the forkhead box gene family, FoxP1 also shows high expression in brain areas involved in vocal learning and production. Here, we investigate FoxP2 and FoxP1 mRNA and protein in adult male budgerigars (Melopsittacus undulatus), a parrot species that exhibits vocal learning as both juveniles and adults. To examine these molecules in adult vocal learners, we compared their expression patterns in the budgerigar striatal nucleus involved in vocal learning, magnocellular nucleus of the medial striatum (MMSt), across birds with different vocal states, such as vocalizing to a female (directed), vocalizing alone (undirected), and non-vocalizing. We found that both FoxP2 mRNA and protein expressions were consistently lower in MMSt than in the adjacent striatum regardless of the vocal states, whereas previous work has shown that songbirds exhibit down-regulation in the homologous region, Area X, only after singing alone. In contrast, FoxP1 levels were high in MMSt compared to the adjacent striatum in all groups. Taken together these results strengthen the general hypothesis that FoxP2 and FoxP1 have specialized expression in vocal nuclei across a range of taxa, and suggest that the adult vocal plasticity seen in budgerigars may be a product of persistent down-regulation of FoxP2 in MMSt. Copyright © 2015 Elsevier B.V. All rights reserved.

  8. Neural FoxP2 and FoxP1 expression in the budgerigar, an avian species with adult vocal learning

    PubMed Central

    Hara, Erina; Perez, Jemima M.; Whitney, Osceola; Chen, Qianqian; White, Stephanie A.; Wright, Timothy F.

    2015-01-01

    Vocal learning underlies acquisition of both language in humans and vocal signals in some avian taxa. These bird groups and humans exhibit convergent developmental phases and associated brain pathways for vocal communication. The transcription factor FoxP2 plays critical roles in vocal learning in humans and songbirds. Another member of the forkhead box gene family, FoxP1 also shows high expression in brain areas involved in vocal learning and production. Here, we investigate FoxP2 and FoxP1 mRNA and protein in adult male budgerigars (Melopsittacus undulatus), a parrot species that exhibits vocal learning as both juveniles and adults. To examine these molecules in adult vocal learners, we compared their expression patterns in the budgerigar striatal nucleus involved in vocal learning, magnocellular nucleus of the medial striatum (MMSt), across birds with different vocal states, such as vocalizing to a female (directed), vocalizing alone (undirected), and non-vocalizing. We found that both FoxP2 mRNA and protein expressions were consistently lower in MMSt than in the adjacent striatum regardless of the vocal states, whereas previous work has shown that songbirds exhibit downregulation in the homologous region, Area X, only after singing alone. In contrast, FoxP1 levels were high in MMSt compared to the adjacent striatum in all groups. Taken together these results strengthen the general hypothesis that FoxP2 and FoxP1 have specialized expression in vocal nuclei across a range of taxa, and suggest that the adult vocal plasticity seen in budgerigars may be a product of persistent down-regulation of FoxP2 in MMSt. PMID:25601574

  9. Singers' Vocal Function Knowledge Levels, Sensorimotor Self-awareness of Vocal Tract, and Impact of Functional Voice Rehabilitation on the Vocal Function Knowledge and Self-awareness of Vocal Tract.

    PubMed

    Sielska-Badurek, Ewelina; Osuch-Wójcikiewicz, Ewa; Sobol, Maria; Kazanecka, Ewa; Niemczyk, Kazimierz

    2017-01-01

    This study investigated vocal function knowledge and vocal tract sensorimotor self-awareness and the impact of functional voice rehabilitation on vocal function knowledge and self-awareness. This is a prospective, randomized study. Twenty singers (study group [SG]) completed a questionnaire before and after functional voice rehabilitation. Twenty additional singers, representing the control group, also completed the questionnaire without functional voice rehabilitation at a 3-month interval. The questionnaire consisted of three parts. The first part evaluated the singers' attitude to the anatomical and physiological knowledge of the vocal tract and their self-esteem of the knowledge level. The second part assessed the theoretical knowledge of the singers' vocal tract physiology. The third part of the questionnaire assessed singers' sensorimotor self-awareness of the vocal tract. The results showed that most singers indicated that knowledge of the vocal tract's anatomy and physiology is useful (59% SG, 67% control group). However, 75% of all participants defined their knowledge of the vocal tract's anatomy and physiology as weak or inadequate. In the SG, vocal function knowledge at the first assessment was 45%. After rehabilitation, the level increased to 67.7%. Vocal tract sensorimotor self-awareness initially was 38.9% in SG but rose to 66.7%. Findings of the study suggest that classical singers lack knowledge about the physiology of the vocal mechanism, especially the breathing patterns. In addition, they have low sensorimotor self-awareness of their vocal tract. The results suggest that singers would benefit from receiving services from phoniatrists and speech-language pathologists during their voice training. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  10. Microvascular lesions of the true vocal fold.

    PubMed

    Postma, G N; Courey, M S; Ossoff, R H

    1998-06-01

    Microvascular lesions, also called varices or capillary ectasias, in contrast to vocal fold polyps with telangiectatic vessels, are relatively small lesions arising from the microcirculation of the vocal fold. Varices are most commonly seen in female professional vocalists and may be secondary to repetitive trauma, hormonal variations, or repeated inflammation. Microvascular lesions may either be asymptomatic or cause frank dysphonia by interrupting the normal vibratory pattern, mass, or closure of the vocal folds. They may also lead to vocal fold hemorrhage, scarring, or polyp formation. Laryngovideostroboscopy is the key in determining the functional significance of vocal fold varices. Management of patients with a varix includes medical therapy, speech therapy, and occasionally surgical vaporization. Indications for surgery are recurrent hemorrhage, enlargement of the varix, development of a mass in conjunction with the varix or hemorrhage, and unacceptable dysphonia after maximal medical and speech therapy due to a functionally significant varix.

  11. In the Beginning Was the Familiar Voice Personally Familiar Voices in the Evolutionary and Contemporary Biology of Communication

    PubMed Central

    Sidtis, Diana; Kreiman, Jody

    2011-01-01

    The human voice is described in dialogic linguistics as an embodiment of self in a social context, contributing to expression, perception and mutual exchange of self, consciousness, inner life, and personhood. While these approaches are subjective and arise from phenomenological perspectives, scientific facts about personal vocal identity, and its role in biological development, support these views. It is our purpose to review studies of the biology of personal vocal identity -- the familiar voice pattern-- as providing an empirical foundation for the view that the human voice is an embodiment of self in the social context. Recent developments in the biology and evolution of communication are concordant with these notions, revealing that familiar voice recognition (also known as vocal identity recognition or individual vocal recognition) or contributed to survival in the earliest vocalizing species. Contemporary ethology documents the crucial role of familiar voices across animal species in signaling and perceiving internal states and personal identities. Neuropsychological studies of voice reveal multimodal cerebral associations arising across brain structures involved in memory, emotion, attention, and arousal in vocal perception and production, such that the voice represents the whole person. Although its roots are in evolutionary biology, human competence for processing layered social and personal meanings in the voice, as well as personal identity in a large repertory of familiar voice patterns, has achieved an immense sophistication. PMID:21710374

  12. Discriminating Simulated Vocal Tremor Source Using Amplitude Modulation Spectra

    PubMed Central

    Carbonell, Kathy M.; Lester, Rosemary A.; Story, Brad H.; Lotto, Andrew J.

    2014-01-01

    Objectives/Hypothesis Sources of vocal tremor are difficult to categorize perceptually and acoustically. This paper describes a preliminary attempt to discriminate vocal tremor sources through the use of spectral measures of the amplitude envelope. The hypothesis is that different vocal tremor sources are associated with distinct patterns of acoustic amplitude modulations. Study Design Statistical categorization methods (discriminant function analysis) were used to discriminate signals from simulated vocal tremor with different sources using only acoustic measures derived from the amplitude envelopes. Methods Simulations of vocal tremor were created by modulating parameters of a vocal fold model corresponding to oscillations of respiratory driving pressure (respiratory tremor), degree of vocal fold adduction (adductory tremor) and fundamental frequency of vocal fold vibration (F0 tremor). The acoustic measures were based on spectral analyses of the amplitude envelope computed across the entire signal and within select frequency bands. Results The signals could be categorized (with accuracy well above chance) in terms of the simulated tremor source using only measures of the amplitude envelope spectrum even when multiple sources of tremor were included. Conclusions These results supply initial support for an amplitude-envelope based approach to identify the source of vocal tremor and provide further evidence for the rich information about talker characteristics present in the temporal structure of the amplitude envelope. PMID:25532813

  13. Neural Representation of a Target Auditory Memory in a Cortico-Basal Ganglia Pathway

    PubMed Central

    Bottjer, Sarah W.

    2013-01-01

    Vocal learning in songbirds, like speech acquisition in humans, entails a period of sensorimotor integration during which vocalizations are evaluated via auditory feedback and progressively refined to achieve an imitation of memorized vocal sounds. This process requires the brain to compare feedback of current vocal behavior to a memory of target vocal sounds. We report the discovery of two distinct populations of neurons in a cortico-basal ganglia circuit of juvenile songbirds (zebra finches, Taeniopygia guttata) during vocal learning: (1) one in which neurons are selectively tuned to memorized sounds and (2) another in which neurons are selectively tuned to self-produced vocalizations. These results suggest that neurons tuned to learned vocal sounds encode a memory of those target sounds, whereas neurons tuned to self-produced vocalizations encode a representation of current vocal sounds. The presence of neurons tuned to memorized sounds is limited to early stages of sensorimotor integration: after learning, the incidence of neurons encoding memorized vocal sounds was greatly diminished. In contrast to this circuit, neurons known to drive vocal behavior through a parallel cortico-basal ganglia pathway show little selective tuning until late in learning. One interpretation of these data is that representations of current and target vocal sounds in the shell circuit are used to compare ongoing patterns of vocal feedback to memorized sounds, whereas the parallel core circuit has a motor-related role in learning. Such a functional subdivision is similar to mammalian cortico-basal ganglia pathways in which associative-limbic circuits mediate goal-directed responses, whereas sensorimotor circuits support motor aspects of learning. PMID:24005299

  14. Conversational Entrainment of Vocal Fry in Young Adult Female American English Speakers.

    PubMed

    Borrie, Stephanie A; Delfino, Christine R

    2017-07-01

    Conversational entrainment, the natural tendency for people to modify their behaviors to more closely match their communication partner, is examined as one possible mechanism modulating the prevalence of vocal fry in the speech of young American women engaged in spoken dialogue. Twenty young adult female American English speakers engaged in two spoken dialogue tasks-one with a young adult female American English conversational partner who exhibited substantial vocal fry and one with a young adult female American English conversational partner who exhibited quantifiably less vocal fry. Dialogues were analyzed for proportion of vocal fry, by speaker, and two measures of communicative success (efficiency and enjoyment). Participants employed significantly more vocal fry when conversing with the partner who exhibited substantial vocal fry than when conversing with the partner who exhibited quantifiably less vocal fry. Further, greater similarity between communication partners in their use of vocal fry tracked with higher scores of communicative efficiency and communicative enjoyment. Conversational entrainment offers a mechanistic framework that may be used to explain, to some degree, the frequency with which vocal fry is employed by young American women engaged in spoken dialogue. Further, young American women who modulated their vocal patterns during dialogue to match those of their conversational partner gained more efficiency and enjoyment from their interactions, demonstrating the cognitive and social benefits of entrainment. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  15. Possible Role of Mother-Daughter Vocal Interactions on the Development of Species-Specific Song in Gibbons

    PubMed Central

    Koda, Hiroki; Lemasson, Alban; Oyakawa, Chisako; Rizaldi; Pamungkas, Joko; Masataka, Nobuo

    2013-01-01

    Mother-infant vocal interactions play a crucial role in the development of human language. However, comparatively little is known about the maternal role during vocal development in nonhuman primates. Here, we report the first evidence of mother-daughter vocal interactions contributing to vocal development in gibbons, a singing and monogamous ape species. Gibbons are well known for their species-specific duets sung between mates, yet little is known about the role of intergenerational duets in gibbon song development. We observed singing interactions between free-ranging mothers and their sub-adult daughters prior to emigration. Daughters sang simultaneously with their mothers at different rates. First, we observed significant acoustic variation between daughters. Co-singing rates between mother and daughter were negatively correlated with the temporal precision of the song’s synchronization. In addition, songs of daughters who co-sang less with their mothers were acoustically more similar to the maternal song than any other adult female’s song. All variables have been reported to be influenced by social relationships of pairs. Therefore those correlations would be mediated by mother-daughter social relationship, which would be modifiable in daughter’s development. Here we hypothesized that daughters who co-sing less often, well-synchronize, and converge acoustically with the maternal acoustic pattern would be at a more advanced stage of social independence in sub-adult females prior to emigration. Second, we observed acoustic matching between mothers and daughters when co-singing, suggesting short-term vocal flexibility. Third, we found that mothers adjusted songs to a more stereotyped pattern when co-singing than when singing alone. This vocal adjustment was stronger for mothers with daughters who co-sang less. These results indicate the presence of socially mediated vocal flexibility in gibbon sub-adults and adults, and that mother-daughter co-singing interactions may enhance vocal development. More comparative work, notably longitudinal and experimental, is now needed to clarify maternal roles during song development. PMID:23951160

  16. Possible role of mother-daughter vocal interactions on the development of species-specific song in gibbons.

    PubMed

    Koda, Hiroki; Lemasson, Alban; Oyakawa, Chisako; Rizaldi; Pamungkas, Joko; Masataka, Nobuo

    2013-01-01

    Mother-infant vocal interactions play a crucial role in the development of human language. However, comparatively little is known about the maternal role during vocal development in nonhuman primates. Here, we report the first evidence of mother-daughter vocal interactions contributing to vocal development in gibbons, a singing and monogamous ape species. Gibbons are well known for their species-specific duets sung between mates, yet little is known about the role of intergenerational duets in gibbon song development. We observed singing interactions between free-ranging mothers and their sub-adult daughters prior to emigration. Daughters sang simultaneously with their mothers at different rates. First, we observed significant acoustic variation between daughters. Co-singing rates between mother and daughter were negatively correlated with the temporal precision of the song's synchronization. In addition, songs of daughters who co-sang less with their mothers were acoustically more similar to the maternal song than any other adult female's song. All variables have been reported to be influenced by social relationships of pairs. Therefore those correlations would be mediated by mother-daughter social relationship, which would be modifiable in daughter's development. Here we hypothesized that daughters who co-sing less often, well-synchronize, and converge acoustically with the maternal acoustic pattern would be at a more advanced stage of social independence in sub-adult females prior to emigration. Second, we observed acoustic matching between mothers and daughters when co-singing, suggesting short-term vocal flexibility. Third, we found that mothers adjusted songs to a more stereotyped pattern when co-singing than when singing alone. This vocal adjustment was stronger for mothers with daughters who co-sang less. These results indicate the presence of socially mediated vocal flexibility in gibbon sub-adults and adults, and that mother-daughter co-singing interactions may enhance vocal development. More comparative work, notably longitudinal and experimental, is now needed to clarify maternal roles during song development.

  17. Diel variation in detection and vocalization rates of king (Rallus elegans) and clapper (Rallus crepitans) rails in intracoastal waterways

    USGS Publications Warehouse

    Stiffler, Lydia L.; Anderson, James T.; Welsh, Amy B.; Harding, Sergio R.; Costanzo, Gary R.; Katzner, Todd

    2017-01-01

    Surveys for secretive marsh birds could be improved with refinements to address regional and species-specific variation in detection probabilities and optimal times of day to survey. Diel variation in relation to naïve occupancy, detection rates, and vocalization rates of King (Rallus elegans) and Clapper (R. crepitans) rails were studied in intracoastal waterways in Virginia, USA. Autonomous acoustic devices recorded vocalizations of King and Clapper rails at 75 locations for 48-hr periods within a marsh complex. Naïve King and Clapper rail occupancy did not vary hourly at either the marsh or the study area level. Combined King and Clapper rail detections and vocalizations varied across marshes, decreased as the sampling season progressed, and, for detections, was greatest during low rising tides (P < 0.01). Hourly variation in vocalization and detection rates did not show a pattern but occurred between 7.8% of pairwise comparisons for detections and 10.5% of pairwise comparisons for vocalizations (P < 0.01). Higher rates of detections and vocalizations occurred during the hours of 00:00–00:59, 05:00–05:59, 14:00–15:59, and lower rates during the hours of 07:00–09:59. Although statistically significant, because there were no patterns in these hourly differences, they may not be biologically relevant and are of little use to management. In fact, these findings demonstrate that surveys for King and Clapper rails in Virginia intracoastal waterways may be effectively conducted throughout the day.

  18. The vocal sac of Hylodidae (Amphibia, Anura): Phylogenetic and functional implications of a unique morphology.

    PubMed

    Elias-Costa, Agustin J; Montesinos, Rachel; Grant, Taran; Faivovich, Julián

    2017-11-01

    Anuran vocal sacs are elastic chambers that recycle exhaled air during vocalizations and are present in males of most species of frogs. Most knowledge of the diversity of vocal sacs relates to external morphology; detailed information on internal anatomy is available for few groups of frogs. Frogs of the family Hylodidae, which is endemic to the Atlantic Forest of Brazil and adjacent Argentina and Paraguay, have three patterns of vocal sac morphology-that is, single, subgular; paired, lateral; and absent. The submandibular musculature and structure of the vocal sac mucosa (the internal wall of the vocal sac) of exemplar species of this family and relatives were studied. In contrast to previous accounts, we found that all species of Crossodactylus and Hylodes possess paired, lateral vocal sacs, with the internal mucosa of each sac being separate from the contralateral one. Unlike all other frogs for which data are available, the mucosa of the vocal sacs in these genera is not supported externally by the mm. intermandibularis and interhyoideus. Rather, the vocal sac mucosa projects through the musculature and is free in the submandibular lymphatic sac. The presence of paired, lateral vocal sacs, the internal separation of the sac mucosae, and their projection through the m. interhyoideus are synapomorphies of the family. Furthermore, the specific configuration of the m. interhyoideus allows asymmetric inflation of paired vocal sacs, a feature only reported in species of these diurnal, stream-dwelling frogs. © 2017 Wiley Periodicals, Inc.

  19. Differential Expression of Glutamate Receptors in Avian Neural Pathways for Learned Vocalization

    PubMed Central

    WADA, KAZUHIRO; SAKAGUCHI, HIRONOBU; JARVIS, ERICH D.; HAGIWARA, MASATOSHI

    2008-01-01

    Learned vocalization, the substrate for human language, is a rare trait. It is found in three distantly related groups of birds—parrots, hummingbirds, and songbirds. These three groups contain cerebral vocal nuclei for learned vocalization not found in their more closely related vocal nonlearning relatives. Here, we cloned 21 receptor subunits/subtypes of all four glutamate receptor families (AMPA, kainate, NMDA, and metabotropic) and examined their expression in vocal nuclei of songbirds. We also examined expression of a subset of these receptors in vocal nuclei of hummingbirds and parrots, as well as in the brains of dove species as examples of close vocal nonlearning relatives. Among the 21 subunits/subtypes, 19 showed higher and/or lower prominent differential expression in songbird vocal nuclei relative to the surrounding brain subdivisions in which the vocal nuclei are located. This included relatively lower levels of all four AMPA subunits in lMAN, strikingly higher levels of the kainite subunit GluR5 in the robust nucleus of the arcopallium (RA), higher and lower levels respectively of the NMDA subunits NR2A and NR2B in most vocal nuclei and lower levels of the metabotropic group I subtypes (mGluR1 and -5) in most vocal nuclei and the group II subtype (mGluR2), showing a unique expression pattern of very low levels in RA and very high levels in HVC. The splice variants of AMPA subunits showed further differential expression in vocal nuclei. Some of the receptor subunits/subtypes also showed differential expression in hummingbird and parrot vocal nuclei. The magnitude of differential expression in vocal nuclei of all three vocal learners was unique compared with the smaller magnitude of differences found for nonvocal areas of vocal learners and vocal nonlearners. Our results suggest that evolution of vocal learning was accompanied by differential expression of a conserved gene family for synaptic transmission and plasticity in vocal nuclei. They also suggest that neural activity and signal transduction in vocal nuclei of vocal learners will be different relative to the surrounding brain areas. PMID:15236466

  20. A sensorimotor area in the songbird brain is required for production of vocalizations in the song learning period of development.

    PubMed

    Piristine, Hande C; Choetso, Tenzin; Gobes, Sharon M H

    2016-11-01

    Sensory feedback is essential for acquiring and maintaining complex motor behaviors, including birdsong. In zebra finches, auditory feedback reaches the song control circuits primarily through the nucleus interfacialis nidopalii (Nif), which provides excitatory input to HVC (proper name)-a premotor region essential for the production of learned vocalizations. Despite being one of the major inputs to the song control pathway, the role of Nif in generating vocalizations is not well understood. To address this, we transiently inactivated Nif in late juvenile zebra finches. Upon Nif inactivation (in both hemispheres or on one side only), birds went from singing stereotyped zebra finch song to uttering highly variable and unstructured vocalizations resembling sub-song, an early juvenile song form driven by a basal ganglia circuit. Simultaneously inactivating Nif and LMAN (lateral magnocellular nucleus of the anterior nidopallium), the output nucleus of a basal ganglia circuit, inhibited song production altogether. These results suggest that Nif is required for generating the premotor drive for song. Permanent Nif lesions, in contrast, have only transient effects on vocal production, with song recovering within a day. The sensorimotor nucleus Nif thus produces a premotor drive to the motor pathway that is acutely required for generating learned vocalizations, but once permanently removed, the song system can compensate for its absence. © 2016 Wiley Periodicals, Inc. Develop Neurobiol 76: 1213-1225, 2016. © 2016 Wiley Periodicals, Inc.

  1. [The three-dimensional simulation of arytenoid cartilage movement].

    PubMed

    Zhang, Jun; Wang, Xuefeng

    2011-08-01

    Exploring the characteristics of arytenoid cartilage movement. Using Pro/ENGINEER (Pro/E) software, the cricoid cartilage, arytenoid cartilage and vocal cords were simulated to the three-dimensional reconstruction, by analyzing the trajectory of arytenoid cartilage in the joint surface from the cricoid cartilage and arytenoid cartilage composition. The 3D animation simulation showed the normal movement patterns of the vocal cords and the characteristics of vocal cords movement in occasion of arytenoid cartilage dislocation vividly. The three-dimensional model has clinical significance for arytenoid cartilage movement disorders.

  2. Learning to detect vocal hyperfunction from ambulatory neck-surface acceleration features: initial results for vocal fold nodules.

    PubMed

    Ghassemi, Marzyeh; Van Stan, Jarrad H; Mehta, Daryush D; Zañartu, Matías; Cheyne, Harold A; Hillman, Robert E; Guttag, John V

    2014-06-01

    Voice disorders are medical conditions that often result from vocal abuse/misuse which is referred to generically as vocal hyperfunction. Standard voice assessment approaches cannot accurately determine the actual nature, prevalence, and pathological impact of hyperfunctional vocal behaviors because such behaviors can vary greatly across the course of an individual's typical day and may not be clearly demonstrated during a brief clinical encounter. Thus, it would be clinically valuable to develop noninvasive ambulatory measures that can reliably differentiate vocal hyperfunction from normal patterns of vocal behavior. As an initial step toward this goal we used an accelerometer taped to the neck surface to provide a continuous, noninvasive acceleration signal designed to capture some aspects of vocal behavior related to vocal cord nodules, a common manifestation of vocal hyperfunction. We gathered data from 12 female adult patients diagnosed with vocal fold nodules and 12 control speakers matched for age and occupation. We derived features from weeklong neck-surface acceleration recordings by using distributions of sound pressure level and fundamental frequency over 5-min windows of the acceleration signal and normalized these features so that intersubject comparisons were meaningful. We then used supervised machine learning to show that the two groups exhibit distinct vocal behaviors that can be detected using the acceleration signal. We were able to correctly classify 22 of the 24 subjects, suggesting that in the future measures of the acceleration signal could be used to detect patients with the types of aberrant vocal behaviors that are associated with hyperfunctional voice disorders.

  3. Automated Assessment of Child Vocalization Development Using LENA.

    PubMed

    Richards, Jeffrey A; Xu, Dongxin; Gilkerson, Jill; Yapanel, Umit; Gray, Sharmistha; Paul, Terrance

    2017-07-12

    To produce a novel, efficient measure of children's expressive vocal development on the basis of automatic vocalization assessment (AVA), child vocalizations were automatically identified and extracted from audio recordings using Language Environment Analysis (LENA) System technology. Assessment was based on full-day audio recordings collected in a child's unrestricted, natural language environment. AVA estimates were derived using automatic speech recognition modeling techniques to categorize and quantify the sounds in child vocalizations (e.g., protophones and phonemes). These were expressed as phone and biphone frequencies, reduced to principal components, and inputted to age-based multiple linear regression models to predict independently collected criterion-expressive language scores. From these models, we generated vocal development AVA estimates as age-standardized scores and development age estimates. AVA estimates demonstrated strong statistical reliability and validity when compared with standard criterion expressive language assessments. Automated analysis of child vocalizations extracted from full-day recordings in natural settings offers a novel and efficient means to assess children's expressive vocal development. More research remains to identify specific mechanisms of operation.

  4. Syllable Durations of Preword and Early Word Vocalizations.

    ERIC Educational Resources Information Center

    Robb, Michael P.; Saxman, John H.

    1990-01-01

    The continuity in development of syllable duration patterns was examined in seven young children as they progressed from preword to multiword periods of vocalization development. Results revealed no systematic increase or decrease in the duration of bisyllables produced by the children as a group, whereas lengthening of final syllables was…

  5. Modeling the biomechanical influence of epilaryngeal stricture on the vocal folds: a low-dimensional model of vocal-ventricular fold coupling.

    PubMed

    Moisik, Scott R; Esling, John H

    2014-04-01

    PURPOSE Physiological and phonetic studies suggest that, at moderate levels of epilaryngeal stricture, the ventricular folds impinge upon the vocal folds and influence their dynamical behavior, which is thought to be responsible for constricted laryngeal sounds. In this work, the authors examine this hypothesis through biomechanical modeling. METHOD The dynamical response of a low-dimensional, lumped-element model of the vocal folds under the influence of vocal-ventricular fold coupling was evaluated. The model was assessed for F0 and cover-mass phase difference. Case studies of simulations of different constricted phonation types and of glottal stop illustrate various additional aspects of model performance. RESULTS Simulated vocal-ventricular fold coupling lowers F0 and perturbs the mucosal wave. It also appears to reinforce irregular patterns of oscillation, and it can enhance laryngeal closure in glottal stop production. CONCLUSION The effects of simulated vocal-ventricular fold coupling are consistent with sounds, such as creaky voice, harsh voice, and glottal stop, that have been observed to involve epilaryngeal stricture and apparent contact between the vocal folds and ventricular folds. This supports the view that vocal-ventricular fold coupling is important in the vibratory dynamics of such sounds and, furthermore, suggests that these sounds may intrinsically require epilaryngeal stricture.

  6. Cause-effect relationship between vocal fold physiology and voice production in a three-dimensional phonation model

    PubMed Central

    Zhang, Zhaoyan

    2016-01-01

    The goal of this study is to better understand the cause-effect relation between vocal fold physiology and the resulting vibration pattern and voice acoustics. Using a three-dimensional continuum model of phonation, the effects of changes in vocal fold stiffness, medial surface thickness in the vertical direction, resting glottal opening, and subglottal pressure on vocal fold vibration and different acoustic measures are investigated. The results show that the medial surface thickness has dominant effects on the vertical phase difference between the upper and lower margins of the medial surface, closed quotient, H1-H2, and higher-order harmonics excitation. The main effects of vocal fold approximation or decreasing resting glottal opening are to lower the phonation threshold pressure, reduce noise production, and increase the fundamental frequency. Increasing subglottal pressure is primarily responsible for vocal intensity increase but also leads to significant increase in noise production and an increased fundamental frequency. Increasing AP stiffness significantly increases the fundamental frequency and slightly reduces noise production. The interaction among vocal fold thickness, stiffness, approximation, and subglottal pressure in the control of F0, vocal intensity, and voice quality is discussed. PMID:27106298

  7. Pathogenesis of vocal fold nodules: new insights from a modelling approach.

    PubMed

    Dejonckere, Philippe H; Kob, Malte

    2009-01-01

    To give new insights into the pathogenesis of vocal fold nodules: (a) why the female/male ratio is so extreme, (b) how an hourglass-shaped vibration pattern - eliciting a localized microtrauma - originates, and (c) what the roles of muscular tension imbalance and of behavioral aspects are. Simulations with a 3-dimensional computer model of the vibrating vocal folds. (1) A slightly incomplete dorsal vocal fold adduction is a first condition for inducing an hourglass vibration pattern. (2) A limited collision zone is only possible with a small degree of curving of the rest position of the vocal fold edges in their ventral portion. This is an anatomical characteristic of the adult female larynx. Muscular fatigue and resulting hypotonia seem to enhance this curving. (3) If both these conditions are fulfilled, a sufficient vibration amplitude is required to achieve a localized impact. (4) This third condition can be obtained by an increased subglottal pressure and/or by a decrease in active stress of the tension forces between the neighboring vocalis masses. These last aspects incorporate muscular tension imbalance (dyskinesia) and behavioral aspects in the modelling process. Decrease in active stress is a possible effect of fatigue, and increase in subglottal pressure a result of effort compensation. Copyright 2009 S. Karger AG, Basel.

  8. Factors associated with voice therapy outcomes in the treatment of presbyphonia.

    PubMed

    Mau, Ted; Jacobson, Barbara H; Garrett, C Gaelyn

    2010-06-01

    Age, vocal fold atrophy, glottic closure pattern, and the burden of medical problems are associated with voice therapy outcomes for presbyphonia. Retrospective. Records of patients seen over a 3-year period at a voice center were screened. Inclusion criteria consisted of age over 55 years, primary complaint of hoarseness, presence of vocal fold atrophy on examination, and absence of laryngeal or neurological pathology. Videostroboscopic examinations on initial presentation were reviewed. Voice therapy outcomes were assessed with the American Speech-Language-Hearing Association National Outcomes Measurement System scale. Statistical analysis was performed with Spearman rank correlation and chi(2) tests. Sixty-seven patients were included in the study. Of the patients, 85% demonstrated improvement with voice therapy. The most common type of glottic closure consisted of a slit gap. Gender or age had no effect on voice therapy outcomes. Larger glottic gaps on initial stroboscopy examination and more pronounced vocal fold atrophy were weakly correlated with less improvement from voice therapy. A weak correlation was also found between the number of chronic medical conditions and poorer outcomes from voice therapy. The degree of clinician-determined improvement in vocal function from voice therapy is independent of patient age but is influenced by the degree of vocal fold atrophy, glottic closure pattern, and the patient's burden of medical problems.

  9. An optical flow-based state-space model of the vocal folds.

    PubMed

    Granados, Alba; Brunskog, Jonas

    2017-06-01

    High-speed movies of the vocal fold vibration are valuable data to reveal vocal fold features for voice pathology diagnosis. This work presents a suitable Bayesian model and a purely theoretical discussion for further development of a framework for continuum biomechanical features estimation. A linear and Gaussian nonstationary state-space model is proposed and thoroughly discussed. The evolution model is based on a self-sustained three-dimensional finite element model of the vocal folds, and the observation model involves a dense optical flow algorithm. The results show that the method is able to capture different deformation patterns between the computed optical flow and the finite element deformation, controlled by the choice of the model tissue parameters.

  10. Using the Natural Language Paradigm (NLP) to increase vocalizations of older adults with cognitive impairments.

    PubMed

    Leblanc, Linda A; Geiger, Kaneen B; Sautter, Rachael A; Sidener, Tina M

    2007-01-01

    The Natural Language Paradigm (NLP) has proven effective in increasing spontaneous verbalizations for children with autism. This study investigated the use of NLP with older adults with cognitive impairments served at a leisure-based adult day program for seniors. Three individuals with limited spontaneous use of functional language participated in a multiple baseline design across participants. Data were collected on appropriate and inappropriate vocalizations with appropriate vocalizations coded as prompted or unprompted during baseline and treatment sessions. All participants experienced increases in appropriate speech during NLP with variable response patterns. Additionally, the two participants with substantial inappropriate vocalizations showed decreases in inappropriate speech. Implications for intervention in day programs are discussed.

  11. Are 50-kHz calls used as play signals in the playful interactions of rats? II. Evidence from the effects of devocalization.

    PubMed

    Kisko, Theresa M; Himmler, Brett T; Himmler, Stephanie M; Euston, David R; Pellis, Sergio M

    2015-02-01

    During playful interactions, juvenile rats emit many 50-kHz ultrasonic vocalizations, which are associated with a positive affective state. In addition, these calls may also serve a communicative role - as play signals that promote playful contact. Consistent with this hypothesis, a previous study found that vocalizations are more frequent prior to playful contact than after contact is terminated. The present study uses devocalized rats to test three predictions arising from the play signals hypothesis. First, if vocalizations are used to facilitate contact, then in pairs of rats in which one is devocalized, the higher frequency of pre-contact calling should only be present when the intact rat is initiating the approach. Second, when both partners in a playing pair are devocalized, the frequency of play should be reduced and the typical pattern of playful wrestling disrupted. Finally, when given a choice to play with a vocal and a non-vocal partner, rats should prefer to play with the one able to vocalize. The second prediction was supported in that the frequency of playful interactions as well as some typical patterns of play was disrupted. Even though the data for the other two predictions did not produce the expected findings, they support the conclusion that, in rats, 50-kHz calls are likely to function to maintain a playful mood and for them to signal to one another during play fighting. Copyright © 2014 Elsevier B.V. All rights reserved.

  12. The Songbird as a Percussionist: Syntactic Rules for Non-Vocal Sound and Song Production in Java Sparrows

    PubMed Central

    Soma, Masayo; Mori, Chihiro

    2015-01-01

    Music and dance are two remarkable human characteristics that are closely related. Communication through integrated vocal and motional signals is also common in the courtship displays of birds. The contribution of songbird studies to our understanding of vocal learning has already shed some light on the cognitive underpinnings of musical ability. Moreover, recent pioneering research has begun to show how animals can synchronize their behaviors with external stimuli, like metronome beats. However, few studies have applied such perspectives to unraveling how animals can integrate multimodal communicative signals that have natural functions. Additionally, studies have rarely asked how well these behaviors are learned. With this in mind, here we cast a spotlight on an unusual animal behavior: non-vocal sound production associated with singing in the Java sparrow (Lonchura oryzivora), a songbird. We show that male Java sparrows coordinate their bill-click sounds with the syntax of their song-note sequences, similar to percussionists. Analysis showed that they produced clicks frequently toward the beginning of songs and before/after specific song notes. We also show that bill-clicking patterns are similar between social fathers and their sons, suggesting that these behaviors might be learned from models or linked to learning-based vocalizations. Individuals untutored by conspecifics also exhibited stereotypical bill-clicking patterns in relation to song-note sequence, indicating that while the production of bill clicking itself is intrinsic, its syncopation appears to develop with songs. This paints an intriguing picture in which non-vocal sounds are integrated with vocal courtship signals in a songbird, a model that we expect will contribute to the further understanding of multimodal communication. PMID:25992841

  13. Subglottal pressure, tracheal airflow, and intrinsic laryngeal muscle activity during rat ultrasound vocalization

    PubMed Central

    2011-01-01

    Vocal production requires complex planning and coordination of respiratory, laryngeal, and vocal tract movements, which are incompletely understood in most mammals. Rats produce a variety of whistles in the ultrasonic range that are of communicative relevance and of importance as a model system, but the sources of acoustic variability were mostly unknown. The goal was to identify sources of fundamental frequency variability. Subglottal pressure, tracheal airflow, and electromyographic (EMG) data from two intrinsic laryngeal muscles were measured during 22-kHz and 50-kHz call production in awake, spontaneously behaving adult male rats. During ultrasound vocalization, subglottal pressure ranged between 0.8 and 1.9 kPa. Pressure differences between call types were not significant. The relation between fundamental frequency and subglottal pressure within call types was inconsistent. Experimental manipulations of subglottal pressure had only small effects on fundamental frequency. Tracheal airflow patterns were also inconsistently associated with frequency. Pressure and flow seem to play a small role in regulation of fundamental frequency. Muscle activity, however, is precisely regulated and very sensitive to alterations, presumably because of effects on resonance properties in the vocal tract. EMG activity of cricothyroid and thyroarytenoid muscle was tonic in calls with slow or no fundamental frequency modulations, like 22-kHz and flat 50-kHz calls. Both muscles showed brief high-amplitude, alternating bursts at rates up to 150 Hz during production of frequency-modulated 50-kHz calls. A differentiated and fine regulation of intrinsic laryngeal muscles is critical for normal ultrasound vocalization. Many features of the laryngeal muscle activation pattern during ultrasound vocalization in rats are shared with other mammals. PMID:21832032

  14. Conjunction of Vocal Production and Perception Regulates Expression of the Immediate Early Gene ZENK in a Novel Cortical Region of Songbirds

    PubMed Central

    Alderete, Tanya L.; Chang, Daniel

    2010-01-01

    The cortical nucleus LMAN (lateral magnocellular nucleus of the anterior nidopallium) provides the output of a basal ganglia pathway that is necessary for acquisition of learned vocal behavior during development in songbirds. LMAN is composed of two subregions, a core and a surrounding shell, that give rise to independent pathways that traverse the forebrain in parallel. The LMANshell pathway forms a recurrent loop that includes a cortical region, the dorsal region of the caudolateral nidopallium (dNCL), hitherto unknown to be involved with learned vocal behavior. Here we show that vocal production strongly induces the IEG product ZENK in dNCL of zebra finches. Hearing tutor song while singing is more effective at inducing expression in dNCL of juvenile birds during the auditory–motor integration stage of vocal learning than is hearing conspecific song. In contrast, hearing conspecific song is relatively more effective at inducing expression in adult birds, regardless of whether they are producing song. Furthermore, ZENK+ neurons in dNCL include projection neurons that are part of the LMANshell recurrent loop and a high proportion of dNCL projection neurons express ZENK in singing juvenile birds that hear tutor song. Thus juvenile birds that are actively refining their vocal pattern to imitate a tutor song show high levels of ZENK induction in dNCL neurons when they are singing while hearing the song of their tutor and low levels when they hear a novel conspecific. This pattern indicates that dNCL is a novel brain region involved with vocal learning and that its function is developmentally regulated. PMID:20107119

  15. Developmental pattern of diacylglycerol lipase-α (DAGLα) immunoreactivity in brain regions important for song learning and control in the zebra finch (Taeniopygia guttata).

    PubMed

    Soderstrom, Ken; Wilson, Ashley R

    2013-11-01

    Zebra finch song is a learned behavior dependent upon successful progress through a sensitive period of late-postnatal development. This learning is associated with maturation of distinct brain nuclei and the fiber tract interconnections between them. We have previously found remarkably distinct and dense CB1 cannabinoid receptor expression within many of these song control brain regions, implying a normal role for endocannabinoid signaling in vocal learning. Activation of CB1 receptors via daily treatments with exogenous agonist during sensorimotor stages of song learning (but not in adulthood) results in persistent alteration of song patterns. Now we are working to understand physiological changes responsible for this cannabinoid-altered vocal learning. We have found that song-altering developmental treatments are associated with changes in expression of endocannabinoid signaling elements, including CB1 receptors and the principal CNS endogenous agonist, 2-AG. Within CNS, 2-AG is produced largely through activity of the α isoform of the enzyme diacylglycerol lipase (DAGLα). To better appreciate the role of 2-AG production in normal vocal development we have determined the spatial distribution of DAGLα expression within zebra finch CNS during vocal development. Early during vocal development at 25 days, DAGLα staining is typically light and of fibroid processes. Staining peaks late in the sensorimotor stage of song learning at 75 days and is characterized by fiber, neuropil and some staining of both small and large cell somata. Results provide insight to the normal role for endocannabinoid signaling in the maturation of brain regions responsible for song learning and vocal-motor output, and suggest mechanisms by which exogenous cannabinoid exposure alters acquisition of this form of vocal communication. Copyright © 2013 Elsevier B.V. All rights reserved.

  16. Two-dimensional vocal tracts with three-dimensional behavior in the numerical generation of vowels.

    PubMed

    Arnela, Marc; Guasch, Oriol

    2014-01-01

    Two-dimensional (2D) numerical simulations of vocal tract acoustics may provide a good balance between the high quality of three-dimensional (3D) finite element approaches and the low computational cost of one-dimensional (1D) techniques. However, 2D models are usually generated by considering the 2D vocal tract as a midsagittal cut of a 3D version, i.e., using the same radius function, wall impedance, glottal flow, and radiation losses as in 3D, which leads to strong discrepancies in the resulting vocal tract transfer functions. In this work, a four step methodology is proposed to match the behavior of 2D simulations with that of 3D vocal tracts with circular cross-sections. First, the 2D vocal tract profile becomes modified to tune the formant locations. Second, the 2D wall impedance is adjusted to fit the formant bandwidths. Third, the 2D glottal flow gets scaled to recover 3D pressure levels. Fourth and last, the 2D radiation model is tuned to match the 3D model following an optimization process. The procedure is tested for vowels /a/, /i/, and /u/ and the obtained results are compared with those of a full 3D simulation, a conventional 2D approach, and a 1D chain matrix model.

  17. Sensorimotor learning in children and adults: Exposure to frequency-altered auditory feedback during speech production.

    PubMed

    Scheerer, N E; Jacobson, D S; Jones, J A

    2016-02-09

    Auditory feedback plays an important role in the acquisition of fluent speech; however, this role may change once speech is acquired and individuals no longer experience persistent developmental changes to the brain and vocal tract. For this reason, we investigated whether the role of auditory feedback in sensorimotor learning differs across children and adult speakers. Participants produced vocalizations while they heard their vocal pitch predictably or unpredictably shifted downward one semitone. The participants' vocal pitches were measured at the beginning of each vocalization, before auditory feedback was available, to assess the extent to which the deviant auditory feedback modified subsequent speech motor commands. Sensorimotor learning was observed in both children and adults, with participants' initial vocal pitch increasing following trials where they were exposed to predictable, but not unpredictable, frequency-altered feedback. Participants' vocal pitch was also measured across each vocalization, to index the extent to which the deviant auditory feedback was used to modify ongoing vocalizations. While both children and adults were found to increase their vocal pitch following predictable and unpredictable changes to their auditory feedback, adults produced larger compensatory responses. The results of the current study demonstrate that both children and adults rapidly integrate information derived from their auditory feedback to modify subsequent speech motor commands. However, these results also demonstrate that children and adults differ in their ability to use auditory feedback to generate compensatory vocal responses during ongoing vocalization. Since vocal variability also differed across the children and adult groups, these results also suggest that compensatory vocal responses to frequency-altered feedback manipulations initiated at vocalization onset may be modulated by vocal variability. Copyright © 2015 IBRO. Published by Elsevier Ltd. All rights reserved.

  18. Learned Vocal Variation Is Associated with Abrupt Cryptic Genetic Change in a Parrot Species Complex

    PubMed Central

    Ribot, Raoul F. H.; Buchanan, Katherine L.; Endler, John A.; Joseph, Leo; Bennett, Andrew T. D.; Berg, Mathew L.

    2012-01-01

    Contact zones between subspecies or closely related species offer valuable insights into speciation processes. A typical feature of such zones is the presence of clinal variation in multiple traits. The nature of these traits and the concordance among clines are expected to influence whether and how quickly speciation will proceed. Learned signals, such as vocalizations in species having vocal learning (e.g. humans, many birds, bats and cetaceans), can exhibit rapid change and may accelerate reproductive isolation between populations. Therefore, particularly strong concordance among clines in learned signals and population genetic structure may be expected, even among continuous populations in the early stages of speciation. However, empirical evidence for this pattern is often limited because differences in vocalisations between populations are driven by habitat differences or have evolved in allopatry. We tested for this pattern in a unique system where we may be able to separate effects of habitat and evolutionary history. We studied geographic variation in the vocalizations of the crimson rosella (Platycercus elegans) parrot species complex. Parrots are well known for their life-long vocal learning and cognitive abilities. We analysed contact calls across a ca 1300 km transect encompassing populations that differed in neutral genetic markers and plumage colour. We found steep clinal changes in two acoustic variables (fundamental frequency and peak frequency position). The positions of the two clines in vocal traits were concordant with a steep cline in microsatellite-based genetic variation, but were discordant with the steep clines in mtDNA, plumage and habitat. Our study provides new evidence that vocal variation, in a species with vocal learning, can coincide with areas of restricted gene flow across geographically continuous populations. Our results suggest that traits that evolve culturally can be strongly associated with reduced gene flow between populations, and therefore may promote speciation, even in the absence of other barriers. PMID:23227179

  19. Integrating perspectives on vocal performance and consistency

    PubMed Central

    Sakata, Jon T.; Vehrencamp, Sandra L.

    2012-01-01

    SUMMARY Recent experiments in divergent fields of birdsong have revealed that vocal performance is important for reproductive success and under active control by distinct neural circuits. Vocal consistency, the degree to which the spectral properties (e.g. dominant or fundamental frequency) of song elements are produced consistently from rendition to rendition, has been highlighted as a biologically important aspect of vocal performance. Here, we synthesize functional, developmental and mechanistic (neurophysiological) perspectives to generate an integrated understanding of this facet of vocal performance. Behavioral studies in the field and laboratory have found that vocal consistency is affected by social context, season and development, and, moreover, positively correlated with reproductive success. Mechanistic investigations have revealed a contribution of forebrain and basal ganglia circuits and sex steroid hormones to the control of vocal consistency. Across behavioral, developmental and mechanistic studies, a convergent theme regarding the importance of vocal practice in juvenile and adult songbirds emerges, providing a basis for linking these levels of analysis. By understanding vocal consistency at these levels, we gain an appreciation for the various dimensions of song control and plasticity and argue that genes regulating the function of basal ganglia circuits and sex steroid hormones could be sculpted by sexual selection. PMID:22189763

  20. A Primary Role for Nucleus Accumbens and Related Limbic Network in Vocal Tics.

    PubMed

    McCairn, Kevin W; Nagai, Yuji; Hori, Yukiko; Ninomiya, Taihei; Kikuchi, Erika; Lee, Ju-Young; Suhara, Tetsuya; Iriki, Atsushi; Minamimoto, Takafumi; Takada, Masahiko; Isoda, Masaki; Matsumoto, Masayuki

    2016-01-20

    Inappropriate vocal expressions, e.g., vocal tics in Tourette syndrome, severely impact quality of life. Neural mechanisms underlying vocal tics remain unexplored because no established animal model representing the condition exists. We report that unilateral disinhibition of the nucleus accumbens (NAc) generates vocal tics in monkeys. Whole-brain PET imaging identified prominent, bilateral limbic cortico-subcortical activation. Local field potentials (LFPs) developed abnormal spikes in the NAc and the anterior cingulate cortex (ACC). Vocalization could occur without obvious LFP spikes, however, when phase-phase coupling of alpha oscillations were accentuated between the NAc, ACC, and the primary motor cortex. These findings contrasted with myoclonic motor tics induced by disinhibition of the dorsolateral putamen, where PET activity was confined to the ipsilateral sensorimotor system and LFP spikes always preceded motor tics. We propose that vocal tics emerge as a consequence of dysrhythmic alpha coupling between critical nodes in the limbic and motor networks. VIDEO ABSTRACT. Copyright © 2016 Elsevier Inc. All rights reserved.

  1. Evolving role of mitomycin-C laryngology

    NASA Astrophysics Data System (ADS)

    Richards, Steven V.; Garrett, C. Gaelyn

    2001-05-01

    Topical mitomycin-C, a chemotherapeutic agent and a fibroblast inhibitor, has been successfully used in larynx, primarily to treat stenosis. Subglottic, tracheal, and anterior glottic stenosis have all shown promising results in a canine model. Less favorable results have been obtained when topical mitomycin-C is used on the vocal folds following surgical excision of mucosa. In the vocal fold studies, laryngeal videostroboscopy revealed diminished mucosal wave vibration in the vocal folds treated with mitomycin-C as well as a more atrophic appearance to the vibratory surface. The tissue treated with mitomycin-C showed fewer fibroblasts and less collagen. However, inflammatory infiltrate was not significantly different between the treated and untreated tissue. These results are consistent with the known suppression of fibroblast proliferation by mitomycin-C. In contrast to the positive effects of mitomycin-C on stenosis, the observed decrease in the healing response in the vocal fold had negative consequences on vocal fold vibratory pattern.

  2. The Human Voice in Speech and Singing

    NASA Astrophysics Data System (ADS)

    Lindblom, Björn; Sundberg, Johan

    This chapter speech describes various aspects of the human voice as a means of communication in speech and singing. From the point of view of function, vocal sounds can be regarded as the end result of a three stage process: (1) the compression of air in the respiratory system, which produces an exhalatory airstream, (2) the vibrating vocal folds' transformation of this air stream to an intermittent or pulsating air stream, which is a complex tone, referred to as the voice source, and (3) the filtering of this complex tone in the vocal tract resonator. The main function of the respiratory system is to generate an overpressure of air under the glottis, or a subglottal pressure. Section 16.1 describes different aspects of the respiratory system of significance to speech and singing, including lung volume ranges, subglottal pressures, and how this pressure is affected by the ever-varying recoil forces. The complex tone generated when the air stream from the lungs passes the vibrating vocal folds can be varied in at least three dimensions: fundamental frequency, amplitude and spectrum. Section 16.2 describes how these properties of the voice source are affected by the subglottal pressure, the length and stiffness of the vocal folds and how firmly the vocal folds are adducted. Section 16.3 gives an account of the vocal tract filter, how its form determines the frequencies of its resonances, and Sect. 16.4 gives an account for how these resonance frequencies or formants shape the vocal sounds by imposing spectrum peaks separated by spectrum valleys, and how the frequencies of these peaks determine vowel and voice qualities. The remaining sections of the chapter describe various aspects of the acoustic signals used for vocal communication in speech and singing. The syllable structure is discussed in Sect. 16.5, the closely related aspects of rhythmicity and timing in speech and singing is described in Sect. 16.6, and pitch and rhythm aspects in Sect. 16.7. The impressive control of all these acoustic characteristics of vocal signals is discussed in Sect. 16.8, while Sect. 16.9 considers expressive aspects of vocal communication.

  3. The Human Voice in Speech and Singing

    NASA Astrophysics Data System (ADS)

    Lindblom, Björn; Sundberg, Johan

    This chapter describes various aspects of the human voice as a means of communication in speech and singing. From the point of view of function, vocal sounds can be regarded as the end result of a three stage process: (1) the compression of air in the respiratory system, which produces an exhalatory airstream, (2) the vibrating vocal folds' transformation of this air stream to an intermittent or pulsating air stream, which is a complex tone, referred to as the voice source, and (3) the filtering of this complex tone in the vocal tract resonator. The main function of the respiratory system is to generate an overpressure of air under the glottis, or a subglottal pressure. Section 16.1 describes different aspects of the respiratory system of significance to speech and singing, including lung volume ranges, subglottal pressures, and how this pressure is affected by the ever-varying recoil forces. The complex tone generated when the air stream from the lungs passes the vibrating vocal folds can be varied in at least three dimensions: fundamental frequency, amplitude and spectrum. Section 16.2 describes how these properties of the voice source are affected by the subglottal pressure, the length and stiffness of the vocal folds and how firmly the vocal folds are adducted. Section 16.3 gives an account of the vocal tract filter, how its form determines the frequencies of its resonances, and Sect. 16.4 gives an account for how these resonance frequencies or formants shape the vocal sounds by imposing spectrum peaks separated by spectrum valleys, and how the frequencies of these peaks determine vowel and voice qualities. The remaining sections of the chapter describe various aspects of the acoustic signals used for vocal communication in speech and singing. The syllable structure is discussed in Sect. 16.5, the closely related aspects of rhythmicity and timing in speech and singing is described in Sect. 16.6, and pitch and rhythm aspects in Sect. 16.7. The impressive control of all these acoustic characteristics of vocal signals is discussed in Sect. 16.8, while Sect. 16.9 considers expressive aspects of vocal communication.

  4. Disrupting vagal feedback affects birdsong motor control.

    PubMed

    Méndez, Jorge M; Dall'asén, Analía G; Goller, Franz

    2010-12-15

    Coordination of different motor systems for sound production involves the use of feedback mechanisms. Song production in oscines is a well-established animal model for studying learned vocal behavior. Whereas the online use of auditory feedback has been studied in the songbird model, very little is known about the role of other feedback mechanisms. Auditory feedback is required for the maintenance of stereotyped adult song. In addition, the use of somatosensory feedback to maintain pressure during song has been demonstrated with experimentally induced fluctuations in air sac pressure. Feedback information mediating this response is thought to be routed to the central nervous system via afferent fibers of the vagus nerve. Here, we tested the effects of unilateral vagotomy on the peripheral motor patterns of song production and the acoustic features. Unilateral vagotomy caused a variety of disruptions and alterations to the respiratory pattern of song, some of which affected the acoustic structure of vocalizations. These changes were most pronounced a few days after nerve resection and varied between individuals. In the most extreme cases, the motor gestures of respiration were so severely disrupted that individual song syllables or the song motif were atypically terminated. Acoustic changes also suggest altered use of the two sound generators and upper vocal tract filtering, indicating that the disruption of vagal feedback caused changes to the motor program of all motor systems involved in song production and modification. This evidence for the use of vagal feedback by the song system with disruption of song during the first days after nerve cut provides a contrast to the longer-term effects of auditory feedback disruption. It suggests a significant role for somatosensory feedback that differs from that of auditory feedback.

  5. Disrupting vagal feedback affects birdsong motor control

    PubMed Central

    Méndez, Jorge M.; Dall'Asén, Analía G.; Goller, Franz

    2010-01-01

    Coordination of different motor systems for sound production involves the use of feedback mechanisms. Song production in oscines is a well-established animal model for studying learned vocal behavior. Whereas the online use of auditory feedback has been studied in the songbird model, very little is known about the role of other feedback mechanisms. Auditory feedback is required for the maintenance of stereotyped adult song. In addition, the use of somatosensory feedback to maintain pressure during song has been demonstrated with experimentally induced fluctuations in air sac pressure. Feedback information mediating this response is thought to be routed to the central nervous system via afferent fibers of the vagus nerve. Here, we tested the effects of unilateral vagotomy on the peripheral motor patterns of song production and the acoustic features. Unilateral vagotomy caused a variety of disruptions and alterations to the respiratory pattern of song, some of which affected the acoustic structure of vocalizations. These changes were most pronounced a few days after nerve resection and varied between individuals. In the most extreme cases, the motor gestures of respiration were so severely disrupted that individual song syllables or the song motif were atypically terminated. Acoustic changes also suggest altered use of the two sound generators and upper vocal tract filtering, indicating that the disruption of vagal feedback caused changes to the motor program of all motor systems involved in song production and modification. This evidence for the use of vagal feedback by the song system with disruption of song during the first days after nerve cut provides a contrast to the longer-term effects of auditory feedback disruption. It suggests a significant role for somatosensory feedback that differs from that of auditory feedback. PMID:21113000

  6. Form and function of long-range vocalizations in a Neotropical fossorial rodent: the Anillaco Tuco-Tuco (Ctenomys sp.)

    PubMed Central

    Valentinuzzi, Veronica S.; Zufiaurre, Emmanuel

    2016-01-01

    The underground environment poses particular communication challenges for subterranean rodents. Some loud and low-pitched acoustic signals that can travel long distances are appropriate for long-range underground communication and have been suggested to be territorial signals. Long-range vocalizations (LRVs) are important in long-distance communication in Ctenomys tuco-tucos. We characterized the LRV of the Anillaco Tuco-Tuco (Ctenomys sp.) using recordings from free-living individuals and described the behavioral context in which this vocalization was produced during laboratory staged encounters between individuals of both sexes. Long-range calls of Anillaco tuco-tucos are low-frequency, broad-band, loud, and long sounds composed by the repetition of two syllable types: series (formed by notes and soft-notes) and individual notes. All vocalizations were initiated with series, but not all had individual notes. Males were heavier than females and gave significantly lower-pitched vocalizations, but acoustic features were independent of body mass in males. The pronounced variation among individuals in the arrangement and number of syllables and the existence of three types of series (dyads, triads, and tetrads), created a diverse collection of syntactic patterns in vocalizations that would provide the opportunity to encode multiple types of information. The existence of complex syntactic patterns and the description of soft-notes represent new aspects of the vocal communication of Ctenomys. Long-distance vocalizations by Anillaco Tuco-Tucos appear to be territorial signals used mostly in male-male interactions. First, emission of LRVs resulted in de-escalation or space-keeping in male-male and male-female encounters in laboratory experiments. Second, these vocalizations were produced most frequently (in the field and in the lab) by males in our study population. Third, males produced LRVs with greater frequency during male-male encounters compared to male-female encounters. Finally, males appear to have larger home ranges that were more spatially segregated than those of females, suggesting that males may have greater need for long-distance signals that advertise their presence. Due to their apparent rarity, the function and acoustic features of LRV in female tuco-tucos remain inadequately known. PMID:27761344

  7. Genetic evidence supports song learning in the three-wattled bellbird Procnias tricarunculata (Cotingidae).

    PubMed

    Saranathan, Vinodkumar; Hamilton, Deborah; Powell, George V N; Kroodsma, Donald E; Prum, Richard O

    2007-09-01

    Vocal learning is thought to have evolved in three clades of birds (parrots, hummingbirds, and oscine passerines), and three clades of mammals (whales, bats, and primates). Behavioural data indicate that, unlike other suboscine passerines, the three-wattled bellbird Procnias tricarunculata (Cotingidae) is capable of vocal learning. Procnias tricarunculata shows conspicuous vocal ontogeny, striking geographical variation in song, and rapid temporal change in song within a population. Deprivation studies of vocal development in P. tricarunculata are impractical. Here, we report evidence from mitochondrial DNA sequences and nuclear microsatellite loci that genetic variation within and among the four allopatric breeding populations of P. tricarunculata is not congruent with variation in vocal behaviour. Sequences of the mitochondrial DNA control region document extensive haplotype sharing among localities and song types, and no phylogenetic resolution of geographical populations or behavioural groups. The vocally differentiated, allopatric breeding populations of P. tricarunculata are only weakly genetically differentiated populations, and are not distinct taxa. Mitochondrial DNA and microsatellite variation show small (2.9% and 13.5%, respectively) but significant correlation with geographical distance, but no significant residual variation by song type. Estimates of the strength of selection that would be needed to maintain the observed geographical pattern in vocal differentiation if songs were genetically based are unreasonably high, further discrediting the hypothesis of a genetic origin of vocal variation. These data support a fourth, phylogenetically independent origin of avian vocal learning in Procnias. Geographical variations in P. tricarunculata vocal behaviour are likely culturally evolved dialects.

  8. Vocal Generalization Depends on Gesture Identity and Sequence

    PubMed Central

    Sober, Samuel J.

    2014-01-01

    Generalization, the brain's ability to transfer motor learning from one context to another, occurs in a wide range of complex behaviors. However, the rules of generalization in vocal behavior are poorly understood, and it is unknown how vocal learning generalizes across an animal's entire repertoire of natural vocalizations and sequences. Here, we asked whether generalization occurs in a nonhuman vocal learner and quantified its properties. We hypothesized that adaptive error correction of a vocal gesture produced in one sequence would generalize to the same gesture produced in other sequences. To test our hypothesis, we manipulated the fundamental frequency (pitch) of auditory feedback in Bengalese finches (Lonchura striata var. domestica) to create sensory errors during vocal gestures (song syllables) produced in particular sequences. As hypothesized, error-corrective learning on pitch-shifted vocal gestures generalized to the same gestures produced in other sequential contexts. Surprisingly, generalization magnitude depended strongly on sequential distance from the pitch-shifted syllables, with greater adaptation for gestures produced near to the pitch-shifted syllable. A further unexpected result was that nonshifted syllables changed their pitch in the direction opposite from the shifted syllables. This apparently antiadaptive pattern of generalization could not be explained by correlations between generalization and the acoustic similarity to the pitch-shifted syllable. These findings therefore suggest that generalization depends on the type of vocal gesture and its sequential context relative to other gestures and may reflect an advantageous strategy for vocal learning and maintenance. PMID:24741046

  9. Reported vocal habits of first-year undergraduate musical theater majors in a preprofessional training program: a 10-year retrospective study.

    PubMed

    Donahue, Erin N; Leborgne, Wendy D; Brehm, Susan Baker; Weinrich, Barbara D

    2014-05-01

    Collegiate-level musical theater performance students are a specialized group of vocal performers, who rely on frequent and optimal voice use for their academic advancement and ultimate livelihood. The purpose of this study was to gather information to develop a greater understanding of vocal health and practice patterns of incoming collegiate-level musical theater performers. Data obtained from questionnaires completed by freshman musical theater majors were retrospectively analyzed to gather information about baseline vocal habits of the participants. Results of a questionnaire were obtained from incoming freshman musical theater students at the Cincinnati Conservatory of Music over a period of 10 years (2002-2011). One hundred eighty-eight participants (female = 90) (male = 98) with an average age of 18.28 years (standard deviation = 0.726) were included. Results specifying participants' self-reported vocal training and practice habits, vocal health and hygiene practices, and current vocal symptoms or contributing factors to potential voice problems are provided. Data obtained from the participants revealed that the potential for vocal problems exists in this group of performers, as over half of the subjects reported at least one current negative vocal symptom. The findings from this study provide information that may be useful for individuals who are involved in the training of vocal performers. Copyright © 2014 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  10. Ultrasonic vocalization changes and FOXP2 expression after experimental stroke.

    PubMed

    Doran, Sarah J; Trammel, Cassandra; Benashaski, Sharon E; Venna, Venugopal Reddy; McCullough, Louise D

    2015-04-15

    Speech impairments affect one in four stroke survivors. However, animal models of post-ischemic vocalization deficits are limited. Male mice vocalize at ultrasonic frequencies when exposed to an estrous female mouse. In this study we assessed vocalization patterns and quantity in male mice after cerebral ischemia. FOXP2, a gene associated with verbal dyspraxia in humans, with known roles in neurogenesis and synaptic plasticity, was also examined after injury. Using a transient middle cerebral artery occlusion (MCAO) model, we assessed correlates of vocal impairment at several time-points after stroke. Further, to identify possible lateralization of vocalization deficits induced by left and right hemispheric strokes were compared. Significant differences in vocalization quantity were observed between stroke and sham animals that persisted for a month after injury. Injury to the left hemisphere reduced early vocalizations more profoundly than those to the right hemisphere. Nuclear expression of Foxp2 was elevated early after stroke (at 6h), but significantly decreased 24h after injury in both the nucleus and the cytoplasm. Neuronal Foxp2 expression increased in stroke mice compared to sham animals 4 weeks after injury. This study demonstrates that quantifiable deficits in ultrasonic vocalizations (USVs) are seen after stroke. USV may be a useful tool to assess chronic behavioral recovery in murine models of stroke. Copyright © 2015 Elsevier B.V. All rights reserved.

  11. A new measure of child vocal reciprocity in children with autism spectrum disorder.

    PubMed

    Harbison, Amy L; Woynaroski, Tiffany G; Tapp, Jon; Wade, Joshua W; Warlaumont, Anne S; Yoder, Paul J

    2018-06-01

    Children's vocal development occurs in the context of reciprocal exchanges with a communication partner who models "speechlike" productions. We propose a new measure of child vocal reciprocity, which we define as the degree to which an adult vocal response increases the probability of an immediately following child vocal response. Vocal reciprocity is likely to be associated with the speechlikeness of vocal communication in young children with autism spectrum disorder (ASD). Two studies were conducted to test the utility of the new measure. The first used simulated vocal samples with randomly sequenced child and adult vocalizations to test the accuracy of the proposed index of child vocal reciprocity. The second was an empirical study of 21 children with ASD who were preverbal or in the early stages of language development. Daylong vocal samples collected in the natural environment were computer analyzed to derive the proposed index of child vocal reciprocity, which was highly stable when derived from two daylong vocal samples and was associated with speechlikeness of vocal communication. This association was significant even when controlling for chance probability of child vocalizations to adult vocal responses, probability of adult vocalizations, or probability of child vocalizations. A valid measure of children's vocal reciprocity might eventually improve our ability to predict which children are on track to develop useful speech and/or are most likely to respond to language intervention. A link to a free, publicly-available software program to derive the new measure of child vocal reciprocity is provided. Autism Res 2018, 11: 903-915. © 2018 International Society for Autism Research, Wiley Periodicals, Inc. Children and adults often engage in back-and-forth vocal exchanges. The extent to which they do so is believed to support children's early speech and language development. Two studies tested a new measure of child vocal reciprocity using computer-generated and real-life vocal samples of young children with autism collected in natural settings. The results provide initial evidence of accuracy, test-retest reliability, and validity of the new measure of child vocal reciprocity. A sound measure of children's vocal reciprocity might improve our ability to predict which children are on track to develop useful speech and/or are most likely to respond to language intervention. A free, publicly-available software program and manuals are provided. © 2018 International Society for Autism Research, Wiley Periodicals, Inc.

  12. The Vocal Repertoire of the Domesticated Zebra Finch: a Data Driven Approach to Decipher the Information-bearing Acoustic Features of Communication Signals

    PubMed Central

    Elie, Julie E.; Theunissen, Frédéric E.

    2018-01-01

    Although a universal code for the acoustic features of animal vocal communication calls may not exist, the thorough analysis of the distinctive acoustical features of vocalization categories is important not only to decipher the acoustical code for a specific species but also to understand the evolution of communication signals and the mechanisms used to produce and understand them. Here, we recorded more than 8,000 examples of almost all the vocalizations of the domesticated zebra finch, Taeniopygia guttata: vocalizations produced to establish contact, to form and maintain pair bonds, to sound an alarm, to communicate distress or to advertise hunger or aggressive intents. We characterized each vocalization type using complete representations that avoided any a priori assumptions on the acoustic code, as well as classical bioacoustics measures that could provide more intuitive interpretations. We then used these acoustical features to rigorously determine the potential information-bearing acoustical features for each vocalization type using both a novel regularized classifier and an unsupervised clustering algorithm. Vocalization categories are discriminated by the shape of their frequency spectrum and by their pitch saliency (noisy to tonal vocalizations) but not particularly by their fundamental frequency. Notably, the spectral shape of zebra finch vocalizations contains peaks or formants that vary systematically across categories and that would be generated by active control of both the vocal organ (source) and the upper vocal tract (filter). PMID:26581377

  13. Management of Vocal Nodules: A Regional Survey of Otolaryngologists and Speech-Language Pathologists.

    ERIC Educational Resources Information Center

    Allen, Marybeth S.; And Others

    1991-01-01

    Survey data from 21 otolaryngologists (70 percent return rate) and 32 speech-language pathologists (46 percent return rate) in Maine found differences in opinions between the 2 professional groups concerning referral patterns and treatment of vocal nodules in children and adults. Attitudinal problems were found to hamper a teamwork approach for…

  14. Is killer whale dialect evolution random?

    PubMed

    Filatova, Olga A; Burdin, Alexandr M; Hoyt, Erich

    2013-10-01

    The killer whale is among the few species in which cultural change accumulates over many generations, leading to cumulative cultural evolution. Killer whales have group-specific vocal repertoires which are thought to be learned rather than being genetically coded. It is supposed that divergence between vocal repertoires of sister groups increases gradually over time due to random learning mistakes and innovations. In this case, the similarity of calls across groups must be correlated with pod relatedness and, consequently, with each other. In this study we tested this prediction by comparing the patterns of call similarity between matrilines of resident killer whales from Eastern Kamchatka. We calculated the similarity of seven components from three call types across 14 matrilines. In contrast to the theoretical predictions, matrilines formed different clusters on the dendrograms made by different calls and even by different components of the same call. We suggest three possible explanations for this phenomenon. First, the lack of agreement between similarity patterns of different components may be the result of constraints in the call structure. Second, it is possible that call components change in time with different speed and/or in different directions. Third, horizontal cultural transmission of call features may occur between matrilines. Copyright © 2013 Elsevier B.V. All rights reserved.

  15. Geographical variation of St. Lucia Parrot flight vocalizations

    USGS Publications Warehouse

    Kleeman, Patrick M.; Gilardi, James D.

    2005-01-01

    Parrots are vocal learners and many species of parrots are capable of learning new calls, even as adults. This capability gives parrots the potential to develop communication systems that can vary dramatically over space. St. Lucia Parrot (Amazona versicolor) flight vocalizations were examined for geographic variation between four different sites on the island of St. Lucia. Spectrographic cross-correlation analysis of a commonly used flight vocalization, the p-chow call, demonstrated quantitative differences between sites. Additionally, the similarity of p-chows decreased as the distance between sites increased. Flight call repertoires also differed among sites; parrots at the Des Bottes and Quilesse sites each used one flight call unique to those sites, while parrots at the Barre de L'Isle site used a flight call that Quilesse parrots gave only while perched. It is unclear whether the vocal variation changed clinally with distance, or whether there were discrete dialect boundaries as in a congener, the Yellow-naped Parrot (Amazona auropalliata, Wright 1996). The geographical scale over which the St. Lucia Parrot's vocal variation occurred was dramatically smaller than that of the Yellow-naped Parrot. Similar patterns of fine-scale vocal variation may be more widespread among other parrot species in the Caribbean than previously documented.

  16. Patterns of call communication between group-housed zebra finches change during the breeding cycle.

    PubMed

    Gill, Lisa F; Goymann, Wolfgang; Ter Maat, Andries; Gahr, Manfred

    2015-10-06

    Vocal signals such as calls play a crucial role for survival and successful reproduction, especially in group-living animals. However, call interactions and call dynamics within groups remain largely unexplored because their relation to relevant contexts or life-history stages could not be studied with individual-level resolution. Using on-bird microphone transmitters, we recorded the vocalisations of individual zebra finches (Taeniopygia guttata) behaving freely in social groups, while females and males previously unknown to each other passed through different stages of the breeding cycle. As birds formed pairs and shifted their reproductive status, their call repertoire composition changed. The recordings revealed that calls occurred non-randomly in fine-tuned vocal interactions and decreased within groups while pair-specific patterns emerged. Call-type combinations of vocal interactions changed within pairs and were associated with successful egg-laying, highlighting a potential fitness relevance of calling dynamics in communication systems.

  17. Experimental and numerical investigation of the sound generation mechanisms of sibilant fricatives using a simplified vocal tract model

    NASA Astrophysics Data System (ADS)

    Yoshinaga, Tsukasa; Nozaki, Kazunori; Wada, Shigeo

    2018-03-01

    The sound generation mechanisms of sibilant fricatives were investigated with experimental measurements and large-eddy simulations using a simplified vocal tract model. The vocal tract geometry was simplified to a three-dimensional rectangular channel, and differences in the geometries while pronouncing fricatives /s/ and /∫/ were expressed by shifting the position of the tongue and its constricted flow channel. Experimental results showed that the characteristic peak frequency of the fricatives decreased when the distance between the tongue and teeth increased. Numerical simulations revealed that the jet flow generated from the constriction impinged on the upper teeth wall and caused the main sound source upstream and downstream from the gap between the teeth. While magnitudes of the sound source decreased with increments of the frequency, amplitudes of the pressure downstream from the constriction increased at the peak frequencies of the corresponding tongue position. These results indicate that the sound pressures at the peak frequencies increased by acoustic resonance in the channel downstream from the constriction, and the different frequency characteristics between /s/ and /∫/ were produced by changing the constriction and the acoustic node positions inside the vocal tract.

  18. Communication patterns within a group of shelter dogs and implications for their welfare.

    PubMed

    Petak, Irena

    2013-01-01

    Keeping shelter dogs in groups provides them with a more socially and physically enriched environment, but eventually it may cause them stress. Understanding dogs' communication could help shelter staff recognize and prevent undesirable communicative patterns and encourage desirable ones. Therefore, the objective of this study was to determine communication patterns in a group of dogs in a shelter. The observed dogs were engaged in different classes of dyadic and group interactions. Certain dogs were frequently initiators of dyadic interactions, and different dogs were the recipients. The predominant form of dyadic interactions was a neutral one, and aggressive behavior was rarely observed. The tendency of certain dogs to interact continuously may represent a nuisance for less social individuals. All of the dogs participated in 3 defined classes of group interactions. At the group level, the dogs frequently interact vocally or olfactorily. A major welfare problem may be very vocal dogs because their vocalizations are noisy and broadcast far-reaching signals. The frequency of some group interactions was reduced by the amount of time the dogs had in the shelter.

  19. A method for assessing the regional vibratory pattern of vocal folds by analysing the video recording of stroboscopy.

    PubMed

    Lee, J S; Kim, E; Sung, M W; Kim, K H; Sung, M Y; Park, K S

    2001-05-01

    Stroboscopy and kymography have been used to examine the motional abnormality of vocal folds and to visualise their regional vibratory pattern. In a previous study (Laryngoscope, 1999), we introduced the conceptual idea of videostrobokymography, in which we applied the concept of kymography on the pre-recorded video images using stroboscopy, and showed its possible clinical application to various disorders in vocal folds. However, a more detailed description about the software and the mathematical formulation used in this system is needed for the reproduction of similar systems. The composition of hardwares, user-interface and detail procedures including mathematical equations in videostrobokymography software is presented in this study. As an initial clinical trial, videostrobokymography was applied to the preoperative and postoperative videostroboscopic images of 15 patients with Reinke's edema. On preoperative examination, videostrobokymograms showed irregular pattern of mucosal wave and, in some patients, a relatively constant glottic gap during phonation. After the operation, the voice quality of all patients was improved in acoustic and aerodynamic assessments, and videostrobokymography showed clearly improved mucosal waves (change in open quotient: mean +/- SD= 0.11 +/- 0.05).

  20. Precise auditory-vocal mirroring in neurons for learned vocal communication.

    PubMed

    Prather, J F; Peters, S; Nowicki, S; Mooney, R

    2008-01-17

    Brain mechanisms for communication must establish a correspondence between sensory and motor codes used to represent the signal. One idea is that this correspondence is established at the level of single neurons that are active when the individual performs a particular gesture or observes a similar gesture performed by another individual. Although neurons that display a precise auditory-vocal correspondence could facilitate vocal communication, they have yet to be identified. Here we report that a certain class of neurons in the swamp sparrow forebrain displays a precise auditory-vocal correspondence. We show that these neurons respond in a temporally precise fashion to auditory presentation of certain note sequences in this songbird's repertoire and to similar note sequences in other birds' songs. These neurons display nearly identical patterns of activity when the bird sings the same sequence, and disrupting auditory feedback does not alter this singing-related activity, indicating it is motor in nature. Furthermore, these neurons innervate striatal structures important for song learning, raising the possibility that singing-related activity in these cells is compared to auditory feedback to guide vocal learning.

  1. Computational Modeling of Fluid–Structure–Acoustics Interaction during Voice Production

    PubMed Central

    Jiang, Weili; Zheng, Xudong; Xue, Qian

    2017-01-01

    The paper presented a three-dimensional, first-principle based fluid–structure–acoustics interaction computer model of voice production, which employed a more realistic human laryngeal and vocal tract geometries. Self-sustained vibrations, important convergent–divergent vibration pattern of the vocal folds, and entrainment of the two dominant vibratory modes were captured. Voice quality-associated parameters including the frequency, open quotient, skewness quotient, and flow rate of the glottal flow waveform were found to be well within the normal physiological ranges. The analogy between the vocal tract and a quarter-wave resonator was demonstrated. The acoustic perturbed flux and pressure inside the glottis were found to be at the same order with their incompressible counterparts, suggesting strong source–filter interactions during voice production. Such high fidelity computational model will be useful for investigating a variety of pathological conditions that involve complex vibrations, such as vocal fold paralysis, vocal nodules, and vocal polyps. The model is also an important step toward a patient-specific surgical planning tool that can serve as a no-risk trial and error platform for different procedures, such as injection of biomaterials and thyroplastic medialization. PMID:28243588

  2. Perceptual connections between prepubertal children's voices in their speaking behavior and their singing behavior.

    PubMed

    Rinta, Tiija Elisabet; Welch, Graham F

    2009-11-01

    Traditionally, children's speaking and singing behaviors have been regarded as two separate sets of behaviors. Nevertheless, according to the voice-scientific view, all vocal functioning is interconnected due to the fact that we exploit the same voice and the same physiological mechanisms in generating all vocalization. The intention of the study was to investigate whether prepubertal children's speaking and singing behaviors are connected perceptually. Voice recordings were conducted with 60 10-year-old children. Each child performed a set of speaking and singing tasks in the voice experiments. Each voice sample was analyzed perceptually with a specially designed perceptual voice assessment protocol. The main finding was that the children's vocal functioning and voice quality in their speaking behavior correlated statistically significantly with those in their singing behavior. The findings imply that children's speaking and singing behaviors are perceptually connected through their vocal functioning and voice quality. Thus, it can be argued that children possess one voice that is used for generating their speaking and singing behaviors.

  3. Paradigms and progress in vocal fold restoration.

    PubMed

    Ford, Charles N

    2008-09-01

    Science advances occur through orderly steps, puzzle-solving leaps, or divergences from the accepted disciplinary matrix that occasionally result in a revolutionary paradigm shift. Key advances must overcome bias, criticism, and rejection. Examples in biological science include use of embryonic stem cells, recognition of Helicobacter pylori in the etiology of ulcer disease, and the evolution of species. Our work in vocal fold restoration reflects these patterns. We progressed through phases of tissue replacement with fillers and biological implants, to current efforts at vocal fold regeneration through tissue engineering, and face challenges of a new "systems biology" paradigm embracing genomics and proteomics.

  4. Experimental analysis of the characteristics of artificial vocal folds.

    PubMed

    Misun, Vojtech; Svancara, Pavel; Vasek, Martin

    2011-05-01

    Specialized literature presents a number of models describing the function of the vocal folds. In most of those models, an emphasis is placed on the air flowing through the glottis and, further, on the effect of the parameters of the air alone (its mass, speed, and so forth). The article focuses on the constructional definition of artificial vocal folds and their experimental analysis. The analysis is conducted for voiced source voice phonation and for the changing mean value of the subglottal pressure. The article further deals with the analysis of the pressure of the airflow through the vocal folds, which is cut (separated) into individual pulses by the vibrating vocal folds. The analysis results show that air pulse characteristics are relevant to voice generation, as they are produced by the flowing air and vibrating vocal folds. A number of artificial vocal folds have been constructed to date, and the aforementioned view of their phonation is confirmed by their analysis. The experiments have confirmed that man is able to consciously affect only two parameters of the source voice, that is, its fundamental frequency and voice intensity. The main forces acting on the vocal folds during phonation are as follows: subglottal air pressure and elastic and inertia forces of the vocal folds' structure. The correctness of the function of the artificial vocal folds is documented by the experimental verification of the spectra of several types of artificial vocal folds. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  5. Spontaneous motor entrainment to music in multiple vocal mimicking species.

    PubMed

    Schachner, Adena; Brady, Timothy F; Pepperberg, Irene M; Hauser, Marc D

    2009-05-26

    The human capacity for music consists of certain core phenomena, including the tendency to entrain, or align movement, to an external auditory pulse [1-3]. This ability, fundamental both for music production and for coordinated dance, has been repeatedly highlighted as uniquely human [4-11]. However, it has recently been hypothesized that entrainment evolved as a by-product of vocal mimicry, generating the strong prediction that only vocal mimicking animals may be able to entrain [12, 13]. Here we provide comparative data demonstrating the existence of two proficient vocal mimicking nonhuman animals (parrots) that entrain to music, spontaneously producing synchronized movements resembling human dance. We also provide an extensive comparative data set from a global video database systematically analyzed for evidence of entrainment in hundreds of species both capable and incapable of vocal mimicry. Despite the higher representation of vocal nonmimics in the database and comparable exposure of mimics and nonmimics to humans and music, only vocal mimics showed evidence of entrainment. We conclude that entrainment is not unique to humans and that the distribution of entrainment across species supports the hypothesis that entrainment evolved as a by-product of selection for vocal mimicry.

  6. Vocal production mechanisms in a non-human primate: morphological data and a model.

    PubMed

    Riede, Tobias; Bronson, Ellen; Hatzikirou, Haralambos; Zuberbühler, Klaus

    2005-01-01

    Human beings are thought to be unique amongst the primates in their capacity to produce rapid changes in the shape of their vocal tracts during speech production. Acoustically, vocal tracts act as resonance chambers, whose geometry determines the position and bandwidth of the formants. Formants provide the acoustic basis for vowels, which enable speakers to refer to external events and to produce other kinds of meaningful communication. Formant-based referential communication is also present in non-human primates, most prominently in Diana monkey alarm calls. Previous work has suggested that the acoustic structure of these calls is the product of a non-uniform vocal tract capable of some degree of articulation. In this study we test this hypothesis by providing morphological measurements of the vocal tract of three adult Diana monkeys, using both radiography and dissection. We use these data to generate a vocal tract computational model capable of simulating the formant structures produced by wild individuals. The model performed best when it combined a non-uniform vocal tract consisting of three different tubes with a number of articulatory manoeuvres. We discuss the implications of these findings for evolutionary theories of human and non-human vocal production.

  7. Homeostasis of Hyaluronic Acid in Normal and Scarred Vocal Folds

    PubMed Central

    Tateya, Ichiro; Tateya, Tomoko; Watanuki, Makoto; Bless, Diane M.

    2015-01-01

    Summary Objectives/Hypothesis Vocal fold scarring is one of the most challenging laryngeal disorders to treat. Hyaluronic acid (HA) is the main component of lamina propria, and it plays an important role in proper vocal fold vibration and is also thought to be important in fetal wound healing without scarring. Although several animal models of vocal fold scarring have been reported, little is known about the way in which HA is maintained in vocal folds. The purpose of this study was to clarify the homeostasis of HA by examining the expression of hyaluronan synthase (Has) and hyaluronidase (Hyal), which produce and digest HA, respectively. Study Design Experimental prospective animal study. Methods Vocal fold stripping was performed on 38 Sprague-Dawley rats. Vocal fold tissue was collected at five time points (3 days–2 months). Expression of HA was examined by immunohistochemistry, and messenger RNA (mRNA) expression of Has and Hyal was examined by real-time polymerase chain reaction and in-situ hybridization. Results In scarred vocal folds, expression of Has1 and Has2 increased at day 3 together with expression of HA and returned to normal at 2 weeks. At 2 months, Has3 and Hyal3 mRNA showed higher expressions than normal. Conclusions Expression patterns of Has and Hyal genes differed between normal, acute-scarred, and chronic-scarred vocal folds, indicating the distinct roles of each enzyme in maintaining HA. Continuous upregulation of Has genes in the acute phase may be necessary to achieve scarless healing of vocal folds. PMID:25499520

  8. Strain Modulations as a Mechanism to Reduce Stress Relaxation in Laryngeal Tissues

    PubMed Central

    Hunter, Eric J.; Siegmund, Thomas; Chan, Roger W.

    2014-01-01

    Vocal fold tissues in animal and human species undergo deformation processes at several types of loading rates: a slow strain involved in vocal fold posturing (on the order of 1 Hz or so), cyclic and faster posturing often found in speech tasks or vocal embellishment (1–10 Hz), and shear strain associated with vocal fold vibration during phonation (100 Hz and higher). Relevant to these deformation patterns are the viscous properties of laryngeal tissues, which exhibit non-linear stress relaxation and recovery. In the current study, a large strain time-dependent constitutive model of human vocal fold tissue is used to investigate effects of phonatory posturing cyclic strain in the range of 1 Hz to 10 Hz. Tissue data for two subjects are considered and used to contrast the potential effects of age. Results suggest that modulation frequency and extent (amplitude), as well as the amount of vocal fold overall strain, all affect the change in stress relaxation with modulation added. Generally, the vocal fold cover reduces the rate of relaxation while the opposite is true for the vocal ligament. Further, higher modulation frequencies appear to reduce the rate of relaxation, primarily affecting the ligament. The potential benefits of cyclic strain, often found in vibrato (around 5 Hz modulation) and intonational inflection, are discussed in terms of vocal effort and vocal pitch maintenance. Additionally, elderly tissue appears to not exhibit these benefits to modulation. The exacerbating effect such modulations may have on certain voice disorders, such as muscle tension dysphonia, are explored. PMID:24614616

  9. Strain modulations as a mechanism to reduce stress relaxation in laryngeal tissues.

    PubMed

    Hunter, Eric J; Siegmund, Thomas; Chan, Roger W

    2014-01-01

    Vocal fold tissues in animal and human species undergo deformation processes at several types of loading rates: a slow strain involved in vocal fold posturing (on the order of 1 Hz or so), cyclic and faster posturing often found in speech tasks or vocal embellishment (1-10 Hz), and shear strain associated with vocal fold vibration during phonation (100 Hz and higher). Relevant to these deformation patterns are the viscous properties of laryngeal tissues, which exhibit non-linear stress relaxation and recovery. In the current study, a large strain time-dependent constitutive model of human vocal fold tissue is used to investigate effects of phonatory posturing cyclic strain in the range of 1 Hz to 10 Hz. Tissue data for two subjects are considered and used to contrast the potential effects of age. Results suggest that modulation frequency and extent (amplitude), as well as the amount of vocal fold overall strain, all affect the change in stress relaxation with modulation added. Generally, the vocal fold cover reduces the rate of relaxation while the opposite is true for the vocal ligament. Further, higher modulation frequencies appear to reduce the rate of relaxation, primarily affecting the ligament. The potential benefits of cyclic strain, often found in vibrato (around 5 Hz modulation) and intonational inflection, are discussed in terms of vocal effort and vocal pitch maintenance. Additionally, elderly tissue appears to not exhibit these benefits to modulation. The exacerbating effect such modulations may have on certain voice disorders, such as muscle tension dysphonia, are explored.

  10. Air Pressure Responses to Sudden Vocal Tract Pressure Bleeds during Production of Stop Consonants: New Evidence of Aeromechanical Regulation.

    ERIC Educational Resources Information Center

    Zajac, David J.; Weissler, Mark C.

    2004-01-01

    Two studies were conducted to evaluate short-latency vocal tract air pressure responses to sudden pressure bleeds during production of voiceless bilabial stop consonants. It was hypothesized that the occurrence of respiratory reflexes would be indicated by distinct patterns of responses as a function of bleed magnitude. In Study 1, 19 adults…

  11. Nonconscious Influence of Masked Stimuli on Response Selection Is Limited to Concrete Stimulus-Response Associations

    ERIC Educational Resources Information Center

    Klapp, Stuart T.; Haas, Brian W.

    2005-01-01

    A pattern-masked arrow negatively biased the "free choice" between 2 manual responses or between 2 vocal responses. This apparently nonconscious influence occurred only when the free-choice trials were intermixed randomly with other trials that terminated in fully visible arrows, which directed a response of the same modality (manual vs. vocal) as…

  12. The Effects of Gesture and Movement Training on the Intonation of Children's Singing in Vocal Warm-Up Sessions

    ERIC Educational Resources Information Center

    Liao, Mei-Ying; Davidson, Jane W.

    2016-01-01

    The main purpose of the current study was to examine the effects of gesture and movement training for beginning children's choirs with regard to improving intonation. It was a between-subjects design with one independent variable Training Technique (TT). One dependent variable was measured: intonation in the singing of vocal pattern warm-up…

  13. Vocal clans in sperm whales (Physeter macrocephalus).

    PubMed Central

    Rendell, L E; Whitehead, H

    2003-01-01

    Cultural transmission may be a significant source of variation in the behaviour of whales and dolphins, especially as regards their vocal signals. We studied variation in the vocal output of 'codas' by sperm whale social groups. Codas are patterns of clicks used by female sperm whales in social circumstances. The coda repertoires of all known social units (n = 18, each consisting of about 11 females and immatures with long-term relationships) and 61 out of 64 groups (about two social units moving together for periods of days) that were recorded in the South Pacific and Caribbean between 1985 and 2000 can be reliably allocated into six acoustic 'clans', five in the Pacific and one in the Caribbean. Clans have ranges that span thousands of kilometres, are sympatric, contain many thousands of whales and most probably result from cultural transmission of vocal patterns. Units seem to form groups preferentially with other units of their own clan. We suggest that this is a rare example of sympatric cultural variation on an oceanic scale. Culture may thus be a more important determinant of sperm whale population structure than genes or geography, a finding that has major implications for our understanding of the species' behavioural and population biology. PMID:12614570

  14. Mobile voice health monitoring using a wearable accelerometer sensor and a smartphone platform.

    PubMed

    Mehta, Daryush D; Zañartu, Matías; Feng, Shengran W; Cheyne, Harold A; Hillman, Robert E

    2012-11-01

    Many common voice disorders are chronic or recurring conditions that are likely to result from faulty and/or abusive patterns of vocal behavior, referred to generically as vocal hyperfunction. An ongoing goal in clinical voice assessment is the development and use of noninvasively derived measures to quantify and track the daily status of vocal hyperfunction so that the diagnosis and treatment of such behaviorally based voice disorders can be improved. This paper reports on the development of a new, versatile, and cost-effective clinical tool for mobile voice monitoring that acquires the high-bandwidth signal from an accelerometer sensor placed on the neck skin above the collarbone. Using a smartphone as the data acquisition platform, the prototype device provides a user-friendly interface for voice use monitoring, daily sensor calibration, and periodic alert capabilities. Pilot data are reported from three vocally normal speakers and three subjects with voice disorders to demonstrate the potential of the device to yield standard measures of fundamental frequency and sound pressure level and model-based glottal airflow properties. The smartphone-based platform enables future clinical studies for the identification of the best set of measures for differentiating between normal and hyperfunctional patterns of voice use.

  15. Mobile voice health monitoring using a wearable accelerometer sensor and a smartphone platform

    PubMed Central

    Mehta, Daryush D.; Zañartu, Matías; Feng, Shengran W.; Cheyne, Harold A.; Hillman, Robert E.

    2012-01-01

    Many common voice disorders are chronic or recurring conditions that are likely to result from faulty and/or abusive patterns of vocal behavior, referred to generically as vocal hyperfunction. An ongoing goal in clinical voice assessment is the development and use of noninvasively derived measures to quantify and track the daily status of vocal hyperfunction so that the diagnosis and treatment of such behaviorally based voice disorders can be improved. This paper reports on the development of a new, versatile, and cost-effective clinical tool for mobile voice monitoring that acquires the high-bandwidth signal from an accelerometer sensor placed on the neck skin above the collarbone. Using a smartphone as the data acquisition platform, the prototype device provides a user-friendly interface for voice use monitoring, daily sensor calibration, and periodic alert capabilities. Pilot data are reported from three vocally normal speakers and three subjects with voice disorders to demonstrate the potential of the device to yield standard measures of fundamental frequency and sound pressure level and model-based glottal airflow properties. The smartphone-based platform enables future clinical studies for the identification of the best set of measures for differentiating between normal and hyperfunctional patterns of voice use. PMID:22875236

  16. Tourette Syndrome

    PubMed Central

    Murray, T. J.

    1982-01-01

    Tourette syndrome (Gilles de la Tourette disease) is a disorder of involuntary muscular tics, vocalizations and compulsive behavior. The tics and muscle movements vary in form and course; the complex repetitive patterns are eventually replaced by other patterns. The vocalization may be in the form of sounds, words or profanities and sometimes echolalia, echopraxia and palilalia. The onset may be from age two to 15 but is usually between ages eight and 12. Recent studies suggest that there is a hypersensitivity of dopamine receptors. Most patients respond well to haloperidol, but other drugs that may be of value include clonidine, pimozide, fluphenazine and trifluoroperazine. PMID:21286050

  17. Three-dimensional optical reconstruction of vocal fold kinematics using high-speed video with a laser projection system

    PubMed Central

    Luegmair, Georg; Mehta, Daryush D.; Kobler, James B.; Döllinger, Michael

    2015-01-01

    Vocal fold kinematics and its interaction with aerodynamic characteristics play a primary role in acoustic sound production of the human voice. Investigating the temporal details of these kinematics using high-speed videoendoscopic imaging techniques has proven challenging in part due to the limitations of quantifying complex vocal fold vibratory behavior using only two spatial dimensions. Thus, we propose an optical method of reconstructing the superior vocal fold surface in three spatial dimensions using a high-speed video camera and laser projection system. Using stereo-triangulation principles, we extend the camera-laser projector method and present an efficient image processing workflow to generate the three-dimensional vocal fold surfaces during phonation captured at 4000 frames per second. Initial results are provided for airflow-driven vibration of an ex vivo vocal fold model in which at least 75% of visible laser points contributed to the reconstructed surface. The method captures the vertical motion of the vocal folds at a high accuracy to allow for the computation of three-dimensional mucosal wave features such as vibratory amplitude, velocity, and asymmetry. PMID:26087485

  18. Contribution to the understanding of the etiology of vocal fold cysts: a functional and histologic study.

    PubMed

    Milutinović, Z; Vasiljević, J

    1992-05-01

    The etiological theories of vocal fold cysts can be divided into two basic groups: those of congenital and acquired cysts. In ongoing practice, the authors had noted that the greater number of cysts appeared at the functionally most active segment of the vocal folds which, on the other hand, has the least number of glands. Also, it had been noted that patients with vocal fold cysts tended to have hyperkinetic patterns of voice production. These observations indicated the possibility of a functional aspect in the etiology of vocal fold cysts, and consideration of such a possibility was the aim of this work. In 37 cases, the exact location of the cyst was established. In addition, the muscular activity of the phonatory apparatus was estimated, patient self-descriptions with respect to talkativeness were taken into account, and histological evaluations were made. The cysts were most frequently found in the area of the junction of the anterior and middle thirds of the free edge of the vocal fold. Muscular activity during speech and phonation was increased in study patients. Sixty-five percent of patients had epidermoid cysts and 35% had retention cysts of the vocal fold. According to study results, the functional aspect of cyst genesis has a marked role in the etiology of vocal fold cysts, which points to the great importance of functional care for cyst patients.

  19. Listen up! Processing of intensity change differs for vocal and nonvocal sounds.

    PubMed

    Schirmer, Annett; Simpson, Elizabeth; Escoffier, Nicolas

    2007-10-24

    Changes in the intensity of both vocal and nonvocal sounds can be emotionally relevant. However, as only vocal sounds directly reflect communicative intent, intensity change of vocal but not nonvocal sounds is socially relevant. Here we investigated whether a change in sound intensity is processed differently depending on its social relevance. To this end, participants listened passively to a sequence of vocal or nonvocal sounds that contained rare deviants which differed from standards in sound intensity. Concurrently recorded event-related potentials (ERPs) revealed a mismatch negativity (MMN) and P300 effect for intensity change. Direction of intensity change was of little importance for vocal stimulus sequences, which recruited enhanced sensory and attentional resources for both loud and soft deviants. In contrast, intensity change in nonvocal sequences recruited more sensory and attentional resources for loud as compared to soft deviants. This was reflected in markedly larger MMN/P300 amplitudes and shorter P300 latencies for the loud as compared to soft nonvocal deviants. Furthermore, while the processing pattern observed for nonvocal sounds was largely comparable between men and women, sex differences for vocal sounds suggest that women were more sensitive to their social relevance. These findings extend previous evidence of sex differences in vocal processing and add to reports of voice specific processing mechanisms by demonstrating that simple acoustic change recruits more processing resources if it is socially relevant.

  20. Identifying Knowledge Gaps in Clinicians Who Evaluate and Treat Vocal Performing Artists in College Health Settings.

    PubMed

    McKinnon-Howe, Leah; Dowdall, Jayme

    2018-05-01

    The goal of this study was to identify knowledge gaps in clinicians who evaluate and treat performing artists for illnesses and injuries that affect vocal function in college health settings. This pilot study utilized a web-based cross-sectional survey design incorporating common clinical scenarios to test knowledge of evaluation and management strategies in the vocal performing artist. A web-based survey was administered to a purposive sample of 28 clinicians to identify the approach utilized to evaluate and treat vocal performing artists in college health settings, and factors that might affect knowledge gaps and influence referral patterns to voice specialists. Twenty-eight clinicians were surveyed, with 36% of respondents incorrectly identifying appropriate vocal hygiene measures, 56% of respondents failing to identify symptoms of vocal fold hemorrhage, 84% failing to identify other indications for referral to a voice specialist, 96% of respondents acknowledging unfamiliarity with the Voice Handicap Index and the Singers Voice Handicap Index, and 68% acknowledging unfamiliarity with the Reflux Symptom Index. The data elucidated specific knowledge gaps in college health providers who are responsible for evaluating and treating common illnesses that affect vocal function, and triaging and referring students experiencing symptoms of potential vocal emergencies. Future work is needed to improve the standard of care for this population. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  1. Characterization of ultrasonic vocalizations of Fragile X mice.

    PubMed

    Belagodu, Amogh P; Johnson, Aaron M; Galvez, Roberto

    2016-09-01

    Fragile X Syndrome (FXS) is the leading form of inherited intellectual disability. It is caused by the transcriptional silencing of FMR1, the gene which codes for the Fragile X Mental Retardation Protein (FMRP). Patients who have FXS exhibit numerous behavioral and cognitive impairments, such as attention-deficit/hyperactivity disorder, obsessive compulsive disorder, and autistic-like behaviors. In addition to these behavioral abnormalities, FXS patients have also been shown to exhibit various deficits in communication such as abnormal sentence structures, increased utterances, repetition of sounds and words, and reduced articulation. These deficits can dramatically hinder communication for FXS patients, exacerbating learning and cognition impairments while decreasing their quality of life. To examine the biological underpinnings of these communication abnormalities, studies have used a mouse model of the Fragile X Syndrome; however, these vocalization studies have resulted in inconsistent findings that often do not correlate with abnormalities observed in FXS patients. Interestingly, a detailed examination of frequency modulated vocalizations that are believed to be a better assessment of rodent communication has never been conducted. The following study used courtship separation to conduct a detailed examination of frequency modulated ultrasonic vocalizations (USV) in FXS mice. Our analyses of frequency modulated USVs demonstrated that adult FXS mice exhibited longer phrases and more motifs. Phrases are vocalizations consisting of multiple frequency modulated ultrasonic vocalizations, while motifs are repeated frequency modulated USV patterns. Fragile X mice had a higher proportion of "u" syllables in all USVs and phrases while their wildtype counterparts preferred isolated "h" syllables. Although the specific importance of these syllables towards communication deficits still needs to be evaluated, these findings in production of USVs are consistent with the repetitive and perseverative speech patterns observed in FXS patients. This study demonstrates that FXS mice can be used to study the underlying biological mechanism(s) mediating FXS vocalization abnormalities. Copyright © 2016 Elsevier B.V. All rights reserved.

  2. Application of motion analysis in the study of the effect of botulinum toxin to rat vocal folds

    NASA Astrophysics Data System (ADS)

    Saadah, Abdul K.; Galatsanos, Nikolas P.; Inagi, K.; Bless, D.

    1997-05-01

    In the past we have proposed a system that measures the deformations of the vocal folds from videostroboscopic images of the larynx, in that system: (1) we extract the boundaries of the vocal folds, (2) we register elastically the vocal fold boundaries in successive frames. This yields the displacement vector field (DVF) between adjacent frames, and (3) we fit using a least-squares approach an affine transformation model to succinctly describe the deformations between adjacent frames. In this paper, we present as an example of the capabilities of this system, an initial study of the deformation changes in rat vocal folds pre and post injection with Botulinum toxin. For this application the generated DVF was segmented into right DVF and left DVF and the deformation of each segment is studied separately.

  3. Vocal mechanics in Darwin's finches: correlation of beak gape and song frequency.

    PubMed

    Podos, Jeffrey; Southall, Joel A; Rossi-Santos, Marcos R

    2004-02-01

    Recent studies of vocal mechanics in songbirds have identified a functional role for the beak in sound production. The vocal tract (trachea and beak) filters harmonic overtones from sounds produced by the syrinx, and birds can fine-tune vocal tract resonance properties through changes in beak gape. In this study, we examine patterns of beak gape during song production in seven species of Darwin's finches of the Galápagos Islands. Our principal goals were to characterize the relationship between beak gape and vocal frequency during song production and to explore the possible influence therein of diversity in beak morphology and body size. Birds were audio and video recorded (at 30 frames s(-1)) as they sang in the field, and 164 song sequences were analyzed. We found that song frequency regressed significantly and positively on beak gape for 38 of 56 individuals and for all seven species examined. This finding provides broad support for a resonance model of vocal tract function in Darwin's finches. Comparison among species revealed significant variation in regression y-intercept values. Body size correlated negatively with y-intercept values, although not at a statistically significant level. We failed to detect variation in regression slopes among finch species, although the regression slopes of Darwin's finch and two North American sparrow species were found to differ. Analysis within one species (Geospiza fortis) revealed significant inter-individual variation in regression parameters; these parameters did not correlate with song frequency features or plumage scores. Our results suggest that patterns of beak use during song production were conserved during the Darwin's finch adaptive radiation, despite the evolution of substantial variation in beak morphology and body size.

  4. Population structure of humpback whales in the western and central South Pacific Ocean as determined by vocal exchange among populations.

    PubMed

    Garland, Ellen C; Goldizen, Anne W; Lilley, Matthew S; Rekdahl, Melinda L; Garrigue, Claire; Constantine, Rochelle; Hauser, Nan Daeschler; Poole, M Michael; Robbins, Jooke; Noad, Michael J

    2015-08-01

    For cetaceans, population structure is traditionally determined by molecular genetics or photographically identified individuals. Acoustic data, however, has provided information on movement and population structure with less effort and cost than traditional methods in an array of taxa. Male humpback whales (Megaptera novaeangliae) produce a continually evolving vocal sexual display, or song, that is similar among all males in a population. The rapid cultural transmission (the transfer of information or behavior between conspecifics through social learning) of different versions of this display between distinct but interconnected populations in the western and central South Pacific region presents a unique way to investigate population structure based on the movement dynamics of a song (acoustic) display. Using 11 years of data, we investigated an acoustically based population structure for the region by comparing stereotyped song sequences among populations and years. We used the Levenshtein distance technique to group previously defined populations into (vocally based) clusters based on the overall similarity of their song display in space and time. We identified the following distinct vocal clusters: western cluster, 1 population off eastern Australia; central cluster, populations around New Caledonia, Tonga, and American Samoa; and eastern region, either a single cluster or 2 clusters, one around the Cook Islands and the other off French Polynesia. These results are consistent with the hypothesis that each breeding aggregation represents a distinct population (each occupied a single, terminal node) in a metapopulation, similar to the current understanding of population structure based on genetic and photo-identification studies. However, the central vocal cluster had higher levels of song-sharing among populations than the other clusters, indicating that levels of vocal connectivity varied within the region. Our results demonstrate the utility and value of using culturally transmitted vocal patterns as a way of defining connectivity to infer population structure. We suggest vocal patterns be incorporated by the International Whaling Commission in conjunction with traditional methods in the assessment of structure. © 2015, Society for Conservation Biology.

  5. [Differences in vocalization and morphology of the syrinx between Carrion crows (Corvus corone) and Jungle crows (C. macrorhynchos)].

    PubMed

    Tsukahara, Naoki; Aoyama, Masato; Sugita, Shoei

    2007-12-01

    The vocal characteristics and the morph of the syrinx in Carrion crows (Corvus corone) and those in Jungle crows (C. macrorhynchos) were compared. The vocalizations of both species of crow were recorded into sonograms and analyzed. The appearance and inner configuration of the syrinx were observed using stereoscopic microscope. In addition, the inside diameter of the syrinx, the sizes of the labia and the attached position of the syringeal muscles were measured. The attached figures of syringeal muscles were different between the two species. The vocalizations of Carrion crows were noisier than possibly because their labias were noticeably smaller than those of Jungle crows. The attachment patterns of the syringeal muscles in Jungle crows suggested that they allow for more flexibility on the inside structure of the syrinx. The inner space of the syrinx in Jungle crows was also wider than those of Carrion crows. These results suggested that Jungle crows may be able to make various vocalizations because of these morphological characteristics.

  6. Vocal Pitch Discrimination in the Motor System

    ERIC Educational Resources Information Center

    D'Ausilio, Alessandro; Bufalari, Ilaria; Salmas, Paola; Busan, Pierpaolo; Fadiga, Luciano

    2011-01-01

    Speech production can be broadly separated into two distinct components: Phonation and Articulation. These two aspects require the efficient control of several phono-articulatory effectors. Speech is indeed generated by the vibration of the vocal-folds in the larynx (F0) followed by "filtering" by articulators, to select certain resonant…

  7. The influence of material anisotropy on vibration at onset in a three-dimensional vocal fold model

    PubMed Central

    Zhang, Zhaoyan

    2014-01-01

    Although vocal folds are known to be anisotropic, the influence of material anisotropy on vocal fold vibration remains largely unknown. Using a linear stability analysis, phonation onset characteristics were investigated in a three-dimensional anisotropic vocal fold model. The results showed that isotropic models had a tendency to vibrate in a swing-like motion, with vibration primarily along the superior-inferior direction. Anterior-posterior (AP) out-of-phase motion was also observed and large vocal fold vibration was confined to the middle third region along the AP length. In contrast, increasing anisotropy or increasing AP-transverse stiffness ratio suppressed this swing-like motion and allowed the vocal fold to vibrate in a more wave-like motion with strong medial-lateral motion over the entire medial surface. Increasing anisotropy also suppressed the AP out-of-phase motion, allowing the vocal fold to vibrate in phase along the entire AP length. Results also showed that such improvement in vibration pattern was the most effective with large anisotropy in the cover layer alone. These numerical predictions were consistent with previous experimental observations using self-oscillating physical models. It was further hypothesized that these differences may facilitate complete glottal closure in finite-amplitude vibration of anisotropic models as observed in recent experiments. PMID:24606284

  8. Proposal for a descriptive guideline of vascular changes in lesions of the vocal folds by the committee on endoscopic laryngeal imaging of the European Laryngological Society.

    PubMed

    Arens, Christoph; Piazza, Cesare; Andrea, Mario; Dikkers, Frederik G; Tjon Pian Gi, Robin E A; Voigt-Zimmermann, Susanne; Peretti, Giorgio

    2016-05-01

    In the last decades new endoscopic tools have been developed to improve the diagnostic work-up of vocal fold lesions in addition to normal laryngoscopy, i.e., contact endoscopy, autofluorescence, narrow band imaging and others. Better contrasted and high definition images offer more details of the epithelial and superficial vascular structure of the vocal folds. Following these developments, particular vascular patterns come into focus during laryngoscopy. The present work aims at a systematic pathogenic description of superficial vascular changes of the vocal folds. Additionally, new nomenclature on vascular lesions of the vocal folds will be presented to harmonize the different terms in the literature. Superficial vascular changes can be divided into longitudinal and perpendicular. Unlike longitudinal vascular lesions, e.g., ectasia, meander and change of direction, perpendicular vascular lesions are characterized by different types of vascular loops. They are primarily observed in recurrent respiratory papillomatosis, and in pre-cancerous and cancerous lesions of the vocal folds. These vascular characteristics play a significant role in the differential diagnosis. Among different parameters, e.g., epithelial changes, increase of volume, stiffness of the vocal fold, vascular lesions play an increasing role in the diagnosis of pre- and cancerous lesions.

  9. Construction and Characterization of a Novel Vocal Fold Bioreactor

    PubMed Central

    Zerdoum, Aidan B.; Tong, Zhixiang; Bachman, Brendan; Jia, Xinqiao

    2014-01-01

    In vitro engineering of mechanically active tissues requires the presentation of physiologically relevant mechanical conditions to cultured cells. To emulate the dynamic environment of vocal folds, a novel vocal fold bioreactor capable of producing vibratory stimulations at fundamental phonation frequencies is constructed and characterized. The device is composed of a function generator, a power amplifier, a speaker selector and parallel vibration chambers. Individual vibration chambers are created by sandwiching a custom-made silicone membrane between a pair of acrylic blocks. The silicone membrane not only serves as the bottom of the chamber but also provides a mechanism for securing the cell-laden scaffold. Vibration signals, generated by a speaker mounted underneath the bottom acrylic block, are transmitted to the membrane aerodynamically by the oscillating air. Eight identical vibration modules, fixed on two stationary metal bars, are housed in an anti-humidity chamber for long-term operation in a cell culture incubator. The vibration characteristics of the vocal fold bioreactor are analyzed non-destructively using a Laser Doppler Vibrometer (LDV). The utility of the dynamic culture device is demonstrated by culturing cellular constructs in the presence of 200-Hz sinusoidal vibrations with a mid-membrane displacement of 40 µm. Mesenchymal stem cells cultured in the bioreactor respond to the vibratory signals by altering the synthesis and degradation of vocal fold-relevant, extracellular matrix components. The novel bioreactor system presented herein offers an excellent in vitro platform for studying vibration-induced mechanotransduction and for the engineering of functional vocal fold tissues. PMID:25145349

  10. Construction and characterization of a novel vocal fold bioreactor.

    PubMed

    Zerdoum, Aidan B; Tong, Zhixiang; Bachman, Brendan; Jia, Xinqiao

    2014-08-01

    In vitro engineering of mechanically active tissues requires the presentation of physiologically relevant mechanical conditions to cultured cells. To emulate the dynamic environment of vocal folds, a novel vocal fold bioreactor capable of producing vibratory stimulations at fundamental phonation frequencies is constructed and characterized. The device is composed of a function generator, a power amplifier, a speaker selector and parallel vibration chambers. Individual vibration chambers are created by sandwiching a custom-made silicone membrane between a pair of acrylic blocks. The silicone membrane not only serves as the bottom of the chamber but also provides a mechanism for securing the cell-laden scaffold. Vibration signals, generated by a speaker mounted underneath the bottom acrylic block, are transmitted to the membrane aerodynamically by the oscillating air. Eight identical vibration modules, fixed on two stationary metal bars, are housed in an anti-humidity chamber for long-term operation in a cell culture incubator. The vibration characteristics of the vocal fold bioreactor are analyzed non-destructively using a Laser Doppler Vibrometer (LDV). The utility of the dynamic culture device is demonstrated by culturing cellular constructs in the presence of 200-Hz sinusoidal vibrations with a mid-membrane displacement of 40 µm. Mesenchymal stem cells cultured in the bioreactor respond to the vibratory signals by altering the synthesis and degradation of vocal fold-relevant, extracellular matrix components. The novel bioreactor system presented herein offers an excellent in vitro platform for studying vibration-induced mechanotransduction and for the engineering of functional vocal fold tissues.

  11. Responses of Middle-Frequency Modulations in Vocal Fundamental Frequency to Different Vocal Intensities and Auditory Feedback.

    PubMed

    Lee, Shao-Hsuan; Fang, Tuan-Jen; Yu, Jen-Fang; Lee, Guo-She

    2017-09-01

    Auditory feedback can make reflexive responses on sustained vocalizations. Among them, the middle-frequency power of F0 (MFP) may provide a sensitive index to access the subtle changes in different auditory feedback conditions. Phonatory airflow temperature was obtained from 20 healthy adults at two vocal intensity ranges under four auditory feedback conditions: (1) natural auditory feedback (NO); (2) binaural speech noise masking (SN); (3) bone-conducted feedback of self-generated voice (BAF); and (4) SN and BAF simultaneously. The modulations of F0 in low-frequency (0.2 Hz-3 Hz), middle-frequency (3 Hz-8 Hz), and high-frequency (8 Hz-25 Hz) bands were acquired using power spectral analysis of F0. Acoustic and aerodynamic analyses were used to acquire vocal intensity, maximum phonation time (MPT), phonatory airflow, and MFP-based vocal efficiency (MBVE). SN and high vocal intensity decreased MFP and raised MBVE and MPT significantly. BAF showed no effect on MFP but significantly lowered MBVE. Moreover, BAF significantly increased the perception of voice feedback and the sensation of vocal effort. Altered auditory feedback significantly changed the middle-frequency modulations of F0. MFP and MBVE could well detect these subtle responses of audio-vocal feedback. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  12. Final Syllable Lengthening (FSL) in infant vocalizations.

    PubMed

    Nathani, Suneeti; Oller, D Kimbrough; Cobo-Lewis, Alan B

    2003-02-01

    Final Syllable Lengthening (FSL) has been extensively examined in infant vocalizations in order to determine whether its basis is biological or learned. Findings suggest there may be a U-shaped developmental trajectory for FSL. The present study sought to verify this pattern and to determine whether vocal maturity and deafness influence FSL. Eight normally hearing infants, aged 0;3 to 1;0, and eight deaf infants, aged 0;8 to 4;0, were examined at three levels of prelinguistic vocal development: precanonical, canonical, and postcanonical. FSL was found at all three levels suggesting a biological basis for this phenomenon. Individual variability was, however, considerable. Reduction in the magnitude of FSL across the three sessions provided some support for a downward trend for FSL in infancy. Findings further indicated that auditory deprivation can significantly affect temporal aspects of infant speech production.

  13. Cetacean vocal learning and communication.

    PubMed

    Janik, Vincent M

    2014-10-01

    The cetaceans are one of the few mammalian clades capable of vocal production learning. Evidence for this comes from synchronous changes in song patterns of baleen whales and experimental work on toothed whales in captivity. While baleen whales like many vocal learners use this skill in song displays that are involved in sexual selection, toothed whales use learned signals in individual recognition and the negotiation of social relationships. Experimental studies demonstrated that dolphins can use learned signals referentially. Studies on wild dolphins demonstrated how this skill appears to be useful in their own communication system, making them an interesting subject for comparative communication studies. Copyright © 2014. Published by Elsevier Ltd.

  14. The vocal monotony of monogamy

    NASA Astrophysics Data System (ADS)

    Thomas, Jeanette

    2003-04-01

    There are four phocids in waters around Antarctica: Weddell, leopard, crabeater, and Ross seals. These four species provide a unique opportunity to examine underwater vocal behavior in species sharing the same ecosystem. Some species live in pack ice, others in factice, but all are restricted to the Antarctic or sub-Antarctic islands. All breed and produce vocalizations under water. Social systems range from polygyny in large breeding colonies, to serial monogamy, to solitary species. The type of mating system influences the number of underwater vocalizations in the repertoire, with monogamous seals producing only a single call, polygynous species producing up to 35 calls, and solitary species an intermediate number of about 10 calls. Breeding occurs during the austral spring and each species carves-out an acoustic niche for communicating, with species using different frequency ranges, temporal patterns, and amplitude changes to convey their species-specific calls and presumably reduce acoustic competition. Some species exhibit geographic variations in their vocalizations around the continent, which may reflect discrete breeding populations. Some seals become silent during a vulnerable time of predation by killer whales, perhaps to avoid detection. Overall, vocalizations of these seals exhibit adaptive characteristics that reflect the co-evolution among species in the same ecosystem.

  15. Characteristics of phonation onset in a two-layer vocal fold model.

    PubMed

    Zhang, Zhaoyan

    2009-02-01

    Characteristics of phonation onset were investigated in a two-layer body-cover continuum model of the vocal folds as a function of the biomechanical and geometric properties of the vocal folds. The analysis showed that an increase in either the body or cover stiffness generally increased the phonation threshold pressure and phonation onset frequency, although the effectiveness of varying body or cover stiffness as a pitch control mechanism varied depending on the body-cover stiffness ratio. Increasing body-cover stiffness ratio reduced the vibration amplitude of the body layer, and the vocal fold motion was gradually restricted to the medial surface, resulting in more effective flow modulation and higher sound production efficiency. The fluid-structure interaction induced synchronization of more than one group of eigenmodes so that two or more eigenmodes may be simultaneously destabilized toward phonation onset. At certain conditions, a slight change in vocal fold stiffness or geometry may cause phonation onset to occur as eigenmode synchronization due to a different pair of eigenmodes, leading to sudden changes in phonation onset frequency, vocal fold vibration pattern, and sound production efficiency. Although observed in a linear stability analysis, a similar mechanism may also play a role in register changes at finite-amplitude oscillations.

  16. Ontogeny of individual and litter identity signaling in grunts of piglets.

    PubMed

    Syrová, Michaela; Policht, Richard; Linhart, Pavel; Špinka, Marek

    2017-11-01

    Many studies have shown that animal vocalizations can signal individual identity and group/family membership. However, much less is known about the ontogeny of identity information-when and how this individual/group distinctiveness in vocalizations arises and how it changes during the animal's life. Recent findings suggest that even species that were thought to have limited vocal plasticity could adjust their calls to sound more similar to each other within a group. It has already been shown that sows can acoustically distinguish their own offspring from alien piglets and that litters differ in their calls. Surprisingly, individual identity in piglet calls has not been reported yet. In this paper, this gap is filled, and it is shown that there is information about piglet identity. Information about litter identity is confirmed as well. Individual identity increased with age, but litter vocal identity did not increase with age. The results were robust as a similar pattern was apparent in two situations differing in arousal: isolation and back-test. This paper argues that, in piglets, increased individual discrimination results from the rapid growth of piglets, which is likely to be associated with growth and diversification of the vocal tract rather than from social effects and vocal plasticity.

  17. Patterns of Occurrence and Marine Mammal Acoustic Behavior in Relation to Navy Sonar Activity Off Jacksonville, Florida.

    PubMed

    Oswald, Julie N; Norris, Thomas F; Yack, Tina M; Ferguson, Elizabeth L; Kumar, Anurag; Nissen, Jene; Bell, Joel

    2016-01-01

    Passive acoustic data collected from marine autonomous recording units deployed off Jacksonville, FL (from 13 September to 8 October 2009 and 3 December 2009 to 8 January 2010), were analyzed for detection of cetaceans and Navy sonar. Cetaceans detected included Balaenoptera acutorostrata, Eubalaena glacialis, B. borealis, Physeter macrocephalus, blackfish, and delphinids. E. glacialis were detected at shallow and, somewhat unexpectedly, deep sites. P. macrocephalus were characterized by a strong diel pattern. B. acutorostrata showed the strongest relationship between sonar activity and vocal behavior. These results provide a preliminary assessment of cetacean occurrence off Jacksonville and new insights on vocal responses to sonar.

  18. High-Speed Imaging Analysis of Register Transitions in Classically and Jazz-Trained Male Voices.

    PubMed

    Dippold, Sebastian; Voigt, Daniel; Richter, Bernhard; Echternach, Matthias

    2015-01-01

    Little data are available concerning register functions in different styles of singing such as classically or jazz-trained voices. Differences between registers seem to be much more audible in jazz singing than classical singing, and so we hypothesized that classically trained singers exhibit a smoother register transition, stemming from more regular vocal fold oscillation patterns. High-speed digital imaging (HSDI) was used for 19 male singers (10 jazz-trained singers, 9 classically trained) who performed a glissando from modal to falsetto register across the register transition. Vocal fold oscillation patterns were analyzed in terms of different parameters of regularity such as relative average perturbation (RAP), correlation dimension (D2) and shimmer. HSDI observations showed more regular vocal fold oscillation patterns during the register transition for the classically trained singers. Additionally, the RAP and D2 values were generally lower and more consistent for the classically trained singers compared to the jazz singers. However, intergroup comparisons showed no statistically significant differences. Some of our results may support the hypothesis that classically trained singers exhibit a smoother register transition from modal to falsetto register. © 2015 S. Karger AG, Basel.

  19. Diversity within a Birdsong

    NASA Astrophysics Data System (ADS)

    Laje, Rodrigo; Mindlin, Gabriel B.

    2002-12-01

    We present a model for the activities of neural circuits in a nucleus found in the brains of songbirds: the robust nucleus of the archistriatum (RA). This is a fore brain song control nucleus responsible for the phasic and precise neural signals driving vocal and respiratory motor neurons during singing. Driving a physical model of the avian vocal organ with the signals generated by the neural model, we produce synthetic songs. This allows us to show that certain connectivity architectures in the RA give rise to a wide range of different vocalizations under simple excitatory instructions.

  20. Paralinguistic mechanisms of production in human "beatboxing": a real-time magnetic resonance imaging study.

    PubMed

    Proctor, Michael; Bresch, Erik; Byrd, Dani; Nayak, Krishna; Narayanan, Shrikanth

    2013-02-01

    Real-time magnetic resonance imaging (rtMRI) was used to examine mechanisms of sound production by an American male beatbox artist. rtMRI was found to be a useful modality with which to study this form of sound production, providing a global dynamic view of the midsagittal vocal tract at frame rates sufficient to observe the movement and coordination of critical articulators. The subject's repertoire included percussion elements generated using a wide range of articulatory and airstream mechanisms. Many of the same mechanisms observed in human speech production were exploited for musical effect, including patterns of articulation that do not occur in the phonologies of the artist's native languages: ejectives and clicks. The data offer insights into the paralinguistic use of phonetic primitives and the ways in which they are coordinated in this style of musical performance. A unified formalism for describing both musical and phonetic dimensions of human vocal percussion performance is proposed. Audio and video data illustrating production and orchestration of beatboxing sound effects are provided in a companion annotated corpus.

  1. Vocal Fold Epithelial Response to Luminal Osmotic Perturbation

    ERIC Educational Resources Information Center

    Sivasankar, Mahalakshmi; Fisher, Kimberly V.

    2007-01-01

    Purpose: Dry-air challenges increase the osmolarity of fluid lining the luminal surface of the proximal airway. The homeostasis of surface fluid is thought to be essential for voice production and laryngeal defense. Therefore, the authors hypothesized that viable vocal fold epithelium would generate a water flux to reduce an osmotic challenge (150…

  2. Patterns of call communication between group-housed zebra finches change during the breeding cycle

    PubMed Central

    Gill, Lisa F; Goymann, Wolfgang; Ter Maat, Andries; Gahr, Manfred

    2015-01-01

    Vocal signals such as calls play a crucial role for survival and successful reproduction, especially in group-living animals. However, call interactions and call dynamics within groups remain largely unexplored because their relation to relevant contexts or life-history stages could not be studied with individual-level resolution. Using on-bird microphone transmitters, we recorded the vocalisations of individual zebra finches (Taeniopygia guttata) behaving freely in social groups, while females and males previously unknown to each other passed through different stages of the breeding cycle. As birds formed pairs and shifted their reproductive status, their call repertoire composition changed. The recordings revealed that calls occurred non-randomly in fine-tuned vocal interactions and decreased within groups while pair-specific patterns emerged. Call-type combinations of vocal interactions changed within pairs and were associated with successful egg-laying, highlighting a potential fitness relevance of calling dynamics in communication systems. DOI: http://dx.doi.org/10.7554/eLife.07770.001 PMID:26441403

  3. A wavelet-based approach for a continuous analysis of phonovibrograms.

    PubMed

    Unger, Jakob; Meyer, Tobias; Doellinger, Michael; Hecker, Dietmar J; Schick, Bernhard; Lohscheller, Joerg

    2012-01-01

    Recently, endoscopic high-speed laryngoscopy has been established for commercial use and constitutes a state-of-the-art technique to examine vocal fold dynamics. Despite overcoming many limitations of commonly applied stroboscopy it has not gained widespread clinical application, yet. A major drawback is a missing methodology of extracting valuable features to support visual assessment or computer-aided diagnosis. In this paper a compact and descriptive feature set is presented. The feature extraction routines are based on two-dimensional color graphs called phonovibrograms (PVG). These graphs contain the full spatio-temporal pattern of vocal fold dynamics and are therefore suited to derive features that comprehensively describe the vibration pattern of vocal folds. Within our approach, clinically relevant features such as glottal closure type, symmetry and periodicity are quantified in a set of 10 descriptive features. The suitability for classification tasks is shown using a clinical data set comprising 50 healthy and 50 paralytic subjects. A classification accuracy of 93.2% has been achieved.

  4. Acoustic Properties of the Voice Source and the Vocal Tract: Are They Perceptually Independent?

    PubMed

    Erickson, Molly L

    2016-11-01

    This study sought to determine whether the properties of the voice source and vocal tract are perceptually independent. Within-subjects design. This study employed a paired-comparison paradigm where listeners heard synthetic voices and rated them as same or different using a visual analog scale. Stimuli were synthesized using three different source slopes and two different formant patterns (mezzo-soprano and soprano) on the vowel /a/ at four pitches: A3, C4, B4, and F5. Whereas formant pattern was the strongest effect, difference in source slope also affected perceived quality difference. Source slope and formant pattern were not independently perceived. These results suggest that when judging laryngeal adduction using perceptual information, judgments may not be accurate when the stimuli are of differing formant patterns. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  5. Selective and Efficient Neural Coding of Communication Signals Depends on Early Acoustic and Social Environment

    PubMed Central

    Amin, Noopur; Gastpar, Michael; Theunissen, Frédéric E.

    2013-01-01

    Previous research has shown that postnatal exposure to simple, synthetic sounds can affect the sound representation in the auditory cortex as reflected by changes in the tonotopic map or other relatively simple tuning properties, such as AM tuning. However, their functional implications for neural processing in the generation of ethologically-based perception remain unexplored. Here we examined the effects of noise-rearing and social isolation on the neural processing of communication sounds such as species-specific song, in the primary auditory cortex analog of adult zebra finches. Our electrophysiological recordings reveal that neural tuning to simple frequency-based synthetic sounds is initially established in all the laminae independent of patterned acoustic experience; however, we provide the first evidence that early exposure to patterned sound statistics, such as those found in native sounds, is required for the subsequent emergence of neural selectivity for complex vocalizations and for shaping neural spiking precision in superficial and deep cortical laminae, and for creating efficient neural representations of song and a less redundant ensemble code in all the laminae. Our study also provides the first causal evidence for ‘sparse coding’, such that when the statistics of the stimuli were changed during rearing, as in noise-rearing, that the sparse or optimal representation for species-specific vocalizations disappeared. Taken together, these results imply that a layer-specific differential development of the auditory cortex requires patterned acoustic input, and a specialized and robust sensory representation of complex communication sounds in the auditory cortex requires a rich acoustic and social environment. PMID:23630587

  6. Phonation threshold pressure predictions using viscoelastic properties up to 1,400 Hz of injectables intended for Reinke's space.

    PubMed

    Klemuk, Sarah A; Lu, Xiaoying; Hoffman, Henry T; Titze, Ingo R

    2010-05-01

    Viscoelastic properties of numerous vocal fold injectables have been reported but not at speaking frequencies. For materials intended for Reinke's space, ramifications of property values are of great concern because of their impact on ease of voice onset. Our objectives were: 1) to measure viscoelastic properties of a new nonresorbing carbomer and well-known vocal fold injectables at vocalization frequencies using established and new instrumentation, and 2) to predict phonation threshold pressures using a computer model with intended placement in Reinke's space. Rheology and phonation threshold pressure calculations. Injectables were evaluated with a traditional rotational rheometer and a new piezo-rotary vibrator. Using these data at vocalization frequencies, phonation threshold pressures (PTP) were calculated for each biomaterial, assuming a low dimensional model with supraglottic coupling and adjusted vocal fold length and thickness at each frequency. Results were normalized to a nominal PTP value. Viscoelastic data were acquired at vocalization frequencies as high as 363 to 1,400 Hz for six new carbomer hydrogels, Hylan B, and Extracel intended for vocal fold Reinke's space injection and for Cymetra (lateral injection). Reliability was confirmed with good data overlap when measuring with either rheometer. PTP predictions ranged from 0.001 to 16 times the nominal PTP value of 0.283 kPa. Accurate viscoelastic measurements of vocal fold injectables are now possible at physiologic frequencies. Hylan B, Extracel, and the new carbomer hydrogels should generate easy vocal onset and sustainable vocalization based on their rheologic properties if injected into Reinke's space. Applications may vary depending on desired longevity of implant. Laryngoscope, 2010.

  7. Emotion in the wilds of nature: The coherence and contagion of fear during threatening group-based outdoors experiences.

    PubMed

    Anderson, Craig L; Monroy, Maria; Keltner, Dacher

    2018-04-01

    Emotional expressions communicate information about the individual's internal state and evoke responses in others that enable coordinated action. The current work investigated the informative and evocative properties of fear vocalizations in a sample of youth from underserved communities and military veterans while white-water rafting. Video-taped footage of participants rafting through white-water rapids was coded for vocal and facial expressions of fear, amusement, pride, and awe, yielding more than 1,300 coded expressions, which were then related to measures of subjective emotion and cortisol response. Consistent with informative properties of emotional expressions, fear vocalizations were positively and significantly related to facial expressions of fear, subjective reports of fear, and individuals' cortisol levels measured after the rafting trip. It is important to note that this coherent pattern was unique to fear vocalizations; vocalizations of amusement, pride, and awe were not significantly related to fear expressions in the face, subjective reports of fear, or cortisol levels. Demonstrating the evocative properties of emotional expression, fear vocalizations of individuals appeared to evoke fear vocalizations in other people in their raft, and cortisol levels of individuals within rafts similarly converged at the end of the trip. We discuss how the study of spontaneous emotion expressions in naturalistic settings can help address basic yet controversial questions about emotions. (PsycINFO Database Record (c) 2018 APA, all rights reserved).

  8. Rules of song development and their use in vocal interactions by birds with large repertoires.

    PubMed

    Geberzahn, Nicole; Hultsch, Henrike

    2004-06-01

    Songbirds are well known for settling their disputes by vocal signals, and their singing plays a dominant role. Most studies on this issue have concentrated on bird species that develop and use small vocal repertoires. In this article we will go farther and focus on examples of how species with large song repertoires make use of their vocal competence. In particular, we will outline the study of interaction rules which have been elucidated by examining time- and pattern-specific relationships between signals exchanged by territorial neighbors. First we present an inquiry into the rules of song learning and development. In birds with large song repertoires, the ontogeny of such rules proceeds along a number of trajectories which help in understanding the often remarkable accomplishments of adult birds. In both approaches, our model species will be the Common Nightingale Luscinia megarhynchos that has been investigated intensively in the field and in the laboratory.

  9. A Neural Code That Is Isometric to Vocal Output and Correlates with Its Sensory Consequences

    PubMed Central

    Vyssotski, Alexei L.; Stepien, Anna E.; Keller, Georg B.; Hahnloser, Richard H. R.

    2016-01-01

    What cortical inputs are provided to motor control areas while they drive complex learned behaviors? We study this question in the nucleus interface of the nidopallium (NIf), which is required for normal birdsong production and provides the main source of auditory input to HVC, the driver of adult song. In juvenile and adult zebra finches, we find that spikes in NIf projection neurons precede vocalizations by several tens of milliseconds and are insensitive to distortions of auditory feedback. We identify a local isometry between NIf output and vocalizations: quasi-identical notes produced in different syllables are preceded by highly similar NIf spike patterns. NIf multiunit firing during song precedes responses in auditory cortical neurons by about 50 ms, revealing delayed congruence between NIf spiking and a neural representation of auditory feedback. Our findings suggest that NIf codes for imminent acoustic events within vocal performance. PMID:27723764

  10. Genetic identification of a hindbrain nucleus essential for innate vocalization.

    PubMed

    Hernandez-Miranda, Luis Rodrigo; Ruffault, Pierre-Louis; Bouvier, Julien C; Murray, Andrew J; Morin-Surun, Marie-Pierre; Zampieri, Niccolò; Cholewa-Waclaw, Justyna B; Ey, Elodie; Brunet, Jean-Francois; Champagnat, Jean; Fortin, Gilles; Birchmeier, Carmen

    2017-07-25

    Vocalization in young mice is an innate response to isolation or mechanical stimulation. Neuronal circuits that control vocalization and breathing overlap and rely on motor neurons that innervate laryngeal and expiratory muscles, but the brain center that coordinates these motor neurons has not been identified. Here, we show that the hindbrain nucleus tractus solitarius (NTS) is essential for vocalization in mice. By generating genetically modified newborn mice that specifically lack excitatory NTS neurons, we show that they are both mute and unable to produce the expiratory drive required for vocalization. Furthermore, the muteness of these newborns results in maternal neglect. We also show that neurons of the NTS directly connect to and entrain the activity of spinal (L1) and nucleus ambiguus motor pools located at positions where expiratory and laryngeal motor neurons reside. These motor neurons control expiratory pressure and laryngeal tension, respectively, thereby establishing the essential biomechanical parameters used for vocalization. In summary, our work demonstrates that the NTS is an obligatory component of the neuronal circuitry that transforms breaths into calls.

  11. Assessment of breathing patterns and respiratory muscle recruitment during singing and speech in quadriplegia.

    PubMed

    Tamplin, Jeanette; Brazzale, Danny J; Pretto, Jeffrey J; Ruehland, Warren R; Buttifant, Mary; Brown, Douglas J; Berlowitz, David J

    2011-02-01

    To explore how respiratory impairment after cervical spinal cord injury affects vocal function, and to explore muscle recruitment strategies used during vocal tasks after quadriplegia. It was hypothesized that to achieve the increased respiratory support required for singing and loud speech, people with quadriplegia use different patterns of muscle recruitment and control strategies compared with control subjects without spinal cord injury. Matched, parallel-group design. Large university-affiliated public hospital. Consenting participants with motor-complete C5-7 quadriplegia (n=6) and able-bodied age-matched controls (n=6) were assessed on physiologic and voice measures during vocal tasks. Not applicable. Standard respiratory function testing, surface electromyographic activity from accessory respiratory muscles, sound pressure levels during vocal tasks, the Voice Handicap Index, and the Perceptual Voice Profile. The group with quadriplegia had a reduced lung capacity (vital capacity, 71% vs 102% of predicted; P=.028), more perceived voice problems (Voice Handicap Index score, 22.5 vs 6.5; P=.046), and greater recruitment of accessory respiratory muscles during both loud and soft volumes (P=.028) than the able-bodied controls. The group with quadriplegia also demonstrated higher accessory muscle activation in changing from soft to loud speech (P=.028). People with quadriplegia have impaired vocal ability and use different muscle recruitment strategies during speech than the able-bodied. These findings will enable us to target specific measurements of respiratory physiology for assessing functional improvements in response to formal therapeutic singing training. Copyright © 2011 American Congress of Rehabilitation Medicine. Published by Elsevier Inc. All rights reserved.

  12. A magnetic resonance imaging-based articulatory and acoustic study of “retroflex” and “bunched” American English ∕r∕

    PubMed Central

    Zhou, Xinhui; Espy-Wilson, Carol Y.; Boyce, Suzanne; Tiede, Mark; Holland, Christy; Choe, Ann

    2008-01-01

    Speakers of rhotic dialects of North American English show a range of different tongue configurations for ∕r∕. These variants produce acoustic profiles that are indistinguishable for the first three formants [Delattre, P., and Freeman, D. C., (1968). “A dialect study of American English r’s by x-ray motion picture,” Linguistics 44, 28–69; Westbury, J. R. et al. (1998), “Differences among speakers in lingual articulation for American English ∕r∕,” Speech Commun. 26, 203–206]. It is puzzling why this should be so, given the very different vocal tract configurations involved. In this paper, two subjects whose productions of “retroflex” ∕r∕ and “bunched” ∕r∕ show similar patterns of F1–F3 but very different spacing between F4 and F5 are contrasted. Using finite element analysis and area functions based on magnetic resonance images of the vocal tract for sustained productions, the results of computer vocal tract models are compared to actual speech recordings. In particular, formant-cavity affiliations are explored using formant sensitivity functions and vocal tract simple-tube models. The difference in F4∕F5 patterns between the subjects is confirmed for several additional subjects with retroflex and bunched vocal tract configurations. The results suggest that the F4∕F5 differences between the variants can be largely explained by differences in whether the long cavity behind the palatal constriction acts as a half- or a quarter-wavelength resonator. PMID:18537397

  13. Birds, primates, and spoken language origins: behavioral phenotypes and neurobiological substrates

    PubMed Central

    Petkov, Christopher I.; Jarvis, Erich D.

    2012-01-01

    Vocal learners such as humans and songbirds can learn to produce elaborate patterns of structurally organized vocalizations, whereas many other vertebrates such as non-human primates and most other bird groups either cannot or do so to a very limited degree. To explain the similarities among humans and vocal-learning birds and the differences with other species, various theories have been proposed. One set of theories are motor theories, which underscore the role of the motor system as an evolutionary substrate for vocal production learning. For instance, the motor theory of speech and song perception proposes enhanced auditory perceptual learning of speech in humans and song in birds, which suggests a considerable level of neurobiological specialization. Another, a motor theory of vocal learning origin, proposes that the brain pathways that control the learning and production of song and speech were derived from adjacent motor brain pathways. Another set of theories are cognitive theories, which address the interface between cognition and the auditory-vocal domains to support language learning in humans. Here we critically review the behavioral and neurobiological evidence for parallels and differences between the so-called vocal learners and vocal non-learners in the context of motor and cognitive theories. In doing so, we note that behaviorally vocal-production learning abilities are more distributed than categorical, as are the auditory-learning abilities of animals. We propose testable hypotheses on the extent of the specializations and cross-species correspondences suggested by motor and cognitive theories. We believe that determining how spoken language evolved is likely to become clearer with concerted efforts in testing comparative data from many non-human animal species. PMID:22912615

  14. American Voice Types: Towards a Vocal Typology for American English

    ERIC Educational Resources Information Center

    McPeek, Tyler

    2013-01-01

    Individual voices are not uniformly similar to others, even when factoring out speaker characteristics such as sex, age, dialect, and so on. Some speakers share common features and can cohere into groups based on gross vocal similarity but, to date, no attempt has been made to describe these features systematically or to generate a taxonomy based…

  15. Infant Vocal-Motor Coordination: Precursor to the Gesture-Speech System?

    ERIC Educational Resources Information Center

    Iverson, Jana M.; Fagan, Mary K.

    2004-01-01

    This study was designed to provide a general picture of infant vocal-motor coordination and test predictions generated by Iverson and Thelen's (1999) model of the development of the gesture-speech system. Forty-seven 6- to 9-month-old infants were videotaped with a primary caregiver during rattle and toy play. Results indicated an age-related…

  16. Source levels of social sounds in migrating humpback whales (Megaptera novaeangliae).

    PubMed

    Dunlop, Rebecca A; Cato, Douglas H; Noad, Michael J; Stokes, Dale M

    2013-07-01

    The source level of an animal sound is important in communication, since it affects the distance over which the sound is audible. Several measurements of source levels of whale sounds have been reported, but the accuracy of many is limited because the distance to the source and the acoustic transmission loss were estimated rather than measured. This paper presents measurements of source levels of social sounds (surface-generated and vocal sounds) of humpback whales from a sample of 998 sounds recorded from 49 migrating humpback whale groups. Sources were localized using a wide baseline five hydrophone array and transmission loss was measured for the site. Social vocalization source levels were found to range from 123 to 183 dB re 1 μPa @ 1 m with a median of 158 dB re 1 μPa @ 1 m. Source levels of surface-generated social sounds ("breaches" and "slaps") were narrower in range (133 to 171 dB re 1 μPa @ 1 m) but slightly higher in level (median of 162 dB re 1 μPa @ 1 m) compared to vocalizations. The data suggest that group composition has an effect on group vocalization source levels in that singletons and mother-calf-singing escort groups tend to vocalize at higher levels compared to other group compositions.

  17. RESPONSES OF MALE TROPICAL MOCKINGBIRDS TO VARIATION IN WITHIN-SONG AND BETWEEN-SONG VERSATILITY

    PubMed Central

    Botero, Carlos A.; Vehrencamp, Sandra L.

    2007-01-01

    Despite their large vocal repertoires and otherwise highly versatile singing style, male mockingbirds sometimes sing in a highly repetitive fashion. We conducted a playback experiment to determine the possible signal value of different syllable presentation patterns during simulated male intrusions in the Tropical Mockingbird (Mimus gilvus) testing the hypothesis that more repetitive singing represents a stronger threat and generates a stronger aggressive response. Responses were measured in terms of approach and singing behavior and were analyzed using McGregor’s (1992) multivariate method. We also introduce the use of survival analysis for analyzing response variables for which subjects do not perform the behavior in question in at least one of the replicates (known as ‘right-censored variables’ in the statistical literature). As predicted by theory, experimental subjects responded more aggressively to songs composed of a single note than to variable ones. However, versatility at the between-song level had an opposite effect as high song switching rates generated stronger responses than low ones. Given the lack of a statistical interaction between within-song versatility and switching rate, we conclude that these two parameters may serve independent purposes and possibly transmit different information. We discuss the possibility that the signal value of variation in vocal versatility lies in the mediation of territorial conflicts, the attraction of female partners and/or the mediation of conflicts over access to reproductive females. PMID:18509510

  18. Vocal exercise may attenuate acute vocal fold inflammation

    PubMed Central

    Abbott, Katherine Verdolini; Li, Nicole Y.K.; Branski, Ryan C.; Rosen, Clark A.; Grillo, Elizabeth; Steinhauer, Kimberly; Hebda, Patricia A.

    2012-01-01

    Objectives/Hypotheses The objective was to assess the utility of selected “resonant voice” exercises for the reduction of acute vocal fold inflammation. The hypothesis was that relatively large-amplitude, low-impact exercises associated with resonant voice would reduce inflammation more than spontaneous speech and possibly more than voice rest. Study Design The study design was prospective, randomized, double-blind. Methods Nine vocally healthy adults underwent a 1-hr vocal loading procedure, followed by randomization to (a) a spontaneous speech condition, (b) a vocal rest condition, or (c) a resonant voice exercise condition. Treatments were monitored in clinic for 4 hr, and continued extra-clinically until the next morning. At baseline, immediately following loading, after the 4-hr in-clinic treatment, and 24 hr post baseline, secretions were suctioned from the vocal folds bilaterally and submitted to enzyme-linked immunosorbent assay (ELISA) to estimate concentrations of key markers of tissue injury and inflammation: IL-1β, IL-6, IL-8, TNF-α, MMP-8, and IL-10. Results Complete data sets were obtained for 3 markers -- IL-1β, IL-6, and MMP-8 -- for one subject in each treatment condition. For those markers, results were poorest at 24-hr follow-up in the spontaneous speech condition, sharply improved in the voice rest condition, and best in the resonant voice condition. Average results for all markers, for all responsive subjects with normal baseline mediator concentrations, revealed an almost identical pattern. Conclusions Some forms of tissue mobilization may be useful to attenuate acute vocal fold inflammation. PMID:23177745

  19. The power of your vocal image.

    PubMed

    McCoy, L A

    1996-03-01

    Your vocal image is the impression that listeners form of you based on the sound of your voice. In a dental office, where the initial patient contact usually occurs over the phone, your vocal image is vitally important. According to social psychologists, people begin to make relatively durable first impressions within six to 12 seconds of perceiving a sensory cue. This means that patients begin to form their impressions of a telephone speaker almost immediately. Based on the qualities of the speaker's voice and how it is used, they'll form impressions related to everything from the speaker's physical and personality characteristics to his or her intellectual ability, and eventually even generalize their impressions to include the office that the speaker represents. If you want to improve your vocal image, you must first be aware of exactly what that image is. There are two factors that combine to create a vocal impression--the speaker's physical vocal tools and the sound that is created by them. The five physical tools involved are the lungs, vocal cords, throat, mouth and ears. At each stage in the sound production process, we can easily fall into negative habits and lazy patterns if we're not careful. Although we can't do much about our physical voice mechanism, we can certainly exercise a great deal of control over how our voice is used. A strong, confident voice is an essential part of effective interpersonal communication. If you want to project an image of confidence and professionalism, don't overlook the subtle benefits of effective vocal power.

  20. Neural Correlates of Vocal Production and Motor Control in Human Heschl's Gyrus

    PubMed Central

    Oya, Hiroyuki; Nourski, Kirill V.; Kawasaki, Hiroto; Larson, Charles R.; Brugge, John F.; Howard, Matthew A.; Greenlee, Jeremy D.W.

    2016-01-01

    The present study investigated how pitch frequency, a perceptually relevant aspect of periodicity in natural human vocalizations, is encoded in Heschl's gyrus (HG), and how this information may be used to influence vocal pitch motor control. We recorded local field potentials from multicontact depth electrodes implanted in HG of 14 neurosurgical epilepsy patients as they vocalized vowel sounds and received brief (200 ms) pitch perturbations at 100 Cents in their auditory feedback. Event-related band power responses to vocalizations showed sustained frequency following responses that tracked voice fundamental frequency (F0) and were significantly enhanced in posteromedial HG during speaking compared with when subjects listened to the playback of their own voice. In addition to frequency following responses, a transient response component within the high gamma frequency band (75–150 Hz) was identified. When this response followed the onset of vocalization, the magnitude of the response was the same for the speaking and playback conditions. In contrast, when this response followed a pitch shift, its magnitude was significantly enhanced during speaking compared with playback. We also observed that, in anterolateral HG, the power of high gamma responses to pitch shifts correlated with the magnitude of compensatory vocal responses. These findings demonstrate a functional parcellation of HG with neural activity that encodes pitch in natural human voice, distinguishes between self-generated and passively heard vocalizations, detects discrepancies between the intended and heard vocalization, and contains information about the resulting behavioral vocal compensations in response to auditory feedback pitch perturbations. SIGNIFICANCE STATEMENT The present study is a significant contribution to our understanding of sensor-motor mechanisms of vocal production and motor control. The findings demonstrate distinct functional parcellation of core and noncore areas within human auditory cortex on Heschl's gyrus that process natural human vocalizations and pitch perturbations in the auditory feedback. In addition, our data provide evidence for distinct roles of high gamma neural oscillations and frequency following responses for processing periodicity in human vocalizations during vocal production and motor control. PMID:26888939

  1. The larynx of roaring and non-roaring cats.

    PubMed

    Hast, M H

    1989-04-01

    Dissections were made of the larynges of 14 species of the cat family, with representative specimens from all genera. It was found that the vocal folds of the larynx of genus Panthera (with the exception of the snow leopard) form the basic structure of a sound generator well-designed to produce a high acoustical energy. Combined with an efficient sound radiator (vocal tract) that can be adjusted in length, a Panthera can use its vocal instrument literally to blow its own horn with a 'roar'. Also, it is proposed that laryngeal morphology can be used as an anatomical character in mammalian taxonomy.

  2. Pitch (F0) and formant profiles of human vowels and vowel-like baboon grunts: The role of vocalizer body size and voice-acoustic allometry

    NASA Astrophysics Data System (ADS)

    Rendall, Drew; Kollias, Sophie; Ney, Christina; Lloyd, Peter

    2005-02-01

    Key voice features-fundamental frequency (F0) and formant frequencies-can vary extensively between individuals. Much of the variation can be traced to differences in the size of the larynx and vocal-tract cavities, but whether these differences in turn simply reflect differences in speaker body size (i.e., neutral vocal allometry) remains unclear. Quantitative analyses were therefore undertaken to test the relationship between speaker body size and voice F0 and formant frequencies for human vowels. To test the taxonomic generality of the relationships, the same analyses were conducted on the vowel-like grunts of baboons, whose phylogenetic proximity to humans and similar vocal production biology and voice acoustic patterns recommend them for such comparative research. For adults of both species, males were larger than females and had lower mean voice F0 and formant frequencies. However, beyond this, F0 variation did not track body-size variation between the sexes in either species, nor within sexes in humans. In humans, formant variation correlated significantly with speaker height but only in males and not in females. Implications for general vocal allometry are discussed as are implications for speech origins theories, and challenges to them, related to laryngeal position and vocal tract length. .

  3. Therapeutic potential of gel-based injectables for vocal fold regeneration

    PubMed Central

    Bartlett, Rebecca S.; Thibeault, Susan L.; Prestwich, Glenn D.

    2012-01-01

    Vocal folds are anatomically and biomechanically unique, thus complicating the design and implementation of tissue engineering strategies for repair and regeneration. Integration of an enhanced understanding of tissue biomechanics, wound healing dynamics and innovative gel-based therapeutics has generated enthusiasm for the notion that an efficacious treatment for vocal fold scarring could be clinically attainable within several years. Fibroblast phenotype and gene expression are mediated by the three-dimensional mechanical and chemical microenvironment at an injury site. Thus, therapeutic approaches need to coordinate spatial and temporal aspects of the wound healing response in an injured vocal tissue to achieve an optimal clinical outcome. Successful gel-based injectables for vocal fold scarring will require a keen understanding of how the native inflammatory response sets into motion the later extracellular matrix remodeling, which in turn will determine the ultimate biomechanical properties of the tissue. We present an overview of the challenges associated with this translation as well as the proposed gel-based injectable solutions. PMID:22456756

  4. Predictive and tempo-flexible synchronization to a visual metronome in monkeys.

    PubMed

    Takeya, Ryuji; Kameda, Masashi; Patel, Aniruddh D; Tanaka, Masaki

    2017-07-21

    Predictive and tempo-flexible synchronization to an auditory beat is a fundamental component of human music. To date, only certain vocal learning species show this behaviour spontaneously. Prior research training macaques (vocal non-learners) to tap to an auditory or visual metronome found their movements to be largely reactive, not predictive. Does this reflect the lack of capacity for predictive synchronization in monkeys, or lack of motivation to exhibit this behaviour? To discriminate these possibilities, we trained monkeys to make synchronized eye movements to a visual metronome. We found that monkeys could generate predictive saccades synchronized to periodic visual stimuli when an immediate reward was given for every predictive movement. This behaviour generalized to novel tempi, and the monkeys could maintain the tempo internally. Furthermore, monkeys could flexibly switch from predictive to reactive saccades when a reward was given for each reactive response. In contrast, when humans were asked to make a sequence of reactive saccades to a visual metronome, they often unintentionally generated predictive movements. These results suggest that even vocal non-learners may have the capacity for predictive and tempo-flexible synchronization to a beat, but that only certain vocal learning species are intrinsically motivated to do it.

  5. An ultra-sparse code underliesthe generation of neural sequences in a songbird

    NASA Astrophysics Data System (ADS)

    Hahnloser, Richard H. R.; Kozhevnikov, Alexay A.; Fee, Michale S.

    2002-09-01

    Sequences of motor activity are encoded in many vertebrate brains by complex spatio-temporal patterns of neural activity; however, the neural circuit mechanisms underlying the generation of these pre-motor patterns are poorly understood. In songbirds, one prominent site of pre-motor activity is the forebrain robust nucleus of the archistriatum (RA), which generates stereotyped sequences of spike bursts during song and recapitulates these sequences during sleep. We show that the stereotyped sequences in RA are driven from nucleus HVC (high vocal centre), the principal pre-motor input to RA. Recordings of identified HVC neurons in sleeping and singing birds show that individual HVC neurons projecting onto RA neurons produce bursts sparsely, at a single, precise time during the RA sequence. These HVC neurons burst sequentially with respect to one another. We suggest that at each time in the RA sequence, the ensemble of active RA neurons is driven by a subpopulation of RA-projecting HVC neurons that is active only at that time. As a population, these HVC neurons may form an explicit representation of time in the sequence. Such a sparse representation, a temporal analogue of the `grandmother cell' concept for object recognition, eliminates the problem of temporal interference during sequence generation and learning attributed to more distributed representations.

  6. Effects of spectral and temporal disruption on cortical encoding of gerbil vocalizations

    PubMed Central

    Ter-Mikaelian, Maria; Semple, Malcolm N.

    2013-01-01

    Animal communication sounds contain spectrotemporal fluctuations that provide powerful cues for detection and discrimination. Human perception of speech is influenced both by spectral and temporal acoustic features but is most critically dependent on envelope information. To investigate the neural coding principles underlying the perception of communication sounds, we explored the effect of disrupting the spectral or temporal content of five different gerbil call types on neural responses in the awake gerbil's primary auditory cortex (AI). The vocalizations were impoverished spectrally by reduction to 4 or 16 channels of band-passed noise. For this acoustic manipulation, an average firing rate of the neuron did not carry sufficient information to distinguish between call types. In contrast, the discharge patterns of individual AI neurons reliably categorized vocalizations composed of only four spectral bands with the appropriate natural token. The pooled responses of small populations of AI cells classified spectrally disrupted and natural calls with an accuracy that paralleled human performance on an analogous speech task. To assess whether discharge pattern was robust to temporal perturbations of an individual call, vocalizations were disrupted by time-reversing segments of variable duration. For this acoustic manipulation, cortical neurons were relatively insensitive to short reversal lengths. Consistent with human perception of speech, these results indicate that the stable representation of communication sounds in AI is more dependent on sensitivity to slow temporal envelopes than on spectral detail. PMID:23761696

  7. Vibratory regime classification of infant phonation.

    PubMed

    Buder, Eugene H; Chorna, Lesya B; Oller, D Kimbrough; Robinson, Rebecca B

    2008-09-01

    Infant phonation is highly variable in many respects, including the basic vibratory patterns by which the vocal tissues create acoustic signals. Previous studies have identified the regular occurrence of nonmodal phonation types in normal infant phonation. The glottis is like many oscillating systems that, because of nonlinear relationships among the elements, may vibrate in ways representing the deterministic patterns classified theoretically within the mathematical framework of nonlinear dynamics. The infant's preverbal vocal explorations present such a variety of phonations that it may be possible to find effectively all the classes of vibration predicted by nonlinear dynamic theory. The current report defines acoustic criteria for an important subset of such vibratory regimes, and demonstrates that analysts can be trained to reliably use these criteria for a classification that includes all instances of infant phonation in the recorded corpora. The method is thus internally comprehensive in the sense that all phonations are classified, but it is not exhaustive in the sense that all vocal qualities are thereby represented. Using the methods thus developed, this study also demonstrates that the distributions of these phonation types vary significantly across sessions of recording in the first year of life, suggesting developmental changes. The method of regime classification is thus capable of tracking changes that may be indicative of maturation of the mechanism, the learning of categories of phonatory control, and the possibly varying use of vocalizations across social contexts.

  8. Multi-voxel Patterns Reveal Functionally Differentiated Networks Underlying Auditory Feedback Processing of Speech

    PubMed Central

    Zheng, Zane Z.; Vicente-Grabovetsky, Alejandro; MacDonald, Ewen N.; Munhall, Kevin G.; Cusack, Rhodri; Johnsrude, Ingrid S.

    2013-01-01

    The everyday act of speaking involves the complex processes of speech motor control. An important component of control is monitoring, detection and processing of errors when auditory feedback does not correspond to the intended motor gesture. Here we show, using fMRI and converging operations within a multi-voxel pattern analysis framework, that this sensorimotor process is supported by functionally differentiated brain networks. During scanning, a real-time speech-tracking system was employed to deliver two acoustically different types of distorted auditory feedback or unaltered feedback while human participants were vocalizing monosyllabic words, and to present the same auditory stimuli while participants were passively listening. Whole-brain analysis of neural-pattern similarity revealed three functional networks that were differentially sensitive to distorted auditory feedback during vocalization, compared to during passive listening. One network of regions appears to encode an ‘error signal’ irrespective of acoustic features of the error: this network, including right angular gyrus, right supplementary motor area, and bilateral cerebellum, yielded consistent neural patterns across acoustically different, distorted feedback types, only during articulation (not during passive listening). In contrast, a fronto-temporal network appears sensitive to the speech features of auditory stimuli during passive listening; this preference for speech features was diminished when the same stimuli were presented as auditory concomitants of vocalization. A third network, showing a distinct functional pattern from the other two, appears to capture aspects of both neural response profiles. Taken together, our findings suggest that auditory feedback processing during speech motor control may rely on multiple, interactive, functionally differentiated neural systems. PMID:23467350

  9. Convergence of pattern generator outputs on a common mechanism of diaphragm motor unit recruitment

    PubMed Central

    Mantilla, Carlos B.; Seven, Yasin B.; Sieck, Gary C.

    2014-01-01

    Motor units are the final element of neuromotor control. In manner analogous to the organization of neuromotor control in other skeletal muscles, diaphragm motor units comprise phrenic motoneurons located in the cervical spinal cord that innervate the diaphragm muscle, the main inspiratory muscle in mammals. Diaphragm motor units play a primary role in sustaining ventilation, but are also active in other non-ventilatory behaviors, including coughing, sneezing, vomiting, defecation and parturition. Diaphragm muscle fibers comprise all fiber types. Thus, diaphragm motor units display substantial differences in contractile and fatigue properties, but importantly properties of the motoneuron and muscle fibers within a motor unit are matched. As in other skeletal muscles, diaphragm motor units are recruited in order such that motor units that display greater fatigue resistance are recruited earlier and more often than more fatigable motor units. The properties of the motor unit population are critical determinants of the function of a skeletal muscle across the range of possible motor tasks. Accordingly, fatigue-resistant motor units are sufficient to generate the forces necessary for ventilatory behaviors whereas more fatigable units are only activated during expulsive behaviors important for airway clearance. Neuromotor control of diaphragm motor units may reflect selective inputs from distinct pattern generators distributed according to the motor unit properties necessary to accomplish these different motor tasks. In contrast, widely-distributed inputs to phrenic motoneurons from various pattern generators (e.g., for breathing, coughing or vocalization) would dictate recruitment order based on intrinsic electrophysiological properties. PMID:24746055

  10. Laryngeal electromyography as a diagnostic tool for Parkinson's disease.

    PubMed

    Zarzur, Ana P; Duprat, André de Campos; Cataldo, Berenice O; Ciampi, Daniel; Fonoff, Erich

    2014-03-01

    To study the laryngeal electromyography pattern in patients with Parkinson's disease (PD) and vocal complaints at different stages of the disease. Cross-sectional cohort study. Ninety-four adults with PD and vocal complaints at different stages of the disease (according to the Hoehn and Yahr scale) underwent laryngeal electromyography. Tremors were not detected on laryngeal electromyography of the cricothyroid and thyroarytenoid muscles even in patients with clinical tremor. Laryngeal electromyography hypercontractility during voice rest was the typical result observed in 91.5% of patients regardless of disease severity. Gender and age of subjects did not correlate with laryngeal electromyography results. Patients with PD presented spontaneous intrinsic laryngeal muscle activity during voice rest, regardless of disease severity. This study was significant because it reported on the use of laryngeal electromyography in a large number of patients with PD and vocal complaints grouped according to PD severity. The patterns observed suggest that laryngeal electromyography is a valuable diagnostic tool for PD even at early phases of the disease. © 2013 The American Laryngological, Rhinological and Otological Society, Inc.

  11. Rhesus macaques recognize unique multi-modal face-voice relations of familiar individuals and not of unfamiliar ones

    PubMed Central

    Habbershon, Holly M.; Ahmed, Sarah Z.; Cohen, Yale E.

    2013-01-01

    Communication signals in non-human primates are inherently multi-modal. However, for laboratory-housed monkeys, there is relatively little evidence in support of the use of multi-modal communication signals in individual recognition. Here, we used a preferential-looking paradigm to test whether laboratory-housed rhesus could “spontaneously” (i.e., in the absence of operant training) use multi-modal communication stimuli to discriminate between known conspecifics. The multi-modal stimulus was a silent movie of two monkeys vocalizing and an audio file of the vocalization from one of the monkeys in the movie. We found that the gaze patterns of those monkeys that knew the individuals in the movie were reliably biased toward the individual that did not produce the vocalization. In contrast, there was not a systematic gaze pattern for those monkeys that did not know the individuals in the movie. These data are consistent with the hypothesis that laboratory-housed rhesus can recognize and distinguish between conspecifics based on auditory and visual communication signals. PMID:23774779

  12. Computation of physiological human vocal fold parameters by mathematical optimization of a biomechanical model

    PubMed Central

    Yang, Anxiong; Stingl, Michael; Berry, David A.; Lohscheller, Jörg; Voigt, Daniel; Eysholdt, Ulrich; Döllinger, Michael

    2011-01-01

    With the use of an endoscopic, high-speed camera, vocal fold dynamics may be observed clinically during phonation. However, observation and subjective judgment alone may be insufficient for clinical diagnosis and documentation of improved vocal function, especially when the laryngeal disease lacks any clear morphological presentation. In this study, biomechanical parameters of the vocal folds are computed by adjusting the corresponding parameters of a three-dimensional model until the dynamics of both systems are similar. First, a mathematical optimization method is presented. Next, model parameters (such as pressure, tension and masses) are adjusted to reproduce vocal fold dynamics, and the deduced parameters are physiologically interpreted. Various combinations of global and local optimization techniques are attempted. Evaluation of the optimization procedure is performed using 50 synthetically generated data sets. The results show sufficient reliability, including 0.07 normalized error, 96% correlation, and 91% accuracy. The technique is also demonstrated on data from human hemilarynx experiments, in which a low normalized error (0.16) and high correlation (84%) values were achieved. In the future, this technique may be applied to clinical high-speed images, yielding objective measures with which to document improved vocal function of patients with voice disorders. PMID:21877808

  13. Gestures, vocalizations, and memory in language origins.

    PubMed

    Aboitiz, Francisco

    2012-01-01

    THIS ARTICLE DISCUSSES THE POSSIBLE HOMOLOGIES BETWEEN THE HUMAN LANGUAGE NETWORKS AND COMPARABLE AUDITORY PROJECTION SYSTEMS IN THE MACAQUE BRAIN, IN AN ATTEMPT TO RECONCILE TWO EXISTING VIEWS ON LANGUAGE EVOLUTION: one that emphasizes hand control and gestures, and the other that emphasizes auditory-vocal mechanisms. The capacity for language is based on relatively well defined neural substrates whose rudiments have been traced in the non-human primate brain. At its core, this circuit constitutes an auditory-vocal sensorimotor circuit with two main components, a "ventral pathway" connecting anterior auditory regions with anterior ventrolateral prefrontal areas, and a "dorsal pathway" connecting auditory areas with parietal areas and with posterior ventrolateral prefrontal areas via the arcuate fasciculus and the superior longitudinal fasciculus. In humans, the dorsal circuit is especially important for phonological processing and phonological working memory, capacities that are critical for language acquisition and for complex syntax processing. In the macaque, the homolog of the dorsal circuit overlaps with an inferior parietal-premotor network for hand and gesture selection that is under voluntary control, while vocalizations are largely fixed and involuntary. The recruitment of the dorsal component for vocalization behavior in the human lineage, together with a direct cortical control of the subcortical vocalizing system, are proposed to represent a fundamental innovation in human evolution, generating an inflection point that permitted the explosion of vocal language and human communication. In this context, vocal communication and gesturing have a common history in primate communication.

  14. Auditory responses in the amygdala to social vocalizations

    NASA Astrophysics Data System (ADS)

    Gadziola, Marie A.

    The underlying goal of this dissertation is to understand how the amygdala, a brain region involved in establishing the emotional significance of sensory input, contributes to the processing of complex sounds. The general hypothesis is that communication calls of big brown bats (Eptesicus fuscus) transmit relevant information about social context that is reflected in the activity of amygdalar neurons. The first specific aim analyzed social vocalizations emitted under a variety of behavioral contexts, and related vocalizations to an objective measure of internal physiological state by monitoring the heart rate of vocalizing bats. These experiments revealed a complex acoustic communication system among big brown bats in which acoustic cues and call structure signal the emotional state of a sender. The second specific aim characterized the responsiveness of single neurons in the basolateral amygdala to a range of social syllables. Neurons typically respond to the majority of tested syllables, but effectively discriminate among vocalizations by varying the response duration. This novel coding strategy underscores the importance of persistent firing in the general functioning of the amygdala. The third specific aim examined the influence of acoustic context by characterizing both the behavioral and neurophysiological responses to natural vocal sequences. Vocal sequences differentially modify the internal affective state of a listening bat, with lower aggression vocalizations evoking the greatest change in heart rate. Amygdalar neurons employ two different coding strategies: low background neurons respond selectively to very few stimuli, whereas high background neurons respond broadly to stimuli but demonstrate variation in response magnitude and timing. Neurons appear to discriminate the valence of stimuli, with aggression sequences evoking robust population-level responses across all sound levels. Further, vocal sequences show improved discrimination among stimuli compared to isolated syllables, and this improved discrimination is expressed in part by the timing of action potentials. Taken together, these data support the hypothesis that big brown bat social vocalizations transmit relevant information about the social context that is encoded within the discharge pattern of amygdalar neurons ultimately responsible for coordinating appropriate social behaviors. I further propose that vocalization-evoked amygdalar activity will have significant impact on subsequent sensory processing and plasticity.

  15. Singing activity-driven Arc expression associated with vocal acoustic plasticity in juvenile songbird.

    PubMed

    Hayase, Shin; Wada, Kazuhiro

    2018-06-23

    Learned vocalization, including birdsong and human speech, is acquired through self-motivated vocal practice during the sensitive period of vocal learning. The zebra finch (Taeniopygia guttata) develops a song characterized by vocal variability and crystalizes a defined song pattern as adulthood. However, it remains unknown how vocal variability is regulated with diurnal singing during the sensorimotor learning period. Here, we investigated the expression of activity-dependent neuroplasticity-related gene Arc during the early plastic song phase to examine its potential association with vocal plasticity. We first confirmed that multiple acoustic features of syllables in the plastic song were dramatically and simultaneously modulated during the first 3 hours of singing in a day and the altered features were maintained until sleep. Concurrently, Arc was intensely induced during morning singing and a subsequent attenuation during afternoon singing in the robust nucleus of the arcopallium (RA) and the interfacial nucleus of the nidopallium (NIf). The singing-driven Arc expression was not altered by circadian rhythm, but rather reduced during the day as juveniles produced more songs. Song stabilization accelerated by testosterone administration in juveniles was accompanied with attenuation of Arc induction in RA and NIf. In contrast, although early-deafened birds produced highly unstable song even at adulthood, singing-driven Arc expression was not different between intact and early-deafened adults. These results suggest a potential functional link between Arc expression in RA and NIf and vocal plasticity during the sensorimotor phase of song learning. Nonetheless, Arc expression did not reflect the quality of bird's own song or auditory feedback. This article is protected by copyright. All rights reserved. This article is protected by copyright. All rights reserved.

  16. Neurotensin neural mRNA expression correlates with vocal communication and other highly-motivated social behaviors in male European starlings.

    PubMed

    Merullo, Devin P; Cordes, Melissa A; Susan DeVries, M; Stevenson, Sharon A; Riters, Lauren V

    2015-11-01

    Vocalizations coordinate social interactions in many species and often are important for behaviors such as mate attraction or territorial defense. Although the neural circuitry underlying vocal communication is well-known for some animal groups, such as songbirds, the motivational processes that regulate vocal signals are not as clearly understood. Neurotensin (NT) is a neuropeptide implicated in motivation that can modulate the activity of dopaminergic neurons. Dopaminergic projections from the ventral tegmental area (VTA) are key to mediating highly motivated, goal-directed behaviors, including sexually-motivated birdsong. However, the role of NT in modifying vocal communication or other social behaviors has not been well-studied. Here in European starlings (Sturnus vulgaris) we analyzed relationships between sexually-motivated song and NT and NT1 receptor (NTSR1) expression in VTA. Additionally, we examined NT and NTSR1 expression in four regions that receive dopaminergic projections from VTA and are involved in courtship song: the medial preoptic nucleus (POM), the lateral septum (LS), Area X, and HVC. Relationships between NT and NTSR1 expression and non-vocal courtship and agonistic behaviors were also examined. NT expression in Area X positively related to sexually-motivated song production. NT expression in POM positively correlated with non-vocal courtship behavior and agonistic behavior. NT expression in POM was greatest in males owning nesting sites, and the opposite pattern was observed for NTSR1 expression in LS. These results are the first to implicate NT in Area X in birdsong, and further highlight NT as a potential neuromodulator for the control of vocal communication and other social behaviors. Copyright © 2015 Elsevier Inc. All rights reserved.

  17. The larynx of roaring and non-roaring cats.

    PubMed Central

    Hast, M H

    1989-01-01

    Dissections were made of the larynges of 14 species of the cat family, with representative specimens from all genera. It was found that the vocal folds of the larynx of genus Panthera (with the exception of the snow leopard) form the basic structure of a sound generator well-designed to produce a high acoustical energy. Combined with an efficient sound radiator (vocal tract) that can be adjusted in length, a Panthera can use its vocal instrument literally to blow its own horn with a 'roar'. Also, it is proposed that laryngeal morphology can be used as an anatomical character in mammalian taxonomy. Images Fig. 1 PMID:2606766

  18. The role of nocturnal vision in mate choice: females prefer conspicuous males in the European tree frog (Hyla arborea)

    PubMed Central

    Gomez, Doris; Richardson, Christina; Lengagne, Thierry; Plenet, Sandrine; Joly, Pierre; Léna, Jean-Paul; Théry, Marc

    2009-01-01

    Nocturnal frog species rely extensively on vocalization for reproduction. But recent studies provide evidence for an important, though long overlooked, role of visual communication. In many species, calling males exhibit a conspicuous pulsing vocal sac, a signal bearing visually important dynamic components. Here, we investigate female preference for male vocal sac coloration—a question hitherto unexplored—and male colour pattern in the European tree frog (Hyla arborea). Under nocturnal conditions, we conducted two-choice experiments involving video playbacks of calling males with identical calls and showing various naturally encountered colour signals, differing in their chromatic and brightness components. We adjusted video colours to match the frogs' visual perception, a crucial aspect not considered in previous experiments. Females prefer males with a colourful sac and a pronounced flank stripe. Both signals probably enhance male conspicuousness and facilitate detection and localization by females. This study provides the first experimental evidence of a preference for specific vocal sac spectral properties in a nocturnal anuran species. Vocal sac coloration is based on carotenoids and may convey information about male quality worthwhile for females to assess. The informative content of the flank stripe remains to be demonstrated. PMID:19324736

  19. Central Nervous System Control of Voice and Swallowing

    PubMed Central

    Ludlow, Christy L.

    2015-01-01

    This review of the central nervous control systems for voice and swallowing has suggested that the traditional concepts of a separation between cortical and limbic and brain stem control should be refined and more integrative. For voice production, a separation of the non-human vocalization system from the human learned voice production system has been posited based primarily on studies of non-human primates. However, recent humans studies of emotionally based vocalizations and human volitional voice production has shown more integration between these two systems than previously proposed. Recent human studies have shown that reflexive vocalization as well as learned voice production not involving speech, involve a common integrative system. On the other hand, recent studies of non-human primates have provided evidence of some cortical activity during vocalization and cortical changes with training during vocal behavior. For swallowing, evidence from the macaque and functional brain imaging in humans indicates that the control for the pharyngeal phase of swallowing is not primarily under brain stem mechanisms as previously proposed. Studies suggest that the initiation and patterning of swallowing for the pharyngeal phase is also under active cortical control for both spontaneous as well as volitional swallowing in awake humans and non-human primates. PMID:26241238

  20. Cicadas impact bird communication in a noisy tropical rainforest

    PubMed Central

    Hall, Robert; Ray, William; Beck, Angela; Zook, James

    2015-01-01

    Many animals communicate through acoustic signaling, and “acoustic space” may be viewed as a limited resource that organisms compete for. If acoustic signals overlap, the information in them is masked, so there should be selection toward strategies that reduce signal overlap. The extent to which animals are able to partition acoustic space in acoustically diverse habitats such as tropical forests is poorly known. Here, we demonstrate that a single cicada species plays a major role in the frequency and timing of acoustic communication in a neotropical wet forest bird community. Using an automated acoustic monitor, we found that cicadas vary the timing of their signals throughout the day and that the frequency range and timing of bird vocalizations closely track these signals. Birds significantly avoid temporal overlap with cicadas by reducing and often shutting down vocalizations at the onset of cicada signals that utilize the same frequency range. When birds do vocalize at the same time as cicadas, the vocalizations primarily occur at nonoverlapping frequencies with cicada signals. Our results greatly improve our understanding of the community dynamics of acoustic signaling and reveal how patterns in biotic noise shape the frequency and timing of bird vocalizations in tropical forests. PMID:26023277

  1. Language-related Cntnap2 gene is differentially expressed in sexually dimorphic song nuclei essential for vocal learning in songbirds

    PubMed Central

    Panaitof, S. Carmen; Abrahams, Brett S.; Dong, Hongmei; Geschwind, Daniel H.; White, Stephanie A.

    2010-01-01

    Multiple studies, involving distinct clinical populations, implicate contactin associated protein-like 2 (CNTNAP2) in aspects of language development and performance. While CNTNAP2 is broadly distributed in developing rodent brain, it shows a striking gradient of frontal cortical enrichment in developing human brain, consistent with a role in patterning circuits that subserve higher cognition and language. To test the hypothesis that CNTNAP2 may be important for learned vocal communication in additional species, we employed in situ hybridization to characterize transcript distribution in the zebra finch, an experimentally tractable songbird for which the neural substrate of this behavior is well-established. Consistent with an important role in learned vocalization, Cntnap2 was enriched or diminished in key song control nuclei relative to adjacent brain tissue. Importantly, this punctuated expression was observed in males, but not females, in accord with the sexual dimorphism of neural circuitry and vocal learning in this species. Ongoing functional work will provide important insights into the relationship between Cntnap2 and vocal communication in songbirds and thereby clarify mechanisms at play in disorders of human cognition and language. PMID:20394055

  2. Automated Vocal Analysis of Children with Hearing Loss and Their Typical and Atypical Peers

    PubMed Central

    VanDam, Mark; Oller, D. Kimbrough; Ambrose, Sophie E.; Gray, Sharmistha; Richards, Jeffrey A.; Xu, Dongxin; Gilkerson, Jill; Silbert, Noah H.; Moeller, Mary Pat

    2014-01-01

    Objectives This study investigated automatic assessment of vocal development in children with hearing loss as compared with children who are typically developing, have language delays, and autism spectrum disorder. Statistical models are examined for performance in a classification model and to predict age within the four groups of children. Design The vocal analysis system analyzed over 1900 whole-day, naturalistic acoustic recordings from 273 toddlers and preschoolers comprising children who were typically developing, hard of hearing, language delayed, or autistic. Results Samples from children who were hard-of-hearing patterned more similarly to those of typically-developing children than to the language-delayed or autistic samples. The statistical models were able to classify children from the four groups examined and estimate developmental age based on automated vocal analysis. Conclusions This work shows a broad similarity between children with hearing loss and typically developing children, although children with hearing loss show some delay in their production of speech. Automatic acoustic analysis can now be used to quantitatively compare vocal development in children with and without speech-related disorders. The work may serve to better distinguish among various developmental disorders and ultimately contribute to improved intervention. PMID:25587667

  3. Gelada vocal sequences follow Menzerath's linguistic law.

    PubMed

    Gustison, Morgan L; Semple, Stuart; Ferrer-I-Cancho, Ramon; Bergman, Thore J

    2016-05-10

    Identifying universal principles underpinning diverse natural systems is a key goal of the life sciences. A powerful approach in addressing this goal has been to test whether patterns consistent with linguistic laws are found in nonhuman animals. Menzerath's law is a linguistic law that states that, the larger the construct, the smaller the size of its constituents. Here, to our knowledge, we present the first evidence that Menzerath's law holds in the vocal communication of a nonhuman species. We show that, in vocal sequences of wild male geladas (Theropithecus gelada), construct size (sequence size in number of calls) is negatively correlated with constituent size (duration of calls). Call duration does not vary significantly with position in the sequence, but call sequence composition does change with sequence size and most call types are abbreviated in larger sequences. We also find that intercall intervals follow the same relationship with sequence size as do calls. Finally, we provide formal mathematical support for the idea that Menzerath's law reflects compression-the principle of minimizing the expected length of a code. Our findings suggest that a common principle underpins human and gelada vocal communication, highlighting the value of exploring the applicability of linguistic laws in vocal systems outside the realm of language.

  4. Perceptual fluency and judgments of vocal aesthetics and stereotypicality.

    PubMed

    Babel, Molly; McGuire, Grant

    2015-05-01

    Research has shown that processing dynamics on the perceiver's end determine aesthetic pleasure. Specifically, typical objects, which are processed more fluently, are perceived as more attractive. We extend this notion of perceptual fluency to judgments of vocal aesthetics. Vocal attractiveness has traditionally been examined with respect to sexual dimorphism and the apparent size of a talker, as reconstructed from the acoustic signal, despite evidence that gender-specific speech patterns are learned social behaviors. In this study, we report on a series of three experiments using 60 voices (30 females) to compare the relationship between judgments of vocal attractiveness, stereotypicality, and gender categorization fluency. Our results indicate that attractiveness and stereotypicality are highly correlated for female and male voices. Stereotypicality and categorization fluency were also correlated for male voices, but not female voices. Crucially, stereotypicality and categorization fluency interacted to predict attractiveness, suggesting the role of perceptual fluency is present, but nuanced, in judgments of human voices. © 2014 Cognitive Science Society, Inc.

  5. A multiscale product approach for an automatic classification of voice disorders from endoscopic high-speed videos.

    PubMed

    Unger, Jakob; Schuster, Maria; Hecker, Dietmar J; Schick, Bernhard; Lohscheller, Joerg

    2013-01-01

    Direct observation of vocal fold vibration is indispensable for a clinical diagnosis of voice disorders. Among current imaging techniques, high-speed videoendoscopy constitutes a state-of-the-art method capturing several thousand frames per second of the vocal folds during phonation. Recently, a method for extracting descriptive features from phonovibrograms, a two-dimensional image containing the spatio-temporal pattern of vocal fold dynamics, was presented. The derived features are closely related to a clinically established protocol for functional assessment of pathologic voices. The discriminative power of these features for different pathologic findings and configurations has not been assessed yet. In the current study, a collective of 220 subjects is considered for two- and multi-class problems of healthy and pathologic findings. The performance of the proposed feature set is compared to conventional feature reduction routines and was found to clearly outperform these. As such, the proposed procedure shows great potential for diagnostical issues of vocal fold disorders.

  6. Linear Classifier with Reject Option for the Detection of Vocal Fold Paralysis and Vocal Fold Edema

    NASA Astrophysics Data System (ADS)

    Kotropoulos, Constantine; Arce, Gonzalo R.

    2009-12-01

    Two distinct two-class pattern recognition problems are studied, namely, the detection of male subjects who are diagnosed with vocal fold paralysis against male subjects who are diagnosed as normal and the detection of female subjects who are suffering from vocal fold edema against female subjects who do not suffer from any voice pathology. To do so, utterances of the sustained vowel "ah" are employed from the Massachusetts Eye and Ear Infirmary database of disordered speech. Linear prediction coefficients extracted from the aforementioned utterances are used as features. The receiver operating characteristic curve of the linear classifier, that stems from the Bayes classifier when Gaussian class conditional probability density functions with equal covariance matrices are assumed, is derived. The optimal operating point of the linear classifier is specified with and without reject option. First results using utterances of the "rainbow passage" are also reported for completeness. The reject option is shown to yield statistically significant improvements in the accuracy of detecting the voice pathologies under study.

  7. The Linked Dual Representation model of vocal perception and production

    PubMed Central

    Hutchins, Sean; Moreno, Sylvain

    2013-01-01

    The voice is one of the most important media for communication, yet there is a wide range of abilities in both the perception and production of the voice. In this article, we review this range of abilities, focusing on pitch accuracy as a particularly informative case, and look at the factors underlying these abilities. Several classes of models have been posited describing the relationship between vocal perception and production, and we review the evidence for and against each class of model. We look at how the voice is different from other musical instruments and review evidence about both the association and the dissociation between vocal perception and production abilities. Finally, we introduce the Linked Dual Representation (LDR) model, a new approach which can account for the broad patterns in prior findings, including trends in the data which might seem to be countervailing. We discuss how this model interacts with higher-order cognition and examine its predictions about several aspects of vocal perception and production. PMID:24204360

  8. Vocalisation Repertoire of Female Bluefin Gurnard (Chelidonichthys kumu) in Captivity: Sound Structure, Context and Vocal Activity.

    PubMed

    Radford, Craig A; Ghazali, Shahriman M; Montgomery, John C; Jeffs, Andrew G

    2016-01-01

    Fish vocalisation is often a major component of underwater soundscapes. Therefore, interpretation of these soundscapes requires an understanding of the vocalisation characteristics of common soniferous fish species. This study of captive female bluefin gurnard, Chelidonichthys kumu, aims to formally characterise their vocalisation sounds and daily pattern of sound production. Four types of sound were produced and characterised, twice as many as previously reported in this species. These sounds fit two aural categories; grunt and growl, the mean peak frequencies for which ranged between 129 to 215 Hz. This species vocalized throughout the 24 hour period at an average rate of (18.5 ± 2.0 sounds fish-1 h-1) with an increase in vocalization rate at dawn and dusk. Competitive feeding did not elevate vocalisation as has been found in other gurnard species. Bluefin gurnard are common in coastal waters of New Zealand, Australia and Japan and, given their vocalization rate, are likely to be significant contributors to ambient underwater soundscape in these areas.

  9. Vocalisation Repertoire of Female Bluefin Gurnard (Chelidonichthys kumu) in Captivity: Sound Structure, Context and Vocal Activity

    PubMed Central

    Radford, Craig A.; Ghazali, Shahriman M.; Montgomery, John C.; Jeffs, Andrew G.

    2016-01-01

    Fish vocalisation is often a major component of underwater soundscapes. Therefore, interpretation of these soundscapes requires an understanding of the vocalisation characteristics of common soniferous fish species. This study of captive female bluefin gurnard, Chelidonichthys kumu, aims to formally characterise their vocalisation sounds and daily pattern of sound production. Four types of sound were produced and characterised, twice as many as previously reported in this species. These sounds fit two aural categories; grunt and growl, the mean peak frequencies for which ranged between 129 to 215 Hz. This species vocalized throughout the 24 hour period at an average rate of (18.5 ± 2.0 sounds fish-1 h-1) with an increase in vocalization rate at dawn and dusk. Competitive feeding did not elevate vocalisation as has been found in other gurnard species. Bluefin gurnard are common in coastal waters of New Zealand, Australia and Japan and, given their vocalization rate, are likely to be significant contributors to ambient underwater soundscape in these areas. PMID:26890124

  10. What's in a voice? Prosody as a test case for the Theory of Mind account of autism.

    PubMed

    Chevallier, Coralie; Noveck, Ira; Happé, Francesca; Wilson, Deirdre

    2011-02-01

    The human voice conveys a variety of information about people's feelings, emotions and mental states. Some of this information relies on sophisticated Theory of Mind (ToM) skills, whilst others are simpler and do not require ToM. This variety provides an interesting test case for the ToM account of autism, which would predict greater impairment as ToM requirements increase. In this paper, we draw on psychological and pragmatic theories to classify vocal cues according to the amount of mindreading required to identify them. Children with a high functioning Autism Spectrum Disorder and matched controls were tested in three experiments where the speakers' state had to be extracted from their vocalizations. Although our results confirm that people with autism have subtle difficulties dealing with vocal cues, they show a pattern of performance that is inconsistent with the view that atypical recognition of vocal cues is caused by impaired ToM. Copyright © 2010 Elsevier Ltd. All rights reserved.

  11. Assessment of vocal cord nodules: a case study in speech processing by using Hilbert-Huang Transform

    NASA Astrophysics Data System (ADS)

    Civera, M.; Filosi, C. M.; Pugno, N. M.; Silvestrini, M.; Surace, C.; Worden, K.

    2017-05-01

    Vocal cord nodules represent a pathological condition for which the growth of unnatural masses on vocal folds affects the patients. Among other effects, changes in the vocal cords’ overall mass and stiffness alter their vibratory behaviour, thus changing the vocal emission generated by them. This causes dysphonia, i.e. abnormalities in the patients’ voice, which can be analysed and inspected via audio signals. However, the evaluation of voice condition through speech processing is not a trivial task, as standard methods based on the Fourier Transform, fail to fit the non-stationary nature of vocal signals. In this study, four audio tracks, provided by a volunteer patient, whose vocal fold nodules have been surgically removed, were analysed using a relatively new technique: the Hilbert-Huang Transform (HHT) via Empirical Mode Decomposition (EMD); specifically, by using the CEEMDAN (Complete Ensemble EMD with Adaptive Noise) algorithm. This method has been applied here to speech signals, which were recorded before removal surgery and during convalescence, to investigate specific trends. Possibilities offered by the HHT are exposed, but also some limitations of decomposing the signals into so-called intrinsic mode functions (IMFs) are highlighted. The results of these preliminary studies are intended to be a basis for the development of new viable alternatives to the softwares currently used for the analysis and evaluation of pathological voice.

  12. Proton density-weighted laryngeal magnetic resonance imaging in systemically dehydrated rats.

    PubMed

    Oleson, Steven; Lu, Kun-Han; Liu, Zhongming; Durkes, Abigail C; Sivasankar, M Preeti

    2018-06-01

    Dehydrated vocal folds are inefficient sound generators. Although systemic dehydration of the body is believed to induce vocal fold dehydration, this causative relationship has not been demonstrated in vivo. Here we investigate the feasibility of using in vivo proton density (PD)-weighted magnetic resonance imaging (MRI) to demonstrate hydration changes in vocal fold tissue following systemic dehydration in rats. Animal study. Sprague-Dawley rats (n = 10) were imaged at baseline and following a 10% reduction in body weight secondary to withholding water. In vivo, high-field (7 T), PD-weighted MRI was used to successfully resolve vocal fold and salivary gland tissue structures. Normalized signal intensities within the vocal fold decreased postdehydration by an average of 11.38% ± 3.95% (mean ± standard error of the mean [SEM], P = .0098) as compared to predehydration levels. The salivary glands experienced a similar decrease in normalized signal intensity by an average of 10.74% ± 4.14% (mean ± SEM, P = .0195) following dehydration. The correlation coefficient (percent change from dehydration) between vocal folds and salivary glands was 0.7145 (P = .0202). Ten percent systemic dehydration induced vocal fold dehydration as assessed by PD-weighted MRI. Changes in the hydration state of vocal fold tissue were highly correlated with that of the salivary glands in dehydrated rats in vivo. These preliminary findings demonstrate the feasibility of using PD-weighted MRI to quantify hydration states of the vocal folds and lay the foundation for further studies that explore more routine and realistic magnitudes of systemic dehydration and rehydration. NA. Laryngoscope, 128:E222-E227, 2018. © 2017 The American Laryngological, Rhinological and Otological Society, Inc.

  13. Syllabic Patterns in the Early Vocalizations of Quichua Children

    ERIC Educational Resources Information Center

    Gildersleeve-Neumann, Christina E.; Davis, Barbara L.; Macneilage, Peter F.

    2013-01-01

    To understand the interactions between production patterns common to children regardless of language environment and the early appearance of production effects based on perceptual learning from the ambient language requires the study of languages with diverse phonological properties. Few studies have evaluated early phonological acquisition…

  14. How do auditory cortex neurons represent communication sounds?

    PubMed

    Gaucher, Quentin; Huetz, Chloé; Gourévitch, Boris; Laudanski, Jonathan; Occelli, Florian; Edeline, Jean-Marc

    2013-11-01

    A major goal in auditory neuroscience is to characterize how communication sounds are represented at the cortical level. The present review aims at investigating the role of auditory cortex in the processing of speech, bird songs and other vocalizations, which all are spectrally and temporally highly structured sounds. Whereas earlier studies have simply looked for neurons exhibiting higher firing rates to particular conspecific vocalizations over their modified, artificially synthesized versions, more recent studies determined the coding capacity of temporal spike patterns, which are prominent in primary and non-primary areas (and also in non-auditory cortical areas). In several cases, this information seems to be correlated with the behavioral performance of human or animal subjects, suggesting that spike-timing based coding strategies might set the foundations of our perceptive abilities. Also, it is now clear that the responses of auditory cortex neurons are highly nonlinear and that their responses to natural stimuli cannot be predicted from their responses to artificial stimuli such as moving ripples and broadband noises. Since auditory cortex neurons cannot follow rapid fluctuations of the vocalizations envelope, they only respond at specific time points during communication sounds, which can serve as temporal markers for integrating the temporal and spectral processing taking place at subcortical relays. Thus, the temporal sparse code of auditory cortex neurons can be considered as a first step for generating high level representations of communication sounds independent of the acoustic characteristic of these sounds. This article is part of a Special Issue entitled "Communication Sounds and the Brain: New Directions and Perspectives". Copyright © 2013 Elsevier B.V. All rights reserved.

  15. In the ear of the beholder: how age shapes emotion processing in nonverbal vocalizations.

    PubMed

    Lima, César F; Alves, Tiago; Scott, Sophie K; Castro, São Luís

    2014-02-01

    It is well established that emotion recognition of facial expressions declines with age, but evidence for age-related differences in vocal emotions is more limited. This is especially true for nonverbal vocalizations such as laughter, sobs, or sighs. In this study, 43 younger adults (M = 22 years) and 43 older ones (M = 61.4 years) provided multiple emotion ratings of nonverbal emotional vocalizations. Contrasting with previous research, which often includes only one positive emotion (happiness) versus several negative ones, we examined 4 positive and 4 negative emotions: achievement/triumph, amusement, pleasure, relief, anger, disgust, fear, and sadness. We controlled for hearing loss and assessed general cognitive decline, cognitive control, verbal intelligence, working memory, current affect, emotion regulation, and personality. Older adults were less sensitive than younger ones to the intended vocal emotions, as indicated by decrements in ratings on the intended emotion scales and accuracy. These effects were similar for positive and negative emotions, and they were independent of age-related differences in cognitive, affective, and personality measures. Regression analyses revealed that younger and older participants' responses could be predicted from the acoustic properties of the temporal, intensity, fundamental frequency, and spectral profile of the vocalizations. The two groups were similarly efficient in using the acoustic cues, but there were differences in the patterns of emotion-specific predictors. This study suggests that ageing produces specific changes on the processing of nonverbal vocalizations. That decrements were not attenuated for positive emotions indicates that they cannot be explained by a positivity effect in older adults. PsycINFO Database Record (c) 2014 APA, all rights reserved.

  16. Teachers' voice use in teaching environments: a field study using ambulatory phonation monitor.

    PubMed

    Lyberg Åhlander, Viveka; Pelegrín García, David; Whitling, Susanna; Rydell, Roland; Löfqvist, Anders

    2014-11-01

    This case-control designed field study examines the vocal behavior in teachers with self-estimated voice problems (VP) and their age- and school-matched voice healthy (VH) colleagues. It was hypothesized that teachers with and teachers without VP use their voices differently regarding fundamental frequency, sound pressure level (SPL), and in relation to the background noise. Teachers with self-estimated VP (n = 14; two males and 12 females) were age and gender matched to VH school colleagues (n = 14; two males and 12 females). The subjects, recruited from an earlier study, had been examined in laryngeal, vocal, hearing, and psychosocial aspects. The fundamental frequency, SPL, and phonation time were recorded with an Ambulatory Phonation Monitor during one representative workday. The teachers reported their activities in a structured diary. The SPL (including teachers' and students' activity and ambient noise) was recorded with a sound level meter; the room temperature and air quality were measured simultaneously. The acoustic properties of the empty classrooms were measured. Teachers with VP behaved vocally different from their VH peers, in particular during teaching sessions. The phonation time was significantly higher in the group with VP, and the number of vibratory cycles differed between the female teachers. The F0 pattern, related to the vocal SPL and room acoustics, differed between the groups. The results suggest a different vocal behavior in subjects with subjective VP and a higher vocal load with fewer possibilities for vocal recovery. Copyright © 2014 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  17. Localization and Divergent Profiles of Estrogen Receptors and Aromatase in the Vocal and Auditory Networks of a Fish with Alternative Mating Tactics

    PubMed Central

    Fergus, Daniel J.; Bass, Andrew H.

    2013-01-01

    Estrogens play a salient role in the development and maintenance of both male and female nervous systems and behaviors. The plainfin midshipman (Porichthys notatus), a teleost fish, has two male reproductive morphs that follow alternative mating tactics and diverge in multiple somatic, hormonal and neural traits, including the central control of morph-specific vocal behaviors. After we identified duplicate estrogen receptors (ERβ1 and ERβ2) in midshipman, we developed antibodies to localize protein expression in the central vocal-acoustic networks and saccule, the auditory division of the inner ear. As in other teleost species, ERβ1 and ERβ2 were robustly expressed in the telencephalon and hypothalamus in vocal-acoustic and other brain regions shown previously to exhibit strong expression of ERα and aromatase (estrogen synthetase, CYP19) in midshipman. Like aromatase, ERβ1 label co-localized with glial fibrillary acidic protein (GFAP) in telencephalic radial glial cells. Quantitative PCR revealed similar patterns of transcript abundance across reproductive morphs for ERβ1, ERβ2, ERα and aromatase in the forebrain and saccule. In contrast, transcript abundance for ERs and aromatase varied significantly between morphs in and around the sexually polymorphic vocal motor nucleus (VMN). Together, the results suggest that VMN is the major estrogen target within the estrogen-sensitive hindbrain vocal network that directly determines the duration, frequency and amplitude of morph-specific vocalizations. Comparable regional differences in steroid receptor abundances likely regulate morph-specific behaviors in males and females of other species exhibiting alternative reproductive tactics. PMID:23460422

  18. Nonlinear dynamic analysis of voices before and after surgical excision of vocal polyps

    NASA Astrophysics Data System (ADS)

    Zhang, Yu; McGilligan, Clancy; Zhou, Liang; Vig, Mark; Jiang, Jack J.

    2004-05-01

    Phase space reconstruction, correlation dimension, and second-order entropy, methods from nonlinear dynamics, are used to analyze sustained vowels generated by patients before and after surgical excision of vocal polyps. Two conventional acoustic perturbation parameters, jitter and shimmer, are also employed to analyze voices before and after surgery. Presurgical and postsurgical analyses of jitter, shimmer, correlation dimension, and second-order entropy are statistically compared. Correlation dimension and second-order entropy show a statistically significant decrease after surgery, indicating reduced complexity and higher predictability of postsurgical voice dynamics. There is not a significant postsurgical difference in shimmer, although jitter shows a significant postsurgical decrease. The results suggest that jitter and shimmer should be applied to analyze disordered voices with caution; however, nonlinear dynamic methods may be useful for analyzing abnormal vocal function and quantitatively evaluating the effects of surgical excision of vocal polyps.

  19. Effects of Speech Output on Maintenance of Requesting and Frequency of Vocalizations in Three Children with Developmental Disabilities.

    PubMed

    Sigafoos, Jeff; Didden, Robert; O'Reilly, Mark

    2003-01-01

    We evaluated the role of digitized speech output on the maintenance of requesting and frequency of vocalizations in three children with developmental disabilities. The children were taught to request access to preferred objects using an augmentative communication speech-generating device (SGD). Following acquisition, rates of requesting and vocalizations were compared across two conditions (speech output on versus speech output off) that were alternated on a session-by-session basis. There were no major or consistent differences across the two conditions for the three children, suggesting that access to preferred objects was the critical variable maintaining use of the SGDs. The results also suggest that feedback in the form of digitized speech from the SGD did not inhibit vocalizations. One child began to speak single words during the latter part of the study, suggesting that in some cases AAC intervention involving SGDs may facilitate speech.

  20. A mixed-effects model approach for the statistical analysis of vocal fold viscoelastic shear properties.

    PubMed

    Xu, Chet C; Chan, Roger W; Sun, Han; Zhan, Xiaowei

    2017-11-01

    A mixed-effects model approach was introduced in this study for the statistical analysis of rheological data of vocal fold tissues, in order to account for the data correlation caused by multiple measurements of each tissue sample across the test frequency range. Such data correlation had often been overlooked in previous studies in the past decades. The viscoelastic shear properties of the vocal fold lamina propria of two commonly used laryngeal research animal species (i.e. rabbit, porcine) were measured by a linear, controlled-strain simple-shear rheometer. Along with published canine and human rheological data, the vocal fold viscoelastic shear moduli of these animal species were compared to those of human over a frequency range of 1-250Hz using the mixed-effects models. Our results indicated that tissues of the rabbit, canine and porcine vocal fold lamina propria were significantly stiffer and more viscous than those of human. Mixed-effects models were shown to be able to more accurately analyze rheological data generated from repeated measurements. Copyright © 2017 Elsevier Ltd. All rights reserved.

  1. Biosimulation of Inflammation and Healing in Surgically Injured Vocal Folds

    PubMed Central

    Li, Nicole Y. K.; Vodovotz, Yoram; Hebda, Patricia A.; Abbott, Katherine Verdolini

    2010-01-01

    Objectives The pathogenesis of vocal fold scarring is complex and remains to be deciphered. The current study is part of research endeavors aimed at applying systems biology approaches to address the complex biological processes involved in the pathogenesis of vocal fold scarring and other lesions affecting the larynx. Methods We developed a computational agent-based model (ABM) to quantitatively characterize multiple cellular and molecular interactions involved in inflammation and healing in vocal fold mucosa after surgical trauma. The ABM was calibrated with empirical data on inflammatory mediators (eg, tumor necrosis factor) and extracellular matrix components (eg, hyaluronan) from published studies on surgical vocal fold injury in the rat population. Results The simulation results reproduced and predicted trajectories seen in the empirical data from the animals. Moreover, the ABM studies suggested that hyaluronan fragments might be the clinical surrogate of tissue damage, a key variable that in these simulations both is enhanced by and further induces inflammation. Conclusions A relatively simple ABM such as the one reported in this study can provide new understanding of laryngeal wound healing and generate working hypotheses for further wet-lab studies. PMID:20583741

  2. Biosimulation of inflammation and healing in surgically injured vocal folds.

    PubMed

    Li, Nicole Y K; Vodovotz, Yoram; Hebda, Patricia A; Abbott, Katherine Verdolini

    2010-06-01

    The pathogenesis of vocal fold scarring is complex and remains to be deciphered. The current study is part of research endeavors aimed at applying systems biology approaches to address the complex biological processes involved in the pathogenesis of vocal fold scarring and other lesions affecting the larynx. We developed a computational agent-based model (ABM) to quantitatively characterize multiple cellular and molecular interactions involved in inflammation and healing in vocal fold mucosa after surgical trauma. The ABM was calibrated with empirical data on inflammatory mediators (eg, tumor necrosis factor) and extracellular matrix components (eg, hyaluronan) from published studies on surgical vocal fold injury in the rat population. The simulation results reproduced and predicted trajectories seen in the empirical data from the animals. Moreover, the ABM studies suggested that hyaluronan fragments might be the clinical surrogate of tissue damage, a key variable that in these simulations both is enhanced by and further induces inflammation. A relatively simple ABM such as the one reported in this study can provide new understanding of laryngeal wound healing and generate working hypotheses for further wet-lab studies.

  3. A New Mechanism of Sound Generation in Songbirds

    NASA Astrophysics Data System (ADS)

    Goller, Franz; Larsen, Ole N.

    1997-12-01

    Our current understanding of the sound-generating mechanism in the songbird vocal organ, the syrinx, is based on indirect evidence and theoretical treatments. The classical avian model of sound production postulates that the medial tympaniform membranes (MTM) are the principal sound generators. We tested the role of the MTM in sound generation and studied the songbird syrinx more directly by filming it endoscopically. After we surgically incapacitated the MTM as a vibratory source, zebra finches and cardinals were not only able to vocalize, but sang nearly normal song. This result shows clearly that the MTM are not the principal sound source. The endoscopic images of the intact songbird syrinx during spontaneous and brain stimulation-induced vocalizations illustrate the dynamics of syringeal reconfiguration before phonation and suggest a different model for sound production. Phonation is initiated by rostrad movement and stretching of the syrinx. At the same time, the syrinx is closed through movement of two soft tissue masses, the medial and lateral labia, into the bronchial lumen. Sound production always is accompanied by vibratory motions of both labia, indicating that these vibrations may be the sound source. However, because of the low temporal resolution of the imaging system, the frequency and phase of labial vibrations could not be assessed in relation to that of the generated sound. Nevertheless, in contrast to the previous model, these observations show that both labia contribute to aperture control and strongly suggest that they play an important role as principal sound generators.

  4. Mapping the Early Language Environment Using All-Day Recordings and Automated Analysis.

    PubMed

    Gilkerson, Jill; Richards, Jeffrey A; Warren, Steven F; Montgomery, Judith K; Greenwood, Charles R; Kimbrough Oller, D; Hansen, John H L; Paul, Terrance D

    2017-05-17

    This research provided a first-generation standardization of automated language environment estimates, validated these estimates against standard language assessments, and extended on previous research reporting language behavior differences across socioeconomic groups. Typically developing children between 2 to 48 months of age completed monthly, daylong recordings in their natural language environments over a span of approximately 6-38 months. The resulting data set contained 3,213 12-hr recordings automatically analyzed by using the Language Environment Analysis (LENA) System to generate estimates of (a) the number of adult words in the child's environment, (b) the amount of caregiver-child interaction, and (c) the frequency of child vocal output. Child vocalization frequency and turn-taking increased with age, whereas adult word counts were age independent after early infancy. Child vocalization and conversational turn estimates predicted 7%-16% of the variance observed in child language assessment scores. Lower socioeconomic status (SES) children produced fewer vocalizations, engaged in fewer adult-child interactions, and were exposed to fewer daily adult words compared with their higher socioeconomic status peers, but within-group variability was high. The results offer new insight into the landscape of the early language environment, with clinical implications for identification of children at-risk for impoverished language environments.

  5. Vibratory Regime Classification of Infant Phonation

    PubMed Central

    Buder, Eugene H.; Chorna, Lesya B.; Oller, D. Kimbrough; Robinson, Rebecca B.

    2008-01-01

    Infant phonation is highly variable in many respects, including the basic vibratory patterns by which the vocal tissues create acoustic signals. Previous studies have identified the regular occurrence of non-modal phonation types in normal infant phonation. The glottis is like many oscillating systems that, because of non-linear relationships among the elements, may vibrate in ways representing the deterministic patterns classified theoretically within the mathematical framework of non-linear dynamics. The infant’s pre-verbal vocal explorations present such a variety of phonations that it may be possible to find effectively all the classes of vibration predicted by non-linear dynamic theory. The current report defines acoustic criteria for an important subset of such vibratory regimes, and demonstrates that analysts can be trained to reliably use these criteria for a classification that includes all instances of infant phonation in the recorded corpora. The method is thus internally comprehensive in the sense that all phonations are classified, but it is not exhaustive in the sense that all vocal qualities are thereby represented. Using the methods thus developed, this study also demonstrates that the distributions of these phonation types vary significantly across sessions of recording in the first year of life, suggesting developmental changes. The method of regime classification is thus capable of tracking changes that may be indicative of maturation of the mechanism, the learning of categories of phonatory control, and the possibly varying use of vocalizations across social contexts. PMID:17509829

  6. Large-Eddy Simulation of Internal Flow through Human Vocal Folds

    NASA Astrophysics Data System (ADS)

    Lasota, Martin; Šidlof, Petr

    2018-06-01

    The phonatory process occurs when air is expelled from the lungs through the glottis and the pressure drop causes flow-induced oscillations of the vocal folds. The flow fields created in phonation are highly unsteady and the coherent vortex structures are also generated. For accuracy it is essential to compute on humanlike computational domain and appropriate mathematical model. The work deals with numerical simulation of air flow within the space between plicae vocales and plicae vestibulares. In addition to the dynamic width of the rima glottidis, where the sound is generated, there are lateral ventriculus laryngis and sacculus laryngis included in the computational domain as well. The paper presents the results from OpenFOAM which are obtained with a large-eddy simulation using second-order finite volume discretization of incompressible Navier-Stokes equations. Large-eddy simulations with different subgrid scale models are executed on structured mesh. In these cases are used only the subgrid scale models which model turbulence via turbulent viscosity and Boussinesq approximation in subglottal and supraglottal area in larynx.

  7. Understanding the intentional acoustic behavior of humpback whales: a production-based approach.

    PubMed

    Cazau, Dorian; Adam, Olivier; Laitman, Jeffrey T; Reidenberg, Joy S

    2013-09-01

    Following a production-based approach, this paper deals with the acoustic behavior of humpback whales. This approach investigates various physical factors, which are either internal (e.g., physiological mechanisms) or external (e.g., environmental constraints) to the respiratory tractus of the whale, for their implications in sound production. This paper aims to describe a functional scenario of this tractus for the generation of vocal sounds. To do so, a division of this tractus into three different configurations is proposed, based on the air recirculation process which determines air sources and laryngeal valves. Then, assuming a vocal function (in sound generation or modification) for several specific anatomical components, an acoustic characterization of each of these configurations is proposed to link different spectral features, namely, fundamental frequencies and formant structures, to specific vocal production mechanisms. A discussion around the question of whether the whale is able to fully exploit the acoustic potential of its respiratory tractus is eventually provided.

  8. Acoustic signatures of sound source-tract coupling.

    PubMed

    Arneodo, Ezequiel M; Perl, Yonatan Sanz; Mindlin, Gabriel B

    2011-04-01

    Birdsong is a complex behavior, which results from the interaction between a nervous system and a biomechanical peripheral device. While much has been learned about how complex sounds are generated in the vocal organ, little has been learned about the signature on the vocalizations of the nonlinear effects introduced by the acoustic interactions between a sound source and the vocal tract. The variety of morphologies among bird species makes birdsong a most suitable model to study phenomena associated to the production of complex vocalizations. Inspired by the sound production mechanisms of songbirds, in this work we study a mathematical model of a vocal organ, in which a simple sound source interacts with a tract, leading to a delay differential equation. We explore the system numerically, and by taking it to the weakly nonlinear limit, we are able to examine its periodic solutions analytically. By these means we are able to explore the dynamics of oscillatory solutions of a sound source-tract coupled system, which are qualitatively different from those of a sound source-filter model of a vocal organ. Nonlinear features of the solutions are proposed as the underlying mechanisms of observed phenomena in birdsong, such as unilaterally produced "frequency jumps," enhancement of resonances, and the shift of the fundamental frequency observed in heliox experiments. ©2011 American Physical Society

  9. Acoustic signatures of sound source-tract coupling

    PubMed Central

    Arneodo, Ezequiel M.; Perl, Yonatan Sanz; Mindlin, Gabriel B.

    2014-01-01

    Birdsong is a complex behavior, which results from the interaction between a nervous system and a biomechanical peripheral device. While much has been learned about how complex sounds are generated in the vocal organ, little has been learned about the signature on the vocalizations of the nonlinear effects introduced by the acoustic interactions between a sound source and the vocal tract. The variety of morphologies among bird species makes birdsong a most suitable model to study phenomena associated to the production of complex vocalizations. Inspired by the sound production mechanisms of songbirds, in this work we study a mathematical model of a vocal organ, in which a simple sound source interacts with a tract, leading to a delay differential equation. We explore the system numerically, and by taking it to the weakly nonlinear limit, we are able to examine its periodic solutions analytically. By these means we are able to explore the dynamics of oscillatory solutions of a sound source-tract coupled system, which are qualitatively different from those of a sound source-filter model of a vocal organ. Nonlinear features of the solutions are proposed as the underlying mechanisms of observed phenomena in birdsong, such as unilaterally produced “frequency jumps,” enhancement of resonances, and the shift of the fundamental frequency observed in heliox experiments. PMID:21599213

  10. High-Resolution, Non-Invasive Imaging of Upper Vocal Tract Articulators Compatible with Human Brain Recordings

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bouchard, Kristofer E.; Conant, David F.; Anumanchipalli, Gopala K.

    A complete neurobiological understanding of speech motor control requires determination of the relationship between simultaneously recorded neural activity and the kinematics of the lips, jaw, tongue, and larynx. Many speech articulators are internal to the vocal tract, and therefore simultaneously tracking the kinematics of all articulators is nontrivial-especially in the context of human electrophysiology recordings. Here, we describe a noninvasive, multi-modal imaging system to monitor vocal tract kinematics, demonstrate this system in six speakers during production of nine American English vowels, and provide new analysis of such data. Classification and regression analysis revealed considerable variability in the articulator-to-acoustic relationship acrossmore » speakers. Non-negative matrix factorization extracted basis sets capturing vocal tract shapes allowing for higher vowel classification accuracy than traditional methods. Statistical speech synthesis generated speech from vocal tract measurements, and we demonstrate perceptual identification. We demonstrate the capacity to predict lip kinematics from ventral sensorimotor cortical activity. These results demonstrate a multi-modal system to non-invasively monitor articulator kinematics during speech production, describe novel analytic methods for relating kinematic data to speech acoustics, and provide the first decoding of speech kinematics from electrocorticography. These advances will be critical for understanding the cortical basis of speech production and the creation of vocal prosthetics.« less

  11. High-Resolution, Non-Invasive Imaging of Upper Vocal Tract Articulators Compatible with Human Brain Recordings

    PubMed Central

    Anumanchipalli, Gopala K.; Dichter, Benjamin; Chaisanguanthum, Kris S.; Johnson, Keith; Chang, Edward F.

    2016-01-01

    A complete neurobiological understanding of speech motor control requires determination of the relationship between simultaneously recorded neural activity and the kinematics of the lips, jaw, tongue, and larynx. Many speech articulators are internal to the vocal tract, and therefore simultaneously tracking the kinematics of all articulators is nontrivial—especially in the context of human electrophysiology recordings. Here, we describe a noninvasive, multi-modal imaging system to monitor vocal tract kinematics, demonstrate this system in six speakers during production of nine American English vowels, and provide new analysis of such data. Classification and regression analysis revealed considerable variability in the articulator-to-acoustic relationship across speakers. Non-negative matrix factorization extracted basis sets capturing vocal tract shapes allowing for higher vowel classification accuracy than traditional methods. Statistical speech synthesis generated speech from vocal tract measurements, and we demonstrate perceptual identification. We demonstrate the capacity to predict lip kinematics from ventral sensorimotor cortical activity. These results demonstrate a multi-modal system to non-invasively monitor articulator kinematics during speech production, describe novel analytic methods for relating kinematic data to speech acoustics, and provide the first decoding of speech kinematics from electrocorticography. These advances will be critical for understanding the cortical basis of speech production and the creation of vocal prosthetics. PMID:27019106

  12. High-Resolution, Non-Invasive Imaging of Upper Vocal Tract Articulators Compatible with Human Brain Recordings

    DOE PAGES

    Bouchard, Kristofer E.; Conant, David F.; Anumanchipalli, Gopala K.; ...

    2016-03-28

    A complete neurobiological understanding of speech motor control requires determination of the relationship between simultaneously recorded neural activity and the kinematics of the lips, jaw, tongue, and larynx. Many speech articulators are internal to the vocal tract, and therefore simultaneously tracking the kinematics of all articulators is nontrivial-especially in the context of human electrophysiology recordings. Here, we describe a noninvasive, multi-modal imaging system to monitor vocal tract kinematics, demonstrate this system in six speakers during production of nine American English vowels, and provide new analysis of such data. Classification and regression analysis revealed considerable variability in the articulator-to-acoustic relationship acrossmore » speakers. Non-negative matrix factorization extracted basis sets capturing vocal tract shapes allowing for higher vowel classification accuracy than traditional methods. Statistical speech synthesis generated speech from vocal tract measurements, and we demonstrate perceptual identification. We demonstrate the capacity to predict lip kinematics from ventral sensorimotor cortical activity. These results demonstrate a multi-modal system to non-invasively monitor articulator kinematics during speech production, describe novel analytic methods for relating kinematic data to speech acoustics, and provide the first decoding of speech kinematics from electrocorticography. These advances will be critical for understanding the cortical basis of speech production and the creation of vocal prosthetics.« less

  13. Geographic isolation drives divergence of uncorrelated genetic and song variation in the Ruddy-capped Nightingale-Thrush (Catharus frantzii; Aves: Turdidae).

    PubMed

    Ortiz-Ramírez, Marco F; Andersen, Michael J; Zaldívar-Riverón, Alejandro; Ornelas, Juan Francisco; Navarro-Sigüenza, Adolfo G

    2016-01-01

    Montane barriers influence the evolutionary history of lineages by promoting isolation of populations. The effects of these historical processes are evident in patterns of differentiation among extant populations, which are often expressed as genetic and behavioral variation between populations. We investigated the effects of geographic barriers on the evolutionary history of a Mesoamerican bird by studying patterns of genetic and vocal variation in the Ruddy-capped Nightingale-Thrush (Turdidae: Catharus frantzii), a non-migratory oscine bird that inhabits montane forests from central Mexico to Panama. We reconstructed the phylogeographic history and estimated divergence times between populations using Bayesian and maximum likelihood methods. We found strong support for the existence of four mitochondrial lineages of C. frantzii corresponding to isolated mountain ranges: Sierra Madre Oriental; Sierra Madre del Sur; the highlands of Chiapas, Guatemala, and El Salvador; and the Talamanca Cordillera. Vocal features in C. frantzii were highly variable among the four observed clades, but vocal variation and genetic variation were uncorrelated. Song variation in C. frantzii suggests that sexual selection and cultural drift could be important factors driving song differentiation in C. frantzii. Copyright © 2015 Elsevier Inc. All rights reserved.

  14. The impact of rate reduction and increased vocal intensity on coarticulation in dysarthria

    NASA Astrophysics Data System (ADS)

    Tjaden, Kris

    2003-04-01

    The dysarthrias are a group of speech disorders resulting from impairment to nervous system structures important for the motor execution of speech. Although numerous studies have examined how dysarthria impacts articulatory movements or changes in vocal tract shape, few studies of dysarthria consider that articulatory events and their acoustic consequences overlap or are coarticulated in connected speech. The impact of rate, loudness, and clarity on coarticulatory patterns in dysarthria also are poorly understood, although these prosodic manipulations frequently are employed as therapy strategies to improve intelligibility in dysarthria and also are known to affect coarticulatory patterns for at least some neurologically healthy speakers. The current study examined the effects of slowed rate and increased vocal intensity on anticipatory coarticulation for speakers with dysarthria secondary to Multiple Sclerosis (MS), as inferred from the acoustic signal. Healthy speakers were studied for comparison purposes. Three repetitions of twelve target words embedded in the carrier phrase ``It's a -- again'' were produced in habitual, loud, and slow speaking conditions. F2 frequencies and first moment coefficients were used to infer coarticulation. Both group and individual speaker trends will be examined in the data analyses.

  15. Focal expression of mutant huntingtin in the songbird basal ganglia disrupts cortico-basal ganglia networks and vocal sequences

    PubMed Central

    Tanaka, Masashi; Singh Alvarado, Jonnathan; Murugan, Malavika; Mooney, Richard

    2016-01-01

    The basal ganglia (BG) promote complex sequential movements by helping to select elementary motor gestures appropriate to a given behavioral context. Indeed, Huntington’s disease (HD), which causes striatal atrophy in the BG, is characterized by hyperkinesia and chorea. How striatal cell loss alters activity in the BG and downstream motor cortical regions to cause these disorganized movements remains unknown. Here, we show that expressing the genetic mutation that causes HD in a song-related region of the songbird BG destabilizes syllable sequences and increases overall vocal activity, but leave the structure of individual syllables intact. These behavioral changes are paralleled by the selective loss of striatal neurons and reduction of inhibitory synapses on pallidal neurons that serve as the BG output. Chronic recordings in singing birds revealed disrupted temporal patterns of activity in pallidal neurons and downstream cortical neurons. Moreover, reversible inactivation of the cortical neurons rescued the disorganized vocal sequences in transfected birds. These findings shed light on a key role of temporal patterns of cortico-BG activity in the regulation of complex motor sequences and show how a genetic mutation alters cortico-BG networks to cause disorganized movements. PMID:26951661

  16. Nonlinear acoustics in the pant-hoot vocalization of common chimpanzees (Pan troglodytes)

    NASA Astrophysics Data System (ADS)

    Riede, Tobias; Arcadi, Adam Clark; Owren, Michael J.

    2003-04-01

    Pant-hoots produced by chimpanzees are multi-call vocalizations. While predominantly harmonically structured, pant-hoots can exhibit acoustic complexity that has recently been found to result from inherent nonlinearity in the vocal-fold dynamics. This complexity reflects abrupt shifts between qualitatively distinct vibration patterns (known as modes), which include but are not limited to simple, synchronous movements by the two vocal folds. Studies with humans in particular have shown that as the amplitude and vibration rate increase, vocal-fold action becomes increasingly susceptible to higher-order synchronizations, desynchronized movements, and irregular behavior. We examined the occurrence of these sorts of nonlinear phenomena in pant-hoots, contrasting quieter and lower-pitched introduction components with loud and high-pitched climax calls in the same sounds. Spectrographic evidence revealed four classic kinds of nonlinear phenomena, including discrete frequency jumps, subharmonics, biphonation, and deterministic chaos. While these events were virtually never found in the introduction, they occurred in more than half of the climax calls. Biphonation was by far the most common. Individual callers varied in the degree to which their climax calls exhibited nonlinear phenomena, but we are consistent in showing more biphonation than any of the other forms. These outcomes demonstrate that understanding these calls requisitely requires an understanding of such events.

  17. Vocal Tract Images Reveal Neural Representations of Sensorimotor Transformation During Speech Imitation

    PubMed Central

    Carey, Daniel; Miquel, Marc E.; Evans, Bronwen G.; Adank, Patti; McGettigan, Carolyn

    2017-01-01

    Abstract Imitating speech necessitates the transformation from sensory targets to vocal tract motor output, yet little is known about the representational basis of this process in the human brain. Here, we address this question by using real-time MR imaging (rtMRI) of the vocal tract and functional MRI (fMRI) of the brain in a speech imitation paradigm. Participants trained on imitating a native vowel and a similar nonnative vowel that required lip rounding. Later, participants imitated these vowels and an untrained vowel pair during separate fMRI and rtMRI runs. Univariate fMRI analyses revealed that regions including left inferior frontal gyrus were more active during sensorimotor transformation (ST) and production of nonnative vowels, compared with native vowels; further, ST for nonnative vowels activated somatomotor cortex bilaterally, compared with ST of native vowels. Using test representational similarity analysis (RSA) models constructed from participants’ vocal tract images and from stimulus formant distances, we found that RSA searchlight analyses of fMRI data showed either type of model could be represented in somatomotor, temporal, cerebellar, and hippocampal neural activation patterns during ST. We thus provide the first evidence of widespread and robust cortical and subcortical neural representation of vocal tract and/or formant parameters, during prearticulatory ST. PMID:28334401

  18. Gelada vocal sequences follow Menzerath’s linguistic law

    PubMed Central

    Gustison, Morgan L.; Semple, Stuart; Ferrer-i-Cancho, Ramon; Bergman, Thore J.

    2016-01-01

    Identifying universal principles underpinning diverse natural systems is a key goal of the life sciences. A powerful approach in addressing this goal has been to test whether patterns consistent with linguistic laws are found in nonhuman animals. Menzerath’s law is a linguistic law that states that, the larger the construct, the smaller the size of its constituents. Here, to our knowledge, we present the first evidence that Menzerath’s law holds in the vocal communication of a nonhuman species. We show that, in vocal sequences of wild male geladas (Theropithecus gelada), construct size (sequence size in number of calls) is negatively correlated with constituent size (duration of calls). Call duration does not vary significantly with position in the sequence, but call sequence composition does change with sequence size and most call types are abbreviated in larger sequences. We also find that intercall intervals follow the same relationship with sequence size as do calls. Finally, we provide formal mathematical support for the idea that Menzerath’s law reflects compression—the principle of minimizing the expected length of a code. Our findings suggest that a common principle underpins human and gelada vocal communication, highlighting the value of exploring the applicability of linguistic laws in vocal systems outside the realm of language. PMID:27091968

  19. Data-driven automated acoustic analysis of human infant vocalizations using neural network tools.

    PubMed

    Warlaumont, Anne S; Oller, D Kimbrough; Buder, Eugene H; Dale, Rick; Kozma, Robert

    2010-04-01

    Acoustic analysis of infant vocalizations has typically employed traditional acoustic measures drawn from adult speech acoustics, such as f(0), duration, formant frequencies, amplitude, and pitch perturbation. Here an alternative and complementary method is proposed in which data-derived spectrographic features are central. 1-s-long spectrograms of vocalizations produced by six infants recorded longitudinally between ages 3 and 11 months are analyzed using a neural network consisting of a self-organizing map and a single-layer perceptron. The self-organizing map acquires a set of holistic, data-derived spectrographic receptive fields. The single-layer perceptron receives self-organizing map activations as input and is trained to classify utterances into prelinguistic phonatory categories (squeal, vocant, or growl), identify the ages at which they were produced, and identify the individuals who produced them. Classification performance was significantly better than chance for all three classification tasks. Performance is compared to another popular architecture, the fully supervised multilayer perceptron. In addition, the network's weights and patterns of activation are explored from several angles, for example, through traditional acoustic measurements of the network's receptive fields. Results support the use of this and related tools for deriving holistic acoustic features directly from infant vocalization data and for the automatic classification of infant vocalizations.

  20. Group cohesion in foraging meerkats: follow the moving 'vocal hot spot'.

    PubMed

    Gall, Gabriella E C; Manser, Marta B

    2017-04-01

    Group coordination, when 'on the move' or when visibility is low, is a challenge faced by many social living animals. While some animals manage to maintain cohesion solely through visual contact, the mechanism of group cohesion through other modes of communication, a necessity when visual contact is reduced, is not yet understood. Meerkats ( Suricata suricatta ), a small, social carnivore, forage as a cohesive group while moving continuously. While foraging, they frequently emit 'close calls', soft close-range contact calls. Variations in their call rates based on their local environment, coupled with individual movement, produce a dynamic acoustic landscape with a moving 'vocal hotspot' of the highest calling activity. We investigated whether meerkats follow such a vocal hotspot by playing back close calls of multiple individuals to foraging meerkats from the front and back edge of the group simultaneously. These two artificially induced vocal hotspots caused the group to spatially elongate and split into two subgroups. We conclude that meerkats use the emergent dynamic call pattern of the group to adjust their movement direction and maintain cohesion. Our study describes a highly flexible mechanism for the maintenance of group cohesion through vocal communication, for mobile species in habitats with low visibility and where movement decisions need to be adjusted continuously to changing environmental conditions.

  1. Vocal Tract Images Reveal Neural Representations of Sensorimotor Transformation During Speech Imitation.

    PubMed

    Carey, Daniel; Miquel, Marc E; Evans, Bronwen G; Adank, Patti; McGettigan, Carolyn

    2017-05-01

    Imitating speech necessitates the transformation from sensory targets to vocal tract motor output, yet little is known about the representational basis of this process in the human brain. Here, we address this question by using real-time MR imaging (rtMRI) of the vocal tract and functional MRI (fMRI) of the brain in a speech imitation paradigm. Participants trained on imitating a native vowel and a similar nonnative vowel that required lip rounding. Later, participants imitated these vowels and an untrained vowel pair during separate fMRI and rtMRI runs. Univariate fMRI analyses revealed that regions including left inferior frontal gyrus were more active during sensorimotor transformation (ST) and production of nonnative vowels, compared with native vowels; further, ST for nonnative vowels activated somatomotor cortex bilaterally, compared with ST of native vowels. Using test representational similarity analysis (RSA) models constructed from participants' vocal tract images and from stimulus formant distances, we found that RSA searchlight analyses of fMRI data showed either type of model could be represented in somatomotor, temporal, cerebellar, and hippocampal neural activation patterns during ST. We thus provide the first evidence of widespread and robust cortical and subcortical neural representation of vocal tract and/or formant parameters, during prearticulatory ST. © The Author 2017. Published by Oxford University Press.

  2. Vocal fold hypomobility secondary to elective endotracheal intubation: a general surgeon's perspective.

    PubMed

    Sariego, Jack

    2010-01-01

    This study was performed retrospectively to evaluate the incidence of documented vocal fold injury as a result of elective endotracheal intubation during general surgical procedures. Medical record review was performed at a single institution and all surgical cases reviewed which required endotracheal intubation in the nonemergent setting between April 1, 2003 and August, 31, 2007. Cases with unexpected and documented vocal fold immobility postoperatively formed the study cohort, and data were gathered regarding diagnosis and procedures performed. Of 23,010 general surgery cases performed during the study period, only seven documented cases of vocal fold paralysis were discovered (0.03%). There were five women and two men in the group; all were adults. Only one patient had a primary diagnosis related to the head and neck. Comorbidities were recorded as well, but there were no statistically significant patterns discerned. Furthermore, during the study period, a total of 31 patients overall (both surgical and nonsurgical) were admitted who carried a primary diagnosis of vocal fold paralysis. Therefore, the study cohort therefore constituted 22.6% of this total. Finally, cohort patients spent a total of 150 days in hospital during the study period; this length of stay (an average of 16.7 hospital days per patient) was significantly longer than the average of 5.1 days, presumably at least in part related to the vocal paralysis. Copyright 2010 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  3. Convergence of pattern generator outputs on a common mechanism of diaphragm motor unit recruitment.

    PubMed

    Mantilla, Carlos B; Seven, Yasin B; Sieck, Gary C

    2014-01-01

    Motor units are the final element of neuromotor control. In manner analogous to the organization of neuromotor control in other skeletal muscles, diaphragm motor units comprise phrenic motoneurons located in the cervical spinal cord that innervate the diaphragm muscle, the main inspiratory muscle in mammals. Diaphragm motor units play a primary role in sustaining ventilation but are also active in other nonventilatory behaviors, including coughing, sneezing, vomiting, defecation, and parturition. Diaphragm muscle fibers comprise all fiber types. Thus, diaphragm motor units display substantial differences in contractile and fatigue properties, but importantly, properties of the motoneuron and muscle fibers within a motor unit are matched. As in other skeletal muscles, diaphragm motor units are recruited in order such that motor units that display greater fatigue resistance are recruited earlier and more often than more fatigable motor units. The properties of the motor unit population are critical determinants of the function of a skeletal muscle across the range of possible motor tasks. Accordingly, fatigue-resistant motor units are sufficient to generate the forces necessary for ventilatory behaviors, whereas more fatigable units are only activated during expulsive behaviors important for airway clearance. Neuromotor control of diaphragm motor units may reflect selective inputs from distinct pattern generators distributed according to the motor unit properties necessary to accomplish these different motor tasks. In contrast, widely distributed inputs to phrenic motoneurons from various pattern generators (e.g., for breathing, coughing, or vocalization) would dictate recruitment order based on intrinsic electrophysiological properties. © 2014 Elsevier B.V. All rights reserved.

  4. Mapping the distribution of language related genes FoxP1, FoxP2, and CntnaP2 in the brains of vocal learning bat species.

    PubMed

    Rodenas-Cuadrado, Pedro M; Mengede, Janine; Baas, Laura; Devanna, Paolo; Schmid, Tobias A; Yartsev, Michael; Firzlaff, Uwe; Vernes, Sonja C

    2018-06-01

    Genes including FOXP2, FOXP1, and CNTNAP2, have been implicated in human speech and language phenotypes, pointing to a role in the development of normal language-related circuitry in the brain. Although speech and language are unique to humans a comparative approach is possible by addressing language-relevant traits in animal systems. One such trait, vocal learning, represents an essential component of human spoken language, and is shared by cetaceans, pinnipeds, elephants, some birds and bats. Given their vocal learning abilities, gregarious nature, and reliance on vocalizations for social communication and navigation, bats represent an intriguing mammalian system in which to explore language-relevant genes. We used immunohistochemistry to detail the distribution of FoxP2, FoxP1, and Cntnap2 proteins, accompanied by detailed cytoarchitectural histology in the brains of two vocal learning bat species; Phyllostomus discolor and Rousettus aegyptiacus. We show widespread expression of these genes, similar to what has been previously observed in other species, including humans. A striking difference was observed in the adult P. discolor bat, which showed low levels of FoxP2 expression in the cortex that contrasted with patterns found in rodents and nonhuman primates. We created an online, open-access database within which all data can be browsed, searched, and high resolution images viewed to single cell resolution. The data presented herein reveal regions of interest in the bat brain and provide new opportunities to address the role of these language-related genes in complex vocal-motor and vocal learning behaviors in a mammalian model system. © 2018 The Authors The Journal of Comparative Neurology Published by Wiley Periodicals, Inc.

  5. An Adapting Auditory-motor Feedback Loop Can Contribute to Generating Vocal Repetition

    PubMed Central

    Brainard, Michael S.; Jin, Dezhe Z.

    2015-01-01

    Consecutive repetition of actions is common in behavioral sequences. Although integration of sensory feedback with internal motor programs is important for sequence generation, if and how feedback contributes to repetitive actions is poorly understood. Here we study how auditory feedback contributes to generating repetitive syllable sequences in songbirds. We propose that auditory signals provide positive feedback to ongoing motor commands, but this influence decays as feedback weakens from response adaptation during syllable repetitions. Computational models show that this mechanism explains repeat distributions observed in Bengalese finch song. We experimentally confirmed two predictions of this mechanism in Bengalese finches: removal of auditory feedback by deafening reduces syllable repetitions; and neural responses to auditory playback of repeated syllable sequences gradually adapt in sensory-motor nucleus HVC. Together, our results implicate a positive auditory-feedback loop with adaptation in generating repetitive vocalizations, and suggest sensory adaptation is important for feedback control of motor sequences. PMID:26448054

  6. A Detailed Motion Analysis of the Angular Velocity Between the Vocal Folds During Throat Clearing Using High-speed Digital Imaging.

    PubMed

    Iwahashi, Toshihiko; Ogawa, Makoto; Hosokawa, Kiyohito; Kato, Chieri; Inohara, Hidenori

    2016-11-01

    To assess the angular velocity between the vocal folds just before the compression phase of throat clearing (TC) using high-speed digital imaging (HSDI) of the larynx. Twenty normal healthy adults (13 males and seven females) were enrolled in the study. Each participant underwent transnasal laryngo-fiberscopy, and was asked to perform weak/strong TC followed by a comfortable, sustained vowel phonation while recording an HSDI movie (4000 frames/s) of the larynx. Using a motion analysis, the changes in the vocal fold angle and angular velocity during vocal fold adduction were assessed. Subsequently, we calculated the average angular velocities in the ranges of 100-80%, 80-20%, and 20-0% from all of the angular changes. The motion analysis demonstrated that the changes in the angular velocity resulted in polynomial-like and sigmoid curves during TC and vowel phonation, respectively. The angular velocities during weak TC were significantly higher in the 20-0%, 80-20%, and 100-80% regions (in order); the 80-20% angular velocity in vocal fold adduction during phonation was highest. The 20-0% angular velocity during strong TC was more than twofold higher than 20-0% angular velocity during phonation. The present results confirmed that the closing motions of the vocal folds accelerate throughout the precompression closing phase of a TC episode, and decelerate just before the impact between the vocal folds at the onset of phonation, suggesting that the vocal fold velocity generated by TC is sufficient to damage the laryngeal tissues. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  7. Effect of resection depth of early glottic cancer on vocal outcome: An optimized finite element simulation

    PubMed Central

    Mau, Ted; Palaparthi, Anil; Riede, Tobias; Titze, Ingo R.

    2015-01-01

    Objectives/Hypothesis To test the hypothesis that subligamental cordectomy produces superior acoustic outcome than subepithelial cordectomy for early (T1-2) glottic cancer that requires complete removal of the superficial lamina propria but does not involve the vocal ligament. Study Design Computer simulation Methods A computational tool for vocal fold surgical planning and simulation (the National Center for Voice and Speech Phonosurgery Optimizer-Simulator) was used to evaluate the acoustic output of alternative vocal fold morphologies. Four morphologies were simulated: normal, subepithelial cordectomy, subligamental cordectomy, and transligamental cordectomy (partial ligament resection). The primary outcome measure was the range of fundamental frequency (F0) and sound pressure level (SPL). A more restricted F0-SPL range was considered less favorable because of reduced acoustic possibilities given the same range of driving subglottic pressure and identical vocal fold posturing. Results Subligamental cordectomy generated solutions covering an F0-SPL range 82% of normal for a rectangular vocal fold. In contrast, transligamental and subepithelial cordectomies produced significantly smaller F0-SPL ranges, 57% and 19% of normal, respectively. Conclusion This study illustrates the use of the Phonosurgery Optimizer-Simulator to test a specific hypothesis regarding the merits of two surgical alternatives. These simulation results provide theoretical support for vocal ligament excision with maximum muscle preservation when superficial lamina propria resection is necessary but the vocal ligament can be spared on oncological grounds. The resection of more tissue may paradoxically allow the eventual recovery of a better speaking voice, assuming glottal width is restored. Application of this conclusion to surgical practice will require confirmatory clinical data. PMID:26010240

  8. Intramodal and Intermodal Functioning of Normal and LD Children

    ERIC Educational Resources Information Center

    Heath, Earl J.; Early, George H.

    1973-01-01

    Assessed were the abilities of 50 normal 5-to 9-year-old children and 30 learning disabled 7-to 9-year-old children to recognize temporal patterns presented visually and auditorially (intramodal abilities) and to vocally produce the patterns whether presentation was visual or auditory (intramodal and cross-modal abilities). (MC)

  9. Aerosol emission during human speech

    NASA Astrophysics Data System (ADS)

    Asadi, Sima; Wexler, Anthony S.; Cappa, Christopher D.; Bouvier, Nicole M.; Barreda-Castanon, Santiago; Ristenpart, William D.

    2017-11-01

    We show that the rate of aerosol particle emission during healthy human speech is strongly correlated with the loudness (amplitude) of vocalization. Emission rates range from approximately 1 to 50 particles per second for quiet to loud amplitudes, regardless of language spoken (English, Spanish, Mandarin, or Arabic). Intriguingly, a small fraction of individuals behave as ``super emitters,'' consistently emitting an order of magnitude more aerosol particles than their peers. We interpret the results in terms of the eggressive flowrate during vocalization, which is known to vary significantly for different types of vocalization and for different individuals. The results suggest that individual speech patterns could affect the probability of airborne disease transmission. The results also provide a possible explanation for the existence of ``super spreaders'' who transmit pathogens much more readily than average and who play a key role in the spread of epidemics.

  10. Sex differences in razorbill (Family: Alcidae) parent-offspring vocal recognition

    NASA Astrophysics Data System (ADS)

    Insley, Stephen J.; Paredes Vela, Rosana; Jones, Ian L.

    2002-05-01

    In this study we examines how a pattern of parental care may result in a sex bias in vocal recognition. In Razorbills (Alca torda), both sexes provide parental care to their chicks while at the nest, after which the male is the sole caregiver for an additional period at sea. Selection pressure acting on recognition behavior is expected to be strongest during the time when males and chicks are together at sea, and as a result, parent-offspring recognition was predicted to be better developed in the male parent, that is, show a paternal bias. In order to test this hypothesis, vocal playback experiments were conducted on breeding Razorbills at the Gannet Islands, Labrador, 2001. The data provide clear evidence of mutual vocal recognition between the male parent and chick but not between the female parent and chick, supporting the hypothesis that parent-offspring recognition is male biased in this species. In addition to acoustic recognition, such a bias could have important social implications for a variety of behavioral and basic life history traits such as cooperation and sex-biased dispersal.

  11. Imaging auditory representations of song and syllables in populations of sensorimotor neurons essential to vocal communication.

    PubMed

    Peh, Wendy Y X; Roberts, Todd F; Mooney, Richard

    2015-04-08

    Vocal communication depends on the coordinated activity of sensorimotor neurons important to vocal perception and production. How vocalizations are represented by spatiotemporal activity patterns in these neuronal populations remains poorly understood. Here we combined intracellular recordings and two-photon calcium imaging in anesthetized adult zebra finches (Taeniopygia guttata) to examine how learned birdsong and its component syllables are represented in identified projection neurons (PNs) within HVC, a sensorimotor region important for song perception and production. These experiments show that neighboring HVC PNs can respond at markedly different times to song playback and that different syllables activate spatially intermingled PNs within a local (~100 μm) region of HVC. Moreover, noise correlations were stronger between PNs that responded most strongly to the same syllable and were spatially graded within and between classes of PNs. These findings support a model in which syllabic and temporal features of song are represented by spatially intermingled PNs functionally organized into cell- and syllable-type networks within local spatial scales in HVC. Copyright © 2015 the authors 0270-6474/15/355589-17$15.00/0.

  12. Frequency Response of Synthetic Vocal Fold Models with Linear and Nonlinear Material Properties

    PubMed Central

    Shaw, Stephanie M.; Thomson, Scott L.; Dromey, Christopher; Smith, Simeon

    2014-01-01

    Purpose The purpose of this study was to create synthetic vocal fold models with nonlinear stress-strain properties and to investigate the effect of linear versus nonlinear material properties on fundamental frequency during anterior-posterior stretching. Method Three materially linear and three materially nonlinear models were created and stretched up to 10 mm in 1 mm increments. Phonation onset pressure (Pon) and fundamental frequency (F0) at Pon were recorded for each length. Measurements were repeated as the models were relaxed in 1 mm increments back to their resting lengths, and tensile tests were conducted to determine the stress-strain responses of linear versus nonlinear models. Results Nonlinear models demonstrated a more substantial frequency response than did linear models and a more predictable pattern of F0 increase with respect to increasing length (although range was inconsistent across models). Pon generally increased with increasing vocal fold length for nonlinear models, whereas for linear models, Pon decreased with increasing length. Conclusions Nonlinear synthetic models appear to more accurately represent the human vocal folds than linear models, especially with respect to F0 response. PMID:22271874

  13. Frequency response of synthetic vocal fold models with linear and nonlinear material properties.

    PubMed

    Shaw, Stephanie M; Thomson, Scott L; Dromey, Christopher; Smith, Simeon

    2012-10-01

    The purpose of this study was to create synthetic vocal fold models with nonlinear stress-strain properties and to investigate the effect of linear versus nonlinear material properties on fundamental frequency (F0) during anterior-posterior stretching. Three materially linear and 3 materially nonlinear models were created and stretched up to 10 mm in 1-mm increments. Phonation onset pressure (Pon) and F0 at Pon were recorded for each length. Measurements were repeated as the models were relaxed in 1-mm increments back to their resting lengths, and tensile tests were conducted to determine the stress-strain responses of linear versus nonlinear models. Nonlinear models demonstrated a more substantial frequency response than did linear models and a more predictable pattern of F0 increase with respect to increasing length (although range was inconsistent across models). Pon generally increased with increasing vocal fold length for nonlinear models, whereas for linear models, Pon decreased with increasing length. Nonlinear synthetic models appear to more accurately represent the human vocal folds than do linear models, especially with respect to F0 response.

  14. Vocal parameters that indicate threat level correlate with FOS immunolabeling in social and vocal control brain regions.

    PubMed

    Ellis, Jesse M S; Riters, Lauren V

    2012-01-01

    Transmitting information via communicative signals is integral to interacting with conspecifics, and some species achieve this task by varying vocalizations to reflect context. Although signal variation is critical to social interactions, the underlying neural control has not been studied. In response to a predator, black-capped chickadees (Poecile atricapilla) produce mobbing calls (chick-a-dee calls) with various parameters, some of which convey information about the threat stimulus. We predicted that vocal parameters indicative of threat would be associated with distinct patterns of neuronal activity within brain areas involved in social behavior and those involved in the sensorimotor control of vocal production. To test this prediction, we measured the syntax and structural aspects of chick-a-dee call production in response to a hawk model and assessed the protein product of the immediate early gene FOS in brain regions implicated in context-specific vocal and social behavior. These regions include the medial preoptic area (POM) and lateral septum (LS), as well as regions involved in vocal motor control, including the dorsomedial nucleus of the intercollicular complex and the HVC. We found correlations linking call rate (previously demonstrated to reflect threat) to labeling in the POM and LS. Labeling in the HVC correlated with the number of D notes per call, which may also signal threat level. Labeling in the call control region dorsomedial nucleus was associated with the structure of D notes and the overall number of notes, but not call rate or type of notes produced. These results suggest that the POM and LS may influence attributes of vocalizations produced in response to predators and that the brain region implicated in song control, the HVC, also influences call production. Because variation in chick-a-dee call rate indicates predator threat, we speculate that these areas could integrate with motor control regions to imbue mobbing signals with additional information about threat level. Copyright © 2011 S. Karger AG, Basel.

  15. An Immersed-Boundary Method for Fluid-Structure Interaction in the Human Larynx

    NASA Astrophysics Data System (ADS)

    Luo, Haoxiang; Zheng, Xudong; Mittal, Rajat; Bielamowicz, Steven

    2006-11-01

    We describe a novel and accurate computational methodology for modeling the airflow and vocal fold dynamics in human larynx. The model is useful in helping us gain deeper insight into the complicated bio-physics of phonation, and may have potential clinical application in design and placement of synthetic implant in vocal fold surgery. The numerical solution of the airflow employs a previously developed immersed-boundary solver. However, in order to incorporate the vocal fold into the model, we have developed a new immersed-boundary method that can simulate the dynamics of the multi-layered, viscoelastic solids. In this method, a finite-difference scheme is used to approximate the derivatives and ghost cells are defined near the boundary. To impose the traction boundary condition, a third-order polynomial is obtained using the weighted least squares fitting to approximate the function locally. Like its analogue for the flow solver, this immersed-boundary method for the solids has the advantage of simple grid generation, and may be easily implemented on parallel computers. In the talk, we will present the simulation results on both the specified vocal fold motion and the flow-induced vocal fold vibration. Supported by NIDCD Grant R01 DC007125-01A1.

  16. Insights into the role of elastin in vocal fold health and disease

    PubMed Central

    Moore, Jaime

    2011-01-01

    Elastic fibers are large, complex and surprisingly poorly understood extracellular matrix (ECM) macromolecules. The elastin fiber, generated from a single human gene - elastin (ELN), is a self assembling integral protein that endows critical mechanic proprieties to elastic tissues and organs such as the skin, lungs, and arteries. The biology of elastic fibers is complex because they have multiple components, a tightly regulated developmental deposition, a multi-step hierarchical assembly and unique biomechanical functions. Elastin is present in vocal folds, where it plays a pivotal role in the quality of phonation. This review article provides an overview of the genesis of elastin and its wide- ranging structure and function. Specific distribution within the vocal fold lamina propria across the lifespan in normal and pathological states and its contribution to vocal fold biomechanics will be examined. Elastin and elastin-derived molecules are increasingly investigated for their application in tissue engineering. The properties of various elastin– based materials will be discussed and their current and future applications evaluated. A new level of understanding of the biomechanical properties of vocal fold elastin composites and their molecular basis should lead to new strategies for elastic fiber repair and regeneration in aging and disease. PMID:21708449

  17. Amplification and spectral shifts of vocalizations inside burrows of the frog Eupsophus calcaratus (Leptodactylidae)

    NASA Astrophysics Data System (ADS)

    Penna, Mario

    2004-08-01

    A variety of animals that communicate by sound emit signals from sites favoring their propagation, thereby increasing the range over which these sounds convey information. A different significance of calling sites has been reported for burrowing frogs Eupsophus emiliopugini from southern Chile: the cavities from which these frogs vocalize amplify conspecific vocalizations generated externally, thus providing a means to enhance the reception of neighbor's vocalizations in chorusing aggregations. In the current study the amplification of vocalizations of a related species, E. calcaratus, is investigated, to explore the extent of sound enhancement reported previously. Advertisement calls broadcast through a loudspeaker placed in the vicinity of a burrow, monitored with small microphones, are amplified by up to 18 dB inside cavities relative to outside. The fundamental resonant frequency of burrows, measured with broadcast noise and pure tones, ranges from 842 to 1836 Hz and is significantly correlated with the burrow's length. Burrows change the spectral envelope of incoming calls by increasing the amplitude of lower relative to higher harmonics. The call amplification effect inside burrows of E. calcaratus parallels the effect reported previously for E. emiliopugini, and indicates that the acoustic properties of calling sites may affect signal reception by burrowing animals.

  18. Amplification and spectral shifts of vocalizations inside burrows of the frog Eupsophus calcaratus (Leptodactylidae).

    PubMed

    Penna, Mario

    2004-08-01

    A variety of animals that communicate by sound emit signals from sites favoring their propagation, thereby increasing the range over which these sounds convey information. A different significance of calling sites has been reported for burrowing frogs Eupsophus emiliopugini from southern Chile: the cavities from which these frogs vocalize amplify conspecific vocalizations generated externally, thus providing a means to enhance the reception of neighbor's vocalizations in chorusing aggregations. In the current study the amplification of vocalizations of a related species, E. calcaratus, is investigated, to explore the extent of sound enhancement reported previously. Advertisement calls broadcast through a loudspeaker placed in the vicinity of a burrow, monitored with small microphones, are amplified by up to 18 dB inside cavities relative to outside. The fundamental resonant frequency of burrows, measured with broadcast noise and pure tones, ranges from 842 to 1836 Hz and is significantly correlated with the burrow's length. Burrows change the spectral envelope of incoming calls by increasing the amplitude of lower relative to higher harmonics. The call amplification effect inside burrows of E. calcaratus parallels the effect reported previously for E. emiliopugini, and indicates that the acoustic properties of calling sites may affect signal reception by burrowing animals.

  19. An Automated Procedure for Evaluating Song Imitation

    PubMed Central

    Mandelblat-Cerf, Yael; Fee, Michale S.

    2014-01-01

    Songbirds have emerged as an excellent model system to understand the neural basis of vocal and motor learning. Like humans, songbirds learn to imitate the vocalizations of their parents or other conspecific “tutors.” Young songbirds learn by comparing their own vocalizations to the memory of their tutor song, slowly improving until over the course of several weeks they can achieve an excellent imitation of the tutor. Because of the slow progression of vocal learning, and the large amounts of singing generated, automated algorithms for quantifying vocal imitation have become increasingly important for studying the mechanisms underlying this process. However, methodologies for quantifying song imitation are complicated by the highly variable songs of either juvenile birds or those that learn poorly because of experimental manipulations. Here we present a method for the evaluation of song imitation that incorporates two innovations: First, an automated procedure for selecting pupil song segments, and, second, a new algorithm, implemented in Matlab, for computing both song acoustic and sequence similarity. We tested our procedure using zebra finch song and determined a set of acoustic features for which the algorithm optimally differentiates between similar and non-similar songs. PMID:24809510

  20. Direct numerical simulation of human phonation

    NASA Astrophysics Data System (ADS)

    Bodony, Daniel; Saurabh, Shakti

    2017-11-01

    The generation and propagation of the human voice in three-dimensions is studied using direct numerical simulation. A full body domain is employed for the purpose of directly computing the sound in the region past the speaker's mouth. The air in the vocal tract is modeled as a compressible and viscous fluid interacting with the elastic vocal folds. The vocal fold tissue material properties are multi-layered, with varying stiffness, and a linear elastic transversely isotropic model is utilized and implemented in a quadratic finite element code. The fluid-solid domains are coupled through a boundary-fitted interface and utilize a Poisson equation-based mesh deformation method. A kinematic constraint based on a specified minimum gap between the vocal folds is applied to prevent collision during glottal closure. Both near VF flow dynamics and far-field acoustics have been studied. A comparison is drawn to current two-dimensional simulations as well as to data from the literature. Near field vocal fold dynamics and glottal flow results are studied and in good agreement with previous three-dimensional phonation studies. Far-field acoustic characteristics, when compared to their two-dimensional counterpart, are shown to be sensitive to the dimensionality. Supported by the National Science Foundation (CAREER Award Number 1150439).

  1. A bioreactor for the dynamic mechanical stimulation of vocal-fold fibroblasts based on vibro-acoustography

    NASA Astrophysics Data System (ADS)

    Chan, Roger W.; Rodriguez, Maritza

    2005-09-01

    During voice production, the vocal folds undergo airflow-induced self-sustained oscillation at a fundamental frequency of around 100-1000 Hz, with an amplitude of around 1-3 mm. The vocal-fold extracellular matrix (ECM), with appropriate tissue viscoelastic properties, is optimally tuned for such vibration. Vocal-fold fibroblasts regulate the gene expressions for key ECM proteins (e.g., collagen, fibronectin, fibromodulin, and hyaluronic acid), and these expressions are affected by the stress fields experi- enced by the fibroblasts. This study attempts to develop a bioreactor for cultivating cells under a micromechanical environment similar to that in vivo, based on the principle of vibro-acoustography. Vocal-fold fibroblasts from primary culture were grown in 3D, biodegradable scaffolds, and were excited dynamically by the radiation force generated by amplitude modulation of two confocal ultrasound beams of slightly different frequencies. Low-frequency acoustic radiation force was applied to the scaffold surface, and its vibratory response was imaged by videostroboscopy. A phantom tissue (standard viscoelastic material) with known elastic modulus was also excited and its vibratory frequency and amplitude were measured by videostroboscopy. Results showed that the bioreactor was capable of delivering mechanical stimuli to the tissue constructs in a physiological frequency range (100-1000 Hz), supporting its potential for vocal-fold tissue engineering applications. [Work supported by NIH Grant R01 DC006101.

  2. [Phoniatric surgery and conservative treatment of vocal cord hematoma].

    PubMed

    Milutinović, Z

    1997-01-01

    Functional-traumatic lesions of the vocal fold include mucous stranding, "nodular" lesions, polyps, cysts, contact hyperplasia and haematoma of the vocal fold. An acute voice overuse may result in bleeding (haematoma) within the vocal fold. This may be in the form of petechial bleeding, or a genuine haematoma develops within the tissues of the vocal fold. Haematoma may also arise as a consequence of prolonged cough, forceful vomiting, lifting of a heavy weight, various effortful activities, etc. Haematoma is usually located close to the vocal fold free edge and therefore disturbs the glottic closure during phonation. The treatment is adapted to the size and localization of haematoma, as well as to the time elapsed from onset of the lesion. Phonosurgery can be used in therapy, as well as corticosteroid treatment. A series of 102 vocal fold haematomas has been treated by phonosurgery (39) and conservative therapy (63). Phonosurgical interventions were performed by an indirect approach, by use of microstroboscopy (28 patients) and videostroboscopy (11 causes). Conservative treatment consisted of corticosteroid therapy. During a 10-year period 1550 phonosurgical operations were performed for benign lesions of the vocal fold, including 39 haematomas (2.5%). It was established that recovery of vibration pattern was significantly faster in the surgery group in comparison to the group of patients treated conservatively. All surgical patients were operated within the first several days after the onset of symptoms. In case of a vocal fold haematoma, it is very important to establish the diagnosis as soon as possible in order to start with the therapy early enough. Within the first several days after the onset (the best within 24-48 hours) a phonosurgical treatment is indicated, preferably by the use of indirect videostroboscopy. If the treatment is started later we use corticoids. However, the results are inferior as compared to surgery. We did not perform direct microlaryngoscopy in these cases, for a lack of function monitoring and possible local trauma to the tissues. In the majority of cases the voice therapy is required as well.

  3. NMDA or non-NMDA receptor antagonism within the amygdaloid central nucleus suppresses the affective dimension of pain in rats: evidence for hemispheric synergy.

    PubMed

    Spuz, Catherine A; Borszcz, George S

    2012-04-01

    The amygdala contributes to generation of affective behaviors to threats. The prototypical threat to an individual is exposure to a noxious stimulus and the amygdaloid central nucleus (CeA) receives nociceptive input that is mediated by glutamatergic neurotransmission. The present study evaluated the contribution of glutamate receptors in CeA to generation of the affective response to acute pain in rats. Vocalizations that occur following a brief noxious tail shock (vocalization afterdischarges) are a validated rodent model of pain affect, and were preferentially suppressed by bilateral injection into CeA of the NMDA receptor antagonist D-2-amino-5-phosphonovalerate (AP5, 1 μg, 2 μg, or 4 μg) or the non-NMDA receptor antagonist 6-Cyano-7-nitroquinoxaline-2,3-dione disodium (CNQX, .25 μg, .5 μg, 1 μg, or 2 μg). Vocalizations that occur during tail shock were suppressed to a lesser degree, whereas spinal motor reflexes (tail flick and hind limb movements) were unaffected by injection of AP5 or CNQX into CeA. Unilateral administration of AP5 or CNQX into CeA of either hemisphere also selectively elevated vocalization thresholds. Bilateral administration of AP5 or CNQX produced greater increases in vocalization thresholds than the same doses of antagonists administered unilaterality into either hemisphere indicating synergistic hemispheric interactions. The amygdala contributes to production of emotional responses to environmental threats. Blocking glutamate neurotransmission within the central nucleus of the amygdala suppressed rats' emotional response to acute painful stimulation. Understanding the neurobiology underlying emotional responses to pain will provide insights into new treatments for pain and its associated affective disorders. Copyright © 2012 American Pain Society. Published by Elsevier Inc. All rights reserved.

  4. Patterns and causes of geographic variation in bat echolocation pulses.

    PubMed

    Jiang, Tinglei; Wu, Hui; Feng, Jiang

    2015-05-01

    Evolutionary biologists have a long-standing interest in how acoustic signals in animals vary geographically, because divergent ecology and sensory perception play an important role in speciation. Geographic comparisons are valuable in determining the factors that influence divergence of acoustic signals. Bats are social mammals and they depend mainly on echolocation pulses to locate prey, to navigate and to communicate. Mounting evidence shows that geographic variation of bat echolocation pulses is common, with a mean 5-10 kHz differences in peak frequency, and a high level of individual variation may be nested in this geographical variation. However, understanding the geographic variation of echolocation pulses in bats is very difficult, because of differences in sample and statistical analysis techniques as well as the variety of factors shaping the vocal geographic evolution. Geographic differences in echolocation pulses of bats generally lack latitudinal, longitudinal and elevational patterns, and little is known about vocal dialects. Evidence is accumulating to support the fact that geographic variation in echolocation pulses of bats may be caused by genetic drift, cultural drift, ecological selection, sexual selection and social selection. Future studies could relate geographic differences in echolocation pulses to social adaptation, vocal learning strategies and patterns of dispersal. In addition, new statistical techniques and acoustic playback experiments may help to illustrate the causes and consequences of the geographic evolution of echolocation pulse in bats. © 2015 International Society of Zoological Sciences, Institute of Zoology/Chinese Academy of Sciences and Wiley Publishing Asia Pty Ltd.

  5. Left-hemisphere activation is associated with enhanced vocal pitch error detection in musicians with absolute pitch

    PubMed Central

    Behroozmand, Roozbeh; Ibrahim, Nadine; Korzyukov, Oleg; Robin, Donald A.; Larson, Charles R.

    2014-01-01

    The ability to process auditory feedback for vocal pitch control is crucial during speaking and singing. Previous studies have suggested that musicians with absolute pitch (AP) develop specialized left-hemisphere mechanisms for pitch processing. The present study adopted an auditory feedback pitch perturbation paradigm combined with ERP recordings to test the hypothesis whether the neural mechanisms of the left-hemisphere enhance vocal pitch error detection and control in AP musicians compared with relative pitch (RP) musicians and non-musicians (NM). Results showed a stronger N1 response to pitch-shifted voice feedback in the right-hemisphere for both AP and RP musicians compared with the NM group. However, the left-hemisphere P2 component activation was greater in AP and RP musicians compared with NMs and also for the AP compared with RP musicians. The NM group was slower in generating compensatory vocal reactions to feedback pitch perturbation compared with musicians, and they failed to re-adjust their vocal pitch after the feedback perturbation was removed. These findings suggest that in the earlier stages of cortical neural processing, the right hemisphere is more active in musicians for detecting pitch changes in voice feedback. In the later stages, the left-hemisphere is more active during the processing of auditory feedback for vocal motor control and seems to involve specialized mechanisms that facilitate pitch processing in the AP compared with RP musicians. These findings indicate that the left hemisphere mechanisms of AP ability are associated with improved auditory feedback pitch processing during vocal pitch control in tasks such as speaking or singing. PMID:24355545

  6. Error-dependent modulation of speech-induced auditory suppression for pitch-shifted voice feedback.

    PubMed

    Behroozmand, Roozbeh; Larson, Charles R

    2011-06-06

    The motor-driven predictions about expected sensory feedback (efference copies) have been proposed to play an important role in recognition of sensory consequences of self-produced motor actions. In the auditory system, this effect was suggested to result in suppression of sensory neural responses to self-produced voices that are predicted by the efference copies during vocal production in comparison with passive listening to the playback of the identical self-vocalizations. In the present study, event-related potentials (ERPs) were recorded in response to upward pitch shift stimuli (PSS) with five different magnitudes (0, +50, +100, +200 and +400 cents) at voice onset during active vocal production and passive listening to the playback. Results indicated that the suppression of the N1 component during vocal production was largest for unaltered voice feedback (PSS: 0 cents), became smaller as the magnitude of PSS increased to 200 cents, and was almost completely eliminated in response to 400 cents stimuli. Findings of the present study suggest that the brain utilizes the motor predictions (efference copies) to determine the source of incoming stimuli and maximally suppresses the auditory responses to unaltered feedback of self-vocalizations. The reduction of suppression for 50, 100 and 200 cents and its elimination for 400 cents pitch-shifted voice auditory feedback support the idea that motor-driven suppression of voice feedback leads to distinctly different sensory neural processing of self vs. non-self vocalizations. This characteristic may enable the audio-vocal system to more effectively detect and correct for unexpected errors in the feedback of self-produced voice pitch compared with externally-generated sounds.

  7. Prenatal exposure to a common organophosphate insecticide delays motor development in a mouse model of idiopathic autism.

    PubMed

    De Felice, Alessia; Scattoni, Maria Luisa; Ricceri, Laura; Calamandrei, Gemma

    2015-01-01

    Autism spectrum disorders are characterized by impaired social and communicative skills and repetitive behaviors. Emerging evidence supported the hypothesis that these neurodevelopmental disorders may result from a combination of genetic susceptibility and exposure to environmental toxins in early developmental phases. This study assessed the effects of prenatal exposure to chlorpyrifos (CPF), a widely diffused organophosphate insecticide endowed with developmental neurotoxicity at sub-toxic doses, in the BTBR T+tf/J mouse strain, a validated model of idiopathic autism that displays several behavioral traits relevant to the autism spectrum. To this aim, pregnant BTBR mice were administered from gestational day 14 to 17 with either vehicle or CPF at a dose of 6 mg/kg/bw by oral gavages. Offspring of both sexes underwent assessment of early developmental milestones, including somatic growth, motor behavior and ultrasound vocalization. To evaluate the potential long-term effects of CPF, two different social behavior patterns typically altered in the BTBR strain (free social interaction with a same-sex companion in females, or interaction with a sexually receptive female in males) were also examined in the two sexes at adulthood. Our findings indicate significant effects of CPF on somatic growth and neonatal motor patterns. CPF treated pups showed reduced weight gain, delayed motor maturation (i.e., persistency of immature patterns such as pivoting at the expenses of coordinated locomotion) and a trend to enhanced ultrasound vocalization. At adulthood, CPF associated alterations were found in males only: the altered pattern of investigation of a sexual partner, previously described in BTBR mice, was enhanced in CPF males, and associated to increased ultrasonic vocalization rate. These findings strengthen the need of future studies to evaluate the role of environmental chemicals in the etiology of neurodevelopment disorders.

  8. Registers in Infant Phonation.

    PubMed

    Buder, Eugene H; McDaniel, Valerie F; Bene, Edina R; Ladmirault, Jennifer; Oller, D Kimbrough

    2018-04-09

    The primary vocal registers of modal, falsetto, and fry have been studied in adults but not per se in infancy. The vocal ligament is thought to play a critical role in the modal-falsetto contrast but is still developing during infancy (Tateya and Tateya, 2015). 41 Cover tissues are also implicated in the modal-fry contrast, but the low fundamental frequency (f o ) cutoff of 70 Hz, shared between genders, suggests a psychoacoustic basis for the contrast. Buder, Chorna, Oller, and Robinson (2008) 6 used the labels of "loft," "modal," and "pulse" for distinct vibratory regimes that appear to be identifiable based on spectrographic inspection of harmonic structure and auditory judgments in infants, but this work did not supply acoustic measurements to verify which of these nominally labeled regimes resembled adult registers. In this report, we identify clear transitions between registers within infant vocalizations and measure these registers and their transitions for f o and relative harmonic amplitudes (H1-H2). By selectively sampling first-year vocalizations, this manuscript quantifies acoustic patterns that correspond to vocal fold vibration types not previously cataloged in infancy. Results support a developmental basis for vocal registers, revealing that a well-developed ligament is not needed for loft-modal quality shifts as seen in harmonic amplitude measures. Results also reveal that a distinctively pulsatile register can occur in infants at a much higher f o than expected on psychoacoustic grounds. Overall results are consistent with cover tissues in infancy that are, for vibratory purposes, highly compliant and readily detached. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  9. Mouse current vocalization threshold measured with a neurospecific nociception assay: The effect of sex, morphine, and isoflurane

    PubMed Central

    Spornick, Nicholas; Guptill, Virginia; Koziol, Deloris; Wesley, Robert; Finkel, Julia; Quezado, Zenaide M.N.

    2012-01-01

    Sine-wave electrical stimulation at frequencies 2000, 250, and 5 Hz to respectively evaluate Aβ, Aδ, and C sensory neurons has recently been added to the armamentarium used to evaluate sensory neurons. We developed an automated nociception assay using sine-wave stimulation methodology to determine current vocalization threshold in response to 2000, 250, and 5 Hz and examine the effects of sex, analgesics, and anesthetics in mice. At baseline, males had significantly higher mean current vocalization thresholds compared with female mice at 2000, 250, and 5 Hz (p ≤ 0.019). By 1 h after intrathecal injections of morphine there were significant increases in current vocalization threshold percent changes from baseline that varied with doses (p = 0.0001) and frequency used (p < 0.0001). Specifically, with increasing doses of morphine, there were significantly greater increases in current vocalization threshold percent changes from baseline in response to 5 Hz compared with 250 and 2000 Hz stimulation in a significantly ordered pattern: 5 Hz > 250 Hz (p < 0.0001) and 250 Hz > 2000 Hz (p = 0.0002). Forty-five minutes after exposure, there were no effects of isoflurane on current vocalization thresholds at any frequency. Therefore, our findings suggest that this automated nociception assay using sine-wave stimulation in mice, can be valuable for measurements of the effects of sex, opioids, and anesthetics on the response to electrical stimuli that preferentially stimulate Aβ, Aδ, and C-sensory fibers in vivo. This investigation suggests the validation of this assay and supports its use to examine mechanisms of nociception in mice. PMID:21864576

  10. A neural circuit mechanism for regulating vocal variability during song learning in zebra finches.

    PubMed

    Garst-Orozco, Jonathan; Babadi, Baktash; Ölveczky, Bence P

    2014-12-15

    Motor skill learning is characterized by improved performance and reduced motor variability. The neural mechanisms that couple skill level and variability, however, are not known. The zebra finch, a songbird, presents a unique opportunity to address this question because production of learned song and induction of vocal variability are instantiated in distinct circuits that converge on a motor cortex analogue controlling vocal output. To probe the interplay between learning and variability, we made intracellular recordings from neurons in this area, characterizing how their inputs from the functionally distinct pathways change throughout song development. We found that inputs that drive stereotyped song-patterns are strengthened and pruned, while inputs that induce variability remain unchanged. A simple network model showed that strengthening and pruning of action-specific connections reduces the sensitivity of motor control circuits to variable input and neural 'noise'. This identifies a simple and general mechanism for learning-related regulation of motor variability.

  11. Interaction of the phencyclidine model of schizophrenia and nicotine on total and categorized ultrasonic vocalizations in rats

    PubMed Central

    Swalve, Natashia; Mulholland, Michele M.; Schulz, Tiffany D.; Li, Ming

    2015-01-01

    Patients with schizophrenia smoke cigarettes at a higher rate than the general population. We hypothesized that a factor in this comorbidity is sensitivity to the reinforcing and reinforcement-enhancement effects of nicotine. Phencyclidine (PCP) was used to model behavioral changes resembling negative symptoms of schizophrenia in rats. USVs in rats have been used to measure emotional states, with 50 kHz USVs indicating positive states and 22 kHz indicating negative. Total and categorized numbers of 22 and 50 kHz ultrasonic vocalizations (USVs) and USVs during a visual stimulus (e.g. a potential measure of reinforcement-enhancement) were examined in rats following .injection ofh PCP (2.0 mg/kg), and/or nicotine (0.2 or 0.4 mg/kg) daily for 7 days. PCP was then discontinued and all rats received nicotine (0.2 mg/kg and 0.4 mg/kg) and PCP (2.0 mg/kg) on 3 challenge days. PCP acutely decreased 50 kHz vocalizations while repeated nicotine potentiated rates of vocalizations, with similar patterns during light presentations. Rats in the PCP and nicotine combination groups made more 50 kHz vocalizations compared to control groups on challenge days. We conclude that PCP may produce a reward deficit that is shown by decreased 50 kHz USVs, and behaviors post-PCP exposure may best model the comorbidity between schizophrenia and nicotine. PMID:26479849

  12. Speech therapy and voice recognition instrument

    NASA Technical Reports Server (NTRS)

    Cohen, J.; Babcock, M. L.

    1972-01-01

    Characteristics of electronic circuit for examining variations in vocal excitation for diagnostic purposes and in speech recognition for determiniog voice patterns and pitch changes are described. Operation of the circuit is discussed and circuit diagram is provided.

  13. Characteristics of fin whale vocalizations recorded on instruments in the northeast Pacific Ocean

    NASA Astrophysics Data System (ADS)

    Weirathmueller, Maria Michelle Josephine

    This thesis focuses on fin whale vocalizations recorded on ocean bottom seismometers (OBSs) in the Northeast Pacific Ocean, using data collected between 2003 and 2013. OBSs are a valuable, and largely untapped resource for the passive acoustic monitoring of large baleen whales. This dissertation is divided into three parts, each of which uses the recordings of fin whale vocalizations to better understand their calling behaviors and distributions. The first study describes the development of a technique to extract source levels of fin whale vocalizations from OBS recordings. Source levels were estimated using data collected on a network of eight OBSs in the Northeast Pacific Ocean. The acoustic pressure levels measured at the instruments were adjusted for the propagation path between the calling whales and the instruments using the call location and estimating losses along the acoustic travel path. A total of 1241 calls were used to estimate an average source level of 189 +/-5.8 dB re 1uPa 1m. This variability is largely attributed to uncertainties in the horizontal and vertical position of the fin whale at the time of each call, and the effect of these uncertainties on subsequent calculations. The second study describes a semi-automated method for obtaining horizontal ranges to vocalizing fin whales using the timing and relative amplitude of multipath arrivals. A matched filter is used to detect fin whale calls and pick the relative times and amplitudes of multipath arrivals. Ray-based propagation models are used to predict multipath times and amplitudes as function of range. Because the direct and first multiple arrivals are not always observed, three hypotheses for the paths of the observed arrivals are considered; the solution is the hypothesis and range that optimizes the fit to the data. Ray-theoretical amplitudes are not accurate and solutions are improved by determining amplitudes from the observations using a bootstrap method. Data from ocean bottom seismometers at two locations are used to assess the method: one on the Juan de Fuca Ridge, a bathymetrically complex mid-ocean ridge environment, and the other at a flat sedimented location in the Cascadia Basin. At both sites, the method is reliable up to 4 km range which is sufficient to enable estimates of call density. The third study explores spatial and temporal trends in fin whale calling patterns. The frequency and inter-pulse interval of fin whale 20 Hz vocalizations were observed over 10 years from 2003-2013 on bottom mounted hydrophones and OBSs in the northeast Pacific Ocean. The instrument locations extended from 40°N and 130°W to 125°W with water depths ranging from 1500-4000 m. The inter-pulse interval (IPI) of fin whale song sequences was observed to increase at a rate of 0.59 seconds/year over the decade of observation. During the same time period, peak frequency decreased at a rate of 0.16 Hz/year. Two primary call patterns were observed. During the earlier years, the more commonly observed pattern had a single frequency and single IPI. In later years, a doublet pattern emerged, with two dominant frequencies and two IPIs. Many call sequences in the intervening years appeared to represent a transitional state between the two patterns. The overall trend was consistent across the entire geographical span, although some regional differences exist.

  14. Quantitative Tools for Examining the Vocalizations of Juvenile Songbirds

    PubMed Central

    Wellock, Cameron D.; Reeke, George N.

    2012-01-01

    The singing of juvenile songbirds is highly variable and not well stereotyped, a feature that makes it difficult to analyze with existing computational techniques. We present here a method suitable for analyzing such vocalizations, windowed spectral pattern recognition (WSPR). Rather than performing pairwise sample comparisons, WSPR measures the typicality of a sample against a large sample set. We also illustrate how WSPR can be used to perform a variety of tasks, such as sample classification, song ontogeny measurement, and song variability measurement. Finally, we present a novel measure, based on WSPR, for quantifying the apparent complexity of a bird's singing. PMID:22701474

  15. Subauditory Speech Recognition based on EMG/EPG Signals

    NASA Technical Reports Server (NTRS)

    Jorgensen, Charles; Lee, Diana Dee; Agabon, Shane; Lau, Sonie (Technical Monitor)

    2003-01-01

    Sub-vocal electromyogram/electro palatogram (EMG/EPG) signal classification is demonstrated as a method for silent speech recognition. Recorded electrode signals from the larynx and sublingual areas below the jaw are noise filtered and transformed into features using complex dual quad tree wavelet transforms. Feature sets for six sub-vocally pronounced words are trained using a trust region scaled conjugate gradient neural network. Real time signals for previously unseen patterns are classified into categories suitable for primitive control of graphic objects. Feature construction, recognition accuracy and an approach for extension of the technique to a variety of real world application areas are presented.

  16. Interactive voice technology: Variations in the vocal utterances of speakers performing a stress-inducing task

    NASA Astrophysics Data System (ADS)

    Mosko, J. D.; Stevens, K. N.; Griffin, G. R.

    1983-08-01

    Acoustical analyses were conducted of words produced by four speakers in a motion stress-inducing situation. The aim of the analyses was to document the kinds of changes that occur in the vocal utterances of speakers who are exposed to motion stress and to comment on the implications of these results for the design and development of voice interactive systems. The speakers differed markedly in the types and magnitudes of the changes that occurred in their speech. For some speakers, the stress-inducing experimental condition caused an increase in fundamental frequency, changes in the pattern of vocal fold vibration, shifts in vowel production and changes in the relative amplitudes of sounds containing turbulence noise. All speakers showed greater variability in the experimental condition than in more relaxed control situation. The variability was manifested in the acoustical characteristics of individual phonetic elements, particularly in speech sound variability observed serve to unstressed syllables. The kinds of changes and variability observed serve to emphasize the limitations of speech recognition systems based on template matching of patterns that are stored in the system during a training phase. There is need for a better understanding of these phonetic modifications and for developing ways of incorporating knowledge about these changes within a speech recognition system.

  17. Cortical Inhibition Reduces Information Redundancy at Presentation of Communication Sounds in the Primary Auditory Cortex

    PubMed Central

    Gaucher, Quentin; Huetz, Chloé; Gourévitch, Boris

    2013-01-01

    In all sensory modalities, intracortical inhibition shapes the functional properties of cortical neurons but also influences the responses to natural stimuli. Studies performed in various species have revealed that auditory cortex neurons respond to conspecific vocalizations by temporal spike patterns displaying a high trial-to-trial reliability, which might result from precise timing between excitation and inhibition. Studying the guinea pig auditory cortex, we show that partial blockage of GABAA receptors by gabazine (GBZ) application (10 μm, a concentration that promotes expansion of cortical receptive fields) increased the evoked firing rate and the spike-timing reliability during presentation of communication sounds (conspecific and heterospecific vocalizations), whereas GABAB receptor antagonists [10 μm saclofen; 10–50 μm CGP55845 (p-3-aminopropyl-p-diethoxymethyl phosphoric acid)] had nonsignificant effects. Computing mutual information (MI) from the responses to vocalizations using either the evoked firing rate or the temporal spike patterns revealed that GBZ application increased the MI derived from the activity of single cortical site but did not change the MI derived from population activity. In addition, quantification of information redundancy showed that GBZ significantly increased redundancy at the population level. This result suggests that a potential role of intracortical inhibition is to reduce information redundancy during the processing of natural stimuli. PMID:23804094

  18. A real-time LPC-based vocal tract area display for voice development.

    PubMed

    Rossiter, D; Howard, D M; Downes, M

    1994-12-01

    This article reports the design and implementation of a graphical display that presents an approximation to vocal tract area in real time for voiced vowel articulation. The acoustic signal is digitally sampled by the system. From these data a set of reflection coefficients is derived using linear predictive coding. A matrix of area coefficients is then determined that approximates the vocal tract area of the user. From this information a graphical display is then generated. The complete cycle of analysis and display is repeated at approximately 20 times/s. Synchronised audio and visual sequences can be recorded and used as dynamic targets for articulatory development. Use of the system is illustrated by diagrams of system output for spoken cardinal vowels and for vowels sung in a trained and untrained style.

  19. Reproduction of mouse-pup ultrasonic vocalizations by nanocrystalline silicon thermoacoustic emitter

    NASA Astrophysics Data System (ADS)

    Kihara, Takashi; Harada, Toshihiro; Kato, Masahiro; Nakano, Kiyoshi; Murakami, Osamu; Kikusui, Takefumi; Koshida, Nobuyoshi

    2006-01-01

    As one of the functional properties of ultrasound generator based on efficient thermal transfer at the nanocrystalline silicon (nc-Si) layer surface, its potential as an ultrasonic simulator of vocalization signals is demonstrated by using the acoustic data of mouse-pup calls. The device composed of a surface-heating thin-film electrode, an nc-Si layer, and a single-crystalline silicon (c-Si) wafer, exhibits an almost completely flat frequency response over a wide range without any mechanical surface vibration systems. It is shown that the fabricated emitter can reproduce digitally recorded ultrasonic mouse-pups vocalizations very accurately in terms of the call duration, frequency dispersion, and sound pressure level. The thermoacoustic nc-Si device provides a powerful physical means for the understanding of ultrasonic communication mechanisms in various living animals.

  20. A Joint Prosodic Origin of Language and Music

    PubMed Central

    Brown, Steven

    2017-01-01

    Vocal theories of the origin of language rarely make a case for the precursor functions that underlay the evolution of speech. The vocal expression of emotion is unquestionably the best candidate for such a precursor, although most evolutionary models of both language and speech ignore emotion and prosody altogether. I present here a model for a joint prosodic precursor of language and music in which ritualized group-level vocalizations served as the ancestral state. This precursor combined not only affective and intonational aspects of prosody, but also holistic and combinatorial mechanisms of phrase generation. From this common stage, there was a bifurcation to form language and music as separate, though homologous, specializations. This separation of language and music was accompanied by their (re)unification in songs with words. PMID:29163276

  1. A Mechanism for Frequency Modulation in Songbirds Shared with Humans

    PubMed Central

    Margoliash, Daniel

    2013-01-01

    In most animals that vocalize, control of fundamental frequency is a key element for effective communication. In humans, subglottal pressure controls vocal intensity but also influences fundamental frequency during phonation. Given the underlying similarities in the biomechanical mechanisms of vocalization in humans and songbirds, songbirds offer an attractive opportunity to study frequency modulation by pressure. Here, we present a novel technique for dynamic control of subsyringeal pressure in zebra finches. By regulating the opening of a custom-built fast valve connected to the air sac system, we achieved partial or total silencing of specific syllables, and could modify syllabic acoustics through more complex manipulations of air sac pressure. We also observed that more nuanced pressure variations over a limited interval during production of a syllable concomitantly affected the frequency of that syllable segment. These results can be explained in terms of a mathematical model for phonation that incorporates a nonlinear description for the vocal source capable of generating the observed frequency modulations induced by pressure variations. We conclude that the observed interaction between pressure and frequency was a feature of the source, not a result of feedback control. Our results indicate that, beyond regulating phonation or its absence, regulation of pressure is important for control of fundamental frequencies of vocalizations. Thus, although there are separate brainstem pathways for syringeal and respiratory control of song production, both can affect airflow and frequency. We hypothesize that the control of pressure and frequency is combined holistically at higher levels of the vocalization pathways. PMID:23825417

  2. A mechanism for frequency modulation in songbirds shared with humans.

    PubMed

    Amador, Ana; Margoliash, Daniel

    2013-07-03

    In most animals that vocalize, control of fundamental frequency is a key element for effective communication. In humans, subglottal pressure controls vocal intensity but also influences fundamental frequency during phonation. Given the underlying similarities in the biomechanical mechanisms of vocalization in humans and songbirds, songbirds offer an attractive opportunity to study frequency modulation by pressure. Here, we present a novel technique for dynamic control of subsyringeal pressure in zebra finches. By regulating the opening of a custom-built fast valve connected to the air sac system, we achieved partial or total silencing of specific syllables, and could modify syllabic acoustics through more complex manipulations of air sac pressure. We also observed that more nuanced pressure variations over a limited interval during production of a syllable concomitantly affected the frequency of that syllable segment. These results can be explained in terms of a mathematical model for phonation that incorporates a nonlinear description for the vocal source capable of generating the observed frequency modulations induced by pressure variations. We conclude that the observed interaction between pressure and frequency was a feature of the source, not a result of feedback control. Our results indicate that, beyond regulating phonation or its absence, regulation of pressure is important for control of fundamental frequencies of vocalizations. Thus, although there are separate brainstem pathways for syringeal and respiratory control of song production, both can affect airflow and frequency. We hypothesize that the control of pressure and frequency is combined holistically at higher levels of the vocalization pathways.

  3. Fluid-Structure Interactions as Flow Propagates Tangentially Over a Flexible Plate with Application to Voiced Speech Production

    NASA Astrophysics Data System (ADS)

    Westervelt, Andrea; Erath, Byron

    2013-11-01

    Voiced speech is produced by fluid-structure interactions that drive vocal fold motion. Viscous flow features influence the pressure in the gap between the vocal folds (i.e. glottis), thereby altering vocal fold dynamics and the sound that is produced. During the closing phases of the phonatory cycle, vortices form as a result of flow separation as air passes through the divergent glottis. It is hypothesized that the reduced pressure within a vortex core will alter the pressure distribution along the vocal fold surface, thereby aiding in vocal fold closure. The objective of this study is to determine the impact of intraglottal vortices on the fluid-structure interactions of voiced speech by investigating how the dynamics of a flexible plate are influenced by a vortex ring passing tangentially over it. A flexible plate, which models the medial vocal fold surface, is placed in a water-filled tank and positioned parallel to the exit of a vortex generator. The physical parameters of plate stiffness and vortex circulation are scaled with physiological values. As vortices propagate over the plate, particle image velocimetry measurements are captured to analyze the energy exchange between the fluid and flexible plate. The investigations are performed over a range of vortex formation numbers, and lateral displacements of the plate from the centerline of the vortex trajectory. Observations show plate oscillations with displacements directly correlated with the vortex core location.

  4. Chironomic stylization of intonation.

    PubMed

    d'Alessandro, Christophe; Rilliard, Albert; Le Beux, Sylvain

    2011-03-01

    Intonation stylization is studied using "chironomy," i.e., the analogy between hand gestures and prosodic movements. An intonation mimicking paradigm is used. The task of the ten subjects is to copy the intonation pattern of sentences with the help of a stylus on a graphic tablet, using a system for real-time manual intonation modification. Gestural imitation is compared to vocal imitation of the same sentences (seven for a male speaker, seven for a female speaker). Distance measures between gestural copies, vocal imitations, and original sentences are computed for performance assessment. Perceptual testing is also used for assessing the quality of gestural copies. The perceptual difference between natural and stylized contours is measured using a mean opinion score paradigm for 15 subjects. The results indicate that intonation contours can be stylized with accuracy by chironomic imitation. The results of vocal imitation and chironomic imitation are comparable, but subjects show better imitation results in vocal imitation. The best stylized contours using chironomy seems perceptually indistinguishable or almost indistinguishable from natural contours, particularly for female speech. This indicates that chironomic stylization is effective, and that hand movements can be analogous to intonation movements. © 2011 Acoustical Society of America

  5. Diagnostic and prognostic contribution of laryngeal electromyography in unilateral vocal-fold immobility in adults.

    PubMed

    Focquet, A; Péréon, Y; Ségura, S; Ferron, C; Malard, O; Espitalier, F

    2017-02-01

    To study the diagnostic and prognostic contribution of laryngeal electromyography in unilateral vocal-fold immobility in adults. A retrospective study included patients with unilateral vocal-fold immobility undergoing laryngeal electromyography between 2007 and 2015. Neurogenic, normal or myogenic findings were compared to the clinical aspect. Prognosis for recovery was assessed from motor unit potentials on laryngeal electromyography, and compared to subsequent progress on laryngoscopy. Sixty-three patients (mean age, 59 years) were initially included; 2 were subsequently excluded from analysis. Mean time from onset of immobility to laryngeal electromyography was 7 months. 85% of the 61 patients showed neurogenic findings, indicating neural lesion; 13% showed normal electromyography, indicating cricoarytenoid joint ankylosis; and 1 patient showed a myogenic pattern. Neurogenic cases were usually secondary to cervical surgery. Thirty-eight patients were followed up. In total, 75% of patients showing reinnervation potentials recovered. The positive predictive value of laryngeal electromyography was 69.2%. Laryngeal electromyography is effective in specifying the origin of unilateral vocal-fold immobility in adults. It also has a prognostic role, lack of reinnervation potentials being a possible indication for early medialization surgery. Copyright © 2016 Elsevier Masson SAS. All rights reserved.

  6. Case-control study of risk factors for spasmodic dysphonia: A comparison with other voice disorders.

    PubMed

    Tanner, Kristine; Roy, Nelson; Merrill, Ray M; Sauder, Cara; Houtz, Daniel R; Smith, Marshall E

    2012-05-01

    This epidemiology study examined risk factors uniquely associated with spasmodic dysphonia (SD). Case-control. A questionnaire was administered to 150 patients with SD (with and without coexisting vocal tremor) and 136 patients with other structural, neurological, and functional voice disorders (excluding SD and vocal tremor). Questions included personal and family medical histories, environmental exposures, trauma, illnesses, voice use habits, and the Short Form 36. Several factors were uniquely associated with SD (α = .05), including: 1) a personal history of cervical dystonia, sinus and throat illnesses, mumps, rubella, dust exposure, and frequent volunteer voice use, 2) a family history of voice disorders, 3) an immediate family history of vocal tremor and meningitis, and 4) an extended family history of head and neck tremor, ocular disease, and meningitis. Vocal tremor coexisted with SD in 29% of cases. Measles and mumps vaccines were protective for SD. SD is likely multifactorial and associated with several endogenous and exogenous factors. Certain viral exposures, voice use patterns, and familial neurological conditions may contribute to the onset of SD later in life. Copyright © 2011 The American Laryngological, Rhinological, and Otological Society, Inc.

  7. Observations of the relationship between noise exposure and preschool teacher voice usage in day-care center environments.

    PubMed

    Lindstrom, Fredric; Waye, Kerstin Persson; Södersten, Maria; McAllister, Anita; Ternström, Sten

    2011-03-01

    Although the relationship between noise exposure and vocal behavior (the Lombard effect) is well established, actual vocal behavior in the workplace is still relatively unexamined. The first purpose of this study was to investigate correlations between noise level and both voice level and voice average fundamental frequency (F₀) for a population of preschool teachers in their normal workplace. The second purpose was to study the vocal behavior of each teacher to investigate whether individual vocal behaviors or certain patterns could be identified. Voice and noise data were obtained for female preschool teachers (n=13) in their workplace, using wearable measurement equipment. Correlations between noise level and voice level, and between voice level and F₀, were calculated for each participant and ranged from 0.07 to 0.87 for voice level and from 0.11 to 0.78 for F₀. The large spread of the correlation coefficients indicates that the teachers react individually to the noise exposure. For example, some teachers increase their voice-to-noise level ratio when the noise is reduced, whereas others do not. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  8. Social functioning and autonomic nervous system sensitivity across vocal and musical emotion in Williams syndrome and autism spectrum disorder.

    PubMed

    Järvinen, Anna; Ng, Rowena; Crivelli, Davide; Neumann, Dirk; Arnold, Andrew J; Woo-VonHoogenstyn, Nicholas; Lai, Philip; Trauner, Doris; Bellugi, Ursula

    2016-01-01

    Both Williams syndrome (WS) and autism spectrum disorders (ASD) are associated with unusual auditory phenotypes with respect to processing vocal and musical stimuli, which may be shaped by the atypical social profiles that characterize the syndromes. Autonomic nervous system (ANS) reactivity to vocal and musical emotional stimuli was examined in 12 children with WS, 17 children with ASD, and 20 typically developing (TD) children, and related to their level of social functioning. The results of this small-scale study showed that after controlling for between-group differences in cognitive ability, all groups showed similar emotion identification performance across conditions. Additionally, in ASD, lower autonomic reactivity to human voice, and in TD, to musical emotion, was related to more normal social functioning. Compared to TD, both clinical groups showed increased arousal to vocalizations. A further result highlighted uniquely increased arousal to music in WS, contrasted with a decrease in arousal in ASD and TD. The ASD and WS groups exhibited arousal patterns suggestive of diminished habituation to the auditory stimuli. The results are discussed in the context of the clinical presentation of WS and ASD. © 2015 Wiley Periodicals, Inc.

  9. Sensory-motor interactions for vocal pitch monitoring in non-primary human auditory cortex.

    PubMed

    Greenlee, Jeremy D W; Behroozmand, Roozbeh; Larson, Charles R; Jackson, Adam W; Chen, Fangxiang; Hansen, Daniel R; Oya, Hiroyuki; Kawasaki, Hiroto; Howard, Matthew A

    2013-01-01

    The neural mechanisms underlying processing of auditory feedback during self-vocalization are poorly understood. One technique used to study the role of auditory feedback involves shifting the pitch of the feedback that a speaker receives, known as pitch-shifted feedback. We utilized a pitch shift self-vocalization and playback paradigm to investigate the underlying neural mechanisms of audio-vocal interaction. High-resolution electrocorticography (ECoG) signals were recorded directly from auditory cortex of 10 human subjects while they vocalized and received brief downward (-100 cents) pitch perturbations in their voice auditory feedback (speaking task). ECoG was also recorded when subjects passively listened to playback of their own pitch-shifted vocalizations. Feedback pitch perturbations elicited average evoked potential (AEP) and event-related band power (ERBP) responses, primarily in the high gamma (70-150 Hz) range, in focal areas of non-primary auditory cortex on superior temporal gyrus (STG). The AEPs and high gamma responses were both modulated by speaking compared with playback in a subset of STG contacts. From these contacts, a majority showed significant enhancement of high gamma power and AEP responses during speaking while the remaining contacts showed attenuated response amplitudes. The speaking-induced enhancement effect suggests that engaging the vocal motor system can modulate auditory cortical processing of self-produced sounds in such a way as to increase neural sensitivity for feedback pitch error detection. It is likely that mechanisms such as efference copies may be involved in this process, and modulation of AEP and high gamma responses imply that such modulatory effects may affect different cortical generators within distinctive functional networks that drive voice production and control.

  10. Left-hemisphere activation is associated with enhanced vocal pitch error detection in musicians with absolute pitch.

    PubMed

    Behroozmand, Roozbeh; Ibrahim, Nadine; Korzyukov, Oleg; Robin, Donald A; Larson, Charles R

    2014-02-01

    The ability to process auditory feedback for vocal pitch control is crucial during speaking and singing. Previous studies have suggested that musicians with absolute pitch (AP) develop specialized left-hemisphere mechanisms for pitch processing. The present study adopted an auditory feedback pitch perturbation paradigm combined with ERP recordings to test the hypothesis whether the neural mechanisms of the left-hemisphere enhance vocal pitch error detection and control in AP musicians compared with relative pitch (RP) musicians and non-musicians (NM). Results showed a stronger N1 response to pitch-shifted voice feedback in the right-hemisphere for both AP and RP musicians compared with the NM group. However, the left-hemisphere P2 component activation was greater in AP and RP musicians compared with NMs and also for the AP compared with RP musicians. The NM group was slower in generating compensatory vocal reactions to feedback pitch perturbation compared with musicians, and they failed to re-adjust their vocal pitch after the feedback perturbation was removed. These findings suggest that in the earlier stages of cortical neural processing, the right hemisphere is more active in musicians for detecting pitch changes in voice feedback. In the later stages, the left-hemisphere is more active during the processing of auditory feedback for vocal motor control and seems to involve specialized mechanisms that facilitate pitch processing in the AP compared with RP musicians. These findings indicate that the left hemisphere mechanisms of AP ability are associated with improved auditory feedback pitch processing during vocal pitch control in tasks such as speaking or singing. Copyright © 2013 Elsevier Inc. All rights reserved.

  11. Sensory-Motor Interactions for Vocal Pitch Monitoring in Non-Primary Human Auditory Cortex

    PubMed Central

    Larson, Charles R.; Jackson, Adam W.; Chen, Fangxiang; Hansen, Daniel R.; Oya, Hiroyuki; Kawasaki, Hiroto; Howard, Matthew A.

    2013-01-01

    The neural mechanisms underlying processing of auditory feedback during self-vocalization are poorly understood. One technique used to study the role of auditory feedback involves shifting the pitch of the feedback that a speaker receives, known as pitch-shifted feedback. We utilized a pitch shift self-vocalization and playback paradigm to investigate the underlying neural mechanisms of audio-vocal interaction. High-resolution electrocorticography (ECoG) signals were recorded directly from auditory cortex of 10 human subjects while they vocalized and received brief downward (−100 cents) pitch perturbations in their voice auditory feedback (speaking task). ECoG was also recorded when subjects passively listened to playback of their own pitch-shifted vocalizations. Feedback pitch perturbations elicited average evoked potential (AEP) and event-related band power (ERBP) responses, primarily in the high gamma (70–150 Hz) range, in focal areas of non-primary auditory cortex on superior temporal gyrus (STG). The AEPs and high gamma responses were both modulated by speaking compared with playback in a subset of STG contacts. From these contacts, a majority showed significant enhancement of high gamma power and AEP responses during speaking while the remaining contacts showed attenuated response amplitudes. The speaking-induced enhancement effect suggests that engaging the vocal motor system can modulate auditory cortical processing of self-produced sounds in such a way as to increase neural sensitivity for feedback pitch error detection. It is likely that mechanisms such as efference copies may be involved in this process, and modulation of AEP and high gamma responses imply that such modulatory effects may affect different cortical generators within distinctive functional networks that drive voice production and control. PMID:23577157

  12. Nonlinear laser scanning microscopy of human vocal folds.

    PubMed

    Miri, Amir K; Tripathy, Umakanta; Mongeau, Luc; Wiseman, Paul W

    2012-02-01

    The purpose of this work was to apply nonlinear laser scanning microscopy (NLSM) for visualizing the morphology of extracellular matrix proteins within human vocal folds. This technique may potentially assist clinicians in making rapid diagnoses of vocal fold tissue disease or damage. Microstructural characterization based on NLSM provides valuable information for better understanding molecular mechanisms and tissue structure. Experimental, ex vivo human vocal fold. A custom-built multimodal nonlinear laser scanning microscope was used to scan fibrillar proteins in three 4% formaldehyde-fixed cadaveric samples. Collagen and elastin, key extracellular matrix proteins in the vocal fold lamina propria, were imaged by two nonlinear microscopy modalities: second harmonic generation (SHG) and two-photon fluorescence (TPF), respectively. An experimental protocol was introduced to characterize the geometrical properties of the imaged fibrous proteins. NLSM revealed the biomorphology of the human vocal fold fibrous proteins. No photobleaching was observed for the incident laser power of ∼60 mW before the excitation objective. Types I and III fibrillar collagen were imaged without label in the tissue by intrinsic SHG. Imaging while rotating the incident laser light-polarization direction confirmed a helical shape for the collagen fibers. The amplitude, periodicity, and overall orientation were then computed for the helically distributed collagen network. The elastin network was simultaneously imaged via TPF and found to have a basket-like structure. In some regions, particularly close to the epithelium, colocalization of both extracellular matrix components were observed. A benchmark study is presented for quantitative real-time, ex vivo, NLSM imaging of the extracellular macromolecules in human vocal fold lamina propria. The results are promising for clinical applications. Copyright © 2011 The American Laryngological, Rhinological, and Otological Society, Inc.

  13. Error-dependent modulation of speech-induced auditory suppression for pitch-shifted voice feedback

    PubMed Central

    2011-01-01

    Background The motor-driven predictions about expected sensory feedback (efference copies) have been proposed to play an important role in recognition of sensory consequences of self-produced motor actions. In the auditory system, this effect was suggested to result in suppression of sensory neural responses to self-produced voices that are predicted by the efference copies during vocal production in comparison with passive listening to the playback of the identical self-vocalizations. In the present study, event-related potentials (ERPs) were recorded in response to upward pitch shift stimuli (PSS) with five different magnitudes (0, +50, +100, +200 and +400 cents) at voice onset during active vocal production and passive listening to the playback. Results Results indicated that the suppression of the N1 component during vocal production was largest for unaltered voice feedback (PSS: 0 cents), became smaller as the magnitude of PSS increased to 200 cents, and was almost completely eliminated in response to 400 cents stimuli. Conclusions Findings of the present study suggest that the brain utilizes the motor predictions (efference copies) to determine the source of incoming stimuli and maximally suppresses the auditory responses to unaltered feedback of self-vocalizations. The reduction of suppression for 50, 100 and 200 cents and its elimination for 400 cents pitch-shifted voice auditory feedback support the idea that motor-driven suppression of voice feedback leads to distinctly different sensory neural processing of self vs. non-self vocalizations. This characteristic may enable the audio-vocal system to more effectively detect and correct for unexpected errors in the feedback of self-produced voice pitch compared with externally-generated sounds. PMID:21645406

  14. Patterns in Early Interaction between Young Preschool Children with Severe Speech and Physical Impairments and Their Parents

    ERIC Educational Resources Information Center

    Sandberg, Annika Dahlgren; Liliedahl, Marie

    2008-01-01

    The aim of this study is to examine whether the asymmetrical pattern of communication usually found between people who use augmentative and alternative communication and their partners using natural speech was also found in the interaction between non-vocal young preschool children with cerebral palsy and their parents. Three parent-child dyads…

  15. Divergent morphological and acoustic traits in sympatric communities of Asian barbets

    PubMed Central

    Tamma, Krishnapriya

    2016-01-01

    The opposing effects of environmental filtering and competitive interactions may influence community assembly and coexistence of related species. Competition, both in the domain of ecological resources, and in the sensory domain (for example, acoustic interference) may also result in sympatric species evolving divergent traits and niches. Delineating these scenarios within communities requires understanding trait distributions and phylogenetic structure within the community, as well as patterns of trait evolution. We report that sympatric assemblages of Asian barbets (frugivorous canopy birds) consist of a random phylogenetic sample of species, but are divergent in both morphological and acoustic traits. Additionally, we find that morphology is more divergent than expected under Brownian evolution, whereas vocal frequency evolution is close to the pattern expected under Brownian motion (i.e. a random walk). Together, these patterns are consistent with a role for competition or competitive exclusion in driving community assembly. Phylogenetic patterns of morphological divergence between related species suggest that these traits are key in species coexistence. Because vocal frequency and size are correlated in barbets, we therefore hypothesize that frequency differences between sympatric barbets are a by-product of their divergent morphologies. PMID:27853589

  16. The respiratory-vocal system of songbirds: anatomy, physiology, and neural control.

    PubMed

    Schmidt, Marc F; Martin Wild, J

    2014-01-01

    This wide-ranging review presents an overview of the respiratory-vocal system in songbirds, which are the only other vertebrate group known to display a degree of respiratory control during song rivalling that of humans during speech; this despite the fact that the peripheral components of both the respiratory and vocal systems differ substantially in the two groups. We first provide a brief description of these peripheral components in songbirds (lungs, air sacs and respiratory muscles, vocal organ (syrinx), upper vocal tract) and then proceed to a review of the organization of central respiratory-related neurons in the spinal cord and brainstem, the latter having an organization fundamentally similar to that of the ventral respiratory group of mammals. The second half of the review describes the nature of the motor commands generated in a specialized "cortical" song control circuit and how these might engage brainstem respiratory networks to shape the temporal structure of song. We also discuss a bilaterally projecting "respiratory-thalamic" pathway that links the respiratory system to "cortical" song control nuclei. This necessary pathway for song originates in the brainstem's primary inspiratory center and is hypothesized to play a vital role in synchronizing song motor commands both within and across hemispheres. © 2014 Elsevier B.V. All rights reserved.

  17. The respiratory-vocal system of songbirds: Anatomy, physiology, and neural control

    PubMed Central

    Schmidt, Marc F.; Wild, J. Martin

    2015-01-01

    This wide-ranging review presents an overview of the respiratory-vocal system in songbirds, which are the only other vertebrate group known to display a degree of respiratory control during song rivalling that of humans during speech; this despite the fact that the peripheral components of both the respiratory and vocal systems differ substantially in the two groups. We first provide a brief description of these peripheral components in songbirds (lungs, air sacs and respiratory muscles, vocal organ (syrinx), upper vocal tract) and then proceed to a review of the organization of central respiratory-related neurons in the spinal cord and brainstem, the latter having an organization fundamentally similar to that of the ventral respiratory group of mammals. The second half of the review describes the nature of the motor commands generated in a specialized “cortical” song control circuit and how these might engage brainstem respiratory networks to shape the temporal structure of song. We also discuss a bilaterally projecting “respiratory-thalamic” pathway that links the respiratory system to “cortical” song control nuclei. This necessary pathway for song originates in the brainstem’s primary inspiratory center and is hypothesized to play a vital role in synchronizing song motor commands both within and across hemispheres. PMID:25194204

  18. [Pursed Lips Inspiration for Vocal Cord Dysfunction].

    PubMed

    Maruyama, Yumiko; Tsukada, Yayoi; Hirai, Nobuyuki; Nakanishi, Yosuke; Yoshizaki, Tomokazu

    2015-01-01

    Paradoxical vocal cord motion (PVCM) during vocal cord dysfunction (VCD) generally occurs spasmodically and transiently. After we had experienced 36 cases of VCD and successfully treated with conservative treatment including "pursed lips inspiration" method, we experienced a boy who had persistent PVCM. It was observed his PVCM vanished when he breathed in through pursed lips, while it appeared again when he stopped pursed lips inspiration. An airway reflex has been reported where the negative pressure in the subglottic space resulting from the inspiratory effort against a narrowed glottis activates the vocal cord adductor. VCD is considered to have both acceleration of laryngeal closure reflex against airway stimuli and active adductive movement of vocal cords against negative pressure in the subglottic space as underlying factors. The pursed lips inspiration method enables VCD patients not only to accomplish slow and light breathing but also to decrease the difference in the pressure between the supra--and subglottic space by occluding the nasal cavity and voluntary puckering up of the mouth which generate negative pressure in the supraglottic space. This is the first report of the pursed lips inspiration method as a treatment for VCD. Pursed lips inspiration is a simple method which is easy to perform anytime, anywhere without any special equipment, and is considered to be worth trying for VCD.

  19. Acoustic, respiratory kinematic and electromyographic effects of vocal training

    NASA Astrophysics Data System (ADS)

    Mendes, Ana Paula De Brito Garcia

    The longitudinal effects of vocal training on the respiratory, phonatory and articulatory systems were investigated in this study. During four semesters, fourteen voice major students were recorded while speaking and singing. Acoustic, temporal, respiratory kinematic and electromyographic parameters were measured to determine changes in the three systems as a function of vocal training. Acoustic measures of the speaking voice included fundamental frequency, sound pressure level (SPL), percent jitter and shimmer, and harmonic-to-noise ratio. Temporal measures included duration of sentences, diphthongs and the closure durations of stop consonants. Acoustic measures of the singing voice included fundamental frequency and sound pressure level of the phonational range, vibrato pulses per second, vibrato amplitude variation and the presence of the singer's formant. Analysis of the data revealed that vocal training had a significant effect on the singing voice. Fundamental frequency and SPL of the 90% level and 90--10% of the phonational range increased significantly during four semesters of vocal training. Physiological data was collected from four subjects during three semesters of vocal training. Respiratory kinematic measures included lung volume, rib cage and abdominal excursions extracted from spoken sung samples. Descriptive statistics revealed that rib cage and abdominal excursions increased from the 1st to the 2nd semester and decrease from the 2nd to the 3rd semester of vocal training. Electromyographic measures of the pectoralis major, rectus abdominis and external obliques muscles revealed that burst duration means decreased from the 1st to the 2nd semester and increased from the 2nd to the 3rd semester. Peak amplitude means increased from the 1st to the 2nd and decreased from the 2nd to the 3rd semester of vocal training. Chest wall excursions and muscle force generation of the three muscles increased as the demanding level and the length of the phonatory tasks increased.

  20. Flow fields and acoustics in a unilateral scarred vocal fold model.

    PubMed

    Murugappan, Shanmugam; Khosla, Sid; Casper, Keith; Oren, Liran; Gutmark, Ephraim

    2009-01-01

    From prior work in an excised canine larynx model, it has been shown that intraglottal vortices form between the vocal folds during the latter part of closing. It has also been shown that the vortices generate a negative pressure between the folds, producing a suction force that causes sudden, rapid closing of the folds. This rapid closing will produce increased loudness and increased higher harmonics. We used a unilateral scarred excised canine larynx model to determine whether the intraglottal vortices and resulting acoustics were changed, compared to those of normal larynges. Acoustic, flow field, and high-speed imaging measurements from 5 normal and 5 unilaterally scarred canine larynges are presented in this report. Scarring was produced by complete resection of the vocal fold mucosa and superficial layer of the lamina propria on the right vocal fold only. Two months later, these dogs were painlessly sacrificed, and testing was done on the excised larynges during phonation. High-speed video imaging was then used to measure vocal fold displacement during different phases. Particle image velocimetry and acoustic measurements were used to describe possible acoustic effects of the vortices. A higher phonation threshold was required to excite the motion of the vocal fold in scarred larynges. As the subglottal pressure increased, the strength of the vortices and the higher harmonics both consistently increased. However, it was seen that increasing the maximum displacement of the scarred fold did not consistently increase the higher harmonics. The improvements that result from increasing subglottal pressure may be due to a combination of increasing the strength of the intraglottal vortices and increasing the maximum displacement of the vocal fold; however, the data in this study suggest that the vortices play a much more important role. The current study indicates that higher subglottal pressures may excite higher harmonics and improve loudness for patients with unilateral vocal fold scarring. This finding implies that therapies that raise the subglottal pressure may be helpful in improving voice quality.

  1. Biomechanical simulation of vocal fold dynamics in adults based on laryngeal high-speed videoendoscopy

    PubMed Central

    Gómez, Pablo; Patel, Rita R.; Alexiou, Christoph; Bohr, Christopher; Schützenberger, Anne

    2017-01-01

    Motivation Human voice is generated in the larynx by the two oscillating vocal folds. Owing to the limited space and accessibility of the larynx, endoscopic investigation of the actual phonatory process in detail is challenging. Hence the biomechanics of the human phonatory process are still not yet fully understood. Therefore, we adapt a mathematical model of the vocal folds towards vocal fold oscillations to quantify gender and age related differences expressed by computed biomechanical model parameters. Methods The vocal fold dynamics are visualized by laryngeal high-speed videoendoscopy (4000 fps). A total of 33 healthy young subjects (16 females, 17 males) and 11 elderly subjects (5 females, 6 males) were recorded. A numerical two-mass model is adapted to the recorded vocal fold oscillations by varying model masses, stiffness and subglottal pressure. For adapting the model towards the recorded vocal fold dynamics, three different optimization algorithms (Nelder–Mead, Particle Swarm Optimization and Simulated Bee Colony) in combination with three cost functions were considered for applicability. Gender differences and age-related kinematic differences reflected by the model parameters were analyzed. Results and conclusion The biomechanical model in combination with numerical optimization techniques allowed phonatory behavior to be simulated and laryngeal parameters involved to be quantified. All three optimization algorithms showed promising results. However, only one cost function seems to be suitable for this optimization task. The gained model parameters reflect the phonatory biomechanics for men and women well and show quantitative age- and gender-specific differences. The model parameters for younger females and males showed lower subglottal pressures, lower stiffness and higher masses than the corresponding elderly groups. Females exhibited higher subglottal pressures, smaller oscillation masses and larger stiffness than the corresponding similar aged male groups. Optimizing numerical models towards vocal fold oscillations is useful to identify underlying laryngeal components controlling the phonatory process. PMID:29121085

  2. Gender Differences in the Recognition of Vocal Emotions

    PubMed Central

    Lausen, Adi; Schacht, Annekathrin

    2018-01-01

    The conflicting findings from the few studies conducted with regard to gender differences in the recognition of vocal expressions of emotion have left the exact nature of these differences unclear. Several investigators have argued that a comprehensive understanding of gender differences in vocal emotion recognition can only be achieved by replicating these studies while accounting for influential factors such as stimulus type, gender-balanced samples, number of encoders, decoders, and emotional categories. This study aimed to account for these factors by investigating whether emotion recognition from vocal expressions differs as a function of both listeners' and speakers' gender. A total of N = 290 participants were randomly and equally allocated to two groups. One group listened to words and pseudo-words, while the other group listened to sentences and affect bursts. Participants were asked to categorize the stimuli with respect to the expressed emotions in a fixed-choice response format. Overall, females were more accurate than males when decoding vocal emotions, however, when testing for specific emotions these differences were small in magnitude. Speakers' gender had a significant impact on how listeners' judged emotions from the voice. The group listening to words and pseudo-words had higher identification rates for emotions spoken by male than by female actors, whereas in the group listening to sentences and affect bursts the identification rates were higher when emotions were uttered by female than male actors. The mixed pattern for emotion-specific effects, however, indicates that, in the vocal channel, the reliability of emotion judgments is not systematically influenced by speakers' gender and the related stereotypes of emotional expressivity. Together, these results extend previous findings by showing effects of listeners' and speakers' gender on the recognition of vocal emotions. They stress the importance of distinguishing these factors to explain recognition ability in the processing of emotional prosody. PMID:29922202

  3. [Developing a canine vocal fold scar model by CO₂ laser and studying the LOX, HSP70 and HA expression in its extra celluar matrix].

    PubMed

    Liang, G T; Duan, B Y; Zhang, Y Y; Luo, S W; Lu, L; Yang, L P; Wang, B R

    2017-01-20

    Objective: Building a canine vocal fold scar model and analyzing the changes in morphology, histopathology and related factors of extra cellular matrix (ECM) of vocal cord healing at different time periods. Method: Five experimental dogs were randomly divided into the control group(one dog) and the experimental group (four dogs). No special treatment was done on the control group, and the experimental group was given CO₂ laser through laryngoscope with minimally invasive surgery on bilateral vocal cords. Observation of the morphological changes of injured vocal cords was made at five different time, pre-operation and 6 h, 3 w, 8 w and 12 w post-operation. HE staining, immunofluorescence, were used respectively to observe the histopathological and ultrastructural changes of each layer of vocal cord. Observation were made on the changing pattern of lysyl oxidase (LOX), heat shock proteins70 (HSP70), and the integrated optical density index (IOD) of Hyaluronic acid (HA) in vocal cord tissues. Result: ①Laryngoscope shows there were mild congestion and edema and inflammatory exudation on the wound surface of vocal cord 3w post-operation. On the wound surface of vocal cord 8 w post-operation, the congestion and edema disappear, the local contraction sink, and no adhesion and granulation form. 12 w post-operation, the surface of the vocal cord was smooth, there were local contractures, vocal fold scar form, and hoarseness in the bilateral vocal cords was obvious. ②HE staining shows 6 h post-operation there were a large number of inflammatory cell infiltration, red blood cell leakage, and cell congestion and edema on the wound surface of vocal cord. 3 w post-operation, there were fibroblast proliferation, angiogenesis, and a large number of fibrous tissues disorderly arranged on the wound surface of vocal cord. 8w post-operation, each layer of fibrous tissues were in hyperplasia and the blood vessels was thicken after on the wound surface of vocal cord, 12 w post-operation, a large number of collagens increases and were in group or fasciculation disorder. There were many irregular gaps in fibers, and blood vessels and glands become rare or disappear. ③The immunofluorescence showed LOX, HSP70 and HA were all localized in the cellular cytoplasm and nucleus. The expression levels were different at different postoperative time, and they were mainly relatively strong expressed in the inflammatory cells, vascular endothelial cells and the glands. ④The comparison of IOD values: The IOD values of LOX were different at different time periods ( P < 0.05). There were no significant differences in IOD of HSP70 between pre-operation and 12 w post-operation, but there are significant differences among other groups ( P < 0.05). There were no significant differences in IOD of HA between pre-operation and 12 w post-operation, but there were significant differences among other groups ( P < 0.01). ⑤The expression of LOX decreased 6h post-operation and increased 3-12 w post-operation. The expression of HSP70 post-operation reached the peak 6 h post-operation and decreased 3-12 w post-operation. The expression of HA decreased 6 h post-operation, increased to the peak 8 w post-operation, and decreased 8-12 w post-operation.⑥The transmission electron microscope showed 3 w to 8 w post-operation, in the intrinsic layer of the vocal cord, the fibroblasts were extremely active, the cells swelled, and the organelles were abundant.12 w post-operation, there were a small number of fibroblasts in the active state, and larger gaps between the fibers and fewer and thinner elastic fibers. Conclusion: CO₂ laser ablation of vocal cords under the Micro Post-Laryngoscope could establish reliable animal model of vocal fold scar, and 12 w reticular the vocal fold scars basically form. LOX, HSP70, HA play different roles at early, middle, and late stages in the vocal fold scar formation, and they can be used as a more sensitive index for vocal fold scar formation. Copyright© by the Editorial Department of Journal of Clinical Otorhinolaryngology Head and Neck Surgery.

  4. Mapping the distribution of language related genes FoxP1, FoxP2, and CntnaP2 in the brains of vocal learning bat species

    PubMed Central

    Rodenas‐Cuadrado, Pedro M.; Mengede, Janine; Baas, Laura; Devanna, Paolo; Schmid, Tobias A.; Yartsev, Michael; Firzlaff, Uwe

    2018-01-01

    Abstract Genes including FOXP2, FOXP1, and CNTNAP2, have been implicated in human speech and language phenotypes, pointing to a role in the development of normal language‐related circuitry in the brain. Although speech and language are unique to humans a comparative approach is possible by addressing language‐relevant traits in animal systems. One such trait, vocal learning, represents an essential component of human spoken language, and is shared by cetaceans, pinnipeds, elephants, some birds and bats. Given their vocal learning abilities, gregarious nature, and reliance on vocalizations for social communication and navigation, bats represent an intriguing mammalian system in which to explore language‐relevant genes. We used immunohistochemistry to detail the distribution of FoxP2, FoxP1, and Cntnap2 proteins, accompanied by detailed cytoarchitectural histology in the brains of two vocal learning bat species; Phyllostomus discolor and Rousettus aegyptiacus. We show widespread expression of these genes, similar to what has been previously observed in other species, including humans. A striking difference was observed in the adult P. discolor bat, which showed low levels of FoxP2 expression in the cortex that contrasted with patterns found in rodents and nonhuman primates. We created an online, open‐access database within which all data can be browsed, searched, and high resolution images viewed to single cell resolution. The data presented herein reveal regions of interest in the bat brain and provide new opportunities to address the role of these language‐related genes in complex vocal‐motor and vocal learning behaviors in a mammalian model system. PMID:29297931

  5. Working Conditions and Workplace Barriers to Vocal Health in Primary School Teachers.

    PubMed

    Munier, Caitriona; Farrell, Rory

    2016-01-01

    The purpose of this study was to identify the working conditions and workplace barriers to vocal health in primary school teachers. The relationship between working conditions and voice is analyzed. This is a survey study in 42 randomized schools from a restricted geographical area. An 85-item questionnaire was administered to 550 primary school teachers in 42 schools in Dublin. It was designed to obtain information on demographics, vocal use patterns, vocal health, work organization, working conditions, and teacher's perceptions of the conditions in teaching that might cause a voice problem. The relationship between voice and overstretched work demands, and voice and class size, was examined. A chi-squared test was run to test the null hypothesis that the variables overstretched work demands and voice and class size and voice are independent. Subjects were given the opportunity to give their opinion on their working conditions and on the availability of advice and support within the workplace. A final question sought their opinion on what should be included in a voice care program. A 55% response rate was obtained (n = 304). It was found with 96.52% confidence that the variables overstretched work demands and voice are related. Likewise, it was found that the variables class size and voice are related with 99.97% confidence. There are workplace barriers to vocal health. The working conditions of primary school teachers need to be fully adapted to promote vocal health. Changes by education and health policy makers are needed to achieve this goal. There is a need for future research which focuses on the working conditions of teachers. Copyright © 2016. Published by Elsevier Inc.

  6. Singing modulates parvalbumin interneurons throughout songbird forebrain vocal control circuitry

    PubMed Central

    Zengin-Toktas, Yildiz

    2017-01-01

    Across species, the performance of vocal signals can be modulated by the social environment. Zebra finches, for example, adjust their song performance when singing to females (‘female-directed’ or FD song) compared to when singing in isolation (‘undirected’ or UD song). These changes are salient, as females prefer the FD song over the UD song. Despite the importance of these performance changes, the neural mechanisms underlying this social modulation remain poorly understood. Previous work in finches has established that expression of the immediate early gene EGR1 is increased during singing and modulated by social context within the vocal control circuitry. Here, we examined whether particular neural subpopulations within those vocal control regions exhibit similar modulations of EGR1 expression. We compared EGR1 expression in neurons expressing parvalbumin (PV), a calcium buffer that modulates network plasticity and homeostasis, among males that performed FD song, males that produced UD song, or males that did not sing. We found that, overall, singing but not social context significantly affected EGR1 expression in PV neurons throughout the vocal control nuclei. We observed differences in EGR1 expression between two classes of PV interneurons in the basal ganglia nucleus Area X. Additionally, we found that singing altered the amount of PV expression in neurons in HVC and Area X and that distinct PV interneuron types in Area X exhibited different patterns of modulation by singing. These data indicate that throughout the vocal control circuitry the singing-related regulation of EGR1 expression in PV neurons may be less influenced by social context than in other neuron types and raise the possibility of cell-type specific differences in plasticity and calcium buffering. PMID:28235074

  7. Multigenerational effects of bisphenol A or ethinyl estradiol exposure on F2 California mice (Peromyscus californicus) pup vocalizations

    PubMed Central

    Johnson, Sarah A.; Farrington, Michelle J.; Murphy, Claire R.; McAllister, Leif A.; Kaur, Sarabjit; Chun, Catherine; Ortega, Madison T.; Marshall, Brittney L.; Hoffmann, Frauke; Ellersieck, Mark R.; Schenk, A. Katrin

    2018-01-01

    Rodent pups use vocalizations to communicate with one or both parents in biparental species, such as California mice (Peromyscus californicus). Previous studies have shown California mice developmentally exposed to endocrine disrupting chemicals, bisphenol A (BPA) or ethinyl estradiol (EE), demonstrate later compromised parental behaviors. Reductions in F1 parental behaviors might also be due to decreased emissions of F2 pup vocalizations. Thus, vocalizations of F2 male and female California mice pups born to F1 parents developmentally exposed to BPA, EE, or controls were examined. Postnatal days (PND) 2–4 were considered early postnatal period, PND 7 and 14 were defined as mid-postnatal period, and PND 21 and 28 were classified as late postnatal period. EE pups showed increased latency to emit the first syllable compared to controls. BPA female pups had decreased syllable duration compared to control and EE female pups during the early postnatal period but enhanced responses compared to controls at late postnatal period; whereas, male BPA and EE pups showed greater syllable duration compared to controls during early postnatal period. In mid-postnatal period, F2 BPA and EE pups emitted greater number of phrases than F2 control pups. Results indicate aspects of vocalizations were disrupted in F2 pups born to F1 parents developmentally exposed to BPA or EE, but their responses were not always identical, suggesting BPA might not activate estrogen receptors to the same extent as EE. Changes in vocalization patterns by F2 pups may be due to multigenerational exposure to BPA or EE and/or reduced parental care received. PMID:29912934

  8. The predictability of frequency-altered auditory feedback changes the weighting of feedback and feedforward input for speech motor control.

    PubMed

    Scheerer, Nichole E; Jones, Jeffery A

    2014-12-01

    Speech production requires the combined effort of a feedback control system driven by sensory feedback, and a feedforward control system driven by internal models. However, the factors that dictate the relative weighting of these feedback and feedforward control systems are unclear. In this event-related potential (ERP) study, participants produced vocalisations while being exposed to blocks of frequency-altered feedback (FAF) perturbations that were either predictable in magnitude (consistently either 50 or 100 cents) or unpredictable in magnitude (50- and 100-cent perturbations varying randomly within each vocalisation). Vocal and P1-N1-P2 ERP responses revealed decreases in the magnitude and trial-to-trial variability of vocal responses, smaller N1 amplitudes, and shorter vocal, P1 and N1 response latencies following predictable FAF perturbation magnitudes. In addition, vocal response magnitudes correlated with N1 amplitudes, vocal response latencies, and P2 latencies. This pattern of results suggests that after repeated exposure to predictable FAF perturbations, the contribution of the feedforward control system increases. Examination of the presentation order of the FAF perturbations revealed smaller compensatory responses, smaller P1 and P2 amplitudes, and shorter N1 latencies when the block of predictable 100-cent perturbations occurred prior to the block of predictable 50-cent perturbations. These results suggest that exposure to large perturbations modulates responses to subsequent perturbations of equal or smaller size. Similarly, exposure to a 100-cent perturbation prior to a 50-cent perturbation within a vocalisation decreased the magnitude of vocal and N1 responses, but increased P1 and P2 latencies. Thus, exposure to a single perturbation can affect responses to subsequent perturbations. © 2014 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.

  9. Control of Vocal and Respiratory Patterns in Birdsong: Dissection of Forebrain and Brainstem Mechanisms Using Temperature

    PubMed Central

    Fee, Michale S.

    2011-01-01

    Learned motor behaviors require descending forebrain control to be coordinated with midbrain and brainstem motor systems. In songbirds, such as the zebra finch, regular breathing is controlled by brainstem centers, but when the adult songbird begins to sing, its breathing becomes tightly coordinated with forebrain-controlled vocalizations. The periods of silence (gaps) between song syllables are typically filled with brief breaths, allowing the bird to sing uninterrupted for many seconds. While substantial progress has been made in identifying the brain areas and pathways involved in vocal and respiratory control, it is not understood how respiratory and vocal control is coordinated by forebrain motor circuits. Here we combine a recently developed technique for localized brain cooling, together with recordings of thoracic air sac pressure, to examine the role of cortical premotor nucleus HVC (proper name) in respiratory-vocal coordination. We found that HVC cooling, in addition to slowing all song timescales as previously reported, also increased the duration of expiratory pulses (EPs) and inspiratory pulses (IPs). Expiratory pulses, like song syllables, were stretched uniformly by HVC cooling, but most inspiratory pulses exhibited non-uniform stretch of pressure waveform such that the majority of stretch occurred late in the IP. Indeed, some IPs appeared to change duration by the earlier or later truncation of an underlying inspiratory event. These findings are consistent with the idea that during singing the temporal structure of EPs is under the direct control of forebrain circuits, whereas that of IPs can be strongly influenced by circuits downstream of HVC, likely in the brainstem. An analysis of the temporal jitter of respiratory and vocal structure suggests that IPs may be initiated by HVC at the end of each syllable and terminated by HVC immediately before the onset of the next syllable. PMID:21980466

  10. Study of human phonation in a full-body domain

    NASA Astrophysics Data System (ADS)

    Saurabh, Shakti; Bodony, Daniel

    2015-11-01

    The generation and propagation of the human voice is studied in two-dimensions using a full-body domain, using direct numerical simulation. The fluid/air in the vocal tract is modeled as a compressible and viscous fluid interacting with the non-linear, viscoelastic vocal folds (VF). The VF tissue material properties are multi-layered, with varying stiffness, and a finite-strain model is utilized and implemented in a quadratic finite element code. The fluid-solid domains are coupled through a boundary-fitted interface and utilize a Poisson equation-based mesh deformation method. The full-body domain includes the near VF region, the vocal tract, a simplified model of the soft palate and mouth, and extends out into the acoustic far-field. A new kind of inflow boundary condition based upon a quasi-one-dimensional formulation with constant sub-glottal volume velocity, which is linked to the VF movement, has been adopted. The sound pressure levels (SPL) measured are realistic and we analyze their connection to the VF dynamics and glottal and vocal tract geometries. Supported by the National Science Foundation (CAREER award number 1150439).

  11. An immersed-boundary method for flow–structure interaction in biological systems with application to phonation

    PubMed Central

    Luo, Haoxiang; Mittal, Rajat; Zheng, Xudong; Bielamowicz, Steven A.; Walsh, Raymond J.; Hahn, James K.

    2008-01-01

    A new numerical approach for modeling a class of flow–structure interaction problems typically encountered in biological systems is presented. In this approach, a previously developed, sharp-interface, immersed-boundary method for incompressible flows is used to model the fluid flow and a new, sharp-interface Cartesian grid, immersed boundary method is devised to solve the equations of linear viscoelasticity that governs the solid. The two solvers are coupled to model flow–structure interaction. This coupled solver has the advantage of simple grid generation and efficient computation on simple, single-block structured grids. The accuracy of the solid-mechanics solver is examined by applying it to a canonical problem. The solution methodology is then applied to the problem of laryngeal aerodynamics and vocal fold vibration during human phonation. This includes a three-dimensional eigen analysis for a multi-layered vocal fold prototype as well as two-dimensional, flow-induced vocal fold vibration in a modeled larynx. Several salient features of the aerodynamics as well as vocal-fold dynamics are presented. PMID:19936017

  12. Self-Organization: Complex Dynamical Systems in the Evolution of Speech

    NASA Astrophysics Data System (ADS)

    Oudeyer, Pierre-Yves

    Human vocalization systems are characterized by complex structural properties. They are combinatorial, based on the systematic reuse of phonemes, and the set of repertoires in human languages is characterized by both strong statistical regularities—universals—and a great diversity. Besides, they are conventional codes culturally shared in each community of speakers. What are the origins of the forms of speech? What are the mechanisms that permitted their evolution in the course of phylogenesis and cultural evolution? How can a shared speech code be formed in a community of individuals? This chapter focuses on the way the concept of self-organization, and its interaction with natural selection, can throw light on these three questions. In particular, a computational model is presented which shows that a basic neural equipment for adaptive holistic vocal imitation, coupling directly motor and perceptual representations in the brain, can generate spontaneously shared combinatorial systems of vocalizations in a society of babbling individuals. Furthermore, we show how morphological and physiological innate constraints can interact with these self-organized mechanisms to account for both the formation of statistical regularities and diversity in vocalization systems.

  13. The neural dynamics of song syntax in songbirds

    NASA Astrophysics Data System (ADS)

    Jin, Dezhe

    2010-03-01

    Songbird is ``the hydrogen atom'' of the neuroscience of complex, learned vocalizations such as human speech. Songs of Bengalese finch consist of sequences of syllables. While syllables are temporally stereotypical, syllable sequences can vary and follow complex, probabilistic syntactic rules, which are rudimentarily similar to grammars in human language. Songbird brain is accessible to experimental probes, and is understood well enough to construct biologically constrained, predictive computational models. In this talk, I will discuss the structure and dynamics of neural networks underlying the stereotypy of the birdsong syllables and the flexibility of syllable sequences. Recent experiments and computational models suggest that a syllable is encoded in a chain network of projection neurons in premotor nucleus HVC (proper name). Precisely timed spikes propagate along the chain, driving vocalization of the syllable through downstream nuclei. Through a computational model, I show that that variable syllable sequences can be generated through spike propagations in a network in HVC in which the syllable-encoding chain networks are connected into a branching chain pattern. The neurons mutually inhibit each other through the inhibitory HVC interneurons, and are driven by external inputs from nuclei upstream of HVC. At a branching point that connects the final group of a chain to the first groups of several chains, the spike activity selects one branch to continue the propagation. The selection is probabilistic, and is due to the winner-take-all mechanism mediated by the inhibition and noise. The model predicts that the syllable sequences statistically follow partially observable Markov models. Experimental results supporting this and other predictions of the model will be presented. We suggest that the syntax of birdsong syllable sequences is embedded in the connection patterns of HVC projection neurons.

  14. The Referent of Children's Early Songs

    ERIC Educational Resources Information Center

    Mang, Esther

    2005-01-01

    Musical creativity during early childhood is readily exemplified in vocal behaviours. This paper is a discussion of observations on children's performance of learned songs and self-generated songs. Longitudinal observations suggest that self-generated songs may be seen as referent-guided improvisation using source materials derived from learned…

  15. A corollary discharge maintains auditory sensitivity during sound production

    NASA Astrophysics Data System (ADS)

    Poulet, James F. A.; Hedwig, Berthold

    2002-08-01

    Speaking and singing present the auditory system of the caller with two fundamental problems: discriminating between self-generated and external auditory signals and preventing desensitization. In humans and many other vertebrates, auditory neurons in the brain are inhibited during vocalization but little is known about the nature of the inhibition. Here we show, using intracellular recordings of auditory neurons in the singing cricket, that presynaptic inhibition of auditory afferents and postsynaptic inhibition of an identified auditory interneuron occur in phase with the song pattern. Presynaptic and postsynaptic inhibition persist in a fictively singing, isolated cricket central nervous system and are therefore the result of a corollary discharge from the singing motor network. Mimicking inhibition in the interneuron by injecting hyperpolarizing current suppresses its spiking response to a 100-dB sound pressure level (SPL) acoustic stimulus and maintains its response to subsequent, quieter stimuli. Inhibition by the corollary discharge reduces the neural response to self-generated sound and protects the cricket's auditory pathway from self-induced desensitization.

  16. [A comparative acoustic study of the speaking and singing voice during the adolescent's break of the voice].

    PubMed

    Amy de la Bretèque, B; Sanchez, S

    2000-01-01

    The observation of the vocal evolution of adolescent singers has shown it takes place in two stages, the singing voice changing after the speaking voice. The same pattern has been encountered and made more explicit with a study of 50 non-singer adolescents. It thus appears that the average pitch of the speaking voice deepening by one octave is not by itself the sign that the break of the voice has ended. This study also shows the individual nature of adolescent vocal evolution and its length (up to two years in one out of four cases).

  17. Absence of deficits in social behaviors and ultrasonic vocalizations in later generations of mice lacking neuroligin4.

    PubMed

    Ey, E; Yang, M; Katz, A M; Woldeyohannes, L; Silverman, J L; Leblond, C S; Faure, P; Torquet, N; Le Sourd, A-M; Bourgeron, T; Crawley, J N

    2012-11-01

    Mutations in NLGN4X have been identified in individuals with autism spectrum disorders and other neurodevelopmental disorders. A previous study reported that adult male mice lacking neuroligin4 (Nlgn4) displayed social approach deficits in the three-chambered test, altered aggressive behaviors and reduced ultrasonic vocalizations. To replicate and extend these findings, independent comprehensive analyses of autism-relevant behavioral phenotypes were conducted in later generations of the same line of Nlgn4 mutant mice at the National Institute of Mental Health in Bethesda, MD, USA and at the Institut Pasteur in Paris, France. Adult social approach was normal in all three genotypes of Nlgn4 mice tested at both sites. Reciprocal social interactions in juveniles were similarly normal across genotypes. No genotype differences were detected in ultrasonic vocalizations in pups separated from the nest or in adults during reciprocal social interactions. Anxiety-like behaviors, self-grooming, rotarod and open field exploration did not differ across genotypes, and measures of developmental milestones and general health were normal. Our findings indicate an absence of autism-relevant behavioral phenotypes in subsequent generations of Nlgn4 mice tested at two locations. Testing environment and methods differed from the original study in some aspects, although the presence of normal sociability was seen in all genotypes when methods taken from Jamain et al. (2008) were used. The divergent results obtained from this study indicate that phenotypes may not be replicable across breeding generations, and highlight the significant roles of environmental, generational and/or procedural factors on behavioral phenotypes. Published 2012. This article is a U.S. Government work and is in the public domain in the USA.

  18. Ontogenetic development of the inner ear saccule and utricle in the Lusitanian toadfish: Potential implications for auditory sensitivity.

    PubMed

    Chaves, Patrícia P; Valdoria, Ciara M C; Amorim, M Clara P; Vasconcelos, Raquel O

    2017-09-01

    Studies addressing structure-function relationships of the fish auditory system during development are sparse compared to other taxa. The Batrachoididae has become an important group to investigate mechanisms of auditory plasticity and evolution of auditory-vocal systems. A recent study reported ontogenetic improvements in the inner ear saccule sensitivity of the Lusitanian toadfish, Halobatrachus didactylus, but whether this results from changes in the sensory morphology remains unknown. We investigated how the macula and organization of auditory receptors in the saccule and utricle change during growth in this species. Inner ear sensory epithelia were removed from the end organs of previously PFA-fixed specimens, from non-vocal posthatch fry (<1.4 cm, standard length) to adults (>23 cm). Epithelia were phalloidin-stained and analysed for area, shape, number and orientation patterns of hair cells (HC), and number and size of saccular supporting cells (SC). Saccular macula area expanded 41x in total, and significantly more (relative to body length) among vocal juveniles (2.3-2.9 cm). Saccular HC number increased 25x but HC density decreased, suggesting that HC addition is slower relative to epithelial growth. While SC density decreased, SC apical area increased, contributing to the epithelial expansion. The utricule revealed increased HC density (striolar region) and less epithelial expansion (5x) with growth, contrasting with the saccule that may have a different developmental pattern due to its larger size and main auditory functions. Both macula shape and HC orientation patterns were already established in the posthatch fry and retained throughout growth in both end organs. We suggest that previously reported ontogenetic improvements in saccular sensitivity might be associated with changes in HC number (not density), size and/or molecular mechanisms controlling HC sensitivity. This is one of the first studies investigating the ontogenetic development of the saccule and utricle in a vocal fish and how it potentially relates to auditory enhancement for acoustic communication. Copyright © 2017 Elsevier B.V. All rights reserved.

  19. Experimental study of the effects of surface mucus viscosity on the glottic cycle.

    PubMed

    Ayache, Stéphane; Ouaknine, Maurice; Dejonkere, Philippe; Prindere, Pierre; Giovanni, Antoine

    2004-03-01

    Numerous clinical findings indicate that viscosity of laryngeal mucosa is a crucial factor in glottal perfomance. Experience using experimental test benches has shown the importance of humidifying air stream used to induce vibration in excised larynges. Nevertheless, there is a lack of knowledge particularly regarding the physicochemical properties of laryngeal mucus. The purpose of this study was to research vocal fold vibration in excised larynges using artificial mucus of precisely known viscosity. Eight freshly harvested porcine larynges were examined. Parameters measured were Fo and vocal fold contact time. Measurements were performed under three conditions: basal (no fluid application on vocal cord surface), after application of a fluid of 60cP viscosity (Visc60), and after application of a fluid of 100cP viscosity (Visc100). Electroglottographic measurements were performed at two different times for each condition: 1 s after airflow onset (T1) and 6 seconds after airflow onset (T2). Statistical analysis consisted of comparing data obtained under each condition at T1 and T2. The results showed a significant decrease in Fo after application of Visc60 and Visc100 fluids and a decrease in Fo at T2. Closure time was significantly higher under Visc60 conditions and under Visc100 conditions than under basal conditions. Application of artificial mucus to the mucosa of the vocal folds lowered vibratory frequency and prolonged the contact phase. Our interpretation of this data is that the presence of mucus on the surface of the vocal folds generated superficial tension and caused adhesion, which is a source of nonlinearity in vocal vibration.

  20. ANALYSIS OF FLOW-STRUCTURE COUPLING IN A MECHANICAL MODEL OF THE VOCAL FOLDS AND THE SUBGLOTTAL SYSTEM.

    PubMed

    Howe, M S; McGowan, R S

    2009-11-01

    An analysis is made of the nonlinear interactions between flow in the subglottal vocal tract and glottis, sound waves in the subglottal system and a mechanical model of the vocal folds. The mean flow through the system is produced by a nominally steady contraction of the lungs, and mechanical experiments frequently involve a 'lung cavity' coupled to an experimental subglottal tube of arbitrary or ill-defined effective length L, on the basis that the actual value of L has little or no influence on excitation of the vocal folds. A simple, self-exciting single mass mathematical model of the vocal folds is used to investigate the sound generated within the subglottal domain and the unsteady volume flux from the glottis for experiments where it is required to suppress feedback of sound from the supraglottal vocal tract. In experiments where the assumed absorption of sound within the sponge-like interior of the lungs is small, the influence of changes in L can be very significant: when the subglottal tube behaves as an open-ended resonator (when L is as large as half the acoustic wavelength) there is predicted to be a mild increase in volume flux magnitude and a small change in waveform. However, the strong appearance of second harmonics of the acoustic field is predicted at intermediate lengths, when L is roughly one quarter of the acoustic wavelength. In cases of large lung damping, however, only modest changes in the volume flux are predicted to occur with variations in L.

  1. Perceptions of Voice Teachers Regarding Students' Vocal Behaviors During Singing and Speaking.

    PubMed

    Beeman, Shellie A

    2017-01-01

    This study examined voice teachers' perceptions of their instruction of healthy singing and speaking voice techniques. An online, researcher-generated questionnaire based on the McClosky technique was administered to college/university voice teachers listed as members in the 2012-2013 College Music Society directory. A majority of participants believed there to be a relationship between the health of the singing voice and the health of the speaking voice. Participants' perception scores were the most positive for variable MBSi, the monitoring of students' vocal behaviors during singing. Perception scores for variable TVB, the teaching of healthy vocal behaviors, and variable MBSp, the monitoring of students' vocal behaviors while speaking, ranked second and third, respectively. Perception scores for variable TVB were primarily associated with participants' familiarity with voice rehabilitation techniques, gender, and familiarity with the McClosky technique. Perception scores for variable MBSi were primarily associated with participants' familiarity with voice rehabilitation techniques, gender, type of student taught, and instruction of a student with a voice disorder. Perception scores for variable MBSp were correlated with the greatest number of characteristics, including participants' familiarity with voice rehabilitation techniques, familiarity with the McClosky technique, type of student taught, years of teaching experience, and instruction of a student with a voice disorder. Voice teachers are purportedly working with injured voices and attempting to include vocal health in their instruction. Although a voice teacher is not obligated to pursue further rehabilitative training, the current study revealed a positive relationship between familiarity with specific rehabilitation techniques and vocal health. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  2. Social, contextual, and individual factors affecting the occurrence and acoustic structure of drumming bouts in wild chimpanzees (Pan troglodytes).

    PubMed

    Babiszewska, Magdalena; Schel, Anne Marijke; Wilke, Claudia; Slocombe, Katie E

    2015-01-01

    The production of structured and repetitive sounds by striking objects is a behavior found not only in humans, but also in a variety of animal species, including chimpanzees (Pan troglodytes). In this study we examined individual and social factors that may influence the frequency with which individuals engage in drumming behavior when producing long distance pant hoot vocalizations, and analyzed the temporal structure of those drumming bouts. Male chimpanzees from Budongo Forest, Uganda, drummed significantly more frequently during travel than feeding or resting and older individuals were significantly more likely to produce drumming bouts than younger ones. In contrast, we found no evidence that the presence of estrus females, high ranking males and preferred social partners in the caller's vicinty had an effect on the frequency with which an individual accompanied their pant hoot vocalization with drumming. Through acoustic analyses, we demonstrated that drumming sequences produced with pant hoots may have contained information on individual identity and that qualitatively, there was individual variation in the complexity of the temporal patterns produced. We conclude that drumming patterns may act as individually distinctive long-distance signals that, together with pant hoot vocalizations, function to coordinate the movement and spacing of dispersed individuals within a community, rather than as signals to group members in the immediate audience. © 2014 Wiley Periodicals, Inc.

  3. Whistle sequences in wild killer whales (Orcinus orca).

    PubMed

    Riesch, Rüdiger; Ford, John K B; Thomsen, Frank

    2008-09-01

    Combining different stereotyped vocal signals into specific sequences increases the range of information that can be transferred between individuals. The temporal emission pattern and the behavioral context of vocal sequences have been described in detail for a variety of birds and mammals. Yet, in cetaceans, the study of vocal sequences is just in its infancy. Here, we provide a detailed analysis of sequences of stereotyped whistles in killer whales off Vancouver Island, British Columbia. A total of 1140 whistle transitions in 192 whistle sequences recorded from resident killer whales were analyzed using common spectrographic analysis techniques. In addition to the stereotyped whistles described by Riesch et al., [(2006). "Stability and group specificity of stereotyped whistles in resident killer whales, Orcinus orca, off British Columbia," Anim. Behav. 71, 79-91.] We found a new and rare stereotyped whistle (W7) as well as two whistle elements, which are closely linked to whistle sequences: (1) stammers and (2) bridge elements. Furthermore, the frequency of occurrence of 12 different stereotyped whistle types within the sequences was not randomly distributed and the transition patterns between whistles were also nonrandom. Finally, whistle sequences were closely tied to close-range behavioral interactions (in particular among males). Hence, we conclude that whistle sequences in wild killer whales are complex signal series and propose that they are most likely emitted by single individuals.

  4. Analysis Of Laryngeal Biomechanics Of Deaf Speakers Utilizing High-Speed Cinematography

    NASA Astrophysics Data System (ADS)

    Metz, Dale E.; Whitehead, Robert L.

    1982-02-01

    Since the formalization of the myoelastic-aerodynamic theory of vocal fold vibration, it has been generally accepted that biomechanical and aerodynamic forces determine the nature of vocal fold vibration patterns, speaking fundamental frequency and vocal intensity. The speech of the deaf is frequently characterized by abnormal voice qualities and aberrant frequency and intensity variations suggesting mismanagement of the biomechanical and aerodynamic forces acting on the larynx. Unfortunately, efforts to remediate these abnormal laryngeal activities are frequently ineffective. It is reasonable to suggest that more effective remedial strategies could be developed if we had a better understanding of the underlying nature of the problems deaf persons experience when trying to control laryngeal functioning for speech purposes. Toward this end, we are employing high speed laryngeal filming procedures in conjunction with glottal impedance, respiratory kinematic and acous-tical measurement procedures to assess abnormal laryngeal functioning of deaf speakers. All data are collected simultaneously and are time-locked to facilitate analysis of specific laryngeal events. This unique combination of instrumentation has provided important insights regarding laryngeal functioning of the deaf. For example, we have observed that deaf speakers may assume abnormal glottal configurations during phonation that pro-hibit normal laryngeal functioning and disturb upper airway dynamics. Also, normal vibratory patterns are frequently disturbed. Instrumentation, data collection protocols, analysis procedures and selected findings will be discussed.

  5. Evidence for cultural dialects in vocal emotion expression: acoustic classification within and across five nations.

    PubMed

    Laukka, Petri; Neiberg, Daniel; Elfenbein, Hillary Anger

    2014-06-01

    The possibility of cultural differences in the fundamental acoustic patterns used to express emotion through the voice is an unanswered question central to the larger debate about the universality versus cultural specificity of emotion. This study used emotionally inflected standard-content speech segments expressing 11 emotions produced by 100 professional actors from 5 English-speaking cultures. Machine learning simulations were employed to classify expressions based on their acoustic features, using conditions where training and testing were conducted on stimuli coming from either the same or different cultures. A wide range of emotions were classified with above-chance accuracy in cross-cultural conditions, suggesting vocal expressions share important characteristics across cultures. However, classification showed an in-group advantage with higher accuracy in within- versus cross-cultural conditions. This finding demonstrates cultural differences in expressive vocal style, and supports the dialect theory of emotions according to which greater recognition of expressions from in-group members results from greater familiarity with culturally specific expressive styles.

  6. Species-specific calls evoke asymmetric activity in the monkey's temporal poles.

    PubMed

    Poremba, Amy; Malloy, Megan; Saunders, Richard C; Carson, Richard E; Herscovitch, Peter; Mishkin, Mortimer

    2004-01-29

    It has often been proposed that the vocal calls of monkeys are precursors of human speech, in part because they provide critical information to other members of the species who rely on them for survival and social interactions. Both behavioural and lesion studies suggest that monkeys, like humans, use the auditory system of the left hemisphere preferentially to process vocalizations. To investigate the pattern of neural activity that might underlie this particular form of functional asymmetry in monkeys, we measured local cerebral metabolic activity while the animals listened passively to species-specific calls compared with a variety of other classes of sound. Within the superior temporal gyrus, significantly greater metabolic activity occurred on the left side than on the right, only in the region of the temporal pole and only in response to monkey calls. This functional asymmetry was absent when these regions were separated by forebrain commissurotomy, suggesting that the perception of vocalizations elicits concurrent interhemispheric interactions that focus the auditory processing within a specialized area of one hemisphere.

  7. Proteomic analysis of a decellularized human vocal fold mucosa scaffold using 2D electrophoresis and high-resolution mass spectrometry.

    PubMed

    Welham, Nathan V; Chang, Zhen; Smith, Lloyd M; Frey, Brian L

    2013-01-01

    Natural biologic scaffolds for tissue engineering are commonly generated by decellularization of tissues and organs. Despite some preclinical and clinical success, in vivo scaffold remodeling and functional outcomes remain variable, presumably due to the influence of unidentified bioactive molecules on the scaffold-host interaction. Here, we used 2D electrophoresis and high-resolution mass spectrometry-based proteomic analyses to evaluate decellularization effectiveness and identify potentially bioactive protein remnants in a human vocal fold mucosa model. We noted proteome, phosphoproteome and O-glycoproteome depletion post-decellularization, and identified >200 unique protein species within the decellularized scaffold. Gene ontology-based enrichment analysis revealed a dominant set of functionally-related ontology terms associated with extracellular matrix assembly, organization, morphology and patterning, consistent with preservation of a tissue-specific niche for later cell seeding and infiltration. We further identified a subset of ontology terms associated with bioactive (some of which are antigenic) cellular proteins, despite histological and immunohistochemical data indicating complete decellularization. These findings demonstrate the value of mass spectrometry-based proteomics in identifying agents potentially responsible for variation in host response to engineered tissues derived from decellularized scaffolds. This work has implications for the manufacturing of biologic scaffolds from any tissue or organ, as well as for prediction and monitoring of the scaffold-host interaction in vivo. Copyright © 2012 Elsevier Ltd. All rights reserved.

  8. Influence of Asymmetric Recurrent Laryngeal Nerve Stimulation on Vibration, Acoustics, and Aerodynamics

    PubMed Central

    Chhetri, Dinesh K.; Neubauer, Juergen; Sofer, Elazar

    2015-01-01

    Objectives/Hypothesis Evaluate the influence of asymmetric recurrent laryngeal nerve (RLN) stimulation on the vibratory phase, acoustics and aerodynamics of phonation. Study Design Basic science study using an in vivo canine model. Methods The RLNs were symmetrically and asymmetrically stimulated over eight graded levels to test a range of vocal fold activation conditions from subtle paresis to paralysis. Vibratory phase, fundamental frequency (F0), subglottal pressure, and airflow were noted at phonation onset. The evaluations were repeated for three levels of symmetric superior laryngeal nerve (SLN) stimulation. Results Asymmetric laryngeal adductor activation from asymmetric left-right RLN stimulation led to a consistent pattern of vibratory phase asymmetry, with the more activated vocal fold leading in the opening phase of the glottal cycle and in mucosal wave amplitude. Vibratory amplitude asymmetry was also observed, with more lateral excursion of the glottis of the less activated side. Onset fundamental frequency was higher with asymmetric activation because the two RLNs were synergistic in decreasing F0, glottal width, and strain. Phonation onset pressure increased and airflow decreased with symmetric RLN activation. Conclusion Asymmetric laryngeal activation from RLN paresis and paralysis has consistent effects on vocal fold vibration, acoustics, and aerodynamics. This information may be useful in diagnosis and management of vocal fold paresis. PMID:24913182

  9. Influence of asymmetric recurrent laryngeal nerve stimulation on vibration, acoustics, and aerodynamics.

    PubMed

    Chhetri, Dinesh K; Neubauer, Juergen; Sofer, Elazar

    2014-11-01

    Evaluate the influence of asymmetric recurrent laryngeal nerve (RLN) stimulation on the vibratory phase, acoustics and aerodynamics of phonation. Basic science study using an in vivo canine model. The RLNs were symmetrically and asymmetrically stimulated over eight graded levels to test a range of vocal fold activation conditions from subtle paresis to paralysis. Vibratory phase, fundamental frequency (F0 ), subglottal pressure, and airflow were noted at phonation onset. The evaluations were repeated for three levels of symmetric superior laryngeal nerve (SLN) stimulation. Asymmetric laryngeal adductor activation from asymmetric left-right RLN stimulation led to a consistent pattern of vibratory phase asymmetry, with the more activated vocal fold leading in the opening phase of the glottal cycle and in mucosal wave amplitude. Vibratory amplitude asymmetry was also observed, with more lateral excursion of the glottis of the less activated side. Onset fundamental frequency was higher with asymmetric activation because the two RLNs were synergistic in decreasing F0 , glottal width, and strain. Phonation onset pressure increased and airflow decreased with symmetric RLN activation. Asymmetric laryngeal activation from RLN paresis and paralysis has consistent effects on vocal fold vibration, acoustics, and aerodynamics. This information may be useful in diagnosis and management of vocal fold paresis. N/A. © 2014 The American Laryngological, Rhinological and Otological Society, Inc.

  10. Underwater audiogram of the California sea lion by the conditioned vocalization technique1

    PubMed Central

    Schusterman, Ronald J.; Balliet, Richard F.; Nixon, James

    1972-01-01

    Conditioning techniques were developed demonstrating that pure tone frequencies under water can exert nearly perfect control over the underwater click vocalizations of the California sea lion (Zalophus californianus). Conditioned vocalizations proved to be a reliable way of obtaining underwater sound detection thresholds in Zalophus at 13 different frequencies, covering a frequency range of 250 to 64,000 Hz. The audiogram generated by these threshold measurements suggests that under water, the range of maximal sensitivity for Zalophus lies between one and 28 kHz with best sensitivity at 16 kHz. Between 28 and 36 kHz there is a loss in sensitivity of 60 dB/octave. However, with relatively intense acoustic signals (> 38 dB re 1 μb underwater), Zalophus will respond to frequencies at least as high as 192 kHz. These results are compared with the underwater hearing of other marine mammals. ImagesFig. 1. PMID:5033891

  11. The impact of extended voice use on the acoustic characteristics of phonation after training and performance of actors from the La MaMa Experimental Theater club.

    PubMed

    Ferrone, Carol; Galgano, Jessica; Ramig, Lorraine Olson

    2011-05-01

    To test the hypothesis that extensive use of La MaMa vocal technique may result in symptoms of vocal abuse, an evaluation of the acoustic and perceptual characteristics of voice for eight performers from the Great Jones Repertory Company of the La MaMa Experimental Theater was conducted. This vocal technique includes wide ranges of frequency from 46 to 2003 Hz and vocal intensity that is sustained at 90-108 dB sound pressure level with a mouth-to-microphone distance of 30 cm for 3-4 hours per performance. The actors rehearsed for 4 hours per day, 5 days per week for 14 weeks before the series of performances. Thirty-nine performances were presented in 6 weeks. Three pretraining, three posttraining, and two postperformance series data collection sessions were carried out for each performer. Speech samples were gathered using the CSL 4500 and analyzed using Real-Time Pitch program and Multidimensional Voice Program. Acoustic analysis was performed on 48 tokens of sustained vowel phonation for each subject. Statistical analysis was performed using the Friedman test of related samples. Perceptual analysis included professional listeners rating voice quality in pretraining, posttraining, and postperformance samples of the Rainbow Passage and sample lines from the plays. The majority of professional listeners (11/12) judged that this technique would result in symptoms of vocal abuse; however, acoustic data revealed statistically stable or improved measurements for all subjects in most dependent acoustic variables when compared with both posttraining and postperformance trials. These findings add support to the notion that a technique that may be perceived as vocally abusive, generating 90-100 dB sound pressure level and sustained over 6 weeks of performances, actually resulted in improved vocal strength and flexibility. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  12. Prefrontal Neuronal Responses during Audiovisual Mnemonic Processing

    PubMed Central

    Hwang, Jaewon

    2015-01-01

    During communication we combine auditory and visual information. Neurophysiological research in nonhuman primates has shown that single neurons in ventrolateral prefrontal cortex (VLPFC) exhibit multisensory responses to faces and vocalizations presented simultaneously. However, whether VLPFC is also involved in maintaining those communication stimuli in working memory or combining stored information across different modalities is unknown, although its human homolog, the inferior frontal gyrus, is known to be important in integrating verbal information from auditory and visual working memory. To address this question, we recorded from VLPFC while rhesus macaques (Macaca mulatta) performed an audiovisual working memory task. Unlike traditional match-to-sample/nonmatch-to-sample paradigms, which use unimodal memoranda, our nonmatch-to-sample task used dynamic movies consisting of both facial gestures and the accompanying vocalizations. For the nonmatch conditions, a change in the auditory component (vocalization), the visual component (face), or both components was detected. Our results show that VLPFC neurons are activated by stimulus and task factors: while some neurons simply responded to a particular face or a vocalization regardless of the task period, others exhibited activity patterns typically related to working memory such as sustained delay activity and match enhancement/suppression. In addition, we found neurons that detected the component change during the nonmatch period. Interestingly, some of these neurons were sensitive to the change of both components and therefore combined information from auditory and visual working memory. These results suggest that VLPFC is not only involved in the perceptual processing of faces and vocalizations but also in their mnemonic processing. PMID:25609614

  13. Subglottal Impedance-Based Inverse Filtering of Voiced Sounds Using Neck Surface Acceleration

    PubMed Central

    Zañartu, Matías; Ho, Julio C.; Mehta, Daryush D.; Hillman, Robert E.; Wodicka, George R.

    2014-01-01

    A model-based inverse filtering scheme is proposed for an accurate, non-invasive estimation of the aerodynamic source of voiced sounds at the glottis. The approach, referred to as subglottal impedance-based inverse filtering (IBIF), takes as input the signal from a lightweight accelerometer placed on the skin over the extrathoracic trachea and yields estimates of glottal airflow and its time derivative, offering important advantages over traditional methods that deal with the supraglottal vocal tract. The proposed scheme is based on mechano-acoustic impedance representations from a physiologically-based transmission line model and a lumped skin surface representation. A subject-specific calibration protocol is used to account for individual adjustments of subglottal impedance parameters and mechanical properties of the skin. Preliminary results for sustained vowels with various voice qualities show that the subglottal IBIF scheme yields comparable estimates with respect to current aerodynamics-based methods of clinical vocal assessment. A mean absolute error of less than 10% was observed for two glottal airflow measures –maximum flow declination rate and amplitude of the modulation component– that have been associated with the pathophysiology of some common voice disorders caused by faulty and/or abusive patterns of vocal behavior (i.e., vocal hyperfunction). The proposed method further advances the ambulatory assessment of vocal function based on the neck acceleration signal, that previously have been limited to the estimation of phonation duration, loudness, and pitch. Subglottal IBIF is also suitable for other ambulatory applications in speech communication, in which further evaluation is underway. PMID:25400531

  14. Experience-dependent modulation of feedback integration during singing: role of the right anterior insula.

    PubMed

    Kleber, Boris; Zeitouni, Anthony G; Friberg, Anders; Zatorre, Robert J

    2013-04-03

    Somatosensation plays an important role in the motor control of vocal functions, yet its neural correlate and relation to vocal learning is not well understood. We used fMRI in 17 trained singers and 12 nonsingers to study the effects of vocal-fold anesthesia on the vocal-motor singing network as a function of singing expertise. Tasks required participants to sing musical target intervals under normal conditions and after anesthesia. At the behavioral level, anesthesia altered pitch accuracy in both groups, but singers were less affected than nonsingers, indicating an experience-dependent effect of the intervention. At the neural level, this difference was accompanied by distinct patterns of decreased activation in singers (cortical and subcortical sensory and motor areas) and nonsingers (subcortical motor areas only) respectively, suggesting that anesthesia affected the higher-level voluntary (explicit) motor and sensorimotor integration network more in experienced singers, and the lower-level (implicit) subcortical motor loops in nonsingers. The right anterior insular cortex (AIC) was identified as the principal area dissociating the effect of expertise as a function of anesthesia by three separate sources of evidence. First, it responded differently to anesthesia in singers (decreased activation) and nonsingers (increased activation). Second, functional connectivity between AIC and bilateral A1, M1, and S1 was reduced in singers but augmented in nonsingers. Third, increased BOLD activity in right AIC in singers was correlated with larger pitch deviation under anesthesia. We conclude that the right AIC and sensory-motor areas play a role in experience-dependent modulation of feedback integration for vocal motor control during singing.

  15. [The prevention of voice disorders in the actor: protocol and follow-up nine months of professional theater].

    PubMed

    Ormezzano, Y; Delale, A; Lamy-Simonian, A

    2011-01-01

    In July 2009, at the beginning of this work, 26 theses addressing professional principles of voice were listed in the database of SUDOC (Système Universitaire de Documentation): 9 related to voices of teachers (about 900,000* professionals in France), 14 theses relating to singers (7500** professionals), and only 3 about the voice of actors (20 000*** professional actors in France in 2006). The latter pertaining to concerning rookie actors (sensibilisation vocale auprès du comédien débutant Bichet, Linda, Bordeaux II, 2006), the mechanical larynx (étude des mécanismes laryngés dans la voix projetée: cas particulier des comédiennes Guerin, Mélanie, Paris VI, 2009), vocal fatigue (Fatigue vocale après une tâche d'utilisation prolongée de la voix chez le comédien Canaan Baggioni, Brigitte, Aix-Marseille II, 2009). Professional actors are plentiful; their training in vocal technique is very heterogeneous, or non-existent: it is not a prerequisite to have a degree to work as an actor! This lack of vocal technique is associated with risk factors specific to the acting profession: numerous travels in air-conditioned vehicles, unsuitable workplaces; dusty or poorly heated, irregular working patterns, excessive demands from directors... All this makes the actors highly susceptible to voice disorders. The protocol for the prevention of voice disorders presented here is holistic and ecological. This work also examines the effectiveness of such a preventive protocol aimed at theatre comedians.

  16. Species-specific loss of sexual dimorphism in vocal effectors accompanies vocal simplification in African clawed frogs (Xenopus)

    PubMed Central

    Leininger, Elizabeth C.; Kitayama, Ken; Kelley, Darcy B.

    2015-01-01

    ABSTRACT Phylogenetic studies can reveal patterns of evolutionary change, including the gain or loss of elaborate courtship traits in males. Male African clawed frogs generally produce complex and rapid courtship vocalizations, whereas female calls are simple and slow. In a few species, however, male vocalizations are also simple and slow, suggesting loss of male-typical traits. Here, we explore features of the male vocal organ that could contribute to loss in two species with simple, slow male calls. In Xenopus boumbaensis, laryngeal morphology is more robust in males than in females. Larynges are larger, have a more complex cartilaginous morphology and contain more muscle fibers. Laryngeal muscle fibers are exclusively fast-twitch in males but are both fast- and slow-twitch in females. The laryngeal electromyogram, a measure of neuromuscular synaptic strength, shows greater potentiation in males than in females. Male-specific physiological features are shared with X. laevis, as well as with a species of the sister clade, Silurana tropicalis, and thus are likely ancestral. In X. borealis, certain aspects of laryngeal morphology and physiology are sexually monomorphic rather than dimorphic. In both sexes, laryngeal muscle fibers are of mixed-twitch type, which limits the production of muscle contractions at rapid intervals. Muscle activity potentiation and discrete tension transients resemble female rather than male X. boumbaensis. The de-masculinization of these laryngeal features suggests an alteration in sensitivity to the gonadal hormones that are known to control the sexual differentiation of the larynx in other Xenopus and Silurana species. PMID:25788725

  17. Acoustic characteristics of phonation in "wet voice" conditions.

    PubMed

    Murugappan, Shanmugam; Boyce, Suzanne; Khosla, Sid; Kelchner, Lisa; Gutmark, Ephraim

    2010-04-01

    A perceptible change in phonation characteristics after a swallow has long been considered evidence that food and/or drink material has entered the laryngeal vestibule and is on the surface of the vocal folds as they vibrate. The current paper investigates the acoustic characteristics of phonation when liquid material is present on the vocal folds, using ex vivo porcine larynges as a model. Consistent with instrumental examinations of swallowing disorders or dysphagia in humans, three liquids of different Varibar viscosity ("thin liquid," "nectar," and "honey") were studied at constant volume. The presence of materials on the folds during phonation was generally found to suppress the higher frequency harmonics and generate intermittent additional frequencies in the low and high end of the acoustic spectrum. Perturbation measures showed a higher percentage of jitter and shimmer when liquid material was present on the folds during phonation, but they were unable to differentiate statistically between the three fluid conditions. The finite correlation dimension and positive Lyapunov exponent measures indicated that the presence of materials on the vocal folds excited a chaotic system. Further, these measures were able to reliably differentiate between the baseline and different types of liquid on the vocal folds.

  18. Acoustic characteristics of phonation in “wet voice” conditions

    PubMed Central

    Murugappan, Shanmugam; Boyce, Suzanne; Khosla, Sid; Kelchner, Lisa; Gutmark, Ephraim

    2010-01-01

    A perceptible change in phonation characteristics after a swallow has long been considered evidence that food and∕or drink material has entered the laryngeal vestibule and is on the surface of the vocal folds as they vibrate. The current paper investigates the acoustic characteristics of phonation when liquid material is present on the vocal folds, using ex vivo porcine larynges as a model. Consistent with instrumental examinations of swallowing disorders or dysphagia in humans, three liquids of different Varibar viscosity (“thin liquid,” “nectar,” and “honey”) were studied at constant volume. The presence of materials on the folds during phonation was generally found to suppress the higher frequency harmonics and generate intermittent additional frequencies in the low and high end of the acoustic spectrum. Perturbation measures showed a higher percentage of jitter and shimmer when liquid material was present on the folds during phonation, but they were unable to differentiate statistically between the three fluid conditions. The finite correlation dimension and positive Lyapunov exponent measures indicated that the presence of materials on the vocal folds excited a chaotic system. Further, these measures were able to reliably differentiate between the baseline and different types of liquid on the vocal folds. PMID:20370039

  19. Automated extraction and classification of time-frequency contours in humpback vocalizations.

    PubMed

    Ou, Hui; Au, Whitlow W L; Zurk, Lisa M; Lammers, Marc O

    2013-01-01

    A time-frequency contour extraction and classification algorithm was created to analyze humpback whale vocalizations. The algorithm automatically extracted contours of whale vocalization units by searching for gray-level discontinuities in the spectrogram images. The unit-to-unit similarity was quantified by cross-correlating the contour lines. A library of distinctive humpback units was then generated by applying an unsupervised, cluster-based learning algorithm. The purpose of this study was to provide a fast and automated feature selection tool to describe the vocal signatures of animal groups. This approach could benefit a variety of applications such as species description, identification, and evolution of song structures. The algorithm was tested on humpback whale song data recorded at various locations in Hawaii from 2002 to 2003. Results presented in this paper showed low probability of false alarm (0%-4%) under noisy environments with small boat vessels and snapping shrimp. The classification algorithm was tested on a controlled set of 30 units forming six unit types, and all the units were correctly classified. In a case study on humpback data collected in the Auau Chanel, Hawaii, in 2002, the algorithm extracted 951 units, which were classified into 12 distinctive types.

  20. Laryngeal airway reconstruction indicates that rodent ultrasonic vocalizations are produced by an edge-tone mechanism

    PubMed Central

    Borgard, Heather L.

    2017-01-01

    Some rodents produce ultrasonic vocalizations (USVs) for social communication using an aerodynamic whistle, a unique vocal production mechanism not found in other animals. The functional anatomy and evolution of this sound production mechanism remains unclear. Using laryngeal airway reconstruction, we identified anatomical specializations critical for USV production. A robust laryngeal cartilaginous framework supports a narrow supraglottal airway. An intralaryngeal airsac-like cavity termed the ventral pouch was present in three muroid rodents (suborder Myomorpha), but was absent in a heteromyid rodent (suborder Castorimorpha) that produces a limited vocal repertoire and no documented USVs. Small lesions to the ventral pouch in laboratory rats caused dramatic changes in USV production, supporting the hypothesis that an interaction between a glottal exit jet and the alar edge generates ultrasonic signals in rodents. The resulting undulating airflow around the alar edge interacts with the resonance of the ventral pouch, which may function as a Helmholtz resonator. The proposed edge-tone mechanism requires control of intrinsic laryngeal muscles and sets the foundation for acoustic variation and diversification among rodents. Our work highlights the importance of anatomical innovations in the evolution of animal sound production mechanisms. PMID:29291091

  1. Development of Vibrational Culture Model Mimicking Vocal Fold Tissues.

    PubMed

    Kim, Dongjoo; Lim, Jae-Yol; Kwon, Soonjo

    2016-10-01

    The vocal folds (VFs) are connective tissues with complex matrix structures that provide the required mechanical properties for voice generation. VF injury leads to changes in tissue structure and properties, resulting in reduced voice quality. However, injury-induced biochemical changes and repair in scarred VF tissues have not been well characterized to date. To treat scarred VFs, it is essential to understand how physiological characteristics of VFs tissue change in response to external perturbation. In this study, we designed a simple vibrational culture model to mimic vibratory microenvironments observed in vivo. This model consists of a flexible culture plate, three linear actuators, a stereo splitter, and a function generator. Human vocal fold fibroblast (hVFF) monolayers were established on the flexible membrane, to which normal phonatory vibrations were delivered from linear actuators and a function generator. The hVFF monolayers were exposed to the vibrational stresses at a frequency of 205 Hz for 2, 6, and 10 h with maximum displacement of 47.1 μm, followed by a 6 h rest. We then observed the changes in cell morphology, cell viability, and gene expression related to extracellular matrix components. In our dynamic culture device mimicking normal phonatory frequencies, cell proliferation increased and expression of hyaluronic acid synthase 2 was downregulated in response to vibrational stresses. The results presented herein will be useful for evaluating cellular responses following VF injuries in the presence or absence of vibrational stresses.

  2. Calibrating passive acoustic monitoring: correcting humpback whale call detections for site-specific and time-dependent environmental characteristics.

    PubMed

    Helble, Tyler A; D'Spain, Gerald L; Campbell, Greg S; Hildebrand, John A

    2013-11-01

    This paper demonstrates the importance of accounting for environmental effects on passive underwater acoustic monitoring results. The situation considered is the reduction in shipping off the California coast between 2008-2010 due to the recession and environmental legislation. The resulting variations in ocean noise change the probability of detecting marine mammal vocalizations. An acoustic model was used to calculate the time-varying probability of detecting humpback whale vocalizations under best-guess environmental conditions and varying noise. The uncorrected call counts suggest a diel pattern and an increase in calling over a two-year period; the corrected call counts show minimal evidence of these features.

  3. Singing-driven gene expression in the developing songbird brain

    PubMed Central

    Johnson, Frank; Whitney, Osceola

    2014-01-01

    Neural and behavioral development arises from an integration of genetic and environmental influences, yet specifying the nature of this interaction remains a primary problem in neuroscience. Here, we review molecular and behavioral studies that focus on the role of singing-driven gene expression during neural and vocal development in the male zebra finch (Taeniopygia guttata), a songbird that learns a species-typical vocal pattern during juvenile development by imitating an adult male tutor. A primary aim of our lab has been to identify naturally-occurring environmental influences that shape the propensity to sing. This ethological approach underlies our theoretical perspective, which is to integrate the significance of singing-driven gene expression into a broader ecological context. PMID:16129463

  4. Singing Fish in an Ocean of Noise: Effects of Boat Noise on the Plainfin Midshipman (Porichthys notatus) in a Natural Ecosystem.

    PubMed

    Cullis-Suzuki, Sarika

    2016-01-01

    When it comes to hearing and vocal communication in fishes, the plainfin midshipman (Porichthys notatus) is perhaps best understood. However, distinctly lacking are studies investigating communication of P. notatus in its natural ecosystems and the effects of noise on wild fish populations. Here, an exploratory look into both is discussed. By monitoring a population of wild P. notatus off British Columbia, Canada, call patterns were distinguished, the function of communicative sounds was identified, and midshipman vocalizations in agonistic encounters with natural predators were evaluated. A preliminary investigation into the effects of boat noise on wild midshipman is also described.

  5. Recursive Vocal Pattern Learning and Generalization in Starlings

    ERIC Educational Resources Information Center

    Bloomfield, Tiffany Corinna

    2012-01-01

    Among known communication systems, human language alone exhibits open-ended productivity of meaning. Interest in the psychological mechanisms supporting this ability, and their evolutionary origins, has resurged following the suggestion that the only uniquely human ability underlying language is a mechanism of recursion. This "Unique…

  6. Social Context-Dependent Activity in Marmoset Frontal Cortex Populations during Natural Conversations

    PubMed Central

    de la Mothe, Lisa; Miller, Cory T.

    2017-01-01

    Communication is an inherently interactive process that weaves together the fabric of both human and nonhuman primate societies. To investigate the properties of the primate brain during active social signaling, we recorded the responses of frontal cortex neurons as freely moving marmosets engaged in conversational exchanges with a visually occluded virtual marmoset. We found that small changes in firing rate (∼1 Hz) occurred across a broadly distributed population of frontal cortex neurons when marmosets heard a conspecific vocalization, and that these changes corresponded to subjects' likelihood of producing or withholding a vocal reply. Although the contributions of individual neurons were relatively small, large populations of neurons were able to clearly distinguish between these social contexts. Most significantly, this social context-dependent change in firing rate was evident even before subjects heard the vocalization, indicating that the probability of a conversational exchange was determined by the state of the frontal cortex at the time a vocalization was heard, and not by a decision driven by acoustic characteristics of the vocalization. We found that changes in neural activity scaled with the length of the conversation, with greater changes in firing rate evident for longer conversations. These data reveal specific and important facets of this neural activity that constrain its possible roles in active social signaling, and we hypothesize that the close coupling between frontal cortex activity and this natural, active primate social-signaling behavior facilitates social-monitoring mechanisms critical to conversational exchanges. SIGNIFICANCE STATEMENT We provide evidence for a novel pattern of neural activity in the frontal cortex of freely moving, naturally behaving, marmoset monkeys that may facilitate natural primate conversations. We discovered small (∼1 Hz), but reliable, changes in neural activity that occurred before marmosets even heard a conspecific vocalization that, as a population, almost perfectly predicted whether subjects would produce a vocalization in response. The change in the state of the frontal cortex persisted throughout the conversation and its magnitude scaled linearly with the length of the interaction. We hypothesize that this social context-dependent change in frontal cortex activity is supported by several mechanisms, such as social arousal and attention, and facilitates social monitoring critical for vocal coordination characteristic of human and nonhuman primate conversations. PMID:28630255

  7. Opposite hemispheric lateralization effects during speaking and singing at motor cortex, insula and cerebellum.

    PubMed

    Riecker, A; Ackermann, H; Wildgruber, D; Dogil, G; Grodd, W

    2000-06-26

    Aside from spoken language, singing represents a second mode of acoustic (auditory-vocal) communication in humans. As a new aspect of brain lateralization, functional magnetic resonance imaging (fMRI) revealed two complementary cerebral networks subserving singing and speaking. Reproduction of a non-lyrical tune elicited activation predominantly in the right motor cortex, the right anterior insula, and the left cerebellum whereas the opposite response pattern emerged during a speech task. In contrast to the hemodynamic responses within motor cortex and cerebellum, activation of the intrasylvian cortex turned out to be bound to overt task performance. These findings corroborate the assumption that the left insula supports the coordination of speech articulation. Similarly, the right insula might mediate temporo-spatial control of vocal tract musculature during overt singing. Both speech and melody production require the integration of sound structure or tonal patterns, respectively, with a speaker's emotions and attitudes. Considering the widespread interconnections with premotor cortex and limbic structures, the insula is especially suited for this task.

  8. IMRT for Image-Guided Single Vocal Cord Irradiation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Osman, Sarah O.S., E-mail: s.osman@erasmusmc.nl; Astreinidou, Eleftheria; Boer, Hans C.J. de

    2012-02-01

    Purpose: We have been developing an image-guided single vocal cord irradiation technique to treat patients with stage T1a glottic carcinoma. In the present study, we compared the dose coverage to the affected vocal cord and the dose delivered to the organs at risk using conventional, intensity-modulated radiotherapy (IMRT) coplanar, and IMRT non-coplanar techniques. Methods and Materials: For 10 patients, conventional treatment plans using two laterally opposed wedged 6-MV photon beams were calculated in XiO (Elekta-CMS treatment planning system). An in-house IMRT/beam angle optimization algorithm was used to obtain the coplanar and non-coplanar optimized beam angles. Using these angles, the IMRTmore » plans were generated in Monaco (IMRT treatment planning system, Elekta-CMS) with the implemented Monte Carlo dose calculation algorithm. The organs at risk included the contralateral vocal cord, arytenoids, swallowing muscles, carotid arteries, and spinal cord. The prescription dose was 66 Gy in 33 fractions. Results: For the conventional plans and coplanar and non-coplanar IMRT plans, the population-averaged mean dose {+-} standard deviation to the planning target volume was 67 {+-} 1 Gy. The contralateral vocal cord dose was reduced from 66 {+-} 1 Gy in the conventional plans to 39 {+-} 8 Gy and 36 {+-} 6 Gy in the coplanar and non-coplanar IMRT plans, respectively. IMRT consistently reduced the doses to the other organs at risk. Conclusions: Single vocal cord irradiation with IMRT resulted in good target coverage and provided significant sparing of the critical structures. This has the potential to improve the quality-of-life outcomes after RT and maintain the same local control rates.« less

  9. Role of N-Methyl-D-Aspartate Receptors in Action-Based Predictive Coding Deficits in Schizophrenia.

    PubMed

    Kort, Naomi S; Ford, Judith M; Roach, Brian J; Gunduz-Bruce, Handan; Krystal, John H; Jaeger, Judith; Reinhart, Robert M G; Mathalon, Daniel H

    2017-03-15

    Recent theoretical models of schizophrenia posit that dysfunction of the neural mechanisms subserving predictive coding contributes to symptoms and cognitive deficits, and this dysfunction is further posited to result from N-methyl-D-aspartate glutamate receptor (NMDAR) hypofunction. Previously, by examining auditory cortical responses to self-generated speech sounds, we demonstrated that predictive coding during vocalization is disrupted in schizophrenia. To test the hypothesized contribution of NMDAR hypofunction to this disruption, we examined the effects of the NMDAR antagonist, ketamine, on predictive coding during vocalization in healthy volunteers and compared them with the effects of schizophrenia. In two separate studies, the N1 component of the event-related potential elicited by speech sounds during vocalization (talk) and passive playback (listen) were compared to assess the degree of N1 suppression during vocalization, a putative measure of auditory predictive coding. In the crossover study, 31 healthy volunteers completed two randomly ordered test days, a saline day and a ketamine day. Event-related potentials during the talk/listen task were obtained before infusion and during infusion on both days, and N1 amplitudes were compared across days. In the case-control study, N1 amplitudes from 34 schizophrenia patients and 33 healthy control volunteers were compared. N1 suppression to self-produced vocalizations was significantly and similarly diminished by ketamine (Cohen's d = 1.14) and schizophrenia (Cohen's d = .85). Disruption of NMDARs causes dysfunction in predictive coding during vocalization in a manner similar to the dysfunction observed in schizophrenia patients, consistent with the theorized contribution of NMDAR hypofunction to predictive coding deficits in schizophrenia. Copyright © 2016 Society of Biological Psychiatry. Published by Elsevier Inc. All rights reserved.

  10. From imitation to meaning: circuit plasticity and the acquisition of a conventionalized semantics

    PubMed Central

    García, Ricardo R.; Zamorano, Francisco; Aboitiz, Francisco

    2014-01-01

    The capacity for language is arguably the most remarkable innovation of the human brain. A relatively recent interpretation prescribes that part of the language-related circuits were co-opted from circuitry involved in hand control—the mirror neuron system (MNS), involved both in the perception and in the execution of voluntary grasping actions. A less radical view is that in early humans, communication was opportunistic and multimodal, using signs, vocalizations or whatever means available to transmit social information. However, one point that is not yet clear under either perspective is how learned communication acquired a semantic property thereby allowing us to name objects and eventually describe our surrounding environment. Here we suggest a scenario involving both manual gestures and learned vocalizations that led to the development of a primitive form of conventionalized reference. This proposal is based on comparative evidence gathered from other species and on neurolinguistic evidence in humans, which points to a crucial role for vocal learning in the early development of language. Firstly, the capacity to direct the attention of others to a common object may have been crucial for developing a consensual referential system. Pointing, which is a ritualized grasping gesture, may have been crucial to this end. Vocalizations also served to generate joint attention among conversants, especially when combined with gaze direction. Another contributing element was the development of pantomimic actions resembling events or animals. In conjunction with this mimicry, the development of plastic neural circuits that support complex, learned vocalizations was probably a significant factor in the evolution of conventionalized semantics in our species. Thus, vocal imitations of sounds, as in onomatopoeias (words whose sound resembles their meaning), are possibly supported by mirror system circuits, and may have been relevant in the acquisition of early meanings. PMID:25152726

  11. Effects of the epilarynx area on vocal fold dynamics and the primary voice signal.

    PubMed

    Döllinger, Michael; Berry, David A; Luegmair, Georg; Hüttner, Björn; Bohr, Christopher

    2012-05-01

    For the analysis of vocal fold dynamics, sub- and supraglottal influences must be taken into account, as recent studies have shown. In this work, we analyze the influence of changes in the epilaryngeal area on vocal fold dynamics. We investigate two excised female larynges in a hemilarynx setup combined with a synthetic vocal tract consisting of hard plastic and simulating the vowel /a/. Eigenmodes, amplitudes, and velocities of the oscillations, the subglottal pressures (P(sub)), and sound pressure levels (SPLs) of the generated signal are investigated as a function of three distinctive epilaryngeal areas (28.4 mm(2), 71.0 mm(2), and 205.9 mm(2)). The results showed that the SPL is independent of the epilarynx cross section and exhibits a nonlinear relation to the insufflated airflow. The P(sub) decreased with an increase in the epilaryngeal area and displayed linear relations to the airflow. The principal eigenfunctions (EEFs) from the vocal fold dynamics exhibited lateral movement for the first EEF and rotational motion for the second EEF. In total, the first two EEFs covered a minimum of 60% of the energy, with an average of more than 50% for the first EEF. Correlations to the epilarynx areas were not found. Maximal values for amplitudes (up to 2.5 mm) and velocities (up to 1.57 mm/ms) changed with varying epilaryngeal area but did not show consistent behavior for both larynges. We conclude that the size of the epilaryngeal area has significant influence on vocal fold dynamics but does not significantly affect the resultant SPL. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  12. Experiments on Analysing Voice Production: Excised (Human, Animal) and In Vivo (Animal) Approaches

    PubMed Central

    Döllinger, Michael; Kobler, James; Berry, David A.; Mehta, Daryush D.; Luegmair, Georg; Bohr, Christopher

    2015-01-01

    Experiments on human and on animal excised specimens as well as in vivo animal preparations are so far the most realistic approaches to simulate the in vivo process of human phonation. These experiments do not have the disadvantage of limited space within the neck and enable studies of the actual organ necessary for phonation, i.e., the larynx. The studies additionally allow the analysis of flow, vocal fold dynamics, and resulting acoustics in relation to well-defined laryngeal alterations. Purpose of Review This paper provides an overview of the applications and usefulness of excised (human/animal) specimen and in vivo animal experiments in voice research. These experiments have enabled visualization and analysis of dehydration effects, vocal fold scarring, bifurcation and chaotic vibrations, three-dimensional vibrations, aerodynamic effects, and mucosal wave propagation along the medial surface. Quantitative data will be shown to give an overview of measured laryngeal parameter values. As yet, a full understanding of all existing interactions in voice production has not been achieved, and thus, where possible, we try to indicate areas needing further study. Recent Findings A further motivation behind this review is to highlight recent findings and technologies related to the study of vocal fold dynamics and its applications. For example, studies of interactions between vocal tract airflow and generation of acoustics have recently shown that airflow superior to the glottis is governed by not only vocal fold dynamics but also by subglottal and supraglottal structures. In addition, promising new methods to investigate kinematics and dynamics have been reported recently, including dynamic optical coherence tomography, X-ray stroboscopy and three-dimensional reconstruction with laser projection systems. Finally, we touch on the relevance of vocal fold dynamics to clinical laryngology and to clinically-oriented research. PMID:26581597

  13. Simulated Birdwatchers’ Playback Affects the Behavior of Two Tropical Birds

    PubMed Central

    Harris, J. Berton C.; Haskell, David G.

    2013-01-01

    Although recreational birdwatchers may benefit conservation by generating interest in birds, they may also have negative effects. One such potentially negative impact is the widespread use of recorded vocalizations, or “playback,” to attract birds of interest, including range-restricted and threatened species. Although playback has been widely used to test hypotheses about the evolution of behavior, no peer-reviewed study has examined the impacts of playback in a birdwatching context on avian behavior. We studied the effects of simulated birdwatchers’ playback on the vocal behavior of Plain-tailed Wrens Thryothorus euophrys and Rufous Antpittas Grallaria rufula in Ecuador. Study species’ vocal behavior was monitored for an hour after playing either a single bout of five minutes of song or a control treatment of background noise. We also studied the effects of daily five minute playback on five groups of wrens over 20 days. In single bout experiments, antpittas made more vocalizations of all types, except for trills, after playback compared to controls. Wrens sang more duets after playback, but did not produce more contact calls. In repeated playback experiments, wren responses were strong at first, but hardly detectable by day 12. During the study, one study group built a nest, apparently unperturbed, near a playback site. The playback-induced habituation and changes in vocal behavior we observed suggest that scientists should consider birdwatching activity when selecting research sites so that results are not biased by birdwatchers’ playback. Increased vocalizations after playback could be interpreted as a negative effect of playback if birds expend energy, become stressed, or divert time from other activities. In contrast, the habituation we documented suggests that frequent, regular birdwatchers’ playback may have minor effects on wren behavior. PMID:24147094

  14. Simulated birdwatchers' playback affects the behavior of two tropical birds.

    PubMed

    Harris, J Berton C; Haskell, David G

    2013-01-01

    Although recreational birdwatchers may benefit conservation by generating interest in birds, they may also have negative effects. One such potentially negative impact is the widespread use of recorded vocalizations, or "playback," to attract birds of interest, including range-restricted and threatened species. Although playback has been widely used to test hypotheses about the evolution of behavior, no peer-reviewed study has examined the impacts of playback in a birdwatching context on avian behavior. We studied the effects of simulated birdwatchers' playback on the vocal behavior of Plain-tailed Wrens Thryothorus euophrys and Rufous Antpittas Grallaria rufula in Ecuador. Study species' vocal behavior was monitored for an hour after playing either a single bout of five minutes of song or a control treatment of background noise. We also studied the effects of daily five minute playback on five groups of wrens over 20 days. In single bout experiments, antpittas made more vocalizations of all types, except for trills, after playback compared to controls. Wrens sang more duets after playback, but did not produce more contact calls. In repeated playback experiments, wren responses were strong at first, but hardly detectable by day 12. During the study, one study group built a nest, apparently unperturbed, near a playback site. The playback-induced habituation and changes in vocal behavior we observed suggest that scientists should consider birdwatching activity when selecting research sites so that results are not biased by birdwatchers' playback. Increased vocalizations after playback could be interpreted as a negative effect of playback if birds expend energy, become stressed, or divert time from other activities. In contrast, the habituation we documented suggests that frequent, regular birdwatchers' playback may have minor effects on wren behavior.

  15. Repeated imitation makes human vocalizations more word-like.

    PubMed

    Edmiston, Pierce; Perlman, Marcus; Lupyan, Gary

    2018-03-14

    People have long pondered the evolution of language and the origin of words. Here, we investigate how conventional spoken words might emerge from imitations of environmental sounds. Does the repeated imitation of an environmental sound gradually give rise to more word-like forms? In what ways do these forms resemble the original sounds that motivated them (i.e. exhibit iconicity)? Participants played a version of the children's game 'Telephone'. The first generation of participants imitated recognizable environmental sounds (e.g. glass breaking, water splashing). Subsequent generations imitated the previous generation of imitations for a maximum of eight generations. The results showed that the imitations became more stable and word-like, and later imitations were easier to learn as category labels. At the same time, even after eight generations, both spoken imitations and their written transcriptions could be matched above chance to the category of environmental sound that motivated them. These results show how repeated imitation can create progressively more word-like forms while continuing to retain a resemblance to the original sound that motivated them, and speak to the possible role of human vocal imitation in explaining the origins of at least some spoken words. © 2018 The Author(s).

  16. Bayesian Ising approximation for learning dictionaries of multispike timing patterns in premotor neurons

    NASA Astrophysics Data System (ADS)

    Hernandez Lahme, Damian; Sober, Samuel; Nemenman, Ilya

    Important questions in computational neuroscience are whether, how much, and how information is encoded in the precise timing of neural action potentials. We recently demonstrated that, in the premotor cortex during vocal control in songbirds, spike timing is far more informative about upcoming behavior than is spike rate (Tang et al, 2014). However, identification of complete dictionaries that relate spike timing patterns with the controled behavior remains an elusive problem. Here we present a computational approach to deciphering such codes for individual neurons in the songbird premotor area RA, an analog of mammalian primary motor cortex. Specifically, we analyze which multispike patterns of neural activity predict features of the upcoming vocalization, and hence are important codewords. We use a recently introduced Bayesian Ising Approximation, which properly accounts for the fact that many codewords overlap and hence are not independent. Our results show which complex, temporally precise multispike combinations are used by individual neurons to control acoustic features of the produced song, and that these code words are different across individual neurons and across different acoustic features. This work was supported, in part, by JSMF Grant 220020321, NSF Grant 1208126, NIH Grant NS084844 and NIH Grant 1 R01 EB022872.

  17. Development of an Acoustic Signal Analysis Tool “Auto-F” Based on the Temperament Scale

    NASA Astrophysics Data System (ADS)

    Modegi, Toshio

    The MIDI interface is originally designed for electronic musical instruments but we consider this music-note based coding concept can be extended for general acoustic signal description. We proposed applying the MIDI technology to coding of bio-medical auscultation sound signals such as heart sounds for retrieving medical records and performing telemedicine. Then we have tried to extend our encoding targets including vocal sounds, natural sounds and electronic bio-signals such as ECG, using Generalized Harmonic Analysis method. Currently, we are trying to separate vocal sounds included in popular songs and encode both vocal sounds and background instrumental sounds into separate MIDI channels. And also, we are trying to extract articulation parameters such as MIDI pitch-bend parameters in order to reproduce natural acoustic sounds using a GM-standard MIDI tone generator. In this paper, we present an overall algorithm of our developed acoustic signal analysis tool, based on those research works, which can analyze given time-based signals on the musical temperament scale. The prominent feature of this tool is producing high-precision MIDI codes, which reproduce the similar signals as the given source signal using a GM-standard MIDI tone generator, and also providing analyzed texts in the XML format.

  18. Musicians' Ratings of Good versus Bad Vocal and String Performances.

    ERIC Educational Resources Information Center

    Geringer, John M.; Madsen, Clifford K.

    1998-01-01

    Continues a line of research attempting to ascertain the focus of musicians' attention when listening to music, particularly whether musicians demonstrate consistent listening patterns across excerpts designed to be perceived as good and bad performances. Indicates that musician-listeners consistently discriminated between good and bad…

  19. Path Models of Vocal Emotion Communication

    PubMed Central

    Bänziger, Tanja; Hosoya, Georg; Scherer, Klaus R.

    2015-01-01

    We propose to use a comprehensive path model of vocal emotion communication, encompassing encoding, transmission, and decoding processes, to empirically model data sets on emotion expression and recognition. The utility of the approach is demonstrated for two data sets from two different cultures and languages, based on corpora of vocal emotion enactment by professional actors and emotion inference by naïve listeners. Lens model equations, hierarchical regression, and multivariate path analysis are used to compare the relative contributions of objectively measured acoustic cues in the enacted expressions and subjective voice cues as perceived by listeners to the variance in emotion inference from vocal expressions for four emotion families (fear, anger, happiness, and sadness). While the results confirm the central role of arousal in vocal emotion communication, the utility of applying an extended path modeling framework is demonstrated by the identification of unique combinations of distal cues and proximal percepts carrying information about specific emotion families, independent of arousal. The statistical models generated show that more sophisticated acoustic parameters need to be developed to explain the distal underpinnings of subjective voice quality percepts that account for much of the variance in emotion inference, in particular voice instability and roughness. The general approach advocated here, as well as the specific results, open up new research strategies for work in psychology (specifically emotion and social perception research) and engineering and computer science (specifically research and development in the domain of affective computing, particularly on automatic emotion detection and synthetic emotion expression in avatars). PMID:26325076

  20. Directionality of dog vocalizations

    NASA Astrophysics Data System (ADS)

    Frommolt, Karl-Heinz; Gebler, Alban

    2004-07-01

    The directionality patterns of sound emission in domestic dogs were measured in an anechoic environment using a microphone array. Mainly long-distance signals from four dogs were investigated. The radiation pattern of the signals differed clearly from an omnidirectional one with average differences in sound-pressure level between the frontal and rear position of 3-7 dB depending from the individual. Frequency dependence of directionality was shown for the range from 250 to 3200 Hz. The results indicate that when studying acoustic communication in mammals, more attention should be paid to the directionality pattern of sound emission.

  1. Higher-order semantic structures in an African Grey parrot's vocalizations: evidence from the hyperspace analog to language (HAL) model.

    PubMed

    Kaufman, Allison B; Colbert-White, Erin N; Burgess, Curt

    2013-09-01

    Previous research has described the significant role that social interaction plays in both the acquisition and use of speech by parrots. The current study analyzed the speech of one home-raised African Grey parrot (Psittacus erithacus erithacus) across three different social contexts: owner interacting with parrot in the same room, owner and parrot interacting out of view in adjacent rooms, and parrot home alone. The purpose was to determine the extent to which the subject's speech reflected an understanding of the contextual substitutability (e.g., the word street can be substituted in context for the word road) of the vocalizations that comprised the units in her repertoire (i.e., global co-occurrence of repertoire units; Burgess in Behav Res Methods Instrum Comput 30:188-198, 1998; Lund and Burgess in Behav Res Methods Instrum Comput 28:203-208, 1996). This was accomplished via the human language model hyperspace analog to language (HAL). HAL is contextually driven and bootstraps language "rules" from input without human intervention. Because HAL does not require human tutelage, it provided an objective measure to empirically examine the parrot's vocalizations. Results indicated that the subject's vocalization patterns did contain global co-occurrence. The presence of this quality in this nonhuman's speech may be strongly indicative of higher-order cognitive skills.

  2. A comparison of a child’s fundamental frequencies in structured elicited vocalizations versus unstructured natural vocalizations: A case study

    PubMed Central

    Hunter, Eric J.

    2009-01-01

    Objectives Building on the concept that task type may influence fundamental frequency (F0) values, the purpose of this case study was to investigate the difference in a child’s F0 during structured, elicited tasks and long-term, unstructured activities. It also explores the possibility that the distribution in children’s F0 may make the standard statistical measures of mean and standard deviation less than ideal metrics. Methods A healthy male child (5 years, 7 months) was evaluated. The child completed four voice tasks used in a previous study of the influence of task type on F0 values: (1) sustaining the vowel /a/; (2) sustaining the vowel, /a/, embedded in a word at the end of a phrase; (3) repeating a sentence; and (4) counting from 1 to 10. The child also wore a National Center for Voice and Speech voice dosimeter, a device that collects voice data over the course of an entire day, during all activities for 34 hours over 4 days. Results Throughout the structured vocal tasks within the clinical environment, the child’s F0, as measured by both the dosimeter and acoustic analysis of microphone data, was similar for all four tasks, with the counting task the most dissimilar. The mean F0 (~257 Hz) matched very closely to the average task results in the literature given for the child’s age group. However, the child’s mean fundamental frequency during the unstructured activities was significantly higher (~376 Hz). Finally, the mode and median of the structured vocal tasks were respectively 260 Hz and 259 Hz (both near the mean), while the unstructured mode and median were respectively 290 Hz and 355 Hz. Conclusions The results of this study suggest that children may produce a notably different voice pattern during clinical observations compared to routine daily activities. In addition, the child’s long-term F0 distribution is not normal. If this distribution is consistent in long-term, unstructured natural vocalization patterns of children, statistical mean would not be a valid measure. Mode and median are suggested as two parameters which convey more accurate information about typical F0 usage. Finally, future research avenues, including further exploration of how children may adapt their F0 to various environments, conversation partners, and activity, are suggested. PMID:19185926

  3. Social Brain Hypothesis: Vocal and Gesture Networks of Wild Chimpanzees

    PubMed Central

    Roberts, Sam G. B.; Roberts, Anna I.

    2016-01-01

    A key driver of brain evolution in primates and humans is the cognitive demands arising from managing social relationships. In primates, grooming plays a key role in maintaining these relationships, but the time that can be devoted to grooming is inherently limited. Communication may act as an additional, more time-efficient bonding mechanism to grooming, but how patterns of communication are related to patterns of sociality is still poorly understood. We used social network analysis to examine the associations between close proximity (duration of time spent within 10 m per hour spent in the same party), grooming, vocal communication, and gestural communication (duration of time and frequency of behavior per hour spent within 10 m) in wild chimpanzees. This study examined hypotheses formulated a priori and the results were not corrected for multiple testing. Chimpanzees had differentiated social relationships, with focal chimpanzees maintaining some level of proximity to almost all group members, but directing gestures at and grooming with a smaller number of preferred social partners. Pairs of chimpanzees that had high levels of close proximity had higher rates of grooming. Importantly, higher rates of gestural communication were also positively associated with levels of proximity, and specifically gestures associated with affiliation (greeting, gesture to mutually groom) were related to proximity. Synchronized low-intensity pant-hoots were also positively related to proximity in pairs of chimpanzees. Further, there were differences in the size of individual chimpanzees' proximity networks—the number of social relationships they maintained with others. Focal chimpanzees with larger proximity networks had a higher rate of both synchronized low- intensity pant-hoots and synchronized high-intensity pant-hoots. These results suggest that in addition to grooming, both gestures and synchronized vocalizations may play key roles in allowing chimpanzees to manage a large and differentiated set of social relationships. Gestures may be important in reducing the aggression arising from being in close proximity to others, allowing for proximity to be maintained for longer and facilitating grooming. Vocalizations may allow chimpanzees to communicate with a larger number of recipients than gestures and the synchronized nature of the pant-hoot calls may facilitate social bonding of more numerous social relationships. As group sizes increased through human evolution, both gestures and synchronized vocalizations may have played important roles in bonding social relationships in a more time-efficient manner than grooming. PMID:27933005

  4. Fourier Analysis and the Rhythm of Conversation.

    ERIC Educational Resources Information Center

    Dabbs, James M., Jr.

    Fourier analysis, a common technique in engineering, breaks down a complex wave form into its simple sine wave components. Communication researchers have recently suggested that this technique may provide an index of the rhythm of conversation, since vocalizing and pausing produce a complex wave form pattern of alternation between two speakers. To…

  5. Does Melody Assist in the Reproduction of Novel Rhythm Patterns?

    ERIC Educational Resources Information Center

    Kinney, Daryl W.; Forsythe, Jere L.

    2013-01-01

    We examined music education majors' ability to reproduce rhythmic stimuli presented in melody and rhythm only conditions. Participants reproduced rhythms of two-measure music examples by immediately echo-performing through a method of their choosing (e.g., clapping, tapping, vocalizing). Forty examples were presented in melody and rhythm only…

  6. Elementary School Teachers' Vocal Dose: Muscle Bioenergetics and Training Implications

    ERIC Educational Resources Information Center

    Smith, Audrey G.; Sandage, Mary J.; Pascoe, David D.; Plexico, Laura W.; Lima, Italo R.; Cao, Guanqun

    2017-01-01

    Purpose: Translating exercise-science methodology for determination of muscle bioenergetics, we hypothesized that the temporal voice-use patterns for classroom and music teachers would indicate a reliance on the immediate energy system for laryngeal skeletal-muscle metabolism. It was hypothesized that the music-teacher group would produce longer…

  7. Treatment of Voice Disorders in Children

    ERIC Educational Resources Information Center

    Hooper, Celia R.

    2004-01-01

    Children with voice disorders do respond to treatment, with vocal hyperfunction being the predominant disorder on the caseload of the pediatric voice clinician. This article reviews the literature in describing what is known about these children and typical disorders, prevention of voice disorders, the need for treatment, the referral patterns of…

  8. On the application of the lattice Boltzmann method to the investigation of glottal flow

    PubMed Central

    Kucinschi, Bogdan R.; Afjeh, Abdollah A.; Scherer, Ronald C.

    2008-01-01

    The production of voice is directly related to the vibration of the vocal folds, which is generated by the interaction between the glottal flow and the tissue of the vocal folds. In the current study, the aerodynamics of the symmetric glottis is investigated numerically for a number of static configurations. The numerical investigation is based on the lattice Boltzmann method (LBM), which is an alternative approach within computational fluid dynamics. Compared to the traditional Navier–Stokes computational fluid dynamics methods, the LBM is relatively easy to implement and can deal with complex geometries without requiring a dedicated grid generator. The multiple relaxation time model was used to improve the numerical stability. The results obtained with LBM were compared to the results provided by a traditional Navier–Stokes solver and experimental data. It was shown that LBM results are satisfactory for all the investigated cases. PMID:18646995

  9. Masculine men articulate less clearly.

    PubMed

    Kempe, Vera; Puts, David A; Cárdenas, Rodrigo A

    2013-12-01

    In previous research, acoustic characteristics of the male voice have been shown to signal various aspects of mate quality and threat potential. But the human voice is also a medium of linguistic communication. The present study explores whether physical and vocal indicators of male mate quality and threat potential are linked to effective communicative behaviors such as vowel differentiation and use of more salient phonetic variants of consonants. We show that physical and vocal indicators of male threat potential, height and formant position, are negatively linked to vowel space size, and that height and levels of circulating testosterone are negatively linked to the use of the aspirated variant of the alveolar stop consonant /t/. Thus, taller, more masculine men display less clarity in their speech and prefer phonetic variants that may be associated with masculine attributes such as toughness. These findings suggest that vocal signals of men's mate quality and/or dominance are not confined to the realm of voice acoustics but extend to other aspects of communicative behavior, even if this means a trade-off with speech patterns that are considered communicatively advantageous, such as clarity and indexical cues to higher social class.

  10. Performance constraints and the production of birdsong

    NASA Astrophysics Data System (ADS)

    Suthers, Roderick A.; Vallet, Eric; Zollinger, Sue Anne

    2004-05-01

    The role of physical and physiological constraints in determining the performance limits on the tempo and frequency bandwidth of birdsong was investigated. One series of experiments examined the mechanism by which a vocal mimic, the northern mockingbird (Mimus polygottos), copied the songs of other species with which it was tutored as a juvenile. Other experiments analyzed the motor basis of special canary (Serinus canaria) syllables eliciting sexual responses from females. In each case, the mechanism of vocalization was determined by measuring the respiratory dynamics and sound produced on each side of the songbirds duplex vocal organ, the syrinx. When mockingbirds copied the songs of other species the accuracy of their copy depended on the accuracy with which they reproduced the motor pattern used by the tutor species. Motor difficulty of various acoustic features was assessed by the accuracy of its copy. The high repetition rate, broadband canary syllables preferred by females required especially demanding bilateral motor skills. The results indicate that constraints on the rate of respiratory ventilation and bilateral syringeal coordination can set an upper limit on syllable repetition rate and frequency bandwidth. [Work supported by NIH and NSF.

  11. Vocal Behavior of the Elusive Purple Frog of India (Nasikabatrachus sahyadrensis), a Fossorial Species Endemic to the Western Ghats

    PubMed Central

    Thomas, Ashish; Suyesh, Robin; Biju, S. D.; Bee, Mark A.

    2014-01-01

    Quantitative descriptions of animal vocalizations can inform an understanding of their evolutionary functions, the mechanisms for their production and perception, and their potential utility in taxonomy, population monitoring, and conservation. The goal of this study was to provide the first acoustical and statistical analysis of the advertisement calls of Nasikabatrachus sahyadrensis. Commonly known as the Indian purple frog, N. sahyadrensis is an endangered species endemic to the Western Ghats of India. As the only known species in its family (Nasikabatrachidae), it has ancient evolutionary ties to frogs restricted to the Seychelles archipelago (Sooglossidae). The role of vocalizations in the behavior of this unique species poses interesting questions, as the animal is fossorial and potentially earless and it breeds explosively above the soil for only about two weeks a year. In this study, we quantified 19 acoustic properties of 208 calls recorded from 10 males. Vocalizations were organized into distinct call groups typically composed of two to six short (59 ms), pulsatile calls, each consisting of about five to seven pulses produced at a rate of about 106 pulses/s. The frequency content of the call consisted of a single dominant peak between 1200–1300 Hz and there was no frequency modulation. The patterns of variation within and among individuals were typical of those seen in other frogs. Few of the properties we measured were related to temperature, body size, or condition, though there was little variation in temperature. Field observations and recordings of captive individuals indicated that males engaged in both antiphonal calling and call overlap with nearby calling neighbors. We discuss our findings in relation to previous work on vocal behavior in other fossorial frogs and in sooglossid frogs. PMID:24516517

  12. OBS records of Whale vocalizations from Lucky-strike segment of the Mid-Atlantic Ridge during 2007-2008

    NASA Astrophysics Data System (ADS)

    Chauhan, A.; Rai, A.; Singh, S. C.; Crawford, W. C.; Escartin, J.; Cannat, M.

    2009-12-01

    Passive seismic experiments to study seismicity require a long term deployment of ocean-bottom seismometers (OBSs). These instruments also record a large amount of non-seismogenic signals such as movement of large ships, air-gun shots, and marine mammal vocalizations. We report a bi-product of our passive seismic experiment (BBMOMAR) conducted around the Lucky-strike hydrothermal field of the slow-spreading mid-Atlantic ridge. Five multi-component ocean-bottom seismometers (recording two horizontal, one vertical and one pressure channel) were deployed during 2007-2008. During 13 months of deployment, abundant vocalizations of marine mammals have been recorded by all the five equipments. By analyzing the frequency content of data and their pattern of occurrence, we conclude that these low-frequency vocalizations (~20-40 Hz) typically corresponds to blue and fin-whales. These signals if not identified, could be mis-interpreted as underwater seismic/hydrothermal activity. Our data show an increase in the number of vocalizations recorded during the winter season relative to the summer. As part of the seismic monitoring of the Lucky-strike site, we anticipate to extend this study to the 2008-2009 and 2009-2010 periods, after the recovery and deployment of the array during the BATHYLUCK09 cruise. Long-term and continuous records of calls of marine mammals provide valuable information that could be used to identify the species, study their seasonal behaviour and their migration paths. Our study suggestes that passive experiments such as ocean-bottom seismometers deployed at key locations, could provide useful secondary infromation about oceanic species besides recording seismicity, which is otherwise not possible without harming or interfering with their activity.

  13. Striatal FoxP2 Is Actively Regulated during Songbird Sensorimotor Learning

    PubMed Central

    Teramitsu, Ikuko; Poopatanapong, Amy; Torrisi, Salvatore; White, Stephanie A.

    2010-01-01

    Background Mutations in the FOXP2 transcription factor lead to language disorders with developmental onset. Accompanying structural abnormalities in cortico-striatal circuitry indicate that at least a portion of the behavioral phenotype is due to organizational deficits. We previously found parallel FoxP2 expression patterns in human and songbird cortico/pallio-striatal circuits important for learned vocalizations, suggesting that FoxP2's function in birdsong may generalize to speech. Methodology/Principal Findings We used zebra finches to address the question of whether FoxP2 is additionally important in the post-organizational function of these circuits. In both humans and songbirds, vocal learning depends on auditory guidance to achieve and maintain optimal vocal output. We tested whether deafening prior to or during the sensorimotor phase of song learning disrupted FoxP2 expression in song circuitry. As expected, the songs of deafened juveniles were abnormal, however basal FoxP2 levels were unaffected. In contrast, when hearing or deaf juveniles sang for two hours in the morning, FoxP2 was acutely down-regulated in the striatal song nucleus, area X. The extent of down-regulation was similar between hearing and deaf birds. Interestingly, levels of FoxP2 and singing were correlated only in hearing birds. Conclusions/Significance Hearing appears to link FoxP2 levels to the amount of vocal practice. As juvenile birds spent more time practicing than did adults, their FoxP2 levels are likely to be low more often. Behaviorally-driven reductions in the mRNA encoding this transcription factor could ultimately affect downstream molecules that function in vocal exploration, especially during sensorimotor learning. PMID:20062527

  14. Nebulized isotonic saline improves voice production in Sjögren's syndrome.

    PubMed

    Tanner, Kristine; Nissen, Shawn L; Merrill, Ray M; Miner, Alison; Channell, Ron W; Miller, Karla L; Elstad, Mark; Kendall, Katherine A; Roy, Nelson

    2015-10-01

    This study examined the effects of a topical vocal fold hydration treatment on voice production over time. Prospective, longitudinal, within-subjects A (baseline), B (treatment), A (withdrawal/reversal), B (treatment) experimental design. Eight individuals with primary Sjögren's syndrome (SS), an autoimmune disease causing laryngeal dryness, completed an 8-week A-B-A-B experiment. Participants performed twice-daily audio recordings of connected speech and sustained vowels and then rated vocal effort, mouth dryness, and throat dryness. Two-week treatment phases introduced twice-daily 9-mL doses of nebulized isotonic saline (0.9% Na(+)Cl(-)). Voice handicap and patient-based measures of SS disease severity were collected before and after each 2-week phase. Connected speech and sustained vowels were analyzed using the Cepstral Spectral Index of Dysphonia (CSID). Acoustic and patient-based ratings during each baseline and treatment phase were analyzed and compared. Baseline CSID and patient-based ratings were in the mild-to-moderate range. CSID measures of voice severity improved by approximately 20% with nebulized saline treatment and worsened during treatment withdrawal. Posttreatment CSID values fell within the normal-to-mild range. Similar patterns were observed in patient-based ratings of vocal effort and dryness. CSID values and patient-based ratings correlated significantly (P < .05). Nebulized isotonic saline improves voice production based on acoustic and patient-based ratings of voice severity. Future work should optimize topical vocal fold hydration treatment formulations, dose, and delivery methodologies for various patient populations. This study lays the groundwork for future topical vocal fold hydration treatment development to manage and possibly prevent dehydration-related voice disorders. 2b. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.

  15. Embryologic innervation of the rat laryngeal musculature--a model for investigation of recurrent laryngeal nerve reinnervation.

    PubMed

    Pitman, Michael J; Berzofsky, Craig E; Alli, Opeyemi; Sharma, Sansar

    2013-12-01

    Optimal management of vocal fold paralysis would entail recurrent laryngeal nerve (RLN) reinnervation resulting in normal vocal fold motion. Unfortunately, RLN reinnervation currently results in a nonfunctional vocal fold due to synkinetic reinnervation. Therapeutic interventions that guide regenerating axons back to the appropriate muscle would prevent synkinesis and restore vocal fold and glottal function. The initial step toward developing these therapies is the elucidation of the embryologic innervation of the larynx. This study aimed to identify the age of occurrence, timing, and pattern of embryologic innervation of the rat larynx, hypothesizing that differences in these parameters exist between distinct laryngeal muscles. Descriptive anatomic study. The larynx of rats aged embryologic day (E) 15, 16, 17, 19, and 21 were harvested and then sectioned. Two rats were used for each age. Sections were colabeled with neuronal class III β-tubulin polyclonal antibody to identify the presence of axons and alexa 488 conjugate α-bungarotoxin to identify the presence of motor endplates. The age at which axons and motor endplates were first present was noted. The position and pattern of the axons and motor endplates was recorded in relation to each other as well as the musculoskeletal anatomy of the larynx. The time at which axons appeared to innervate the medial thyroarytenoid (MTA) muscle, lateral thyroarytenoid (LTA) muscle, and the posterior cricoarytenoid (PCA) muscle was documented. Findings in the rat suggest the RLN reaches the larynx and begins branching by E15. Axons branch dorsally first and reach the PCA muscle before the other muscles. Branching toward the MTA muscle occurs only after axons have reached the LTA muscle. By E19, RLN axons have been guided to and selected their respective muscles with formation of neuromuscular junctions (NMJs) in the PCA, LTA and MTA muscles, though the formation of NMJs in the MTA muscle was comparatively delayed. This study describes the embryologic innervation of the rat larynx and suggests that there are distinct differences in the age of occurrence, timing, and pattern of innervation of the PCA, LTA, and MTA muscles of the rat. These findings lay the foundation for studies investigating the role of guidance cues in RLN axon guidance and the utility of these cues in the treatment of RLN injury via the stimulation of functional, nonsynkinetic reinnervation. Copyright © 2013 The American Laryngological, Rhinological and Otological Society, Inc.

  16. Variation in vocal-motor development in infant siblings of children with autism.

    PubMed

    Iverson, Jana M; Wozniak, Robert H

    2007-01-01

    In this study we examined early motor, vocal, and communicative development in a group of younger siblings of children diagnosed with autism (Infant Siblings). Infant Siblings and no-risk comparison later-born infants were videotaped at home with a primary caregiver each month from 5 to 14 months, with follow-up at 18 months. As a group, Infant Siblings were delayed in the onset of early developmental milestones and spent significantly less time in a greater number of postures, suggestive of relative postural instability. In addition, they demonstrated attenuated patterns of change in rhythmic arm activity around the time of reduplicated babble onset; and they were highly likely to exhibit delayed language development at 18 months.

  17. JS-X syndrome: A multiple congenital malformation with vocal cord paralysis, ear deformity, hearing loss, shoulder musculature underdevelopment, and X-linked recessive inheritance.

    PubMed

    Hoeve, Hans L J; Brooks, Alice S; Smit, Liesbeth S

    2015-07-01

    We report on a family with a not earlier described multiple congenital malformation. Several male family members suffer from laryngeal obstruction caused by bilateral vocal cord paralysis, outer and middle ear deformity with conductive and sensorineural hearing loss, facial dysmorphisms, and underdeveloped shoulder musculature. The affected female members only have middle ear deformity and hearing loss. The pedigree is suggestive of an X-linked recessive inheritance pattern. SNP-array revealed a deletion and duplication on Xq28 in the affected family members. A possible aetiology is a neurocristopathy with most symptoms expressed in structures derived from branchial arches. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  18. Spatial and temporal trends in fin whale vocalizations recorded in the NE Pacific Ocean between 2003-2013

    PubMed Central

    Weirathmueller, Michelle J.; Stafford, Kathleen M.; Wilcock, William S. D.; Hilmo, Rose S.; Dziak, Robert P.; Tréhu, Anne M.

    2017-01-01

    In order to study the long-term stability of fin whale (Balaenoptera physalus) singing behavior, the frequency and inter-pulse interval of fin whale 20 Hz vocalizations were observed over 10 years from 2003–2013 from bottom mounted hydrophones and seismometers in the northeast Pacific Ocean. The instrument locations extended from 40°N to 48°N and 130°W to 125°W with water depths ranging from 1500–4000 m. The inter-pulse interval (IPI) of fin whale song sequences was observed to increase at a rate of 0.54 seconds/year over the decade of observation. During the same time period, peak frequency decreased at a rate of 0.17 Hz/year. Two primary call patterns were observed. During the earlier years, the more commonly observed pattern had a single frequency and single IPI. In later years, a doublet pattern emerged, with two dominant frequencies and IPIs. Many call sequences in the intervening years appeared to represent a transitional state between the two patterns. The overall trend was consistent across the entire geographical span, although some regional differences exist. Understanding changes in acoustic behavior over long time periods is needed to help establish whether acoustic characteristics can be used to help determine population identity in a widely distributed, difficult to study species such as the fin whale. PMID:29073230

  19. Spatial and temporal trends in fin whale vocalizations recorded in the NE Pacific Ocean between 2003-2013.

    PubMed

    Weirathmueller, Michelle J; Stafford, Kathleen M; Wilcock, William S D; Hilmo, Rose S; Dziak, Robert P; Tréhu, Anne M

    2017-01-01

    In order to study the long-term stability of fin whale (Balaenoptera physalus) singing behavior, the frequency and inter-pulse interval of fin whale 20 Hz vocalizations were observed over 10 years from 2003-2013 from bottom mounted hydrophones and seismometers in the northeast Pacific Ocean. The instrument locations extended from 40°N to 48°N and 130°W to 125°W with water depths ranging from 1500-4000 m. The inter-pulse interval (IPI) of fin whale song sequences was observed to increase at a rate of 0.54 seconds/year over the decade of observation. During the same time period, peak frequency decreased at a rate of 0.17 Hz/year. Two primary call patterns were observed. During the earlier years, the more commonly observed pattern had a single frequency and single IPI. In later years, a doublet pattern emerged, with two dominant frequencies and IPIs. Many call sequences in the intervening years appeared to represent a transitional state between the two patterns. The overall trend was consistent across the entire geographical span, although some regional differences exist. Understanding changes in acoustic behavior over long time periods is needed to help establish whether acoustic characteristics can be used to help determine population identity in a widely distributed, difficult to study species such as the fin whale.

  20. MATERNAL ANXIETY SYMPTOMS AND MOTHER–INFANT SELF- AND INTERACTIVE CONTINGENCY

    PubMed Central

    Beebe, Beatrice; Steele, Miriam; Jaffe, Joseph; Buck, Karen A.; Chen, Henian; Cohen, Patricia; Kaitz, Marsha; Markese, Sara; Andrews, Howard; Margolis, Amy; Feldstein, Stanley

    2014-01-01

    Associations of maternal self-report anxiety-related symptoms with mother–infant 4-month face-to-face play were investigated in 119 pairs. Attention, affect, spatial orientation, and touch were coded from split-screen videotape on a 1-s time base. Self- and interactive contingency were assessed by time-series methods. Because anxiety symptoms signal emotional dysregulation, we expected to find atypical patterns of mother–infant interactive contingencies, and of degree of stability/lability within an individual’s own rhythms of behavior (self-contingencies). Consistent with our optimum midrange model, maternal anxiety-related symptoms biased the interaction toward interactive contingencies that were both heightened (vigilant) in some modalities and lowered (withdrawn) in others; both may be efforts to adapt to stress. Infant self-contingency was lowered (“destabilized”) with maternal anxiety symptoms; however, maternal self-contingency was both lowered in some modalities and heightened (overly stable) in others. Interactive contingency patterns were characterized by intermodal discrepancies, confusing forms of communication. For example, mothers vigilantly monitored infants visually, but withdrew from contingently coordinating with infants emotionally, as if mothers were “looking through” them. This picture fits descriptions of mothers with anxiety symptoms as overaroused/fearful, leading to vigilance, but dealing with their fear through emotional distancing. Infants heightened facial affect coordination (vigilance), but dampened vocal affect coordination (withdrawal), with mother’s face—a pattern of conflict. The maternal and infant patterns together generated a mutual ambivalence. PMID:25983359

  1. Analysis and localization of blue whale vocalizations in the Solomon Sea using waveform amplitude data.

    PubMed

    Frank, Scott D; Ferris, Aaron N

    2011-08-01

    During the Woodlark Basin seismic experiment in eastern Papua New Guinea (1999-2000), an ocean-bottom seismic array recorded marine mammal vocalizations along with target earthquake signals. The array consisted of 14 instruments, 7 of which were three-component seismometers with a fourth component hydrophone. They were deployed at 2.0-3.2 km water depth and operated from September 1999 through February 2000. While whale vocalizations were recorded throughout the deployment, this study focuses on 3 h from December 21, 1999 during which the signals are particularly clear. The recordings show a blue whale song composed of a three-unit phrase. That song does not match vocalization characteristics of other known Pacific subpopulations and may represent a previously undocumented blue whale song. Animal tracking and source level estimates are obtained with a Bayesian inversion method that generates probabilistic source locations. The Bayesian method is augmented to include travel time estimates from seismometers and hydrophones and acoustic signal amplitude. Tracking results show the whale traveled northeasterly over the course of 3 h, covering approximately 27 km. The path followed the edge of the Woodlark Basin along a shelf that separates the shallow waters of the Trobriand platform from the deep waters of the basin.

  2. How do you say 'hello'? Personality impressions from brief novel voices.

    PubMed

    McAleer, Phil; Todorov, Alexander; Belin, Pascal

    2014-01-01

    On hearing a novel voice, listeners readily form personality impressions of that speaker. Accurate or not, these impressions are known to affect subsequent interactions; yet the underlying psychological and acoustical bases remain poorly understood. Furthermore, hitherto studies have focussed on extended speech as opposed to analysing the instantaneous impressions we obtain from first experience. In this paper, through a mass online rating experiment, 320 participants rated 64 sub-second vocal utterances of the word 'hello' on one of 10 personality traits. We show that: (1) personality judgements of brief utterances from unfamiliar speakers are consistent across listeners; (2) a two-dimensional 'social voice space' with axes mapping Valence (Trust, Likeability) and Dominance, each driven by differing combinations of vocal acoustics, adequately summarises ratings in both male and female voices; and (3) a positive combination of Valence and Dominance results in increased perceived male vocal Attractiveness, whereas perceived female vocal Attractiveness is largely controlled by increasing Valence. Results are discussed in relation to the rapid evaluation of personality and, in turn, the intent of others, as being driven by survival mechanisms via approach or avoidance behaviours. These findings provide empirical bases for predicting personality impressions from acoustical analyses of short utterances and for generating desired personality impressions in artificial voices.

  3. Elephant low-frequency vocalizations propagate in the ground and seismic playbacks of these vocalizations are detectable by wild African elephants (Loxodonta africana)

    NASA Astrophysics Data System (ADS)

    O'Connell-Rodwell, Caitlin E.; Wood, Jason D.; Gunther, Roland; Klemperer, Simon; Rodwell, Timothy C.; Puria, Sunil; Sapolsky, Robert; Kinzley, Colleen; Arnason, Byron T.; Hart, Lynette A.

    2004-05-01

    Seismic correlates of low-frequency vocalizations in African and Asian elephants propagate in the ground at different velocities, with the potential of traveling farther than their airborne counterparts. A semblance technique applied to linear moveouts on narrow-bandpass-filtered data, coupled with forward modeling, demonstrates that the complex waves observed are the interference of an air wave and a Rayleigh wave traveling at the appropriate velocities. The Rayleigh wave appears to be generated at or close to the elephant, either by coupling through the elephant's body or through the air near the body to the ground. Low-frequency elephant vocalizations were reproduced seismically and played back to both a captive elephant and to elephant breeding herds in the wild, monitoring the elephants' behavioral responses, spacing between herd members and time spent at the water hole as an index of heightened vigilance. Breeding herds detected and responded appropriately to seismically transmitted elephant warning calls. The captive studies promise to elucidate a vibrotactile threshold of sensitivity for the elephant foot. Elephants may benefit from the exploitation of seismic cues as an additional communication modality, thus expanding their signaling repertoire and extending their range of potential communication and eavesdropping beyond that possible with airborne sound.

  4. Individual killer whale vocal variation during intra-group behavioral dynamics

    NASA Astrophysics Data System (ADS)

    Grebner, Dawn M.

    The scientific goal of this dissertation was to carefully study the signal structure of killer whale communications and vocal complexity and link them to behavioral circumstances. The overall objective of this research sought to provide insight into killer whale call content and usage which may be conveying information to conspecifics in order to maintain group cohesion. Data were collected in the summers of 2006 and 2007 in Johnstone Strait, British Columbia. For both individuals and small groups, vocalizations were isolated using a triangular hydrophone array and the behavioral movement patterns were captured by a theodolite and video camera positioned on a cliff overlooking the hyrophone locations. This dissertation is divided into four analysis chapters. In Chapter 3, discriminant analysis was used to validate the four N04 call subtypes which were originally parsed due to variations in slope segments. The first two functions of the discriminant analysis explained 97% of the variability. Most of the variability for the N04 call was found in the front convex and the terminal portions of the call, while very little variability was found in the center region of the call. This research revealed that individual killer whales produced multiple subtypes of the N04 call. No correlations of behaviors to acoustic parameters obtained were found. The aim of the Chapter 4 was to determine if killer whale calling behavior varied prior to and after the animals had joined. Pulsed call rates were found to be greater pre- compared to post-joining events. Two-way vocal exchanges were more common occurring 74% of the time during pre-joining events. In Chapter 5, initiated and first response to calls varied between age/sex class groups when mothers were separated from an offspring. Solo mothers and calves initiated pulsed calls more often than they responded. Most of the no vocal responses were due to mothers who were foraging. Finally, observations of the frequency split in N04 calls discussed in Chapter 6 showed that the higher frequency component (HFC) was always associated with sideband 7 (SB7) of the lower frequency component (LFC). Insight into Northern Resident killer whale intra-group vocal dynamics would aid our understanding of vocal behaviors of many other marine mammal species that rely on vocal exchanges for prey capture, group movement or survival. This is the first study to focus on killer whale vocal content and usage as it pertains to intra-group dynamics for (1) mother and offspring separations and (2) for all individuals prior to joining events, as well as (3) individual usage in a diverging pulsed call. It is also the first time the N04 call has been parsed into subtypes.

  5. Study of non-linear deformation of vocal folds in simulations of human phonation

    NASA Astrophysics Data System (ADS)

    Saurabh, Shakti; Bodony, Daniel

    2014-11-01

    Direct numerical simulation is performed on a two-dimensional compressible, viscous fluid interacting with a non-linear, viscoelastic solid as a model for the generation of the human voice. The vocal fold (VF) tissues are modeled as multi-layered with varying stiffness in each layer and using a finite-strain Standard Linear Solid (SLS) constitutive model implemented in a quadratic finite element code and coupled to a high-order compressible Navier-Stokes solver through a boundary-fitted fluid-solid interface. The large non-linear mesh deformation is handled using an elliptic/poisson smoothening technique. Supra-glottal flow shows asymmetry in the flow, which in turn has a coupling effect on the motion of the VF. The fully compressible simulations gives direct insight into the sound produced as pressure distributions and the vocal fold deformation helps study the unsteady vortical flow resulting from the fluid-structure interaction along the full phonation cycle. Supported by the National Science Foundation (CAREER Award Number 1150439).

  6. Medial surface dynamics of an in vivo canine vocal fold during phonation

    NASA Astrophysics Data System (ADS)

    Döllinger, Michael; Berry, David A.; Berke, Gerald S.

    2005-05-01

    Quantitative measurement of the medial surface dynamics of the vocal folds is important for understanding how sound is generated within the larynx. Building upon previous excised hemilarynx studies, the present study extended the hemilarynx methodology to the in vivo canine larynx. Through use of an in vivo model, the medial surface dynamics of the vocal fold were examined as a function of active thyroarytenoid muscle contraction. Data were collected using high-speed digital imaging at a sampling frequency of 2000 Hz, and a spatial resolution of 1024×1024 pixels. Chest-like and fry-like vibrations were observed, but could not be distinguished based on the input stimulation current to the recurrent laryngeal nerve. The subglottal pressure did distinguish the registers, as did an estimate of the thyroarytenoid muscle activity. Upon quantification of the three-dimensional motion, the method of Empirical Eigenfunctions was used to extract the underlying modes of vibration, and to investigate mechanisms of sustained oscillation. Results were compared with previous findings from excised larynx experiments and theoretical models. .

  7. Optogenetic control of contractile function in skeletal muscle

    PubMed Central

    Bruegmann, Tobias; van Bremen, Tobias; Vogt, Christoph C.; Send, Thorsten; Fleischmann, Bernd K.; Sasse, Philipp

    2015-01-01

    Optogenetic stimulation allows activation of cells with high spatial and temporal precision. Here we show direct optogenetic stimulation of skeletal muscle from transgenic mice expressing the light-sensitive channel Channelrhodopsin-2 (ChR2). Largest tetanic contractions are observed with 5-ms light pulses at 30 Hz, resulting in 84% of the maximal force induced by electrical stimulation. We demonstrate the utility of this approach by selectively stimulating with a light guide individual intralaryngeal muscles in explanted larynges from ChR2-transgenic mice, which enables selective opening and closing of the vocal cords. Furthermore, systemic injection of adeno-associated virus into wild-type mice provides sufficient ChR2 expression for optogenetic opening of the vocal cords. Thus, direct optogenetic stimulation of skeletal muscle generates large force and provides the distinct advantage of localized and cell-type-specific activation. This technology could be useful for therapeutic purposes, such as restoring the mobility of the vocal cords in patients suffering from laryngeal paralysis. PMID:26035411

  8. Pedagogical efficiency of melodic contour mapping technology as it relates to vocal timbre in singers of classical music repertoire.

    PubMed

    Barnes-Burroughs, Kathryn; Anderson, Edward E; Hughes, Thomas; Lan, William Y; Dent, Karl; Arnold, Sue; Dolter, Gerald; McNeil, Kathy

    2007-11-01

    The purpose of this investigation was to ascertain the pedagogical viability of computer-generated melodic contour mapping systems in the classical singing studio, as perceived by their resulting effect (if any) on vocal timbre when a singer's head and neck remained in a normal singing posture. The evaluation of data gathered during the course of the study indicates that the development of consistent vocal timbre produced by the classical singing student may be enhanced through visual/kinesthetic response to melodic contour inversion mapping, as it balances the singer's perception of melodic intervals in standard musical notation. Unexpectedly, it was discovered that the system, in its natural melodic contour mode, may also be useful for teaching a student to sing a consistent legato line. The results of the study also suggest that the continued development of this new technology for the general teaching studio, designed to address standard musical notation and a singer's visual/kinesthetic response to it, may indeed be useful.

  9. Two distinct modes of forebrain circuit dynamics underlie temporal patterning in the vocalizations of young songbirds

    PubMed Central

    Aronov, Dmitriy; Veit, Lena; Goldberg, Jesse H.; Fee, Michale S.

    2011-01-01

    Accurate timing is a critical aspect of motor control, yet the temporal structure of many mature behaviors emerges during learning from highly variable exploratory actions. How does a developing brain acquire the precise control of timing in behavioral sequences? To investigate the development of timing, we analyzed the songs of young juvenile zebra finches. These highly variable vocalizations, akin to human babbling, gradually develop into temporally-stereotyped adult songs. We find that the durations of syllables and silences in juvenile singing are formed by a mixture of two distinct modes of timing – a random mode producing broadly-distributed durations early in development, and a stereotyped mode underlying the gradual emergence of stereotyped durations. Using lesions, inactivations, and localized brain cooling we investigated the roles of neural dynamics within two premotor cortical areas in the production of these temporal modes. We find that LMAN (lateral magnocellular nucleus of the nidopallium) is required specifically for the generation of the random mode of timing, and that mild cooling of LMAN causes an increase in the durations produced by this mode. On the contrary, HVC (used as a proper name) is required specifically for producing the stereotyped mode of timing, and its cooling causes a slowing of all stereotyped components. These results show that two neural pathways contribute to the timing of juvenile songs, and suggest an interesting organization in the forebrain, whereby different brain areas are specialized for the production of distinct forms of neural dynamics. PMID:22072687

  10. Artificially lengthened and constricted vocal tract in vocal training methods.

    PubMed

    Bele, Irene Velsvik

    2005-01-01

    It is common practice in vocal training to make use of vocal exercise techniques that involve partial occlusion of the vocal tract. Various techniques are used; some of them form an occlusion within the front part of the oral cavity or at the lips. Another vocal exercise technique involves lengthening the vocal tract; for example, the method of phonation into small tubes. This essay presents some studies made on the effects of various vocal training methods that involve an artificially lengthened and constricted vocal tract. The influence of sufficient acoustic impedance on vocal fold vibration and economical voice production is presented.

  11. Directed functional connectivity matures with motor learning in a cortical pattern generator.

    PubMed

    Day, Nancy F; Terleski, Kyle L; Nykamp, Duane Q; Nick, Teresa A

    2013-02-01

    Sequential motor skills may be encoded by feedforward networks that consist of groups of neurons that fire in sequence (Abeles 1991; Long et al. 2010). However, there has been no evidence of an anatomic map of activation sequence in motor control circuits, which would be potentially detectable as directed functional connectivity of coactive neuron groups. The proposed pattern generator for birdsong, the HVC (Long and Fee 2008; Vu et al. 1994), contains axons that are preferentially oriented in the rostrocaudal axis (Nottebohm et al. 1982; Stauffer et al. 2012). We used four-tetrode recordings to assess the activity of ensembles of single neurons along the rostrocaudal HVC axis in anesthetized zebra finches. We found an axial, polarized neural network in which sequential activity is directionally organized along the rostrocaudal axis in adult males, who produce a stereotyped song. Principal neurons fired in rostrocaudal order and with interneurons that were rostral to them, suggesting that groups of excitatory neurons fire at the leading edge of travelling waves of inhibition. Consistent with the synchronization of neurons by caudally travelling waves of inhibition, the activity of interneurons was more coherent in the orthogonal mediolateral axis than in the rostrocaudal axis. If directed functional connectivity within the HVC is important for stereotyped, learned song, then it may be lacking in juveniles, which sing a highly variable song. Indeed, we found little evidence for network directionality in juveniles. These data indicate that a functionally directed network within the HVC matures during sensorimotor learning and may underlie vocal patterning.

  12. Directed functional connectivity matures with motor learning in a cortical pattern generator

    PubMed Central

    Day, Nancy F.; Terleski, Kyle L.; Nykamp, Duane Q.

    2013-01-01

    Sequential motor skills may be encoded by feedforward networks that consist of groups of neurons that fire in sequence (Abeles 1991; Long et al. 2010). However, there has been no evidence of an anatomic map of activation sequence in motor control circuits, which would be potentially detectable as directed functional connectivity of coactive neuron groups. The proposed pattern generator for birdsong, the HVC (Long and Fee 2008; Vu et al. 1994), contains axons that are preferentially oriented in the rostrocaudal axis (Nottebohm et al. 1982; Stauffer et al. 2012). We used four-tetrode recordings to assess the activity of ensembles of single neurons along the rostrocaudal HVC axis in anesthetized zebra finches. We found an axial, polarized neural network in which sequential activity is directionally organized along the rostrocaudal axis in adult males, who produce a stereotyped song. Principal neurons fired in rostrocaudal order and with interneurons that were rostral to them, suggesting that groups of excitatory neurons fire at the leading edge of travelling waves of inhibition. Consistent with the synchronization of neurons by caudally travelling waves of inhibition, the activity of interneurons was more coherent in the orthogonal mediolateral axis than in the rostrocaudal axis. If directed functional connectivity within the HVC is important for stereotyped, learned song, then it may be lacking in juveniles, which sing a highly variable song. Indeed, we found little evidence for network directionality in juveniles. These data indicate that a functionally directed network within the HVC matures during sensorimotor learning and may underlie vocal patterning. PMID:23175804

  13. Can vocal conditioning trigger a semiotic ratchet in marmosets?

    PubMed

    Turesson, Hjalmar K; Ribeiro, Sidarta

    2015-01-01

    The complexity of human communication has often been taken as evidence that our language reflects a true evolutionary leap, bearing little resemblance to any other animal communication system. The putative uniqueness of the human language poses serious evolutionary and ethological challenges to a rational explanation of human communication. Here we review ethological, anatomical, molecular, and computational results across several species to set boundaries for these challenges. Results from animal behavior, cognitive psychology, neurobiology, and semiotics indicate that human language shares multiple features with other primate communication systems, such as specialized brain circuits for sensorimotor processing, the capability for indexical (pointing) and symbolic (referential) signaling, the importance of shared intentionality for associative learning, affective conditioning and parental scaffolding of vocal production. The most substantial differences lie in the higher human capacity for symbolic compositionality, fast vertical transmission of new symbols across generations, and irreversible accumulation of novel adaptive behaviors (cultural ratchet). We hypothesize that increasingly-complex vocal conditioning of an appropriate animal model may be sufficient to trigger a semiotic ratchet, evidenced by progressive sign complexification, as spontaneous contact calls become indexes, then symbols and finally arguments (strings of symbols). To test this hypothesis, we outline a series of conditioning experiments in the common marmoset (Callithrix jacchus). The experiments are designed to probe the limits of vocal communication in a prosocial, highly vocal primate 35 million years far from the human lineage, so as to shed light on the mechanisms of semiotic complexification and cultural transmission, and serve as a naturalistic behavioral setting for the investigation of language disorders.

  14. Can vocal conditioning trigger a semiotic ratchet in marmosets?

    PubMed Central

    Turesson, Hjalmar K.; Ribeiro, Sidarta

    2015-01-01

    The complexity of human communication has often been taken as evidence that our language reflects a true evolutionary leap, bearing little resemblance to any other animal communication system. The putative uniqueness of the human language poses serious evolutionary and ethological challenges to a rational explanation of human communication. Here we review ethological, anatomical, molecular, and computational results across several species to set boundaries for these challenges. Results from animal behavior, cognitive psychology, neurobiology, and semiotics indicate that human language shares multiple features with other primate communication systems, such as specialized brain circuits for sensorimotor processing, the capability for indexical (pointing) and symbolic (referential) signaling, the importance of shared intentionality for associative learning, affective conditioning and parental scaffolding of vocal production. The most substantial differences lie in the higher human capacity for symbolic compositionality, fast vertical transmission of new symbols across generations, and irreversible accumulation of novel adaptive behaviors (cultural ratchet). We hypothesize that increasingly-complex vocal conditioning of an appropriate animal model may be sufficient to trigger a semiotic ratchet, evidenced by progressive sign complexification, as spontaneous contact calls become indexes, then symbols and finally arguments (strings of symbols). To test this hypothesis, we outline a series of conditioning experiments in the common marmoset (Callithrix jacchus). The experiments are designed to probe the limits of vocal communication in a prosocial, highly vocal primate 35 million years far from the human lineage, so as to shed light on the mechanisms of semiotic complexification and cultural transmission, and serve as a naturalistic behavioral setting for the investigation of language disorders. PMID:26500583

  15. Improvement of plastic optical fiber microphone based on moisture pattern sensing in devoiced breath

    NASA Astrophysics Data System (ADS)

    Taki, Tomohito; Honma, Satoshi; Morisawa, Masayuki; Muto, Shinzo

    2008-03-01

    Conversation is the most practical and common form in communication. However, people with a verbal handicap feel a difficulty to produce words due to variations in vocal chords. This research leads to develop a new devoiced microphone system based on distinguishes between the moisture patterns for each devoiced breaths, using a plastic optical fiber (POF) moisture sensor. In the experiment, five POF-type moisture sensors with fast response were fabricated by coating swell polymer with a slightly larger refractive index than that of fiber core and were set in front of mouth. When these sensors are exposed into humid air produced by devoiced breath, refractive index in cladding layer decreases by swelling and then the POF sensor heads change to guided type. Based on the above operation principle, the output light intensities from the five sensors set in front of mouth change each other. Using above mentioned output light intensity patterns, discernment of devoiced vowels in Japanese (a,i,u,e,o) was tried by means of DynamicProgramming-Matching (DP-matching) method. As the result, distinction rate over 90% was obtained to Japanese devoiced vowels. Therefore, using this system and a voice synthesizer, development of new microphone for the person with a functional disorder in the vocal chords seems to be possible.

  16. Bioacoustic and multi-locus DNA data of Ninox owls support high incidence of extinction and recolonisation on small, low-lying islands across Wallacea.

    PubMed

    Gwee, Chyi Yin; Christidis, Les; Eaton, James A; Norman, Janette A; Trainor, Colin R; Verbelen, Philippe; Rheindt, Frank E

    2017-04-01

    Known for their rich biodiversity and high level of endemism, the islands of Wallacea serve as natural laboratories for the study of spatio-temporal evolution and patterns of species diversification. Our study focuses on the owl genus Ninox, particularly the Southern Boobook (N. novaeseelandiae) and Moluccan Boobook (N. squamipila) complexes, which are widely distributed across Australasia. We conducted bioacoustic and multi-locus DNA analyses of 24 Ninox owl taxa to evaluate relationships and levels of divergence within the two complexes and ultimately assess the relationship between patterns of taxonomic differentiation and bioclimatic factors. We found that taxa that are vocally and/or genetically distinct from populations on the Australian mainland are found on islands that are significantly larger and higher in altitude than taxa that are vocally and/or genetically indistinct from populations on the Australian mainland. This pattern suggests that taxa occurring on small, low-lying Wallacean islands are likely to be recent colonisers that have dispersed from Australia. Overall, our observations demonstrate that the genus Ninox is likely to have colonised the Wallacean region multiple times as small, low-lying islands undergo frequent extinction, whereas populations on large and high-altitude islands are more resilient. Copyright © 2017 Elsevier Inc. All rights reserved.

  17. Neural Processing of Vocal Emotion and Identity

    ERIC Educational Resources Information Center

    Spreckelmeyer, Katja N.; Kutas, Marta; Urbach, Thomas; Altenmuller, Eckart; Munte, Thomas F.

    2009-01-01

    The voice is a marker of a person's identity which allows individual recognition even if the person is not in sight. Listening to a voice also affords inferences about the speaker's emotional state. Both these types of personal information are encoded in characteristic acoustic feature patterns analyzed within the auditory cortex. In the present…

  18. Linguistic Significance of Babbling: Evidence from a Tracheostomized Infant.

    ERIC Educational Resources Information Center

    Locke, John L.; Pearson, Dawn M.

    1990-01-01

    Examines the phonetic patterns and linguistic development of an infant who was tracheostomized during the period that infants normally begin to produce syllabic vocalization. It was found that the infant had developed only a tenth of the canonical syllables expected in normally developing infants, a small inventory of consonant-like segments, and…

  19. A Survey of the Research on Sex Differences in Nonverbal Communication.

    ERIC Educational Resources Information Center

    Blahna, Loretta J.

    Although the bulk of recent research on nonverbal communication has involved studies of the functions of nonverbal behavior (emotion conveying, regulation, and adaption), a few studies have focused on the differences in nonverbal communication variables between men and women. These differences have been found in vocal patterns, intensities, length…

  20. Vocal and Gestural Productions of 24-Month-Old Children with Sex Chromosome Trisomies

    ERIC Educational Resources Information Center

    Zampini, Laura; Draghi, Lara; Silibello, Gaia; Dall'Ara, Francesca; Rigamonti, Claudia; Suttora, Chiara; Zanchi, Paola; Salerni, Nicoletta; Lalatta, Faustina; Vizziello, Paola

    2018-01-01

    Background: Children with sex chromosome trisomies (SCT) frequently show problems in language development. However, a clear description of the communicative patterns of these children is still lacking. Aims: To describe the first stages of language development in children with SCT in comparison with those in typically developing (TD) children. The…

  1. Prosody Signals the Emergence of Intentional Communication in the First Year of Life: Evidence from Catalan-Babbling Infants

    ERIC Educational Resources Information Center

    Esteve-Gibert, Nuria; Prieto, Pilar

    2013-01-01

    There is considerable debate about whether early vocalizations mimic the target language and whether prosody signals emergent intentional communication. A longitudinal corpus of four Catalan-babbling infants was analyzed to investigate whether children use different prosodic patterns to distinguish communicative from investigative vocalizations…

  2. Young Listeners' Music Style Preferences: Patterns Related to Cultural Identification and Language Use

    ERIC Educational Resources Information Center

    Brittin, Ruth V.

    2014-01-01

    Listeners ("N" = 543) in grades 4, 5, and 6 rated their preference for 10 instrumental and vocal selections from various styles, including four popular music selections with versions performed in English, Spanish, or an Asian language. Participants estimated their identification with Spanish/Hispanic/Latino and Asian cultures, the number…

  3. Control Networks in Paediatric Tourette Syndrome Show Immature and Anomalous Patterns of Functional Connectivity

    ERIC Educational Resources Information Center

    Church, Jessica A.; Fair, Damien A.; Dosenbach, Nico U. F.; Cohen, Alexander L.; Miezin, Francis M.; Petersen, Steven E.; Schlaggar, Bradley L.

    2009-01-01

    Tourette syndrome (TS) is a developmental disorder characterized by unwanted, repetitive behaviours that manifest as stereotyped movements and vocalizations called "tics". Operating under the hypothesis that the brain's control systems may be impaired in TS, we measured resting-state functional connectivity MRI (rs-fcMRI) between 39 previously…

  4. Effect of Acting Experience on Emotion Expression and Recognition in Voice: Non-Actors Provide Better Stimuli than Expected.

    PubMed

    Jürgens, Rebecca; Grass, Annika; Drolet, Matthis; Fischer, Julia

    Both in the performative arts and in emotion research, professional actors are assumed to be capable of delivering emotions comparable to spontaneous emotional expressions. This study examines the effects of acting training on vocal emotion depiction and recognition. We predicted that professional actors express emotions in a more realistic fashion than non-professional actors. However, professional acting training may lead to a particular speech pattern; this might account for vocal expressions by actors that are less comparable to authentic samples than the ones by non-professional actors. We compared 80 emotional speech tokens from radio interviews with 80 re-enactments by professional and inexperienced actors, respectively. We analyzed recognition accuracies for emotion and authenticity ratings and compared the acoustic structure of the speech tokens. Both play-acted conditions yielded similar recognition accuracies and possessed more variable pitch contours than the spontaneous recordings. However, professional actors exhibited signs of different articulation patterns compared to non-trained speakers. Our results indicate that for emotion research, emotional expressions by professional actors are not better suited than those from non-actors.

  5. Interaction between telencephalic signals and respiratory dynamics in songbirds

    PubMed Central

    Méndez, Jorge M.; Mindlin, Gabriel B.

    2012-01-01

    The mechanisms by which telencephalic areas affect motor activities are largely unknown. They could either take over motor control from downstream motor circuits or interact with the intrinsic dynamics of these circuits. Both models have been proposed for telencephalic control of respiration during learned vocal behavior in birds. The interactive model postulates that simple signals from the telencephalic song control areas are sufficient to drive the nonlinear respiratory network into producing complex temporal sequences. We tested this basic assumption by electrically stimulating telencephalic song control areas and analyzing the resulting respiratory patterns in zebra finches and in canaries. We found strong evidence for interaction between the rhythm of stimulation and the intrinsic respiratory rhythm, including naturally emerging subharmonic behavior and integration of lateralized telencephalic input. The evidence for clear interaction in our experimental paradigm suggests that telencephalic vocal control also uses a similar mechanism. Furthermore, species differences in the response of the respiratory system to stimulation show parallels to differences in the respiratory patterns of song, suggesting that the interactive production of respiratory rhythms is manifested in species-specific specialization of the involved circuitry. PMID:22402649

  6. Preschoolers' real-time coordination of vocal and facial emotional information.

    PubMed

    Berman, Jared M J; Chambers, Craig G; Graham, Susan A

    2016-02-01

    An eye-tracking methodology was used to examine the time course of 3- and 5-year-olds' ability to link speech bearing different acoustic cues to emotion (i.e., happy-sounding, neutral, and sad-sounding intonation) to photographs of faces reflecting different emotional expressions. Analyses of saccadic eye movement patterns indicated that, for both 3- and 5-year-olds, sad-sounding speech triggered gaze shifts to a matching (sad-looking) face from the earliest moments of speech processing. However, it was not until approximately 800ms into a happy-sounding utterance that preschoolers began to use the emotional cues from speech to identify a matching (happy-looking) face. Complementary analyses based on conscious/controlled behaviors (children's explicit points toward the faces) indicated that 5-year-olds, but not 3-year-olds, could successfully match happy-sounding and sad-sounding vocal affect to a corresponding emotional face. Together, the findings clarify developmental patterns in preschoolers' implicit versus explicit ability to coordinate emotional cues across modalities and highlight preschoolers' greater sensitivity to sad-sounding speech as the auditory signal unfolds in time. Copyright © 2015 Elsevier Inc. All rights reserved.

  7. Effect of Levodopa + Carbidopa on the Laryngeal Electromyographic Pattern in Parkinson Disease.

    PubMed

    Noffs, Gustavo; de Campos Duprat, André; Zarzur, Ana Paula; Cury, Rubens Gisbert; Cataldo, Berenice Oliveira; Fonoff, Erich

    2017-05-01

    Vocal impairment is one of the main debilitating symptoms of Parkinson disease (PD). The effect of levodopa on vocal function remains unclear. This study aimed to determine the effect of levodopa on electromyographic patterns of the laryngeal muscle in patients with PD. This is a prospective interventional trial. Nineteen patients with PD-diagnosed by laryngeal electromyography-were enrolled. Cricothyroid and thyroarytenoid (TA) muscle activities were measured at rest and during muscle contraction (phonation), when participants were on and off medication (12 hours after the last levodopa dose). Prevalence of resting hypertonia in the cricothyroid muscle was similar in the off and on states (7 of 19, P = 1.00). Eight patients off medication and four patients on medication had hypertonic TA muscle at rest (P = 0.289). No electromyographic alterations were observed during phonation for either medication states. Despite a tendency for increased rest tracings in the TA muscle when participants were on medication, no association was found between laryngeal electromyography findings and levodopa + carbidopa administration. Copyright © 2017. Published by Elsevier Inc.

  8. The effects of preventive vocal hygiene education on the vocal hygiene habits and perceptual vocal characteristics of training singers.

    PubMed

    Broaddus-Lawrence, P L; Treole, K; McCabe, R B; Allen, R L; Toppin, L

    2000-03-01

    The purpose of the present study was to determine the effects of vocal hygiene education on the vocal hygiene behaviors and perceptual vocal characteristics of untrained singers. Eleven adult untrained singers served as subjects. They attended four 1-hour class sessions on vocal hygiene, including anatomy and physiology of the phonatory mechanism, vocally abusive behaviors, voice disorders commonly seen in singers, and measures to prevent voice disorders. Pre- and postinstruction surveys were used to record subjects' vocal abuses and their perceptions of their speaking and singing voice. They also rated their perceived value of vocal hygiene education. Results revealed minimal changes in vocal hygiene behaviors and perceptual voice characteristics. The subjects did report a high degree of benefit and learning, however.

  9. Acoustic Analysis and Electroglottography in Elite Vocal Performers.

    PubMed

    Villafuerte-Gonzalez, Rocio; Valadez-Jimenez, Victor M; Sierra-Ramirez, Jose A; Ysunza, Pablo Antonio; Chavarria-Villafuerte, Karen; Hernandez-Lopez, Xochiquetzal

    2017-05-01

    Acoustic analysis of voice (AAV) and electroglottography (EGG) have been used for assessing vocal quality in patients with voice disorders. The effectiveness of these procedures for detecting mild disturbances in vocal quality in elite vocal performers has been controversial. To compare acoustic parameters obtained by AAV and EGG before and after vocal training to determine the effectiveness of these procedures for detecting vocal improvements in elite vocal performers. Thirty-three elite vocal performers were studied. The study group included 14 males and 19 females, ages 18-40 years, without a history of voice disorders. Acoustic parameters were obtained through AAV and EGG before and after vocal training using the Linklater method. Nonsignificant differences (P > 0.05) were found between values of fundamental frequency (F 0 ), shimmer, and jitter obtained by both procedures before vocal training. Mean F 0 was similar after vocal training. Jitter percentage as measured by AAV showed nonsignificant differences (P > 0.05) before and after vocal training. Shimmer percentage as measured by AAV demonstrated a significant reduction (P < 0.05) after vocal training. As measured by EGG after vocal training, shimmer and jitter were significantly reduced (P < 0.05); open quotient was significantly increased (P < 0.05); and irregularity was significantly reduced (P < 0.05). AAV and EGG were effective for detecting improvements in vocal function after vocal training in male and female elite vocal performers undergoing vocal training. EGG demonstrated better efficacy for detecting improvements and provided additional parameters as compared to AAV. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  10. Functional Connectivity Associated with Acoustic Stability During Vowel Production: Implications for Vocal-Motor Control

    PubMed Central

    2015-01-01

    Abstract Vowels provide the acoustic foundation of communication through speech and song, but little is known about how the brain orchestrates their production. Positron emission tomography was used to study regional cerebral blood flow (rCBF) during sustained production of the vowel /a/. Acoustic and blood flow data from 13, normal, right-handed, native speakers of American English were analyzed to identify CBF patterns that predicted the stability of the first and second formants of this vowel. Formants are bands of resonance frequencies that provide vowel identity and contribute to voice quality. The results indicated that formant stability was directly associated with blood flow increases and decreases in both left- and right-sided brain regions. Secondary brain regions (those associated with the regions predicting formant stability) were more likely to have an indirect negative relationship with first formant variability, but an indirect positive relationship with second formant variability. These results are not definitive maps of vowel production, but they do suggest that the level of motor control necessary to produce stable vowels is reflected in the complexity of an underlying neural system. These results also extend a systems approach to functional image analysis, previously applied to normal and ataxic speech rate that is solely based on identifying patterns of brain activity associated with specific performance measures. Understanding the complex relationships between multiple brain regions and the acoustic characteristics of vocal stability may provide insight into the pathophysiology of the dysarthrias, vocal disorders, and other speech changes in neurological and psychiatric disorders. PMID:25295385

  11. Northern Elephant Seals Memorize the Rhythm and Timbre of Their Rivals' Voices.

    PubMed

    Mathevon, Nicolas; Casey, Caroline; Reichmuth, Colleen; Charrier, Isabelle

    2017-08-07

    The evolutionary origin of rhythm perception, a cognitive ability essential to musicality, remains unresolved [1-5]. The ability to perceive and memorize rhythmic sounds is widely shared among humans [6] but seems rare among other mammals [7, 8]. Although the perception of temporal metrical patterns has been found in a few species, this ability has only been demonstrated through behavioral training [9] (but see [10] for an example of spontaneous tempo coordination in a bonobo), and there is no experimental evidence to indicate its biological function. Furthermore, there is no example of a non-human mammal able to remember and recognize auditory rhythmic patterns among a wide range of tempi. In the northern elephant seal Mirounga angustirostris, the calls of mature males comprise a rhythmic series of pulses, with the call of each individual characterized by its tempo and timbre; these individual vocal signatures are stable over years and across contexts [11]. Here, we report that northern elephant seal males routinely memorize and recognize the unique tempo and timbre of their rivals' voices and use this rhythmic information to individually identify competitors, which facilitates navigation within the social network of the rookery. By performing playbacks with natural and modified vocalizations, we show that males are sensitive to call rhythm disruption independently of modification of spectral features and that they use both temporal and spectral cues to identify familiar rivals. While spectral features of calls typically encode individual identity in mammalian vocalizations [12], this is the first example of this phenomenon involving sound rhythm. Copyright © 2017 Elsevier Ltd. All rights reserved.

  12. An agent-based model of dialect evolution in killer whales.

    PubMed

    Filatova, Olga A; Miller, Patrick J O

    2015-05-21

    The killer whale is one of the few animal species with vocal dialects that arise from socially learned group-specific call repertoires. We describe a new agent-based model of killer whale populations and test a set of vocal-learning rules to assess which mechanisms may lead to the formation of dialect groupings observed in the wild. We tested a null model with genetic transmission and no learning, and ten models with learning rules that differ by template source (mother or matriline), variation type (random errors or innovations) and type of call change (no divergence from kin vs. divergence from kin). The null model without vocal learning did not produce the pattern of group-specific call repertoires we observe in nature. Learning from either mother alone or the entire matriline with calls changing by random errors produced a graded distribution of the call phenotype, without the discrete call types observed in nature. Introducing occasional innovation or random error proportional to matriline variance yielded more or less discrete and stable call types. A tendency to diverge from the calls of related matrilines provided fast divergence of loose call clusters. A pattern resembling the dialect diversity observed in the wild arose only when rules were applied in combinations and similar outputs could arise from different learning rules and their combinations. Our results emphasize the lack of information on quantitative features of wild killer whale dialects and reveal a set of testable questions that can draw insights into the cultural evolution of killer whale dialects. Copyright © 2015 Elsevier Ltd. All rights reserved.

  13. Recognizing vocal emotions in Mandarin Chinese: a validated database of Chinese vocal emotional stimuli.

    PubMed

    Liu, Pan; Pell, Marc D

    2012-12-01

    To establish a valid database of vocal emotional stimuli in Mandarin Chinese, a set of Chinese pseudosentences (i.e., semantically meaningless sentences that resembled real Chinese) were produced by four native Mandarin speakers to express seven emotional meanings: anger, disgust, fear, sadness, happiness, pleasant surprise, and neutrality. These expressions were identified by a group of native Mandarin listeners in a seven-alternative forced choice task, and items reaching a recognition rate of at least three times chance performance in the seven-choice task were selected as a valid database and then subjected to acoustic analysis. The results demonstrated expected variations in both perceptual and acoustic patterns of the seven vocal emotions in Mandarin. For instance, fear, anger, sadness, and neutrality were associated with relatively high recognition, whereas happiness, disgust, and pleasant surprise were recognized less accurately. Acoustically, anger and pleasant surprise exhibited relatively high mean f0 values and large variation in f0 and amplitude; in contrast, sadness, disgust, fear, and neutrality exhibited relatively low mean f0 values and small amplitude variations, and happiness exhibited a moderate mean f0 value and f0 variation. Emotional expressions varied systematically in speech rate and harmonics-to-noise ratio values as well. This validated database is available to the research community and will contribute to future studies of emotional prosody for a number of purposes. To access the database, please contact pan.liu@mail.mcgill.ca.

  14. Reinforcement of Infant Vocalizations through Contingent Vocal Imitation

    ERIC Educational Resources Information Center

    Pelaez, Martha; Virues-Ortega, Javier; Gewirtz, Jacob L.

    2011-01-01

    Maternal vocal imitation of infant vocalizations is highly prevalent during face-to-face interactions of infants and their caregivers. Although maternal vocal imitation has been associated with later verbal development, its potentially reinforcing effect on infant vocalizations has not been explored experimentally. This study examined the…

  15. Limiting parental interaction during vocal development affects acoustic call structure in marmoset monkeys

    PubMed Central

    2018-01-01

    Human vocal development is dependent on learning by imitation through social feedback between infants and caregivers. Recent studies have revealed that vocal development is also influenced by parental feedback in marmoset monkeys, suggesting vocal learning mechanisms in nonhuman primates. Marmoset infants that experience more contingent vocal feedback than their littermates develop vocalizations more rapidly, and infant marmosets with limited parental interaction exhibit immature vocal behavior beyond infancy. However, it is yet unclear whether direct parental interaction is an obligate requirement for proper vocal development because all monkeys in the aforementioned studies were able to produce the adult call repertoire after infancy. Using quantitative measures to compare distinct call parameters and vocal sequence structure, we show that social interaction has a direct impact not only on the maturation of the vocal behavior but also on acoustic call structures during vocal development. Monkeys with limited parental interaction during development show systematic differences in call entropy, a measure for maturity, compared with their normally raised siblings. In addition, different call types were occasionally uttered in motif-like sequences similar to those exhibited by vocal learners, such as birds and humans, in early vocal development. These results indicate that a lack of parental interaction leads to long-term disturbances in the acoustic structure of marmoset vocalizations, suggesting an imperative role for social interaction in proper primate vocal development. PMID:29651461

  16. Limiting parental interaction during vocal development affects acoustic call structure in marmoset monkeys.

    PubMed

    Gultekin, Yasemin B; Hage, Steffen R

    2018-04-01

    Human vocal development is dependent on learning by imitation through social feedback between infants and caregivers. Recent studies have revealed that vocal development is also influenced by parental feedback in marmoset monkeys, suggesting vocal learning mechanisms in nonhuman primates. Marmoset infants that experience more contingent vocal feedback than their littermates develop vocalizations more rapidly, and infant marmosets with limited parental interaction exhibit immature vocal behavior beyond infancy. However, it is yet unclear whether direct parental interaction is an obligate requirement for proper vocal development because all monkeys in the aforementioned studies were able to produce the adult call repertoire after infancy. Using quantitative measures to compare distinct call parameters and vocal sequence structure, we show that social interaction has a direct impact not only on the maturation of the vocal behavior but also on acoustic call structures during vocal development. Monkeys with limited parental interaction during development show systematic differences in call entropy, a measure for maturity, compared with their normally raised siblings. In addition, different call types were occasionally uttered in motif-like sequences similar to those exhibited by vocal learners, such as birds and humans, in early vocal development. These results indicate that a lack of parental interaction leads to long-term disturbances in the acoustic structure of marmoset vocalizations, suggesting an imperative role for social interaction in proper primate vocal development.

  17. Do you hear what I see? Vocalization relative to visual detection rates of Hawaiian hoary bats (Lasiurus cinereus semotus)

    USGS Publications Warehouse

    Gorresen, Paulo Marcos; Cryan, Paul; Montoya-Aiona, Kristina; Bonaccorso, Frank

    2017-01-01

    Bats vocalize during flight as part of the sensory modality called echolocation, but very little is known about whether flying bats consistently call. Occasional vocal silence during flight when bats approach prey or conspecifics has been documented for relatively few species and situations. Bats flying alone in clutter-free airspace are not known to forgo vocalization, yet prior observations suggested possible silent behavior in certain, unexpected situations. Determining when, why, and where silent behavior occurs in bats will help evaluate major assumptions of a primary monitoring method for bats used in ecological research, management, and conservation. In this study, we recorded flight activity of Hawaiian hoary bats (Lasiurus cinereus semotus) under seminatural conditions using both thermal video cameras and acoustic detectors. Simultaneous video and audio recordings from 20 nights of observation at 10 sites were analyzed for correspondence between detection methods, with a focus on video observations in three distance categories for which accompanying vocalizations were detected. Comparison of video and audio detections revealed that a high proportion of Hawaiian hoary bats “seen” on video were not simultaneously “heard.” On average, only about one in three visual detections within a night had an accompanying call detection, but this varied greatly among nights. Bats flying on curved flight paths and individuals nearer the cameras were more likely to be detected by both methods. Feeding and social calls were detected, but no clear pattern emerged from the small number of observations involving closely interacting bats. These results may indicate that flying Hawaiian hoary bats often forgo echolocation, or do not always vocalize in a way that is detectable with common sampling and monitoring methods. Possible reasons for the low correspondence between visual and acoustic detections range from methodological to biological and include a number of biases associated with the propagation and detection of sound, cryptic foraging strategies, or conspecific presence. Silent flight behavior may be more prevalent in echolocating bats than previously appreciated, has profound implications for ecological research, and deserves further characterization and study.

  18. The effect of resonance tubes on glottal contact quotient with and without task instruction: a comparison of trained and untrained voices.

    PubMed

    Gaskill, Christopher S; Quinney, Dana M

    2012-05-01

    Phonation into narrow tubes or straws has been used as a voice training and voice therapy technique and belongs to a group of techniques known as semi-occluded vocal tract exercises. The use of what are called resonance tubes has received renewed attention in the voice research literature, in both theoretical and empirical studies. The assumption is that the partially occluded and lengthened vocal tract alters supraglottal acoustics in such a way as to allow phonation near a lowered first vocal tract formant, which has been suggested as a way to bring about a more efficient glottal closure pattern for sustained oscillation. In this study, two groups of male participants, 10 with no vocal training and 10 with classical vocal training, phonated into a resonance tube for approximately 1 minute. Electroglottography was used to estimate glottal contact quotient (CQ) during spoken /a/ vowels before tube phonation, during tube phonation, and again during spoken /a/ vowels after tube phonation. Half of each group of participants was made to keep pitch and loudness consistent for all phases of the experiment, replicating the method of a previous study by this author. The other half was instructed to practice phonating into the resonance tube before collecting data and was encouraged to find a pitch and loudness combination that maximized ease of phonation and a sense of forward oral resonance. Glottal CQ altered considerably from baseline for almost all participants during tube phonation, with a larger variability than that during vowel production. Small differences in glottal CQ were found as a function of training and instruction, with most participants' CQ increasing during tube phonation. A small post-tube phonation effect was found primarily for the trained and instructed group. Secondary single-subject analyses revealed large intersubject variation, highlighting the highly individualized response to the resonance tube task. Continued study of resonance tubes is recommended, comparing both male and female as well as vocally trained and untrained participants. Future studies should continue to examine systematic variations in task instruction, length of practice, and resonance tube dimensions. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  19. Histopathologic study of human vocal fold mucosa unphonated over a decade.

    PubMed

    Sato, Kiminori; Umeno, Hirohito; Ono, Takeharu; Nakashima, Tadashi

    2011-12-01

    Mechanotransduction caused by vocal fold vibration could possibly be an important factor in the maintenance of extracellular matrices and layered structure of the human adult vocal fold mucosa as a vibrating tissue after the layered structure has been completed. Vocal fold stellate cells (VFSCs) in the human maculae flavae of the vocal fold mucosa are inferred to be involved in the metabolism of extracellular matrices of the vocal fold mucosa. Maculae flavae are also considered to be an important structure in the growth and development of the human vocal fold mucosa. Tension caused by phonation (vocal fold vibration) is hypothesized to stimulate the VFSCs to accelerate production of extracellular matrices. A human adult vocal fold mucosa unphonated over a decade was investigated histopathologically. Vocal fold mucosa unphonated for 11 years and 2 months of a 64-year-old male with cerebral hemorrhage was investigated by light and electron microscopy. The vocal fold mucosae (including maculae flavae) were atrophic. The vocal fold mucosa did not have a vocal ligament, Reinke's space or a layered structure. The lamina propria appeared as a uniform structure. Morphologically, the VFSCs synthesized fewer extracellular matrices, such as fibrous protein and glycosaminoglycan. Consequently, VFSCs appeared to decrease their level of activity.

  20. The Vocal Repertoire of Adult and Neonate Giant Otters (Pteronura brasiliensis)

    PubMed Central

    Mumm, Christina A. S.; Knörnschild, Mirjam

    2014-01-01

    Animals use vocalizations to exchange information about external events, their own physical or motivational state, or about individuality and social affiliation. Infant babbling can enhance the development of the full adult vocal repertoire by providing ample opportunity for practice. Giant otters are very social and frequently vocalizing animals. They live in highly cohesive groups, generally including a reproductive pair and their offspring born in different years. This basic social structure may vary in the degree of relatedness of the group members. Individuals engage in shared group activities and different social roles and thus, the social organization of giant otters provides a basis for complex and long-term individual relationships. We recorded and analysed the vocalizations of adult and neonate giant otters from wild and captive groups. We classified the adult vocalizations according to their acoustic structure, and described their main behavioural context. Additionally, we present the first description of vocalizations uttered in babbling bouts of new born giant otters. We expected to find 1) a sophisticated vocal repertoire that would reflect the species’ complex social organisation, 2) that giant otter vocalizations have a clear relationship between signal structure and function, and 3) that the vocal repertoire of new born giant otters would comprise age-specific vocalizations as well as precursors of the adult repertoire. We found a vocal repertoire with 22 distinct vocalization types produced by adults and 11 vocalization types within the babbling bouts of the neonates. A comparison within the otter subfamily suggests a relation between vocal and social complexity, with the giant otters being the socially and vocally most complex species. PMID:25391142

  1. Asymmetric vibration in a two-layer vocal fold model with left-right stiffness asymmetry: Experiment and simulation

    PubMed Central

    Zhang, Zhaoyan; Hieu Luu, Trung

    2012-01-01

    Vibration characteristics of a self-oscillating two-layer vocal fold model with left-right asymmetry in body-layer stiffness were experimentally and numerically investigated. Two regimes of distinct vibratory pattern were identified as a function of left-right stiffness mismatch. In the first regime with extremely large left-right stiffness mismatch, phonation onset resulted from an eigenmode synchronization process that involved only eigenmodes of the soft fold. Vocal fold vibration in this regime was dominated by a large-amplitude vibration of the soft fold, and phonation frequency was determined by the properties of the soft fold alone. The stiff fold was only enslaved to vibrate at a much reduced amplitude. In the second regime with small left-right stiffness mismatch, eigenmodes of both folds actively participated in the eigenmode synchronization process. The two folds vibrated with comparable amplitude, but the stiff fold consistently led the soft fold in phase for all conditions. A qualitatively good agreement was obtained between experiment and simulation, although the simulations generally underestimated phonation threshold pressure and onset frequency. The clinical implications of the results of this study are also discussed. PMID:22978891

  2. Asymmetric vibration in a two-layer vocal fold model with left-right stiffness asymmetry: experiment and simulation.

    PubMed

    Zhang, Zhaoyan; Luu, Trung Hieu

    2012-09-01

    Vibration characteristics of a self-oscillating two-layer vocal fold model with left-right asymmetry in body-layer stiffness were experimentally and numerically investigated. Two regimes of distinct vibratory pattern were identified as a function of left-right stiffness mismatch. In the first regime with extremely large left-right stiffness mismatch, phonation onset resulted from an eigenmode synchronization process that involved only eigenmodes of the soft fold. Vocal fold vibration in this regime was dominated by a large-amplitude vibration of the soft fold, and phonation frequency was determined by the properties of the soft fold alone. The stiff fold was only enslaved to vibrate at a much reduced amplitude. In the second regime with small left-right stiffness mismatch, eigenmodes of both folds actively participated in the eigenmode synchronization process. The two folds vibrated with comparable amplitude, but the stiff fold consistently led the soft fold in phase for all conditions. A qualitatively good agreement was obtained between experiment and simulation, although the simulations generally underestimated phonation threshold pressure and onset frequency. The clinical implications of the results of this study are also discussed.

  3. Multivariate sensitivity to voice during auditory categorization.

    PubMed

    Lee, Yune Sang; Peelle, Jonathan E; Kraemer, David; Lloyd, Samuel; Granger, Richard

    2015-09-01

    Past neuroimaging studies have documented discrete regions of human temporal cortex that are more strongly activated by conspecific voice sounds than by nonvoice sounds. However, the mechanisms underlying this voice sensitivity remain unclear. In the present functional MRI study, we took a novel approach to examining voice sensitivity, in which we applied a signal detection paradigm to the assessment of multivariate pattern classification among several living and nonliving categories of auditory stimuli. Within this framework, voice sensitivity can be interpreted as a distinct neural representation of brain activity that correctly distinguishes human vocalizations from other auditory object categories. Across a series of auditory categorization tests, we found that bilateral superior and middle temporal cortex consistently exhibited robust sensitivity to human vocal sounds. Although the strongest categorization was in distinguishing human voice from other categories, subsets of these regions were also able to distinguish reliably between nonhuman categories, suggesting a general role in auditory object categorization. Our findings complement the current evidence of cortical sensitivity to human vocal sounds by revealing that the greatest sensitivity during categorization tasks is devoted to distinguishing voice from nonvoice categories within human temporal cortex. Copyright © 2015 the American Physiological Society.

  4. The devil is in the detail: Quantifying vocal variation in a complex, multi-levelled, and rapidly evolving display.

    PubMed

    Garland, Ellen C; Rendell, Luke; Lilley, Matthew S; Poole, M Michael; Allen, Jenny; Noad, Michael J

    2017-07-01

    Identifying and quantifying variation in vocalizations is fundamental to advancing our understanding of processes such as speciation, sexual selection, and cultural evolution. The song of the humpback whale (Megaptera novaeangliae) presents an extreme example of complexity and cultural evolution. It is a long, hierarchically structured vocal display that undergoes constant evolutionary change. Obtaining robust metrics to quantify song variation at multiple scales (from a sound through to population variation across the seascape) is a substantial challenge. Here, the authors present a method to quantify song similarity at multiple levels within the hierarchy. To incorporate the complexity of these multiple levels, the calculation of similarity is weighted by measurements of sound units (lower levels within the display) to bridge the gap in information between upper and lower levels. Results demonstrate that the inclusion of weighting provides a more realistic and robust representation of song similarity at multiple levels within the display. This method permits robust quantification of cultural patterns and processes that will also contribute to the conservation management of endangered humpback whale populations, and is applicable to any hierarchically structured signal sequence.

  5. Vocal fold vibration and voice source aperiodicity in 'dist' tones: a study of a timbral ornament in rock singing.

    PubMed

    Borch, D Zangger; Sundberg, J; Lindestad, P A; Thalén, M

    2004-01-01

    The acoustic characteristics of so-called 'dist' tones, commonly used in singing rock music, are analyzed in a case study. In an initial experiment a professional rock singer produced examples of 'dist' tones. The tones were found to contain aperiodicity, SPL at 0.3 m varied between 90 and 96 dB, and subglottal pressure varied in the range of 20-43 cm H2O, a doubling yielding, on average, an SPL increase of 2.3 dB. In a second experiment, the associated vocal fold vibration patterns were recorded by digital high-speed imaging of the same singer. Inverse filtering of the simultaneously recorded audio signal showed that the aperiodicity was caused by a low frequency modulation of the flow glottogram pulse amplitude. This modulation was produced by an aperiodic or periodic vibration of the supraglottic mucosa. This vibration reduced the pulse amplitude by obstructing the airway for some of the pulses produced by the apparently periodically vibrating vocal folds. The supraglottic mucosa vibration can be assumed to be driven by the high airflow produced by the elevated subglottal pressure.

  6. Applicability of Cone Beam Computed Tomography to the Assessment of the Vocal Tract before and after Vocal Exercises in Normal Subjects.

    PubMed

    Garcia, Elisângela Zacanti; Yamashita, Hélio Kiitiro; Garcia, Davi Sousa; Padovani, Marina Martins Pereira; Azevedo, Renata Rangel; Chiari, Brasília Maria

    2016-01-01

    Cone beam computed tomography (CBCT), which represents an alternative to traditional computed tomography and magnetic resonance imaging, may be a useful instrument to study vocal tract physiology related to vocal exercises. This study aims to evaluate the applicability of CBCT to the assessment of variations in the vocal tract of healthy individuals before and after vocal exercises. Voice recordings and CBCT images before and after vocal exercises performed by 3 speech-language pathologists without vocal complaints were collected and compared. Each participant performed 1 type of exercise, i.e., Finnish resonance tube technique, prolonged consonant "b" technique, or chewing technique. The analysis consisted of an acoustic analysis and tomographic imaging. Modifications of the vocal tract settings following vocal exercises were properly detected by CBCT, and changes in the acoustic parameters were, for the most part, compatible with the variations detected in image measurements. CBCT was shown to be capable of properly assessing the changes in vocal tract settings promoted by vocal exercises. © 2017 S. Karger AG, Basel.

  7. Dependence of phonation threshold pressure on vocal tract acoustics and vocal fold tissue mechanics.

    PubMed

    Chan, Roger W; Titze, Ingo R

    2006-04-01

    Analytical and computer simulation studies have shown that the acoustic impedance of the vocal tract as well as the viscoelastic properties of vocal fold tissues are critical for determining the dynamics and the energy transfer mechanism of vocal fold oscillation. In the present study, a linear, small-amplitude oscillation theory was revised by taking into account the propagation of a mucosal wave and the inertive reactance (inertance) of the supraglottal vocal tract as the major energy transfer mechanisms for flow-induced self-oscillation of the vocal fold. Specifically, analytical results predicted that phonation threshold pressure (Pth) increases with the viscous shear properties of the vocal fold, but decreases with vocal tract inertance. This theory was empirically tested using a physical model of the larynx, where biological materials (fat, hyaluronic acid, and fibronectin) were implanted into the vocal fold cover to investigate the effect of vocal fold tissue viscoelasticity on Pth. A uniform-tube supraglottal vocal tract was also introduced to examine the effect of vocal tract inertance on Pth. Results showed that Pth decreased with the inertive impedance of the vocal tract and increased with the viscous shear modulus (G") or dynamic viscosity (eta') of the vocal fold cover, consistent with theoretical predictions. These findings supported the potential biomechanical benefits of hyaluronic acid as a surgical bioimplant for repairing voice disorders involving the superficial layer of the lamina propria, such as scarring, sulcus vocalis, atrophy, and Reinke's edema.

  8. Noise Pollution Filters Bird Communities Based on Vocal Frequency

    PubMed Central

    Francis, Clinton D.; Ortega, Catherine P.; Cruz, Alexander

    2011-01-01

    Background Human-generated noise pollution now permeates natural habitats worldwide, presenting evolutionarily novel acoustic conditions unprecedented to most landscapes. These acoustics not only harm humans, but threaten wildlife, and especially birds, via changes to species densities, foraging behavior, reproductive success, and predator-prey interactions. Explanations for negative effects of noise on birds include disruption of acoustic communication through energetic masking, potentially forcing species that rely upon acoustic communication to abandon otherwise suitable areas. However, this hypothesis has not been adequately tested because confounding stimuli often co-vary with noise and are difficult to separate from noise exposure. Methodology/Principal Findings Using a natural experiment that controls for confounding stimuli, we evaluate whether species vocal features or urban-tolerance classifications explain their responses to noise measured through habitat use. Two data sets representing nesting and abundance responses reveal that noise filters bird communities nonrandomly. Signal duration and urban tolerance failed to explain species-specific responses, but birds with low-frequency signals that are more susceptible to masking from noise avoided noisy areas and birds with higher frequency vocalizations remained. Signal frequency was also negatively correlated with body mass, suggesting that larger birds may be more sensitive to noise due to the link between body size and vocal frequency. Conclusions/Significance Our findings suggest that acoustic masking by noise may be a strong selective force shaping the ecology of birds worldwide. Larger birds with lower frequency signals may be excluded from noisy areas, whereas smaller species persist via transmission of higher frequency signals. We discuss our findings as they relate to interspecific relationships among body size, vocal amplitude and frequency and suggest that they are immediately relevant to the global problem of increases in noise by providing critical insight as to which species traits influence tolerance of these novel acoustics. PMID:22096517

  9. Psychopathology in a Swedish Population of School Children with Tic Disorders

    ERIC Educational Resources Information Center

    Khalifa, Najah; Von Knorring, Anne-Liis

    2006-01-01

    Objective: To examine patterns of psychiatric comorbid disorders and associated problems in a school population of children with tic disorders. Method: From a total population of 4,479 children, 25 with Tourette's disorder (TD), 34 with chronic motor tics (CMT), 24 with chronic vocal tics (CVT), and 214 with transient tics (TT) during the past…

  10. Functional Analysis and Treatment of Arranging and Ordering by Individuals with an Autism Spectrum Disorder

    ERIC Educational Resources Information Center

    Rodriguez, Nicole M.; Thompson, Rachel H.; Schlichenmeyer, Kevin; Stocco, Corey S.

    2012-01-01

    Of the diagnostic features of autism, relatively little research has been devoted to restricted and repetitive behavior, particularly topographically complex forms of restricted and repetitive behavior such as rigidity in routines or compulsive-like behavior (e.g., arranging objects in patterns or rows). Like vocal or motor stereotypy,…

  11. The role of auditory and kinaesthetic feedback mechanisms on phonatory stability in children.

    PubMed

    Rathna Kumar, S B; Azeem, Suhail; Choudhary, Abhishek Kumar; Prakash, S G R

    2013-12-01

    Auditory feedback plays an important role in phonatory control. When auditory feedback is disrupted, various changes are observed in vocal motor control. Vocal intensity and fundamental frequency (F0) levels tend to increase in response to auditory masking. Because of the close reflexive links between the auditory and phonatory systems, it is likely that phonatory stability may be disrupted when auditory feedback is disrupted or altered. However, studies on phonatory stability under auditory masking condition in adult subjects showed that most of the subjects maintained normal levels of phonatory stability. The authors in the earlier investigations suggested that auditory feedback is not the sole contributor to vocal motor control and phonatory stability, a complex neuromuscular reflex system known as kinaesthetic feedback may play a role in controlling phonatory stability when auditory feedback is disrupted or lacking. This proposes the need to further investigate this phenomenon as to whether children show similar patterns of phonatory stability under auditory masking since their neuromotor systems are still at developmental stage, less mature and are less resistant to altered auditory feedback than adults. A total of 40 normal hearing and speaking children (20 male and 20 female) between the age group of 6 and 8 years participated as subjects. The acoustic parameters such as shimmer, jitter and harmonic-to-noise ratio (HNR) were measures and compared between no masking condition (0 dB ML) and masking condition (90 dB ML). Despite the neuromotor systems being less mature in children and less resistant than adults to altered auditory feedback, most of the children in the study demonstrated increased phonatory stability which was reflected by reduced shimmer, jitter and increased HNR values. This study implicates that most of the children demonstrate well established patterns of kinaesthetic feedback, which might have allowed them to maintain normal levels of vocal motor control even in the presence of disturbed auditory feedback. Hence, it can be concluded that children also exhibit kinaesthetic feedback mechanism to control phonatory stability when auditory feedback is disrupted which in turn highlights the importance of kinaesthetic feedback to be included in the therapeutic/intervention approaches for children with hearing and neurogenic speech deficits.

  12. The objective vocal quality, vocal risk factors, vocal complaints, and corporal pain in Dutch female students training to be speech-language pathologists during the 4 years of study.

    PubMed

    Van Lierde, Kristiane M; D'haeseleer, Evelien; Wuyts, Floris L; De Ley, Sophia; Geldof, Ruben; De Vuyst, Julie; Sofie, Claeys

    2010-09-01

    The purpose of the present cross-sectional study was to determine the objective vocal quality and the vocal characteristics (vocal risk factors, vocal and corporal complaints) in 197 female students in speech-language pathology during the 4 years of study. The objective vocal quality was measured by means of the Dysphonia Severity Index (DSI). Perceptual voice assessment, the Voice Handicap Index (VHI), questionnaires addressing vocal risks, and vocal and corporal complaints during and/or after voice usage were performed. Speech-language pathology (SLP) students have a borderline vocal quality corresponding to a DSI% of 68. The analysis of variance revealed no significant change of the objective vocal quality between the first bachelor year and the master year. No psychosocial handicapping effect of the voice was observed by means of the VHI total, though there was an effect at the functional VHI level in addition to some vocal complaints. Ninety-three percent of the student SLPs reported the presence of corporal pain during and/or after speaking. In particular, sore throat and headache were mentioned as the prevalent corporal pain symptoms. A longitudinal study of the objective vocal quality of the same subjects during their career as an SLP might provide new insights. 2010 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  13. University Vocal Training and Vocal Health of Music Educators and Music Therapists

    ERIC Educational Resources Information Center

    Baker, Vicki D.; Cohen, Nicki

    2017-01-01

    The purpose of this study was to describe the university vocal training and vocal health of music educators and music therapists. The participants (N = 426), music educators (n = 351) and music therapists (n = 75), completed a survey addressing demographics, vocal training, voice usage, and vocal health. Both groups reported singing at least 50%…

  14. Monkey vocal tracts are speech-ready.

    PubMed

    Fitch, W Tecumseh; de Boer, Bart; Mathur, Neil; Ghazanfar, Asif A

    2016-12-01

    For four decades, the inability of nonhuman primates to produce human speech sounds has been claimed to stem from limitations in their vocal tract anatomy, a conclusion based on plaster casts made from the vocal tract of a monkey cadaver. We used x-ray videos to quantify vocal tract dynamics in living macaques during vocalization, facial displays, and feeding. We demonstrate that the macaque vocal tract could easily produce an adequate range of speech sounds to support spoken language, showing that previous techniques based on postmortem samples drastically underestimated primate vocal capabilities. Our findings imply that the evolution of human speech capabilities required neural changes rather than modifications of vocal anatomy. Macaques have a speech-ready vocal tract but lack a speech-ready brain to control it.

  15. Vocal Dose Measures: Quantifying Accumulated Vibration Exposure in Vocal Fold Tissues

    PubMed Central

    Titze, Ingo R.; Švec, Jan G.; Popolo, Peter S.

    2011-01-01

    To measure the exposure to self-induced tissue vibration in speech, three vocal doses were defined and described: distance dose, which accumulates the distance that tissue particles of the vocal folds travel in an oscillatory trajectory; energy dissipation dose, which accumulates the total amount of heat dissipated over a unit volume of vocal fold tissues; and time dose, which accumulates the total phonation time. These doses were compared to a previously used vocal dose measure, the vocal loading index, which accumulates the number of vibration cycles of the vocal folds. Empirical rules for viscosity and vocal fold deformation were used to calculate all the doses from the fundamental frequency (F0) and sound pressure level (SPL) values of speech. Six participants were asked to read in normal, monotone, and exaggerated speech and the doses associated with these vocalizations were calculated. The results showed that large F0 and SPL variations in speech affected the dose measures, suggesting that accumulation of phonation time alone is insufficient. The vibration exposure of the vocal folds in normal speech was related to the industrial limits for hand-transmitted vibration, in which the safe distance dose was derived to be about 500 m. This limit was found rather low for vocalization; it was related to a comparable time dose of about 17 min of continuous vocalization, or about 35 min of continuous reading with normal breathing and unvoiced segments. The voicing pauses in normal speech and dialogue effectively prolong the safe time dose. The derived safety limits for vocalization will likely require refinement based on a more detailed knowledge of the differences in hand and vocal fold tissue morphology and their response to vibrational stress, and on the effect of recovery of the vocal fold tissue during voicing pauses. PMID:12959470

  16. Vocal dose measures: quantifying accumulated vibration exposure in vocal fold tissues.

    PubMed

    Titze, Ingo R; Svec, Jan G; Popolo, Peter S

    2003-08-01

    To measure the exposure to self-induced tissue vibration in speech, three vocal doses were defined and described: distance dose, which accumulates the distance that tissue particles of the vocal folds travel in an oscillatory trajectory; energy dissipation dose, which accumulates the total amount of heat dissipated over a unit volume of vocal fold tissues; and time dose, which accumulates the total phonation time. These doses were compared to a previously used vocal dose measure, the vocal loading index, which accumulates the number of vibration cycles of the vocal folds. Empirical rules for viscosity and vocal fold deformation were used to calculate all the doses from the fundamental frequency (F0) and sound pressure level (SPL) values of speech. Six participants were asked to read in normal, monotone, and exaggerated speech and the doses associated with these vocalizations were calculated. The results showed that large F0 and SPL variations in speech affected the dose measures, suggesting that accumulation of phonation time alone is insufficient. The vibration exposure of the vocal folds in normal speech was related to the industrial limits for hand-transmitted vibration, in which the safe distance dose was derived to be about 500 m. This limit was found rather low for vocalization; it was related to a comparable time dose of about 17 min of continuous vocalization, or about 35 min of continuous reading with normal breathing and unvoiced segments. The voicing pauses in normal speech and dialogue effectively prolong the safe time dose. The derived safety limits for vocalization will likely require refinement based on a more detailed knowledge of the differences in hand and vocal fold tissue morphology and their response to vibrational stress, and on the effect of recovery of the vocal fold tissue during voicing pauses.

  17. Audio-vocal interaction in single neurons of the monkey ventrolateral prefrontal cortex.

    PubMed

    Hage, Steffen R; Nieder, Andreas

    2015-05-06

    Complex audio-vocal integration systems depend on a strong interconnection between the auditory and the vocal motor system. To gain cognitive control over audio-vocal interaction during vocal motor control, the PFC needs to be involved. Neurons in the ventrolateral PFC (VLPFC) have been shown to separately encode the sensory perceptions and motor production of vocalizations. It is unknown, however, whether single neurons in the PFC reflect audio-vocal interactions. We therefore recorded single-unit activity in the VLPFC of rhesus monkeys (Macaca mulatta) while they produced vocalizations on command or passively listened to monkey calls. We found that 12% of randomly selected neurons in VLPFC modulated their discharge rate in response to acoustic stimulation with species-specific calls. Almost three-fourths of these auditory neurons showed an additional modulation of their discharge rates either before and/or during the monkeys' motor production of vocalization. Based on these audio-vocal interactions, the VLPFC might be well positioned to combine higher order auditory processing with cognitive control of the vocal motor output. Such audio-vocal integration processes in the VLPFC might constitute a precursor for the evolution of complex learned audio-vocal integration systems, ultimately giving rise to human speech. Copyright © 2015 the authors 0270-6474/15/357030-11$15.00/0.

  18. Vocalization-Induced Enhancement of the Auditory Cortex Responsiveness during Voice F0 Feedback Perturbation

    PubMed Central

    Behroozmand, Roozbeh; Karvelis, Laura; Liu, Hanjun; Larson, Charles R.

    2009-01-01

    Objective The present study investigated whether self-vocalization enhances auditory neural responsiveness to voice pitch feedback perturbation and how this vocalization-induced neural modulation can be affected by the extent of the feedback deviation. Method Event related potentials (ERPs) were recorded in 15 subjects in response to +100, +200 and +500 cents pitch-shifted voice auditory feedback during active vocalization and passive listening to the playback of the self-produced vocalizations. Result The amplitude of the evoked P1 (latency: 73.51 ms) and P2 (latency: 199.55 ms) ERP components in response to feedback perturbation were significantly larger during vocalization than listening. The difference between P2 peak amplitudes during vocalization vs. listening was shown to be significantly larger for +100 than +500 cents stimulus. Conclusion Results indicate that the human auditory cortex is more responsive to voice F0 feedback perturbations during vocalization than passive listening. Greater vocalization-induced enhancement of the auditory responsiveness to smaller feedback perturbations may imply that the audio-vocal system detects and corrects for errors in vocal production that closely match the expected vocal output. Significance Findings of this study support previous suggestions regarding the enhanced auditory sensitivity to feedback alterations during self-vocalization, which may serve the purpose of feedback-based monitoring of one’s voice. PMID:19520602

  19. Limiting parental feedback disrupts vocal development in marmoset monkeys

    PubMed Central

    Gultekin, Yasemin B.; Hage, Steffen R.

    2017-01-01

    Vocalizations of human infants undergo dramatic changes across the first year by becoming increasingly mature and speech-like. Human vocal development is partially dependent on learning by imitation through social feedback between infants and caregivers. Recent studies revealed similar developmental processes being influenced by parental feedback in marmoset monkeys for apparently innate vocalizations. Marmosets produce infant-specific vocalizations that disappear after the first postnatal months. However, it is yet unclear whether parental feedback is an obligate requirement for proper vocal development. Using quantitative measures to compare call parameters and vocal sequence structure we show that, in contrast to normally raised marmosets, marmosets that were separated from parents after the third postnatal month still produced infant-specific vocal behaviour at subadult stages. These findings suggest a significant role of social feedback on primate vocal development until the subadult stages and further show that marmoset monkeys are a compelling model system for early human vocal development. PMID:28090084

  20. Modeling vocalization with ECoG cortical activity recorded during vocal production in the macaque monkey.

    PubMed

    Fukushima, Makoto; Saunders, Richard C; Fujii, Naotaka; Averbeck, Bruno B; Mishkin, Mortimer

    2014-01-01

    Vocal production is an example of controlled motor behavior with high temporal precision. Previous studies have decoded auditory evoked cortical activity while monkeys listened to vocalization sounds. On the other hand, there have been few attempts at decoding motor cortical activity during vocal production. Here we recorded cortical activity during vocal production in the macaque with a chronically implanted electrocorticographic (ECoG) electrode array. The array detected robust activity in motor cortex during vocal production. We used a nonlinear dynamical model of the vocal organ to reduce the dimensionality of `Coo' calls produced by the monkey. We then used linear regression to evaluate the information in motor cortical activity for this reduced representation of calls. This simple linear model accounted for circa 65% of the variance in the reduced sound representations, supporting the feasibility of using the dynamical model of the vocal organ for decoding motor cortical activity during vocal production.

Top