Sample records for syllable final position

  1. The Role of the Syllable in the Segmentation of Cairene Spoken Arabic

    ERIC Educational Resources Information Center

    Aquil, Rajaa

    2012-01-01

    The syllable as a perceptual unit has been investigated cross linguistically. In Cairene Arabic syllables fall into three categories, light CV, heavy CVC/CVV and superheavy CVCC/CVVC. However, heavy syllables in Cariene Arabic have varied weight depending on their position in a word, whether internal or final. The present paper investigates the…

  2. Syllabic encoding during overt speech production in Cantonese: Evidence from temporal brain responses.

    PubMed

    Wong, Andus Wing-Kuen; Wang, Jie; Ng, Tin-Yan; Chen, Hsuan-Chih

    2016-10-01

    The time course of phonological encoding in overt Cantonese disyllabic word production was investigated using a picture-word interference task with concurrent recording of the event-related brain potentials (ERPs). Participants were asked to name aloud individually presented pictures and ignore a distracting Chinese character. Participants' naming responses were faster, relative to an unrelated control, when the distractor overlapped with the target's word-initial or word-final syllables. Furthermore, ERP waves in the syllable-related conditions were more positive-going than those in the unrelated control conditions from 500ms to 650ms post target onset (i.e., a late positivity). The mean and peak amplitudes of this late positivity correlated with the size of phonological facilitation. More importantly, the onset of the late positivity associated with word-initial syllable priming was 44ms earlier than that associated with word-final syllable priming, suggesting that phonological encoding in overt speech runs incrementally and the encoding duration for one syllable unit is approximately 44ms. Although the size of effective phonological units might vary across languages, as suggested by previous speech production studies, the present data indicate that the incremental nature of phonological encoding is a universal mechanism. Copyright © 2016 Elsevier B.V. All rights reserved.

  3. Perceptual Processing of Mandarin Nasals by L1 and L2 Mandarin Speakers

    ERIC Educational Resources Information Center

    Lai, Yi-hsiu

    2012-01-01

    Nasals are cross-linguistically susceptible to change, especially in the syllable final position. Acoustic reports on Mandarin nasal production have recently shown that the syllable-final distinction is frequently dropped. Few studies, however, have addressed the issue of perceptual processing in Mandarin nasals for L1 and L2 speakers of Mandarin…

  4. Syllable Structure Universals and Native Language Interference in Second Language Perception and Production: Positional Asymmetry and Perceptual Links to Accentedness

    PubMed Central

    Cheng, Bing; Zhang, Yang

    2015-01-01

    The present study investigated how syllable structure differences between the first Language (L1) and the second language (L2) affect L2 consonant perception and production at syllable-initial and syllable-final positions. The participants were Mandarin-speaking college students who studied English as a second language. Monosyllabic English words were used in the perception test. Production was recorded from each Chinese subject and rated for accentedness by two native speakers of English. Consistent with previous studies, significant positional asymmetry effects were found across speech sound categories in terms of voicing, place of articulation, and manner of articulation. Furthermore, significant correlations between perception and accentedness ratings were found at the syllable onset position but not for the coda. Many exceptions were also found, which could not be solely accounted for by differences in L1–L2 syllabic structures. The results show a strong effect of language experience at the syllable level, which joins force with acoustic, phonetic, and phonemic properties of individual consonants in influencing positional asymmetry in both domains of L2 segmental perception and production. The complexities and exceptions call for further systematic studies on the interactions between syllable structure universals and native language interference with refined theoretical models to specify the links between perception and production in second language acquisition. PMID:26635699

  5. Phonological awareness of English by Chinese and Korean bilinguals

    NASA Astrophysics Data System (ADS)

    Chung, Hyunjoo; Schmidt, Anna; Cheng, Tse-Hsuan

    2002-05-01

    This study examined non-native speakers phonological awareness of spoken English. Chinese speaking adults, Korean speaking adults, and English speaking adults were tested. The L2 speakers had been in the US for less than 6 months. Chinese and Korean allow no consonant clusters and have limited numbers of consonants allowable in syllable final position, whereas English allows a variety of clusters and various consonants in syllable final position. Subjects participated in eight phonological awareness tasks (4 replacement tasks and 4 deletion tasks) based on English phonology. In addition, digit span was measured. Preliminary analysis indicates that Chinese and Korean speaker errors appear to reflect L1 influences (such as orthography, phonotactic constraints, and phonology). All three groups of speakers showed more difficulty with manipulation of rime than onset, especially with postvocalic nasals. Results will be discussed in terms of syllable structure, L1 influence, and association with short term memory.

  6. The perception of intonation questions and statements in Cantonese.

    PubMed

    Ma, Joan K-Y; Ciocca, Valter; Whitehill, Tara L

    2011-02-01

    In tone languages there are potential conflicts in the perception of lexical tone and intonation, as both depend mainly on the differences in fundamental frequency (F0) patterns. The present study investigated the acoustic cues associated with the perception of sentences as questions or statements in Cantonese, as a function of the lexical tone in sentence final position. Cantonese listeners performed intonation identification tasks involving complete sentences, isolated final syllables, and sentences without the final syllable (carriers). Sensitivity (d' scores) were similar for complete sentences and final syllables but were significantly lower for carriers. Sensitivity was also affected by tone identity. These findings show that the perception of questions and statements relies primarily on the F0 characteristics of the final syllables (local F0 cues). A measure of response bias (c) provided evidence for a general bias toward the perception of statements. Logistic regression analyses showed that utterances were accurately classified as questions or statements by using average F0 and F0 interval. Average F0 of carriers (global F0 cue) was also found to be a reliable secondary cue. These findings suggest that the use of F0 cues for the perception of intonation question in tonal languages is likely to be language-specific.

  7. Effects of Word Position on the Acoustic Realization of Vietnamese Final Consonants.

    PubMed

    Tran, Thi Thuy Hien; Vallée, Nathalie; Granjon, Lionel

    2018-05-28

    A variety of studies have shown differences between phonetic features of consonants according to their prosodic and/or syllable (onset vs. coda) positions. However, differences are not always found, and interactions between the various factors involved are complex and not well understood. Our study compares acoustical characteristics of coda consonants in Vietnamese taking into account their position within words. Traditionally described as monosyllabic, Vietnamese is partially polysyllabic at the lexical level. In this language, tautosyllabic consonant sequences are prohibited, and adjacent consonants are only found at syllable boundaries either within polysyllabic words (CVC.CVC) or across monosyllabic words (CVC#CVC). This study is designed to examine whether or not syllable boundary types (interword vs. intraword) have an effect on the acoustic realization of codas. The results show significant acoustic differences in consonant realizations according to syllable boundary type, suggesting different coarticulation patterns between nuclei and codas. In addition, as Vietnamese voiceless stops are generally unreleased in coda position, with no burst to carry consonantal information, our results show that a vowel's second half contains acoustic cues which are available to aid in the discrimination of place of articulation of the vowel's following consonant. © 2018 S. Karger AG, Basel.

  8. Automatic initial and final segmentation in cleft palate speech of Mandarin speakers

    PubMed Central

    Liu, Yin; Yin, Heng; Zhang, Junpeng; Zhang, Jing; Zhang, Jiang

    2017-01-01

    The speech unit segmentation is an important pre-processing step in the analysis of cleft palate speech. In Mandarin, one syllable is composed of two parts: initial and final. In cleft palate speech, the resonance disorders occur at the finals and the voiced initials, while the articulation disorders occur at the unvoiced initials. Thus, the initials and finals are the minimum speech units, which could reflect the characteristics of cleft palate speech disorders. In this work, an automatic initial/final segmentation method is proposed. It is an important preprocessing step in cleft palate speech signal processing. The tested cleft palate speech utterances are collected from the Cleft Palate Speech Treatment Center in the Hospital of Stomatology, Sichuan University, which has the largest cleft palate patients in China. The cleft palate speech data includes 824 speech segments, and the control samples contain 228 speech segments. The syllables are extracted from the speech utterances firstly. The proposed syllable extraction method avoids the training stage, and achieves a good performance for both voiced and unvoiced speech. Then, the syllables are classified into with “quasi-unvoiced” or with “quasi-voiced” initials. Respective initial/final segmentation methods are proposed to these two types of syllables. Moreover, a two-step segmentation method is proposed. The rough locations of syllable and initial/final boundaries are refined in the second segmentation step, in order to improve the robustness of segmentation accuracy. The experiments show that the initial/final segmentation accuracies for syllables with quasi-unvoiced initials are higher than quasi-voiced initials. For the cleft palate speech, the mean time error is 4.4ms for syllables with quasi-unvoiced initials, and 25.7ms for syllables with quasi-voiced initials, and the correct segmentation accuracy P30 for all the syllables is 91.69%. For the control samples, P30 for all the syllables is 91.24%. PMID:28926572

  9. Automatic initial and final segmentation in cleft palate speech of Mandarin speakers.

    PubMed

    He, Ling; Liu, Yin; Yin, Heng; Zhang, Junpeng; Zhang, Jing; Zhang, Jiang

    2017-01-01

    The speech unit segmentation is an important pre-processing step in the analysis of cleft palate speech. In Mandarin, one syllable is composed of two parts: initial and final. In cleft palate speech, the resonance disorders occur at the finals and the voiced initials, while the articulation disorders occur at the unvoiced initials. Thus, the initials and finals are the minimum speech units, which could reflect the characteristics of cleft palate speech disorders. In this work, an automatic initial/final segmentation method is proposed. It is an important preprocessing step in cleft palate speech signal processing. The tested cleft palate speech utterances are collected from the Cleft Palate Speech Treatment Center in the Hospital of Stomatology, Sichuan University, which has the largest cleft palate patients in China. The cleft palate speech data includes 824 speech segments, and the control samples contain 228 speech segments. The syllables are extracted from the speech utterances firstly. The proposed syllable extraction method avoids the training stage, and achieves a good performance for both voiced and unvoiced speech. Then, the syllables are classified into with "quasi-unvoiced" or with "quasi-voiced" initials. Respective initial/final segmentation methods are proposed to these two types of syllables. Moreover, a two-step segmentation method is proposed. The rough locations of syllable and initial/final boundaries are refined in the second segmentation step, in order to improve the robustness of segmentation accuracy. The experiments show that the initial/final segmentation accuracies for syllables with quasi-unvoiced initials are higher than quasi-voiced initials. For the cleft palate speech, the mean time error is 4.4ms for syllables with quasi-unvoiced initials, and 25.7ms for syllables with quasi-voiced initials, and the correct segmentation accuracy P30 for all the syllables is 91.69%. For the control samples, P30 for all the syllables is 91.24%.

  10. Prosodic structure shapes the temporal realization of intonation and manual gesture movements.

    PubMed

    Esteve-Gibert, Núria; Prieto, Pilar

    2013-06-01

    Previous work on the temporal coordination between gesture and speech found that the prominence in gesture coordinates with speech prominence. In this study, the authors investigated the anchoring regions in speech and pointing gesture that align with each other. The authors hypothesized that (a) in contrastive focus conditions, the gesture apex is anchored in the intonation peak and (b) the upcoming prosodic boundary influences the timing of gesture and intonation movements. Fifteen Catalan speakers pointed at a screen while pronouncing a target word with different metrical patterns in a contrastive focus condition and followed by a phrase boundary. A total of 702 co-speech deictic gestures were acoustically and gesturally analyzed. Intonation peaks and gesture apexes showed parallel behavior with respect to their position within the accented syllable: They occurred at the end of the accented syllable in non-phrase-final position, whereas they occurred well before the end of the accented syllable in phrase-final position. Crucially, the position of intonation peaks and gesture apexes was correlated and was bound by prosodic structure. The results refine the phonological synchronization rule (McNeill, 1992), showing that gesture apexes are anchored in intonation peaks and that gesture and prosodic movements are bound by prosodic phrasing.

  11. On the edge of language acquisition: inherent constraints on encoding multisyllabic sequences in the neonate brain.

    PubMed

    Ferry, Alissa L; Fló, Ana; Brusini, Perrine; Cattarossi, Luigi; Macagno, Francesco; Nespor, Marina; Mehler, Jacques

    2016-05-01

    To understand language, humans must encode information from rapid, sequential streams of syllables - tracking their order and organizing them into words, phrases, and sentences. We used Near-Infrared Spectroscopy (NIRS) to determine whether human neonates are born with the capacity to track the positions of syllables in multisyllabic sequences. After familiarization with a six-syllable sequence, the neonate brain responded to the change (as shown by an increase in oxy-hemoglobin) when the two edge syllables switched positions but not when two middle syllables switched positions (Experiment 1), indicating that they encoded the syllables at the edges of sequences better than those in the middle. Moreover, when a 25 ms pause was inserted between the middle syllables as a segmentation cue, neonates' brains were sensitive to the change (Experiment 2), indicating that subtle cues in speech can signal a boundary, with enhanced encoding of the syllables located at the edges of that boundary. These findings suggest that neonates' brains can encode information from multisyllabic sequences and that this encoding is constrained. Moreover, subtle segmentation cues in a sequence of syllables provide a mechanism with which to accurately encode positional information from longer sequences. Tracking the order of syllables is necessary to understand language and our results suggest that the foundations for this encoding are present at birth. © 2015 John Wiley & Sons Ltd.

  12. Spoken Word Recognition of Chinese Words in Continuous Speech

    ERIC Educational Resources Information Center

    Yip, Michael C. W.

    2015-01-01

    The present study examined the role of positional probability of syllables played in recognition of spoken word in continuous Cantonese speech. Because some sounds occur more frequently at the beginning position or ending position of Cantonese syllables than the others, so these kinds of probabilistic information of syllables may cue the locations…

  13. Infants Learn Phonotactic Regularities from Brief Auditory Experience.

    ERIC Educational Resources Information Center

    Chambers, Kyle E.; Onishi, Kristine H.; Fisher, Cynthia

    2003-01-01

    Two experiments investigated whether novel phonotactic regularities, not present in English, could be acquired by 16.5-month-olds from brief auditory experience. Subjects listened to consonant-vowel-consonant syllables in which particular consonants were artificially restricted to either initial or final position. Findings in a subsequent…

  14. Syllable Durations of Preword and Early Word Vocalizations.

    ERIC Educational Resources Information Center

    Robb, Michael P.; Saxman, John H.

    1990-01-01

    The continuity in development of syllable duration patterns was examined in seven young children as they progressed from preword to multiword periods of vocalization development. Results revealed no systematic increase or decrease in the duration of bisyllables produced by the children as a group, whereas lengthening of final syllables was…

  15. Final Syllable Lengthening (FSL) in Infant Vocalizations.

    ERIC Educational Resources Information Center

    Nathani, Suneeti; Oller, D. Kimbrough; Cobo-Lewis, Alan B.

    2003-01-01

    Sought to verify research findings that suggest there may be a U-shaped developmental trajectory for final syllable lengthening (FSL). Attempted to determine whether vocal maturity and deafness influence FSL . Eight normally hearing infants and eight deaf infants were examined at three levels of prelinguistic vocal development. (Author/VWL)

  16. Cross-linguistic differences in the use of durational cues for the segmentation of a novel language.

    PubMed

    Ordin, Mikhail; Polyanskaya, Leona; Laka, Itziar; Nespor, Marina

    2017-07-01

    It is widely accepted that duration can be exploited as phonological phrase final lengthening in the segmentation of a novel language, i.e., in extracting discrete constituents from continuous speech. The use of final lengthening for segmentation and its facilitatory effect has been claimed to be universal. However, lengthening in the world languages can also mark lexically stressed syllables. Stress-induced lengthening can potentially be in conflict with right edge phonological phrase boundary lengthening. Thus the processing of durational cues in segmentation can be dependent on the listener's linguistic background, e.g., on the specific correlates and unmarked location of lexical stress in the native language of the listener. We tested this prediction and found that segmentation by both German and Basque speakers is facilitated when lengthening is aligned with the word final syllable and is not affected by lengthening on either the penultimate or the antepenultimate syllables. Lengthening of the word final syllable, however, does not help Italian and Spanish speakers to segment continuous speech, and lengthening of the antepenultimate syllable impedes their performance. We have also found a facilitatory effect of penultimate lengthening on segmentation by Italians. These results confirm our hypothesis that processing of lengthening cues is not universal, and interpretation of lengthening as a phonological phrase final boundary marker in a novel language of exposure can be overridden by the phonology of lexical stress in the native language of the listener.

  17. Exploring Assessment Demands and Task Supports in Early Childhood Phonological Awareness Assessments

    ERIC Educational Resources Information Center

    Cassano, Christina M.; Steiner, Lilly

    2016-01-01

    Phonological awareness is assessed in various ways in both research studies and early childhood classrooms. The measures used to assess phonological awareness are related closely, although they differ in the linguistic unit used (e.g., word, syllable, onset-rime, or phoneme), the position of the linguistic unit (e.g., initial, medial, final), the…

  18. Effects of prosody and position on the timing of deictic gestures.

    PubMed

    Rusiewicz, Heather Leavy; Shaiman, Susan; Iverson, Jana M; Szuminsky, Neil

    2013-04-01

    In this study, the authors investigated the hypothesis that the perceived tight temporal synchrony of speech and gesture is evidence of an integrated spoken language and manual gesture communication system. It was hypothesized that experimental manipulations of the spoken response would affect the timing of deictic gestures. The authors manipulated syllable position and contrastive stress in compound words in multiword utterances by using a repeated-measures design to investigate the degree of synchronization of speech and pointing gestures produced by 15 American English speakers. Acoustic measures were compared with the gesture movement recorded via capacitance. Although most participants began a gesture before the target word, the temporal parameters of the gesture changed as a function of syllable position and prosody. Syllables with contrastive stress in the 2nd position of compound words were the longest in duration and also most consistently affected the timing of gestures, as measured by several dependent measures. Increasing the stress of a syllable significantly affected the timing of a corresponding gesture, notably for syllables in the 2nd position of words that would not typically be stressed. The findings highlight the need to consider the interaction of gestures and spoken language production from a motor-based perspective of coordination.

  19. The Frame Constraint on Experimentally Elicited Speech Errors in Japanese.

    PubMed

    Saito, Akie; Inoue, Tomoyoshi

    2017-06-01

    The so-called syllable position effect in speech errors has been interpreted as reflecting constraints posed by the frame structure of a given language, which is separately operating from linguistic content during speech production. The effect refers to the phenomenon that when a speech error occurs, replaced and replacing sounds tend to be in the same position within a syllable or word. Most of the evidence for the effect comes from analyses of naturally occurring speech errors in Indo-European languages, and there are few studies examining the effect in experimentally elicited speech errors and in other languages. This study examined whether experimentally elicited sound errors in Japanese exhibits the syllable position effect. In Japanese, the sub-syllabic unit known as "mora" is considered to be a basic sound unit in production. Results showed that the syllable position effect occurred in mora errors, suggesting that the frame constrains the ordering of sounds during speech production.

  20. Stress Domain Effects in French Phonology and Phonological Development.

    PubMed

    Rose, Yvan; Dos Santos, Christophe

    In this paper, we discuss two distinct data sets. The first relates to the so-called allophonic process of closed-syllable laxing in Québec French, which targets final (stressed) vowels even though these vowels are arguably syllabified in open syllables in lexical representations. The second is found in the forms produced by a first language learner of European French, who displays an asymmetry in her production of CVC versus CVCV target (adult) forms. The former display full preservation (with concomitant manner harmony) of both consonants. The latter undergoes deletion of the initial syllable if the consonants are not manner-harmonic in the input. We argue that both patterns can be explained through a phonological process of prosodic strengthening targeting the head of the prosodic domain which, in the contexts described above, yields the incorporation of final consonants into the coda of the stressed syllable.

  1. Acquisition of Malay word recognition skills: lessons from low-progress early readers.

    PubMed

    Lee, Lay Wah; Wheldall, Kevin

    2011-02-01

    Malay is a consistent alphabetic orthography with complex syllable structures. The focus of this research was to investigate word recognition performance in order to inform reading interventions for low-progress early readers. Forty-six Grade 1 students were sampled and 11 were identified as low-progress readers. The results indicated that both syllable awareness and phoneme blending were significant predictors of word recognition, suggesting that both syllable and phonemic grain-sizes are important in Malay word recognition. Item analysis revealed a hierarchical pattern of difficulty based on the syllable and the phonic structure of the words. Error analysis identified the sources of errors to be errors due to inefficient syllable segmentation, oversimplification of syllables, insufficient grapheme-phoneme knowledge and inefficient phonemic code assembly. Evidence also suggests that direct instruction in syllable segmentation, phonemic awareness and grapheme-phoneme correspondence is necessary for low-progress readers to acquire word recognition skills. Finally, a logical sequence to teach grapheme-phoneme decoding in Malay is suggested. Copyright © 2010 John Wiley & Sons, Ltd.

  2. Influence of Initial and Final Consonants on Vowel Duration in CVC Syllables.

    ERIC Educational Resources Information Center

    Naeser, Margaret A.

    This study investigates the influence of initial and final consonants /p, b, s, z/ on the duration of four vowels /I, i, u, ae/ in 64 CVC syllables uttered by eight speakers of English from the same dialect area. The CVC stimuli were presented to the subjects in a frame sentence from a master tape. Subjects repeated each sentence immediately after…

  3. Estimating feedforward vs. feedback control of speech production through kinematic analyses of unperturbed articulatory movements.

    PubMed

    Kim, Kwang S; Max, Ludo

    2014-01-01

    To estimate the contributions of feedforward vs. feedback control systems in speech articulation, we analyzed the correspondence between initial and final kinematics in unperturbed tongue and jaw movements for consonant-vowel (CV) and vowel-consonant (VC) syllables. If movement extents and endpoints are highly predictable from early kinematic information, then the movements were most likely completed without substantial online corrections (feedforward control); if the correspondence between early kinematics and final amplitude or position is low, online adjustments may have altered the planned trajectory (feedback control) (Messier and Kalaska, 1999). Five adult speakers produced CV and VC syllables with high, mid, or low vowels while movements of the tongue and jaw were tracked electromagnetically. The correspondence between the kinematic parameters peak acceleration or peak velocity and movement extent as well as between the articulators' spatial coordinates at those kinematic landmarks and movement endpoint was examined both for movements across different target distances (i.e., across vowel height) and within target distances (i.e., within vowel height). Taken together, results suggest that jaw and tongue movements for these CV and VC syllables are mostly under feedforward control but with feedback-based contributions. One type of feedback-driven compensatory adjustment appears to regulate movement duration based on variation in peak acceleration. Results from a statistical model based on multiple regression are presented to illustrate how the relative strength of these feedback contributions can be estimated.

  4. Consonant acquisition in the Malay language: a cross-sectional study of preschool aged Malay children.

    PubMed

    Phoon, Hooi San; Abdullah, Anna Christina; Lee, Lay Wah; Murugaiah, Puvaneswary

    2014-05-01

    To date, there has been little research done on phonological acquisition in the Malay language of typically developing Malay-speaking children. This study serves to fill this gap by providing a systematic description of Malay consonant acquisition in a large cohort of preschool-aged children between 4- and 6-years-old. In the study, 326 Malay-dominant speaking children were assessed using a picture naming task that elicited 53 single words containing all the primary consonants in Malay. Two main analyses were conducted to study their consonant acquisition: (1) age of customary and mastery production of consonants; and (2) consonant accuracy. Results revealed that Malay children acquired all the syllable-initial and syllable-final consonants before 4;06-years-old, with the exception of syllable-final /s/, /h/ and /l/ which were acquired after 5;06-years-old. The development of Malay consonants increased gradually from 4- to 6 years old, with female children performing better than male children. The accuracy of consonants based on manner of articulation showed that glides, affricates, nasals, and stops were higher than fricatives and liquids. In general, syllable-initial consonants were more accurate than syllable-final consonants while consonants in monosyllabic and disyllabic words were more accurate than polysyllabic words. These findings will provide significant information for speech-language pathologists for assessing Malay-speaking children and designing treatment objectives that reflect the course of phonological development in Malay.

  5. Brain correlates of stuttering and syllable production. A PET performance-correlation analysis.

    PubMed

    Fox, P T; Ingham, R J; Ingham, J C; Zamarripa, F; Xiong, J H; Lancaster, J L

    2000-10-01

    To distinguish the neural systems of normal speech from those of stuttering, PET images of brain blood flow were probed (correlated voxel-wise) with per-trial speech-behaviour scores obtained during PET imaging. Two cohorts were studied: 10 right-handed men who stuttered and 10 right-handed, age- and sex-matched non-stuttering controls. Ninety PET blood flow images were obtained in each cohort (nine per subject as three trials of each of three conditions) from which r-value statistical parametric images (SPI¿r¿) were computed. Brain correlates of stutter rate and syllable rate showed striking differences in both laterality and sign (i.e. positive or negative correlations). Stutter-rate correlates, both positive and negative, were strongly lateralized to the right cerebral and left cerebellar hemispheres. Syllable correlates in both cohorts were bilateral, with a bias towards the left cerebral and right cerebellar hemispheres, in keeping with the left-cerebral dominance for language and motor skills typical of right-handed subjects. For both stutters and syllables, the brain regions that were correlated positively were those of speech production: the mouth representation in the primary motor cortex; the supplementary motor area; the inferior lateral premotor cortex (Broca's area); the anterior insula; and the cerebellum. The principal difference between syllable-rate and stutter-rate positive correlates was hemispheric laterality. A notable exception to this rule was that cerebellar positive correlates for syllable rate were far more extensive in the stuttering cohort than in the control cohort, which suggests a specific role for the cerebellum in enabling fluent utterances in persons who stutter. Stutters were negatively correlated with right-cerebral regions (superior and middle temporal gyrus) associated with auditory perception and processing, regions which were positively correlated with syllables in both the stuttering and control cohorts. These findings support long-held theories that the brain correlates of stuttering are the speech-motor regions of the non-dominant (right) cerebral hemisphere, and extend this theory to include the non-dominant (left) cerebellar hemisphere. The present findings also indicate a specific role of the cerebellum in the fluent utterances of persons who stutter. Support is also offered for theories that implicate auditory processing problems in stuttering.

  6. Attention effects on the processing of task-relevant and task-irrelevant speech sounds and letters

    PubMed Central

    Mittag, Maria; Inauri, Karina; Huovilainen, Tatu; Leminen, Miika; Salo, Emma; Rinne, Teemu; Kujala, Teija; Alho, Kimmo

    2013-01-01

    We used event-related brain potentials (ERPs) to study effects of selective attention on the processing of attended and unattended spoken syllables and letters. Participants were presented with syllables randomly occurring in the left or right ear and spoken by different voices and with a concurrent foveal stream of consonant letters written in darker or lighter fonts. During auditory phonological (AP) and non-phonological tasks, they responded to syllables in a designated ear starting with a vowel and spoken by female voices, respectively. These syllables occurred infrequently among standard syllables starting with a consonant and spoken by male voices. During visual phonological and non-phonological tasks, they responded to consonant letters with names starting with a vowel and to letters written in dark fonts, respectively. These letters occurred infrequently among standard letters with names starting with a consonant and written in light fonts. To examine genuine effects of attention and task on ERPs not overlapped by ERPs associated with target processing or deviance detection, these effects were studied only in ERPs to auditory and visual standards. During selective listening to syllables in a designated ear, ERPs to the attended syllables were negatively displaced during both phonological and non-phonological auditory tasks. Selective attention to letters elicited an early negative displacement and a subsequent positive displacement (Pd) of ERPs to attended letters being larger during the visual phonological than non-phonological task suggesting a higher demand for attention during the visual phonological task. Active suppression of unattended speech during the AP and non-phonological tasks and during the visual phonological tasks was suggested by a rejection positivity (RP) to unattended syllables. We also found evidence for suppression of the processing of task-irrelevant visual stimuli in visual ERPs during auditory tasks involving left-ear syllables. PMID:24348324

  7. Asymmetries in Generalizing Alternations to and from Initial Syllables

    ERIC Educational Resources Information Center

    Becker, Michael; Nevins, Andrew; Levine, Jonathan

    2012-01-01

    In the English lexicon, laryngeal alternations in the plural (e.g. "leaf" ~ "leaves") impact monosyllables more than finally stressed polysyllables. This is the opposite of what happens typologically, and would thereby run contrary to the predictions of "initial-syllable faithfulness." Despite the lexical pattern, in a wug test we found…

  8. Phonological Acquisition of Korean Consonants in Conversational Speech Produced by Young Korean Children

    ERIC Educational Resources Information Center

    Kim, Minjung; Kim, Soo-Jin; Stoel-Gammon, Carol

    2017-01-01

    This study investigates the phonological acquisition of Korean consonants using conversational speech samples collected from sixty monolingual typically developing Korean children aged two, three, and four years. Phonemic acquisition was examined for syllable-initial and syllable-final consonants. Results showed that Korean children acquired stops…

  9. Feeling backwards? How temporal order in speech affects the time course of vocal emotion recognition

    PubMed Central

    Rigoulot, Simon; Wassiliwizky, Eugen; Pell, Marc D.

    2013-01-01

    Recent studies suggest that the time course for recognizing vocal expressions of basic emotion in speech varies significantly by emotion type, implying that listeners uncover acoustic evidence about emotions at different rates in speech (e.g., fear is recognized most quickly whereas happiness and disgust are recognized relatively slowly; Pell and Kotz, 2011). To investigate whether vocal emotion recognition is largely dictated by the amount of time listeners are exposed to speech or the position of critical emotional cues in the utterance, 40 English participants judged the meaning of emotionally-inflected pseudo-utterances presented in a gating paradigm, where utterances were gated as a function of their syllable structure in segments of increasing duration from the end of the utterance (i.e., gated syllable-by-syllable from the offset rather than the onset of the stimulus). Accuracy for detecting six target emotions in each gate condition and the mean identification point for each emotion in milliseconds were analyzed and compared to results from Pell and Kotz (2011). We again found significant emotion-specific differences in the time needed to accurately recognize emotions from speech prosody, and new evidence that utterance-final syllables tended to facilitate listeners' accuracy in many conditions when compared to utterance-initial syllables. The time needed to recognize fear, anger, sadness, and neutral from speech cues was not influenced by how utterances were gated, although happiness and disgust were recognized significantly faster when listeners heard the end of utterances first. Our data provide new clues about the relative time course for recognizing vocally-expressed emotions within the 400–1200 ms time window, while highlighting that emotion recognition from prosody can be shaped by the temporal properties of speech. PMID:23805115

  10. Contextual Variability in American English Dark-L

    ERIC Educational Resources Information Center

    Oxley, Judith; Roussel, Nancye; Buckingham, Hugh

    2007-01-01

    This paper presents a four-subject study that examines the relative influence of syllable position and stress, together with vowel context on the colouring of the dark-l characteristic of speakers of General American English. Most investigators report lighter /l/ tokens in syllable onsets and darker tokens in coda positions. The present study…

  11. The neural dynamics of song syntax in songbirds

    NASA Astrophysics Data System (ADS)

    Jin, Dezhe

    2010-03-01

    Songbird is ``the hydrogen atom'' of the neuroscience of complex, learned vocalizations such as human speech. Songs of Bengalese finch consist of sequences of syllables. While syllables are temporally stereotypical, syllable sequences can vary and follow complex, probabilistic syntactic rules, which are rudimentarily similar to grammars in human language. Songbird brain is accessible to experimental probes, and is understood well enough to construct biologically constrained, predictive computational models. In this talk, I will discuss the structure and dynamics of neural networks underlying the stereotypy of the birdsong syllables and the flexibility of syllable sequences. Recent experiments and computational models suggest that a syllable is encoded in a chain network of projection neurons in premotor nucleus HVC (proper name). Precisely timed spikes propagate along the chain, driving vocalization of the syllable through downstream nuclei. Through a computational model, I show that that variable syllable sequences can be generated through spike propagations in a network in HVC in which the syllable-encoding chain networks are connected into a branching chain pattern. The neurons mutually inhibit each other through the inhibitory HVC interneurons, and are driven by external inputs from nuclei upstream of HVC. At a branching point that connects the final group of a chain to the first groups of several chains, the spike activity selects one branch to continue the propagation. The selection is probabilistic, and is due to the winner-take-all mechanism mediated by the inhibition and noise. The model predicts that the syllable sequences statistically follow partially observable Markov models. Experimental results supporting this and other predictions of the model will be presented. We suggest that the syntax of birdsong syllable sequences is embedded in the connection patterns of HVC projection neurons.

  12. Possible-word constraints in Cantonese speech segmentation.

    PubMed

    Yip, Michael C

    2004-03-01

    A Cantonese syllable-spotting experiment was conducted to examine whether the Possible-Word Constraint (PWC), proposed by Norris, McQueen, Cutler, and Butterfield (1997), can apply in Cantonese speech segmentation. In the experiment, listeners were asked to spot out the target Cantonese syllable from a series of nonsense sound strings. Results suggested that listeners found it more difficult to spot out the target syllable [kDm1] in the nonsense sound strings that attached with a single consonant [tkDm1] than in the nonsense sound strings that attached either with a vowel [a:kDm1] or a pseudo-syllable [khow1kDm1]. Finally, the current set of results further supported that the PWC appears to be a language-universal mechanism in segmenting continuous speech.

  13. Phrase-Final Syllable Lengthening and Intonation in Early Child Speech.

    ERIC Educational Resources Information Center

    Snow, David

    1994-01-01

    To test opposing theories about the relationship between intonation and syllable timing, these boundary features were compared in a longitudinal study of 9 children's speech development between the mean ages of 16 and 25 months. Results suggest that young children acquire the skills that control intonation earlier than they do skills of final…

  14. Diminutives facilitate word segmentation in natural speech: cross-linguistic evidence.

    PubMed

    Kempe, Vera; Brooks, Patricia J; Gillis, Steven; Samson, Graham

    2007-06-01

    Final-syllable invariance is characteristic of diminutives (e.g., doggie), which are a pervasive feature of the child-directed speech registers of many languages. Invariance in word endings has been shown to facilitate word segmentation (Kempe, Brooks, & Gillis, 2005) in an incidental-learning paradigm in which synthesized Dutch pseudonouns were used. To broaden the cross-linguistic evidence for this invariance effect and to increase its ecological validity, adult English speakers (n=276) were exposed to naturally spoken Dutch or Russian pseudonouns presented in sentence contexts. A forced choice test was given to assess target recognition, with foils comprising unfamiliar syllable combinations in Experiments 1 and 2 and syllable combinations straddling word boundaries in Experiment 3. A control group (n=210) received the recognition test with no prior exposure to targets. Recognition performance improved with increasing final-syllable rhyme invariance, with larger increases for the experimental group. This confirms that word ending invariance is a valid segmentation cue in artificial, as well as naturalistic, speech and that diminutives may aid segmentation in a number of languages.

  15. Influence of syllable structure on L2 auditory word learning.

    PubMed

    Hamada, Megumi; Goya, Hideki

    2015-04-01

    This study investigated the role of syllable structure in L2 auditory word learning. Based on research on cross-linguistic variation of speech perception and lexical memory, it was hypothesized that Japanese L1 learners of English would learn English words with an open-syllable structure without consonant clusters better than words with a closed-syllable structure and consonant clusters. Two groups of college students (Japanese group, N = 22; and native speakers of English, N = 21) learned paired English pseudowords and pictures. The pseudoword types differed in terms of the syllable structure and consonant clusters (congruent vs. incongruent) and the position of consonant clusters (coda vs. onset). Recall accuracy was higher for the pseudowords in the congruent type and the pseudowords with the coda-consonant clusters. The syllable structure effect was obtained from both participant groups, disconfirming the hypothesized cross-linguistic influence on L2 auditory word learning.

  16. An Adapting Auditory-motor Feedback Loop Can Contribute to Generating Vocal Repetition

    PubMed Central

    Brainard, Michael S.; Jin, Dezhe Z.

    2015-01-01

    Consecutive repetition of actions is common in behavioral sequences. Although integration of sensory feedback with internal motor programs is important for sequence generation, if and how feedback contributes to repetitive actions is poorly understood. Here we study how auditory feedback contributes to generating repetitive syllable sequences in songbirds. We propose that auditory signals provide positive feedback to ongoing motor commands, but this influence decays as feedback weakens from response adaptation during syllable repetitions. Computational models show that this mechanism explains repeat distributions observed in Bengalese finch song. We experimentally confirmed two predictions of this mechanism in Bengalese finches: removal of auditory feedback by deafening reduces syllable repetitions; and neural responses to auditory playback of repeated syllable sequences gradually adapt in sensory-motor nucleus HVC. Together, our results implicate a positive auditory-feedback loop with adaptation in generating repetitive vocalizations, and suggest sensory adaptation is important for feedback control of motor sequences. PMID:26448054

  17. An Advantage for Perceptual Edges in Young Infants' Memory for Speech

    ERIC Educational Resources Information Center

    Hochmann, Jean-Rémy; Langus, Alan; Mehler, Jacques

    2016-01-01

    Models of language acquisition are constrained by the information that learners can extract from their input. Experiment 1 investigated whether 3-month-old infants are able to encode a repeated, unsegmented sequence of five syllables. Event-related-potentials showed that infants reacted to a change of the initial or the final syllable, but not to…

  18. Perception of resyllabification in French.

    PubMed

    Gaskell, M Gareth; Spinelli, Elsa; Meunier, Fanny

    2002-07-01

    In three experiments, we examined the effects of phonological resyllabification processes on the perception of French speech. Enchainment involves the resyllabification of a word-final consonant across a syllable boundary (e.g., in chaque avion, the /k/ crosses the syllable boundary to become syllable initial). Liaison involves a further process of realization of a latent consonant, alongside resyllabification (e.g., the /t/ in petit avion). If the syllable is a dominant unit of perception in French (Mehler, Dommergues, Frauenfelder, & Segui, 1981), these processes should cause problems for recognition of the following word. A cross-modal priming experiment showed no cost attached to either type of resyllabification in terms of reduced activation of the following word. Furthermore, word- and sequence-monitoring experiments again showed no cost and suggested that the recognition of vowel-initial words may be facilitated when they are preceded by a word that had undergone resyllabification through enchainment or liaison. We examine the sources of information that could underpin facilitation and propose a refinement of the syllable's role in the perception of French speech.

  19. Primary phonological planning units in spoken word production are language-specific: Evidence from an ERP study.

    PubMed

    Wang, Jie; Wong, Andus Wing-Kuen; Wang, Suiping; Chen, Hsuan-Chih

    2017-07-19

    It is widely acknowledged in Germanic languages that segments are the primary planning units at the phonological encoding stage of spoken word production. Mixed results, however, have been found in Chinese, and it is still unclear what roles syllables and segments play in planning Chinese spoken word production. In the current study, participants were asked to first prepare and later produce disyllabic Mandarin words upon picture prompts and a response cue while electroencephalogram (EEG) signals were recorded. Each two consecutive pictures implicitly formed a pair of prime and target, whose names shared the same word-initial atonal syllable or the same word-initial segments, or were unrelated in the control conditions. Only syllable repetition induced significant effects on event-related brain potentials (ERPs) after target onset: a widely distributed positivity in the 200- to 400-ms interval and an anterior positivity in the 400- to 600-ms interval. We interpret these to reflect syllable-size representations at the phonological encoding and phonetic encoding stages. Our results provide the first electrophysiological evidence for the distinct role of syllables in producing Mandarin spoken words, supporting a language specificity hypothesis about the primary phonological units in spoken word production.

  20. Neural Encoding and Integration of Learned Probabilistic Sequences in Avian Sensory-Motor Circuitry

    PubMed Central

    Brainard, Michael S.

    2013-01-01

    Many complex behaviors, such as human speech and birdsong, reflect a set of categorical actions that can be flexibly organized into variable sequences. However, little is known about how the brain encodes the probabilities of such sequences. Behavioral sequences are typically characterized by the probability of transitioning from a given action to any subsequent action (which we term “divergence probability”). In contrast, we hypothesized that neural circuits might encode the probability of transitioning to a given action from any preceding action (which we term “convergence probability”). The convergence probability of repeatedly experienced sequences could naturally become encoded by Hebbian plasticity operating on the patterns of neural activity associated with those sequences. To determine whether convergence probability is encoded in the nervous system, we investigated how auditory-motor neurons in vocal premotor nucleus HVC of songbirds encode different probabilistic characterizations of produced syllable sequences. We recorded responses to auditory playback of pseudorandomly sequenced syllables from the bird's repertoire, and found that variations in responses to a given syllable could be explained by a positive linear dependence on the convergence probability of preceding sequences. Furthermore, convergence probability accounted for more response variation than other probabilistic characterizations, including divergence probability. Finally, we found that responses integrated over >7–10 syllables (∼700–1000 ms) with the sign, gain, and temporal extent of integration depending on convergence probability. Our results demonstrate that convergence probability is encoded in sensory-motor circuitry of the song-system, and suggest that encoding of convergence probability is a general feature of sensory-motor circuits. PMID:24198363

  1. The role of prominence in determining the scope of boundary-related lengthening in Greek.

    PubMed

    Katsika, Argyro

    2016-03-01

    This study aims at examining and accounting for the scope of the temporal effect of phrase boundaries. Previous research has indicated that there is an interaction between boundary-related lengthening and prominence such that the former extends towards the nearby prominent syllable. However, it is unclear whether this interaction is due to lexical stress and/or phrasal prominence (marked by pitch accent) and how far towards the prominent syllable the effect extends. Here, we use an electromagnetic articulography (EMA) study of Greek to examine the scope of boundary-related lengthening as a function of lexical stress and pitch accent separately. Boundaries are elicited by the means of a variety of syntactic constructions.. The results show an effect of lexical stress. Phrase-final lengthening affects the articulatory gestures of the phrase-final syllable that are immediately adjacent to the boundary in words with final stress, but is initiated earlier within phrase-final words with non-final stress. Similarly, the articulatory configurations during inter-phrasal pauses reach their point of achievement later in words with final stress than in words with non-final stress. These effects of stress hold regardless of whether the phrase-final word is accented or de-accented. Phrase-initial lengthening, on the other hand, is consistently detected on the phrase-initial constriction, independently of where the stress is within the preceding, phrase-final, word. These results indicate that the lexical aspect of prominence plays a role in determining the scope of boundary-related lengthening in Greek. Based on these results, a gestural account of prosodic boundaries in Greek is proposed in which lexical and phrasal prosody interact in a systematic and coordinated fashion. The cross-linguistic dimensions of this account and its implications for prosodic structure are discussed.

  2. Phrase-Final Words in Greek Storytelling Speech: A Study on the Effect of a Culturally-Specific Prosodic Feature on Short-Term Memory.

    PubMed

    Loutrari, Ariadne; Tselekidou, Freideriki; Proios, Hariklia

    2018-02-27

    Prosodic patterns of speech appear to make a critical contribution to memory-related processing. We considered the case of a previously unexplored prosodic feature of Greek storytelling and its effect on free recall in thirty typically developing children between the ages of 10 and 12 years, using short ecologically valid auditory stimuli. The combination of a falling pitch contour and, more notably, extensive final-syllable vowel lengthening, which gives rise to the prosodic feature in question, led to statistically significantly higher performance in comparison to neutral phrase-final prosody. Number of syllables in target words did not reveal substantial difference in performance. The current study presents a previously undocumented culturally-specific prosodic pattern and its effect on short-term memory.

  3. Bilateral coordination and the motor basis of female preference for sexual signals in canary song

    PubMed Central

    Suthers, Roderick A.; Vallet, Eric; Kreutzer, Michel

    2012-01-01

    SUMMARY The preference of female songbirds for particular traits in the songs of courting males has received considerable attention, but the relationship of preferred traits to male quality is poorly understood. Female domestic canaries (Serinus canaria, Linnaeus) preferentially solicit copulation with males that sing special high repetition rate, wide-band, multi-note syllables, called ‘sexy’ or A-syllables. Syllables are separated by minibreaths but each note is produced by pulsatile expiration, allowing high repetition rates and long duration phrases. The wide bandwidth is achieved by including two notes produced sequentially on opposite sides of the syrinx, in which the left and right sides are specialized for low or high frequencies, respectively. The emphasis of low frequencies is facilitated by a positive relationship between syllable repetition rate and the bandwidth of the fundamental frequency of notes sung by the left syrinx, such that bandwidth increases with increasing syllable repetition rate. The temporal offset between notes prevents cheating by unilaterally singing a note on the left side with a low fundamental frequency and prominent higher harmonics. The syringeal and respiratory motor patterns by which sexy syllables are produced support the hypothesis that these syllables provide a sensitive vocal–auditory indicator of a male's performance limit for the rapid, precisely coordinated interhemispheric switching, which is essential for many sensory and motor processes involving specialized contributions from each cerebral hemisphere. PMID:22875764

  4. Verbal Positional Memory in 7-Month-Olds

    ERIC Educational Resources Information Center

    Benavides-Varela, Silvia; Mehler, Jacques

    2015-01-01

    Verbal memory is a fundamental prerequisite for language learning. This study investigated 7-month-olds' (N = 62) ability to remember the identity and order of elements in a multisyllabic word. The results indicate that infants detect changes in the order of edge syllables, or the identity of the middle syllables, but fail to encode the order…

  5. On the Edge of Language Acquisition: Inherent Constraints on Encoding Multisyllabic Sequences in the Neonate Brain

    ERIC Educational Resources Information Center

    Ferry, Alissa L.; Fló, Ana; Brusini, Perrine; Cattarossi, Luigi; Macagno, Francesco; Nespor, Marina; Mehler, Jacques

    2016-01-01

    To understand language, humans must encode information from rapid, sequential streams of syllables--tracking their order and organizing them into words, phrases, and sentences. We used Near-Infrared Spectroscopy (NIRS) to determine whether human neonates are born with the capacity to track the positions of syllables in multisyllabic sequences.…

  6. The Different Time Course of Phonotactic Constraint Learning in Children and Adults: Evidence from Speech Errors

    ERIC Educational Resources Information Center

    Smalle, Eleonore H. M.; Muylle, Merel; Szmalec, Arnaud; Duyck, Wouter

    2017-01-01

    Speech errors typically respect the speaker's implicit knowledge of language-wide phonotactics (e.g., /t/ cannot be a syllable onset in the English language). Previous work demonstrated that adults can learn novel experimentally induced phonotactic constraints by producing syllable strings in which the allowable position of a phoneme depends on…

  7. For a Psycholinguistic Model of Handwriting Production: Testing the Syllable-Bigram Controversy

    ERIC Educational Resources Information Center

    Kandel, Sonia; Peereman, Ronald; Grosjacques, Geraldine; Fayol, Michel

    2011-01-01

    This study examined the theoretical controversy on the impact of syllables and bigrams in handwriting production. French children and adults wrote words on a digitizer so that we could collect data on the local, online processing of handwriting production. The words differed in the position of the lowest frequency bigram. In one condition, it…

  8. Infant Directed Speech Enhances Statistical Learning in Newborn Infants: An ERP Study

    PubMed Central

    Teinonen, Tuomas; Tervaniemi, Mari; Huotilainen, Minna

    2016-01-01

    Statistical learning and the social contexts of language addressed to infants are hypothesized to play important roles in early language development. Previous behavioral work has found that the exaggerated prosodic contours of infant-directed speech (IDS) facilitate statistical learning in 8-month-old infants. Here we examined the neural processes involved in on-line statistical learning and investigated whether the use of IDS facilitates statistical learning in sleeping newborns. Event-related potentials (ERPs) were recorded while newborns were exposed to12 pseudo-words, six spoken with exaggerated pitch contours of IDS and six spoken without exaggerated pitch contours (ADS) in ten alternating blocks. We examined whether ERP amplitudes for syllable position within a pseudo-word (word-initial vs. word-medial vs. word-final, indicating statistical word learning) and speech register (ADS vs. IDS) would interact. The ADS and IDS registers elicited similar ERP patterns for syllable position in an early 0–100 ms component but elicited different ERP effects in both the polarity and topographical distribution at 200–400 ms and 450–650 ms. These results provide the first evidence that the exaggerated pitch contours of IDS result in differences in brain activity linked to on-line statistical learning in sleeping newborns. PMID:27617967

  9. Effects of blocking and presentation on the recognition of word and nonsense syllables in noise

    NASA Astrophysics Data System (ADS)

    Benkí, José R.

    2003-10-01

    Listener expectations may have significant effects on spoken word recognition, modulating word similarity effects from the lexicon. This study investigates the effect of blocking by lexical status on the recognition of word and nonsense syllables in noise. 240 phonemically matched word and nonsense CVC syllables [Boothroyd and Nittrouer, J. Acoust. Soc. Am. 84, 101-108 (1988)] were presented to listeners at different S/N ratios for identification. In the mixed condition, listeners were presented with blocks containing both words and nonwords, while listeners in the blocked condition were presented with the trials in blocks containing either words or nonwords. The targets were presented in isolation with 50 ms of preceding and following noise. Preliminary results indicate no effect of blocking on accuracy for either word or nonsense syllables; results from neighborhood density analyses will be presented. Consistent with previous studies, a j-factor analysis indicates that words are perceived as containing at least 0.5 fewer independent units than nonwords in both conditions. Relative to previous work on syllables presented in a frame sentence [Benkí, J. Acoust. Soc. Am. 113, 1689-1705 (2003)], initial consonants were perceived significantly less accurately, while vowels and final consonants were perceived at comparable rates.

  10. Perceptual invariance of coarticulated vowels over variations in speaking rate.

    PubMed

    Stack, Janet W; Strange, Winifred; Jenkins, James J; Clarke, William D; Trent, Sonja A

    2006-04-01

    This study examined the perception and acoustics of a large corpus of vowels spoken in consonant-vowel-consonant syllables produced in citation-form (lists) and spoken in sentences at normal and rapid rates by a female adult. Listeners correctly categorized the speaking rate of sentence materials as normal or rapid (2% errors) but did not accurately classify the speaking rate of the syllables when they were excised from the sentences (25% errors). In contrast, listeners accurately identified the vowels produced in sentences spoken at both rates when presented the sentences and when presented the excised syllables blocked by speaking rate or randomized. Acoustical analysis showed that formant frequencies at syllable midpoint for vowels in sentence materials showed "target undershoot" relative to citation-form values, but little change over speech rate. Syllable durations varied systematically with vowel identity, speaking rate, and voicing of final consonant. Vowel-inherent-spectral-change was invariant in direction of change over rate and context for most vowels. The temporal location of maximum F1 frequency further differentiated spectrally adjacent lax and tense vowels. It was concluded that listeners were able to utilize these rate- and context-independent dynamic spectrotemporal parameters to identify coarticulated vowels, even when sentential information about speaking rate was not available.

  11. Examination of the Locus of Positional Effects on Children's Production of Plural -s: Considerations From Local and Global Speech Planning.

    PubMed

    Theodore, Rachel M; Demuth, Katherine; Shattuck-Hufnagel, Stefanie

    2015-06-01

    Prosodic and articulatory factors influence children's production of inflectional morphemes. For example, plural -s is produced more reliably in utterance-final compared to utterance-medial position (i.e., the positional effect), which has been attributed to the increased planning time in utterance-final position. In previous investigations of plural -s, utterance-medial plurals were followed by a stop consonant (e.g., dogsbark), inducing high articulatory complexity. We examined whether the positional effect would be observed if the utterance-medial context were simplified to a following vowel. An elicited imitation task was used to collect productions of plural nouns from 2-year-old children. Nouns were elicited utterance-medially and utterance-finally, with the medial plural followed by either a stressed or an unstressed vowel. Acoustic analysis was used to identify evidence of morpheme production. The positional effect was absent when the morpheme was followed by a vowel (e.g., dogseat). However, it returned when the vowel-initial word contained 2 syllables (e.g., dogsarrive), suggesting that the increased processing load in the latter condition negated the facilitative effect of the easy articulatory context. Children's productions of grammatical morphemes reflect a rich interaction between emerging levels of linguistic competence, raising considerations for diagnosis and rehabilitation of language disorders.

  12. Stress priming in picture naming: an SOA study.

    PubMed

    Schiller, Niels O; Fikkert, Paula; Levelt, Clara C

    2004-01-01

    This study investigates whether or not the representation of lexical stress information can be primed during speech production. In four experiments, we attempted to prime the stress position of bisyllabic target nouns (picture names) having initial and final stress with auditory prime words having either the same or different stress as the target (e.g., WORtel-MOtor vs. koSTUUM-MOtor; capital letters indicate stressed syllables in prime-target pairs). Furthermore, half of the prime words were semantically related, the other half unrelated. Overall, picture names were not produced faster when the prime word had the same stress as the target than when the prime had different stress, i.e., there was no stress-priming effect in any experiment. This result would not be expected if stress were stored in the lexicon. However, targets with initial stress were responded to faster than final-stress targets. The reason for this effect was neither the quality of the pictures nor frequency of occurrence or voice-key characteristics. We hypothesize here that this stress effect is a genuine encoding effect, i.e., words with stress on the second syllable take longer to be encoded because their stress pattern is irregular with respect to the lexical distribution of bisyllabic stress patterns, even though it can be regular with respect to metrical stress rules in Dutch. The results of the experiments are discussed in the framework of models of phonological encoding.

  13. An Experimental Approach to Debuccalization and Supplementary Gestures

    ERIC Educational Resources Information Center

    O'Brien, Jeremy

    2012-01-01

    Debuccalization is a weakening phenomenon whereby various consonants reduce to laryngeals. Examples include Spanish s-aspiration (s becomes h word-finally) and English t-glottalization (t becomes glottal stop syllable-finally). Previous analyses of debuccalization view it as a lenition process that deletes or manipulates formal phonological…

  14. Audio-visual onset differences are used to determine syllable identity for ambiguous audio-visual stimulus pairs

    PubMed Central

    ten Oever, Sanne; Sack, Alexander T.; Wheat, Katherine L.; Bien, Nina; van Atteveldt, Nienke

    2013-01-01

    Content and temporal cues have been shown to interact during audio-visual (AV) speech identification. Typically, the most reliable unimodal cue is used more strongly to identify specific speech features; however, visual cues are only used if the AV stimuli are presented within a certain temporal window of integration (TWI). This suggests that temporal cues denote whether unimodal stimuli belong together, that is, whether they should be integrated. It is not known whether temporal cues also provide information about the identity of a syllable. Since spoken syllables have naturally varying AV onset asynchronies, we hypothesize that for suboptimal AV cues presented within the TWI, information about the natural AV onset differences can aid in speech identification. To test this, we presented low-intensity auditory syllables concurrently with visual speech signals, and varied the stimulus onset asynchronies (SOA) of the AV pair, while participants were instructed to identify the auditory syllables. We revealed that specific speech features (e.g., voicing) were identified by relying primarily on one modality (e.g., auditory). Additionally, we showed a wide window in which visual information influenced auditory perception, that seemed even wider for congruent stimulus pairs. Finally, we found a specific response pattern across the SOA range for syllables that were not reliably identified by the unimodal cues, which we explained as the result of the use of natural onset differences between AV speech signals. This indicates that temporal cues not only provide information about the temporal integration of AV stimuli, but additionally convey information about the identity of AV pairs. These results provide a detailed behavioral basis for further neuro-imaging and stimulation studies to unravel the neurofunctional mechanisms of the audio-visual-temporal interplay within speech perception. PMID:23805110

  15. An ERP Study of Good Production vis-à-vis Poor Perception of Tones in Cantonese: Implications for Top-Down Speech Processing

    PubMed Central

    Law, Sam-Po; Fung, Roxana; Kung, Carmen

    2013-01-01

    This study investigated a theoretically challenging dissociation between good production and poor perception of tones among neurologically unimpaired native speakers of Cantonese. The dissociation is referred to as the near-merger phenomenon in sociolinguistic studies of sound change. In a passive oddball paradigm, lexical and nonlexical syllables of the T1/T6 and T4/T6 contrasts were presented to elicit the mismatch negativity (MMN) and P3a from two groups of participants, those who could produce and distinguish all tones in the language (Control) and those who could produce all tones but specifically failed to distinguish between T4 and T6 in perception (Dissociation). The presence of MMN to T1/T6 and null response to T4/T6 of lexical syllables in the dissociation group confirmed the near-merger phenomenon. The observation that the control participants exhibited a statistically reliable MMN to lexical syllables of T1/T6, weaker responses to nonlexical syllables of T1/T6 and lexical syllables of T4/T6, and finally null response to nonlexical syllables of T4/T6, suggests the involvement of top-down processing in speech perception. Furthermore, the stronger P3a response of the control group, compared with the dissociation group in the same experimental conditions, may be taken to indicate higher cognitive capability in attention switching, auditory attention or memory in the control participants. This cognitive difference, together with our speculation that constant top-down predictions without complete bottom-up analysis of acoustic signals in speech recognition may reduce one’s sensitivity to small acoustic contrasts, account for the occurrence of dissociation in some individuals but not others. PMID:23342146

  16. Audio-visual onset differences are used to determine syllable identity for ambiguous audio-visual stimulus pairs.

    PubMed

    Ten Oever, Sanne; Sack, Alexander T; Wheat, Katherine L; Bien, Nina; van Atteveldt, Nienke

    2013-01-01

    Content and temporal cues have been shown to interact during audio-visual (AV) speech identification. Typically, the most reliable unimodal cue is used more strongly to identify specific speech features; however, visual cues are only used if the AV stimuli are presented within a certain temporal window of integration (TWI). This suggests that temporal cues denote whether unimodal stimuli belong together, that is, whether they should be integrated. It is not known whether temporal cues also provide information about the identity of a syllable. Since spoken syllables have naturally varying AV onset asynchronies, we hypothesize that for suboptimal AV cues presented within the TWI, information about the natural AV onset differences can aid in speech identification. To test this, we presented low-intensity auditory syllables concurrently with visual speech signals, and varied the stimulus onset asynchronies (SOA) of the AV pair, while participants were instructed to identify the auditory syllables. We revealed that specific speech features (e.g., voicing) were identified by relying primarily on one modality (e.g., auditory). Additionally, we showed a wide window in which visual information influenced auditory perception, that seemed even wider for congruent stimulus pairs. Finally, we found a specific response pattern across the SOA range for syllables that were not reliably identified by the unimodal cues, which we explained as the result of the use of natural onset differences between AV speech signals. This indicates that temporal cues not only provide information about the temporal integration of AV stimuli, but additionally convey information about the identity of AV pairs. These results provide a detailed behavioral basis for further neuro-imaging and stimulation studies to unravel the neurofunctional mechanisms of the audio-visual-temporal interplay within speech perception.

  17. An ERP study of good production vis-à-vis poor perception of tones in Cantonese: implications for top-down speech processing.

    PubMed

    Law, Sam-Po; Fung, Roxana; Kung, Carmen

    2013-01-01

    This study investigated a theoretically challenging dissociation between good production and poor perception of tones among neurologically unimpaired native speakers of Cantonese. The dissociation is referred to as the near-merger phenomenon in sociolinguistic studies of sound change. In a passive oddball paradigm, lexical and nonlexical syllables of the T1/T6 and T4/T6 contrasts were presented to elicit the mismatch negativity (MMN) and P3a from two groups of participants, those who could produce and distinguish all tones in the language (Control) and those who could produce all tones but specifically failed to distinguish between T4 and T6 in perception (Dissociation). The presence of MMN to T1/T6 and null response to T4/T6 of lexical syllables in the dissociation group confirmed the near-merger phenomenon. The observation that the control participants exhibited a statistically reliable MMN to lexical syllables of T1/T6, weaker responses to nonlexical syllables of T1/T6 and lexical syllables of T4/T6, and finally null response to nonlexical syllables of T4/T6, suggests the involvement of top-down processing in speech perception. Furthermore, the stronger P3a response of the control group, compared with the dissociation group in the same experimental conditions, may be taken to indicate higher cognitive capability in attention switching, auditory attention or memory in the control participants. This cognitive difference, together with our speculation that constant top-down predictions without complete bottom-up analysis of acoustic signals in speech recognition may reduce one's sensitivity to small acoustic contrasts, account for the occurrence of dissociation in some individuals but not others.

  18. Vowel reduction in word-final position by early and late Spanish-English bilinguals.

    PubMed

    Byers, Emily; Yavas, Mehmet

    2017-01-01

    Vowel reduction is a prominent feature of American English, as well as other stress-timed languages. As a phonological process, vowel reduction neutralizes multiple vowel quality contrasts in unstressed syllables. For bilinguals whose native language is not characterized by large spectral and durational differences between tonic and atonic vowels, systematically reducing unstressed vowels to the central vowel space can be problematic. Failure to maintain this pattern of stressed-unstressed syllables in American English is one key element that contributes to a "foreign accent" in second language speakers. Reduced vowels, or "schwas," have also been identified as particularly vulnerable to the co-articulatory effects of adjacent consonants. The current study examined the effects of adjacent sounds on the spectral and temporal qualities of schwa in word-final position. Three groups of English-speaking adults were tested: Miami-based monolingual English speakers, early Spanish-English bilinguals, and late Spanish-English bilinguals. Subjects performed a reading task to examine their schwa productions in fluent speech when schwas were preceded by consonants from various points of articulation. Results indicated that monolingual English and late Spanish-English bilingual groups produced targeted vowel qualities for schwa, whereas early Spanish-English bilinguals lacked homogeneity in their vowel productions. This extends prior claims that schwa is targetless for F2 position for native speakers to highly-proficient bilingual speakers. Though spectral qualities lacked homogeneity for early Spanish-English bilinguals, early bilinguals produced schwas with near native-like vowel duration. In contrast, late bilinguals produced schwas with significantly longer durations than English monolinguals or early Spanish-English bilinguals. Our results suggest that the temporal properties of a language are better integrated into second language phonologies than spectral qualities. Finally, we examined the role of nonstructural variables (e.g. linguistic history measures) in predicting native-like vowel duration. These factors included: Age of L2 learning, amount of L1 use, and self-reported bilingual dominance. Our results suggested that different sociolinguistic factors predicted native-like reduced vowel duration than predicted native-like vowel qualities across multiple phonetic environments.

  19. Vowel reduction in word-final position by early and late Spanish-English bilinguals

    PubMed Central

    2017-01-01

    Vowel reduction is a prominent feature of American English, as well as other stress-timed languages. As a phonological process, vowel reduction neutralizes multiple vowel quality contrasts in unstressed syllables. For bilinguals whose native language is not characterized by large spectral and durational differences between tonic and atonic vowels, systematically reducing unstressed vowels to the central vowel space can be problematic. Failure to maintain this pattern of stressed-unstressed syllables in American English is one key element that contributes to a “foreign accent” in second language speakers. Reduced vowels, or “schwas,” have also been identified as particularly vulnerable to the co-articulatory effects of adjacent consonants. The current study examined the effects of adjacent sounds on the spectral and temporal qualities of schwa in word-final position. Three groups of English-speaking adults were tested: Miami-based monolingual English speakers, early Spanish-English bilinguals, and late Spanish-English bilinguals. Subjects performed a reading task to examine their schwa productions in fluent speech when schwas were preceded by consonants from various points of articulation. Results indicated that monolingual English and late Spanish-English bilingual groups produced targeted vowel qualities for schwa, whereas early Spanish-English bilinguals lacked homogeneity in their vowel productions. This extends prior claims that schwa is targetless for F2 position for native speakers to highly-proficient bilingual speakers. Though spectral qualities lacked homogeneity for early Spanish-English bilinguals, early bilinguals produced schwas with near native-like vowel duration. In contrast, late bilinguals produced schwas with significantly longer durations than English monolinguals or early Spanish-English bilinguals. Our results suggest that the temporal properties of a language are better integrated into second language phonologies than spectral qualities. Finally, we examined the role of nonstructural variables (e.g. linguistic history measures) in predicting native-like vowel duration. These factors included: Age of L2 learning, amount of L1 use, and self-reported bilingual dominance. Our results suggested that different sociolinguistic factors predicted native-like reduced vowel duration than predicted native-like vowel qualities across multiple phonetic environments. PMID:28384234

  20. Compensatory lengthening in the Spanish of Havana, Cuba: Acoustic analyses of word-internal, post-nuclear /l/ and /r/

    NASA Astrophysics Data System (ADS)

    Carlson, Kristin M.

    Given the geographic, demographic, and historical importance of Cuba vis-a-vis the dissemination of language and culture throughout the Hispanic Caribbean, one would naturally anticipate a larger corpus of scientifically-noteworthy linguistic publications on Cuban Spanish, which is far from the actual case. Moreover, the gemination of an onset positionally subsequent to the deletion of a syllable-final liquid (generally termed liquid gemination in the literature) has been repeatedly claimed yet remarkably unsubstantiated as a pervasive characteristic of Cuban Spanish, particularly of the western dialect region (cf. Alfaraz (2000, 2007, 2008), Casanellas and Alamo (1985), Choy Lopez (1985, 1988, 1989), Costa Sanchez (1987), Darias Concepcion (2001, 2005), Dohotaru (2002, 2007), Figueroa Esteva and Dohotaru (1994), Garcia Perez (2006), Garcia Riveron (1991), Haden and Matluck (1973, 1974, 1977), Isbǎsescu (1965, 1968), Lamb (1968), Levina (1970), Montero Bernal (1990, 2002, 2007a, b), Ringer Uber (1986), Ruiz Hernandez (1978), Sosa (1974), Terrell (1976), Trista and Valdes (1978), Valdes Acosta (1980), and Vera Riveron (2000)). As a result, in the interest of supplementing all antecedent work concerning the allophony of final liquids as well as affording a more descriptively-precise account of the allophony of word-internal, post-nuclear /l/ and /[Special character omitted]/ in Cuban Spanish in addition to expressly addressing the need for empirical data-collection and analysis processes, the present investigation was specifically designed and implemented to acoustically investigate the phenomenon of gemination as it is purported to occur in the Spanish of the region of Havana, Cuba: more specifically, (1) to acoustically examine the qualitative and quantitative patternings of post-nuclear /l/ and /[Special character omitted]/ within the word; and (2) to statistically evaluate the relationship between gemination and eight independent variables: gender, age group, educational level, morphological significance of the syllable closed by the liquid phoneme, position of the syllable closed by the liquid phoneme relative to stress placement, voicing specification, manner of articulation, and place of articulation of the surface realization of the immediately-following onset. Qualitative acoustic analyses of the 1,895 tokens of word-internal, post-nuclear /l/ (n=469) and /[Special character omitted]/ (n=1,426) extracted from the corpus of informal data demonstrated seven allophones for the lateral liquid phoneme (O, [l], [d[Special character omitted

  1. Medial-Vowel Writing Difficulty in Korean Syllabic Writing: A Characteristic Sign of Alzheimer's Disease

    PubMed Central

    Yoon, Ji Hye; Jeong, Yong

    2018-01-01

    Background and Purpose Korean-speaking patients with a brain injury may show agraphia that differs from that of English-speaking patients due to the unique features of Hangul syllabic writing. Each grapheme in Hangul must be arranged from left to right and/or top to bottom within a square space to form a syllable, which requires greater visuospatial abilities than when writing the letters constituting an alphabetic writing system. Among the Hangul grapheme positions within a syllable, the position of a vowel is important because it determines the writing direction and the whole configuration in Korean syllabic writing. Due to the visuospatial characteristics of the Hangul vowel, individuals with early-onset Alzheimer's disease (EOAD) may experiences differences between the difficulties of writing Hangul vowels and consonants due to prominent visuospatial dysfunctions caused by parietal lesions. Methods Eighteen patients with EOAD and 18 age-and-education-matched healthy adults participated in this study. The participants were requested to listen to and write 30 monosyllabic characters that consisted of an initial consonant, medial vowel, and final consonant with a one-to-one phoneme-to-grapheme correspondence. We measured the writing time for each grapheme, the pause time between writing the initial consonant and the medial vowel (P1), and the pause time between writing the medial vowel and the final consonant (P2). Results All grapheme writing and pause times were significantly longer in the EOAD group than in the controls. P1 was also significantly longer than P2 in the EOAD group. Conclusions Patients with EOAD might require a higher judgment ability and longer processing time for determining the visuospatial grapheme position before writing medial vowels. This finding suggests that a longer pause time before writing medial vowels is an early marker of visuospatial dysfunction in patients with EOAD. PMID:29504296

  2. Medial-Vowel Writing Difficulty in Korean Syllabic Writing: A Characteristic Sign of Alzheimer's Disease.

    PubMed

    Yoon, Ji Hye; Jeong, Yong; Na, Duk L

    2018-04-01

    Korean-speaking patients with a brain injury may show agraphia that differs from that of English-speaking patients due to the unique features of Hangul syllabic writing. Each grapheme in Hangul must be arranged from left to right and/or top to bottom within a square space to form a syllable, which requires greater visuospatial abilities than when writing the letters constituting an alphabetic writing system. Among the Hangul grapheme positions within a syllable, the position of a vowel is important because it determines the writing direction and the whole configuration in Korean syllabic writing. Due to the visuospatial characteristics of the Hangul vowel, individuals with early-onset Alzheimer's disease (EOAD) may experiences differences between the difficulties of writing Hangul vowels and consonants due to prominent visuospatial dysfunctions caused by parietal lesions. Eighteen patients with EOAD and 18 age-and-education-matched healthy adults participated in this study. The participants were requested to listen to and write 30 monosyllabic characters that consisted of an initial consonant, medial vowel, and final consonant with a one-to-one phoneme-to-grapheme correspondence. We measured the writing time for each grapheme, the pause time between writing the initial consonant and the medial vowel (P1), and the pause time between writing the medial vowel and the final consonant (P2). All grapheme writing and pause times were significantly longer in the EOAD group than in the controls. P1 was also significantly longer than P2 in the EOAD group. Patients with EOAD might require a higher judgment ability and longer processing time for determining the visuospatial grapheme position before writing medial vowels. This finding suggests that a longer pause time before writing medial vowels is an early marker of visuospatial dysfunction in patients with EOAD. Copyright © 2018 Korean Neurological Association.

  3. Final Syllable Lengthening (FSL) in infant vocalizations.

    PubMed

    Nathani, Suneeti; Oller, D Kimbrough; Cobo-Lewis, Alan B

    2003-02-01

    Final Syllable Lengthening (FSL) has been extensively examined in infant vocalizations in order to determine whether its basis is biological or learned. Findings suggest there may be a U-shaped developmental trajectory for FSL. The present study sought to verify this pattern and to determine whether vocal maturity and deafness influence FSL. Eight normally hearing infants, aged 0;3 to 1;0, and eight deaf infants, aged 0;8 to 4;0, were examined at three levels of prelinguistic vocal development: precanonical, canonical, and postcanonical. FSL was found at all three levels suggesting a biological basis for this phenomenon. Individual variability was, however, considerable. Reduction in the magnitude of FSL across the three sessions provided some support for a downward trend for FSL in infancy. Findings further indicated that auditory deprivation can significantly affect temporal aspects of infant speech production.

  4. Time course of syllabic and sub-syllabic processing in Mandarin word production: Evidence from the picture-word interference paradigm.

    PubMed

    Wang, Jie; Wong, Andus Wing-Kuen; Chen, Hsuan-Chih

    2017-06-05

    The time course of phonological encoding in Mandarin monosyllabic word production was investigated by using the picture-word interference paradigm. Participants were asked to name pictures in Mandarin while visual distractor words were presented before, at, or after picture onset (i.e., stimulus-onset asynchrony/SOA = -100, 0, or +100 ms, respectively). Compared with the unrelated control, the distractors sharing atonal syllables with the picture names significantly facilitated the naming responses at -100- and 0-ms SOAs. In addition, the facilitation effect of sharing word-initial segments only appeared at 0-ms SOA, and null effects were found for sharing word-final segments. These results indicate that both syllables and subsyllabic units play important roles in Mandarin spoken word production and more critically that syllabic processing precedes subsyllabic processing. The current results lend strong support to the proximate units principle (O'Seaghdha, Chen, & Chen, 2010), which holds that the phonological structure of spoken word production is language-specific and that atonal syllables are the proximate phonological units in Mandarin Chinese. On the other hand, the significance of word-initial segments over word-final segments suggests that serial processing of segmental information seems to be universal across Germanic languages and Chinese, which remains to be verified in future studies.

  5. Rhythmic speech and stuttering reduction in a syllable-timed language.

    PubMed

    Law, Thomas; Packman, Ann; Onslow, Mark; To, Carol K-S; Tong, Michael C-F; Lee, Kathy Y-S

    2018-06-06

    Speaking rhythmically, also known as syllable-timed speech (STS), has been known for centuries to be a fluency-inducing condition for people who stutter. Cantonese is a tonal syllable-timed language and it has been shown that, of all languages, Cantonese is the most rhythmic (Mok, 2009). However, it is not known if STS reduces stuttering in Cantonese as it does in English. This is the first study to investigate the effects of STS on stuttering in a syllable-timed language. Nineteen native Cantonese-speaking adults who stutter were engaged in conversational tasks in Cantonese under two conditions: one in their usual speaking style and one using STS. The speakers' percentage syllables stuttered (%SS) and speech rhythmicity were rated. The rhythmicity ratings were used to estimate the extent to which speakers were using STS in the syllable-timed condition. Results revealed a statistically significant reduction in %SS in the STS condition; however, this reduction was not as large as in previous studies in other languages and the amount of stuttering reduction varied across speakers. The rhythmicity ratings showed that some speakers were perceived to be speaking more rhythmically than others and that the perceived rhythmicity correlated positively with reductions in stuttering. The findings were unexpected, as it was anticipated that speakers of a highly rhythmic language such as Cantonese would find STS easy to use and that the consequent reductions in stuttering would be great, even greater perhaps than in a stress-timed language such as English. The theoretical and clinical implications of the findings are discussed.

  6. LISTENERS’ PERCEPTION OF “COMPENSATORY SHORTENING

    PubMed Central

    Fowler, Carol A.; Thompson, Jaqueline M.

    2010-01-01

    English exhibits “compensatory shortening” whereby a stressed syllable followed by an unstressed syllable is measured to be shorter than the same stressed syllable alone. This “anticipatory” shortening is much greater than “backward” shortening whereby an unstressed syllable is measured to shorten a following stressed syllable. We speculated that measured shortening reflects, not true shortening, but coarticulatory hiding. Hence, we asked whether listeners are sensitive to parts of stressed syllables hidden by following or preceding unstressed syllables. In two experiments, we found the point of subjective equality, that is the durational difference between a stressed syllable in isolation and one followed by an unstressed syllable, at which listeners cannot tell which is longer. A third experiment found the point of subjective equality for stressed monosyllables and disyllables with a weak-strong stress pattern. In all experiments, the points of subjective equality occurred when stressed syllables in disyllables were measured to be shorter than those in monosyllables as if listeners hear the coarticulatory onsest or continuation of a stressed syllable within unstressed syllables. PMID:20139461

  7. Automated classification of mouse pup isolation syllables: from cluster analysis to an Excel-based "mouse pup syllable classification calculator".

    PubMed

    Grimsley, Jasmine M S; Gadziola, Marie A; Wenstrup, Jeffrey J

    2012-01-01

    Mouse pups vocalize at high rates when they are cold or isolated from the nest. The proportions of each syllable type produced carry information about disease state and are being used as behavioral markers for the internal state of animals. Manual classifications of these vocalizations identified 10 syllable types based on their spectro-temporal features. However, manual classification of mouse syllables is time consuming and vulnerable to experimenter bias. This study uses an automated cluster analysis to identify acoustically distinct syllable types produced by CBA/CaJ mouse pups, and then compares the results to prior manual classification methods. The cluster analysis identified two syllable types, based on their frequency bands, that have continuous frequency-time structure, and two syllable types featuring abrupt frequency transitions. Although cluster analysis computed fewer syllable types than manual classification, the clusters represented well the probability distributions of the acoustic features within syllables. These probability distributions indicate that some of the manually classified syllable types are not statistically distinct. The characteristics of the four classified clusters were used to generate a Microsoft Excel-based mouse syllable classifier that rapidly categorizes syllables, with over a 90% match, into the syllable types determined by cluster analysis.

  8. The speech focus position effect on jaw-finger coordination in a pointing task.

    PubMed

    Rochet-Capellan, Amélie; Laboissière, Rafael; Galván, Arturo; Schwartz, Jean-Luc

    2008-12-01

    This article investigates jaw-finger coordination in a task involving pointing to a target while naming it with a CVCV (e.g., /papa/) versus CVCV (e.g., /papa/) word. According to the authors' working hypothesis, the pointing apex (gesture extremum) would be synchronized with the apex of the jaw-opening gesture corresponding to the stressed syllable. Jaw and finger motions were recorded using Optotrak (Northern Digital, Waterloo, Ontario, Canada). The effects of stress position on jaw-finger coordination were tested across different target positions (near vs. far) and different consonants in the target word (/t/ vs. /p/). Twenty native Portuguese Brazilian speakers participated in the experiment (all conditions). Jaw response starts earlier, and finger-target alignment period is longer for CVCV words than for CVCV ones. The apex of the jaw-opening gesture for the stressed syllable appears synchronized with the onset of the finger-target alignment period (corresponding to the pointing apex) for CVCV words and with the offset of that period for CVCV words. For both stress conditions, the stressed syllable occurs within the finger-target alignment period because of tight finger-jaw coordination. This result is interpreted as evidence for an anchoring of the speech deictic site (part of speech that shows) in the pointing gesture.

  9. Atypical mismatch negativity to distressful voices associated with conduct disorder symptoms.

    PubMed

    Hung, An-Yi; Ahveninen, Jyrki; Cheng, Yawei

    2013-09-01

    Although a general consensus holds that emotional reactivity in youth with conduct disorder (CD) symptoms arises as one of the main causes of successive aggression, it remains to be determined whether automatic emotional processing is altered in this population. We measured auditory event-related potentials (ERP) in 20 young offenders and 20 controls, screened for DSM-IV criteria of CD and evaluated using the youth version of Hare Psychopathy Checklist (PCL:YV), State-Trait Anxiety Inventory (STAI) and Barrett Impulsiveness Scale (BIS-11). In an oddball design, sadly or fearfully spoken 'deviant' syllables were randomly presented within a train of emotionally neutral 'standard' syllables. In young offenders meeting with CD criteria, the ERP component mismatch negativity (MMN), presumed to reflect preattentive auditory change detection, was significantly stronger for fearful than sad syllables. No MMN differences for fearful versus sad syllables were observed in controls. Analyses of nonvocal deviants, matched spectrally with the fearful and sad sounds, supported our interpretation that the MMN abnormalities in juvenile offenders were related to the emotional content of sounds, instead of purely acoustic factors. Further, in the young offenders with CD symptoms, strong MMN amplitudes to fearful syllables were associated with high impulsive tendencies (PCL:YV, Factor 2). Higher trait and state anxiety, assessed by STAI, were positively correlated with P3a amplitudes to fearful and sad syllables, respectively. The differences in group-interaction MMN/P3a patterns to emotional syllables and nonvocal sounds could be speculated to suggest that there is a distinct processing route for preattentive processing of species-specific emotional information in human auditory cortices. Our results suggest that youths with CD symptoms may process distressful voices in an atypical fashion already at the preattentive level. This auditory processing abnormality correlated with increased impulsivity and anxiety. Our results may help to shed light on the neural mechanisms of aggression. © 2013 The Authors. Journal of Child Psychology and Psychiatry © 2013 Association for Child and Adolescent Mental Health.

  10. Planning and Articulation in Incremental Word Production: Syllable-Frequency Effects in English

    ERIC Educational Resources Information Center

    Cholin, Joana; Dell, Gary S.; Levelt, Willem J. M.

    2011-01-01

    We investigated the role of syllables during speech planning in English by measuring syllable-frequency effects. So far, syllable-frequency effects in English have not been reported. English has poorly defined syllable boundaries, and thus the syllable might not function as a prominent unit in English speech production. Speakers produced either…

  11. A Sociolinguistic Analysis of Final /s/ in Miami Cuban Spanish

    ERIC Educational Resources Information Center

    Lynch, Andrew

    2009-01-01

    This study analyzes the variation of syllable- and word-final /s/ among two generations of Cubans in Miami, Florida (USA): older, early exile immigrants who arrived in Miami as adults in the 1960s and 1970s, and young Miami-born Cubans whose maternal and paternal grandparents immigrated to Miami from Cuba prior to 1980. Since sibilant weakening is…

  12. Adult perceptions of phonotactic violations in Japanese

    NASA Astrophysics Data System (ADS)

    Fais, Laurel; Kajikawa, Sachiyo; Werker, Janet; Amano, Shigeaki

    2004-05-01

    Adult Japanese speakers ``hear'' epenthetic vowels in productions of Japanese-like words that violate the canonical CVCVCV form by containing internal consonant clusters (CVCCV) [Dupoux et al., J. Exp. Psychol. 25, 1568-1578 (1999)]. Given this finding, this research examined how Japanese adults rated the goodness of Japanese-like words produced without a vowel in the final syllable (CVC), and words produced without vowels in the penultimate and final syllables (CVCC). Furthermore, in some of these contexts, voiceless vowels may appear in fluent, casual Japanese productions, especially in the Kanto dialect, and in some, such voiceless vowels may not appear. Results indicate that both Kanto and Kinki speakers rated CVC productions for contexts in which voiceless vowels are not allowed as the worst; they rated CVC and CVCC contexts in which voiceless vowel productions are allowed as better. In these latter contexts, the CVC words, which result from the loss of one, final, vowel, are judged to be better than the CVCC words, which result from the loss of two (final and penultimate) vowels. These results mirror the relative seriousness of the phonotactic violations and indicate listeners have tacit knowledge of these regularities in their language.

  13. The Role of Lexical Stress on the Use of Vocal Fry in Young Adult Female Speakers.

    PubMed

    Gibson, Todd A

    2017-01-01

    Vocal fry is a voice register often used by young adult women for sociolinguistic purposes. Some acoustic correlates of lexical stress, however, appear incompatible with the use of vocal fry. The objective of this study was to systematically examine the role of lexical stress in the use of vocal fry by young adult women. This is a semi-randomized controlled laboratory study. Fifty female undergraduate students were recorded repeating one-, two-, three-, and four-syllable nonwords that conformed to English phonotactics. Nonwords were presented in order from shorter to longer lengths, with stimuli randomized within syllable length. Perceptual analyses of recordings were augmented by acoustic analyses to identify each syllable in which vocal fry occurred. Eighty-six percent of participants produced at least one episode of vocal fry. Vocal fry was more likely to occur in unstressed than stressed position, and the likelihood increased as distance from the stressed syllable increased. There was considerable variability in the use of vocal fry. Frequent and infrequent users varied on the degree to which they used vocal fry in single-syllable nonwords. Vocal fry use persists among young adult women even in the absence of syntactic and pragmatic influences. Lexical stress appeared to dramatically reduce the use of vocal fry. Patterns of vocal fry use appeared to be different for frequent and infrequent users of this vocal register. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  14. Neural Recruitment for the Production of Native and Novel Speech Sounds

    PubMed Central

    Moser, Dana; Fridriksson, Julius; Bonilha, Leonardo; Healy, Eric W.; Baylis, Gordon; Baker, Julie; Rorden, Chris

    2010-01-01

    Two primary areas of damage have been implicated in apraxia of speech (AOS) based on the time post-stroke: (1) the left inferior frontal gyrus (IFG) in acute patients, and (2) the left anterior insula (aIns) in chronic patients. While AOS is widely characterized as a disorder in motor speech planning, little is known about the specific contributions of each of these regions in speech. The purpose of this study was to investigate cortical activation during speech production with a specific focus on the aIns and the IFG in normal adults. While undergoing sparse fMRI, 30 normal adults completed a 30-minute speech-repetition task consisting of three-syllable nonwords that contained either (a) English (native) syllables or (b) Non-English (novel) syllables. When the novel syllable productions were compared to the native syllable productions, greater neural activation was observed in the aIns and IFG, particularly during the first 10 minutes of the task when novelty was the greatest. Although activation in the aIns remained high throughout the task for novel productions, greater activation was clearly demonstrated when the initial 10 minutes were compared to the final 10 minutes of the task. These results suggest increased activity within an extensive neural network, including the aIns and IFG, when the motor speech system is taxed, such as during the production of novel speech. We speculate that the amount of left aIns recruitment during speech production may be related to the internal construction of the motor speech unit such that the degree of novelty/automaticity would result in more or less demands respectively. The role of the IFG as a storehouse and integrative processor for previously acquired routines is also discussed. PMID:19385020

  15. Monitoring Syllable Boundaries during Speech Production

    ERIC Educational Resources Information Center

    Jansma, Bernadette M.; Schiller, Niels O.

    2004-01-01

    This study investigated the encoding of syllable boundary information during speech production in Dutch. Based on Levelt's model of phonological encoding, we hypothesized segments and syllable boundaries to be encoded in an incremental way. In a self-monitoring experiment, decisions about the syllable affiliation (first or second syllable) of a…

  16. Using the self-select paradigm to delineate the nature of speech motor programming.

    PubMed

    Wright, David L; Robin, Don A; Rhee, Jooyhun; Vaculin, Amber; Jacks, Adam; Guenther, Frank H; Fox, Peter T

    2009-06-01

    The authors examined the involvement of 2 speech motor programming processes identified by S. T. Klapp (1995, 2003) during the articulation of utterances differing in syllable and sequence complexity. According to S. T. Klapp, 1 process, INT, resolves the demands of the programmed unit, whereas a second process, SEQ, oversees the serial order demands of longer sequences. A modified reaction time paradigm was used to assess INT and SEQ demands. Specifically, syllable complexity was dependent on syllable structure, whereas sequence complexity involved either repeated or unique syllabi within an utterance. INT execution was slowed when articulating single syllables in the form CCCV compared to simpler CV syllables. Planning unique syllables within a multisyllabic utterance rather than repetitions of the same syllable slowed INT but not SEQ. The INT speech motor programming process, important for mental syllabary access, is sensitive to changes in both syllable structure and the number of unique syllables in an utterance.

  17. Orthographic vs. Phonologic Syllables in Handwriting Production

    ERIC Educational Resources Information Center

    Kandel, Sonia; Herault, Lucie; Grosjacques, Geraldine; Lambert, Eric; Fayol, Michel

    2009-01-01

    French children program the words they write syllable by syllable. We examined whether the syllable the children use to segment words is determined phonologically (i.e., is derived from speech production processes) or orthographically. Third, 4th and 5th graders wrote on a digitiser words that were mono-syllables phonologically (e.g.…

  18. Auditory-Visual Speech Integration by Adults with and without Language-Learning Disabilities

    ERIC Educational Resources Information Center

    Norrix, Linda W.; Plante, Elena; Vance, Rebecca

    2006-01-01

    Auditory and auditory-visual (AV) speech perception skills were examined in adults with and without language-learning disabilities (LLD). The AV stimuli consisted of congruent consonant-vowel syllables (auditory and visual syllables matched in terms of syllable being produced) and incongruent McGurk syllables (auditory syllable differed from…

  19. Who do you love, your mother or your horse? An event-related brain potential analysis of tone processing in Mandarin Chinese.

    PubMed

    Brown-Schmidt, Sarah; Canseco-Gonzalez, Enriqueta

    2004-03-01

    In Mandarin Chinese, word meaning is partially determined by lexical tone (Wang, 1973). Previous studies suggest that lexical tone is processed as linguistic information and not as pure tonal information (Gandour, 1998; Van Lanker & Fromkin, 1973). The current study explored the online processing of lexical tones. Event-related potentials were obtained from 25 Mandarin speakers while they listened to normal and anomalous sentences containing one of three types of semantic anomalies created by manipulating the tone, the syllable, or both tone and syllable (double-anomaly) of sentence-final words. We hypothesized N400 effects elicited by all three types of anomalies and the largest by the double-anomaly. As expected, all three elicited N400 effects starting approximately 150 ms poststimulus and continuing until 1000 ms in some areas. Surprisingly, onset of the double-anomaly effect was approximately 50 ms later than the rest. Delayed detection of errors in this condition may be responsible for the apparent delay. Slight differences between syllable and tone conditions may be due to the relative timing of these acoustic cues.

  20. Using the Self-Select Paradigm to Delineate the Nature of Speech Motor Programming

    PubMed Central

    Wright, David L.; Robin, Don A.; Rhee, Jooyhun; Vaculin, Amber; Jacks, Adam; Guenther, Frank H.; Fox, Peter T.

    2015-01-01

    Purpose The authors examined the involvement of 2 speech motor programming processes identified by S. T. Klapp (1995, 2003) during the articulation of utterances differing in syllable and sequence complexity. According to S. T. Klapp, 1 process, INT, resolves the demands of the programmed unit, whereas a second process, SEQ, oversees the serial order demands of longer sequences. Method A modified reaction time paradigm was used to assess INT and SEQ demands. Specifically, syllable complexity was dependent on syllable structure, whereas sequence complexity involved either repeated or unique syllabi within an utterance. Results INT execution was slowed when articulating single syllables in the form CCCV compared to simpler CV syllables. Planning unique syllables within a multisyllabic utterance rather than repetitions of the same syllable slowed INT but not SEQ. Conclusions The INT speech motor programming process, important for mental syllabary access, is sensitive to changes in both syllable structure and the number of unique syllables in an utterance. PMID:19474396

  1. Vowel Deletion in Latvian.

    ERIC Educational Resources Information Center

    Karins, A. Krisjanis

    1995-01-01

    Investigates variable deletion of short vowels in word-final unstressed syllables in Latvian spoken in Riga. Affected vowels were almost always inflectional endings and results indicated that internal phonological and prosodic factors (especially distance from main word stress) were the strongest constraints on vowel deletion, along with the…

  2. Highly Complex Syllable Structure: A Typological Study of Its Phonological Characteristics and Diachronic Development

    ERIC Educational Resources Information Center

    Easterday, Shelece Michelle

    2017-01-01

    The syllable is a natural unit of organization in spoken language. Strong cross-linguistic tendencies in syllable size and shape are often explained in terms of a universal preference for the CV structure, a type which is also privileged in abstract models of the syllable. Syllable patterns such as those found in Itelmen "qsa?txt??"…

  3. Speaking rate affects the perception of duration as a suprasegmental lexical-stress cue.

    PubMed

    Reinisch, Eva; Jesse, Alexandra; McQueen, James M

    2011-06-01

    Three categorization experiments investigated whether the speaking rate of a preceding sentence influences durational cues to the perception of suprasegmental lexical-stress patterns. Dutch two-syllable word fragments had to be judged as coming from one of two longer words that matched the fragment segmentally but differed in lexical stress placement. Word pairs contrasted primary stress on either the first versus the second syllable or the first versus the third syllable. Duration of the initial or the second syllable of the fragments and rate of the preceding context (fast vs. slow) were manipulated. Listeners used speaking rate to decide about the degree of stress on initial syllables whether the syllables' absolute durations were informative about stress (Experiment Ia) or not (Experiment Ib). Rate effects on the second syllable were visible only when the initial syllable was ambiguous in duration with respect to the preceding rate context (Experiment 2). Absolute second syllable durations contributed little to stress perception (Experiment 3). These results suggest that speaking rate is used to disambiguate words and that rate-modulated stress cues are more important on initial than noninitial syllables. Speaking rate affects perception of suprasegmental information.

  4. An Avian Basal Ganglia-Forebrain Circuit Contributes Differentially to Syllable Versus Sequence Variability of Adult Bengalese Finch Song

    PubMed Central

    Hampton, Cara M.; Sakata, Jon T.; Brainard, Michael S.

    2009-01-01

    Behavioral variability is important for motor skill learning but continues to be present and actively regulated even in well-learned behaviors. In adult songbirds, two types of song variability can persist and are modulated by social context: variability in syllable structure and variability in syllable sequencing. The degree to which the control of both types of adult variability is shared or distinct remains unknown. The output of a basal ganglia-forebrain circuit, LMAN (the lateral magnocellular nucleus of the anterior nidopallium), has been implicated in song variability. For example, in adult zebra finches, neurons in LMAN actively control the variability of syllable structure. It is unclear, however, whether LMAN contributes to variability in adult syllable sequencing because sequence variability in adult zebra finch song is minimal. In contrast, Bengalese finches retain variability in both syllable structure and syllable sequencing into adulthood. We analyzed the effects of LMAN lesions on the variability of syllable structure and sequencing and on the social modulation of these forms of variability in adult Bengalese finches. We found that lesions of LMAN significantly reduced the variability of syllable structure but not of syllable sequencing. We also found that LMAN lesions eliminated the social modulation of the variability of syllable structure but did not detect significant effects on the modulation of sequence variability. These results show that LMAN contributes differentially to syllable versus sequence variability of adult song and suggest that these forms of variability are regulated by distinct neural pathways. PMID:19357331

  5. Subtlety of Ambient-Language Effects in Babbling: A Study of English- and Chinese-Learning Infants at 8, 10, and 12 Months

    PubMed Central

    Lee, Chia-Cheng; Jhang, Yuna; Chen, Li-mei; Relyea, George; Oller, D. Kimbrough

    2016-01-01

    Prior research on ambient-language effects in babbling has often suggested infants produce language-specific phonological features within the first year. These results have been questioned in research failing to find such effects and challenging the positive findings on methodological grounds. We studied English- and Chinese-learning infants at 8, 10, and 12 months and found listeners could not detect ambient-language effects in the vast majority of infant utterances, but only in items deemed to be words or to contain canonical syllables that may have made them sound like words with language-specific shapes. Thus, the present research suggests the earliest ambient-language effects may be found in emerging lexical items or in utterances influenced by language-specific features of lexical items. Even the ambient-language effects for infant canonical syllables and words were very small compared with ambient-language effects for meaningless but phonotactically well-formed syllable sequences spoken by adult native speakers of English and Chinese. PMID:28496393

  6. Segmental transition of the first syllables of words in Japanese children who stutter: Comparison between word and sentence production.

    PubMed

    Matsumoto, Sachiyo; Ito, Tomohiko

    2016-01-01

    Matsumoto-Shimamori, Ito, Fukuda, & Fukuda (2011) proposed the hypothesis that in Japanese, the transition from the core vowels (i.e. syllable nucleus) of the first syllables of words to the following segments affected the occurrence of stuttering. Moreover, in this transition position, an inter-syllabic transition precipitated more stuttering than an intra-syllabic one (Shimamori & Ito, 2007, 2008). However, these studies have only used word production tasks. The purpose of this study was to investigate whether the same results could be obtained in sentence production tasks. Participants were 28 Japanese school-age children who stutter, ranging in age from 7;3 to 12;7. The frequency of stuttering on words with an inter-syllabic transition was significantly higher than on those having an intra-syllabic transition, not only in isolated words but in the first words of sentences. These results suggested that Matsumoto et al.'s hypothesis could be applicable to the results of sentence production tasks.

  7. Acoustic and Perceptual Effects of Dysarthria in Greek with a Focus on Lexical Stress

    NASA Astrophysics Data System (ADS)

    Papakyritsis, Ioannis

    The field of motor speech disorders in Greek is substantially underresearched. Additionally, acoustic studies on lexical stress in dysarthria are generally very rare (Kim et al. 2010). This dissertation examined the acoustic and perceptual effects of Greek dysarthria focusing on lexical stress. Additional possibly deviant speech characteristics were acoustically analyzed. Data from three dysarthric participants and matched controls was analyzed using a case study design. The analysis of lexical stress was based on data drawn from a single word repetition task that included pairs of disyllabic words differentiated by stress location. This data was acoustically analyzed in terms of the use of the acoustic cues for Greek stress. The ability of the dysarthric participants to signal stress in single words was further assessed in a stress identification task carried out by 14 naive Greek listeners. Overall, the acoustic and perceptual data indicated that, although all three dysarthric speakers presented with some difficulty in the patterning of stressed and unstressed syllables, each had different underlying problems that gave rise to quite distinct patterns of deviant speech characteristics. The atypical use of lexical stress cues in Anna's data obscured the prominence relations of stressed and unstressed syllables to the extent that the position of lexical stress was usually not perceptually transparent. Chris and Maria on the other hand, did not have marked difficulties signaling lexical stress location, although listeners were not 100% successful in the stress identification task. For the most part, Chris' atypical phonation patterns and Maria's very slow rate of speech did not interfere with lexical stress signaling. The acoustic analysis of the lexical stress cues was generally in agreement with the participants' performance in the stress identification task. Interestingly, in all three dysarthric participants, but more so in Anna, targets stressed on the 1st syllable were more impervious to error judgments of lexical stress location than targets stressed on the 2nd syllable, although the acoustic metrics did not always suggest a more appropriate use of lexical stress cues in 1st syllable position. The findings contribute to our limited knowledge of the speech characteristics of dysarthria across different languages.

  8. Comparisons of stuttering frequency during and after speech initiation in unaltered feedback, altered auditory feedback and choral speech conditions.

    PubMed

    Saltuklaroglu, Tim; Kalinowski, Joseph; Robbins, Mary; Crawcour, Stephen; Bowers, Andrew

    2009-01-01

    Stuttering is prone to strike during speech initiation more so than at any other point in an utterance. The use of auditory feedback (AAF) has been found to produce robust decreases in the stuttering frequency by creating an electronic rendition of choral speech (i.e., speaking in unison). However, AAF requires users to self-initiate speech before it can go into effect and, therefore, it might not be as helpful as true choral speech during speech initiation. To examine how AAF and choral speech differentially enhance fluency during speech initiation and in subsequent portions of utterances. Ten participants who stuttered read passages without altered feedback (NAF), under four AAF conditions and under a true choral speech condition. Each condition was blocked into ten 10 s trials separated by 5 s intervals so each trial required 'cold' speech initiation. In the first analysis, comparisons of stuttering frequencies were made across conditions. A second, finer grain analysis involved examining stuttering frequencies on the initial syllable, the subsequent four syllables produced and the five syllables produced immediately after the midpoint of each trial. On average, AAF reduced stuttering by approximately 68% relative to the NAF condition. Stuttering frequencies on the initial syllables were considerably higher than on the other syllables analysed (0.45 and 0.34 for NAF and AAF conditions, respectively). After the first syllable was produced, stuttering frequencies dropped precipitously and remained stable. However, this drop in stuttering frequency was significantly greater (approximately 84%) in the AAF conditions than in the NAF condition (approximately 66%) with frequencies on the last nine syllables analysed averaging 0.15 and 0.05 for NAF and AAF conditions, respectively. In the true choral speech condition, stuttering was virtually (approximately 98%) eliminated across all utterances and all syllable positions. Altered auditory feedback effectively inhibits stuttering immediately after speech has been initiated. However, unlike a true choral signal, which is exogenously initiated and offers the most complete fluency enhancement, AAF requires speech to be initiated by the user and 'fed back' before it can directly inhibit stuttering. It is suggested that AAF can be a viable clinical option for those who stutter and should often be used in combination with therapeutic techniques, particularly those that aid speech initiation. The substantially higher rate of stuttering occurring on initiation supports a hypothesis that overt stuttering events help 'release' and 'inhibit' central stuttering blocks. This perspective is examined in the context of internal models and mirror neurons.

  9. Using syllable-timed speech to treat preschool children who stutter: a multiple baseline experiment.

    PubMed

    Trajkovski, Natasha; Andrews, Cheryl; Onslow, Mark; Packman, Ann; O'Brian, Sue; Menzies, Ross

    2009-03-01

    This report presents the results of an experimental investigation of the effects of a syllable-timed speech treatment on three stuttering preschool children. Syllable-timed speech involves speaking with minimal differentiation in linguistic stress across syllables. Three children were studied in a multiple baseline across participants design, with percent syllables stuttered (%SS) as the dependent variable. In the week following the initial clinic visit, each child decreased their beyond-clinic stuttering by 40%, 49% and 32%, respectively. These reductions are only evident in the time series after the introduction of the syllable-timed speech treatment procedure. Participants required a mean of six clinic visits, of approximately 30-60 min in duration, to reach and sustain a beyond-clinic %SS below 1.0. The results suggest that clinical trials of the treatment are warranted. The reader will be able to summarize, discuss and evaluate: (1) The nature, impact and treatment options available for early stuttering. (2) The syllable-timed speech treatment protocol administered. (3) The advantages of syllable-timed speech treatment for early stuttering. (4) The questions that further research needs to answer about the syllable-timed speech treatment.

  10. Metrical expectations from preceding prosody influence perception of lexical stress

    PubMed Central

    Brown, Meredith; Salverda, Anne Pier; Dilley, Laura C.; Tanenhaus, Michael K.

    2015-01-01

    Two visual-world experiments tested the hypothesis that expectations based on preceding prosody influence the perception of suprasegmental cues to lexical stress. The results demonstrate that listeners’ consideration of competing alternatives with different stress patterns (e.g., ‘jury/gi’raffe) can be influenced by the fundamental frequency and syllable timing patterns across material preceding a target word. When preceding stressed syllables distal to the target word shared pitch and timing characteristics with the first syllable of the target word, pictures of alternatives with primary lexical stress on the first syllable (e.g., jury) initially attracted more looks than alternatives with unstressed initial syllables (e.g., giraffe). This effect was modulated when preceding unstressed syllables had pitch and timing characteristics similar to the initial syllable of the target word, with more looks to alternatives with unstressed initial syllables (e.g., giraffe) than to those with stressed initial syllables (e.g., jury). These findings suggest that expectations about the acoustic realization of upcoming speech include information about metrical organization and lexical stress, and that these expectations constrain the initial interpretation of suprasegmental stress cues. These distal prosody effects implicate on-line probabilistic inferences about the sources of acoustic-phonetic variation during spoken-word recognition. PMID:25621583

  11. Lingual Kinematics during Rapid Syllable Repetition in Parkinson's Disease

    ERIC Educational Resources Information Center

    Wong, Min Ney; Murdoch, Bruce E.; Whelan, Brooke-Mai

    2012-01-01

    Background: Rapid syllable repetition tasks are commonly used in the assessment of motor speech disorders. However, little is known about the articulatory kinematics during rapid syllable repetition in individuals with Parkinson's disease (PD). Aims: To investigate and compare lingual kinematics during rapid syllable repetition in dysarthric…

  12. Syllable Structure in Arabic Varieties with a Focus on Superheavy Syllables

    ERIC Educational Resources Information Center

    Bamakhramah, Majdi A.

    2010-01-01

    This thesis has two broad goals. The first is to contribute to the study of Arabic phonology particularly syllable structure and syllabification. This will be achieved through examining phenomena related to syllable structure and syllabic weight such as syllabification, stress assignment, epenthesis, syncope, and sonority in three different…

  13. The Effect of Anatomic Factors on Tongue Position Variability during Consonants

    ERIC Educational Resources Information Center

    Rudy, Krista; Yunusova, Yana

    2013-01-01

    Purpose: This study sought to investigate the effect of palate morphology and anthropometric measures of the head on positional variability of the tongue during consonants. Method: An electromagnetic tracking system was used to record tongue movements of 21 adults. Each talker produced a series of symmetrical VCV syllables containing one of the…

  14. Cognitive Conflict and Inhibition in Primed Dichotic Listening

    ERIC Educational Resources Information Center

    Saetrevik, Bjorn; Specht, Karsten

    2009-01-01

    In previous behavioral studies, a prime syllable was presented just prior to a dichotic syllable pair, with instructions to ignore the prime and report one syllable from the dichotic pair. When the prime matched one of the syllables in the dichotic pair, response selection was biased towards selecting the unprimed target. The suggested mechanism…

  15. On the Locus of the Syllable Frequency Effect in Speech Production

    ERIC Educational Resources Information Center

    Laganaro, Marina; Alario, F. -Xavier

    2006-01-01

    The observation of a syllable frequency effect in naming latencies has been an argument in favor of a functional role of stored syllables in speech production. Accordingly, various theoretical models postulate that a repository of syllable representations is accessed during phonetic encoding. However, the direct empirical evidence for locating the…

  16. Temporal order processing of syllables in the left parietal lobe.

    PubMed

    Moser, Dana; Baker, Julie M; Sanchez, Carmen E; Rorden, Chris; Fridriksson, Julius

    2009-10-07

    Speech processing requires the temporal parsing of syllable order. Individuals suffering from posterior left hemisphere brain injury often exhibit temporal processing deficits as well as language deficits. Although the right posterior inferior parietal lobe has been implicated in temporal order judgments (TOJs) of visual information, there is limited evidence to support the role of the left inferior parietal lobe (IPL) in processing syllable order. The purpose of this study was to examine whether the left inferior parietal lobe is recruited during temporal order judgments of speech stimuli. Functional magnetic resonance imaging data were collected on 14 normal participants while they completed the following forced-choice tasks: (1) syllable order of multisyllabic pseudowords, (2) syllable identification of single syllables, and (3) gender identification of both multisyllabic and monosyllabic speech stimuli. Results revealed increased neural recruitment in the left inferior parietal lobe when participants made judgments about syllable order compared with both syllable identification and gender identification. These findings suggest that the left inferior parietal lobe plays an important role in processing syllable order and support the hypothesized role of this region as an interface between auditory speech and the articulatory code. Furthermore, a breakdown in this interface may explain some components of the speech deficits observed after posterior damage to the left hemisphere.

  17. Temporal Order Processing of Syllables in the Left Parietal Lobe

    PubMed Central

    Baker, Julie M.; Sanchez, Carmen E.; Rorden, Chris; Fridriksson, Julius

    2009-01-01

    Speech processing requires the temporal parsing of syllable order. Individuals suffering from posterior left hemisphere brain injury often exhibit temporal processing deficits as well as language deficits. Although the right posterior inferior parietal lobe has been implicated in temporal order judgments (TOJs) of visual information, there is limited evidence to support the role of the left inferior parietal lobe (IPL) in processing syllable order. The purpose of this study was to examine whether the left inferior parietal lobe is recruited during temporal order judgments of speech stimuli. Functional magnetic resonance imaging data were collected on 14 normal participants while they completed the following forced-choice tasks: (1) syllable order of multisyllabic pseudowords, (2) syllable identification of single syllables, and (3) gender identification of both multisyllabic and monosyllabic speech stimuli. Results revealed increased neural recruitment in the left inferior parietal lobe when participants made judgments about syllable order compared with both syllable identification and gender identification. These findings suggest that the left inferior parietal lobe plays an important role in processing syllable order and support the hypothesized role of this region as an interface between auditory speech and the articulatory code. Furthermore, a breakdown in this interface may explain some components of the speech deficits observed after posterior damage to the left hemisphere. PMID:19812331

  18. Metrical expectations from preceding prosody influence perception of lexical stress.

    PubMed

    Brown, Meredith; Salverda, Anne Pier; Dilley, Laura C; Tanenhaus, Michael K

    2015-04-01

    Two visual-world experiments tested the hypothesis that expectations based on preceding prosody influence the perception of suprasegmental cues to lexical stress. The results demonstrate that listeners' consideration of competing alternatives with different stress patterns (e.g., 'jury/gi'raffe) can be influenced by the fundamental frequency and syllable timing patterns across material preceding a target word. When preceding stressed syllables distal to the target word shared pitch and timing characteristics with the first syllable of the target word, pictures of alternatives with primary lexical stress on the first syllable (e.g., jury) initially attracted more looks than alternatives with unstressed initial syllables (e.g., giraffe). This effect was modulated when preceding unstressed syllables had pitch and timing characteristics similar to the initial syllable of the target word, with more looks to alternatives with unstressed initial syllables (e.g., giraffe) than to those with stressed initial syllables (e.g., jury). These findings suggest that expectations about the acoustic realization of upcoming speech include information about metrical organization and lexical stress and that these expectations constrain the initial interpretation of suprasegmental stress cues. These distal prosody effects implicate online probabilistic inferences about the sources of acoustic-phonetic variation during spoken-word recognition. (c) 2015 APA, all rights reserved.

  19. The Basis of the Syllable Hierarchy: Articulatory Pressures or Universal Phonological Constraints?

    PubMed

    Zhao, Xu; Berent, Iris

    2018-02-01

    Across languages, certain syllable types are systematically preferred to others (e.g., [Formula: see text] lbif, where [Formula: see text] indicates a preference). Previous research has shown that these preferences are active in the brains of individual speakers, they are evident even when none of these syllable types exists in participants' language, and even when the stimuli are presented in print. These results suggest that the syllable hierarchy cannot be reduced to either lexical or auditory/phonetic pressures. Here, we examine whether the syllable hierarchy is due to articulatory pressures. According to the motor embodiment view, the perception of a linguistic stimulus requires simulating its production; dispreferred syllables (e.g., lbif) are universally disliked because their production is harder to simulate. To address this possibility, we assessed syllable preferences while articulation was mechanically suppressed. Our four experiments each found significant effects of suppression. Remarkably, people remained sensitive to the syllable hierarchy regardless of suppression. Specifically, results with auditory materials (Experiments 1-2) showed strong effects of syllable structure irrespective of suppression. Moreover, syllable structure uniquely accounted for listeners' behavior even when controlling for several phonetic characteristics of our auditory materials. Results with printed stimuli (Experiments 3-4) were more complex, as participants in these experiments relied on both phonological and graphemic information. Nonetheless, readers were sensitive to most of the syllable hierarchy (e.g., [Formula: see text]), and these preferences emerged when articulation was suppressed, and even when the statistical properties of our materials were controlled via a regression analysis. Together, these findings indicate that speakers possess broad grammatical preferences that are irreducible to either sensory or motor factors.

  20. Support for context effects on segmentation and segments depends on the context.

    PubMed

    Heffner, Christopher C; Newman, Rochelle S; Idsardi, William J

    2017-04-01

    Listeners must adapt to differences in speech rate across talkers and situations. Speech rate adaptation effects are strong for adjacent syllables (i.e., proximal syllables). For studies that have assessed adaptation effects on speech rate information more than one syllable removed from a point of ambiguity in speech (i.e., distal syllables), the difference in strength between different types of ambiguity is stark. Studies of word segmentation have shown large shifts in perception as a result of distal rate manipulations, while studies of segmental perception have shown only weak, or even nonexistent, effects. However, no study has standardized methods and materials to study context effects for both types of ambiguity simultaneously. Here, a set of sentences was created that differed as minimally as possible except for whether the sentences were ambiguous to the voicing of a consonant or ambiguous to the location of a word boundary. The sentences were then rate-modified to slow down the distal context speech rate to various extents, dependent on three different definitions of distal context that were adapted from previous experiments, along with a manipulation of proximal context to assess whether proximal effects were comparable across ambiguity types. The results indicate that the definition of distal influenced the extent of distal rate effects strongly for both segments and segmentation. They also establish the presence of distal rate effects on word-final segments for the first time. These results were replicated, with some caveats regarding the perception of individual segments, in an Internet-based sample recruited from Mechanical Turk.

  1. Two-Month-Old Infants' Sensitivity to Changes in Arbitrary Syllable-Object Pairings: The Role of Temporal Synchrony

    ERIC Educational Resources Information Center

    Gogate, Lakshmi J.; Prince, Christopher G.; Matatyaho, Dalit J.

    2009-01-01

    To explore early lexical development, the authors examined infants' sensitivity to changes in spoken syllables and objects given different temporal relations between syllable-object pairings. In Experiment 1, they habituated 2-month-olds to 1 syllable, /tah/ or /gah/, paired with an object in "synchronous" (utterances coincident with object…

  2. The Effect of the Number of Syllables on Handwriting Production

    ERIC Educational Resources Information Center

    Lambert, Eric; Kandel, Sonia; Fayol, Michel; Esperet, Eric

    2008-01-01

    Four experiments examined whether motor programming in handwriting production can be modulated by the syllable structure of the word to be written. This study manipulated the number of syllables. The items, words and pseudo-words, had 2, 3 or 4 syllables. French adults copied them three times. We measured the latencies between the visual…

  3. Syllable timing and pausing: evidence from Cantonese.

    PubMed

    Perry, Conrad; Wong, Richard Kwok-Shing; Matthews, Stephen

    2009-01-01

    We examined the relationship between the acoustic duration of syllables and the silent pauses that follow them in Cantonese. The results showed that at major syntactic junctures, acoustic plus silent pause durations were quite similar for a number of different syllable types whose acoustic durations differed substantially. In addition, it appeared that CV: syllables, which had the longest acoustic duration of all syllable types that were examined, were also the least likely to have silent pauses after them. These results suggest that cross-language differences between the probability that silent pauses are used at major syntactic junctures might potentially be explained by the accuracy at which timing slots can be assigned for syllables, rather than more complex explanations that have been proposed.

  4. Assessing Plural Morphology in Children Acquiring /S/-Leniting Dialects of Spanish

    ERIC Educational Resources Information Center

    Miller, Karen

    2014-01-01

    Purpose: To examine the production of plural morphology in children acquiring a dialect of Spanish with syllable-final /s/ lenition with the goal of comparing how plural marker omissions in the speech of these children compare with plural marker omissions in children with language impairment acquiring other varieties of Spanish. Method: Three…

  5. Searching for Syllabic Coding Units in Speech Perception

    ERIC Educational Resources Information Center

    Dumay, Nicolas; Content, Alain

    2012-01-01

    Two auditory priming experiments tested whether the effect of final phonological overlap relies on syllabic representations. Amount of shared phonemic information and syllabic status of the overlap between nonword primes and targets were varied orthogonally. In the related conditions, CV.CCVC items shared the last syllable (e.g., vi.klyd-p[image…

  6. Tip-of-the-tongue states reveal age differences in the syllable frequency effect.

    PubMed

    Farrell, Meagan T; Abrams, Lise

    2011-01-01

    Syllable frequency has been shown to facilitate production in some languages but has yielded inconsistent results in English and has never been examined in older adults. Tip-of-the-tongue (TOT) states represent a unique type of production failure where the phonology of a word is unable to be retrieved, suggesting that the frequency of phonological forms, like syllables, may influence the occurrence of TOT states. In the current study, we investigated the role of first-syllable frequency on TOT incidence and resolution in young (18-26 years of age), young-old (60-74 years of age), and old-old (75-89 years of age) adults. Data from 3 published studies were compiled, where TOTs were elicited by presenting definition-like questions and asking participants to respond with "Know," "Don't Know," or "TOT." Young-old and old-old adults, but not young adults, experienced more TOTs for words beginning with low-frequency first syllables relative to high-frequency first syllables. Furthermore, age differences in TOT incidence occurred only for words with low-frequency first syllables. In contrast, when a prime word with the same first syllable as the target was presented during TOT states, all age groups resolved more TOTs for words beginning with low-frequency syllables. These findings support speech production models that allow for bidirectional activation between conceptual, lexical, and phonological forms of words. Furthermore, the age-specific effects of syllable frequency provide insight into the progression of age-linked changes to phonological processes. (PsycINFO Database Record (c) 2010 APA, all rights reserved).

  7. Learning of Syllable-Object Relations by Preverbal Infants: The Role of Temporal Synchrony and Syllable Distinctiveness

    ERIC Educational Resources Information Center

    Gogate, Lakshmi J.

    2010-01-01

    The role of temporal synchrony and syllable distinctiveness in preverbal infants' learning of word-object relations was investigated. In Experiment 1, 7- and 8-month-olds (N=64) were habituated under conditions where two "similar-sounding" syllables, /tah/ and /gah/, were spoken simultaneously with the motions of one of two sets of…

  8. A Prominence Account of Syllable Reduction in Early Speech Development: The Child's Prosodic Phonology of "Tiger" and "Giraffe."

    ERIC Educational Resources Information Center

    Snow, David

    1998-01-01

    This paper tested a theory of syllable prominence with 11 children (ages 11 to 26 months). The theory proposes that syllable prominence is a product of two orthogonal suprasegmental systems: stress/accent peaks and phrase boundaries. Use of the developed prominence scale found it parsimoniously accounted for observed biases in syllable omissions…

  9. The Ortho-Syllable as a Processing Unit in Handwriting: The Mute E Effect

    ERIC Educational Resources Information Center

    Lambert, Eric; Sausset, Solen; Rigalleau, François

    2015-01-01

    Some research on written production has focused on the role of the syllable as a processing unit. However, the precise nature of this syllable unit has yet to be elucidated. The present study examined whether the nature of this processing unit is orthographic (i.e., an ortho-syllable) or phonological. We asked French adults to copy three-syllable…

  10. Tone classification of syllable-segmented Thai speech based on multilayer perception

    NASA Astrophysics Data System (ADS)

    Satravaha, Nuttavudh; Klinkhachorn, Powsiri; Lass, Norman

    2002-05-01

    Thai is a monosyllabic tonal language that uses tone to convey lexical information about the meaning of a syllable. Thus to completely recognize a spoken Thai syllable, a speech recognition system not only has to recognize a base syllable but also must correctly identify a tone. Hence, tone classification of Thai speech is an essential part of a Thai speech recognition system. Thai has five distinctive tones (``mid,'' ``low,'' ``falling,'' ``high,'' and ``rising'') and each tone is represented by a single fundamental frequency (F0) pattern. However, several factors, including tonal coarticulation, stress, intonation, and speaker variability, affect the F0 pattern of a syllable in continuous Thai speech. In this study, an efficient method for tone classification of syllable-segmented Thai speech, which incorporates the effects of tonal coarticulation, stress, and intonation, as well as a method to perform automatic syllable segmentation, were developed. Acoustic parameters were used as the main discriminating parameters. The F0 contour of a segmented syllable was normalized by using a z-score transformation before being presented to a tone classifier. The proposed system was evaluated on 920 test utterances spoken by 8 speakers. A recognition rate of 91.36% was achieved by the proposed system.

  11. Syllable-related breathing in infants in the second year of life.

    PubMed

    Parham, Douglas F; Buder, Eugene H; Oller, D Kimbrough; Boliek, Carol A

    2011-08-01

    This study explored whether breathing behaviors of infants within the 2nd year of life differ between tidal breathing and breathing supporting single unarticulated syllables and canonical/articulated syllables. Vocalizations and breathing kinematics of 9 infants between 53 and 90 weeks of age were recorded. A strict selection protocol was used to identify analyzable breath cycles. Syllables were categorized on the basis of consensus coding. Inspiratory and expiratory durations, excursions, and slopes were calculated for the 3 breath cycle types and were normalized using mean tidal breath measures. Tidal breathing cycles were significantly different from syllable-related cycles on all breathing measures. There were no significant differences between unarticulated syllable cycles and canonical syllable cycles, even after controlling for utterance duration and sound pressure level. Infants in the 2nd year of life exhibit clear differences between tidal breathing and speech-related breathing, but categorically distinct breath support for syllable types with varying articulatory demands was not evident in the present findings. Speech development introduces increasingly complex utterances, so older infants may produce detectable articulation-related adaptations of breathing kinematics. For younger infants, breath support may vary systematically among utterance types, due more to phonatory variations than to articulatory demands.

  12. Syllable-Related Breathing in Infants in the Second Year of Life

    PubMed Central

    Parham, Douglas F.; Buder, Eugene H.; Oller, D. Kimbrough; Boliek, Carol A.

    2010-01-01

    Purpose This study explored whether breathing behaviors of infants within the second year of life differ between tidal breathing and breathing supporting single unarticulated syllables and canonical/articulated syllables. Method Vocalizations and breathing kinematics of nine infants between 53 and 90 weeks of age were recorded. A strict selection protocol was used to identify analyzable breath cycles. Syllables were categorized based on consensus coding. Inspiratory and expiratory durations, excursions, and slopes were calculated for the three breath cycle types and normalized using mean tidal breath measures. Results Tidal breathing cycles were significantly different from syllable-related cycles on all breathing measures. There were no significant differences between unarticulated syllable cycles and canonical syllable cycles, even after controlling for utterance duration and sound pressure level. Conclusions Infants in the second year of life exhibit clear differences between tidal breathing and speech-related breathing, but categorically distinct breath support for syllable types with varying articulatory demands was not evident in the current findings. Speech development introduces increasingly complex utterances, so older infants may produce detectable articulation-related adaptations of breathing kinematics. For younger infants, breath support may vary systematically among utterance types, due more to phonatory variations than to articulatory demands. PMID:21173390

  13. Investigating the Phonological Similarity Effect: Syllable Structure and the Position of Common Phonemes

    ERIC Educational Resources Information Center

    Nimmo, Lisa M.; Roodenrys, Steven

    2004-01-01

    The aim of the present research was to determine whether the effect that phonological similarity has on immediate serial recall is influenced by the consistency and position of phonemes within words. In comparison to phonologically dissimilar lists, when the stimulus lists rhyme there is a facilitative effect on the recall of item information and…

  14. Syllable acoustics, temporal patterns, and call composition vary with behavioral context in Mexican free-tailed bats

    PubMed Central

    Bohn, Kirsten M.; Schmidt-French, Barbara; Ma, Sean T.; Pollak, George D.

    2008-01-01

    Recent research has shown that some bat species have rich vocal repertoires with diverse syllable acoustics. Few studies, however, have compared vocalizations across different behavioral contexts or examined the temporal emission patterns of vocalizations. In this paper, a comprehensive examination of the vocal repertoire of Mexican free-tailed bats, T. brasiliensis, is presented. Syllable acoustics and temporal emission patterns for 16 types of vocalizations including courtship song revealed three main findings. First, although in some cases syllables are unique to specific calls, other syllables are shared among different calls. Second, entire calls associated with one behavior can be embedded into more complex vocalizations used in entirely different behavioral contexts. Third, when different calls are composed of similar syllables, distinctive temporal emission patterns may facilitate call recognition. These results indicate that syllable acoustics alone do not likely provide enough information for call recognition; rather, the acoustic context and temporal emission patterns of vocalizations may affect meaning. PMID:19045674

  15. Words, rules, and mechanisms of language acquisition.

    PubMed

    Endress, Ansgar D; Bonatti, Luca L

    2016-01-01

    We review recent artificial language learning studies, especially those following Endress and Bonatti (Endress AD, Bonatti LL. Rapid learning of syllable classes from a perceptually continuous speech stream. Cognition 2007, 105:247-299), suggesting that humans can deploy a variety of learning mechanisms to acquire artificial languages. Several experiments provide evidence for multiple learning mechanisms that can be deployed in fluent speech: one mechanism encodes the positions of syllables within words and can be used to extract generalization, while the other registers co-occurrence statistics of syllables and can be used to break a continuum into its components. We review dissociations between these mechanisms and their potential role in language acquisition. We then turn to recent criticisms of the multiple mechanisms hypothesis and show that they are inconsistent with the available data. Our results suggest that artificial and natural language learning is best understood by dissecting the underlying specialized learning abilities, and that these data provide a rare opportunity to link important language phenomena to basic psychological mechanisms. For further resources related to this article, please visit the WIREs website. © 2015 Wiley Periodicals, Inc.

  16. Beyond Single Syllables: The Effect of First Syllable Frequency and Orthographic Similarity on Eye Movements during Silent Reading

    ERIC Educational Resources Information Center

    Hawelka, Stefan; Schuster, Sarah; Gagl, Benjamin; Hutzler, Florian

    2013-01-01

    The study assessed the eye movements of 60 adult German readers during silent reading of target words, consisting of two and three syllables, embedded in sentences. The first objective was to assess whether the inhibitory effect of first syllable frequency, which was up to now primarily shown for isolated words, generalises to natural reading. The…

  17. Generating Enthusiasm with Generative Phonology.

    ERIC Educational Resources Information Center

    Dickerson, Wayne B.

    This paper attempts a systematic approach to the teaching of word stress in the ESL classroom. Stress assignment rules from Chomsky and Halle and from Ross are used to establish the SISL Principle (Stress Initial Strong Left), for final weak-syllable words. On the basis of spelling, this rule can be applied correctly to 95 out of 100 cases. (AM)

  18. Stress Regularity or Consistency? Reading Aloud Italian Polysyllables with Different Stress Patterns

    ERIC Educational Resources Information Center

    Burani, Cristina; Arduino, Lisa S.

    2004-01-01

    Stress assignment to three- and four-syllable Italian words is not predictable by rule, but needs lexical look-up. The present study investigated whether stress assignment to low-frequency Italian words is determined by stress regularity, or by the number of words sharing the final phonological segment and the stress pattern (stress neighborhood…

  19. Twenty-Four-Month-Olds' Perception of Word-Medial Onsets and Codas

    ERIC Educational Resources Information Center

    Wang, Yuanyuan; Seidl, Amanda

    2016-01-01

    Recent work has shown that children have detailed phonological representations of consonants at both word-initial and word-final edges. Nonetheless, it remains unclear whether onsets and codas are equally represented by young learners since word edges are isomorphic with syllable edges in this work. The current study sought to explore toddler's…

  20. Variable Input and the Acquisition of Plural Morphology

    ERIC Educational Resources Information Center

    Miller, Karen L.; Schmitt, Cristina

    2012-01-01

    The present article examines the effect of variable input on the acquisition of plural morphology in two varieties of Spanish: Chilean Spanish, where the plural marker is sometimes omitted due to a phonological process of syllable final /s/ lenition, and Mexican Spanish (of Mexico City), with no such lenition process. The goal of the study is to…

  1. The right ear advantage revisited: speech lateralisation in dichotic listening using consonant-vowel and vowel-consonant syllables.

    PubMed

    Sætrevik, Bjørn

    2012-01-01

    The dichotic listening task is typically administered by presenting a consonant-vowel (CV) syllable to each ear and asking the participant to report the syllable heard most clearly. The results tend to show more reports of the right ear syllable than of the left ear syllable, an effect called the right ear advantage (REA). The REA is assumed to be due to the crossing over of auditory fibres and the processing of language stimuli being lateralised to left temporal areas. However, the tendency for most dichotic listening experiments to use only CV syllable stimuli limits the extent to which the conclusions can be generalised to also apply to other speech phonemes. The current study re-examines the REA in dichotic listening by using both CV and vowel-consonant (VC) syllables and combinations thereof. Results showed a replication of the REA response pattern for both CV and VC syllables, thus indicating that the general assumption of left-side localisation of processing can be applied for both types of stimuli. Further, on trials where a CV is presented in one ear and a VC is presented in the other ear, the CV is selected more often than the VC, indicating that these phonemes have an acoustic or processing advantage.

  2. A comparison of aphasic and non-brain-injured adults on a dichotic CV-syllable listening task.

    PubMed

    Shanks, J; Ryan, W

    1976-06-01

    A dichotic CV-syllable listening task was administered to a group of eleven non-brain-injured adults and to a group of eleven adult aphasics. The results of this study may be summarized as follows: 1)The group of non-brain-injured adults showed a slight right ear advantage for dichotically presented CV-syllables. 2)In comparison with the control group the asphasic group showed a bilateral deficit in response to the dichotic CV-syllables, superimposed on a non-significant right ear advantage. 3) The asphasic group demonstrated a great deal of intersubject variability on the dichotic task with six aphasics showing a right ear preference for the stimuli. The non-brain-injured subjects performed more homogeneously on the task. 4) The two subgroups of aphasics, a right ear advantage group and a left ear advantage group, performed significantly different on the dichotic listening task. 5) Single correct data analysis proved valuable by deleting accuracy of report for an examination of trials in which there was true competition for the single left hemispheric speech processor. These results were analyzed in terms of a functional model of auditory processing. In view of this model, the bilateral deficit in dichotic performance of the asphasic group was accounted for by the presence of a lesion within the dominant left hemisphere, where the speech signals from both ears converge for final processing. The right ear advantage shown by one asphasic subgroup was explained by a lesion interfering with the corpus callosal pathways from the left hemisphere; the left ear advantage observed within the other subgroup was explained by a lesion in the area of the auditory processor of the left hemisphere.

  3. Phonological Similarity in American Sign Language.

    ERIC Educational Resources Information Center

    Hildebrandt, Ursula; Corina, David

    2002-01-01

    Investigates deaf and hearing subjects' ratings of American Sign Language (ASL) signs to assess whether linguistic experience shapes judgments of sign similarity. Findings are consistent with linguistic theories that posit movement and location as core structural elements of syllable structure in ASL. (Author/VWL)

  4. Growth and splitting of neural sequences in songbird vocal development

    PubMed Central

    Okubo, Tatsuo S.; Mackevicius, Emily L.; Payne, Hannah L.; Lynch, Galen F.; Fee, Michale S.

    2015-01-01

    Neural sequences are a fundamental feature of brain dynamics underlying diverse behaviors, but the mechanisms by which they develop during learning remain unknown. Songbirds learn vocalizations composed of syllables; in adult birds, each syllable is produced by a different sequence of action potential bursts in the premotor cortical area HVC. Here we carried out recordings of large populations of HVC neurons in singing juvenile birds throughout learning to examine the emergence of neural sequences. Early in vocal development, HVC neurons begin producing rhythmic bursts, temporally locked to a ‘prototype’ syllable. Different neurons are active at different latencies relative to syllable onset to form a continuous sequence. Through development, as new syllables emerge from the prototype syllable, initially highly overlapping burst sequences become increasingly distinct. We propose a mechanistic model in which multiple neural sequences can emerge from the growth and splitting of a common precursor sequence. PMID:26618871

  5. The influence of phonological context on the sound errors of a speaker with Wernicke's aphasia.

    PubMed

    Goldmann, R E; Schwartz, M F; Wilshire, C E

    2001-09-01

    A corpus of phonological errors produced in narrative speech by a Wernicke's aphasic speaker (R.W.B.) was tested for context effects using two new methods for establishing chance baselines. A reliable anticipatory effect was found using the second method, which estimated chance from the distance between phoneme repeats in the speech sample containing the errors. Relative to this baseline, error-source distances were shorter than expected for anticipations, but not perseverations. R.W.B.'s anticipation/perseveration ratio measured intermediate between a nonaphasic error corpus and that of a more severe aphasic speaker (both reported in Schwartz et al., 1994), supporting the view that the anticipatory bias correlates to severity. Finally, R.W.B's anticipations favored word-initial segments, although errors and sources did not consistently share word or syllable position. Copyright 2001 Academic Press.

  6. The Effect of Uni- and Bilateral Thalamic Deep Brain Stimulation on Speech in Patients With Essential Tremor: Acoustics and Intelligibility.

    PubMed

    Becker, Johannes; Barbe, Michael T; Hartinger, Mariam; Dembek, Till A; Pochmann, Jil; Wirths, Jochen; Allert, Niels; Mücke, Doris; Hermes, Anne; Meister, Ingo G; Visser-Vandewalle, Veerle; Grice, Martine; Timmermann, Lars

    2017-04-01

    Deep brain stimulation (DBS) of the ventral intermediate nucleus (VIM) is performed to suppress medically-resistant essential tremor (ET). However, stimulation induced dysarthria (SID) is a common side effect, limiting the extent to which tremor can be suppressed. To date, the exact pathogenesis of SID in VIM-DBS treated ET patients is unknown. We investigate the effect of inactivated, uni- and bilateral VIM-DBS on speech production in patients with ET. We employ acoustic measures, tempo, and intelligibility ratings and patient's self-estimated speech to quantify SID, with a focus on comparing bilateral to unilateral stimulation effects and the effect of electrode position on speech. Sixteen German ET patients participated in this study. Each patient was acoustically recorded with DBS-off, unilateral-right-hemispheric-DBS-on, unilateral-left-hemispheric-DBS-on, and bilateral-DBS-on during an oral diadochokinesis task and a read German standard text. To capture the extent of speech impairment, we measured syllable duration and intensity ratio during the DDK task. Naïve listeners rated speech tempo and speech intelligibility of the read text on a 5-point-scale. Patients had to rate their "ability to speak". We found an effect of bilateral compared to unilateral and inactivated stimulation on syllable durations and intensity ratio, as well as on external intelligibility ratings and patients' VAS scores. Additionally, VAS scores are associated with more laterally located active contacts. For speech ratings, we found an effect of syllable duration such that tempo and intelligibility was rated worse for speakers exhibiting greater syllable durations. Our data confirms that SID is more pronounced under bilateral compared to unilateral stimulation. Laterally located electrodes are associated with more severe SID according to patient's self-ratings. We can confirm the relation between diadochokinetic rate and SID in that listener's tempo and intelligibility ratings can be predicted by measured syllable durations from DDK tasks. © 2017 International Neuromodulation Society.

  7. The relative roles of cultural drift and acoustic adaptation in shaping syllable repertoires of island bird populations change with time since colonization.

    PubMed

    Potvin, Dominique A; Clegg, Sonya M

    2015-02-01

    In birds, song divergence often precedes and facilitates divergence of other traits. We assessed the relative roles of cultural drift, innovation, and acoustic adaptation in divergence of island bird dialects, using silvereyes (Zosterops lateralis). In recently colonized populations, syllable diversity was not significantly lower than source populations, shared syllables between populations decreased with increasing number of founder events, and dialect variation displayed contributions from both habitat features and drift. The breadth of multivariate space occupied by recently colonized Z. l. lateralis populations was comparable to evolutionarily old forms that have diverged over thousands to hundreds of thousands of years. In evolutionarily old subspecies, syllable diversity was comparable to the mainland and the amount of variation in syllable composition explained by habitat features increased by two- to threefold compared to recently colonized populations. Together these results suggest that cultural drift influences syllable repertoires in recently colonized populations, but innovation likely counters syllable loss from colonization. In evolutionarily older populations, the influence of acoustic adaptation increases, possibly favoring a high diversity of syllables. These results suggest that the relative importance of cultural drift and acoustic adaptation changes with time since colonization in island bird populations, highlighting the value of considering multiple mechanisms and timescale of divergence when investigating island song divergence. © 2014 The Author(s). Evolution © 2014 The Society for the Study of Evolution.

  8. The transition from the core vowels to the following segments in Japanese children who stutter: the second, third and fourth syllables.

    PubMed

    Matsumoto-Shimamori, Sachiyo; Ito, Tomohiko; Fukuda, Suzy E; Fukuda, Shinji

    2011-09-01

    Shimamori and Ito (2007 , Syllable weight and phonological encoding in Japanese children who stutter. Japanese Journal of Special Education, 44, 451-462; 2008, Syllable weight and frequency of stuttering: Comparison between children who stutter with and without a family history of stuttering. Japanese Journal of Special Education, 45, 437-445; 2009, Difference in frequency of stuttering between light and heavy syllables in the production of monosyllables: From the viewpoint of phonetic transition. The Japanese Journal of Logopedics and Phoniatrics, 50, 116-122 (in Japanese)) proposed the hypothesis that in Japanese the transition from the core vowels (CVs) to the following segments affected the occurrence of stuttering. However, the transition we investigated was in the first syllables only, and the effect of the transition in second, third and fourth syllables was not addressed. The purpose of this study was to investigate whether the transition from the CVs in the second, third and fourth syllables affected the occurrence of stuttering. The participants were 21 Japanese children. A non-word naming task and a non-word reading task were used. The frequency of stuttering was not significantly different where the number of transitions from the CVs differed on either task. These results suggest that the transition from the CVs in the second, third and fourth syllables does not have a significant effect on the occurrence of stuttering in Japanese.

  9. Text-to-phonemic transcription and parsing into mono-syllables of English text

    NASA Astrophysics Data System (ADS)

    Jusgir Mullick, Yugal; Agrawal, S. S.; Tayal, Smita; Goswami, Manisha

    2004-05-01

    The present paper describes a program that converts the English text (entered through the normal computer keyboard) into its phonemic representation and then parses it into mono-syllables. For every letter a set of context based rules is defined in lexical order. A default rule is also defined separately for each letter. Beginning from the first letter of the word the rules are checked and the most appropriate rule is applied on the letter to find its actual orthographic representation. If no matching rule is found, then the default rule is applied. Current rule sets the next position to be analyzed. Proceeding in the same manner orthographic representation for each word can be found. For example, ``reading'' is represented as ``rEdiNX'' by applying the following rules: r-->r move 1 position ahead ead-->Ed move 3 position ahead i-->i move 1 position ahead ng-->NX move 2 position ahead, i.e., end of word. The phonemic representations obtained from the above procedure are parsed to get mono-syllabic representation for various combinations such as CVC, CVCC, CV, CVCVC, etc. For example, the above phonemic representation will be parsed as rEdiNX---> /rE/ /diNX/. This study is a part of developing TTS for Indian English.

  10. Contextual Variability and Exemplar Strength in Phonotactic Learning

    ERIC Educational Resources Information Center

    Denby, Thomas; Schecter, Jeffrey; Arn, Sean; Dimov, Svetlin; Goldrick, Matthew

    2018-01-01

    Phonotactics--constraints on the position and combination of speech sounds within syllables--are subject to statistical differences that gradiently affect speaker and listener behavior (e.g., Vitevitch & Luce, 1999). What statistical properties drive the acquisition of such constraints? Because they are naturally highly correlated, previous…

  11. Contrasting the effects of duration and number of syllables on the perceptual normalization of lexical tones

    NASA Astrophysics Data System (ADS)

    Ciocca, Valter; Francis, Alexander L.; Yau, Teresa S.-K.

    2004-05-01

    In tonal languages, syllabic fundamental frequency (F0) patterns (``lexical tones'') convey lexical meaning. Listeners need to relate such pitch patterns to the pitch range of a speaker (``tone normalization'') to accurately identify lexical tones. This study investigated the amount of tonal information required to perform tone normalization. A target CV syllable, perceived as either a high level, a low level, or a mid level Cantonese tone, was preceded by a four-syllable carrier sentence whose F0 was shifted (1 semitone), or not shifted. Four conditions were obtained by gating one, two, three, or four syllables from the onset of the target. Presentation rate (normal versus fast) was set such that the duration of the one, two, and three syllable conditions (normal carrier) was equal to that of the two, three, and four syllable conditions (fast carrier). Results suggest that tone normalization is largely accomplished within 250 ms or so prior to target onset, independent of the number of syllables; additional tonal information produces a relatively small increase in tone normalization. Implications for models of lexical tone normalization will be discussed. [Work supported by the RGC of the Hong Kong SAR, Project No. HKU 7193/00H.

  12. Universal and Language-Specific Constraints on Phonemic Awareness: Evidence from Russian-Hebrew Bilingual Children

    ERIC Educational Resources Information Center

    Saiegh-Haddad, Elinor; Kogan, Nadya; Walters, Joel

    2010-01-01

    The study tested phonemic awareness in the two languages of Russian (L1)-Hebrew (L2) sequential bilingual children (N = 20) using phoneme deletion tasks where the phoneme to be deleted occurred word initial, word final, as a singleton, or part of a cluster, in long and short words and stressed and unstressed syllables. The experiments were…

  13. Masked syllable priming effects in word and picture naming in Chinese.

    PubMed

    You, Wenping; Zhang, Qingfang; Verdonschot, Rinus G

    2012-01-01

    Four experiments investigated the role of the syllable in Chinese spoken word production. Chen, Chen and Ferrand (2003) reported a syllable priming effect when primes and targets shared the first syllable using a masked priming paradigm in Chinese. Our Experiment 1 was a direct replication of Chen et al.'s (2003) Experiment 3 employing CV (e.g., ,/ba2.ying2/, strike camp) and CVG (e.g., ,/bai2.shou3/, white haired) syllable types. Experiment 2 tested the syllable priming effect using different syllable types: e.g., CV (,/qi4.qiu2/, balloon) and CVN (,/qing1.ting2/, dragonfly). Experiment 3 investigated this issue further using line drawings of common objects as targets that were preceded either by a CV (e.g., ,/qi3/, attempt), or a CVN (e.g., ,/qing2/, affection) prime. Experiment 4 further examined the priming effect by a comparison between CV or CVN priming and an unrelated priming condition using CV-NX (e.g., ,/mi2.ni3/, mini) and CVN-CX (e.g., ,/min2.ju1/, dwellings) as target words. These four experiments consistently found that CV targets were named faster when preceded by CV primes than when they were preceded by CVG, CVN or unrelated primes, whereas CVG or CVN targets showed the reverse pattern. These results indicate that the priming effect critically depends on the match between the structure of the prime and that of the first syllable of the target. The effect obtained in this study was consistent across different stimuli and different tasks (word and picture naming), and provides more conclusive and consistent data regarding the role of the syllable in Chinese speech production.

  14. Rudimentary Reading Repertoires via Stimulus Equivalence and Recombination of Minimal Verbal Units

    PubMed Central

    Matos, Maria Amelia; Avanzi, Alessandra Lopes; McIlvane, William J

    2006-01-01

    We report a study with sixteen low-SES Brazilian children that sought to establish a repertoire of relations involving dictated words, printed words, and corresponding pictures. Children were taught: (1) in response to dictated words, to select corresponding pictures; (2) in response to syllables presented in both visual and auditory formats, to select words which contained a corresponding syllable in either the first or the last position; (3) in response to dictated-word samples, to “construct” corresponding printed words via arranging their constituent syllabic components; and (4) in response to printed word samples, to construct identical printed words by arranging their syllabic constituents. After training on the first two types of tasks, children were given tests for potentially emergent relations involving printed words and pictures. Almost all exhibited relations consistent with stimulus equivalence. They also displayed emergent naming performances––not only with training words but also with new words that were recombinations of their constituent syllables. The present work was inspired by Sidman's stimulus equivalence paradigm and by Skinner's functional analysis of verbal relations, particularly as applied to conceptions of minimal behavioral units and creativity (i.e., behavioral flexibility) in the analytical units applied to verbal relations. PMID:22477340

  15. Deconstructing the Southeast Asian Sesquisyllable: A Gestural Account

    ERIC Educational Resources Information Center

    Butler, Becky Ann

    2014-01-01

    This dissertation explores a purportedly unusual word type known as the "sesquisyllable," which has long been considered characteristic of mainland Southeast Asian languages. Sesquisyllables are traditionally defined as "one and a half" syllables, or as one "major" syllable preceded by one "minor" syllable,…

  16. Serial Position Effects in Nonword Repetition

    ERIC Educational Resources Information Center

    Gupta, P.; Lipinski, J.; Abbs, B.; Lin, P.H.

    2005-01-01

    A growing body of research has emphasized the linkage between performance in immediate serial recall of lists, nonword repetition, and word learning. Recently, it has been reported that primacy and recency effects are obtained in repetition of individual syllables within nonwords (Gupta, in press). Five experiments examined whether such…

  17. Flexibility of orthographic and graphomotor coordination during a handwritten copy task: effect of time pressure

    PubMed Central

    Sausset, Solen; Lambert, Eric; Olive, Thierry

    2013-01-01

    The coordination of the various processes involved in language production is a subject of keen debate in writing research. Some authors hold that writing processes can be flexibly coordinated according to task demands, whereas others claim that process coordination is entirely inflexible. For instance, orthographic planning has been shown to be resource-dependent during handwriting, but inflexible in typing, even under time pressure. The present study therefore went one step further in studying flexibility in the coordination of orthographic processing and graphomotor execution, by measuring the impact of time pressure during a handwritten copy task. Orthographic and graphomotor processes were observed via syllable processing. Writers copied out two- and three-syllable words three times in a row, with and without time pressure. Latencies and letter measures at syllable boundaries were analyzed. We hypothesized that if coordination is flexible and varies according to task demands, it should be modified by time pressure, affecting both latency before execution and duration of execution. We therefore predicted that the extent of syllable processing before execution would be reduced under time pressure and, as a consequence, syllable effects during execution would be more salient. Results showed, however, that time pressure interacted neither with syllable number nor with syllable structure. Accordingly, syllable processing appears to remain the same regardless of time pressure. The flexibility of process coordination during handwriting is discussed, as is the operationalization of time pressure constraints. PMID:24319435

  18. The Basis of the Syllable Hierarchy: Articulatory Pressures or Universal Phonological Constraints?

    ERIC Educational Resources Information Center

    Zhao, Xu; Berent, Iris

    2018-01-01

    Across languages, certain syllable types are systematically preferred to others (e.g., "blif" ? "bnif" ? "bdif" ? "lbif" where ? indicates a preference). Previous research has shown that these preferences are active in the brains of individual speakers, they are evident even when none of these syllable types…

  19. Two genetic loci control syllable sequences of ultrasonic courtship vocalizations in inbred mice

    PubMed Central

    2011-01-01

    Background The ultrasonic vocalizations (USV) of courting male mice are known to possess a phonetic structure with a complex combination of several syllables. The genetic mechanisms underlying the syllable sequence organization were investigated. Results This study compared syllable sequence organization in two inbred strains of mice, 129S4/SvJae (129) and C57BL6J (B6), and demonstrated that they possessed two mutually exclusive phenotypes. The 129S4/SvJae (129) strain frequently exhibited a "chevron-wave" USV pattern, which was characterized by the repetition of chevron-type syllables. The C57BL/6J strain produced a "staccato" USV pattern, which was characterized by the repetition of short-type syllables. An F1 strain obtained by crossing the 129S4/SvJae and C57BL/6J strains produced only the staccato phenotype. The chevron-wave and staccato phenotypes reappeared in the F2 generations, following the Mendelian law of independent assortment. Conclusions These results suggest that two genetic loci control the organization of syllable sequences. These loci were occupied by the staccato and chevron-wave alleles in the B6 and 129 mouse strains, respectively. Recombination of these alleles might lead to the diversity of USV patterns produced by mice. PMID:22018021

  20. Learning novel words: detail and vulnerability of initial representations for children with specific language impairment and typically developing peers.

    PubMed

    Alt, Mary; Suddarth, Rachael

    2012-01-01

    This study examines the phonological representations that children with specific language impairment (SLI) and typically developing peers (TD) have during the initial process of word learning. The goals of this study were to determine if children with SLI attended to different components of words than peers, and whether they were more vulnerable to interference than peers. Forty 7- and 8-year-old children, half with SLI, took part in a fast mapping, word learning task. In addition to producing the word, there was a mispronunciation detection task that included mispronunciations of the target word in the initial position, final position or that modified the word's syllable structure. Children with SLI showed a different learning profile than peers, demonstrating stronger representations of the word-initial phonemes, but less information about word-final phonemes. They were more prone to interference overall, but especially from word-final foils. Children with SLI did not demonstrate less-defined phonological representations, but did attend to different features than TD children, perhaps in an attempt to compensate for problems learning longer words. The greatest weakness of children with SLI appears to be their susceptibility to interference, particularly for word-final information. Readers will be able to: (1) explain what children attend to when learning new words; (2) state the pattern of recognition and production performance for both children with SLI and their typical language peers; and (3) identify specific parts of novel words that are most susceptible to interference in children with SLI. © 2011 Elsevier Inc. All rights reserved.

  1. No, There Is No 150 ms Lead of Visual Speech on Auditory Speech, but a Range of Audiovisual Asynchronies Varying from Small Audio Lead to Large Audio Lag

    PubMed Central

    Schwartz, Jean-Luc; Savariaux, Christophe

    2014-01-01

    An increasing number of neuroscience papers capitalize on the assumption published in this journal that visual speech would be typically 150 ms ahead of auditory speech. It happens that the estimation of audiovisual asynchrony in the reference paper is valid only in very specific cases, for isolated consonant-vowel syllables or at the beginning of a speech utterance, in what we call “preparatory gestures”. However, when syllables are chained in sequences, as they are typically in most parts of a natural speech utterance, asynchrony should be defined in a different way. This is what we call “comodulatory gestures” providing auditory and visual events more or less in synchrony. We provide audiovisual data on sequences of plosive-vowel syllables (pa, ta, ka, ba, da, ga, ma, na) showing that audiovisual synchrony is actually rather precise, varying between 20 ms audio lead and 70 ms audio lag. We show how more complex speech material should result in a range typically varying between 40 ms audio lead and 200 ms audio lag, and we discuss how this natural coordination is reflected in the so-called temporal integration window for audiovisual speech perception. Finally we present a toy model of auditory and audiovisual predictive coding, showing that visual lead is actually not necessary for visual prediction. PMID:25079216

  2. Automatic Syllabification in English: A Comparison of Different Algorithms

    ERIC Educational Resources Information Center

    Marchand, Yannick; Adsett, Connie R.; Damper, Robert I.

    2009-01-01

    Automatic syllabification of words is challenging, not least because the syllable is not easy to define precisely. Consequently, no accepted standard algorithm for automatic syllabification exists. There are two broad approaches: rule-based and data-driven. The rule-based method effectively embodies some theoretical position regarding the…

  3. Work Papers of the Summer Institute of Linguistics, University of North Dakota Session. Volume 41.

    ERIC Educational Resources Information Center

    Marlett, Stephen A., Ed.

    The collection of seven working papers in linguistics includes: "Proceeding from Syllable Inventory to Phonemic Inventory in the Analysis of Liangshan Yi (Andy Eatough); "Empty Consonants in Root-Medial Position" (Stephen A. Marlett); "What Is Literature? A Definition Based on Prototypes" (Jim Meyer); "Seri…

  4. Onset of Duplicated Syllable Babbling and Unimanual Handedness in Infancy: Evidence for Developmental Change in Hemispheric Specialization?

    ERIC Educational Resources Information Center

    Ramsay, Douglas S.

    1984-01-01

    Examines the possible developmental relationship between unimanual handedness and duplicated syllable babbling. Thirty infants were tested at weekly intervals between five months of age and eight weeks after the onset of duplicated syllable babbling. Results suggest developmental change in hemispheric specialization or at least asymmetrical…

  5. Syllable-Related Breathing in Infants in the Second Year of Life

    ERIC Educational Resources Information Center

    Parham, Douglas F.; Buder, Eugene H.; Oller, D. Kimbrough; Boliek, Carol A.

    2011-01-01

    Purpose: This study explored whether breathing behaviors of infants within the 2nd year of life differ between tidal breathing and breathing supporting single unarticulated syllables and canonical/articulated syllables. Method: Vocalizations and breathing kinematics of 9 infants between 53 and 90 weeks of age were recorded. A strict selection…

  6. Intersensory Redundancy and Seven-Month-Old Infants' Memory for Arbitrary Syllable-Object Relations.

    ERIC Educational Resources Information Center

    Gogate, Lakshmi J.; Bahrick, Lorraine E.

    Seven-month-old infants require redundant information such as temporal synchrony to learn arbitrary syllable-object relations. Infants learned the relations between spoken syllables, /a/ and /i/, and two moving objects only when temporal synchrony was present during habituation. Two experiments examined infants' memory for these relations. In…

  7. Clinical Application of the Mean Babbling Level and Syllable Structure Level

    ERIC Educational Resources Information Center

    Morris, Sherrill R.

    2010-01-01

    Purpose: This clinical exchange reviews two independent phonological assessment measures: mean babbling level (MBL) and syllable structure level (SSL). Both measures summarize phonetic inventory and syllable shape in a calculated average and have been used in research to describe the phonological abilities of children ages 9 to 36 months. An…

  8. Rise Time Perception and Detection of Syllable Stress in Adults with Developmental Dyslexia

    ERIC Educational Resources Information Center

    Leong, Victoria; Hamalainen, Jarmo; Soltesz, Fruzsina; Goswami, Usha

    2011-01-01

    Introduction: The perception of syllable stress has not been widely studied in developmental dyslexia, despite strong evidence for auditory rhythmic perceptual difficulties. Here we investigate the hypothesis that perception of sound rise time is related to the perception of syllable stress in adults with developmental dyslexia. Methods: A…

  9. Lexical Stress and Reading: A Study with Children

    ERIC Educational Resources Information Center

    Palma, Nicolas Gutierrez; Reyes, Alfonso Palma

    2004-01-01

    Introduction: Stress in Spanish is associated with an orthographic mark that indicates stress, but there are also other clues that point to it. Most words have the same stress (on the penultimate syllable), and closed syllables (syllables ending in a consonant) attract the stress. In this paper we study these clues, and consequently the function…

  10. Spatial and Temporal Properties of Gestures in North American English /r/

    ERIC Educational Resources Information Center

    Campbell, Fiona; Gick, Bryan; Wilson, Ian; Vatikiotis-Bateson, Eric

    2010-01-01

    Systematic syllable-based variation has been observed in the relative spatial and temporal properties of supralaryngeal gestures in a number of complex segments. Generally, more anterior gestures tend to appear at syllable peripheries while less anterior gestures occur closer to syllable peaks. Because previous studies compared only two gestures,…

  11. The Influence of Lexical Status and Neighborhood Density on Children's Nonword Repetition

    ERIC Educational Resources Information Center

    Metsala, Jamie L.; Chisholm, Gina M.

    2010-01-01

    This study examined effects of lexical status and neighborhood density of constituent syllables on children's nonword repetition and interactions with nonword length. Lexical status of the target syllable impacted repetition accuracy for the longest nonwords. In addition, children made more errors that changed a nonword syllable to a word syllable…

  12. Syllables and bigrams: orthographic redundancy and syllabic units affect visual word recognition at different processing levels.

    PubMed

    Conrad, Markus; Carreiras, Manuel; Tamm, Sascha; Jacobs, Arthur M

    2009-04-01

    Over the last decade, there has been increasing evidence for syllabic processing during visual word recognition. If syllabic effects prove to be independent from orthographic redundancy, this would seriously challenge the ability of current computational models to account for the processing of polysyllabic words. Three experiments are presented to disentangle effects of the frequency of syllabic units and orthographic segments in lexical decision. In Experiment 1 the authors obtained an inhibitory syllable frequency effect that was unaffected by the presence or absence of a bigram trough at the syllable boundary. In Experiments 2 and 3 an inhibitory effect of initial syllable frequency but a facilitative effect of initial bigram frequency emerged when manipulating 1 of the 2 measures and controlling for the other in Spanish words starting with consonant-vowel syllables. The authors conclude that effects of syllable frequency and letter-cluster frequency are independent and arise at different processing levels of visual word recognition. Results are discussed within the framework of an interactive activation model of visual word recognition. (c) 2009 APA, all rights reserved.

  13. Lexical tone and stuttering loci in Mandarin: evidence from preschool children who stutter.

    PubMed

    Chou, Fang-Chi; Zebrowski, Patricia; Yang, Shu-Lan

    2015-02-01

    The purpose of this study was to examine the relationship between stuttering loci and lexical tone in Mandarin-speaking preschoolers. Conversational samples from 20 Taiwanese children who stutter (CWS; M = 4:9; range = 3:2-6:4) were analysed for frequency and type of speech disfluency and lexical tone associated with stuttering-like disfluencies (SLDs). Results indicated that SLDs were significantly more likely to be produced on Mandarin syllables carrying Tone 3 and Tone 4 syllables compared to syllables carrying either Tone 1 or Tone 2. Post-hoc analyses revealed: (1) no significant differences in the stuttering frequencies between Tone 1 and Tone 2, or between Tone 3 and Tone 4, and (2) a higher incidence of stuttering on syllables carrying Tone 3 and Tone 4 embedded in conflicting (as opposed to compatible) tonal contexts. Results suggest that the higher incidence of stuttering on Mandarin syllables carrying either Tone 3 or 4 may be attributed to the increased level of speech motor demand underlying rapid F0 change both within and across syllables.

  14. The Role of Secondary-Stressed and Unstressed-Unreduced Syllables in Word Recognition: Acoustic and Perceptual Studies with Russian Learners of English.

    PubMed

    Banzina, Elina; Dilley, Laura C; Hewitt, Lynne E

    2016-08-01

    The importance of secondary-stressed (SS) and unstressed-unreduced (UU) syllable accuracy for spoken word recognition in English is as yet unclear. An acoustic study first investigated Russian learners' of English production of SS and UU syllables. Significant vowel quality and duration reductions in Russian-spoken SS and UU vowels were found, likely due to a transfer of native phonological features. Next, a cross-modal phonological priming technique combined with a lexical decision task assessed the effect of inaccurate SS and UU syllable productions on native American English listeners' speech processing. Inaccurate UU vowels led to significant inhibition of lexical access, while reduced SS vowels revealed less interference. The results have implications for understanding the role of SS and UU syllables for word recognition and English pronunciation instruction.

  15. Assessment of rhythmic entrainment at multiple timescales in dyslexia: evidence for disruption to syllable timing.

    PubMed

    Leong, Victoria; Goswami, Usha

    2014-02-01

    Developmental dyslexia is associated with rhythmic difficulties, including impaired perception of beat patterns in music and prosodic stress patterns in speech. Spoken prosodic rhythm is cued by slow (<10 Hz) fluctuations in speech signal amplitude. Impaired neural oscillatory tracking of these slow amplitude modulation (AM) patterns is one plausible source of impaired rhythm tracking in dyslexia. Here, we characterise the temporal profile of the dyslexic rhythm deficit by examining rhythmic entrainment at multiple speech timescales. Adult dyslexic participants completed two experiments aimed at testing the perception and production of speech rhythm. In the perception task, participants tapped along to the beat of 4 metrically-regular nursery rhyme sentences. In the production task, participants produced the same 4 sentences in time to a metronome beat. Rhythmic entrainment was assessed using both traditional rhythmic indices and a novel AM-based measure, which utilised 3 dominant AM timescales in the speech signal each associated with a different phonological grain-sized unit (0.9-2.5 Hz, prosodic stress; 2.5-12 Hz, syllables; 12-40 Hz, phonemes). The AM-based measure revealed atypical rhythmic entrainment by dyslexic participants to syllable patterns in speech, in perception and production. In the perception task, both groups showed equally strong phase-locking to Syllable AM patterns, but dyslexic responses were entrained to a significantly earlier oscillatory phase angle than controls. In the production task, dyslexic utterances showed shorter syllable intervals, and differences in Syllable:Phoneme AM cross-frequency synchronisation. Our data support the view that rhythmic entrainment at slow (∼5 Hz, Syllable) rates is atypical in dyslexia, suggesting that neural mechanisms for syllable perception and production may also be atypical. These syllable timing deficits could contribute to the atypical development of phonological representations for spoken words, the central cognitive characteristic of developmental dyslexia across languages. Copyright © 2013 The Authors. Published by Elsevier B.V. All rights reserved.

  16. Assessment of rhythmic entrainment at multiple timescales in dyslexia: Evidence for disruption to syllable timing☆

    PubMed Central

    Leong, Victoria; Goswami, Usha

    2014-01-01

    Developmental dyslexia is associated with rhythmic difficulties, including impaired perception of beat patterns in music and prosodic stress patterns in speech. Spoken prosodic rhythm is cued by slow (<10 Hz) fluctuations in speech signal amplitude. Impaired neural oscillatory tracking of these slow amplitude modulation (AM) patterns is one plausible source of impaired rhythm tracking in dyslexia. Here, we characterise the temporal profile of the dyslexic rhythm deficit by examining rhythmic entrainment at multiple speech timescales. Adult dyslexic participants completed two experiments aimed at testing the perception and production of speech rhythm. In the perception task, participants tapped along to the beat of 4 metrically-regular nursery rhyme sentences. In the production task, participants produced the same 4 sentences in time to a metronome beat. Rhythmic entrainment was assessed using both traditional rhythmic indices and a novel AM-based measure, which utilised 3 dominant AM timescales in the speech signal each associated with a different phonological grain-sized unit (0.9–2.5 Hz, prosodic stress; 2.5–12 Hz, syllables; 12–40 Hz, phonemes). The AM-based measure revealed atypical rhythmic entrainment by dyslexic participants to syllable patterns in speech, in perception and production. In the perception task, both groups showed equally strong phase-locking to Syllable AM patterns, but dyslexic responses were entrained to a significantly earlier oscillatory phase angle than controls. In the production task, dyslexic utterances showed shorter syllable intervals, and differences in Syllable:Phoneme AM cross-frequency synchronisation. Our data support the view that rhythmic entrainment at slow (∼5 Hz, Syllable) rates is atypical in dyslexia, suggesting that neural mechanisms for syllable perception and production may also be atypical. These syllable timing deficits could contribute to the atypical development of phonological representations for spoken words, the central cognitive characteristic of developmental dyslexia across languages. This article is part of a Special Issue entitled . PMID:23916752

  17. Female Presence and Estrous State Influence Mouse Ultrasonic Courtship Vocalizations

    PubMed Central

    Hanson, Jessica L.; Hurley, Laura M.

    2012-01-01

    The laboratory mouse is an emerging model for context-dependent vocal signaling and reception. Mouse ultrasonic vocalizations are robustly produced in social contexts. In adults, male vocalization during courtship has become a model of interest for signal-receiver interactions. These vocalizations can be grouped into syllable types that are consistently produced by different subspecies and strains of mice. Vocalizations are unique to individuals, vary across development, and depend on social housing conditions. The behavioral significance of different syllable types, including the contexts in which different vocalizations are made and the responses listeners have to different types of vocalizations, is not well understood. We examined the effect of female presence and estrous state on male vocalizations by exploring the use of syllable types and the parameters of syllables during courtship. We also explored correlations between vocalizations and other behaviors. These experimental manipulations produced four main findings: 1) vocalizations varied among males, 2) the production of USVs and an increase in the use of a specific syllable type were temporally related to mounting behavior, 3) the frequency (kHz), bandwidth, and duration of syllables produced by males were influenced by the estrous phase of female partners, and 4) syllable types changed when females were removed. These findings show that mouse ultrasonic courtship vocalizations are sensitive to changes in female phase and presence, further demonstrating the context-sensitivity of these calls. PMID:22815817

  18. Influence of Syllable Structure on L2 Auditory Word Learning

    ERIC Educational Resources Information Center

    Hamada, Megumi; Goya, Hideki

    2015-01-01

    This study investigated the role of syllable structure in L2 auditory word learning. Based on research on cross-linguistic variation of speech perception and lexical memory, it was hypothesized that Japanese L1 learners of English would learn English words with an open-syllable structure without consonant clusters better than words with a…

  19. InfoSyll: A Syllabary Providing Statistical Information on Phonological and Orthographic Syllables

    ERIC Educational Resources Information Center

    Chetail, Fabienne; Mathey, Stephanie

    2010-01-01

    There is now a growing body of evidence in various languages supporting the claim that syllables are functional units of visual word processing. In the perspective of modeling the processing of polysyllabic words and the activation of syllables, current studies investigate syllabic effects with subtle manipulations. We present here a syllabary of…

  20. The Effects of Background Noise on Dichotic Listening to Consonant-Vowel Syllables

    ERIC Educational Resources Information Center

    Sequeira, Sarah Dos Santos; Specht, Karsten; Hamalainen, Heikki; Hugdahl, Kenneth

    2008-01-01

    Lateralization of verbal processing is frequently studied with the dichotic listening technique, yielding a so called right ear advantage (REA) to consonant-vowel (CV) syllables. However, little is known about how background noise affects the REA. To address this issue, we presented CV-syllables either in silence or with traffic background noise…

  1. Implicit Segmentation of a Stream of Syllables Based on Transitional Probabilities: An MEG Study

    ERIC Educational Resources Information Center

    Teinonen, Tuomas; Huotilainen, Minna

    2012-01-01

    Statistical segmentation of continuous speech, i.e., the ability to utilise transitional probabilities between syllables in order to detect word boundaries, is reflected in the brain's auditory event-related potentials (ERPs). The N1 and N400 ERP components are typically enhanced for word onsets compared to random syllables during active…

  2. EEG Correlates of Song Prosody: A New Look at the Relationship between Linguistic and Musical Rhythm

    PubMed Central

    Gordon, Reyna L.; Magne, Cyrille L.; Large, Edward W.

    2011-01-01

    Song composers incorporate linguistic prosody into their music when setting words to melody, a process called “textsetting.” Composers tend to align the expected stress of the lyrics with strong metrical positions in the music. The present study was designed to explore the idea that temporal alignment helps listeners to better understand song lyrics by directing listeners’ attention to instances where strong syllables occur on strong beats. Three types of textsettings were created by aligning metronome clicks with all, some or none of the strong syllables in sung sentences. Electroencephalographic recordings were taken while participants listened to the sung sentences (primes) and performed a lexical decision task on subsequent words and pseudowords (targets, presented visually). Comparison of misaligned and well-aligned sentences showed that temporal alignment between strong/weak syllables and strong/weak musical beats were associated with modulations of induced beta and evoked gamma power, which have been shown to fluctuate with rhythmic expectancies. Furthermore, targets that followed well-aligned primes elicited greater induced alpha and beta activity, and better lexical decision task performance, compared with targets that followed misaligned and varied sentences. Overall, these findings suggest that alignment of linguistic stress and musical meter in song enhances musical beat tracking and comprehension of lyrics by synchronizing neural activity with strong syllables. This approach may begin to explain the mechanisms underlying the relationship between linguistic and musical rhythm in songs, and how rhythmic attending facilitates learning and recall of song lyrics. Moreover, the observations reported here coincide with a growing number of studies reporting interactions between the linguistic and musical dimensions of song, which likely stem from shared neural resources for processing music and speech. PMID:22144972

  3. Subband-Based Group Delay Segmentation of Spontaneous Speech into Syllable-Like Units

    NASA Astrophysics Data System (ADS)

    Nagarajan, T.; Murthy, H. A.

    2004-12-01

    In the development of a syllable-centric automatic speech recognition (ASR) system, segmentation of the acoustic signal into syllabic units is an important stage. Although the short-term energy (STE) function contains useful information about syllable segment boundaries, it has to be processed before segment boundaries can be extracted. This paper presents a subband-based group delay approach to segment spontaneous speech into syllable-like units. This technique exploits the additive property of the Fourier transform phase and the deconvolution property of the cepstrum to smooth the STE function of the speech signal and make it suitable for syllable boundary detection. By treating the STE function as a magnitude spectrum of an arbitrary signal, a minimum-phase group delay function is derived. This group delay function is found to be a better representative of the STE function for syllable boundary detection. Although the group delay function derived from the STE function of the speech signal contains segment boundaries, the boundaries are difficult to determine in the context of long silences, semivowels, and fricatives. In this paper, these issues are specifically addressed and algorithms are developed to improve the segmentation performance. The speech signal is first passed through a bank of three filters, corresponding to three different spectral bands. The STE functions of these signals are computed. Using these three STE functions, three minimum-phase group delay functions are derived. By combining the evidence derived from these group delay functions, the syllable boundaries are detected. Further, a multiresolution-based technique is presented to overcome the problem of shift in segment boundaries during smoothing. Experiments carried out on the Switchboard and OGI-MLTS corpora show that the error in segmentation is at most 25 milliseconds for 67% and 76.6% of the syllable segments, respectively.

  4. The Effect of Orthography on the Lexical Encoding of Palatalized Consonants in L2 Russian.

    PubMed

    Simonchyk, Ala; Darcy, Isabelle

    2018-03-01

    The current study investigated the potential facilitative or inhibiting effects of orthography on the lexical encoding of palatalized consonants in L2 Russian. We hypothesized that learners with stable knowledge of orthographic and metalinguistic representations of palatalized consonants would display more accurate lexical encoding of the plain/palatalized contrast. The participants of the study were 40 American learners of Russian. Ten Russian native speakers served as a control group. The materials of the study comprised 20 real words, familiar to the participants, with target coronal consonants alternating in word-final and intervocalic positions. The participants performed three tasks: written picture naming, metalinguistic, and auditory word-picture matching. Results showed that learners were not entirely familiar with the grapheme-phoneme correspondences in L2 Russian. Even though they spelled almost all of these familiar Russian words accurately, they were able to identify the plain/palatalized status of the target consonants in these words with about 80% accuracy on a metalinguistic task. The effect of orthography on the lexical encoding was found to be dependent on the syllable position of the target consonants. In intervocalic position, learners erroneously relied on vowels following the target consonants rather than the consonants themselves to encode words with plain/palatalized consonants. In word-final position, although learners possessed the orthographic and metalinguistic knowledge of the difference in the palatalization status of the target consonants-and hence had established some aspects of the lexical representations for the words-those representations appeared to lack in phonological granularity and detail, perhaps due to the lack of perceptual salience.

  5. Beyond Decoding Deficit: Inhibitory Effect of Positional Syllable Frequency in Dyslexic Spanish Children

    ERIC Educational Resources Information Center

    Luque, Juan L.; López-Zamora, Miguel; Álvarez, Carlos J.; Bordoy, Soraya

    2013-01-01

    This study explores whether activation and inhibition word processes contribute to the characteristic speed deficits found in transparent orthographies (Wimmer, "Appl Psycholinguist" 14:1-33, 1993). A second and fourth grade sample of normal school readers and dyslexic school readers participated in a lexical decision task. Words were…

  6. Word Onset Patterns and Lexical Stress in English

    ERIC Educational Resources Information Center

    Kelly, Michael H.

    2004-01-01

    Theories of English phonology regard syllable onset patterns as irrelevant to the assignment of lexical stress. This paper describes three studies that challenge this position. Study 1 tested whether stress patterns on a large sample of disyllabic English words varied as a function of word onset. The incidence of trochaic stress increased…

  7. The role of stimulus-specific adaptation in songbird syntax generation

    NASA Astrophysics Data System (ADS)

    Wittenbach, Jason D.

    Sequential behaviors are an important part of the behavioral repertoire of many animals and understanding how neural circuits encode and generate such sequences is a long-standing question in neuroscience. The Bengalese finch is a useful model system for studying variable action sequences. The songs of these birds consist of well-defined vocal elements (syllables) that are strung together to form sequences. The ordering of the syllables within the sequence is variable but not random - it shows complex statistical patterns (syntax). While often thought to be first-order, the syntax of the Bengalese finch song shows a distinct form of history dependence where the probability of repeating a syllable decreases as a function of the number of repetitions that have already occurred. Current models of the Bengalese finch song control circuitry offer no explanation for this repetition adaptation. The Bengalese finch also uses real-time auditory feedback to control the song syntax. Considering these facts, we hypothesize that repetition adaptation in the Bengalese finch syntax may be caused by stimulus-specific adaptation - a wide-spread phenomenon where neural responses to a specific stimulus become weaker with repeated presentations of the same stimulus. We begin by proposing a computational model for the song-control circuit where an auditory feedback signal that undergoes stimulus-specific adaptation helps drive repeated syllables. We show that this model does indeed capture the repetition adaptation observed in Bengalese finch syntax; along the way, we derive a new probabilistic model for repetition adaptation. Key predictions of our model are analyzed in light of experiments performed by collaborators. Next we extend the model in order to predict how the syntax will change as a function of brain temperature. These predictions are compared to experimental results from collaborators where portions of the Bengalese finch song circuit are cooled in awake and behaving birds. Finally we show that repetition adaptation persists even in a simplified dynamical system model when a parameter controlling the repeat probability changes slowly over repetitions.

  8. Syllable Frequency Effects in Visual Word Recognition: Developmental Approach in French Children

    ERIC Educational Resources Information Center

    Maionchi-Pino, Norbert; Magnan, Annie; Ecalle, Jean

    2010-01-01

    This study investigates the syllable's role in the normal reading acquisition of French children at three grade levels (1st, 3rd, and 5th), using a modified version of Cole, Magnan, and Grainger's (1999) paradigm. We focused on the effects of syllable frequency and word frequency. The results suggest that from the first to third years of reading…

  9. A Longitudinal Study of Handwriting Skills in Pre-Schoolers: The Acquisition of Syllable Oriented Programming Strategies

    ERIC Educational Resources Information Center

    Soler Vilageliu, Olga; Kandel, Sonia

    2012-01-01

    Previous studies have shown the relevance of the syllable as a programming unit in handwriting production, both in adults and elementary school children. This longitudinal study focuses on the acquisition of writing skills in a group of preschoolers. It examines how and when the syllable structure of the word starts regulating motor programming in…

  10. The Effect of Modern Standard Arabic Orthography on Speech Production by Arab Children with Hearing Loss

    ERIC Educational Resources Information Center

    Most, Tova; Levin, Iris; Sarsour, Marwa

    2008-01-01

    This article examined the effect of Modern Standard Arabic orthography on speech production quality (syllable stress and vowels) by 23 Arabic-speaking children with severe or profound hearing loss aged 8-12 years. Children produced 15 one-syllable minimal pairs of words that differed in vowel length (short vs. long) and 20 two-syllable minimal…

  11. A Report of On-Going Research Aimed at Developing Unweighted and Weighted Syllable Lists.

    ERIC Educational Resources Information Center

    Sakiey, Elizabeth

    Knowing which syllables are most commonly used should aid in linguistic research and in the preparation of curriculum materials, particularly in reading. A research project has been undertaken to develop unweighted and weighted (by the frequency of the words in which they appear) syllable lists. At present, two of the project's three phases are…

  12. The Effect of Syllable Repetition Rate on Vocal Characteristics

    ERIC Educational Resources Information Center

    Topbas, Oya; Orlikoff, Robert F.; St. Louis, Kenneth O.

    2012-01-01

    This study examined whether mean vocal fundamental frequency ("F"[subscript 0]) or speech sound pressure level (SPL) varies with changes in syllable repetition rate. Twenty-four young adults (12 M and 12 F) repeated the syllables/p[inverted v]/,/p[inverted v]t[schwa]/, and/p[inverted v]t[schwa]k[schwa]/at a modeled "slow" rate of approximately one…

  13. An acoustical study of English word stress produced by Americans and Koreans

    NASA Astrophysics Data System (ADS)

    Yang, Byunggon

    2002-05-01

    Acoustical correlates of stress can be divided into duration, intensity, and fundamental frequency. This study examined the acoustical difference in the first two syllables of stressed English words produced by ten American and Korean speakers. The Korean subjects scored very high in TOEFL. They read, at a normal speed, a fable from which the acoustical parameters of eight words were analyzed. In order to make the data comparison meaningful, each parameter was collected at 100 dynamic time points proportional to the total duration of the two syllables. Then, the ratio of the parameter sum of the first rime to that of the second rime was calculated to determine the relative prominence of the syllables. Results showed that the durations of the first two syllables were almost comparable between the Americans and Koreans. However, statistically significant differences showed up in the diphthong pronunciations and in the words with the second syllable stressed. Also, remarkably high r-squared values were found between pairs of the three acoustical parameters, which suggests that either one or a combination of two or more parameters may account for the prominence of a syllable within a word. [Work supported by Korea Science Foundation R01-1999-00229.

  14. Stereotypy and variability of social calls among clustering female big-footed myotis (Myotis macrodactylus).

    PubMed

    Xiao, Yan-Hong; Wang, Lei; Hoyt, Joseph R; Jiang, Ting-Lei; Lin, Ai-Qing; Feng, Jiang

    2018-03-18

    Echolocating bats have developed advanced auditory perception systems, predominantly using acoustic signaling to communicate with each other. They can emit a diverse range of social calls in complex behavioral contexts. This study examined the vocal repertoire of five pregnant big-footed myotis bats (Myotis macrodactylus). In the process of clustering, the last individual to return to the colony (LI) emitted social calls that correlated with behavior, as recorded on a PC-based digital recorder. These last individuals could emit 10 simple monosyllabic and 27 complex multisyllabic types of calls, constituting four types of syllables. The social calls were composed of highly stereotyped syllables, hierarchically organized by a common set of syllables. However, intra-specific variation was also found in the number of syllables, syllable order and patterns of syllable repetition across call renditions. Data were obtained to characterize the significant individual differences that existed in the maximum frequency and duration of calls. Time taken to return to the roost was negatively associated with the diversity of social calls. Our findings indicate that variability in social calls may be an effective strategy taken by individuals during reintegration into clusters of female M. macrodactylus.

  15. Hierarchical Organization of Auditory and Motor Representations in Speech Perception: Evidence from Searchlight Similarity Analysis

    PubMed Central

    Evans, Samuel; Davis, Matthew H.

    2015-01-01

    How humans extract the identity of speech sounds from highly variable acoustic signals remains unclear. Here, we use searchlight representational similarity analysis (RSA) to localize and characterize neural representations of syllables at different levels of the hierarchically organized temporo-frontal pathways for speech perception. We asked participants to listen to spoken syllables that differed considerably in their surface acoustic form by changing speaker and degrading surface acoustics using noise-vocoding and sine wave synthesis while we recorded neural responses with functional magnetic resonance imaging. We found evidence for a graded hierarchy of abstraction across the brain. At the peak of the hierarchy, neural representations in somatomotor cortex encoded syllable identity but not surface acoustic form, at the base of the hierarchy, primary auditory cortex showed the reverse. In contrast, bilateral temporal cortex exhibited an intermediate response, encoding both syllable identity and the surface acoustic form of speech. Regions of somatomotor cortex associated with encoding syllable identity in perception were also engaged when producing the same syllables in a separate session. These findings are consistent with a hierarchical account of how variable acoustic signals are transformed into abstract representations of the identity of speech sounds. PMID:26157026

  16. The effects of age, education, and ethnicity on verbal fluency.

    PubMed

    Kempler, D; Teng, E L; Dick, M; Taussig, I M; Davis, D S

    1998-11-01

    A group of 317 healthy participants between 54 and 99 years of age performed a verbal fluency task. The participants included Chinese, Hispanic, and Vietnamese immigrants, as well as White and African American English speakers. They were given 1 min to name as many animals as possible in their native language. The results showed that more animal names were produced by younger people and those with more education. Language background was also an important factor: The Vietnamese produced the most animal names and the Spanish speakers produced the fewest. The exaggerated difference between these two groups is attributed to the fact that Vietnamese animal names are short (predominantly 1 syllable) while the Spanish animal names are longer than any other language in this study (2 and 3 syllables per word). Finally, although the ethnic groups named different animals, and appeared to vary in the variety of animal names they used, these factors did not affect overall verbal fluency performance.

  17. Imitation of contrastive lexical stress in children with speech delay

    NASA Astrophysics Data System (ADS)

    Vick, Jennell C.; Moore, Christopher A.

    2005-09-01

    This study examined the relationship between acoustic correlates of stress in trochaic (strong-weak), spondaic (strong-strong), and iambic (weak-strong) nonword bisyllables produced by children (30-50) with normal speech acquisition and children with speech delay. Ratios comparing the acoustic measures (vowel duration, rms, and f0) of the first syllable to the second syllable were calculated to evaluate the extent to which each phonetic parameter was used to mark stress. In addition, a calculation of the variability of jaw movement in each bisyllable was made. Finally, perceptual judgments of accuracy of stress production were made. Analysis of perceptual judgments indicated a robust difference between groups: While both groups of children produced errors in imitating the contrastive lexical stress models (~40%), the children with normal speech acquisition tended to produce trochaic forms in substitution for other stress types, whereas children with speech delay showed no preference for trochees. The relationship between segmental acoustic parameters, kinematic variability, and the ratings of stress by trained listeners will be presented.

  18. [Acoustic characteristics of adductor spasmodic dysphonia].

    PubMed

    Yang, Yang; Wang, Li-Ping

    2008-06-01

    To explore the acoustic characteristics of adductor spasmodic dysphonia. The acoustic characteristics, including acoustic signal of recorded voice, three-dimensional sonogram patterns and subjective assessment of voice, between 10 patients (7 women, 3 men) with adductor spasmodic dysphonia and 10 healthy volunteers (5 women, 5 men), were compared. The main clinical manifestation of adductor spasmodic dysphonia included the disorders of sound quality, rhyme and fluency. It demonstrated the tension dysphonia when reading, acoustic jitter, momentary fluctuation of frequency and volume, voice squeezing, interruption, voice prolongation, and losing normal chime. Among 10 patients, there were 1 mild dysphonia (abnormal syllable number < 25%), 6 moderate dysphonia (abnormal syllable number 25%-49%), 1 severe dysphonia (abnormal syllable number 50%-74%) and 2 extremely severe dysphonia (abnormal syllable number > or = 75%). The average reading time in 10 patients was 49 s, with reading time extension and aphasia area interruption in acoustic signals, whereas the average reading time in health control group was 30 s, without voice interruption. The aphasia ratio averaged 42%. The respective symptom syllable in different patients demonstrated in the three-dimensional sonogram. There were voice onset time prolongation, irregular, interrupted and even absent vowel formants. The consonant of symptom syllables displayed absence or prolongation of friction murmur in the block-friction murmur occasionally. The acoustic characteristics of adductor spasmodic dysphonia is the disorders of sound quality, rhyme and fluency. The three-dimensional sonogram of the symptom syllables show distinctive changes of proportional vowels or consonant phonemes.

  19. The Hyphen as a Syllabification Cue in Reading Bisyllabic and Multisyllabic Words among Finnish 1st and 2nd Graders

    ERIC Educational Resources Information Center

    Häikiö, Tuomo; Bertram, Raymond; Hyönä, Jukka

    2016-01-01

    Finnish ABC books present words with hyphens inserted at syllable boundaries. Syllabification by hyphens is abandoned in the 2nd grade for bisyllabic words, but continues for words with three or more syllables. The current eye movement study investigated how and to what extent syllable hyphens in bisyllabic ("kah-vi" "cof-fee")…

  20. The Transition from the Core Vowels to the Following Segments in Japanese Children Who Stutter: The Second, Third and Fourth Syllables

    ERIC Educational Resources Information Center

    Matsumoto-Shimamori, Sachiyo; Ito, Tomohiko; Fukuda, Suzy E.; Fukuda, Shinji

    2011-01-01

    Shimamori and Ito (2007, Syllable weight and phonological encoding in Japanese children who stutter. "Japanese Journal of Special Education", 44, 451-462; 2008, Syllable weight and frequency of stuttering: Comparison between children who stutter with and without a family history of stuttering. "Japanese Journal of Special Education", 45, 437-445;…

  1. Speech Recognition: Acoustic Phonetic and Lexical Knowledge Representation.

    DTIC Science & Technology

    1984-02-01

    be very powerful. We have also conducted a number of experiments examining the functional loads carried by segments in stressed versus unstressed...syllables. We found that the stressed * syllables provide a significantly greater amount of constraining power than unstressed syllables. This implies that...we started to investigate the constraints imposed by the stress pattern of words. Preliminary results indicate that knowledge about the stress pattern

  2. The Role of Syllables in Intermediate-Depth Stress-Timed Languages: Masked Priming Evidence in European Portuguese

    ERIC Educational Resources Information Center

    Campos, Ana Duarte; Mendes Oliveira, Helena; Soares, Ana Paula

    2018-01-01

    The role of syllables as a sublexical unit in visual word recognition and reading is well established in deep and shallow syllable-timed languages such as French and Spanish, respectively. However, its role in intermediate stress-timed languages remains unclear. This paper aims to overcome this gap by studying for the first time the role of…

  3. The Role of Secondary-Stressed and Unstressed-Unreduced Syllables in Word Recognition: Acoustic and Perceptual Studies with Russian Learners of English

    ERIC Educational Resources Information Center

    Banzina, Elina; Dilley, Laura C.; Hewitt, Lynne E.

    2016-01-01

    The importance of secondary-stressed (SS) and unstressed-unreduced (UU) syllable accuracy for spoken word recognition in English is as yet unclear. An acoustic study first investigated Russian learners' of English production of SS and UU syllables. Significant vowel quality and duration reductions in Russian-spoken SS and UU vowels were found,…

  4. Pitch Perception in Tone Language-Speaking Adults With and Without Autism Spectrum Disorders

    PubMed Central

    Cheng, Stella T. T.; Lam, Gary Y. H.

    2017-01-01

    Enhanced low-level pitch perception has been universally reported in autism spectrum disorders (ASD). This study examined whether tone language speakers with ASD exhibit this advantage. The pitch perception skill of 20 Cantonese-speaking adults with ASD was compared with that of 20 neurotypical individuals. Participants discriminated pairs of real syllable, pseudo-syllable (syllables that do not conform the phonotactic rules or are accidental gaps), and non-speech (syllables with attenuated high-frequency segmental content) stimuli contrasting pitch levels. The results revealed significantly higher discrimination ability in both groups for the non-speech stimuli than for the pseudo-syllables with one semitone difference. No significant group differences were noted. Different from previous findings, post hoc analysis found that enhanced pitch perception was observed in a subgroup of participants with ASD showing no history of delayed speech onset. The tone language experience may have modulated the pitch processing mechanism in the speakers in both ASD and non-ASD groups. PMID:28616150

  5. Stochastic Time Models of Syllable Structure

    PubMed Central

    Shaw, Jason A.; Gafos, Adamantios I.

    2015-01-01

    Drawing on phonology research within the generative linguistics tradition, stochastic methods, and notions from complex systems, we develop a modelling paradigm linking phonological structure, expressed in terms of syllables, to speech movement data acquired with 3D electromagnetic articulography and X-ray microbeam methods. The essential variable in the models is syllable structure. When mapped to discrete coordination topologies, syllabic organization imposes systematic patterns of variability on the temporal dynamics of speech articulation. We simulated these dynamics under different syllabic parses and evaluated simulations against experimental data from Arabic and English, two languages claimed to parse similar strings of segments into different syllabic structures. Model simulations replicated several key experimental results, including the fallibility of past phonetic heuristics for syllable structure, and exposed the range of conditions under which such heuristics remain valid. More importantly, the modelling approach consistently diagnosed syllable structure proving resilient to multiple sources of variability in experimental data including measurement variability, speaker variability, and contextual variability. Prospects for extensions of our modelling paradigm to acoustic data are also discussed. PMID:25996153

  6. The Breadth of Coarticulatory Units in Children and Adults

    ERIC Educational Resources Information Center

    Goffman, Lisa; Smith, Anne; Heisler, Lori; Ho, Michael

    2008-01-01

    Purpose: To assess, in children and adults, the breadth of coarticulatory movements associated with a single rounded vowel. Method: Upper and lower lip movements were recorded from 8 young adults and 8 children (aged 4-5 years). A single rounded versus unrounded vowel was embedded in the medial position of pairs of 7-word/7-syllable sentences.…

  7. Probability of Equivalence Formation: Familiar Stimuli and Training Sequence

    ERIC Educational Resources Information Center

    Arntzen, Erik

    2004-01-01

    The present study was conducted to show how responding in accord with equivalence relations changes as a function of position of familiar stimuli, pictures, and with the use of nonsense syllables in an MTO-training structure. Fifty college students were tested for responding in accord with equivalence in an AB, CB, DB, and EB training structure.…

  8. The Frame Constraint on Experimentally Elicited Speech Errors in Japanese

    ERIC Educational Resources Information Center

    Saito, Akie; Inoue, Tomoyoshi

    2017-01-01

    The so-called syllable position effect in speech errors has been interpreted as reflecting constraints posed by the frame structure of a given language, which is separately operating from linguistic content during speech production. The effect refers to the phenomenon that when a speech error occurs, replaced and replacing sounds tend to be in the…

  9. Tonal Targets in Early Child English, Spanish, and Catalan

    ERIC Educational Resources Information Center

    Astruc, Lluisa; Payne, Elinor; Post, Brechtje; Vanrell, Maria del Mar; Prieto, Pilar

    2013-01-01

    This study analyses the scaling and alignment of low and high intonational targets in the speech of 27 children--nine English-speaking, nine Catalan-speaking and nine Spanish-speaking--between the ages of two and six years. We compared the intonational patterns of words controlled for number of syllables and stress position in the child speech to…

  10. The Nature of the Phonological Processing in French Dyslexic Children: Evidence for the Phonological Syllable and Linguistic Features' Role in Silent Reading and Speech Discrimination

    ERIC Educational Resources Information Center

    Maionchi-Pino, Norbert; Magnan, Annie; Ecalle, Jean

    2010-01-01

    This study investigated the status of phonological representations in French dyslexic children (DY) compared with reading level- (RL) and chronological age-matched (CA) controls. We focused on the syllable's role and on the impact of French linguistic features. In Experiment 1, we assessed oral discrimination abilities of pairs of syllables that…

  11. Effects of age and hearing loss on recognition of unaccented and accented multisyllabic words.

    PubMed

    Gordon-Salant, Sandra; Yeni-Komshian, Grace H; Fitzgibbons, Peter J; Cohen, Julie I

    2015-02-01

    The effects of age and hearing loss on recognition of unaccented and accented words of varying syllable length were investigated. It was hypothesized that with increments in length of syllables, there would be atypical alterations in syllable stress in accented compared to native English, and that these altered stress patterns would be sensitive to auditory temporal processing deficits with aging. Sets of one-, two-, three-, and four-syllable words with the same initial syllable were recorded by one native English and two Spanish-accented talkers. Lists of these words were presented in isolation and in sentence contexts to younger and older normal-hearing listeners and to older hearing-impaired listeners. Hearing loss effects were apparent for unaccented and accented monosyllabic words, whereas age effects were observed for recognition of accented multisyllabic words, consistent with the notion that altered syllable stress patterns with accent are sensitive for revealing effects of age. Older listeners also exhibited lower recognition scores for moderately accented words in sentence contexts than in isolation, suggesting that the added demands on working memory for words in sentence contexts impact recognition of accented speech. The general pattern of results suggests that hearing loss, age, and cognitive factors limit the ability to recognize Spanish-accented speech.

  12. Effects of age and hearing loss on recognition of unaccented and accented multisyllabic words

    PubMed Central

    Gordon-Salant, Sandra; Yeni-Komshian, Grace H.; Fitzgibbons, Peter J.; Cohen, Julie I.

    2015-01-01

    The effects of age and hearing loss on recognition of unaccented and accented words of varying syllable length were investigated. It was hypothesized that with increments in length of syllables, there would be atypical alterations in syllable stress in accented compared to native English, and that these altered stress patterns would be sensitive to auditory temporal processing deficits with aging. Sets of one-, two-, three-, and four-syllable words with the same initial syllable were recorded by one native English and two Spanish-accented talkers. Lists of these words were presented in isolation and in sentence contexts to younger and older normal-hearing listeners and to older hearing-impaired listeners. Hearing loss effects were apparent for unaccented and accented monosyllabic words, whereas age effects were observed for recognition of accented multisyllabic words, consistent with the notion that altered syllable stress patterns with accent are sensitive for revealing effects of age. Older listeners also exhibited lower recognition scores for moderately accented words in sentence contexts than in isolation, suggesting that the added demands on working memory for words in sentence contexts impact recognition of accented speech. The general pattern of results suggests that hearing loss, age, and cognitive factors limit the ability to recognize Spanish-accented speech. PMID:25698021

  13. Effects of obstruent consonants on the F0 contour

    NASA Astrophysics Data System (ADS)

    Hanson, Helen M.

    2003-10-01

    When a vowel follows an obstruent consonant, the fundamental frequency in the first few tens of milliseconds of the vowel is influenced by the voicing characteristics of the consonant. The goal of the research reported here is to model this influence, with the intention of improving generation of F0 contours in rule-based speech synthesis. Data have been recorded from 10 subjects. Stops, fricatives, and the nasal /m/ were paired with the vowels /i,opena/ to form CVm syllables. The syllables mVm served as baselines with which to compare the obstruents. The target syllables were embedded in carrier sentences. Intonation was varied so that each target syllable was produced with either a high, low, or no pitch accent. Results vary among subjects, but in general, obstruent effects on F0 primarily occur when the syllable carries a high pitch. In that case, F0 is increased relative to the baseline following voiceless obstruents, but F0 closely follows the baseline following voiced obstruents. After voiceless obstruents, F0 may be increased for up to 80 ms following voicing onset. When a syllable carries a low or no pitch accent, F0 is increased slightly following all obstruents. [Work supported by NIH Grant No. DC04331.

  14. Hierarchical Organization of Auditory and Motor Representations in Speech Perception: Evidence from Searchlight Similarity Analysis.

    PubMed

    Evans, Samuel; Davis, Matthew H

    2015-12-01

    How humans extract the identity of speech sounds from highly variable acoustic signals remains unclear. Here, we use searchlight representational similarity analysis (RSA) to localize and characterize neural representations of syllables at different levels of the hierarchically organized temporo-frontal pathways for speech perception. We asked participants to listen to spoken syllables that differed considerably in their surface acoustic form by changing speaker and degrading surface acoustics using noise-vocoding and sine wave synthesis while we recorded neural responses with functional magnetic resonance imaging. We found evidence for a graded hierarchy of abstraction across the brain. At the peak of the hierarchy, neural representations in somatomotor cortex encoded syllable identity but not surface acoustic form, at the base of the hierarchy, primary auditory cortex showed the reverse. In contrast, bilateral temporal cortex exhibited an intermediate response, encoding both syllable identity and the surface acoustic form of speech. Regions of somatomotor cortex associated with encoding syllable identity in perception were also engaged when producing the same syllables in a separate session. These findings are consistent with a hierarchical account of how variable acoustic signals are transformed into abstract representations of the identity of speech sounds. © The Author 2015. Published by Oxford University Press.

  15. Perception of steady-state vowels and vowelless syllables by adults and children

    NASA Astrophysics Data System (ADS)

    Nittrouer, Susan

    2005-04-01

    Vowels can be produced as long, isolated, and steady-state, but that is not how they are found in natural speech. Instead natural speech consists of almost continuously changing (i.e., dynamic) acoustic forms from which mature listeners recover underlying phonetic form. Some theories suggest that children need steady-state information to recognize vowels (and so learn vowel systems), even though that information is sparse in natural speech. The current study examined whether young children can recover vowel targets from dynamic forms, or whether they need steady-state information. Vowel recognition was measured for adults and children (3, 5, and 7 years) for natural productions of /dæd/, /dUd/ /æ/, /U/ edited to make six stimulus sets: three dynamic (whole syllables; syllables with middle 50-percent replaced by cough; syllables with all but the first and last three pitch periods replaced by cough), and three steady-state (natural, isolated vowels; reiterated pitch periods from those vowels; reiterated pitch periods from the syllables). Adults scored nearly perfectly on all but first/last three pitch period stimuli. Children performed nearly perfectly only when the entire syllable was heard, and performed similarly (near 80%) for all other stimuli. Consequently, children need dynamic forms to perceive vowels; steady-state forms are not preferred.

  16. Multiple functional units in the preattentive segmentation of speech in Japanese: evidence from word illusions.

    PubMed

    Nakamura, Miyoko; Kolinsky, Régine

    2014-12-01

    We explored the functional units of speech segmentation in Japanese using dichotic presentation and a detection task requiring no intentional sublexical analysis. Indeed, illusory perception of a target word might result from preattentive migration of phonemes, morae, or syllables from one ear to the other. In Experiment I, Japanese listeners detected targets presented in hiragana and/or kanji. Phoneme migrations did occur, suggesting that orthography-independent sublexical constituents play some role in segmentation. However, syllable and especially mora migrations were more numerous. This pattern of results was not observed in French speakers (Experiment 2), suggesting that it reflects native segmentation in Japanese. To control for the intervention of kanji representations (many words are written in kanji, and one kanji often corresponds to one syllable), in Experiment 3, Japanese listeners were presented with target loanwords that can be written only in katakana. Again, phoneme migrations occurred, while the first mora and syllable led to similar rates of illusory percepts. No migration occurred for the second, "special" mora (/J/ or/N/), probably because this constitutes the latter part of a heavy syllable. Overall, these findings suggest that multiple units, such as morae, syllables, and even phonemes, function independently of orthographic knowledge in Japanese preattentive speech segmentation.

  17. [Two cases of fluent aphasia with selective difficulty of syllable identification].

    PubMed

    Endo, K; Suzuki, K; Yamadori, A; Fujii, T; Tobita, M; Ohtake, H

    1999-10-01

    We report two aphasic patients who could discriminate Japanese syllables but could not identify them. Case 1 was a 51-year-old right handed woman with 12-year education. Case 2 was a 50-year-old right handed man with 9-year education. They developed fluent aphasia after a cerebral infarction. Brain MRI of case 1 revealed widely distributed lesions including inferior frontal, superior temporal, angular and supramarginal gyri. Lesions revealed by Brain CT in case 2 included the left superior and middle temporal, angular and supramarginal gyri. Both showed severe impairment of repetition and confrontation naming. No difference of performance was present between repetition of single syllables and polysyllabic words. On the contrary, oral reading of Kana characters were preserved. We examined their ability to perceive syllables in detail. In the discrimination task, they judged whether a pair of heard syllables was same or different. Case 1 was correct in 85% of the tasks and case 2 in 98%. In an identification task, they heard a syllable and chose a corresponding Kana, Kanji, or picture out of 10 respective candidates. Case 1 was correct only in 30% and case 2 in 50% of these tasks. On the other hand, selection of a correct target in response to a polysyllabic word was much better, i.e. 70% in case 1 and 90% in case 2. Based on these data we concluded that (1) syllabic identification is a different process from syllabic discrimination, and (2) comprehension of a polysyllabic word can be achieved even when the precise phonological analysis of continuent syllables are impaired.

  18. Role of the motor system in language knowledge.

    PubMed

    Berent, Iris; Brem, Anna-Katharine; Zhao, Xu; Seligson, Erica; Pan, Hong; Epstein, Jane; Stern, Emily; Galaburda, Albert M; Pascual-Leone, Alvaro

    2015-02-17

    All spoken languages express words by sound patterns, and certain patterns (e.g., blog) are systematically preferred to others (e.g., lbog). What principles account for such preferences: does the language system encode abstract rules banning syllables like lbog, or does their dislike reflect the increased motor demands associated with speech production? More generally, we ask whether linguistic knowledge is fully embodied or whether some linguistic principles could potentially be abstract. To address this question, here we gauge the sensitivity of English speakers to the putative universal syllable hierarchy (e.g., blif ≻ bnif ≻ bdif ≻ lbif) while undergoing transcranial magnetic stimulation (TMS) over the cortical motor representation of the left orbicularis oris muscle. If syllable preferences reflect motor simulation, then worse-formed syllables (e.g., lbif) should (i) elicit more errors; (ii) engage more strongly motor brain areas; and (iii) elicit stronger effects of TMS on these motor regions. In line with the motor account, we found that repetitive TMS pulses impaired participants' global sensitivity to the number of syllables, and functional MRI confirmed that the cortical stimulation site was sensitive to the syllable hierarchy. Contrary to the motor account, however, ill-formed syllables were least likely to engage the lip sensorimotor area and they were least impaired by TMS. Results suggest that speech perception automatically triggers motor action, but this effect is not causally linked to the computation of linguistic structure. We conclude that the language and motor systems are intimately linked, yet distinct. Language is designed to optimize motor action, but its knowledge includes principles that are disembodied and potentially abstract.

  19. Role of the motor system in language knowledge

    PubMed Central

    Berent, Iris; Brem, Anna-Katharine; Zhao, Xu; Seligson, Erica; Pan, Hong; Epstein, Jane; Stern, Emily; Galaburda, Albert M.; Pascual-Leone, Alvaro

    2015-01-01

    All spoken languages express words by sound patterns, and certain patterns (e.g., blog) are systematically preferred to others (e.g., lbog). What principles account for such preferences: does the language system encode abstract rules banning syllables like lbog, or does their dislike reflect the increased motor demands associated with speech production? More generally, we ask whether linguistic knowledge is fully embodied or whether some linguistic principles could potentially be abstract. To address this question, here we gauge the sensitivity of English speakers to the putative universal syllable hierarchy (e.g., blif≻bnif≻bdif≻lbif) while undergoing transcranial magnetic stimulation (TMS) over the cortical motor representation of the left orbicularis oris muscle. If syllable preferences reflect motor simulation, then worse-formed syllables (e.g., lbif) should (i) elicit more errors; (ii) engage more strongly motor brain areas; and (iii) elicit stronger effects of TMS on these motor regions. In line with the motor account, we found that repetitive TMS pulses impaired participants’ global sensitivity to the number of syllables, and functional MRI confirmed that the cortical stimulation site was sensitive to the syllable hierarchy. Contrary to the motor account, however, ill-formed syllables were least likely to engage the lip sensorimotor area and they were least impaired by TMS. Results suggest that speech perception automatically triggers motor action, but this effect is not causally linked to the computation of linguistic structure. We conclude that the language and motor systems are intimately linked, yet distinct. Language is designed to optimize motor action, but its knowledge includes principles that are disembodied and potentially abstract. PMID:25646465

  20. Place preference and vocal learning rely on distinct reinforcers in songbirds.

    PubMed

    Murdoch, Don; Chen, Ruidong; Goldberg, Jesse H

    2018-04-30

    In reinforcement learning (RL) agents are typically tasked with maximizing a single objective function such as reward. But it remains poorly understood how agents might pursue distinct objectives at once. In machines, multiobjective RL can be achieved by dividing a single agent into multiple sub-agents, each of which is shaped by agent-specific reinforcement, but it remains unknown if animals adopt this strategy. Here we use songbirds to test if navigation and singing, two behaviors with distinct objectives, can be differentially reinforced. We demonstrate that strobe flashes aversively condition place preference but not song syllables. Brief noise bursts aversively condition song syllables but positively reinforce place preference. Thus distinct behavior-generating systems, or agencies, within a single animal can be shaped by correspondingly distinct reinforcement signals. Our findings suggest that spatially segregated vocal circuits can solve a credit assignment problem associated with multiobjective learning.

  1. The development of phonological skills in late and early talkers

    PubMed Central

    KEHOE, Margaret; CHAPLIN, Elisa; MUDRY, Pauline; FRIEND, Margaret

    2016-01-01

    This study examined the relationship between phonological and lexical development in a group of French-speaking children (n=30), aged 29 months. The participants were divided into three sub-groups based on the number of words in their expressive vocabulary : low vocabulary (below the 15th percentile) (<< late-talkers >>) ; average-sized vocabulary (40-60th percentile) (<< middle group >>) and advanced vocabulary (above the 90th percentile) (<< precocious >> or “early talkers”). The phonological abilities (e.g., phonemic inventory, percentage of correct consonants, and phonological processes) of the three groups were compared. The comparison was based on analyses of spontaneous language samples. Most findings were consistent with previous results found in English-speaking children, indicating that the phonological abilities of late talkers are less well developed than those of children with average-sized vocabularies which in turn are less well-developed than those of children with advanced vocabularies. Nevertheless, several phonological measures were not related to vocabulary size, in particular those concerning syllable-final position. These findings differ from those obtained in English. The article finally discusses the clinical implications of the findings for children with delayed language development. PMID:26924855

  2. Authorship Discovery in Blogs Using Bayesian Classification with Corrective Scaling

    DTIC Science & Technology

    2008-06-01

    4 2.3 W. Fucks ’ Diagram of n-Syllable Word Frequencies . . . . . . . . . . . . . . 5 3.1 Confusion Matrix for All Test Documents of 500...of the books which scholars believed he had. • Wilhelm Fucks discriminated between authors using the average number of syllables per word and average...distance between equal-syllabled words [8]. Fucks , too, concluded that a study such as his reveals a “possibility of a quantitative classification

  3. Are syllabification and resyllabification strategies phonotactically directed in French children with dyslexia? A preliminary report.

    PubMed

    Maïonchi-Pino, Norbert; de Cara, Bruno; Ecalle, Jean; Magnan, Annie

    2012-04-01

    In this study, the authors queried whether French-speaking children with dyslexia were sensitive to consonant sonority and position within syllable boundaries to influence a phonological syllable-based segmentation in silent reading. Participants included 15 French-speaking children with dyslexia, compared with 30 chronological age-matched and reading level-matched controls. Children were tested with an audiovisual recognition task. A target pseudoword (TOLPUDE) was simultaneously presented visually and auditorily and then was compared with a printed test pseudoword that either was identical or differed after the coda deletion (TOPUDE) or the onset deletion (TOLUDE). The intervocalic consonant sequences had either a sonorant coda-sonorant onset (TOR.LADE), sonorant coda-obstruent onset (TOL.PUDE), obstruent coda-sonorant onset (DOT.LIRE), or obstruent coda-obstruent onset (BIC.TADE) sonority profile. All children processed identity better than they processed deletion, especially with the optimal sonorant coda-obstruent onset sonority profile. However, children preserved syllabification (coda deletion; TO.PUDE) rather than resyllabification (onset deletion; TO.LUDE) with intervocalic consonant sequence reductions, especially when sonorant codas were deleted but the optimal intersyllable contact was respected. It was surprising to find that although children with dyslexia generally exhibit phonological and acoustic-phonetic impairments (voicing), they showed sensitivity to the optimal sonority profile and a preference for preserved syllabification. The authors proposed a sonority-modulated explanation to account for phonological syllable-based processing. Educational implications are discussed.

  4. Impaired extraction of speech rhythm from temporal modulation patterns in speech in developmental dyslexia

    PubMed Central

    Leong, Victoria; Goswami, Usha

    2014-01-01

    Dyslexia is associated with impaired neural representation of the sound structure of words (phonology). The “phonological deficit” in dyslexia may arise in part from impaired speech rhythm perception, thought to depend on neural oscillatory phase-locking to slow amplitude modulation (AM) patterns in the speech envelope. Speech contains AM patterns at multiple temporal rates, and these different AM rates are associated with phonological units of different grain sizes, e.g., related to stress, syllables or phonemes. Here, we assess the ability of adults with dyslexia to use speech AMs to identify rhythm patterns (RPs). We study 3 important temporal rates: “Stress” (~2 Hz), “Syllable” (~4 Hz) and “Sub-beat” (reduced syllables, ~14 Hz). 21 dyslexics and 21 controls listened to nursery rhyme sentences that had been tone-vocoded using either single AM rates from the speech envelope (Stress only, Syllable only, Sub-beat only) or pairs of AM rates (Stress + Syllable, Syllable + Sub-beat). They were asked to use the acoustic rhythm of the stimulus to identity the original nursery rhyme sentence. The data showed that dyslexics were significantly poorer at detecting rhythm compared to controls when they had to utilize multi-rate temporal information from pairs of AMs (Stress + Syllable or Syllable + Sub-beat). These data suggest that dyslexia is associated with a reduced ability to utilize AMs <20 Hz for rhythm recognition. This perceptual deficit in utilizing AM patterns in speech could be underpinned by less efficient neuronal phase alignment and cross-frequency neuronal oscillatory synchronization in dyslexia. Dyslexics' perceptual difficulties in capturing the full spectro-temporal complexity of speech over multiple timescales could contribute to the development of impaired phonological representations for words, the cognitive hallmark of dyslexia across languages. PMID:24605099

  5. Effects of Diet on Early Stage Cortical Perception and Discrimination of Syllables Differing in Voice-Onset Time: A Longitudinal ERP Study in 3 and 6 Month Old Infants

    ERIC Educational Resources Information Center

    Pivik, R. T.; Andres, Aline; Badger, Thomas M.

    2012-01-01

    The influence of diet on cortical processing of syllables was examined at 3 and 6 months in 239 infants who were breastfed or fed milk or soy-based formula. Event-related potentials to syllables differing in voice-onset-time were recorded from placements overlying brain areas specialized for language processing. P1 component amplitude and latency…

  6. Effects of syllable structure in aphasic errors: implications for a new model of speech production.

    PubMed

    Romani, Cristina; Galluzzi, Claudia; Bureca, Ivana; Olson, Andrew

    2011-03-01

    Current models of word production assume that words are stored as linear sequences of phonemes which are structured into syllables only at the moment of production. This is because syllable structure is always recoverable from the sequence of phonemes. In contrast, we present theoretical and empirical evidence that syllable structure is lexically represented. Storing syllable structure would have the advantage of making representations more stable and resistant to damage. On the other hand, re-syllabifications affect only a minimal part of phonological representations and occur only in some languages and depending on speech register. Evidence for these claims comes from analyses of aphasic errors which not only respect phonotactic constraints, but also avoid transformations which move the syllabic structure of the word further away from the original structure, even when equating for segmental complexity. This is true across tasks, types of errors, and, crucially, types of patients. The same syllabic effects are shown by apraxic patients and by phonological patients who have more central difficulties in retrieving phonological representations. If syllable structure was only computed after phoneme retrieval, it would have no way to influence the errors of phonological patients. Our results have implications for psycholinguistic and computational models of language as well as for clinical and educational practices. Copyright © 2010 Elsevier Inc. All rights reserved.

  7. Curvilinear relationship between phonological working memory load and social-emotional modulation

    PubMed Central

    Mano, Quintino R.; Brown, Gregory G.; Bolden, Khalima; Aupperle, Robin; Sullivan, Sarah; Paulus, Martin P.; Stein, Murray B.

    2015-01-01

    Accumulating evidence suggests that working memory load is an important factor for the interplay between cognitive and facial-affective processing. However, it is unclear how distraction caused by perception of faces interacts with load-related performance. We developed a modified version of the delayed match-to-sample task wherein task-irrelevant facial distracters were presented early in the rehearsal of pseudoword memoranda that varied incrementally in load size (1-syllable, 2-syllables, or 3-syllables). Facial distracters displayed happy, sad, or neutral expressions in Experiment 1 (N=60) and happy, fearful, or neutral expressions in Experiment 2 (N=29). Facial distracters significantly disrupted task performance in the intermediate load condition (2-syllable) but not in the low or high load conditions (1- and 3-syllables, respectively), an interaction replicated and generalised in Experiment 2. All facial distracters disrupted working memory in the intermediate load condition irrespective of valence, suggesting a primary and general effect of distraction caused by faces. However, sad and fearful faces tended to be less disruptive than happy faces, suggesting a secondary and specific valence effect. Working memory appears to be most vulnerable to social-emotional information at intermediate loads. At low loads, spare capacity is capable of accommodating the combinatorial load (1-syllable plus facial distracter), whereas high loads maximised capacity and deprived facial stimuli from occupying working memory slots to cause disruption. PMID:22928750

  8. Imaging auditory representations of song and syllables in populations of sensorimotor neurons essential to vocal communication.

    PubMed

    Peh, Wendy Y X; Roberts, Todd F; Mooney, Richard

    2015-04-08

    Vocal communication depends on the coordinated activity of sensorimotor neurons important to vocal perception and production. How vocalizations are represented by spatiotemporal activity patterns in these neuronal populations remains poorly understood. Here we combined intracellular recordings and two-photon calcium imaging in anesthetized adult zebra finches (Taeniopygia guttata) to examine how learned birdsong and its component syllables are represented in identified projection neurons (PNs) within HVC, a sensorimotor region important for song perception and production. These experiments show that neighboring HVC PNs can respond at markedly different times to song playback and that different syllables activate spatially intermingled PNs within a local (~100 μm) region of HVC. Moreover, noise correlations were stronger between PNs that responded most strongly to the same syllable and were spatially graded within and between classes of PNs. These findings support a model in which syllabic and temporal features of song are represented by spatially intermingled PNs functionally organized into cell- and syllable-type networks within local spatial scales in HVC. Copyright © 2015 the authors 0270-6474/15/355589-17$15.00/0.

  9. Sing that Tune: Infants’ Perception of Melody and Lyrics and the Facilitation of Phonetic Recognition in Songs

    PubMed Central

    Lebedeva, Gina C.; Kuhl, Patricia K.

    2010-01-01

    To better understand how infants process complex auditory input, this study investigated whether 11-month-old infants perceive the pitch (melodic) or the phonetic (lyric) components within songs as more salient, and whether melody facilitates phonetic recognition. Using a preferential looking paradigm, uni-dimensional and multi-dimensional songs were tested; either the pitch or syllable order of the stimuli varied. As a group, infants detected a change in pitch order in a 4-note sequence when the syllables were redundant (Experiment 1), but did not detect the identical pitch change with variegated syllables (Experiment 2). Infants were better able to detect a change in syllable order in a sung sequence (Experiment 2) than the identical syllable change in a spoken sequence (Experiment 1). These results suggest that by 11 months, infants cannot “ignore” phonetic information in the context of perceptually salient pitch variation. Moreover, the increased phonetic recognition in song contexts mirrors findings that demonstrate advantages of infant-directed speech. Findings are discussed in terms of how stimulus complexity interacts with the perception of sung speech in infancy. PMID:20472295

  10. Copy Counts

    ERIC Educational Resources Information Center

    Beaumont, Lee R.

    1970-01-01

    The level of difficulty of straight copy, which is used to measure typewriting speed, is influenced by syllable intensity (the average number of syllables per word), stroke intensity (average number of strokes per word), and high-frequency words. (CH)

  11. The limits of metrical segmentation: intonation modulates infants' extraction of embedded trochees.

    PubMed

    Zahner, Katharina; Schönhuber, Muna; Braun, Bettina

    2016-11-01

    We tested German nine-month-olds' reliance on pitch and metrical stress for segmentation. In a headturn-preference paradigm, infants were familiarized with trisyllabic words (weak-strong-weak (WSW) stress pattern) in sentence-contexts. The words were presented in one of three naturally occurring intonation conditions: one in which high pitch was aligned with the stressed syllable and two misalignment conditions (with high pitch preceding vs. following the stressed syllable). Infants were tested on the SW unit of the WSW carriers. Experiment 1 showed recognition only when the stressed syllable was high-pitched. Intonation of test items (similar vs. dissimilar to familiarization) had no influence (Experiment 2). Thus, German nine-month-olds perceive stressed syllables as word onsets only when high-pitched, although they already generalize over different pitch contours. Different mechanisms underlying this pattern of results are discussed.

  12. Alternating motion rate as an index of speech motor disorder in traumatic brain injury.

    PubMed

    Wang, Yu-Tsai; Kent, Ray D; Duffy, Joseph R; Thomas, Jack E; Weismer, Gary

    2004-01-01

    The task of syllable alternating motion rate (AMR) (also called diadochokinesis) is suitable for examining speech disorders of varying degrees of severity and in individuals with varying levels of linguistic and cognitive ability. However, very limited information on this task has been published for subjects with traumatic brain injury (TBI). This study is a quantitative and qualitative acoustic analysis of AMR in seven subjects with TBI. The primary goal was to use acoustic analyses to assess speech motor control disturbances for the group as a whole and for individual patients. Quantitative analyses included measures of syllable rate, syllable and intersyllable gap durations, energy maxima, and voice onset time (VOT). Qualitative analyses included classification of features evident in spectrograms and waveforms to provide a more detailed description. The TBI group had (1) a slowed syllable rate due mostly to lengthened syllables and, to a lesser degree, lengthened intersyllable gaps, (2) highly correlated syllable rates between AMR and conversation, (3) temporal and energy maxima irregularities within repetition sequences, (4) normal median VOT values but with large variation, and (5) a number of speech production abnormalities revealed by qualitative analysis, including explosive speech quality, breathy voice quality, phonatory instability, multiple or missing stop bursts, continuous voicing, and spirantization. The relationships between these findings and TBI speakers' neurological status and dysarthria types are also discussed. It was concluded that acoustic analyses of the AMR task provides specific information on motor speech limitations in individuals with TBI.

  13. Vocal Generalization Depends on Gesture Identity and Sequence

    PubMed Central

    Sober, Samuel J.

    2014-01-01

    Generalization, the brain's ability to transfer motor learning from one context to another, occurs in a wide range of complex behaviors. However, the rules of generalization in vocal behavior are poorly understood, and it is unknown how vocal learning generalizes across an animal's entire repertoire of natural vocalizations and sequences. Here, we asked whether generalization occurs in a nonhuman vocal learner and quantified its properties. We hypothesized that adaptive error correction of a vocal gesture produced in one sequence would generalize to the same gesture produced in other sequences. To test our hypothesis, we manipulated the fundamental frequency (pitch) of auditory feedback in Bengalese finches (Lonchura striata var. domestica) to create sensory errors during vocal gestures (song syllables) produced in particular sequences. As hypothesized, error-corrective learning on pitch-shifted vocal gestures generalized to the same gestures produced in other sequential contexts. Surprisingly, generalization magnitude depended strongly on sequential distance from the pitch-shifted syllables, with greater adaptation for gestures produced near to the pitch-shifted syllable. A further unexpected result was that nonshifted syllables changed their pitch in the direction opposite from the shifted syllables. This apparently antiadaptive pattern of generalization could not be explained by correlations between generalization and the acoustic similarity to the pitch-shifted syllable. These findings therefore suggest that generalization depends on the type of vocal gesture and its sequential context relative to other gestures and may reflect an advantageous strategy for vocal learning and maintenance. PMID:24741046

  14. Functional asymmetry and effective connectivity of the auditory system during speech perception is modulated by the place of articulation of the consonant- A 7T fMRI study

    PubMed Central

    Specht, Karsten; Baumgartner, Florian; Stadler, Jörg; Hugdahl, Kenneth; Pollmann, Stefan

    2014-01-01

    To differentiate between stop-consonants, the auditory system has to detect subtle place of articulation (PoA) and voice-onset time (VOT) differences between stop-consonants. How this differential processing is represented on the cortical level remains unclear. The present functional magnetic resonance (fMRI) study takes advantage of the superior spatial resolution and high sensitivity of ultra-high-field 7 T MRI. Subjects were attentively listening to consonant–vowel (CV) syllables with an alveolar or bilabial stop-consonant and either a short or long VOT. The results showed an overall bilateral activation pattern in the posterior temporal lobe during the processing of the CV syllables. This was however modulated strongest by PoA such that syllables with an alveolar stop-consonant showed stronger left lateralized activation. In addition, analysis of underlying functional and effective connectivity revealed an inhibitory effect of the left planum temporale (PT) onto the right auditory cortex (AC) during the processing of alveolar CV syllables. Furthermore, the connectivity result indicated also a directed information flow from the right to the left AC, and further to the left PT for all syllables. These results indicate that auditory speech perception relies on an interplay between the left and right ACs, with the left PT as modulator. Furthermore, the degree of functional asymmetry is determined by the acoustic properties of the CV syllables. PMID:24966841

  15. Ixpantepec Nieves Mixtec Word Prosody

    NASA Astrophysics Data System (ADS)

    Carroll, Lucien Serapio

    This dissertation presents a phonological description and acoustic analysis of the word prosody of Ixpantepec Nieves Mixtec, which involves both a complex tone system and a default stress system. The analysis of Nieves Mixtec word prosody is complicated by a close association between morphological structure and prosodic structure, and by the interactions between word prosody and phonation type, which has both contrastive and non-contrastive roles in the phonology. I contextualize these systems within the phonology of Nieves Mixtec as a whole, within the literature on other Mixtec varieties, and within the literature on cross-linguistic prosodic typology. The literature on prosodic typology indicates that stress is necessarily defined abstractly, as structured prominence realized differently in each language. Descriptions of stress in other Mixtec varieties widely report default stress on the initial syllable of the canonical bimoraic root, though some descriptions suggest final stress or mobile stress. I first present phonological evidence---from distributional restrictions, phonological processes, and loanword adaptation---that Nieves Mixtec word prosody does involve a stress system, based on trochaic feet aligned to the root. I then present an acoustic study comparing stressed syllables to unstressed syllables, for ten potential acoustic correlates of stress. The results indicate that the acoustic correlates of stress in Nieves Mixtec include segmental duration, intensity and periodicity. Building on analyses of other Mixtec tone systems, I show that the distribution of tone and the tone processes in Nieves Mixtec support an analysis in which morae may bear H, M or L tone, where M tone is underlyingly unspecified, and each morpheme may sponsor a final +H or +L floating tone. Bimoraic roots thus host up to two linked tones and one floating tone, while monomoraic clitics host just one linked tone and one floating tone, and tonal morphemes are limited to a single floating tone. I then present three studies describing the acoustic realization of tone and comparing the realization of tone in different prosodic types. The findings of these studies include a strong directional asymmetry in tonal coarticulation, increased duration at the word or phrase boundary, phonation differences among the tone categories, and F0 differences between the glottalization categories.

  16. Investigating the Retention and Time-Course of Phonotactic Constraint Learning From Production Experience

    PubMed Central

    Warker, Jill A.

    2013-01-01

    Adults can rapidly learn artificial phonotactic constraints such as /f/ only occurs at the beginning of syllables by producing syllables that contain those constraints. This implicit learning is then reflected in their speech errors. However, second-order constraints in which the placement of a phoneme depends on another characteristic of the syllable (e.g., if the vowel is /æ/, /f/ occurs at the beginning of syllables and /s/ occurs at the end of syllables but if the vowel is /I/, the reverse is true) require a longer learning period. Two experiments question the transience of second-order learning and whether consolidation plays a role in learning phonological dependencies. Using speech errors as a measure of learning, Experiment 1 investigated the durability of learning, and Experiment 2 investigated the time-course of learning. Experiment 1 found that learning is still present in speech errors a week later. Experiment 2 looked at whether more time in the form of a consolidation period or more experience in the form of more trials was necessary for learning to be revealed in speech errors. Both consolidation and more trials led to learning; however, consolidation provided a more substantial benefit. PMID:22686839

  17. The effect of deep brain stimulation on the speech motor system.

    PubMed

    Mücke, Doris; Becker, Johannes; Barbe, Michael T; Meister, Ingo; Liebhart, Lena; Roettger, Timo B; Dembek, Till; Timmermann, Lars; Grice, Martine

    2014-08-01

    Chronic deep brain stimulation of the nucleus ventralis intermedius is an effective treatment for individuals with medication-resistant essential tremor. However, these individuals report that stimulation has a deleterious effect on their speech. The present study investigates one important factor leading to these effects: the coordination of oral and glottal articulation. Sixteen native-speaking German adults with essential tremor, between 26 and 86 years old, with and without chronic deep brain stimulation of the nucleus ventralis intermedius and 12 healthy, age-matched subjects were recorded performing a fast syllable repetition task (/papapa/, /tatata/, /kakaka/). Syllable duration and voicing-to-syllable ratio as well as parameters related directly to consonant production, voicing during constriction, and frication during constriction were measured. Voicing during constriction was greater in subjects with essential tremor than in controls, indicating a perseveration of voicing into the voiceless consonant. Stimulation led to fewer voiceless intervals (voicing-to-syllable ratio), indicating a reduced degree of glottal abduction during the entire syllable cycle. Stimulation also induced incomplete oral closures (frication during constriction), indicating imprecise oral articulation. The detrimental effect of stimulation on the speech motor system can be quantified using acoustic measures at the subsyllabic level.

  18. Design of Phoneme MIDI Codes Using the MIDI Encoding Tool “Auto-F” and Realizing Voice Synthesizing Functions Based on Musical Sounds

    NASA Astrophysics Data System (ADS)

    Modegi, Toshio

    Using our previously developed audio to MIDI code converter tool “Auto-F”, from given vocal acoustic signals we can create MIDI data, which enable to playback the voice-like signals with a standard MIDI synthesizer. Applying this tool, we are constructing a MIDI database, which consists of previously converted simple harmonic structured MIDI codes from a set of 71 Japanese male and female syllable recorded signals. And we are developing a novel voice synthesizing system based on harmonically synthesizing musical sounds, which can generate MIDI data and playback voice signals with a MIDI synthesizer by giving Japanese plain (kana) texts, referring to the syllable MIDI code database. In this paper, we propose an improved MIDI converter tool, which can produce temporally higher-resolution MIDI codes. Then we propose an algorithm separating a set of 20 consonant and vowel phoneme MIDI codes from 71 syllable MIDI converted codes in order to construct a voice synthesizing system. And, we present the evaluation results of voice synthesizing quality between these separated phoneme MIDI codes and their original syllable MIDI codes by our developed 4-syllable word listening tests.

  19. Selective Influences of Precision and Power Grips on Speech Categorization.

    PubMed

    Tiainen, Mikko; Tiippana, Kaisa; Vainio, Martti; Peromaa, Tarja; Komeilipoor, Naeem; Vainio, Lari

    2016-01-01

    Recent studies have shown that articulatory gestures are systematically associated with specific manual grip actions. Here we show that executing such actions can influence performance on a speech-categorization task. Participants watched and/or listened to speech stimuli while executing either a power or a precision grip. Grip performance influenced the syllable categorization by increasing the proportion of responses of the syllable congruent with the executed grip (power grip-[ke] and precision grip-[te]). Two follow-up experiments indicated that the effect was based on action-induced bias in selecting the syllable.

  20. Articulation in schoolchildren and adults with neurofibromatosis type 1.

    PubMed

    Cosyns, Marjan; Mortier, Geert; Janssens, Sandra; Bogaert, Famke; D'Hondt, Stephanie; Van Borsel, John

    2012-01-01

    Several authors mentioned the occurrence of articulation problems in the neurofibromatosis type 1 (NF1) population. However, few studies have undertaken a detailed analysis of the articulation skills of NF1 patients, especially in schoolchildren and adults. Therefore, the aim of the present study was to examine in depth the articulation skills of NF1 schoolchildren and adults, both phonetically and phonologically. Speech samples were collected from 43 Flemish NF1 patients (14 children and 29 adults), ranging in age between 7 and 53 years, using a standardized speech test in which all Flemish single speech sounds and most clusters occur in all their permissible syllable positions. Analyses concentrated on consonants only and included a phonetic inventory, a phonetic, and a phonological analysis. It was shown that phonetic inventories were incomplete in 16.28% (7/43) of participants, in which totally correct realizations of the sibilants /ʃ/ and/or /ʒ/ were missing. Phonetic analysis revealed that distortions were the predominant phonetic error type. Sigmatismus stridens, multiple ad- or interdentality, and, in children, rhotacismus non vibrans were frequently observed. From a phonological perspective, the most common error types were substitution and syllable structure errors. Particularly, devoicing, cluster simplification, and, in children, deletion of the final consonant of words were perceived. Further, it was demonstrated that significantly more men than women presented with an incomplete phonetic inventory, and that girls tended to display more articulation errors than boys. Additionally, children exhibited significantly more articulation errors than adults, suggesting that although the articulation skills of NF1 patients evolve positively with age, articulation problems do not resolve completely from childhood to adulthood. As such, the articulation errors made by NF1 adults may be regarded as residual articulation disorders. It can be concluded that the speech of NF1 patients is characterized by mild articulation disorders at an age where this is no longer expected. Readers will be able to describe neurofibromatosis type 1 (NF1) and explain the articulation errors displayed by schoolchildren and adults with this genetic syndrome. © 2011 Elsevier Inc. All rights reserved.

  1. Rhythmic patterning in Malaysian and Singapore English.

    PubMed

    Tan, Rachel Siew Kuang; Low, Ee-Ling

    2014-06-01

    Previous work on the rhythm of Malaysian English has been based on impressionistic observations. This paper utilizes acoustic analysis to measure the rhythmic patterns of Malaysian English. Recordings of the read speech and spontaneous speech of 10 Malaysian English speakers were analyzed and compared with recordings of an equivalent sample of Singaporean English speakers. Analysis was done using two rhythmic indexes, the PVI and VarcoV. It was found that although the rhythm of read speech of the Singaporean speakers was syllable-based as described by previous studies, the rhythm of the Malaysian speakers was even more syllable-based. Analysis of the syllables in specific utterances showed that Malaysian speakers did not reduce vowels as much as Singaporean speakers in cases of syllables in utterances. Results of the spontaneous speech confirmed the findings for the read speech; that is, the same rhythmic patterning was found which normally triggers vowel reductions.

  2. Positive modulation of a neutral declarative memory by a threatening social event.

    PubMed

    Fernández, Rodrigo S; Bavassi, Luz; Campos, Jorge; Allegri, Ricardo F; Molina, Victor A; Forcato, Cecilia; Pedreira, María E

    2015-12-01

    Memories can be altered by negative or arousing experiences due to the activation of the stress-responsive sympatho-adrenal-medullary axis (SYM). Here, we used a neutral declarative memory that was acquired during multi-trial training to determine the effect of a threatening event on memory without emotional valence. To this end, participants received a new threatening social protocol before learning pairs of meaningless syllables and were tested either 15 min, 2 days or 8 days after acquisition. We first demonstrated that this threatening social situation activates not only the SYM axis (Experiment 1) and the hypothalamus-pituitary-adrenal axis (HPA; Experiment 2), but also, it improves the acquisition or early consolidation of the syllable pairs (Experiment 3). This improvement is not a transient effect; it can be observed after the memory is consolidated. Furthermore, this modulation increases the persistence of memory (Experiment 4). Thus, it is possible to affect memories with specific events that contain unrelated content and a different valence. Copyright © 2015 Elsevier Inc. All rights reserved.

  3. Early sound patterns in the speech of two Brazilian Portuguese speakers.

    PubMed

    Teixeira, Elizabeth Reis; Davis, Barbara L

    2002-06-01

    Sound patterns in the speech of two Brazilian-Portuguese speaking children are compared with early production patterns in English-learning children as well as English and Brazilian-Portuguese (BP) characteristics. The relationship between production system effects and ambient language influences in the acquisition of early sound patterns is of primary interest, as English and BP are characterized by differing phonological systems. Results emphasize the primacy of production system effects in early acquisition, although even the earliest word forms show evidence of perceptual effects from the ambient language in both BP children. Use of labials and coronals and low and midfront vowels in simple syllable shapes is consistent with acquisition data for this period across languages. However, potential ambient language influences include higher frequencies of dorsals, use of multisyllabic words, and different phone types in syllable-offset position. These results suggest that to fully understand early acquisition of sound systems one must account for both production system effects and perceptual effects from the ambient language.

  4. Perceptual, durational and tongue displacement measures following articulation therapy for rhotic sound errors.

    PubMed

    Bressmann, Tim; Harper, Susan; Zhylich, Irina; Kulkarni, Gajanan V

    2016-01-01

    Outcomes of articulation therapy for rhotic errors are usually assessed perceptually. However, our understanding of associated changes of tongue movement is limited. This study described perceptual, durational and tongue displacement changes over 10 sessions of articulation therapy for /ɹ/ in six children. Four of the participants also received ultrasound biofeedback of their tongue shape. Speech and tongue movement were recorded pre-therapy, after 5 sessions, in the final session and at a one month follow-up. Perceptually, listeners perceived improvement and classified more productions as /ɹ/ in the final and follow-up assessments. The durations of VɹV syllables at the midway point of the therapy were longer. Cumulative tongue displacement increased in the final session. The average standard deviation was significantly higher in the middle and final assessments. The duration and tongue displacement measures illustrated how articulation therapy affected tongue movement and may be useful for outcomes research about articulation therapy.

  5. Dissociable Effects on Birdsong of Androgen Signaling in Cortex-Like Brain Regions of Canaries

    PubMed Central

    2017-01-01

    The neural basis of how learned vocalizations change during development and in adulthood represents a major challenge facing cognitive neuroscience. This plasticity in the degree to which learned vocalizations can change in both humans and songbirds is linked to the actions of sex steroid hormones during ontogeny but also in adulthood in the context of seasonal changes in birdsong. We investigated the role of steroid hormone signaling in the brain on distinct features of birdsong using adult male canaries (Serinus canaria), which show extensive seasonal vocal plasticity as adults. Specifically, we bilaterally implanted the potent androgen receptor antagonist flutamide in two key brain regions that control birdsong. We show that androgen signaling in the motor cortical-like brain region, the robust nucleus of the arcopallium (RA), controls syllable and trill bandwidth stereotypy, while not significantly affecting higher order features of song such syllable-type usage (i.e., how many times each syllable type is used) or syllable sequences. In contrast, androgen signaling in the premotor cortical-like brain region, HVC (proper name), controls song variability by increasing the variability of syllable-type usage and syllable sequences, while having no effect on syllable or trill bandwidth stereotypy. Other aspects of song, such as the duration of trills and the number of syllables per song, were also differentially affected by androgen signaling in HVC versus RA. These results implicate androgens in regulating distinct features of complex motor output in a precise and nonredundant manner. SIGNIFICANCE STATEMENT Vocal plasticity is linked to the actions of sex steroid hormones, but the precise mechanisms are unclear. We investigated this question in adult male canaries (Serinus canaria), which show extensive vocal plasticity throughout their life. We show that androgens in two cortex-like vocal control brain regions regulate distinct aspects of vocal plasticity. For example, in HVC (proper name), androgens regulate variability in syntax but not phonology, whereas androgens in the robust nucleus of the arcopallium (RA) regulate variability in phonology but not syntax. Temporal aspects of song were also differentially affected by androgen signaling in HVC versus RA. Thus, androgen signaling may reduce vocal plasticity by acting in a nonredundant and precise manner in the brain. PMID:28821656

  6. Comparison of spectrographic records of two syllables pronounced from scripts in hiragana and romaji by students with different familiarity with English.

    PubMed

    Ototake, Harumi; Yamada, Jun

    2005-10-01

    The same syllables /mu/ and /ra/ written in Japanese hiragana and romaji given on a standard speeded naming task elicited phonetically or acoustically different responses in a syllabic hiragana condition and a romaji condition. The participants were two groups of Japanese college students (ns = 15 and 16) with different familiarity with English as a second language. The results suggested that the phonetic reality of syllables represented in these scripts can differ, depending on the interaction between the kind of script and speakers' orthographic familiarity.

  7. Social Vocalizations of Big Brown Bats Vary with Behavioral Context

    PubMed Central

    Gadziola, Marie A.; Grimsley, Jasmine M. S.; Faure, Paul A.; Wenstrup, Jeffrey J.

    2012-01-01

    Bats are among the most gregarious and vocal mammals, with some species demonstrating a diverse repertoire of syllables under a variety of behavioral contexts. Despite extensive characterization of big brown bat (Eptesicus fuscus) biosonar signals, there have been no detailed studies of adult social vocalizations. We recorded and analyzed social vocalizations and associated behaviors of captive big brown bats under four behavioral contexts: low aggression, medium aggression, high aggression, and appeasement. Even limited to these contexts, big brown bats possess a rich repertoire of social vocalizations, with 18 distinct syllable types automatically classified using a spectrogram cross-correlation procedure. For each behavioral context, we describe vocalizations in terms of syllable acoustics, temporal emission patterns, and typical syllable sequences. Emotion-related acoustic cues are evident within the call structure by context-specific syllable types or variations in the temporal emission pattern. We designed a paradigm that could evoke aggressive vocalizations while monitoring heart rate as an objective measure of internal physiological state. Changes in the magnitude and duration of elevated heart rate scaled to the level of evoked aggression, confirming the behavioral state classifications assessed by vocalizations and behavioral displays. These results reveal a complex acoustic communication system among big brown bats in which acoustic cues and call structure signal the emotional state of a caller. PMID:22970247

  8. Exploring the role of hand gestures in learning novel phoneme contrasts and vocabulary in a second language

    PubMed Central

    Kelly, Spencer D.; Hirata, Yukari; Manansala, Michael; Huang, Jessica

    2014-01-01

    Co-speech hand gestures are a type of multimodal input that has received relatively little attention in the context of second language learning. The present study explored the role that observing and producing different types of gestures plays in learning novel speech sounds and word meanings in an L2. Naïve English-speakers were taught two components of Japanese—novel phonemic vowel length contrasts and vocabulary items comprised of those contrasts—in one of four different gesture conditions: Syllable Observe, Syllable Produce, Mora Observe, and Mora Produce. Half of the gestures conveyed intuitive information about syllable structure, and the other half, unintuitive information about Japanese mora structure. Within each Syllable and Mora condition, half of the participants only observed the gestures that accompanied speech during training, and the other half also produced the gestures that they observed along with the speech. The main finding was that participants across all four conditions had similar outcomes in two different types of auditory identification tasks and a vocabulary test. The results suggest that hand gestures may not be well suited for learning novel phonetic distinctions at the syllable level within a word, and thus, gesture-speech integration may break down at the lowest levels of language processing and learning. PMID:25071646

  9. Attention Is Required for Knowledge-Based Sequential Grouping: Insights from the Integration of Syllables into Words.

    PubMed

    Ding, Nai; Pan, Xunyi; Luo, Cheng; Su, Naifei; Zhang, Wen; Zhang, Jianfeng

    2018-01-31

    How the brain groups sequential sensory events into chunks is a fundamental question in cognitive neuroscience. This study investigates whether top-down attention or specific tasks are required for the brain to apply lexical knowledge to group syllables into words. Neural responses tracking the syllabic and word rhythms of a rhythmic speech sequence were concurrently monitored using electroencephalography (EEG). The participants performed different tasks, attending to either the rhythmic speech sequence or a distractor, which was another speech stream or a nonlinguistic auditory/visual stimulus. Attention to speech, but not a lexical-meaning-related task, was required for reliable neural tracking of words, even when the distractor was a nonlinguistic stimulus presented cross-modally. Neural tracking of syllables, however, was reliably observed in all tested conditions. These results strongly suggest that neural encoding of individual auditory events (i.e., syllables) is automatic, while knowledge-based construction of temporal chunks (i.e., words) crucially relies on top-down attention. SIGNIFICANCE STATEMENT Why we cannot understand speech when not paying attention is an old question in psychology and cognitive neuroscience. Speech processing is a complex process that involves multiple stages, e.g., hearing and analyzing the speech sound, recognizing words, and combining words into phrases and sentences. The current study investigates which speech-processing stage is blocked when we do not listen carefully. We show that the brain can reliably encode syllables, basic units of speech sounds, even when we do not pay attention. Nevertheless, when distracted, the brain cannot group syllables into multisyllabic words, which are basic units for speech meaning. Therefore, the process of converting speech sound into meaning crucially relies on attention. Copyright © 2018 the authors 0270-6474/18/381178-11$15.00/0.

  10. [Developmental changes in reading ability of Japanese elementary school children--analysis of 4 kana reading tasks].

    PubMed

    Kobayashi, Tomoka; Inagaki, Masumi; Gunji, Atsuko; Yatabe, Kiyomi; Kaga, Makiko; Goto, Takaaki; Koike, Toshihide; Wakamiya, Eiji; Koeda, Tatsuya

    2010-01-01

    Five hundred and twenty-eight Japanese elementary school children aged from 6 (Grade 1) to 12 (Grade 6) were tested for their abilities to read Hiragana characters, words, and short sentences. They were typically developing children whom the classroom teachers judged to have no problems with reading and writing in Japanese. Each child was asked to read four tasks which were written in Hiragana script: single mora reading task, four syllable non-word reading task, four syllable word reading task, and short sentence reading task. The total articulation time for reading and performance in terms of accuracy were measured for each task. Developmental changes in these variables were evaluated. The articulation time was significantly longer for the first graders, and it gradually shortened as they moved through to the upper grades in all tasks. The articulation time reached a plateau in the 4th grade for the four syllable word and short sentence reading tasks, while it did so for the single mora and four syllable non-word reading tasks in the 5th grade. The articulation times for the four syllable word and short sentence reading tasks correlated strongly. There were very few clear errors for all tasks, and the number of such errors significantly changed between the school grades only for the single mora and four syllable word reading tasks. It was noted that more than half of the children read the beginning portion of the word or phrase twice or more, in order to read it accurately, and developmental changes were also seen in this pattern of reading. This study revealed that the combination of these reading tasks may function as a screening test for reading disorders such as developmental dyslexia in children below the age of ten or eleven years old.

  11. Prevalence of phonological disorders and phonological processes in typical and atypical phonological development.

    PubMed

    Ceron, Marizete Ilha; Gubiani, Marileda Barichello; Oliveira, Camila Rosa de; Gubiani, Marieli Barichello; Keske-Soares, Márcia

    2017-05-08

    To determine the occurrence of phonological disorders by age, gender and school type, and analyze the phonological processes observed in typical and atypical phonological development across different age groups. The sample consisted of 866 children aged between 3:0 and 8:11 years, recruited from public and private schools in the city of Santa Maria/RS. A phonological evaluation was performed to analyze the operative phonological processes. 15.26% (n = 132) of the sample presented atypical phonological acquisition (phonological disorders). Phonological impairments were more frequent in public school students across all age groups. Phonological alterations were most frequent between ages 4 -to 6, and more prevalent in males than females in all but the youngest age group. The most common phonological processes in typical phonological acquisition were: cluster reduction; nonlateral liquid deletion in coda; nonlateral liquid substitution in onset; semivocalization of lateral liquids in coda; and unstressed syllable deletion. In children with phonological disorders, the most common phonological processes were: lateral and nonlateral liquid substitution in onset position; nonlateral liquid deletion; fronting of fricatives in onset position; unstressed syllable deletion; semivocalization of nonlateral liquid in coda; and nonlateral liquid deletion in coda position. Phonological processes were highly prevalent in the present sample, and occurred more often in boys than in girls. Information regarding the type and frequency of phonological processes in both typical phonological acquisition and phonological disorders may contribute to early diagnosis and increase the efficiency of treatment planning.

  12. Syllable Transposition Effects in Korean Word Recognition

    ERIC Educational Resources Information Center

    Lee, Chang H.; Kwon, Youan; Kim, Kyungil; Rastle, Kathleen

    2015-01-01

    Research on the impact of letter transpositions in visual word recognition has yielded important clues about the nature of orthographic representations. This study investigated the impact of syllable transpositions on the recognition of Korean multisyllabic words. Results showed that rejection latencies in visual lexical decision for…

  13. Oral and Hand Movement Speeds are Associated with Expressive Language Ability in Children with Speech Sound Disorder

    PubMed Central

    Peter, Beate

    2013-01-01

    This study tested the hypothesis that children with speech sound disorder have generalized slowed motor speeds. It evaluated associations among oral and hand motor speeds and measures of speech (articulation and phonology) and language (receptive vocabulary, sentence comprehension, sentence imitation), in 11 children with moderate to severe SSD and 11 controls. Syllable durations from a syllable repetition task served as an estimate of maximal oral movement speed. In two imitation tasks, nonwords and clapped rhythms, unstressed vowel durations and quarter-note clap intervals served as estimates of oral and hand movement speed, respectively. Syllable durations were significantly correlated with vowel durations and hand clap intervals. Sentence imitation was correlated with all three timed movement measures. Clustering on syllable repetition durations produced three clusters that also differed in sentence imitation scores. Results are consistent with limited movement speeds across motor systems and SSD subtypes defined by motor speeds as a corollary of expressive language abilities. PMID:22411590

  14. Song convergence in multiple urban populations of silvereyes (Zosterops lateralis)

    PubMed Central

    Potvin, Dominique A; Parris, Kirsten M

    2012-01-01

    Recent studies have revealed differences between urban and rural vocalizations of numerous bird species. These differences include frequency shifts, amplitude shifts, altered song speed, and selective meme use. If particular memes sung by urban populations are adapted to the urban soundscape, “urban-typical” calls, memes, or repertoires should be consistently used in multiple urban populations of the same species, regardless of geographic location. We tested whether songs or contact calls of silvereyes (Zosterops lateralis) might be subject to such convergent cultural evolution by comparing syllable repertoires of geographically dispersed urban and rural population pairs throughout southeastern Australia. Despite frequency and tempo differences between urban and rural calls, call repertoires were similar between habitat types. However, certain song syllables were used more frequently by birds from urban than rural populations. Partial redundancy analysis revealed that both geographic location and habitat characteristics were important predictors of syllable repertoire composition. These findings suggest convergent cultural evolution: urban populations modify both song and call syllables from their local repertoire in response to noise. PMID:22957198

  15. Song convergence in multiple urban populations of silvereyes (Zosterops lateralis).

    PubMed

    Potvin, Dominique A; Parris, Kirsten M

    2012-08-01

    Recent studies have revealed differences between urban and rural vocalizations of numerous bird species. These differences include frequency shifts, amplitude shifts, altered song speed, and selective meme use. If particular memes sung by urban populations are adapted to the urban soundscape, "urban-typical" calls, memes, or repertoires should be consistently used in multiple urban populations of the same species, regardless of geographic location. We tested whether songs or contact calls of silvereyes (Zosterops lateralis) might be subject to such convergent cultural evolution by comparing syllable repertoires of geographically dispersed urban and rural population pairs throughout southeastern Australia. Despite frequency and tempo differences between urban and rural calls, call repertoires were similar between habitat types. However, certain song syllables were used more frequently by birds from urban than rural populations. Partial redundancy analysis revealed that both geographic location and habitat characteristics were important predictors of syllable repertoire composition. These findings suggest convergent cultural evolution: urban populations modify both song and call syllables from their local repertoire in response to noise.

  16. The role of syllables in sign language production.

    PubMed

    Baus, Cristina; Gutiérrez, Eva; Carreiras, Manuel

    2014-01-01

    The aim of the present study was to investigate the functional role of syllables in sign language and how the different phonological combinations influence sign production. Moreover, the influence of age of acquisition was evaluated. Deaf signers (native and non-native) of Catalan Signed Language (LSC) were asked in a picture-sign interference task to sign picture names while ignoring distractor-signs with which they shared two phonological parameters (out of three of the main sign parameters: Location, Movement, and Handshape). The results revealed a different impact of the three phonological combinations. While no effect was observed for the phonological combination Handshape-Location, the combination Handshape-Movement slowed down signing latencies, but only in the non-native group. A facilitatory effect was observed for both groups when pictures and distractors shared Location-Movement. Importantly, linguistic models have considered this phonological combination to be a privileged unit in the composition of signs, as syllables are in spoken languages. Thus, our results support the functional role of syllable units during phonological articulation in sign language production.

  17. Oral and hand movement speeds are associated with expressive language ability in children with speech sound disorder.

    PubMed

    Peter, Beate

    2012-12-01

    This study tested the hypothesis that children with speech sound disorder have generalized slowed motor speeds. It evaluated associations among oral and hand motor speeds and measures of speech (articulation and phonology) and language (receptive vocabulary, sentence comprehension, sentence imitation), in 11 children with moderate to severe SSD and 11 controls. Syllable durations from a syllable repetition task served as an estimate of maximal oral movement speed. In two imitation tasks, nonwords and clapped rhythms, unstressed vowel durations and quarter-note clap intervals served as estimates of oral and hand movement speed, respectively. Syllable durations were significantly correlated with vowel durations and hand clap intervals. Sentence imitation was correlated with all three timed movement measures. Clustering on syllable repetition durations produced three clusters that also differed in sentence imitation scores. Results are consistent with limited movement speeds across motor systems and SSD subtypes defined by motor speeds as a corollary of expressive language abilities.

  18. Electrophysiological evidence of statistical learning of long-distance dependencies in 8-month-old preterm and full-term infants.

    PubMed

    Kabdebon, C; Pena, M; Buiatti, M; Dehaene-Lambertz, G

    2015-09-01

    Using electroencephalography, we examined 8-month-old infants' ability to discover a systematic dependency between the first and third syllables of successive words, concatenated into a monotonous speech stream, and to subsequently generalize this regularity to new items presented in isolation. Full-term and preterm infants, while exposed to the stream, displayed a significant entrainment (phase-locking) to the syllabic and word frequencies, demonstrating that they were sensitive to the word unit. The acquisition of the systematic dependency defining words was confirmed by the significantly different neural responses to rule-words and part-words subsequently presented during the test phase. Finally, we observed a correlation between syllabic entrainment during learning and the difference in phase coherence between the test conditions (rule-words vs part-words) suggesting that temporal processing of the syllable unit might be crucial in linguistic learning. No group difference was observed suggesting that non-adjacent statistical computations are already robust at 8 months, even in preterm infants, and thus develop during the first year of life, earlier than expected from behavioral studies. Copyright © 2015 Elsevier Inc. All rights reserved.

  19. Investigating speech perception in children with dyslexia: is there evidence of a consistent deficit in individuals?

    PubMed Central

    Messaoud-Galusi, Souhila; Hazan, Valerie; Rosen, Stuart

    2012-01-01

    Purpose The claim that speech perception abilities are impaired in dyslexia was investigated in a group of 62 dyslexic children and 51 average readers matched in age. Method To test whether there was robust evidence of speech perception deficits in children with dyslexia, speech perception in noise and quiet was measured using eight different tasks involving the identification and discrimination of a complex and highly natural synthetic ‘pea’-‘bee’ contrast (copy synthesised from natural models) and the perception of naturally-produced words. Results Children with dyslexia, on average, performed more poorly than average readers in the synthetic syllables identification task in quiet and in across-category discrimination (but not when tested using an adaptive procedure). They did not differ from average readers on two tasks of word recognition in noise or identification of synthetic syllables in noise. For all tasks, a majority of individual children with dyslexia performed within norms. Finally, speech perception generally did not correlate with pseudo-word reading or phonological processing, the core skills related to dyslexia. Conclusions On the tasks and speech stimuli we used, most children with dyslexia do not appear to show a consistent deficit in speech perception. PMID:21930615

  20. The phonological abilities of Cantonese-speaking children with hearing loss.

    PubMed

    Dodd, B J; So, L K

    1994-06-01

    Little is known about the acquisition of phonology by children with hearing loss who learn languages other than English. In this study, the phonological abilities of 12 Cantonese-speaking children (ages 4:2 to 6:11) with prelingual hearing impairment are described. All but 3 children had almost complete syllable-initial consonant repertoires; all but 2 had complete syllable-final consonant and vowel repertoires; and only 1 child failed to produce all nine tones. Children's perception of single words was assessed using sets of words that included tone, consonant, and semantic distractors. Although the performance of the subjects was not age appropriate, they nevertheless most often chose the target, with most errors observed for the tone distractor. The phonological rules used included those that characterize the speech of younger hearing children acquiring Cantonese (e.g., cluster reduction, stopping, and deaspiration). However, most children also used at least one unusual phonological rule (e.g., frication, addition, initial consonant deletion, and/or backing). These rules are common in the speech of Cantonese-speaking children diagnosed as phonologically disordered. The influence of the ambient language on children's patterns of phonological errors is discussed.

  1. The Separate and Cumulative Effects of TBI and PTSD on Cognitive Function and Emotional Control

    DTIC Science & Technology

    2012-04-01

    indicate an altered profile of persistent hyper- arousal , exaggerated startle responses (Fani et al., 2012; Pole, 2007), larger eye-blink, eye pupil...were each compared. This separation kept variables such as word frequency, valence, arousal , and other properties as consistent as possible across...number of syllables and frequency. Only high arousal Negative and Positive words were used and arousal and valence ratings for Neutral, Negative and

  2. Consciousness-Raising and Prepositions

    ERIC Educational Resources Information Center

    Hendricks, Monica

    2010-01-01

    For a variety of reasons, learning English prepositions is notoriously difficult and a slow, gradual process for English as a Second Language (ESL) students. To begin, English prepositions typically are short, single-syllable or two-syllable words that are seldom stressed when speaking and therefore often not articulated clearly or heard…

  3. Effect of attentional load on audiovisual speech perception: evidence from ERPs.

    PubMed

    Alsius, Agnès; Möttönen, Riikka; Sams, Mikko E; Soto-Faraco, Salvador; Tiippana, Kaisa

    2014-01-01

    Seeing articulatory movements influences perception of auditory speech. This is often reflected in a shortened latency of auditory event-related potentials (ERPs) generated in the auditory cortex. The present study addressed whether this early neural correlate of audiovisual interaction is modulated by attention. We recorded ERPs in 15 subjects while they were presented with auditory, visual, and audiovisual spoken syllables. Audiovisual stimuli consisted of incongruent auditory and visual components known to elicit a McGurk effect, i.e., a visually driven alteration in the auditory speech percept. In a Dual task condition, participants were asked to identify spoken syllables whilst monitoring a rapid visual stream of pictures for targets, i.e., they had to divide their attention. In a Single task condition, participants identified the syllables without any other tasks, i.e., they were asked to ignore the pictures and focus their attention fully on the spoken syllables. The McGurk effect was weaker in the Dual task than in the Single task condition, indicating an effect of attentional load on audiovisual speech perception. Early auditory ERP components, N1 and P2, peaked earlier to audiovisual stimuli than to auditory stimuli when attention was fully focused on syllables, indicating neurophysiological audiovisual interaction. This latency decrement was reduced when attention was loaded, suggesting that attention influences early neural processing of audiovisual speech. We conclude that reduced attention weakens the interaction between vision and audition in speech.

  4. Truncation Without Shape Constraints: The Latter Stages of Prosodic Acquisition.

    ERIC Educational Resources Information Center

    Kehoe, Margaret M.

    2000-01-01

    Evaluates the claim of uniform size and shape restrictions in prosodic development using a cross-sectional database of English-speaking children's multisyllabic word productions. Suggests children's increasing faithfulness to unstressed syllables can be explained by different constraint rankings that relate to edge alignment, syllable structure,…

  5. Intensity Accents in French 2 Year Olds' Speech.

    ERIC Educational Resources Information Center

    Allen, George D.

    The acoustic features and functions of accentuation in French are discussed, and features of accentuation in the speech of French 2-year-olds are explored. The four major acoustic features used to signal accentual distinctions are fundamental frequency of voicing, duration of segments and syllables, intensity of segments and syllables, and…

  6. Quantitative Investigations in Hungarian Phonotactics and Syllable Structure

    ERIC Educational Resources Information Center

    Grimes, Stephen M.

    2010-01-01

    This dissertation investigates statistical properties of segment collocation and syllable geometry of the Hungarian language. A corpus and dictionary based approach to studying language phonologies is outlined. In order to conduct research on Hungarian, a phonological lexicon was created by compiling existing dictionaries and corpora and using a…

  7. The Influence of Syllable Onset Complexity and Syllable Frequency on Speech Motor Control

    ERIC Educational Resources Information Center

    Riecker, Axel; Brendel, Bettina; Ziegler, Wolfram; Erb, Michael; Ackermann, Hermann

    2008-01-01

    Functional imaging studies have delineated a "minimal network for overt speech production," encompassing mesiofrontal structures (supplementary motor area, anterior cingulate gyrus), bilateral pre- and postcentral convolutions, extending rostrally into posterior parts of the inferior frontal gyrus (IFG) of the language-dominant hemisphere, left…

  8. Oral-Motor and Motor-Speech Characteristics of Children with Autism.

    ERIC Educational Resources Information Center

    Adams, Lynn

    1998-01-01

    This study compared the oral-motor and motor-speech characteristics of four young children with autism and four nonautistic children. Three tasks requiring oral motor movements, simple syllable productions, and complex syllable productions were utilized. Significant differences were found in scores on the oral-motor movements and the…

  9. A Vowel Is a Vowel: Generalizing Newly Learned Phonotactic Constraints to New Contexts

    ERIC Educational Resources Information Center

    Chambers, Kyle E.; Onishi, Kristine H.; Fisher, Cynthia

    2010-01-01

    Adults can learn novel phonotactic constraints from brief listening experience. We investigated the representations underlying phonotactic learning by testing generalization to syllables containing new vowels. Adults heard consonant-vowel-consonant study syllables in which particular consonants were artificially restricted to the onset or coda…

  10. Incidence of Dysarthria in Children with Cerebellar Tumors: A Prospective Study

    ERIC Educational Resources Information Center

    Richter, S.; Schoch, B.; Ozimek, A.; Gorissen, B.; Hein-Kropp, C.; Kaiser, O.; Hovel, M.; Wieland, R.; Gizewski, E.; Timmann, D.

    2005-01-01

    The present study investigated dysarthric symptoms in children with cerebellar tumors. Ten children with cerebellar tumors and 10 orthopedic control children were tested prior and one week after surgery. Clinical dysarthric symptoms were quantified in spontaneous speech. Syllable durations were analyzed in syllable repetition and sentence…

  11. Possible-Word Constraints in Cantonese Speech Segmentation

    ERIC Educational Resources Information Center

    Yip, Michael C. W.

    2004-01-01

    A Cantonese syllable-spotting experiment was conducted to examine whether the Possible-Word Constraint (PWC), proposed by Norris, McQueen, Cutler, and Butterfield (1997), can apply in Cantonese speech segmentation. In the experiment, listeners were asked to spot out the target Cantonese syllable from a series of nonsense sound strings. Results…

  12. ERP evidence for implicit L2 word stress knowledge in listeners of a fixed-stress language.

    PubMed

    Kóbor, Andrea; Honbolygó, Ferenc; Becker, Angelika B C; Schild, Ulrike; Csépe, Valéria; Friedrich, Claudia K

    2018-06-01

    Languages with contrastive stress, such as English or German, distinguish some words only via the stress status of their syllables, such as "CONtent" and "conTENT" (capitals indicate a stressed syllable). Listeners with a fixed-stress native language, such as Hungarian, have difficulties in explicitly discriminating variation of the stress position in a second language (L2). However, Event-Related Potentials (ERPs) indicate that Hungarian listeners implicitly notice variation from their native fixed-stress pattern. Here we used ERPs to investigate Hungarian listeners' implicit L2 processing. In a cross-modal word fragment priming experiment, we presented spoken stressed and unstressed German word onsets (primes) followed by printed versions of initially stressed and initially unstressed German words (targets). ERPs reflected stress priming exerted by both prime types. This indicates that Hungarian listeners implicitly linked German words with the stress status of the primes. Thus, the formerly described explicit stress discrimination difficulty associated with a fixed-stress native language does not generalize to implicit aspects of L2 word stress processing. Copyright © 2018 Elsevier B.V. All rights reserved.

  13. Learning metathesis: Evidence for syllable structure constraints.

    PubMed

    Finley, Sara

    2017-02-01

    One of the major questions in the cognitive science of language is whether the perceptual and phonological motivations for the rules and patterns that govern the sounds of language are a part of the psychological reality of grammatical representations. This question is particularly important in the study of phonological patterns - systematic constraints on the representation of sounds, because phonological patterns tend to be grounded in phonetic constraints. This paper focuses on phonological metathesis, which occurs when two adjacent sounds switch positions (e.g., ca st pronounced as ca ts ). While many cases of phonological metathesis appear to be motivated by constraints on syllable structure, it is possible that these metathesis patterns are merely artifacts of historical change, and do not represent the linguistic knowledge of the speaker (Blevins & Garrett, 1998). Participants who were exposed to a metathesis pattern that can be explained in terms of structural or perceptual improvement were less likely to generalize to metathesis patterns that did not show the same improvements. These results support a substantively biased theory in which phonological patterns are encoded in terms of structurally motivated constraints.

  14. Learning metathesis: Evidence for syllable structure constraints

    PubMed Central

    Finley, Sara

    2016-01-01

    One of the major questions in the cognitive science of language is whether the perceptual and phonological motivations for the rules and patterns that govern the sounds of language are a part of the psychological reality of grammatical representations. This question is particularly important in the study of phonological patterns – systematic constraints on the representation of sounds, because phonological patterns tend to be grounded in phonetic constraints. This paper focuses on phonological metathesis, which occurs when two adjacent sounds switch positions (e.g., cast pronounced as cats). While many cases of phonological metathesis appear to be motivated by constraints on syllable structure, it is possible that these metathesis patterns are merely artifacts of historical change, and do not represent the linguistic knowledge of the speaker (Blevins & Garrett, 1998). Participants who were exposed to a metathesis pattern that can be explained in terms of structural or perceptual improvement were less likely to generalize to metathesis patterns that did not show the same improvements. These results support a substantively biased theory in which phonological patterns are encoded in terms of structurally motivated constraints. PMID:28082764

  15. A Foxp2 Mutation Implicated in Human Speech Deficits Alters Sequencing of Ultrasonic Vocalizations in Adult Male Mice.

    PubMed

    Chabout, Jonathan; Sarkar, Abhra; Patel, Sheel R; Radden, Taylor; Dunson, David B; Fisher, Simon E; Jarvis, Erich D

    2016-01-01

    Development of proficient spoken language skills is disrupted by mutations of the FOXP2 transcription factor. A heterozygous missense mutation in the KE family causes speech apraxia, involving difficulty producing words with complex learned sequences of syllables. Manipulations in songbirds have helped to elucidate the role of this gene in vocal learning, but findings in non-human mammals have been limited or inconclusive. Here, we performed a systematic study of ultrasonic vocalizations (USVs) of adult male mice carrying the KE family mutation. Using novel statistical tools, we found that Foxp2 heterozygous mice did not have detectable changes in USV syllable acoustic structure, but produced shorter sequences and did not shift to more complex syntax in social contexts where wildtype animals did. Heterozygous mice also displayed a shift in the position of their rudimentary laryngeal motor cortex (LMC) layer-5 neurons. Our findings indicate that although mouse USVs are mostly innate, the underlying contributions of FoxP2 to sequencing of vocalizations are conserved with humans.

  16. A Foxp2 Mutation Implicated in Human Speech Deficits Alters Sequencing of Ultrasonic Vocalizations in Adult Male Mice

    PubMed Central

    Chabout, Jonathan; Sarkar, Abhra; Patel, Sheel R.; Radden, Taylor; Dunson, David B.; Fisher, Simon E.; Jarvis, Erich D.

    2016-01-01

    Development of proficient spoken language skills is disrupted by mutations of the FOXP2 transcription factor. A heterozygous missense mutation in the KE family causes speech apraxia, involving difficulty producing words with complex learned sequences of syllables. Manipulations in songbirds have helped to elucidate the role of this gene in vocal learning, but findings in non-human mammals have been limited or inconclusive. Here, we performed a systematic study of ultrasonic vocalizations (USVs) of adult male mice carrying the KE family mutation. Using novel statistical tools, we found that Foxp2 heterozygous mice did not have detectable changes in USV syllable acoustic structure, but produced shorter sequences and did not shift to more complex syntax in social contexts where wildtype animals did. Heterozygous mice also displayed a shift in the position of their rudimentary laryngeal motor cortex (LMC) layer-5 neurons. Our findings indicate that although mouse USVs are mostly innate, the underlying contributions of FoxP2 to sequencing of vocalizations are conserved with humans. PMID:27812326

  17. L2 Spelling Errors in Italian Children with Dyslexia.

    PubMed

    Palladino, Paola; Cismondo, Dhebora; Ferrari, Marcella; Ballagamba, Isabella; Cornoldi, Cesare

    2016-05-01

    The present study aimed to investigate L2 spelling skills in Italian children by administering an English word dictation task to 13 children with dyslexia (CD), 13 control children (comparable in age, gender, schooling and IQ) and a group of 10 children with an English learning difficulty, but no L1 learning disorder. Patterns of difficulties were examined for accuracy and type of errors, in spelling dictated short and long words (i.e. disyllables and three syllables). Notably, CD were poor in spelling English words. Furthermore, their errors were mainly related with phonological representation of words, as they made more 'phonologically' implausible errors than controls. In addition, CD errors were more frequent for short than long words. Conversely, the three groups did not differ in the number of plausible ('non-phonological') errors, that is, words that were incorrectly written, but whose reading could correspond to the dictated word via either Italian or English rules. Error analysis also showed syllable position differences in the spelling patterns of CD, children with and English learning difficulty and control children. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.

  18. Neural Representations Used by Brain Regions Underlying Speech Production

    ERIC Educational Resources Information Center

    Segawa, Jennifer Anne

    2013-01-01

    Speech utterances are phoneme sequences but may not always be represented as such in the brain. For instance, electropalatography evidence indicates that as speaking rate increases, gestures within syllables are manipulated separately but those within consonant clusters act as one motor unit. Moreover, speech error data suggest that a syllable's…

  19. ERP measures of syllable processing in 1 year olds: infant diet- and gender-related differences

    USDA-ARS?s Scientific Manuscript database

    Language skills are generally better in females than males, but the basis for these differences has not been determined. To investigate whether variations in infant diet contribute to these differences, cortical responses to the syllable /pa/ (ERPs;124 sites) were examined in healthy 12-month-old, f...

  20. Topography of Syllable Change-Detection Electrophysiological Indices in Children and Adults with Reading Disabilities

    ERIC Educational Resources Information Center

    Hommet, Caroline; Vidal, Julie; Roux, Sylvie; Blanc, Romuald; Barthez, Marie Anne; De Becque, Brigitte; Barthelemy, Catherine; Bruneau, Nicole; Gomot, Marie

    2009-01-01

    Introduction: Developmental dyslexia (DD) is a frequent language-based learning disorder. The predominant etiological view postulates that reading problems originate from a phonological impairment. Method: We studied mismatch negativity (MMN) and Late Discriminative Negativity (LDN) to syllables change in both children (n = 12; 8-12 years) and…

  1. A Cross-Language Study of Laryngeal-Oral Coordination across Varying Prosodic and Syllable-Structure Conditions

    ERIC Educational Resources Information Center

    Hoole, Philip; Bombien, Lasse

    2017-01-01

    Purpose: The purpose of this study is to use prosodic and syllable-structure variation to probe the underlying representation of laryngeal kinematics in languages traditionally considered to differ in voicing typology (German vs. Dutch and French). Method: Transillumination and videofiberendoscopic filming were used to investigate the devoicing…

  2. Prosodic Transfer: From Chinese Lexical Tone to English Pitch Accent

    ERIC Educational Resources Information Center

    Ploquin, Marie

    2013-01-01

    Chinese tones are associated with a syllable to convey meaning, English pitch accents are prominence markers associated with stressed syllables. As both are created by pitch modulation, their pitch contours can be quite similar. The experiment reported here examines whether native speakers of Chinese produce, when speaking English, the Chinese…

  3. Impaired Perception of Syllable Stress in Children with Dyslexia: A Longitudinal Study

    ERIC Educational Resources Information Center

    Goswami, Usha; Mead, Natasha; Fosker, Tim; Huss, Martina; Barnes, Lisa; Leong, Victoria

    2013-01-01

    Prosodic patterning is a key structural element of spoken language. However, the potential role of prosodic awareness in the phonological difficulties that characterise children with developmental dyslexia has been little studied. Here we report the first longitudinal study of sensitivity to syllable stress in children with dyslexia, enabling the…

  4. Clinician Percent Syllables Stuttered, Clinician Severity Ratings and Speaker Severity Ratings: Are They Interchangeable?

    ERIC Educational Resources Information Center

    Karimi, Hamid; Jones, Mark; O'Brian, Sue; Onslow, Mark

    2014-01-01

    Background: At present, percent syllables stuttered (%SS) is the gold standard outcome measure for behavioural stuttering treatment research. However, ordinal severity rating (SR) procedures have some inherent advantages over that method. Aims: To establish the relationship between Clinician %SS, Clinician SR and self-reported Speaker SR. To…

  5. Syllable Structure in Dysfunctional Portuguese Children's Speech

    ERIC Educational Resources Information Center

    Candeias, Sara; Perdigao, Fernando

    2010-01-01

    The goal of this work is to investigate whether children with speech dysfunctions (SD) show a deficit in planning some Portuguese syllable structures (PSS) in continuous speech production. Knowledge of which aspects of speech production are affected by SD is necessary for efficient improvement in the therapy techniques. The case-study is focused…

  6. A Syllable Segmentation, Letter-Sound, and Initial-Sound Intervention with Students Who Are Deaf or Hard of Hearing and Use Sign Language

    ERIC Educational Resources Information Center

    Tucci, Stacey L.; Easterbrooks, Susan R.

    2015-01-01

    This study investigated children's acquisition of three aspects of an early literacy curriculum, "Foundations for Literacy" ("Foundations"), designed specifically for prekindergarten students who are deaf or hard of hearing (DHH): syllable segmentation, identification of letter-sound correspondences, and initial-sound…

  7. Speech Perception Deficits in Poor Readers: Auditory Processing or Phonological Coding?

    ERIC Educational Resources Information Center

    Mody, Maria; And Others

    1997-01-01

    Forty second-graders, 20 good and 20 poor readers, completed a /ba/-/da/ temporal order judgment (TOJ) task. The groups did not differ in TOJ when /ba/ and /da/ were paired with more easily discriminated syllables. Poor readers' difficulties with /ba/-/da/ reflected perceptual confusion between phonetically similar syllables rather than difficulty…

  8. Structure of Preschool Phonological Sensitivity: Overlapping Sensitivity to Rhyme, Words, Syllables, and Phonemes.

    ERIC Educational Resources Information Center

    Anthony, Jason L.; Lonigan, Christopher J.; Burgess, Stephen R.; Driscoll, Kimberly; Phillips, Beth M.; Cantor, Brenlee G.

    2002-01-01

    This study examined relations among sensitivity to words, syllables, rhymes, and phonemes in older and younger preschoolers. Confirmatory factor analyses found that a one-factor model best explained the date from both groups of children. Only variance common to all phonological sensitivity skills was related to print knowledge and rudimentary…

  9. Telehealth Delivery of Rapid Syllable Transitions (ReST) Treatment for Childhood Apraxia of Speech

    ERIC Educational Resources Information Center

    Thomas, Donna C.; McCabe, Patricia; Ballard, Kirrie J.; Lincoln, Michelle

    2016-01-01

    Background: Rapid Syllable Transitions (ReST) treatment uses pseudo-word targets with varying lexical stress to target simultaneously articulation, prosodic accuracy and coarticulatory transitions in childhood apraxia of speech (CAS). The treatment is efficacious for the acquisition of imitated pseudo-words, and generalization of skill to…

  10. Investigating the Retention and Time Course of Phonotactic Constraint Learning from Production Experience

    ERIC Educational Resources Information Center

    Warker, Jill A.

    2013-01-01

    Adults can rapidly learn artificial phonotactic constraints such as /"f"/ "occurs only at the beginning of syllables" by producing syllables that contain those constraints. This implicit learning is then reflected in their speech errors. However, second-order constraints in which the placement of a phoneme depends on another…

  11. Input Frequency and the Acquisition of Syllable Structure in Polish

    ERIC Educational Resources Information Center

    Jarosz, Gaja; Calamaro, Shira; Zentz, Jason

    2017-01-01

    This article examines phonological development and its relationship to input statistics. Using novel data from a longitudinal corpus of spontaneous child speech in Polish, we evaluate and compare the predictions of a variety of input-based phonotactic models for syllable structure acquisition. We find that many commonly examined input statistics…

  12. Absolute and Relative Reliability of Percentage of Syllables Stuttered and Severity Rating Scales

    ERIC Educational Resources Information Center

    Karimi, Hamid; O'Brian, Sue; Onslow, Mark; Jones, Mark

    2014-01-01

    Purpose: Percentage of syllables stuttered (%SS) and severity rating (SR) scales are measures in common use to quantify stuttering severity and its changes during basic and clinical research conditions. However, their reliability has not been assessed with indices measuring both relative and absolute reliability. This study was designed to provide…

  13. Geographically pervasive effects of urban noise on frequency and syllable rate of songs and calls in silvereyes (Zosterops lateralis).

    PubMed

    Potvin, Dominique A; Parris, Kirsten M; Mulder, Raoul A

    2011-08-22

    Recent studies in the Northern Hemisphere have shown that songbirds living in noisy urban environments sing at higher frequencies than their rural counterparts. However, several aspects of this phenomenon remain poorly understood. These include the geographical scale over which such patterns occur (most studies have compared local populations), and whether they involve phenotypic plasticity or microevolutionary change. We conducted a field study of silvereye (Zosterops lateralis) vocalizations over more than 1 million km(2) of urban and rural south-eastern Australia, and compared possible effects of urban noise on songs (which are learned) and contact calls (which are innate). Across 14 paired urban and rural populations, silvereyes consistently sang both songs and contact calls at higher frequencies in urban environments. Syllable rate (syllables per second) decreased in urban environments, consistent with the hypothesis that reflective structures degrade song and encourage longer intervals between syllables. This comprehensive study is, to our knowledge, the first to demonstrate varied adaptations of urban bird vocalizations over a vast geographical area, and to provide insight into the mechanism responsible for these changes.

  14. Phase-Specific Vocalizations of Male Mice at the Initial Encounter during the Courtship Sequence

    PubMed Central

    Matsumoto, Yui K.; Okanoya, Kazuo

    2016-01-01

    Mice produce ultrasonic vocalizations featuring a variety of syllables. Vocalizations are observed during social interactions. In particular, males produce numerous syllables during courtship. Previous studies have shown that vocalizations change according to sexual behavior, suggesting that males vary their vocalizations depending on the phase of the courtship sequence. To examine this process, we recorded large sets of mouse vocalizations during male–female interactions and acoustically categorized these sounds into 12 vocal types. We found that males emitted predominantly short syllables during the first minute of interaction, more long syllables in the later phases, and mainly harmonic sounds during mounting. These context- and time-dependent changes in vocalization indicate that vocal communication during courtship in mice consists of at least three stages and imply that each vocalization type has a specific role in a phase of the courtship sequence. Our findings suggest that recording for a sufficiently long time and taking the phase of courtship into consideration could provide more insights into the role of vocalization in mouse courtship behavior in future study. PMID:26841117

  15. The role of syllables in sign language production

    PubMed Central

    Baus, Cristina; Gutiérrez, Eva; Carreiras, Manuel

    2014-01-01

    The aim of the present study was to investigate the functional role of syllables in sign language and how the different phonological combinations influence sign production. Moreover, the influence of age of acquisition was evaluated. Deaf signers (native and non-native) of Catalan Signed Language (LSC) were asked in a picture-sign interference task to sign picture names while ignoring distractor-signs with which they shared two phonological parameters (out of three of the main sign parameters: Location, Movement, and Handshape). The results revealed a different impact of the three phonological combinations. While no effect was observed for the phonological combination Handshape-Location, the combination Handshape-Movement slowed down signing latencies, but only in the non-native group. A facilitatory effect was observed for both groups when pictures and distractors shared Location-Movement. Importantly, linguistic models have considered this phonological combination to be a privileged unit in the composition of signs, as syllables are in spoken languages. Thus, our results support the functional role of syllable units during phonological articulation in sign language production. PMID:25431562

  16. Duration of the speech disfluencies of beginning stutterers.

    PubMed

    Zebrowski, P M

    1991-06-01

    This study compared the duration of within-word disfluencies and the number of repeated units per instance of sound/syllable and whole-word repetitions of beginning stutterers to those produced by age- and sex-matched nonstuttering children. Subjects were 10 stuttering children [9 males and 1 female; mean age 4:1 (years:months); age range 3:2-5:1), and 10 nonstuttering children (9 males and 1 female; mean age 4:0; age range: 2:10-5:1). Mothers of the stuttering children reported that their children had been stuttering for 1 year or less. One 300-word conversational speech sample from each of the stuttering and nonstuttering children was analyzed for (a) mean duration of sound/syllable repetition and sound prolongation, (b) mean number of repeated units per instance of sound/syllable and whole-word repetition, and (c) various related measures of the frequency of all between- and within-word speech disfluencies. There were no significant between-group differences for either the duration of acoustically measured sound/syllable repetitions and sound prolongations or the number of repeated units per instance of sound/syllable and whole-word repetition. Unlike frequency and type of speech disfluency produced, average duration of within-word disfluencies and number of repeated units per repetition do not differentiate the disfluent speech of beginning stutterers and their nonstuttering peers. Additional analyses support findings from previous perceptual work that type and frequency of speech disfluency, not duration, are the principal characteristics listeners use in distinguishing these two talker groups.

  17. Hemispheric asymmetry in auditory processing of speech envelope modulations in prereading children.

    PubMed

    Vanvooren, Sophie; Poelmans, Hanne; Hofmann, Michael; Ghesquière, Pol; Wouters, Jan

    2014-01-22

    The temporal envelope of speech is an important cue contributing to speech intelligibility. Theories about the neural foundations of speech perception postulate that the left and right auditory cortices are functionally specialized in analyzing speech envelope information at different time scales: the right hemisphere is thought to be specialized in processing syllable rate modulations, whereas a bilateral or left hemispheric specialization is assumed for phoneme rate modulations. Recently, it has been found that this functional hemispheric asymmetry is different in individuals with language-related disorders such as dyslexia. Most studies were, however, performed in adults and school-aged children, and only a little is known about how neural auditory processing at these specific rates manifests and develops in very young children before reading acquisition. Yet, studying hemispheric specialization for processing syllable and phoneme rate modulations in preliterate children may reveal early neural markers for dyslexia. In the present study, human cortical evoked potentials to syllable and phoneme rate modulations were measured in 5-year-old children at high and low hereditary risk for dyslexia. The results demonstrate a right hemispheric preference for processing syllable rate modulations and a symmetric pattern for phoneme rate modulations, regardless of hereditary risk for dyslexia. These results suggest that, while hemispheric specialization for processing syllable rate modulations seems to be mature in prereading children, hemispheric specialization for phoneme rate modulation processing may still be developing. These findings could have important implications for the development of phonological and reading skills.

  18. Shared processing of planning articulatory gestures and grasping.

    PubMed

    Vainio, L; Tiainen, M; Tiippana, K; Vainio, M

    2014-07-01

    It has been proposed that articulatory gestures are shaped by tight integration in planning mouth and hand acts. This hypothesis is supported by recent behavioral evidence showing that response selection between the precision and power grip is systematically influenced by simultaneous articulation of a syllable. For example, precision grip responses are performed relatively fast when the syllable articulation employs the tongue tip (e.g., [te]), whereas power grip responses are performed relatively fast when the syllable articulation employs the tongue body (e.g., [ke]). However, this correspondence effect, and other similar effects that demonstrate the interplay between grasping and articulatory gestures, has been found when the grasping is performed during overt articulation. The present study demonstrates that merely reading the syllables silently (Experiment 1) or hearing them (Experiment 2) results in a similar correspondence effect. The results suggest that the correspondence effect is based on integration in planning articulatory gestures and grasping rather than requiring an overt articulation of the syllables. We propose that this effect reflects partially overlapped planning of goal shapes of the two distal effectors: a vocal tract shape for articulation and a hand shape for grasping. In addition, the paper shows a pitch-grip correspondence effect in which the precision grip is associated with a high-pitched vocalization of the auditory stimuli and the power grip is associated with a low-pitched vocalization. The underlying mechanisms of this phenomenon are discussed in relation to the articulation-grip correspondence.

  19. Effect of attentional load on audiovisual speech perception: evidence from ERPs

    PubMed Central

    Alsius, Agnès; Möttönen, Riikka; Sams, Mikko E.; Soto-Faraco, Salvador; Tiippana, Kaisa

    2014-01-01

    Seeing articulatory movements influences perception of auditory speech. This is often reflected in a shortened latency of auditory event-related potentials (ERPs) generated in the auditory cortex. The present study addressed whether this early neural correlate of audiovisual interaction is modulated by attention. We recorded ERPs in 15 subjects while they were presented with auditory, visual, and audiovisual spoken syllables. Audiovisual stimuli consisted of incongruent auditory and visual components known to elicit a McGurk effect, i.e., a visually driven alteration in the auditory speech percept. In a Dual task condition, participants were asked to identify spoken syllables whilst monitoring a rapid visual stream of pictures for targets, i.e., they had to divide their attention. In a Single task condition, participants identified the syllables without any other tasks, i.e., they were asked to ignore the pictures and focus their attention fully on the spoken syllables. The McGurk effect was weaker in the Dual task than in the Single task condition, indicating an effect of attentional load on audiovisual speech perception. Early auditory ERP components, N1 and P2, peaked earlier to audiovisual stimuli than to auditory stimuli when attention was fully focused on syllables, indicating neurophysiological audiovisual interaction. This latency decrement was reduced when attention was loaded, suggesting that attention influences early neural processing of audiovisual speech. We conclude that reduced attention weakens the interaction between vision and audition in speech. PMID:25076922

  20. Reduced frontal theta oscillations indicate altered crossmodal prediction error processing in schizophrenia

    PubMed Central

    Keil, Julian; Balz, Johanna; Gallinat, Jürgen; Senkowski, Daniel

    2016-01-01

    Our brain generates predictions about forthcoming stimuli and compares predicted with incoming input. Failures in predicting events might contribute to hallucinations and delusions in schizophrenia (SZ). When a stimulus violates prediction, neural activity that reflects prediction error (PE) processing is found. While PE processing deficits have been reported in unisensory paradigms, it is unknown whether SZ patients (SZP) show altered crossmodal PE processing. We measured high-density electroencephalography and applied source estimation approaches to investigate crossmodal PE processing generated by audiovisual speech. In SZP and healthy control participants (HC), we used an established paradigm in which high- and low-predictive visual syllables were paired with congruent or incongruent auditory syllables. We examined crossmodal PE processing in SZP and HC by comparing differences in event-related potentials and neural oscillations between incongruent and congruent high- and low-predictive audiovisual syllables. In both groups event-related potentials between 206 and 250 ms were larger in high- compared with low-predictive syllables, suggesting intact audiovisual incongruence detection in the auditory cortex of SZP. The analysis of oscillatory responses revealed theta-band (4–7 Hz) power enhancement in high- compared with low-predictive syllables between 230 and 370 ms in the frontal cortex of HC but not SZP. Thus aberrant frontal theta-band oscillations reflect crossmodal PE processing deficits in SZ. The present study suggests a top-down multisensory processing deficit and highlights the role of dysfunctional frontal oscillations for the SZ psychopathology. PMID:27358314

  1. White Matter Integrity and Treatment-Based Change in Speech Performance in Minimally Verbal Children with Autism Spectrum Disorder.

    PubMed

    Chenausky, Karen; Kernbach, Julius; Norton, Andrea; Schlaug, Gottfried

    2017-01-01

    We investigated the relationship between imaging variables for two language/speech-motor tracts and speech fluency variables in 10 minimally verbal (MV) children with autism. Specifically, we tested whether measures of white matter integrity-fractional anisotropy (FA) of the arcuate fasciculus (AF) and frontal aslant tract (FAT)-were related to change in percent syllable-initial consonants correct, percent items responded to, and percent syllable insertion errors (from best baseline to post 25 treatment sessions). Twenty-three MV children with autism spectrum disorder (ASD) received Auditory-Motor Mapping Training (AMMT), an intonation-based treatment to improve fluency in spoken output, and we report on seven who received a matched control treatment. Ten of the AMMT participants were able to undergo a magnetic resonance imaging study at baseline; their performance on baseline speech production measures is compared to that of the other two groups. No baseline differences were found between groups. A canonical correlation analysis (CCA) relating FA values for left- and right-hemisphere AF and FAT to speech production measures showed that FA of the left AF and right FAT were the largest contributors to the synthetic independent imaging-related variable. Change in percent syllable-initial consonants correct and percent syllable-insertion errors were the largest contributors to the synthetic dependent fluency-related variable. Regression analyses showed that FA values in left AF significantly predicted change in percent syllable-initial consonants correct, no FA variables significantly predicted change in percent items responded to, and FA of right FAT significantly predicted change in percent syllable-insertion errors. Results are consistent with previously identified roles for the AF in mediating bidirectional mapping between articulation and acoustics, and the FAT in its relationship to speech initiation and fluency. They further suggest a division of labor between the hemispheres, implicating the left hemisphere in accuracy of speech production and the right hemisphere in fluency in this population. Changes in response rate are interpreted as stemming from factors other than the integrity of these two fiber tracts. This study is the first to document the existence of a subgroup of MV children who experience increases in syllable- insertion errors as their speech develops in response to therapy.

  2. The Effect of Pitch Peak Alignment on Sentence Type Identification in Russian

    ERIC Educational Resources Information Center

    Makarova, Veronika

    2007-01-01

    This paper reports the results of an experimental phonetic study examining pitch peak alignment in production and perception of three-syllable one-word sentences with phonetic rising-falling pitch movement by speakers of Russian. The first part of the study (Experiment 1) utilizes 22 one-word three-syllable utterances read by five female speakers…

  3. Fluctuations in Unimanual Hand Preference in Infants Following the Onset of Duplicated Syllable Babbling.

    ERIC Educational Resources Information Center

    Ramsay, Douglas S.

    1985-01-01

    Infants were tested for unimanual handedness at weekly intervals for a 14-week period beginning with the week of onset of duplicated syllable babbling. Group analyses indicating effects of sex and/or birth order on fluctuations and date review for individual infants suggested considerable variability across infants in occurrence and/or timing of…

  4. Differential Memory of Picture and Word Stimuli in a Paired-Associate Learning Task.

    ERIC Educational Resources Information Center

    Bartels, Laura Grand; Feinbloom, Jessica

    Ten concrete nouns represented in either a pictorial or a linguistic mode and accompanied by ten nonsense syllables were shown to 77 college students in a study of how pictorial stimuli varied in recall and recognition tasks. The group receiving pictorial stimuli recalled and recognized significantly more nonsense syllables than did the group…

  5. Development of Phase Locking and Frequency Representation in the Infant Frequency-Following Response

    ERIC Educational Resources Information Center

    Van Dyke, Katlyn B.; Lieberman, Rachel; Presacco, Alessandro; Anderson, Samira

    2017-01-01

    Purpose: This study investigates the development of phase locking and frequency representation in infants using the frequency-following response to consonant-vowel syllables. Method: The frequency-following response was recorded in 56 infants and 15 young adults to 2 speech syllables (/ba/ and /ga/), which were presented in randomized order to the…

  6. Production of Syllable Stress in Speakers with Autism Spectrum Disorders

    ERIC Educational Resources Information Center

    Paul, Rhea; Bianchi, Nancy; Augustyn, Amy; Klin, Ami; Volkmar, Fred R.

    2008-01-01

    This paper reports a study of the ability to reproduce stress in a nonsense syllable imitation task by adolescent speakers with autism spectrum disorders (ASD), as compared to typically developing (TD) age-mates. Results are reported for both raters' judgments of the subjects' stress production, as well as acoustic measures of pitch range and…

  7. Syllable Onset Intervals as an Indicator of Discourse and Syntactic Boundaries in Taiwan Mandarin

    ERIC Educational Resources Information Center

    Fon, Janice; Johnson, Keith

    2004-01-01

    This study looks at the syllable onset interval (SOI) patterning in Taiwan Mandarin spontaneous speech and its relationship to discourse and syntactic units. Monologs were elicited by asking readers to tell stories depicted in comic strips and were transcribed and segmented into Discourse Segment Units (Grosz & Sidner, 1986), clauses, and…

  8. Relations among Detection of Syllable Stress, Speech Abnormalities, and Communicative Ability in Adults with Autism Spectrum Disorders

    ERIC Educational Resources Information Center

    Kargas, Niko; López, Beatriz; Morris, Paul; Reddy, Vasudevi

    2016-01-01

    Purpose: To date, the literature on perception of affective, pragmatic, and grammatical prosody abilities in autism spectrum disorders (ASD) has been sparse and contradictory. It is interesting to note that the primary perception of syllable stress within the word structure, which is crucial for all prosody functions, remains relatively unexplored…

  9. Interaction between Phonemic Abilities and Syllable Congruency Effect in Young Readers

    ERIC Educational Resources Information Center

    Chetail, Fabienne; Mathey, Stephanie

    2013-01-01

    This study investigated whether and to what extent phonemic abilities of young readers (Grade 5) influence syllabic effects in reading. More precisely, the syllable congruency effect was tested in the lexical decision task combined with masked priming in eleven-year-old children. Target words were preceded by a pseudo-word prime sharing the first…

  10. A Componential Approach to Training Reading Skills.

    DTIC Science & Technology

    1983-03-17

    1 syllable, mixed vowels A2 16 one-syll., 4 two-syll., mixed vowels A3 14 one-syll., 6 two-syll., mixed vowels A4 All two-syllable, mixed vowels* B1 ...06520 I ERIC Facility-Acquisitions I Dr. John S. Brown 4833 Rugby Avenue XEROX Palo Alto Research Center Bethesda, MD 20014 3333 Coyote Road Palo Alto, CA

  11. Enhanced Passive and Active Processing of Syllables in Musician Children

    ERIC Educational Resources Information Center

    Chobert, Julie; Marie, Celine; Francois, Clement; Schon, Daniele; Besson, Mireille

    2011-01-01

    The aim of this study was to examine the influence of musical expertise in 9-year-old children on passive (as reflected by MMN) and active (as reflected by discrimination accuracy) processing of speech sounds. Musician and nonmusician children were presented with a sequence of syllables that included standards and deviants in vowel frequency,…

  12. Boosting Reading Fluency: An Intervention Case Study at Subword Level

    ERIC Educational Resources Information Center

    Kairaluoma, Leila; Ahonen, Timo; Aro, Mikko; Holopainen, Leena

    2007-01-01

    This study is an intervention case study of fluency in Finnish-speaking children with dyslexia. Two 7-year-old children, a girl and a boy, were selected from the Jyvaskyla Longitudinal Study of Dyslexia. The intervention emphasised syllables as reading units, and proceeded from reading syllables to reading words and text. Letter knowledge, reading…

  13. The Influence of Phonological Mechanisms in Written Spelling of Profoundly Deaf Children

    ERIC Educational Resources Information Center

    Colombo, Lucia; Arfe, Barbara; Bronte, Tiziana

    2012-01-01

    In the present study, the effect of phonological and working memory mechanisms involved in spelling Italian single words was explored in two groups of children matched for grade level: a group of normally hearing children and a group of pre-verbally deaf children, with severe-to-profound hearing loss. Three-syllable and four-syllable familiar…

  14. Length Effects Turn Out To Be Syllable Structure Effects: Response to Roelofs (2002).

    ERIC Educational Resources Information Center

    Santiago, Julio; MacKay, Donald G.; Palma, Alfonso

    2002-01-01

    Responds to a commentary written in response a research study conducted by the author (Santiago et al., 2000) that suggests that a reanalysis of the data on syllable structure effects that takes word length into account leads to a conclusion that is the opposite of what the study found. (Author/VWL)

  15. A Forced-Attention Dichotic Listening fMRI Study on 113 Subjects

    ERIC Educational Resources Information Center

    Kompus, Kristiina; Specht, Karsten; Ersland, Lars; Juvodden, Hilde T.; van Wageningen, Heidi; Hugdahl, Kenneth; Westerhausen, Rene

    2012-01-01

    We report fMRI and behavioral data from 113 subjects on attention and cognitive control using a variant of the classic dichotic listening paradigm with pairwise presentations of consonant-vowel syllables. The syllable stimuli were presented in a block-design while subjects were in the MR scanner. The subjects were instructed to pay attention to…

  16. Vowels, Syllables, and Letter Names: Differences between Young Children's Spelling in English and Portuguese

    ERIC Educational Resources Information Center

    Pollo, Tatiana Cury; Kessler, Brett; Treiman, Rebecca

    2005-01-01

    Young Portuguese-speaking children have been reported to produce more vowel- and syllable-oriented spellings than have English speakers. To investigate the extent and source of such differences, we analyzed children's vocabulary and found that Portuguese words have more vowel letter names and a higher vowel-consonant ratio than do English words.…

  17. A Retrospective Video Analysis of Canonical Babbling and Volubility in Infants with Fragile X Syndrome at 9 – 12 Months of Age

    PubMed Central

    Belardi, Katie; Watson, Linda R.; Faldowski, Richard A.; Hazlett, Heather; Crais, Elizabeth; Baranek, Grace T.; McComish, Cara; Patten, Elena; Oller, D. Kimbrough

    2017-01-01

    An infant’s vocal capacity develops significantly during the first year of life. Research suggests early measures of pre-speech development, such as canonical babbling and volubility, can differentiate typical versus disordered development. This study offers a new contribution by comparing early vocal development in 10 infants with Fragile X syndrome and 14 with typical development. Results suggest infants with Fragile X syndrome produce fewer syllables and have significantly lower canonical babbling ratios (i.e., canonical syllables/total syllables) compared to infants who are typically developing. Furthermore, the particular measures of babbling were strong predictors of group membership, adding evidence regarding the possible utility of these markers in early identification. PMID:28247019

  18. Cultural variation in savannah sparrow, Passerculus sandwichensis, songs: an analysis using the meme concept.

    PubMed

    Burnell

    1998-10-01

    I used the meme concept to investigate patterns of cultural variation among the songs of eight, geographically distinct populations of savannah sparrows. Memes composed of only one syllable were geographically widespread and randomly distributed among populations, but memes of two-, three- and four-syllables became progressively more restricted in their geographical distribution. Thus, the populations were memetically more similar with respect to one-syllable memes and more divergent with respect to larger memes. These results suggest that differences in memetic mutation rates and susceptibility to loss by memetic drift could be sufficient to create the observed pattern of greater divergence among populations for large memes. Copyright 1998 The Association for the Study of Animal Behaviour.

  19. Developmental Role of Static, Dynamic, and Contextual Cues in Speech Perception

    ERIC Educational Resources Information Center

    Hicks, Candace Bourland; Ohde, Ralph N.

    2005-01-01

    The purpose of the current study was to examine the role of syllable duration context as well as static and dynamic acoustic properties in child and adult speech perception. Ten adults and eleven 4?5-year-old children identified a syllable as [ba] or [wa] (stop-glide contrast) in 3 conditions differing in synthetic continua. The 1st condition…

  20. Preterm and Term Infants' Perception of Temporally Coordinated Syllable-Object Pairings: Implications for Lexical Development

    ERIC Educational Resources Information Center

    Gogate, Lakshmi; Maganti, Madhavilatha; Perenyi, Agnes

    2014-01-01

    Purpose: This experimental study examined term infants (n = 34) and low-risk near-term preterm infants (gestational age 32-36 weeks) at 2 months chronological age (n = 34) and corrected age (n = 16). The study investigated whether the preterm infants presented with a delay in their sensitivity to synchronous syllable-object pairings when compared…

  1. Evidence and control of bifurcations in a respiratory system.

    PubMed

    Goldin, Matías A; Mindlin, Gabriel B

    2013-12-01

    We studied the pressure patterns used by domestic canaries in the production of birdsong. Acoustically different sound elements ("syllables") were generated by qualitatively different pressure gestures. We found that some ubiquitous transitions between syllables can be interpreted as bifurcations of a low dimensional dynamical system. We interpreted these results as evidence supporting a model in which different timescales interact nonlinearly.

  2. Tracking the Speech Signal--Time-Locked MEG Signals during Perception of Ultra-Fast and Moderately Fast Speech in Blind and in Sighted Listeners

    ERIC Educational Resources Information Center

    Hertrich, Ingo; Dietrich, Susanne; Ackermann, Hermann

    2013-01-01

    Blind people can learn to understand speech at ultra-high syllable rates (ca. 20 syllables/s), a capability associated with hemodynamic activation of the central-visual system. To further elucidate the neural mechanisms underlying this skill, magnetoencephalographic (MEG) measurements during listening to sentence utterances were cross-correlated…

  3. A Randomized Controlled Trial for Children with Childhood Apraxia of Speech Comparing Rapid Syllable Transition Treatment and the Nuffield Dyspraxia Programme-Third Edition

    ERIC Educational Resources Information Center

    Murray, Elizabeth; McCabe, Patricia; Ballard, Kirrie J.

    2015-01-01

    Purpose: This randomized controlled trial compared the experimental Rapid Syllable Transition (ReST) treatment to the Nuffield Dyspraxia Programme-Third Edition (NDP3; Williams & Stephens, 2004), used widely in clinical practice in Australia and the United Kingdom. Both programs aim to improve speech motor planning/programming for children…

  4. Authorship Attribution in the E-mail Domain: A Study of the Effect of Size of Author Corpus and Topic on Accuracy of Identification

    DTIC Science & Technology

    2011-03-01

    Wilhelm Fucks discriminated between authors using the average number of syllables per word and average distance between equal-syllabled words. He concluded...363–390, 1939. [4] Conrad Mascol. Curves of pauline and pseudo-pauline style. Unitarian Review, 1888. [5] Wilhelm Fucks . On mathmatical analysis of

  5. The Locus Equation as an Index of Coarticulation in Syllables Produced by Speakers with Profound Hearing Loss

    ERIC Educational Resources Information Center

    McCaffrey Morrison, Helen

    2008-01-01

    Locus equations (LEs) were derived from consonant-vowel-consonant (CVC) syllables produced by four speakers with profound hearing loss. Group data indicated that LE functions obtained for the separate CVC productions initiated by /b/, /d/, and /g/ were less well-separated in acoustic space than those obtained from speakers with normal hearing. A…

  6. Rhyming Words and Onset-Rime Constituents: An Inquiry into Structural Breaking Points and Emergent Boundaries in the Syllable

    ERIC Educational Resources Information Center

    Geudens, Astrid; Sandra, Dominiek; Martensen, Heike

    2005-01-01

    Geudens and Sandra, in their 2003 study, investigated the special role of onsets and rimes in Dutch-speaking children's explicit phonological awareness. In the current study, we tapped implicit phonological knowledge using forced-choice similarity judgment (Experiment 1) and recall of syllable lists (Experiment 2). In Experiment 1, Dutch-speaking…

  7. Evidence for a Preserved Sensitivity to Orthographic Redundancy and an Impaired Access to Phonological Syllables in French Developmental Dyslexics

    ERIC Educational Resources Information Center

    Doignon-Camus, Nadège; Seigneuric, Alix; Perrier, Emeline; Sisti, Aurélie; Zagar, Daniel

    2013-01-01

    To evaluate the orthographic and phonological processing skills of developmental dyslexics, we (a) examined their abilities to exploit properties of orthographic redundancy and (b) tested whether their phonological deficit extends to spelling-to-sound connections for large-grain size units such as syllables. To assess the processing skills in…

  8. Still Not Adult-Like: Lexical Stress Contrastivity in Word Productions of Eight- to Eleven-Year-Olds

    ERIC Educational Resources Information Center

    Arciuli, Joanne; Ballard, Kirrie J.

    2017-01-01

    Lexical stress is the contrast between strong and weak syllables within words. Ballard et al. (2012) examined the amount of stress contrastivity across adjacent syllables in word productions of typically developing three- to seven-year-olds and adults. Here, eight- to eleven-year-olds are compared with the adults from Ballard et al. using acoustic…

  9. The perception of syllable affiliation of singleton stops in repetitive speech.

    PubMed

    de Jong, Kenneth J; Lim, Byung-Jin; Nagao, Kyoko

    2004-01-01

    Stetson (1951) noted that repeating singleton coda consonants at fast speech rates makes them be perceived as onset consonants affiliated with a following vowel. The current study documents the perception of rate-induced resyllabification, as well as what temporal properties give rise to the perception of syllable affiliation. Stimuli were extracted from a previous study of repeated stop + vowel and vowel + stop syllables (de Jong, 2001a, 2001b). Forced-choice identification tasks show that slow repetitions are clearly distinguished. As speakers increase rate, they reach a point after which listeners disagree as to the affiliation of the stop. This pattern is found for voiced and voiceless consonants using different stimulus extraction techniques. Acoustic models of the identifications indicate that the sudden shift in syllabification occurs with the loss of an acoustic hiatus between successive syllables. Acoustic models of the fast rate identifications indicate various other qualities, such as consonant voicing, affect the probability that the consonants will be perceived as onsets. These results indicate a model of syllabic affiliation where specific juncture-marking aspects of the signal dominate parsing, and in their absence other differences provide additional, weaker cues to syllabic affiliation.

  10. Polar-phase indices of perioral muscle reciprocity during syllable production in Parkinson's disease.

    PubMed

    Chu, Shin Ying; Barlow, Steven M; Lee, Jaehoon; Wang, Jingyan

    2017-12-01

    This research characterised perioral muscle reciprocity and amplitude ratio in lower lip during bilabial syllable production [pa] at three rates to understand the neuromotor dynamics and scaling of motor speech patterns in individuals with Parkinson's disease (PD). Electromyographic (EMG) signals of the orbicularis oris superior [OOS], orbicularis oris inferior [OOI] and depressor labii inferioris [DLI] were recorded during syllable production and expressed as polar-phase notations. PD participants exhibited the general features of reciprocity between OOS, OOI and DLI muscles as reflected in the EMG during syllable production. The control group showed significantly higher integrated EMG amplitude ratio in the DLI:OOS muscle pairs than PD participants. No speech rate effects were found in EMG muscle reciprocity and amplitude magnitude across all muscle pairs. Similar patterns of muscle reciprocity in PD and controls suggest that corticomotoneuronal output to the facial nucleus and respective perioral muscles is relatively well-preserved in our cohort of mild idiopathic PD participants. Reduction of EMG amplitude ratio among PD participants is consistent with the putative reduction in the thalamocortical activation characteristic of this disease which limits motor cortex drive from generating appropriate commands which contributes to bradykinesia and hypokinesia of the orofacial mechanism.

  11. Contrast effects on stop consonant identification.

    PubMed

    Diehl, R L; Elman, J L; McCusker, S B

    1978-11-01

    Changes in the identification of speech sounds following selective adaptation are usually attributed to a reduction in sensitivity of auditory feature detectors. An alternative explanation of these effects is based on the notion of response contrast. In several experiments, subjects identified the initial segment of synthetic consonant-vowel syllables as either the voiced stop [b] or the voiceless stop [ph]. Each test syllable had a value of voice onset time (VOT) that placed it near the English voiced-voiceless boundary. When the test syllables were preceded by a single clear [b] (VOT = -100 msec), subjects tended to identify them as [ph], whereas when they were preceded by an unambiguous [ph] (VOT = 100 msec), the syllables were predominantly labeled [b]. This contrast effect occurred even when the contextual stimuli were velar and the test stimuli were bilabial, which suggests a featural rather than a phonemic basis for the effect. To discount the possibility that these might be instances of single-trial sensory adaptation, we conducted a similar experiment in which the contextual stimuli followed the test items. Reliable contrast effects were still obtained. In view of these results, it appears likely that response contrast accounts for at least some component of the adaptation effects reported in the literature.

  12. Rhythmic syllable-related activity in a songbird motor thalamic nucleus necessary for learned vocalizations

    PubMed Central

    Danish, Husain H.; Aronov, Dmitriy; Fee, Michale S.

    2017-01-01

    Birdsong is a complex behavior that exhibits hierarchical organization. While the representation of singing behavior and its hierarchical organization has been studied in some detail in avian cortical premotor circuits, our understanding of the role of the thalamus in adult birdsong is incomplete. Using a combination of behavioral and electrophysiological studies, we seek to expand on earlier work showing that the thalamic nucleus Uvaeformis (Uva) is necessary for the production of stereotyped, adult song in zebra finch (Taeniopygia guttata). We confirm that complete bilateral lesions of Uva abolish singing in the ‘directed’ social context, but find that in the ‘undirected’ social context, such lesions result in highly variable vocalizations similar to early babbling song in juvenile birds. Recordings of neural activity in Uva reveal strong syllable-related modulation, maximally active prior to syllable onsets and minimally active prior to syllable offsets. Furthermore, both song and Uva activity exhibit a pronounced coherent modulation at 10Hz—a pattern observed in downstream premotor areas in adult and, even more prominently, in juvenile birds. These findings are broadly consistent with the idea that Uva is critical in the sequential activation of behavioral modules in HVC. PMID:28617829

  13. Measures of native and non-native rhythm in a quantity language.

    PubMed

    Stockmal, Verna; Markus, Dace; Bond, Dzintra

    2005-01-01

    The traditional phonetic classification of language rhythm as stress-timed or syllable-timed is attributed to Pike. Recently, two different proposals have been offered for describing the rhythmic structure of languages from acoustic-phonetic measurements. Ramus has suggested a metric based on the proportion of vocalic intervals and the variability (SD) of consonantal intervals. Grabe has proposed Pairwise Variability Indices (nPVI, rPVI) calculated from the differences in vocalic and consonantal durations between successive syllables. We have calculated both the Ramus and Grabe metrics for Latvian, traditionally considered a syllable rhythm language, and for Latvian as spoken by Russian learners. Native speakers and proficient learners were very similar whereas low-proficiency learners showed high variability on some properties. The metrics did not provide an unambiguous classification of Latvian.

  14. [Stimuli phrases of adductor spasmodic dysphonia phonatory break in mandarin Chinese].

    PubMed

    Ge, Pingjiang; Ren, Qingyi; Chen, Zhipeng; Cheng, Qiuhui; Sheng, Xiaoli; Wang, Ling; Chen, Shaohua; Zhang, Siyi

    2015-12-01

    To investigate the characteristics of adductor spasmodic dysphonia phonatory break in mandarin Chinese and select the stimuli phrases. Thirty-eight patients with adductor spasmodic dysphonia were involved in this study. Standard phrase " fù mŭ xīn" and a speech corpus in mandarin Chinese with 229 syllables covering all vowel and constant of mandarin Chinese were selected. Every patient read the phrases above twice in normal speed and comfortable voice. Two auditory perpetual speech pathologists marked phonatory break syllables respectively. The frequency of phonatory break syllables and their located phrases were calculated, rated and described. The phrases including the most phonatory break syllables were selected as stimuli phrases, the phonatory break frequency of which was also higher than that of standard phrase "fù mŭ xīn". Phonatory break happened in the reading of all patients. The average number of phonatory break syllables was 14 (3-33). Phonatroy break occurred when saying 177 (77.3%) syllables in the speech corpus. The syllables "guŏ, rén, zāng, diàn, chē, gè, guăn, a, bā, ne, de" broke in 23.1%-41.0% patients. These syllables belonged to the phrases "pĭng guŏ, huŏ chē, shì de, nĭ shì gè hăo rén, wŏ mén shì yŏu zŏng shì bă qĭn shì nong dé hĕn zāng, wŏ mén nà biān yŏu wăng qiú yùn dong chăng, cān gŭan, jiŭ bā hé yī gè miàn bāo dìan, tā shì duō me kāng kăi a,wŏ yīng gāi zài xìn lĭ xiĕ yī xiē shén mē ne?". Thirty-seven patients (97.3%) had phonatory break in above mentioned words. Ratios of these words phonatory break also were more than "fù mŭ xīn". Adductor spasmodic dysphonic patients exhibited different degrees of phonatory break in mandarine Chinese. The phrases" shì de, pĭng guŏ, huŏ chē, nĭ shì gè hăo rén, wŏ mén nà biān yŏu wăng qiú yùn dong chăng, cān gŭan, jiŭ bā hé yī gè miàn bāo dìan, tā shì duō me kāng kăi a" were recommended as stimuli phrases for adductor spasmodic dysphonia evaluation.

  15. The Parsing Syllable Envelopes Test for Assessment of Amplitude Modulation Discrimination Skills in Children: Development, Normative Data, and Test-Retest Reliability Studies.

    PubMed

    Cameron, Sharon; Chong-White, Nicky; Mealings, Kiri; Beechey, Tim; Dillon, Harvey; Young, Taegan

    2018-02-01

    Intensity peaks and valleys in the acoustic signal are salient cues to syllable structure, which is accepted to be a crucial early step in phonological processing. As such, the ability to detect low-rate (envelope) modulations in signal amplitude is essential to parse an incoming speech signal into smaller phonological units. The Parsing Syllable Envelopes (ParSE) test was developed to quantify the ability of children to recognize syllable boundaries using an amplitude modulation detection paradigm. The envelope of a 750-msec steady-state /a/ vowel is modulated into two or three pseudo-syllables using notches with modulation depths varying between 0% and 100% along an 11-step continuum. In an adaptive three-alternative forced-choice procedure, the participant identified whether one, two, or three pseudo-syllables were heard. Development of the ParSE stimuli and test protocols, and collection of normative and test-retest reliability data. Eleven adults (aged 23 yr 10 mo to 50 yr 9 mo, mean 32 yr 10 mo) and 134 typically developing, primary-school children (aged 6 yr 0 mo to 12 yr 4 mo, mean 9 yr 3 mo). There were 73 males and 72 females. Data were collected using a touchscreen computer. Psychometric functions (PFs) were automatically fit to individual data by the ParSE software. Performance was related to the modulation depth at which syllables can be detected with 88% accuracy (referred to as the upper boundary of the uncertainty region [UBUR]). A shallower PF slope reflected a greater level of uncertainty. Age effects were determined based on raw scores. z Scores were calculated to account for the effect of age on performance. Outliers, and individual data for which the confidence interval of the UBUR exceeded a maximum allowable value, were removed. Nonparametric tests were used as the data were skewed toward negative performance. Across participants, the performance criterion (UBUR) was met with a median modulation depth of 42%. The effect of age on the UBUR was significant (p < 0.00001). The UBUR ranged from 50% modulation depth for 6-yr-olds to 25% for adults. Children aged 6-10 had significantly higher uncertainty region boundaries than adults. A skewed distribution toward negative performance occurred (p = 0.00007). There was no significant difference in performance on the ParSE between males and females (p = 0.60). Test-retest z scores were strongly correlated (r = 0.68, p < 0.0000001). The ParSE normative data show that the ability to identify syllable boundaries based on changes in amplitude modulation improves with age, and that some children in the general population have performance much worse than their age peers. The test is suitable for use in planned studies in a clinical population. American Academy of Audiology

  16. The Functional Unit in Phonological Encoding: Evidence for Moraic Representation in Native Japanese Speakers

    ERIC Educational Resources Information Center

    Kureta, Yoichi; Fushimi, Takao; Tatsumi, Itaru F.

    2006-01-01

    Speech production studies have shown that the phonological form of a word is made up of phonemic segments in stress-timed languages (e.g., Dutch) and of syllables in syllable timed languages (e.g., Chinese). To clarify the functional unit of mora-timed languages, the authors asked native Japanese speakers to perform an implicit priming task (A. S.…

  17. Proximate Units in Word Production: Phonological Encoding Begins with Syllables in Mandarin Chinese but with Segments in English

    ERIC Educational Resources Information Center

    O'Seaghdha, Padraig G.; Chen, Jenn-Yeu; Chen, Train-Min

    2010-01-01

    In Mandarin Chinese, speakers benefit from fore-knowledge of what the first syllable but not of what the first phonemic segment of a disyllabic word will be (Chen, Chen, & Dell, 2002), contrasting with findings in English, Dutch, and other Indo-European languages, and challenging the generality of current theories of word production. In this…

  18. The Acquisition of Sight-Singing Skills in Second-Grade General Music: Effects of Using Solfege and of Relating Tonal Patterns to Songs

    ERIC Educational Resources Information Center

    Reifinger, James L., Jr.

    2012-01-01

    This study was designed to examine two aspects of sight-singing instruction: (1) solfege syllables versus the syllable "loo" for singing patterns and (2) the use of related songs (songs that began with tonal patterns being studied) as compared with unrelated songs. Second-grade students (N = 193) enrolled in general music classes…

  19. Effect of Syllable Congruency in Sixth Graders in the Lexical Decision Task with Masked Priming

    ERIC Educational Resources Information Center

    Chetail, Fabienne; Mathey, Stephanie

    2012-01-01

    The aim of this study was to investigate the role of the syllable in visual recognition of French words in Grade 6. To do so, the syllabic congruency effect was examined in the lexical decision task combined with masked priming. Target words were preceded by pseudoword primes sharing the first letters that either corresponded to the syllable…

  20. Conditioning of Attitudes Using a Backward Conditioning Paradigm. Language, Personality, Social and Cross-Cultural Study and Measurement of the Human A-R-D (Motivational) System.

    ERIC Educational Resources Information Center

    Brewer, Barbara A.; Gross, Michael C.

    In order to test whether meaning will transfer when a backward conditioning paradigm is utilized, Staats' language conditioning procedure, including the pairing of unconditioned stimulus (UCS) evaluative words with conditioned stimulus (CS) nonsense syllables, was modified so that the UCS words preceded the CS nonsense syllables on each trial.…

  1. A Study of Mexican Free-Tailed Bat Chirp Syllables: Bayesian Functional Mixed Models for Nonstationary Acoustic Time Series.

    PubMed

    Martinez, Josue G; Bohn, Kirsten M; Carroll, Raymond J; Morris, Jeffrey S

    2013-06-01

    We describe a new approach to analyze chirp syllables of free-tailed bats from two regions of Texas in which they are predominant: Austin and College Station. Our goal is to characterize any systematic regional differences in the mating chirps and assess whether individual bats have signature chirps. The data are analyzed by modeling spectrograms of the chirps as responses in a Bayesian functional mixed model. Given the variable chirp lengths, we compute the spectrograms on a relative time scale interpretable as the relative chirp position, using a variable window overlap based on chirp length. We use 2D wavelet transforms to capture correlation within the spectrogram in our modeling and obtain adaptive regularization of the estimates and inference for the regions-specific spectrograms. Our model includes random effect spectrograms at the bat level to account for correlation among chirps from the same bat, and to assess relative variability in chirp spectrograms within and between bats. The modeling of spectrograms using functional mixed models is a general approach for the analysis of replicated nonstationary time series, such as our acoustical signals, to relate aspects of the signals to various predictors, while accounting for between-signal structure. This can be done on raw spectrograms when all signals are of the same length, and can be done using spectrograms defined on a relative time scale for signals of variable length in settings where the idea of defining correspondence across signals based on relative position is sensible.

  2. Speed-difficulty trade-off in speech: Chinese versus English

    PubMed Central

    Sun, Yao; Latash, Elizaveta M.; Mikaelian, Irina L.

    2011-01-01

    This study continues the investigation of the previously described speed-difficulty trade-off in picture description tasks. In particular, we tested a hypothesis that the Mandarin Chinese and American English are similar in showing logarithmic dependences between speech time and index of difficulty (ID), while they differ significantly in the amount of time needed to describe simple pictures, this difference increases for more complex pictures, and it is associated with a proportional difference in the number of syllables used. Subjects (eight Chinese speakers and eight English speakers) were tested in pairs. One subject (the Speaker) described simple pictures, while the other subject (the Performer) tried to reproduce the pictures based on the verbal description as quickly as possible with a set of objects. The Chinese speakers initiated speech production significantly faster than the English speakers. Speech time scaled linearly with ln(ID) in all subjects, but the regression coefficient was significantly higher in the English speakers as compared with the Chinese speakers. The number of errors was somewhat lower in the Chinese participants (not significantly). The Chinese pairs also showed a shorter delay between the initiation of speech and initiation of action by the Performer, shorter movement time by the Performer, and shorter overall performance time. The number of syllables scaled with ID, and the Chinese speakers used significantly smaller numbers of syllables. Speech rate was comparable between the two groups, about 3 syllables/s; it dropped for more complex pictures (higher ID). When asked to reproduce the same pictures without speaking, movement time scaled linearly with ln(ID); the Chinese performers were slower than the English performers. We conclude that natural languages show a speed-difficulty trade-off similar to Fitts’ law; the trade-offs in movement and speech production are likely to originate at a cognitive level. The time advantage of the Chinese participants originates not from similarity of the simple pictures and Chinese written characters and not from more sloppy performance. It is linked to using fewer syllables to transmit the same information. We suggest that natural languages may differ by informational density defined as the amount of information transmitted by a given number of syllables. PMID:21479658

  3. Experience-Related Structural Changes of Degenerated Occipital White Matter in Late-Blind Humans – A Diffusion Tensor Imaging Study

    PubMed Central

    Dietrich, Susanne; Hertrich, Ingo; Kumar, Vinod; Ackermann, Hermann

    2015-01-01

    Late-blind humans can learn to understand speech at ultra-fast syllable rates (ca. 20 syllables/s), a capability associated with hemodynamic activation of the central-visual system. Thus, the observed functional cross-modal recruitment of occipital cortex might facilitate ultra-fast speech processing in these individuals. To further elucidate the structural prerequisites of this skill, diffusion tensor imaging (DTI) was conducted in late-blind subjects differing in their capability of understanding ultra-fast speech. Fractional anisotropy (FA) was determined as a quantitative measure of the directionality of water diffusion, indicating fiber tract characteristics that might be influenced by blindness as well as the acquired perceptual skills. Analysis of the diffusion images revealed reduced FA in late-blind individuals relative to sighted controls at the level of the optic radiations at either side and the right-hemisphere dorsal thalamus (pulvinar). Moreover, late-blind subjects showed significant positive correlations between FA and the capacity of ultra-fast speech comprehension within right-hemisphere optic radiation and thalamus. Thus, experience-related structural alterations occurred in late-blind individuals within visual pathways that, presumably, are linked to higher order frontal language areas. PMID:25830371

  4. Spoken verb processing in Spanish: An analysis using a new online resource

    PubMed Central

    Rivera, Semilla M.; Bates, Elizabeth A.; Orozco-Figueroa, Araceli; Wicha, Nicole Y. Y.

    2012-01-01

    Verbs are one of the basic building blocks of grammar, yet few studies have examined the grammatical, morphological, and phonological factors contributing to lexical access and production of Spanish verb inflection. This report describes an online data set that incorporates psycholinguistic dimensions for 50 of the most common early-acquired Spanish verbs. Using this data set, predictors of response time (RT) from stimulus onset and mean differences at offset are examined. Native Spanish speakers, randomly assigned to one of two tasks, listened to prerecorded verbs and either repeated the verb (single word shadowing) or produced its corresponding pronoun. Factors such as stimulus duration, number of syllables, syllable stress position, and specific levels of initial phoneme facilitated both shadowing of a verb and production of its pronoun. Higher frequency verbs facilitated faster verb repetition, whereas verbs with alternative pronouns increased RT to pronoun production. Mean differences at offset (stimulus duration is removed) indicated that listeners begin speaking earlier when the verb is longer and multisyllabic compared to shorter, monosyllabic words. These results highlight the association between psycholinguistic factors and RT measures of verb processing, in particular, features unique to languages like Spanish, such as alternative pronoun and tense. PMID:23002318

  5. Learning multiple rules simultaneously: Affixes are more salient than reduplications.

    PubMed

    Gervain, Judit; Endress, Ansgar D

    2017-04-01

    Language learners encounter numerous opportunities to learn regularities, but need to decide which of these regularities to learn, because some are not productive in their native language. Here, we present an account of rule learning based on perceptual and memory primitives (Endress, Dehaene-Lambertz, & Mehler, Cognition, 105(3), 577-614, 2007; Endress, Nespor, & Mehler, Trends in Cognitive Sciences, 13(8), 348-353, 2009), suggesting that learners preferentially learn regularities that are more salient to them, and that the pattern of salience reflects the frequency of language features across languages. We contrast this view with previous artificial grammar learning research, which suggests that infants "choose" the regularities they learn based on rational, Bayesian criteria (Frank & Tenenbaum, Cognition, 120(3), 360-371, 2013; Gerken, Cognition, 98(3)B67-B74, 2006, Cognition, 115(2), 362-366, 2010). In our experiments, adult participants listened to syllable strings starting with a syllable reduplication and always ending with the same "affix" syllable, or to syllable strings starting with this "affix" syllable and ending with the "reduplication". Both affixation and reduplication are frequently used for morphological marking across languages. We find three crucial results. First, participants learned both regularities simultaneously. Second, affixation regularities seemed easier to learn than reduplication regularities. Third, regularities in sequence offsets were easier to learn than regularities at sequence onsets. We show that these results are inconsistent with previous Bayesian rule learning models, but mesh well with the perceptual or memory primitives view. Further, we show that the pattern of salience revealed in our experiments reflects the distribution of regularities across languages. Ease of acquisition might thus be one determinant of the frequency of regularities across languages.

  6. Recombinative generalization of within-syllable units in nonreading adults with mental retardation.

    PubMed

    Saunders, Kathryn J; O'Donnell, Jennifer; Vaidya, Manish; Williams, Dean C

    2003-01-01

    Two adults with mental retardation demonstrated the recombination of within-syllable units (onsets and rimes) using a spoken-to-printed-word matching-to-sample (MTS) procedure. Further testing with 1 participant showed comprehension of the printed words. Printed-word naming was minimal before, but greater after, comprehension tests. The findings suggest that these procedures hold promise for further basic and applied analyses of word-attack skills.

  7. The Link between Text Difficulty, Reading Speed and Exploration of Printed Text during Shared Book Reading

    ERIC Educational Resources Information Center

    Roy-Charland, Annie; Perron, Melanie; Turgeon, Krystle-Lee; Hoffman, Nichola; Chamberland, Justin A.

    2016-01-01

    In the current study the reading speed of the narration and the difficulty of the text was manipulated and links were explored with children's attention to the printed text in shared book reading. Thirty-nine children (24 grade 1 and 15 grade 2) were presented easy and difficult books at slow (syllable by syllable) or fast (adult reading speed)…

  8. The mismatch negativity as a measure of auditory stream segregation in a simulated "cocktail-party" scenario: effect of age.

    PubMed

    Getzmann, Stephan; Näätänen, Risto

    2015-11-01

    With age the ability to understand speech in multitalker environments usually deteriorates. The central auditory system has to perceptually segregate and group the acoustic input into sequences of distinct auditory objects. The present study used electrophysiological measures to study effects of age on auditory stream segregation in a multitalker scenario. Younger and older adults were presented with streams of short speech stimuli. When a single target stream was presented, the occurrence of a rare (deviant) syllable among a frequent (standard) syllable elicited the mismatch negativity (MMN), an electrophysiological correlate of automatic deviance detection. The presence of a second, concurrent stream consisting of the deviant syllable of the target stream reduced the MMN amplitude, especially when located nearby the target stream. The decrease in MMN amplitude indicates that the rare syllable of the target stream was less perceived as deviant, suggesting reduced stream segregation with decreasing stream distance. Moreover, the presence of a concurrent stream increased the MMN peak latency of the older group but not that of the younger group. The results provide neurophysiological evidence for the effects of concurrent speech on auditory processing in older adults, suggesting that older adults need more time for stream segregation in the presence of concurrent speech. Copyright © 2015 Elsevier Inc. All rights reserved.

  9. ERP index of the morphological family size effect during word recognition.

    PubMed

    Kwon, Youan; Nam, Kichun; Lee, Yoonhyoung

    2012-12-01

    The purpose of this study was to examine whether the N400 is affected by the semantic richness of associated neighboring word members or by the density of the orthographic syllable neighborhood. Another purpose of this study was to investigate the source of the different LPC in respect to the semantic richness. To do so, the density of the syllable neighborhood and the size of the morphological family of a word were orthogonally manipulated. ERPs from 24 participants were collected during a go/no-go semantic categorization task. The results showed that the N400 effect was mainly influenced by the density of the syllable neighborhood rather than by the morphological family size. The results also showed that words with a larger morphological family size generate significantly larger LPC than words with a smaller morphological family size. The present study did not support the assumption that the main source of the N400 effect is the semantic richness of the associated neighbors. The present results suggest that the N400 is more sensitive to the density of the syllable neighborhood and LPC is sensitive to the density of the semantic neighborhood reflected by the morphological family size. Copyright © 2012 Elsevier Ltd. All rights reserved.

  10. The validity of a simple outcome measure to assess stuttering therapy.

    PubMed

    Huinck, Wendy; Rietveld, Toni

    2007-01-01

    The validity of a simple and not time-consuming self-assessment (SA) Scale was tested to establish progress after or during stuttering therapy. The scores on the SA scale were related to (1) objective measures (percentage of stuttered syllables, and syllables per minute) and (2) (self-)evaluation tests (self-evaluation questionnaires and perceptual evaluations or judgments of disfluency, naturalness and comfort by naïve listeners). Data were collected from two groups of stutterers at four measurement times: pretherapy, posttherapy, 12 months after therapy and 24 months after therapy. The first group attended the Comprehensive Stuttering Program: an integrated program based on fluency shaping techniques, and the second group participated in a Dutch group therapy: the Doetinchem Method that focuses on emotions and cognitions related to stuttering. Results showed similar score patterns on the SA scale, the self-evaluation questionnaires, the objective measures over time, and significant correlations between the SA scale and syllables per minute, percentage of stuttered syllables, Struggle subscale of the Perceptions of Stuttering Inventory and judged fluency on the T1-T2 difference scores. We concluded that the validity of the SA measure was proved and therefore encourage the use of such an instrument when (stuttering) treatment efficacy is studied.

  11. A population memetics approach to cultural evolution in chaffinch song: meme diversity within populations.

    PubMed

    Lynch, A; Baker, A J

    1993-04-01

    We investigated cultural evolution in populations of common chaffinches (Fringilla coelebs) in the Atlantic islands (Azores, Madeira, Canaries) and neighboring continental regions (Morocco, Iberia) by employing a population memetics approach. To quantify variability within populations, we used the concept of a song meme, defined as a single syllable or a series of linked syllables capable of being transmitted. The frequency distribution of memes within populations generally fit a neutral model in which there is an equilibrium between mutation, migration, and drift, which suggests that memes are functionally equivalent. The diversity of memes of single syllables is significantly greater in the Azores compared to all other regions, consistent with higher population densities of chaffinches there. On the other hand, memes of two to five syllables have greater diversity in Atlantic island and Moroccan populations compared to their Iberian counterparts. This higher diversity emanates from a looser syntax and increased recombination in songs, presumably because of relaxed selection for distinctive songs in these peripheral and depauperate avifaunas. We urge comparative population memetic studies of other species of songbirds and predict that they will lead to a formulation of a general theory for the cultural evolution of bird song analogous to population genetics theory for biological traits.

  12. English Listeners Use Suprasegmental Cues to Lexical Stress Early During Spoken-Word Recognition

    PubMed Central

    Poellmann, Katja; Kong, Ying-Yee

    2017-01-01

    Purpose We used an eye-tracking technique to investigate whether English listeners use suprasegmental information about lexical stress to speed up the recognition of spoken words in English. Method In a visual world paradigm, 24 young English listeners followed spoken instructions to choose 1 of 4 printed referents on a computer screen (e.g., “Click on the word admiral”). Displays contained a critical pair of words (e.g., ˈadmiral–ˌadmiˈration) that were segmentally identical for their first 2 syllables but differed suprasegmentally in their 1st syllable: One word began with primary lexical stress, and the other began with secondary lexical stress. All words had phrase-level prominence. Listeners' relative proportion of eye fixations on these words indicated their ability to differentiate them over time. Results Before critical word pairs became segmentally distinguishable in their 3rd syllables, participants fixated target words more than their stress competitors, but only if targets had initial primary lexical stress. The degree to which stress competitors were fixated was independent of their stress pattern. Conclusions Suprasegmental information about lexical stress modulates the time course of spoken-word recognition. Specifically, suprasegmental information on the primary-stressed syllable of words with phrase-level prominence helps in distinguishing the word from phonological competitors with secondary lexical stress. PMID:28056135

  13. Perceptual-center modeling is affected by including acoustic rate-of-change modulations.

    PubMed

    Harsin, C A

    1997-02-01

    This study investigated the acoustic correlates of perceptual centers (p-centers) in CV and VC syllables and developed an acoustic p-center model. In Part 1, listeners located syllables' p-centers by a method-of-adjustment procedure. The CV syllables contained the consonants /s/,/r/,/n/,/t/,/d/,/k/, and /g/; the VCs, the consonants /s/,/r/, and /n/. The vowel in all syllables was /a/. The results of this experiment replicated and extended previous findings regarding the effects of phonetic variation on p-centers. In Part 2, a digital signal processing procedure was used to acoustically model p-center perception. Each stimulus was passed through a six-band digital filter, and the outputs were processed to derive low-frequency modulation components. These components were weighted according to a perceived modulation magnitude function and recombined to create six psychoacoustic envelopes containing modulation energies from 3 to 47 Hz. In this analysis, p-centers were found to be highly correlated with the time-weighted function of the rate-of-change in the psychoacoustic envelopes, multiplied by the psychoacoustic envelope magnitude increment. The results were interpreted as suggesting (1) the probable role of low-frequency energy modulations in p-center perception, and (2) the presence of perceptual processes that integrate multiple articulatory events into a single syllabic event.

  14. Perception and the Temporal Properties of Speech.

    DTIC Science & Technology

    1993-01-11

    conditions. In the embedded condition, phoneme sequences equivalent to these words formed the second syllable of a two-syllable word. In the unembedded ... unembedded in the sequence "warm lips". These priming sequences were based on the sequences used in Experiment 2. Each combinable priming sequence in...unrelated, to the embedded or unembedded prime word. The probes used in this experiment were identical to the ones used in Experiment 2. Subjects were tested

  15. The Whole Is More than the Sum of Its Parts--Audiovisual Processing of Phonemes Investigated with ERPs

    ERIC Educational Resources Information Center

    Hessler, Dorte; Jonkers, Roel; Stowe, Laurie; Bastiaanse, Roelien

    2013-01-01

    In the current ERP study, an active oddball task was carried out, testing pure tones and auditory, visual and audiovisual syllables. For pure tones, an MMN, an N2b, and a P3 were found, confirming traditional findings. Auditory syllables evoked an N2 and a P3. We found that the amplitude of the P3 depended on the distance between standard and…

  16. Sight-Singing Scores of High School Choristers with Extensive Training in Movable Solfege Syllables and Curwen Hand Signs

    ERIC Educational Resources Information Center

    McClung, Alan C.

    2008-01-01

    Randomly chosen high school choristers with extensive training in solfege syllables and Curwen hand signs (N = 38) are asked to sight-sing two melodies, one while using Curwen hand signs and the other without. Out of a perfect score of 16, the mean score with hand signs was 10.37 (SD = 4.23), and without hand signs, 10.84 (SD = 3.96). A…

  17. A new, morphologically cryptic bush-cricket discovered on the basis of its song in the Carpathian Mountains (Insecta, Orthoptera, Tettigoniidae).

    PubMed

    Iorgu, Ionuţ Ştefan; Iorgu, Elena Iulia; Szövényi, Gergely; Orci, Kirill Márk

    2017-01-01

    A new, morphologically cryptic species of phaneropterine bush-crickets is described from the grasslands of the Romanian Eastern Carpathians. Despite the morphological and acoustic similarities with the recently described Isophya nagyi Szövényi, Puskás & Orci, I. bucovinensis sp. n. is characterized by a peculiar male calling song, with faster syllable repetition rate (160-220 syllables per minute, at 22-27°C) and less complex syllable structure (composed of only two elements instead of three observable in I. nagyi ). The morphological description of the new species is supplemented with an oscillographic and spectrographic analysis of the male calling song and male-female pair-forming acoustic duet. An acoustic signal-based identification key is provided for all the presently known species of the Isophya camptoxypha species group, including the new species.

  18. A new, morphologically cryptic bush-cricket discovered on the basis of its song in the Carpathian Mountains (Insecta, Orthoptera, Tettigoniidae)

    PubMed Central

    Iorgu, Ionuţ Ştefan; Iorgu, Elena Iulia; Szövényi, Gergely; Orci, Kirill Márk

    2017-01-01

    Abstract A new, morphologically cryptic species of phaneropterine bush-crickets is described from the grasslands of the Romanian Eastern Carpathians. Despite the morphological and acoustic similarities with the recently described Isophya nagyi Szövényi, Puskás & Orci, I. bucovinensis sp. n. is characterized by a peculiar male calling song, with faster syllable repetition rate (160–220 syllables per minute, at 22–27°C) and less complex syllable structure (composed of only two elements instead of three observable in I. nagyi). The morphological description of the new species is supplemented with an oscillographic and spectrographic analysis of the male calling song and male–female pair-forming acoustic duet. An acoustic signal-based identification key is provided for all the presently known species of the Isophya camptoxypha species group, including the new species. PMID:28769716

  19. Surprise! Infants consider possible bases of generalization for a single input example.

    PubMed

    Gerken, LouAnn; Dawson, Colin; Chatila, Razanne; Tenenbaum, Josh

    2015-01-01

    Infants have been shown to generalize from a small number of input examples. However, existing studies allow two possible means of generalization. One is via a process of noting similarities shared by several examples. Alternatively, generalization may reflect an implicit desire to explain the input. The latter view suggests that generalization might occur when even a single input example is surprising, given the learner's current model of the domain. To test the possibility that infants are able to generalize based on a single example, we familiarized 9-month-olds with a single three-syllable input example that contained either one surprising feature (syllable repetition, Experiment 1) or two features (repetition and a rare syllable, Experiment 2). In both experiments, infants generalized only to new strings that maintained all of the surprising features from familiarization. This research suggests that surprise can promote very rapid generalization. © 2014 John Wiley & Sons Ltd.

  20. The influence of musical experience on lateralisation of auditory processing.

    PubMed

    Spajdel, Marián; Jariabková, Katarína; Riecanský, Igor

    2007-11-01

    The influence of musical experience on free-recall dichotic listening to environmental sounds, two-tone sequences, and consonant-vowel (CV) syllables was investigated. A total of 60 healthy right-handed participants were divided into two groups according to their active musical competence ("musicians" and "non-musicians"). In both groups, we found a left ear advantage (LEA) for nonverbal stimuli (environmental sounds and two-tone sequences) and a right ear advantage (REA) for CV syllables. Dichotic listening to environmental sounds was uninfluenced by musical experience. The total accuracy of recall for two-tone sequences was higher in musicians than in non-musicians but the lateralisation was similar in both groups. For CV syllables a lower REA was found in male but not female musicians in comparison to non-musicians. The results indicate a specific sex-dependent effect of musical experience on lateralisation of phonological auditory processing.

  1. The stop voicing contrast in French: From citation speech to sentencial speech

    NASA Astrophysics Data System (ADS)

    Abdelli-Beruh, Nassima; Demaio, Eileen; Hisagi, Miwako

    2004-05-01

    This study explores the influence of speaking style on the salience of the acoustic correlates to the stop voicing distinction in French. Monolingual French speakers produced twenty-one C_vC_ syllables in citation speech, in minimal pairs and in sentence-length utterances (/pa/-/a/ context: /il a di pa C_vC_ a lui/; /pas/-/s/ context: /il a di pas C_vC_ sa~ lui/). Prominent stress was on the C_vC_. Voicing-related differences in percentages of closure voicing, durations of aspiration, closure, and vowel were analyzed as a function of these three speaking styles. Results show that the salience of the acoustic-phonetic segments present when the syllables are uttered in isolation or in minimal pairs is different than when the syllables are spoken in a sentence. These results are in agreement with findings in English.

  2. Word length, set size, and lexical factors: Re-examining what causes the word length effect.

    PubMed

    Guitard, Dominic; Gabel, Andrew J; Saint-Aubin, Jean; Surprenant, Aimée M; Neath, Ian

    2018-04-19

    The word length effect, better recall of lists of short (fewer syllables) than long (more syllables) words has been termed a benchmark effect of working memory. Despite this, experiments on the word length effect can yield quite different results depending on set size and stimulus properties. Seven experiments are reported that address these 2 issues. Experiment 1 replicated the finding of a preserved word length effect under concurrent articulation for large stimulus sets, which contrasts with the abolition of the word length effect by concurrent articulation for small stimulus sets. Experiment 2, however, demonstrated that when the short and long words are equated on more dimensions, concurrent articulation abolishes the word length effect for large stimulus sets. Experiment 3 shows a standard word length effect when output time is equated, but Experiments 4-6 show no word length effect when short and long words are equated on increasingly more dimensions that previous demonstrations have overlooked. Finally, Experiment 7 compared recall of a small and large neighborhood words that were equated on all the dimensions used in Experiment 6 (except for those directly related to neighborhood size) and a neighborhood size effect was still observed. We conclude that lexical factors, rather than word length per se, are better predictors of when the word length effect will occur. (PsycINFO Database Record (c) 2018 APA, all rights reserved).

  3. Factors that enhance English-speaking speech-language pathologists' transcription of Cantonese-speaking children's consonants.

    PubMed

    Lockart, Rebekah; McLeod, Sharynne

    2013-08-01

    To investigate speech-language pathology students' ability to identify errors and transcribe typical and atypical speech in Cantonese, a nonnative language. Thirty-three English-speaking speech-language pathology students completed 3 tasks in an experimental within-subjects design. Task 1 (baseline) involved transcribing English words. In Task 2, students transcribed 25 words spoken by a Cantonese adult. An average of 59.1% consonants was transcribed correctly (72.9% when Cantonese-English transfer patterns were allowed). There was higher accuracy on shared English and Cantonese syllable-initial consonants /m,n,f,s,h,j,w,l/ and syllable-final consonants. In Task 3, students identified consonant errors and transcribed 100 words spoken by Cantonese-speaking children under 4 additive conditions: (1) baseline, (2) +adult model, (3) +information about Cantonese phonology, and (4) all variables (2 and 3 were counterbalanced). There was a significant improvement in the students' identification and transcription scores for conditions 2, 3, and 4, with a moderate effect size. Increased skill was not based on listeners' proficiency in speaking another language, perceived transcription skill, musicality, or confidence with multilingual clients. Speech-language pathology students, with no exposure to or specific training in Cantonese, have some skills to identify errors and transcribe Cantonese. Provision of a Cantonese-adult model and information about Cantonese phonology increased students' accuracy in transcribing Cantonese speech.

  4. Prosodic Similarity Effects in Short-Term Memory in Developmental Dyslexia.

    PubMed

    Goswami, Usha; Barnes, Lisa; Mead, Natasha; Power, Alan James; Leong, Victoria

    2016-11-01

    Children with developmental dyslexia are characterized by phonological difficulties across languages. Classically, this 'phonological deficit' in dyslexia has been investigated with tasks using single-syllable words. Recently, however, several studies have demonstrated difficulties in prosodic awareness in dyslexia. Potential prosodic effects in short-term memory have not yet been investigated. Here we create a new instrument based on three-syllable words that vary in stress patterns, to investigate whether prosodic similarity (the same prosodic pattern of stressed and unstressed syllables) exerts systematic effects on short-term memory. We study participants with dyslexia and age-matched and younger reading-level-matched typically developing controls. We find that all participants, including dyslexic participants, show prosodic similarity effects in short-term memory. All participants exhibited better retention of words that differed in prosodic structure, although participants with dyslexia recalled fewer words accurately overall compared to age-matched controls. Individual differences in prosodic memory were predicted by earlier vocabulary abilities, by earlier sensitivity to syllable stress and by earlier phonological awareness. To our knowledge, this is the first demonstration of prosodic similarity effects in short-term memory. The implications of a prosodic similarity effect for theories of lexical representation and of dyslexia are discussed. © 2016 The Authors. Dyslexia published by John Wiley & Sons Ltd. © 2016 The Authors. Dyslexia published by John Wiley & Sons Ltd.

  5. Infant vocalizations and the early diagnosis of severe hearing impairment.

    PubMed

    Eilers, R E; Oller, D K

    1994-02-01

    To determine whether late onset of canonical babbling could be used as a criterion to determine risk of hearing impairment, we obtained vocalization samples longitudinally from 94 infants with normal hearing and 37 infants with severe to profound hearing impairment. Parents were instructed to report the onset of canonical babbling (the production of well-formed syllables such as "da," "na," "bee," "yaya"). Verification that the infants were producing canonical syllables was collected in laboratory audio recordings. Infants with normal hearing produced canonical vocalizations before 11 months of age (range, 3 to 10 months; mode, 7 months); infants who were deaf failed to produce canonical syllables until 11 months of age or older, often well into the third year of life (range, 11 to 49 months; mode, 24 months). The correlation between age at onset of the canonical stage and age at auditory amplification was 0.68, indicating that early identification and fitting of hearing aids is of significant benefit to infants learning language. The fact that there is no overlap in the distribution of the onset of canonical babbling between infants with normal hearing and infants with hearing impairment means that the failure of otherwise healthy infants to produce canonical syllables before 11 months of age should be considered a serious risk factor for hearing impairment and, when observed, should result in immediate referral for audiologic evaluation.

  6. A Neural Basis for Interindividual Differences in the McGurk Effect, a Multisensory Speech Illusion

    PubMed Central

    Nath, Audrey R.; Beauchamp, Michael S.

    2011-01-01

    The McGurk effect is a compelling illusion in which humans perceive mismatched audiovisual speech as a completely different syllable. However, some normal individuals do not experience the illusion, reporting that the stimulus sounds the same with or without visual input. Converging evidence suggests that the left superior temporal sulcus (STS) is critical for audiovisual integration during speech perception. We used blood-oxygen level dependent functional magnetic resonance imaging (BOLD fMRI) to measure brain activity as McGurk perceivers and non-perceivers were presented with congruent audiovisual syllables, McGurk audiovisual syllables, and non-McGurk incongruent syllables. The inferior frontal gyrus showed an effect of stimulus condition (greater responses for incongruent stimuli) but not susceptibility group, while the left auditory cortex showed an effect of susceptibility group (greater response in susceptible individuals) but not stimulus condition. Only one brain region, the left STS, showed a significant effect of both susceptibility and stimulus condition. The amplitude of the response in the left STS was significantly correlated with the likelihood of perceiving the McGurk effect: a weak STS response meant that a subject was less likely to perceive the McGurk effect, while a strong response meant that a subject was more likely to perceive it. These results suggest that the left STS is a key locus for interindividual differences in speech perception. PMID:21787869

  7. Vocal exploration is locally regulated during song learning

    PubMed Central

    Ravbar, Primoz; Parra, Lucas C.; Lipkind, Dina; Tchernichovski, Ofer

    2012-01-01

    Exploratory variability is essential for sensory-motor learning, but it is not known how and at what time scales it is regulated. We manipulated song learning in zebra finches to experimentally control the requirements for vocal exploration in different parts of their song. We first trained birds to perform a one-syllable song, and once they mastered it we added a new syllable to the song model. Remarkably, when practicing the modified song, birds rapidly alternated between high and low acoustic variability to confine vocal exploration to the newly added syllable. Further, even within syllables, acoustic variability changed independently across song elements that were only milliseconds apart. Analysis of the entire vocal output during learning revealed that the variability of each song element decreased as it approached the target, correlating with momentary local distance from the target and less so with the overall distance. We conclude that vocal error is computed locally in sub-syllabic time scales and that song elements can be learned and crystalized independently. Songbirds have dedicated brain circuitry for vocal babbling in the anterior forebrain pathway (AFP), which generates exploratory song patterns that drive premotor neurons at the song nucleus RA (robust nucleus of the arcopallium). We hypothesize that either AFP adjusts the gain of vocal exploration in fine time scales, or that the sensitivity of RA premotor neurons to AFP/HVC inputs varies across song elements. PMID:22399765

  8. Within- and across-language spectral and temporal variability of vowels in different phonetic and prosodic contexts: Russian and Japanese

    NASA Astrophysics Data System (ADS)

    Gilichinskaya, Yana D.; Hisagi, Miwako; Law, Franzo F.; Berkowitz, Shari; Ito, Kikuyo

    2005-04-01

    Contextual variability of vowels in three languages with large vowel inventories was examined previously. Here, variability of vowels in two languages with small inventories (Russian, Japanese) was explored. Vowels were produced by three female speakers of each language in four contexts: (Vba) disyllables and in 3-syllable nonsense words (gaC1VC2a) embedded within carrier sentences; contexts included bilabial stops (bVp) in normal rate sentences and alveolar stops (dVt) in both normal and rapid rate sentences. Dependent variables were syllable durations and formant frequencies at syllable midpoint. Results showed very little variation across consonant and rate conditions in formants for /i/ in both languages. Japanese short /u, o, a/ showed fronting (F2 increases) in alveolar context relative to labial context (1.3-2.0 Barks), which was more pronounced in rapid sentences. Fronting of Japanese long vowels was less pronounced (0.3 to 0.9 Barks). Japanese long/short vowel ratios varied with speaking style (syllables versus sentences) and speaking rate. All Russian vowels except /i/ were fronted in alveolar vs labial context (1.1-3.1 Barks) but showed little change in either spectrum or duration with speaking rate. Comparisons of these patterns of variability with American English, French and German vowel results will be discussed.

  9. A Study of Mexican Free-Tailed Bat Chirp Syllables: Bayesian Functional Mixed Models for Nonstationary Acoustic Time Series

    PubMed Central

    MARTINEZ, Josue G.; BOHN, Kirsten M.; CARROLL, Raymond J.

    2013-01-01

    We describe a new approach to analyze chirp syllables of free-tailed bats from two regions of Texas in which they are predominant: Austin and College Station. Our goal is to characterize any systematic regional differences in the mating chirps and assess whether individual bats have signature chirps. The data are analyzed by modeling spectrograms of the chirps as responses in a Bayesian functional mixed model. Given the variable chirp lengths, we compute the spectrograms on a relative time scale interpretable as the relative chirp position, using a variable window overlap based on chirp length. We use 2D wavelet transforms to capture correlation within the spectrogram in our modeling and obtain adaptive regularization of the estimates and inference for the regions-specific spectrograms. Our model includes random effect spectrograms at the bat level to account for correlation among chirps from the same bat, and to assess relative variability in chirp spectrograms within and between bats. The modeling of spectrograms using functional mixed models is a general approach for the analysis of replicated nonstationary time series, such as our acoustical signals, to relate aspects of the signals to various predictors, while accounting for between-signal structure. This can be done on raw spectrograms when all signals are of the same length, and can be done using spectrograms defined on a relative time scale for signals of variable length in settings where the idea of defining correspondence across signals based on relative position is sensible. PMID:23997376

  10. Mapping the cortical representation of speech sounds in a syllable repetition task.

    PubMed

    Markiewicz, Christopher J; Bohland, Jason W

    2016-11-01

    Speech repetition relies on a series of distributed cortical representations and functional pathways. A speaker must map auditory representations of incoming sounds onto learned speech items, maintain an accurate representation of those items in short-term memory, interface that representation with the motor output system, and fluently articulate the target sequence. A "dorsal stream" consisting of posterior temporal, inferior parietal and premotor regions is thought to mediate auditory-motor representations and transformations, but the nature and activation of these representations for different portions of speech repetition tasks remains unclear. Here we mapped the correlates of phonetic and/or phonological information related to the specific phonemes and syllables that were heard, remembered, and produced using a series of cortical searchlight multi-voxel pattern analyses trained on estimates of BOLD responses from individual trials. Based on responses linked to input events (auditory syllable presentation), predictive vowel-level information was found in the left inferior frontal sulcus, while syllable prediction revealed significant clusters in the left ventral premotor cortex and central sulcus and the left mid superior temporal sulcus. Responses linked to output events (the GO signal cueing overt production) revealed strong clusters of vowel-related information bilaterally in the mid to posterior superior temporal sulcus. For the prediction of onset and coda consonants, input-linked responses yielded distributed clusters in the superior temporal cortices, which were further informative for classifiers trained on output-linked responses. Output-linked responses in the Rolandic cortex made strong predictions for the syllables and consonants produced, but their predictive power was reduced for vowels. The results of this study provide a systematic survey of how cortical response patterns covary with the identity of speech sounds, which will help to constrain and guide theoretical models of speech perception, speech production, and phonological working memory. Copyright © 2016 Elsevier Inc. All rights reserved.

  11. Acoustic-Emergent Phonology in the Amplitude Envelope of Child-Directed Speech

    PubMed Central

    Leong, Victoria; Goswami, Usha

    2015-01-01

    When acquiring language, young children may use acoustic spectro-temporal patterns in speech to derive phonological units in spoken language (e.g., prosodic stress patterns, syllables, phonemes). Children appear to learn acoustic-phonological mappings rapidly, without direct instruction, yet the underlying developmental mechanisms remain unclear. Across different languages, a relationship between amplitude envelope sensitivity and phonological development has been found, suggesting that children may make use of amplitude modulation (AM) patterns within the envelope to develop a phonological system. Here we present the Spectral Amplitude Modulation Phase Hierarchy (S-AMPH) model, a set of algorithms for deriving the dominant AM patterns in child-directed speech (CDS). Using Principal Components Analysis, we show that rhythmic CDS contains an AM hierarchy comprising 3 core modulation timescales. These timescales correspond to key phonological units: prosodic stress (Stress AM, ~2 Hz), syllables (Syllable AM, ~5 Hz) and onset-rime units (Phoneme AM, ~20 Hz). We argue that these AM patterns could in principle be used by naïve listeners to compute acoustic-phonological mappings without lexical knowledge. We then demonstrate that the modulation statistics within this AM hierarchy indeed parse the speech signal into a primitive hierarchically-organised phonological system comprising stress feet (proto-words), syllables and onset-rime units. We apply the S-AMPH model to two other CDS corpora, one spontaneous and one deliberately-timed. The model accurately identified 72–82% (freely-read CDS) and 90–98% (rhythmically-regular CDS) stress patterns, syllables and onset-rime units. This in-principle demonstration that primitive phonology can be extracted from speech AMs is termed Acoustic-Emergent Phonology (AEP) theory. AEP theory provides a set of methods for examining how early phonological development is shaped by the temporal modulation structure of speech across languages. The S-AMPH model reveals a crucial developmental role for stress feet (AMs ~2 Hz). Stress feet underpin different linguistic rhythm typologies, and speech rhythm underpins language acquisition by infants in all languages. PMID:26641472

  12. Software for objective comparison of vocal acoustic features over weeks of audio recording: KLFromRecordingDays

    NASA Astrophysics Data System (ADS)

    Soderstrom, Ken; Alalawi, Ali

    KLFromRecordingDays allows measurement of Kullback-Leibler (KL) distances between 2D probability distributions of vocal acoustic features. Greater KL distance measures reflect increased phonological divergence across the vocalizations compared. The software has been used to compare *.wav file recordings made by Sound Analysis Recorder 2011 of songbird vocalizations pre- and post-drug and surgical manipulations. Recordings from individual animals in *.wav format are first organized into subdirectories by recording day and then segmented into individual syllables uttered and acoustic features of these syllables using Sound Analysis Pro 2011 (SAP). KLFromRecordingDays uses syllable acoustic feature data output by SAP to a MySQL table to generate and compare "template" (typically pre-treatment) and "target" (typically post-treatment) probability distributions. These distributions are a series of virtual 2D plots of the duration of each syllable (as x-axis) to each of 13 other acoustic features measured by SAP for that syllable (as y-axes). Differences between "template" and "target" probability distributions for each acoustic feature are determined by calculating KL distance, a measure of divergence of the target 2D distribution pattern from that of the template. KL distances and the mean KL distance across all acoustic features are calculated for each recording day and output to an Excel spreadsheet. Resulting data for individual subjects may then be pooled across treatment groups and graphically summarized and used for statistical comparisons. Because SAP-generated MySQL files are accessed directly, data limits associated with spreadsheet output are avoided, and the totality of vocal output over weeks may be objectively analyzed all at once. The software has been useful for measuring drug effects on songbird vocalizations and assessing recovery from damage to regions of vocal motor cortex. It may be useful in studies employing other species, and as part of speech therapies tracking progress in producing distinct speech sounds in isolation.

  13. Why Early Tactile Speech Aids May Have Failed: No Perceptual Integration of Tactile and Auditory Signals.

    PubMed

    Rizza, Aurora; Terekhov, Alexander V; Montone, Guglielmo; Olivetti-Belardinelli, Marta; O'Regan, J Kevin

    2018-01-01

    Tactile speech aids, though extensively studied in the 1980's and 1990's, never became a commercial success. A hypothesis to explain this failure might be that it is difficult to obtain true perceptual integration of a tactile signal with information from auditory speech: exploitation of tactile cues from a tactile aid might require cognitive effort and so prevent speech understanding at the high rates typical of everyday speech. To test this hypothesis, we attempted to create true perceptual integration of tactile with auditory information in what might be considered the simplest situation encountered by a hearing-impaired listener. We created an auditory continuum between the syllables /BA/ and /VA/, and trained participants to associate /BA/ to one tactile stimulus and /VA/ to another tactile stimulus. After training, we tested if auditory discrimination along the continuum between the two syllables could be biased by incongruent tactile stimulation. We found that such a bias occurred only when the tactile stimulus was above, but not when it was below its previously measured tactile discrimination threshold. Such a pattern is compatible with the idea that the effect is due to a cognitive or decisional strategy, rather than to truly perceptual integration. We therefore ran a further study (Experiment 2), where we created a tactile version of the McGurk effect. We extensively trained two Subjects over 6 days to associate four recorded auditory syllables with four corresponding apparent motion tactile patterns. In a subsequent test, we presented stimulation that was either congruent or incongruent with the learnt association, and asked Subjects to report the syllable they perceived. We found no analog to the McGurk effect, suggesting that the tactile stimulation was not being perceptually integrated with the auditory syllable. These findings strengthen our hypothesis according to which tactile aids failed because integration of tactile cues with auditory speech occurred at a cognitive or decisional level, rather than truly at a perceptual level.

  14. Acoustic-Emergent Phonology in the Amplitude Envelope of Child-Directed Speech.

    PubMed

    Leong, Victoria; Goswami, Usha

    2015-01-01

    When acquiring language, young children may use acoustic spectro-temporal patterns in speech to derive phonological units in spoken language (e.g., prosodic stress patterns, syllables, phonemes). Children appear to learn acoustic-phonological mappings rapidly, without direct instruction, yet the underlying developmental mechanisms remain unclear. Across different languages, a relationship between amplitude envelope sensitivity and phonological development has been found, suggesting that children may make use of amplitude modulation (AM) patterns within the envelope to develop a phonological system. Here we present the Spectral Amplitude Modulation Phase Hierarchy (S-AMPH) model, a set of algorithms for deriving the dominant AM patterns in child-directed speech (CDS). Using Principal Components Analysis, we show that rhythmic CDS contains an AM hierarchy comprising 3 core modulation timescales. These timescales correspond to key phonological units: prosodic stress (Stress AM, ~2 Hz), syllables (Syllable AM, ~5 Hz) and onset-rime units (Phoneme AM, ~20 Hz). We argue that these AM patterns could in principle be used by naïve listeners to compute acoustic-phonological mappings without lexical knowledge. We then demonstrate that the modulation statistics within this AM hierarchy indeed parse the speech signal into a primitive hierarchically-organised phonological system comprising stress feet (proto-words), syllables and onset-rime units. We apply the S-AMPH model to two other CDS corpora, one spontaneous and one deliberately-timed. The model accurately identified 72-82% (freely-read CDS) and 90-98% (rhythmically-regular CDS) stress patterns, syllables and onset-rime units. This in-principle demonstration that primitive phonology can be extracted from speech AMs is termed Acoustic-Emergent Phonology (AEP) theory. AEP theory provides a set of methods for examining how early phonological development is shaped by the temporal modulation structure of speech across languages. The S-AMPH model reveals a crucial developmental role for stress feet (AMs ~2 Hz). Stress feet underpin different linguistic rhythm typologies, and speech rhythm underpins language acquisition by infants in all languages.

  15. Why Early Tactile Speech Aids May Have Failed: No Perceptual Integration of Tactile and Auditory Signals

    PubMed Central

    Rizza, Aurora; Terekhov, Alexander V.; Montone, Guglielmo; Olivetti-Belardinelli, Marta; O’Regan, J. Kevin

    2018-01-01

    Tactile speech aids, though extensively studied in the 1980’s and 1990’s, never became a commercial success. A hypothesis to explain this failure might be that it is difficult to obtain true perceptual integration of a tactile signal with information from auditory speech: exploitation of tactile cues from a tactile aid might require cognitive effort and so prevent speech understanding at the high rates typical of everyday speech. To test this hypothesis, we attempted to create true perceptual integration of tactile with auditory information in what might be considered the simplest situation encountered by a hearing-impaired listener. We created an auditory continuum between the syllables /BA/ and /VA/, and trained participants to associate /BA/ to one tactile stimulus and /VA/ to another tactile stimulus. After training, we tested if auditory discrimination along the continuum between the two syllables could be biased by incongruent tactile stimulation. We found that such a bias occurred only when the tactile stimulus was above, but not when it was below its previously measured tactile discrimination threshold. Such a pattern is compatible with the idea that the effect is due to a cognitive or decisional strategy, rather than to truly perceptual integration. We therefore ran a further study (Experiment 2), where we created a tactile version of the McGurk effect. We extensively trained two Subjects over 6 days to associate four recorded auditory syllables with four corresponding apparent motion tactile patterns. In a subsequent test, we presented stimulation that was either congruent or incongruent with the learnt association, and asked Subjects to report the syllable they perceived. We found no analog to the McGurk effect, suggesting that the tactile stimulation was not being perceptually integrated with the auditory syllable. These findings strengthen our hypothesis according to which tactile aids failed because integration of tactile cues with auditory speech occurred at a cognitive or decisional level, rather than truly at a perceptual level. PMID:29875719

  16. Oral-diadochokinesis rates across languages: English and Hebrew norms.

    PubMed

    Icht, Michal; Ben-David, Boaz M

    2014-01-01

    Oro-facial and speech motor control disorders represent a variety of speech and language pathologies. Early identification of such problems is important and carries clinical implications. A common and simple tool for gauging the presence and severity of speech motor control impairments is oral-diadochokinesis (oral-DDK). Surprisingly, norms for adult performance are missing from the literature. The goals of this study were: (1) to establish a norm for oral-DDK rate for (young to middle-age) adult English speakers, by collecting data from the literature (five studies, N=141); (2) to investigate the possible effect of language (and culture) on oral-DDK performance, by analyzing studies conducted in other languages (five studies, N=140), alongside the English norm; and (3) to find a new norm for adult Hebrew speakers, by testing 115 speakers. We first offer an English norm with a mean of 6.2syllables/s (SD=.8), and a lower boundary of 5.4syllables/s that can be used to indicate possible abnormality. Next, we found significant differences between four tested languages (English, Portuguese, Farsi and Greek) in oral-DDK rates. Results suggest the need to set language and culture sensitive norms for the application of the oral-DDK task world-wide. Finally, we found the oral-DDK performance for adult Hebrew speakers to be 6.4syllables/s (SD=.8), not significantly different than the English norms. This implies possible phonological similarities between English and Hebrew. We further note that no gender effects were found in our study. We recommend using oral-DDK as an important tool in the speech language pathologist's arsenal. Yet, application of this task should be done carefully, comparing individual performance to a set norm within the specific language. Readers will be able to: (1) identify the Speech-Language Pathologist assessment process using the oral-DDK task, by comparing an individual performance to the present English norm, (2) describe the impact of language on oral-DDK performance, and (3) accurately detect Hebrew speakers' patients using this tool. Copyright © 2014 Elsevier Inc. All rights reserved.

  17. Speech recognition: Acoustic-phonetic knowledge acquisition and representation

    NASA Astrophysics Data System (ADS)

    Zue, Victor W.

    1988-09-01

    The long-term research goal is to develop and implement speaker-independent continuous speech recognition systems. It is believed that the proper utilization of speech-specific knowledge is essential for such advanced systems. This research is thus directed toward the acquisition, quantification, and representation, of acoustic-phonetic and lexical knowledge, and the application of this knowledge to speech recognition algorithms. In addition, we are exploring new speech recognition alternatives based on artificial intelligence and connectionist techniques. We developed a statistical model for predicting the acoustic realization of stop consonants in various positions in the syllable template. A unification-based grammatical formalism was developed for incorporating this model into the lexical access algorithm. We provided an information-theoretic justification for the hierarchical structure of the syllable template. We analyzed segmented duration for vowels and fricatives in continuous speech. Based on contextual information, we developed durational models for vowels and fricatives that account for over 70 percent of the variance, using data from multiple, unknown speakers. We rigorously evaluated the ability of human spectrogram readers to identify stop consonants spoken by many talkers and in a variety of phonetic contexts. Incorporating the declarative knowledge used by the readers, we developed a knowledge-based system for stop identification. We achieved comparable system performance to that to the readers.

  18. Some phonatory and resonatory characteristics of the rock, pop, soul, and Swedish dance band styles of singing.

    PubMed

    Borch, D Zangger; Sundberg, Johan

    2011-09-01

    This investigation aims at describing voice function of four nonclassical styles of singing, Rock, Pop, Soul, and Swedish Dance Band. A male singer, professionally experienced in performing in these genres, sang representative tunes, both with their original lyrics and on the syllable /pae/. In addition, he sang tones in a triad pattern ranging from the pitch Bb2 to the pitch C4 on the syllable /pae/ in pressed and neutral phonation. An expert panel was successful in classifying the samples, thus suggesting that the samples were representative of the various styles. Subglottal pressure was estimated from oral pressure during the occlusion for the consonant [p]. Flow glottograms were obtained from inverse filtering. The four lowest formant frequencies differed between the styles. The mean of the subglottal pressure and the mean of the normalized amplitude quotient (NAQ), that is, the ratio between the flow pulse amplitude and the product of period and maximum flow declination rate, were plotted against the mean of fundamental frequency. In these graphs, Rock and Swedish Dance Band assumed opposite extreme positions with respect to subglottal pressure and mean phonation frequency, whereas the mean NAQ values differed less between the styles. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  19. Dynamic spectral structure specifies vowels for children and adultsa

    PubMed Central

    Nittrouer, Susan

    2008-01-01

    When it comes to making decisions regarding vowel quality, adults seem to weight dynamic syllable structure more strongly than static structure, although disagreement exists over the nature of the most relevant kind of dynamic structure: spectral change intrinsic to the vowel or structure arising from movements between consonant and vowel constrictions. Results have been even less clear regarding the signal components children use in making vowel judgments. In this experiment, listeners of four different ages (adults, and 3-, 5-, and 7-year-old children) were asked to label stimuli that sounded either like steady-state vowels or like CVC syllables which sometimes had middle sections masked by coughs. Four vowel contrasts were used, crossed for type (front/back or closed/open) and consonant context (strongly or only slightly constraining of vowel tongue position). All listeners recognized vowel quality with high levels of accuracy in all conditions, but children were disproportionately hampered by strong coarticulatory effects when only steady-state formants were available. Results clarified past studies, showing that dynamic structure is critical to vowel perception for all aged listeners, but particularly for young children, and that it is the dynamic structure arising from vocal-tract movement between consonant and vowel constrictions that is most important. PMID:17902868

  20. Background noise can enhance cortical auditory evoked potentials under certain conditions

    PubMed Central

    Papesh, Melissa A.; Billings, Curtis J.; Baltzell, Lucas S.

    2017-01-01

    Objective To use cortical auditory evoked potentials (CAEPs) to understand neural encoding in background noise and the conditions under which noise enhances CAEP responses. Methods CAEPs from 16 normal-hearing listeners were recorded using the speech syllable/ba/presented in quiet and speech-shaped noise at signal-to-noise ratios of 10 and 30 dB. The syllable was presented binaurally and monaurally at two presentation rates. Results The amplitudes of N1 and N2 peaks were often significantly enhanced in the presence of low-level background noise relative to quiet conditions, while P1 and P2 amplitudes were consistently reduced in noise. P1 and P2 amplitudes were significantly larger during binaural compared to monaural presentations, while N1 and N2 peaks were similar between binaural and monaural conditions. Conclusions Methodological choices impact CAEP peaks in very different ways. Negative peaks can be enhanced by background noise in certain conditions, while positive peaks are generally enhanced by binaural presentations. Significance Methodological choices significantly impact CAEPs acquired in quiet and in noise. If CAEPs are to be used as a tool to explore signal encoding in noise, scientists must be cognizant of how differences in acquisition and processing protocols selectively shape CAEP responses. PMID:25453611

  1. Familiar units prevail over statistical cues in word segmentation.

    PubMed

    Poulin-Charronnat, Bénédicte; Perruchet, Pierre; Tillmann, Barbara; Peereman, Ronald

    2017-09-01

    In language acquisition research, the prevailing position is that listeners exploit statistical cues, in particular transitional probabilities between syllables, to discover words of a language. However, other cues are also involved in word discovery. Assessing the weight learners give to these different cues leads to a better understanding of the processes underlying speech segmentation. The present study evaluated whether adult learners preferentially used known units or statistical cues for segmenting continuous speech. Before the exposure phase, participants were familiarized with part-words of a three-word artificial language. This design allowed the dissociation of the influence of statistical cues and familiar units, with statistical cues favoring word segmentation and familiar units favoring (nonoptimal) part-word segmentation. In Experiment 1, performance in a two-alternative forced choice (2AFC) task between words and part-words revealed part-word segmentation (even though part-words were less cohesive in terms of transitional probabilities and less frequent than words). By contrast, an unfamiliarized group exhibited word segmentation, as usually observed in standard conditions. Experiment 2 used a syllable-detection task to remove the likely contamination of performance by memory and strategy effects in the 2AFC task. Overall, the results suggest that familiar units overrode statistical cues, ultimately questioning the need for computation mechanisms of transitional probabilities (TPs) in natural language speech segmentation.

  2. Characteristics of the acquisition of sonorant consonants orthography in Brazilian children from a São Paulo municipality.

    PubMed

    Vaz, Suellen; Pezarini, Isabela de Oliveira; Paschoal, Larissa; Chacon, Lourenço

    2015-01-01

    To describe the spelling performance of children with regard to the record of sonorant consonants in Brazilian Portuguese language, to verify if the errors in their records were influenced by the accent in the word, and to categorize the kinds of errors found. For this current survey, 801 text productions were selected as a result of the development of 14 different thematic proposals, prepared by 76 children from the first grade of primary school, in 2001, coming from two schools of a city from São Paulo, Brazil. Of these productions, all words with sonorant consonants in a syllabic position of simple onset were selected. They were then organized as they appeared as pre-tonic, tonic, and post-tonic syllables, unstressed and tonic monosyllables. The following was observed: the number of hits was extremely higher than that of errors; higher occurrence of errors in non-accented syllables; higher occurrence of phonological substitutions followed by omissions and, at last, orthographic substitutions; and higher number of substitutions that involved graphemes referring to the sonorant class. Considering the distribution of orthographic data between hits and errors, as well as their relationship with phonetic-phonological aspects, may contribute to the comprehension of school difficulties, which are usually found in the first years of literacy instruction.

  3. Implementation of Three Text to Speech Systems for Kurdish Language

    NASA Astrophysics Data System (ADS)

    Bahrampour, Anvar; Barkhoda, Wafa; Azami, Bahram Zahir

    Nowadays, concatenative method is used in most modern TTS systems to produce artificial speech. The most important challenge in this method is choosing appropriate unit for creating database. This unit must warranty smoothness and high quality speech, and also, creating database for it must reasonable and inexpensive. For example, syllable, phoneme, allophone, and, diphone are appropriate units for all-purpose systems. In this paper, we implemented three synthesis systems for Kurdish language based on syllable, allophone, and diphone and compare their quality using subjective testing.

  4. Production of sentence-final intonation contours by hearing-impaired children.

    PubMed

    Allen, G D; Arndorfer, P M

    2000-04-01

    Studies of intonation in the hearing impaired (HI) are often concerned with either objective measures or listener perceptions. Less often has the focus been on how these two aspects of communication interrelate. This study examined the relationship between certain acoustic parameters and listeners' perceptions of intonation contours produced by HI children. Six severe-to-profound HI children and 6 normal-hearing (NH) children, ages 7;9 to 14;7, were individually tape recorded while reading 10 declarative sentences and 10 phonemically matched interrogative sentences within the context of a script. Each sentence ended with a carefully chosen disyllabic (target) word. Twelve adult listeners, inexperienced with the speech of the HI, listened to a randomized audio tape presentation of all of these productions and categorized each one as a statement, question, or other. Fundamental frequency (F0) and duration measurements were obtained for the target (final) word of each sentence, and intensity measures were recorded for each entire sentence. Acoustic analysis showed that all 6 of the NH children and 4 of the 6 HI children produced acoustically different intonation contours for declarative versus interrogative sentences. The HI children's productions were, in general, similar to the NH children, in that they used F0, duration, and intensity cues to mark the distinction. Their contrastive use of these acoustic cues, however, was less pronounced than for the NH children. Analysis of listener responses indicated that, although listeners were able to differentiate between some of the declarative and interrogative sentences produced by these 4 HI children, judgments corresponded with their intended type less often for the HI than for the NH children. (Judgments of NH children's utterances were 100% correct.) Multiple logistic regression of listeners' responses to the HI children's utterances showed that 4 acoustic measures, all derived from the sentence-final word, were significantly predictive: (1) sentence-final F0, (2) slope between the target word's initial and final F0, (3) duration of the target word, and (4) dB difference between the target word's 1st and 2nd syllables. Results were similar for the NH children's data, except that the ratio of the 2 syllables' durations was significant, rather than total word duration. These findings differ in several important ways from previously published data for HI children's intonation contours and suggest that many HI children have the ability to benefit substantially from training in the production of intonation.

  5. Sleep underpins the plasticity of language production.

    PubMed

    Gaskell, M Gareth; Warker, Jill; Lindsay, Shane; Frost, Rebecca; Guest, James; Snowdon, Reza; Stackhouse, Abigail

    2014-07-01

    The constraints that govern acceptable phoneme combinations in speech perception and production have considerable plasticity. We addressed whether sleep influences the acquisition of new constraints and their integration into the speech-production system. Participants repeated sequences of syllables in which two phonemes were artificially restricted to syllable onset or syllable coda, depending on the vowel in that sequence. After 48 sequences, participants either had a 90-min nap or remained awake. Participants then repeated 96 sequences so implicit constraint learning could be examined, and then were tested for constraint generalization in a forced-choice task. The sleep group, but not the wake group, produced speech errors at test that were consistent with restrictions on the placement of phonemes in training. Furthermore, only the sleep group generalized their learning to new materials. Polysomnography data showed that implicit constraint learning was associated with slow-wave sleep. These results show that sleep facilitates the integration of new linguistic knowledge with existing production constraints. These data have relevance for systems-consolidation models of sleep. © The Author(s) 2014.

  6. Visual Influences on Perception of Speech and Nonspeech Vocal-Tract Events

    PubMed Central

    Brancazio, Lawrence; Best, Catherine T.; Fowler, Carol A.

    2009-01-01

    We report four experiments designed to determine whether visual information affects judgments of acoustically-specified nonspeech events as well as speech events (the “McGurk effect”). Previous findings have shown only weak McGurk effects for nonspeech stimuli, whereas strong effects are found for consonants. We used click sounds that serve as consonants in some African languages, but that are perceived as nonspeech by American English listeners. We found a significant McGurk effect for clicks presented in isolation that was much smaller than that found for stop-consonant-vowel syllables. In subsequent experiments, we found strong McGurk effects, comparable to those found for English syllables, for click-vowel syllables, and weak effects, comparable to those found for isolated clicks, for excised release bursts of stop consonants presented in isolation. We interpret these findings as evidence that the potential contributions of speech-specific processes on the McGurk effect are limited, and discuss the results in relation to current explanations for the McGurk effect. PMID:16922061

  7. Interaction in planning vocalizations and grasping.

    PubMed

    Tiainen, Mikko; Tiippana, Kaisa; Vainio, Martti; Komeilipoor, Naeem; Vainio, Lari

    2017-08-01

    Previous studies have shown a congruency effect between manual grasping and syllable articulation. For instance, a power grip is associated with syllables whose articulation involves the tongue body and/or large mouth aperture ([kɑ]) whereas a precision grip is associated with articulations that involve the tongue tip and/or small mouth aperture ([ti]). Previously, this effect has been observed in manual reaction times. The primary aim of the current study was to investigate whether this congruency effect also takes place in vocal responses and to investigate involvement of action selection processes in the effect. The congruency effect was found in vocal and manual responses regardless of whether or not the syllable or grip was known a priori, suggesting that the effect operates with minimal or absent action selection processes. In addition, the effect was observed in vocal responses even when the grip was only prepared but not performed, suggesting that merely planning a grip response primes the corresponding articulatory response. These results support the view that articulation and grasping are processed in a partially overlapping network.

  8. Discrete Circuits Support Generalized versus Context-Specific Vocal Learning in the Songbird.

    PubMed

    Tian, Lucas Y; Brainard, Michael S

    2017-12-06

    Motor skills depend on the reuse of individual gestures in multiple sequential contexts (e.g., a single phoneme in different words). Yet optimal performance requires that a given gesture be modified appropriately depending on the sequence in which it occurs. To investigate the neural architecture underlying such context-dependent modifications, we studied Bengalese finch song, which, like speech, consists of variable sequences of "syllables." We found that when birds are instructed to modify a syllable in one sequential context, learning generalizes across contexts; however, if unique instruction is provided in different contexts, learning is specific for each context. Using localized inactivation of a cortical-basal ganglia circuit specialized for song, we show that this balance between generalization and specificity reflects a hierarchical organization of neural substrates. Primary motor circuitry encodes a core syllable representation that contributes to generalization, while top-down input from cortical-basal ganglia circuitry biases this representation to enable context-specific learning. Copyright © 2017 Elsevier Inc. All rights reserved.

  9. Seeking an Anchorage. Stability and Variability in Tonal Alignment of Rising Prenuclear Pitch Accents in Cypriot Greek.

    PubMed

    Themistocleous, Charalambos

    2016-12-01

    Although tonal alignment constitutes a quintessential property of pitch accents, its exact characteristics remain unclear. This study, by exploring the timing of the Cypriot Greek L*+H prenuclear pitch accent, examines the predictions of three hypotheses about tonal alignment: the invariance hypothesis, the segmental anchoring hypothesis, and the segmental anchorage hypothesis. The study reports on two experiments: the first of which manipulates the syllable patterns of the stressed syllable, and the second of which modifies the distance of the L*+H from the following pitch accent. The findings on the alignment of the low tone (L) are illustrative of the segmental anchoring hypothesis predictions: the L persistently aligns inside the onset consonant, a few milliseconds before the stressed vowel. However, the findings on the alignment of the high tone (H) are both intriguing and unexpected: the alignment of the H depends on the number of unstressed syllables that follow the prenuclear pitch accent. The 'wandering' of the H over multiple syllables is extremely rare among languages, and casts doubt on the invariance hypothesis and the segmental anchoring hypothesis, as well as indicating the need for a modified version of the segmental anchorage hypothesis. To address the alignment of the H, we suggest that it aligns within a segmental anchorage-the area that follows the prenuclear pitch accent-in such a way as to protect the paradigmatic contrast between the L*+H prenuclear pitch accent and the L+H* nuclear pitch accent.

  10. SyllabO+: A new tool to study sublexical phenomena in spoken Quebec French.

    PubMed

    Bédard, Pascale; Audet, Anne-Marie; Drouin, Patrick; Roy, Johanna-Pascale; Rivard, Julie; Tremblay, Pascale

    2017-10-01

    Sublexical phonotactic regularities in language have a major impact on language development, as well as on speech processing and production throughout the entire lifespan. To understand the impact of phonotactic regularities on speech and language functions at the behavioral and neural levels, it is essential to have access to oral language corpora to study these complex phenomena in different languages. Yet, probably because of their complexity, oral language corpora remain less common than written language corpora. This article presents the first corpus and database of spoken Quebec French syllables and phones: SyllabO+. This corpus contains phonetic transcriptions of over 300,000 syllables (over 690,000 phones) extracted from recordings of 184 healthy adult native Quebec French speakers, ranging in age from 20 to 97 years. To ensure the representativeness of the corpus, these recordings were made in both formal and familiar communication contexts. Phonotactic distributional statistics (e.g., syllable and co-occurrence frequencies, percentages, percentile ranks, transition probabilities, and pointwise mutual information) were computed from the corpus. An open-access online application to search the database was developed, and is available at www.speechneurolab.ca/syllabo . In this article, we present a brief overview of the corpus, as well as the syllable and phone databases, and we discuss their practical applications in various fields of research, including cognitive neuroscience, psycholinguistics, neurolinguistics, experimental psychology, phonetics, and phonology. Nonacademic practical applications are also discussed, including uses in speech-language pathology.

  11. The perception of sentence stress in cochlear implant recipients.

    PubMed

    Meister, Hartmut; Landwehr, Markus; Pyschny, Verena; Wagner, Petra; Walger, Martin

    2011-01-01

    Sentence stress is a vital attribute of speech since it indicates the importance of specific words within an utterance. Basic acoustic correlates of stress are syllable duration, intensity, and fundamental frequency (F0). Objectives of the study were to determine cochlear implant (CI) users' perception of the acoustic correlates and to uncover which cues are used for stress identification. Several experiments addressed the discrimination of changes in syllable duration, intensity, and F0 as well as stress identification based on these cues. Moreover, the discrimination of combined cues and identification of stress in conversational speech was examined. Both natural utterances and artificial manipulations of the acoustic cues were used as stimuli. Discrimination of syllable duration did not differ significantly between CI recipients and a control group of normal-hearing listeners. In contrast, CI users performed significantly worse on tasks of discrimination and stress identification based on F0 as well as on intensity. Results from these measurements were significantly correlated with the ability to identify stress in conversational speech. Discrimination performance for covarying F0 and intensity changes was more strongly correlated to identification performance than was found for discrimination of either F0 or intensity alone. Syllable duration was not related to stress identification in natural utterances. The outcome emphasizes the importance of both F0 and intensity for CI users' identification of sentence-based stress. Both cues were used separately for stress perception, but combining the cues provided extra benefit for most of the subjects.

  12. Left and right basal ganglia and frontal activity during language generation: contributions to lexical, semantic, and phonological processes.

    PubMed

    Crosson, Bruce; Benefield, Hope; Cato, M Allison; Sadek, Joseph R; Moore, Anna Bacon; Wierenga, Christina E; Gopinath, Kaundinya; Soltysik, David; Bauer, Russell M; Auerbach, Edward J; Gökçay, Didem; Leonard, Christiana M; Briggs, Richard W

    2003-11-01

    fMRI was used to determine the frontal, basal ganglia, and thalamic structures engaged by three facets of language generation: lexical status of generated items, the use of semantic vs. phonological information during language generation, and rate of generation. During fMRI, 21 neurologically normal subjects performed four tasks: generation of nonsense syllables given beginning and ending consonant blends, generation of words given a rhyming word, generation of words given a semantic category at a fast rate (matched to the rate of nonsense syllable generation), and generation of words given a semantic category at a slow rate (matched to the rate of generating of rhyming words). Components of a left pre-SMA-dorsal caudate nucleus-ventral anterior thalamic loop were active during word generation from rhyming or category cues but not during nonsense syllable generation. Findings indicate that this loop is involved in retrieving words from pre-existing lexical stores. Relatively diffuse activity in the right basal ganglia (caudate nucleus and putamen) also was found during word-generation tasks but not during nonsense syllable generation. Given the relative absence of right frontal activity during the word generation tasks, we suggest that the right basal ganglia activity serves to suppress right frontal activity, preventing right frontal structures from interfering with language production. Current findings establish roles for the left and the right basal ganglia in word generation. Hypotheses are discussed for future research to help refine our understanding of basal ganglia functions in language generation.

  13. Attentional modulation of informational masking on early cortical representations of speech signals.

    PubMed

    Zhang, Changxin; Arnott, Stephen R; Rabaglia, Cristina; Avivi-Reich, Meital; Qi, James; Wu, Xihong; Li, Liang; Schneider, Bruce A

    2016-01-01

    To recognize speech in a noisy auditory scene, listeners need to perceptually segregate the target talker's voice from other competing sounds (stream segregation). A number of studies have suggested that the attentional demands placed on listeners increase as the acoustic properties and informational content of the competing sounds become more similar to that of the target voice. Hence we would expect attentional demands to be considerably greater when speech is masked by speech than when it is masked by steady-state noise. To investigate the role of attentional mechanisms in the unmasking of speech sounds, event-related potentials (ERPs) were recorded to a syllable masked by noise or competing speech under both active (the participant was asked to respond when the syllable was presented) or passive (no response was required) listening conditions. The results showed that the long-latency auditory response to a syllable (/bi/), presented at different signal-to-masker ratios (SMRs), was similar in both passive and active listening conditions, when the masker was a steady-state noise. In contrast, a switch from the passive listening condition to the active one, when the masker was two-talker speech, significantly enhanced the ERPs to the syllable. These results support the hypothesis that the need to engage attentional mechanisms in aid of scene analysis increases as the similarity (both acoustic and informational) between the target speech and the competing background sounds increases. Copyright © 2015 Elsevier B.V. All rights reserved.

  14. Syllabic discrimination in premature human infants prior to complete formation of cortical layers

    PubMed Central

    Mahmoudzadeh, Mahdi; Dehaene-Lambertz, Ghislaine; Fournier, Marc; Kongolo, Guy; Goudjil, Sabrina; Dubois, Jessica; Grebe, Reinhard; Wallois, Fabrice

    2013-01-01

    The ontogeny of linguistic functions in the human brain remains elusive. Although some auditory capacities are described before term, whether and how such immature cortical circuits might process speech are unknown. Here we used functional optical imaging to evaluate the cerebral responses to syllables at the earliest age at which cortical responses to external stimuli can be recorded in humans (28- to 32-wk gestational age). At this age, the cortical organization in layers is not completed. Many neurons are still located in the subplate and in the process of migrating to their final location. Nevertheless, we observed several points of similarity with the adult linguistic network. First, whereas syllables elicited larger right than left responses, the posterior temporal region escaped this general pattern, showing faster and more sustained responses over the left than over the right hemisphere. Second, discrimination responses to a change of phoneme (ba vs. ga) and a change of human voice (male vs. female) were already present and involved inferior frontal areas, even in the youngest infants (29-wk gestational age). Third, whereas both types of changes elicited responses in the right frontal region, the left frontal region only reacted to a change of phoneme. These results demonstrate a sophisticated organization of perisylvian areas at the very onset of cortical circuitry, 3 mo before term. They emphasize the influence of innate factors on regions involved in linguistic processing and social communication in humans. PMID:23440196

  15. Phonological perception by birds: budgerigars can perceive lexical stress.

    PubMed

    Hoeschele, Marisa; Fitch, W Tecumseh

    2016-05-01

    Metrical phonology is the perceptual "strength" in language of some syllables relative to others. The ability to perceive lexical stress is important, as it can help a listener segment speech and distinguish the meaning of words and sentences. Despite this importance, there has been little comparative work on the perception of lexical stress across species. We used a go/no-go operant paradigm to train human participants and budgerigars (Melopsittacus undulatus) to distinguish trochaic (stress-initial) from iambic (stress-final) two-syllable nonsense words. Once participants learned the task, we presented both novel nonsense words, and familiar nonsense words that had certain cues removed (e.g., pitch, duration, loudness, or vowel quality) to determine which cues were most important in stress perception. Members of both species learned the task and were then able to generalize to novel exemplars, showing categorical learning rather than rote memorization. Tests using reduced stimuli showed that humans could identify stress patterns with amplitude and pitch alone, but not with only duration or vowel quality. Budgerigars required more than one cue to be present and had trouble if vowel quality or amplitude were missing as cues. The results suggest that stress patterns in human speech can be decoded by other species. Further comparative stress-perception research with more species could help to determine what species characteristics predict this ability. In addition, tests with a variety of stimuli could help to determine how much this ability depends on general pattern learning processes versus vocalization-specific cues.

  16. A Mis-recognized Medical Vocabulary Correction System for Speech-based Electronic Medical Record

    PubMed Central

    Seo, Hwa Jeong; Kim, Ju Han; Sakabe, Nagamasa

    2002-01-01

    Speech recognition as an input tool for electronic medical record (EMR) enables efficient data entry at the point of care. However, the recognition accuracy for medical vocabulary is much poorer than that for doctor-patient dialogue. We developed a mis-recognized medical vocabulary correction system based on syllable-by-syllable comparison of speech text against medical vocabulary database. Using specialty medical vocabulary, the algorithm detects and corrects mis-recognized medical vocabularies in narrative text. Our preliminary evaluation showed 94% of accuracy in mis-recognized medical vocabulary correction.

  17. [Analysis of the speech discrimination scores of patients with congenital unilateral microtia and external auditory canal atresia in noise].

    PubMed

    Zhang, Y; Li, D D; Chen, X W

    2017-06-20

    Objective: Case-control study analysis of the speech discrimination of unilateral microtia and external auditory canal atresia patients with normal hearing subjects in quiet and noisy environment. To understand the speech recognition results of patients with unilateral external auditory canal atresia and provide scientific basis for clinical early intervention. Method: Twenty patients with unilateral congenital microtia malformation combined external auditory canal atresia, 20 age matched normal subjects as control group. All subjects used Mandarin speech audiometry material, to test the speech discrimination scores (SDS) in quiet and noisy environment in sound field. Result: There's no significant difference of speech discrimination scores under the condition of quiet between two groups. There's a statistically significant difference when the speech signal in the affected side and noise in the nomalside (single syllable, double syllable, statements; S/N=0 and S/N=-10) ( P <0.05). There's no significant difference of speech discrimination scores when the speech signal in the nomalside and noise in the affected side. There's a statistically significant difference in condition of the signal and noise in the same side when used one-syllable word recognition (S/N=0 and S/N=-5) ( P <0.05), while double syllable word and statement has no statistically significant difference ( P >0.05). Conclusion: The speech discrimination scores of unilateral congenital microtia malformation patients with external auditory canal atresia under the condition of noise is lower than the normal subjects. Copyright© by the Editorial Department of Journal of Clinical Otorhinolaryngology Head and Neck Surgery.

  18. A Measure of the Auditory-perceptual Quality of Strain from Electroglottographic Analysis of Continuous Dysphonic Speech: Application to Adductor Spasmodic Dysphonia.

    PubMed

    Somanath, Keerthan; Mau, Ted

    2016-11-01

    (1) To develop an automated algorithm to analyze electroglottographic (EGG) signal in continuous dysphonic speech, and (2) to identify EGG waveform parameters that correlate with the auditory-perceptual quality of strain in the speech of patients with adductor spasmodic dysphonia (ADSD). Software development with application in a prospective controlled study. EGG was recorded from 12 normal speakers and 12 subjects with ADSD reading excerpts from the Rainbow Passage. Data were processed by a new algorithm developed with the specific goal of analyzing continuous dysphonic speech. The contact quotient, pulse width, a new parameter peak skew, and various contact closing slope quotient and contact opening slope quotient measures were extracted. EGG parameters were compared between normal and ADSD speech. Within the ADSD group, intra-subject comparison was also made between perceptually strained syllables and unstrained syllables. The opening slope quotient SO7525 distinguished strained syllables from unstrained syllables in continuous speech within individual subjects with ADSD. The standard deviations, but not the means, of contact quotient, EGGW50, peak skew, and SO7525 were different between normal and ADSD speakers. The strain-stress pattern in continuous speech can be visualized as color gradients based on the variation of EGG parameter values. EGG parameters may provide a within-subject measure of vocal strain and serve as a marker for treatment response. The addition of EGG to multidimensional assessment may lead to improved characterization of the voice disturbance in ADSD. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  19. A measure of the auditory-perceptual quality of strain from electroglottographic analysis of continuous dysphonic speech: Application to adductor spasmodic dysphonia

    PubMed Central

    Somanath, Keerthan; Mau, Ted

    2016-01-01

    Objectives (1) To develop an automated algorithm to analyze electroglottographic (EGG) signal in continuous, dysphonic speech, and (2) to identify EGG waveform parameters that correlate with the auditory-perceptual quality of strain in the speech of patients with adductor spasmodic dysphonia (ADSD). Study Design Software development with application in a prospective controlled study. Methods EGG was recorded from 12 normal speakers and 12 subjects with ADSD reading excerpts from the Rainbow Passage. Data were processed by a new algorithm developed with the specific goal of analyzing continuous dysphonic speech. The contact quotient (CQ), pulse width (EGGW), a new parameter peak skew, and various contact closing slope quotient (SC) and contact opening slope quotient (SO) measures were extracted. EGG parameters were compared between normal and ADSD speech. Within the ADSD group, intra-subject comparison was also made between perceptually strained syllables and unstrained syllables. Results The opening slope quotient SO7525 distinguished strained syllables from unstrained syllables in continuous speech within individual ADSD subjects. The standard deviations, but not the means, of CQ, EGGW50, peak skew, and SO7525 were different between normal and ADSD speakers. The strain-stress pattern in continuous speech can be visualized as color gradients based on the variation of EGG parameter values. Conclusions EGG parameters may provide a within-subject measure of vocal strain and serve as a marker for treatment response. The addition of EGG to multi-dimensional assessment may lead to improved characterization of the voice disturbance in ADSD. PMID:26739857

  20. Lexical stress encoding in single word production estimated by event-related brain potentials.

    PubMed

    Schiller, Niels O

    2006-09-27

    An event-related brain potentials (ERPs) experiment was carried out to investigate the time course of lexical stress encoding in language production. Native speakers of Dutch viewed a series of pictures corresponding to bisyllabic names which were either stressed on the first or on the second syllable and made go/no-go decisions on the lexical stress location of those picture names. Behavioral results replicated a pattern that was observed earlier, i.e. faster button-press latencies to initial as compared to final stress targets. The electrophysiological results indicated that participants could make a lexical stress decision significantly earlier when picture names had initial than when they had final stress. Moreover, the present data suggest the time course of lexical stress encoding during single word form formation in language production. When word length is corrected for, the temporal interval for lexical stress encoding specified by the current ERP results falls into the time window previously identified for phonological encoding in language production.

  1. Acoustic cues to perception of word stress by English, Mandarin, and Russian speakers.

    PubMed

    Chrabaszcz, Anna; Winn, Matthew; Lin, Candise Y; Idsardi, William J

    2014-08-01

    This study investigated how listeners' native language affects their weighting of acoustic cues (such as vowel quality, pitch, duration, and intensity) in the perception of contrastive word stress. Native speakers (N = 45) of typologically diverse languages (English, Russian, and Mandarin) performed a stress identification task on nonce disyllabic words with fully crossed combinations of each of the 4 cues in both syllables. The results revealed that although the vowel quality cue was the strongest cue for all groups of listeners, pitch was the second strongest cue for the English and the Mandarin listeners but was virtually disregarded by the Russian listeners. Duration and intensity cues were used by the Russian listeners to a significantly greater extent compared with the English and Mandarin participants. Compared with when cues were noncontrastive across syllables, cues were stronger when they were in the iambic contour than when they were in the trochaic contour. Although both English and Russian are stress languages and Mandarin is a tonal language, stress perception performance of the Mandarin listeners but not of the Russian listeners is more similar to that of the native English listeners, both in terms of weighting of the acoustic cues and the cues' relative strength in different word positions. The findings suggest that tuning of second-language prosodic perceptions is not entirely predictable by prosodic similarities across languages.

  2. Acoustic Cues to Perception of Word Stress by English, Mandarin, and Russian Speakers

    PubMed Central

    Chrabaszcz, Anna; Winn, Matthew; Lin, Candise Y.; Idsardi, William J.

    2017-01-01

    Purpose This study investigated how listeners’ native language affects their weighting of acoustic cues (such as vowel quality, pitch, duration, and intensity) in the perception of contrastive word stress. Method Native speakers (N = 45) of typologically diverse languages (English, Russian, and Mandarin) performed a stress identification task on nonce disyllabic words with fully crossed combinations of each of the 4 cues in both syllables. Results The results revealed that although the vowel quality cue was the strongest cue for all groups of listeners, pitch was the second strongest cue for the English and the Mandarin listeners but was virtually disregarded by the Russian listeners. Duration and intensity cues were used by the Russian listeners to a significantly greater extent compared with the English and Mandarin participants. Compared with when cues were noncontrastive across syllables, cues were stronger when they were in the iambic contour than when they were in the trochaic contour. Conclusions Although both English and Russian are stress languages and Mandarin is a tonal language, stress perception performance of the Mandarin listeners but not of the Russian listeners is more similar to that of the native English listeners, both in terms of weighting of the acoustic cues and the cues’ relative strength in different word positions. The findings suggest that tuning of second-language prosodic perceptions is not entirely predictable by prosodic similarities across languages. PMID:24686836

  3. Dichotic listening performance predicts language comprehension.

    PubMed

    Asbjørnsen, Arve E; Helland, Turid

    2006-05-01

    Dichotic listening performance is considered a reliable and valid procedure for the assessment of language lateralisation in the brain. However, the documentation of a relationship between language functions and dichotic listening performance is sparse, although it is accepted that dichotic listening measures language perception. In particular, language comprehension should show close correspondence to perception of language stimuli. In the present study, we tested samples of reading-impaired and normally achieving children between 10 and 13 years of age with tests of reading skills, language comprehension, and dichotic listening to consonant-vowel (CV) syllables. A high correlation between the language scores and the dichotic listening performance was expected. However, since the left ear score is believed to be an error when assessing language laterality, covariation was expected for the right ear scores only. In addition, directing attention to one ear input was believed to reduce the influence of random factors, and thus show a more concise estimate of left hemisphere language capacity. Thus, a stronger correlation between language comprehension skills and the dichotic listening performance when attending to the right ear was expected. The analyses yielded a positive correlation between the right ear score in DL and language comprehension, an effect that was stronger when attending to the right ear. The present results confirm the assumption that dichotic listening with CV syllables measures an aspect of language perception and language skills that is related to general language comprehension.

  4. The cerebral control of speech tempo: opposite relationship between speaking rate and BOLD signal changes at striatal and cerebellar structures.

    PubMed

    Riecker, Axel; Kassubek, Jan; Gröschel, Klaus; Grodd, Wolfgang; Ackermann, Hermann

    2006-01-01

    So far, only sparse data on the cerebral organization of speech motor control are available. In order to further delineate the neural basis of articulatory functions, fMRI measurements were performed during self-paced syllable repetitions at six different frequencies (2-6 Hz). Bilateral hemodynamic main effects, calculated across all syllable rates considered, emerged within sensorimotor cortex, putamen, thalamus and cerebellum. At the level of the caudatum and the anterior insula, activation was found restricted to the left side. The computation of rate-to-response functions of the BOLD signal revealed a negative linear relationship between syllable frequency and response magnitude within the striatum whereas cortical areas and cerebellar hemispheres exhibited an opposite activation pattern. Dysarthric patients with basal ganglia disorders show unimpaired or even accelerated speaking rate whereas, in contrast, cerebellar dysfunctions give rise to slowed speech tempo which does not fall below a rate of about 3 Hz. The observed rate-to-response profiles of the BOLD signal thus might help to elucidate the pathophysiological mechanisms of dysarthric deficits in central motor disorders.

  5. Scores on Riley's stuttering severity instrument versions three and four for samples of different length and for different types of speech material.

    PubMed

    Todd, Helena; Mirawdeli, Avin; Costelloe, Sarah; Cavenagh, Penny; Davis, Stephen; Howell, Peter

    2014-12-01

    Riley stated that the minimum speech sample length necessary to compute his stuttering severity estimates was 200 syllables. This was investigated. Procedures supplied for the assessment of readers and non-readers were examined to see whether they give equivalent scores. Recordings of spontaneous speech samples from 23 young children (aged between 2 years 8 months and 6 years 3 months) and 31 older children (aged between 10 years 0 months and 14 years 7 months) were made. Riley's severity estimates were scored on extracts of different lengths. The older children provided spontaneous and read samples, which were scored for severity according to reader and non-reader procedures. Analysis of variance supported the use of 200-syllable-long samples as the minimum necessary for obtaining severity scores. There was no significant difference in SSI-3 scores for the older children when the reader and non-reader procedures were used. Samples that are 200-syllables long are the minimum that is appropriate for obtaining stable Riley's severity scores. The procedural variants provide similar severity scores.

  6. McGurk stimuli for the investigation of multisensory integration in cochlear implant users: The Oldenburg Audio Visual Speech Stimuli (OLAVS).

    PubMed

    Stropahl, Maren; Schellhardt, Sebastian; Debener, Stefan

    2017-06-01

    The concurrent presentation of different auditory and visual syllables may result in the perception of a third syllable, reflecting an illusory fusion of visual and auditory information. This well-known McGurk effect is frequently used for the study of audio-visual integration. Recently, it was shown that the McGurk effect is strongly stimulus-dependent, which complicates comparisons across perceivers and inferences across studies. To overcome this limitation, we developed the freely available Oldenburg audio-visual speech stimuli (OLAVS), consisting of 8 different talkers and 12 different syllable combinations. The quality of the OLAVS set was evaluated with 24 normal-hearing subjects. All 96 stimuli were characterized based on their stimulus disparity, which was obtained from a probabilistic model (cf. Magnotti & Beauchamp, 2015). Moreover, the McGurk effect was studied in eight adult cochlear implant (CI) users. By applying the individual, stimulus-independent parameters of the probabilistic model, the predicted effect of stronger audio-visual integration in CI users could be confirmed, demonstrating the validity of the new stimulus material.

  7. Syllable-constituent perception by hearing-aid users: Common factors in quiet and noise

    PubMed Central

    Miller, James D.; Watson, Charles S.; Leek, Marjorie R.; Dubno, Judy R.; Wark, David J.; Souza, Pamela E.; Gordon-Salant, Sandra; Ahlstrom, Jayne B.

    2017-01-01

    The abilities of 59 adult hearing-aid users to hear phonetic details were assessed by measuring their abilities to identify syllable constituents in quiet and in differing levels of noise (12-talker babble) while wearing their aids. The set of sounds consisted of 109 frequently occurring syllable constituents (45 onsets, 28 nuclei, and 36 codas) spoken in varied phonetic contexts by eight talkers. In nominal quiet, a speech-to-noise ratio (SNR) of 40 dB, scores of individual listeners ranged from about 23% to 85% correct. Averaged over the range of SNRs commonly encountered in noisy situations, scores of individual listeners ranged from about 10% to 71% correct. The scores in quiet and in noise were very strongly correlated, R = 0.96. This high correlation implies that common factors play primary roles in the perception of phonetic details in quiet and in noise. Otherwise said, hearing-aid users' problems perceiving phonetic details in noise appear to be tied to their problems perceiving phonetic details in quiet and vice versa. PMID:28464618

  8. Comparison of audio and audiovisual measures of adult stuttering: Implications for clinical trials.

    PubMed

    O'Brian, Sue; Jones, Mark; Onslow, Mark; Packman, Ann; Menzies, Ross; Lowe, Robyn

    2015-04-15

    This study investigated whether measures of percentage syllables stuttered (%SS) and stuttering severity ratings with a 9-point scale differ when made from audiovisual compared with audio-only recordings. Four experienced speech-language pathologists measured %SS and assigned stuttering severity ratings to 10-minute audiovisual and audio-only recordings of 36 adults. There was a mean 18% increase in %SS scores when samples were presented in audiovisual compared with audio-only mode. This result was consistent across both higher and lower %SS scores and was found to be directly attributable to counts of stuttered syllables rather than the total number of syllables. There was no significant difference between stuttering severity ratings made from the two modes. In clinical trials research, when using %SS as the primary outcome measure, audiovisual samples would be preferred as long as clear, good quality, front-on images can be easily captured. Alternatively, stuttering severity ratings may be a more valid measure to use as they correlate well with %SS and values are not influenced by the presentation mode.

  9. Word-level prominence in Persian: An Experimental Study.

    PubMed

    Sadeghi, Vahid

    2017-12-01

    Previous literature on the phonetics of stress in Persian has reported that fundamental frequency is the only reliable acoustic correlate of stress, and that stressed and unstressed syllables are not differentiated from each other in the absence of accentuation. In this study, the effects of lexical stress on duration, overall intensity and spectral tilt were examined in Persian both in the accented and unaccented conditions. Results showed that syllable duration is consistently affected by stress in Persian in both the accented and unaccented conditions across all vowel types. Unlike duration, the results for overall intensity and spectral tilt were significant only in the accented condition, suggesting that measures of intensity are not a correlate of stress in Persian but they are mainly caused by the presence of a pitch movement. The findings are phonologically interpreted as suggesting that word-level prominence in Persian is typologically similar to 'stress accent' languages, in which multiple phonetic cues are used to signal the prominence contrast in the accented condition, and stressed and unstressed syllables are different from each other even when the word is not pitch-accented.

  10. Musical expertise and foreign speech perception

    PubMed Central

    Martínez-Montes, Eduardo; Hernández-Pérez, Heivet; Chobert, Julie; Morgado-Rodríguez, Lisbet; Suárez-Murias, Carlos; Valdés-Sosa, Pedro A.; Besson, Mireille

    2013-01-01

    The aim of this experiment was to investigate the influence of musical expertise on the automatic perception of foreign syllables and harmonic sounds. Participants were Cuban students with high level of expertise in music or in visual arts and with the same level of general education and socio-economic background. We used a multi-feature Mismatch Negativity (MMN) design with sequences of either syllables in Mandarin Chinese or harmonic sounds, both comprising deviants in pitch contour, duration and Voice Onset Time (VOT) or equivalent that were either far from (Large deviants) or close to (Small deviants) the standard. For both Mandarin syllables and harmonic sounds, results were clear-cut in showing larger MMNs to pitch contour deviants in musicians than in visual artists. Results were less clear for duration and VOT deviants, possibly because of the specific characteristics of the stimuli. Results are interpreted as reflecting similar processing of pitch contour in speech and non-speech sounds. The implications of these results for understanding the influence of intense musical training from childhood to adulthood and of genetic predispositions for music on foreign language perception are discussed. PMID:24294193

  11. Invented Spelling, Word Stress, and Syllable Awareness in Relation to Reading Difficulties in Children.

    PubMed

    Mehta, Sheena; Ding, Yi; Ness, Molly; Chen, Eric C

    2018-06-01

    The study assessed the clinical utility of an invented spelling tool and determined whether invented spelling with linguistic manipulation at segmental and supra-segmental levels can be used to better identify reading difficulties. We conducted linguistic manipulation by using real and nonreal words, incorporating word stress, alternating the order of consonants and vowels, and alternating the number of syllables. We recruited 60 third-grade students, of which half were typical readers and half were poor readers. The invented spelling task consistently differentiated those with reading difficulties from typical readers. It explained unique variance in conventional spelling, but not in word reading. Word stress explained unique variance in both word reading and conventional spelling, highlighting the importance of addressing phonological awareness at the supra-segmental level. Poor readers had poorer performance when spelling both real and nonreal words and demonstrated substantial difficulty in detecting word stress. Poor readers struggled with spelling words with double consonants at the beginning and ending of words, and performed worse on spelling two- and three-syllable words than typical readers. Practical implications for early identification and instruction are discussed.

  12. Verbal cues effectively orient children's auditory attention in a CV-syllable dichotic listening paradigm.

    PubMed

    Phélip, Marion; Donnot, Julien; Vauclair, Jacques

    2015-12-18

    In their groundbreaking work featuring verbal dichotic listening tasks, Mondor and Bryden showed that tone cues do not enhance children's attentional orienting, in contrast to adults. The magnitude of the children's right-ear advantage was not attenuated when their attention was directed to the left ear. Verbal cues did, however, appear to favour the orientation of attention at around 10 years, although stimulus-onset asynchronies (SOAs), which ranged between 450 and 750 ms, were not rigorously controlled. The aim of our study was therefore to investigate the role of both types of cues in a typical CV-syllable dichotic listening task administered to 8- to 10-year-olds, applying a protocol as similar as possible to that used by Mondor and Bryden, but controlling for SOA as well as for cued ear. Results confirmed that verbal cues are more effective than tone cues in orienting children's attention. However, in contrast to adults, no effect of SOA was observed. We discuss the relative difficulty young children have processing CV syllables, as well as the role of top-down processes in attentional orienting abilities.

  13. Musical expertise and foreign speech perception.

    PubMed

    Martínez-Montes, Eduardo; Hernández-Pérez, Heivet; Chobert, Julie; Morgado-Rodríguez, Lisbet; Suárez-Murias, Carlos; Valdés-Sosa, Pedro A; Besson, Mireille

    2013-01-01

    The aim of this experiment was to investigate the influence of musical expertise on the automatic perception of foreign syllables and harmonic sounds. Participants were Cuban students with high level of expertise in music or in visual arts and with the same level of general education and socio-economic background. We used a multi-feature Mismatch Negativity (MMN) design with sequences of either syllables in Mandarin Chinese or harmonic sounds, both comprising deviants in pitch contour, duration and Voice Onset Time (VOT) or equivalent that were either far from (Large deviants) or close to (Small deviants) the standard. For both Mandarin syllables and harmonic sounds, results were clear-cut in showing larger MMNs to pitch contour deviants in musicians than in visual artists. Results were less clear for duration and VOT deviants, possibly because of the specific characteristics of the stimuli. Results are interpreted as reflecting similar processing of pitch contour in speech and non-speech sounds. The implications of these results for understanding the influence of intense musical training from childhood to adulthood and of genetic predispositions for music on foreign language perception are discussed.

  14. Performance constraints and the production of birdsong

    NASA Astrophysics Data System (ADS)

    Suthers, Roderick A.; Vallet, Eric; Zollinger, Sue Anne

    2004-05-01

    The role of physical and physiological constraints in determining the performance limits on the tempo and frequency bandwidth of birdsong was investigated. One series of experiments examined the mechanism by which a vocal mimic, the northern mockingbird (Mimus polygottos), copied the songs of other species with which it was tutored as a juvenile. Other experiments analyzed the motor basis of special canary (Serinus canaria) syllables eliciting sexual responses from females. In each case, the mechanism of vocalization was determined by measuring the respiratory dynamics and sound produced on each side of the songbirds duplex vocal organ, the syrinx. When mockingbirds copied the songs of other species the accuracy of their copy depended on the accuracy with which they reproduced the motor pattern used by the tutor species. Motor difficulty of various acoustic features was assessed by the accuracy of its copy. The high repetition rate, broadband canary syllables preferred by females required especially demanding bilateral motor skills. The results indicate that constraints on the rate of respiratory ventilation and bilateral syringeal coordination can set an upper limit on syllable repetition rate and frequency bandwidth. [Work supported by NIH and NSF.

  15. Effects of Online Augmented Kinematic and Perceptual Feedback on Treatment of Speech Movements in Apraxia of Speech

    PubMed Central

    McNeil, M.R.; Katz, W.F.; Fossett, T.R.D.; Garst, D.M.; Szuminsky, N.J.; Carter, G.; Lim, K.Y.

    2010-01-01

    Apraxia of speech (AOS) is a motor speech disorder characterized by disturbed spatial and temporal parameters of movement. Research on motor learning suggests that augmented feedback may provide a beneficial effect for training movement. This study examined the effects of the presence and frequency of online augmented visual kinematic feedback (AVKF) and clinician-provided perceptual feedback on speech accuracy in 2 adults with acquired AOS. Within a single-subject multiple-baseline design, AVKF was provided using electromagnetic midsagittal articulography (EMA) in 2 feedback conditions (50 or 100%). Articulator placement was specified for speech motor targets (SMTs). Treated and baselined SMTs were in the initial or final position of single-syllable words, in varying consonant-vowel or vowel-consonant contexts. SMTs were selected based on each participant's pre-assessed erred productions. Productions were digitally recorded and online perceptual judgments of accuracy (including segment and intersegment distortions) were made. Inter- and intra-judge reliability for perceptual accuracy was high. Results measured by visual inspection and effect size revealed positive acquisition and generalization effects for both participants. Generalization occurred across vowel contexts and to untreated probes. Results of the frequency manipulation were confounded by presentation order. Maintenance of learned and generalized effects were demonstrated for 1 participant. These data provide support for the role of augmented feedback in treating speech movements that result in perceptually accurate speech production. Future investigations will explore the independent contributions of each feedback type (i.e. kinematic and perceptual) in producing efficient and effective training of SMTs in persons with AOS. PMID:20424468

  16. The genus Phlesirtes Bolivar, 1922 (Orthoptera: Tettigoniidae: Conocephalinae, Conocephalini; Karniellina), a review of the genus with data on its bioacoustics and the description of new species.

    PubMed

    Hemp, Claudia; Heller, Klaus-Gerhard

    2017-03-22

    Species of the genus Phlesirtes are reviewed in this paper and 16 new species described. A key is provided to the species of Phlesirtes. Among the subtribe Karniellina of Conocephalini members of the genus Phlesirtes occupy habitats of montane to afroalpine grasslands in East Africa. All recorded species produced long-lasting, trill-like calling songs, consisting of sequences of continuously repeated syllables or syllable pairs with the peak of the carrier frequency in the low ultrasound.

  17. Feet and syllables in elephants and missiles: a reappraisal.

    PubMed

    Zonneveld, Wim; van der Pas, Brigit; de Bree, Elise

    2007-01-01

    Using data from a case study presented in Chiat (1989), Marshall and Chiat (2003) compare two different approaches to account for the realization of intervocalic consonants in child phonology: "coda capture theory" and the "foot domain account". They argue in favour of the latter account. In this note, we present a reappraisal of this argument using the same data. We conclude that acceptance of the foot domain account, in the specific way developed by the authors, is unmotivated for both theoretical and empirical reasons. We maintain that syllable-based coda capture is (still) the better approach to account for the relevant facts.

  18. Oral-diadochokinetic rates for Hebrew-speaking healthy ageing population: non-word versus real-word repetition.

    PubMed

    Ben-David, Boaz M; Icht, Michal

    2017-05-01

    Oral-diadochokinesis (oral-DDK) tasks are extensively used in the evaluation of motor speech abilities. Currently, validated normative data for older adults (aged 65 years and older) are missing in Hebrew. The effect of task stimuli (non-word versus real-word repetition) is also non-clear in the population of older adult Hebrew speakers. (1) To establish a norm for oral-DDK rate for older adult (aged 65 years and older) Hebrew speakers, and to investigate the possible effect of age and gender on performance rate; and (2) to examine the effects of stimuli (non-word versus real word) on oral-DDK rates. In experiment 1, 88 healthy older Hebrew speakers (60-95 years, 48 females and 40 males) were audio-recorded while performing an oral-DDK task (repetition of /pataka/), and repetition rates (syllables/s) were coded. In experiment 2, the effect of real-word repetition was evaluated. Sixty-eight older Hebrew speakers (aged 66-95 years, 43 females and 25 males) were asked to repeat 'pataka' (non-word) and 'bodeket' (Hebrew real word). Experiment 1: Oral-DDK performance for older adult Hebrew speakers was 5.07 syllables/s (SD = 1.16 syllables/s), across age groups and gender. Comparison of this data with Hebrew norms for younger adults (and equivalent data in English) shows the following gradient of oral-DDK rates: ages 15-45 > 65-74 > 75-86 years. Gender was not a significant factor in our data. Experiment 2: Repetition of real words was faster than that of non-words, by 13.5%. The paper provides normative values for oral-DDK rates for older Hebrew speakers. The data show the large impact of ageing on oro-motor functions. The analysis further indicates that speech and language pathologists should consider separate norms for clients of 65-74 years and those of 75-86 years. Hebrew rates were found to be different from English norms for the oldest group, shedding light on the impact of language on these norms. Finally, the data support using a dual-protocol (real- and non-word repetition) with older adults to improve differential diagnosis of normal and pathological ageing in this task. © 2016 Royal College of Speech and Language Therapists.

  19. Intra-oral pressure-based voicing control of electrolaryngeal speech with intra-oral vibrator.

    PubMed

    Takahashi, Hirokazu; Nakao, Masayuki; Kikuchi, Yataro; Kaga, Kimitaka

    2008-07-01

    In normal speech, coordinated activities of intrinsic laryngeal muscles suspend a glottal sound at utterance of voiceless consonants, automatically realizing a voicing control. In electrolaryngeal speech, however, the lack of voicing control is one of the causes of unclear voice, voiceless consonants tending to be misheard as the corresponding voiced consonants. In the present work, we developed an intra-oral vibrator with an intra-oral pressure sensor that detected utterance of voiceless phonemes during the intra-oral electrolaryngeal speech, and demonstrated that an intra-oral pressure-based voicing control could improve the intelligibility of the speech. The test voices were obtained from one electrolaryngeal speaker and one normal speaker. We first investigated on the speech analysis software how a voice onset time (VOT) and first formant (F1) transition of the test consonant-vowel syllables contributed to voiceless/voiced contrasts, and developed an adequate voicing control strategy. We then compared the intelligibility of consonant-vowel syllables among the intra-oral electrolaryngeal speech with and without online voicing control. The increase of intra-oral pressure, typically with a peak ranging from 10 to 50 gf/cm2, could reliably identify utterance of voiceless consonants. The speech analysis and intelligibility test then demonstrated that a short VOT caused the misidentification of the voiced consonants due to a clear F1 transition. Finally, taking these results together, the online voicing control, which suspended the prosthetic tone while the intra-oral pressure exceeded 2.5 gf/cm2 and during the 35 milliseconds that followed, proved efficient to improve the voiceless/voiced contrast.

  20. The observation of manual grasp actions affects the control of speech: a combined behavioral and Transcranial Magnetic Stimulation study.

    PubMed

    Gentilucci, Maurizio; Campione, Giovanna Cristina; Dalla Volta, Riccardo; Bernardis, Paolo

    2009-12-01

    Does the mirror system affect the control of speech? This issue was addressed in behavioral and Transcranial Magnetic Stimulation (TMS) experiments. In behavioral experiment 1, participants pronounced the syllable /da/ while observing (1) a hand grasping large and small objects with power and precision grasps, respectively, (2) a foot interacting with large and small objects and (3) differently sized objects presented alone. Voice formant 1 was higher when observing power as compared to precision grasp, whereas it remained unaffected by observation of the different types of foot interaction and objects alone. In TMS experiment 2, we stimulated hand motor cortex, while participants observed the two types of grasp. Motor Evoked Potentials (MEPs) of hand muscles active during the two types of grasp were greater when observing power than precision grasp. In experiments 3-5, TMS was applied to tongue motor cortex of participants silently pronouncing the syllable /da/ and simultaneously observing power and precision grasps, pantomimes of the two types of grasps, and differently sized objects presented alone. Tongue MEPs were greater when observing power than precision grasp either executed or pantomimed. Finally, in TMS experiment 6, the observation of foot interaction with large and small objects did not modulate tongue MEPs. We hypothesized that grasp observation activated motor commands to the mouth as well as to the hand that were congruent with the hand kinematics implemented in the observed type of grasp. The commands to the mouth selectively affected postures of phonation organs and consequently basic features of phonological units.

  1. True katydids (Pseudophyllinae) from Guadeloupe: acoustic signals and functional considerations of song production.

    PubMed

    Stumpner, Andreas; Dann, Angela; Schink, Matthias; Gubert, Silvia; Hugel, Sylvain

    2013-01-01

    Guadeloupe, the largest of the Leeward Islands, harbors three species of Pseudophyllinae (Orthoptera: Tettigoniidae) belonging to distinct tribes. This study examined the basic aspects of sound production and acousto-vibratory behavior of these species. As the songs of many Pseudophyllinae are complex and peak at high frequencies, they require high quality recordings. Wild specimens were therefore recorded ex situ. Collected specimens were used in structure-function experiments. Karukerana aguilari Bonfils (Pterophyllini) is a large species with a mirror in each tegmen and conspicuous folds over the mirror. It sings 4-6 syllables, each comprising 10-20 pulses, with several peaks in the frequency spectrum between 4 and 20 kHz. The song is among the loudest in Orthoptera (> 125 dB SPL in 10 cm distance). The folds are protective and have no function in song production. Both mirrors may work independently in sound radiation. Nesonotus reticulatus (Fabricius) (Cocconotini) produces verses from two syllables at irregular intervals. The song peaks around 20 kHz. While singing, the males often produce a tremulation signal with the abdomen at about 8-10 Hz. To our knowledge, it is the first record of simultaneous calling song and tremulation in Orthoptera. Other males reply to the tremulation with their own tremulation. Xerophyllopteryx fumosa (Brunner von Wattenwyl) (Pleminiini) is a large, bark-like species, producing a syllable of around 20 pulses. The syllables are produced with irregular rhythms (often two with shorter intervals). The song peaks around 2-3 kHz and 10 kHz. The hind wings are relatively thick and are held between the half opened tegmina during singing. Removal of the hind wings reduces song intensity by about 5 dB, especially of the low frequency component, suggesting that the hind wings have a role in amplifying the song.

  2. Limited role of phonology in reading Chinese two-character compounds: evidence from an ERP study.

    PubMed

    Wong, A W-K; Wu, Y; Chen, H-C

    2014-01-03

    This study investigates the role of phonology in reading logographic Chinese. Specifically, whether phonological information is obligatorily activated in reading Chinese two-character compounds was examined using the masked-priming paradigm with event-related potential (ERP) recordings. Twenty-two native Cantonese Chinese speakers participated in a lexical decision experiment. The targets were visually presented Chinese two-character strings and the participants were asked to judge whether the target in each trial was a legitimate compound word in Chinese. Each target was preceded by a briefly presented word prime. The prime and target shared an identical constituent character in the Character-related condition, a syllable in the Syllable-related condition, were semantically related in the Semantic-related condition, and were unrelated (both phonologically and semantically) in the control condition. The prime–target relationship was manipulated to probe the effects of word-form (i.e., character- or syllable-relatedness) and word-semantic relatedness on phonological (as indexed by an N250 ERP component) and semantic (as indexed by an N400 ERP component) processing. Significant and comparable facilitation effects in reaction time, relative to the control, were observed in the Character-related and the Semantic-related conditions. Furthermore, a significant reduction in ERP amplitudes (N250), relative to the control, was obtained in the Character-related condition in the time window of 150-250 ms post target. In addition, attenuation in ERP amplitudes was found in the Semantic-related condition in the window of 250-500 ms (N400). However, no significant results (neither behavioral nor ERP) were found in the Syllable-related condition. These results suggest that phonological activation is not mandatory and the role of phonology is minimal at best in reading Chinese two-character compounds.

  3. Control of Vocal and Respiratory Patterns in Birdsong: Dissection of Forebrain and Brainstem Mechanisms Using Temperature

    PubMed Central

    Fee, Michale S.

    2011-01-01

    Learned motor behaviors require descending forebrain control to be coordinated with midbrain and brainstem motor systems. In songbirds, such as the zebra finch, regular breathing is controlled by brainstem centers, but when the adult songbird begins to sing, its breathing becomes tightly coordinated with forebrain-controlled vocalizations. The periods of silence (gaps) between song syllables are typically filled with brief breaths, allowing the bird to sing uninterrupted for many seconds. While substantial progress has been made in identifying the brain areas and pathways involved in vocal and respiratory control, it is not understood how respiratory and vocal control is coordinated by forebrain motor circuits. Here we combine a recently developed technique for localized brain cooling, together with recordings of thoracic air sac pressure, to examine the role of cortical premotor nucleus HVC (proper name) in respiratory-vocal coordination. We found that HVC cooling, in addition to slowing all song timescales as previously reported, also increased the duration of expiratory pulses (EPs) and inspiratory pulses (IPs). Expiratory pulses, like song syllables, were stretched uniformly by HVC cooling, but most inspiratory pulses exhibited non-uniform stretch of pressure waveform such that the majority of stretch occurred late in the IP. Indeed, some IPs appeared to change duration by the earlier or later truncation of an underlying inspiratory event. These findings are consistent with the idea that during singing the temporal structure of EPs is under the direct control of forebrain circuits, whereas that of IPs can be strongly influenced by circuits downstream of HVC, likely in the brainstem. An analysis of the temporal jitter of respiratory and vocal structure suggests that IPs may be initiated by HVC at the end of each syllable and terminated by HVC immediately before the onset of the next syllable. PMID:21980466

  4. Case file audit of Lidcombe program outcomes in a student-led stuttering clinic.

    PubMed

    McCulloch, Julia; Swift, Michelle C; Wagnitz, Bianca

    2017-04-01

    The current study aimed to benchmark clinical outcomes for preschool-aged clients (2;0-5;11 years old) that attended a student-led clinic and undertook the Lidcombe Program. A case file audit was undertaken for all preschool clients who attended the clinic between February 2008 and February 2013 and commenced the Lidcombe Program. Clients were grouped according to Stage 1 completion. A mixed ANOVA was used to test for differences between the groups in initial and final percentage syllables stuttered (%SS). Associations between case variable factors and treatment duration were investigated using Pearson correlations. Clients who completed Stage 1 had final %SS and severity rating (SR) scores comparable to the literature; however, the median Stage 1 duration was greater. Over half of the clients (57%) withdrew prior to completing Stage 1. These clients had a significantly higher %SS at final treatment session than their completing peers. Initial %SS and SR scores were the only case variables associated with treatment duration. Students can achieve the same short-term treatment outcomes for children who stutter using the Lidcombe Program as the current published literature; however, treatment duration is greater and may impact completion. Implications of this for clinical education are discussed.

  5. Estimating consumer familiarity with health terminology: a context-based approach.

    PubMed

    Zeng-Treitler, Qing; Goryachev, Sergey; Tse, Tony; Keselman, Alla; Boxwala, Aziz

    2008-01-01

    Effective health communication is often hindered by a "vocabulary gap" between language familiar to consumers and jargon used in medical practice and research. To present health information to consumers in a comprehensible fashion, we need to develop a mechanism to quantify health terms as being more likely or less likely to be understood by typical members of the lay public. Prior research has used approaches including syllable count, easy word list, and frequency count, all of which have significant limitations. In this article, we present a new method that predicts consumer familiarity using contextual information. The method was applied to a large query log data set and validated using results from two previously conducted consumer surveys. We measured the correlation between the survey result and the context-based prediction, syllable count, frequency count, and log normalized frequency count. The correlation coefficient between the context-based prediction and the survey result was 0.773 (p < 0.001), which was higher than the correlation coefficients between the survey result and the syllable count, frequency count, and log normalized frequency count (p < or = 0.012). The context-based approach provides a good alternative to the existing term familiarity assessment methods.

  6. Perceptual rate normalization in naturally produced bilabial stops

    NASA Astrophysics Data System (ADS)

    Nagao, Kyoko; de Jong, Kenneth

    2003-10-01

    The perception of voicing categories is affected by the speaking rate, so that listeners' category boundaries on a VOT continuum shift to a lower value when the syllable duration decreases (Miller and Volaitis, 1989; Volaitis and Miller, 1992). Previous rate normalization effects have been found using computer-generated stimuli. This study examines the effect of speech rate on voicing categorization in naturally produced speech. Four native speakers of American English repeated syllables (/bi/ and /pi/) at increasing rates in time with a metronome. Three-syllable stimuli were spliced from the repetitive speech. These stimuli contained natural decreases in VOT with faster speech rates. Besides, this rate effect on VOT was larger for /p/ than /b/, so that VOT values for /b/ and /p/ overlapped at the fastest rates. Eighteen native listeners of American English were presented with 168 stimuli and asked to identify the consonant. Perceptual category boundaries occur at VOT values 15 ms shorter than the values reported for synthesized stimuli. This difference may be due to the extraordinarily wide range of VOT values in previous studies. The values found in the current study closely match the actual division point for /b/ and /p/. The underlying mechanism of perceptual normalization will be discussed.

  7. Exploring vocal recovery after cranial nerve injury in Bengalese finches.

    PubMed

    Urbano, Catherine M; Peterson, Jennifer R; Cooper, Brenton G

    2013-02-08

    Songbirds and humans use auditory feedback to acquire and maintain their vocalizations. The Bengalese finch (Lonchura striata domestica) is a songbird species that rapidly modifies its vocal output to adhere to an internal song memory. In this species, the left side of the bipartite vocal organ is specialized for producing louder, higher frequencies (≥2.2kHz) and denervation of the left vocal muscles eliminates these notes. Thus, the return of higher frequency notes after cranial nerve injury can be used as a measure of vocal recovery. Either the left or right side of the syrinx was denervated by resection of the tracheosyringeal portion of the hypoglossal nerve. Histologic analyses of syringeal muscle tissue showed significant muscle atrophy in the denervated side. After left nerve resection, songs were mainly composed of lower frequency syllables, but three out of five birds recovered higher frequency syllables. Right nerve resection minimally affected phonology, but it did change song syntax; syllable sequence became abnormally stereotyped after right nerve resection. Therefore, damage to the neuromuscular control of sound production resulted in reduced motor variability, and Bengalese finches are a potential model for functional vocal recovery following cranial nerve injury. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.

  8. Psychophysics of the McGurk and Other Audiovisual Speech Integration Effects

    PubMed Central

    Jiang, Jintao; Bernstein, Lynne E.

    2011-01-01

    When the auditory and visual components of spoken audiovisual nonsense syllables are mismatched, perceivers produce four different types of perceptual responses, auditory correct, visual correct, fusion (the so-called McGurk effect), and combination (i.e., two consonants are reported). Here, quantitative measures were developed to account for the distribution of types of perceptual responses to 384 different stimuli from four talkers. The measures included mutual information, the presented acoustic signal versus the acoustic signal recorded with the presented video, and the correlation between the presented acoustic and video stimuli. In Experiment 1, open-set perceptual responses were obtained for acoustic /bA/ or /lA/ dubbed to video /bA, dA, gA, vA, zA, lA, wA, ΔA/. The talker, the video syllable, and the acoustic syllable significantly influenced the type of response. In Experiment 2, the best predictors of response category proportions were a subset of the physical stimulus measures, with the variance accounted for in the perceptual response category proportions between 17% and 52%. That audiovisual stimulus relationships can account for response distributions supports the possibility that internal representations are based on modality-specific stimulus relationships. PMID:21574741

  9. Electromagnetic articulography assessment of articulatory function in adults with dysarthria following traumatic brain injury.

    PubMed

    Kuruvilla, Mili; Murdoch, Bruce; Goozèe, Justine

    2007-06-01

    To explore articulatory kinematic differences between normal and dysarthric speakers post-traumatic brain injury (TBI) during syllable and sentence productions. A comparison between the control, mild (MTBI) and severe TBI groups for all measured kinematic parameters was carried out using the Kruskal Wallis test. Ten participants with a severe TBI and six post-MTBI formed the experimental group. The control group consisted of 14 age and sex matched non-neurologically impaired speakers. Articulatory kinematic profiles for the three groups were obtained using the Electromagnetic Articulograph (EMA) while repeating sentence and syllable embedded /t/ and /k/ productions at a habitual rate and loudness level. Significant differences between the severe TBI and control group were identified only for the release phase of the /t/ sentence productions wherein an increase in mean maximum acceleration was observed for the severe TBI group. While a simple syllable repetition task at a moderate rate was unable to differentiate the three groups, a complex sentence production task precipitated an increase in mean maximum acceleration which may be indicative of increased articulatory effort and impaired speech motor control even at a convenient rate for the severe group.

  10. A Daily Oscillation in the Fundamental Frequency and Amplitude of Harmonic Syllables of Zebra Finch Song

    PubMed Central

    Wood, William E.; Osseward, Peter J.; Roseberry, Thomas K.; Perkel, David J.

    2013-01-01

    Complex motor skills are more difficult to perform at certain points in the day (for example, shortly after waking), but the daily trajectory of motor-skill error is more difficult to predict. By undertaking a quantitative analysis of the fundamental frequency (FF) and amplitude of hundreds of zebra finch syllables per animal per day, we find that zebra finch song follows a previously undescribed daily oscillation. The FF and amplitude of harmonic syllables rises across the morning, reaching a peak near mid-day, and then falls again in the late afternoon until sleep. This oscillation, although somewhat variable, is consistent across days and across animals and does not require serotonin, as animals with serotonergic lesions maintained daily oscillations. We hypothesize that this oscillation is driven by underlying physiological factors which could be shared with other taxa. Song production in zebra finches is a model system for studying complex learned behavior because of the ease of gathering comprehensive behavioral data and the tractability of the underlying neural circuitry. The daily oscillation that we describe promises to reveal new insights into how time of day affects the ability to accomplish a variety of complex learned motor skills. PMID:24312654

  11. Precursors to Natural Grammar Learning: Preliminary Evidence from 4-Month-Old Infants

    PubMed Central

    Friederici, Angela D.; Mueller, Jutta L.; Oberecker, Regine

    2011-01-01

    When learning a new language, grammar—although difficult—is very important, as grammatical rules determine the relations between the words in a sentence. There is evidence that very young infants can detect rules determining the relation between neighbouring syllables in short syllable sequences. A critical feature of all natural languages, however, is that many grammatical rules concern the dependency relation between non-neighbouring words or elements in a sentence i.e. between an auxiliary and verb inflection as in is singing. Thus, the issue of when and how children begin to recognize such non-adjacent dependencies is fundamental to our understanding of language acquisition. Here, we use brain potential measures to demonstrate that the ability to recognize dependencies between non-adjacent elements in a novel natural language is observable by the age of 4 months. Brain responses indicate that 4-month-old German infants discriminate between grammatical and ungrammatical dependencies in auditorily presented Italian sentences after only brief exposure to correct sentences of the same type. As the grammatical dependencies are realized by phonologically distinct syllables the present data most likely reflect phonologically based implicit learning mechanisms which can serve as a precursor to later grammar learning. PMID:21445341

  12. Distress vocalization sequences broadcasted by bats carry redundant information.

    PubMed

    Hechavarría, Julio C; Beetz, M Jerome; Macias, Silvio; Kössl, Manfred

    2016-07-01

    Distress vocalizations (also known as alarm or screams) are an important component of the vocal repertoire of a number of animal species, including bats, humans, monkeys and birds, among others. Although the behavioral relevance of distress vocalizations is undeniable, at present, little is known about the rules that govern vocalization production when in alarmful situations. In this article, we show that when distressed, bats of the species Carollia perspicillata produce repetitive vocalization sequences in which consecutive syllables are likely to be similar to one another regarding their physical attributes. The uttered distress syllables are broadband (12-73 kHz) with most of their energy focussing at 23 kHz. Distress syllables are short (~4 ms), their average sound pressure level is close to 70 dB SPL, and they are produced at high repetition rates (every 14 ms). We discuss that, because of their physical attributes, bat distress vocalizations could serve a dual purpose: (1) advertising threatful situations to conspecifics, and (2) informing the threatener that the bats are ready to defend themselves. We also discuss possible advantages of advertising danger/discomfort using repetitive utterances, a calling strategy that appears to be ubiquitous across the animal kingdom.

  13. Age effects on acquisition of word stress in Spanish-English bilinguals

    NASA Astrophysics Data System (ADS)

    Guion, Susan G.; Clark, J. J.; Harada, Tetsuo

    2003-10-01

    Based on studies of syntactic and semantic learning, it has been proposed that certain aspects of second language learning may be more adversely affected by delays in language learning than others. Here, this proposal is extended to the phonological domain in which the acquisition of English word stress patterns by early (AOA <6 years) and late (AOA >14 years) Spanish-English bilinguals is investigated. The knowledge of English word stress was investigated by three behavioral tasks. In a production task, participants produced two syllable nonwords in both noun and verb sentence frames. In a perception task, participants indicated a preference for first or last syllable stress on the nonwords. Real words that were phonologically similar to the test items were also collected from each participant. Regression analyses and ANOVAs were conducted to determine the effect of syllable structure, lexical class, and stress pattern of phonologically similar words on the data from the production and perception tasks. Early bilinguals patterned similarly to the native English participants. Late bilinguals showed little evidence of learning prosodically based stress patterns but did show evidence of application of distributional patterns based on lexical class and analogy in stress assignment. [Research supported by NIH.

  14. GreekLex 2: A comprehensive lexical database with part-of-speech, syllabic, phonological, and stress information

    PubMed Central

    van Heuven, Walter J. B.; Pitchford, Nicola J.; Ledgeway, Timothy

    2017-01-01

    Databases containing lexical properties on any given orthography are crucial for psycholinguistic research. In the last ten years, a number of lexical databases have been developed for Greek. However, these lack important part-of-speech information. Furthermore, the need for alternative procedures for calculating syllabic measurements and stress information, as well as combination of several metrics to investigate linguistic properties of the Greek language are highlighted. To address these issues, we present a new extensive lexical database of Modern Greek (GreekLex 2) with part-of-speech information for each word and accurate syllabification and orthographic information predictive of stress, as well as several measurements of word similarity and phonetic information. The addition of detailed statistical information about Greek part-of-speech, syllabification, and stress neighbourhood allowed novel analyses of stress distribution within different grammatical categories and syllabic lengths to be carried out. Results showed that the statistical preponderance of stress position on the pre-final syllable that is reported for Greek language is dependent upon grammatical category. Additionally, analyses showed that a proportion higher than 90% of the tokens in the database would be stressed correctly solely by relying on stress neighbourhood information. The database and the scripts for orthographic and phonological syllabification as well as phonetic transcription are available at http://www.psychology.nottingham.ac.uk/greeklex/. PMID:28231303

  15. The perceptual chunking of speech: a demonstration using ERPs.

    PubMed

    Gilbert, Annie C; Boucher, Victor J; Jemel, Boutheina

    2015-04-07

    In tasks involving the learning of verbal or non-verbal sequences, groupings are spontaneously produced. These groupings are generally marked by a lengthening of final elements and have been attributed to a domain-general perceptual chunking linked to working memory. Yet, no study has shown how this domain-general chunking applies to speech processing, partly because of the traditional view that chunking involves a conceptual recoding of meaningful verbal items like words (Miller, 1956). The present study provides a demonstration of the perceptual chunking of speech by way of two experiments using evoked Positive Shifts (PSs), which capture on-line neural responses to marks of various groups. We observed listeners׳ response to utterances (Experiment 1) and meaningless series of syllables (Experiment 2) containing changing intonation and temporal marks, while also examining how these marks affect the recognition of heard items. The results show that, across conditions - and irrespective of the presence of meaningful items - PSs are specifically evoked by groups marked by lengthening. Moreover, this on-line detection of marks corresponds to characteristic grouping effects on listeners' immediate recognition of heard items, which suggests chunking effects linked to working memory. These findings bear out a perceptual chunking of speech input in terms of groups marked by lengthening, which constitute the defining marks of a domain-general chunking. Copyright © 2015 Elsevier B.V. All rights reserved.

  16. GreekLex 2: A comprehensive lexical database with part-of-speech, syllabic, phonological, and stress information.

    PubMed

    Kyparissiadis, Antonios; van Heuven, Walter J B; Pitchford, Nicola J; Ledgeway, Timothy

    2017-01-01

    Databases containing lexical properties on any given orthography are crucial for psycholinguistic research. In the last ten years, a number of lexical databases have been developed for Greek. However, these lack important part-of-speech information. Furthermore, the need for alternative procedures for calculating syllabic measurements and stress information, as well as combination of several metrics to investigate linguistic properties of the Greek language are highlighted. To address these issues, we present a new extensive lexical database of Modern Greek (GreekLex 2) with part-of-speech information for each word and accurate syllabification and orthographic information predictive of stress, as well as several measurements of word similarity and phonetic information. The addition of detailed statistical information about Greek part-of-speech, syllabification, and stress neighbourhood allowed novel analyses of stress distribution within different grammatical categories and syllabic lengths to be carried out. Results showed that the statistical preponderance of stress position on the pre-final syllable that is reported for Greek language is dependent upon grammatical category. Additionally, analyses showed that a proportion higher than 90% of the tokens in the database would be stressed correctly solely by relying on stress neighbourhood information. The database and the scripts for orthographic and phonological syllabification as well as phonetic transcription are available at http://www.psychology.nottingham.ac.uk/greeklex/.

  17. [P300 event-related potentials in stutterers pre and post treatment: a pilot study].

    PubMed

    Andrade, Claudia Regina Furquim de; Sassi, Fernanda Chiarion; Matas, Carla Gentile; Neves, Ivone Ferreira; Martins, Vanessa Oliveira

    2007-01-01

    P300 event-related potential has been used as an instrument to establish the diagnosis of several disorders as well as to assess therapeutic outcomes. to investigate the relationship between stuttering amelioration and cerebral activity. P300 event-related potentials were obtained in three adult males, all stutterers, aged 20 to 31 years, pre and post-treatment, verifying changes in wave amplitude and latency between waves. results indicate a significant positive correlation between the reduction in the percentage of stuttered syllables and the improvement in wave amplitude for the right ear. stutterers can exhibit different patterns of interhemispheric activity with a tonal P300 task after undergoing a fluency-enhancing program.

  18. A higher sensory brain region is involved in reversing reinforcement-induced vocal changes in a songbird.

    PubMed

    Canopoli, Alessandro; Herbst, Joshua A; Hahnloser, Richard H R

    2014-05-14

    Many animals exhibit flexible behaviors that they can adjust to increase reward or avoid harm (learning by positive or aversive reinforcement). But what neural mechanisms allow them to restore their original behavior (motor program) after reinforcement is withdrawn? One possibility is that motor restoration relies on brain areas that have a role in memorization but no role in either motor production or in sensory processing relevant for expressing the behavior and its refinement. We investigated the role of a higher auditory brain area in the songbird for modifying and restoring the stereotyped adult song. We exposed zebra finches to aversively reinforcing white noise stimuli contingent on the pitch of one of their stereotyped song syllables. In response, birds significantly changed the pitch of that syllable to avoid the aversive reinforcer. After we withdrew reinforcement, birds recovered their original song within a few days. However, we found that large bilateral lesions in the caudal medial nidopallium (NCM, a high auditory area) impaired recovery of the original pitch even several weeks after withdrawal of the reinforcing stimuli. Because NCM lesions spared both successful noise-avoidance behavior and birds' auditory discrimination ability, our results show that NCM is not needed for directed motor changes or for auditory discriminative processing, but is implied in memorizing or recalling the memory of the recent song target. Copyright © 2014 the authors 0270-6474/14/347018-09$15.00/0.

  19. Computer game as a tool for training the identification of phonemic length.

    PubMed

    Pennala, Riitta; Richardson, Ulla; Ylinen, Sari; Lyytinen, Heikki; Martin, Maisa

    2014-12-01

    Computer-assisted training of Finnish phonemic length was conducted with 7-year-old Russian-speaking second-language learners of Finnish. Phonemic length plays a different role in these two languages. The training included game activities with two- and three-syllable word and pseudo-word minimal pairs with prototypical vowel durations. The lowest accuracy scores were recorded for two-syllable words. Accuracy scores were higher for the minimal pairs with larger rather than smaller differences in duration. Accuracy scores were lower for long duration than for short duration. The ability to identify quantity degree was generalized to stimuli used in the identification test in two of the children. Ideas for improving the game are introduced.

  20. Synergetic Organization in Speech Rhythm

    NASA Astrophysics Data System (ADS)

    Cummins, Fred

    The Speech Cycling Task is a novel experimental paradigm developed together with Robert Port and Keiichi Tajima at Indiana University. In a task of this sort, subjects repeat a phrase containing multiple prominent, or stressed, syllables in time with an auditory metronome, which can be simple or complex. A phase-based collective variable is defined in the acoustic speech signal. This paper reports on two experiments using speech cycling which together reveal many of the hallmarks of hierarchically coupled oscillatory processes. The first experiment requires subjects to place the final stressed syllable of a small phrase at specified phases within the overall Phrase Repetition Cycle (PRC). It is clearly demonstrated that only three patterns, characterized by phases around 1/3, 1/2 or 2/3 are reliably produced, and these points are attractors for other target phases. The system is thus multistable, and the attractors correspond to stable couplings between the metrical foot and the PRC. A second experiment examines the behavior of these attractors at increased rates. Faster rates lead to mode jumps between attractors. Previous experiments have also illustrated hysteresis as the system moves from one mode to the next. The dynamical organization is particularly interesting from a modeling point of view, as there is no single part of the speech production system which cycles at the level of either the metrical foot or the phrase repetition cycle. That is, there is no continuous kinematic observable in the system. Nonetheless, there is strong evidence that the oscopic behavior of the entire production system is correctly described as hierarchically coupled oscillators. There are many parallels between this organization and the forms of inter-limb coupling observed in locomotion and rhythmic manual tasks.

  1. From Mimicry to Language: A Neuroanatomically Based Evolutionary Model of the Emergence of Vocal Language

    PubMed Central

    Poliva, Oren

    2016-01-01

    The auditory cortex communicates with the frontal lobe via the middle temporal gyrus (auditory ventral stream; AVS) or the inferior parietal lobule (auditory dorsal stream; ADS). Whereas the AVS is ascribed only with sound recognition, the ADS is ascribed with sound localization, voice detection, prosodic perception/production, lip-speech integration, phoneme discrimination, articulation, repetition, phonological long-term memory and working memory. Previously, I interpreted the juxtaposition of sound localization, voice detection, audio-visual integration and prosodic analysis, as evidence that the behavioral precursor to human speech is the exchange of contact calls in non-human primates. Herein, I interpret the remaining ADS functions as evidence of additional stages in language evolution. According to this model, the role of the ADS in vocal control enabled early Homo (Hominans) to name objects using monosyllabic calls, and allowed children to learn their parents' calls by imitating their lip movements. Initially, the calls were forgotten quickly but gradually were remembered for longer periods. Once the representations of the calls became permanent, mimicry was limited to infancy, and older individuals encoded in the ADS a lexicon for the names of objects (phonological lexicon). Consequently, sound recognition in the AVS was sufficient for activating the phonological representations in the ADS and mimicry became independent of lip-reading. Later, by developing inhibitory connections between acoustic-syllabic representations in the AVS and phonological representations of subsequent syllables in the ADS, Hominans became capable of concatenating the monosyllabic calls for repeating polysyllabic words (i.e., developed working memory). Finally, due to strengthening of connections between phonological representations in the ADS, Hominans became capable of encoding several syllables as a single representation (chunking). Consequently, Hominans began vocalizing and mimicking/rehearsing lists of words (sentences). PMID:27445676

  2. Dynamic Spectral Structure Specifies Vowels for Adults and Children

    PubMed Central

    Nittrouer, Susan; Lowenstein, Joanna H.

    2014-01-01

    The dynamic specification account of vowel recognition suggests that formant movement between vowel targets and consonant margins is used by listeners to recognize vowels. This study tested that account by measuring contributions to vowel recognition of dynamic (i.e., time-varying) spectral structure and coarticulatory effects on stationary structure. Adults and children (four-and seven-year-olds) were tested with three kinds of consonant-vowel-consonant syllables: (1) unprocessed; (2) sine waves that preserved both stationary coarticulated and dynamic spectral structure; and (3) vocoded signals that primarily preserved that stationary, but not dynamic structure. Sections of two lengths were removed from syllable middles: (1) half the vocalic portion; and (2) all but the first and last three pitch periods. Adults performed accurately with unprocessed and sine-wave signals, as long as half the syllable remained; their recognition was poorer for vocoded signals, but above chance. Seven-year-olds performed more poorly than adults with both sorts of processed signals, but disproportionately worse with vocoded than sine-wave signals. Most four-year-olds were unable to recognize vowels at all with vocoded signals. Conclusions were that both dynamic and stationary coarticulated structures support vowel recognition for adults, but children attend to dynamic spectral structure more strongly because early phonological organization favors whole words. PMID:25536845

  3. Chinese children's early knowledge about writing.

    PubMed

    Zhang, Lan; Yin, Li; Treiman, Rebecca

    2017-09-01

    Much research on literacy development has focused on learners of alphabetic writing systems. Researchers have hypothesized that children learn about the formal characteristics of writing before they learn about the relations between units of writing and units of speech. We tested this hypothesis by examining young Chinese children's understanding of writing. Mandarin-speaking 2- to 5-year-olds completed a graphic task, which tapped their knowledge about the formal characteristics of writing, and a phonological task, which tapped their knowledge about the correspondence between Chinese characters and syllables. The 3- to 5-year-olds performed better on the graphic task than the phonological task, indicating that learning how writing appears visually begins earlier than learning that writing corresponds to linguistic units, even in a writing system in which written units correspond to syllables. Statement of contribution What is already known on this subject? Learning about writing's visual form, how it looks, is an important part of emergent literacy. Knowledge of how writing symbolizes linguistic units may emerge later. What does this study add? We test the hypothesis that Chinese children learn about writing's visual form earlier than its symbolic nature. Chinese 3- to 5- year-olds know more about visual features than character-syllable links. Results show learning of the visual appearance of a notation system is developmentally precocious. © 2016 The British Psychological Society.

  4. The effects of study task on prestimulus subsequent memory effects in the hippocampus.

    PubMed

    de Chastelaine, Marianne; Rugg, Michael D

    2015-11-01

    Functional magnetic resonance imaging (fMRI) was employed to examine the effects of a study task manipulation on pre-stimulus activity in the hippocampus predictive of later successful recollection. Eighteen young participants were scanned while making either animacy or syllable judgments on visually presented study words. Cues presented before each word denoted which judgment should be made. Following the study phase, a surprise recognition memory test was administered in which each test item had to be endorsed as "Remembered," "Known," or "New." As expected, "deep" animacy judgments led to better memory for study items than did "shallow" syllable judgments. In both study tasks, pre-stimulus subsequent recollection effects were evident in the interval between the cue and the study item in bilateral anterior hippocampus. However, the direction of the effects differed according to the study task: whereas pre-stimulus hippocampal activity on animacy trials was greater for later recollected items than items judged old on the basis of familiarity (replicating prior findings), these effects reversed for syllable trials. We propose that the direction of pre-stimulus hippocampal subsequent memory effects depends on whether an optimal pre-stimulus task set facilitates study processing that is conducive or unconducive to the formation of contextually rich episodic memories. © 2015 Wiley Periodicals, Inc.

  5. Acoustic and laryngographic measures of the laryngeal reflexes of linguistic prominence and vocal effort in German1

    PubMed Central

    Mooshammer, Christine

    2010-01-01

    This study uses acoustic and physiological measures to compare laryngeal reflexes of global changes in vocal effort to the effects of modulating such aspects of linguistic prominence as sentence accent, induced by focus variation, and word stress. Seven speakers were recorded by using a laryngograph. The laryngographic pulses were preprocessed to normalize time and amplitude. The laryngographic pulse shape was quantified using open and skewness quotients and also by applying a functional version of the principal component analysis. Acoustic measures included the acoustic open quotient and spectral balance in the vowel ∕e∕ during the test syllable. The open quotient and the laryngographic pulse shape indicated a significantly shorter open phase for loud speech than for soft speech. Similar results were found for lexical stress, suggesting that lexical stress and loud speech are produced with a similar voice source mechanism. Stressed syllables were distinguished from unstressed syllables by their open phase and pulse shape, even in the absence of sentence accent. Evidence for laryngeal involvement in signaling focus, independent of fundamental frequency changes, was not as consistent across speakers. Acoustic results on various spectral balance measures were generally much less consistent compared to results from laryngographic data. PMID:20136226

  6. Size variability of handwriting in healthy Korean older adults.

    PubMed

    Yoon, Ji Hye; Kim, Hyanghee; Kim, Jungwan; Park, Eunjeong; Kim, Soo Ryon

    2014-04-01

    The aim of the present study was to delineate how age-related deterioration affects the handwriting of healthy elderly (HE) subjects. A total of 235 HE (54 males, 181 females) aged 57-91 years participated as subjects in the study. In order to compare the area of handwriting, we divided the participants into two groups: (i) aged 57-74 years; and (ii) aged 75-91 years. The writing stimulus was a four-syllabic word with one-to-one grapheme-to-phoneme correspondence. The size of each syllable in the target word was measured using a software program. Alignment of the word to baseline was assessed using a multiple-choice checklist. As compared with handwriting by the younger group, the older group showed greater variability in the size of the written syllables within the word (P = 0.023). The handwriting was characterized by unequal size among syllables and non-perpendicular alignment, which could be explained by several factors. First, the variability might have resulted from irregular fine movement motor control in older adults. Second, the deterioration of visual feedback and visuomotor integration in normal aging might have affected handwriting performance. In conclusion, variability of handwriting can be sensitive in predicting the aging process. © 2013 Japan Geriatrics Society.

  7. Effects of syllable-initial voicing and speaking rate on the temporal characteristics of monosyllabic words.

    PubMed

    Allen, J S; Miller, J L

    1999-10-01

    Two speech production experiments tested the validity of the traditional method of creating voice-onset-time (VOT) continua for perceptual studies in which the systematic increase in VOT across the continuum is accompanied by a concomitant decrease in the duration of the following vowel. In experiment 1, segmental durations were measured for matched monosyllabic words beginning with either a voiced stop (e.g., big, duck, gap) or a voiceless stop (e.g., pig, tuck, cap). Results from four talkers showed that the change from voiced to voiceless stop produced not only an increase in VOT, but also a decrease in vowel duration. However, the decrease in vowel duration was consistently less than the increase in VOT. In experiment 2, results from four new talkers replicated these findings at two rates of speech, as well as highlighted the contrasting temporal effects on vowel duration of an increase in VOT due to a change in syllable-initial voicing versus a change in speaking rate. It was concluded that the traditional method of creating VOT continua for perceptual experiments, although not perfect, approximates natural speech by capturing the basic trade-off between VOT and vowel duration in syllable-initial voiced versus voiceless stop consonants.

  8. The influence of orthographic experience on the development of phonological preparation in spoken word production.

    PubMed

    Li, Chuchu; Wang, Min

    2017-08-01

    Three sets of experiments using the picture naming tasks with the form preparation paradigm investigated the influence of orthographic experience on the development of phonological preparation unit in spoken word production in native Mandarin-speaking children. Participants included kindergarten children who have not received formal literacy instruction, Grade 1 children who are comparatively more exposed to the alphabetic pinyin system and have very limited Chinese character knowledge, Grades 2 and 4 children who have better character knowledge and more exposure to characters, and skilled adult readers who have the most advanced character knowledge and most exposure to characters. Only Grade 1 children showed the form preparation effect in the same initial consonant condition (i.e., when a list of target words shared the initial consonant). Both Grade 4 children and adults showed the preparation effect when the initial syllable (but not tone) among target words was shared. Kindergartners and Grade 2 children only showed the preparation effect when the initial syllable including tonal information was shared. These developmental changes in phonological preparation could be interpreted as a joint function of the modification of phonological representation and attentional shift. Extensive pinyin experience encourages speakers to attend to and select onset phoneme in phonological preparation, whereas extensive character experience encourages speakers to prepare spoken words in syllables.

  9. Paired variability indices in assessing speech rhythm in Spanish/English bilingual language acquisition

    NASA Astrophysics Data System (ADS)

    Work, Richard; Andruski, Jean; Casielles, Eugenia; Kim, Sahyang; Nathan, Geoff

    2005-04-01

    Traditionally, English is classified as a stress-timed language while Spanish is classified as syllable-timed. Examining the contrasting development of rhythmic patterns in bilingual first language acquisition should provide information on how this differentiation takes place. As part of a longitudinal study, speech samples were taken of a Spanish/English bilingual child of Argentinean parents living in the Midwestern United States between the ages of 1;8 and 3;2. Spanish is spoken at home and English input comes primarily from an English day care the child attends 5 days a week. The parents act as interlocutors for Spanish recordings with a native speaker interacting with the child for the English recordings. Following the work of Grabe, Post and Watson (1999) and Grabe and Low (2002) a normalized Pairwise Variability Index (PVI) is used which compares, in utterances of minimally four syllables, the durations of vocalic intervals in successive syllables. Comparisons are then made between the rhythmic patterns of the child's productions within each language over time and between languages at comparable MLUs. Comparisons are also made with the rhythmic patterns of the adult productions of each language. Results will be analyzed for signs of native speaker-like rhythmic production in the child.

  10. On the function of stress rhythms in speech: evidence of a link with grouping effects on serial memory.

    PubMed

    Boucher, Victor J

    2006-01-01

    Language learning requires a capacity to recall novel series of speech sounds. Research shows that prosodic marks create grouping effects enhancing serial recall. However, any restriction on memory affecting the reproduction of prosody would limit the set of patterns that could be learned and subsequently used in speech. By implication, grouping effects of prosody would also be limited to reproducible patterns. This view of the role of prosody and the contribution of memory processes in the organization of prosodic patterns is examined by evaluating the correspondence between a reported tendency to restrict stress intervals in speech and size limits on stress-grouping effects. French speech is used where stress defines the endpoints of groups. In Experiment 1, 40 speakers recalled novel series of syllables containing stress-groups of varying size. Recall was not enhanced by groupings exceeding four syllables, which corresponded to a restriction on the reproducibility of stress-groups. In Experiment 2, the subjects produced given sentences containing phrases of differing length. The results show a strong tendency to insert stress within phrases that exceed four syllables. Since prosody can arise in the recall of syntactically unstructured lists, the results offer initial support for viewing memory processes as a factor of stress-rhythm organization.

  11. Speech-feature discrimination in children with Asperger syndrome as determined with the multi-feature mismatch negativity paradigm.

    PubMed

    Kujala, T; Kuuluvainen, S; Saalasti, S; Jansson-Verkasalo, E; von Wendt, L; Lepistö, T

    2010-09-01

    Asperger syndrome, belonging to the autistic spectrum of disorders, involves deficits in social interaction and prosodic use of language but normal development of formal language abilities. Auditory processing involves both hyper- and hypoactive reactivity to acoustic changes. Responses composed of mismatch negativity (MMN) and obligatory components were recorded for five types of deviations in syllables (vowel, vowel duration, consonant, syllable frequency, syllable intensity) with the multi-feature paradigm from 8-12-year old children with Asperger syndrome. Children with Asperger syndrome had larger MMNs for intensity and smaller MMNs for frequency changes than typically developing children, whereas no MMN group differences were found for the other deviant stimuli. Furthermore, children with Asperger syndrome performed more poorly than controls in Comprehension of Instructions subtest of a language test battery. Cortical speech-sound discrimination is aberrant in children with Asperger syndrome. This is evident both as hypersensitive and depressed neural reactions to speech-sound changes, and is associated with features (frequency, intensity) which are relevant for prosodic processing. The multi-feature MMN paradigm, which includes variation and thereby resembles natural speech hearing circumstances, suggests abnormal pattern of speech discrimination in Asperger syndrome, including both hypo- and hypersensitive responses for speech features. 2010 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.

  12. Electrophysiological and hemodynamic mismatch responses in rats listening to human speech syllables.

    PubMed

    Mahmoudzadeh, Mahdi; Dehaene-Lambertz, Ghislaine; Wallois, Fabrice

    2017-01-01

    Speech is a complex auditory stimulus which is processed according to several time-scales. Whereas consonant discrimination is required to resolve rapid acoustic events, voice perception relies on slower cues. Humans, right from preterm ages, are particularly efficient to encode temporal cues. To compare the capacities of preterms to those observed in other mammals, we tested anesthetized adult rats by using exactly the same paradigm as that used in preterm neonates. We simultaneously recorded neural (using ECoG) and hemodynamic responses (using fNIRS) to series of human speech syllables and investigated the brain response to a change of consonant (ba vs. ga) and to a change of voice (male vs. female). Both methods revealed concordant results, although ECoG measures were more sensitive than fNIRS. Responses to syllables were bilateral, but with marked right-hemispheric lateralization. Responses to voice changes were observed with both methods, while only ECoG was sensitive to consonant changes. These results suggest that rats more effectively processed the speech envelope than fine temporal cues in contrast with human preterm neonates, in whom the opposite effects were observed. Cross-species comparisons constitute a very valuable tool to define the singularities of the human brain and species-specific bias that may help human infants to learn their native language.

  13. Discrepancies between bilinguals' performance on the Spanish and English versions of the WAIS Digit Span task: Cross-cultural implications.

    PubMed

    López, Enrique; Steiner, Alexander J; Hardy, David J; IsHak, Waguih W; Anderson, W Brantley

    2016-01-01

    This study explored within-subjects differences in the performance of 40 bilingual participants on the English and Spanish versions of the Wechsler Adult Intelligence Scale (WAIS) Digit Span task. To test the linguistic hypothesis that individuals would perform worse in Spanish because of its syllabic demand, we compared the number of syllables correctly recalled by each participant for every correct trial. Our analysis of the correct number of syllables remembered per trial showed that participants performed significantly better (i.e., recalling more syllables) in Spanish than in English on the total score. Findings suggest the Spanish version of the Digit Span (total score) was significantly more difficult than the English version utilizing traditional scoring methods. Moreover, the Forward Trial, rather than the Backward Trial, was more likely to show group differences between both language versions. Additionally, the Spanish trials of the Digit Span were correlated with language comprehension and verbal episodic memory measures, whereas the English trials of the Digit Span were correlated with confrontational naming and verbal fluency tasks. The results suggest that more research is necessary to further investigate other cognitive factors, rather than just syllabic demand, that might contribute to performance and outcome differences on the WAIS Digit Span in Spanish-English bilinguals.

  14. Acquisition of English word stress patterns in early and late bilinguals

    NASA Astrophysics Data System (ADS)

    Guion, Susan G.

    2004-05-01

    Given early acquisition of prosodic knowledge as demonstrated by infants' sensitivity to native language accentual patterns, the question of whether learners can acquire new prosodic patterns across the life span arises. Acquisition of English stress by early and late Spanish-English and Korean-English bilinguals was investigated. In a production task, two-syllable nonwords were produced in noun and verb sentence frames. In a perception task, preference for first or last syllable stress on the nonwords was indicated. Also, real words that were phonologically similar to the nonwords were collected. Logistic regression analyses and ANOVAs were conducted to determine the effect of three factors (syllable structure, lexical class, and stress patterns of phonologically similar words) on the production and perception responses. In all three groups, stress patterns of phonologically similar real words predicted stress on nonwords. For the two other factors, early bilinguals patterned similarly to the native-English participants. Late Spanish-English bilinguals demonstrated less learning of stress patterns based on syllabic structure, and late Korean-English bilinguals demonstrated less learning of stress patterns based on lexical class than native-English speakers. Thus, compared to native speakers, late bilinguals' ability to abstract stress patterns is reduced and affected by the first language. [Work supported by NIH.

  15. Retention of Implicit Sequence Learning in Persons who Stutter and Persons with Parkinson's Disease

    PubMed Central

    Smits-Bandstra, Sarah; Gracco, Vincent

    2014-01-01

    This study investigated the retention of implicit sequence learning in 14 persons with Parkinson's disease (PPD), 14 persons who stutter (PWS) and 14 control participants. Participants completed a nonsense syllable serial reaction time task in a 120-minute session. Participants named aloud four syllables in response to four visual stimuli. The syllables formed a repeating 8-item sequence not made known to participants. After one week, participants completed a 60-minute retention session that included an explicit learning questionnaire and a sequence generation task. PPD showed retention of general learning equivalent to controls but PWS's reaction times were significantly slower on early trials of the retention test relative to other groups. Controls showed implicit learning during the initial session that was retained on the retention test. In contrast, PPD and PWS did not demonstrate significant implicit learning until the retention test suggesting intact, but delayed, learning and retention of implicit sequencing skills. All groups demonstrated similar limited explicit sequence knowledge. Performance differences between PWS and PPD relative to controls during the initial session and on early retention trials indicated possible dysfunction of the cortico-striato-thalamo-cortical loop. The etiological implications for stuttering, and clinical implications for both populations, of this dysfunction are discussed. PMID:23844763

  16. Compensation for pitch-shifted auditory feedback during the production of Mandarin tone sequences

    NASA Astrophysics Data System (ADS)

    Xu, Yi; Larson, Charles R.; Bauer, Jay J.; Hain, Timothy C.

    2004-08-01

    Recent research has found that while speaking, subjects react to perturbations in pitch of voice auditory feedback by changing their voice fundamental frequency (F0) to compensate for the perceived pitch-shift. The long response latencies (150-200 ms) suggest they may be too slow to assist in on-line control of the local pitch contour patterns associated with lexical tones on a syllable-to-syllable basis. In the present study, we introduced pitch-shifted auditory feedback to native speakers of Mandarin Chinese while they produced disyllabic sequences /ma ma/ with different tonal combinations at a natural speaking rate. Voice F0 response latencies (100-150 ms) to the pitch perturbations were shorter than syllable durations reported elsewhere. Response magnitudes increased from 50 cents during static tone to 85 cents during dynamic tone productions. Response latencies and peak times decreased in phrases involving a dynamic change in F0. The larger response magnitudes and shorter latency and peak times in tasks requiring accurate, dynamic control of F0, indicate this automatic system for regulation of voice F0 may be task-dependent. These findings suggest that auditory feedback may be used to help regulate voice F0 during production of bi-tonal Mandarin phrases.

  17. True Katydids (Pseudophyllinae) from Guadeloupe: Acoustic Signals and Functional Considerations of Song Production

    PubMed Central

    Stumpner, Andreas; Dann, Angela; Schink, Matthias; Gubert, Silvia; Hugel, Sylvain

    2013-01-01

    Guadeloupe, the largest of the Leeward Islands, harbors three species of Pseudophyllinae (Orthoptera: Tettigoniidae) belonging to distinct tribes. This study examined the basic aspects of sound production and acousto-vibratory behavior of these species. As the songs of many Pseudophyllinae are complex and peak at high frequencies, they require high quality recordings. Wild specimens were therefore recorded ex situ. Collected specimens were used in structure-function experiments. Karukerana aguilari Bonfils (Pterophyllini) is a large species with a mirror in each tegmen and conspicuous folds over the mirror. It sings 4–6 syllables, each comprising 10–20 pulses, with several peaks in the frequency spectrum between 4 and 20 kHz. The song is among the loudest in Orthoptera (> 125 dB SPL in 10 cm distance). The folds are protective and have no function in song production. Both mirrors may work independently in sound radiation. Nesonotus reticulatus (Fabricius) (Cocconotini) produces verses from two syllables at irregular intervals. The song peaks around 20 kHz. While singing, the males often produce a tremulation signal with the abdomen at about 8–10 Hz. To our knowledge, it is the first record of simultaneous calling song and tremulation in Orthoptera. Other males reply to the tremulation with their own tremulation. Xerophyllopteryx fumosa (Brunner von Wattenwyl) (Pleminiini) is a large, bark-like species, producing a syllable of around 20 pulses. The syllables are produced with irregular rhythms (often two with shorter intervals). The song peaks around 2–3 kHz and 10 kHz. The hind wings are relatively thick and are held between the half opened tegmina during singing. Removal of the hind wings reduces song intensity by about 5 dB, especially of the low frequency component, suggesting that the hind wings have a role in amplifying the song. PMID:24785151

  18. Speech Perception in Older Hearing Impaired Listeners: Benefits of Perceptual Training

    PubMed Central

    Woods, David L.; Doss, Zoe; Herron, Timothy J.; Arbogast, Tanya; Younus, Masood; Ettlinger, Marc; Yund, E. William

    2015-01-01

    Hearing aids (HAs) only partially restore the ability of older hearing impaired (OHI) listeners to understand speech in noise, due in large part to persistent deficits in consonant identification. Here, we investigated whether adaptive perceptual training would improve consonant-identification in noise in sixteen aided OHI listeners who underwent 40 hours of computer-based training in their homes. Listeners identified 20 onset and 20 coda consonants in 9,600 consonant-vowel-consonant (CVC) syllables containing different vowels (/ɑ/, /i/, or /u/) and spoken by four different talkers. Consonants were presented at three consonant-specific signal-to-noise ratios (SNRs) spanning a 12 dB range. Noise levels were adjusted over training sessions based on d’ measures. Listeners were tested before and after training to measure (1) changes in consonant-identification thresholds using syllables spoken by familiar and unfamiliar talkers, and (2) sentence reception thresholds (SeRTs) using two different sentence tests. Consonant-identification thresholds improved gradually during training. Laboratory tests of d’ thresholds showed an average improvement of 9.1 dB, with 94% of listeners showing statistically significant training benefit. Training normalized consonant confusions and improved the thresholds of some consonants into the normal range. Benefits were equivalent for onset and coda consonants, syllables containing different vowels, and syllables presented at different SNRs. Greater training benefits were found for hard-to-identify consonants and for consonants spoken by familiar than unfamiliar talkers. SeRTs, tested with simple sentences, showed less elevation than consonant-identification thresholds prior to training and failed to show significant training benefit, although SeRT improvements did correlate with improvements in consonant thresholds. We argue that the lack of SeRT improvement reflects the dominant role of top-down semantic processing in processing simple sentences and that greater transfer of benefit would be evident in the comprehension of more unpredictable speech material. PMID:25730330

  19. An EMA analysis of the effect of increasing word length on consonant production in apraxia of speech: a case study.

    PubMed

    Bartle, Carly J; Goozée, Justine V; Murdoch, Bruce E

    2007-03-01

    The effect of increasing word length on the articulatory dynamics (i.e. duration, distance, maximum acceleration, maximum deceleration, and maximum velocity) of consonant production in acquired apraxia of speech was investigated using electromagnetic articulography (EMA). Tongue-tip and tongue-back movement of one apraxic patient was recorded using the AG-200 EMA system during word-initial consonant productions in one, two, and three syllable words. Significantly deviant articulatory parameters were recorded for each of the target consonants during one, two, and three syllables words. Word length effects were most evident during the release phase of target consonant productions. The results are discussed with respect to theories of speech motor control as they relate to AOS.

  20. Does Kaniso activate CASINO?: input coding schemes and phonology in visual-word recognition.

    PubMed

    Acha, Joana; Perea, Manuel

    2010-01-01

    Most recent input coding schemes in visual-word recognition assume that letter position coding is orthographic rather than phonological in nature (e.g., SOLAR, open-bigram, SERIOL, and overlap). This assumption has been drawn - in part - by the fact that the transposed-letter effect (e.g., caniso activates CASINO) seems to be (mostly) insensitive to phonological manipulations (e.g., Perea & Carreiras, 2006, 2008; Perea & Pérez, 2009). However, one could argue that the lack of a phonological effect in prior research was due to the fact that the manipulation always occurred in internal letter positions - note that phonological effects tend to be stronger for the initial syllable (Carreiras, Ferrand, Grainger, & Perea, 2005). To reexamine this issue, we conducted a masked priming lexical decision experiment in which we compared the priming effect for transposed-letter pairs (e.g., caniso-CASINO vs. caviro-CASINO) and for pseudohomophone transposed-letter pairs (kaniso-CASINO vs. kaviro-CASINO). Results showed a transposed-letter priming effect for the correctly spelled pairs, but not for the pseudohomophone pairs. This is consistent with the view that letter position coding is (primarily) orthographic in nature.

  1. An evaluation of dynamic lip-tooth characteristics during speech and smile in adolescents.

    PubMed

    Ackerman, Marc B; Brensinger, Colleen; Landis, J Richard

    2004-02-01

    This retrospective study was conducted to measure lip-tooth characteristics of adolescents. Pretreatment video clips of 1242 consecutive patients were screened for Class-I skeletal and dental patterns. After all inclusion criteria were applied, the final sample consisted of 50 patients (27 boys, 23 girls) with a mean age of 12.5 years. The raw digital video stream of each patient was edited to select a single image frame representing the patient saying the syllable "chee" and a second single image representing the patient's posed social smile and saved as part of a 12-frame image sequence. Each animation image was analyzed using a SmileMesh computer application to measure the smile index (the ratio of the intercommissure width divided by the interlabial gap), intercommissure width (mm), interlabial gap (mm), percent incisor below the intercommissure line, and maximum incisor exposure (mm). The data were analyzed using SAS (version 8.1). All recorded differences in linear measures had to be > or = 2 mm. The results suggest that anterior tooth display at speech and smile should be recorded independently but evaluated as part of a dynamic range. Asking patients to say "cheese" and then smile is no longer a valid method to elicit the parameters of anterior tooth display. When planning the vertical positions of incisors during orthodontic treatment, the orthodontist should view the dynamics of anterior tooth display as a continuum delineated by the time points of rest, speech, posed social smile, and a Duchenne smile.

  2. Male mice ultrasonic vocalizations enhance female sexual approach and hypothalamic kisspeptin neuron activity.

    PubMed

    Asaba, Akari; Osakada, Takuya; Touhara, Kazushige; Kato, Masahiro; Mogi, Kazutaka; Kikusui, Takefumi

    2017-08-01

    Vocal communication in animals is important for ensuring reproductive success. Male mice emit song-like "ultrasonic vocalizations (USVs)" when they encounter female mice, and females show approach to the USVs. However, it is unclear whether USVs of male mice trigger female behavioral and endocrine responses in reproduction. In this study, we first investigated the relationship between the number of deliveries in breeding pairs for 4months and USVs syllables emitted from those paired males during 3min of sexual encounter with unfamiliar female mice. There was a positive correlation between these two indices, which suggests that breeding pairs in which males could emit USVs more frequently had more offspring. Further, we examined the effect of USVs of male mice on female sexual behavior. Female mice showed more approach behavior towards vocalizing males than devocalized males. Finally, to determine whether USVs of male mice could activate the neural system governing reproductive function in female mice, the activation of kisspeptin neurons, key neurons to drive gonadotropin-releasing hormone neurons in the hypothalamus, was examined using dual-label immunocytochemistry with cAMP response element-binding protein phosphorylation (pCREB). In the arcuate nucleus (Arc), the number of kisspeptin neurons expressing pCREB significantly increased after exposure to USVs of male as compared with noise exposure group. In conclusion, our results suggest that USVs of male mice promote fertility in female mice by activating both their approaching behavior and central kisspeptin neurons. Copyright © 2017 Elsevier Inc. All rights reserved.

  3. Frustration in the pattern formation of polysyllabic words

    NASA Astrophysics Data System (ADS)

    Hayata, Kazuya

    2016-12-01

    A novel frustrated system is given for the analysis of (m + 1)-syllabled vocal sounds for languages with the m-vowel system, where the varieties of vowels are assumed to be m (m > 2). The necessary and sufficient condition for observing the sound frustration is that the configuration of m vowels in an m-syllabled word has a preference for the ‘repulsive’ type, in which there is no duplication of an identical vowel. For languages that meet this requirement, no (m + 1)-syllabled word can in principle select the present type because at most m different vowels are available and consequently the duplicated use of an identical vowel is inevitable. For languages showing a preference for the ‘attractive’ type, where an identical vowel aggregates in a word, there arises no such conflict. In this paper, we first elucidate for Arabic with m = 3 how to deal with the conflicting situation, where a statistical approach based on the chi-square testing is employed. In addition to the conventional three-vowel system, analyses are made also for Russian, where a polysyllabic word contains both a stressed and an indeterminate vowel. Through the statistical analyses the selection scheme for quadrisyllabic configurations is found to be strongly dependent on the parts of speech as well as the gender of nouns. In order to emphasize the relevance to the sound model of binary oppositions, analyzed results of Greek verbs are also given.

  4. Longitudinal Evaluation of Community Support Project to Improve Oral Function in Japanese Elderly.

    PubMed

    Sakayori, Takaharu; Maki, Yoshinobu; Ohkubo, Mai; Ishida, Ryo; Hirata, SoIchiro; Ishii, Takuo

    2016-01-01

    Change in oral function was evaluated longitudinally in elderly persons participating in an Exercises for Healthy Oral Function program implemented as part of the Long-Term Care Prevention Project. The participants comprised high-risk and healthy persons aged 65 yr or over. A questionnaire was used to classify them into two groups ('every day or sometimes' or 'rarely') at the end of the study for a comparison of change in repetitive saliva swallowing test (RSST) scores and oral diadochokinesis between 3 time points: at before, at immediately after, and at 1 yr after completion of the program. The average RSST score showed a decrease at 1 yr after intervention, but the difference was not statistically significant. Oral diadochokinesis showed a significant increase for all syllables upon completion of the program compared with at the beginning. This was followed by a significant decrease at 1 yr later compared with at the time of completion in the 'rarely' group for all syllables, but not in the 'every day or sometimes' group. In addition, the number of repetitions was significantly lower in the 'rarely' group than in the 'every day or sometimes' group for all syllables at 1 yr after completion. The results of the present study suggest that Long-Term Care Prevention Projects are necessary to maintain and improve oral function.

  5. How prior expectations shape multisensory perception.

    PubMed

    Gau, Remi; Noppeney, Uta

    2016-01-01

    The brain generates a representation of our environment by integrating signals from a common source, but segregating signals from different sources. This fMRI study investigated how the brain arbitrates between perceptual integration and segregation based on top-down congruency expectations and bottom-up stimulus-bound congruency cues. Participants were presented audiovisual movies of phonologically congruent, incongruent or McGurk syllables that can be integrated into an illusory percept (e.g. "ti" percept for visual «ki» with auditory /pi/). They reported the syllable they perceived. Critically, we manipulated participants' top-down congruency expectations by presenting McGurk stimuli embedded in blocks of congruent or incongruent syllables. Behaviorally, participants were more likely to fuse audiovisual signals into an illusory McGurk percept in congruent than incongruent contexts. At the neural level, the left inferior frontal sulcus (lIFS) showed increased activations for bottom-up incongruent relative to congruent inputs. Moreover, lIFS activations were increased for physically identical McGurk stimuli, when participants segregated the audiovisual signals and reported their auditory percept. Critically, this activation increase for perceptual segregation was amplified when participants expected audiovisually incongruent signals based on prior sensory experience. Collectively, our results demonstrate that the lIFS combines top-down prior (in)congruency expectations with bottom-up (in)congruency cues to arbitrate between multisensory integration and segregation. Copyright © 2015 Elsevier Inc. All rights reserved.

  6. Can Chunk Size Differences Explain Developmental Changes in Lexical Learning?

    PubMed Central

    Smalle, Eleonore H. M.; Bogaerts, Louisa; Simonis, Morgane; Duyck, Wouter; Page, Michael P. A.; Edwards, Martin G.; Szmalec, Arnaud

    2016-01-01

    In three experiments, we investigated Hebb repetition learning (HRL) differences between children and adults, as a function of the type of item (lexical vs. sub-lexical) and the level of item-overlap between sequences. In a first experiment, it was shown that when non-repeating and repeating (Hebb) sequences of words were all permutations of the same words, HRL was slower than when the sequences shared no words. This item-overlap effect was observed in both children and adults. In a second experiment, we used syllable sequences and we observed reduced HRL due to item-overlap only in children. The findings are explained within a chunking account of the HRL effect on the basis of which we hypothesize that children, compared with adults, chunk syllable sequences in smaller units. By hypothesis, small chunks are more prone to interference from anagram representations included in the filler sequences, potentially explaining the item-overlap effect in children. This hypothesis was tested in a third experiment with adults where we experimentally manipulated the chunk size by embedding pauses in the syllable sequences. Interestingly, we showed that imposing a small chunk size caused adults to show the same behavioral effects as those observed in children. Departing from the analogy between verbal HRL and lexical development, the results are discussed in light of the less-is-more hypothesis of age-related differences in language acquisition. PMID:26779065

  7. The homogeneity of audibility and prosody of Zulu words for speech reception threshold (SRT) testing.

    PubMed

    Panday, Seema; Kathard, Harsha; Pillay, Mershen; Govender, Cyril

    2009-01-01

    The aim of this investigation was to determine which of 58 preselected Zulu words developed by Panday et al. (2007) could be used for Speech Reception Threshold (SRT) testing. To realize this aim the homogeneity of audibility of 58 bisyllabic Zulu low tone verbs was measured, followed by an analysis of the prosodic features of the selected words. The words were digitally recorded by a Zulu first language male speaker and presented at 6 intensity levels to 30 Zulu first language speakers (18-25 years, mean age of 21.5 years), whose hearing was normal. Homogeneity of audibility was determined by employing logistic regression analysis. Twenty eight words met the criterion of homogeneity of audibility. This was evidenced by a mean slope of 50% at 5.98%/dB. The prosodic features of the twenty eight words were further analyzed using a computerized speech laboratory system. The findings confirmed that the pitch contours of the words followed the prosodic pattern apparent within Zulu linguistic structure. Eighty nine percent of the Zulu verbs were found to have a difference in the pitch pattern between the two syllables i.e. the first syllable was low in pitch, while the second syllable was high in pitch. It emerged that the twenty eight words could be used for establishing SRT within a normal hearing Zulu speaking population. Further research within clinical populations is recommended.

  8. A Musical Approach to Speech Melody

    PubMed Central

    Chow, Ivan; Brown, Steven

    2018-01-01

    We present here a musical approach to speech melody, one that takes advantage of the intervallic precision made possible with musical notation. Current phonetic and phonological approaches to speech melody either assign localized pitch targets that impoverish the acoustic details of the pitch contours and/or merely highlight a few salient points of pitch change, ignoring all the rest of the syllables. We present here an alternative model using musical notation, which has the advantage of representing the pitch of all syllables in a sentence as well as permitting a specification of the intervallic excursions among syllables and the potential for group averaging of pitch use across speakers. We tested the validity of this approach by recording native speakers of Canadian English reading unfamiliar test items aloud, spanning from single words to full sentences containing multiple intonational phrases. The fundamental-frequency trajectories of the recorded items were converted from hertz into semitones, averaged across speakers, and transcribed into musical scores of relative pitch. Doing so allowed us to quantify local and global pitch-changes associated with declarative, imperative, and interrogative sentences, and to explore the melodic dynamics of these sentence types. Our basic observation is that speech is atonal. The use of a musical score ultimately has the potential to combine speech rhythm and melody into a unified representation of speech prosody, an important analytical feature that is not found in any current linguistic approach to prosody. PMID:29556206

  9. Extricating Manual and Non-Manual Features for Subunit Level Medical Sign Modelling in Automatic Sign Language Classification and Recognition.

    PubMed

    R, Elakkiya; K, Selvamani

    2017-09-22

    Subunit segmenting and modelling in medical sign language is one of the important studies in linguistic-oriented and vision-based Sign Language Recognition (SLR). Many efforts were made in the precedent to focus the functional subunits from the view of linguistic syllables but the problem is implementing such subunit extraction using syllables is not feasible in real-world computer vision techniques. And also, the present recognition systems are designed in such a way that it can detect the signer dependent actions under restricted and laboratory conditions. This research paper aims at solving these two important issues (1) Subunit extraction and (2) Signer independent action on visual sign language recognition. Subunit extraction involved in the sequential and parallel breakdown of sign gestures without any prior knowledge on syllables and number of subunits. A novel Bayesian Parallel Hidden Markov Model (BPaHMM) is introduced for subunit extraction to combine the features of manual and non-manual parameters to yield better results in classification and recognition of signs. Signer independent action aims in using a single web camera for different signer behaviour patterns and for cross-signer validation. Experimental results have proved that the proposed signer independent subunit level modelling for sign language classification and recognition has shown improvement and variations when compared with other existing works.

  10. Phoneme categorization and discrimination in younger and older adults: a comparative analysis of perceptual, lexical, and attentional factors.

    PubMed

    Mattys, Sven L; Scharenborg, Odette

    2014-03-01

    This study investigates the extent to which age-related language processing difficulties are due to a decline in sensory processes or to a deterioration of cognitive factors, specifically, attentional control. Two facets of attentional control were examined: inhibition of irrelevant information and divided attention. Younger and older adults were asked to categorize the initial phoneme of spoken syllables ("Was it m or n?"), trying to ignore the lexical status of the syllables. The phonemes were manipulated to range in eight steps from m to n. Participants also did a discrimination task on syllable pairs ("Were the initial sounds the same or different?"). Categorization and discrimination were performed under either divided attention (concurrent visual-search task) or focused attention (no visual task). The results showed that even when the younger and older adults were matched on their discrimination scores: (1) the older adults had more difficulty inhibiting lexical knowledge than did younger adults, (2) divided attention weakened lexical inhibition in both younger and older adults, and (3) divided attention impaired sound discrimination more in older than younger listeners. The results confirm the independent and combined contribution of sensory decline and deficit in attentional control to language processing difficulties associated with aging. The relative weight of these variables and their mechanisms of action are discussed in the context of theories of aging and language. (c) 2014 APA, all rights reserved.

  11. Lexical tone and stuttering in Cantonese.

    PubMed

    Law, Thomas; Packman, Ann; Onslow, Mark; To, Carol K-S; Tong, Michael C-F; Lee, Kathy Y-S

    2018-01-01

    Cantonese is a tone language, in which the variation of the fundamental frequency contour of a syllable can change meaning. There are six different lexical tones in Cantonese. While research with Western languages has shown an association between stuttering and syllabic stress, nothing is known about whether stuttering in Cantonese speakers is associated with one or more of the six lexical tones. Such an association has been reported in conversational speech in Mandarin, which is also a tone language, but which varies markedly from Cantonese. Twenty-four native Cantonese-speaking adults who stutter participated in this study, ranging in age from 18-33 years. There were 18 men and 6 women. Participants read aloud 13 Cantonese syllables, each of which was produced with six contrastive lexical tones. All 78 syllables were embedded in the same carrier sentence, to reduce the influence of suprasegmental or linguistic stress, and were presented in random order. No significant differences were found for stuttering moments across the six lexical tones. It is suggested that this is because lexical tones, at least in Cantonese, do not place the task demands on the speech motor system that typify varying syllabic stress in Western languages: variations not only in fundamental frequency, but also in duration and intensity. The findings of this study suggest that treatments for adults who stutter in Western languages, such as speech restructuring, can be used with Cantonese speakers without undue attention to lexical tone.

  12. Effects of gender and regional dialect on prosodic patterns in American English

    PubMed Central

    Clopper, Cynthia G.; Smiljanic, Rajka

    2011-01-01

    While cross-dialect prosodic variation has been well established for many languages, most variationist research on regional dialects of American English has focused on the vowel system. The current study was designed to explore prosodic variation in read speech in two regional varieties of American English: Southern and Midland. Prosodic dialect variation was analyzed in two domains: speaking rate and the phonetic expression of pitch movements associated with accented and phrase-final syllables. The results revealed significant effects of regional dialect on the distributions of pauses, pitch accents, and phrasal-boundary tone combinations. Significant effects of talker gender were also observed on the distributions of pitch accents and phrasal-boundary tone combinations. The findings from this study demonstrate that regional and gender identity features are encoded in part through prosody, and provide further motivation for the close examination of prosodic patterns across regional and social varieties of American English. PMID:21686317

  13. Prosodic differences between declaratives and interrogatives in infant-directed speech.

    PubMed

    Geffen, Susan; Mintz, Toben H

    2017-07-01

    In many languages, declaratives and interrogatives differ in word order properties, and in syntactic organization more broadly. Thus, in order to learn the distinct syntactic properties of the two sentence types, learners must first be able to distinguish them using non-syntactic information. Prosodic information is often assumed to be a useful basis for this type of discrimination, although no systematic studies of the prosodic cues available to infants have been reported. Analysis of maternal speech in three Standard American English-speaking mother-infant dyads found that polar interrogatives differed from declaratives on the patterning of pitch and duration on the final two syllables, but wh-questions did not. Thus, while prosody is unlikely to aid discrimination of declaratives from wh-questions, infant-directed speech provides prosodic information that infants could use to distinguish declaratives and polar interrogatives. We discuss how learners could leverage this information to identify all question forms, in the context of syntax acquisition.

  14. Early sound symbolism for vowel sounds.

    PubMed

    Spector, Ferrinne; Maurer, Daphne

    2013-01-01

    Children and adults consistently match some words (e.g., kiki) to jagged shapes and other words (e.g., bouba) to rounded shapes, providing evidence for non-arbitrary sound-shape mapping. In this study, we investigated the influence of vowels on sound-shape matching in toddlers, using four contrasting pairs of nonsense words differing in vowel sound (/i/ as in feet vs. /o/ as in boat) and four rounded-jagged shape pairs. Crucially, we used reduplicated syllables (e.g., kiki vs. koko) rather than confounding vowel sound with consonant context and syllable variability (e.g., kiki vs. bouba). Toddlers consistently matched words with /o/ to rounded shapes and words with /i/ to jagged shapes (p < 0.01). The results suggest that there may be naturally biased correspondences between vowel sound and shape.

  15. Early sound symbolism for vowel sounds

    PubMed Central

    Spector, Ferrinne; Maurer, Daphne

    2013-01-01

    Children and adults consistently match some words (e.g., kiki) to jagged shapes and other words (e.g., bouba) to rounded shapes, providing evidence for non-arbitrary sound–shape mapping. In this study, we investigated the influence of vowels on sound–shape matching in toddlers, using four contrasting pairs of nonsense words differing in vowel sound (/i/ as in feet vs. /o/ as in boat) and four rounded–jagged shape pairs. Crucially, we used reduplicated syllables (e.g., kiki vs. koko) rather than confounding vowel sound with consonant context and syllable variability (e.g., kiki vs. bouba). Toddlers consistently matched words with /o/ to rounded shapes and words with /i/ to jagged shapes (p < 0.01). The results suggest that there may be naturally biased correspondences between vowel sound and shape. PMID:24349684

  16. Tone and prosodic organization in Cherokee nouns

    NASA Astrophysics Data System (ADS)

    Johnson, Keith; Haag, Marcia

    2005-04-01

    Preliminary observations in the speech of one speaker of Cherokee led us to postulate three factors affecting tone in Cherokee. (1) Tone may be lexically specified with distinctive low, low fall, low rise, and high tones. (2) There is a metrically determined high fall pattern which may be distributed over not more than 2 syllables from the right edge of a prosodic domain. (3) Intonational domains may be associated with discourse functions, marked by high fall, or by pitch range upstep. This paper tests these observations in recordings of word lists and sentences produced by five additional speakers. The analysis we give, positing both lexical tone and metrical prosodic accent, is not unique in descriptions of language, but is different from the usual description of Cherokee. [Work supported by NSF.

  17. The effects of age, viewing distance, display type, font type, colour contrast and number of syllables on the legibility of Korean characters.

    PubMed

    Kong, Yong-Ku; Lee, Inseok; Jung, Myung-Chul; Song, Young-Woong

    2011-05-01

    This study evaluated the effects of age (20s and 60s), viewing distance (50 cm, 200 cm), display type (paper, monitor), font type (Gothic, Ming), colour contrast (black letters on white background, white letters on black background) and number of syllables (one, two) on the legibility of Korean characters by using the four legibility measures (minimum letter size for 100% correctness, maximum letter size for 0% correctness, minimum letter size for the least discomfort and maximum letter size for the most discomfort). Ten subjects in each age group read the four letters presented on a slide (letter size varied from 80 pt to 2 pt). Subjects also subjectively rated the reading discomfort of the letters on a 4-point scale (1 = no discomfort, 4 = most discomfort). According to the ANOVA procedure, age, viewing distance and font type significantly affected the four dependent variables (p < 0.05), while the main effect of colour contrast was not statistically significant for any measures. Two-syllable letters had smaller letters than one-syllable letters in the two correctness measures. The younger group could see letter sizes two times smaller than the old group could and the viewing distance of 50 cm showed letters about three times smaller than those at a 200 cm viewing distance. The Gothic fonts were smaller than the Ming fonts. Monitors were smaller than paper for correctness and maximum letter size for the most discomfort. From a comparison of the results for correctness and discomfort, people generally preferred larger letter sizes to those that they could read. The findings of this study may provide basic information for setting a global standard of letter size or font type to improve the legibility of characters written in Korean. STATEMENT OF RELEVANCE: Results obtained in this study will provide basic information and guidelines for setting standards of letter size and font type to improve the legibility of characters written in Korean. Also, the results might offer useful information for people who are working on design of visual displays.

  18. Behavioural and physiological effects of population density on domesticated Zebra Finches (Taeniopygia guttata) held in aviaries.

    PubMed

    Poot, Hanneke; ter Maat, Andries; Trost, Lisa; Schwabl, Ingrid; Jansen, René F; Gahr, Manfred

    2012-02-01

    Zebra Finches (Taeniopygia guttata) are highly social and monogamous birds that display relatively low levels of aggression and coordinate group life mainly by means of vocal communication. In the wild, small groups may congregate to larger flocks of up to 150-350 birds. Little is known, however, about possible effects of population density on development in captivity. Investigating density effects on physiology and behaviour might be helpful in identifying optimal group size, in order to optimise Zebra Finch wellbeing. A direct effect of population density on development and reproduction was found: birds in lower density conditions produced significantly more and larger (body mass, tarsus length) surviving offspring than birds in high density conditions. Furthermore, offspring in low density aviaries produced slightly longer song motifs and more different syllables than their tutors, whereas offspring in high density aviaries produced shorter motifs and a smaller or similar number of different syllables than their tutors. Aggression levels within the populations were low throughout the experiment, but the number of aggressive interactions was significantly higher in high density aviaries. Baseline corticosterone levels did not differ significantly between high- and low density aviaries for either adult or offspring birds. On day 15 post hatching, brood size and baseline corticosterone levels were positively correlated. On days 60 and 100 post hatching this correlation was no longer present. The results of this study prove that population density affects various aspects of Zebra Finch development, with birds living in low population density conditions having an advantage over those living under higher population density conditions. Copyright © 2011 Elsevier Inc. All rights reserved.

  19. Popular song and lyrics synchronization and its application to music information retrieval

    NASA Astrophysics Data System (ADS)

    Chen, Kai; Gao, Sheng; Zhu, Yongwei; Sun, Qibin

    2006-01-01

    An automatic synchronization system of the popular song and its lyrics is presented in the paper. The system includes two main components: a) automatically detecting vocal/non-vocal in the audio signal and b) automatically aligning the acoustic signal of the song with its lyric using speech recognition techniques and positioning the boundaries of the lyrics in its acoustic realization at the multiple levels simultaneously (e.g. the word / syllable level and phrase level). The GMM models and a set of HMM-based acoustic model units are carefully designed and trained for the detection and alignment. To eliminate the severe mismatch due to the diversity of musical signal and sparse training data available, the unsupervised adaptation technique such as maximum likelihood linear regression (MLLR) is exploited for tailoring the models to the real environment, which improves robustness of the synchronization system. To further reduce the effect of the missed non-vocal music on alignment, a novel grammar net is build to direct the alignment. As we know, this is the first automatic synchronization system only based on the low-level acoustic feature such as MFCC. We evaluate the system on a Chinese song dataset collecting from 3 popular singers. We obtain 76.1% for the boundary accuracy at the syllable level (BAS) and 81.5% for the boundary accuracy at the phrase level (BAP) using fully automatic vocal/non-vocal detection and alignment. The synchronization system has many applications such as multi-modality (audio and textual) content-based popular song browsing and retrieval. Through the study, we would like to open up the discussion of some challenging problems when developing a robust synchronization system for largescale database.

  20. A neuroimaging study of conflict during word recognition.

    PubMed

    Riba, Jordi; Heldmann, Marcus; Carreiras, Manuel; Münte, Thomas F

    2010-08-04

    Using functional magnetic resonance imaging the neural activity associated with error commission and conflict monitoring in a lexical decision task was assessed. In a cohort of 20 native speakers of Spanish conflict was introduced by presenting words with high and low lexical frequency and pseudo-words with high and low syllabic frequency for the first syllable. Erroneous versus correct responses showed activation in the frontomedial and left inferior frontal cortex. A similar pattern was found for correctly classified words of low versus high lexical frequency and for correctly classified pseudo-words of high versus low syllabic frequency. Conflict-related activations for language materials largely overlapped with error-induced activations. The effect of syllabic frequency underscores the role of sublexical processing in visual word recognition and supports the view that the initial syllable mediates between the letter and word level.

  1. Stress errors in a case of developmental surface dyslexia in Filipino.

    PubMed

    Dulay, Katrina May; Hanley, J Richard

    2015-01-01

    This paper reports the case of a dyslexic boy (L.A.) whose impaired reading of Filipino is consistent with developmental surface dyslexia. Filipino has a transparent alphabetic orthography with stress typically falling on the penultimate syllable of multisyllabic words. However, exceptions to the typical stress pattern are not marked in the Filipino orthography. L.A. read words with typical stress patterns as accurately as controls, but made many more stress errors than controls when reading Filipino words with atypical stress. He regularized the pronunciation of many of these words by incorrectly placing the stress on the penultimate syllable. Since he also read nonwords as accurately and quickly as controls and performed well on tests of phonological awareness, L.A. appears to present a clear case of developmental surface dyslexia in a transparent orthography.

  2. Verbal implicit sequence learning in persons who stutter and persons with Parkinson's disease.

    PubMed

    Smits-Bandstra, Sarah; Gracco, Vincent

    2013-01-01

    The authors investigated the integrity of implicit learning systems in 14 persons with Parkinson's disease (PPD), 14 persons who stutter (PWS), and 14 control participants. In a 120-min session participants completed a verbal serial reaction time task, naming aloud 4 syllables in response to 4 visual stimuli. Unbeknownst to participants, the syllables formed a repeating 8-item sequence. PWS and PPD demonstrated slower reaction times for early but not late learning trials relative to controls reflecting delays but not deficiencies in general learning. PPD also demonstrated less accuracy in general learning relative to controls. All groups demonstrated similar limited explicit sequence knowledge. Both PWS and PPD demonstrated significantly less implicit sequence learning relative to controls, suggesting that stuttering may be associated with compromised functional integrity of the cortico-striato-thalamo-cortical loop.

  3. Acoustic foundations of the speech-to-song illusion.

    PubMed

    Tierney, Adam; Patel, Aniruddh D; Breen, Mara

    2018-06-01

    In the "speech-to-song illusion," certain spoken phrases are heard as highly song-like when isolated from context and repeated. This phenomenon occurs to a greater degree for some stimuli than for others, suggesting that particular cues prompt listeners to perceive a spoken phrase as song. Here we investigated the nature of these cues across four experiments. In Experiment 1, participants were asked to rate how song-like spoken phrases were after each of eight repetitions. Initial ratings were correlated with the consistency of an underlying beat and within-syllable pitch slope, while rating change was linked to beat consistency, within-syllable pitch slope, and melodic structure. In Experiment 2, the within-syllable pitch slope of the stimuli was manipulated, and this manipulation changed the extent to which participants heard certain stimuli as more musical than others. In Experiment 3, the extent to which the pitch sequences of a phrase fit a computational model of melodic structure was altered, but this manipulation did not have a significant effect on musicality ratings. In Experiment 4, the consistency of intersyllable timing was manipulated, but this manipulation did not have an effect on the change in perceived musicality after repetition. Our methods provide a new way of studying the causal role of specific acoustic features in the speech-to-song illusion via subtle acoustic manipulations of speech, and show that listeners can rapidly (and implicitly) assess the degree to which nonmusical stimuli contain musical structure. (PsycINFO Database Record (c) 2018 APA, all rights reserved).

  4. Spatiotemporal frequency characteristics of cerebral oscillations during the perception of fundamental frequency contour changes in one-syllable intonation.

    PubMed

    Ueno, Sanae; Okumura, Eiichi; Remijn, Gerard B; Yoshimura, Yuko; Kikuchi, Mitsuru; Shitamichi, Kiyomi; Nagao, Kikuko; Mochiduki, Masayuki; Haruta, Yasuhiro; Hayashi, Norio; Munesue, Toshio; Tsubokawa, Tsunehisa; Oi, Manabu; Nakatani, Hideo; Higashida, Haruhiro; Minabe, Yoshio

    2012-05-02

    Accurate perception of fundamental frequency (F0) contour changes in the human voice is important for understanding a speaker's intonation, and consequently also his/her attitude. In this study, we investigated the neural processes involved in the perception of F0 contour changes in the Japanese one-syllable interjection "ne" in 21 native-Japanese listeners. A passive oddball paradigm was applied in which "ne" with a high falling F0 contour, used when urging a reaction from the listener, was randomly presented as a rare deviant among a frequent "ne" syllable with a flat F0 contour (i.e., meaningless intonation). We applied an adaptive spatial filtering method to the neuromagnetic time course recorded by whole-head magnetoencephalography (MEG) and estimated the spatiotemporal frequency dynamics of event-related cerebral oscillatory changes in the oddball paradigm. Our results demonstrated a significant elevation of beta band event-related desynchronization (ERD) in the right temporal and frontal areas, in time windows from 100 to 300 and from 300 to 500 ms after the onset of deviant stimuli (high falling F0 contour). This is the first study to reveal detailed spatiotemporal frequency characteristics of cerebral oscillations during the perception of intonational (not lexical) F0 contour changes in the human voice. The results further confirmed that the right hemisphere is associated with perception of intonational F0 contour information in the human voice, especially in early time windows. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  5. Neural Correlates of Interindividual Differences in Children’s Audiovisual Speech Perception

    PubMed Central

    Nath, Audrey R.; Fava, Eswen E.; Beauchamp, Michael S.

    2011-01-01

    Children use information from both the auditory and visual modalities to aid in understanding speech. A dramatic illustration of this multisensory integration is the McGurk effect, an illusion in which an auditory syllable is perceived differently when it is paired with an incongruent mouth movement. However, there are significant interindividual differences in McGurk perception: some children never perceive the illusion, while others always do. Because converging evidence suggests that the posterior superior temporal sulcus (STS) is a critical site for multisensory integration, we hypothesized that activity within the STS would predict susceptibility to the McGurk effect. To test this idea, we used blood-oxygen level dependent functional magnetic resonance imaging (BOLD fMRI) in seventeen children aged 6 to 12 years to measure brain responses to three audiovisual stimulus categories: McGurk incongruent, non-McGurk incongruent and congruent syllables. Two separate analysis approaches, one using independent functional localizers and another using whole-brain voxel-based regression, showed differences in the left STS between perceivers and non-perceivers. The STS of McGurk perceivers responded significantly more than non-perceivers to McGurk syllables, but not to other stimuli, and perceivers’ hemodynamic responses in the STS were significantly prolonged. In addition to the STS, weaker differences between perceivers and non-perceivers were observed in the FFA and extrastriate visual cortex. These results suggest that the STS is an important source of interindividual variability in children’s audiovisual speech perception. PMID:21957257

  6. Where do interjections come from? A psycholinguistic analysis of Shaw's Pygmalion.

    PubMed

    O'Connell, Daniel C; Kowal, Sabine

    2005-09-01

    Starting from our recent findings regarding emotional and initializing functions of interjections in TV and radio interviews (Kowal & O'Connell, 2004b; O'Connell & Kowal, in press; O'Connell, Kowal, & Ageneau, 2005), we used the book and script of Shaw (1916/1969) and the audiotape of the motion picture (Pascal, Asquith, & Howard, 1938) Pygmalion to investigate how actors use interjections to express emotions. The following hypotheses were tested: (1) The actors use the written cues selectively in their oral performance by substituting, adding, and deleting interjections; (2) primary interjections added by the actors are less conventional than those in the written text; (3) durations and number of syllables of Eliza Doolittle's spoken renditions of her signature interjection ah-ah-ah-ow-ow-ow-oo do not correlate with the length in letters and syllables of the written versions; and (4) there is no evidence for Ameka's (1992b, 1994) characterization of interjections as temporally isolated, i.e., preceded and followed by silent pauses, in consequence of their syntactic isolation. Our findings confirmed all the hypotheses except for one unexpectedly significant correlation between number of syllables in Eliza Doolittle's signature interjection in the written version and duration in seconds of the spoken version thereof. The common thread throughout these data is the actor's need to personalize emotions in a dramatic performance--by means of interjections other than those provided in the written text. In this process of personalization, the emotional and initializing functions of interjections are confirmed.

  7. Speech processing asymmetry revealed by dichotic listening and functional brain imaging.

    PubMed

    Hugdahl, Kenneth; Westerhausen, René

    2016-12-01

    In this article, we review research in our laboratory from the last 25 to 30 years on the neuronal basis for laterality of speech perception focusing on the upper, posterior parts of the temporal lobes, and its functional and structural connections to other brain regions. We review both behavioral and brain imaging data, with a focus on dichotic listening experiments, and using a variety of imaging modalities. The data have come in most parts from healthy individuals and from studies on normally functioning brain, although we also review a few selected clinical examples. We first review and discuss the structural model for the explanation of the right-ear advantage (REA) and left hemisphere asymmetry for auditory language processing. A common theme across many studies have been our interest in the interaction between bottom-up, stimulus-driven, and top-down, instruction-driven, aspects of hemispheric asymmetry, and how perceptual factors interact with cognitive factors to shape asymmetry of auditory language information processing. In summary, our research have shown laterality for the initial processing of consonant-vowel syllables, first observed as a behavioral REA when subjects are required to report which syllable of a dichotic syllable-pair they perceive. In subsequent work we have corroborated the REA with brain imaging, and have shown that the REA is modulated through both bottom-up manipulations of stimulus properties, like sound intensity, and top-down manipulations of cognitive properties, like attention focus. Copyright © 2015 Elsevier Ltd. All rights reserved.

  8. Developmental weighting shifts for noise components of fricative-vowel syllables.

    PubMed

    Nittrouer, S; Miller, M E

    1997-07-01

    Previous studies have convincingly shown that the weight assigned to vocalic formant transitions in decisions of fricative identity for fricative-vowel syllables decreases with development. Although these same studies suggested a developmental increase in the weight assigned to the noise spectrum, the role of the aperiodic-noise portions of the signals in these fricative decisions have not been as well-studied. The purpose of these experiments was to examine more closely developmental shifts in the weight assigned to the aperiodic-noise components of the signals in decisions of syllable-initial fricative identity. Two experiments used noises varying along continua from a clear /s/ percept to a clear /[symbol: see text]/ percept. In experiment 1, these noises were created by combining /s/ and /[symbol: see text]/ noises produced by a human vocal tract at different amplitude ratios, a process that resulted in stimuli differing primarily in the amplitude of a relatively low-frequency (roughly 2.2-kHz) peak. In experiment 2, noises that varied only in the amplitude of a similar low-frequency peak were created with a software synthesizer. Both experiments used synthetic /a/ and /u/ portions, and efforts were made to minimize possible contributions of vocalic formant transitions to fricative labeling. Children and adults labeled the resulting stimuli as /s/ vowel or /[symbol: see text]/ vowel. Combined results of the two experiments showed that children's responses were less influenced than those of adults by the amplitude of the low-frequency peak of fricative noises.

  9. A Mechanism for Frequency Modulation in Songbirds Shared with Humans

    PubMed Central

    Margoliash, Daniel

    2013-01-01

    In most animals that vocalize, control of fundamental frequency is a key element for effective communication. In humans, subglottal pressure controls vocal intensity but also influences fundamental frequency during phonation. Given the underlying similarities in the biomechanical mechanisms of vocalization in humans and songbirds, songbirds offer an attractive opportunity to study frequency modulation by pressure. Here, we present a novel technique for dynamic control of subsyringeal pressure in zebra finches. By regulating the opening of a custom-built fast valve connected to the air sac system, we achieved partial or total silencing of specific syllables, and could modify syllabic acoustics through more complex manipulations of air sac pressure. We also observed that more nuanced pressure variations over a limited interval during production of a syllable concomitantly affected the frequency of that syllable segment. These results can be explained in terms of a mathematical model for phonation that incorporates a nonlinear description for the vocal source capable of generating the observed frequency modulations induced by pressure variations. We conclude that the observed interaction between pressure and frequency was a feature of the source, not a result of feedback control. Our results indicate that, beyond regulating phonation or its absence, regulation of pressure is important for control of fundamental frequencies of vocalizations. Thus, although there are separate brainstem pathways for syringeal and respiratory control of song production, both can affect airflow and frequency. We hypothesize that the control of pressure and frequency is combined holistically at higher levels of the vocalization pathways. PMID:23825417

  10. Automatic processing of tones and speech stimuli in children with specific language impairment.

    PubMed

    Uwer, Ruth; Albrecht, Ronald; von Suchodoletz, W

    2002-08-01

    It is well known from behavioural experiments that children with specific language impairment (SLI) have difficulties discriminating consonant-vowel (CV) syllables such as /ba/, /da/, and /ga/. Mismatch negativity (MMN) is an auditory event-related potential component that represents the outcome of an automatic comparison process. It could, therefore, be a promising tool for assessing central auditory processing deficits for speech and non-speech stimuli in children with SLI. MMN is typically evoked by occasionally occurring 'deviant' stimuli in a sequence of identical 'standard' sounds. In this study MMN was elicited using simple tone stimuli, which differed in frequency (1000 versus 1200 Hz) and duration (175 versus 100 ms) and to digitized CV syllables which differed in place of articulation (/ba/, /da/, and /ga/) in children with expressive and receptive SLI and healthy control children (n=21 in each group, 46 males and 17 females; age range 5 to 10 years). Mean MMN amplitudes between groups were compared. Additionally, the behavioural discrimination performance was assessed. Children with SLI had attenuated MMN amplitudes to speech stimuli, but there was no significant difference between the two diagnostic subgroups. MMN to tone stimuli did not differ between the groups. Children with SLI made more errors in the discrimination task, but discrimination scores did not correlate with MMN amplitudes. The present data suggest that children with SLI show a specific deficit in automatic discrimination of CV syllables differing in place of articulation, whereas the processing of simple tone differences seems to be unimpaired.

  11. Complex Sequencing Rules of Birdsong Can be Explained by Simple Hidden Markov Processes

    PubMed Central

    Katahira, Kentaro; Suzuki, Kenta; Okanoya, Kazuo; Okada, Masato

    2011-01-01

    Complex sequencing rules observed in birdsongs provide an opportunity to investigate the neural mechanism for generating complex sequential behaviors. To relate the findings from studying birdsongs to other sequential behaviors such as human speech and musical performance, it is crucial to characterize the statistical properties of the sequencing rules in birdsongs. However, the properties of the sequencing rules in birdsongs have not yet been fully addressed. In this study, we investigate the statistical properties of the complex birdsong of the Bengalese finch (Lonchura striata var. domestica). Based on manual-annotated syllable labeles, we first show that there are significant higher-order context dependencies in Bengalese finch songs, that is, which syllable appears next depends on more than one previous syllable. We then analyze acoustic features of the song and show that higher-order context dependencies can be explained using first-order hidden state transition dynamics with redundant hidden states. This model corresponds to hidden Markov models (HMMs), well known statistical models with a large range of application for time series modeling. The song annotation with these models with first-order hidden state dynamics agreed well with manual annotation, the score was comparable to that of a second-order HMM, and surpassed the zeroth-order model (the Gaussian mixture model; GMM), which does not use context information. Our results imply that the hierarchical representation with hidden state dynamics may underlie the neural implementation for generating complex behavioral sequences with higher-order dependencies. PMID:21915345

  12. A mechanism for frequency modulation in songbirds shared with humans.

    PubMed

    Amador, Ana; Margoliash, Daniel

    2013-07-03

    In most animals that vocalize, control of fundamental frequency is a key element for effective communication. In humans, subglottal pressure controls vocal intensity but also influences fundamental frequency during phonation. Given the underlying similarities in the biomechanical mechanisms of vocalization in humans and songbirds, songbirds offer an attractive opportunity to study frequency modulation by pressure. Here, we present a novel technique for dynamic control of subsyringeal pressure in zebra finches. By regulating the opening of a custom-built fast valve connected to the air sac system, we achieved partial or total silencing of specific syllables, and could modify syllabic acoustics through more complex manipulations of air sac pressure. We also observed that more nuanced pressure variations over a limited interval during production of a syllable concomitantly affected the frequency of that syllable segment. These results can be explained in terms of a mathematical model for phonation that incorporates a nonlinear description for the vocal source capable of generating the observed frequency modulations induced by pressure variations. We conclude that the observed interaction between pressure and frequency was a feature of the source, not a result of feedback control. Our results indicate that, beyond regulating phonation or its absence, regulation of pressure is important for control of fundamental frequencies of vocalizations. Thus, although there are separate brainstem pathways for syringeal and respiratory control of song production, both can affect airflow and frequency. We hypothesize that the control of pressure and frequency is combined holistically at higher levels of the vocalization pathways.

  13. Singing-Related Activity in Anterior Forebrain of Male Zebra Finches Reflects Courtship Motivation for Target Females

    PubMed Central

    Iwasaki, Mai; Poulsen, Thomas M.; Oka, Kotaro; Hessler, Neal A.

    2013-01-01

    A critical function of singing by male songbirds is to attract a female mate. Previous studies have suggested that the anterior forebrain system is involved in this courtship behavior. Neural activity in this system, including the striatal Area X, is strikingly dependent on the function of male singing. When males sing to attract a female bird rather than while alone, less variable neural activity results in less variable song spectral features, which may be attractive to the female. These characteristics of neural activity and singing thus may reflect a male's motivation for courtship. Here, we compared the variability of neural activity and song features between courtship singing directed to a female with whom a male had previously formed a pair-bond or to other females. Surprisingly, across all units, there was no clear tendency for a difference in variability of neural activity or song features between courtship of paired females, nonpaired females, or dummy females. However, across the population of recordings, there was a significant relationship between the relative variability of syllable frequency and neural activity: when syllable frequency was less variable to paired than nonpaired females, neural activity was also less variable (and vice-versa). These results show that the lower variability of neural activity and syllable frequency during directed singing is not a binary distinction from undirected singing, but can vary in intensity, possibly related to the relative preference of a male for his singing target. PMID:24312344

  14. Objective and subjective assessment of tracheoesophageal prosthesis voice outcome.

    PubMed

    D'Alatri, Lucia; Bussu, Francesco; Scarano, Emanuele; Paludetti, Gaetano; Marchese, Maria Raffaella

    2012-09-01

    To investigate the relationships between objective measures and the results of subjective assessment of voice quality and speech intelligibility in patients submitted to total laryngectomy and tracheoesophageal (TE) puncture. Retrospective. Twenty patients implanted with voice prosthesis were studied. After surgery, the entire sample performed speech rehabilitation. The assessment protocol included maximum phonation time (MPT), number of syllables per deep breath, acoustic analysis of the sustained vowel /a/ and of a bisyllabic word, perceptual evaluation (pleasantness and intelligibility%), and self-assessment. The correlation between pleasantness and intelligibility% was statistically significant. Both the latter were significantly correlated with the acoustic signal type, the number of formant peaks, and the F2-F1 difference. The intelligibility% and number of formant peaks were significantly correlated with the MPT and number of syllables per deep breath. Moreover, significant correlations were found between the number of formant peaks and both intelligibility% and pleasantness. The higher the number of syllables per deep breath and the longer the MPT, significantly higher was the number of formant peaks and the intelligibility%. The study failed to show significant correlation between patient's self-assessment of voice quality and both pleasantness and communication effectiveness. The multidimensional assessment seems to be a reliable tool to evaluate the TE functional outcome. Particularly, the results showed that both pleasantness and intelligibility of TE speech are correlated to the availability of expired air and the function of the vocal tract. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  15. Effects of measurement method and transcript availability on inexperienced raters' stuttering frequency scores.

    PubMed

    Chakraborty, Nalanda; Logan, Kenneth J

    To examine the effects of measurement method and transcript availability on the accuracy, reliability, and efficiency of inexperienced raters' stuttering frequency measurements. 44 adults, all inexperienced at evaluating stuttered speech, underwent 20 min of preliminary training in stuttering measurement and then analyzed a series of sentences, with and without access to transcripts of sentence stimuli, using either a syllable-based analysis (SBA) or an utterance-based analysis (UBA). Participants' analyses were compared between groups and to a composite analysis from two experienced evaluators. Stuttering frequency scores from the SBA and UBA groups differed significantly from the experienced evaluators' scores; however, UBA scores were significantly closer to the experienced evaluators' scores and were completed significantly faster than the SBA scores. Transcript availability facilitated scoring accuracy and efficiency in both groups. The internal reliability of stuttering frequency scores was acceptable for the SBA and UBA groups; however, the SBA group demonstrated only modest point-by-point agreement with ratings from the experienced evaluators. Given its accuracy and efficiency advantages over syllable-based analysis, utterance-based fluency analysis appears to be an appropriate context for introducing stuttering frequency measurement to raters who have limited experience in stuttering measurement. To address accuracy gaps between experienced and inexperienced raters, however, use of either analysis must be supplemented with training activities that expose inexperienced raters to the decision-making processes used by experienced raters when identifying stuttered syllables. Copyright © 2018 Elsevier Inc. All rights reserved.

  16. Phonemic accuracy development in children with cochlear implants up to five years of age by using Levenshtein distance.

    PubMed

    Faes, Jolien; Gillis, Joris; Gillis, Steven

    2016-01-01

    Phonemic accuracy of children with cochlear implants (CI) is often reported to be lower in comparison with normally hearing (NH) age-matched children. In this study, we compare phonemic accuracy development in the spontaneous speech of Dutch-speaking children with CI and NH age-matched peers. A dynamic cost model of Levenshtein distance is used to compute the accuracy of each word token. We set up a longitudinal design with monthly data for comparisons up to age two and a cross-sectional design with yearly data between three and five years of age. The main finding is that phonemic accuracy steadily increases throughout the period studied. Children with CI's accuracy is lower than that of their NH age mates, but this difference is not statistically significant in the earliest stages of lexical development. But accuracy of children with CI initially improves significantly less steeply than that of NH peers. Furthermore, the number of syllables in the target word and target word's complexity influence children's accuracy, as longer and more complex target words are less accurately produced. Up to age four, children with CI are significantly less accurate than NH children with increasing word length and word complexity. This difference has disappeared at age five. Finally, hearing age is shown to influence accuracy development of children with CI, while age of implant activation is not. This article informs the reader about phonemic accuracy development in children. The reader will be able to (a) discuss different metrics to measure phonemic accuracy development, (b) discuss phonemic accuracy of children with CI up to five years of age and compare them with NH children, (c) discuss the influence of target word's complexity and target word's syllable length on phonemic accuracy, (d) discuss the influence of hearing experience and age of implantation on phonemic accuracy of children with CI. Copyright © 2015 Elsevier Inc. All rights reserved.

  17. Basal ganglia function, stuttering, sequencing, and repair in adult songbirds.

    PubMed

    Kubikova, Lubica; Bosikova, Eva; Cvikova, Martina; Lukacova, Kristina; Scharff, Constance; Jarvis, Erich D

    2014-10-13

    A pallial-basal-ganglia-thalamic-pallial loop in songbirds is involved in vocal motor learning. Damage to its basal ganglia part, Area X, in adult zebra finches has been noted to have no strong effects on song and its function is unclear. Here we report that neurotoxic damage to adult Area X induced changes in singing tempo and global syllable sequencing in all animals, and considerably increased syllable repetition in birds whose song motifs ended with minor repetitions before lesioning. This stuttering-like behavior started at one month, and improved over six months. Unexpectedly, the lesioned region showed considerable recovery, including immigration of newly generated or repaired neurons that became active during singing. The timing of the recovery and stuttering suggest that immature recovering activity of the circuit might be associated with stuttering. These findings indicate that even after juvenile learning is complete, the adult striatum plays a role in higher level organization of learned vocalizations.

  18. Temporal attractors for speech onsets

    NASA Astrophysics Data System (ADS)

    Port, Robert; Oglesbee, Eric

    2003-10-01

    When subjects say a single syllable like da in time with a metronome, what is the easiest relationship? Superimposed on the metronome pulse, of course. The second easiest way is probably to locate the syllable halfway between pulses. We tested these hypotheses by having subjects repeat da at both phase angles at a range of metronome rates. The vowel onset (or P-center) was automatically obtained for each token. In-phase targets were produced close to the metronome onset for rates as fast as 3 per second. Antiphase targets were accurate at slow rates (~2/s) but tended to slip to inphase timing with faster metronomes. These results resemble the findings of Haken et al. [Biol. Cybern. 51, 347-356 (1985)] for oscillatory finger motions. Results suggest a strong attractor for speech onsets at zero phase and a weaker attractor at phase 0.5 that may disappear as rate is increased.

  19. Perception of suprasegmental features of speech by children with cochlear implants and children with hearing AIDS.

    PubMed

    Most, Tova; Peled, Miriam

    2007-01-01

    This study assessed perception of suprasegmental features of speech by 30 prelingual children with sensorineural hearing loss. Ten children had cochlear implants (CIs), and 20 children wore hearing aids (HA): 10 with severe hearing loss and 10 with profound hearing loss. Perception of intonation, syllable stress, word emphasis, and word pattern was assessed. Results revealed that the two HA groups significantly outperformed the CI group in perceiving both intonation and stress. Within each group, word pattern was perceived best, and then intonation and emphasis, with syllable stress perceived poorest. No significant correlation emerged between age at implantation and perception of the various suprasegmental features, possibly due to participants' relatively late age at implantation. Results indicated that CI use did not show an advantage over HA use in the perception of suprasegmental features of speech. Future research should continue to explore variables that might improve this perception.

  20. Speech rate reduction and "nasality" in normal speakers.

    PubMed

    Brancewicz, T M; Reich, A R

    1989-12-01

    This study explored the effects of reduced speech rate on nasal/voice accelerometric measures and nasality ratings. Nasal/voice accelerometric measures were obtained from normal adults for various speech stimuli and speaking rates. Stimuli included three sentences (one obstruent-loaded, one semivowel-loaded, and one containing a single nasal), and /pv/ syllable trains.. Speakers read the stimuli at their normal rate, half their normal rate, and as slowly as possible. In addition, a computer program paced each speaker at rates of 1, 2, and 3 syllables per second. The nasal/voice accelerometric values revealed significant stimulus effects but no rate effects. The nasality ratings of experienced listeners, evaluated as a function of stimulus and speaking rate, were compared to the accelerometric measures. The nasality scale values demonstrated small, but statistically significant, stimulus and rate effects. However, the nasality percepts were poorly correlated with the nasal/voice accelerometric measures.

  1. Mind the gap: Neural coding of species identity in birdsong prosody.

    PubMed

    Araki, Makoto; Bandi, M M; Yazaki-Sugiyama, Yoko

    2016-12-09

    Juvenile songbirds learn vocal communication from adult tutors of the same species but not from adults of other species. How species-specific learning emerges from the basic features of song prosody remains unknown. In the zebra finch auditory cortex, we discovered a class of neurons that register the silent temporal gaps between song syllables and are distinct from neurons encoding syllable morphology. Behavioral learning and neuronal coding of temporal gap structure resisted song tutoring from other species: Zebra finches fostered by Bengalese finch parents learned Bengalese finch song morphology transposed onto zebra finch temporal gaps. During the vocal learning period, temporal gap neurons fired selectively to zebra finch song. The innate temporal coding of intersyllable silent gaps suggests a neuronal barcode for conspecific vocal learning and social communication in acoustically diverse environments. Copyright © 2016, American Association for the Advancement of Science.

  2. Knockout of Foxp2 disrupts vocal development in mice.

    PubMed

    Castellucci, Gregg A; McGinley, Matthew J; McCormick, David A

    2016-03-16

    The FOXP2 gene is important for the development of proper speech motor control in humans. However, the role of the gene in general vocal behavior in other mammals, including mice, is unclear. Here, we track the vocal development of Foxp2 heterozygous knockout (Foxp2+/-) mice and their wildtype (WT) littermates from juvenile to adult ages, and observe severe abnormalities in the courtship song of Foxp2+/- mice. In comparison to their WT littermates, Foxp2+/- mice vocalized less, produced shorter syllable sequences, and possessed an abnormal syllable inventory. In addition, Foxp2+/- song also exhibited irregular rhythmic structure, and its development did not follow the consistent trajectories observed in WT vocalizations. These results demonstrate that the Foxp2 gene is critical for normal vocal behavior in juvenile and adult mice, and that Foxp2 mutant mice may provide a tractable model system for the study of the gene's role in general vocal motor control.

  3. Singing proficiency in congenital amusia: imitation helps.

    PubMed

    Tremblay-Champoux, Alexandra; Dalla Bella, Simone; Phillips-Silver, Jessica; Lebrun, Marie-Andrée; Peretz, Isabelle

    2010-09-01

    Singing out of tune characterizes congenital amusia. Here, we examine whether an aid to memory improves singing by studying vocal imitation in 11 amusic adults and 11 matched controls. Participants sang a highly familiar melody on the original lyrics and on the syllable /la/ in three conditions. First, they sang the melody from memory. Second, they sang it after hearing a model, and third, they sang in unison with the model. Results show that amusic individuals benefit from singing by imitation, whether singing after the model or in unison with the model. The amusics who were the most impaired in memory benefited most, particularly when singing on the syllable /la/. Nevertheless, singing remains poor on the pitch dimension; rhythm was intact and unaffected by imitation. These results point to memory as a source of impairment in poor singing, and to imitation as a possible aid for poor singers.

  4. A common neural circuit mechanism for internally guided and externally reinforced forms of motor learning.

    PubMed

    Hisey, Erin; Kearney, Matthew Gene; Mooney, Richard

    2018-04-01

    The complex skills underlying verbal and musical expression can be learned without external punishment or reward, indicating their learning is internally guided. The neural mechanisms that mediate internally guided learning are poorly understood, but a circuit comprising dopamine-releasing neurons in the midbrain ventral tegmental area (VTA) and their targets in the basal ganglia are important to externally reinforced learning. Juvenile zebra finches copy a tutor song in a process that is internally guided and, in adulthood, can learn to modify the fundamental frequency (pitch) of a target syllable in response to external reinforcement with white noise. Here we combined intersectional genetic ablation of VTA neurons, reversible blockade of dopamine receptors in the basal ganglia, and singing-triggered optogenetic stimulation of VTA terminals to establish that a common VTA-basal ganglia circuit enables internally guided song copying and externally reinforced syllable pitch learning.

  5. Typography manipulations can affect priming of word stem completion in older and younger adults.

    PubMed

    Gibson, J M; Brooks, J O; Friedman, L; Yesavage, J A

    1993-12-01

    The experiments reported here investigated whether changes of typography affected priming of word stem completion performance in older and younger adults. Across all experiments, the typeface in which a word appeared at presentation either did or did not match that of its 3-letter stem at test. In Experiment 1, no significant evidence of a typography effect was found when words were presented with a sentence judgment or letter judgment task. However, subsequent experiments revealed that, in both older and younger adults, only words presented with a syllable judgment task gave rise to the typography effect (Experiments 2-4). Specifically, performance was greater, when the presentation and test typeface matched than when they did not. Experiment 5, which used stem-cued recall, did not reveal a difference between syllable and letter judgment tasks. These findings highlight the complex nature of word stem completion performance.

  6. Dissociating word stem completion and cued recall as a function of divided attention at retrieval.

    PubMed

    Clarke, A J Benjamin; Butler, Laurie T

    2008-10-01

    The aim of this study was to investigate the widely held, but largely untested, view that implicit memory (repetition priming) reflects an automatic form of retrieval. Specifically, in Experiment 1 we explored whether a secondary task (syllable monitoring), performed during retrieval, would disrupt performance on explicit (cued recall) and implicit (stem completion) memory tasks equally. Surprisingly, despite substantial memory and secondary costs to cued recall when performed with a syllable-monitoring task, the same manipulation had no effect on stem completion priming or on secondary task performance. In Experiment 2 we demonstrated that even when using a particularly demanding version of the stem completion task that incurred secondary task costs, the corresponding disruption to implicit memory performance was minimal. Collectively, the results are consistent with the view that implicit memory retrieval requires little or no processing capacity and is not seemingly susceptible to the effects of dividing attention at retrieval.

  7. Vietnamese Document Representation and Classification

    NASA Astrophysics Data System (ADS)

    Nguyen, Giang-Son; Gao, Xiaoying; Andreae, Peter

    Vietnamese is very different from English and little research has been done on Vietnamese document classification, or indeed, on any kind of Vietnamese language processing, and only a few small corpora are available for research. We created a large Vietnamese text corpus with about 18000 documents, and manually classified them based on different criteria such as topics and styles, giving several classification tasks of different difficulty levels. This paper introduces a new syllable-based document representation at the morphological level of the language for efficient classification. We tested the representation on our corpus with different classification tasks using six classification algorithms and two feature selection techniques. Our experiments show that the new representation is effective for Vietnamese categorization, and suggest that best performance can be achieved using syllable-pair document representation, an SVM with a polynomial kernel as the learning algorithm, and using Information gain and an external dictionary for feature selection.

  8. Seeking Temporal Predictability in Speech: Comparing Statistical Approaches on 18 World Languages.

    PubMed

    Jadoul, Yannick; Ravignani, Andrea; Thompson, Bill; Filippi, Piera; de Boer, Bart

    2016-01-01

    Temporal regularities in speech, such as interdependencies in the timing of speech events, are thought to scaffold early acquisition of the building blocks in speech. By providing on-line clues to the location and duration of upcoming syllables, temporal structure may aid segmentation and clustering of continuous speech into separable units. This hypothesis tacitly assumes that learners exploit predictability in the temporal structure of speech. Existing measures of speech timing tend to focus on first-order regularities among adjacent units, and are overly sensitive to idiosyncrasies in the data they describe. Here, we compare several statistical methods on a sample of 18 languages, testing whether syllable occurrence is predictable over time. Rather than looking for differences between languages, we aim to find across languages (using clearly defined acoustic, rather than orthographic, measures), temporal predictability in the speech signal which could be exploited by a language learner. First, we analyse distributional regularities using two novel techniques: a Bayesian ideal learner analysis, and a simple distributional measure. Second, we model higher-order temporal structure-regularities arising in an ordered series of syllable timings-testing the hypothesis that non-adjacent temporal structures may explain the gap between subjectively-perceived temporal regularities, and the absence of universally-accepted lower-order objective measures. Together, our analyses provide limited evidence for predictability at different time scales, though higher-order predictability is difficult to reliably infer. We conclude that temporal predictability in speech may well arise from a combination of individually weak perceptual cues at multiple structural levels, but is challenging to pinpoint.

  9. Sensorimotor speech disorders in Parkinson's disease: Programming and execution deficits.

    PubMed

    Ortiz, Karin Zazo; Brabo, Natalia Casagrande; Minett, Thais Soares C

    2016-01-01

    Dysfunction in the basal ganglia circuits is a determining factor in the physiopathology of the classic signs of Parkinson's disease (PD) and hypokinetic dysarthria is commonly related to PD. Regarding speech disorders associated with PD, the latest four-level framework of speech complicates the traditional view of dysarthria as a motor execution disorder. Based on findings that dysfunctions in basal ganglia can cause speech disorders, and on the premise that the speech deficits seen in PD are not related to an execution motor disorder alone but also to a disorder at the motor programming level, the main objective of this study was to investigate the presence of sensorimotor disorders of programming (besides the execution disorders previously described) in PD patients. A cross-sectional study was conducted in a sample of 60 adults matched for gender, age and education: 30 adult patients diagnosed with idiopathic PD (PDG) and 30 healthy adults (CG). All types of articulation errors were reanalyzed to investigate the nature of these errors. Interjections, hesitations and repetitions of words or sentences (during discourse) were considered typical disfluencies; blocking, episodes of palilalia (words or syllables) were analyzed as atypical disfluencies. We analysed features including successive self-initiated trial, phoneme distortions, self-correction, repetition of sounds and syllables, prolonged movement transitions, additions or omissions of sounds and syllables, in order to identify programming and/or execution failures. Orofacial agility was also investigated. The PDG had worse performance on all sensorimotor speech tasks. All PD patients had hypokinetic dysarthria. The clinical characteristics found suggest both execution and programming sensorimotor speech disorders in PD patients.

  10. Cascaded processing in written compound word production

    PubMed Central

    Bertram, Raymond; Tønnessen, Finn Egil; Strömqvist, Sven; Hyönä, Jukka; Niemi, Pekka

    2015-01-01

    In this study we investigated the intricate interplay between central linguistic processing and peripheral motor processes during typewriting. Participants had to typewrite two-constituent (noun-noun) Finnish compounds in response to picture presentation while their typing behavior was registered. As dependent measures we used writing onset time to assess what processes were completed before writing and inter-key intervals to assess what processes were going on during writing. It was found that writing onset time was determined by whole word frequency rather than constituent frequencies, indicating that compound words are retrieved as whole orthographic units before writing is initiated. In addition, we found that the length of the first syllable also affects writing onset time, indicating that the first syllable is fully prepared before writing commences. The inter-key interval results showed that linguistic planning is not fully ready before writing, but cascades into the motor execution phase. More specifically, inter-key intervals were largest at syllable and morpheme boundaries, supporting the view that additional linguistic planning takes place at these boundaries. Bigram and trigram frequency also affected inter-key intervals with shorter intervals corresponding to higher frequencies. This can be explained by stronger memory traces for frequently co-occurring letter sequences in the motor memory for typewriting. These frequency effects were even larger in the second than in the first constituent, indicating that low-level motor memory starts to become more important during the course of writing compound words. We discuss our results in the light of current models of morphological processing and written word production. PMID:25954182

  11. Cascaded processing in written compound word production.

    PubMed

    Bertram, Raymond; Tønnessen, Finn Egil; Strömqvist, Sven; Hyönä, Jukka; Niemi, Pekka

    2015-01-01

    In this study we investigated the intricate interplay between central linguistic processing and peripheral motor processes during typewriting. Participants had to typewrite two-constituent (noun-noun) Finnish compounds in response to picture presentation while their typing behavior was registered. As dependent measures we used writing onset time to assess what processes were completed before writing and inter-key intervals to assess what processes were going on during writing. It was found that writing onset time was determined by whole word frequency rather than constituent frequencies, indicating that compound words are retrieved as whole orthographic units before writing is initiated. In addition, we found that the length of the first syllable also affects writing onset time, indicating that the first syllable is fully prepared before writing commences. The inter-key interval results showed that linguistic planning is not fully ready before writing, but cascades into the motor execution phase. More specifically, inter-key intervals were largest at syllable and morpheme boundaries, supporting the view that additional linguistic planning takes place at these boundaries. Bigram and trigram frequency also affected inter-key intervals with shorter intervals corresponding to higher frequencies. This can be explained by stronger memory traces for frequently co-occurring letter sequences in the motor memory for typewriting. These frequency effects were even larger in the second than in the first constituent, indicating that low-level motor memory starts to become more important during the course of writing compound words. We discuss our results in the light of current models of morphological processing and written word production.

  12. A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech.

    PubMed

    Magnotti, John F; Beauchamp, Michael S

    2017-02-01

    Audiovisual speech integration combines information from auditory speech (talker's voice) and visual speech (talker's mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in audiovisual speech perception is deciding whether auditory and visual speech have the same source, a process known as causal inference. A well-known illusion, the McGurk Effect, consists of incongruent audiovisual syllables, such as auditory "ba" + visual "ga" (AbaVga), that are integrated to produce a fused percept ("da"). This illusion raises two fundamental questions: first, given the incongruence between the auditory and visual syllables in the McGurk stimulus, why are they integrated; and second, why does the McGurk effect not occur for other, very similar syllables (e.g., AgaVba). We describe a simplified model of causal inference in multisensory speech perception (CIMS) that predicts the perception of arbitrary combinations of auditory and visual speech. We applied this model to behavioral data collected from 60 subjects perceiving both McGurk and non-McGurk incongruent speech stimuli. The CIMS model successfully predicted both the audiovisual integration observed for McGurk stimuli and the lack of integration observed for non-McGurk stimuli. An identical model without causal inference failed to accurately predict perception for either form of incongruent speech. The CIMS model uses causal inference to provide a computational framework for studying how the brain performs one of its most important tasks, integrating auditory and visual speech cues to allow us to communicate with others.

  13. Seeking Temporal Predictability in Speech: Comparing Statistical Approaches on 18 World Languages

    PubMed Central

    Jadoul, Yannick; Ravignani, Andrea; Thompson, Bill; Filippi, Piera; de Boer, Bart

    2016-01-01

    Temporal regularities in speech, such as interdependencies in the timing of speech events, are thought to scaffold early acquisition of the building blocks in speech. By providing on-line clues to the location and duration of upcoming syllables, temporal structure may aid segmentation and clustering of continuous speech into separable units. This hypothesis tacitly assumes that learners exploit predictability in the temporal structure of speech. Existing measures of speech timing tend to focus on first-order regularities among adjacent units, and are overly sensitive to idiosyncrasies in the data they describe. Here, we compare several statistical methods on a sample of 18 languages, testing whether syllable occurrence is predictable over time. Rather than looking for differences between languages, we aim to find across languages (using clearly defined acoustic, rather than orthographic, measures), temporal predictability in the speech signal which could be exploited by a language learner. First, we analyse distributional regularities using two novel techniques: a Bayesian ideal learner analysis, and a simple distributional measure. Second, we model higher-order temporal structure—regularities arising in an ordered series of syllable timings—testing the hypothesis that non-adjacent temporal structures may explain the gap between subjectively-perceived temporal regularities, and the absence of universally-accepted lower-order objective measures. Together, our analyses provide limited evidence for predictability at different time scales, though higher-order predictability is difficult to reliably infer. We conclude that temporal predictability in speech may well arise from a combination of individually weak perceptual cues at multiple structural levels, but is challenging to pinpoint. PMID:27994544

  14. Visual processing affects the neural basis of auditory discrimination.

    PubMed

    Kislyuk, Daniel S; Möttönen, Riikka; Sams, Mikko

    2008-12-01

    The interaction between auditory and visual speech streams is a seamless and surprisingly effective process. An intriguing example is the "McGurk effect": The acoustic syllable /ba/ presented simultaneously with a mouth articulating /ga/ is typically heard as /da/ [McGurk, H., & MacDonald, J. Hearing lips and seeing voices. Nature, 264, 746-748, 1976]. Previous studies have demonstrated the interaction of auditory and visual streams at the auditory cortex level, but the importance of these interactions for the qualitative perception change remained unclear because the change could result from interactions at higher processing levels as well. In our electroencephalogram experiment, we combined the McGurk effect with mismatch negativity (MMN), a response that is elicited in the auditory cortex at a latency of 100-250 msec by any above-threshold change in a sequence of repetitive sounds. An "odd-ball" sequence of acoustic stimuli consisting of frequent /va/ syllables (standards) and infrequent /ba/ syllables (deviants) was presented to 11 participants. Deviant stimuli in the unisensory acoustic stimulus sequence elicited a typical MMN, reflecting discrimination of acoustic features in the auditory cortex. When the acoustic stimuli were dubbed onto a video of a mouth constantly articulating /va/, the deviant acoustic /ba/ was heard as /va/ due to the McGurk effect and was indistinguishable from the standards. Importantly, such deviants did not elicit MMN, indicating that the auditory cortex failed to discriminate between the acoustic stimuli. Our findings show that visual stream can qualitatively change the auditory percept at the auditory cortex level, profoundly influencing the auditory cortex mechanisms underlying early sound discrimination.

  15. How brain asymmetry relates to performance – a large-scale dichotic listening study

    PubMed Central

    Hirnstein, Marco; Hugdahl, Kenneth; Hausmann, Markus

    2014-01-01

    All major mental functions including language, spatial and emotional processing are lateralized but how strongly and to which hemisphere is subject to inter- and intraindividual variation. Relatively little, however, is known about how the degree and direction of lateralization affect how well the functions are carried out, i.e., how lateralization and task performance are related. The present study therefore examined the relationship between lateralization and performance in a dichotic listening task for which we had data available from 1839 participants. In this task, consonant-vowel syllables are presented simultaneously to the left and right ear, such that each ear receives a different syllable. When asked which of the two they heard best, participants typically report more syllables from the right ear, which is a marker of left-hemispheric speech dominance. We calculated the degree of lateralization (based on the difference between correct left and right ear reports) and correlated it with overall response accuracy (left plus right ear reports). In addition, we used reference models to control for statistical interdependency between left and right ear reports. The results revealed a u-shaped relationship between degree of lateralization and overall accuracy: the stronger the left or right ear advantage, the better the overall accuracy. This u-shaped asymmetry-performance relationship consistently emerged in males, females, right-/non-right-handers, and different age groups. Taken together, the present study demonstrates that performance on lateralized language functions depends on how strongly these functions are lateralized. The present study further stresses the importance of controlling for statistical interdependency when examining asymmetry-performance relationships in general. PMID:24427151

  16. Babbling in children with neurodevelopmental disability and validity of a simplified way of measuring canonical babbling ratio.

    PubMed

    Nyman, Anna; Lohmander, Anette

    2018-01-01

    Babbling is an important precursor to speech, but has not yet been thoroughly investigated in children with neurodevelopmental disabilities. Canonical babbling ratio (CBR) is a commonly used but time-consuming measure for quantifying babbling. The aim of this study was twofold: to validate a simplified version of the CBR (CBR UTTER ), and to use this measure to determine if early precursors to speech and language development could be detected in children with different neurodevelopmental disabilities. Two different data sets were used. In Part I, CBR UTTER was compared to two other CBR measures using previously obtained phonetic transcriptions of 3571 utterances from 38 audio recordings of 12-18 month old children with and without cleft palate. In CBR UTTER , number of canonical utterances was divided by total number of utterances. In CBR syl , number of canonical syllables was divided by total number of syllables. In CBR utt , number of canonical syllables was divided by total number of utterances. High agreement was seen between CBR UTTER and CBR syl , suggesting CBR UTTER as an alternative. In Part II, babbling in children with neurodevelopmental disability was examined. Eighteen children aged 12-22 months with Down syndrome, cerebral palsy or developmental delay were audio-video recorded during interaction with a parent. Recordings were analysed by observation of babbling, consonant production, calculation of CBR UTTER , and compared to data from controls. The study group showed significantly lower occurrence of all variables, except for of plosives. The long-term relevance of the findings for the speech and language development of the children needs to be investigated.

  17. Learning to breathe and sing: development of respiratory-vocal coordination in young songbirds

    PubMed Central

    Veit, Lena; Aronov, Dmitriy

    2011-01-01

    How do animals with learned vocalizations coordinate vocal production with respiration? Songbirds such as the zebra finch learn their songs, beginning with highly variable babbling vocalizations known as subsong. After several weeks of practice, zebra finches are able to produce a precisely timed pattern of syllables and silences, precisely coordinated with expiratory and inspiratory pulses (Franz M, Goller F. J Neurobiol 51: 129–141, 2002). While respiration in adult song is well described, relatively little is known about respiratory patterns in subsong or about the processes by which respiratory and vocal patterns become coordinated. To address these questions, we recorded thoracic air sac pressure in juvenile zebra finches prior to the appearance of any consistent temporal or acoustic structure in their songs. We found that subsong contains brief inspiratory pulses (50 ms) alternating with longer pulses of sustained expiratory pressure (50–500 ms). In striking contrast to adult song, expiratory pulses often contained multiple (0–8) variably timed syllables separated by expiratory gaps and were only partially vocalized. During development, expiratory pulses became shorter and more stereotyped in duration with shorter and fewer nonvocalized parts. These developmental changes eventually resulted in the production of a single syllable per expiratory pulse and a single inspiratory pulse filling each gap, forming a coordinated sequence similar to that of adult song. To examine the role of forebrain song-control nuclei in the development of respiratory patterns, we performed pressure recordings before and after lesions of nucleus HVC (proper name) and found that this manipulation reverses the developmental trends in measures of the respiratory pattern. PMID:21697438

  18. Learning to breathe and sing: development of respiratory-vocal coordination in young songbirds.

    PubMed

    Veit, Lena; Aronov, Dmitriy; Fee, Michale S

    2011-10-01

    How do animals with learned vocalizations coordinate vocal production with respiration? Songbirds such as the zebra finch learn their songs, beginning with highly variable babbling vocalizations known as subsong. After several weeks of practice, zebra finches are able to produce a precisely timed pattern of syllables and silences, precisely coordinated with expiratory and inspiratory pulses (Franz M, Goller F. J Neurobiol 51: 129-141, 2002). While respiration in adult song is well described, relatively little is known about respiratory patterns in subsong or about the processes by which respiratory and vocal patterns become coordinated. To address these questions, we recorded thoracic air sac pressure in juvenile zebra finches prior to the appearance of any consistent temporal or acoustic structure in their songs. We found that subsong contains brief inspiratory pulses (50 ms) alternating with longer pulses of sustained expiratory pressure (50-500 ms). In striking contrast to adult song, expiratory pulses often contained multiple (0-8) variably timed syllables separated by expiratory gaps and were only partially vocalized. During development, expiratory pulses became shorter and more stereotyped in duration with shorter and fewer nonvocalized parts. These developmental changes eventually resulted in the production of a single syllable per expiratory pulse and a single inspiratory pulse filling each gap, forming a coordinated sequence similar to that of adult song. To examine the role of forebrain song-control nuclei in the development of respiratory patterns, we performed pressure recordings before and after lesions of nucleus HVC (proper name) and found that this manipulation reverses the developmental trends in measures of the respiratory pattern.

  19. Revisiting the "enigma" of musicians with dyslexia: Auditory sequencing and speech abilities.

    PubMed

    Zuk, Jennifer; Bishop-Liebler, Paula; Ozernov-Palchik, Ola; Moore, Emma; Overy, Katie; Welch, Graham; Gaab, Nadine

    2017-04-01

    Previous research has suggested a link between musical training and auditory processing skills. Musicians have shown enhanced perception of auditory features critical to both music and speech, suggesting that this link extends beyond basic auditory processing. It remains unclear to what extent musicians who also have dyslexia show these specialized abilities, considering often-observed persistent deficits that coincide with reading impairments. The present study evaluated auditory sequencing and speech discrimination in 52 adults comprised of musicians with dyslexia, nonmusicians with dyslexia, and typical musicians. An auditory sequencing task measuring perceptual acuity for tone sequences of increasing length was administered. Furthermore, subjects were asked to discriminate synthesized syllable continua varying in acoustic components of speech necessary for intraphonemic discrimination, which included spectral (formant frequency) and temporal (voice onset time [VOT] and amplitude envelope) features. Results indicate that musicians with dyslexia did not significantly differ from typical musicians and performed better than nonmusicians with dyslexia for auditory sequencing as well as discrimination of spectral and VOT cues within syllable continua. However, typical musicians demonstrated superior performance relative to both groups with dyslexia for discrimination of syllables varying in amplitude information. These findings suggest a distinct profile of speech processing abilities in musicians with dyslexia, with specific weaknesses in discerning amplitude cues within speech. Because these difficulties seem to remain persistent in adults with dyslexia despite musical training, this study only partly supports the potential for musical training to enhance the auditory processing skills known to be crucial for literacy in individuals with dyslexia. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  20. Sensory Constraints on Birdsong Syntax: Neural Responses to Swamp Sparrow Songs with Accelerated Trill Rates

    PubMed Central

    Prather, JF; Peters, S; Mooney, R; Nowicki, S

    2013-01-01

    Both sensory and motor mechanisms can constrain behavioral performance. Sensory mechanisms may be especially important for constraining behaviors that depend on experience, such as learned birdsongs. Swamp sparrows learn to sing by imitating the song of a tutor, but sparrows fail to accurately imitate artificial tutor songs with abnormally accelerated trills, instead singing brief and rapid trills interrupted by silent gaps. This “broken syntax” has been proposed to arise from vocal-motor limitations. Here we consider whether sensory limitations exist that could also contribute to broken syntax. We tested this idea by recording auditory-evoked activity of sensorimotor neurons in the swamp sparrow’s brain that are known to be important for the learning, performance and perception of song. In freely behaving adult sparrows that sang songs with normal syntax, neurons were detected that exhibited precisely time-locked activity to each repetition of the syllable in a trill when presented at a natural rate. Those cells failed to faithfully follow syllables presented at an accelerated rate, however, and their failure to respond to consecutive syllables increased as a function of trill rate. This “flickering” auditory representation in animals performing normal syntax reveals a central constraint on the sensory processing of rapid trills. Furthermore, because these neurons are implicated in both song learning and perception, and because auditory flickering began to occur at accelerated trill rates previously associated with the emergence of broken song syntax, these sensory constraints may contribute to the emergence of broken syntax. PMID:23976787

Top